Sponsored Content
Operating Systems Linux /root filesystem size is full Post 302075529 by hegemaro on Sunday 4th of June 2006 06:36:28 PM
Old 06-04-2006
It has been my experiece that the smit log and shell histories sneak up growing very large and numerous over time. Also, do you have automated processes that cleanup or archive key files such as passwd by making copies? It could easily be one very large file or a huge number of little files once thought innocuous.

Good luck. I'd be interested in what you find.
This User Gave Thanks to hegemaro For This Post:
 

8 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Filesystem Full

I noticed that whenever something is printed from my workstation, the available disk space in the /dev/dsk/c0t0d0s0 decreases considerably. Hence, after using my workstation for sometime, I encounter an error message: "Filesystem Full" that prevents me from printing any further. Is there a way to... (16 Replies)
Discussion started by: ilak1008
16 Replies

2. SCO

Maximum size of root filesystem in SCO 5.0.7

Does anyone know the maximum size of the root filesystem in SCO 5.0.7.? It used to be 1GB max. SCO 5.0.7 with Development System does not fit in 1GB. Is there a limit or has it been removed and the documenation just was not updated? TIA, Dan (2 Replies)
Discussion started by: njsco
2 Replies

3. UNIX for Dummies Questions & Answers

filesystem full

my root filesystem is eventually full "/dev/rdsk/c1d0s0" as a result i cannot boot to the operating system, i booted into the fail safe mode to check the space using df -h command i discover that it is eventually full. Also to my amazement i found that i cannot see the filesystem which mounted on... (1 Reply)
Discussion started by: seyiisq
1 Replies

4. Red Hat

/ filesystem getting full

Hi All, How do I increase the root filesystem? It's getting full. / 90% Here's the break down, below 232 dev 5624 tmp *6764 bin 16860 root *19680 sbin *20436 lib64 28329 boot *47992 etc 150012 var *254540 lib 651708 home *2445044 usr (5 Replies)
Discussion started by: itik
5 Replies

5. UNIX for Dummies Questions & Answers

filesystem is full

Hello everybody, a very basic question. Inspite of me deleting huge files in a filesystem(AIX 5.3) in oracle folder, the filesystem when i check using df -k still shows 100% full. Does that mean there is a process still pointing to the files which i deleted. how do i work around this. Thanks!... (3 Replies)
Discussion started by: karthikosu
3 Replies

6. Solaris

Filesystem Full

In our shop we have to run a batch cycle. Every so often while we are running batch we get a filesystem full situation that causes batch to stop or slow down. Anyway, the practiced procedure is to look for large files and zip them. This takes a lot of time. We are in a sun solaris environment. What... (1 Reply)
Discussion started by: Harleyrci
1 Replies

7. Solaris

How to resolve Filesystem Full?

I have experienced Filesystem full (%iused = 100%) as below. I have tried to remove a lot of garbage but it still become full soon. # df -F ufs -o i Filesystem       iused    ifree  %iused  Mount /dev/dsk/c0t0d0s0  512000     0   100%  / /dev/dsk/c0t0d0s3   1887   382113   0%  /cms Other... (4 Replies)
Discussion started by: SwordCar
4 Replies

8. UNIX for Advanced & Expert Users

Physical disk IO size smaller than fragment block filesystem size ?

Hello, in one default UFS filesystem we have 8K block size (bsize) and 1K fragmentsize (fsize). At this scenary I thought all "FileSytem IO" will be 8K (or greater) but never smaller than the fragment size (1K). If a UFS fragment/blocksize is allwasy several ADJACENTS sectors on disk (in a ... (4 Replies)
Discussion started by: rarino2
4 Replies
clfmerge(1)							     logtools							       clfmerge(1)

NAME
clfmerge - merge Common-Log Format web logs based on time-stamps SYNOPSIS
clfmerge [--help | -h] [-b size] [-d] [file names] DESCRIPTION
The clfmerge program is designed to avoid using sort to merge multiple web log files. Web logs for big sites consist of multiple files in the >100M size range from a number of machines. For such files it is not practical to use a program such as gnusort to merge the files because the data is not always entirely in order (so the merge option of gnusort doesn't work so well), but it is not in random order (so doing a complete sort would be a waste). Also the date field that is being sorted on is not particularly easy to specify for gnusort (I have seen it done but it was messy). This program is designed to simply and quickly sort multiple large log files with no need for temporary storage space or overly large buf- fers in memory (the memory footprint is generally only a few megs). OVERVIEW
It will take a number (from 0 to n) of file-names on the command line, it will open them for reading and read CLF format web log data from them all. Lines which don't appear to be in CLF format (NB they aren't parsed fully, only minimal parsing to determine the date is per- formed) will be rejected and displayed on standard-error. If zero files are specified then there will be no error, it will just silently output nothing, this is for scripts which use the find com- mand to find log files and which can't be counted on to find any log files, it saves doing an extra check in your shell scripts. If one file is specified then the data will be read into a 1000 line buffer and it will be removed from the buffer (and displayed on stan- dard output) in date order. This is to handle the case of web servers which date entries on the connection time but write them to the log at completion time and thus generate log files that aren't in order (Netscape web server does this - I haven't checked what other web servers do). If more than one file is specified then a line will be read from each file, the file that had the earliest time stamp will be read from until it returns a time stamp later than one of the other files. Then the file with the earlier time stamp will be read. With multiple files the buffer size is 1000 lines or 100 * the number of files (whichever is larger). When the buffer becomes full the first line will be removed and displayed on standard output. OPTIONS
-b buffer-size Specify the buffer-size to use, if 0 is specified then it means to disable the sliding-window sorting of the data which improves the speed. -d Set domain-name mangling to on. This means that if a line starts with as the name of the site that was requested then that would be removed from the start of the line and the GET / would be changed to GET http://www.company.com/ which allows programs like Webal- izer to produce good graphs for large hosting sites. Also it will make the domain name in lower case. EXIT STATUS
0 No errors 1 Bad parameters 2 Can't open one of the specified files 3 Can't write to output AUTHOR
This program, its manual page, and the Debian package were written by Russell Coker <russell@coker.com.au>. SEE ALSO
clfsplit(1),clfdomainsplit(1) Russell Coker <;russell@coker.com.au> 0.06 clfmerge(1)
All times are GMT -4. The time now is 01:22 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy