Sponsored Content
Operating Systems Linux Red Hat Disk is Full but really does not contain huge data Post 302661423 by kalpeer on Monday 25th of June 2012 09:04:11 AM
Old 06-25-2012
MySQL Disk is Full but really does not contain huge data

Hi All,

My disk usage show 100 % . When I check “df –kh” it shows my root partition is full. But when I run the “du –skh /” shows only 7 GB is used.

Code:
Filesystem Size Used Avail Use% Mounted on
/dev/sda1 30G 28G 260MB 100% /

How I can identify who is using the 20 GB of memory.
Os: Centos 5.5

Thanks,
Kalai

Last edited by zaxxon; 06-25-2012 at 10:08 AM.. Reason: code tags
 

9 More Discussions You Might Find Interesting

1. UNIX for Advanced & Expert Users

disk full

Please solve the following NOTICE HTFS:No space on dev hd(1/42) (2 Replies)
Discussion started by: msuheel
2 Replies

2. Shell Programming and Scripting

How to extract data from a huge file?

Hi, I have a huge file of bibliographic records in some standard format.I need a script to do some repeatable task as follows: 1. Needs to create folders as the strings starts with "item_*" from the input file 2. Create a file "contents" in each folders having "license.txt(tab... (5 Replies)
Discussion started by: srsahu75
5 Replies

3. Linux

Disk full 100%

one of my servers / was full by 100% i cleard some space, now though i have enough space on / partition still df is showing disk usage as 100% am not able to create any single txt file ? why so ? (3 Replies)
Discussion started by: bryanabhay
3 Replies

4. AIX

Huge difference in reported Disk usage between ls,df and du

IBM RS6000 F50 AIX 4.3.2 i am having trouble in calculating the actual size of a set of directories and reconciling the results with the actual Hard Disk space used I have 33GB disk which is showing 7.8GB used, a byte count of the files in the directory/sub-dirs i`m interested in is 48GB,... (4 Replies)
Discussion started by: cooperuf
4 Replies

5. Shell Programming and Scripting

Split a huge data into few different files?!

Input file data contents: >seq_1 MSNQSPPQSQRPGHSHSHSHSHAGLASSTSSHSNPSANASYNLNGPRTGGDQRYRASVDA >seq_2 AGAAGRGWGRDVTAAASPNPRNGGGRPASDLLSVGNAGGQASFASPETIDRWFEDLQHYE >seq_3 ATLEEMAAASLDANFKEELSAIEQWFRVLSEAERTAALYSLLQSSTQVQMRFFVTVLQQM ARADPITALLSPANPGQASMEAQMDAKLAAMGLKSPASPAVRQYARQSLSGDTYLSPHSA... (7 Replies)
Discussion started by: patrick87
7 Replies

6. UNIX for Advanced & Expert Users

Disk Space full

I was tryin to copy a large file under /tmp location. I guess the disk space got full and i got fork error. Then I tried removing some files but the shell did not let me do anything bash> rm apache22.tar bash: fork: Not enough space bash> pwd /tmp bash> vmstat 1 bash: fork: Not... (3 Replies)
Discussion started by: mohtashims
3 Replies

7. Shell Programming and Scripting

Aggregation of huge data

Hi Friends, I have a file with sample amount data as follows: -89990.3456 8788798.990000128 55109787.20 -12455558989.90876 I need to exclude the '-' symbol in order to treat all values as an absolute one and then I need to sum up.The record count is around 1 million. How... (8 Replies)
Discussion started by: Ravichander
8 Replies

8. Solaris

The Fastest for copy huge data

Dear Experts, I would like to know what's the best method for copy data around 3 mio (spread in a hundred folders, size each file around 1kb) between 2 servers? I already tried using Rsync and tar command. But using these command is too long. Please advice. Thanks Edy (11 Replies)
Discussion started by: edydsuranta
11 Replies

9. Shell Programming and Scripting

Disk full alerts

i want to create 1 script to monitor 1 particular filesystem out of the diferent filesystems. if disk space of that particular filesystem increases by 80% it sends an alert mail to an email id ---------- Post updated at 04:18 PM ---------- Previous update was at 04:17 PM ---------- no. I am... (1 Reply)
Discussion started by: rakeshhhhhhhh
1 Replies
vdf(8)							      System Manager's Manual							    vdf(8)

NAME
vdf - Display disk information for AdvFS domains and filesets SYNOPSIS
/sbin/advfs/vdf [-k] [-l] domain | domain#fileset OPTIONS
Display disk blocks as 1024-byte blocks instead of the default of 512-byte blocks. When the -l option is specified, the default informa- tion for both the domain and filesets is reformatted to show the relationships between them. For example, any domain metadata displayed is the total metadata shared by filesets in the domain. OPERANDS
The full path name of an AdvFS file domain. When a domain argument is specified, the default display contains information about the number of disk blocks allocated to the domain, the number of disk blocks in use by the domain, and the number of disk blocks that are available to the domain. The name of an AdvFS fileset in an AdvFS file domain. When a domain#fileset argument is specified, the default display contains information about the number of disk blocks allocated to the fileset, the number of disk blocks in use by the fileset, and the number of disk blocks that are available to the fileset. This information is in the same format as that displayed by the df command. DESCRIPTION
The vdf utility is a script that reformats output from the showfdmn, showfsets, shfragbf, and df utilities in order to display information about the disk usage of AdvFS file domains and filesets. In addition, the utility computes and displays the sizes of metadata files in a domain or fileset. NOTES
The disk space used by clone filesets is not calculated. If clone filesets are present in the specified domain, the utility displays the following warning message: Clone fileset(s) in this domain; totals may not be accurate RESTRICTIONS
You must be the root user to use this command. The command cannot be used on filesets that are NFS mounted. All filesets in a domain must be mounted in order to calculate the disk usage of the domain. EXIT STATUS
The utility returns a value of 0 (zero) on successful completion. It returns a value of 1 on failure. EXAMPLES
The following example shows the summary information for a fileset named testfs in a domain named test. The output is the same as for the df command: # vdf test#testfs Filesystem 512-blocks Used Available Capacity Mounted on test#testfs 60000 20800 39200 35% /mnt The following example shows the summary information for a domain named usr_domain where the number of blocks used for meta- data and data equal 89% of the domain's capacity: # vdf usr_domain Domain 512-blocks Metadata Used Available Capacity usr_domain 65536 11219 47549 6768 89% The following example shows a detailed display of the domain that contains the fileset test#testfs. There are two filesets in the domain; each has a quota limit of 60,000 blocks: # vdf -l test#testfs Domain 512-blocks Metadata Used Available Capacity test 266240 5824 29128 231288 13% Fileset QuotaLimit Used Available Capacity testfs 60000 20800 39200 35% testfs2 60000 8328 51672 14% The following example shows a detailed display of a fileset that is the only fileset in the domain and that the fileset has a quota limit that is greater than the domain can hold: # vdf -l test#testfs Domain 512-blocks Metadata Used Avail- able Capacity test 266240 5824 29128 231288 13% Fileset QuotaLimit Used Available Capacity testfs 300000 20800 231288 8% FILES
SEE ALSO
Commands: df(1), showfdmn(8), showfsets(8), shfragbf(8) Files: advfs(4) vdf(8)
All times are GMT -4. The time now is 04:09 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy