09-19-2013
Delete the file which crossed 2GB
Hi ,
I wants to create the bash script for deleting the specified 2gb file and wants to take the backup before doing that.
please help me how to do the same,I use RHEL5 server
10 More Discussions You Might Find Interesting
1. UNIX for Dummies Questions & Answers
I am working on HP-Unix.
I have a 600 MB file in compressed form.
During decompression, when file size reaches
2GB, decompression aborts.
What should be done? (3 Replies)
Discussion started by: Nadeem Mistry
3 Replies
2. Programming
My C++ program returns 'Disk Full' Message when I tried to manage a file larger than 2Gb. the process is very simple: based on a TXT file, the process combine the information generating another temporary file (generating the error) to fillup a database.
My FS, during the process, reaches 40%...... (4 Replies)
Discussion started by: ASOliveira
4 Replies
3. Shell Programming and Scripting
Can an expert kindly write an efficient Linux ksh script that will split a large 2 GB text file into two?
Here is a couple of sample record from that text file:
"field1","field2","field3",11,22,33,44
"TG","field2b","field3b",1,2,3,4
The above rows are delimited by commas.
This script is to... (2 Replies)
Discussion started by: ihot
2 Replies
4. UNIX for Dummies Questions & Answers
Hi experts,
How do i find which are the filesystems which has crossed 90% capacity in solaris box.
thanks
Shaan:) (4 Replies)
Discussion started by: shaan_dmp
4 Replies
5. UNIX for Dummies Questions & Answers
Hi All,
We are running HP rp7400 box with hpux 11iv1.
Recently, we changed 3 kernel parameters
a) msgseg from 32560 to 32767
b) msgmnb from 65536 to 65535
c) msgssz from 128 to 256
Then we noticed that all application debug file size increase upto 2GB then it stops. So far we did not... (1 Reply)
Discussion started by: mhbd
1 Replies
6. AIX
I am trying to execute a database dump to a file, but can't seem to get around the 2GB file size. I have tried setting the user limit to -1, but no luck. (4 Replies)
Discussion started by: markper
4 Replies
7. Programming
Hi,
I've created a simple application that is supposed to fill up a file with messages up
to the size I pass as parameter.
The problem is that once the file reaches the 2GB size, it stops growing.
The flow of the application, for what is worth, is as follows:
while ( bytes written <... (7 Replies)
Discussion started by: emitrax
7 Replies
8. Linux
I am not able to unzip file greater then 2gb,
Any suggestions how to do that in linux?
Regards,
Manoj (5 Replies)
Discussion started by: manoj.solaris
5 Replies
9. UNIX for Advanced & Expert Users
Hi,
I am executing a SQL query and the output is more than 2GB. Hence the process is failing. How can I have a file created more than 2GB ?
Thanks,
Risshanth (1 Reply)
Discussion started by: risshanth
1 Replies
10. HP-UX
Greetings,
I'm attempting to dump a filesystem from a RHEL5 Linux server to a VXFS filesystem on an HP-UX server. The VXFS filesystem is large file enabled and I've confirmed that I can copy/scp a file >2GB to the filesystem.
# fsadm -F vxfs /os_dumps
largefiles
# mkfs -F vxfs -m... (12 Replies)
Discussion started by: bkimura
12 Replies
LEARN ABOUT DEBIAN
bup-on
bup-on(1) General Commands Manual bup-on(1)
NAME
bup-on - run a bup server locally and client remotely
SYNOPSIS
bup on <hostname> index ...
bup on <hostname> save ...
bup on <hostname> split ...
DESCRIPTION
bup on runs the given bup command on the given host using ssh. It runs a bup server on the local machine, so that commands like bup save
on the remote machine can back up to the local machine. (You don't need to provide a --remote option to bup save in order for this to
work.)
See bup-index(1), bup-save(1), and so on for details of how each subcommand works.
This 'reverse mode' operation is useful when the machine being backed up isn't supposed to be able to ssh into the backup server. For
example, your backup server can be hidden behind a one-way firewall on a private or dynamic IP address; using an ssh key, it can be autho-
rized to ssh into each of your important machines. After connecting to each destination machine, it initiates a backup, receiving the
resulting data and storing in its local repository.
For example, if you run several virtual private Linux machines on a remote hosting provider, you could back them up to a local (much less
expensive) computer in your basement.
EXAMPLES
# First index the files on the remote server
$ bup on myserver index -vux /etc
bup server: reading from stdin.
Indexing: 2465, done.
bup: merging indexes (186668/186668), done.
bup server: done
# Now save the files from the remote server to the
# local $BUP_DIR
$ bup on myserver save -n myserver-backup /etc
bup server: reading from stdin.
bup server: command: 'list-indexes'
PackIdxList: using 7 indexes.
Saving: 100.00% (241/241k, 648/648 files), done.
bup server: received 55 objects.
Indexing objects: 100% (55/55), done.
bup server: command: 'quit'
bup server: done
# Now we can look at the resulting repo on the local
# machine
$ bup ftp 'cat /myserver-backup/latest/etc/passwd'
root:x:0:0:root:/root:/bin/bash
daemon:x:1:1:daemon:/usr/sbin:/bin/sh
bin:x:2:2:bin:/bin:/bin/sh
sys:x:3:3:sys:/dev:/bin/sh
sync:x:4:65534:sync:/bin:/bin/sync
...
SEE ALSO
bup-index(1), bup-save(1), bup-split(1)
BUP
Part of the bup(1) suite.
AUTHORS
Avery Pennarun <apenwarr@gmail.com>.
Bup unknown- bup-on(1)