Sponsored Content
Top Forums Shell Programming and Scripting How to tar large amount of files? Post 302472811 by jim mcnamara on Thursday 18th of November 2010 07:21:19 AM
Old 11-18-2010
Assuming HPUX tar allows the u option
Code:
tar -cvf backup.tar VOICE* 
tar -uvf backup.tar SUBSCR* 
tar -uvf backup.tar DEL*

 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

Using TAR on a large file

I am runnign Solaris 8. I am trying to TAR a large file (13 gig) and it is complaining about the size. Any suggestions? (4 Replies)
Discussion started by: hshapiro
4 Replies

2. Linux

shmat() Failure While Using a Large Amount of Shared Memory

Hi, I'm developing a data processing pipeline with multiple stages, with data being moved between the stages using shared memory segments. The size of the data is typically of the order of hundreds of megabytes, and there are typically a few tens of main shared memory segments each of size... (2 Replies)
Discussion started by: theicarusagenda
2 Replies

3. Programming

Read/Write a fairly large amount of data to a file as fast as possible

Hi, I'm trying to figure out the best solution to the following problem, and I'm not yet that much experienced like you. :-) Basically I have to read a fairly large file, composed of "messages" , in order to display all of them through an user interface (made with QT). The messages that... (3 Replies)
Discussion started by: emitrax
3 Replies

4. AIX

amount of memory allocated to large page

We just set up a system to use large pages. I want to know if there is a command to see how much of the memory is being used for large pages. For example if we have a system with 8GB of RAm assigned and it has been set to use 4GB for large pages is there a command to show that 4GB of the *GB is... (1 Reply)
Discussion started by: daveisme
1 Replies

5. Solaris

Tar too large to archive. Use E function modifier.

hey all, i am trying to tar up a folder with sub folders the over all size will be about 70gb but when i use the normal command tar -cvf tar -cvf CLPSI_PRU_Escrow_31994.tar CLPSI_PRU_Escrow_31994 i get an error tar: CLPSI_PRU_Escrow_31994/dump1/PROD_SAE_jria3_dump.5 too large to archive. ... (9 Replies)
Discussion started by: dshakey
9 Replies

6. UNIX for Dummies Questions & Answers

questing regarding tar large number of files

I want to tar large number of files about 150k. i am using the find command as below to create a file with all file names. & then trying to use the tar -I command as below. # find . -type f -name "gpi*" > include-file # tar -I include-file -cvf newfile.tar This i got from one of the posts... (2 Replies)
Discussion started by: crux123
2 Replies

7. Shell Programming and Scripting

tar command to explore multiple layers of tar and tar.gz files

Hi all, I have a tar file and inside that tar file is a folder with additional tar.gz files. What I want to do is look inside the first tar file and then find the second tar file I'm looking for, look inside that tar.gz file to find a certain directory. I'm encountering issues by trying to... (1 Reply)
Discussion started by: bashnewbee
1 Replies

8. UNIX for Dummies Questions & Answers

Compress 1st 20 Large different File using tar

Sorry guys.. dont know where to put this.. currently I am cleaning up system dump on our aix machine and I wanted to set zero the 1st 20 large file but before doing that I wanted to create some backup. is there any command that can compress all these file same time? tar syntax file? ... (2 Replies)
Discussion started by: thermometer
2 Replies

9. Shell Programming and Scripting

Perl : Large amount of data put into an array

This basic code works. I have a very long list, almost 10000 lines that I am building into the array. Each line has either 2 or 3 fields as shown in the code snippit. The array elements are static (for a few reasons that out of scope of this question) the list has to be "built in". It... (5 Replies)
Discussion started by: sumguy
5 Replies

10. Shell Programming and Scripting

How to make awk command faster for large amount of data?

I have nginx web server logs with all requests that were made and I'm filtering them by date and time. Each line has the following structure: 127.0.0.1 - xyz.com GET 123.ts HTTP/1.1 (200) 0.000 s 3182 CoreMedia/1.0.0.15F79 (iPhone; U; CPU OS 11_4 like Mac OS X; pt_br) These text files are... (21 Replies)
Discussion started by: brenoasrm
21 Replies
GPODDER-BACKUP(1)						   User Commands						 GPODDER-BACKUP(1)

NAME
gpodder-backup - Backup and restore utility for gPodder user data SYNOPSIS
gpodder-backup [--create|--extract] <archive.gpo.tar.gz> [options] gpodder-backup --purge DESCRIPTION
This utility can be used to create a dump of the current gPodder data (configuration files + downloads), optionally replacing the real con- tents of the download folder with zero-byte files (for submitting your data to a bug report without having to transfer lots of data). OPTIONS
--version show program's version number and exit -h, --help show this help message and exit -c <FILE>, --create=<FILE> Create a new archive -x <FILE>, --extract=<FILE> Extract an existing archive -f, --fake-downloads Store downloads as zero-byte files in backup -n, --no-covers Do not include cover files in archive -D <DIR>, --destination=<DIR> Extract downloads in different folder -P, --purge Remove current data (can be combined with --extract) USAGE FOR BUG REPORTING
This command is useful if you want to report a bug in gPodder: gpodder-backup --create bug123.gpo.tar.gz --fake-downloads Backup your current data to file bug123.gpo.tar.gz, but don't store download data (create zero-size dummy files instead). You can then attach bug123.gpo.tar.gz to the bug report. If you are instructed to try gPodder from a "clean state", you can use the following command (be sure to backup your data before!): gpodder-backup --purge Remove all gPodder data, so you can start from a clean state EXAMPLES
gpodder-backup --create today.gpo.tar.gz Backup your current data to file today.gpo.tar.gz gpodder-backup --extract mybackup.gpo.tar.gz Restore (without purging) the contents of mybackup.gpo.tar.gz gpodder-backup --extract default.gpo.tar.gz --purge Remove current data, then restore the contents of default.gpo.tar.gz gpodder-backup --purge Remove all gPodder data, so you can start from a clean state EXTRACTING FAKED DOWNLOADS
Please note that any existing downloads will be overwritten with zero-byte files when using the --extract option of the gpodder-backup utility with a backup created with --fake-downloads AUTHOR
gpodder-backup was written by Thomas Perl (thp@gpodder.org) gpodder-backup 1.0 December 2010 GPODDER-BACKUP(1)
All times are GMT -4. The time now is 12:43 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy