We are taking backup of our application data(cobol file system, AIX/unix) before and after EOD job runs. The data size is approximately 260 GB in biggest branch. To reduce the backup time, 5 parallel execution is scheduled through control-m which backups up the files in 5 different *.gz. The job takes approximately 90 minutes to complete. Backup is done locally and later on it's passed to different location for retention.
Issue:
Each execution takes approximately 10% CPU which is putting heavy load on the server. This is causing the issue as the server is hosting multiple branches.
Is there anyway, we can improve the backup further? Is there any new features in GTAR to fasten the backup or any new backup command which can replace GTAR? Any suggestion would be really appreciated.
Hi All,
We have a gtar file and we are trying to untar the file with the option
gtar -xvzf <filename>
The gtar gets us till the end and throws the error message as highlighed below
mfcp/XFHFCD2.CPY
mfcp/XFHFCD3.CPY
mfcp/XFHFCD.CPY
gzip: stdin: unexpected end of file
gtar:... (1 Reply)
I am trying to write a very large file, 570 gb, to a tape using gtar like this :
gtar czxf /dev/rmt/1 ./*
I get a message:
off_t value 570635451556 too large (max=68719476735)
It is writing to tape, but will it be good?
Thanks (1 Reply)
My data is something like shown below.
date1 date2 aaa bbbb ccccc
date3 date4 dddd eeeeeee ffffffffff ggggg hh
I want the output like this
date1date2 aaa eeeeee
I serached in the forum but didn't find the exact matching solution. Please help. (7 Replies)
Hi all,
will gtar zcvf command work in csh and tcsh shells? Becuase when i'm executing one script in bash and ksh, it's working fine. But it's not working in csh and tcsh shells. We have to run multiple scripts in tcsh, so we can not change the shell while executing these scripts. One of my... (2 Replies)
Hello All,
I am preparing a script to view or Extract contents of a tape drive using gtar.But facing a strange issue while trying to extract files using gtar.
If running script using sudo the getting the below error.
################
/usr/local/lib /usr/X11/lib /usr/X11R6/lib... (1 Reply)
Ok so once again im back with what is probably a beginner question although somewhat more complicated (for me) than the last.
Background:
A client has a daily backup which is carried out via rsync.
Due to this, when they move a file around that file is then coppied a second time.
On top of... (4 Replies)
Hi Everyone,
we are running rsync with --backup mode, Are there any rsync options to remove backup folders on successful deployment?
Thanks in adv. (0 Replies)
We are taking backup of our application data(cobol file system, AIX/unix) before and after EOD job runs. The data size is approximately 260 GB in biggest branch. To reduce the backup time, 5 parallel execution is scheduled through control-m which backups up the files in 5 different *.gz. The job... (8 Replies)
Hi All!
i am trying to copy files from a SCO Openserver 5.0.6 to a NAS Server using NFS. I have a cron job that takes 1.5 hours to run and most of the data is static. I would like to find a faster way. In the event I needed to running manually or avoid an issue with taking down the servers... (9 Replies)
Hi all
I have a unix based firewall, which creates a daily backup file on the device.
I need a script to scp this file over to a remote server.
I can get this working daily using a basic script and a cron job.
However, I only want it to send the latest config back up file and currently... (4 Replies)
Discussion started by: jimmyzoom
4 Replies
LEARN ABOUT DEBIAN
svn-fast-backup
svn-fast-backup(1) General Commands Manual svn-fast-backup(1)NAME
svn-fast-backup - very fast backup for Subversion fsfs repositories.
SYNOPSIS
svn-fast-backup [-q] [-k{N|all}] [-f] [-t] [-s] repos_path backup_dir
DESCRIPTION
svn-fast-backup uses rsync snapshots for very fast backup of a Subversion fsfs repository at repos_path to backup_dir/repos-rev, the latest
revision number in the repository. Multiple fsfs backups share data via hardlinks, so old backups are almost free, since a newer revision
of a repository is almost a complete superset of an older revision.
This is good for replacing incremental log-dump+restore-style backups because it is just as space-conserving and even faster; there is no
inter-backup state (old backups are essentially caches); each backup directory is self-contained. It has the same command-line interface
as svn-hot-backup(1) (if you use --force), but only works for fsfs repositories.
svn-fast-backup keeps 64 backups by default and deletes backups older than these; this can be adjusted with the -k option.
OPTIONS -h, --help
Shows some brief help text.
-q, --quiet
Quieter-than-usual operation.
-k, --keep=N
Keep a specified number of backups; the default is to keep 64.
-k, --keep=all
Do not delete any old backups at all.
-f, --force
Make a new backup even if one with the current revision exists.
-t, --trace
Show actions.
-s, --simulate
Don't perform actions.
AUTHOR
Voluntary contributions made by many individuals. Copyright (C) 2006 CollabNet.
2006-11-09 svn-fast-backup(1)