I'm using nas4free as a SAN and am having troubles getting a backup of it's data to work properly. I've posted in the nas4free forums, but haven't received much help.
Here is the code I'm using:
It is scheduled to run as a cron once per day via the nas4free gui. The script runs, and the daily rsync works, but the log file is empty and the weekly commands (in the if statement) do not seem to be working. When I run the tar individually, I get the following:
Quote:
syl-s-030:~# tar -zcfv /mnt/syl-s-012/Backups/CaseData/Weekly/date.tar.gz /mnt/syl-s-012/Backups/CaseData/Daily/
tar: Removing leading '/' from member names
/: write failed, filesystem is full
tar: Write error
I know that the embedded nas4free doesn't have room to install any additional applications, but the source and destination are mounted shares.
So my questions are this.
Does tar use some kind of temp directory that is local? If so, can it be changed?
My logging syntax might be incorrect. If so, how can I create a log of the scripts output that can be emailed?
I have been using the hostdump.sh backup script for over a year now and have recently run into a problem. I'm now getting the following error at the end of my jobs;
/bin/mt -f: error fsf'ing tape.
This script uses the native 'ufsdump'. So, I try to go back and read the last dump on the tape... (11 Replies)
Hello everyone my ? is about backups. I'am running SCO OS 505 and currently backing up the hole HD. Well the back up is taking too long and this is becoming a problem for the users since we are a 24-7 bussines, I whant stop backing up every thing on the HD. What are the most important files and... (1 Reply)
I've been handed the task of backing up some of our system files on a Solaris box. No probs. Zipped the logs that needed backing up but my superiors do not want it on tape, they want it spanned on CD's. The CD-Writer is available on a MS box. FTP'd the zipped logs across too the MS Machine but now... (1 Reply)
I am currently bringing up an offsite location, right now I am in the process of copying some data offsite (about 400GB).
The problem I see is that running a single rsync for everything is not using the available bandwidth and testing shows that I double in speed for each instance of Rsync I am... (3 Replies)
I'm wanting to do remote backups of the entire /home/* directory structure, which is where HG puts all www files and sets the user and group independently of any other user ( so /home/user1 will be using group user1, instead of a generic apache group)
The problem I'm running into is that only... (0 Replies)
Hello,
I've got multiple AIX LPARs running on VIOS, within a blade environment. I need to dump a mksysb backup to backup rootvg and a couple of other volume groups.
mksysb -i "destination"; works however I'd like to make sure its being done correctly.
on the other volume groups, ive... (2 Replies)
I wish to copy all the files & folder under /web/Transfer_Files/data/ on mymac1 (Linux) to remote server mybank.intra.com (Solaris 10) /tmp/ location
I am using Ansible tool synchronize module which triggers the unix rsync command as below:rsync --delay-updates -F --compress --archive --rsh=ssh... (2 Replies)
Discussion started by: mohtashims
2 Replies
LEARN ABOUT DEBIAN
svn-fast-backup
svn-fast-backup(1) General Commands Manual svn-fast-backup(1)NAME
svn-fast-backup - very fast backup for Subversion fsfs repositories.
SYNOPSIS
svn-fast-backup [-q] [-k{N|all}] [-f] [-t] [-s] repos_path backup_dir
DESCRIPTION
svn-fast-backup uses rsync snapshots for very fast backup of a Subversion fsfs repository at repos_path to backup_dir/repos-rev, the latest
revision number in the repository. Multiple fsfs backups share data via hardlinks, so old backups are almost free, since a newer revision
of a repository is almost a complete superset of an older revision.
This is good for replacing incremental log-dump+restore-style backups because it is just as space-conserving and even faster; there is no
inter-backup state (old backups are essentially caches); each backup directory is self-contained. It has the same command-line interface
as svn-hot-backup(1) (if you use --force), but only works for fsfs repositories.
svn-fast-backup keeps 64 backups by default and deletes backups older than these; this can be adjusted with the -k option.
OPTIONS -h, --help
Shows some brief help text.
-q, --quiet
Quieter-than-usual operation.
-k, --keep=N
Keep a specified number of backups; the default is to keep 64.
-k, --keep=all
Do not delete any old backups at all.
-f, --force
Make a new backup even if one with the current revision exists.
-t, --trace
Show actions.
-s, --simulate
Don't perform actions.
AUTHOR
Voluntary contributions made by many individuals. Copyright (C) 2006 CollabNet.
2006-11-09 svn-fast-backup(1)