Sponsored Content
Operating Systems OS X (Apple) Rsync for back up, external HD Post 302824629 by drl on Friday 21st of June 2013 08:17:57 AM
Old 06-21-2013
Hi.

For backup of files, I use rsnapshot, possibly in your repositories, more information at rsnapshot . It uses rsync as part of a larger organization process. I have used it for more than a year, it allows recovery of individual files, and uses hard-links to create a very compact backup, yet still "addressable" by date-time of collection. You can set up cron jobs to do hourly, daily, weekly, monthly backups from the remote gathering computer. I backup at 4-hour intervals.

For a very quick backup of everything, I often used a tar.gz file -- it quickly restores my files to updated, and newly installed systems with no trouble (although you may not want every old file).

For cloning to create a bootable image, I suggest you look at Clonezilla - About -- it will use dd as a last resort, but will first check to see if it can copy by omitting unused sectors. (I use software RAID, which causes a dd copy.)

Always experiment before you commit important data.

Best wishes ... cheers, drl
 

8 More Discussions You Might Find Interesting

1. Solaris

External USB

Is it possible to install Solaris 10 on an external USB drive? I'd like to dual boot Linux and Solaris 10. Thanks! (2 Replies)
Discussion started by: otterit
2 Replies

2. UNIX for Dummies Questions & Answers

rsync back?

Hi, I am in the process of developing backup script using rsync, my code is bellow: EXCLUDE_DIR="/home/kannanpg/mscdr/src/ex1" rsync -az -e ssh -v --exclude $EXCLUDE_DIR --delete $HOSTTOBACKUP:$SOURCE $DR_BACKUP_DIR/daily.0 >$tempfile 2>&1 even my exclude dir is coping.. what is wrong in... (1 Reply)
Discussion started by: redlotus72
1 Replies

3. AIX

back to back printing in UNIX

Hi , Can you suggest me how to back to back printing in UNIX? Is there any way? Kindly advise. Regards Vijaya Amirtha Raj (3 Replies)
Discussion started by: amirthraj_12
3 Replies

4. Shell Programming and Scripting

Rsync to an external list of URLs

I'm going to have a text file formatted something like this: some_name http://www.someurl.com/ another_name http://www.anotherurl.com/ third_name http://www.thirdurl.com/ I need to write a script that can rsync from a file path I'll set, to each URL in the list. Any ideas? (8 Replies)
Discussion started by: ibsen
8 Replies

5. IP Networking

Back-to-Back Connection using HBAs

Hi every body, Is it possible to connect two servers Back-to-Back (Point-to-Point) using HBA adapters & using Fiber. Note it is direct connection & there is no switches between the servers. I'm concern about using HBA adapters, it is possible or not. Thanks in advance. :) (3 Replies)
Discussion started by: aldowsary
3 Replies

6. Shell Programming and Scripting

Bash script to back up with rsync

I'm trying to write a script to back up with rsync, the script I am posting below: nomServer="shiva horus isis" dirshiva="/etc /faturamento" shivaListExc="/usr/local/bin/shivaListExclRsync" dumpDir="$dumpFile/Shiva" dateBkp=`date +%A` for nServer in ${nomServer} do ... (3 Replies)
Discussion started by: ajmoreti
3 Replies

7. UNIX for Dummies Questions & Answers

External HDD

I need to get an external HDD for a SUN server running Solaris 10. The Western Digital that I have will not recognize and when I went looking for drivers WD only has them for MAC and Windows. Is there a External HDD that is known to work with Unix? (24 Replies)
Discussion started by: SIFT3R
24 Replies

8. Shell Programming and Scripting

Rsync Error: rsync: link_stat failed: No such file or directory (2)

I wish to copy all the files & folder under /web/Transfer_Files/data/ on mymac1 (Linux) to remote server mybank.intra.com (Solaris 10) /tmp/ location I am using Ansible tool synchronize module which triggers the unix rsync command as below:rsync --delay-updates -F --compress --archive --rsh=ssh... (2 Replies)
Discussion started by: mohtashims
2 Replies
EXPIRE_BACKUPS(1)						       S3QL							 EXPIRE_BACKUPS(1)

NAME
expire_backups - Intelligently expire old backups SYNOPSIS
expire_backups [options] <age> [<age> ...] DESCRIPTION
The expire_backups command intelligently remove old backups that are no longer needed. To define what backups you want to keep for how long, you define a number of age ranges. expire_backups ensures that you will have at least one backup in each age range at all times. It will keep exactly as many backups as are required for that and delete any backups that become redundant. Age ranges are specified by giving a list of range boundaries in terms of backup cycles. Every time you create a new backup, the existing backups age by one cycle. Example: when expire_backups is called with the age range definition 1 3 7 14 31, it will guarantee that you always have the following backups available: 1. A backup that is 0 to 1 cycles old (i.e, the most recent backup) 2. A backup that is 1 to 3 cycles old 3. A backup that is 3 to 7 cycles old 4. A backup that is 7 to 14 cycles old 5. A backup that is 14 to 31 cycles old Note If you do backups in fixed intervals, then one cycle will be equivalent to the backup interval. The advantage of specifying the age ranges in terms of backup cycles rather than days or weeks is that it allows you to gracefully handle irregular backup intervals. Imagine that for some reason you do not turn on your computer for one month. Now all your backups are at least a month old, and if you had specified the above backup strategy in terms of absolute ages, they would all be deleted! Specifying age ranges in terms of backup cycles avoids these sort of problems. expire_backups usage is simple. It requires backups to have names of the forms year-month-day_hour:minute:seconds (YYYY-MM-DD_HH:mm:ss) and works on all backups in the current directory. So for the above backup strategy, the correct invocation would be: expire_backups.py 1 3 7 14 31 When storing your backups on an S3QL file system, you probably want to specify the --use-s3qlrm option as well. This tells expire_backups to use the s3qlrm command to delete directories. expire_backups uses a "state file" to keep track which backups are how many cycles old (since this cannot be inferred from the dates con- tained in the directory names). The standard name for this state file is .expire_backups.dat. If this file gets damaged or deleted, expire_backups no longer knows the ages of the backups and refuses to work. In this case you can use the --reconstruct-state option to try to reconstruct the state from the backup dates. However, the accuracy of this reconstruction depends strongly on how rigorous you have been with making backups (it is only completely correct if the time between subsequent backups has always been exactly the same), so it's gener- ally a good idea not to tamper with the state file. OPTIONS
The expire_backups command accepts the following options: --quiet be really quiet --debug activate debugging output --version just print program version and exit --state <file> File to save state information in (default: ".expire_backups.dat") -n Dry run. Just show which backups would be deleted. --reconstruct-state Try to reconstruct a missing state file from backup dates. --use-s3qlrm Use s3qlrm command to delete backups. EXIT STATUS
expire_backups returns exit code 0 if the operation succeeded and 1 if some error occured. SEE ALSO
expire_backups is shipped as part of S3QL, http://code.google.com/p/s3ql/. COPYRIGHT
2008-2011, Nikolaus Rath 1.11.1 August 27, 2014 EXPIRE_BACKUPS(1)
All times are GMT -4. The time now is 03:41 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy