Sponsored Content
Full Discussion: Ufsdump
Top Forums UNIX for Dummies Questions & Answers Ufsdump Post 3712 by loitschix on Tuesday 10th of July 2001 08:52:06 AM
Old 07-10-2001
MySQL ufsdump...

hi.

it`s no problem to make an ufs-dump to a mounted filesystem, but the meaning is this: if you want a "consistent" state of datas, no process have to write to a file you wanna back up; so when you are really shure NO process is writing to the datas you want to back up, you can run "ufs-dump" unmounted. (check if there are processes with "fuser <dir>" or freeware "lsof").

when you are not shure, go to single-user mode to "kick" all user-processes from the system. but when you wanna be really shure, umount the filesystems to backup, cause there are always systemprocesses writing to filesystems like /var.

best regards,
alex...
 

10 More Discussions You Might Find Interesting

1. UNIX for Dummies Questions & Answers

netinstall with ufsdump

Hello all, I was wondering is it possible to install Solaris over the network whit jumpstart and just ufsretore a dump file? The clinet's disk would totally be empty. I know that I can use the rules files, would I put the command to uferestore from the server in the pre install rules file. I... (0 Replies)
Discussion started by: larryase
0 Replies

2. UNIX for Dummies Questions & Answers

ufsdump

hi i would like to backup my OS what is the exact command to do? beside backup to tapes can i back up to a directory in another server? will the command be the same? thanks. :confused: (4 Replies)
Discussion started by: legato
4 Replies

3. Solaris

ufsdump

Hi, I want to how to take backup of File system From Disk to Disk using ufsdump? Wheather we can take the backup of root file system from one disk to another if tape drive is not available? wheather it is reliable to restore root file system? (2 Replies)
Discussion started by: manoj.solaris
2 Replies

4. UNIX for Dummies Questions & Answers

ufsdump issue

I ran this dump (as root) and got the following............. # ufsdump 0f /dev/rmt/0n / DUMP: Writing 32 Kilobyte records DUMP: Date of this level 0 dump: Tue 11 Dec 2007 06:07:35 AM PST DUMP: Date of last level 0 dump: the epoch DUMP: Dumping /dev/rdsk/c2t1d0s0 (gambler:/) to... (1 Reply)
Discussion started by: shorty
1 Replies

5. UNIX for Advanced & Expert Users

ufsdump and restore

I just completed a level 0 ufsdump of the following files:- ufsdump 0uf /dev/rmt/0n / ufsdump 0uf /dev/rmt/0n /usr ufsdump 0uf /dev/rmt/0n /export/home ufsdump 0uf /dev/rmt/0n /oracle ufsdump 0uf /dev/rmt/0n /sapr3 I need to restore ALL the files onto a different machine and continue... (1 Reply)
Discussion started by: charleywasharo
1 Replies

6. Solaris

Using ufsdump and ufsrestore

HI Gurus, I have a sunfire V445 server running SAP ECC 6.0 with an Oracle database on Solaris 9 (SunOS 5.9). I recently completed a ufsdump to tape of the following files:- /, /usr, /oracle, /export, /sapr3, I want to restore these files from tape onto a different server of the same... (5 Replies)
Discussion started by: charleywasharo
5 Replies

7. UNIX for Dummies Questions & Answers

Ufsdump

Hi, I have 5 soraris boxes and i am trying to backup all to SAN.Which backup is more suitable tar or ufsdump? Also pls mention what are the important folders i need to back up. Thanks (3 Replies)
Discussion started by: solaris5.10
3 Replies

8. UNIX for Dummies Questions & Answers

How to do a ufsdump remote?

People i have 2 sun sparc solaris 10. I have one filesystem to 98% and i need to do a ufsdump of that filesystem. THe problem is that i don't have space to allocate the dump of the filesystem. Thats why i want to do it from the other sun and bring the ufs dump to de /FREE filesystem. ... (4 Replies)
Discussion started by: enkei17
4 Replies

9. Solaris

ufsdump

Experts, Before patching am advised to take backup : so am going with: ufsdump -0uf /dev/rmt0 / ---> to take the whole / bkp to tape. some servers have /var in diff slice, In this case whether i need to take backup of /var also in tape? ufsdump -ouf /dev/rmt0 /var ---> to take... (3 Replies)
Discussion started by: fizan
3 Replies

10. Solaris

Ufsdump and migration

Hi all, I have a Solaris 10 running on a M4000. I wanted to migrate it to M10 q1) can I do a ufsdump on / in the m4000 and restore it on the m10 ? q2) how do we ufsrestore on the naked M10 without any OS being install/running in the 1st place ? q3) if the answer to the... (5 Replies)
Discussion started by: javanoob
5 Replies
EXPIRE_BACKUPS(1)						       S3QL							 EXPIRE_BACKUPS(1)

NAME
expire_backups - Intelligently expire old backups SYNOPSIS
expire_backups [options] <age> [<age> ...] DESCRIPTION
The expire_backups command intelligently remove old backups that are no longer needed. To define what backups you want to keep for how long, you define a number of age ranges. expire_backups ensures that you will have at least one backup in each age range at all times. It will keep exactly as many backups as are required for that and delete any backups that become redundant. Age ranges are specified by giving a list of range boundaries in terms of backup cycles. Every time you create a new backup, the existing backups age by one cycle. Example: when expire_backups is called with the age range definition 1 3 7 14 31, it will guarantee that you always have the following backups available: 1. A backup that is 0 to 1 cycles old (i.e, the most recent backup) 2. A backup that is 1 to 3 cycles old 3. A backup that is 3 to 7 cycles old 4. A backup that is 7 to 14 cycles old 5. A backup that is 14 to 31 cycles old Note If you do backups in fixed intervals, then one cycle will be equivalent to the backup interval. The advantage of specifying the age ranges in terms of backup cycles rather than days or weeks is that it allows you to gracefully handle irregular backup intervals. Imagine that for some reason you do not turn on your computer for one month. Now all your backups are at least a month old, and if you had specified the above backup strategy in terms of absolute ages, they would all be deleted! Specifying age ranges in terms of backup cycles avoids these sort of problems. expire_backups usage is simple. It requires backups to have names of the forms year-month-day_hour:minute:seconds (YYYY-MM-DD_HH:mm:ss) and works on all backups in the current directory. So for the above backup strategy, the correct invocation would be: expire_backups.py 1 3 7 14 31 When storing your backups on an S3QL file system, you probably want to specify the --use-s3qlrm option as well. This tells expire_backups to use the s3qlrm command to delete directories. expire_backups uses a "state file" to keep track which backups are how many cycles old (since this cannot be inferred from the dates con- tained in the directory names). The standard name for this state file is .expire_backups.dat. If this file gets damaged or deleted, expire_backups no longer knows the ages of the backups and refuses to work. In this case you can use the --reconstruct-state option to try to reconstruct the state from the backup dates. However, the accuracy of this reconstruction depends strongly on how rigorous you have been with making backups (it is only completely correct if the time between subsequent backups has always been exactly the same), so it's gener- ally a good idea not to tamper with the state file. OPTIONS
The expire_backups command accepts the following options: --quiet be really quiet --debug activate debugging output --version just print program version and exit --state <file> File to save state information in (default: ".expire_backups.dat") -n Dry run. Just show which backups would be deleted. --reconstruct-state Try to reconstruct a missing state file from backup dates. --use-s3qlrm Use s3qlrm command to delete backups. EXIT STATUS
expire_backups returns exit code 0 if the operation succeeded and 1 if some error occured. SEE ALSO
expire_backups is shipped as part of S3QL, http://code.google.com/p/s3ql/. COPYRIGHT
2008-2011, Nikolaus Rath 1.11.1 August 27, 2014 EXPIRE_BACKUPS(1)
All times are GMT -4. The time now is 03:19 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy