Sponsored Content
Full Discussion: NFS mounts query
Operating Systems Linux Red Hat NFS mounts query Post 302875071 by ikn3 on Monday 18th of November 2013 12:16:31 PM
Old 11-18-2013
Thanks, Peasant,
I guess that will be the simplest solution to this problem. Was trying to make sure this could be done automatically. But guess have to do manual mounts when we face issues.

With the current setup. There is a problem with the backups not initiating from node2. Even when both nodes are running only node1 can start the backups. Not sure what is wrong. My fstab entries for both nodes for reference.

Node1: /dev/mapper/mpath7 /backup ext3 defaults 1 2
Node2: 172.16.184.220:/backup /backup nfs defaults 1 2

The IP above is for node1.
 

10 More Discussions You Might Find Interesting

1. Solaris

new mounts

hi, i have currently below mounts in solaris box and i want to create new mount points. please let me know how can i do it? bash-3.00# df -h Filesystem size used avail capacity Mounted on / 1000M 350M 609M 37% / /dev 1000M 350M ... (3 Replies)
Discussion started by: rags_s11
3 Replies

2. UNIX for Dummies Questions & Answers

Mounts

Hi, I'm new to Linux and to this forum too. Now, I need some info. I have an application which writes some data onto one mount(logs and others). Now, I want to have some convention or script where if the mount(where the application is writing data) reaches certain amount of memory or if it... (1 Reply)
Discussion started by: krisdasword
1 Replies

3. AIX

NFS mounts and user permissions

We need to allow ordinary users to preform NFS mounts on a AIX server without giving them root access to the server. Is there a way to give an ordinary users root access on a tem basis or a script to allow them to preform NFS mounts? (4 Replies)
Discussion started by: daveisme
4 Replies

4. Solaris

Query regarding NFS Mount...

hi All, This is the first time that i would be doing a NFS mount. Its solaris 10 machine on which i would be doing this. Can any of you gurus help me with the steps that should be taken. (3 Replies)
Discussion started by: sankasu
3 Replies

5. AIX

nfs query

Hi, I have shared the file system and given access to 2 clients Examples client A client b but after sharing and mounting on client side when i am issue command share it is showing me only single ip of client A , it does.nt showing me ip address of client b and After mouting as RW ,... (6 Replies)
Discussion started by: manoj.solaris
6 Replies

6. Solaris

Max. number of NFS mounts

Hi, I was wondering, whether there is a limit regarding the max number of nfs mounts in Oracle Solaris 10 (newest update). The data center plans to migrate from a fibre channel based storage environment (hitachi) to a nfs based storage environment (netapp). Regarding the Solaris 10 database... (1 Reply)
Discussion started by: schms
1 Replies

7. Red Hat

NFS Query

Hi, My query is on NFS, I have shared the /test folder on Server A, whose owner is user "A" uid=101, gid=202 and mounted on Server B as /test1, whose owner is "A" uid=204, gid=204, Now My query is if uid and gid of users of both servers are different how user A will have full access to... (1 Reply)
Discussion started by: manoj.solaris
1 Replies

8. Shell Programming and Scripting

find command exclude nfs mounts

Gentleman and Ladies, I am having some difficulty in regards to the find utility. I am trying to not traverse nfs mounts when searching for files or directories. This command does not seem to work. It still returns directories that live on nfs shares. The man page says: -fstype type... (6 Replies)
Discussion started by: jaysunn
6 Replies

9. Solaris

NFS mounts not automounting on boot

I have several Solaris 11.2 zones. when I reboot them I have to go in and do mountall to mount the NFS mounts. any ideas where to troubleshoot why they are not automounting? (2 Replies)
Discussion started by: os2mac
2 Replies

10. AIX

Sysmirror 7.1.3 Resource Group NFS mounts

Hello all, I'm working to fix a two-node SysMirror cluster that uses NFS mounts from a NetApp appliance as the data repository. Currently all the NFS mounts/unmounts are called from the application controller scripts, and since file collection isn't currently working, (One fight at at time... (3 Replies)
Discussion started by: ZekesGarage
3 Replies
STOREBACKUPDEL(1)					User Contributed Perl Documentation					 STOREBACKUPDEL(1)

NAME
storeBackupDel.pl - this program deletes backups created by storeBackup SYNOPSIS
storeBackupDel.pl [-f configFile] [--print] [-b backupDirectory] [-S series] [--doNotDelete] [--deleteNotFinishedDirs] [-L lockFile] [--keepAll timePeriod] [--keepWeekday entry] [--keepFirstOfYear] [--keepLastOfYear] [--keepFirstOfMonth] [--keepLastOfMonth] [--keepFirstOfWeek] [--keepLastOfWeek] [--keepDuplicate] [--keepMinNumber] [--keepMaxNumber] [-l logFile [--plusLogStdout] [--suppressTime] [-m maxFilelen] [[-n noOfOldFiles] | [--saveLogs] [--compressWith compressprog]] WARNING
!!! USAGE IN PARALLEL WITH storeBackup.pl CAN DESTROY YOUR BACKUPS !!! OPTIONS
--file, -f configuration file (instead of parameters) --print print configuration read from configuration file and stop --backupDir, -b top level directory of all backups (must exist) --series, -S directory of backup series same parameter as in storeBackup / relative path from backupDir, default is 'default' --lockFile, -L lock file, if exists, new instances will finish if an old is already running, default is $lockFile --doNotDelete test only, do not delete any backup --deleteNotFinishedDirs delete old backups which where not finished this will not happen if doNotDelete is set --keepAll keep backups which are not older than the specified amount of time. This is like a default value for all days in --keepWeekday. Begins deleting at the end of the script the time range has to be specified in format 'dhms', e.g. 10d4h means 10 days and 4 hours default = $keepAll; --keepWeekday keep backups for the specified days for the specified amount of time. Overwrites the default values chosen in --keepAll. 'Mon,Wed:40d Sat:60d10m' means: keep backups of Mon and Wed 40days + 5mins keep backups of Sat 60days + 10mins keep backups of the rest of the days like spcified in --keepAll (default $keepAll) if you also use the 'archive flag' it means to not delete the affected directories via --keepMaxNumber: a10d4h means 10 days and 4 hours and 'archive flag' e.g. 'Mon,Wed:a40d Sat:60d10m' means: keep backups of Mon and Wed 40days + 5mins + 'archive' keep backups of Sat 60days + 10mins keep backups of the rest of the days like specified in --keepAll (default $keepAll) --keepFirstOfYear do not delete the first backup of a year format is timePeriod with possible 'archive flag' --keepLastOfYear do not delete the last backup of a year format is timePeriod with possible 'archive flag' --keepFirstOfMonth do not delete the first backup of a month format is timePeriod with possible 'archive flag' --keepLastOfMonth do not delete the last backup of a month format is timePeriod with possible 'archive flag' --firstDayOfWeek default: 'Sun'. This value is used for calculating --keepFirstOfWeek and --keepLastOfWeek --keepFirstOfWeek do not delete the first backup of a week format is timePeriod with possible 'archive flag' --keepLastOfWeek do not delete the last backup of a week format is timePeriod with possible 'archive flag' --keepDuplicate keep multiple backups of one day up to timePeriod format is timePeriod, 'archive flag' is not possible default = $keepDuplicate; --keepMinNumber Keep that miminum of backups. Multiple backups of one day are counted as one backup. Default is 10. --keepMaxNumber Try to keep only that maximum of backups. If you have more backups, the following sequence of deleting will happen: - delete all duplicates of a day, beginning with the old once, except the oldest of every day - if this is not enough, delete the rest of the backups beginning with the oldest, but *never* a backup with the 'archive flag' or the last backup --keepRelative, -R Alternative deletion scheme. If you use this option, all other keep options are ignored. Preserves backups depending on their *relative* age. Example: -R '1d 7d 2m 3m' will (try to) ensure that there is always - One backup between 1 day and 7 days old - One backup between 5 days and 2 months old - One backup between 2 months and 3 months old If there is no backup for a specified timespan (e.g. because the last backup was done more than 2 weeks ago) the next older backup will be used for this timespan. --logFile, -l log file (default is STDOUT) --plusLogStdout if you specify a log file with --logFile you can additionally print the output to STDOUT with this flag --suppressTime suppress output of time in logfile --maxFilelen, -m maximal length of file, default = 1e6 --noOfOldFiles, -n number of old log files, default = 5 --saveLogs save log files with date and time instead of deleting the old (with [-noOldFiles]) --compressWith compress saved log files (e.g. with 'gzip -9') default is 'bzip2' COPYRIGHT
Copyright (c) 2003-2008 by Heinz-Josef Claes (see README). Published under the GNU General Public License v3 or any later version perl v5.14.2 2012-06-16 STOREBACKUPDEL(1)
All times are GMT -4. The time now is 06:49 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy