Sponsored Content
Operating Systems Solaris How to increase the /var (UFS) filesystem and root disk under veritas control? Post 303042103 by amity on Saturday 14th of December 2019 10:37:57 PM
Old 12-14-2019
How to increase the /var (UFS) filesystem and root disk under veritas control?

I need to increase the /var (UFS) filesystem and root disk under veritas control or root disk is encapsulated
Code:
# df -k /var
Filesystem            kbytes    used   avail capacity  Mounted on
/dev/vx/dsk/var      13241195 12475897 674524    96%    /var

# fstyp /dev/vx/dsk/var
ufs
# pkginfo -l VRTSvxvm
   PKGINST:  VRTSvxvm
      NAME:  VERITAS Volume Manager, Binaries
  CATEGORY:  system
      ARCH:  sparc
   VERSION:  3.5,REV=06.21.2002.23.14
   BASEDIR:  /
    VENDOR:  VERITAS Software
      DESC:  Virtual Disk Subsystem
    PSTAMP:  VERITAS-3.5s:21-Jun-2002
  INSTDATE:  Jul 13 2005 16:01
   HOTLINE:  800-342-0652
     EMAIL:  support@veritas.com
    STATUS:  completely installed
     FILES:      599 installed pathnames
                  24 shared pathnames
                   9 linked files
                  76 directories
                 336 executables
              158828 blocks used (approx)

# vxassist -g rootdg maxgrow var
Volume var can be extended by 124742928 to: 166315264 (83123Mb+1280 sectors)
# vxdg -g rootdg free
DISK         DEVICE       TAG          OFFSET    LENGTH    FLAGS
rootdisk     c1t0d0s2     c1t0d0       0         61055     -
rootdisk     c1t0d0s2     c1t0d0       161983467 124634705 -

Please let me know the right process as I know vxresize command and done increasing the vxfs filesystem many times but I think for increasing the ufs filesystem in veritas control has some different process.

Moderator's Comments:
Mod Comment
Use code tags please.

Last edited by Peasant; 12-15-2019 at 01:32 AM..
 

10 More Discussions You Might Find Interesting

1. Solaris

Veritas root disk mirroring

Hi there, My task is to replace the two 73 G disks with two 143 G disks , which has vxvm 4.1 running on it. I would like to know whether the steps iam following are correct. 1. Break the sub-disks, plexes of the root mirror. 2. Remove the sub-disks,plexes of the root mirror. 3. Remove one of... (10 Replies)
Discussion started by: Jartan
10 Replies

2. Solaris

how to increase the filesystem size under veritas control

Hi all, aloe:root-> df �k Filesystem 1024-blocks Used Available Capacity Mounted on /dev/vx/dsk/rootvol 8254263 2064133 6107588 26% / /proc 0 0 0 0% /proc mnttab ... (2 Replies)
Discussion started by: krishna176
2 Replies

3. Solaris

How to increase or decrease inode number of the particular UFS filesystem

Hi Gurus I want to know the command & tips regarding, how to increase or decrease inode number of the particular ufs filesystem. Is it possible to do it in a live/production environment. Regards (3 Replies)
Discussion started by: girish.batra
3 Replies

4. AIX

Remove internal disk from Veritas Control

I installed new internal disks in my p570. They will be part of a new AIX vg. Unfortunately, we have Veritas Volume Manager running on this AIX 5.2 ml 10 box. Veritas has grabbed control of the disks. I want AIX LVM to control the disks. I cannot get these disks free of Veritas: <lspv... (2 Replies)
Discussion started by: BobSmith
2 Replies

5. UNIX for Advanced & Expert Users

Veritas VxVM & Vxfs to SVM and UFS

Hi, I have 2 hosts with Veritas VxVM and VxFS (5.0 MP3_RP2). I need to use the application filesystem's luns from these 2 hosts and mount it on another 2 hosts that are running Solaris 9 09/05 and SVM. Is there resources online or has someone tried this? (0 Replies)
Discussion started by: xor
0 Replies

6. Solaris

Veritas VxVM & Vxfs to SVM and UFS

Hi, I have 2 hosts with Veritas VxVM and VxFS (5.0 MP3_RP2). I need to use the application filesystem's luns from these 2 hosts and mount it on another 2 hosts that are running Solaris 9 09/05 and SVM. Is there resources online or has someone tried this? (4 Replies)
Discussion started by: xor
4 Replies

7. Solaris

Increase root filesystem on solaris zone using zpool

I have a solaris zone of 12 GB and i have to increase the / filesystem to 31GB as requested. Earlier I had expanded filesystems other than / by setting quota to new value like "zfs set quota=new value mountpoint" but I am not sure whether its a good practice in zfs because by default in my... (5 Replies)
Discussion started by: vikkash
5 Replies

8. Solaris

Migration of system having UFS root FS with zones root to ZFS root FS

Hi All After downloading ZFS documentation from oracle site, I am able to successfully migrate UFS root FS without zones to ZFS root FS. But in case of UFS root file system with zones , I am successfully able to migrate global zone to zfs root file system but zone are still in UFS root file... (2 Replies)
Discussion started by: sb200
2 Replies

9. Solaris

how to restore backup from veritas vm in ufs file system

Hi all I have a DLT tape in that tape backup is there is in veritas volume format and i want to restore it in ufs file system how can i do it? right now i don't have veritas file system setup. i have only ufs file sysytem please help some production data is to be restore. backup was taken... (0 Replies)
Discussion started by: nikhil kasar
0 Replies

10. UNIX for Beginners Questions & Answers

How to extend a disk in veritas volume manager in veritas cluster?

Hi Experts, I wanted to extend a veritas file system which is running on veritas cluster and mounted on node2 system. #hastatus -sum -- System State Frozen A node1 running 0 A node2 running 0 -- Group State -- Group System Probed ... (1 Reply)
Discussion started by: Skmanojkum
1 Replies
ufs(7FS)																  ufs(7FS)

NAME
ufs - UFS file system SYNOPSIS
#include <sys/param.h> #include <sys/types.h> #include <sys/fs/ufs_fs.h> #include <sys/fs/ufs_inode.h> UFS is the default disk-based file system for the Solaris environment. The UFS file system is hierarchical, starting with its root direc- tory (/) and continuing downward through a number of directories. The root of a UFS file system is inode 2. A UFS file system's root con- tents replace the contents of the directory upon which it is mounted. Subsequent sections of this manpage provide details of the UFS file systems. State Flags (fs_state and fs_clean) UFS uses state flags to identify the state of the file system. fs_state is FSOKAY - fs_time. fs_time is the timestamp that indicates when the last system write occurred. fs_state is updated whenever fs_clean changes. Some fs_clean values are: FSCLEAN Indicates an undamaged, cleanly unmounted file system. FSACTIVE Indicates a mounted file system that has modified data in memory. A mounted file system with this state flag indi- cates that user data or metadata would be lost if power to the system is interrupted. FSSTABLE Indicates an idle mounted file system. A mounted file system with this state flag indicates that neither user data nor metadata would be lost if power to the system is interrupted. FSBAD Indicates that this file system contains inconsistent file system data. FSLOG Indicates that the file system has logging enabled. A file system with this flag set is either mounted or unmounted. If a file system has logging enabled, the only flags that it can have are FSLOG or FSBAD. A non-logging file system can have FSACTIVE, FSSTABLE, or FSCLEAN. It is not necessary to run the fsck command on unmounted file systems with a state of FSCLEAN, FSSTABLE, or FSLOG. mount(2) returns ENOSPC if an attempt is made to mount a UFS file system with a state of FSACTIVE for read/write access. As an additional safeguard, fs_clean should be trusted only if fs_state contains a value equal to FSOKAY - fs_time, where FSOKAY is a con- stant integer defined in the /usr/include/sys/fs/ufs_fs.h file. Otherwise, fs_clean is treated as though it contains the value of FSAC- TIVE. Extended Fundamental Types (EFT) Extended Fundamental Types (EFT) provide 32-bit user ID (UID), group ID (GID), and device numbers. If a UID or GID contains an extended value, the short variable (ic_suid, ic_sgid) contains the value 65535 and the corresponding UID or GID is in ic_uid or ic_gid. Because numbers for block and character devices are stored in the first direct block pointer of the inode (ic_db[0]) and the disk block addresses are already 32 bit values, no special encoding exists for device numbers (unlike UID or GID fields). Multiterabyte File System A multiterabyte file system enables creation of a UFS file system up to approximately 16 terabytes of usable space, minus approximately one percent overhead. A sparse file can have a logical size of one terabyte. However, the actual amount of data that can be stored in a file is approximately one percent less than one terabyte because of file system overhead. On-disk format changes for a multiterabyte UFS file system include: o The magic number in the superblock changes from FS_MAGIC to MTB_UFS_MAGIC. For more information, see the /usr/include/sys/fs/ufs_fs file. o The fs_logbno unit is a sector for UFS that is less than 1 terabyte in size and fragments for a multiterabyte UFS file system. UFS Logging UFS logging bundles the multiple metadata changes that comprise a complete UFS operation into a transaction. Sets of transactions are recorded in an on-disk log and are applied to the actual UFS file system's metadata. UFS logging provides two advantages: 1. A file system that is consistent with the transaction log eliminates the need to run fsck after a system crash or an unclean shutdown. 2. UFS logging often provides a significant performance improvement. This is because a file system with logging enabled converts multiple updates to the same data into single updates, thereby reducing the number of overhead disk operations. The UFS log is allocated from free blocks on the file system, and is sized at approximately 1 Mbyte per 1 Gbyte of file system, up to a maximum of 64 Mbytes. The log is continually flushed as it fills up. The log is also flushed when the file system is unmounted or as a result of a lockfs command. Mounting UFS File Systems You can mount a UFS file system in various ways using syntax similar to the following: 1. Use mount from the command line: # mount -F ufs /dev/dsk/c0t0d0s7 /export/home 2. Include an entry in the /etc/vfstab file to mount the file system at boot time: /dev/dsk/c0t0d0s7 /dev/rdsk/c0t0d0s7 /export/home ufs 2 yes - For more information on mounting UFS file systems, see mount_ufs(1M). See attributes(5) for a description of the following attributes: +-----------------------------+-----------------------------+ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | +-----------------------------+-----------------------------+ |Interface Stability |Unstable | +-----------------------------+-----------------------------+ df(1M), fsck(1M), fsck_ufs(1M), fstyp(1M), mkfs_ufs(1M), newfs(1M), ufsdump(1M), ufsrestore(1M), tunefs(1M), mount(2), attributes(5) NOTES
For information about internal UFS structures, see newfs(1M) and mkfs_ufs(1M). For information about the ufsdump and ufsrestore commands, see ufsdump(1M), ufsrestore(1M), and /usr/include/protocols/dumprestore.h. If you experience difficulty in allocating space on the ufs filesystem, if may be due to framentation. Fragmentation can occur when you do not have sufficient free blocks to satisfy an allocation request even though df(1M) indicates that enough free space is available. (This may occur because df only uses the available fragment count to calculate available space, but the file system requires contiguous sets of fragments for most allocations). If you suspect that you have exhausted contiguous fragments on your file system, you can use the fstyp(1M) utility with the -v option. In the fstyp output, look at the nbfree (number of blocks free) and nffree (number of frag- ments free) fields. On unmounted filesystems, you can use fsck(1M) and observe the last line of output, which reports, among other items, the number of fragments and the degree of fragmentation. To correct a fragmentation problem, run ufsdump(1M) and ufsrestore(1M) on the ufs filesystem. 25 Jun 2003 ufs(7FS)
All times are GMT -4. The time now is 11:07 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy