Sponsored Content
Full Discussion: Disk I/O Issue
Operating Systems AIX Disk I/O Issue Post 302312661 by ukatru on Saturday 2nd of May 2009 02:05:34 PM
Old 05-02-2009
Disk I/O Issue

We have a filesystem which contains 8 hard disks but i am facing disk I/O issue becuase data is not spreading across all the disks.Is there any way i can check how data is spreading and any parameter we need to change to spread ata across all disks.

OS--AIX 5.3
 

10 More Discussions You Might Find Interesting

1. UNIX for Advanced & Expert Users

Solaris 8 disk/mirroring issue

Hello! I recently inherited system administration duties for a SUN v880 box. The system has 6 physical hard disks.. In doing some basic maintenance, I found they're configured for mirroring. I ran the metastat and metadb commands, and many of the mirrors are showing they are in need of... (5 Replies)
Discussion started by: ghuber
5 Replies

2. AIX

disk issue

Hi, I have a AIX 4.3 box here with problems cause of a disk in below VG. volume group: workvg lspv hdisk4 000166789869ab2d workvg hdisk5 000166789869b96b workvg now hdisk4 disk has failed and cause the quorum was set, workvg became varied off. I have to replace the... (2 Replies)
Discussion started by: karthikosu
2 Replies

3. UNIX for Advanced & Expert Users

partition disk issue

hi guys, I've got a strange issue, may be one of you has experienced this. SunOS 5.10 Generic_118833-33 sun4u sparc SUNW,Sun-Fire-V440 everything is mirrored. My issue is that I have a umpty directory but seems to have data on. Let me show you # df -h /data Filesystem size used... (10 Replies)
Discussion started by: moustik
10 Replies

4. Solaris

Disk Suite issue

Solaris 9 We had a problem server where only root was not mirrored (before my time). When I tried to mirror it, the live root slice bailed with errors at 97% so it couldn't be mirrored. It's a matched pair of boxes (nfs1 and nfs2) and they are interchangeable with regards to the NFS... (0 Replies)
Discussion started by: BOFH
0 Replies

5. AIX

Disk I/O Issue using LVM

We have a filesystem which contains 8 hard disks but i am facing disk I/O issue becuase data is not spreading across all the disks.Is there any way i can check how data is spreading and any parameter we need to change to spread ata across all disks. OS--AIX 5.3 (1 Reply)
Discussion started by: ukatru
1 Replies

6. Filesystems, Disks and Memory

Issue available disk space while using xdd

Good morning, I seem to be running into an issue with some drives I have attached to my solaris server. The drives are attached correctly, the partitions are arranged with fdisk, the ext3 filesystem is setup using mkfs, and finally the drive is mounted. When I use xdd to perform read/write... (3 Replies)
Discussion started by: mrpogo07
3 Replies

7. UNIX Desktop Questions & Answers

Issue with disk space usage

Issue with disk space usage I have the following line in my "df -h" output: Filesystem Size Used Avail Capacity Mounted on /dev/ad4s1a 496M 495M -39M 109% / What is the issue with having 9% excess utilisation? How can I find out what this partition is... (2 Replies)
Discussion started by: figaro
2 Replies

8. Solaris

T5220 disk mapping issue

Hi, More a Sun T5220 problem then a Solaris 10 problem, but perhaps someone had a similar issue. For starters the output with 1 disk in slot 0 of the server. It points to PhyNum 5, where I would expect PhyNum 0. {0} ok probe-scsi MPT Version 1.05, Firmware Version 1.22.00.00 Target... (2 Replies)
Discussion started by: ejdv
2 Replies

9. Solaris

Server disk issue need help

Hello all, Our Solaris 9? Sun Fire 480R backup server(in another city) is throwing disk errors such as these repeatedly. WARNING: vxvm:vxio: Subdisk rootdisk-02 block 24037056: Uncorrectable read error WARNING: vxvm:vxio: Subdisk rootdisk-02 block 7767072: Uncorrectable write error ... (18 Replies)
Discussion started by: RyanV
18 Replies

10. Solaris

Solaris 11 disk issue

I have 2 disks in my system.I recently added a zpool to the disk, but today I changed my mind and deleted the zpool , zpool destroy -f extra The zpool is now deleted and I want to partition the disk, so I delete the only partition on the disk. Now when I run format again, format... (13 Replies)
Discussion started by: cbtshare
13 Replies
vxdarestore(1M) 														   vxdarestore(1M)

NAME
vxdarestore - restore simple or nopriv disk access records SYNOPSIS
/etc/vx/bin/vxdarestore DESCRIPTION
The vxdarestore utility is used to restore persistent simple or nopriv disk access (da) records that have failed due to changing the naming scheme used by vxconfigd from c#t#d#-based to enclosure-based. The use of vxdarestore is required if you use the vxdiskadm command to change from the c#t#d#-based to the enclosure-based naming scheme. As a result, some existing persistent simple or nopriv disks go into the "error" state and the VxVM objects on those disks fail. vxdarestore may be used to restore the disk access records that have failed. The utility also recovers the VxVM objects on the failed disk access records. Note: vxdarestore may only be run when vxconfigd is using the enclosure-based naming scheme. Note: You can use the command vxdisk list da_name to discover whether a disk access record is persistent. The record is non-persistent if the flags field includes the flag autoconfig; otherwise it is persistent. The following sections describe how to use the vxdarestore utility under various conditions. Persistent Simple/Nopriv Disks in the rootdg Disk Group If all persistent simple or nopriv disks in the rootdg disk group go into the "error" state, use the following procedure: 1. Use the vxdiskadm command to change back to the c#t#d# based naming scheme. 2. Either shut down and reboot the host, or run the following command: vxconfigd -kr reset 3. If you want to use the enclosure-based naming scheme, add a non-persistent simple disk to the rootdg disk group, use vxdiskadm to change to the enclosure-based naming scheme, and then run vxdarestore. Note: If not all the disks in rootdg go into the error state, simply running vxdarestore restores those disks in the error state and the objects that that they contain. Persistent Simple/Nopriv Disks in Disk Groups other than rootdg If all disk access records in an imported disk group consist only of persistent simple and/or nopriv disks, the disk group is put in the "online dgdisabled" state after changing to the enclosure-based naming scheme. For such disk groups, perform the following steps: 1. Deport the disk group using the following command: vxdg deport diskgroup 2. Run the vxdarestore command. 3. Re-import the disk group using the following command: vxdg import diskgroup NOTES
Use of the vxdarestore command is not required in the following cases: o If there are no persistent simple or nopriv disk access records on an HP-UX host. o If all devices on which simple or nopriv disks are present are not automatically configurable by VxVM. For example, third-party drivers export devices that are not automatically configured by VxVM. VxVM objects on simple/nopriv disks created from such disks are not affected by switching to the enclosure based naming scheme. The vxdarestore command does not handle the following cases: o If the enclosure-based naming scheme is in use and the vxdmpadm command is used to change the name of an enclosure, the disk access names of all devices in that enclosure are also changed. As a result, any persistent simple/nopriv disks in the enclosure are put into the "error" state, and VxVM objects configured on those disks fail. o If the enclosure-based naming scheme is in use and the system is rebooted after making hardware configuration changes to the host. This may change the disk access names and cause some persistent simple/nopriv disks to be put into the "error" state. o If the enclosure-based naming scheme is in use, the device discovery layer claims some disks under the JBOD category, and the vxdd- ladm rmjbod command is used to remove remove support for the JBOD category for disks from a particular vendor. As a result of the consequent name change, disks with persistent disk access records are put into the "error" state, and VxVM objects configured on those disks fail. EXIT CODES
A zero exit status is returned if the operation is successful or if no actions were necessary. An exit status of 1 is returned if vxdare- store is run while vxconfigd is using the c#t#d# naming scheme. An exit status of 2 is returned if vxconfigd is not running. SEE ALSO
vxconfigd(1M), vxdg(1M), vxdisk(1M), vxdiskadm(1M), vxdmpadm(1M), vxintro(1M), vxreattach(1M), vxrecover(1M) VxVM 5.0.31.1 24 Mar 2008 vxdarestore(1M)
All times are GMT -4. The time now is 06:07 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy