Sponsored Content
Full Discussion: VXVM in POWERHA 6.1
Operating Systems AIX VXVM in POWERHA 6.1 Post 302911107 by sunnybee on Wednesday 30th of July 2014 05:52:51 AM
Old 07-30-2014
Hi,

I have attached logs. VG created in VXVM not in LVM. While adding VG in Resource group VGs(vxvm) are not listing.

hdisk19,hdisk20 are using for creating VG.


regards
 

10 More Discussions You Might Find Interesting

1. AIX

strange route issue in powerha 5.4?

hi all we upgraded hacmp(powerha) from 5.2 to 5.4 recently . during the failover testing, we found a strange network issue. after standby node took service ip address (172.15.100.8) online at standby NIC, we were able to log in the standby node by telnet 172.15.100.8 which stays at... (1 Reply)
Discussion started by: rs6000er
1 Replies

2. AIX

PowerHA - assistant can't find network

I've configured multiple clusters using PowerHA 5.5, but never using the 'Two-node cluster configuration assistant'. I removed all the clusters, clean setup, and gave the assistant a try. I've filled in all the necessary options: Communication Path to Takeover Node Application Server ... (11 Replies)
Discussion started by: Koen87
11 Replies

3. AIX

PowerHA, same subnet for persistent and service ip

I am new in AIX and please forgive my poor english. I know that AIX allow same subnet IPs for different interfaces, which result in multipath routing / route striping. My question is, is there any best practice for the persistent and service IP with same subnet to stay on same interface, or... (5 Replies)
Discussion started by: skeyeung
5 Replies

4. AIX

PowerHA Disk on VIO Server

Hello AIX GURU's Can anybody tell me the steps to crate shared VG (enhanced concurent) for my Lpars from VIO server? my questions are: 1. Should I crate Enhanced Concurent VG in VIO and map it using virtual Scsi to Lpar? or 2. Can I just create virtual SCSI in VIO and map to Lpar and... (1 Reply)
Discussion started by: Vit0_Corleone
1 Replies

5. AIX

PowerHA with 3rd node for communication

Looking to find PowerHA with 3rd node used for communication (voting node) only. Does anyone use such configuration with IBM products? (2 Replies)
Discussion started by: gogogo
2 Replies

6. AIX

PowerHA HACMP on VIOS servers

Few questions regarding Power HA ( previously known as HACMP) and VIOS POWERVM IVM ( IBM Virtualization I/O Server ) Is it possible to create HACMP cluster between two VIOS servers Physical Machine_1 VIOS_SERVER_1 LPAR_1 SHARED_DISK_XX VIOS_SERVER_2 Physical Machine_2 LPAR_2... (6 Replies)
Discussion started by: filosophizer
6 Replies

7. AIX

Powerha on AIX 6.1.9

hello ive installed powerha 7.1.3 on two servers aix 6.1.9 6100-09-03-1415 work with dmx4 EMC storgae. after sync cluster ( terminate with OK ) ive see that the repository disk upper only in one machine : hdiskpower60 00c7f6b59fc60d9d caavg_private active... (1 Reply)
Discussion started by: ariec
1 Replies

8. AIX

PowerHA; Import flash copy of VG?

Hey guys, wondering if this is possible to accomplish. PowerHA SystemMirror7.1 active/passive cluster. Restored a clustered system in test and upgraded to the latest version of PowerHA. Things are working great. However in the interest of time during cut over to the new system, I was hoping to... (4 Replies)
Discussion started by: j_aix
4 Replies

9. AIX

PowerHa 7.1 Migration check failed

Hello Guys, I was trying to upgrade the PowerHA from 6.1.0.9 to 7.1.3 in prod cluster. Have snapshot migration and run climgcheck. 1. Run climgcheck on node1(kul1pqcmur01) and it success 2. Upgrade the cluster sw 3. Run climgcheck on node2(kul1pqcmur02) and it failed with below errors ... (6 Replies)
Discussion started by: novaothers
6 Replies

10. AIX

Upgrading PowerHA from 7.1.0 to 7.1.3

Hi All, As per the IBM upgrade/support matrix IBM Techdocs Technote: PowerHA for AIX Version Compatibility Matrix we can't do online upgrade or rolling migration from powerha v7.1.0 to v7.1.3. on AIX61_TL9_SP4, So we are following the steps as below ... 1 ) Bring down the cluster 2 )... (2 Replies)
Discussion started by: linux.amrit
2 Replies
vxsited(1M)															       vxsited(1M)

NAME
vxsited - site monitoring daemon SYNOPSIS
/etc/vx/bin/vxsited [mail_address...] DESCRIPTION
The vxsited daemon monitors Veritas Volume Manager (VxVM) for disks being attached, and reattaches a detached site if the disks that belong to that site become accessible. vxsited analyzes the output of the vxnotify command, and waits for a failed disk to attach. When a disk is attached, vxsited attempts to online the disk, and tries to reattach the failed site. If a site is successfully reattached, vxsited starts recovery using vxrecover, and sends mail to root (by default) or to other specified users. Mail Notification By default, vxsited sends mail to root with information about the disk status of any attempts to reattach the site. To send mail to other users, add the user login name to the line that starts vxsited in the startup script, /etc/init.d/vxvm-recover, and reboot the system. For example, if the line appears as: nohup vxsited root & and you want mail also to be sent to user1 and user2, change the line to read: nohup vxsited root user1 user2 & Alternatively, kill the vxsite process, and restart it from the command line with the required mail addresses as arguments. The mail notification has a format that is similar to the following: Subject : Volume Manager site reattach on host hostname Reattached site sitename in disk-group diskgroup Reattachment Procedure If a disk from a detached site becomes accessible again, vxsited checks whether the relocation daemon, vxrelocd, is running. If vxrelocd is running, vxsited attempts to reattach the site. The relocation daemon can then try to relocate the failed subdisks using space on the available disks in the disk group. If the failed objects are successfully relocated, vxrelocd changes the state of the site to RECOVER, and starts the recovery of volumes at the site. When all the plexes at a site have been recovered, the plexes are put into the ACTIVE state, and the state of the site is set to ACTIVE. If vxrelocd is not running, vxsited only reattaches a site when all the disks from that site become accessible. After successfully reat- taching a site, vxsited changes the site state to ACTIVE, and initiates recovery using vxrecover. When all the plexes from a site have been recovered, the plexes are put into the ACTIVE state, and the state of the site is set to ACTIVE. vxsited does not attempt to reattach a site that has been explicitly detached by an administrator. The state OFFLINE is set for sites that have been detached by using the following command: vxdg -g dg_name detachsite sitename Disabling vxsited If you do not want a site to be recovered automatically, kill the vxsited daemon, and prevent it from restarting. To kill the daemon, run the following command from the command line, and locate the process table entry for vxsited: ps -ef Execute the command: kill -9 PID Substitute the process ID of the vxsited process for PID. To prevent vxsited from being restarted, comment out the line that starts vxsited in the startup script /sbin/init.d/vxvm-recover. FILES
/sbin/init.d/vxvm-recover The startup file for vxsited. SEE ALSO
kill(1), mailx(1), ps(1), vxdg(1M), vxrelocd(1M), vxintro(1M), vxnotify(1M), vxrecover(1M) VxVM 5.0.31.1 24 Mar 2008 vxsited(1M)
All times are GMT -4. The time now is 09:40 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy