Sponsored Content
Full Discussion: Sun Cluster-Metaset problem
Homework and Emergencies Emergency UNIX and Linux Support Sun Cluster-Metaset problem Post 302407707 by madanmeer on Friday 26th of March 2010 06:41:50 AM
Old 03-26-2010
Sun Cluster-Metaset problem

Code:
Hi Guys
         i have two servers which are n cluster prd1 and prd2.. I  already have a metaset configured in it..My job was to create a 1.8 tb  LUN in my storage which is having Hardware raid5 pre-configured..i  Created 1.8 tb with raid 5 in my storage and it got detected in both t  servers prd1 prd2 and i got my DID id same in both..fine with that..its  like /dev/rdsk/c4t600A0B80004836600000057E4BAC04EFd0    /dev/did/rdsk/d8...wen i try to attach it to my metaset im not able to  do that because i think its raw disk and its takin default slice as  S0..so i tried to create a new file system on it but its sayin no such  file or directory...il post my output of that disk...slice 8 is present i  dont kno how it has come...

---------- Post updated at 05:41 AM ---------- Previous update was at 05:40 AM ----------

here is t files..

 

9 More Discussions You Might Find Interesting

1. High Performance Computing

SUN Cluster Vs Veritas Cluster

Dear All, Can anyone explain about Pros and Cons of SUN and Veritas Cluster ? Any comparison chart is highly appreciated. Regards, RAA (4 Replies)
Discussion started by: RAA
4 Replies

2. Solaris

unable to mount metaset on cluster node

Dear all, I have created a shared metaset(500gb) having 3 hosts in which 2 hosts are in cluster and 1 is non cluster. I have taken the ownership in cluster node from non cluster node but the problem is i am unable to mount the file system it is giving error "/dev/md/eccdb-ds/d100 or /eccdb-ds... (1 Reply)
Discussion started by: spandhan
1 Replies

3. Solaris

sun cluster 3.2 + oracle 11g HA problem C748634 ,C135343

Dear Cluster Guru, I had installed sc 3.2 on 2 box sunfire 240 (solaris 10 update 7 ,each node 2 gb ram) and use scsi juke box as share storage . i had test cluster application : 1. apache HA 2 oracleHA (oracle 9i) all running well (switch over between... (3 Replies)
Discussion started by: hadibn
3 Replies

4. Solaris

Sun Cluster-Metaset problem

Hi Guys i have two servers which are n cluster prd1 and prd2.. I already have a metaset configured in it..My job was to create a 1.8 tb LUN in my storage which is having Hardware raid5 pre-configured..i Created 1.8 tb with raid 5 in my storage and it got detected in both t servers prd1... (4 Replies)
Discussion started by: madanmeer
4 Replies

5. Solaris

Facing Problem with metaset in SVM

hi all, i am using solaris 5.10 on sun blade 150 and i am trying to configure diskset in sun volume manager. When i fire the following command, it says some rpc related error. bash-3.00# metaset -s kingston -a -h u15_9 metaset: u15_9: metad client create: RPC: Program not registered how to... (4 Replies)
Discussion started by: kingston
4 Replies

6. Solaris

Sun cluster and Veritas cluster question.

Yesterday my customer told me to expect a vcs upgrade to happen in the future. He also plans to stop using HDS and move to EMC. Am thinking how to migrate to sun cluster setup instead. My plan as follows leave the existing vcs intact as a fallback plan. Then install and build suncluster on... (5 Replies)
Discussion started by: sparcguy
5 Replies

7. Solaris

SVM metaset on 2 node Solaris cluster storage replicated to non-clustered Solaris node

Hi, Is it possible to have a Solaris cluster of 2 nodes at SITE-A using SVM and creating metaset using say 2 LUNs (on SAN). Then replicating these 2 LUNs to remote site SITE-B via storage based replication and then using these LUNs by importing them as a metaset on a server at SITE-B which is... (0 Replies)
Discussion started by: dn2011
0 Replies

8. Solaris

sun cluster problem

Hello everyone, Can you please help me understand what happened; here is the problem: We have a sun cluster composed by two nodes (E4900; sSolaris 9), suddenly one of them were inaccessible by PUTTY, however we were able to ping on it. The cluster haven't basculated and after 20 minutes the... (0 Replies)
Discussion started by: adilyos
0 Replies

9. Solaris

Sun cluster 4.0 - zone cluster failover doubt

Hello experts - I am planning to install a Sun cluster 4.0 zone cluster fail-over. few basic doubts. (1) Where should i install the cluster s/w binaries ?. ( global zone or the container zone where i am planning to install the zone fail-over) (2) Or should i perform the installation on... (0 Replies)
Discussion started by: NVA
0 Replies
i2o_bs(7D)							      Devices								i2o_bs(7D)

NAME
i2o_bs - Block Storage OSM for I2O SYNOPSIS
disk@local target id#:a through u disk@local target id#:a through u raw DESCRIPTION
The I2O Block Storage OSM abstraction (BSA, which also is referred to as block storage class) layer is the primary interface that Solaris operating environments use to access block storage devices. A block storage device provides random access to a permanent storage medium. The i2o_bs device driver uses I2O Block Storage class messages to control the block device; and provides the same functionality (ioctls, for example) that is present in the Solaris device driver like 'cmdk, dadk' on x86 for disk. The maximum size disk supported by i2o_bs is the same as what is available on x86. The i2o_bs is currently implemented version 1.5 of Intelligent IO specification. The block files access the disk using the system's normal buffering mechanism and are read and written without regard to physical disk records. There is also a "raw" interface that provides for direct transmission between the disk and the user's read or write buffer. A single read or write call usually results in one I/O operation; raw I/O is therefore considerably more efficient when many bytes are transmitted. The names of the block files are found in /dev/dsk; the names of the raw files are found in /dev/rdsk. I2O associates each block storage device with a unique ID called a local target id that is assigned by I2O hardware. This information can be acquired by the block storage OSM through I2O Block Storage class messages. For Block Storage OSM, nodes are created in /devices/pci#/pci# which include the local target ID as one component of device name that the node refers to. However the /dev names and the names in /dev/dsk and /dev/rdsk do not encode the local target id in any part of the name. For example, you might have the following: /devices/ /dev/dsk name --------------------------------------------------------------- /devices/pci@0,0/pci101e,0@10,1/disk@10:a /dev/dsk/c1d0s0 I/O requests to the disk must have an offset and transfer length that is a multiple of 512 bytes or the driver returns an EINVAL error. Slice 0 is normally used for the root file system on a disk, slice 1 is used as a paging area (for example, swap), and slice 2 for backing up the entire fdisk partition for Solaris software. Other slices may be used for usr file systems or system reserved area. Fdisk partition 0 is to access the entire disk and is generally used by the fdisk(1M) program. FILES
/dev/dsk/cndn[s|p]n block device /dev/rdsk/cndn[s|p]n raw device where: cn controller n dn instance number sn UNIX system slice n (0-15) pn fdisk partition(0) /kernel/drv/i2o_bs i2o_bs driver /kernel/drv/i2o_bs.conf Configuration file ATTRIBUTES
See attributes(5) for descriptions of the following attributes: +-----------------------------+-----------------------------+ |ATTRIBUTE TYPE |ATTRIBUTE VALUE +-----------------------------+-----------------------------+ |Architecture |x86 | +-----------------------------+-----------------------------+ SEE ALSO
fdisk(1M), format(1M)mount(1M),lseek(2), read(2), write(2), readdir(3C), vfstab(4), acct.h(3HEAD), attributes(5), dkio(7I) SunOS 5.10 21 Jul 1998 i2o_bs(7D)
All times are GMT -4. The time now is 07:18 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy