Sponsored Content
Operating Systems Solaris Where are my other disks on T5220 ? Post 302987839 by Corona688 on Friday 16th of December 2016 01:03:36 PM
Old 12-16-2016
A true hardware RAID will need to be told what you want it to use the disks for before it will show them, even when you're just asking for JBOD.
 

10 More Discussions You Might Find Interesting

1. Solaris

kstat on T5220

We used kstat on netra240/440 to check if the link is full duplex. So we do something like: bash-3.00# kstat -pc net bge:1::/'link_duplex'/ | cut -f2 2 The output 2 indicates the link is full duplex. This kind of check worked all right for Netra 240/440 systems. However, the nxge... (2 Replies)
Discussion started by: illcar
2 Replies

2. Solaris

ILOM and sc on T5220

I 've 2 different T5220 and both are showing different prompts when i ssh to the console. 1. Shows sc> ALOM 2. Shows -> ILOM as far as i know the T5220 comes with ILOM , so why i 'm seeing 2 different type of console for same tyep of hardware ? and i found that we... (6 Replies)
Discussion started by: fugitive
6 Replies

3. Solaris

SC on T5220 not responding

Hi, SC on one of my T5220 is not responding .. it does not show not let me type anything .. what could be the reason for it ? though when i powercycle the box from sc it shows me the system coming up but after that does not let me login thru sc sc> console -f Enter #. to return to ALOM. (2 Replies)
Discussion started by: fugitive
2 Replies

4. Solaris

LDOMs on T5220

We're just about to migrate a large application from an Enterprise 6900 to a pair of T5220s utilising LDOMs. Does anyone have any experience of LDOMs on this kit and can provide any recommendations or pitfalls to avoid? I've heard that use of LDOMs can have an impact on I/O speeds as it's all... (9 Replies)
Discussion started by: JerryHone
9 Replies

5. Solaris

T5220 server

Hi, We have two T5220 servers in our datacentre one server is showing 64 processors online and another showing 32 processors online two are newly purchased, there is no hardware faults in this server one server has 64gb memory and it shows 64 processors, and other having 32gb memory and it... (5 Replies)
Discussion started by: rajasekg
5 Replies

6. Solaris

Set up new T5220 remotely

I have a new T5220 racked at our remote office and the guys who racked it does not know how to configure the SP on the system .. is there any way i can do it remotely ? (1 Reply)
Discussion started by: fugitive
1 Replies

7. Solaris

Jumpstart Issues With T5220

I'm trying to use a V210 (sun4u) configured as a jumpstart server to flash install a number of T5220's (sunv4). The Facts: V210 - Solaris 10 (September 2010 Release / not patched) T5220's - Archive's created using prior (unknown) version of solaris 10. Flar -i <archive name> shows the... (6 Replies)
Discussion started by: Celtic_Monkey
6 Replies

8. Solaris

T5220

Hi Guys and Gals, Wondering if anyone could help me, tricky one this. Got a T5220 with a graphics card. When you power it on, it outputs its output to the screen and you can type commands with the keyboard etc. But no matter what version of Solaris 10 you use, when you do a boot cdrom... (1 Reply)
Discussion started by: callmebob
1 Replies

9. Solaris

T5220

Hello guys, can anyone give me some advice on how to perform clean installation of Solaris 10 on T5220 ? I am connected with serial management port on the machine, i have the ok promt and now i should eject the dvdrom to insert Solaris dvd into and start the clean install. So how to do that... (3 Replies)
Discussion started by: tonijel
3 Replies

10. Solaris

Storage issue on T5220

We have one box T5220 which was assigned 2x100G Luns but i have a strange problem if i use powermt display it only shows me one card but cfgadm, fcinfo commands shows 2 dual port cards as up. What could be the reason for it. I already did a re-configure reboot but that didn't help. Can anyone help... (1 Reply)
Discussion started by: fugitive
1 Replies
BIOCTL(8)						    BSD System Manager's Manual 						 BIOCTL(8)

NAME
bioctl -- RAID management interface SYNOPSIS
bioctl device command [arg [...]] DESCRIPTION
RAID device drivers which support management functionality can register their services with the bio(4) driver. bioctl then can be used to manage the RAID controller's properties. COMMANDS
The following commands are supported: show [disks | volumes] Without any argument by default bioctl will show information about all volumes and the logical disks used on them. If disks is specified, only information about physical disks will be shown. If volumes is specified, only information about the volumes will be shown. alarm [disable | enable | silence | test] Control the RAID card's alarm functionality, if supported. By default if no argument is specified, its current state will be shown. Optionally the disable, enable, silence, or test arguments may be specified to enable, disable, silence, or test the RAID card's alarm. blink start channel:target.lun | stop channel:target.lun Instruct the device at channel:target.lun to start or cease blinking, if there's ses(4) support in the enclosure. hotspare add channel:target.lun | remove channel:target.lun Create or remove a hot-spare drive at location channel:target.lun. passthru add DISKID channel:target.lun | remove channel:target.lun Create or remove a pass-through device. The DISKID argument specifies the disk that will be used for the new device, and it will be created at the location channel:target.lun. NOTE: Removing a pass-through device that has a mounted filesys- tem will lead to undefined behaviour. check start VOLID | stop VOLID Start or stop consistency volume check in the volume with index VOLID. NOTE: Not many RAID controllers support this fea- ture. create volume VOLID DISKIDs [SIZE] STRIPE RAID_LEVEL channel:target.lun Create a volume at index VOLID. The DISKIDs argument will specify the first and last disk, i.e.: 0-3 will use the disks 0, 1, 2, and 3. The SIZE argument is optional and may be specified if not all available disk space is wanted (also dependent of the RAID_LEVEL). The volume will have a stripe size defined in the STRIPE argument and it will be located at channel:target.lun. remove volume VOLID channel:target.lun Remove a volume at index VOLID and located at channel:target.lun. NOTE: Removing a RAID volume that has a mounted filesystem will lead to undefined behaviour. EXAMPLES
The following command, executed from the command line, shows the status of the volumes and its logical disks on the RAID controller: $ bioctl arcmsr0 show Volume Status Size Device/Label RAID Level Stripe ================================================================= 0 Building 468G sd0 ARC-1210-VOL#00 RAID 6 128KB 0% done 0:0 Online 234G 0:0.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:1 Online 234G 0:1.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:2 Online 234G 0:2.0 noencl <WDC WD2500YS-01SHB1 20.06C06> 0:3 Online 234G 0:3.0 noencl <WDC WD2500YS-01SHB1 20.06C06> To create a RAID 5 volume on the SCSI 0:15.0 location on the disks 0, 1, 2, and 3, with stripe size of 64Kb on the first volume ID, using all available free space on the disks: $ bioctl arcmsr0 create volume 0 0-3 64 5 0:15.0 To remove the volume 0 previously created at the SCSI 0:15.0 location: $ bioctl arcmsr0 remove volume 0 0:15.0 SEE ALSO
arcmsr(4), bio(4), cac(4), ciss(4), mfi(4) HISTORY
The bioctl command first appeared in OpenBSD 3.8, it was rewritten for NetBSD 5.0. AUTHORS
The bioctl interface was written by Marco Peereboom <marco@openbsd.org> and was rewritten with multiple features by Juan Romero Pardines <xtraeme@NetBSD.org>. BSD
March 16, 2008 BSD
All times are GMT -4. The time now is 09:40 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy