Sponsored Content
Operating Systems Solaris 6120 Array. Additional physical Disks and ZFS Post 302281184 by DukeNuke2 on Wednesday 28th of January 2009 10:22:03 AM
Old 01-28-2009
some arrays with other raid level can be edited (raid1) but not the "stripped" raid levels (raid0 and 5).
 

10 More Discussions You Might Find Interesting

1. Filesystems, Disks and Memory

how to assign same mount point for file systems mounted on physical disks

We have 6 hard disks attached to the hardware. Of this 2 hard disks are of 9 GB each. Now I want combine both the same in such a way that i see a combined entry in the output of df -k . The steps I follow are 1. Create partition on hard disks (Using format partition) 2. Run newfs -v for... (6 Replies)
Discussion started by: Hitesh Shah
6 Replies

2. UNIX for Dummies Questions & Answers

monitoring the state of physical disks

Hello, I would like to know if there are commands that can be used to monitor the state of physical disks (including RAID) under AIX and SUN unix platforms? Thank you in advance. (4 Replies)
Discussion started by: VeroL
4 Replies

3. UNIX for Dummies Questions & Answers

Vertias (physical disks monitoring)

Can anyone tell me what Vertias is? Is it free? What is it used for exactly? Thank you in advance. (1 Reply)
Discussion started by: VeroL
1 Replies

4. Filesystems, Disks and Memory

How do I check 4 physical damaged on Linux hard disks?

How do I check for physical damage on red hat linux hard disks? I tried smartctl /dev/sdb but it came back so fast saying it was ok. Is there a better linux command to check for bad sectors or physical disks in linux? Is there a good way such as with parted or something else? I normally in HP... (4 Replies)
Discussion started by: taekwondo
4 Replies

5. AIX

How to determine the physical volume fo the disks

This is the report I got running the comand rptconf, but I would like to know what is the capacity of the disks installed into our server power 6 with AIX System Model: IBM,7778-23X Machine Serial Number: 1066D5A Processor Type: PowerPC_POWER6 Processor Implementation Mode: POWER 6... (6 Replies)
Discussion started by: cucosss
6 Replies

6. Solaris

ZFS : Can arc size value exceed Physical RAM ?

Hi, kstat -p -m zfs -n arcstats -s size returns zfs:0:arcstats:size 8177310584 this values is approx (7.61 GB) but my Physical Memory size is only 6144 Megabytes. Can this happen ? if yes, then how can I find free memory on the system. BTW, I ran the kstat commands from a Non... (2 Replies)
Discussion started by: sapre_amit
2 Replies

7. Red Hat

Create volume using LVM over 2 physical disks

I wanted to know how we can combine volumes over 2 physical drives. # fdisk -l Disk /dev/sda: 42.9 GB, 42949672960 bytes 255 heads, 63 sectors/track, 5221 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 ... (16 Replies)
Discussion started by: ikn3
16 Replies

8. Solaris

ZFS rpool physical disk move

I'd like to finish setting up this system and then move the secondary or primary disk to another system that is the exact same hardware. I've done things like this in the past with ufs and disk suite mirroring just fine. But I have yet to do it with a zfs root pool mirror. Are there any... (1 Reply)
Discussion started by: Metasin
1 Replies

9. Solaris

Pls. help with Sun Array 6120

Hi, I am on Sun Edge 6120 Disk array. When I do port listmap. The failover status below means I have to take action. I never used Sun Array before. Please advice. port targetid addr_type lun volume owner access u1p1 1 hard 0 v0 ... (0 Replies)
Discussion started by: samnyc
0 Replies

10. Solaris

Exporting physical disk to ldom or ZFS volume

Generally, this is what we do:- On primary, export 2 LUNs (add-vdsdev). On primary, assign these disks to the ldom in question (add-vdisk). On ldom, created mirrored zpool from these two disks. On one server (which is older) we have:- On primary, create mirrored zpool from the two LUNs.... (4 Replies)
Discussion started by: psychocandy
4 Replies
VGREDUCE(8)                                                   System Manager's Manual                                                  VGREDUCE(8)

NAME
vgreduce - reduce a volume group SYNOPSIS
vgreduce [-a|--all] [-A|--autobackup y|n] [-d|--debug] [-h|-?|--help] [--removemissing] [-t|--test] [-v|--verbose] VolumeGroupName [Physi- calVolumePath...] DESCRIPTION
vgreduce allows you to remove one or more unused physical volumes from a volume group. OPTIONS
See lvm for common options. -a, --all Removes all empty physical volumes if none are given on command line. --removemissing Removes all missing physical volumes from the volume group, if there are no logical volumes allocated on those. This resumes normal operation of the volume group (new logical volumes may again be created, changed and so on). If this is not possible (there are logical volumes referencing the missing physical volumes) and you cannot or do not want to remove them manually, you can run this option with --force to have vgreduce remove any partial LVs. Any logical volumes and dependent snapshots that were partly on the missing disks get removed completely. This includes those parts that lie on disks that are still present. If your logical volumes spanned several disks including the ones that are lost, you might want to try to salvage data first by acti- vating your logical volumes with --partial as described in lvm (8). SEE ALSO
lvm(8), vgextend(8) Sistina Software UK LVM TOOLS 2.02.95(2) (2012-03-06) VGREDUCE(8)
All times are GMT -4. The time now is 09:23 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy