Sponsored Content
Full Discussion: Help with mirroring rpool
Operating Systems Solaris Help with mirroring rpool Post 302995712 by gull04 on Tuesday 11th of April 2017 08:17:01 AM
Old 04-11-2017
Hi Guys,

Am I missing something here, shouldn't this be running against slice two (s2)?

Regards

Gull04
This User Gave Thanks to gull04 For This Post:
 

10 More Discussions You Might Find Interesting

1. Solaris

rpool issue

My system is showing following .. how can i remove this rpool .. ? zpool status -v rpool pool: rpool state: UNAVAIL status: One or more devices could not be opened. There are insufficient replicas for the pool to continue functioning. action: Attach the missing device and... (0 Replies)
Discussion started by: fugitive
0 Replies

2. Solaris

How to change hostid saved in rpool?

Hi, We have a Sun M5000. I am now trying to boot the second system domain by using the boot disk (a mirrored boot disk actually) of the first domain (if succeed then no OS installation is needed for the second domain). I got the following errors: SPARC Enterprise M5000 Server, using Domain... (21 Replies)
Discussion started by: aixlover
21 Replies

3. Solaris

ZFS rpool physical disk move

I'd like to finish setting up this system and then move the secondary or primary disk to another system that is the exact same hardware. I've done things like this in the past with ufs and disk suite mirroring just fine. But I have yet to do it with a zfs root pool mirror. Are there any... (1 Reply)
Discussion started by: Metasin
1 Replies

4. Solaris

Create rpool manually

Hi All; My server's root partition was encapsulated with VxVM, I try to convert it to ZFS. I successfully de-encapsulated root. Now I try to mirror 2 root disks using ZFS. But I receive following error: # zpool create rpool mirror c0t0d0s0 c0t1d0s0 invalid vdev specification use '-f' to... (6 Replies)
Discussion started by: reseki
6 Replies

5. Solaris

How to access content of problematic rpool ?

Hi guys, I have a Solaris 10 x86 server with 2 physical disks. It had a ZFS rpool on the first disk c0t0d0s0. The server stopped booting with some error like "Cannot find bootfile". I booted with Solaris DVD and re-installed the operating system on the second disk with a different pool... (0 Replies)
Discussion started by: ARPcPro
0 Replies

6. Solaris

Resize rpool

Hi, I have rpool about 500G. So i want to use 210G from rpool and assign mount point as /database. I seek in google and couldn't found it. Does anyone know how to achieve it? Thanks. (2 Replies)
Discussion started by: mzainal
2 Replies

7. Solaris

Zfs rpool size

Hi everyone, I am doing housekeeping of my Solaris 11 for zfs snapshot to reduce the snapshot size. I have already cleared the / file system, however the rpool size still not reduced. Filesystem Size Used Available Capacity Mounted on rpool/ROOT/solaris-2 98G 6.9G ... (2 Replies)
Discussion started by: freshmeat
2 Replies

8. Solaris

Need help to understand zfs rpool space allocation

Hi, I am unable to understand that, in one of my servers while df -kh Filesystem Size Used Available Capacity Mounted on rpool/ROOT/solaris-2 98G 11G 29G 29% / Even the Root FS filled on 40gb and system becomes unstable. it is showing... (4 Replies)
Discussion started by: anuragr
4 Replies

9. Solaris

Trying to create ZFS slice on rpool

I have a 240GB disk as rpool. I have installed Solaris 11.3 to a partition which is 110GB. Now I have another 130GB which is unallocated. I want to use that additional space as a temporary folder to be shared between Solaris and Linux. The additional space had no /dev/dsk/c2t4... entry so I used... (8 Replies)
Discussion started by: kebabbert
8 Replies

10. Solaris

How to increase the size of rpool in Solaris 11?

Hi Please let me know how to increase the size of rpool in solaris 11 in ldom. I know how to map the new LUN to LDOM after that please let me know the procedure to increase the rpool and how to identify new disk in OBP level of ldom as I need to set the new/lun to be my new boot device. (5 Replies)
Discussion started by: hb00
5 Replies
SYSTEMD.SLICE(5)						   systemd.slice						  SYSTEMD.SLICE(5)

NAME
systemd.slice - Slice unit configuration SYNOPSIS
slice.slice DESCRIPTION
A unit configuration file whose name ends in ".slice" encodes information about a slice unit. A slice unit is a concept for hierarchically managing resources of a group of processes. This management is performed by creating a node in the Linux Control Group (cgroup) tree. Units that manage processes (primarily scope and service units) may be assigned to a specific slice. For each slice, certain resource limits may be set that apply to all processes of all units contained in that slice. Slices are organized hierarchically in a tree. The name of the slice encodes the location in the tree. The name consists of a dash-separated series of names, which describes the path to the slice from the root slice. The root slice is named -.slice. Example: foo-bar.slice is a slice that is located within foo.slice, which in turn is located in the root slice -.slice. Note that slice units cannot be templated, nor is possible to add multiple names to a slice unit by creating additional symlinks to its unit file. By default, service and scope units are placed in system.slice, virtual machines and containers registered with systemd-machined(1) are found in machine.slice, and user sessions handled by systemd-logind(1) in user.slice. See systemd.special(5) for more information. See systemd.unit(5) for the common options of all unit configuration files. The common configuration items are configured in the generic [Unit] and [Install] sections. The slice specific configuration options are configured in the [Slice] section. Currently, only generic resource control settings as described in systemd.resource-control(5) are allowed. See the New Control Group Interfaces[1] for an introduction on how to make use of slice units from programs. IMPLICIT DEPENDENCIES
The following dependencies are implicitly added: o Slice units automatically gain dependencies of type After= and Requires= on their immediate parent slice unit. DEFAULT DEPENDENCIES
The following dependencies are added unless DefaultDependencies=no is set: o Slice units will automatically have dependencies of type Conflicts= and Before= on shutdown.target. These ensure that slice units are removed prior to system shutdown. Only slice units involved with late system shutdown should disable DefaultDependencies= option. SEE ALSO
systemd(1), systemd.unit(5), systemd.resource-control(5), systemd.service(5), systemd.scope(5), systemd.special(7), systemd.directives(7) NOTES
1. New Control Group Interfaces https://www.freedesktop.org/wiki/Software/systemd/ControlGroupInterface/ systemd 237 SYSTEMD.SLICE(5)
All times are GMT -4. The time now is 04:01 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy