07-18-2007
just run the "format" commando....
9 More Discussions You Might Find Interesting
1. UNIX for Dummies Questions & Answers
Hello !
Does anyone knows which de limit of files in the ufs filesystem in the Solaris 2.6 ??
Danke,
Witt (2 Replies)
Discussion started by: witt
2 Replies
2. Solaris
Hi there,
I am trying to mount a SAN volume (which is mapped to solaris sparc) partitioned with ufs filesystem onto a linux (intel processor 64bit) server.
*I have re-compiled the linux kernel t support ufs fstype with ro mount support.
filesystem on solaris:... (3 Replies)
Discussion started by: ilan
3 Replies
3. Solaris
I have two Unix (Solaris) hosts that are both attached to our SAN. They are both presented with the same luns.
What I want to do is have the same UFS filesystem mounted on both hosts at the same time. What I am trying to accomplish is creating a zone that will reside on both hosts but only... (2 Replies)
Discussion started by: beaker457
2 Replies
4. BSD
Hi, I'm new to BSD and would like to create a dual-boot between Solaris Express Community Edition and FreeBSD.
I would just like to know if the Solaris UFS file system can be written to by BSD?
I know that BSD uses UFS2, but I'm hoping that it is backwards compatible with UFS1 provided that... (0 Replies)
Discussion started by: Johnny SSH
0 Replies
5. Solaris
Hi Gurus
I want to know the command & tips regarding, how to increase or decrease inode number of the particular ufs filesystem. Is it possible to do it in a live/production environment.
Regards (3 Replies)
Discussion started by: girish.batra
3 Replies
6. Solaris
What can I fix this issue? I have ran below commands but everything is same.:confused:
WARNING: Last shutdown is later than time on time-of-day chip: check date.
The / file system (/dev/rdsk/c1t0d0s0) is being checked
WARNING - unable to repair the / filesystem. Run fsck manually (fsck -F... (4 Replies)
Discussion started by: getrue
4 Replies
7. UNIX for Dummies Questions & Answers
Hello, I've got an Ultrix 4.2 ISO file (gxemul doesn't recognize it when I type gxemul -X -e 3max -d rootdisk.img -d Ultrix4_2.iso)
So I read some other posts which show that this can be done via a UFS CDROM image. But I don't know how to create one. Anyway, I think the ISO file could be in... (0 Replies)
Discussion started by: lucky7456969
0 Replies
8. Solaris
I'm prompted to start this thread following my attempt to help on this thread here (see my posts).
I was proposing the OP deep checked a Solaris ufs filesystem using:
# fsck -o full <filesystem node>
however this option does not appear to be valid on Solaris 10.
I've used... (8 Replies)
Discussion started by: hicksd8
8 Replies
9. Solaris
I need to increase the /var (UFS) filesystem and root disk under veritas control or root disk is encapsulated
# df -k /var
Filesystem kbytes used avail capacity Mounted on
/dev/vx/dsk/var 13241195 12475897 674524 96% /var
# fstyp /dev/vx/dsk/var
ufs
# pkginfo... (1 Reply)
Discussion started by: amity
1 Replies
LEARN ABOUT X11R4
queuedefs
queuedefs(4) File Formats queuedefs(4)
NAME
queuedefs - queue description file for at, batch, and cron
SYNOPSIS
/etc/cron.d/queuedefs
DESCRIPTION
The queuedefs file describes the characteristics of the queues managed by cron(1M). Each non-comment line in this file describes one queue.
The format of the lines are as follows:
q.[njobj][nicen][nwaitw]
The fields in this line are:
q The name of the queue. a is the default queue for jobs started by at(1); b is the default queue for jobs started by batch (see
at(1)); c is the default queue for jobs run from a crontab(1) file.
njob The maximum number of jobs that can be run simultaneously in that queue; if more than njob jobs are ready to run, only the first
njob jobs will be run, and the others will be run as jobs that are currently running terminate. The default value is 100.
nice The nice(1) value to give to all jobs in that queue that are not run with a user ID of super-user. The default value is 2.
nwait The number of seconds to wait before rescheduling a job that was deferred because more than njob jobs were running in that job's
queue, or because the system-wide limit of jobs executing has been reached. The default value is 60.
Lines beginning with # are comments, and are ignored.
EXAMPLES
Example 1: A sample file.
#
#
a.4j1n
b.2j2n90w
This file specifies that the a queue, for at jobs, can have up to 4 jobs running simultaneously; those jobs will be run with a nice value
of 1. As no nwait value was given, if a job cannot be run because too many other jobs are running cron will wait 60 seconds before trying
again to run it.
The b queue, for batch(1) jobs, can have up to 2 jobs running simultaneously; those jobs will be run with a nice(1) value of 2. If a job
cannot be run because too many other jobs are running, cron(1M) will wait 90 seconds before trying again to run it. All other queues can
have up to 100 jobs running simultaneously; they will be run with a nice value of 2, and if a job cannot be run because too many other jobs
are running cron will wait 60 seconds before trying again to run it.
FILES
/etc/cron.d/queuedefs queue description file for at, batch, and cron.
SEE ALSO
at(1), crontab(1), nice(1), cron(1M)
SunOS 5.10 1 Mar 1994 queuedefs(4)