Thanks for reaching out.
I'm primarily a Linux administrator and don't dabble much with HP-UX so if you need additional info, please let me know.
The HP-UX server is attached to EMC storage. Our Linux servers were previously backing up to a legacy Sun Solaris server but we've run out of space there so I'm trying to shift the scripts to now backup to the HP server. I've created the logical volume and filesystem from scratch. As mentioned, everything seems to be working as expected with the exception of using dump from Linux to this filesystem. The Linux servers are using the dump options "0uf". I've tried 0auf to no avail. Thanks again for reaching out.
- Bill
---------- Post updated at 12:11 PM ---------- Previous update was at 12:08 PM ----------
Keep in mind that the exact same script works flawlessly to both a Solaris server and another Linux server. As soon as I change one of the variables to point to the HP-UX server, it craps out after 2GB every time. The dump is over SSH. I've also tried RSH but got the same results. Thanks.
---------- Post updated at 12:20 PM ---------- Previous update was at 12:11 PM ----------
Proof that the filesystem in question does in fact support large files:
(I've also scp'd an 8GB file from the same Linux server to the filesystem)
Last edited by Scott; 11-09-2011 at 02:24 PM..
Reason: Code tags
I am working on HP-Unix.
I have a 600 MB file in compressed form.
During decompression, when file size reaches
2GB, decompression aborts.
What should be done? (3 Replies)
I want to have a permanent file created - and limit the size that this file can grow.. I want a circular file..
ie max size of file is 10 mb.. and if any new data written to file the oldest data removed..
How can I do this?
I am on solaris 9 x86 (3 Replies)
Can anybody help me?
How to increase file size limit in aix 5.2? I have already specified in /etc/security/limits file :
default:
fsize = -1
core = 2097151
cpu = -1
data = -1
rss = -1
stack = -1
nofiles = 2000 (2 Replies)
Hi,
I have a problem writing or copying a file 2GB or larger to either the second or third disk on my C8000. I've searched this forum and found some good information on this but still nothing to solve the problem.
I'm running hpux 11i, JFS3.3 and disk version 4 (from fstyp) on all 3 disks.
... (2 Replies)
Any idea how to get around this limit? I have a 42GB database backup file (.dmp) taking up disk space because neither tar nor cpio are able to put it onto a tape. I am on a SUN Solaris using SunOS 5.8. I would appreciate whatever help can be provided. Thanks! (9 Replies)
Hi All,
We are running HP rp7400 box with hpux 11iv1.
Recently, we changed 3 kernel parameters
a) msgseg from 32560 to 32767
b) msgmnb from 65536 to 65535
c) msgssz from 128 to 256
Then we noticed that all application debug file size increase upto 2GB then it stops. So far we did not... (1 Reply)
Hi All,
I want to store 32KB of file in Oracle DB into CLOB field. I am not able to insert more than 32KB of file into CLOB. So i want to put a limit on the file size. I am using k shell.
My file size will dynamically increase its size, i want to check the file size if it is more than 32KB... (1 Reply)
Hi,
Am trying to run zip -r on a 2.4G directory and it is failing with the error below. I believe this is because of the 2G limit of the zip program.
server101(oper01)/u01/temp$: date
Thu Mar 15 12:53:44 NZDT 2012
server101(oper01)/u01/temp$: ls -l
total 8
drwxr-x--x 4 oracle dba ... (1 Reply)
Discussion started by: newbie_01
1 Replies
LEARN ABOUT CENTOS
unsquashfs
UNSQUASHFS(1) uncompress squashfs filesystems UNSQUASHFS(1)NAME
mksquashfs - tool to uncompress squashfs filesystems
SYNOPSIS
unsquashfs [OPTIONS] FILESYSTEM [directories or files to extract]
DESCRIPTION
Squashfs is a highly compressed read-only filesystem for Linux. It uses zlib compression to compress both files, inodes and directories.
Inodes in the system are very small and all blocks are packed to minimize data overhead. Block sizes greater than 4K are supported up to a
maximum of 64K.
Squashfs is intended for general read-only filesystem use, for archival use (i.e. in cases where a .tar.gz file may be used), and in con-
strained block device/memory systems (e.g. embedded systems) where low overhead is needed.
OPTIONS -v, -version
print version, licence and copyright information.
-d PATHNAME, -dest PATHNAME
unsquash to PATHNAME, default "squashfs-root".
-n, -no-progress
don't display the progress bar.
-no, -no-xattrs
don't extract xattrs in file system.
-x, -xattrs
extract xattrs in file system (default).
-p NUMBER, -processors NUMBER
use NUMBER processors. By default will use number of processors available.
-i, -info
print files as they are unsquashed.
-li, -linfo
print files as they are unsquashed with file attributes (like ls -l output).
-l, -ls
list filesystem, but don't unsquash.
-ll, -lls
list filesystem with file attributes (like ls -l output), but don't unsquash.
-f, -force
if file already exists then overwrite.
-s, -stat
display filesystem superblock information.
-e EXTRACT_FILE, -ef EXTRACT_FILE
list of directories or files to extract. One per line.
-da SIZE, -data-queue SIZE
Set data queue to SIZE Mbytes. Default 256 Mbytes.
-fr SIZE, -frag-queue SIZE
Set fragment queue to SIZE Mbytes. Default 256 Mbytes.
-r, -regex
treat extract names as POSIX regular expressions rather than use the default shell wildcard expansion (globbing).
Decompressors available
gzip
lzo
xz
SEE ALSO mksquashfs(1)HOMEPAGE
More information about unsquashfs and the squashfs filesystem can be found at <http://squashfs.sourceforge.net/>.
AUTHOR
squashfs was written by Phillip Lougher <plougher@users.sourceforge.net>.
This manual page was written by Daniel Baumann <daniel.baumann@progress-technologies.net>.
4.2 2012-06-30 UNSQUASHFS(1)