10-31-2013
No, but you could put them in a zip -9 and mount that. Disk is cheap, and compression is pretty nice in a CPU much faster than disk (never mind fiber and wire net remote disk). There are many file systems with compression. The fluffy stuff in -g unstripped unoptimized files compresses great. (Sometimes optimizations grow the object, like loop unrolling and subroutine inlining.)
Any stable libraries can be created without debug information.
You can trace a lot of calls using truss without -g or such. The load of being traced can be applied after the app is running. You can truss after an app is started, without source code, recompile or restart.
10 More Discussions You Might Find Interesting
1. HP-UX
When i execute bdf, /home direcory seems 100% full. But when i check /home with 'du', total used memory is 30 MB in 1,4 Gb. how can I reduce this 100% to its real state? (11 Replies)
Discussion started by: akyuceisik
11 Replies
2. Shell Programming and Scripting
HPUX 11iv2 #!/bin/sh
Hi all. I have a script that results in the creation of an ascii file which is ultimately emailed out to several people. The email wraps each line so I would like to reduce the font size of the ascii file. I looked at nroff and also tr but it wasn't clear to me how to do... (2 Replies)
Discussion started by: lyoncc
2 Replies
3. Shell Programming and Scripting
Hello,
I want to compress any given file or directory. I used
1)gzip
2)zip
But when I do "ls -l". I found that the zipped file is in fact greater in size than the original file.
Can you please tell me the commands which will show me the difference in its size. (2 Replies)
Discussion started by: nsharath
2 Replies
4. Shell Programming and Scripting
Hello everyone,
I would like to filter and search for files in my curr dir where the blocks used by those files are over a certain number (i.e. 30), when I try this command
find . -name "c*" -size +30 -exec ls -ls {} \;
I get a list of files, the first column is the block size, right?
... (1 Reply)
Discussion started by: gio001
1 Replies
5. SuSE
Hello,
I do not know Linux. It is a black box.
We have 2 virtuals servers (SAPVM01 and SAPVM06) in one physical server.
The first virtual system (SAPVM01) has a total size of 420 Gb and a free space of 170 GB. A SAP system is running.
The second virtual system (SAPVM06) has a total... (3 Replies)
Discussion started by: daniel04
3 Replies
6. Solaris
I'm trying to reduce hard drive size (number of cylinders) in SPARC Solaris. Its easy to change last cylinder of last slice, but that cannot be done for slice2/backupslice because it insists on whole disk. If I try to change disk type/geometry, all slices get replaced with some 'default'... (1 Reply)
Discussion started by: orange47
1 Replies
7. Shell Programming and Scripting
Dear all,
i have a lot of .pdf files that i need to reduce size with pdf2ps and ps2pdf app. I need a script which i can reduce file size of all .pdf files in every subfolder of WORKDIR folder.
folder tree like:
WORKDIR
SUBBWORK DIR1
SUB_SUB_WORKDIR1
... (1 Reply)
Discussion started by: migor78
1 Replies
8. UNIX for Dummies Questions & Answers
My volume group of size 200 gb.
out of which only 100 gb is used by 2 logical volumes /dev/vg00/lvol0 and /dev/vg00/lvol0 respectively (both are 50 gb each).
Whenever i use vgreduce command to reduce the size of volume group i get below error.
# vgreduce vg00 -a
Physical volume... (16 Replies)
Discussion started by: pinga123
16 Replies
9. Red Hat
Hi,
inode size reached its 100% in /var
Due to this i'am getting the error No space left on device
my crond process is stopped and when i want to restart it
it is showing the below error
Starting crond: crond: can't open or create /var/run/crond.pid: No space left on device
df -i o/p
... (3 Replies)
Discussion started by: Mohamed Thamim
3 Replies
10. Red Hat
We have several dozen Redhat 5, 6 and 7 servers that are running Oracle databases. On some databases we are using automatic memory management, which uses shared memory. On other databases we are use manual memory management, which does not use shared memory.
When I see that a server is swapping... (2 Replies)
Discussion started by: gandolf989
2 Replies
hdik(8) BSD System Manager's Manual hdik(8)
NAME
hdik -- lightweight tool to attach and mount disk images in-kernel
SYNOPSIS
hdik imagefile [options]
DESCRIPTION
hdik is a lightweight tool that can be used to attach disk images in-kernel (i.e. without a user-land process to provide the backing store).
Only a subset of disk images can be mounted in this manner including read/write disk images, UDIF disk images that use zlib compression,
shadowed disk images, and sparse disk images.
hdik is intended for use in situations where linking against the DiskImages framework is problematic or an extremely lightweight mechanism
for attaching a disk image is needed.
You can specify that the image should not be processed by Disk Arbitration by specifying the -nomount option. You can also specify that the
image be mounted with a shadow file by using the -shadow option.
The following argument must be specified:
imagefile the disk image to be mounted.
OPTIONS
-shadow [shadowfile]
Use a shadow file in conjunction with the data in the image. This option prevents modification of the original image and allows
read-only images to be used as read/write images. When blocks are being read from the image, blocks present in the shadow file
override blocks in the base image. When blocks are being written, the writes will be redirected to the shadow file. If not
specified, -shadow defaults to <imagename>.shadow. If the shadow file does not exist, it is created.
-nomount Suppress automatic mounting of the image or partitions on it. This will result in /dev entries being created, but will not mount
any volumes.
-drivekey keyname=value
Specify a key/value pair for the IOHDIXHDDrive object created (shows up in the IOKit registry of devices which is viewable with
ioreg(8)).
SEE ALSO
hdiutil(1), diskarbitrationd(8), diskutil(8), ioreg(8)
Mac OS X 29 Apr 2003 Mac OS X