Sponsored Content
Operating Systems AIX AIX flag to reduce size of shared file Post 302866047 by Abhi04 on Monday 21st of October 2013 06:47:10 AM
Old 10-21-2013
AIX flag to reduce size of shared file

I am using xlC (Version: 11.01.0000.0011).

While build i am using "-g" to have debug information in build.
there are many object files (>500) due to which resultant shared file (.so) will have huge size.

I can't reduce optimization level.
Is there any way or flag is present by using which i can reduce size of shared file (.so)?
 

10 More Discussions You Might Find Interesting

1. HP-UX

How to reduce fil system size seen in bdf!

When i execute bdf, /home direcory seems 100% full. But when i check /home with 'du', total used memory is 30 MB in 1,4 Gb. how can I reduce this 100% to its real state? (11 Replies)
Discussion started by: akyuceisik
11 Replies

2. Shell Programming and Scripting

How to reduce font size in a file

HPUX 11iv2 #!/bin/sh Hi all. I have a script that results in the creation of an ascii file which is ultimately emailed out to several people. The email wraps each line so I would like to reduce the font size of the ascii file. I looked at nroff and also tr but it wasn't clear to me how to do... (2 Replies)
Discussion started by: lyoncc
2 Replies

3. Shell Programming and Scripting

command to reduce size of file/directory???

Hello, I want to compress any given file or directory. I used 1)gzip 2)zip But when I do "ls -l". I found that the zipped file is in fact greater in size than the original file. Can you please tell me the commands which will show me the difference in its size. (2 Replies)
Discussion started by: nsharath
2 Replies

4. Shell Programming and Scripting

find and size flag

Hello everyone, I would like to filter and search for files in my curr dir where the blocks used by those files are over a certain number (i.e. 30), when I try this command find . -name "c*" -size +30 -exec ls -ls {} \; I get a list of files, the first column is the block size, right? ... (1 Reply)
Discussion started by: gio001
1 Replies

5. SuSE

Reduce Size of serveur in LINUX-Suse

Hello, I do not know Linux. It is a black box. We have 2 virtuals servers (SAPVM01 and SAPVM06) in one physical server. The first virtual system (SAPVM01) has a total size of 420 Gb and a free space of 170 GB. A SAP system is running. The second virtual system (SAPVM06) has a total... (3 Replies)
Discussion started by: daniel04
3 Replies

6. Solaris

reduce hard drive size

I'm trying to reduce hard drive size (number of cylinders) in SPARC Solaris. Its easy to change last cylinder of last slice, but that cannot be done for slice2/backupslice because it insists on whole disk. If I try to change disk type/geometry, all slices get replaced with some 'default'... (1 Reply)
Discussion started by: orange47
1 Replies

7. Shell Programming and Scripting

reduce pdf file size through multiple folders

Dear all, i have a lot of .pdf files that i need to reduce size with pdf2ps and ps2pdf app. I need a script which i can reduce file size of all .pdf files in every subfolder of WORKDIR folder. folder tree like: WORKDIR SUBBWORK DIR1 SUB_SUB_WORKDIR1 ... (1 Reply)
Discussion started by: migor78
1 Replies

8. UNIX for Dummies Questions & Answers

Unable to reduce the size of volume group?

My volume group of size 200 gb. out of which only 100 gb is used by 2 logical volumes /dev/vg00/lvol0 and /dev/vg00/lvol0 respectively (both are 50 gb each). Whenever i use vgreduce command to reduce the size of volume group i get below error. # vgreduce vg00 -a Physical volume... (16 Replies)
Discussion started by: pinga123
16 Replies

9. Red Hat

How to reduce inode size of /var?

Hi, inode size reached its 100% in /var Due to this i'am getting the error No space left on device my crond process is stopped and when i want to restart it it is showing the below error Starting crond: crond: can't open or create /var/run/crond.pid: No space left on device df -i o/p ... (3 Replies)
Discussion started by: Mohamed Thamim
3 Replies

10. Red Hat

Does it make sense to reduce the total shared memory

We have several dozen Redhat 5, 6 and 7 servers that are running Oracle databases. On some databases we are using automatic memory management, which uses shared memory. On other databases we are use manual memory management, which does not use shared memory. When I see that a server is swapping... (2 Replies)
Discussion started by: gandolf989
2 Replies
dwz(1)							      General Commands Manual							    dwz(1)

NAME
dwz - DWARF optimization and duplicate removal tool SYNOPSIS
dwz [OPTION...] [FILES] DESCRIPTION
dwz is a program that attempts to optimize DWARF debugging information contained in ELF shared libraries and ELF executables for size, by replacing DWARF information representation with equivalent smaller representation where possible and by reducing the amount of duplication using techniques from DWARF standard appendix E - creating DW_TAG_partial_unit compilation units (CUs) for duplicated information and using DW_TAG_imported_unit to import it into each CU that needs it. The tool handles DWARF 32-bit format debugging sections of versions 2, 3 and 4 and GNU extensions on top of those, though using DWARF 4 or worst case DWARF 3 is strongly recommended. The tool has two main modes of operation, without the -m option it attempts to optimize DWARF debugging information in each given object (executable or shared library) individually, with the -m option it afterwards attempts to optimize even more by moving DWARF debugging information entries (DIEs), strings and macro descriptions duplicated in more than one object into a newly created ELF ET_REL object whose filename is given as -m option argument. The debug sections in the executables and shared libraries specified on the command line are then modified again, referring to the entities in the newly created object. OPTIONS
-m FILE --multifile FILE Multifile mode. After processing all named executables and shared libraries, attempt to create ELF object FILE and put debugging information duplicated in more than one object there, afterwards optimize each named executable or shared library even further if possible. -h --hardlink Look for executables or shared libraries hardlinked together, instead of rewriting them individually rewrite just one of them and hardlink the rest to the first one again. -M NAME --multifile-name NAME Specify the name of the common file that should be put into the .gnu_debugaltlink section alongside with its build ID. By default dwz puts there the argument of the -m option. -r --relative Specify that the name of the common file to be put into the .gnu_debugaltlink section is supposed to be relative path from the directory containing the executable or shared library to the file named in the argument of the -m option. Either -M or -r option can be specified, but not both. -q --quiet Silence up some of the most common messages. -o FILE --output FILE This option instructs dwz not to overwrite the specified file, but instead store the new content into FILE. Nothing is written if dwz exits with non-zero exit code. Can be used only with a single executable or shared library (if there are no arguments at all, a.out is assumed). -l COUNT --low-mem-die-limit COUNT Handle executables or shared libraries containing more than COUNT debugging information entries in their .debug_info section using a slower and more memory usage friendly mode and don't attempt to optimize that object in multifile mode. The default is 10 million DIEs. There is a risk that for very large amounts of debugging information in a single shared library or executable there might not be enough memory (especially when dwz tool is 32-bit binary, it might run out of available virtual address space even sooner). -L COUNT --max-die-limit COUNT Don't attempt to optimize executables or shared libraries containing more than COUNT DIEs at all. The default is 50 million DIEs. -? --help Print short help and exit. ARGUMENTS
Command-line arguments should be the executables, shared libraries or their stripped to file separate debug information objects. EXAMPLES
$ dwz -m .dwz/foobar-1.2.debug -rh bin/foo.debug bin/foo2.debug foo/lib/libbar.so.debug will attempt to optimize debugging information in bin/foo.debug, bin/foo2.debug and lib/libbar.so.debug (by modifying the files in place) and when beneficial also will create .dwz/foobar-1.2.debug file. .gnu_debugaltlink section in the first two files will refer to ../.dwz/foobar-1.2.debug and in the last file to ../../.dwz/foobar-1.2.debug. If e.g. bin/foo.debug and bin/foo2.debug were hardlinked together initially, they will be hardlinked again and for multifile optimizations considered just as a single file rather than two. $ dwz -o foo.dwz foo will not modify foo but instead store the ELF object with optimized debugging information if successful into foo.dwz file it creates. $ dwz *.debug foo/*.debug will attempt to optimize debugging information in *.debug and foo/*.debug files, optimizing each file individually in place. $ dwz is equivalent to dwz a.out command. SEE ALSO
http://dwarfstd.org/doc/DWARF4.pdf , gdb(1). AUTHORS
Jakub Jelinek <jakub@redhat.com>. 15 June 2012 dwz(1)
All times are GMT -4. The time now is 07:54 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy