Sponsored Content
Operating Systems Solaris Oracle memory usage on Solaris box Post 302458239 by prabumohan on Thursday 30th of September 2010 07:02:17 AM
Old 09-30-2010
Code:
root@sunfire130 # echo "::memstat" | mdb -k
Page Summary                Pages                MB  %Tot
------------     ----------------  ----------------  ----
Kernel                     572519              4472   14%
Anon                      3087015             24117   75%
Exec and libs               64314               502    2%
Page cache                   9661                75    0%
Free (cachelist)           340233              2658    8%
Free (freelist)             28226               220    1%

Total                     4101968             32046
Physical                  4096728             32005

Code:
sh memory.sh
process: -csh
                shared: 0 KB
                tot private: 632 KB
process: asm_lms0_+ASM2
                shared: 126984 KB
                tot private: 12448 KB
process: asm_dbw0_+ASM2
                shared: 126984 KB
                tot private: 2488 KB
process: asm_lmd0_+ASM2
                shared: 126984 KB
                tot private: 14040 KB
process: xargs
                shared: 0 KB
                tot private: 120 KB
process: asm_lck0_+ASM2
                shared: 126984 KB
                tot private: 3136 KB
process: asm_gmon_+ASM2
                shared: 126984 KB
                tot private: 3648 KB
process: asm_pmon_+ASM2
                shared: 126984 KB
                tot private: 2328 KB
process: sqlplus
                shared: 0 KB
                tot private: 3320 KB
process: asm_mman_+ASM2
                shared: 126984 KB
                tot private: 2024 KB
process: /usr/lib/ssh/sshd
                shared: 0 KB
                tot private: 552 KB
process: sh
                shared: 0 KB
                tot private: 152 KB
process: oracle
                shared: 126984 KB
                tot private: 10551136 KB
process: nawk
                shared: 0 KB
                tot private: 392 KB
process: asm_lmon_+ASM2
                shared: 126984 KB
                tot private: 4320 KB
process: asm_o000_+ASM2
                shared: 126984 KB
                tot private: 832 KB
process: asm_ckpt_+ASM2
                shared: 126984 KB
                tot private: 1728 KB
process: asm_psp0_+ASM2
                shared: 126984 KB
                tot private: 1192 KB
process: asm_rbal_+ASM2
                shared: 126984 KB
                tot private: 2880 KB
process: asm_lgwr_+ASM2
                shared: 126984 KB
                tot private: 2280 KB
process: asm_smon_+ASM2
                shared: 126984 KB
                tot private: 1696 KB
process: asm_diag_+ASM2
                shared: 126984 KB
                tot private: 2896 KB
process: /usr/bin/ssh
                shared: 0 KB
                tot private: 82136 KB

 

10 More Discussions You Might Find Interesting

1. Filesystems, Disks and Memory

Memory usage in the box

Hello: Environment is: Oracle 817 on IBM RS/6000 AIX 433 I have 4GB RAM on the box and Page/Swap is about the same. Presently I am using close to 1GB of RAM towards 5 instances of ORACLE production environments. How can I know, how much of memory/RAM is used for : Oracle Processes , I... (2 Replies)
Discussion started by: ST2000
2 Replies

2. Solaris

oracle process and LWP on solaris 8 box

Hi all, Without oracle tunning I understand that oracle processes always run on a single CPU at a time (the prstat output always show 1 CPU number) even on multi-processor hardware. Is it correct? Regards, (1 Reply)
Discussion started by: solea
1 Replies

3. UNIX for Dummies Questions & Answers

Comman/Tool to check memory leak in solaris box

Hello Guys, Please let me know about commands or tools that are available to check memory leak in solaris box. Thanks in Advance (1 Reply)
Discussion started by: rajeshlbryan
1 Replies

4. UNIX for Dummies Questions & Answers

memory usage : kernel, oracle, other users

Hello, I'm looking for a command or script that could permit me to know what is memory usage group by HP-UX kernel, oracle processes and other users (specific software). A command like 'top' but for all processes and not only for 12 max could be very useful Thanks in advance (4 Replies)
Discussion started by: thierryUX
4 Replies

5. HP-UX

how can I find cpu usage memory usage swap usage and logical volume usage

how can I find cpu usage memory usage swap usage and I want to know CPU usage above X% and contiue Y times and memory usage above X % and contiue Y times my final destination is monitor process logical volume usage above X % and number of Logical voluage above can I not to... (3 Replies)
Discussion started by: alert0919
3 Replies

6. Solaris

Memory usage in Solaris - memory not freed?

Hi, I'm running a multi-process software system on a Solaris 8 machine. When I monitor the memory usage, I see that the free memory is dropping rapidly, but I can't detect a process that uses this memory. I'm using "top" to get the free memory and the memory usage of processes. Thanks. (3 Replies)
Discussion started by: gewurtz
3 Replies

7. Solaris

Memory usage in a Solaris Container

Hi All, We have a server with Solaris 10 installed. The total memory of the server is 64GB. In order to check the memory info of the server I use "top" utility that gives me total and free memory in real-time. I have also installed a Sun container (non-global zone) on top of parent operating... (4 Replies)
Discussion started by: Slayer
4 Replies

8. Solaris

Solaris Memory usage by oracle, java

Hello Gurus, In Office, I have unix Solaris Server. Oracle and Java is installed on Solaris Server. Oracle when starts then oracle loads SGA, PGA memory and graps solaris server memory. In the same manner java and other applications grabs memory for them. I want memory distribution of... (4 Replies)
Discussion started by: get_chandrakant
4 Replies

9. Solaris

Solaris memory usage

One of my Solaris Unix server has total RAM 128G . top & vmstat command shows free memory is 86G and usage is 42G. ps -eo pid,rss,commshows memory usage by process. when i sum all the memory usage by the process, it shows 1.9TB as shown below. $ ps -eo pid,pmem,vsz,rss,comm | sort -rnk2 |... (5 Replies)
Discussion started by: baladelaware73
5 Replies

10. Solaris

How to find out the memory usage in Solaris?

Hi All, In one of the solaris box aslert got triggered as ... (Used_Real_Mem_Pct=93.0 Used_Swap_Space_Pct=75.0 )] when i see the usage by vmstat and sar i am not able to relate the alert with the free memory and swap memory please help to understand the vmstat output as below.. kthr ... (4 Replies)
Discussion started by: Riverstone
4 Replies
CORE(5) 						     Linux Programmer's Manual							   CORE(5)

NAME
core - core dump file DESCRIPTION
The default action of certain signals is to cause a process to terminate and produce a core dump file, a disk file containing an image of the process's memory at the time of termination. This image can be used in a debugger (e.g., gdb(1)) to inspect the state of the program at the time that it terminated. A list of the signals which cause a process to dump core can be found in signal(7). A process can set its soft RLIMIT_CORE resource limit to place an upper limit on the size of the core dump file that will be produced if it receives a "core dump" signal; see getrlimit(2) for details. There are various circumstances in which a core dump file is not produced: * The process does not have permission to write the core file. (By default the core file is called core, and is created in the current working directory. See below for details on naming.) Writing the core file will fail if the directory in which it is to be created is nonwritable, or if a file with the same name exists and is not writable or is not a regular file (e.g., it is a directory or a symbolic link). * A (writable, regular) file with the same name as would be used for the core dump already exists, but there is more than one hard link to that file. * The file system where the core dump file would be created is full; or has run out of inodes; or is mounted read-only; or the user has reached their quota for the file system. * The directory in which the core dump file is to be created does not exist. * The RLIMIT_CORE (core file size) or RLIMIT_FSIZE (file size) resource limits for the process are set to zero; see getrlimit(2) and the documentation of the shell's ulimit command (limit in csh(1)). * The binary being executed by the process does not have read permission enabled. * The process is executing a set-user-ID (set-group-ID) program that is owned by a user (group) other than the real user (group) ID of the process. (However, see the description of the prctl(2) PR_SET_DUMPABLE operation, and the description of the /proc/sys/fs/suid_dumpable file in proc(5).) Naming of core dump files By default, a core dump file is named core, but the /proc/sys/kernel/core_pattern file (since Linux 2.6 and 2.4.21) can be set to define a template that is used to name core dump files. The template can contain % specifiers which are substituted by the following values when a core file is created: %% a single % character %p PID of dumped process %u (numeric) real UID of dumped process %g (numeric) real GID of dumped process %s number of signal causing dump %t time of dump, expressed as seconds since the Epoch, 1970-01-01 00:00:00 +0000 (UTC) %h hostname (same as nodename returned by uname(2)) %e executable filename (without path prefix) %c core file size soft resource limit of crashing process (since Linux 2.6.24) A single % at the end of the template is dropped from the core filename, as is the combination of a % followed by any character other than those listed above. All other characters in the template become a literal part of the core filename. The template may include '/' charac- ters, which are interpreted as delimiters for directory names. The maximum size of the resulting core filename is 128 bytes (64 bytes in kernels before 2.6.19). The default value in this file is "core". For backward compatibility, if /proc/sys/kernel/core_pattern does not include "%p" and /proc/sys/kernel/core_uses_pid (see below) is nonzero, then .PID will be appended to the core filename. Since version 2.4, Linux has also provided a more primitive method of controlling the name of the core dump file. If the /proc/sys/ker- nel/core_uses_pid file contains the value 0, then a core dump file is simply named core. If this file contains a nonzero value, then the core dump file includes the process ID in a name of the form core.PID. Piping core dumps to a program Since kernel 2.6.19, Linux supports an alternate syntax for the /proc/sys/kernel/core_pattern file. If the first character of this file is a pipe symbol (|), then the remainder of the line is interpreted as a program to be executed. Instead of being written to a disk file, the core dump is given as standard input to the program. Note the following points: * The program must be specified using an absolute pathname (or a pathname relative to the root directory, /), and must immediately follow the '|' character. * The process created to run the program runs as user and group root. * Command-line arguments can be supplied to the program (since kernel 2.6.24), delimited by white space (up to a total line length of 128 bytes). * The command-line arguments can include any of the % specifiers listed above. For example, to pass the PID of the process that is being dumped, specify %p in an argument. Controlling which mappings are written to the core dump Since kernel 2.6.23, the Linux-specific /proc/PID/coredump_filter file can be used to control which memory segments are written to the core dump file in the event that a core dump is performed for the process with the corresponding process ID. The value in the file is a bit mask of memory mapping types (see mmap(2)). If a bit is set in the mask, then memory mappings of the corre- sponding type are dumped; otherwise they are not dumped. The bits in this file have the following meanings: bit 0 Dump anonymous private mappings. bit 1 Dump anonymous shared mappings. bit 2 Dump file-backed private mappings. bit 3 Dump file-backed shared mappings. bit 4 (since Linux 2.6.24) Dump ELF headers. bit 5 (since Linux 2.6.28) Dump private huge pages. bit 6 (since Linux 2.6.28) Dump shared huge pages. By default, the following bits are set: 0, 1, 4 (if the CONFIG_CORE_DUMP_DEFAULT_ELF_HEADERS kernel configuration option is enabled), and 5. The value of this file is displayed in hexadecimal. (The default value is thus displayed as 33.) Memory-mapped I/O pages such as frame buffer are never dumped, and virtual DSO pages are always dumped, regardless of the coredump_filter value. A child process created via fork(2) inherits its parent's coredump_filter value; the coredump_filter value is preserved across an execve(2). It can be useful to set coredump_filter in the parent shell before running a program, for example: $ echo 0x7 > /proc/self/coredump_filter $ ./some_program This file is only provided if the kernel was built with the CONFIG_ELF_CORE configuration option. NOTES
The gdb(1) gcore command can be used to obtain a core dump of a running process. If a multithreaded process (or, more precisely, a process that shares its memory with another process by being created with the CLONE_VM flag of clone(2)) dumps core, then the process ID is always appended to the core filename, unless the process ID was already included else- where in the filename via a %p specification in /proc/sys/kernel/core_pattern. (This is primarily useful when employing the LinuxThreads implementation, where each thread of a process has a different PID.) EXAMPLE
The program below can be used to demonstrate the use of the pipe syntax in the /proc/sys/kernel/core_pattern file. The following shell session demonstrates the use of this program (compiled to create an executable named core_pattern_pipe_test): $ cc -o core_pattern_pipe_test core_pattern_pipe_test.c $ su Password: # echo '|$PWD/core_pattern_pipe_test %p UID=%u GID=%g sig=%s' > /proc/sys/kernel/core_pattern # exit $ sleep 100 ^ # type control-backslash Quit (core dumped) $ cat core.info argc=5 argc[0]=</home/mtk/core_pattern_pipe_test> argc[1]=<20575> argc[2]=<UID=1000> argc[3]=<GID=100> argc[4]=<sig=3> Total bytes in core dump: 282624 Program source /* core_pattern_pipe_test.c */ #define _GNU_SOURCE #include <sys/stat.h> #include <fcntl.h> #include <limits.h> #include <stdio.h> #include <stdlib.h> #include <unistd.h> #define BUF_SIZE 1024 int main(int argc, char *argv[]) { int tot, j; ssize_t nread; char buf[BUF_SIZE]; FILE *fp; char cwd[PATH_MAX]; /* Change our current working directory to that of the crashing process */ snprintf(cwd, PATH_MAX, "/proc/%s/cwd", argv[1]); chdir(cwd); /* Write output to file "core.info" in that directory */ fp = fopen("core.info", "w+"); if (fp == NULL) exit(EXIT_FAILURE); /* Display command-line arguments given to core_pattern pipe program */ fprintf(fp, "argc=%d ", argc); for (j = 0; j < argc; j++) fprintf(fp, "argc[%d]=<%s> ", j, argv[j]); /* Count bytes in standard input (the core dump) */ tot = 0; while ((nread = read(STDIN_FILENO, buf, BUF_SIZE)) > 0) tot += nread; fprintf(fp, "Total bytes in core dump: %d ", tot); exit(EXIT_SUCCESS); } SEE ALSO
bash(1), gdb(1), getrlimit(2), mmap(2), prctl(2), sigaction(2), elf(5), proc(5), pthreads(7), signal(7) COLOPHON
This page is part of release 3.27 of the Linux man-pages project. A description of the project, and information about reporting bugs, can be found at http://www.kernel.org/doc/man-pages/. Linux 2010-02-27 CORE(5)
All times are GMT -4. The time now is 04:20 PM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy