Sponsored Content
Operating Systems HP-UX Bad performance but Low CPU loading? Post 302416356 by methyl on Monday 26th of April 2010 08:33:48 AM
Old 04-26-2010
"midaemon" is part of MeasureWare. Whether you need it running 24/7 depends on whether you are using MeasureWare to monitor your system performance in detail and whether you have enough disc space to hold the files it creates in /var/opt/perf/datafiles .
The on/off switch for "midaemon" is in /etc/rc.config.d/mwa .
If it is not running it will start itself for the duration of a "glance" session.

Using "glance" itself the following options are useful clues about performance:
m = memory report
t = system tables
B = global waits
i = i/o by filesystem


The figures you post don't show a system under great stress. The disc utilisation is so low that I wonder if the active portion of the database fits into its own buffers or whether you have very slow discs. Normally I would expect a database system to be disc-bound. I can't imagine that anyone would configure Informix to use raw disc nowadays but that might be an explanation.

Last edited by methyl; 04-26-2010 at 10:43 AM.. Reason: spellin
 

10 More Discussions You Might Find Interesting

1. Programming

CPU Loading

How can I measure CPU loading (like performance monitor in Windows OS). I use Solaris but I would like to write portable code. Besides, I have to write programm to load CPU with known percent. How can I use CPU in 30% for example. Thanks for any ideas. (6 Replies)
Discussion started by: serge
6 Replies

2. UNIX for Dummies Questions & Answers

comparing Huge Files - Performance is very bad

Hi All, Can you please help me in resolving the following problem? My requirement is like this: 1) I have two files YESTERDAY_FILE and TODAY_FILE. Each one is having nearly two million data. 2) I need to check each record of TODAY_FILE in YESTERDAY_FILE. If exists we can skip that by... (5 Replies)
Discussion started by: madhukalyan
5 Replies

3. AIX

Bad performance when log in with putty

Hello guys! I'm n00b in AIX and I'm sticked in a problem. (my English is poor enough, but I hope you can understand me :P). So.. I'm trying to connect to an AIX machine with putty, and .. 'using username xxx' appears after 2 sec (OK), but 'xxx@ip's password' appears after 1:15 min. After... (6 Replies)
Discussion started by: combat2k
6 Replies

4. Solaris

Low average cpu utilization.

Hi to all, i have an app on solaris 5.8 writed in C++ (3.2.1) that use multi threading. Hardware has 8 cpu. When i run my app i note that the average of cpu go at least at 40%, and the performance are not so higher.. There is a cpu limitation on solaris, that dedicate only a part of cpu... (3 Replies)
Discussion started by: Moodie
3 Replies

5. Solaris

Sun T1000 application low performance

Hello All. I have Sun T1000 server with Solaris 10. On T1000 installed EMC smarts, application for monitoring network devices via SNMP + SNMP. So, Smarts has own DB (contains object - devices and relationships), file takes 30 mb, now, all queries to DB works very slow, so Smarts works too slow,... (5 Replies)
Discussion started by: hemulll
5 Replies

6. Solaris

By loading I have received the following issue: BAD PBR SIGN.

Hi folks. By disk cloning on Solaris x86, I used a command dd. I pulled out the source and inserted the new disk. By loading I have received the following issue: BAD PBR SIGN. :( (5 Replies)
Discussion started by: wolfgang
5 Replies

7. Solaris

Performance (iops) becomes bad, what is the reason?

I have written a virtual HBA driver named "xmp_vhba". A scsi disk is attached on it. as shown below: xmp_vhba, instance #0 disk, instance #11 But the performance became very bad when we read/write the scsi disk using the vdbench(a read/write io tool). What is the reason? ... (7 Replies)
Discussion started by: ForgetChen
7 Replies

8. AIX

High Runqueue (R) LOW CPU LOW I/O Low Network Low memory usage

Hello All I have a system running AIX 61 shared uncapped partition (with 11 physical processors, 24 Virtual 72GB of Memory) . The output from NMON, vmstat show a high run queue (60+) for continous periods of time intervals, but NO paging, relatively low I/o (6000) , CPU % is 40, Low network.... (9 Replies)
Discussion started by: IL-Malti
9 Replies

9. UNIX for Dummies Questions & Answers

CPU with long hours in top, is this bad?

Hi, We have a Solaris server that has about 43 Oracle databases on it and we also have the Oracle Enterprise Manager - emagent that is used to monitor these databases When running top, the emagent is showing as one of the top process. Excerpts from running top shows something as below: ... (3 Replies)
Discussion started by: newbie_01
3 Replies

10. AIX

AIX lpar bad disk I/O performance - 4k per IO limitation ?

Hi Guys, I have fresh new installed VIO 2.2.3.70 on a p710, 3 physical SAS disks, rootvg on hdisk0 and 3 VIO clients through vscsi, AIX7.1tl4 AIX6.1tl9 RHEL6.5ppc, each lpar has its rootvg installed on a LV on datavg (hdisk2) mapped to vhost0,1,2 There is no vg on hdisk1, I use it for my... (1 Reply)
Discussion started by: frenchy59
1 Replies
unix_master(9r) 														   unix_master(9r)

NAME
unix_master - General: Forces execution onto the master CPU SYNOPSIS
void unix_master( void ); ARGUMENTS
None DESCRIPTION
The unix_master routine forces execution of the kernel thread onto the master CPU (also called the boot CPU). In other words, unix_master binds the kernel thread to the master CPU. To release the kernel thread from the bind to the master CPU, call the unix_release routine. You can make recursive calls to unix_master as long as you make an equal number of calls to unix_release. The unix_master routine provides another way besides the simple and complex lock routines to make a kernel module symmetric multiprocessing (SMP) safe. Although calling unix_master is not optimal for performance on an SMP CPU, it does provide third-party kernel module writers with an easy way to make their modules SMP safe without using the lock routines. NOTES
Device drivers should not directly call the unix_master and unix_release routines. One exception to this recommendation is when you want a device driver's kernel threads to run only on the master CPU. This situation occurs when your driver creates and starts its own kernel threads and you set the d_funnel member of the associated dsent structure to the value DEV_FUNNEL. In this case, each kernel thread must call unix_master once to ensure that the kernel thread runs only on the master CPU. Remember to make a corresponding call to unix_release. CAUTIONS
To avoid deadlock, do not call the unix_master routine under the following circumstances: When holding a simple lock In the driver's inter- rupt service routine RETURN VALUES
None SEE ALSO
routines: unix_release(9r) unix_master(9r)
All times are GMT -4. The time now is 01:19 AM.
Unix & Linux Forums Content Copyright 1993-2022. All Rights Reserved.
Privacy Policy