Bug 45471
Summary: | Increasing load average over several kernel revisions | ||
---|---|---|---|
Product: | Process Management | Reporter: | Konstantin Svist (fry.kun) |
Component: | Other | Assignee: | process_other |
Status: | RESOLVED INVALID | ||
Severity: | normal | CC: | alan |
Priority: | P1 | ||
Hardware: | All | ||
OS: | Linux | ||
Kernel Version: | Subsystem: | ||
Regression: | Yes | Bisected commit-id: | |
Attachments: | lspci -nnn -vvv |
Description
Konstantin Svist
2012-08-03 01:08:12 UTC
Created attachment 76721 [details]
lspci -nnn -vvv
Version 3.4.2-1 still has load ~5-6, so the latest regression is between that and 3.4.6-1 Still checking other versions Load from ~1 to ~6 seems to originate somewhere between 2.6.38.6-26.rc1.fc15 and 2.6.43.8-1.fc15 -- can't find easily installable versions inbetween those so far After testing many different kernel versions, I'm more confused than ever. Same kernel version on F14 and F16 (2.6.35.14-97.fc14.x86_64) gives me very different results. The main culprit seems to be nginx process, which (according to sysprof) calls __close_nocancel -> kernel. In F14, it takes up ~5%, but in F16 (same kernel!) it takes up ~57% (most of which seems to be used on __mutex_lock_common.clone.5 35% and do_raw_spin_lock 13%) Load averages: f14: 4.70, 5.60, 5.95 f16: 36.99, 36.37, 32.72 Can anyone help? Excuse me, I saw that bug and it's not my case at all - my system is NOT idle and the load actually IS high, as referenced by top and sysprof By looking at vmstat output, it seems in retrospect that old kernels had calculated the load average incorrectly (or differently). With 40 tasks running, load average was reported in the single digits. Latest kernel versions (3.6, at least) seems to average out number of tasks running over time period and reports that as load average. |