hdtodd
June 17, 2019, 12:18am
#1
I just updated 5.1.5 → 5.1.7, but I’ve seen the same issue on both installs: high ambient background load with no user tasks active:
hdtodd@Pi-6:~> uname -a
Linux Pi-6 5.1.7-1-default #1 SMP Tue Jun 4 07:56:54 UTC 2019 (55f2451) aarch64 aarch64 aarch64 GNU/Linux
hdtodd@Pi-6:~> uptime
16:11:36 up 0:13, 1 user, load average: 4.00, 3.82, 2.49
hdtodd@Pi-6:~>
If I run “top”, I see no culprit – nothing using more than 1% CPU and most using nothing. But performance is sluggish, so I think it’s more than just a load-calculation error.
Anyone else seeing this? Any idea what’s causing it and how to fix it?
hdtodd:
I just updated 5.1.5 → 5.1.7, but I’ve seen the same issue on both installs: high ambient background load with no user tasks active:
hdtodd@Pi-6:~> uname -a
Linux Pi-6 5.1.7-1-default #1 SMP Tue Jun 4 07:56:54 UTC 2019 (55f2451) aarch64 aarch64 aarch64 GNU/Linux
hdtodd@Pi-6:~> uptime
16:11:36 up 0:13, 1 user, load average: 4.00, 3.82, 2.49
hdtodd@Pi-6:~>
If I run “top”, I see no culprit – nothing using more than 1% CPU and most using nothing. But performance is sluggish, so I think it’s more than just a load-calculation error.
Anyone else seeing this? Any idea what’s causing it and how to fix it?
Hi
What about iostat, vmstat, iotop… not using swap?
iostat -x 5
vmstat 1
ps awwlx --sort=vsz (sorts by virtual sizes, low to high)
iotop
In your output, the 15 min time has not passed yet…
Install glances (python3-Glances), this should give the best overview but will add load…
What about disk i/o scheduler?
cat /sys/block/<disk_eg_mmcblk0>/queue/scheduler
hdtodd
June 17, 2019, 1:30am
#3
malcolmlewis:
Hi
What about iostat, vmstat, iotop… not using swap?
iostat -x 5
vmstat 1
ps awwlx --sort=vsz (sorts by virtual sizes, low to high)
iotop
In your output, the 15 min time has not passed yet…
Install glances (python3-Glances), this should give the best overview but will add load…
What about disk i/o scheduler?
cat /sys/block/<disk_eg_mmcblk0>/queue/scheduler
Thanks for the quick response, Malcolm!
After an hour …
Pi-6:/home/hdtodd # uptime
17:08:41 up 1:11, 1 user, load average: 4.00, 4.00, 4.00
Virtual sizes …
Pi-6:/home/hdtodd # ps awwlx --sort=-vsz
F UID PID PPID PRI NI VSZ RSS WCHAN STAT TTY TIME COMMAND
5 474 559 1 20 0 721304 3828 ep_pol Ssl ? 0:00 /usr/sbin/nscd
5 502 1952 1951 20 0 171992 3212 do_sig S ? 0:00 (sd-pam)
4 0 1 0 20 0 166620 9928 ep_pol Ss ? 0:22 /usr/lib/systemd/systemd --switched-root --system --deserialize 32
4 0 383 1 20 0 86688 1324 do_sel Ss ? 0:00 /usr/sbin/lvmetad -f
5 472 1600 1 20 0 78148 2104 do_sel S ? 0:00 /usr/sbin/chronyd
4 0 367 1 20 0 49624 16336 ep_pol Ss ? 0:02 /usr/lib/systemd/systemd-journald
4 0 394 1 20 0 29812 8176 ep_pol Ss ? 0:03 /usr/lib/systemd/systemd-udevd
4 0 1591 1 20 0 29492 7816 ep_pol Ss ? 0:00 /usr/sbin/cupsd -l
4 0 2113 1958 20 0 17708 5944 do_sys S pts/0 0:00 sudo -s
5 0 530 1 16 -4 17672 1640 ep_pol S<sl ? 0:00 /sbin/auditd
4 502 1951 1 20 0 16868 8132 ep_pol Ss ? 0:00 /usr/lib/systemd/systemd --user
4 0 1947 1634 20 0 16212 7532 do_sys Ss ? 0:00 sshd: hdtodd [priv]
5 502 1957 1947 20 0 16212 4508 - S ? 0:00 sshd: hdtodd@pts/0
4 0 614 1 20 0 15092 5880 ep_pol Ss ? 0:01 /usr/lib/systemd/systemd-logind
4 499 550 1 20 0 13548 5860 ep_pol Ss ? 0:02 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only
4 0 1634 1 20 0 12784 6504 do_sel Ss ? 0:00 /usr/sbin/sshd -D
4 0 664 1 20 0 12096 6944 do_sel Ss ? 0:00 /usr/sbin/wpa_supplicant -c /etc/wpa_supplicant/wpa_supplicant.conf -u -t -f /var/log/wpa_supplicant.log
0 502 1958 1957 20 0 10396 6864 do_wai Ss pts/0 0:00 -bash
4 0 2114 2113 20 0 10340 6808 do_wai S pts/0 0:00 /bin/bash
4 0 660 1 20 0 9860 6036 do_sys SLs ? 0:00 /usr/sbin/wickedd-nanny --systemd --foreground
4 0 655 1 20 0 9816 6252 do_sys SLs ? 0:00 /usr/sbin/wickedd --systemd --foreground
4 0 551 1 20 0 9700 6096 do_sys SLs ? 0:00 /usr/lib/wicked/bin/wickedd-dhcp6 --systemd --foreground
4 0 552 1 20 0 9700 6132 do_sys SLs ? 0:00 /usr/lib/wicked/bin/wickedd-dhcp4 --systemd --foreground
4 0 556 1 20 0 9696 5852 do_sys SLs ? 0:00 /usr/lib/wicked/bin/wickedd-auto4 --systemd --foreground
4 0 1658 1 20 0 7384 2444 hrtime Ss ? 0:00 /usr/sbin/cron -n
0 0 2236 2114 20 0 7184 2884 - R+ pts/0 0:00 ps awwlx --sort=-vsz
4 0 1597 1 20 0 2560 464 do_sel Ss+ tty1 0:00 /sbin/agetty -o -p -- \u --noclear tty1 linux
1 0 2 0 20 0 0 0 kthrea S ? 0:00 [kthreadd]
(rest have VSZ = 0)
VMSTAT:
Pi-6:/home/hdtodd # vmstat 1
procs -----------memory---------- ---swap-- -----io---- -system-- ------cpu-----
r b swpd free buff cache si so bi bo in cs us sy id wa st
0 0 0 525328 50456 291296 0 0 21 3 2046 34 1 0 99 0 0
0 0 0 525328 50456 291296 0 0 0 0 8158 144 0 0 100 0 0
0 0 0 525328 50456 291296 0 0 0 0 8329 316 0 0 100 0 0 ...
Not using MMC, using SDA:
Pi-6:/home/hdtodd # cat /sys/block/sda/queue/scheduler
[none] kyber bfq
IOTOP shows no activity.
And then, finally:
Pi-6:/home/hdtodd # uptime
17:22:48 up 1:25, 1 user, load average: 4.21, 4.14, 4.05
Pi-6:/home/hdtodd # iostat -x 5
Linux 5.1.7-1-default (Pi-6) 06/16/19 _aarch64_ (4 CPU)
avg-cpu: %user %nice %system %iowait %steal %idle
0.75 0.26 0.38 0.04 0.00 98.56
Device r/s w/s rkB/s wkB/s rrqm/s wrqm/s %rrqm %wrqm r_await w_await aqu-sz rareq-sz wareq-sz svctm %util
sda 6.11 0.78 75.53 17.44 0.00 0.58 0.00 42.64 0.86 1.15 0.00 12.37 22.22 1.46 1.01
avg-cpu: %user %nice %system %iowait %steal %idle
0.00 0.00 0.00 0.00 0.00 100.00
Device r/s w/s rkB/s wkB/s rrqm/s wrqm/s %rrqm %wrqm r_await w_await aqu-sz rareq-sz wareq-sz svctm %util
sda 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
Note 100% idle, but load >4 (so maybe it really is the load calculation!), but the system is very sluggish.
If I’m the only one seeing this, don’t worry about tracking it down: might get fixed in the next release. It just seemed odd, and I couldn’t find any sign of what was causing the problem … thought it might be an issue in the kernel.
David
hdtodd:
Thanks for the quick response, Malcolm!
After an hour …
Pi-6:/home/hdtodd # uptime
17:08:41 up 1:11, 1 user, load average: 4.00, 4.00, 4.00
Virtual sizes …
Pi-6:/home/hdtodd # ps awwlx --sort=-vsz
F UID PID PPID PRI NI VSZ RSS WCHAN STAT TTY TIME COMMAND
5 474 559 1 20 0 721304 3828 ep_pol Ssl ? 0:00 /usr/sbin/nscd
5 502 1952 1951 20 0 171992 3212 do_sig S ? 0:00 (sd-pam)
4 0 1 0 20 0 166620 9928 ep_pol Ss ? 0:22 /usr/lib/systemd/systemd --switched-root --system --deserialize 32
4 0 383 1 20 0 86688 1324 do_sel Ss ? 0:00 /usr/sbin/lvmetad -f
5 472 1600 1 20 0 78148 2104 do_sel S ? 0:00 /usr/sbin/chronyd
4 0 367 1 20 0 49624 16336 ep_pol Ss ? 0:02 /usr/lib/systemd/systemd-journald
4 0 394 1 20 0 29812 8176 ep_pol Ss ? 0:03 /usr/lib/systemd/systemd-udevd
4 0 1591 1 20 0 29492 7816 ep_pol Ss ? 0:00 /usr/sbin/cupsd -l
4 0 2113 1958 20 0 17708 5944 do_sys S pts/0 0:00 sudo -s
5 0 530 1 16 -4 17672 1640 ep_pol S<sl ? 0:00 /sbin/auditd
4 502 1951 1 20 0 16868 8132 ep_pol Ss ? 0:00 /usr/lib/systemd/systemd --user
4 0 1947 1634 20 0 16212 7532 do_sys Ss ? 0:00 sshd: hdtodd [priv]
5 502 1957 1947 20 0 16212 4508 - S ? 0:00 sshd: hdtodd@pts/0
4 0 614 1 20 0 15092 5880 ep_pol Ss ? 0:01 /usr/lib/systemd/systemd-logind
4 499 550 1 20 0 13548 5860 ep_pol Ss ? 0:02 /usr/bin/dbus-daemon --system --address=systemd: --nofork --nopidfile --systemd-activation --syslog-only
4 0 1634 1 20 0 12784 6504 do_sel Ss ? 0:00 /usr/sbin/sshd -D
4 0 664 1 20 0 12096 6944 do_sel Ss ? 0:00 /usr/sbin/wpa_supplicant -c /etc/wpa_supplicant/wpa_supplicant.conf -u -t -f /var/log/wpa_supplicant.log
0 502 1958 1957 20 0 10396 6864 do_wai Ss pts/0 0:00 -bash
4 0 2114 2113 20 0 10340 6808 do_wai S pts/0 0:00 /bin/bash
4 0 660 1 20 0 9860 6036 do_sys SLs ? 0:00 /usr/sbin/wickedd-nanny --systemd --foreground
4 0 655 1 20 0 9816 6252 do_sys SLs ? 0:00 /usr/sbin/wickedd --systemd --foreground
4 0 551 1 20 0 9700 6096 do_sys SLs ? 0:00 /usr/lib/wicked/bin/wickedd-dhcp6 --systemd --foreground
4 0 552 1 20 0 9700 6132 do_sys SLs ? 0:00 /usr/lib/wicked/bin/wickedd-dhcp4 --systemd --foreground
4 0 556 1 20 0 9696 5852 do_sys SLs ? 0:00 /usr/lib/wicked/bin/wickedd-auto4 --systemd --foreground
4 0 1658 1 20 0 7384 2444 hrtime Ss ? 0:00 /usr/sbin/cron -n
0 0 2236 2114 20 0 7184 2884 - R+ pts/0 0:00 ps awwlx --sort=-vsz
4 0 1597 1 20 0 2560 464 do_sel Ss+ tty1 0:00 /sbin/agetty -o -p -- \u --noclear tty1 linux
1 0 2 0 20 0 0 0 kthrea S ? 0:00 [kthreadd]
(rest have VSZ = 0)
VMSTAT:
Pi-6:/home/hdtodd # vmstat 1
procs -----------memory---------- ---swap-- -----io---- -system-- ------cpu-----
r b swpd free buff cache si so bi bo in cs us sy id wa st
0 0 0 525328 50456 291296 0 0 21 3 2046 34 1 0 99 0 0
0 0 0 525328 50456 291296 0 0 0 0 8158 144 0 0 100 0 0
0 0 0 525328 50456 291296 0 0 0 0 8329 316 0 0 100 0 0 ...
Not using MMC, using SDA:
Pi-6:/home/hdtodd # cat /sys/block/sda/queue/scheduler
[none] kyber bfq
IOTOP shows no activity.
And then, finally:
Pi-6:/home/hdtodd # uptime
17:22:48 up 1:25, 1 user, load average: 4.21, 4.14, 4.05
Pi-6:/home/hdtodd # iostat -x 5
Linux 5.1.7-1-default (Pi-6) 06/16/19 _aarch64_ (4 CPU)
avg-cpu: %user %nice %system %iowait %steal %idle
0.75 0.26 0.38 0.04 0.00 98.56
Device r/s w/s rkB/s wkB/s rrqm/s wrqm/s %rrqm %wrqm r_await w_await aqu-sz rareq-sz wareq-sz svctm %util
sda 6.11 0.78 75.53 17.44 0.00 0.58 0.00 42.64 0.86 1.15 0.00 12.37 22.22 1.46 1.01
avg-cpu: %user %nice %system %iowait %steal %idle
0.00 0.00 0.00 0.00 0.00 100.00
Device r/s w/s rkB/s wkB/s rrqm/s wrqm/s %rrqm %wrqm r_await w_await aqu-sz rareq-sz wareq-sz svctm %util
sda 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
Note 100% idle, but load >4 (so maybe it really is the load calculation!), but the system is very sluggish.
If I’m the only one seeing this, don’t worry about tracking it down: might get fixed in the next release. It just seemed odd, and I couldn’t find any sign of what was causing the problem … thought it might be an issue in the kernel.
David
Hi
I’m guessing it’s a bug with uptime and load calculation with cpu’s…
I get
uptime
19:09:01 up 1:12, 1 user, load average: 4.00, 4.00, 4.00
Which in IMHO should all be zero…
hdtodd
June 17, 2019, 4:52pm
#5
malcolmlewis:
Hi
I’m guessing it’s a bug with uptime and load calculation with cpu’s…
I get
uptime
19:09:01 up 1:12, 1 user, load average: 4.00, 4.00, 4.00
Which in IMHO should all be zero…
Yes, I’m concluding the same thing. Over time, the “uptime” reported load levels out at 4 (on a quad-core), but idle is reported 100% by other tools. Still seems sluggish (going in via SSH), but I agree that it’s a bug in “uptime”.