VERY slow performance on igb+FreeBSD8.2+mpd5.6
Коньков Евгений
kes-kes at yandex.ru
Sat Feb 25 21:29:49 UTC 2012
Здравствуйте, Коньков.
Вы писали 25 февраля 2012 г., 22:00:36:
КЕ> Здравствуйте, Volodymyr.
КЕ> Вы писали 24 февраля 2012 г., 10:10:12:
VK>> Коньков Евгений wrote:
>>>
>>> #uname FreeBSD 8.3-PRERELEASE #2 r231881: Thu Feb 23 00:53:28 UTC 2012
>>> и Version 5.6 (root@ 10:03 20-Feb-2012)
>>> http://www.speedtest.net/result/1790445113.png
>>> try to reconnect to mpd 10-20times and you get next:
>>> http://www.speedtest.net/result/1790454801.png
VK>> Used server differs in your images. Would you please track down assigned
VK>> IP's?
КЕ> I have load video on youtube.
КЕ> When low performance occur top -SHP shows next:
КЕ> PID USERNAME PRI NICE SIZE RES STATE C TIME WCPU COMMAND
КЕ> 11 root 171 ki31 0K 64K CPU2 2 222:21 100.00% idle{idle: cpu
КЕ> 11 root 171 ki31 0K 64K CPU1 1 222:11 100.00% idle{idle: cpu
КЕ> 11 root 171 ki31 0K 64K CPU3 3 222:07 100.00% idle{idle: cpu
КЕ> 11 root 171 ki31 0K 64K RUN 0 221:39 98.73% idle{idle: cpu0
КЕ> 12 root -32 - 0K 624K WAIT 1 1:30 0.00% intr{swi4: cloc
КЕ> 0 root -68 0 0K 384K - 3 1:03 0.00% kernel{dummynet
КЕ> 0 root -16 0 0K 384K sched 0 0:44 0.00% kernel{swapper}
КЕ> 12 root -44 - 0K 624K WAIT 3 0:31 0.00% intr{swi1: neti
КЕ> 12 root -44 - 0K 624K WAIT 3 0:11 0.00% intr{swi1: neti
КЕ> 12 root -68 - 0K 624K WAIT 2 0:07 0.00% intr{irq263: ig
КЕ> 12 root -44 - 0K 624K WAIT 1 0:06 0.00% intr{swi1: neti
КЕ> 13 root -16 - 0K 64K sleep 2 0:06 0.00% ng_queue{ng_que
КЕ> 13 root -16 - 0K 64K sleep 0 0:06 0.00% ng_queue{ng_que
КЕ> 13 root -16 - 0K 64K sleep 2 0:06 0.00% ng_queue{ng_que
КЕ> 13 root -16 - 0K 64K sleep 1 0:06 0.00% ng_queue{ng_que
КЕ> 12 root -68 - 0K 624K WAIT 3 0:05 0.00% intr{irq276: re
КЕ> it seems like no ticks are given to new subsystem
КЕ> what is comming on you can see on video (there are also shown netstat, vmstat etc)
КЕ> see video #3 this is best video then #2 then #1
КЕ> http://youtu.be/f90nMtNdKB8
when setting up values:
net.isr.bindthreads: 1
net.isr.direct: 1
net.isr.direct_force: 1
seems have another problem: dummynet take 1/4 of system CPU even when
no traffic at all (<10Kbit/s)
bwm-ng v0.6 (probing every 0.500s), press 'h' for help
input: getifaddrs type: rate
| iface Rx Tx Total
==============================================================================
igb0: 2.96 Kb/s 1.93 Kb/s 4.89 Kb/s
igb1: 2.09 Kb/s 1.93 Kb/s 4.02 Kb/s
igb2: 0.00 b/s 0.00 b/s 0.00 b/s
igb3: 2.80 Kb/s 0.00 b/s 2.80 Kb/s
re0: 956.18 b/s 9.28 Kb/s 10.21 Kb/s
------------------------------------------------------------------------------
total: 8.78 Kb/s 13.13 Kb/s 21.91 Kb/s
last pid: 37916; load averages: 0.01, 0.04, 0.06 up 0+00:11:43 23:20:40
148 processes: 5 running, 104 sleeping, 39 waiting
CPU 0: 0.0% user, 0.0% nice, 0.0% system, 0.0% interrupt, 100% idle
CPU 1: 0.0% user, 0.0% nice, 97.3% system, 0.0% interrupt, 2.7% idle
CPU 2: 0.0% user, 0.0% nice, 0.0% system, 0.0% interrupt, 100% idle
CPU 3: 0.0% user, 0.0% nice, 0.0% system, 0.0% interrupt, 100% idle
Mem: 92M Active, 14M Inact, 260M Wired, 308K Cache, 23M Buf, 3474M Free
Swap: 4096M Total, 4096M Free
PID USERNAME PRI NICE SIZE RES STATE C TIME WCPU COMMAND
11 root 171 ki31 0K 64K CPU3 3 11:37 100.00% idle{idle: cpu3}
11 root 171 ki31 0K 64K CPU2 2 11:31 100.00% idle{idle: cpu2}
11 root 171 ki31 0K 64K RUN 0 11:29 100.00% idle{idle: cpu0}
0 root -68 0 0K 384K - 1 0:05 82.23% kernel{dummynet}
11 root 171 ki31 0K 64K CPU1 1 11:34 20.12% idle{idle: cpu1}
12 root -32 - 0K 624K WAIT 0 0:05 0.10% intr{swi4: clock}
0 root -16 0 0K 384K sched 0 0:44 0.00% kernel{swapper}
12 root -68 - 0K 624K WAIT 2 0:02 0.00% intr{irq263: igb1:que}
12 root -68 - 0K 624K WAIT 1 0:01 0.00% intr{irq262: igb1:que}
13 root -16 - 0K 64K sleep 1 0:01 0.00% ng_queue{ng_queue3}
13 root -16 - 0K 64K sleep 1 0:01 0.00% ng_queue{ng_queue1}
13 root -16 - 0K 64K sleep 0 0:01 0.00% ng_queue{ng_queue2}
13 root -16 - 0K 64K sleep 1 0:01 0.00% ng_queue{ng_queue0}
12 root -68 - 0K 624K WAIT 3 0:00 0.00% intr{irq276: re0}
12 root -68 - 0K 624K WAIT 0 0:00 0.00% intr{irq261: igb1:que}
12 root -68 - 0K 624K WAIT 3 0:00 0.00% intr{irq264: igb1:que}
14 root -16 - 0K 16K - 1 0:00 0.00% yarrow
5292 root 44 -10 43876K 13760K select 1 0:00 0.00% mpd5{mpd5}
12 root -68 - 0K 624K WAIT 1 0:00 0.00% intr{irq257: igb0:que}
4070 root 44 0 5248K 3212K select 2 0:00 0.00% devd
12 root -68 - 0K 624K WAIT 2 0:00 0.00% intr{irq258: igb0:que}
2301 root 44 0 14548K 6968K select 2 0:00 0.00% bgpd
5226 bind 44 0 59228K 30092K ucond 0 0:00 0.00% named{named}
# ipfw show | grep queue
26275 1193 227000 queue 54 ip from any not 80,110 to any in recv vlan492
# ipfw show | grep pipe
# ipfw pipe show
00051: 170.000 Mbit/s 0 ms burst 0
q131123 50 sl. 0 flows (1 buckets) sched 65587 weight 0 lmax 0 pri 0
GRED w_q 0.001999 min_th 10 max_th 30 max_p 0.099991
sched 65587 type FIFO flags 0x1 64 buckets 0 active
mask: 0x00 0xffffffff/0x0000 -> 0x00000000/0x0000
BKT Prot ___Source IP/port____ ____Dest. IP/port____ Tot_pkt/bytes Pkt/Byte Drp
00054: 60.000 Mbit/s 0 ms burst 0
q131126 50 sl. 0 flows (1 buckets) sched 65590 weight 0 lmax 0 pri 0
GRED w_q 0.001999 min_th 10 max_th 30 max_p 0.099991
sched 65590 type FIFO flags 0x1 64 buckets 0 active
mask: 0x00 0x00000000/0x0000 -> 0xffffffff/0x0000
BKT Prot ___Source IP/port____ ____Dest. IP/port____ Tot_pkt/bytes Pkt/Byte Drp
00053: 60.000 Mbit/s 0 ms burst 0
q131125 50 sl. 0 flows (1 buckets) sched 65589 weight 0 lmax 0 pri 0
GRED w_q 0.001999 min_th 10 max_th 30 max_p 0.099991
sched 65589 type FIFO flags 0x1 64 buckets 0 active
mask: 0x00 0xffffffff/0x0000 -> 0x00000000/0x0000
BKT Prot ___Source IP/port____ ____Dest. IP/port____ Tot_pkt/bytes Pkt/Byte Drp
00052: 170.000 Mbit/s 0 ms burst 0
q131124 50 sl. 0 flows (1 buckets) sched 65588 weight 0 lmax 0 pri 0
GRED w_q 0.001999 min_th 10 max_th 30 max_p 0.099991
sched 65588 type FIFO flags 0x1 64 buckets 0 active
mask: 0x00 0x00000000/0x0000 -> 0xffffffff/0x0000
BKT Prot ___Source IP/port____ ____Dest. IP/port____ Tot_pkt/bytes Pkt/Byte Drp
# netstat -m
65778/1557/67335 mbufs in use (current/cache/total)
65776/790/66566/262144 mbuf clusters in use (current/cache/total/max)
65776/784 mbuf+clusters out of packet secondary zone in use (current/cache)
0/44/44/12800 4k (page size) jumbo clusters in use (current/cache/total/max)
0/0/0/6400 9k jumbo clusters in use (current/cache/total/max)
0/0/0/3200 16k jumbo clusters in use (current/cache/total/max)
147996K/2145K/150141K bytes allocated to network (current/cache/total)
0/0/0 requests for mbufs denied (mbufs/clusters/mbuf+clusters)
0/0/0 requests for jumbo clusters denied (4k/9k/16k)
0/0/0 sfbufs in use (current/peak/max)
0 requests for sfbufs denied
0 requests for sfbufs delayed
0 requests for I/O initiated by sendfile
0 calls to protocol drain routines
--
С уважением,
Коньков mailto:kes-kes at yandex.ru
More information about the freebsd-questions
mailing list