Fwd: igb driver tx hangs when out of mbuf clusters
Karim Fodil-Lemelin
fodillemlinkarim at gmail.com
Tue Feb 8 02:21:46 UTC 2011
2011/2/7 Pyun YongHyeon <pyunyh at gmail.com>
> On Mon, Feb 07, 2011 at 05:33:47PM -0500, Karim Fodil-Lemelin wrote:
> > Subject: Re: igb driver tx hangs when out of mbuf clusters
> >
> > > To: Lev Serebryakov <lev at serebryakov.spb.ru>
> > > Cc: freebsd-net at freebsd.org
> > >
> > >
> > > 2011/2/7 Lev Serebryakov <lev at serebryakov.spb.ru>
> > >
> > > Hello, Karim.
> > >> You wrote 7 февраля 2011 г., 19:58:04:
> > >>
> > >>
> > >> > The issue is with the igb driver from 7.4 RC3 r218406. If the driver
> > >> runs
> > >> > out of mbuf clusters it simply stops receiving even after the
> clusters
> > >> have
> > >> > been freed.
> > >> It looks like my problems with em0 (see thread "em0 hangs without
> > >> any messages like "Watchdog timeout", only down/up reset it.")...
> > >> Codebase for em and igb is somewhat common...
> > >>
> > >> --
> > >> // Black Lion AKA Lev Serebryakov <lev at serebryakov.spb.ru>
> > >>
> > >> I agree.
> > >
> > > Do you get missed packets in mac_stats (sysctl dev.em | grep missed)?
> > >
> > > I might not have mentioned but I can also 'fix' the problem by doing
> > > ifconfig igb0 down/up.
> > >
> > > I will try using POLLING to 'automatize' the reset as you mentioned in
> your
> > > thread.
> > >
> > > Karim.
> > >
> > >
> > Follow up on tests with POLLING: The problem is still occurring although
> it
> > takes more time ... Outputs of sysctl dev.igb0 and netstat -m will
> follow:
> >
> > 9219/99426/108645 mbufs in use (current/cache/total)
> > 9217/90783/100000/100000 mbuf clusters in use (current/cache/total/max)
>
> Do you see network processes are stuck in keglim state? If you see
> that I think that's not trivial to solve. You wouldn't even kill
> that process if it is under keglim state unless some more mbuf
> clusters are freed from other places.
>
No keglim state, here is a snapshot of top -SH while the problem is
happening:
12 root 171 ki31 0K 8K CPU5 5 19:27 100.00% idle:
cpu5
10 root 171 ki31 0K 8K CPU7 7 19:26 100.00% idle:
cpu7
14 root 171 ki31 0K 8K CPU3 3 19:25 100.00% idle:
cpu3
11 root 171 ki31 0K 8K CPU6 6 19:25 100.00% idle:
cpu6
13 root 171 ki31 0K 8K CPU4 4 19:24 100.00% idle:
cpu4
15 root 171 ki31 0K 8K CPU2 2 19:22 100.00% idle:
cpu2
16 root 171 ki31 0K 8K CPU1 1 19:18 100.00% idle:
cpu1
17 root 171 ki31 0K 8K RUN 0 19:12 100.00% idle:
cpu0
18 root -32 - 0K 8K WAIT 6 0:04 0.10% swi4:
clock s
20 root -44 - 0K 8K WAIT 4 0:08 0.00% swi1: net
29 root -68 - 0K 8K - 0 0:02 0.00% igb0 que
35 root -68 - 0K 8K - 2 0:02 0.00% em1 taskq
28 root -68 - 0K 8K WAIT 5 0:01 0.00% irq256:
igb0
keep in mind that num_queues has been forced to 1.
>
> I think both igb(4) and em(4) pass received frame to upper stack
> before allocating new RX buffer. If driver fails to allocate new RX
> buffer driver will try to refill RX buffers in next run. Under
> extreme resource shortage case, this situation can produce no more
> RX buffers in RX descriptor ring and this will take the box out of
> network. Other drivers avoid that situation by allocating new RX
> buffer before passing received frame to upper stack. If RX buffer
> allocation fails driver will just reuse old RX buffer without
> passing received frame to upper stack. That does not completely
> solve the keglim issue though. I think you should have enough mbuf
> cluters to avoid keglim.
>
> However the output above indicates you have enough free mbuf
> clusters. So I guess igb(4) encountered zero available RX buffer
> situation in past but failed to refill the RX buffer again. I guess
> driver may be able to periodically check available RX buffers.
> Jack may have better idea if this was the case.(CCed)
>
That is exactly the pattern. The driver runs out of clusters but they
eventually get consumed and freed although the driver refuses to process any
new frames. It is, on the other hand, perfectly capable of sending out
packets.
> > 0/640 mbuf+clusters out of packet secondary zone in use (current/cache)
> > 0/12800/12800/12800 4k (page size) jumbo clusters in use
> > (current/cache/total/max)
> > 0/0/0/6400 9k jumbo clusters in use (current/cache/total/max)
> > 0/0/0/3200 16k jumbo clusters in use (current/cache/total/max)
> > 20738K/257622K/278361K bytes allocated to network (current/cache/total)
> > 0/291/0 requests for mbufs denied (mbufs/clusters/mbuf+clusters)
> > 0/0/0 requests for jumbo clusters denied (4k/9k/16k)
> > 0/5/6656 sfbufs in use (current/peak/max)
> > 0 requests for sfbufs denied
> > 0 requests for sfbufs delayed
> > 0 requests for I/O initiated by sendfile
> > 0 calls to protocol drain routines
> >
> > dev.igb.0.%desc: Intel(R) PRO/1000 Network Connection version - 2.0.7
> > dev.igb.0.%driver: igb
> > dev.igb.0.%location: slot=0 function=0
> > dev.igb.0.%pnpinfo: vendor=0x8086 device=0x10a7 subvendor=0x8086
> > subdevice=0x0000 class=0x020000
> > dev.igb.0.%parent: pci7
> > dev.igb.0.nvm: -1
> > dev.igb.0.flow_control: 3
> > dev.igb.0.enable_aim: 1
> > dev.igb.0.rx_processing_limit: 100
> > dev.igb.0.link_irq: 4
> > dev.igb.0.dropped: 0
> > dev.igb.0.tx_dma_fail: 0
> > dev.igb.0.rx_overruns: 464
> > dev.igb.0.watchdog_timeouts: 0
> > dev.igb.0.device_control: 1490027073
> > dev.igb.0.rx_control: 67141658
> > dev.igb.0.interrupt_mask: 0
> > dev.igb.0.extended_int_mask: 0
> > dev.igb.0.tx_buf_alloc: 14
> > dev.igb.0.rx_buf_alloc: 34
> > dev.igb.0.fc_high_water: 29488
> > dev.igb.0.fc_low_water: 29480
> > dev.igb.0.queue0.interrupt_rate: 111111
> > dev.igb.0.queue0.txd_head: 877
> > dev.igb.0.queue0.txd_tail: 877
> > dev.igb.0.queue0.no_desc_avail: 0
> > dev.igb.0.queue0.tx_packets: 92013
> > dev.igb.0.queue0.rxd_head: 570
> > dev.igb.0.queue0.rxd_tail: 570
> > dev.igb.0.queue0.rx_packets: 163386
> > dev.igb.0.queue0.rx_bytes: 240260310
> > dev.igb.0.queue0.lro_queued: 0
> > dev.igb.0.queue0.lro_flushed: 0
> > dev.igb.0.mac_stats.excess_coll: 0
> > dev.igb.0.mac_stats.single_coll: 0
> > dev.igb.0.mac_stats.multiple_coll: 0
> > dev.igb.0.mac_stats.late_coll: 0
> > dev.igb.0.mac_stats.collision_count: 0
> > dev.igb.0.mac_stats.symbol_errors: 0
> > dev.igb.0.mac_stats.sequence_errors: 0
> > dev.igb.0.mac_stats.defer_count: 0
> > dev.igb.0.mac_stats.missed_packets: 3104
> > dev.igb.0.mac_stats.recv_no_buff: 4016
> > dev.igb.0.mac_stats.recv_undersize: 0
> > dev.igb.0.mac_stats.recv_fragmented: 0
> > dev.igb.0.mac_stats.recv_oversize: 0
> > dev.igb.0.mac_stats.recv_jabber: 0
> > dev.igb.0.mac_stats.recv_errs: 0
> > dev.igb.0.mac_stats.crc_errs: 0
> > dev.igb.0.mac_stats.alignment_errs: 0
> > dev.igb.0.mac_stats.coll_ext_errs: 0
> > dev.igb.0.mac_stats.xon_recvd: 0
> > dev.igb.0.mac_stats.xon_txd: 346
> > dev.igb.0.mac_stats.xoff_recvd: 0
> > dev.igb.0.mac_stats.xoff_txd: 3450
> > dev.igb.0.mac_stats.total_pkts_recvd: 166515
> > dev.igb.0.mac_stats.good_pkts_recvd: 163411
> > dev.igb.0.mac_stats.bcast_pkts_recvd: 0
> > dev.igb.0.mac_stats.mcast_pkts_recvd: 51
> > dev.igb.0.mac_stats.rx_frames_64: 10
> > dev.igb.0.mac_stats.rx_frames_65_127: 1601
> > dev.igb.0.mac_stats.rx_frames_128_255: 53
> > dev.igb.0.mac_stats.rx_frames_256_511: 42
> > dev.igb.0.mac_stats.rx_frames_512_1023: 18
> > dev.igb.0.mac_stats.rx_frames_1024_1522: 161687
> > dev.igb.0.mac_stats.good_octets_recvd: 240948229
> > dev.igb.0.mac_stats.good_octets_txd: 5947150
> > dev.igb.0.mac_stats.total_pkts_txd: 95809
> > dev.igb.0.mac_stats.good_pkts_txd: 92013
> > dev.igb.0.mac_stats.bcast_pkts_txd: 1516
> > dev.igb.0.mac_stats.mcast_pkts_txd: 1817
> > dev.igb.0.mac_stats.tx_frames_64: 90302
> > dev.igb.0.mac_stats.tx_frames_65_127: 1711
> > dev.igb.0.mac_stats.tx_frames_128_255: 0
> > dev.igb.0.mac_stats.tx_frames_256_511: 0
> > dev.igb.0.mac_stats.tx_frames_512_1023: 0
> > dev.igb.0.mac_stats.tx_frames_1024_1522: 0
> > dev.igb.0.mac_stats.tso_txd: 0
> > dev.igb.0.mac_stats.tso_ctx_fail: 0
> > dev.igb.0.interrupts.asserts: 5584
> > dev.igb.0.interrupts.rx_pkt_timer: 163411
> > dev.igb.0.interrupts.rx_abs_timer: 163386
> > dev.igb.0.interrupts.tx_pkt_timer: 92013
> > dev.igb.0.interrupts.tx_abs_timer: 0
> > dev.igb.0.interrupts.tx_queue_empty: 92013
> > dev.igb.0.interrupts.tx_queue_min_thresh: 0
> > dev.igb.0.interrupts.rx_desc_min_thresh: 19
> > dev.igb.0.interrupts.rx_overrun: 0
> > dev.igb.0.host.breaker_tx_pkt: 0
> > dev.igb.0.host.host_tx_pkt_discard: 0
> > dev.igb.0.host.rx_pkt: 0
> > dev.igb.0.host.breaker_rx_pkts: 0
> > dev.igb.0.host.breaker_rx_pkt_drop: 0
> > dev.igb.0.host.tx_good_pkt: 0
> > dev.igb.0.host.breaker_tx_pkt_drop: 0
> > dev.igb.0.host.rx_good_bytes: 240948229
> > dev.igb.0.host.tx_good_bytes: 5947150
> > dev.igb.0.host.length_errors: 0
> > dev.igb.0.host.serdes_violation_pkt: 0
> > dev.igb.0.host.header_redir_missed: 0
>
More information about the freebsd-net
mailing list