Message ID | 49CBC8CC.1090807@cosmosbay.com |
---|---|
State | RFC, archived |
Delegated to: | David Miller |
Headers | show |
From: Eric Dumazet <dada1@cosmosbay.com> Date: Thu, 26 Mar 2009 19:26:20 +0100 > I dont have NIU hardware but it seems this driver could suffer from > starvation of high numbered RX queues under flood. I suspect other > multiqueue drivers might have same problem. > > With a standard budget of 64, we can consume all credit when handling > first queue(s), and last queues might discard packets. > > Solve this by rotating the starting point, so that we garantee to scan at > least one new queue at each niu_poll_core() invocation, even under stress. > > Also, change logic calling niu_sync_rx_discard_stats() to take > into account the device qlen, not bounded by budget (which could be 0) > > Signed-off-by: Eric Dumazet <dada1@cosmosbay.com> This change looks good to me, thanks for spotting this and suggesting the fix. It only hits people who don't have usable MSI-X interrupts, and hey if you have one of these cards what are you doing sticking it into such a machine :-) Anyways, I'll apply this directly, thanks a lot! -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
From: David Miller <davem@davemloft.net> Date: Fri, 27 Mar 2009 00:55:46 -0700 (PDT) > From: Eric Dumazet <dada1@cosmosbay.com> > Date: Thu, 26 Mar 2009 19:26:20 +0100 > > > I dont have NIU hardware but it seems this driver could suffer from > > starvation of high numbered RX queues under flood. I suspect other > > multiqueue drivers might have same problem. > > > > With a standard budget of 64, we can consume all credit when handling > > first queue(s), and last queues might discard packets. > > > > Solve this by rotating the starting point, so that we garantee to scan at > > least one new queue at each niu_poll_core() invocation, even under stress. > > > > Also, change logic calling niu_sync_rx_discard_stats() to take > > into account the device qlen, not bounded by budget (which could be 0) > > > > Signed-off-by: Eric Dumazet <dada1@cosmosbay.com> > > This change looks good to me, thanks for spotting this and > suggesting the fix. > > It only hits people who don't have usable MSI-X interrupts, > and hey if you have one of these cards what are you doing > sticking it into such a machine :-) > > Anyways, I'll apply this directly, thanks a lot! Actually, I take that back... I want to do this slightly differently. That niu_poll_core() is used in both the MSI-X and non-MSI case. I always meant to fix this driver to not scan over all the RX rings when we use MSI-X and the event can only be for one specific queue. Then we can add your rotating logic to the non-MSI-X interrupt paths. -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
David Miller a écrit : >> Anyways, I'll apply this directly, thanks a lot! > > Actually, I take that back... I want to do this slightly > differently. > > That niu_poll_core() is used in both the MSI-X and non-MSI > case. > > I always meant to fix this driver to not scan over all the > RX rings when we use MSI-X and the event can only be for > one specific queue. > > Then we can add your rotating logic to the non-MSI-X interrupt > paths. Great, I'll take care of this. Thanks -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/drivers/net/niu.c b/drivers/net/niu.c index 50c1112..d62d186 100644 --- a/drivers/net/niu.c +++ b/drivers/net/niu.c @@ -3726,8 +3726,8 @@ static int niu_rx_work(struct niu *np, struct rx_ring_info *rp, int budget) np->dev->name, rp->rx_channel, (unsigned long long) stat, qlen); rcr_done = work_done = 0; - qlen = min(qlen, budget); - while (work_done < qlen) { + budget = min(qlen, budget); + while (work_done < budget) { rcr_done += niu_process_rx_pkt(np, rp); work_done++; } @@ -3759,6 +3759,7 @@ static int niu_poll_core(struct niu *np, struct niu_ldg *lp, int budget) u32 tx_vec = (v0 >> 32); u32 rx_vec = (v0 & 0xffffffff); int i, work_done = 0; + unsigned int cur_rx_ring; niudbg(INTR, "%s: niu_poll_core() v0[%016llx]\n", np->dev->name, (unsigned long long) v0); @@ -3770,17 +3771,24 @@ static int niu_poll_core(struct niu *np, struct niu_ldg *lp, int budget) nw64(LD_IM0(LDN_TXDMA(rp->tx_channel)), 0); } + cur_rx_ring = np->last_rx_ring; for (i = 0; i < np->num_rx_rings; i++) { - struct rx_ring_info *rp = &np->rx_rings[i]; + struct rx_ring_info *rp; + if (++cur_rx_ring >= np->num_rx_rings) + cur_rx_ring = 0; + rp = &np->rx_rings[cur_rx_ring]; if (rx_vec & (1 << rp->rx_channel)) { int this_work_done; this_work_done = niu_rx_work(np, rp, budget); - - budget -= this_work_done; - work_done += this_work_done; + if (this_work_done) { + budget -= this_work_done; + work_done += this_work_done; + if (budget <= 0) + np->last_rx_ring = cur_rx_ring; + } } nw64(LD_IM0(LDN_RXDMA(rp->rx_channel)), 0); } @@ -4497,6 +4505,7 @@ static int niu_alloc_channels(struct niu *np) } np->num_rx_rings = parent->rxchan_per_port[port]; + np->last_rx_ring = 0; np->num_tx_rings = parent->txchan_per_port[port]; np->dev->real_num_tx_queues = np->num_tx_rings; diff --git a/drivers/net/niu.h b/drivers/net/niu.h index 8754e44..113fe1d 100644 --- a/drivers/net/niu.h +++ b/drivers/net/niu.h @@ -3266,6 +3266,7 @@ struct niu { struct tx_ring_info *tx_rings; int num_rx_rings; int num_tx_rings; + int last_rx_ring; struct niu_ldg ldg[NIU_NUM_LDG]; int num_ldg;
I dont have NIU hardware but it seems this driver could suffer from starvation of high numbered RX queues under flood. I suspect other multiqueue drivers might have same problem. With a standard budget of 64, we can consume all credit when handling first queue(s), and last queues might discard packets. Solve this by rotating the starting point, so that we garantee to scan at least one new queue at each niu_poll_core() invocation, even under stress. Also, change logic calling niu_sync_rx_discard_stats() to take into account the device qlen, not bounded by budget (which could be 0) Signed-off-by: Eric Dumazet <dada1@cosmosbay.com> -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html