From patchwork Fri Jul 24 09:57:59 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Li RongQing X-Patchwork-Id: 1335552 X-Patchwork-Delegate: davem@davemloft.net Return-Path: X-Original-To: patchwork-incoming-netdev@ozlabs.org Delivered-To: patchwork-incoming-netdev@ozlabs.org Authentication-Results: ozlabs.org; spf=pass (sender SPF authorized) smtp.mailfrom=vger.kernel.org (client-ip=23.128.96.18; helo=vger.kernel.org; envelope-from=netdev-owner@vger.kernel.org; receiver=) Authentication-Results: ozlabs.org; dmarc=none (p=none dis=none) header.from=baidu.com Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by ozlabs.org (Postfix) with ESMTP id 4BCl515whWz9sSn for ; Fri, 24 Jul 2020 19:58:13 +1000 (AEST) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727900AbgGXJ6N (ORCPT ); Fri, 24 Jul 2020 05:58:13 -0400 Received: from mx133-tc.baidu.com ([61.135.168.133]:25238 "EHLO tc-sys-mailedm03.tc.baidu.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726114AbgGXJ6M (ORCPT ); Fri, 24 Jul 2020 05:58:12 -0400 Received: from localhost (cp01-cos-dev01.cp01.baidu.com [10.92.119.46]) by tc-sys-mailedm03.tc.baidu.com (Postfix) with ESMTP id AD77E4500035; Fri, 24 Jul 2020 17:57:59 +0800 (CST) From: Li RongQing To: netdev@vger.kernel.org, intel-wired-lan@lists.osuosl.org, magnus.karlsson@intel.com Subject: [PATCH 2/2] ice/xdp: not adjust rx buffer for copy mode xdp Date: Fri, 24 Jul 2020 17:57:59 +0800 Message-Id: <1595584679-30652-2-git-send-email-lirongqing@baidu.com> X-Mailer: git-send-email 1.7.1 In-Reply-To: <1595584679-30652-1-git-send-email-lirongqing@baidu.com> References: <1595584679-30652-1-git-send-email-lirongqing@baidu.com> Sender: netdev-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org ice_rx_buf_adjust_pg_offset in copy mode xdp can lead to data corruption, like the following flow: 1. first skb is not for xsk, and forwarded to another device or socket queue 2. seconds skb is for xsk, copy data to xsk memory, and page of skb->data is released 3. rx_buff is reusable since only first skb is in it, but ice_rx_buf_adjust_pg_offset will make that page_offset is set to first skb data 4. then reuse rx buffer, first skb which still is living will be corrupted. so adjust rx buffer page offset when xdp memory type is MEM_TYPE_XSK_BUFF_POOL, or map type is not BPF_MAP_TYPE_XSKMAP which means that memory will be released immediately Fixes: 2d4238f55697 ("ice: Add support for AF_XDP") Signed-off-by: Li RongQing --- drivers/net/ethernet/intel/ice/ice_txrx.c | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/drivers/net/ethernet/intel/ice/ice_txrx.c b/drivers/net/ethernet/intel/ice/ice_txrx.c index abdb137c8bb7..6ceb1a0c33ae 100644 --- a/drivers/net/ethernet/intel/ice/ice_txrx.c +++ b/drivers/net/ethernet/intel/ice/ice_txrx.c @@ -1169,7 +1169,10 @@ int ice_clean_rx_irq(struct ice_ring *rx_ring, int budget) goto construct_skb; if (xdp_res & (ICE_XDP_TX | ICE_XDP_REDIR)) { xdp_xmit |= xdp_res; - ice_rx_buf_adjust_pg_offset(rx_buf, xdp.frame_sz); + + if (xdp.rxq->mem.type == MEM_TYPE_XSK_BUFF_POOL || + xdp_get_map_type_no_direct() != BPF_MAP_TYPE_XSKMAP) + ice_rx_buf_adjust_pg_offset(rx_buf, xdp.frame_sz); } else { rx_buf->pagecnt_bias++; }