NACK: [PATCH 2/4] net: mana: Refactor RX buffer allocation code to prepare for various MTU

Roxana Nicolescu roxana.nicolescu at canonical.com
Wed Apr 19 08:33:36 UTC 2023


On 18/04/2023 17:56, Tim Gardner wrote:
> From: Haiyang Zhang <haiyangz at microsoft.com>
>
> BugLink: https://bugs.launchpad.net/bugs/2016898
>
> Move out common buffer allocation code from mana_process_rx_cqe() and
> mana_alloc_rx_wqe() to helper functions.
> Refactor related variables so they can be changed in one place, and buffer
> sizes are in sync.
>
> Signed-off-by: Haiyang Zhang <haiyangz at microsoft.com>
> Reviewed-by: Jesse Brandeburg <jesse.brandeburg at intel.com>
> Signed-off-by: David S. Miller <davem at davemloft.net>
> (cherry picked from commit a2917b23497e4205db32271e4e06e142a9f8a6aa linux-next)
> Signed-off-by: Tim Gardner <tim.gardner at canonical.com>
> ---
>   drivers/net/ethernet/microsoft/mana/mana_en.c | 154 ++++++++++--------
>   include/net/mana/mana.h                       |   6 +-
>   2 files changed, 91 insertions(+), 69 deletions(-)
>
> diff --git a/drivers/net/ethernet/microsoft/mana/mana_en.c b/drivers/net/ethernet/microsoft/mana/mana_en.c
> index a1b7905ed2f7..af0c0ee95d87 100644
> --- a/drivers/net/ethernet/microsoft/mana/mana_en.c
> +++ b/drivers/net/ethernet/microsoft/mana/mana_en.c
> @@ -1282,14 +1282,64 @@ static void mana_rx_skb(void *buf_va, struct mana_rxcomp_oob *cqe,
>   	u64_stats_update_end(&rx_stats->syncp);
>   
>   drop:
> -	WARN_ON_ONCE(rxq->xdp_save_page);
> -	rxq->xdp_save_page = virt_to_page(buf_va);
> +	WARN_ON_ONCE(rxq->xdp_save_va);
> +	/* Save for reuse */
> +	rxq->xdp_save_va = buf_va;
>   
>   	++ndev->stats.rx_dropped;
>   
>   	return;
>   }
>   
> +static void *mana_get_rxfrag(struct mana_rxq *rxq, struct device *dev,
> +			     dma_addr_t *da, bool is_napi)
> +{
> +	struct page *page;
> +	void *va;
> +
> +	/* Reuse XDP dropped page if available */
> +	if (rxq->xdp_save_va) {
> +		va = rxq->xdp_save_va;
> +		rxq->xdp_save_va = NULL;
> +	} else {
> +		page = dev_alloc_page();
> +		if (!page)
> +			return NULL;
> +
> +		va = page_to_virt(page);
> +	}
> +
> +	*da = dma_map_single(dev, va + XDP_PACKET_HEADROOM, rxq->datasize,
> +			     DMA_FROM_DEVICE);
> +
> +	if (dma_mapping_error(dev, *da)) {
> +		put_page(virt_to_head_page(va));
> +		return NULL;
> +	}
> +
> +	return va;
> +}
> +
> +/* Allocate frag for rx buffer, and save the old buf */
> +static void mana_refill_rxoob(struct device *dev, struct mana_rxq *rxq,
> +			      struct mana_recv_buf_oob *rxoob, void **old_buf)
> +{
> +	dma_addr_t da;
> +	void *va;
> +
> +	va = mana_get_rxfrag(rxq, dev, &da, true);
> +
> +	if (!va)
> +		return;
> +
> +	dma_unmap_single(dev, rxoob->sgl[0].address, rxq->datasize,
> +			 DMA_FROM_DEVICE);
> +	*old_buf = rxoob->buf_va;
> +
> +	rxoob->buf_va = va;
> +	rxoob->sgl[0].address = da;
> +}
> +
>   static void mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq,
>   				struct gdma_comp *cqe)
>   {
> @@ -1299,10 +1349,8 @@ static void mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq,
>   	struct mana_recv_buf_oob *rxbuf_oob;
>   	struct mana_port_context *apc;
>   	struct device *dev = gc->dev;
> -	void *new_buf, *old_buf;
> -	struct page *new_page;
> +	void *old_buf = NULL;
>   	u32 curr, pktlen;
> -	dma_addr_t da;
>   
>   	apc = netdev_priv(ndev);
>   
> @@ -1345,40 +1393,11 @@ static void mana_process_rx_cqe(struct mana_rxq *rxq, struct mana_cq *cq,
>   	rxbuf_oob = &rxq->rx_oobs[curr];
>   	WARN_ON_ONCE(rxbuf_oob->wqe_inf.wqe_size_in_bu != 1);
>   
> -	/* Reuse XDP dropped page if available */
> -	if (rxq->xdp_save_page) {
> -		new_page = rxq->xdp_save_page;
> -		rxq->xdp_save_page = NULL;
> -	} else {
> -		new_page = alloc_page(GFP_ATOMIC);
> -	}
> -
> -	if (new_page) {
> -		da = dma_map_page(dev, new_page, XDP_PACKET_HEADROOM, rxq->datasize,
> -				  DMA_FROM_DEVICE);
> -
> -		if (dma_mapping_error(dev, da)) {
> -			__free_page(new_page);
> -			new_page = NULL;
> -		}
> -	}
> -
> -	new_buf = new_page ? page_to_virt(new_page) : NULL;
> -
> -	if (new_buf) {
> -		dma_unmap_page(dev, rxbuf_oob->buf_dma_addr, rxq->datasize,
> -			       DMA_FROM_DEVICE);
> -
> -		old_buf = rxbuf_oob->buf_va;
> -
> -		/* refresh the rxbuf_oob with the new page */
> -		rxbuf_oob->buf_va = new_buf;
> -		rxbuf_oob->buf_dma_addr = da;
> -		rxbuf_oob->sgl[0].address = rxbuf_oob->buf_dma_addr;
> -	} else {
> -		old_buf = NULL; /* drop the packet if no memory */
> -	}
> +	mana_refill_rxoob(dev, rxq, rxbuf_oob, &old_buf);
>   
> +	/* Unsuccessful refill will have old_buf == NULL.
> +	 * In this case, mana_rx_skb() will drop the packet.
> +	 */
>   	mana_rx_skb(old_buf, oob, rxq);
>   
>   drop:
> @@ -1659,8 +1678,8 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
>   
>   	mana_deinit_cq(apc, &rxq->rx_cq);
>   
> -	if (rxq->xdp_save_page)
> -		__free_page(rxq->xdp_save_page);
> +	if (rxq->xdp_save_va)
> +		put_page(virt_to_head_page(rxq->xdp_save_va));
>   
>   	for (i = 0; i < rxq->num_rx_buf; i++) {
>   		rx_oob = &rxq->rx_oobs[i];
> @@ -1668,10 +1687,10 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
>   		if (!rx_oob->buf_va)
>   			continue;
>   
> -		dma_unmap_page(dev, rx_oob->buf_dma_addr, rxq->datasize,
> -			       DMA_FROM_DEVICE);
> +		dma_unmap_single(dev, rx_oob->sgl[0].address,
> +				 rx_oob->sgl[0].size, DMA_FROM_DEVICE);
>   
> -		free_page((unsigned long)rx_oob->buf_va);
> +		put_page(virt_to_head_page(rx_oob->buf_va));
>   		rx_oob->buf_va = NULL;
>   	}
>   
> @@ -1681,6 +1700,26 @@ static void mana_destroy_rxq(struct mana_port_context *apc,
>   	kfree(rxq);
>   }
>   
> +static int mana_fill_rx_oob(struct mana_recv_buf_oob *rx_oob, u32 mem_key,
> +			    struct mana_rxq *rxq, struct device *dev)
> +{
> +	dma_addr_t da;
> +	void *va;
> +
> +	va = mana_get_rxfrag(rxq, dev, &da, false);
> +
> +	if (!va)
> +		return -ENOMEM;
> +
> +	rx_oob->buf_va = va;
> +
> +	rx_oob->sgl[0].address = da;
> +	rx_oob->sgl[0].size = rxq->datasize;
> +	rx_oob->sgl[0].mem_key = mem_key;
> +
> +	return 0;
> +}
> +
>   #define MANA_WQE_HEADER_SIZE 16
>   #define MANA_WQE_SGE_SIZE 16
>   
> @@ -1690,9 +1729,8 @@ static int mana_alloc_rx_wqe(struct mana_port_context *apc,
>   	struct gdma_context *gc = apc->ac->gdma_dev->gdma_context;
>   	struct mana_recv_buf_oob *rx_oob;
>   	struct device *dev = gc->dev;
> -	struct page *page;
> -	dma_addr_t da;
>   	u32 buf_idx;
> +	int ret;
>   
>   	WARN_ON(rxq->datasize == 0 || rxq->datasize > PAGE_SIZE);
>   
> @@ -1703,25 +1741,12 @@ static int mana_alloc_rx_wqe(struct mana_port_context *apc,
>   		rx_oob = &rxq->rx_oobs[buf_idx];
>   		memset(rx_oob, 0, sizeof(*rx_oob));
>   
> -		page = alloc_page(GFP_KERNEL);
> -		if (!page)
> -			return -ENOMEM;
> -
> -		da = dma_map_page(dev, page, XDP_PACKET_HEADROOM, rxq->datasize,
> -				  DMA_FROM_DEVICE);
> -
> -		if (dma_mapping_error(dev, da)) {
> -			__free_page(page);
> -			return -ENOMEM;
> -		}
> -
> -		rx_oob->buf_va = page_to_virt(page);
> -		rx_oob->buf_dma_addr = da;
> -
>   		rx_oob->num_sge = 1;
> -		rx_oob->sgl[0].address = rx_oob->buf_dma_addr;
> -		rx_oob->sgl[0].size = rxq->datasize;
> -		rx_oob->sgl[0].mem_key = apc->ac->gdma_dev->gpa_mkey;
> +
> +		ret = mana_fill_rx_oob(rx_oob, apc->ac->gdma_dev->gpa_mkey, rxq,
> +				       dev);
> +		if (ret)
> +			return ret;
>   
>   		rx_oob->wqe_req.sgl = rx_oob->sgl;
>   		rx_oob->wqe_req.num_sge = rx_oob->num_sge;
> @@ -1780,9 +1805,10 @@ static struct mana_rxq *mana_create_rxq(struct mana_port_context *apc,
>   	rxq->ndev = ndev;
>   	rxq->num_rx_buf = RX_BUFFERS_PER_QUEUE;
>   	rxq->rxq_idx = rxq_idx;
> -	rxq->datasize = ALIGN(MAX_FRAME_SIZE, 64);
>   	rxq->rxobj = INVALID_MANA_HANDLE;
>   
> +	rxq->datasize = ALIGN(ETH_FRAME_LEN, 64);
> +
>   	err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size);
>   	if (err)
>   		goto out;
> diff --git a/include/net/mana/mana.h b/include/net/mana/mana.h
> index bb11a6535d80..037bcabf6b98 100644
> --- a/include/net/mana/mana.h
> +++ b/include/net/mana/mana.h
> @@ -36,9 +36,6 @@ enum TRI_STATE {
>   
>   #define COMP_ENTRY_SIZE 64
>   
> -#define ADAPTER_MTU_SIZE 1500
> -#define MAX_FRAME_SIZE (ADAPTER_MTU_SIZE + 14)
> -
>   #define RX_BUFFERS_PER_QUEUE 512
>   
>   #define MAX_SEND_BUFFERS_PER_QUEUE 256
> @@ -282,7 +279,6 @@ struct mana_recv_buf_oob {
>   	struct gdma_wqe_request wqe_req;
>   
>   	void *buf_va;
> -	dma_addr_t buf_dma_addr;
>   
>   	/* SGL of the buffer going to be sent has part of the work request. */
>   	u32 num_sge;
> @@ -322,7 +318,7 @@ struct mana_rxq {
>   
>   	struct bpf_prog __rcu *bpf_prog;
>   	struct xdp_rxq_info xdp_rxq;
> -	struct page *xdp_save_page;
> +	void *xdp_save_va; /* for reusing */
>   	bool xdp_flush;
>   	int xdp_rc; /* XDP redirect return code */
>   
It needs some context adjustment for jammy and kinetic.

Roxana



More information about the kernel-team mailing list