From patchwork Mon Sep 10 08:24:39 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ilias Apalodimas X-Patchwork-Id: 146291 Delivered-To: patch@linaro.org Received: by 2002:a2e:1648:0:0:0:0:0 with SMTP id 8-v6csp2224000ljw; Mon, 10 Sep 2018 01:24:56 -0700 (PDT) X-Google-Smtp-Source: ANB0VdY3/fSGEIs9fzsvnRHQnxwtY+OQB+V4C5sKkIHpOW7USezG2YM7KVoCas0WvgJI8YByMsNU X-Received: by 2002:a63:1823:: with SMTP id y35-v6mr21389087pgl.438.1536567896245; Mon, 10 Sep 2018 01:24:56 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1536567896; cv=none; d=google.com; s=arc-20160816; b=wbNOfaGksfMrSSWgBfhWDGvr2KogNokv4bQNG2JISIdoOe1LfDVv4V3hb8UWnuRBGb iGP1MQyMFYbVvGnD5Ynzb5rR4Mxww9OlsuqJm4uAnqhsLHr5CFWtQDhZmwjMPfWZ96zc WnlMJWnHW8X6cdhiVvROe/baBEL/3NyzvpGeFyuxgHtmXZTm15kgs6yUGHuwhqIuITR6 bZBdT2QuvSib10iq86ZyV83w00Xwg9A8WMRPYRHFNDgyAbeJCtTTREY6aoayg701eCd+ eORGhy75QvQZxg0mhjBKDlxbO2k+r/vGA50p95Pl0BxDydw0U78z0aTZ5gMs33Mlr121 +84A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:dkim-signature; bh=+6r/Vczd8izUYdc9rxIdfjoAWn+U4i97H4htsH1eCxE=; b=CdrXHed06ExIj2nqjSydWsL8ONVB7QHOjJOmcxR8azR9O++ywbyu78ig9koZdhpUVH iMM+GJ4B9RiiCTkYfb5BmpkLmkV4vOABr8p+fktzC7XEl6FA+G9pPieIuiThqM/+FwrV vVmglk7eoflJR/aRDXEwfhKVGzY6gsE8amQbupwROdHRzgHo9VnWVePqS3VfB/TEOaWx P3PX5NL4+rnxUm0WNX6dc0R5Dz6Yo7kFoE2X5E5MJVe0FZQwRKTqYgSnbwWFUplfBPMx jEpKLMTKZep5Zl4P2HkdyI5S7JtpYEjhlvgaQEqSAKmAYFIEpBQlI5HPByQATvgnR7oI HZZg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=Sg4wYvLm; spf=pass (google.com: best guess record for domain of netdev-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id z127-v6si16982147pgz.370.2018.09.10.01.24.56; Mon, 10 Sep 2018 01:24:56 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of netdev-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b=Sg4wYvLm; spf=pass (google.com: best guess record for domain of netdev-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727944AbeIJNRp (ORCPT + 10 others); Mon, 10 Sep 2018 09:17:45 -0400 Received: from mail-wm0-f65.google.com ([74.125.82.65]:53999 "EHLO mail-wm0-f65.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727784AbeIJNRo (ORCPT ); Mon, 10 Sep 2018 09:17:44 -0400 Received: by mail-wm0-f65.google.com with SMTP id b19-v6so20589704wme.3 for ; Mon, 10 Sep 2018 01:24:52 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=+6r/Vczd8izUYdc9rxIdfjoAWn+U4i97H4htsH1eCxE=; b=Sg4wYvLm0a70oVA3M1C8KUNLvRbrZLJ1WM+yW/zZLiT1Gv7s0eBCUYC3WPNwXjq9Rb En3DD65Gf/jCKKoCXxdA1TtY7lItXJbt0t5hJdKG06san7IHrcA+QEbdwa6bOvXyKYyM Kip+D/s9dZImnBFb9C7VSO4H4hbRfOflS8svs= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=+6r/Vczd8izUYdc9rxIdfjoAWn+U4i97H4htsH1eCxE=; b=lWg6LqPv55gXcXWRupjN7hKKKnugu6K9QB/7r7OeOUGvBE6h746Mux5Uttt1HzAR1Y /edA1gKjfRLrWCCNmKibQvZJtuVbANbFLfaeokQTn3H9jotVVUrI+jj7NkF4+7o/owjE 9BzEX1O3g+2sqXZg2tctZ+9GW7JaYYAgPKfdQkFNn+/n/RK7c3ladkvBQgyWt5T2U9aE 4uunf0cb8/DTv3Yxr31V6W8ERkDPm9QnlFqjh3NsuQYr0Iog5fc23hIt605N6+qrkHq/ jL4iDt8pUabrSS9hO+5CmeKhM6pONrZy2bx+Y0XfOHcv8XKym3iRhQwq17IZUxkrUyPL xgHg== X-Gm-Message-State: APzg51BXUxtJZ53KbaDqDsHxE+Croas/8KOTs00o3lofxV2MEFmaLEVJ YMCUuDfxDNIRyD4i+mRydPy4+v7nD4o= X-Received: by 2002:a1c:cc1a:: with SMTP id h26-v6mr12669853wmb.112.1536567890996; Mon, 10 Sep 2018 01:24:50 -0700 (PDT) Received: from localhost.localdomain ([2a02:587:4607:6a00:2e56:dcff:fe9a:8f06]) by smtp.gmail.com with ESMTPSA id o6-v6sm10877529wmc.15.2018.09.10.01.24.49 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Mon, 10 Sep 2018 01:24:50 -0700 (PDT) From: Ilias Apalodimas To: netdev@vger.kernel.org, jaswinder.singh@linaro.org Cc: ard.biesheuvel@linaro.org, masami.hiramatsu@linaro.org, arnd@arndb.de, mykyta.iziumtsev@linaro.org, bjorn.topel@intel.com, magnus.karlsson@intel.com, Ilias Apalodimas Subject: [net-next, PATCH 1/2, v1] net: socionext: different approach on DMA Date: Mon, 10 Sep 2018 11:24:39 +0300 Message-Id: <1536567880-15097-2-git-send-email-ilias.apalodimas@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1536567880-15097-1-git-send-email-ilias.apalodimas@linaro.org> References: <1536567880-15097-1-git-send-email-ilias.apalodimas@linaro.org> Sender: netdev-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org Current driver dynamically allocates an skb and maps it as DMA rx buffer. A following patch introduces AF_XDP functionality, so we need a different allocation scheme. Buffers are allocated dynamically and mapped into hardware. During the Rx operation the driver uses build_skb() to produce the necessary buffers for the network stack Signed-off-by: Ilias Apalodimas --- drivers/net/ethernet/socionext/netsec.c | 239 +++++++++++++++++--------------- 1 file changed, 130 insertions(+), 109 deletions(-) -- 2.7.4 diff --git a/drivers/net/ethernet/socionext/netsec.c b/drivers/net/ethernet/socionext/netsec.c index 7aa5ebb..666fee2 100644 --- a/drivers/net/ethernet/socionext/netsec.c +++ b/drivers/net/ethernet/socionext/netsec.c @@ -296,6 +296,11 @@ struct netsec_rx_pkt_info { bool err_flag; }; +static void netsec_rx_fill(struct netsec_priv *priv, u16 from, u16 num); + +static void *netsec_alloc_rx_data(struct netsec_priv *priv, + dma_addr_t *dma_addr, u16 *len); + static void netsec_write(struct netsec_priv *priv, u32 reg_addr, u32 val) { writel(val, priv->ioaddr + reg_addr); @@ -556,34 +561,10 @@ static const struct ethtool_ops netsec_ethtool_ops = { /************* NETDEV_OPS FOLLOW *************/ -static struct sk_buff *netsec_alloc_skb(struct netsec_priv *priv, - struct netsec_desc *desc) -{ - struct sk_buff *skb; - - if (device_get_dma_attr(priv->dev) == DEV_DMA_COHERENT) { - skb = netdev_alloc_skb_ip_align(priv->ndev, desc->len); - } else { - desc->len = L1_CACHE_ALIGN(desc->len); - skb = netdev_alloc_skb(priv->ndev, desc->len); - } - if (!skb) - return NULL; - - desc->addr = skb->data; - desc->dma_addr = dma_map_single(priv->dev, desc->addr, desc->len, - DMA_FROM_DEVICE); - if (dma_mapping_error(priv->dev, desc->dma_addr)) { - dev_kfree_skb_any(skb); - return NULL; - } - return skb; -} static void netsec_set_rx_de(struct netsec_priv *priv, struct netsec_desc_ring *dring, u16 idx, - const struct netsec_desc *desc, - struct sk_buff *skb) + const struct netsec_desc *desc) { struct netsec_de *de = dring->vaddr + DESC_SZ * idx; u32 attr = (1 << NETSEC_RX_PKT_OWN_FIELD) | @@ -602,59 +583,6 @@ static void netsec_set_rx_de(struct netsec_priv *priv, dring->desc[idx].dma_addr = desc->dma_addr; dring->desc[idx].addr = desc->addr; dring->desc[idx].len = desc->len; - dring->desc[idx].skb = skb; -} - -static struct sk_buff *netsec_get_rx_de(struct netsec_priv *priv, - struct netsec_desc_ring *dring, - u16 idx, - struct netsec_rx_pkt_info *rxpi, - struct netsec_desc *desc, u16 *len) -{ - struct netsec_de de = {}; - - memcpy(&de, dring->vaddr + DESC_SZ * idx, DESC_SZ); - - *len = de.buf_len_info >> 16; - - rxpi->err_flag = (de.attr >> NETSEC_RX_PKT_ER_FIELD) & 1; - rxpi->rx_cksum_result = (de.attr >> NETSEC_RX_PKT_CO_FIELD) & 3; - rxpi->err_code = (de.attr >> NETSEC_RX_PKT_ERR_FIELD) & - NETSEC_RX_PKT_ERR_MASK; - *desc = dring->desc[idx]; - return desc->skb; -} - -static struct sk_buff *netsec_get_rx_pkt_data(struct netsec_priv *priv, - struct netsec_rx_pkt_info *rxpi, - struct netsec_desc *desc, - u16 *len) -{ - struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; - struct sk_buff *tmp_skb, *skb = NULL; - struct netsec_desc td; - int tail; - - *rxpi = (struct netsec_rx_pkt_info){}; - - td.len = priv->ndev->mtu + 22; - - tmp_skb = netsec_alloc_skb(priv, &td); - - tail = dring->tail; - - if (!tmp_skb) { - netsec_set_rx_de(priv, dring, tail, &dring->desc[tail], - dring->desc[tail].skb); - } else { - skb = netsec_get_rx_de(priv, dring, tail, rxpi, desc, len); - netsec_set_rx_de(priv, dring, tail, &td, tmp_skb); - } - - /* move tail ahead */ - dring->tail = (dring->tail + 1) % DESC_NUM; - - return skb; } static int netsec_clean_tx_dring(struct netsec_priv *priv, int budget) @@ -721,19 +649,29 @@ static int netsec_process_tx(struct netsec_priv *priv, int budget) return done; } +static void nsetsec_adv_desc(u16 *idx) +{ + *idx = *idx + 1; + if (unlikely(*idx >= DESC_NUM)) + *idx = 0; +} + static int netsec_process_rx(struct netsec_priv *priv, int budget) { struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; struct net_device *ndev = priv->ndev; - struct netsec_rx_pkt_info rx_info; - int done = 0; - struct netsec_desc desc; struct sk_buff *skb; - u16 len; + int done = 0; while (done < budget) { u16 idx = dring->tail; struct netsec_de *de = dring->vaddr + (DESC_SZ * idx); + struct netsec_desc *desc = &dring->desc[idx]; + struct netsec_rx_pkt_info rpi; + dma_addr_t dma_handle; + void *buf_addr; + u16 pkt_len; + u16 desc_len; if (de->attr & (1U << NETSEC_RX_PKT_OWN_FIELD)) break; @@ -744,28 +682,62 @@ static int netsec_process_rx(struct netsec_priv *priv, int budget) */ dma_rmb(); done++; - skb = netsec_get_rx_pkt_data(priv, &rx_info, &desc, &len); - if (unlikely(!skb) || rx_info.err_flag) { + + pkt_len = de->buf_len_info >> 16; + rpi.err_code = (de->attr >> NETSEC_RX_PKT_ERR_FIELD) & + NETSEC_RX_PKT_ERR_MASK; + rpi.err_flag = (de->attr >> NETSEC_RX_PKT_ER_FIELD) & 1; + if (rpi.err_flag) { netif_err(priv, drv, priv->ndev, - "%s: rx fail err(%d)\n", - __func__, rx_info.err_code); + "%s: rx fail err(%d)\n", __func__, + rpi.err_code); ndev->stats.rx_dropped++; + nsetsec_adv_desc(&dring->tail); + /* reuse buffer page frag */ + netsec_rx_fill(priv, idx, 1); continue; } + rpi.rx_cksum_result = (de->attr >> NETSEC_RX_PKT_CO_FIELD) & 3; - dma_unmap_single(priv->dev, desc.dma_addr, desc.len, - DMA_FROM_DEVICE); - skb_put(skb, len); + dma_sync_single_for_cpu(priv->dev, desc->dma_addr, pkt_len, + DMA_FROM_DEVICE); + + prefetch(desc->addr); + buf_addr = netsec_alloc_rx_data(priv, &dma_handle, &desc_len); + if (unlikely(!buf_addr)) + break; + + skb = build_skb(desc->addr, desc->len); + if (unlikely(!skb)) { + dma_unmap_single(priv->dev, dma_handle, desc_len, + DMA_TO_DEVICE); + skb_free_frag(buf_addr); + netif_err(priv, drv, priv->ndev, + "rx failed to alloc skb\n"); + break; + } + dma_unmap_single_attrs(priv->dev, desc->dma_addr, desc->len, + DMA_TO_DEVICE, DMA_ATTR_SKIP_CPU_SYNC); + + /* Update the descriptor with fresh buffers */ + desc->len = desc_len; + desc->dma_addr = dma_handle; + desc->addr = buf_addr; + + skb_put(skb, pkt_len); skb->protocol = eth_type_trans(skb, priv->ndev); if (priv->rx_cksum_offload_flag && - rx_info.rx_cksum_result == NETSEC_RX_CKSUM_OK) + rpi.rx_cksum_result == NETSEC_RX_CKSUM_OK) skb->ip_summed = CHECKSUM_UNNECESSARY; if (napi_gro_receive(&priv->napi, skb) != GRO_DROP) { ndev->stats.rx_packets++; - ndev->stats.rx_bytes += len; + ndev->stats.rx_bytes += pkt_len; } + + netsec_rx_fill(priv, idx, 1); + nsetsec_adv_desc(&dring->tail); } return done; @@ -928,7 +900,10 @@ static void netsec_uninit_pkt_dring(struct netsec_priv *priv, int id) dma_unmap_single(priv->dev, desc->dma_addr, desc->len, id == NETSEC_RING_RX ? DMA_FROM_DEVICE : DMA_TO_DEVICE); - dev_kfree_skb(desc->skb); + if (id == NETSEC_RING_RX) + skb_free_frag(desc->addr); + else if (id == NETSEC_RING_TX) + dev_kfree_skb(desc->skb); } memset(dring->desc, 0, sizeof(struct netsec_desc) * DESC_NUM); @@ -953,50 +928,96 @@ static void netsec_free_dring(struct netsec_priv *priv, int id) dring->desc = NULL; } +static void *netsec_alloc_rx_data(struct netsec_priv *priv, + dma_addr_t *dma_handle, u16 *desc_len) +{ + size_t len = priv->ndev->mtu + ETH_HLEN + VLAN_HLEN * 2 + NET_SKB_PAD + + NET_IP_ALIGN; + dma_addr_t mapping; + void *buf; + + len = SKB_DATA_ALIGN(len); + len += SKB_DATA_ALIGN(sizeof(struct skb_shared_info)); + + buf = napi_alloc_frag(len); + if (!buf) + return NULL; + + mapping = dma_map_single(priv->dev, buf, len, DMA_FROM_DEVICE); + if (unlikely(dma_mapping_error(priv->dev, mapping))) + goto err_out; + + *dma_handle = mapping; + *desc_len = len; + + return buf; + +err_out: + skb_free_frag(buf); + return NULL; +} + +static void netsec_rx_fill(struct netsec_priv *priv, u16 from, u16 num) +{ + struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; + u16 idx = from; + + while (num) { + netsec_set_rx_de(priv, dring, idx, &dring->desc[idx]); + idx++; + if (idx >= DESC_NUM) + idx = 0; + num--; + } +} + static int netsec_alloc_dring(struct netsec_priv *priv, enum ring_id id) { struct netsec_desc_ring *dring = &priv->desc_ring[id]; - int ret = 0; dring->vaddr = dma_zalloc_coherent(priv->dev, DESC_SZ * DESC_NUM, &dring->desc_dma, GFP_KERNEL); - if (!dring->vaddr) { - ret = -ENOMEM; + if (!dring->vaddr) goto err; - } dring->desc = kcalloc(DESC_NUM, sizeof(*dring->desc), GFP_KERNEL); - if (!dring->desc) { - ret = -ENOMEM; + if (!dring->desc) goto err; - } return 0; err: netsec_free_dring(priv, id); - return ret; + return -ENOMEM; } static int netsec_setup_rx_dring(struct netsec_priv *priv) { struct netsec_desc_ring *dring = &priv->desc_ring[NETSEC_RING_RX]; - struct netsec_desc desc; - struct sk_buff *skb; - int n; + int i; - desc.len = priv->ndev->mtu + 22; + for (i = 0; i < DESC_NUM; i++) { + struct netsec_desc *desc = &dring->desc[i]; + dma_addr_t dma_handle; + void *buf; + u16 len; - for (n = 0; n < DESC_NUM; n++) { - skb = netsec_alloc_skb(priv, &desc); - if (!skb) { + buf = netsec_alloc_rx_data(priv, &dma_handle, &len); + if (!buf) { netsec_uninit_pkt_dring(priv, NETSEC_RING_RX); - return -ENOMEM; + goto err_out; } - netsec_set_rx_de(priv, dring, n, &desc, skb); + desc->dma_addr = dma_handle; + desc->addr = buf; + desc->len = len; } + netsec_rx_fill(priv, 0, DESC_NUM); + return 0; + +err_out: + return -ENOMEM; } static int netsec_netdev_load_ucode_region(struct netsec_priv *priv, u32 reg,