From patchwork Tue May 5 14:08:26 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186203 Delivered-To: patch@linaro.org Received: by 2002:a92:8d81:0:0:0:0:0 with SMTP id w1csp1173989ill; Tue, 5 May 2020 07:10:42 -0700 (PDT) X-Google-Smtp-Source: APiQypK1z6MbkD/xtWjQ3xkilFEJcwZ9DpGs7FtAN7h4XGrBw84yYSGOgHLzEGJziCu3CNlGFxkb X-Received: by 2002:aa7:dcc3:: with SMTP id w3mr2618568edu.231.1588687842471; Tue, 05 May 2020 07:10:42 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588687842; cv=none; d=google.com; s=arc-20160816; b=W3skmrsZa56VjQ7DylvESp++G/sBRITQXOx7PbE7yxevoNn65xtOgMPBVQ0XHacDdV AjwW/9LibvUYPiflO4Pqs1RlgV48I9v5Kgi/iSfuIwJyvKGh2EYzGPQu+/pTgR+CKpHx jMwvM0R54KoQMK7kLUfhr6tPEx+nIg/XYWdEEEWOBuZgmMQB1gqke5qGxTKhpBGolNIV X+75PGXTjfuig0aGavZW/SsFPqg4/TdQ2iTJiA+illWf71F+Kh9fxKrXGglIFZPy/k0L ALi1ZvOkYQU7QnrUDTjuVB0+HDFAfXUrLrmnL02NRcZOcFd7S55snmlkh+dS/jiYtD0P CKVQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:message-id:date:cc:to :from; bh=8r0Q4kfzhK41jYbcooGCNP7Rnsd9i5Ujrc8swspwP1I=; b=rkZaZFOV9cRWTTyfMIoaI6nicwOGg7u78Fi332XHJYBiIZg8g1ANGJW9w6YE/RJu+D KL8hmeQlGQ17uhn9+Mf1sZ0zQxxvhbH6TAgW+W40q5z3HACvUm37beezvwbWYDctdy9y pcvufowabqGieolvOUdsXwfrH1Ik+uj+/vZJIPdoJls30DLcSVqr0MBPOZQtQnSvjhfa I+m139rnysFkQEbeJno4IXaoOTvcJuTrK7X+JrrgGeCC/4e9GxEvaeGHlm8KrM0gLg9b 0nQqWSuggL2sd6T0vOTXJ29tm9O1urzJsB3ExqvjouTNXoEcrYEhb1wUO/eLLwnHz9ww rezQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id la2si1203106ejb.371.2020.05.05.07.10.41; Tue, 05 May 2020 07:10:42 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id D4F191C2AA; Tue, 5 May 2020 16:10:40 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id 4D8851C1C8 for ; Tue, 5 May 2020 16:10:40 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id AC2D820084C; Tue, 5 May 2020 16:10:39 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id B556A20084D; Tue, 5 May 2020 16:10:37 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id 3BD9A4028F; Tue, 5 May 2020 22:10:35 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com Cc: Hemant Agrawal Date: Tue, 5 May 2020 19:38:26 +0530 Message-Id: <20200505140832.646-1-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH 1/7] common/dpaax: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- drivers/common/dpaax/dpaa_of.h | 15 +++++++++++++++ drivers/common/dpaax/dpaax_iova_table.h | 4 ++++ drivers/common/dpaax/rte_common_dpaax_version.map | 2 +- 3 files changed, 20 insertions(+), 1 deletion(-) -- 2.17.1 diff --git a/drivers/common/dpaax/dpaa_of.h b/drivers/common/dpaax/dpaa_of.h index 960b421766..38d91a1afe 100644 --- a/drivers/common/dpaax/dpaa_of.h +++ b/drivers/common/dpaax/dpaa_of.h @@ -24,6 +24,7 @@ #include #include #include +#include #ifndef OF_INIT_DEFAULT_PATH #define OF_INIT_DEFAULT_PATH "/proc/device-tree" @@ -102,6 +103,7 @@ struct dt_file { uint64_t buf[OF_FILE_BUF_MAX >> 3]; }; +__rte_internal const struct device_node *of_find_compatible_node( const struct device_node *from, const char *type __rte_unused, @@ -113,32 +115,44 @@ const struct device_node *of_find_compatible_node( dev_node != NULL; \ dev_node = of_find_compatible_node(dev_node, type, compatible)) +__rte_internal const void *of_get_property(const struct device_node *from, const char *name, size_t *lenp) __attribute__((nonnull(2))); +__rte_internal bool of_device_is_available(const struct device_node *dev_node); + +__rte_internal const struct device_node *of_find_node_by_phandle(uint64_t ph); +__rte_internal const struct device_node *of_get_parent(const struct device_node *dev_node); +__rte_internal const struct device_node *of_get_next_child(const struct device_node *dev_node, const struct device_node *prev); +__rte_internal const void *of_get_mac_address(const struct device_node *np); #define for_each_child_node(parent, child) \ for (child = of_get_next_child(parent, NULL); child != NULL; \ child = of_get_next_child(parent, child)) + +__rte_internal uint32_t of_n_addr_cells(const struct device_node *dev_node); uint32_t of_n_size_cells(const struct device_node *dev_node); +__rte_internal const uint32_t *of_get_address(const struct device_node *dev_node, size_t idx, uint64_t *size, uint32_t *flags); +__rte_internal uint64_t of_translate_address(const struct device_node *dev_node, const uint32_t *addr) __attribute__((nonnull)); +__rte_internal bool of_device_is_compatible(const struct device_node *dev_node, const char *compatible); @@ -146,6 +160,7 @@ bool of_device_is_compatible(const struct device_node *dev_node, * subsystem that is device-tree-dependent. Eg. Qman/Bman, config layers, etc. * The path should usually be "/proc/device-tree". */ +__rte_internal int of_init_path(const char *dt_path); /* of_finish() allows a controlled tear-down of the device-tree layer, eg. if a diff --git a/drivers/common/dpaax/dpaax_iova_table.h b/drivers/common/dpaax/dpaax_iova_table.h index fc3b9e7a8f..230fba8ba0 100644 --- a/drivers/common/dpaax/dpaax_iova_table.h +++ b/drivers/common/dpaax/dpaax_iova_table.h @@ -61,9 +61,13 @@ extern struct dpaax_iova_table *dpaax_iova_table_p; #define DPAAX_MEM_SPLIT_MASK_OFF (DPAAX_MEM_SPLIT - 1) /**< Offset */ /* APIs exposed */ +__rte_internal int dpaax_iova_table_populate(void); +__rte_internal void dpaax_iova_table_depopulate(void); +__rte_internal int dpaax_iova_table_update(phys_addr_t paddr, void *vaddr, size_t length); +__rte_internal void dpaax_iova_table_dump(void); static inline void *dpaax_iova_table_get_va(phys_addr_t paddr) __rte_hot; diff --git a/drivers/common/dpaax/rte_common_dpaax_version.map b/drivers/common/dpaax/rte_common_dpaax_version.map index f72eba761d..ad2b2b3fec 100644 --- a/drivers/common/dpaax/rte_common_dpaax_version.map +++ b/drivers/common/dpaax/rte_common_dpaax_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: dpaax_iova_table_depopulate; From patchwork Tue May 5 14:08:27 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186204 Delivered-To: patch@linaro.org Received: by 2002:a92:8d81:0:0:0:0:0 with SMTP id w1csp1174100ill; Tue, 5 May 2020 07:10:49 -0700 (PDT) X-Google-Smtp-Source: APiQypLCy05pqOC6ID4VN35eseR1EDCLhSV+VtKrpWdXEDl9Y7Sez/X2S0BS6foIXOX/kHaYP/ng X-Received: by 2002:a17:906:4048:: with SMTP id y8mr2878374ejj.258.1588687849158; Tue, 05 May 2020 07:10:49 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588687849; cv=none; d=google.com; s=arc-20160816; b=kuZW9iRb00a5FLnvGOAo6/A2cCunts1XkeLgfPhRYvnWebQJmUrs3JxUt2sl3jfC2r Fkg/7waHjUgGTQqPHcXOOStIaOyF48KEUJyPyc+mDMb0Y7FcJvXOwn8Kuz+i36Bs/MhZ s6KwRvwnXiHYnexKTIv6LKQPs6z9p15hLmbh3dacR/dvjvldJPMtHSEBisbEVZLRf2eK V/urjF4bBSchTCdweK/DXIfSekT46wgRKUNoR2VSe0rXlD6zd3Q26eU5+i6/PsEXIvHa vWuC0lXD7pL8BlXP2T2p/KuJz9DNUQ53txcdE5qKaZi6eQxsDBToZp7QEVhghNg1Yf5T kJxA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=Pjg3Z7ob3LLOZ0iSLiKrYklLVPnyNHmb03KcLM5XSpM=; b=0lONRwB850EKzWQ6rsLZW1nly729SS+2N2Qkv1atNs4Q7ATWyses3R6cz0EzXJsXu+ 2heT92KG6W8l6zMiVdKmsnFrk1lQhi4/dnBY+Ng9KOdwOQYJh8SGmS2V1mhuqbT4pY7d HlN3QS50sRY5r4KQJZ7zNdQ0EmjYFWK7UTYRz5r+WEVUJD3DRn8jQyQaA0F/OoI2ujeL icgMGVpVKT7KZ3G1SOMzOeWWHBreNEbeJZWkP+qN9YoHFUAtq0I5gD/ThByfhRpNbTlq 0hDX8EV5ejFpDkM8k5RMS+/xhlFody8/JHY8PcM58nfIn8zMKHb4OPqbytuF2FgRKFm6 FK8Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id dn8si1219074edb.142.2020.05.05.07.10.48; Tue, 05 May 2020 07:10:49 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 2C1FB1D5C1; Tue, 5 May 2020 16:10:42 +0200 (CEST) Received: from inva020.nxp.com (inva020.nxp.com [92.121.34.13]) by dpdk.org (Postfix) with ESMTP id 761F11C2AA for ; Tue, 5 May 2020 16:10:40 +0200 (CEST) Received: from inva020.nxp.com (localhost [127.0.0.1]) by inva020.eu-rdc02.nxp.com (Postfix) with ESMTP id 053881A07AF; Tue, 5 May 2020 16:10:40 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva020.eu-rdc02.nxp.com (Postfix) with ESMTP id 582C81A07BB; Tue, 5 May 2020 16:10:38 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id CDDAD402A8; Tue, 5 May 2020 22:10:35 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com Cc: Hemant Agrawal Date: Tue, 5 May 2020 19:38:27 +0530 Message-Id: <20200505140832.646-2-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200505140832.646-1-hemant.agrawal@nxp.com> References: <20200505140832.646-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH 2/7] mempool/dpaa2: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- drivers/mempool/dpaa/rte_mempool_dpaa_version.map | 2 +- drivers/mempool/dpaa2/dpaa2_hw_mempool.h | 1 + drivers/mempool/dpaa2/rte_mempool_dpaa2_version.map | 9 +++++++-- 3 files changed, 9 insertions(+), 3 deletions(-) -- 2.17.1 diff --git a/drivers/mempool/dpaa/rte_mempool_dpaa_version.map b/drivers/mempool/dpaa/rte_mempool_dpaa_version.map index 9eebaf7ffd..142547ee38 100644 --- a/drivers/mempool/dpaa/rte_mempool_dpaa_version.map +++ b/drivers/mempool/dpaa/rte_mempool_dpaa_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: rte_dpaa_bpid_info; diff --git a/drivers/mempool/dpaa2/dpaa2_hw_mempool.h b/drivers/mempool/dpaa2/dpaa2_hw_mempool.h index fa0f2280d5..53fa1552d1 100644 --- a/drivers/mempool/dpaa2/dpaa2_hw_mempool.h +++ b/drivers/mempool/dpaa2/dpaa2_hw_mempool.h @@ -61,6 +61,7 @@ struct dpaa2_bp_info { extern struct dpaa2_bp_info *rte_dpaa2_bpid_info; +__rte_internal int rte_dpaa2_mbuf_alloc_bulk(struct rte_mempool *pool, void **obj_table, unsigned int count); diff --git a/drivers/mempool/dpaa2/rte_mempool_dpaa2_version.map b/drivers/mempool/dpaa2/rte_mempool_dpaa2_version.map index cd4bc88273..686b024624 100644 --- a/drivers/mempool/dpaa2/rte_mempool_dpaa2_version.map +++ b/drivers/mempool/dpaa2/rte_mempool_dpaa2_version.map @@ -1,10 +1,15 @@ DPDK_20.0 { global: - rte_dpaa2_bpid_info; - rte_dpaa2_mbuf_alloc_bulk; rte_dpaa2_mbuf_from_buf_addr; rte_dpaa2_mbuf_pool_bpid; local: *; }; + +INTERNAL { + global: + + rte_dpaa2_bpid_info; + rte_dpaa2_mbuf_alloc_bulk; +}; From patchwork Tue May 5 14:08:28 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186205 Delivered-To: patch@linaro.org Received: by 2002:a92:8d81:0:0:0:0:0 with SMTP id w1csp1174284ill; Tue, 5 May 2020 07:11:01 -0700 (PDT) X-Google-Smtp-Source: APiQypLufhXuihZhRCg2Pk2gkB4Xfu3s0yVy4Li2BLx70gei4qKAKyqk9XghYazBWWmq98oTor/4 X-Received: by 2002:a17:906:e5a:: with SMTP id q26mr2652338eji.261.1588687861193; Tue, 05 May 2020 07:11:01 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588687861; cv=none; d=google.com; s=arc-20160816; b=YDAqaxLiXcyBDsZB18FRBgfYP96d5cBFEaPq3/Y80WPEHloeUn/l0KW0Ywyr/fjlW5 q04i1aBxEjQXNplv48FYY3+0tIFKZNzhrAxdH385CglmjSVyB5tBV8kF2px1hdO8JCsO KebIkNyACCS4+sMZF3DcuU7Vw/CoyDJrGflK/eNVrNkjEoF5wutf9lAHzm/i9kNDcDAH 7PZ2+A1NmuVmHaBmVyY9mMkMeQgOmAgvo5vQMe3bYK4knj4Gn0+eNwL6agP5ogtCxOsZ HzcKkU7wuZAp05ObnfQkSq0zrgRRbvL0Pb3wIKBhhYf1vUsXEWv/TCfV+5/o3Y9Lvg1b tCEg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=OhgxIX+OGiioziOuErdc89u82RoU1bs7PgXu1voPzPs=; b=RBRloubNpr99aKsBgk1Xf7eL4epipvqVQMX9Z3MGSqejmuIh8XcVn/ai/xAFU9RrpP ryshEsdkYbIR23/9DZGRzeMj+qsCdtk+W6gCgjILsNadQU41RR6XlCJ/80gtdguMP42V 6/czTzQgrcx3sVYuwqJtQ8tPkzlkLrsVZB1ZOK247dwTXzjWVevPsb9Ir7IlxdeOkDC7 nDTG4YeBiXWgAebCg+nCJnJNa892+MWHNqkQ0bx/IsJu0SuCSlhzWOKwTfuvX0Jff48f ROPfgoIqyMJYQ/YRaDfQs1CDZjJW3ao63pHU6E1VNq+3818gEEFVKBN6k2L38jNxz5qr 7wwQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id a35si1295005edf.377.2020.05.05.07.11.00; Tue, 05 May 2020 07:11:01 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 8C5231D5C8; Tue, 5 May 2020 16:10:43 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id 77BCB1D156 for ; Tue, 5 May 2020 16:10:41 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 1B69420084F; Tue, 5 May 2020 16:10:41 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 13A95200848; Tue, 5 May 2020 16:10:39 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id 717FF402BF; Tue, 5 May 2020 22:10:36 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com Cc: Hemant Agrawal Date: Tue, 5 May 2020 19:38:28 +0530 Message-Id: <20200505140832.646-3-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200505140832.646-1-hemant.agrawal@nxp.com> References: <20200505140832.646-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH 3/7] bus/fslmc: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- drivers/bus/fslmc/fslmc_vfio.h | 4 ++ drivers/bus/fslmc/mc/fsl_dpbp.h | 6 +++ drivers/bus/fslmc/mc/fsl_dpci.h | 3 ++ drivers/bus/fslmc/mc/fsl_dpcon.h | 2 + drivers/bus/fslmc/mc/fsl_dpdmai.h | 8 ++++ drivers/bus/fslmc/mc/fsl_dpio.h | 9 ++++ drivers/bus/fslmc/mc/fsl_dpmng.h | 2 + drivers/bus/fslmc/mc/fsl_mc_cmd.h | 1 + drivers/bus/fslmc/portal/dpaa2_hw_dpio.h | 5 +++ drivers/bus/fslmc/portal/dpaa2_hw_pvt.h | 8 ++++ .../bus/fslmc/qbman/include/fsl_qbman_debug.h | 3 ++ .../fslmc/qbman/include/fsl_qbman_portal.h | 41 +++++++++++++++++++ drivers/bus/fslmc/rte_bus_fslmc_version.map | 3 +- drivers/bus/fslmc/rte_fslmc.h | 4 ++ 14 files changed, 97 insertions(+), 2 deletions(-) -- 2.17.1 diff --git a/drivers/bus/fslmc/fslmc_vfio.h b/drivers/bus/fslmc/fslmc_vfio.h index c988121294..609e48aea3 100644 --- a/drivers/bus/fslmc/fslmc_vfio.h +++ b/drivers/bus/fslmc/fslmc_vfio.h @@ -41,7 +41,11 @@ typedef struct fslmc_vfio_container { } fslmc_vfio_container; extern char *fslmc_container; + +__rte_internal int rte_dpaa2_intr_enable(struct rte_intr_handle *intr_handle, int index); + +__rte_internal int rte_dpaa2_intr_disable(struct rte_intr_handle *intr_handle, int index); int rte_dpaa2_vfio_setup_intr(struct rte_intr_handle *intr_handle, diff --git a/drivers/bus/fslmc/mc/fsl_dpbp.h b/drivers/bus/fslmc/mc/fsl_dpbp.h index 9d405b42c4..7b537a21be 100644 --- a/drivers/bus/fslmc/mc/fsl_dpbp.h +++ b/drivers/bus/fslmc/mc/fsl_dpbp.h @@ -14,6 +14,7 @@ struct fsl_mc_io; +__rte_internal int dpbp_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpbp_id, @@ -42,10 +43,12 @@ int dpbp_destroy(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint32_t obj_id); +__rte_internal int dpbp_enable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpbp_disable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -55,6 +58,7 @@ int dpbp_is_enabled(struct fsl_mc_io *mc_io, uint16_t token, int *en); +__rte_internal int dpbp_reset(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -70,6 +74,7 @@ struct dpbp_attr { uint16_t bpid; }; +__rte_internal int dpbp_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -88,6 +93,7 @@ int dpbp_get_api_version(struct fsl_mc_io *mc_io, uint16_t *major_ver, uint16_t *minor_ver); +__rte_internal int dpbp_get_num_free_bufs(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpci.h b/drivers/bus/fslmc/mc/fsl_dpci.h index a0ee5bfe69..81fd3438aa 100644 --- a/drivers/bus/fslmc/mc/fsl_dpci.h +++ b/drivers/bus/fslmc/mc/fsl_dpci.h @@ -181,6 +181,7 @@ struct dpci_rx_queue_cfg { int order_preservation_en; }; +__rte_internal int dpci_set_rx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -228,6 +229,7 @@ int dpci_get_api_version(struct fsl_mc_io *mc_io, uint16_t *major_ver, uint16_t *minor_ver); +__rte_internal int dpci_set_opr(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -235,6 +237,7 @@ int dpci_set_opr(struct fsl_mc_io *mc_io, uint8_t options, struct opr_cfg *cfg); +__rte_internal int dpci_get_opr(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpcon.h b/drivers/bus/fslmc/mc/fsl_dpcon.h index af81d51195..7caa6c68a1 100644 --- a/drivers/bus/fslmc/mc/fsl_dpcon.h +++ b/drivers/bus/fslmc/mc/fsl_dpcon.h @@ -20,6 +20,7 @@ struct fsl_mc_io; */ #define DPCON_INVALID_DPIO_ID (int)(-1) +__rte_internal int dpcon_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpcon_id, @@ -77,6 +78,7 @@ struct dpcon_attr { uint8_t num_priorities; }; +__rte_internal int dpcon_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpdmai.h b/drivers/bus/fslmc/mc/fsl_dpdmai.h index 40469cc139..e7e8a5dda9 100644 --- a/drivers/bus/fslmc/mc/fsl_dpdmai.h +++ b/drivers/bus/fslmc/mc/fsl_dpdmai.h @@ -23,11 +23,13 @@ struct fsl_mc_io; */ #define DPDMAI_ALL_QUEUES (uint8_t)(-1) +__rte_internal int dpdmai_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpdmai_id, uint16_t *token); +__rte_internal int dpdmai_close(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -54,10 +56,12 @@ int dpdmai_destroy(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint32_t object_id); +__rte_internal int dpdmai_enable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpdmai_disable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -82,6 +86,7 @@ struct dpdmai_attr { uint8_t num_of_queues; }; +__rte_internal int dpdmai_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -148,6 +153,7 @@ struct dpdmai_rx_queue_cfg { }; +__rte_internal int dpdmai_set_rx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -168,6 +174,7 @@ struct dpdmai_rx_queue_attr { uint32_t fqid; }; +__rte_internal int dpdmai_get_rx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -184,6 +191,7 @@ struct dpdmai_tx_queue_attr { uint32_t fqid; }; +__rte_internal int dpdmai_get_tx_queue(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpio.h b/drivers/bus/fslmc/mc/fsl_dpio.h index 3158f53191..92e97db94b 100644 --- a/drivers/bus/fslmc/mc/fsl_dpio.h +++ b/drivers/bus/fslmc/mc/fsl_dpio.h @@ -13,11 +13,13 @@ struct fsl_mc_io; +__rte_internal int dpio_open(struct fsl_mc_io *mc_io, uint32_t cmd_flags, int dpio_id, uint16_t *token); +__rte_internal int dpio_close(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -57,10 +59,12 @@ int dpio_destroy(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint32_t object_id); +__rte_internal int dpio_enable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpio_disable(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); @@ -70,10 +74,12 @@ int dpio_is_enabled(struct fsl_mc_io *mc_io, uint16_t token, int *en); +__rte_internal int dpio_reset(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token); +__rte_internal int dpio_set_stashing_destination(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -84,12 +90,14 @@ int dpio_get_stashing_destination(struct fsl_mc_io *mc_io, uint16_t token, uint8_t *sdest); +__rte_internal int dpio_add_static_dequeue_channel(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, int dpcon_id, uint8_t *channel_index); +__rte_internal int dpio_remove_static_dequeue_channel(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, @@ -119,6 +127,7 @@ struct dpio_attr { uint32_t clk; }; +__rte_internal int dpio_get_attributes(struct fsl_mc_io *mc_io, uint32_t cmd_flags, uint16_t token, diff --git a/drivers/bus/fslmc/mc/fsl_dpmng.h b/drivers/bus/fslmc/mc/fsl_dpmng.h index 36c387af27..cdd8506625 100644 --- a/drivers/bus/fslmc/mc/fsl_dpmng.h +++ b/drivers/bus/fslmc/mc/fsl_dpmng.h @@ -34,6 +34,7 @@ struct mc_version { uint32_t revision; }; +__rte_internal int mc_get_version(struct fsl_mc_io *mc_io, uint32_t cmd_flags, struct mc_version *mc_ver_info); @@ -48,6 +49,7 @@ struct mc_soc_version { uint32_t pvr; }; +__rte_internal int mc_get_soc_version(struct fsl_mc_io *mc_io, uint32_t cmd_flags, struct mc_soc_version *mc_platform_info); diff --git a/drivers/bus/fslmc/mc/fsl_mc_cmd.h b/drivers/bus/fslmc/mc/fsl_mc_cmd.h index ac919610cf..06ea41a3b2 100644 --- a/drivers/bus/fslmc/mc/fsl_mc_cmd.h +++ b/drivers/bus/fslmc/mc/fsl_mc_cmd.h @@ -80,6 +80,7 @@ enum mc_cmd_status { #define MC_CMD_HDR_FLAGS_MASK 0xFF00FF00 +__rte_internal int mc_send_command(struct fsl_mc_io *mc_io, struct mc_command *cmd); static inline uint64_t mc_encode_cmd_header(uint16_t cmd_id, diff --git a/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h b/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h index 2829c93806..7c5966241a 100644 --- a/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h +++ b/drivers/bus/fslmc/portal/dpaa2_hw_dpio.h @@ -36,20 +36,25 @@ extern uint8_t dpaa2_eqcr_size; extern struct dpaa2_io_portal_t dpaa2_io_portal[RTE_MAX_LCORE]; /* Affine a DPIO portal to current processing thread */ +__rte_internal int dpaa2_affine_qbman_swp(void); /* Affine additional DPIO portal to current crypto processing thread */ +__rte_internal int dpaa2_affine_qbman_ethrx_swp(void); /* allocate memory for FQ - dq storage */ +__rte_internal int dpaa2_alloc_dq_storage(struct queue_storage_info_t *q_storage); /* free memory for FQ- dq storage */ +__rte_internal void dpaa2_free_dq_storage(struct queue_storage_info_t *q_storage); /* free the enqueue response descriptors */ +__rte_internal uint32_t dpaa2_free_eq_descriptors(void); diff --git a/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h b/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h index 368fe7c688..33b191f823 100644 --- a/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h +++ b/drivers/bus/fslmc/portal/dpaa2_hw_pvt.h @@ -426,11 +426,19 @@ void set_swp_active_dqs(uint16_t dpio_index, struct qbman_result *dqs) { rte_global_active_dqs_list[dpio_index].global_active_dqs = dqs; } +__rte_internal struct dpaa2_dpbp_dev *dpaa2_alloc_dpbp_dev(void); + +__rte_internal void dpaa2_free_dpbp_dev(struct dpaa2_dpbp_dev *dpbp); + +__rte_internal int dpaa2_dpbp_supported(void); +__rte_internal struct dpaa2_dpci_dev *rte_dpaa2_alloc_dpci_dev(void); + +__rte_internal void rte_dpaa2_free_dpci_dev(struct dpaa2_dpci_dev *dpci); #endif diff --git a/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h b/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h index e010b1b6ae..328f2022fc 100644 --- a/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h +++ b/drivers/bus/fslmc/qbman/include/fsl_qbman_debug.h @@ -24,7 +24,10 @@ uint8_t verb; uint8_t reserved2[29]; }; +__rte_internal int qbman_fq_query_state(struct qbman_swp *s, uint32_t fqid, struct qbman_fq_query_np_rslt *r); + +__rte_internal uint32_t qbman_fq_state_frame_count(const struct qbman_fq_query_np_rslt *r); uint32_t qbman_fq_state_byte_count(const struct qbman_fq_query_np_rslt *r); diff --git a/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h b/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h index 88f0a99686..7ac0f82106 100644 --- a/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h +++ b/drivers/bus/fslmc/qbman/include/fsl_qbman_portal.h @@ -117,6 +117,7 @@ uint32_t qbman_swp_interrupt_read_status(struct qbman_swp *p); * @p: the given software portal object. * @mask: The value to set in SWP_ISR register. */ +__rte_internal void qbman_swp_interrupt_clear_status(struct qbman_swp *p, uint32_t mask); /** @@ -286,6 +287,7 @@ void qbman_swp_push_get(struct qbman_swp *s, uint8_t channel_idx, int *enabled); * rather by specifying the index (from 0 to 15) that has been mapped to the * desired channel. */ +__rte_internal void qbman_swp_push_set(struct qbman_swp *s, uint8_t channel_idx, int enable); /* ------------------- */ @@ -325,6 +327,7 @@ enum qbman_pull_type_e { * default/starting state. * @d: the pull dequeue descriptor to be cleared. */ +__rte_internal void qbman_pull_desc_clear(struct qbman_pull_desc *d); /** @@ -340,6 +343,7 @@ void qbman_pull_desc_clear(struct qbman_pull_desc *d); * the caller provides in 'storage_phys'), and 'stash' controls whether or not * those writes to main-memory express a cache-warming attribute. */ +__rte_internal void qbman_pull_desc_set_storage(struct qbman_pull_desc *d, struct qbman_result *storage, uint64_t storage_phys, @@ -349,6 +353,7 @@ void qbman_pull_desc_set_storage(struct qbman_pull_desc *d, * @d: the pull dequeue descriptor to be set. * @numframes: number of frames to be set, must be between 1 and 16, inclusive. */ +__rte_internal void qbman_pull_desc_set_numframes(struct qbman_pull_desc *d, uint8_t numframes); /** @@ -372,6 +377,7 @@ void qbman_pull_desc_set_token(struct qbman_pull_desc *d, uint8_t token); * qbman_pull_desc_set_fq() - Set fqid from which the dequeue command dequeues. * @fqid: the frame queue index of the given FQ. */ +__rte_internal void qbman_pull_desc_set_fq(struct qbman_pull_desc *d, uint32_t fqid); /** @@ -407,6 +413,7 @@ void qbman_pull_desc_set_rad(struct qbman_pull_desc *d, int rad); * Return 0 for success, and -EBUSY if the software portal is not ready * to do pull dequeue. */ +__rte_internal int qbman_swp_pull(struct qbman_swp *s, struct qbman_pull_desc *d); /* -------------------------------- */ @@ -421,12 +428,14 @@ int qbman_swp_pull(struct qbman_swp *s, struct qbman_pull_desc *d); * only once, so repeated calls can return a sequence of DQRR entries, without * requiring they be consumed immediately or in any particular order. */ +__rte_internal const struct qbman_result *qbman_swp_dqrr_next(struct qbman_swp *p); /** * qbman_swp_prefetch_dqrr_next() - prefetch the next DQRR entry. * @s: the software portal object. */ +__rte_internal void qbman_swp_prefetch_dqrr_next(struct qbman_swp *s); /** @@ -435,6 +444,7 @@ void qbman_swp_prefetch_dqrr_next(struct qbman_swp *s); * @s: the software portal object. * @dq: the DQRR entry to be consumed. */ +__rte_internal void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct qbman_result *dq); /** @@ -442,6 +452,7 @@ void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct qbman_result *dq); * @s: the software portal object. * @dqrr_index: the DQRR index entry to be consumed. */ +__rte_internal void qbman_swp_dqrr_idx_consume(struct qbman_swp *s, uint8_t dqrr_index); /** @@ -450,6 +461,7 @@ void qbman_swp_dqrr_idx_consume(struct qbman_swp *s, uint8_t dqrr_index); * * Return dqrr index. */ +__rte_internal uint8_t qbman_get_dqrr_idx(const struct qbman_result *dqrr); /** @@ -460,6 +472,7 @@ uint8_t qbman_get_dqrr_idx(const struct qbman_result *dqrr); * * Return dqrr entry object. */ +__rte_internal struct qbman_result *qbman_get_dqrr_from_idx(struct qbman_swp *s, uint8_t idx); /* ------------------------------------------------- */ @@ -485,6 +498,7 @@ struct qbman_result *qbman_get_dqrr_from_idx(struct qbman_swp *s, uint8_t idx); * Return 1 for getting a valid dequeue result, or 0 for not getting a valid * dequeue result. */ +__rte_internal int qbman_result_has_new_result(struct qbman_swp *s, struct qbman_result *dq); @@ -497,8 +511,10 @@ int qbman_result_has_new_result(struct qbman_swp *s, * Return 1 for getting a valid dequeue result, or 0 for not getting a valid * dequeue result. */ +__rte_internal int qbman_check_command_complete(struct qbman_result *dq); +__rte_internal int qbman_check_new_result(struct qbman_result *dq); /* -------------------------------------------------------- */ @@ -624,6 +640,7 @@ int qbman_result_is_FQPN(const struct qbman_result *dq); * * Return the state field. */ +__rte_internal uint8_t qbman_result_DQ_flags(const struct qbman_result *dq); /** @@ -658,6 +675,7 @@ static inline int qbman_result_DQ_is_pull_complete( * * Return seqnum. */ +__rte_internal uint16_t qbman_result_DQ_seqnum(const struct qbman_result *dq); /** @@ -667,6 +685,7 @@ uint16_t qbman_result_DQ_seqnum(const struct qbman_result *dq); * * Return odpid. */ +__rte_internal uint16_t qbman_result_DQ_odpid(const struct qbman_result *dq); /** @@ -699,6 +718,7 @@ uint32_t qbman_result_DQ_frame_count(const struct qbman_result *dq); * * Return the frame queue context. */ +__rte_internal uint64_t qbman_result_DQ_fqd_ctx(const struct qbman_result *dq); /** @@ -707,6 +727,7 @@ uint64_t qbman_result_DQ_fqd_ctx(const struct qbman_result *dq); * * Return the frame descriptor. */ +__rte_internal const struct qbman_fd *qbman_result_DQ_fd(const struct qbman_result *dq); /* State-change notifications (FQDAN/CDAN/CSCN/...). */ @@ -717,6 +738,7 @@ const struct qbman_fd *qbman_result_DQ_fd(const struct qbman_result *dq); * * Return the state in the notifiation. */ +__rte_internal uint8_t qbman_result_SCN_state(const struct qbman_result *scn); /** @@ -850,6 +872,7 @@ struct qbman_eq_response { * default/starting state. * @d: the given enqueue descriptor. */ +__rte_internal void qbman_eq_desc_clear(struct qbman_eq_desc *d); /* Exactly one of the following descriptor "actions" should be set. (Calling @@ -870,6 +893,7 @@ void qbman_eq_desc_clear(struct qbman_eq_desc *d); * @response_success: 1 = enqueue with response always; 0 = enqueue with * rejections returned on a FQ. */ +__rte_internal void qbman_eq_desc_set_no_orp(struct qbman_eq_desc *d, int respond_success); /** * qbman_eq_desc_set_orp() - Set order-resotration in the enqueue descriptor @@ -881,6 +905,7 @@ void qbman_eq_desc_set_no_orp(struct qbman_eq_desc *d, int respond_success); * @incomplete: indiates whether this is the last fragments using the same * sequeue number. */ +__rte_internal void qbman_eq_desc_set_orp(struct qbman_eq_desc *d, int respond_success, uint16_t opr_id, uint16_t seqnum, int incomplete); @@ -915,6 +940,7 @@ void qbman_eq_desc_set_orp_nesn(struct qbman_eq_desc *d, uint16_t opr_id, * data structure.) 'stash' controls whether or not the write to main-memory * expresses a cache-warming attribute. */ +__rte_internal void qbman_eq_desc_set_response(struct qbman_eq_desc *d, uint64_t storage_phys, int stash); @@ -929,6 +955,7 @@ void qbman_eq_desc_set_response(struct qbman_eq_desc *d, * result "storage" before issuing an enqueue, and use any non-zero 'token' * value. */ +__rte_internal void qbman_eq_desc_set_token(struct qbman_eq_desc *d, uint8_t token); /** @@ -944,6 +971,7 @@ void qbman_eq_desc_set_token(struct qbman_eq_desc *d, uint8_t token); * @d: the enqueue descriptor * @fqid: the id of the frame queue to be enqueued. */ +__rte_internal void qbman_eq_desc_set_fq(struct qbman_eq_desc *d, uint32_t fqid); /** @@ -953,6 +981,7 @@ void qbman_eq_desc_set_fq(struct qbman_eq_desc *d, uint32_t fqid); * @qd_bin: the queuing destination bin * @qd_prio: the queuing destination priority. */ +__rte_internal void qbman_eq_desc_set_qd(struct qbman_eq_desc *d, uint32_t qdid, uint16_t qd_bin, uint8_t qd_prio); @@ -978,6 +1007,7 @@ void qbman_eq_desc_set_eqdi(struct qbman_eq_desc *d, int enable); * held-active (order-preserving) FQ, whether the FQ should be parked instead of * being rescheduled.) */ +__rte_internal void qbman_eq_desc_set_dca(struct qbman_eq_desc *d, int enable, uint8_t dqrr_idx, int park); @@ -987,6 +1017,7 @@ void qbman_eq_desc_set_dca(struct qbman_eq_desc *d, int enable, * * Return the fd pointer. */ +__rte_internal struct qbman_fd *qbman_result_eqresp_fd(struct qbman_result *eqresp); /** @@ -997,6 +1028,7 @@ struct qbman_fd *qbman_result_eqresp_fd(struct qbman_result *eqresp); * This value is set into the response id before the enqueue command, which, * get overwritten by qbman once the enqueue command is complete. */ +__rte_internal void qbman_result_eqresp_set_rspid(struct qbman_result *eqresp, uint8_t val); /** @@ -1009,6 +1041,7 @@ void qbman_result_eqresp_set_rspid(struct qbman_result *eqresp, uint8_t val); * copied into the enqueue response to determine if the command has been * completed, and response has been updated. */ +__rte_internal uint8_t qbman_result_eqresp_rspid(struct qbman_result *eqresp); /** @@ -1017,6 +1050,7 @@ uint8_t qbman_result_eqresp_rspid(struct qbman_result *eqresp); * * Return 0 when command is sucessful. */ +__rte_internal uint8_t qbman_result_eqresp_rc(struct qbman_result *eqresp); /** @@ -1043,6 +1077,7 @@ int qbman_swp_enqueue(struct qbman_swp *s, const struct qbman_eq_desc *d, * * Return the number of enqueued frames, -EBUSY if the EQCR is not ready. */ +__rte_internal int qbman_swp_enqueue_multiple(struct qbman_swp *s, const struct qbman_eq_desc *d, const struct qbman_fd *fd, @@ -1060,6 +1095,7 @@ int qbman_swp_enqueue_multiple(struct qbman_swp *s, * * Return the number of enqueued frames, -EBUSY if the EQCR is not ready. */ +__rte_internal int qbman_swp_enqueue_multiple_fd(struct qbman_swp *s, const struct qbman_eq_desc *d, struct qbman_fd **fd, @@ -1076,6 +1112,7 @@ int qbman_swp_enqueue_multiple_fd(struct qbman_swp *s, * * Return the number of enqueued frames, -EBUSY if the EQCR is not ready. */ +__rte_internal int qbman_swp_enqueue_multiple_desc(struct qbman_swp *s, const struct qbman_eq_desc *d, const struct qbman_fd *fd, @@ -1117,12 +1154,14 @@ struct qbman_release_desc { * default/starting state. * @d: the qbman release descriptor. */ +__rte_internal void qbman_release_desc_clear(struct qbman_release_desc *d); /** * qbman_release_desc_set_bpid() - Set the ID of the buffer pool to release to * @d: the qbman release descriptor. */ +__rte_internal void qbman_release_desc_set_bpid(struct qbman_release_desc *d, uint16_t bpid); /** @@ -1141,6 +1180,7 @@ void qbman_release_desc_set_rcdi(struct qbman_release_desc *d, int enable); * * Return 0 for success, -EBUSY if the release command ring is not ready. */ +__rte_internal int qbman_swp_release(struct qbman_swp *s, const struct qbman_release_desc *d, const uint64_t *buffers, unsigned int num_buffers); @@ -1166,6 +1206,7 @@ int qbman_swp_release_thresh(struct qbman_swp *s, unsigned int thresh); * Return 0 for success, or negative error code if the acquire command * fails. */ +__rte_internal int qbman_swp_acquire(struct qbman_swp *s, uint16_t bpid, uint64_t *buffers, unsigned int num_buffers); diff --git a/drivers/bus/fslmc/rte_bus_fslmc_version.map b/drivers/bus/fslmc/rte_bus_fslmc_version.map index fe45575046..0eb40d39eb 100644 --- a/drivers/bus/fslmc/rte_bus_fslmc_version.map +++ b/drivers/bus/fslmc/rte_bus_fslmc_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: dpaa2_affine_qbman_ethrx_swp; @@ -101,7 +101,6 @@ DPDK_20.0 { rte_fslmc_driver_unregister; rte_fslmc_get_device_count; rte_fslmc_object_register; - rte_fslmc_vfio_dmamap; rte_global_active_dqs_list; rte_mcp_ptr_list; diff --git a/drivers/bus/fslmc/rte_fslmc.h b/drivers/bus/fslmc/rte_fslmc.h index 96ba8dc259..5078b48ee1 100644 --- a/drivers/bus/fslmc/rte_fslmc.h +++ b/drivers/bus/fslmc/rte_fslmc.h @@ -162,6 +162,7 @@ RTE_DECLARE_PER_LCORE(struct dpaa2_portal_dqrr, dpaa2_held_bufs); * A pointer to a rte_dpaa2_driver structure describing the driver * to be registered. */ +__rte_internal void rte_fslmc_driver_register(struct rte_dpaa2_driver *driver); /** @@ -171,6 +172,7 @@ void rte_fslmc_driver_register(struct rte_dpaa2_driver *driver); * A pointer to a rte_dpaa2_driver structure describing the driver * to be unregistered. */ +__rte_internal void rte_fslmc_driver_unregister(struct rte_dpaa2_driver *driver); /** Helper for DPAA2 device registration from driver (eth, crypto) instance */ @@ -189,6 +191,7 @@ RTE_PMD_EXPORT_NAME(nm, __COUNTER__) * A pointer to a rte_dpaa_object structure describing the mc object * to be registered. */ +__rte_internal void rte_fslmc_object_register(struct rte_dpaa2_object *object); /** @@ -200,6 +203,7 @@ void rte_fslmc_object_register(struct rte_dpaa2_object *object); * >=0 for count; 0 indicates either no device of the said type scanned or * invalid device type. */ +__rte_internal uint32_t rte_fslmc_get_device_count(enum rte_dpaa2_dev_type device_type); /** Helper for DPAA2 object registration */ From patchwork Tue May 5 14:08:29 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186206 Delivered-To: patch@linaro.org Received: by 2002:a92:8d81:0:0:0:0:0 with SMTP id w1csp1174466ill; Tue, 5 May 2020 07:11:13 -0700 (PDT) X-Google-Smtp-Source: APiQypKiTOb5pFO7ij/Z3EU1t8+Kc19znk1Dhj3iq44NNDKZX1jTMWnMQOat2L/1xtrs283XCRRE X-Received: by 2002:a17:906:804a:: with SMTP id x10mr3042667ejw.86.1588687872934; Tue, 05 May 2020 07:11:12 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588687872; cv=none; d=google.com; s=arc-20160816; b=IlgU+BPiA12riiKAYsLcUq1hyv0gEVRlV8Qq6q2lRNNWQofALTRwTZ1nbRCKux9EcW vfTobjoYa3Fg8RZt6aGOYRl6QO6sFNycV6AlrKkSyhnlcvLQLJdLwINogkyV9ACG2amA iDXXLRXjj+irhergZ/SnRCIJG0z6i4O50NFXRYL5t9Rz5bZOC43PrgGasqLG2zcU3eBZ 60+wfyMOXD4b6Q0oA7Su7sF4z0AtGG7E/1/RM3nPR+2BnR7QI/lDZSQRl1gzRcTXCMTT em3M5RbD/k83YVSXYMPctgxCC3wHqf8G1G3DpyoSUDQriSo6ZFyn/cxHB+aoEANRgQcS mEaA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=nQnmjqVxhmeMU3CHpEVBxgkXns2m6LYHP8hN4vdCxBs=; b=UUmtqD0M9stpeQ5wx/xsZec2etT0lJl5LfQ5pLZ/XjEPFokUiMzInR9eOwir2dPpvg NeXSdMG3pUlARjXd5ArMRGXyBhikKWpgI48A93ys6DDNA/WidvHuQPebV+b7SVOlf4Zq qwDsHKSUqsEz+4x+dvpia43DedFORFF9XWUDeK1JKJTckuqGJpSFnvQNKBpJhIoiTQKs qYYD+ppH+nt7asjpAENCbfRmRczvsmqGVcJq/S2c4caH29clh1+xNzV5nJUhthJzozWS BbwM/mTKS8QmVpKwFhwICfUmBFTzY5U7a9kcNWiNncD0xIfAFB0k/IOTBk13O7iUc3pc OXZQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id q24si1143665ejb.184.2020.05.05.07.11.12; Tue, 05 May 2020 07:11:12 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 0BC2E1D5E5; Tue, 5 May 2020 16:10:45 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id E79611D156 for ; Tue, 5 May 2020 16:10:41 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id A2CA3200854; Tue, 5 May 2020 16:10:41 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 9CB1F200823; Tue, 5 May 2020 16:10:39 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id 24C8A402DB; Tue, 5 May 2020 22:10:37 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com Cc: Hemant Agrawal Date: Tue, 5 May 2020 19:38:29 +0530 Message-Id: <20200505140832.646-4-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200505140832.646-1-hemant.agrawal@nxp.com> References: <20200505140832.646-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH 4/7] bus/dpaa: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- drivers/bus/dpaa/include/fsl_bman.h | 6 +++++ drivers/bus/dpaa/include/fsl_fman.h | 27 +++++++++++++++++++ drivers/bus/dpaa/include/fsl_qman.h | 32 +++++++++++++++++++++++ drivers/bus/dpaa/include/fsl_usd.h | 6 +++++ drivers/bus/dpaa/include/netcfg.h | 2 ++ drivers/bus/dpaa/rte_bus_dpaa_version.map | 7 +---- drivers/bus/dpaa/rte_dpaa_bus.h | 5 ++++ 7 files changed, 79 insertions(+), 6 deletions(-) -- 2.17.1 diff --git a/drivers/bus/dpaa/include/fsl_bman.h b/drivers/bus/dpaa/include/fsl_bman.h index f9cd972153..82da2fcfe0 100644 --- a/drivers/bus/dpaa/include/fsl_bman.h +++ b/drivers/bus/dpaa/include/fsl_bman.h @@ -264,12 +264,14 @@ int bman_shutdown_pool(u32 bpid); * the structure provided by the caller can be released or reused after the * function returns. */ +__rte_internal struct bman_pool *bman_new_pool(const struct bman_pool_params *params); /** * bman_free_pool - Deallocates a Buffer Pool object * @pool: the pool object to release */ +__rte_internal void bman_free_pool(struct bman_pool *pool); /** @@ -279,6 +281,7 @@ void bman_free_pool(struct bman_pool *pool); * The returned pointer refers to state within the pool object so must not be * modified and can no longer be read once the pool object is destroyed. */ +__rte_internal const struct bman_pool_params *bman_get_params(const struct bman_pool *pool); /** @@ -289,6 +292,7 @@ const struct bman_pool_params *bman_get_params(const struct bman_pool *pool); * @flags: bit-mask of BMAN_RELEASE_FLAG_*** options * */ +__rte_internal int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num, u32 flags); @@ -302,6 +306,7 @@ int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num, * The return value will be the number of buffers obtained from the pool, or a * negative error code if a h/w error or pool starvation was encountered. */ +__rte_internal int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num, u32 flags); @@ -317,6 +322,7 @@ int bman_query_pools(struct bm_pool_state *state); * * Return the number of the free buffers */ +__rte_internal u32 bman_query_free_buffers(struct bman_pool *pool); /** diff --git a/drivers/bus/dpaa/include/fsl_fman.h b/drivers/bus/dpaa/include/fsl_fman.h index 5705ebfdce..6c87c8db0d 100644 --- a/drivers/bus/dpaa/include/fsl_fman.h +++ b/drivers/bus/dpaa/include/fsl_fman.h @@ -7,6 +7,8 @@ #ifndef __FSL_FMAN_H #define __FSL_FMAN_H +#include + #ifdef __cplusplus extern "C" { #endif @@ -43,18 +45,23 @@ struct fm_status_t { } __rte_packed; /* Set MAC address for a particular interface */ +__rte_internal int fman_if_add_mac_addr(struct fman_if *p, uint8_t *eth, uint8_t addr_num); /* Remove a MAC address for a particular interface */ +__rte_internal void fman_if_clear_mac_addr(struct fman_if *p, uint8_t addr_num); /* Get the FMAN statistics */ +__rte_internal void fman_if_stats_get(struct fman_if *p, struct rte_eth_stats *stats); /* Reset the FMAN statistics */ +__rte_internal void fman_if_stats_reset(struct fman_if *p); /* Get all of the FMAN statistics */ +__rte_internal void fman_if_stats_get_all(struct fman_if *p, uint64_t *value, int n); /* Set ignore pause option for a specific interface */ @@ -64,32 +71,43 @@ void fman_if_set_rx_ignore_pause_frames(struct fman_if *p, bool enable); void fman_if_conf_max_frame_len(struct fman_if *p, unsigned int max_frame_len); /* Enable/disable Rx promiscuous mode on specified interface */ +__rte_internal void fman_if_promiscuous_enable(struct fman_if *p); +__rte_internal void fman_if_promiscuous_disable(struct fman_if *p); /* Enable/disable Rx on specific interfaces */ +__rte_internal void fman_if_enable_rx(struct fman_if *p); +__rte_internal void fman_if_disable_rx(struct fman_if *p); /* Enable/disable loopback on specific interfaces */ +__rte_internal void fman_if_loopback_enable(struct fman_if *p); +__rte_internal void fman_if_loopback_disable(struct fman_if *p); /* Set buffer pool on specific interface */ +__rte_internal void fman_if_set_bp(struct fman_if *fm_if, unsigned int num, int bpid, size_t bufsize); /* Get Flow Control threshold parameters on specific interface */ +__rte_internal int fman_if_get_fc_threshold(struct fman_if *fm_if); /* Enable and Set Flow Control threshold parameters on specific interface */ +__rte_internal int fman_if_set_fc_threshold(struct fman_if *fm_if, u32 high_water, u32 low_water, u32 bpid); /* Get Flow Control pause quanta on specific interface */ +__rte_internal int fman_if_get_fc_quanta(struct fman_if *fm_if); /* Set Flow Control pause quanta on specific interface */ +__rte_internal int fman_if_set_fc_quanta(struct fman_if *fm_if, u16 pause_quanta); /* Set default error fqid on specific interface */ @@ -99,35 +117,44 @@ void fman_if_set_err_fqid(struct fman_if *fm_if, uint32_t err_fqid); int fman_if_get_ic_params(struct fman_if *fm_if, struct fman_if_ic_params *icp); /* Set IC transfer params */ +__rte_internal int fman_if_set_ic_params(struct fman_if *fm_if, const struct fman_if_ic_params *icp); /* Get interface fd->offset value */ +__rte_internal int fman_if_get_fdoff(struct fman_if *fm_if); /* Set interface fd->offset value */ +__rte_internal void fman_if_set_fdoff(struct fman_if *fm_if, uint32_t fd_offset); /* Get interface SG enable status value */ +__rte_internal int fman_if_get_sg_enable(struct fman_if *fm_if); /* Set interface SG support mode */ +__rte_internal void fman_if_set_sg(struct fman_if *fm_if, int enable); /* Get interface Max Frame length (MTU) */ uint16_t fman_if_get_maxfrm(struct fman_if *fm_if); /* Set interface Max Frame length (MTU) */ +__rte_internal void fman_if_set_maxfrm(struct fman_if *fm_if, uint16_t max_frm); /* Set interface next invoked action for dequeue operation */ void fman_if_set_dnia(struct fman_if *fm_if, uint32_t nia); /* discard error packets on rx */ +__rte_internal void fman_if_discard_rx_errors(struct fman_if *fm_if); +__rte_internal void fman_if_set_mcast_filter_table(struct fman_if *p); +__rte_internal void fman_if_reset_mcast_filter_table(struct fman_if *p); int fman_if_add_hash_mac_addr(struct fman_if *p, uint8_t *eth); diff --git a/drivers/bus/dpaa/include/fsl_qman.h b/drivers/bus/dpaa/include/fsl_qman.h index 1b3342e7e6..4411bb0a79 100644 --- a/drivers/bus/dpaa/include/fsl_qman.h +++ b/drivers/bus/dpaa/include/fsl_qman.h @@ -1314,6 +1314,7 @@ struct qman_cgr { #define QMAN_CGR_MODE_FRAME 0x00000001 #ifdef CONFIG_FSL_QMAN_FQ_LOOKUP +__rte_internal void qman_set_fq_lookup_table(void **table); #endif @@ -1322,6 +1323,7 @@ void qman_set_fq_lookup_table(void **table); */ int qman_get_portal_index(void); +__rte_internal u32 qman_portal_dequeue(struct rte_event ev[], unsigned int poll_limit, void **bufs); @@ -1333,6 +1335,7 @@ u32 qman_portal_dequeue(struct rte_event ev[], unsigned int poll_limit, * processed via qman_poll_***() functions). Returns zero for success, or * -EINVAL if the current CPU is sharing a portal hosted on another CPU. */ +__rte_internal int qman_irqsource_add(u32 bits); /** @@ -1340,6 +1343,7 @@ int qman_irqsource_add(u32 bits); * takes portal (fq specific) as input rather than using the thread affined * portal. */ +__rte_internal int qman_fq_portal_irqsource_add(struct qman_portal *p, u32 bits); /** @@ -1350,6 +1354,7 @@ int qman_fq_portal_irqsource_add(struct qman_portal *p, u32 bits); * instead be processed via qman_poll_***() functions. Returns zero for success, * or -EINVAL if the current CPU is sharing a portal hosted on another CPU. */ +__rte_internal int qman_irqsource_remove(u32 bits); /** @@ -1357,6 +1362,7 @@ int qman_irqsource_remove(u32 bits); * takes portal (fq specific) as input rather than using the thread affined * portal. */ +__rte_internal int qman_fq_portal_irqsource_remove(struct qman_portal *p, u32 bits); /** @@ -1369,6 +1375,7 @@ int qman_fq_portal_irqsource_remove(struct qman_portal *p, u32 bits); */ u16 qman_affine_channel(int cpu); +__rte_internal unsigned int qman_portal_poll_rx(unsigned int poll_limit, void **bufs, struct qman_portal *q); @@ -1380,6 +1387,7 @@ unsigned int qman_portal_poll_rx(unsigned int poll_limit, * * This function will issue a volatile dequeue command to the QMAN. */ +__rte_internal int qman_set_vdq(struct qman_fq *fq, u16 num, uint32_t vdqcr_flags); /** @@ -1390,6 +1398,7 @@ int qman_set_vdq(struct qman_fq *fq, u16 num, uint32_t vdqcr_flags); * is issued. It will keep returning NULL until there is no packet available on * the DQRR. */ +__rte_internal struct qm_dqrr_entry *qman_dequeue(struct qman_fq *fq); /** @@ -1401,6 +1410,7 @@ struct qm_dqrr_entry *qman_dequeue(struct qman_fq *fq); * This will consume the DQRR enrey and make it available for next volatile * dequeue. */ +__rte_internal void qman_dqrr_consume(struct qman_fq *fq, struct qm_dqrr_entry *dq); @@ -1414,6 +1424,7 @@ void qman_dqrr_consume(struct qman_fq *fq, * this function will return -EINVAL, otherwise the return value is >=0 and * represents the number of DQRR entries processed. */ +__rte_internal int qman_poll_dqrr(unsigned int limit); /** @@ -1460,6 +1471,7 @@ void qman_start_dequeues(void); * (SDQCR). The requested pools are limited to those the portal has dequeue * access to. */ +__rte_internal void qman_static_dequeue_add(u32 pools, struct qman_portal *qm); /** @@ -1507,6 +1519,7 @@ void qman_dca(const struct qm_dqrr_entry *dq, int park_request); * function must be called from the same CPU as that which processed the DQRR * entry in the first place. */ +__rte_internal void qman_dca_index(u8 index, int park_request); /** @@ -1564,6 +1577,7 @@ void qman_set_dc_ern(qman_cb_dc_ern handler, int affine); * a frame queue object based on that, rather than assuming/requiring that it be * Out of Service. */ +__rte_internal int qman_create_fq(u32 fqid, u32 flags, struct qman_fq *fq); /** @@ -1582,6 +1596,7 @@ void qman_destroy_fq(struct qman_fq *fq, u32 flags); * qman_fq_fqid - Queries the frame queue ID of a FQ object * @fq: the frame queue object to query */ +__rte_internal u32 qman_fq_fqid(struct qman_fq *fq); /** @@ -1594,6 +1609,7 @@ u32 qman_fq_fqid(struct qman_fq *fq); * This captures the state, as seen by the driver, at the time the function * executes. */ +__rte_internal void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags); /** @@ -1630,6 +1646,7 @@ void qman_fq_state(struct qman_fq *fq, enum qman_fq_state *state, u32 *flags); * context_a.address fields and will leave the stashing fields provided by the * user alone, otherwise it will zero out the context_a.stashing fields. */ +__rte_internal int qman_init_fq(struct qman_fq *fq, u32 flags, struct qm_mcc_initfq *opts); /** @@ -1659,6 +1676,7 @@ int qman_schedule_fq(struct qman_fq *fq); * caller should be prepared to accept the callback as the function is called, * not only once it has returned. */ +__rte_internal int qman_retire_fq(struct qman_fq *fq, u32 *flags); /** @@ -1668,6 +1686,7 @@ int qman_retire_fq(struct qman_fq *fq, u32 *flags); * The frame queue must be retired and empty, and if any order restoration list * was released as ERNs at the time of retirement, they must all be consumed. */ +__rte_internal int qman_oos_fq(struct qman_fq *fq); /** @@ -1701,6 +1720,7 @@ int qman_query_fq_has_pkts(struct qman_fq *fq); * @fq: the frame queue object to be queried * @np: storage for the queried FQD fields */ +__rte_internal int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np); /** @@ -1708,6 +1728,7 @@ int qman_query_fq_np(struct qman_fq *fq, struct qm_mcr_queryfq_np *np); * @fq: the frame queue object to be queried * @frm_cnt: number of frames in the queue */ +__rte_internal int qman_query_fq_frm_cnt(struct qman_fq *fq, u32 *frm_cnt); /** @@ -1738,6 +1759,7 @@ int qman_query_wq(u8 query_dedicated, struct qm_mcr_querywq *wq); * callback, or by waiting for the QMAN_FQ_STATE_VDQCR bit to disappear from the * "flags" retrieved from qman_fq_state(). */ +__rte_internal int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr); /** @@ -1773,8 +1795,10 @@ int qman_volatile_dequeue(struct qman_fq *fq, u32 flags, u32 vdqcr); * of an already busy hardware resource by throttling many of the to-be-dropped * enqueues "at the source". */ +__rte_internal int qman_enqueue(struct qman_fq *fq, const struct qm_fd *fd, u32 flags); +__rte_internal int qman_enqueue_multi(struct qman_fq *fq, const struct qm_fd *fd, u32 *flags, int frames_to_send); @@ -1788,6 +1812,7 @@ int qman_enqueue_multi(struct qman_fq *fq, const struct qm_fd *fd, u32 *flags, * This API is similar to qman_enqueue_multi(), but it takes fd which needs * to be processed by different frame queues. */ +__rte_internal int qman_enqueue_multi_fq(struct qman_fq *fq[], const struct qm_fd *fd, u32 *flags, int frames_to_send); @@ -1876,6 +1901,7 @@ int qman_shutdown_fq(u32 fqid); * @fqid: the base FQID of the range to deallocate * @count: the number of FQIDs in the range */ +__rte_internal int qman_reserve_fqid_range(u32 fqid, unsigned int count); static inline int qman_reserve_fqid(u32 fqid) { @@ -1895,6 +1921,7 @@ static inline int qman_reserve_fqid(u32 fqid) * than requested (though alignment will be as requested). If @partial is zero, * the return value will either be 'count' or negative. */ +__rte_internal int qman_alloc_pool_range(u32 *result, u32 count, u32 align, int partial); static inline int qman_alloc_pool(u32 *result) { @@ -1942,6 +1969,7 @@ void qman_seed_pool_range(u32 id, unsigned int count); * any unspecified parameters) will be used rather than a modify hw hardware * (which only modifies the specified parameters). */ +__rte_internal int qman_create_cgr(struct qman_cgr *cgr, u32 flags, struct qm_mcc_initcgr *opts); @@ -1964,6 +1992,7 @@ int qman_create_cgr_to_dcp(struct qman_cgr *cgr, u32 flags, u16 dcp_portal, * is executed. This must be excuted on the same affine portal on which it was * created. */ +__rte_internal int qman_delete_cgr(struct qman_cgr *cgr); /** @@ -1980,6 +2009,7 @@ int qman_delete_cgr(struct qman_cgr *cgr); * unspecified parameters) will be used rather than a modify hw hardware (which * only modifies the specified parameters). */ +__rte_internal int qman_modify_cgr(struct qman_cgr *cgr, u32 flags, struct qm_mcc_initcgr *opts); @@ -2008,6 +2038,7 @@ int qman_query_congestion(struct qm_mcr_querycongestion *congestion); * than requested (though alignment will be as requested). If @partial is zero, * the return value will either be 'count' or negative. */ +__rte_internal int qman_alloc_cgrid_range(u32 *result, u32 count, u32 align, int partial); static inline int qman_alloc_cgrid(u32 *result) { @@ -2021,6 +2052,7 @@ static inline int qman_alloc_cgrid(u32 *result) * @id: the base CGR ID of the range to deallocate * @count: the number of CGR IDs in the range */ +__rte_internal void qman_release_cgrid_range(u32 id, unsigned int count); static inline void qman_release_cgrid(u32 id) { diff --git a/drivers/bus/dpaa/include/fsl_usd.h b/drivers/bus/dpaa/include/fsl_usd.h index 263d9bb976..30ec63a09d 100644 --- a/drivers/bus/dpaa/include/fsl_usd.h +++ b/drivers/bus/dpaa/include/fsl_usd.h @@ -58,6 +58,7 @@ int bman_allocate_raw_portal(struct dpaa_raw_portal *portal); int bman_free_raw_portal(struct dpaa_raw_portal *portal); /* Obtain thread-local UIO file-descriptors */ +__rte_internal int qman_thread_fd(void); int bman_thread_fd(void); @@ -66,8 +67,12 @@ int bman_thread_fd(void); * processing is complete. As such, it is essential to call this before going * into another blocking read/select/poll. */ +__rte_internal void qman_thread_irq(void); + +__rte_internal void bman_thread_irq(void); +__rte_internal void qman_fq_portal_thread_irq(struct qman_portal *qp); void qman_clear_irq(void); @@ -77,6 +82,7 @@ int qman_global_init(void); int bman_global_init(void); /* Direct portal create and destroy */ +__rte_internal struct qman_portal *fsl_qman_fq_portal_create(int *fd); int fsl_qman_fq_portal_destroy(struct qman_portal *qp); int fsl_qman_fq_portal_init(struct qman_portal *qp); diff --git a/drivers/bus/dpaa/include/netcfg.h b/drivers/bus/dpaa/include/netcfg.h index bf7bfae8cb..d7d1befd24 100644 --- a/drivers/bus/dpaa/include/netcfg.h +++ b/drivers/bus/dpaa/include/netcfg.h @@ -46,11 +46,13 @@ struct netcfg_interface { * cfg_file: FMC config XML file * Returns the configuration information in newly allocated memory. */ +__rte_internal struct netcfg_info *netcfg_acquire(void); /* cfg_ptr: configuration information pointer. * Frees the resources allocated by the configuration layer. */ +__rte_internal void netcfg_release(struct netcfg_info *cfg_ptr); #ifdef RTE_LIBRTE_DPAA_DEBUG_DRIVER diff --git a/drivers/bus/dpaa/rte_bus_dpaa_version.map b/drivers/bus/dpaa/rte_bus_dpaa_version.map index e6ca4361e0..f4947fac41 100644 --- a/drivers/bus/dpaa/rte_bus_dpaa_version.map +++ b/drivers/bus/dpaa/rte_bus_dpaa_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: bman_acquire; @@ -13,7 +13,6 @@ DPDK_20.0 { dpaa_logtype_pmd; dpaa_netcfg; dpaa_svr_family; - fman_ccsr_map_fd; fman_dealloc_bufs_mask_hi; fman_dealloc_bufs_mask_lo; fman_if_add_mac_addr; @@ -51,7 +50,6 @@ DPDK_20.0 { qm_channel_pool1; qman_alloc_cgrid_range; qman_alloc_pool_range; - qman_clear_irq; qman_create_cgr; qman_create_fq; qman_dca_index; @@ -87,10 +85,7 @@ DPDK_20.0 { qman_volatile_dequeue; rte_dpaa_driver_register; rte_dpaa_driver_unregister; - rte_dpaa_mem_ptov; rte_dpaa_portal_fq_close; rte_dpaa_portal_fq_init; rte_dpaa_portal_init; - - local: *; }; diff --git a/drivers/bus/dpaa/rte_dpaa_bus.h b/drivers/bus/dpaa/rte_dpaa_bus.h index 373aca9785..d4aee132ef 100644 --- a/drivers/bus/dpaa/rte_dpaa_bus.h +++ b/drivers/bus/dpaa/rte_dpaa_bus.h @@ -158,6 +158,7 @@ rte_dpaa_mem_vtop(void *vaddr) * A pointer to a rte_dpaa_driver structure describing the driver * to be registered. */ +__rte_internal void rte_dpaa_driver_register(struct rte_dpaa_driver *driver); /** @@ -167,6 +168,7 @@ void rte_dpaa_driver_register(struct rte_dpaa_driver *driver); * A pointer to a rte_dpaa_driver structure describing the driver * to be unregistered. */ +__rte_internal void rte_dpaa_driver_unregister(struct rte_dpaa_driver *driver); /** @@ -178,10 +180,13 @@ void rte_dpaa_driver_unregister(struct rte_dpaa_driver *driver); * @return * 0 in case of success, error otherwise */ +__rte_internal int rte_dpaa_portal_init(void *arg); +__rte_internal int rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq); +__rte_internal int rte_dpaa_portal_fq_close(struct qman_fq *fq); /** From patchwork Tue May 5 14:08:30 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186207 Delivered-To: patch@linaro.org Received: by 2002:a92:8d81:0:0:0:0:0 with SMTP id w1csp1174608ill; Tue, 5 May 2020 07:11:21 -0700 (PDT) X-Google-Smtp-Source: APiQypJgsUaJnSv2MHiGhmsTTybe8qlUC+MZRIJ1EwhxuESO1QVCeDxLY2+HHXOj6wl2ZH2AuPTC X-Received: by 2002:aa7:d455:: with SMTP id q21mr2737205edr.235.1588687881609; Tue, 05 May 2020 07:11:21 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588687881; cv=none; d=google.com; s=arc-20160816; b=W9AdVPZCjNnDiEnh7jGyVR5YklumW4vRbB0uKjpE9pMsl5MndrwSM1IUSnByPdxZ2L UKzYz+AdRUdFGfpIswaDqEEKYSW59iIeDc9D18rV20ktTl/FjyyxdLLNCJPl71REkB7Q F29svsGdEuMLfLIsth7NQ06h3YWaFS55vTWNNmxZln58EW6AdT1Gd5DXyVKTRBq2Rxjr 7a6d1QFgnfByjY/Rz8I4ibz+1775bhCNID9gsqRPE5GfgFNWTYdw1CM3sB4J2sL9Jd5A HiNMfyQbvi2RfmsNJ+7IxaWo2Gs+yB2BcH/iYJ2SRAzeyZpU8tFOgpSGqQJksjkuGwFf oUcw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=W9fiL/Qqj+qNRpu1Sf45b0PzaStfE/lwa5kvaffXr5o=; b=IHs0owco3SSy8Wi2wUtMJYkdEA21YCv2+610EH9PsVc3i2S01TgqZX8uHJttmE+TsV xbZCakcMqDMzrqo0aYwjUv8NeNZcwOhwTt2dYgEim0c7Xp4gcAstYQf5Q4wu4jj4fyCV WWj8Qf5l/tPmkk3lDHnnmKwdUE9VW4NPj84N10Wpu8G2NGdBwnbw2zQNTZ9ZWDOObXBi 1aFf3SuwaGggcym4x7TKnLIv/fcg7BJj9jUGzEko/dM7CWaF2VMTZRQh3pXpkykKxkl9 9GXtPRZnRKFyU/YO5t7pPD7WYUWmBlOehzO+QzLGbSodPWSKvKEE/oNmZokX6IzdnUza JRTA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id e2si1463089edr.204.2020.05.05.07.11.21; Tue, 05 May 2020 07:11:21 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 42D9E1D5F4; Tue, 5 May 2020 16:10:46 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id 06E151D5A6 for ; Tue, 5 May 2020 16:10:42 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id DBCFA200848; Tue, 5 May 2020 16:10:41 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 3B2AA200846; Tue, 5 May 2020 16:10:40 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id B712040302; Tue, 5 May 2020 22:10:37 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com Cc: Hemant Agrawal Date: Tue, 5 May 2020 19:38:30 +0530 Message-Id: <20200505140832.646-5-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200505140832.646-1-hemant.agrawal@nxp.com> References: <20200505140832.646-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH 5/7] crypto: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- drivers/crypto/dpaa2_sec/dpaa2_sec_event.h | 5 +++-- drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map | 2 +- drivers/crypto/dpaa_sec/dpaa_sec_event.h | 8 ++++---- drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map | 4 +--- 4 files changed, 9 insertions(+), 10 deletions(-) -- 2.17.1 diff --git a/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h b/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h index c779d5d837..675cbbb81d 100644 --- a/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h +++ b/drivers/crypto/dpaa2_sec/dpaa2_sec_event.h @@ -6,12 +6,13 @@ #ifndef _DPAA2_SEC_EVENT_H_ #define _DPAA2_SEC_EVENT_H_ -int -dpaa2_sec_eventq_attach(const struct rte_cryptodev *dev, +__rte_internal +int dpaa2_sec_eventq_attach(const struct rte_cryptodev *dev, int qp_id, struct dpaa2_dpcon_dev *dpcon, const struct rte_event *event); +__rte_internal int dpaa2_sec_eventq_detach(const struct rte_cryptodev *dev, int qp_id); diff --git a/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map b/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map index 5952d645fd..1352f576e5 100644 --- a/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map +++ b/drivers/crypto/dpaa2_sec/rte_pmd_dpaa2_sec_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: dpaa2_sec_eventq_attach; diff --git a/drivers/crypto/dpaa_sec/dpaa_sec_event.h b/drivers/crypto/dpaa_sec/dpaa_sec_event.h index 8d1a018096..0b09fa8f75 100644 --- a/drivers/crypto/dpaa_sec/dpaa_sec_event.h +++ b/drivers/crypto/dpaa_sec/dpaa_sec_event.h @@ -6,14 +6,14 @@ #ifndef _DPAA_SEC_EVENT_H_ #define _DPAA_SEC_EVENT_H_ -int -dpaa_sec_eventq_attach(const struct rte_cryptodev *dev, +__rte_internal +int dpaa_sec_eventq_attach(const struct rte_cryptodev *dev, int qp_id, uint16_t ch_id, const struct rte_event *event); -int -dpaa_sec_eventq_detach(const struct rte_cryptodev *dev, +__rte_internal +int dpaa_sec_eventq_detach(const struct rte_cryptodev *dev, int qp_id); #endif /* _DPAA_SEC_EVENT_H_ */ diff --git a/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map b/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map index 8580fa13db..aed07fb371 100644 --- a/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map +++ b/drivers/crypto/dpaa_sec/rte_pmd_dpaa_sec_version.map @@ -1,8 +1,6 @@ -DPDK_20.0 { +INTERNAL { global: dpaa_sec_eventq_attach; dpaa_sec_eventq_detach; - - local: *; }; From patchwork Tue May 5 14:08:31 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186208 Delivered-To: patch@linaro.org Received: by 2002:a92:8d81:0:0:0:0:0 with SMTP id w1csp1174913ill; Tue, 5 May 2020 07:11:38 -0700 (PDT) X-Google-Smtp-Source: APiQypL7anEpLbfowAgEh/Nd+ZX9h+fIx2h8bYNz0UvCnaMw6zhdbRCyN0EJeB2WooXNvn/ufEug X-Received: by 2002:a17:906:4903:: with SMTP id b3mr2796798ejq.80.1588687898089; Tue, 05 May 2020 07:11:38 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588687898; cv=none; d=google.com; s=arc-20160816; b=l98FAlZ9un5IHcc3rogp8Z6vXE7KVlYe6iqLpVJjVaKVTsbRwZIj6D9RKFTrD3PL7K u0mhNYH+I8uoCyxK+sldyzpbJd+lDq2vg8eZhX4n4XuUIN116LohSWjwY5MWo4zw67PU HaT5wM/bI9e6kCIF5vO6cfzTKjeFobwqMGkF7AMKEwDbRTfhZov4EZvenntMpAf/jDRV y4PfmZC0lAxs6XTXABNsxeuujMSHX/ckSTu7H3PYuDAcOehqE2y18FVjxMrM+BA/BE4f w/mGnS7fPYiaE96Gc87++b0zLCozu1DXye5K9Ry61zBM5VCteaZ43dc8t8WXEYWRQHb1 EUCA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=w22UwGszSSkeWPTziVzrpcy2wg4fLAGXia+QK04doVQ=; b=MLuSHJItadUs+caNYitQDjCeeeyC1b2/oHFlYRmBBguR/awHw7/z3IdppherjaBDFp 48AjCnmqlKWhjPq23if9rJ/7ohfNHnn0O4WcZmvMfSwE+aNvt/YEEz64oVoxcHXb1/yB tq/j7KiqtqiDYWpjyzzwN1KNFum00DGYQe1hmc6fp5iNzl915GKP0uCwYI07HmlNayMA WFqL45MyRz5FiE3aNcrS0DmIXt92QOYUbUnJvhwPLxOHcfYXn/JENdMfcoRu21DDJ6tq PLe+IysGsrN2n2EaNExePVgrilSetYxujrChhzJSNKtp6+XpBq7y/Sqr3Ajuv58yT3hm zerw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id l5si1162342ejg.164.2020.05.05.07.11.37; Tue, 05 May 2020 07:11:38 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 850F61D601; Tue, 5 May 2020 16:10:47 +0200 (CEST) Received: from inva021.nxp.com (inva021.nxp.com [92.121.34.21]) by dpdk.org (Postfix) with ESMTP id 8D7561D5CB for ; Tue, 5 May 2020 16:10:43 +0200 (CEST) Received: from inva021.nxp.com (localhost [127.0.0.1]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id 71C53200846; Tue, 5 May 2020 16:10:43 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva021.eu-rdc02.nxp.com (Postfix) with ESMTP id C59E7200823; Tue, 5 May 2020 16:10:41 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id 556B64030B; Tue, 5 May 2020 22:10:38 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com Cc: Hemant Agrawal Date: Tue, 5 May 2020 19:38:31 +0530 Message-Id: <20200505140832.646-6-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200505140832.646-1-hemant.agrawal@nxp.com> References: <20200505140832.646-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH 6/7] net/dpaa2: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- drivers/net/dpaa2/dpaa2_ethdev.h | 2 ++ drivers/net/dpaa2/rte_pmd_dpaa2_version.map | 2 +- 2 files changed, 3 insertions(+), 1 deletion(-) -- 2.17.1 diff --git a/drivers/net/dpaa2/dpaa2_ethdev.h b/drivers/net/dpaa2/dpaa2_ethdev.h index 2c49a7f01f..c7fb6539ff 100644 --- a/drivers/net/dpaa2/dpaa2_ethdev.h +++ b/drivers/net/dpaa2/dpaa2_ethdev.h @@ -164,11 +164,13 @@ int dpaa2_remove_flow_dist(struct rte_eth_dev *eth_dev, int dpaa2_attach_bp_list(struct dpaa2_dev_priv *priv, void *blist); +__rte_internal int dpaa2_eth_eventq_attach(const struct rte_eth_dev *dev, int eth_rx_queue_id, struct dpaa2_dpcon_dev *dpcon, const struct rte_event_eth_rx_adapter_queue_conf *queue_conf); +__rte_internal int dpaa2_eth_eventq_detach(const struct rte_eth_dev *dev, int eth_rx_queue_id); diff --git a/drivers/net/dpaa2/rte_pmd_dpaa2_version.map b/drivers/net/dpaa2/rte_pmd_dpaa2_version.map index f2bb793319..78ea74e058 100644 --- a/drivers/net/dpaa2/rte_pmd_dpaa2_version.map +++ b/drivers/net/dpaa2/rte_pmd_dpaa2_version.map @@ -1,4 +1,4 @@ -DPDK_20.0 { +INTERNAL { global: dpaa2_eth_eventq_attach; From patchwork Tue May 5 14:08:32 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Hemant Agrawal X-Patchwork-Id: 186209 Delivered-To: patch@linaro.org Received: by 2002:a92:8d81:0:0:0:0:0 with SMTP id w1csp1175070ill; Tue, 5 May 2020 07:11:46 -0700 (PDT) X-Google-Smtp-Source: APiQypKclWRN6P5EwWSKJnFni55FDBeV5u9rOJc6ASmGzSy8NGBEBhW3ZWPQK6eceLChy/BKpDdw X-Received: by 2002:aa7:d6c7:: with SMTP id x7mr2780880edr.313.1588687906783; Tue, 05 May 2020 07:11:46 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1588687906; cv=none; d=google.com; s=arc-20160816; b=AtK+YGF573e0juWvRKmQjCVWOsox/lawi39RLH+ULuGDlMMscVj/74JEoC0j8Nu7e4 fMlu/cJR9zhuUbuecTFza4jo2GZuyb+hWB9IefaBFS1In6kFnR+IdpjMjQlbyHRtQXIZ c25kxCmPzZSnmtQA6eieyaoDy4CZqaIV1ZxgP01/YQG+Y97P7/gPY27ja6dnJALRQtul rJe7F7thsfuPWQFzKzOSuEAeCzIFfpF/45KFm9o0ncO3p2BlZa8SCXORITL+8HI46soX ZwHKospjZlbR7Qqc5jxPFqxSpgPDpHh3n5jjnUmyOITHYx8Cx6xERph29hZRbd1dWYEH MOIw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=sender:errors-to:list-subscribe:list-help:list-post:list-archive :list-unsubscribe:list-id:precedence:subject:references:in-reply-to :message-id:date:cc:to:from; bh=ZQozAEGeMirPdJGvkDUwmP/Rpu10vPRdSD9Gx75Dm7A=; b=QMnYfCBKIUSC0Kk+ce/xHIh/REQkfn98WktrNXWgi79TjLSgoi9jNFPfF0g90O1fCw B3mdGSGikj/4h5Ukii86vDoUkOdE29rVTcmOi5klp0PHpt2sgBgyoGW+CBG5vBOsIK0k j38IOlputVkaf+dnsmYIq+HtyXMh2nJxBfV3fEwsa8wPjNsEW8SV1KpR/UwUPyn5w9HM gfWKqT5l3Gu/bQqCe0LXj1tBQkVQgUM/90er+kEGbw1nanypv3b/08qvtWlLwjd1z77y Xiob8e390mmRnXLR2DFI64A/AeCtLN+xpi82iu5CZeZjfHtGer2+n1jE+Th6JEQBdEd5 B0JQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Return-Path: Received: from dpdk.org (dpdk.org. [92.243.14.124]) by mx.google.com with ESMTP id f25si1221669eja.442.2020.05.05.07.11.46; Tue, 05 May 2020 07:11:46 -0700 (PDT) Received-SPF: pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) client-ip=92.243.14.124; Authentication-Results: mx.google.com; spf=pass (google.com: domain of dev-bounces@dpdk.org designates 92.243.14.124 as permitted sender) smtp.mailfrom=dev-bounces@dpdk.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=nxp.com Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id CF20F1D5B5; Tue, 5 May 2020 16:10:48 +0200 (CEST) Received: from inva020.nxp.com (inva020.nxp.com [92.121.34.13]) by dpdk.org (Postfix) with ESMTP id AB0321D5CF for ; Tue, 5 May 2020 16:10:43 +0200 (CEST) Received: from inva020.nxp.com (localhost [127.0.0.1]) by inva020.eu-rdc02.nxp.com (Postfix) with ESMTP id 71E291A07B1; Tue, 5 May 2020 16:10:43 +0200 (CEST) Received: from invc005.ap-rdc01.nxp.com (invc005.ap-rdc01.nxp.com [165.114.16.14]) by inva020.eu-rdc02.nxp.com (Postfix) with ESMTP id C5A4B1A0764; Tue, 5 May 2020 16:10:41 +0200 (CEST) Received: from bf-netperf1.ap.freescale.net (bf-netperf1.ap.freescale.net [10.232.133.63]) by invc005.ap-rdc01.nxp.com (Postfix) with ESMTP id F0B264030E; Tue, 5 May 2020 22:10:38 +0800 (SGT) From: Hemant Agrawal To: dev@dpdk.org, david.marchand@redhat.com Cc: Hemant Agrawal Date: Tue, 5 May 2020 19:38:32 +0530 Message-Id: <20200505140832.646-7-hemant.agrawal@nxp.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200505140832.646-1-hemant.agrawal@nxp.com> References: <20200505140832.646-1-hemant.agrawal@nxp.com> X-Virus-Scanned: ClamAV using ClamSMTP Subject: [dpdk-dev] [PATCH 7/7] net/dpaa: move internal symbols into INTERNAL section X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch moves the internal symbols to INTERNAL sections so that any change in them is not reported as ABI breakage. Signed-off-by: Hemant Agrawal --- devtools/libabigail.abignore | 1 + drivers/net/dpaa/dpaa_ethdev.h | 2 ++ drivers/net/dpaa/rte_pmd_dpaa_version.map | 9 ++++++++- 3 files changed, 11 insertions(+), 1 deletion(-) -- 2.17.1 diff --git a/devtools/libabigail.abignore b/devtools/libabigail.abignore index 986a527715..0db7f5e0b9 100644 --- a/devtools/libabigail.abignore +++ b/devtools/libabigail.abignore @@ -29,3 +29,4 @@ type_kind = enum name = rte_eth_event_type changed_enumerators = RTE_ETH_EVENT_MAX + diff --git a/drivers/net/dpaa/dpaa_ethdev.h b/drivers/net/dpaa/dpaa_ethdev.h index af9fc2105d..7393a9df05 100644 --- a/drivers/net/dpaa/dpaa_ethdev.h +++ b/drivers/net/dpaa/dpaa_ethdev.h @@ -160,12 +160,14 @@ struct dpaa_if_stats { uint64_t tund; /**