From patchwork Wed Mar 4 11:23:29 2015 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Maxim Uvarov X-Patchwork-Id: 45407 Return-Path: X-Original-To: linaro@patches.linaro.org Delivered-To: linaro@patches.linaro.org Received: from mail-la0-f72.google.com (mail-la0-f72.google.com [209.85.215.72]) by ip-10-151-82-157.ec2.internal (Postfix) with ESMTPS id CFCEB21486 for ; Wed, 4 Mar 2015 11:23:54 +0000 (UTC) Received: by labmn12 with SMTP id mn12sf5578080lab.2 for ; Wed, 04 Mar 2015 03:23:53 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:delivered-to:from:to:date:message-id:subject :precedence:list-id:list-unsubscribe:list-archive:list-post :list-help:list-subscribe:mime-version:content-type :content-transfer-encoding:errors-to:sender:x-original-sender :x-original-authentication-results:mailing-list; bh=keRuJ49Lgx2r2XYxqSC8HAESxCO0TfEomDfSONWY8dY=; b=di1nWwYHzBH9UMjQzT926n8q+F9N2r+fgeXRd5Ony+5+wuMEUnu8rvAWlJSCFjvzFa NryVO9+5OsdYJc5pKaFDM0BAbxAymVZrJFMTi0fxswFY3/dA2T/Ho6zyw5GMTdrulyzW 5u5e8n9KQodhne3QgiSX5HGUvcXcSTxDoqdXunpERs1R6zMkVWyrWMtrh7TriXsgm7W4 gIxpY1ssTxFaGW3cNkKZspiPZgX7KvGu5UL4OfI/rmc/MXo7orLqnZRgZ85hcxEP9Xdg 7zRMD7znVW+YJS+dM8UZbk+z3cB7FzbdpG6Bevfs/4pGGcY1Fk5wGYE9j9KV7R7/mpPZ 70kQ== X-Gm-Message-State: ALoCoQntuLysE4cwhZHqJ/LXMb58lYWrgoSP+eHMYa2rF7HjAeLZJJuM+ZaIyVkmDVJh4QuTd62m X-Received: by 10.112.130.70 with SMTP id oc6mr594826lbb.13.1425468233857; Wed, 04 Mar 2015 03:23:53 -0800 (PST) X-BeenThere: patchwork-forward@linaro.org Received: by 10.152.115.140 with SMTP id jo12ls169510lab.19.gmail; Wed, 04 Mar 2015 03:23:53 -0800 (PST) X-Received: by 10.153.8.135 with SMTP id dk7mr2758333lad.93.1425468233709; Wed, 04 Mar 2015 03:23:53 -0800 (PST) Received: from mail-la0-f54.google.com (mail-la0-f54.google.com. [209.85.215.54]) by mx.google.com with ESMTPS id f11si2329891lbm.55.2015.03.04.03.23.53 for (version=TLSv1.2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Wed, 04 Mar 2015 03:23:53 -0800 (PST) Received-SPF: pass (google.com: domain of patch+caf_=patchwork-forward=linaro.org@linaro.org designates 209.85.215.54 as permitted sender) client-ip=209.85.215.54; Received: by labgf13 with SMTP id gf13so19189374lab.10 for ; Wed, 04 Mar 2015 03:23:53 -0800 (PST) X-Received: by 10.152.115.136 with SMTP id jo8mr3111424lab.32.1425468233609; Wed, 04 Mar 2015 03:23:53 -0800 (PST) X-Forwarded-To: patchwork-forward@linaro.org X-Forwarded-For: patch@linaro.org patchwork-forward@linaro.org Delivered-To: patch@linaro.org Received: by 10.112.35.133 with SMTP id h5csp1154531lbj; Wed, 4 Mar 2015 03:23:52 -0800 (PST) X-Received: by 10.229.207.198 with SMTP id fz6mr5113469qcb.27.1425468232136; Wed, 04 Mar 2015 03:23:52 -0800 (PST) Received: from ip-10-35-177-41.ec2.internal (lists.linaro.org. [54.225.227.206]) by mx.google.com with ESMTPS id b200si3167230qhc.93.2015.03.04.03.23.50 (version=TLSv1 cipher=RC4-SHA bits=128/128); Wed, 04 Mar 2015 03:23:52 -0800 (PST) Received-SPF: none (google.com: lng-odp-bounces@lists.linaro.org does not designate permitted sender hosts) client-ip=54.225.227.206; Received: from localhost ([127.0.0.1] helo=ip-10-35-177-41.ec2.internal) by ip-10-35-177-41.ec2.internal with esmtp (Exim 4.76) (envelope-from ) id 1YT7P2-0006Vm-Sc; Wed, 04 Mar 2015 11:23:48 +0000 Received: from mail-lb0-f175.google.com ([209.85.217.175]) by ip-10-35-177-41.ec2.internal with esmtp (Exim 4.76) (envelope-from ) id 1YT7Ow-0006Vh-Dv for lng-odp@lists.linaro.org; Wed, 04 Mar 2015 11:23:42 +0000 Received: by lbjf15 with SMTP id f15so15481662lbj.2 for ; Wed, 04 Mar 2015 03:23:36 -0800 (PST) X-Received: by 10.112.51.114 with SMTP id j18mr2729186lbo.97.1425468216434; Wed, 04 Mar 2015 03:23:36 -0800 (PST) Received: from maxim-Aspire-VN7-791.Dlink (broadband-46-188-125-102.2com.net. [46.188.125.102]) by mx.google.com with ESMTPSA id xu8sm709007lab.21.2015.03.04.03.23.34 (version=TLSv1.2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Wed, 04 Mar 2015 03:23:35 -0800 (PST) From: Maxim Uvarov To: lng-odp@lists.linaro.org Date: Wed, 4 Mar 2015 14:23:29 +0300 Message-Id: <1425468209-22480-1-git-send-email-maxim.uvarov@linaro.org> X-Mailer: git-send-email 1.9.1 X-Topics: patch Subject: [lng-odp] [PATCHv5] linux-generic: mmap: jumbo frames support X-BeenThere: lng-odp@lists.linaro.org X-Mailman-Version: 2.1.14 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: , List-Help: , List-Subscribe: , MIME-Version: 1.0 Errors-To: lng-odp-bounces@lists.linaro.org Sender: lng-odp-bounces@lists.linaro.org X-Removed-Original-Auth: Dkim didn't pass. X-Original-Sender: maxim.uvarov@linaro.org X-Original-Authentication-Results: mx.google.com; spf=pass (google.com: domain of patch+caf_=patchwork-forward=linaro.org@linaro.org designates 209.85.215.54 as permitted sender) smtp.mail=patch+caf_=patchwork-forward=linaro.org@linaro.org Mailing-list: list patchwork-forward@linaro.org; contact patchwork-forward+owners@linaro.org X-Google-Group-Id: 836684582541 Support for jumbo frames for linux-generic with unsegmented buffers. Test for pkio is also adjusted to work with 9*1024=9216 bytes packets. https://bugs.linaro.org/show_bug.cgi?id=509 Signed-off-by: Maxim Uvarov --- v5: - test_4_jumbo_pkts -> test_jumbo - do not use stack for jumbo packet, simple allocate it. v4: - fix work on real interfaces (make check under root) - better define jumbo packet payload size platform/linux-generic/odp_packet_socket.c | 2 +- test/validation/odp_pktio.c | 95 +++++++++++++++++++++++------- test/validation/odp_pktio_run | 4 +- 3 files changed, 77 insertions(+), 24 deletions(-) diff --git a/platform/linux-generic/odp_packet_socket.c b/platform/linux-generic/odp_packet_socket.c index 55c212e..4dcb111 100644 --- a/platform/linux-generic/odp_packet_socket.c +++ b/platform/linux-generic/odp_packet_socket.c @@ -587,7 +587,7 @@ static inline unsigned pkt_mmap_v2_tx(int sock, struct ring *ring, static void mmap_fill_ring(struct ring *ring, unsigned blocks) { ring->req.tp_block_size = getpagesize() << 2; - ring->req.tp_frame_size = TPACKET_ALIGNMENT << 7; + ring->req.tp_frame_size = 9216 + TPACKET_HDRLEN; ring->req.tp_block_nr = blocks; ring->req.tp_frame_nr = ring->req.tp_block_size / diff --git a/test/validation/odp_pktio.c b/test/validation/odp_pktio.c index 8df367d..f8b9ecc 100644 --- a/test/validation/odp_pktio.c +++ b/test/validation/odp_pktio.c @@ -15,6 +15,10 @@ #define PKT_BUF_NUM 32 #define PKT_BUF_SIZE 1856 +#define PKT_BUF_JUMBO_SIZE 9216 +#define PKT_BUF_JUMBO_MAX_PAYLOAD (PKT_BUF_JUMBO_SIZE -\ + (ODPH_UDPHDR_LEN +\ + ODPH_IPV4HDR_LEN + ODPH_ETHHDR_LEN)) #define MAX_NUM_IFACES 2 #define TEST_SEQ_INVALID ((uint32_t)~0) #define TEST_SEQ_MAGIC 0x92749451 @@ -33,12 +37,21 @@ typedef struct { odp_queue_t inq; } pktio_info_t; -/** structure of test packet UDP payload */ -typedef struct { +typedef struct ODP_PACKED { uint32be_t magic; uint32be_t seq; +} pkt_head_t; + +/** structure of test packet UDP payload */ +typedef struct ODP_PACKED { + pkt_head_t head; + char data[PKT_BUF_JUMBO_MAX_PAYLOAD - sizeof(pkt_head_t) - + sizeof(uint32be_t)]; + uint32be_t magic2; } pkt_test_data_t; +static int test_jumbo; + /** default packet pool */ odp_pool_t default_pkt_pool = ODP_POOL_INVALID; @@ -59,14 +72,18 @@ static void pktio_pkt_set_macs(odp_packet_t pkt, CU_ASSERT(ret == ODPH_ETHADDR_LEN); } +static uint32_t pkt_payload_len(void) +{ + return test_jumbo ? sizeof(pkt_test_data_t) : sizeof(pkt_head_t); +} + static int pktio_pkt_set_seq(odp_packet_t pkt) { static uint32_t tstseq; size_t l4_off; - pkt_test_data_t data; + pkt_test_data_t *data; + uint32_t len = pkt_payload_len(); - data.magic = TEST_SEQ_MAGIC; - data.seq = tstseq; l4_off = odp_packet_l4_offset(pkt); if (!l4_off) { @@ -74,9 +91,16 @@ static int pktio_pkt_set_seq(odp_packet_t pkt) return -1; } + data = calloc(1, len); + CU_ASSERT_FATAL(data != NULL); + + data->head.magic = TEST_SEQ_MAGIC; + data->magic2 = TEST_SEQ_MAGIC; + data->head.seq = tstseq; + odp_packet_copydata_in(pkt, l4_off+ODPH_UDPHDR_LEN, - sizeof(data), &data); - + len, data); + free(data); tstseq++; return 0; @@ -85,18 +109,30 @@ static int pktio_pkt_set_seq(odp_packet_t pkt) static uint32_t pktio_pkt_seq(odp_packet_t pkt) { size_t l4_off; - pkt_test_data_t data; + uint32_t seq = TEST_SEQ_INVALID; + pkt_test_data_t *data; + uint32_t len = pkt_payload_len(); l4_off = odp_packet_l4_offset(pkt); - if (l4_off) { - odp_packet_copydata_out(pkt, l4_off+ODPH_UDPHDR_LEN, - sizeof(data), &data); + if (!l4_off) + return TEST_SEQ_INVALID; - if (data.magic == TEST_SEQ_MAGIC) - return data.seq; + data = calloc(1, len); + CU_ASSERT_FATAL(data != NULL); + + odp_packet_copydata_out(pkt, l4_off+ODPH_UDPHDR_LEN, + len, data); + + if (data->head.magic == TEST_SEQ_MAGIC) { + if (test_jumbo && data->magic2 != TEST_SEQ_MAGIC) { + free(data); + return TEST_SEQ_INVALID; + } + seq = data->head.seq; } - return TEST_SEQ_INVALID; + free(data); + return seq; } static odp_packet_t pktio_create_packet(void) @@ -107,7 +143,7 @@ static odp_packet_t pktio_create_packet(void) odph_udphdr_t *udp; char *buf; uint16_t seq; - size_t payload_len = sizeof(pkt_test_data_t); + size_t payload_len = pkt_payload_len(); uint8_t mac[ODPH_ETHADDR_LEN] = {0}; pkt = odp_packet_alloc(default_pkt_pool, payload_len + ODPH_UDPHDR_LEN + @@ -187,8 +223,8 @@ static int default_pool_create(void) return -1; memset(¶ms, 0, sizeof(params)); - params.pkt.seg_len = PKT_BUF_SIZE; - params.pkt.len = PKT_BUF_SIZE; + params.pkt.seg_len = PKT_BUF_JUMBO_SIZE; + params.pkt.len = PKT_BUF_JUMBO_SIZE; params.pkt.num = PKT_BUF_NUM; params.type = ODP_POOL_PACKET; @@ -208,15 +244,24 @@ static odp_pktio_t create_pktio(const char *iface) odp_pool_param_t params; memset(¶ms, 0, sizeof(params)); - params.pkt.seg_len = PKT_BUF_SIZE; - params.pkt.len = PKT_BUF_SIZE; + if (test_jumbo) { + params.pkt.seg_len = PKT_BUF_JUMBO_SIZE; + params.pkt.len = PKT_BUF_JUMBO_SIZE; + + } else { + params.pkt.seg_len = PKT_BUF_SIZE; + params.pkt.len = PKT_BUF_SIZE; + } params.pkt.num = PKT_BUF_NUM; params.type = ODP_POOL_PACKET; snprintf(pool_name, sizeof(pool_name), "pkt_pool_%s", iface); + pool = odp_pool_lookup(pool_name); - if (pool == ODP_POOL_INVALID) - pool = odp_pool_create(pool_name, ODP_SHM_NULL, ¶ms); + if (pool != ODP_POOL_INVALID) + odp_pool_destroy(pool); + + pool = odp_pool_create(pool_name, ODP_SHM_NULL, ¶ms); CU_ASSERT(pool != ODP_POOL_INVALID); pktio = odp_pktio_open(iface, pool); @@ -450,6 +495,13 @@ static void test_odp_pktio_sched_multi(void) pktio_test_txrx(ODP_QUEUE_TYPE_SCHED, 4); } +static void test_odp_pktio_jumbo(void) +{ + test_jumbo = 1; + test_odp_pktio_sched_multi(); + test_jumbo = 0; +} + static void test_odp_pktio_mtu(void) { int ret; @@ -668,6 +720,7 @@ CU_TestInfo pktio_tests[] = { {"pktio poll multi", test_odp_pktio_poll_multi}, {"pktio sched queues", test_odp_pktio_sched_queue}, {"pktio sched multi", test_odp_pktio_sched_multi}, + {"pktio jumbo frames", test_odp_pktio_jumbo}, {"pktio mtu", test_odp_pktio_mtu}, {"pktio promisc mode", test_odp_pktio_promisc}, {"pktio mac", test_odp_pktio_mac}, diff --git a/test/validation/odp_pktio_run b/test/validation/odp_pktio_run index 08288e6..b9d7e3c 100755 --- a/test/validation/odp_pktio_run +++ b/test/validation/odp_pktio_run @@ -56,8 +56,8 @@ setup_env1() echo "pktio: error: unable to create veth pair" exit $TEST_SKIPPED fi - ip link set $IF0 up - ip link set $IF1 up + ip link set $IF0 mtu 9216 up + ip link set $IF1 mtu 9216 up # network needs a little time to come up sleep 1