From patchwork Sun Aug 11 22:59:12 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ard Biesheuvel X-Patchwork-Id: 171041 Delivered-To: patch@linaro.org Received: by 2002:a92:d204:0:0:0:0:0 with SMTP id y4csp2114662ily; Sun, 11 Aug 2019 15:59:33 -0700 (PDT) X-Google-Smtp-Source: APXvYqyTBm0LW/rVupjrRbNunHV4sE10ZvC14LpyLrl8Vsq2yyAh8Tl4L6mftzkwmbuhA37V8PXV X-Received: by 2002:a17:902:ba8e:: with SMTP id k14mr9338692pls.10.1565564373558; Sun, 11 Aug 2019 15:59:33 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1565564373; cv=none; d=google.com; s=arc-20160816; b=KnJ55nBRMdMjKrQmhjvwXZW42pO21mLJq0YUE6JYDkKpwv8XdoGKV71VNZI6fNza56 mKuwJAxYIcZZocEVujb/1hEQv1mE5ju1Xiemfgxv2T8szWhVgcCS8qe1sAmMDsWM2o74 OSpN/KIbc8mtSI7SH74W5AR1aVqg9GGQUW9phJSc6ZVNSkHPwdwubGMUW2CEV7uI7MOA mBDpmgF7aWOMcBCAGK1/LDu6HLq+bVHlqZFiUf6D26ueFZxlMWKV0h3wcWYwS+S8ENwH fU9AXPSx2iQiFymk+BV9ERDK3P4ue1w0g2m+ncgK008nhJ/8/r7dGZTIUsG3rnShm9Aw zmYA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from:dkim-signature; bh=AWsY2soOMxAD12H34LvTvFzff/WUbKiZ2Byzr98+9CE=; b=DoWMXjc0YzdjVagWNTF2sJg4dCSbXX8e2kzVv5/PaT/Qe1JmMEdoSJZbE3kBijpsCQ /jFAgxzgRQZQTg3Pb2iHUCuPlyu8lwIWY6j7k9ZZoWiSSSBeQD3enRDeh+kAjJtYXCIe 4CJQTRj2ieT87CTJnXBgaWR69E7g7GxwOU7ywiqIdcJgvFivcl6+aHVmoL+rxHv7ajVy IHc61KFnU7VPc4zQu4/e6kCtUEUqOIw5TY4wUbvP9ASm0RRRDy47d28i4f/SAV9NS8C9 EVx8kmv/xCqm/Hj+YHdnEmPmZNxt/tujS/dSIaMz1UNxm4y8ujQeDXuZqZn4twiS9I0A FJfQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b="wy/cf3sM"; spf=pass (google.com: best guess record for domain of linux-crypto-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-crypto-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id v189si60508878pgd.289.2019.08.11.15.59.33; Sun, 11 Aug 2019 15:59:33 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-crypto-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org header.s=google header.b="wy/cf3sM"; spf=pass (google.com: best guess record for domain of linux-crypto-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-crypto-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726144AbfHKW7c (ORCPT + 3 others); Sun, 11 Aug 2019 18:59:32 -0400 Received: from mail-wm1-f67.google.com ([209.85.128.67]:38097 "EHLO mail-wm1-f67.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726164AbfHKW7c (ORCPT ); Sun, 11 Aug 2019 18:59:32 -0400 Received: by mail-wm1-f67.google.com with SMTP id m125so6123494wmm.3 for ; Sun, 11 Aug 2019 15:59:30 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=AWsY2soOMxAD12H34LvTvFzff/WUbKiZ2Byzr98+9CE=; b=wy/cf3sMMgiSvzTXFMbmS7h0Mhvph+S/NHLixX1zQdYlr683eAAoVKqRmwqhsqquR9 12euTzNqgruvc8jguJyQStdac7TcEHZxSpfSRERPeUDdMGfRVhLdnaKYAyQzb3I0uaZY rNVTQgiQslhz9pmrfrRbKI9ToZk4stlk+NxEQCueNd9OY+B/dc7+8jGb406/AC8yeQK5 mY9+2IsXhzPeAqbO7EdLcGbCuQHvqn13LMzIvxuGGuAlT2Xv8pN6gPQKCk+ozzoHoQmU llw449Gne0ww0PTYYyltJ4nA6jdCvuLWKM9CfBdC+H3hIeM9jMzDnOlisIHnxZzzafXS ahaA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=AWsY2soOMxAD12H34LvTvFzff/WUbKiZ2Byzr98+9CE=; b=hWeXwMmNn2ksZP2NXPUhVz0RcCnyWxyJDUXMHtR++vQbgTD8WxD8I5CnhcutB/YhHx 4CtATJ1+RXtcHP1NOQFLmufhKvfRnN5OQ1LdMjtwUGD7fFtSC3CDzR4V2J0PmMwuehhE fkqmNX39J220JMeMPfEFkGckTlfzoS0b1A5jKFhuz28d10tQV5syeWFHhoj0NbCjgcdU CxODzAH6l7DroIohlfG2mMlO91hU5CU1MBHCUTZWJoUPIkTXuSIwTTdGzxj9d1gYr/nF NhR0vIQoNfsZXqvFSOLsxR9D7KM8vGjmdxE5RdgKcPnLTBnjuRxyuEnCYZwuzN+/q691 P4Dg== X-Gm-Message-State: APjAAAV/N8Jd9/QRz69LsPVOnebTJHg/YO7jYEzXj7CN5ViwR7gx69n+ HzhxCyO14vM780Hyk6JB2chciAAmj3tTMA== X-Received: by 2002:a1c:61d4:: with SMTP id v203mr23504532wmb.164.1565564369426; Sun, 11 Aug 2019 15:59:29 -0700 (PDT) Received: from localhost.localdomain ([2a02:587:a407:da00:5df1:4fcc:7bd1:4860]) by smtp.gmail.com with ESMTPSA id a17sm5930888wmm.47.2019.08.11.15.59.27 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Sun, 11 Aug 2019 15:59:28 -0700 (PDT) From: Ard Biesheuvel To: linux-crypto@vger.kernel.org Cc: herbert@gondor.apana.org.au, ebiggers@kernel.org, Ard Biesheuvel , Nick Desaulniers Subject: [PATCH v2 3/3] crypto: arm64/aegis128 - implement plain NEON version Date: Mon, 12 Aug 2019 01:59:12 +0300 Message-Id: <20190811225912.19412-4-ard.biesheuvel@linaro.org> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20190811225912.19412-1-ard.biesheuvel@linaro.org> References: <20190811225912.19412-1-ard.biesheuvel@linaro.org> Sender: linux-crypto-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-crypto@vger.kernel.org Provide a version of the core AES transform to the aegis128 SIMD code that does not rely on the special AES instructions, but uses plain NEON instructions instead. This allows the SIMD version of the aegis128 driver to be used on arm64 systems that do not implement those instructions (which are not mandatory in the architecture), such as the Raspberry Pi 3. Since GCC makes a mess of this when using the tbl/tbx intrinsics to perform the sbox substitution, preload the Sbox into v16..v31 in this case and use inline asm to emit the tbl/tbx instructions. Clang does not support this approach, nor does it require it, since it does a much better job at code generation, so there we use the intrinsics as usual. Cc: Nick Desaulniers Signed-off-by: Ard Biesheuvel --- crypto/Makefile | 9 ++- crypto/aegis128-neon-inner.c | 65 ++++++++++++++++++++ crypto/aegis128-neon.c | 8 ++- 3 files changed, 80 insertions(+), 2 deletions(-) -- 2.17.1 Acked-by: Nick Desaulniers diff --git a/crypto/Makefile b/crypto/Makefile index 99a9fa9087d1..0d2cdd523fd9 100644 --- a/crypto/Makefile +++ b/crypto/Makefile @@ -98,7 +98,14 @@ CFLAGS_aegis128-neon-inner.o += -mfpu=crypto-neon-fp-armv8 aegis128-$(CONFIG_CRYPTO_AEGIS128_SIMD) += aegis128-neon.o aegis128-neon-inner.o endif ifeq ($(ARCH),arm64) -CFLAGS_aegis128-neon-inner.o += -ffreestanding -mcpu=generic+crypto +aegis128-cflags-y := -ffreestanding -mcpu=generic+crypto +aegis128-cflags-$(CONFIG_CC_IS_GCC) += -ffixed-q16 -ffixed-q17 -ffixed-q18 \ + -ffixed-q19 -ffixed-q20 -ffixed-q21 \ + -ffixed-q22 -ffixed-q23 -ffixed-q24 \ + -ffixed-q25 -ffixed-q26 -ffixed-q27 \ + -ffixed-q28 -ffixed-q29 -ffixed-q30 \ + -ffixed-q31 +CFLAGS_aegis128-neon-inner.o += $(aegis128-cflags-y) CFLAGS_REMOVE_aegis128-neon-inner.o += -mgeneral-regs-only aegis128-$(CONFIG_CRYPTO_AEGIS128_SIMD) += aegis128-neon.o aegis128-neon-inner.o endif diff --git a/crypto/aegis128-neon-inner.c b/crypto/aegis128-neon-inner.c index 3d8043c4832b..ed55568afd1b 100644 --- a/crypto/aegis128-neon-inner.c +++ b/crypto/aegis128-neon-inner.c @@ -17,6 +17,8 @@ #include +extern int aegis128_have_aes_insn; + void *memcpy(void *dest, const void *src, size_t n); void *memset(void *s, int c, size_t n); @@ -24,6 +26,8 @@ struct aegis128_state { uint8x16_t v[5]; }; +extern const uint8x16x4_t crypto_aes_sbox[]; + static struct aegis128_state aegis128_load_state_neon(const void *state) { return (struct aegis128_state){ { @@ -49,6 +53,46 @@ uint8x16_t aegis_aes_round(uint8x16_t w) { uint8x16_t z = {}; +#ifdef CONFIG_ARM64 + if (!__builtin_expect(aegis128_have_aes_insn, 1)) { + static const uint8x16_t shift_rows = { + 0x0, 0x5, 0xa, 0xf, 0x4, 0x9, 0xe, 0x3, + 0x8, 0xd, 0x2, 0x7, 0xc, 0x1, 0x6, 0xb, + }; + static const uint8x16_t ror32by8 = { + 0x1, 0x2, 0x3, 0x0, 0x5, 0x6, 0x7, 0x4, + 0x9, 0xa, 0xb, 0x8, 0xd, 0xe, 0xf, 0xc, + }; + uint8x16_t v; + + // shift rows + w = vqtbl1q_u8(w, shift_rows); + + // sub bytes + if (!IS_ENABLED(CONFIG_CC_IS_GCC)) { + v = vqtbl4q_u8(crypto_aes_sbox[0], w); + v = vqtbx4q_u8(v, crypto_aes_sbox[1], w - 0x40); + v = vqtbx4q_u8(v, crypto_aes_sbox[2], w - 0x80); + v = vqtbx4q_u8(v, crypto_aes_sbox[3], w - 0xc0); + } else { + asm("tbl %0.16b, {v16.16b-v19.16b}, %1.16b" : "=w"(v) : "w"(w)); + w -= 0x40; + asm("tbx %0.16b, {v20.16b-v23.16b}, %1.16b" : "+w"(v) : "w"(w)); + w -= 0x40; + asm("tbx %0.16b, {v24.16b-v27.16b}, %1.16b" : "+w"(v) : "w"(w)); + w -= 0x40; + asm("tbx %0.16b, {v28.16b-v31.16b}, %1.16b" : "+w"(v) : "w"(w)); + } + + // mix columns + w = (v << 1) ^ (uint8x16_t)(((int8x16_t)v >> 7) & 0x1b); + w ^= (uint8x16_t)vrev32q_u16((uint16x8_t)v); + w ^= vqtbl1q_u8(v ^ w, ror32by8); + + return w; + } +#endif + /* * We use inline asm here instead of the vaeseq_u8/vaesmcq_u8 intrinsics * to force the compiler to issue the aese/aesmc instructions in pairs. @@ -73,10 +117,27 @@ struct aegis128_state aegis128_update_neon(struct aegis128_state st, return st; } +static inline __attribute__((always_inline)) +void preload_sbox(void) +{ + if (!IS_ENABLED(CONFIG_ARM64) || + !IS_ENABLED(CONFIG_CC_IS_GCC) || + __builtin_expect(aegis128_have_aes_insn, 1)) + return; + + asm("ld1 {v16.16b-v19.16b}, [%0], #64 \n\t" + "ld1 {v20.16b-v23.16b}, [%0], #64 \n\t" + "ld1 {v24.16b-v27.16b}, [%0], #64 \n\t" + "ld1 {v28.16b-v31.16b}, [%0] \n\t" + :: "r"(crypto_aes_sbox)); +} + void crypto_aegis128_update_neon(void *state, const void *msg) { struct aegis128_state st = aegis128_load_state_neon(state); + preload_sbox(); + st = aegis128_update_neon(st, vld1q_u8(msg)); aegis128_save_state_neon(st, state); @@ -88,6 +149,8 @@ void crypto_aegis128_encrypt_chunk_neon(void *state, void *dst, const void *src, struct aegis128_state st = aegis128_load_state_neon(state); uint8x16_t msg; + preload_sbox(); + while (size >= AEGIS_BLOCK_SIZE) { uint8x16_t s = st.v[1] ^ (st.v[2] & st.v[3]) ^ st.v[4]; @@ -120,6 +183,8 @@ void crypto_aegis128_decrypt_chunk_neon(void *state, void *dst, const void *src, struct aegis128_state st = aegis128_load_state_neon(state); uint8x16_t msg; + preload_sbox(); + while (size >= AEGIS_BLOCK_SIZE) { msg = vld1q_u8(src) ^ st.v[1] ^ (st.v[2] & st.v[3]) ^ st.v[4]; st = aegis128_update_neon(st, msg); diff --git a/crypto/aegis128-neon.c b/crypto/aegis128-neon.c index c1c0a1686f67..751f9c195aa4 100644 --- a/crypto/aegis128-neon.c +++ b/crypto/aegis128-neon.c @@ -14,9 +14,15 @@ void crypto_aegis128_encrypt_chunk_neon(void *state, void *dst, const void *src, void crypto_aegis128_decrypt_chunk_neon(void *state, void *dst, const void *src, unsigned int size); +int aegis128_have_aes_insn __ro_after_init; + bool crypto_aegis128_have_simd(void) { - return cpu_have_feature(cpu_feature(AES)); + if (cpu_have_feature(cpu_feature(AES))) { + aegis128_have_aes_insn = 1; + return true; + } + return IS_ENABLED(CONFIG_ARM64); } void crypto_aegis128_update_simd(union aegis_block *state, const void *msg)