// aria_simd.cpp - written and placed in the public domain by // Jeffrey Walton, Uri Blumenthal and Marcel Raad. // // This source file uses intrinsics to gain access to ARMv7a and // ARMv8a NEON instructions. A separate source file is needed // because additional CXXFLAGS are required to enable the // appropriate instructions sets in some build configurations. #include "pch.h" #include "config.h" #include "misc.h" #if (CRYPTOPP_SSSE3_AVAILABLE) # include #endif #if (CRYPTOPP_ARM_NEON_HEADER) # include #endif #if (CRYPTOPP_ARM_ACLE_HEADER) # include # include #endif // Squash MS LNK4221 and libtool warnings extern const char ARIA_SIMD_FNAME[] = __FILE__; NAMESPACE_BEGIN(CryptoPP) NAMESPACE_BEGIN(ARIATab) extern const word32 S1[256]; extern const word32 S2[256]; extern const word32 X1[256]; extern const word32 X2[256]; extern const word32 KRK[3][4]; NAMESPACE_END NAMESPACE_END ANONYMOUS_NAMESPACE_BEGIN using CryptoPP::byte; using CryptoPP::word32; inline byte ARIA_BRF(const word32 x, const int y) { return static_cast(GETBYTE(x, y)); } ANONYMOUS_NAMESPACE_END NAMESPACE_BEGIN(CryptoPP) using CryptoPP::ARIATab::S1; using CryptoPP::ARIATab::S2; using CryptoPP::ARIATab::X1; using CryptoPP::ARIATab::X2; using CryptoPP::ARIATab::KRK; #if (CRYPTOPP_ARM_NEON_AVAILABLE) template inline void ARIA_GSRK_NEON(const uint32x4_t X, const uint32x4_t Y, byte RK[16]) { enum { Q1 = (4-(N/32)) % 4, Q2 = (3-(N/32)) % 4, R = N % 32 }; vst1q_u8(RK, vreinterpretq_u8_u32( veorq_u32(X, veorq_u32( vshrq_n_u32(vextq_u32(Y, Y, Q1), R), vshlq_n_u32(vextq_u32(Y, Y, Q2), 32-R))))); } void ARIA_UncheckedSetKey_Schedule_NEON(byte* rk, word32* ws, unsigned int keylen) { const uint32x4_t w0 = vld1q_u32(ws+ 0); const uint32x4_t w1 = vld1q_u32(ws+ 8); const uint32x4_t w2 = vld1q_u32(ws+12); const uint32x4_t w3 = vld1q_u32(ws+16); ARIA_GSRK_NEON<19>(w0, w1, rk + 0); ARIA_GSRK_NEON<19>(w1, w2, rk + 16); ARIA_GSRK_NEON<19>(w2, w3, rk + 32); ARIA_GSRK_NEON<19>(w3, w0, rk + 48); ARIA_GSRK_NEON<31>(w0, w1, rk + 64); ARIA_GSRK_NEON<31>(w1, w2, rk + 80); ARIA_GSRK_NEON<31>(w2, w3, rk + 96); ARIA_GSRK_NEON<31>(w3, w0, rk + 112); ARIA_GSRK_NEON<67>(w0, w1, rk + 128); ARIA_GSRK_NEON<67>(w1, w2, rk + 144); ARIA_GSRK_NEON<67>(w2, w3, rk + 160); ARIA_GSRK_NEON<67>(w3, w0, rk + 176); ARIA_GSRK_NEON<97>(w0, w1, rk + 192); if (keylen > 16) { ARIA_GSRK_NEON<97>(w1, w2, rk + 208); ARIA_GSRK_NEON<97>(w2, w3, rk + 224); if (keylen > 24) { ARIA_GSRK_NEON< 97>(w3, w0, rk + 240); ARIA_GSRK_NEON<109>(w0, w1, rk + 256); } } } void ARIA_ProcessAndXorBlock_NEON(const byte* xorBlock, byte* outBlock, const byte *rk, word32 *t) { outBlock[ 0] = (byte)(X1[ARIA_BRF(t[0],3)] ); outBlock[ 1] = (byte)(X2[ARIA_BRF(t[0],2)]>>8); outBlock[ 2] = (byte)(S1[ARIA_BRF(t[0],1)] ); outBlock[ 3] = (byte)(S2[ARIA_BRF(t[0],0)] ); outBlock[ 4] = (byte)(X1[ARIA_BRF(t[1],3)] ); outBlock[ 5] = (byte)(X2[ARIA_BRF(t[1],2)]>>8); outBlock[ 6] = (byte)(S1[ARIA_BRF(t[1],1)] ); outBlock[ 7] = (byte)(S2[ARIA_BRF(t[1],0)] ); outBlock[ 8] = (byte)(X1[ARIA_BRF(t[2],3)] ); outBlock[ 9] = (byte)(X2[ARIA_BRF(t[2],2)]>>8); outBlock[10] = (byte)(S1[ARIA_BRF(t[2],1)] ); outBlock[11] = (byte)(S2[ARIA_BRF(t[2],0)] ); outBlock[12] = (byte)(X1[ARIA_BRF(t[3],3)] ); outBlock[13] = (byte)(X2[ARIA_BRF(t[3],2)]>>8); outBlock[14] = (byte)(S1[ARIA_BRF(t[3],1)] ); outBlock[15] = (byte)(S2[ARIA_BRF(t[3],0)] ); // 'outBlock' and 'xorBlock' may be unaligned. if (xorBlock != NULLPTR) { vst1q_u8(outBlock, veorq_u8( vld1q_u8(xorBlock), veorq_u8( vld1q_u8(outBlock), vrev32q_u8(vld1q_u8((rk)))))); } else { vst1q_u8(outBlock, veorq_u8( vld1q_u8(outBlock), vrev32q_u8(vld1q_u8(rk)))); } } #endif // CRYPTOPP_ARM_NEON_AVAILABLE #if (CRYPTOPP_SSSE3_AVAILABLE) void ARIA_ProcessAndXorBlock_SSSE3(const byte* xorBlock, byte* outBlock, const byte *rk, word32 *t) { const __m128i MASK = _mm_set_epi8(12,13,14,15, 8,9,10,11, 4,5,6,7, 0,1,2,3); outBlock[ 0] = (byte)(X1[ARIA_BRF(t[0],3)] ); outBlock[ 1] = (byte)(X2[ARIA_BRF(t[0],2)]>>8); outBlock[ 2] = (byte)(S1[ARIA_BRF(t[0],1)] ); outBlock[ 3] = (byte)(S2[ARIA_BRF(t[0],0)] ); outBlock[ 4] = (byte)(X1[ARIA_BRF(t[1],3)] ); outBlock[ 5] = (byte)(X2[ARIA_BRF(t[1],2)]>>8); outBlock[ 6] = (byte)(S1[ARIA_BRF(t[1],1)] ); outBlock[ 7] = (byte)(S2[ARIA_BRF(t[1],0)] ); outBlock[ 8] = (byte)(X1[ARIA_BRF(t[2],3)] ); outBlock[ 9] = (byte)(X2[ARIA_BRF(t[2],2)]>>8); outBlock[10] = (byte)(S1[ARIA_BRF(t[2],1)] ); outBlock[11] = (byte)(S2[ARIA_BRF(t[2],0)] ); outBlock[12] = (byte)(X1[ARIA_BRF(t[3],3)] ); outBlock[13] = (byte)(X2[ARIA_BRF(t[3],2)]>>8); outBlock[14] = (byte)(S1[ARIA_BRF(t[3],1)] ); outBlock[15] = (byte)(S2[ARIA_BRF(t[3],0)] ); // 'outBlock' and 'xorBlock' may be unaligned. if (xorBlock != NULLPTR) { _mm_storeu_si128(M128_CAST(outBlock), _mm_xor_si128( _mm_loadu_si128(CONST_M128_CAST(xorBlock)), _mm_xor_si128( _mm_loadu_si128(CONST_M128_CAST(outBlock)), _mm_shuffle_epi8(_mm_load_si128(CONST_M128_CAST(rk)), MASK))) ); } else { _mm_storeu_si128(M128_CAST(outBlock), _mm_xor_si128(_mm_loadu_si128(CONST_M128_CAST(outBlock)), _mm_shuffle_epi8(_mm_load_si128(CONST_M128_CAST(rk)), MASK))); } } #endif // CRYPTOPP_SSSE3_AVAILABLE NAMESPACE_END