diff options
-rw-r--r-- | cpu.h | 2 | ||||
-rw-r--r-- | gcm.cpp | 16 |
2 files changed, 9 insertions, 9 deletions
@@ -17,7 +17,7 @@ # pragma GCC diagnostic ignored "-Wsign-conversion"
#endif
-// ARM32/ARM64 Headers
+// ARM32 and ARM64 Headers
#if (CRYPTOPP_BOOL_ARM32 || CRYPTOPP_BOOL_ARM64)
# if defined(__GNUC__)
# include <stdint.h>
@@ -87,7 +87,7 @@ inline uint64x2_t PMULL_11(const uint64x2_t a, const uint64x2_t b) return r;
}
-inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b, unsigned int c)
+inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b, unsigned int c)
{
uint64x2_t r;
__asm __volatile("ext %0.16b, %1.16b, %2.16b, %3 \n\t"
@@ -97,7 +97,7 @@ inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b, unsigned int c) // https://github.com/weidai11/cryptopp/issues/366
template <unsigned int C>
-inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b)
+inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b)
{
uint64x2_t r;
__asm __volatile("ext %0.16b, %1.16b, %2.16b, %3 \n\t"
@@ -131,14 +131,14 @@ inline uint64x2_t PMULL_11(const uint64x2_t a, const uint64x2_t b) vgetq_lane_u64(vreinterpretq_u64_u8(b),1)));
}
-inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b, unsigned int c)
+inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b, unsigned int c)
{
return (uint64x2_t)vextq_u8(vreinterpretq_u8_u64(a), vreinterpretq_u8_u64(b), c);
}
// https://github.com/weidai11/cryptopp/issues/366
template <unsigned int C>
-inline uint64x2_t VEXT_8(uint64x2_t a, uint64x2_t b)
+inline uint64x2_t VEXT_U8(uint64x2_t a, uint64x2_t b)
{
return (uint64x2_t)vextq_u8(vreinterpretq_u8_u64(a), vreinterpretq_u8_u64(b), C);
}
@@ -297,13 +297,13 @@ static const unsigned int s_clmulTableSizeInBlocks = 8; inline uint64x2_t PMULL_Reduce(uint64x2_t c0, uint64x2_t c1, uint64x2_t c2, const uint64x2_t &r)
{
// See comments fo CLMUL_Reduce
- c1 = veorq_u64(c1, VEXT_8<8>(vdupq_n_u64(0), c0));
+ c1 = veorq_u64(c1, VEXT_U8<8>(vdupq_n_u64(0), c0));
c1 = veorq_u64(c1, PMULL_01(c0, r));
- c0 = VEXT_8<8>(c0, vdupq_n_u64(0));
+ c0 = VEXT_U8<8>(c0, vdupq_n_u64(0));
c0 = vshlq_n_u64(veorq_u64(c0, c1), 1);
c0 = PMULL_00(c0, r);
c2 = veorq_u64(c2, c0);
- c2 = veorq_u64(c2, VEXT_8<8>(c1, vdupq_n_u64(0)));
+ c2 = veorq_u64(c2, VEXT_U8<8>(c1, vdupq_n_u64(0)));
c1 = vshrq_n_u64(vcombine_u64(vget_low_u64(c1), vget_low_u64(c2)), 63);
c2 = vshlq_n_u64(c2, 1);
@@ -750,7 +750,7 @@ size_t GCM_Base::AuthenticateBlocks(const byte *data, size_t len) #if CRYPTOPP_BOOL_SSE2_ASM_AVAILABLE || defined(CRYPTOPP_X64_MASM_AVAILABLE)
+ HasSSE2()
//#elif CRYPTOPP_BOOL_NEON_INTRINSICS_AVAILABLE
-// + HasNEON()
+// + HasNEON()
#endif
)
{
|