summaryrefslogtreecommitdiff
path: root/lsh256.cpp
diff options
context:
space:
mode:
authorJeffrey Walton <noloader@gmail.com>2021-04-16 23:09:08 -0400
committerJeffrey Walton <noloader@gmail.com>2021-04-16 23:09:08 -0400
commit7d6a259140d463228332fd45d879da3bff9e18a1 (patch)
tree4ed34a8e73d43a1a226a03e3e967d671139f96fc /lsh256.cpp
parent0b281bc1ee15adb741d95ccf14dd87e474ab4cba (diff)
downloadcryptopp-git-7d6a259140d463228332fd45d879da3bff9e18a1.tar.gz
Update LSH-512
Fix a couple of typos
Diffstat (limited to 'lsh256.cpp')
-rw-r--r--lsh256.cpp117
1 files changed, 64 insertions, 53 deletions
diff --git a/lsh256.cpp b/lsh256.cpp
index 0d0c65ab..a41f918f 100644
--- a/lsh256.cpp
+++ b/lsh256.cpp
@@ -10,7 +10,13 @@
#include "lsh.h"
#include "misc.h"
-#if defined(__SSE2__) || defined(_M_X64)
+#ifndef CRYPTOPP_DISABLE_ASM
+# if defined(__SSE2__) || defined(_M_X64)
+# define CRYPTOPP_LSH256_ASM_AVAILABLE 1
+# endif
+#endif
+
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
# include <emmintrin.h>
# define M128_CAST(x) ((__m128i *)(void *)(x))
# define CONST_M128_CAST(x) ((const __m128i *)(const void *)(x))
@@ -187,56 +193,61 @@ inline void load_msg_blk(LSH256_Internal* i_state, const lsh_u32* msgblk)
CRYPTOPP_ASSERT(i_state != NULLPTR);
CRYPTOPP_ASSERT(msgblk != NULLPTR);
-#if defined(__SSE2__) || defined(_M_X64)
- _mm_storeu_si128(M128_CAST(i_state->submsg_e_l+0),
+ lsh_u32* submsg_e_l = i_state->submsg_e_l;
+ lsh_u32* submsg_e_r = i_state->submsg_e_r;
+ lsh_u32* submsg_o_l = i_state->submsg_o_l;
+ lsh_u32* submsg_o_r = i_state->submsg_o_r;
+
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
+ _mm_storeu_si128(M128_CAST(submsg_e_l+0),
_mm_loadu_si128(CONST_M128_CAST(msgblk+0)));
- _mm_storeu_si128(M128_CAST(i_state->submsg_e_l+4),
+ _mm_storeu_si128(M128_CAST(submsg_e_l+4),
_mm_loadu_si128(CONST_M128_CAST(msgblk+4)));
- _mm_storeu_si128(M128_CAST(i_state->submsg_e_r+0),
+ _mm_storeu_si128(M128_CAST(submsg_e_r+0),
_mm_loadu_si128(CONST_M128_CAST(msgblk+8)));
- _mm_storeu_si128(M128_CAST(i_state->submsg_e_r+4),
+ _mm_storeu_si128(M128_CAST(submsg_e_r+4),
_mm_loadu_si128(CONST_M128_CAST(msgblk+12)));
- _mm_storeu_si128(M128_CAST(i_state->submsg_o_l+0),
+ _mm_storeu_si128(M128_CAST(submsg_o_l+0),
_mm_loadu_si128(CONST_M128_CAST(msgblk+16)));
- _mm_storeu_si128(M128_CAST(i_state->submsg_o_l+4),
+ _mm_storeu_si128(M128_CAST(submsg_o_l+4),
_mm_loadu_si128(CONST_M128_CAST(msgblk+20)));
- _mm_storeu_si128(M128_CAST(i_state->submsg_o_r+0),
+ _mm_storeu_si128(M128_CAST(submsg_o_r+0),
_mm_loadu_si128(CONST_M128_CAST(msgblk+24)));
- _mm_storeu_si128(M128_CAST(i_state->submsg_o_r+4),
+ _mm_storeu_si128(M128_CAST(submsg_o_r+4),
_mm_loadu_si128(CONST_M128_CAST(msgblk+28)));
#else
- i_state->submsg_e_l[0] = loadLE32(msgblk[0]);
- i_state->submsg_e_l[1] = loadLE32(msgblk[1]);
- i_state->submsg_e_l[2] = loadLE32(msgblk[2]);
- i_state->submsg_e_l[3] = loadLE32(msgblk[3]);
- i_state->submsg_e_l[4] = loadLE32(msgblk[4]);
- i_state->submsg_e_l[5] = loadLE32(msgblk[5]);
- i_state->submsg_e_l[6] = loadLE32(msgblk[6]);
- i_state->submsg_e_l[7] = loadLE32(msgblk[7]);
- i_state->submsg_e_r[0] = loadLE32(msgblk[8]);
- i_state->submsg_e_r[1] = loadLE32(msgblk[9]);
- i_state->submsg_e_r[2] = loadLE32(msgblk[10]);
- i_state->submsg_e_r[3] = loadLE32(msgblk[11]);
- i_state->submsg_e_r[4] = loadLE32(msgblk[12]);
- i_state->submsg_e_r[5] = loadLE32(msgblk[13]);
- i_state->submsg_e_r[6] = loadLE32(msgblk[14]);
- i_state->submsg_e_r[7] = loadLE32(msgblk[15]);
- i_state->submsg_o_l[0] = loadLE32(msgblk[16]);
- i_state->submsg_o_l[1] = loadLE32(msgblk[17]);
- i_state->submsg_o_l[2] = loadLE32(msgblk[18]);
- i_state->submsg_o_l[3] = loadLE32(msgblk[19]);
- i_state->submsg_o_l[4] = loadLE32(msgblk[20]);
- i_state->submsg_o_l[5] = loadLE32(msgblk[21]);
- i_state->submsg_o_l[6] = loadLE32(msgblk[22]);
- i_state->submsg_o_l[7] = loadLE32(msgblk[23]);
- i_state->submsg_o_r[0] = loadLE32(msgblk[24]);
- i_state->submsg_o_r[1] = loadLE32(msgblk[25]);
- i_state->submsg_o_r[2] = loadLE32(msgblk[26]);
- i_state->submsg_o_r[3] = loadLE32(msgblk[27]);
- i_state->submsg_o_r[4] = loadLE32(msgblk[28]);
- i_state->submsg_o_r[5] = loadLE32(msgblk[29]);
- i_state->submsg_o_r[6] = loadLE32(msgblk[30]);
- i_state->submsg_o_r[7] = loadLE32(msgblk[31]);
+ submsg_e_l[0] = loadLE32(msgblk[0]);
+ submsg_e_l[1] = loadLE32(msgblk[1]);
+ submsg_e_l[2] = loadLE32(msgblk[2]);
+ submsg_e_l[3] = loadLE32(msgblk[3]);
+ submsg_e_l[4] = loadLE32(msgblk[4]);
+ submsg_e_l[5] = loadLE32(msgblk[5]);
+ submsg_e_l[6] = loadLE32(msgblk[6]);
+ submsg_e_l[7] = loadLE32(msgblk[7]);
+ submsg_e_r[0] = loadLE32(msgblk[8]);
+ submsg_e_r[1] = loadLE32(msgblk[9]);
+ submsg_e_r[2] = loadLE32(msgblk[10]);
+ submsg_e_r[3] = loadLE32(msgblk[11]);
+ submsg_e_r[4] = loadLE32(msgblk[12]);
+ submsg_e_r[5] = loadLE32(msgblk[13]);
+ submsg_e_r[6] = loadLE32(msgblk[14]);
+ submsg_e_r[7] = loadLE32(msgblk[15]);
+ submsg_o_l[0] = loadLE32(msgblk[16]);
+ submsg_o_l[1] = loadLE32(msgblk[17]);
+ submsg_o_l[2] = loadLE32(msgblk[18]);
+ submsg_o_l[3] = loadLE32(msgblk[19]);
+ submsg_o_l[4] = loadLE32(msgblk[20]);
+ submsg_o_l[5] = loadLE32(msgblk[21]);
+ submsg_o_l[6] = loadLE32(msgblk[22]);
+ submsg_o_l[7] = loadLE32(msgblk[23]);
+ submsg_o_r[0] = loadLE32(msgblk[24]);
+ submsg_o_r[1] = loadLE32(msgblk[25]);
+ submsg_o_r[2] = loadLE32(msgblk[26]);
+ submsg_o_r[3] = loadLE32(msgblk[27]);
+ submsg_o_r[4] = loadLE32(msgblk[28]);
+ submsg_o_r[5] = loadLE32(msgblk[29]);
+ submsg_o_r[6] = loadLE32(msgblk[30]);
+ submsg_o_r[7] = loadLE32(msgblk[31]);
#endif
}
@@ -249,7 +260,7 @@ inline void msg_exp_even(LSH256_Internal* i_state)
lsh_u32* submsg_o_l = i_state->submsg_o_l;
lsh_u32* submsg_o_r = i_state->submsg_o_r;
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(submsg_e_l+0), _mm_add_epi32(
_mm_shuffle_epi32(_mm_loadu_si128(CONST_M128_CAST(submsg_o_l+0)), _MM_SHUFFLE(3,2,1,0)),
_mm_shuffle_epi32(_mm_loadu_si128(CONST_M128_CAST(submsg_e_l+0)), _MM_SHUFFLE(1,0,2,3))));
@@ -299,7 +310,7 @@ inline void msg_exp_odd(LSH256_Internal* i_state)
lsh_u32* submsg_o_l = i_state->submsg_o_l;
lsh_u32* submsg_o_r = i_state->submsg_o_r;
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(submsg_o_l+0), _mm_add_epi32(
_mm_shuffle_epi32(_mm_loadu_si128(CONST_M128_CAST(submsg_e_l+0)), _MM_SHUFFLE(3,2,1,0)),
_mm_shuffle_epi32(_mm_loadu_si128(CONST_M128_CAST(submsg_o_l+0)), _MM_SHUFFLE(1,0,2,3))));
@@ -356,7 +367,7 @@ inline void msg_add_even(lsh_u32* cv_l, lsh_u32* cv_r, LSH256_Internal* i_state)
lsh_u32* submsg_e_l = i_state->submsg_e_l;
lsh_u32* submsg_e_r = i_state->submsg_e_r;
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(cv_l), _mm_xor_si128(
_mm_loadu_si128(CONST_M128_CAST(cv_l)),
_mm_loadu_si128(CONST_M128_CAST(submsg_e_l))));
@@ -390,7 +401,7 @@ inline void msg_add_odd(lsh_u32* cv_l, lsh_u32* cv_r, LSH256_Internal* i_state)
lsh_u32* submsg_o_l = i_state->submsg_o_l;
lsh_u32* submsg_o_r = i_state->submsg_o_r;
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(cv_l), _mm_xor_si128(
_mm_loadu_si128(CONST_M128_CAST(cv_l)),
_mm_loadu_si128(CONST_M128_CAST(submsg_o_l))));
@@ -420,7 +431,7 @@ inline void add_blk(lsh_u32* cv_l, const lsh_u32* cv_r)
CRYPTOPP_ASSERT(cv_l != NULLPTR);
CRYPTOPP_ASSERT(cv_r != NULLPTR);
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(cv_l), _mm_add_epi32(
_mm_loadu_si128(CONST_M128_CAST(cv_l)),
_mm_loadu_si128(CONST_M128_CAST(cv_r))));
@@ -449,7 +460,7 @@ inline void rotate_blk(lsh_u32 cv[8])
_mm_roti_epi32(_mm_loadu_si128(CONST_M128_CAST(cv)), R));
_mm_storeu_si128(M128_CAST(cv+4),
_mm_roti_epi32(_mm_loadu_si128(CONST_M128_CAST(cv+4)), R));
-#elif defined(__SSE2__) || defined(_M_X64)
+#elif defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(cv), _mm_or_si128(
_mm_slli_epi32(_mm_loadu_si128(CONST_M128_CAST(cv)), R),
_mm_srli_epi32(_mm_loadu_si128(CONST_M128_CAST(cv)), 32-R)));
@@ -473,7 +484,7 @@ inline void xor_with_const(lsh_u32* cv_l, const lsh_u32* const_v)
CRYPTOPP_ASSERT(cv_l != NULLPTR);
CRYPTOPP_ASSERT(const_v != NULLPTR);
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(cv_l), _mm_xor_si128(
_mm_loadu_si128(CONST_M128_CAST(cv_l)),
_mm_loadu_si128(CONST_M128_CAST(const_v))));
@@ -509,7 +520,7 @@ inline void word_perm(lsh_u32* cv_l, lsh_u32* cv_r)
CRYPTOPP_ASSERT(cv_l != NULLPTR);
CRYPTOPP_ASSERT(cv_r != NULLPTR);
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(cv_l+0), _mm_shuffle_epi32(
_mm_loadu_si128(CONST_M128_CAST(cv_l+0)), _MM_SHUFFLE(3,1,0,2)));
_mm_storeu_si128(M128_CAST(cv_l+4), _mm_shuffle_epi32(
@@ -616,7 +627,7 @@ inline void compress(LSH256_Context* ctx, const lsh_u32 pdMsgBlk[MSG_BLK_WORD_LE
inline void load_iv(word32* cv_l, word32* cv_r, const word32* iv)
{
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
// The IV's are aligned so we can use _mm_load_si128.
_mm_storeu_si128(M128_CAST(cv_l+ 0), _mm_load_si128(CONST_M128_CAST(iv+ 0)));
_mm_storeu_si128(M128_CAST(cv_l+ 4), _mm_load_si128(CONST_M128_CAST(iv+ 4)));
@@ -664,7 +675,7 @@ inline void fin(LSH256_Context* ctx)
{
CRYPTOPP_ASSERT(ctx != NULLPTR);
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
_mm_storeu_si128(M128_CAST(ctx->cv_l+0), _mm_xor_si128(
_mm_loadu_si128(CONST_M128_CAST(ctx->cv_l+0)),
_mm_loadu_si128(CONST_M128_CAST(ctx->cv_r+0))));
@@ -848,7 +859,7 @@ NAMESPACE_BEGIN(CryptoPP)
std::string LSH256_Base::AlgorithmProvider() const
{
-#if defined(__SSE2__) || defined(_M_X64)
+#if defined(CRYPTOPP_LSH256_ASM_AVAILABLE)
return "SSE2";
#else
return "C++";