summaryrefslogtreecommitdiff
path: root/REORG.TODO/crypt/crypt.c
diff options
context:
space:
mode:
Diffstat (limited to 'REORG.TODO/crypt/crypt.c')
-rw-r--r--REORG.TODO/crypt/crypt.c115
1 files changed, 115 insertions, 0 deletions
diff --git a/REORG.TODO/crypt/crypt.c b/REORG.TODO/crypt/crypt.c
new file mode 100644
index 0000000000..17f61e35c2
--- /dev/null
+++ b/REORG.TODO/crypt/crypt.c
@@ -0,0 +1,115 @@
+/*
+ * UFC-crypt: ultra fast crypt(3) implementation
+ *
+ * Copyright (C) 1991-2017 Free Software Foundation, Inc.
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with this library; see the file COPYING.LIB. If not,
+ * see <http://www.gnu.org/licenses/>.
+ *
+ * @(#)crypt.c 2.25 12/20/96
+ *
+ * Semiportable C version
+ *
+ */
+
+#include "crypt-private.h"
+
+#ifdef _UFC_32_
+
+/*
+ * 32 bit version
+ */
+
+#define SBA(sb, v) (*(long32*)((char*)(sb)+(v)))
+
+void
+_ufc_doit_r (ufc_long itr, struct crypt_data * __restrict __data,
+ ufc_long *res)
+{
+ int i;
+ long32 s, *k;
+ long32 *sb01 = (long32*)__data->sb0;
+ long32 *sb23 = (long32*)__data->sb2;
+ long32 l1, l2, r1, r2;
+
+ l1 = (long32)res[0]; l2 = (long32)res[1];
+ r1 = (long32)res[2]; r2 = (long32)res[3];
+
+ while(itr--) {
+ k = (long32*)__data->keysched;
+ for(i=8; i--; ) {
+ s = *k++ ^ r1;
+ l1 ^= SBA(sb01, s & 0xffff); l2 ^= SBA(sb01, (s & 0xffff)+4);
+ l1 ^= SBA(sb01, s >>= 16 ); l2 ^= SBA(sb01, (s )+4);
+ s = *k++ ^ r2;
+ l1 ^= SBA(sb23, s & 0xffff); l2 ^= SBA(sb23, (s & 0xffff)+4);
+ l1 ^= SBA(sb23, s >>= 16 ); l2 ^= SBA(sb23, (s )+4);
+
+ s = *k++ ^ l1;
+ r1 ^= SBA(sb01, s & 0xffff); r2 ^= SBA(sb01, (s & 0xffff)+4);
+ r1 ^= SBA(sb01, s >>= 16 ); r2 ^= SBA(sb01, (s )+4);
+ s = *k++ ^ l2;
+ r1 ^= SBA(sb23, s & 0xffff); r2 ^= SBA(sb23, (s & 0xffff)+4);
+ r1 ^= SBA(sb23, s >>= 16 ); r2 ^= SBA(sb23, (s )+4);
+ }
+ s=l1; l1=r1; r1=s; s=l2; l2=r2; r2=s;
+ }
+ res[0] = l1; res[1] = l2; res[2] = r1; res[3] = r2;
+}
+
+#endif
+
+#ifdef _UFC_64_
+
+/*
+ * 64 bit version
+ */
+
+#define SBA(sb, v) (*(long64*)((char*)(sb)+(v)))
+
+void
+_ufc_doit_r (ufc_long itr, struct crypt_data * __restrict __data,
+ ufc_long *res)
+{
+ int i;
+ long64 l, r, s, *k;
+ long64 *sb01 = (long64*)__data->sb0;
+ long64 *sb23 = (long64*)__data->sb2;
+
+ l = (((long64)res[0]) << 32) | ((long64)res[1]);
+ r = (((long64)res[2]) << 32) | ((long64)res[3]);
+
+ while(itr--) {
+ k = (long64*)__data->keysched;
+ for(i=8; i--; ) {
+ s = *k++ ^ r;
+ l ^= SBA(sb23, (s ) & 0xffff);
+ l ^= SBA(sb23, (s >>= 16) & 0xffff);
+ l ^= SBA(sb01, (s >>= 16) & 0xffff);
+ l ^= SBA(sb01, (s >>= 16) );
+
+ s = *k++ ^ l;
+ r ^= SBA(sb23, (s ) & 0xffff);
+ r ^= SBA(sb23, (s >>= 16) & 0xffff);
+ r ^= SBA(sb01, (s >>= 16) & 0xffff);
+ r ^= SBA(sb01, (s >>= 16) );
+ }
+ s=l; l=r; r=s;
+ }
+
+ res[0] = l >> 32; res[1] = l & 0xffffffff;
+ res[2] = r >> 32; res[3] = r & 0xffffffff;
+}
+
+#endif