-rw-r--r-- 7412 lib25519-20220726/crypto_hashblocks/sha512/avx/inner.c raw
#include <immintrin.h>
#include "inner.h"
#define uint64 crypto_uint64
static uint64 load_bigendian(const unsigned char *x)
{
return __builtin_bswap64(*(uint64 *) x);
}
static void store_bigendian(unsigned char *x,uint64 u)
{
*(uint64 *) x = __builtin_bswap64(u);
}
#define SHR(x,c) ((x) >> (c))
#define ROTR(x,c) (((x) >> (c)) | ((x) << (64 - (c))))
#define sigma0(x) (ROTR(x, 1) ^ ROTR(x, 8) ^ SHR(x,7))
#define sigma1(x) (ROTR(x,19) ^ ROTR(x,61) ^ SHR(x,6))
#define Ch(x,y,z) (z ^ (x & (y ^ z)))
#define Maj(x,y,z) ((x & (y ^ z)) ^ (y & z))
#define Sigma0(x) (ROTR(x,28) ^ ROTR(x,34) ^ ROTR(x,39))
#define Sigma1(x) (ROTR(x,14) ^ ROTR(x,18) ^ ROTR(x,41))
#define ALIGNED __attribute((aligned(32)))
#define load64(x) (*(uint64 *) (x))
#define store256(x,y) (*(volatile __m256i *) (x) = (y))
#define bigendian64 _mm256_set_epi8(8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,0,1,2,3,4,5,6,7)
#define PREEXPANDx4(X0,X9,X1) \
X0 = _mm256_add_epi64(X0, \
_mm256_srli_epi64(X1,1) ^ _mm256_slli_epi64(X1,63) ^ \
_mm256_srli_epi64(X1,8) ^ _mm256_slli_epi64(X1,56) ^ \
_mm256_srli_epi64(X1,7) \
); \
X0 = _mm256_add_epi64(X0,X9);
#define POSTEXPANDx4(X0,W0,W2,W14) \
W0 = ( \
_mm256_extracti128_si256(X0,0)); \
W0 = _mm_add_epi64(W0, \
_mm_srli_epi64(W14,19) ^ _mm_slli_epi64(W14,45) ^ \
_mm_srli_epi64(W14,61) ^ _mm_slli_epi64(W14,3) ^ \
_mm_srli_epi64(W14,6)); \
W2 = ( \
_mm256_extracti128_si256(X0,1)); \
W2 = _mm_add_epi64(W2, \
_mm_srli_epi64(W0,19) ^ _mm_slli_epi64(W0,45) ^ \
_mm_srli_epi64(W0,61) ^ _mm_slli_epi64(W0,3) ^ \
_mm_srli_epi64(W0,6)); \
X0 = _mm256_insertf128_si256(_mm256_castsi128_si256(W0),W2,1);
#define ROUND0(i,r0,r1,r2,r3,r4,r5,r6,r7) \
r7 += load64(&wc[i]); \
r7 += Ch(r4,r5,r6); \
r7 += Sigma1(r4); \
r3 += r7; \
r7 += Maj(r2,r0,r1); \
r7 += Sigma0(r0); \
#define ROUND1(i,r0,r1,r2,r3,r4,r5,r6,r7) \
r7 += load64(&wc[i]); \
r7 += Ch(r4,r5,r6); \
r7 += Sigma1(r4); \
r3 += r7; \
r7 += Maj(r0,r1,r2); \
r7 += Sigma0(r0); \
int inner(unsigned char *statebytes,const unsigned char *in,unsigned int inlen,const uint64 *constants)
{
ALIGNED uint64 state[8];
ALIGNED uint64 w[20];
ALIGNED uint64 wc[16]; /* w[i]+constants[i] */
uint64 r0,r1,r2,r3,r4,r5,r6,r7;
__m128i W0,W2,W4,W6,W8,W10,W12,W14;
__m256i X0,X1,X4,X5,X8,X9,X12,X13;
__m256i D0,D4,D8,D12;
int i;
state[0] = r0 = load_bigendian(statebytes);
state[1] = r1 = load_bigendian(statebytes+8);
state[2] = r2 = load_bigendian(statebytes+16);
state[3] = r3 = load_bigendian(statebytes+24);
state[4] = r4 = load_bigendian(statebytes+32);
state[5] = r5 = load_bigendian(statebytes+40);
state[6] = r6 = load_bigendian(statebytes+48);
state[7] = r7 = load_bigendian(statebytes+56);
do {
X0 = _mm256_loadu_si256((void *) (in+0));
X0 = _mm256_shuffle_epi8(X0,bigendian64);
D0 = _mm256_loadu_si256((void *) &constants[0]);
D0 = _mm256_add_epi64(X0,D0);
store256(&wc[0],D0);
store256(&w[0],X0);
store256(&w[16],X0);
X4 = _mm256_loadu_si256((void *) (in+32));
X4 = _mm256_shuffle_epi8(X4,bigendian64);
D4 = _mm256_loadu_si256((void *) &constants[4]);
D4 = _mm256_add_epi64(X4,D4);
store256(&wc[4],D4);
store256(&w[4],X4);
ROUND0(0,r0,r1,r2,r3,r4,r5,r6,r7)
ROUND1(1,r7,r0,r1,r2,r3,r4,r5,r6)
X8 = _mm256_loadu_si256((void *) (in+64));
X8 = _mm256_shuffle_epi8(X8,bigendian64);
D8 = _mm256_loadu_si256((void *) &constants[8]);
D8 = _mm256_add_epi64(X8,D8);
store256(&wc[8],D8);
store256(&w[8],X8);
ROUND0(2,r6,r7,r0,r1,r2,r3,r4,r5)
ROUND1(3,r5,r6,r7,r0,r1,r2,r3,r4)
ROUND0(4,r4,r5,r6,r7,r0,r1,r2,r3)
ROUND1(5,r3,r4,r5,r6,r7,r0,r1,r2)
X12 = _mm256_loadu_si256((void *) (in+96));
X12 = _mm256_shuffle_epi8(X12,bigendian64);
D12 = _mm256_loadu_si256((void *) &constants[12]);
D12 = _mm256_add_epi64(X12,D12);
store256(&wc[12],D12);
store256(&w[12],X12);
ROUND0(6,r2,r3,r4,r5,r6,r7,r0,r1)
ROUND1(7,r1,r2,r3,r4,r5,r6,r7,r0)
ROUND0(8,r0,r1,r2,r3,r4,r5,r6,r7)
ROUND1(9,r7,r0,r1,r2,r3,r4,r5,r6)
for (i = 4;i > 0;--i) {
constants += 16;
X1 = _mm256_loadu_si256((void *) (w+1));
X9 = _mm256_loadu_si256((void *) (w+9));
PREEXPANDx4(X0,X9,X1)
W14 = _mm_loadu_si128((void *) (w+14));
POSTEXPANDx4(X0,W0,W2,W14)
D0 = _mm256_loadu_si256((void *) &constants[0]);
D0 = _mm256_add_epi64(X0,D0);
store256(&wc[0],D0);
store256(w+16,X0);
store256(w+0,X0);
ROUND0(10,r6,r7,r0,r1,r2,r3,r4,r5)
ROUND1(11,r5,r6,r7,r0,r1,r2,r3,r4)
ROUND0(12,r4,r5,r6,r7,r0,r1,r2,r3)
ROUND1(13,r3,r4,r5,r6,r7,r0,r1,r2)
X5 = _mm256_loadu_si256((void *) (w+5));
X13 = _mm256_loadu_si256((void *) (w+13));
PREEXPANDx4(X4,X13,X5)
W2 = _mm_loadu_si128((void *) (w+2));
POSTEXPANDx4(X4,W4,W6,W2)
D4 = _mm256_loadu_si256((void *) &constants[4]);
D4 = _mm256_add_epi64(X4,D4);
store256(&wc[4],D4);
store256(w+4,X4);
ROUND0(14,r2,r3,r4,r5,r6,r7,r0,r1)
ROUND1(15,r1,r2,r3,r4,r5,r6,r7,r0)
ROUND0(0,r0,r1,r2,r3,r4,r5,r6,r7)
ROUND1(1,r7,r0,r1,r2,r3,r4,r5,r6)
X9 = _mm256_loadu_si256((void *) (w+9));
X1 = _mm256_loadu_si256((void *) (w+1));
PREEXPANDx4(X8,X1,X9)
W6 = _mm_loadu_si128((void *) (w+6));
POSTEXPANDx4(X8,W8,W10,W6)
D8 = _mm256_loadu_si256((void *) &constants[8]);
D8 = _mm256_add_epi64(X8,D8);
store256(&wc[8],D8);
store256(w+8,X8);
ROUND0(2,r6,r7,r0,r1,r2,r3,r4,r5)
ROUND1(3,r5,r6,r7,r0,r1,r2,r3,r4)
ROUND0(4,r4,r5,r6,r7,r0,r1,r2,r3)
ROUND1(5,r3,r4,r5,r6,r7,r0,r1,r2)
X13 = _mm256_loadu_si256((void *) (w+13));
X5 = _mm256_loadu_si256((void *) (w+5));
PREEXPANDx4(X12,X5,X13)
W10 = _mm_loadu_si128((void *) (w+10));
POSTEXPANDx4(X12,W12,W14,W10)
D12 = _mm256_loadu_si256((void *) &constants[12]);
D12 = _mm256_add_epi64(X12,D12);
store256(&wc[12],D12);
store256(w+12,X12);
ROUND0(6,r2,r3,r4,r5,r6,r7,r0,r1)
ROUND1(7,r1,r2,r3,r4,r5,r6,r7,r0)
ROUND0(8,r0,r1,r2,r3,r4,r5,r6,r7)
ROUND1(9,r7,r0,r1,r2,r3,r4,r5,r6)
}
{
ROUND0(10,r6,r7,r0,r1,r2,r3,r4,r5)
ROUND1(11,r5,r6,r7,r0,r1,r2,r3,r4)
ROUND0(12,r4,r5,r6,r7,r0,r1,r2,r3)
ROUND1(13,r3,r4,r5,r6,r7,r0,r1,r2)
ROUND0(14,r2,r3,r4,r5,r6,r7,r0,r1)
ROUND1(15,r1,r2,r3,r4,r5,r6,r7,r0)
}
constants -= 64;
r0 += state[0]; state[0] = r0;
r1 += state[1]; state[1] = r1;
r2 += state[2]; state[2] = r2;
r3 += state[3]; state[3] = r3;
r4 += state[4]; state[4] = r4;
r5 += state[5]; state[5] = r5;
r6 += state[6]; state[6] = r6;
r7 += state[7]; state[7] = r7;
in += 128;
inlen -= 128;
} while (inlen >= 128);
for (i = 0;i < 8;++i)
store_bigendian(statebytes+8*i,state[i]);
return inlen;
}