summaryrefslogtreecommitdiffstats
path: root/xmrstak/backend/nvidia/nvcc_code/cuda_keccak.hpp
diff options
context:
space:
mode:
authorpsychocrypt <psychocrypt@users.noreply.github.com>2017-09-29 20:32:31 +0200
committerpsychocrypt <psychocrypt@users.noreply.github.com>2017-09-30 23:46:08 +0200
commitcc429b68fadc502b981fd0acd64a5ff6e2ae1d15 (patch)
tree3fb23fc4db15dbdd08af4c7ea20134b9d82e58fd /xmrstak/backend/nvidia/nvcc_code/cuda_keccak.hpp
parente5b0319d5a9f58762fa934ad700113908940cb31 (diff)
downloadxmr-stak-cc429b68fadc502b981fd0acd64a5ff6e2ae1d15.zip
xmr-stak-cc429b68fadc502b981fd0acd64a5ff6e2ae1d15.tar.gz
group files
- move source code to `src` - categorize files and move to group folder - change upper case class files to lower case - change C++ header to `*.hpp`
Diffstat (limited to 'xmrstak/backend/nvidia/nvcc_code/cuda_keccak.hpp')
-rw-r--r--xmrstak/backend/nvidia/nvcc_code/cuda_keccak.hpp197
1 files changed, 197 insertions, 0 deletions
diff --git a/xmrstak/backend/nvidia/nvcc_code/cuda_keccak.hpp b/xmrstak/backend/nvidia/nvcc_code/cuda_keccak.hpp
new file mode 100644
index 0000000..99c6516
--- /dev/null
+++ b/xmrstak/backend/nvidia/nvcc_code/cuda_keccak.hpp
@@ -0,0 +1,197 @@
+#if __CUDA_ARCH__ >= 350
+ __forceinline__ __device__ uint64_t cuda_rotl64(const uint64_t value, const int offset)
+ {
+ uint2 result;
+ if(offset >= 32)
+ {
+ asm("shf.l.wrap.b32 %0, %1, %2, %3;" : "=r"(result.x) : "r"(__double2loint(__longlong_as_double(value))), "r"(__double2hiint(__longlong_as_double(value))), "r"(offset));
+ asm("shf.l.wrap.b32 %0, %1, %2, %3;" : "=r"(result.y) : "r"(__double2hiint(__longlong_as_double(value))), "r"(__double2loint(__longlong_as_double(value))), "r"(offset));
+ }
+ else
+ {
+ asm("shf.l.wrap.b32 %0, %1, %2, %3;" : "=r"(result.x) : "r"(__double2hiint(__longlong_as_double(value))), "r"(__double2loint(__longlong_as_double(value))), "r"(offset));
+ asm("shf.l.wrap.b32 %0, %1, %2, %3;" : "=r"(result.y) : "r"(__double2loint(__longlong_as_double(value))), "r"(__double2hiint(__longlong_as_double(value))), "r"(offset));
+ }
+ return __double_as_longlong(__hiloint2double(result.y, result.x));
+ }
+ #define rotl64_1(x, y) (cuda_rotl64((x), (y)))
+#else
+ #define rotl64_1(x, y) ((x) << (y) | ((x) >> (64 - (y))))
+#endif
+
+#define rotl64_2(x, y) rotl64_1(((x) >> 32) | ((x) << 32), (y))
+#define bitselect(a, b, c) ((a) ^ ((c) & ((b) ^ (a))))
+
+__device__ __forceinline__ void cn_keccakf2(uint64_t *s)
+{
+ uint8_t i;
+
+ for(i = 0; i < 24; ++i)
+ {
+ uint64_t bc[5], tmpxor[5], tmp1, tmp2;
+
+ tmpxor[0] = s[0] ^ s[5] ^ s[10] ^ s[15] ^ s[20];
+ tmpxor[1] = s[1] ^ s[6] ^ s[11] ^ s[16] ^ s[21];
+ tmpxor[2] = s[2] ^ s[7] ^ s[12] ^ s[17] ^ s[22];
+ tmpxor[3] = s[3] ^ s[8] ^ s[13] ^ s[18] ^ s[23];
+ tmpxor[4] = s[4] ^ s[9] ^ s[14] ^ s[19] ^ s[24];
+
+ bc[0] = tmpxor[0] ^ rotl64_1(tmpxor[2], 1);
+ bc[1] = tmpxor[1] ^ rotl64_1(tmpxor[3], 1);
+ bc[2] = tmpxor[2] ^ rotl64_1(tmpxor[4], 1);
+ bc[3] = tmpxor[3] ^ rotl64_1(tmpxor[0], 1);
+ bc[4] = tmpxor[4] ^ rotl64_1(tmpxor[1], 1);
+
+ tmp1 = s[1] ^ bc[0];
+
+ s[0] ^= bc[4];
+ s[1] = rotl64_2(s[6] ^ bc[0], 12);
+ s[6] = rotl64_1(s[9] ^ bc[3], 20);
+ s[9] = rotl64_2(s[22] ^ bc[1], 29);
+ s[22] = rotl64_2(s[14] ^ bc[3], 7);
+ s[14] = rotl64_1(s[20] ^ bc[4], 18);
+ s[20] = rotl64_2(s[2] ^ bc[1], 30);
+ s[2] = rotl64_2(s[12] ^ bc[1], 11);
+ s[12] = rotl64_1(s[13] ^ bc[2], 25);
+ s[13] = rotl64_1(s[19] ^ bc[3], 8);
+ s[19] = rotl64_2(s[23] ^ bc[2], 24);
+ s[23] = rotl64_2(s[15] ^ bc[4], 9);
+ s[15] = rotl64_1(s[4] ^ bc[3], 27);
+ s[4] = rotl64_1(s[24] ^ bc[3], 14);
+ s[24] = rotl64_1(s[21] ^ bc[0], 2);
+ s[21] = rotl64_2(s[8] ^ bc[2], 23);
+ s[8] = rotl64_2(s[16] ^ bc[0], 13);
+ s[16] = rotl64_2(s[5] ^ bc[4], 4);
+ s[5] = rotl64_1(s[3] ^ bc[2], 28);
+ s[3] = rotl64_1(s[18] ^ bc[2], 21);
+ s[18] = rotl64_1(s[17] ^ bc[1], 15);
+ s[17] = rotl64_1(s[11] ^ bc[0], 10);
+ s[11] = rotl64_1(s[7] ^ bc[1], 6);
+ s[7] = rotl64_1(s[10] ^ bc[4], 3);
+ s[10] = rotl64_1(tmp1, 1);
+
+ tmp1 = s[0]; tmp2 = s[1]; s[0] = bitselect(s[0] ^ s[2], s[0], s[1]); s[1] = bitselect(s[1] ^ s[3], s[1], s[2]); s[2] = bitselect(s[2] ^ s[4], s[2], s[3]); s[3] = bitselect(s[3] ^ tmp1, s[3], s[4]); s[4] = bitselect(s[4] ^ tmp2, s[4], tmp1);
+ tmp1 = s[5]; tmp2 = s[6]; s[5] = bitselect(s[5] ^ s[7], s[5], s[6]); s[6] = bitselect(s[6] ^ s[8], s[6], s[7]); s[7] = bitselect(s[7] ^ s[9], s[7], s[8]); s[8] = bitselect(s[8] ^ tmp1, s[8], s[9]); s[9] = bitselect(s[9] ^ tmp2, s[9], tmp1);
+ tmp1 = s[10]; tmp2 = s[11]; s[10] = bitselect(s[10] ^ s[12], s[10], s[11]); s[11] = bitselect(s[11] ^ s[13], s[11], s[12]); s[12] = bitselect(s[12] ^ s[14], s[12], s[13]); s[13] = bitselect(s[13] ^ tmp1, s[13], s[14]); s[14] = bitselect(s[14] ^ tmp2, s[14], tmp1);
+ tmp1 = s[15]; tmp2 = s[16]; s[15] = bitselect(s[15] ^ s[17], s[15], s[16]); s[16] = bitselect(s[16] ^ s[18], s[16], s[17]); s[17] = bitselect(s[17] ^ s[19], s[17], s[18]); s[18] = bitselect(s[18] ^ tmp1, s[18], s[19]); s[19] = bitselect(s[19] ^ tmp2, s[19], tmp1);
+ tmp1 = s[20]; tmp2 = s[21]; s[20] = bitselect(s[20] ^ s[22], s[20], s[21]); s[21] = bitselect(s[21] ^ s[23], s[21], s[22]); s[22] = bitselect(s[22] ^ s[24], s[22], s[23]); s[23] = bitselect(s[23] ^ tmp1, s[23], s[24]); s[24] = bitselect(s[24] ^ tmp2, s[24], tmp1);
+ s[0] ^= keccakf_rndc[i];
+ }
+}
+
+__device__ __forceinline__ void cn_keccakf(uint64_t *s)
+{
+ uint64_t bc[5], tmpxor[5], tmp1, tmp2;
+
+ tmpxor[0] = s[0] ^ s[5];
+ tmpxor[1] = s[1] ^ s[6] ^ 0x8000000000000000ULL;
+ tmpxor[2] = s[2] ^ s[7];
+ tmpxor[3] = s[3] ^ s[8];
+ tmpxor[4] = s[4] ^ s[9];
+
+ bc[0] = tmpxor[0] ^ rotl64_1(tmpxor[2], 1);
+ bc[1] = tmpxor[1] ^ rotl64_1(tmpxor[3], 1);
+ bc[2] = tmpxor[2] ^ rotl64_1(tmpxor[4], 1);
+ bc[3] = tmpxor[3] ^ rotl64_1(tmpxor[0], 1);
+ bc[4] = tmpxor[4] ^ rotl64_1(tmpxor[1], 1);
+
+ tmp1 = s[1] ^ bc[0];
+
+ s[0] ^= bc[4];
+ s[1] = rotl64_2(s[6] ^ bc[0], 12);
+ s[6] = rotl64_1(s[9] ^ bc[3], 20);
+ s[9] = rotl64_2(bc[1], 29);
+ s[22] = rotl64_2(bc[3], 7);
+ s[14] = rotl64_1(bc[4], 18);
+ s[20] = rotl64_2(s[2] ^ bc[1], 30);
+ s[2] = rotl64_2(bc[1], 11);
+ s[12] = rotl64_1(bc[2], 25);
+ s[13] = rotl64_1(bc[3], 8);
+ s[19] = rotl64_2(bc[2], 24);
+ s[23] = rotl64_2(bc[4], 9);
+ s[15] = rotl64_1(s[4] ^ bc[3], 27);
+ s[4] = rotl64_1(bc[3], 14);
+ s[24] = rotl64_1(bc[0], 2);
+ s[21] = rotl64_2(s[8] ^ bc[2], 23);
+ s[8] = rotl64_2(0x8000000000000000ULL ^ bc[0], 13);
+ s[16] = rotl64_2(s[5] ^ bc[4], 4);
+ s[5] = rotl64_1(s[3] ^ bc[2], 28);
+ s[3] = rotl64_1(bc[2], 21);
+ s[18] = rotl64_1(bc[1], 15);
+ s[17] = rotl64_1(bc[0], 10);
+ s[11] = rotl64_1(s[7] ^ bc[1], 6);
+ s[7] = rotl64_1(bc[4], 3);
+ s[10] = rotl64_1(tmp1, 1);
+
+ tmp1 = s[0]; tmp2 = s[1]; s[0] = bitselect(s[0] ^ s[2], s[0], s[1]); s[1] = bitselect(s[1] ^ s[3], s[1], s[2]); s[2] = bitselect(s[2] ^ s[4], s[2], s[3]); s[3] = bitselect(s[3] ^ tmp1, s[3], s[4]); s[4] = bitselect(s[4] ^ tmp2, s[4], tmp1);
+ tmp1 = s[5]; tmp2 = s[6]; s[5] = bitselect(s[5] ^ s[7], s[5], s[6]); s[6] = bitselect(s[6] ^ s[8], s[6], s[7]); s[7] = bitselect(s[7] ^ s[9], s[7], s[8]); s[8] = bitselect(s[8] ^ tmp1, s[8], s[9]); s[9] = bitselect(s[9] ^ tmp2, s[9], tmp1);
+ tmp1 = s[10]; tmp2 = s[11]; s[10] = bitselect(s[10] ^ s[12], s[10], s[11]); s[11] = bitselect(s[11] ^ s[13], s[11], s[12]); s[12] = bitselect(s[12] ^ s[14], s[12], s[13]); s[13] = bitselect(s[13] ^ tmp1, s[13], s[14]); s[14] = bitselect(s[14] ^ tmp2, s[14], tmp1);
+ tmp1 = s[15]; tmp2 = s[16]; s[15] = bitselect(s[15] ^ s[17], s[15], s[16]); s[16] = bitselect(s[16] ^ s[18], s[16], s[17]); s[17] = bitselect(s[17] ^ s[19], s[17], s[18]); s[18] = bitselect(s[18] ^ tmp1, s[18], s[19]); s[19] = bitselect(s[19] ^ tmp2, s[19], tmp1);
+ tmp1 = s[20]; tmp2 = s[21]; s[20] = bitselect(s[20] ^ s[22], s[20], s[21]); s[21] = bitselect(s[21] ^ s[23], s[21], s[22]); s[22] = bitselect(s[22] ^ s[24], s[22], s[23]); s[23] = bitselect(s[23] ^ tmp1, s[23], s[24]); s[24] = bitselect(s[24] ^ tmp2, s[24], tmp1);
+ s[0] ^= 0x0000000000000001;
+
+ for(int i = 1; i < 24; ++i)
+ {
+ tmpxor[0] = s[0] ^ s[5] ^ s[10] ^ s[15] ^ s[20];
+ tmpxor[1] = s[1] ^ s[6] ^ s[11] ^ s[16] ^ s[21];
+ tmpxor[2] = s[2] ^ s[7] ^ s[12] ^ s[17] ^ s[22];
+ tmpxor[3] = s[3] ^ s[8] ^ s[13] ^ s[18] ^ s[23];
+ tmpxor[4] = s[4] ^ s[9] ^ s[14] ^ s[19] ^ s[24];
+
+ bc[0] = tmpxor[0] ^ rotl64_1(tmpxor[2], 1);
+ bc[1] = tmpxor[1] ^ rotl64_1(tmpxor[3], 1);
+ bc[2] = tmpxor[2] ^ rotl64_1(tmpxor[4], 1);
+ bc[3] = tmpxor[3] ^ rotl64_1(tmpxor[0], 1);
+ bc[4] = tmpxor[4] ^ rotl64_1(tmpxor[1], 1);
+
+ tmp1 = s[1] ^ bc[0];
+
+ s[0] ^= bc[4];
+ s[1] = rotl64_2(s[6] ^ bc[0], 12);
+ s[6] = rotl64_1(s[9] ^ bc[3], 20);
+ s[9] = rotl64_2(s[22] ^ bc[1], 29);
+ s[22] = rotl64_2(s[14] ^ bc[3], 7);
+ s[14] = rotl64_1(s[20] ^ bc[4], 18);
+ s[20] = rotl64_2(s[2] ^ bc[1], 30);
+ s[2] = rotl64_2(s[12] ^ bc[1], 11);
+ s[12] = rotl64_1(s[13] ^ bc[2], 25);
+ s[13] = rotl64_1(s[19] ^ bc[3], 8);
+ s[19] = rotl64_2(s[23] ^ bc[2], 24);
+ s[23] = rotl64_2(s[15] ^ bc[4], 9);
+ s[15] = rotl64_1(s[4] ^ bc[3], 27);
+ s[4] = rotl64_1(s[24] ^ bc[3], 14);
+ s[24] = rotl64_1(s[21] ^ bc[0], 2);
+ s[21] = rotl64_2(s[8] ^ bc[2], 23);
+ s[8] = rotl64_2(s[16] ^ bc[0], 13);
+ s[16] = rotl64_2(s[5] ^ bc[4], 4);
+ s[5] = rotl64_1(s[3] ^ bc[2], 28);
+ s[3] = rotl64_1(s[18] ^ bc[2], 21);
+ s[18] = rotl64_1(s[17] ^ bc[1], 15);
+ s[17] = rotl64_1(s[11] ^ bc[0], 10);
+ s[11] = rotl64_1(s[7] ^ bc[1], 6);
+ s[7] = rotl64_1(s[10] ^ bc[4], 3);
+ s[10] = rotl64_1(tmp1, 1);
+
+ tmp1 = s[0]; tmp2 = s[1]; s[0] = bitselect(s[0] ^ s[2], s[0], s[1]); s[1] = bitselect(s[1] ^ s[3], s[1], s[2]); s[2] = bitselect(s[2] ^ s[4], s[2], s[3]); s[3] = bitselect(s[3] ^ tmp1, s[3], s[4]); s[4] = bitselect(s[4] ^ tmp2, s[4], tmp1);
+ tmp1 = s[5]; tmp2 = s[6]; s[5] = bitselect(s[5] ^ s[7], s[5], s[6]); s[6] = bitselect(s[6] ^ s[8], s[6], s[7]); s[7] = bitselect(s[7] ^ s[9], s[7], s[8]); s[8] = bitselect(s[8] ^ tmp1, s[8], s[9]); s[9] = bitselect(s[9] ^ tmp2, s[9], tmp1);
+ tmp1 = s[10]; tmp2 = s[11]; s[10] = bitselect(s[10] ^ s[12], s[10], s[11]); s[11] = bitselect(s[11] ^ s[13], s[11], s[12]); s[12] = bitselect(s[12] ^ s[14], s[12], s[13]); s[13] = bitselect(s[13] ^ tmp1, s[13], s[14]); s[14] = bitselect(s[14] ^ tmp2, s[14], tmp1);
+ tmp1 = s[15]; tmp2 = s[16]; s[15] = bitselect(s[15] ^ s[17], s[15], s[16]); s[16] = bitselect(s[16] ^ s[18], s[16], s[17]); s[17] = bitselect(s[17] ^ s[19], s[17], s[18]); s[18] = bitselect(s[18] ^ tmp1, s[18], s[19]); s[19] = bitselect(s[19] ^ tmp2, s[19], tmp1);
+ tmp1 = s[20]; tmp2 = s[21]; s[20] = bitselect(s[20] ^ s[22], s[20], s[21]); s[21] = bitselect(s[21] ^ s[23], s[21], s[22]); s[22] = bitselect(s[22] ^ s[24], s[22], s[23]); s[23] = bitselect(s[23] ^ tmp1, s[23], s[24]); s[24] = bitselect(s[24] ^ tmp2, s[24], tmp1);
+ s[0] ^= keccakf_rndc[i];
+ }
+}
+
+__device__ __forceinline__ void cn_keccak(const uint8_t * __restrict__ in, uint32_t len, uint8_t * __restrict__ md)
+{
+ uint64_t st[25];
+
+ MEMSET8(st + 8, 0x00, 25 - 8);
+ memcpy(st, in, len);
+ ((uint8_t*)st)[len] = 0x01;
+ st[16] = 0x8000000000000000ULL;
+
+ cn_keccakf(st);
+
+ MEMCPY8(md, st, 25);
+ return;
+}
OpenPOWER on IntegriCloud