|
|
|
/**
|
|
|
|
* BMW-256 CUDA Implementation - tpruvot 2015
|
|
|
|
*
|
|
|
|
* Not optimal but close to the sph version and easier to adapt.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <stdio.h>
|
|
|
|
#include <memory.h>
|
|
|
|
|
|
|
|
#define SPH_64 1
|
|
|
|
#define USE_MIDSTATE
|
|
|
|
|
|
|
|
extern "C" {
|
|
|
|
#include "sph/sph_bmw.h"
|
|
|
|
}
|
|
|
|
|
|
|
|
#include "cuda_helper.h"
|
|
|
|
|
|
|
|
__constant__ uint64_t c_PaddedMessage80[16]; // padded message (80 bytes + padding)
|
|
|
|
|
|
|
|
#ifndef USE_MIDSTATE
|
|
|
|
__constant__ static sph_u32 IV256[16] = {
|
|
|
|
0x40414243, 0x44454647, 0x48494A4B, 0x4C4D4E4F,
|
|
|
|
0x50515253, 0x54555657, 0x58595A5B, 0x5C5D5E5F,
|
|
|
|
0x60616263, 0x64656667, 0x68696A6B, 0x6C6D6E6F,
|
|
|
|
0x70717273, 0x74757677, 0x78797A7B, 0x7C7D7E7F
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
|
|
|
__constant__ static sph_u32 final_s[16] = {
|
|
|
|
0xaaaaaaa0, 0xaaaaaaa1, 0xaaaaaaa2, 0xaaaaaaa3,
|
|
|
|
0xaaaaaaa4, 0xaaaaaaa5, 0xaaaaaaa6, 0xaaaaaaa7,
|
|
|
|
0xaaaaaaa8, 0xaaaaaaa9, 0xaaaaaaaa, 0xaaaaaaab,
|
|
|
|
0xaaaaaaac, 0xaaaaaaad, 0xaaaaaaae, 0xaaaaaaaf
|
|
|
|
};
|
|
|
|
|
|
|
|
static sph_bmw_small_context* d_midstate[MAX_GPUS];
|
|
|
|
|
|
|
|
#define I16_16 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15
|
|
|
|
#define I16_17 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16
|
|
|
|
#define I16_18 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17
|
|
|
|
#define I16_19 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18
|
|
|
|
#define I16_20 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19
|
|
|
|
#define I16_21 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20
|
|
|
|
#define I16_22 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21
|
|
|
|
#define I16_23 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22
|
|
|
|
#define I16_24 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23
|
|
|
|
#define I16_25 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24
|
|
|
|
#define I16_26 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25
|
|
|
|
#define I16_27 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26
|
|
|
|
#define I16_28 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27
|
|
|
|
#define I16_29 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28
|
|
|
|
#define I16_30 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29
|
|
|
|
#define I16_31 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30
|
|
|
|
|
|
|
|
//#define M16_16 0, 1, 3, 4, 7, 10, 11
|
|
|
|
//#define M16_17 1, 2, 4, 5, 8, 11, 12
|
|
|
|
#define M16_18 2, 3, 5, 6, 9, 12, 13
|
|
|
|
#define M16_19 3, 4, 6, 7, 10, 13, 14
|
|
|
|
#define M16_20 4, 5, 7, 8, 11, 14, 15
|
|
|
|
#define M16_21 5, 6, 8, 9, 12, 15, 16
|
|
|
|
#define M16_22 6, 7, 9, 10, 13, 0, 1
|
|
|
|
#define M16_23 7, 8, 10, 11, 14, 1, 2
|
|
|
|
#define M16_24 8, 9, 11, 12, 15, 2, 3
|
|
|
|
#define M16_25 9, 10, 12, 13, 0, 3, 4
|
|
|
|
#define M16_26 10, 11, 13, 14, 1, 4, 5
|
|
|
|
#define M16_27 11, 12, 14, 15, 2, 5, 6
|
|
|
|
#define M16_28 12, 13, 15, 16, 3, 6, 7
|
|
|
|
#define M16_29 13, 14, 0, 1, 4, 7, 8
|
|
|
|
#define M16_30 14, 15, 1, 2, 5, 8, 9
|
|
|
|
#define M16_31 15, 16, 2, 3, 6, 9, 10
|
|
|
|
|
|
|
|
#define ss0(x) (((x) >> 1) ^ ((x) << 3) ^ ROTL32(x, 4) ^ ROTL32(x, 19))
|
|
|
|
#define ss1(x) (((x) >> 1) ^ ((x) << 2) ^ ROTL32(x, 8) ^ ROTL32(x, 23))
|
|
|
|
#define ss2(x) (((x) >> 2) ^ ((x) << 1) ^ ROTL32(x, 12) ^ ROTL32(x, 25))
|
|
|
|
#define ss3(x) (((x) >> 2) ^ ((x) << 2) ^ ROTL32(x, 15) ^ ROTL32(x, 29))
|
|
|
|
#define ss4(x) (((x) >> 1) ^ (x))
|
|
|
|
#define ss5(x) (((x) >> 2) ^ (x))
|
|
|
|
|
|
|
|
#define rs1(x) ROTL32(x, 3)
|
|
|
|
#define rs2(x) ROTL32(x, 7)
|
|
|
|
#define rs3(x) ROTL32(x, 13)
|
|
|
|
#define rs4(x) ROTL32(x, 16)
|
|
|
|
#define rs5(x) ROTL32(x, 19)
|
|
|
|
#define rs6(x) ROTL32(x, 23)
|
|
|
|
#define rs7(x) ROTL32(x, 27)
|
|
|
|
|
|
|
|
#define MAKE_W(tt, i0, op01, i1, op12, i2, op23, i3, op34, i4) \
|
|
|
|
tt((data[i0] ^ h[i0]) op01 (data[i1] ^ h[i1]) op12 (data[i2] ^ h[i2]) op23 (data[i3] ^ h[i3]) op34 (data[i4] ^ h[i4]))
|
|
|
|
//#define Ws0 MAKE_W(SPH_T32, 5, -, 7, +, 10, +, 13, +, 14)
|
|
|
|
//#define Ws1 MAKE_W(SPH_T32, 6, -, 8, +, 11, +, 14, -, 15)
|
|
|
|
//#define Ws2 MAKE_W(SPH_T32, 0, +, 7, +, 9, -, 12, +, 15)
|
|
|
|
//#define Ws3 MAKE_W(SPH_T32, 0, -, 1, +, 8, -, 10, +, 13)
|
|
|
|
//#define Ws4 MAKE_W(SPH_T32, 1, +, 2, +, 9, -, 11, -, 14)
|
|
|
|
//#define Ws5 MAKE_W(SPH_T32, 3, -, 2, +, 10, -, 12, +, 15)
|
|
|
|
//#define Ws6 MAKE_W(SPH_T32, 4, -, 0, -, 3, -, 11, +, 13)
|
|
|
|
//#define Ws7 MAKE_W(SPH_T32, 1, -, 4, -, 5, -, 12, -, 14)
|
|
|
|
//#define Ws8 MAKE_W(SPH_T32, 2, -, 5, -, 6, +, 13, -, 15)
|
|
|
|
//#define Ws9 MAKE_W(SPH_T32, 0, -, 3, +, 6, -, 7, +, 14)
|
|
|
|
//#define Ws10 MAKE_W(SPH_T32, 8, -, 1, -, 4, -, 7, +, 15)
|
|
|
|
//#define Ws11 MAKE_W(SPH_T32, 8, -, 0, -, 2, -, 5, +, 9)
|
|
|
|
//#define Ws12 MAKE_W(SPH_T32, 1, +, 3, -, 6, -, 9, +, 10)
|
|
|
|
//#define Ws13 MAKE_W(SPH_T32, 2, +, 4, +, 7, +, 10, +, 11)
|
|
|
|
//#define Ws14 MAKE_W(SPH_T32, 3, -, 5, +, 8, -, 11, -, 12)
|
|
|
|
//#define Ws15 MAKE_W(SPH_T32, 12, -, 4, -, 6, -, 9, +, 13)
|
|
|
|
|
|
|
|
__device__
|
|
|
|
static void gpu_compress_small(const sph_u32 *data, const sph_u32 h[16], sph_u32 dh[16])
|
|
|
|
{
|
|
|
|
// FOLD MAKE_Qas;
|
|
|
|
|
|
|
|
sph_u32 dx[16];
|
|
|
|
for (int i=0; i<16; i++)
|
|
|
|
dx[i] = data[i] ^ h[i];
|
|
|
|
|
|
|
|
sph_u32 qt[32];
|
|
|
|
qt[ 0] = dx[ 5] - dx[7] + dx[10] + dx[13] + dx[14]; // Ws0
|
|
|
|
qt[ 1] = dx[ 6] - dx[8] + dx[11] + dx[14] - dx[15]; // Ws1
|
|
|
|
qt[ 2] = dx[ 0] + dx[7] + dx[ 9] - dx[12] + dx[15]; // Ws2
|
|
|
|
qt[ 3] = dx[ 0] - dx[1] + dx[ 8] - dx[10] + dx[13]; // Ws3
|
|
|
|
qt[ 4] = dx[ 1] + dx[2] + dx[ 9] - dx[11] - dx[14]; // Ws4;
|
|
|
|
qt[ 5] = dx[ 3] - dx[2] + dx[10] - dx[12] + dx[15]; // Ws5;
|
|
|
|
qt[ 6] = dx[ 4] - dx[0] - dx[ 3] - dx[11] + dx[13]; // Ws6;
|
|
|
|
qt[ 7] = dx[ 1] - dx[4] - dx[ 5] - dx[12] - dx[14]; // Ws7;
|
|
|
|
qt[ 8] = dx[ 2] - dx[5] - dx[ 6] + dx[13] - dx[15]; // Ws8;
|
|
|
|
qt[ 9] = dx[ 0] - dx[3] + dx[ 6] - dx[ 7] + dx[14]; // Ws9;
|
|
|
|
qt[10] = dx[ 8] - dx[1] - dx[ 4] - dx[ 7] + dx[15]; // Ws10;
|
|
|
|
qt[11] = dx[ 8] - dx[0] - dx[ 2] - dx[ 5] + dx[ 9]; // Ws11;
|
|
|
|
qt[12] = dx[ 1] + dx[3] - dx[ 6] - dx[ 9] + dx[10]; // Ws12;
|
|
|
|
qt[13] = dx[ 2] + dx[4] + dx[ 7] + dx[10] + dx[11]; // Ws13;
|
|
|
|
qt[14] = dx[ 3] - dx[5] + dx[ 8] - dx[11] - dx[12]; // Ws14;
|
|
|
|
qt[15] = dx[12] - dx[4] - dx[ 6] - dx[ 9] + dx[13]; // Ws15;
|
|
|
|
|
|
|
|
qt[ 0] = ss0(qt[ 0]) + h[ 1];
|
|
|
|
qt[ 1] = ss1(qt[ 1]) + h[ 2];
|
|
|
|
qt[ 2] = ss2(qt[ 2]) + h[ 3];
|
|
|
|
qt[ 3] = ss3(qt[ 3]) + h[ 4];
|
|
|
|
qt[ 4] = ss4(qt[ 4]) + h[ 5];
|
|
|
|
|
|
|
|
qt[ 5] = ss0(qt[ 5]) + h[ 6];
|
|
|
|
qt[ 6] = ss1(qt[ 6]) + h[ 7];
|
|
|
|
qt[ 7] = ss2(qt[ 7]) + h[ 8];
|
|
|
|
qt[ 8] = ss3(qt[ 8]) + h[ 9];
|
|
|
|
qt[ 9] = ss4(qt[ 9]) + h[10];
|
|
|
|
|
|
|
|
qt[10] = ss0(qt[10]) + h[11];
|
|
|
|
qt[11] = ss1(qt[11]) + h[12];
|
|
|
|
qt[12] = ss2(qt[12]) + h[13];
|
|
|
|
qt[13] = ss3(qt[13]) + h[14];
|
|
|
|
qt[14] = ss4(qt[14]) + h[15];
|
|
|
|
|
|
|
|
qt[15] = ss0(qt[15]) + h[ 0];
|
|
|
|
|
|
|
|
//MAKE_Qbs;
|
|
|
|
#define Ks(j) ((sph_u32)(0x05555555UL * j))
|
|
|
|
#define Qs(j) (qt[j])
|
|
|
|
|
|
|
|
#define expand1s_in(i16, \
|
|
|
|
i0, i1, i2, i3, i4, i5, i6, i7, i8, i9, i10, i11, i12, i13, i14, i15, \
|
|
|
|
i0m, i1m, i3m, i4m, i7m, i10m, i11m) \
|
|
|
|
(ss1(qt[i0]) + ss2(qt[i1]) + ss3(qt[i2]) + ss0(qt[i3]) + ss1(qt[i4]) + ss2(qt[i5]) + ss3(qt[i6]) + ss0(qt[i7]) \
|
|
|
|
+ ss1(qt[i8]) + ss2(qt[i9]) + ss3(qt[i10]) + ss0(qt[i11]) + ss1(qt[i12]) + ss2(qt[i13]) + ss3(qt[i14]) + ss0(qt[i15]) \
|
|
|
|
+ ((ROTL32(data[i0m], i1m) + ROTL32(data[i3m], i4m) - ROTL32(data[i10m], i11m) + Ks(i16)) ^ h[i7m]))
|
|
|
|
|
|
|
|
qt[16] = expand1s_in(16, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 0, 1, 3, 4, 7, 10, 11);
|
|
|
|
qt[17] = expand1s_in(17, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 1, 2, 4, 5, 8, 11, 12);
|
|
|
|
|
|
|
|
#define expand2s_inner(qf, i16, \
|
|
|
|
i0, i1, i2, i3, i4, i5, i6, i7, i8, i9, i10, i11, i12, i13, i14, i15, \
|
|
|
|
i0m, i1m, i3m, i4m, i7m, i10m, i11m) \
|
|
|
|
(qf(i0) + rs1(qf(i1)) + qf(i2) + rs2(qf(i3)) \
|
|
|
|
+ qf(i4) + rs3(qf(i5)) + qf(i6) + rs4(qf(i7)) + qf(i8) + rs5(qf(i9)) + qf(i10) + rs6(qf(i11)) \
|
|
|
|
+ qf(i12) + rs7(qf(i13)) + ss4(qf(i14)) + ss5(qf(i15)) \
|
|
|
|
+ ((ROTL32(data[i0m], i1m) + ROTL32(data[i3m], i4m) - ROTL32(data[i10m], i11m) + Ks(i16)) ^ h[i7m]))
|
|
|
|
|
|
|
|
#ifdef _MSC_VER
|
|
|
|
#define LPAR (
|
|
|
|
#define expand2s(i16) \
|
|
|
|
expand2s_(Qs, i16, I16_ ## i16, M16_ ## i16)
|
|
|
|
#define expand2s_(qf, i16, ix, iy) \
|
|
|
|
expand2s_inner LPAR qf, i16, ix, iy)
|
|
|
|
#else
|
|
|
|
#define expand2s_(i16, ix, iy) \
|
|
|
|
expand2s_inner(Qs, i16, ix, iy)
|
|
|
|
#define expand2s(i16) \
|
|
|
|
expand2s_(i16, I16_ ## i16, M16_ ## i16)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
qt[18] = expand2s(18);
|
|
|
|
qt[19] = expand2s(19);
|
|
|
|
qt[20] = expand2s(20);
|
|
|
|
qt[21] = expand2s(21);
|
|
|
|
qt[22] = expand2s(22);
|
|
|
|
qt[23] = expand2s(23);
|
|
|
|
qt[24] = expand2s(24);
|
|
|
|
qt[25] = expand2s(25);
|
|
|
|
qt[26] = expand2s(26);
|
|
|
|
qt[27] = expand2s(27);
|
|
|
|
qt[28] = expand2s(28);
|
|
|
|
qt[29] = expand2s(29);
|
|
|
|
qt[30] = expand2s(30);
|
|
|
|
qt[31] = expand2s(31);
|
|
|
|
|
|
|
|
sph_u32 xl, xh;
|
|
|
|
xl = Qs(16) ^ Qs(17) ^ Qs(18) ^ Qs(19) ^ Qs(20) ^ Qs(21) ^ Qs(22) ^ Qs(23);
|
|
|
|
|
|
|
|
xh = xl ^ Qs(24) ^ Qs(25) ^ Qs(26) ^ Qs(27) ^ Qs(28) ^ Qs(29) ^ Qs(30) ^ Qs(31);
|
|
|
|
|
|
|
|
dh[ 0] = ((xh << 5) ^ (Qs(16) >> 5) ^ data[ 0]) + (xl ^ Qs(24) ^ Qs(0));
|
|
|
|
dh[ 1] = ((xh >> 7) ^ (Qs(17) << 8) ^ data[ 1]) + (xl ^ Qs(25) ^ Qs(1));
|
|
|
|
dh[ 2] = ((xh >> 5) ^ (Qs(18) << 5) ^ data[ 2]) + (xl ^ Qs(26) ^ Qs(2));
|
|
|
|
dh[ 3] = ((xh >> 1) ^ (Qs(19) << 5) ^ data[ 3]) + (xl ^ Qs(27) ^ Qs(3));
|
|
|
|
dh[ 4] = ((xh >> 3) ^ (Qs(20) << 0) ^ data[ 4]) + (xl ^ Qs(28) ^ Qs(4));
|
|
|
|
dh[ 5] = ((xh << 6) ^ (Qs(21) >> 6) ^ data[ 5]) + (xl ^ Qs(29) ^ Qs(5));
|
|
|
|
dh[ 6] = ((xh >> 4) ^ (Qs(22) << 6) ^ data[ 6]) + (xl ^ Qs(30) ^ Qs(6));
|
|
|
|
dh[ 7] = ((xh >> 11) ^ (Qs(23) << 2) ^ data[ 7]) + (xl ^ Qs(31) ^ Qs(7));
|
|
|
|
|
|
|
|
dh[ 8] = ROTL32(dh[4], 9) + (xh ^ Qs(24) ^ data[ 8]) + ((xl << 8) ^ Qs(23) ^ Qs( 8));
|
|
|
|
dh[ 9] = ROTL32(dh[5], 10) + (xh ^ Qs(25) ^ data[ 9]) + ((xl >> 6) ^ Qs(16) ^ Qs( 9));
|
|
|
|
dh[10] = ROTL32(dh[6], 11) + (xh ^ Qs(26) ^ data[10]) + ((xl << 6) ^ Qs(17) ^ Qs(10));
|
|
|
|
dh[11] = ROTL32(dh[7], 12) + (xh ^ Qs(27) ^ data[11]) + ((xl << 4) ^ Qs(18) ^ Qs(11));
|
|
|
|
dh[12] = ROTL32(dh[0], 13) + (xh ^ Qs(28) ^ data[12]) + ((xl >> 3) ^ Qs(19) ^ Qs(12));
|
|
|
|
dh[13] = ROTL32(dh[1], 14) + (xh ^ Qs(29) ^ data[13]) + ((xl >> 4) ^ Qs(20) ^ Qs(13));
|
|
|
|
dh[14] = ROTL32(dh[2], 15) + (xh ^ Qs(30) ^ data[14]) + ((xl >> 7) ^ Qs(21) ^ Qs(14));
|
|
|
|
dh[15] = ROTL32(dh[3], 16) + (xh ^ Qs(31) ^ data[15]) + ((xl >> 2) ^ Qs(22) ^ Qs(15));
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef USE_MIDSTATE
|
|
|
|
|
|
|
|
__device__
|
|
|
|
static void gpu_bmw256_init(sph_bmw_small_context *sc)
|
|
|
|
{
|
|
|
|
memcpy(sc->H, IV256, sizeof sc->H);
|
|
|
|
sc->ptr = 0;
|
|
|
|
sc->bit_count = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
__device__
|
|
|
|
static void gpu_bmw256(sph_bmw_small_context *sc, const void *data, size_t len)
|
|
|
|
{
|
|
|
|
sph_u32 htmp[16];
|
|
|
|
sph_u32 *h1, *h2;
|
|
|
|
unsigned char *buf = sc->buf;
|
|
|
|
size_t ptr = sc->ptr;
|
|
|
|
|
|
|
|
sc->bit_count += (sph_u64)len << 3;
|
|
|
|
|
|
|
|
h1 = sc->H;
|
|
|
|
h2 = htmp;
|
|
|
|
while (len > 0) {
|
|
|
|
size_t clen;
|
|
|
|
|
|
|
|
clen = (sizeof sc->buf) - ptr;
|
|
|
|
if (clen > len)
|
|
|
|
clen = len;
|
|
|
|
memcpy(buf + ptr, data, clen);
|
|
|
|
data = (const unsigned char *)data + clen;
|
|
|
|
len -= clen;
|
|
|
|
ptr += clen;
|
|
|
|
if (ptr == sizeof sc->buf) {
|
|
|
|
sph_u32 *ht;
|
|
|
|
|
|
|
|
gpu_compress_small((sph_u32 *) buf, h1, h2);
|
|
|
|
ht = h1;
|
|
|
|
h1 = h2;
|
|
|
|
h2 = ht;
|
|
|
|
ptr = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
sc->ptr = ptr;
|
|
|
|
if (h1 != sc->H)
|
|
|
|
memcpy(sc->H, h1, sizeof sc->H);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define sph_enc64le(ptr, x) \
|
|
|
|
*((uint64_t*)(ptr)) = x
|
|
|
|
#define sph_enc64le_aligned sph_enc64le
|
|
|
|
|
|
|
|
__device__
|
|
|
|
static void gpu_bmw256_close(sph_bmw_small_context *sc, uint2 *out)
|
|
|
|
{
|
|
|
|
unsigned char *buf = sc->buf;
|
|
|
|
size_t ptr = sc->ptr;
|
|
|
|
|
|
|
|
buf[ptr ++] = 0x80;
|
|
|
|
sph_u32 *h = sc->H;
|
|
|
|
|
|
|
|
sph_u32 h1[16];
|
|
|
|
if (ptr > (sizeof sc->buf) - 8) {
|
|
|
|
memset(buf + ptr, 0, (sizeof sc->buf) - ptr);
|
|
|
|
gpu_compress_small((sph_u32 *) buf, h, h1);
|
|
|
|
ptr = 0;
|
|
|
|
h = h1;
|
|
|
|
}
|
|
|
|
memset(buf + ptr, 0, sizeof(sc->buf) - 8 - ptr);
|
|
|
|
|
|
|
|
sph_enc64le_aligned(buf + sizeof(sc->buf) - 8, SPH_T64(sc->bit_count));
|
|
|
|
|
|
|
|
sph_u32 h2[16];
|
|
|
|
gpu_compress_small((sph_u32 *) buf, h, h2);
|
|
|
|
gpu_compress_small(h2, final_s, h1);
|
|
|
|
|
|
|
|
uint64_t* h64 = (uint64_t*) (&h1[8]);
|
|
|
|
#pragma unroll
|
|
|
|
for (int i = 0; i < 4; i++) {
|
|
|
|
out[i] = vectorize(h64[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
__global__ /* __launch_bounds__(256, 3) */
|
|
|
|
void bmw256_gpu_hash_80(uint32_t threads, uint32_t startNonce, uint64_t *g_hash, sph_bmw256_context *d_midstate, int swap)
|
|
|
|
{
|
|
|
|
uint32_t thread = (blockDim.x * blockIdx.x + threadIdx.x);
|
|
|
|
if (thread < threads)
|
|
|
|
{
|
|
|
|
uint32_t nonce = startNonce + thread;
|
|
|
|
nonce = swap ? cuda_swab32(nonce): nonce;
|
|
|
|
|
|
|
|
#ifndef USE_MIDSTATE
|
|
|
|
uint2 hash[10];
|
|
|
|
#pragma unroll
|
|
|
|
for(int i=0;i<9;i++)
|
|
|
|
hash[i] = vectorize(c_PaddedMessage80[i]);
|
|
|
|
hash[9] = make_uint2(c_PaddedMessage80[9], nonce);
|
|
|
|
|
|
|
|
sph_bmw256_context ctx;
|
|
|
|
gpu_bmw256_init(&ctx);
|
|
|
|
gpu_bmw256(&ctx, (void*) hash, 80);
|
|
|
|
#else
|
|
|
|
sph_bmw256_context ctx;
|
|
|
|
ctx.ptr = 16; ctx.bit_count = 640;
|
|
|
|
uint2 *buf = (uint2 *) ctx.buf;
|
|
|
|
buf[0] = vectorize(c_PaddedMessage80[8]);
|
|
|
|
buf[1] = make_uint2(c_PaddedMessage80[9], nonce);
|
|
|
|
#pragma unroll
|
|
|
|
for(int i=0;i<16;i++)
|
|
|
|
ctx.H[i] = d_midstate->H[i];
|
|
|
|
#endif
|
|
|
|
gpu_bmw256_close(&ctx, (uint2*) &g_hash[thread << 2]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
__host__
|
|
|
|
void bmw256_cpu_hash_80(int thr_id, uint32_t threads, uint32_t startNonce, uint32_t *d_outputHash, int swap)
|
|
|
|
{
|
|
|
|
const uint32_t threadsperblock = 256;
|
|
|
|
|
|
|
|
dim3 grid((threads + threadsperblock-1)/threadsperblock);
|
|
|
|
dim3 block(threadsperblock);
|
|
|
|
|
|
|
|
bmw256_gpu_hash_80<<<grid, block>>>(threads, startNonce, (uint64_t*)d_outputHash, d_midstate[thr_id], swap);
|
|
|
|
}
|
|
|
|
|
|
|
|
__host__
|
|
|
|
void bmw256_setBlock_80(int thr_id, void *pdata)
|
|
|
|
{
|
|
|
|
uint64_t PaddedMessage[16];
|
|
|
|
memcpy(PaddedMessage, pdata, 80);
|
|
|
|
memset(&PaddedMessage[10], 0, 48);
|
|
|
|
CUDA_SAFE_CALL(cudaMemcpyToSymbol(c_PaddedMessage80, PaddedMessage, sizeof(PaddedMessage), 0, cudaMemcpyHostToDevice));
|
|
|
|
|
|
|
|
sph_bmw256_context ctx;
|
|
|
|
sph_bmw256_init(&ctx);
|
|
|
|
sph_bmw256(&ctx, (void*) PaddedMessage, 80);
|
|
|
|
CUDA_SAFE_CALL(cudaMemcpy(d_midstate[thr_id], &ctx, sizeof(sph_bmw256_context), cudaMemcpyHostToDevice));
|
|
|
|
}
|
|
|
|
|
|
|
|
__host__
|
|
|
|
void bmw256_midstate_init(int thr_id, uint32_t threads)
|
|
|
|
{
|
|
|
|
cudaMalloc(&d_midstate[thr_id], sizeof(sph_bmw256_context));
|
|
|
|
}
|
|
|
|
|
|
|
|
__host__
|
|
|
|
void bmw256_midstate_free(int thr_id)
|
|
|
|
{
|
|
|
|
cudaFree(d_midstate[thr_id]);
|
|
|
|
}
|