mirror of https://github.com/GOSTSec/ccminer
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
957 lines
28 KiB
957 lines
28 KiB
/* $Id: bmw.c 227 2010-06-16 17:28:38Z tp $ */ |
|
/* |
|
* BMW implementation. |
|
* |
|
* ==========================(LICENSE BEGIN)============================ |
|
* |
|
* Copyright (c) 2007-2010 Projet RNRT SAPHIR |
|
* |
|
* Permission is hereby granted, free of charge, to any person obtaining |
|
* a copy of this software and associated documentation files (the |
|
* "Software"), to deal in the Software without restriction, including |
|
* without limitation the rights to use, copy, modify, merge, publish, |
|
* distribute, sublicense, and/or sell copies of the Software, and to |
|
* permit persons to whom the Software is furnished to do so, subject to |
|
* the following conditions: |
|
* |
|
* The above copyright notice and this permission notice shall be |
|
* included in all copies or substantial portions of the Software. |
|
* |
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, |
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. |
|
* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY |
|
* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, |
|
* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE |
|
* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
|
* |
|
* ===========================(LICENSE END)============================= |
|
* |
|
* @author Thomas Pornin <thomas.pornin@cryptolog.com> |
|
*/ |
|
|
|
#include <stddef.h> |
|
#include <string.h> |
|
#include <limits.h> |
|
|
|
#include "sph_bmw.h" |
|
|
|
#if SPH_SMALL_FOOTPRINT && !defined SPH_SMALL_FOOTPRINT_BMW |
|
#define SPH_SMALL_FOOTPRINT_BMW 1 |
|
#endif |
|
|
|
#ifdef _MSC_VER |
|
#pragma warning (disable: 4146) |
|
#endif |
|
|
|
static const sph_u32 IV224[] = { |
|
SPH_C32(0x00010203), SPH_C32(0x04050607), |
|
SPH_C32(0x08090A0B), SPH_C32(0x0C0D0E0F), |
|
SPH_C32(0x10111213), SPH_C32(0x14151617), |
|
SPH_C32(0x18191A1B), SPH_C32(0x1C1D1E1F), |
|
SPH_C32(0x20212223), SPH_C32(0x24252627), |
|
SPH_C32(0x28292A2B), SPH_C32(0x2C2D2E2F), |
|
SPH_C32(0x30313233), SPH_C32(0x34353637), |
|
SPH_C32(0x38393A3B), SPH_C32(0x3C3D3E3F) |
|
}; |
|
|
|
static const sph_u32 IV256[] = { |
|
SPH_C32(0x40414243), SPH_C32(0x44454647), |
|
SPH_C32(0x48494A4B), SPH_C32(0x4C4D4E4F), |
|
SPH_C32(0x50515253), SPH_C32(0x54555657), |
|
SPH_C32(0x58595A5B), SPH_C32(0x5C5D5E5F), |
|
SPH_C32(0x60616263), SPH_C32(0x64656667), |
|
SPH_C32(0x68696A6B), SPH_C32(0x6C6D6E6F), |
|
SPH_C32(0x70717273), SPH_C32(0x74757677), |
|
SPH_C32(0x78797A7B), SPH_C32(0x7C7D7E7F) |
|
}; |
|
|
|
#if SPH_64 |
|
|
|
static const sph_u64 IV384[] = { |
|
SPH_C64(0x0001020304050607), SPH_C64(0x08090A0B0C0D0E0F), |
|
SPH_C64(0x1011121314151617), SPH_C64(0x18191A1B1C1D1E1F), |
|
SPH_C64(0x2021222324252627), SPH_C64(0x28292A2B2C2D2E2F), |
|
SPH_C64(0x3031323334353637), SPH_C64(0x38393A3B3C3D3E3F), |
|
SPH_C64(0x4041424344454647), SPH_C64(0x48494A4B4C4D4E4F), |
|
SPH_C64(0x5051525354555657), SPH_C64(0x58595A5B5C5D5E5F), |
|
SPH_C64(0x6061626364656667), SPH_C64(0x68696A6B6C6D6E6F), |
|
SPH_C64(0x7071727374757677), SPH_C64(0x78797A7B7C7D7E7F) |
|
}; |
|
|
|
static const sph_u64 IV512[] = { |
|
SPH_C64(0x8081828384858687), SPH_C64(0x88898A8B8C8D8E8F), |
|
SPH_C64(0x9091929394959697), SPH_C64(0x98999A9B9C9D9E9F), |
|
SPH_C64(0xA0A1A2A3A4A5A6A7), SPH_C64(0xA8A9AAABACADAEAF), |
|
SPH_C64(0xB0B1B2B3B4B5B6B7), SPH_C64(0xB8B9BABBBCBDBEBF), |
|
SPH_C64(0xC0C1C2C3C4C5C6C7), SPH_C64(0xC8C9CACBCCCDCECF), |
|
SPH_C64(0xD0D1D2D3D4D5D6D7), SPH_C64(0xD8D9DADBDCDDDEDF), |
|
SPH_C64(0xE0E1E2E3E4E5E6E7), SPH_C64(0xE8E9EAEBECEDEEEF), |
|
SPH_C64(0xF0F1F2F3F4F5F6F7), SPH_C64(0xF8F9FAFBFCFDFEFF) |
|
}; |
|
|
|
#endif |
|
|
|
#define XCAT(x, y) XCAT_(x, y) |
|
#define XCAT_(x, y) x ## y |
|
|
|
#define LPAR ( |
|
|
|
#define I16_16 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 |
|
#define I16_17 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16 |
|
#define I16_18 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17 |
|
#define I16_19 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18 |
|
#define I16_20 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19 |
|
#define I16_21 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20 |
|
#define I16_22 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21 |
|
#define I16_23 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22 |
|
#define I16_24 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23 |
|
#define I16_25 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24 |
|
#define I16_26 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25 |
|
#define I16_27 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26 |
|
#define I16_28 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27 |
|
#define I16_29 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28 |
|
#define I16_30 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29 |
|
#define I16_31 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30 |
|
|
|
#define M16_16 0, 1, 3, 4, 7, 10, 11 |
|
#define M16_17 1, 2, 4, 5, 8, 11, 12 |
|
#define M16_18 2, 3, 5, 6, 9, 12, 13 |
|
#define M16_19 3, 4, 6, 7, 10, 13, 14 |
|
#define M16_20 4, 5, 7, 8, 11, 14, 15 |
|
#define M16_21 5, 6, 8, 9, 12, 15, 16 |
|
#define M16_22 6, 7, 9, 10, 13, 0, 1 |
|
#define M16_23 7, 8, 10, 11, 14, 1, 2 |
|
#define M16_24 8, 9, 11, 12, 15, 2, 3 |
|
#define M16_25 9, 10, 12, 13, 0, 3, 4 |
|
#define M16_26 10, 11, 13, 14, 1, 4, 5 |
|
#define M16_27 11, 12, 14, 15, 2, 5, 6 |
|
#define M16_28 12, 13, 15, 16, 3, 6, 7 |
|
#define M16_29 13, 14, 0, 1, 4, 7, 8 |
|
#define M16_30 14, 15, 1, 2, 5, 8, 9 |
|
#define M16_31 15, 16, 2, 3, 6, 9, 10 |
|
|
|
#define ss0(x) (((x) >> 1) ^ SPH_T32((x) << 3) \ |
|
^ SPH_ROTL32(x, 4) ^ SPH_ROTL32(x, 19)) |
|
#define ss1(x) (((x) >> 1) ^ SPH_T32((x) << 2) \ |
|
^ SPH_ROTL32(x, 8) ^ SPH_ROTL32(x, 23)) |
|
#define ss2(x) (((x) >> 2) ^ SPH_T32((x) << 1) \ |
|
^ SPH_ROTL32(x, 12) ^ SPH_ROTL32(x, 25)) |
|
#define ss3(x) (((x) >> 2) ^ SPH_T32((x) << 2) \ |
|
^ SPH_ROTL32(x, 15) ^ SPH_ROTL32(x, 29)) |
|
#define ss4(x) (((x) >> 1) ^ (x)) |
|
#define ss5(x) (((x) >> 2) ^ (x)) |
|
#define rs1(x) SPH_ROTL32(x, 3) |
|
#define rs2(x) SPH_ROTL32(x, 7) |
|
#define rs3(x) SPH_ROTL32(x, 13) |
|
#define rs4(x) SPH_ROTL32(x, 16) |
|
#define rs5(x) SPH_ROTL32(x, 19) |
|
#define rs6(x) SPH_ROTL32(x, 23) |
|
#define rs7(x) SPH_ROTL32(x, 27) |
|
|
|
#define Ks(j) SPH_T32((sph_u32)(j) * SPH_C32(0x05555555)) |
|
|
|
#define add_elt_s(mf, hf, j0m, j1m, j3m, j4m, j7m, j10m, j11m, j16) \ |
|
(SPH_T32(SPH_ROTL32(mf(j0m), j1m) + SPH_ROTL32(mf(j3m), j4m) \ |
|
- SPH_ROTL32(mf(j10m), j11m) + Ks(j16)) ^ hf(j7m)) |
|
|
|
#define expand1s_inner(qf, mf, hf, i16, \ |
|
i0, i1, i2, i3, i4, i5, i6, i7, i8, \ |
|
i9, i10, i11, i12, i13, i14, i15, \ |
|
i0m, i1m, i3m, i4m, i7m, i10m, i11m) \ |
|
SPH_T32(ss1(qf(i0)) + ss2(qf(i1)) + ss3(qf(i2)) + ss0(qf(i3)) \ |
|
+ ss1(qf(i4)) + ss2(qf(i5)) + ss3(qf(i6)) + ss0(qf(i7)) \ |
|
+ ss1(qf(i8)) + ss2(qf(i9)) + ss3(qf(i10)) + ss0(qf(i11)) \ |
|
+ ss1(qf(i12)) + ss2(qf(i13)) + ss3(qf(i14)) + ss0(qf(i15)) \ |
|
+ add_elt_s(mf, hf, i0m, i1m, i3m, i4m, i7m, i10m, i11m, i16)) |
|
|
|
#define expand1s(qf, mf, hf, i16) \ |
|
expand1s_(qf, mf, hf, i16, I16_ ## i16, M16_ ## i16) |
|
#define expand1s_(qf, mf, hf, i16, ix, iy) \ |
|
expand1s_inner LPAR qf, mf, hf, i16, ix, iy) |
|
|
|
#define expand2s_inner(qf, mf, hf, i16, \ |
|
i0, i1, i2, i3, i4, i5, i6, i7, i8, \ |
|
i9, i10, i11, i12, i13, i14, i15, \ |
|
i0m, i1m, i3m, i4m, i7m, i10m, i11m) \ |
|
SPH_T32(qf(i0) + rs1(qf(i1)) + qf(i2) + rs2(qf(i3)) \ |
|
+ qf(i4) + rs3(qf(i5)) + qf(i6) + rs4(qf(i7)) \ |
|
+ qf(i8) + rs5(qf(i9)) + qf(i10) + rs6(qf(i11)) \ |
|
+ qf(i12) + rs7(qf(i13)) + ss4(qf(i14)) + ss5(qf(i15)) \ |
|
+ add_elt_s(mf, hf, i0m, i1m, i3m, i4m, i7m, i10m, i11m, i16)) |
|
|
|
#define expand2s(qf, mf, hf, i16) \ |
|
expand2s_(qf, mf, hf, i16, I16_ ## i16, M16_ ## i16) |
|
#define expand2s_(qf, mf, hf, i16, ix, iy) \ |
|
expand2s_inner LPAR qf, mf, hf, i16, ix, iy) |
|
|
|
#if SPH_64 |
|
|
|
#define sb0(x) (((x) >> 1) ^ SPH_T64((x) << 3) \ |
|
^ SPH_ROTL64(x, 4) ^ SPH_ROTL64(x, 37)) |
|
#define sb1(x) (((x) >> 1) ^ SPH_T64((x) << 2) \ |
|
^ SPH_ROTL64(x, 13) ^ SPH_ROTL64(x, 43)) |
|
#define sb2(x) (((x) >> 2) ^ SPH_T64((x) << 1) \ |
|
^ SPH_ROTL64(x, 19) ^ SPH_ROTL64(x, 53)) |
|
#define sb3(x) (((x) >> 2) ^ SPH_T64((x) << 2) \ |
|
^ SPH_ROTL64(x, 28) ^ SPH_ROTL64(x, 59)) |
|
#define sb4(x) (((x) >> 1) ^ (x)) |
|
#define sb5(x) (((x) >> 2) ^ (x)) |
|
#define rb1(x) SPH_ROTL64(x, 5) |
|
#define rb2(x) SPH_ROTL64(x, 11) |
|
#define rb3(x) SPH_ROTL64(x, 27) |
|
#define rb4(x) SPH_ROTL64(x, 32) |
|
#define rb5(x) SPH_ROTL64(x, 37) |
|
#define rb6(x) SPH_ROTL64(x, 43) |
|
#define rb7(x) SPH_ROTL64(x, 53) |
|
|
|
#define Kb(j) SPH_T64((sph_u64)(j) * SPH_C64(0x0555555555555555)) |
|
|
|
#if SPH_SMALL_FOOTPRINT_BMW |
|
|
|
static const sph_u64 Kb_tab[] = { |
|
Kb(16), Kb(17), Kb(18), Kb(19), Kb(20), Kb(21), Kb(22), Kb(23), |
|
Kb(24), Kb(25), Kb(26), Kb(27), Kb(28), Kb(29), Kb(30), Kb(31) |
|
}; |
|
|
|
#define rol_off(mf, j, off) \ |
|
SPH_ROTL64(mf(((j) + (off)) & 15), (((j) + (off)) & 15) + 1) |
|
|
|
#define add_elt_b(mf, hf, j) \ |
|
(SPH_T64(rol_off(mf, j, 0) + rol_off(mf, j, 3) \ |
|
- rol_off(mf, j, 10) + Kb_tab[j]) ^ hf(((j) + 7) & 15)) |
|
|
|
#define expand1b(qf, mf, hf, i) \ |
|
SPH_T64(sb1(qf((i) - 16)) + sb2(qf((i) - 15)) \ |
|
+ sb3(qf((i) - 14)) + sb0(qf((i) - 13)) \ |
|
+ sb1(qf((i) - 12)) + sb2(qf((i) - 11)) \ |
|
+ sb3(qf((i) - 10)) + sb0(qf((i) - 9)) \ |
|
+ sb1(qf((i) - 8)) + sb2(qf((i) - 7)) \ |
|
+ sb3(qf((i) - 6)) + sb0(qf((i) - 5)) \ |
|
+ sb1(qf((i) - 4)) + sb2(qf((i) - 3)) \ |
|
+ sb3(qf((i) - 2)) + sb0(qf((i) - 1)) \ |
|
+ add_elt_b(mf, hf, (i) - 16)) |
|
|
|
#define expand2b(qf, mf, hf, i) \ |
|
SPH_T64(qf((i) - 16) + rb1(qf((i) - 15)) \ |
|
+ qf((i) - 14) + rb2(qf((i) - 13)) \ |
|
+ qf((i) - 12) + rb3(qf((i) - 11)) \ |
|
+ qf((i) - 10) + rb4(qf((i) - 9)) \ |
|
+ qf((i) - 8) + rb5(qf((i) - 7)) \ |
|
+ qf((i) - 6) + rb6(qf((i) - 5)) \ |
|
+ qf((i) - 4) + rb7(qf((i) - 3)) \ |
|
+ sb4(qf((i) - 2)) + sb5(qf((i) - 1)) \ |
|
+ add_elt_b(mf, hf, (i) - 16)) |
|
|
|
#else |
|
|
|
#define add_elt_b(mf, hf, j0m, j1m, j3m, j4m, j7m, j10m, j11m, j16) \ |
|
(SPH_T64(SPH_ROTL64(mf(j0m), j1m) + SPH_ROTL64(mf(j3m), j4m) \ |
|
- SPH_ROTL64(mf(j10m), j11m) + Kb(j16)) ^ hf(j7m)) |
|
|
|
#define expand1b_inner(qf, mf, hf, i16, \ |
|
i0, i1, i2, i3, i4, i5, i6, i7, i8, \ |
|
i9, i10, i11, i12, i13, i14, i15, \ |
|
i0m, i1m, i3m, i4m, i7m, i10m, i11m) \ |
|
SPH_T64(sb1(qf(i0)) + sb2(qf(i1)) + sb3(qf(i2)) + sb0(qf(i3)) \ |
|
+ sb1(qf(i4)) + sb2(qf(i5)) + sb3(qf(i6)) + sb0(qf(i7)) \ |
|
+ sb1(qf(i8)) + sb2(qf(i9)) + sb3(qf(i10)) + sb0(qf(i11)) \ |
|
+ sb1(qf(i12)) + sb2(qf(i13)) + sb3(qf(i14)) + sb0(qf(i15)) \ |
|
+ add_elt_b(mf, hf, i0m, i1m, i3m, i4m, i7m, i10m, i11m, i16)) |
|
|
|
#define expand1b(qf, mf, hf, i16) \ |
|
expand1b_(qf, mf, hf, i16, I16_ ## i16, M16_ ## i16) |
|
#define expand1b_(qf, mf, hf, i16, ix, iy) \ |
|
expand1b_inner LPAR qf, mf, hf, i16, ix, iy) |
|
|
|
#define expand2b_inner(qf, mf, hf, i16, \ |
|
i0, i1, i2, i3, i4, i5, i6, i7, i8, \ |
|
i9, i10, i11, i12, i13, i14, i15, \ |
|
i0m, i1m, i3m, i4m, i7m, i10m, i11m) \ |
|
SPH_T64(qf(i0) + rb1(qf(i1)) + qf(i2) + rb2(qf(i3)) \ |
|
+ qf(i4) + rb3(qf(i5)) + qf(i6) + rb4(qf(i7)) \ |
|
+ qf(i8) + rb5(qf(i9)) + qf(i10) + rb6(qf(i11)) \ |
|
+ qf(i12) + rb7(qf(i13)) + sb4(qf(i14)) + sb5(qf(i15)) \ |
|
+ add_elt_b(mf, hf, i0m, i1m, i3m, i4m, i7m, i10m, i11m, i16)) |
|
|
|
#define expand2b(qf, mf, hf, i16) \ |
|
expand2b_(qf, mf, hf, i16, I16_ ## i16, M16_ ## i16) |
|
#define expand2b_(qf, mf, hf, i16, ix, iy) \ |
|
expand2b_inner LPAR qf, mf, hf, i16, ix, iy) |
|
|
|
#endif |
|
|
|
#endif |
|
|
|
#define MAKE_W(tt, i0, op01, i1, op12, i2, op23, i3, op34, i4) \ |
|
tt((M(i0) ^ H(i0)) op01 (M(i1) ^ H(i1)) op12 (M(i2) ^ H(i2)) \ |
|
op23 (M(i3) ^ H(i3)) op34 (M(i4) ^ H(i4))) |
|
|
|
#define Ws0 MAKE_W(SPH_T32, 5, -, 7, +, 10, +, 13, +, 14) |
|
#define Ws1 MAKE_W(SPH_T32, 6, -, 8, +, 11, +, 14, -, 15) |
|
#define Ws2 MAKE_W(SPH_T32, 0, +, 7, +, 9, -, 12, +, 15) |
|
#define Ws3 MAKE_W(SPH_T32, 0, -, 1, +, 8, -, 10, +, 13) |
|
#define Ws4 MAKE_W(SPH_T32, 1, +, 2, +, 9, -, 11, -, 14) |
|
#define Ws5 MAKE_W(SPH_T32, 3, -, 2, +, 10, -, 12, +, 15) |
|
#define Ws6 MAKE_W(SPH_T32, 4, -, 0, -, 3, -, 11, +, 13) |
|
#define Ws7 MAKE_W(SPH_T32, 1, -, 4, -, 5, -, 12, -, 14) |
|
#define Ws8 MAKE_W(SPH_T32, 2, -, 5, -, 6, +, 13, -, 15) |
|
#define Ws9 MAKE_W(SPH_T32, 0, -, 3, +, 6, -, 7, +, 14) |
|
#define Ws10 MAKE_W(SPH_T32, 8, -, 1, -, 4, -, 7, +, 15) |
|
#define Ws11 MAKE_W(SPH_T32, 8, -, 0, -, 2, -, 5, +, 9) |
|
#define Ws12 MAKE_W(SPH_T32, 1, +, 3, -, 6, -, 9, +, 10) |
|
#define Ws13 MAKE_W(SPH_T32, 2, +, 4, +, 7, +, 10, +, 11) |
|
#define Ws14 MAKE_W(SPH_T32, 3, -, 5, +, 8, -, 11, -, 12) |
|
#define Ws15 MAKE_W(SPH_T32, 12, -, 4, -, 6, -, 9, +, 13) |
|
|
|
#if SPH_SMALL_FOOTPRINT_BMW |
|
|
|
#define MAKE_Qas do { \ |
|
unsigned u; \ |
|
sph_u32 Ws[16]; \ |
|
Ws[ 0] = Ws0; \ |
|
Ws[ 1] = Ws1; \ |
|
Ws[ 2] = Ws2; \ |
|
Ws[ 3] = Ws3; \ |
|
Ws[ 4] = Ws4; \ |
|
Ws[ 5] = Ws5; \ |
|
Ws[ 6] = Ws6; \ |
|
Ws[ 7] = Ws7; \ |
|
Ws[ 8] = Ws8; \ |
|
Ws[ 9] = Ws9; \ |
|
Ws[10] = Ws10; \ |
|
Ws[11] = Ws11; \ |
|
Ws[12] = Ws12; \ |
|
Ws[13] = Ws13; \ |
|
Ws[14] = Ws14; \ |
|
Ws[15] = Ws15; \ |
|
for (u = 0; u < 15; u += 5) { \ |
|
qt[u + 0] = SPH_T32(ss0(Ws[u + 0]) + H(u + 1)); \ |
|
qt[u + 1] = SPH_T32(ss1(Ws[u + 1]) + H(u + 2)); \ |
|
qt[u + 2] = SPH_T32(ss2(Ws[u + 2]) + H(u + 3)); \ |
|
qt[u + 3] = SPH_T32(ss3(Ws[u + 3]) + H(u + 4)); \ |
|
qt[u + 4] = SPH_T32(ss4(Ws[u + 4]) + H(u + 5)); \ |
|
} \ |
|
qt[15] = SPH_T32(ss0(Ws[15]) + H(0)); \ |
|
} while (0) |
|
|
|
#define MAKE_Qbs do { \ |
|
qt[16] = expand1s(Qs, M, H, 16); \ |
|
qt[17] = expand1s(Qs, M, H, 17); \ |
|
qt[18] = expand2s(Qs, M, H, 18); \ |
|
qt[19] = expand2s(Qs, M, H, 19); \ |
|
qt[20] = expand2s(Qs, M, H, 20); \ |
|
qt[21] = expand2s(Qs, M, H, 21); \ |
|
qt[22] = expand2s(Qs, M, H, 22); \ |
|
qt[23] = expand2s(Qs, M, H, 23); \ |
|
qt[24] = expand2s(Qs, M, H, 24); \ |
|
qt[25] = expand2s(Qs, M, H, 25); \ |
|
qt[26] = expand2s(Qs, M, H, 26); \ |
|
qt[27] = expand2s(Qs, M, H, 27); \ |
|
qt[28] = expand2s(Qs, M, H, 28); \ |
|
qt[29] = expand2s(Qs, M, H, 29); \ |
|
qt[30] = expand2s(Qs, M, H, 30); \ |
|
qt[31] = expand2s(Qs, M, H, 31); \ |
|
} while (0) |
|
|
|
#else |
|
|
|
#define MAKE_Qas do { \ |
|
qt[ 0] = SPH_T32(ss0(Ws0 ) + H( 1)); \ |
|
qt[ 1] = SPH_T32(ss1(Ws1 ) + H( 2)); \ |
|
qt[ 2] = SPH_T32(ss2(Ws2 ) + H( 3)); \ |
|
qt[ 3] = SPH_T32(ss3(Ws3 ) + H( 4)); \ |
|
qt[ 4] = SPH_T32(ss4(Ws4 ) + H( 5)); \ |
|
qt[ 5] = SPH_T32(ss0(Ws5 ) + H( 6)); \ |
|
qt[ 6] = SPH_T32(ss1(Ws6 ) + H( 7)); \ |
|
qt[ 7] = SPH_T32(ss2(Ws7 ) + H( 8)); \ |
|
qt[ 8] = SPH_T32(ss3(Ws8 ) + H( 9)); \ |
|
qt[ 9] = SPH_T32(ss4(Ws9 ) + H(10)); \ |
|
qt[10] = SPH_T32(ss0(Ws10) + H(11)); \ |
|
qt[11] = SPH_T32(ss1(Ws11) + H(12)); \ |
|
qt[12] = SPH_T32(ss2(Ws12) + H(13)); \ |
|
qt[13] = SPH_T32(ss3(Ws13) + H(14)); \ |
|
qt[14] = SPH_T32(ss4(Ws14) + H(15)); \ |
|
qt[15] = SPH_T32(ss0(Ws15) + H( 0)); \ |
|
} while (0) |
|
|
|
#define MAKE_Qbs do { \ |
|
qt[16] = expand1s(Qs, M, H, 16); \ |
|
qt[17] = expand1s(Qs, M, H, 17); \ |
|
qt[18] = expand2s(Qs, M, H, 18); \ |
|
qt[19] = expand2s(Qs, M, H, 19); \ |
|
qt[20] = expand2s(Qs, M, H, 20); \ |
|
qt[21] = expand2s(Qs, M, H, 21); \ |
|
qt[22] = expand2s(Qs, M, H, 22); \ |
|
qt[23] = expand2s(Qs, M, H, 23); \ |
|
qt[24] = expand2s(Qs, M, H, 24); \ |
|
qt[25] = expand2s(Qs, M, H, 25); \ |
|
qt[26] = expand2s(Qs, M, H, 26); \ |
|
qt[27] = expand2s(Qs, M, H, 27); \ |
|
qt[28] = expand2s(Qs, M, H, 28); \ |
|
qt[29] = expand2s(Qs, M, H, 29); \ |
|
qt[30] = expand2s(Qs, M, H, 30); \ |
|
qt[31] = expand2s(Qs, M, H, 31); \ |
|
} while (0) |
|
|
|
#endif |
|
|
|
#define MAKE_Qs do { \ |
|
MAKE_Qas; \ |
|
MAKE_Qbs; \ |
|
} while (0) |
|
|
|
#define Qs(j) (qt[j]) |
|
|
|
#if SPH_64 |
|
|
|
#define Wb0 MAKE_W(SPH_T64, 5, -, 7, +, 10, +, 13, +, 14) |
|
#define Wb1 MAKE_W(SPH_T64, 6, -, 8, +, 11, +, 14, -, 15) |
|
#define Wb2 MAKE_W(SPH_T64, 0, +, 7, +, 9, -, 12, +, 15) |
|
#define Wb3 MAKE_W(SPH_T64, 0, -, 1, +, 8, -, 10, +, 13) |
|
#define Wb4 MAKE_W(SPH_T64, 1, +, 2, +, 9, -, 11, -, 14) |
|
#define Wb5 MAKE_W(SPH_T64, 3, -, 2, +, 10, -, 12, +, 15) |
|
#define Wb6 MAKE_W(SPH_T64, 4, -, 0, -, 3, -, 11, +, 13) |
|
#define Wb7 MAKE_W(SPH_T64, 1, -, 4, -, 5, -, 12, -, 14) |
|
#define Wb8 MAKE_W(SPH_T64, 2, -, 5, -, 6, +, 13, -, 15) |
|
#define Wb9 MAKE_W(SPH_T64, 0, -, 3, +, 6, -, 7, +, 14) |
|
#define Wb10 MAKE_W(SPH_T64, 8, -, 1, -, 4, -, 7, +, 15) |
|
#define Wb11 MAKE_W(SPH_T64, 8, -, 0, -, 2, -, 5, +, 9) |
|
#define Wb12 MAKE_W(SPH_T64, 1, +, 3, -, 6, -, 9, +, 10) |
|
#define Wb13 MAKE_W(SPH_T64, 2, +, 4, +, 7, +, 10, +, 11) |
|
#define Wb14 MAKE_W(SPH_T64, 3, -, 5, +, 8, -, 11, -, 12) |
|
#define Wb15 MAKE_W(SPH_T64, 12, -, 4, -, 6, -, 9, +, 13) |
|
|
|
#if SPH_SMALL_FOOTPRINT_BMW |
|
|
|
#define MAKE_Qab do { \ |
|
unsigned u; \ |
|
sph_u64 Wb[16]; \ |
|
Wb[ 0] = Wb0; \ |
|
Wb[ 1] = Wb1; \ |
|
Wb[ 2] = Wb2; \ |
|
Wb[ 3] = Wb3; \ |
|
Wb[ 4] = Wb4; \ |
|
Wb[ 5] = Wb5; \ |
|
Wb[ 6] = Wb6; \ |
|
Wb[ 7] = Wb7; \ |
|
Wb[ 8] = Wb8; \ |
|
Wb[ 9] = Wb9; \ |
|
Wb[10] = Wb10; \ |
|
Wb[11] = Wb11; \ |
|
Wb[12] = Wb12; \ |
|
Wb[13] = Wb13; \ |
|
Wb[14] = Wb14; \ |
|
Wb[15] = Wb15; \ |
|
for (u = 0; u < 15; u += 5) { \ |
|
qt[u + 0] = SPH_T64(sb0(Wb[u + 0]) + H(u + 1)); \ |
|
qt[u + 1] = SPH_T64(sb1(Wb[u + 1]) + H(u + 2)); \ |
|
qt[u + 2] = SPH_T64(sb2(Wb[u + 2]) + H(u + 3)); \ |
|
qt[u + 3] = SPH_T64(sb3(Wb[u + 3]) + H(u + 4)); \ |
|
qt[u + 4] = SPH_T64(sb4(Wb[u + 4]) + H(u + 5)); \ |
|
} \ |
|
qt[15] = SPH_T64(sb0(Wb[15]) + H(0)); \ |
|
} while (0) |
|
|
|
#define MAKE_Qbb do { \ |
|
unsigned u; \ |
|
for (u = 16; u < 18; u ++) \ |
|
qt[u] = expand1b(Qb, M, H, u); \ |
|
for (u = 18; u < 32; u ++) \ |
|
qt[u] = expand2b(Qb, M, H, u); \ |
|
} while (0) |
|
|
|
#else |
|
|
|
#define MAKE_Qab do { \ |
|
qt[ 0] = SPH_T64(sb0(Wb0 ) + H( 1)); \ |
|
qt[ 1] = SPH_T64(sb1(Wb1 ) + H( 2)); \ |
|
qt[ 2] = SPH_T64(sb2(Wb2 ) + H( 3)); \ |
|
qt[ 3] = SPH_T64(sb3(Wb3 ) + H( 4)); \ |
|
qt[ 4] = SPH_T64(sb4(Wb4 ) + H( 5)); \ |
|
qt[ 5] = SPH_T64(sb0(Wb5 ) + H( 6)); \ |
|
qt[ 6] = SPH_T64(sb1(Wb6 ) + H( 7)); \ |
|
qt[ 7] = SPH_T64(sb2(Wb7 ) + H( 8)); \ |
|
qt[ 8] = SPH_T64(sb3(Wb8 ) + H( 9)); \ |
|
qt[ 9] = SPH_T64(sb4(Wb9 ) + H(10)); \ |
|
qt[10] = SPH_T64(sb0(Wb10) + H(11)); \ |
|
qt[11] = SPH_T64(sb1(Wb11) + H(12)); \ |
|
qt[12] = SPH_T64(sb2(Wb12) + H(13)); \ |
|
qt[13] = SPH_T64(sb3(Wb13) + H(14)); \ |
|
qt[14] = SPH_T64(sb4(Wb14) + H(15)); \ |
|
qt[15] = SPH_T64(sb0(Wb15) + H( 0)); \ |
|
} while (0) |
|
|
|
#define MAKE_Qbb do { \ |
|
qt[16] = expand1b(Qb, M, H, 16); \ |
|
qt[17] = expand1b(Qb, M, H, 17); \ |
|
qt[18] = expand2b(Qb, M, H, 18); \ |
|
qt[19] = expand2b(Qb, M, H, 19); \ |
|
qt[20] = expand2b(Qb, M, H, 20); \ |
|
qt[21] = expand2b(Qb, M, H, 21); \ |
|
qt[22] = expand2b(Qb, M, H, 22); \ |
|
qt[23] = expand2b(Qb, M, H, 23); \ |
|
qt[24] = expand2b(Qb, M, H, 24); \ |
|
qt[25] = expand2b(Qb, M, H, 25); \ |
|
qt[26] = expand2b(Qb, M, H, 26); \ |
|
qt[27] = expand2b(Qb, M, H, 27); \ |
|
qt[28] = expand2b(Qb, M, H, 28); \ |
|
qt[29] = expand2b(Qb, M, H, 29); \ |
|
qt[30] = expand2b(Qb, M, H, 30); \ |
|
qt[31] = expand2b(Qb, M, H, 31); \ |
|
} while (0) |
|
|
|
#endif |
|
|
|
#define MAKE_Qb do { \ |
|
MAKE_Qab; \ |
|
MAKE_Qbb; \ |
|
} while (0) |
|
|
|
#define Qb(j) (qt[j]) |
|
|
|
#endif |
|
|
|
#define FOLD(type, mkQ, tt, rol, mf, qf, dhf) do { \ |
|
type qt[32], xl, xh; \ |
|
mkQ; \ |
|
xl = qf(16) ^ qf(17) ^ qf(18) ^ qf(19) \ |
|
^ qf(20) ^ qf(21) ^ qf(22) ^ qf(23); \ |
|
xh = xl ^ qf(24) ^ qf(25) ^ qf(26) ^ qf(27) \ |
|
^ qf(28) ^ qf(29) ^ qf(30) ^ qf(31); \ |
|
dhf( 0) = tt(((xh << 5) ^ (qf(16) >> 5) ^ mf( 0)) \ |
|
+ (xl ^ qf(24) ^ qf( 0))); \ |
|
dhf( 1) = tt(((xh >> 7) ^ (qf(17) << 8) ^ mf( 1)) \ |
|
+ (xl ^ qf(25) ^ qf( 1))); \ |
|
dhf( 2) = tt(((xh >> 5) ^ (qf(18) << 5) ^ mf( 2)) \ |
|
+ (xl ^ qf(26) ^ qf( 2))); \ |
|
dhf( 3) = tt(((xh >> 1) ^ (qf(19) << 5) ^ mf( 3)) \ |
|
+ (xl ^ qf(27) ^ qf( 3))); \ |
|
dhf( 4) = tt(((xh >> 3) ^ (qf(20) << 0) ^ mf( 4)) \ |
|
+ (xl ^ qf(28) ^ qf( 4))); \ |
|
dhf( 5) = tt(((xh << 6) ^ (qf(21) >> 6) ^ mf( 5)) \ |
|
+ (xl ^ qf(29) ^ qf( 5))); \ |
|
dhf( 6) = tt(((xh >> 4) ^ (qf(22) << 6) ^ mf( 6)) \ |
|
+ (xl ^ qf(30) ^ qf( 6))); \ |
|
dhf( 7) = tt(((xh >> 11) ^ (qf(23) << 2) ^ mf( 7)) \ |
|
+ (xl ^ qf(31) ^ qf( 7))); \ |
|
dhf( 8) = tt(rol(dhf(4), 9) + (xh ^ qf(24) ^ mf( 8)) \ |
|
+ ((xl << 8) ^ qf(23) ^ qf( 8))); \ |
|
dhf( 9) = tt(rol(dhf(5), 10) + (xh ^ qf(25) ^ mf( 9)) \ |
|
+ ((xl >> 6) ^ qf(16) ^ qf( 9))); \ |
|
dhf(10) = tt(rol(dhf(6), 11) + (xh ^ qf(26) ^ mf(10)) \ |
|
+ ((xl << 6) ^ qf(17) ^ qf(10))); \ |
|
dhf(11) = tt(rol(dhf(7), 12) + (xh ^ qf(27) ^ mf(11)) \ |
|
+ ((xl << 4) ^ qf(18) ^ qf(11))); \ |
|
dhf(12) = tt(rol(dhf(0), 13) + (xh ^ qf(28) ^ mf(12)) \ |
|
+ ((xl >> 3) ^ qf(19) ^ qf(12))); \ |
|
dhf(13) = tt(rol(dhf(1), 14) + (xh ^ qf(29) ^ mf(13)) \ |
|
+ ((xl >> 4) ^ qf(20) ^ qf(13))); \ |
|
dhf(14) = tt(rol(dhf(2), 15) + (xh ^ qf(30) ^ mf(14)) \ |
|
+ ((xl >> 7) ^ qf(21) ^ qf(14))); \ |
|
dhf(15) = tt(rol(dhf(3), 16) + (xh ^ qf(31) ^ mf(15)) \ |
|
+ ((xl >> 2) ^ qf(22) ^ qf(15))); \ |
|
} while (0) |
|
|
|
#define FOLDs FOLD(sph_u32, MAKE_Qs, SPH_T32, SPH_ROTL32, M, Qs, dH) |
|
|
|
#if SPH_64 |
|
|
|
#define FOLDb FOLD(sph_u64, MAKE_Qb, SPH_T64, SPH_ROTL64, M, Qb, dH) |
|
|
|
#endif |
|
|
|
static void |
|
compress_small(const unsigned char *data, const sph_u32 h[16], sph_u32 dh[16]) |
|
{ |
|
#if SPH_LITTLE_FAST |
|
#define M(x) sph_dec32le_aligned(data + 4 * (x)) |
|
#else |
|
sph_u32 mv[16]; |
|
|
|
mv[ 0] = sph_dec32le_aligned(data + 0); |
|
mv[ 1] = sph_dec32le_aligned(data + 4); |
|
mv[ 2] = sph_dec32le_aligned(data + 8); |
|
mv[ 3] = sph_dec32le_aligned(data + 12); |
|
mv[ 4] = sph_dec32le_aligned(data + 16); |
|
mv[ 5] = sph_dec32le_aligned(data + 20); |
|
mv[ 6] = sph_dec32le_aligned(data + 24); |
|
mv[ 7] = sph_dec32le_aligned(data + 28); |
|
mv[ 8] = sph_dec32le_aligned(data + 32); |
|
mv[ 9] = sph_dec32le_aligned(data + 36); |
|
mv[10] = sph_dec32le_aligned(data + 40); |
|
mv[11] = sph_dec32le_aligned(data + 44); |
|
mv[12] = sph_dec32le_aligned(data + 48); |
|
mv[13] = sph_dec32le_aligned(data + 52); |
|
mv[14] = sph_dec32le_aligned(data + 56); |
|
mv[15] = sph_dec32le_aligned(data + 60); |
|
#define M(x) (mv[x]) |
|
#endif |
|
#define H(x) (h[x]) |
|
#define dH(x) (dh[x]) |
|
|
|
FOLDs; |
|
|
|
#undef M |
|
#undef H |
|
#undef dH |
|
} |
|
|
|
static const sph_u32 final_s[16] = { |
|
SPH_C32(0xaaaaaaa0), SPH_C32(0xaaaaaaa1), SPH_C32(0xaaaaaaa2), |
|
SPH_C32(0xaaaaaaa3), SPH_C32(0xaaaaaaa4), SPH_C32(0xaaaaaaa5), |
|
SPH_C32(0xaaaaaaa6), SPH_C32(0xaaaaaaa7), SPH_C32(0xaaaaaaa8), |
|
SPH_C32(0xaaaaaaa9), SPH_C32(0xaaaaaaaa), SPH_C32(0xaaaaaaab), |
|
SPH_C32(0xaaaaaaac), SPH_C32(0xaaaaaaad), SPH_C32(0xaaaaaaae), |
|
SPH_C32(0xaaaaaaaf) |
|
}; |
|
|
|
static void |
|
bmw32_init(sph_bmw_small_context *sc, const sph_u32 *iv) |
|
{ |
|
memcpy(sc->H, iv, sizeof sc->H); |
|
sc->ptr = 0; |
|
#if SPH_64 |
|
sc->bit_count = 0; |
|
#else |
|
sc->bit_count_high = 0; |
|
sc->bit_count_low = 0; |
|
#endif |
|
} |
|
|
|
static void |
|
bmw32(sph_bmw_small_context *sc, const void *data, size_t len) |
|
{ |
|
unsigned char *buf; |
|
size_t ptr; |
|
sph_u32 htmp[16]; |
|
sph_u32 *h1, *h2; |
|
#if !SPH_64 |
|
sph_u32 tmp; |
|
#endif |
|
|
|
#if SPH_64 |
|
sc->bit_count += (sph_u64)len << 3; |
|
#else |
|
tmp = sc->bit_count_low; |
|
sc->bit_count_low = SPH_T32(tmp + ((sph_u32)len << 3)); |
|
if (sc->bit_count_low < tmp) |
|
sc->bit_count_high ++; |
|
sc->bit_count_high += len >> 29; |
|
#endif |
|
buf = sc->buf; |
|
ptr = sc->ptr; |
|
h1 = sc->H; |
|
h2 = htmp; |
|
while (len > 0) { |
|
size_t clen; |
|
|
|
clen = (sizeof sc->buf) - ptr; |
|
if (clen > len) |
|
clen = len; |
|
memcpy(buf + ptr, data, clen); |
|
data = (const unsigned char *)data + clen; |
|
len -= clen; |
|
ptr += clen; |
|
if (ptr == sizeof sc->buf) { |
|
sph_u32 *ht; |
|
|
|
compress_small(buf, h1, h2); |
|
ht = h1; |
|
h1 = h2; |
|
h2 = ht; |
|
ptr = 0; |
|
} |
|
} |
|
sc->ptr = ptr; |
|
if (h1 != sc->H) |
|
memcpy(sc->H, h1, sizeof sc->H); |
|
} |
|
|
|
static void |
|
bmw32_close(sph_bmw_small_context *sc, unsigned ub, unsigned n, |
|
void *dst, size_t out_size_w32) |
|
{ |
|
unsigned char *buf, *out; |
|
size_t ptr, u, v; |
|
unsigned z; |
|
sph_u32 h1[16], h2[16], *h; |
|
|
|
buf = sc->buf; |
|
ptr = sc->ptr; |
|
z = 0x80 >> n; |
|
buf[ptr ++] = ((ub & -z) | z) & 0xFF; |
|
h = sc->H; |
|
if (ptr > (sizeof sc->buf) - 8) { |
|
memset(buf + ptr, 0, (sizeof sc->buf) - ptr); |
|
compress_small(buf, h, h1); |
|
ptr = 0; |
|
h = h1; |
|
} |
|
memset(buf + ptr, 0, (sizeof sc->buf) - 8 - ptr); |
|
#if SPH_64 |
|
sph_enc64le_aligned(buf + (sizeof sc->buf) - 8, |
|
SPH_T64(sc->bit_count + n)); |
|
#else |
|
sph_enc32le_aligned(buf + (sizeof sc->buf) - 8, |
|
sc->bit_count_low + n); |
|
sph_enc32le_aligned(buf + (sizeof sc->buf) - 4, |
|
SPH_T32(sc->bit_count_high)); |
|
#endif |
|
compress_small(buf, h, h2); |
|
for (u = 0; u < 16; u ++) |
|
sph_enc32le_aligned(buf + 4 * u, h2[u]); |
|
compress_small(buf, final_s, h1); |
|
out = dst; |
|
for (u = 0, v = 16 - out_size_w32; u < out_size_w32; u ++, v ++) |
|
sph_enc32le(out + 4 * u, h1[v]); |
|
} |
|
|
|
#if SPH_64 |
|
|
|
static void |
|
compress_big(const unsigned char *data, const sph_u64 h[16], sph_u64 dh[16]) |
|
{ |
|
#if SPH_LITTLE_FAST |
|
#define M(x) sph_dec64le_aligned(data + 8 * (x)) |
|
#else |
|
sph_u64 mv[16]; |
|
|
|
mv[ 0] = sph_dec64le_aligned(data + 0); |
|
mv[ 1] = sph_dec64le_aligned(data + 8); |
|
mv[ 2] = sph_dec64le_aligned(data + 16); |
|
mv[ 3] = sph_dec64le_aligned(data + 24); |
|
mv[ 4] = sph_dec64le_aligned(data + 32); |
|
mv[ 5] = sph_dec64le_aligned(data + 40); |
|
mv[ 6] = sph_dec64le_aligned(data + 48); |
|
mv[ 7] = sph_dec64le_aligned(data + 56); |
|
mv[ 8] = sph_dec64le_aligned(data + 64); |
|
mv[ 9] = sph_dec64le_aligned(data + 72); |
|
mv[10] = sph_dec64le_aligned(data + 80); |
|
mv[11] = sph_dec64le_aligned(data + 88); |
|
mv[12] = sph_dec64le_aligned(data + 96); |
|
mv[13] = sph_dec64le_aligned(data + 104); |
|
mv[14] = sph_dec64le_aligned(data + 112); |
|
mv[15] = sph_dec64le_aligned(data + 120); |
|
#define M(x) (mv[x]) |
|
#endif |
|
#define H(x) (h[x]) |
|
#define dH(x) (dh[x]) |
|
|
|
FOLDb; |
|
|
|
#undef M |
|
#undef H |
|
#undef dH |
|
} |
|
|
|
static const sph_u64 final_b[16] = { |
|
SPH_C64(0xaaaaaaaaaaaaaaa0), SPH_C64(0xaaaaaaaaaaaaaaa1), |
|
SPH_C64(0xaaaaaaaaaaaaaaa2), SPH_C64(0xaaaaaaaaaaaaaaa3), |
|
SPH_C64(0xaaaaaaaaaaaaaaa4), SPH_C64(0xaaaaaaaaaaaaaaa5), |
|
SPH_C64(0xaaaaaaaaaaaaaaa6), SPH_C64(0xaaaaaaaaaaaaaaa7), |
|
SPH_C64(0xaaaaaaaaaaaaaaa8), SPH_C64(0xaaaaaaaaaaaaaaa9), |
|
SPH_C64(0xaaaaaaaaaaaaaaaa), SPH_C64(0xaaaaaaaaaaaaaaab), |
|
SPH_C64(0xaaaaaaaaaaaaaaac), SPH_C64(0xaaaaaaaaaaaaaaad), |
|
SPH_C64(0xaaaaaaaaaaaaaaae), SPH_C64(0xaaaaaaaaaaaaaaaf) |
|
}; |
|
|
|
static void |
|
bmw64_init(sph_bmw_big_context *sc, const sph_u64 *iv) |
|
{ |
|
memcpy(sc->H, iv, sizeof sc->H); |
|
sc->ptr = 0; |
|
sc->bit_count = 0; |
|
} |
|
|
|
static void |
|
bmw64(sph_bmw_big_context *sc, const void *data, size_t len) |
|
{ |
|
unsigned char *buf; |
|
size_t ptr; |
|
sph_u64 htmp[16]; |
|
sph_u64 *h1, *h2; |
|
|
|
sc->bit_count += (sph_u64)len << 3; |
|
buf = sc->buf; |
|
ptr = sc->ptr; |
|
h1 = sc->H; |
|
h2 = htmp; |
|
while (len > 0) { |
|
size_t clen; |
|
|
|
clen = (sizeof sc->buf) - ptr; |
|
if (clen > len) |
|
clen = len; |
|
memcpy(buf + ptr, data, clen); |
|
data = (const unsigned char *)data + clen; |
|
len -= clen; |
|
ptr += clen; |
|
if (ptr == sizeof sc->buf) { |
|
sph_u64 *ht; |
|
|
|
compress_big(buf, h1, h2); |
|
ht = h1; |
|
h1 = h2; |
|
h2 = ht; |
|
ptr = 0; |
|
} |
|
} |
|
sc->ptr = ptr; |
|
if (h1 != sc->H) |
|
memcpy(sc->H, h1, sizeof sc->H); |
|
} |
|
|
|
static void |
|
bmw64_close(sph_bmw_big_context *sc, unsigned ub, unsigned n, |
|
void *dst, size_t out_size_w64) |
|
{ |
|
unsigned char *buf, *out; |
|
size_t ptr, u, v; |
|
unsigned z; |
|
sph_u64 h1[16], h2[16], *h; |
|
|
|
buf = sc->buf; |
|
ptr = sc->ptr; |
|
z = 0x80 >> n; |
|
buf[ptr ++] = ((ub & -z) | z) & 0xFF; |
|
h = sc->H; |
|
if (ptr > (sizeof sc->buf) - 8) { |
|
memset(buf + ptr, 0, (sizeof sc->buf) - ptr); |
|
compress_big(buf, h, h1); |
|
ptr = 0; |
|
h = h1; |
|
} |
|
memset(buf + ptr, 0, (sizeof sc->buf) - 8 - ptr); |
|
sph_enc64le_aligned(buf + (sizeof sc->buf) - 8, |
|
SPH_T64(sc->bit_count + n)); |
|
compress_big(buf, h, h2); |
|
for (u = 0; u < 16; u ++) |
|
sph_enc64le_aligned(buf + 8 * u, h2[u]); |
|
compress_big(buf, final_b, h1); |
|
out = dst; |
|
for (u = 0, v = 16 - out_size_w64; u < out_size_w64; u ++, v ++) |
|
sph_enc64le(out + 8 * u, h1[v]); |
|
} |
|
|
|
#endif |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw224_init(void *cc) |
|
{ |
|
bmw32_init(cc, IV224); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw224(void *cc, const void *data, size_t len) |
|
{ |
|
bmw32(cc, data, len); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw224_close(void *cc, void *dst) |
|
{ |
|
sph_bmw224_addbits_and_close(cc, 0, 0, dst); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw224_addbits_and_close(void *cc, unsigned ub, unsigned n, void *dst) |
|
{ |
|
bmw32_close(cc, ub, n, dst, 7); |
|
sph_bmw224_init(cc); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw256_init(void *cc) |
|
{ |
|
bmw32_init(cc, IV256); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw256(void *cc, const void *data, size_t len) |
|
{ |
|
bmw32(cc, data, len); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw256_close(void *cc, void *dst) |
|
{ |
|
sph_bmw256_addbits_and_close(cc, 0, 0, dst); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw256_addbits_and_close(void *cc, unsigned ub, unsigned n, void *dst) |
|
{ |
|
bmw32_close(cc, ub, n, dst, 8); |
|
sph_bmw256_init(cc); |
|
} |
|
|
|
#if SPH_64 |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw384_init(void *cc) |
|
{ |
|
bmw64_init(cc, IV384); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw384(void *cc, const void *data, size_t len) |
|
{ |
|
bmw64(cc, data, len); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw384_close(void *cc, void *dst) |
|
{ |
|
sph_bmw384_addbits_and_close(cc, 0, 0, dst); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw384_addbits_and_close(void *cc, unsigned ub, unsigned n, void *dst) |
|
{ |
|
bmw64_close(cc, ub, n, dst, 6); |
|
sph_bmw384_init(cc); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw512_init(void *cc) |
|
{ |
|
bmw64_init(cc, IV512); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw512(void *cc, const void *data, size_t len) |
|
{ |
|
bmw64(cc, data, len); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw512_close(void *cc, void *dst) |
|
{ |
|
sph_bmw512_addbits_and_close(cc, 0, 0, dst); |
|
} |
|
|
|
/* see sph_bmw.h */ |
|
void |
|
sph_bmw512_addbits_and_close(void *cc, unsigned ub, unsigned n, void *dst) |
|
{ |
|
bmw64_close(cc, ub, n, dst, 8); |
|
sph_bmw512_init(cc); |
|
} |
|
|
|
#endif
|
|
|