|
| 1 | +/* |
| 2 | + * FIPS 180-2 256 implementation based on code by Oliver Gay |
| 3 | + * under a BSD-style license. http://ouah.org/ogay/sha2/ |
| 4 | + * |
| 5 | + * Copyright (C) 2005-2023 Olivier Gay <olivier.gay@a3.epfl.ch> |
| 6 | + * All rights reserved. |
| 7 | + * |
| 8 | + * Redistribution and use in source and binary forms, with or without |
| 9 | + * modification, are permitted provided that the following conditions |
| 10 | + * are met: |
| 11 | + * 1. Redistributions of source code must retain the above copyright |
| 12 | + * notice, this list of conditions and the following disclaimer. |
| 13 | + * 2. Redistributions in binary form must reproduce the above copyright |
| 14 | + * notice, this list of conditions and the following disclaimer in the |
| 15 | + * documentation and/or other materials provided with the distribution. |
| 16 | + * 3. Neither the name of the project nor the names of its contributors |
| 17 | + * may be used to endorse or promote products derived from this software |
| 18 | + * without specific prior written permission. |
| 19 | + * |
| 20 | + * THIS SOFTWARE IS PROVIDED BY THE PROJECT AND CONTRIBUTORS ``AS IS'' AND |
| 21 | + * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| 22 | + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| 23 | + * ARE DISCLAIMED. IN NO EVENT SHALL THE PROJECT OR CONTRIBUTORS BE LIABLE |
| 24 | + * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
| 25 | + * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS |
| 26 | + * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) |
| 27 | + * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT |
| 28 | + * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY |
| 29 | + * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF |
| 30 | + * SUCH DAMAGE. |
| 31 | + */ |
| 32 | + |
| 33 | +#if 0 |
| 34 | +#define UNROLL_LOOPS /* Enable loops unrolling */ |
| 35 | +#endif |
| 36 | + |
| 37 | +#include "sha256.h" |
| 38 | + |
| 39 | +#define SHFR(x, n) (x >> n) |
| 40 | +#define ROTR(x, n) ((x >> n) | (x << ((sizeof (x) << 3) - n))) |
| 41 | +#define ROTL(x, n) ((x << n) | (x >> ((sizeof (x) << 3) - n))) |
| 42 | +#define CH(x, y, z) ((x & y) ^ (~x & z)) |
| 43 | +#define MAJ(x, y, z) ((x & y) ^ (x & z) ^ (y & z)) |
| 44 | + |
| 45 | +#define SHA256_F1(x) (ROTR(x, 2) ^ ROTR(x, 13) ^ ROTR(x, 22)) |
| 46 | +#define SHA256_F2(x) (ROTR(x, 6) ^ ROTR(x, 11) ^ ROTR(x, 25)) |
| 47 | +#define SHA256_F3(x) (ROTR(x, 7) ^ ROTR(x, 18) ^ SHFR(x, 3)) |
| 48 | +#define SHA256_F4(x) (ROTR(x, 17) ^ ROTR(x, 19) ^ SHFR(x, 10)) |
| 49 | + |
| 50 | +#define UNPACK32(x, str) \ |
| 51 | +{ \ |
| 52 | + *((str) + 3) = (uint8_t) ((x) ); \ |
| 53 | + *((str) + 2) = (uint8_t) ((x) >> 8); \ |
| 54 | + *((str) + 1) = (uint8_t) ((x) >> 16); \ |
| 55 | + *((str) + 0) = (uint8_t) ((x) >> 24); \ |
| 56 | +} |
| 57 | + |
| 58 | +#define PACK32(str, x) \ |
| 59 | +{ \ |
| 60 | + *(x) = ((uint32_t) *((str) + 3) ) \ |
| 61 | + | ((uint32_t) *((str) + 2) << 8) \ |
| 62 | + | ((uint32_t) *((str) + 1) << 16) \ |
| 63 | + | ((uint32_t) *((str) + 0) << 24); \ |
| 64 | +} |
| 65 | + |
| 66 | +#define UNPACK64(x, str) \ |
| 67 | +{ \ |
| 68 | + *((str) + 7) = (uint8_t) ((x) ); \ |
| 69 | + *((str) + 6) = (uint8_t) ((x) >> 8); \ |
| 70 | + *((str) + 5) = (uint8_t) ((x) >> 16); \ |
| 71 | + *((str) + 4) = (uint8_t) ((x) >> 24); \ |
| 72 | + *((str) + 3) = (uint8_t) ((x) >> 32); \ |
| 73 | + *((str) + 2) = (uint8_t) ((x) >> 40); \ |
| 74 | + *((str) + 1) = (uint8_t) ((x) >> 48); \ |
| 75 | + *((str) + 0) = (uint8_t) ((x) >> 56); \ |
| 76 | +} |
| 77 | + |
| 78 | +#define PACK64(str, x) \ |
| 79 | +{ \ |
| 80 | + *(x) = ((uint64_t) *((str) + 7) ) \ |
| 81 | + | ((uint64_t) *((str) + 6) << 8) \ |
| 82 | + | ((uint64_t) *((str) + 5) << 16) \ |
| 83 | + | ((uint64_t) *((str) + 4) << 24) \ |
| 84 | + | ((uint64_t) *((str) + 3) << 32) \ |
| 85 | + | ((uint64_t) *((str) + 2) << 40) \ |
| 86 | + | ((uint64_t) *((str) + 1) << 48) \ |
| 87 | + | ((uint64_t) *((str) + 0) << 56); \ |
| 88 | +} |
| 89 | + |
| 90 | +/* Macros used for loops unrolling */ |
| 91 | + |
| 92 | +#define SHA256_SCR(i) \ |
| 93 | +{ \ |
| 94 | + w[i] = SHA256_F4(w[i - 2]) + w[i - 7] \ |
| 95 | + + SHA256_F3(w[i - 15]) + w[i - 16]; \ |
| 96 | +} |
| 97 | + |
| 98 | +#define SHA256_EXP(a, b, c, d, e, f, g, h, j) \ |
| 99 | +{ \ |
| 100 | + t1 = wv[h] + SHA256_F2(wv[e]) + CH(wv[e], wv[f], wv[g]) \ |
| 101 | + + sha256_k[j] + w[j]; \ |
| 102 | + t2 = SHA256_F1(wv[a]) + MAJ(wv[a], wv[b], wv[c]); \ |
| 103 | + wv[d] += t1; \ |
| 104 | + wv[h] = t1 + t2; \ |
| 105 | +} |
| 106 | + |
| 107 | +static const uint32_t sha256_h0[8] = |
| 108 | + {0x6a09e667, 0xbb67ae85, 0x3c6ef372, 0xa54ff53a, |
| 109 | + 0x510e527f, 0x9b05688c, 0x1f83d9ab, 0x5be0cd19}; |
| 110 | + |
| 111 | +static const uint32_t sha256_k[64] = |
| 112 | + {0x428a2f98, 0x71374491, 0xb5c0fbcf, 0xe9b5dba5, |
| 113 | + 0x3956c25b, 0x59f111f1, 0x923f82a4, 0xab1c5ed5, |
| 114 | + 0xd807aa98, 0x12835b01, 0x243185be, 0x550c7dc3, |
| 115 | + 0x72be5d74, 0x80deb1fe, 0x9bdc06a7, 0xc19bf174, |
| 116 | + 0xe49b69c1, 0xefbe4786, 0x0fc19dc6, 0x240ca1cc, |
| 117 | + 0x2de92c6f, 0x4a7484aa, 0x5cb0a9dc, 0x76f988da, |
| 118 | + 0x983e5152, 0xa831c66d, 0xb00327c8, 0xbf597fc7, |
| 119 | + 0xc6e00bf3, 0xd5a79147, 0x06ca6351, 0x14292967, |
| 120 | + 0x27b70a85, 0x2e1b2138, 0x4d2c6dfc, 0x53380d13, |
| 121 | + 0x650a7354, 0x766a0abb, 0x81c2c92e, 0x92722c85, |
| 122 | + 0xa2bfe8a1, 0xa81a664b, 0xc24b8b70, 0xc76c51a3, |
| 123 | + 0xd192e819, 0xd6990624, 0xf40e3585, 0x106aa070, |
| 124 | + 0x19a4c116, 0x1e376c08, 0x2748774c, 0x34b0bcb5, |
| 125 | + 0x391c0cb3, 0x4ed8aa4a, 0x5b9cca4f, 0x682e6ff3, |
| 126 | + 0x748f82ee, 0x78a5636f, 0x84c87814, 0x8cc70208, |
| 127 | + 0x90befffa, 0xa4506ceb, 0xbef9a3f7, 0xc67178f2}; |
| 128 | + |
| 129 | +/* SHA-2 internal function */ |
| 130 | + |
| 131 | +static void sha256_transf(sha256_ctx *ctx, const uint8_t *message, |
| 132 | + uint64_t block_nb) |
| 133 | +{ |
| 134 | + uint32_t w[64]; |
| 135 | + uint32_t wv[8]; |
| 136 | + uint32_t t1, t2; |
| 137 | + const uint8_t *sub_block; |
| 138 | + uint64_t i; |
| 139 | + |
| 140 | +#ifndef UNROLL_LOOPS |
| 141 | + int j; |
| 142 | +#endif |
| 143 | + |
| 144 | + for (i = 0; i < block_nb; i++) { |
| 145 | + sub_block = message + (i << 6); |
| 146 | + |
| 147 | +#ifndef UNROLL_LOOPS |
| 148 | + for (j = 0; j < 16; j++) { |
| 149 | + PACK32(&sub_block[j << 2], &w[j]); |
| 150 | + } |
| 151 | + |
| 152 | + for (j = 16; j < 64; j++) { |
| 153 | + SHA256_SCR(j); |
| 154 | + } |
| 155 | + |
| 156 | + for (j = 0; j < 8; j++) { |
| 157 | + wv[j] = ctx->h[j]; |
| 158 | + } |
| 159 | + |
| 160 | + for (j = 0; j < 64; j++) { |
| 161 | + t1 = wv[7] + SHA256_F2(wv[4]) + CH(wv[4], wv[5], wv[6]) |
| 162 | + + sha256_k[j] + w[j]; |
| 163 | + t2 = SHA256_F1(wv[0]) + MAJ(wv[0], wv[1], wv[2]); |
| 164 | + wv[7] = wv[6]; |
| 165 | + wv[6] = wv[5]; |
| 166 | + wv[5] = wv[4]; |
| 167 | + wv[4] = wv[3] + t1; |
| 168 | + wv[3] = wv[2]; |
| 169 | + wv[2] = wv[1]; |
| 170 | + wv[1] = wv[0]; |
| 171 | + wv[0] = t1 + t2; |
| 172 | + } |
| 173 | + |
| 174 | + for (j = 0; j < 8; j++) { |
| 175 | + ctx->h[j] += wv[j]; |
| 176 | + } |
| 177 | +#else |
| 178 | + PACK32(&sub_block[ 0], &w[ 0]); PACK32(&sub_block[ 4], &w[ 1]); |
| 179 | + PACK32(&sub_block[ 8], &w[ 2]); PACK32(&sub_block[12], &w[ 3]); |
| 180 | + PACK32(&sub_block[16], &w[ 4]); PACK32(&sub_block[20], &w[ 5]); |
| 181 | + PACK32(&sub_block[24], &w[ 6]); PACK32(&sub_block[28], &w[ 7]); |
| 182 | + PACK32(&sub_block[32], &w[ 8]); PACK32(&sub_block[36], &w[ 9]); |
| 183 | + PACK32(&sub_block[40], &w[10]); PACK32(&sub_block[44], &w[11]); |
| 184 | + PACK32(&sub_block[48], &w[12]); PACK32(&sub_block[52], &w[13]); |
| 185 | + PACK32(&sub_block[56], &w[14]); PACK32(&sub_block[60], &w[15]); |
| 186 | + |
| 187 | + SHA256_SCR(16); SHA256_SCR(17); SHA256_SCR(18); SHA256_SCR(19); |
| 188 | + SHA256_SCR(20); SHA256_SCR(21); SHA256_SCR(22); SHA256_SCR(23); |
| 189 | + SHA256_SCR(24); SHA256_SCR(25); SHA256_SCR(26); SHA256_SCR(27); |
| 190 | + SHA256_SCR(28); SHA256_SCR(29); SHA256_SCR(30); SHA256_SCR(31); |
| 191 | + SHA256_SCR(32); SHA256_SCR(33); SHA256_SCR(34); SHA256_SCR(35); |
| 192 | + SHA256_SCR(36); SHA256_SCR(37); SHA256_SCR(38); SHA256_SCR(39); |
| 193 | + SHA256_SCR(40); SHA256_SCR(41); SHA256_SCR(42); SHA256_SCR(43); |
| 194 | + SHA256_SCR(44); SHA256_SCR(45); SHA256_SCR(46); SHA256_SCR(47); |
| 195 | + SHA256_SCR(48); SHA256_SCR(49); SHA256_SCR(50); SHA256_SCR(51); |
| 196 | + SHA256_SCR(52); SHA256_SCR(53); SHA256_SCR(54); SHA256_SCR(55); |
| 197 | + SHA256_SCR(56); SHA256_SCR(57); SHA256_SCR(58); SHA256_SCR(59); |
| 198 | + SHA256_SCR(60); SHA256_SCR(61); SHA256_SCR(62); SHA256_SCR(63); |
| 199 | + |
| 200 | + wv[0] = ctx->h[0]; wv[1] = ctx->h[1]; |
| 201 | + wv[2] = ctx->h[2]; wv[3] = ctx->h[3]; |
| 202 | + wv[4] = ctx->h[4]; wv[5] = ctx->h[5]; |
| 203 | + wv[6] = ctx->h[6]; wv[7] = ctx->h[7]; |
| 204 | + |
| 205 | + SHA256_EXP(0,1,2,3,4,5,6,7, 0); SHA256_EXP(7,0,1,2,3,4,5,6, 1); |
| 206 | + SHA256_EXP(6,7,0,1,2,3,4,5, 2); SHA256_EXP(5,6,7,0,1,2,3,4, 3); |
| 207 | + SHA256_EXP(4,5,6,7,0,1,2,3, 4); SHA256_EXP(3,4,5,6,7,0,1,2, 5); |
| 208 | + SHA256_EXP(2,3,4,5,6,7,0,1, 6); SHA256_EXP(1,2,3,4,5,6,7,0, 7); |
| 209 | + SHA256_EXP(0,1,2,3,4,5,6,7, 8); SHA256_EXP(7,0,1,2,3,4,5,6, 9); |
| 210 | + SHA256_EXP(6,7,0,1,2,3,4,5,10); SHA256_EXP(5,6,7,0,1,2,3,4,11); |
| 211 | + SHA256_EXP(4,5,6,7,0,1,2,3,12); SHA256_EXP(3,4,5,6,7,0,1,2,13); |
| 212 | + SHA256_EXP(2,3,4,5,6,7,0,1,14); SHA256_EXP(1,2,3,4,5,6,7,0,15); |
| 213 | + SHA256_EXP(0,1,2,3,4,5,6,7,16); SHA256_EXP(7,0,1,2,3,4,5,6,17); |
| 214 | + SHA256_EXP(6,7,0,1,2,3,4,5,18); SHA256_EXP(5,6,7,0,1,2,3,4,19); |
| 215 | + SHA256_EXP(4,5,6,7,0,1,2,3,20); SHA256_EXP(3,4,5,6,7,0,1,2,21); |
| 216 | + SHA256_EXP(2,3,4,5,6,7,0,1,22); SHA256_EXP(1,2,3,4,5,6,7,0,23); |
| 217 | + SHA256_EXP(0,1,2,3,4,5,6,7,24); SHA256_EXP(7,0,1,2,3,4,5,6,25); |
| 218 | + SHA256_EXP(6,7,0,1,2,3,4,5,26); SHA256_EXP(5,6,7,0,1,2,3,4,27); |
| 219 | + SHA256_EXP(4,5,6,7,0,1,2,3,28); SHA256_EXP(3,4,5,6,7,0,1,2,29); |
| 220 | + SHA256_EXP(2,3,4,5,6,7,0,1,30); SHA256_EXP(1,2,3,4,5,6,7,0,31); |
| 221 | + SHA256_EXP(0,1,2,3,4,5,6,7,32); SHA256_EXP(7,0,1,2,3,4,5,6,33); |
| 222 | + SHA256_EXP(6,7,0,1,2,3,4,5,34); SHA256_EXP(5,6,7,0,1,2,3,4,35); |
| 223 | + SHA256_EXP(4,5,6,7,0,1,2,3,36); SHA256_EXP(3,4,5,6,7,0,1,2,37); |
| 224 | + SHA256_EXP(2,3,4,5,6,7,0,1,38); SHA256_EXP(1,2,3,4,5,6,7,0,39); |
| 225 | + SHA256_EXP(0,1,2,3,4,5,6,7,40); SHA256_EXP(7,0,1,2,3,4,5,6,41); |
| 226 | + SHA256_EXP(6,7,0,1,2,3,4,5,42); SHA256_EXP(5,6,7,0,1,2,3,4,43); |
| 227 | + SHA256_EXP(4,5,6,7,0,1,2,3,44); SHA256_EXP(3,4,5,6,7,0,1,2,45); |
| 228 | + SHA256_EXP(2,3,4,5,6,7,0,1,46); SHA256_EXP(1,2,3,4,5,6,7,0,47); |
| 229 | + SHA256_EXP(0,1,2,3,4,5,6,7,48); SHA256_EXP(7,0,1,2,3,4,5,6,49); |
| 230 | + SHA256_EXP(6,7,0,1,2,3,4,5,50); SHA256_EXP(5,6,7,0,1,2,3,4,51); |
| 231 | + SHA256_EXP(4,5,6,7,0,1,2,3,52); SHA256_EXP(3,4,5,6,7,0,1,2,53); |
| 232 | + SHA256_EXP(2,3,4,5,6,7,0,1,54); SHA256_EXP(1,2,3,4,5,6,7,0,55); |
| 233 | + SHA256_EXP(0,1,2,3,4,5,6,7,56); SHA256_EXP(7,0,1,2,3,4,5,6,57); |
| 234 | + SHA256_EXP(6,7,0,1,2,3,4,5,58); SHA256_EXP(5,6,7,0,1,2,3,4,59); |
| 235 | + SHA256_EXP(4,5,6,7,0,1,2,3,60); SHA256_EXP(3,4,5,6,7,0,1,2,61); |
| 236 | + SHA256_EXP(2,3,4,5,6,7,0,1,62); SHA256_EXP(1,2,3,4,5,6,7,0,63); |
| 237 | + |
| 238 | + ctx->h[0] += wv[0]; ctx->h[1] += wv[1]; |
| 239 | + ctx->h[2] += wv[2]; ctx->h[3] += wv[3]; |
| 240 | + ctx->h[4] += wv[4]; ctx->h[5] += wv[5]; |
| 241 | + ctx->h[6] += wv[6]; ctx->h[7] += wv[7]; |
| 242 | +#endif /* !UNROLL_LOOPS */ |
| 243 | + } |
| 244 | +} |
| 245 | + |
| 246 | +/* SHA-256 functions */ |
| 247 | + |
| 248 | +void sha256(const uint8_t *message, uint64_t len, uint8_t *digest) |
| 249 | +{ |
| 250 | + sha256_ctx ctx; |
| 251 | + |
| 252 | + sha256_init(&ctx); |
| 253 | + sha256_update(&ctx, message, len); |
| 254 | + sha256_final(&ctx, digest); |
| 255 | +} |
| 256 | + |
| 257 | +void sha256_init(sha256_ctx *ctx) |
| 258 | +{ |
| 259 | +#ifndef UNROLL_LOOPS |
| 260 | + int i; |
| 261 | + for (i = 0; i < 8; i++) { |
| 262 | + ctx->h[i] = sha256_h0[i]; |
| 263 | + } |
| 264 | +#else |
| 265 | + ctx->h[0] = sha256_h0[0]; ctx->h[1] = sha256_h0[1]; |
| 266 | + ctx->h[2] = sha256_h0[2]; ctx->h[3] = sha256_h0[3]; |
| 267 | + ctx->h[4] = sha256_h0[4]; ctx->h[5] = sha256_h0[5]; |
| 268 | + ctx->h[6] = sha256_h0[6]; ctx->h[7] = sha256_h0[7]; |
| 269 | +#endif /* !UNROLL_LOOPS */ |
| 270 | + |
| 271 | + ctx->len = 0; |
| 272 | + ctx->tot_len = 0; |
| 273 | +} |
| 274 | + |
| 275 | +void sha256_update(sha256_ctx *ctx, const uint8_t *message, uint64_t len) |
| 276 | +{ |
| 277 | + uint64_t block_nb; |
| 278 | + uint64_t new_len, rem_len, tmp_len; |
| 279 | + const uint8_t *shifted_message; |
| 280 | + |
| 281 | + tmp_len = SHA256_BLOCK_SIZE - ctx->len; |
| 282 | + rem_len = len < tmp_len ? len : tmp_len; |
| 283 | + |
| 284 | + memcpy(&ctx->block[ctx->len], message, rem_len); |
| 285 | + |
| 286 | + if (ctx->len + len < SHA256_BLOCK_SIZE) { |
| 287 | + ctx->len += len; |
| 288 | + return; |
| 289 | + } |
| 290 | + |
| 291 | + new_len = len - rem_len; |
| 292 | + block_nb = new_len / SHA256_BLOCK_SIZE; |
| 293 | + |
| 294 | + shifted_message = message + rem_len; |
| 295 | + |
| 296 | + sha256_transf(ctx, ctx->block, 1); |
| 297 | + sha256_transf(ctx, shifted_message, block_nb); |
| 298 | + |
| 299 | + rem_len = new_len % SHA256_BLOCK_SIZE; |
| 300 | + |
| 301 | + memcpy(ctx->block, &shifted_message[block_nb << 6], rem_len); |
| 302 | + |
| 303 | + ctx->len = rem_len; |
| 304 | + ctx->tot_len += (block_nb + 1) << 6; |
| 305 | +} |
| 306 | + |
| 307 | +void sha256_final(sha256_ctx *ctx, uint8_t *digest) |
| 308 | +{ |
| 309 | + uint64_t block_nb; |
| 310 | + uint64_t pm_len; |
| 311 | + uint64_t len_b; |
| 312 | + uint64_t tot_len; |
| 313 | + |
| 314 | +#ifndef UNROLL_LOOPS |
| 315 | + int i; |
| 316 | +#endif |
| 317 | + |
| 318 | + block_nb = (1 + ((SHA256_BLOCK_SIZE - 9) |
| 319 | + < (ctx->len % SHA256_BLOCK_SIZE))); |
| 320 | + |
| 321 | + tot_len = ctx->tot_len + ctx->len; |
| 322 | + ctx->tot_len = tot_len; |
| 323 | + |
| 324 | + len_b = tot_len << 3; |
| 325 | + pm_len = block_nb << 6; |
| 326 | + |
| 327 | + memset(ctx->block + ctx->len, 0, pm_len - ctx->len); |
| 328 | + ctx->block[ctx->len] = 0x80; |
| 329 | + UNPACK64(len_b, ctx->block + pm_len - 8); |
| 330 | + |
| 331 | + sha256_transf(ctx, ctx->block, block_nb); |
| 332 | + |
| 333 | +#ifndef UNROLL_LOOPS |
| 334 | + for (i = 0 ; i < 8; i++) { |
| 335 | + UNPACK32(ctx->h[i], &digest[i << 2]); |
| 336 | + } |
| 337 | +#else |
| 338 | + UNPACK32(ctx->h[0], &digest[ 0]); |
| 339 | + UNPACK32(ctx->h[1], &digest[ 4]); |
| 340 | + UNPACK32(ctx->h[2], &digest[ 8]); |
| 341 | + UNPACK32(ctx->h[3], &digest[12]); |
| 342 | + UNPACK32(ctx->h[4], &digest[16]); |
| 343 | + UNPACK32(ctx->h[5], &digest[20]); |
| 344 | + UNPACK32(ctx->h[6], &digest[24]); |
| 345 | + UNPACK32(ctx->h[7], &digest[28]); |
| 346 | +#endif /* !UNROLL_LOOPS */ |
| 347 | +} |
0 commit comments