2 * Copyright (C) 2007 Michael Niedermayer <michaelni@gmx.at>
3 * Copyright (C) 2009 Konstantin Shishkov
4 * based on public domain SHA-1 code by Steve Reid <steve@edmweb.com>
5 * and on BSD-licensed SHA-2 code by Aaron D. Gifford
7 * This file is part of Libav.
9 * Libav is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU Lesser General Public
11 * License as published by the Free Software Foundation; either
12 * version 2.1 of the License, or (at your option) any later version.
14 * Libav is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17 * Lesser General Public License for more details.
19 * You should have received a copy of the GNU Lesser General Public
20 * License along with Libav; if not, write to the Free Software
21 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
26 #include "attributes.h"
30 #include "intreadwrite.h"
34 typedef struct AVSHA {
35 uint8_t digest_len; ///< digest length in 32-bit words
36 uint64_t count; ///< number of bytes in buffer
37 uint8_t buffer[64]; ///< 512-bit buffer of input values used in hash updating
38 uint32_t state[8]; ///< current hash value
39 /** function used to update hash for 512-bit input block */
40 void (*transform)(uint32_t *state, const uint8_t buffer[64]);
43 struct AVSHA *av_sha_alloc(void)
45 return av_mallocz(sizeof(struct AVSHA));
48 #define rol(value, bits) (((value) << (bits)) | ((value) >> (32 - (bits))))
50 /* (R0+R1), R2, R3, R4 are the different operations used in SHA1 */
51 #define blk0(i) (block[i] = AV_RB32(buffer + 4 * (i)))
52 #define blk(i) (block[i] = rol(block[i-3] ^ block[i-8] ^ block[i-14] ^ block[i-16], 1))
54 #define R0(v,w,x,y,z,i) z += ((w&(x^y))^y) + blk0(i) + 0x5A827999 + rol(v, 5); w = rol(w, 30);
55 #define R1(v,w,x,y,z,i) z += ((w&(x^y))^y) + blk (i) + 0x5A827999 + rol(v, 5); w = rol(w, 30);
56 #define R2(v,w,x,y,z,i) z += ( w^x ^y) + blk (i) + 0x6ED9EBA1 + rol(v, 5); w = rol(w, 30);
57 #define R3(v,w,x,y,z,i) z += (((w|x)&y)|(w&x)) + blk (i) + 0x8F1BBCDC + rol(v, 5); w = rol(w, 30);
58 #define R4(v,w,x,y,z,i) z += ( w^x ^y) + blk (i) + 0xCA62C1D6 + rol(v, 5); w = rol(w, 30);
60 /* Hash a single 512-bit block. This is the core of the algorithm. */
62 static void sha1_transform(uint32_t state[5], const uint8_t buffer[64])
65 unsigned int i, a, b, c, d, e;
73 for (i = 0; i < 80; i++) {
76 t = AV_RB32(buffer + 4 * i);
78 t = rol(block[i-3] ^ block[i-8] ^ block[i-14] ^ block[i-16], 1);
83 t += ((b&(c^d))^d) + 0x5A827999;
85 t += ( b^c ^d) + 0x6ED9EBA1;
88 t += (((b|c)&d)|(b&c)) + 0x8F1BBCDC;
90 t += ( b^c ^d) + 0xCA62C1D6;
99 for (i = 0; i < 15; i += 5) {
100 R0(a, b, c, d, e, 0 + i);
101 R0(e, a, b, c, d, 1 + i);
102 R0(d, e, a, b, c, 2 + i);
103 R0(c, d, e, a, b, 3 + i);
104 R0(b, c, d, e, a, 4 + i);
106 R0(a, b, c, d, e, 15);
107 R1(e, a, b, c, d, 16);
108 R1(d, e, a, b, c, 17);
109 R1(c, d, e, a, b, 18);
110 R1(b, c, d, e, a, 19);
111 for (i = 20; i < 40; i += 5) {
112 R2(a, b, c, d, e, 0 + i);
113 R2(e, a, b, c, d, 1 + i);
114 R2(d, e, a, b, c, 2 + i);
115 R2(c, d, e, a, b, 3 + i);
116 R2(b, c, d, e, a, 4 + i);
118 for (; i < 60; i += 5) {
119 R3(a, b, c, d, e, 0 + i);
120 R3(e, a, b, c, d, 1 + i);
121 R3(d, e, a, b, c, 2 + i);
122 R3(c, d, e, a, b, 3 + i);
123 R3(b, c, d, e, a, 4 + i);
125 for (; i < 80; i += 5) {
126 R4(a, b, c, d, e, 0 + i);
127 R4(e, a, b, c, d, 1 + i);
128 R4(d, e, a, b, c, 2 + i);
129 R4(c, d, e, a, b, 3 + i);
130 R4(b, c, d, e, a, 4 + i);
140 static const uint32_t K256[64] = {
141 0x428a2f98, 0x71374491, 0xb5c0fbcf, 0xe9b5dba5,
142 0x3956c25b, 0x59f111f1, 0x923f82a4, 0xab1c5ed5,
143 0xd807aa98, 0x12835b01, 0x243185be, 0x550c7dc3,
144 0x72be5d74, 0x80deb1fe, 0x9bdc06a7, 0xc19bf174,
145 0xe49b69c1, 0xefbe4786, 0x0fc19dc6, 0x240ca1cc,
146 0x2de92c6f, 0x4a7484aa, 0x5cb0a9dc, 0x76f988da,
147 0x983e5152, 0xa831c66d, 0xb00327c8, 0xbf597fc7,
148 0xc6e00bf3, 0xd5a79147, 0x06ca6351, 0x14292967,
149 0x27b70a85, 0x2e1b2138, 0x4d2c6dfc, 0x53380d13,
150 0x650a7354, 0x766a0abb, 0x81c2c92e, 0x92722c85,
151 0xa2bfe8a1, 0xa81a664b, 0xc24b8b70, 0xc76c51a3,
152 0xd192e819, 0xd6990624, 0xf40e3585, 0x106aa070,
153 0x19a4c116, 0x1e376c08, 0x2748774c, 0x34b0bcb5,
154 0x391c0cb3, 0x4ed8aa4a, 0x5b9cca4f, 0x682e6ff3,
155 0x748f82ee, 0x78a5636f, 0x84c87814, 0x8cc70208,
156 0x90befffa, 0xa4506ceb, 0xbef9a3f7, 0xc67178f2
160 #define Ch(x,y,z) (((x) & ((y) ^ (z))) ^ (z))
161 #define Maj(x,y,z) ((((x) | (y)) & (z)) | ((x) & (y)))
163 #define Sigma0_256(x) (rol((x), 30) ^ rol((x), 19) ^ rol((x), 10))
164 #define Sigma1_256(x) (rol((x), 26) ^ rol((x), 21) ^ rol((x), 7))
165 #define sigma0_256(x) (rol((x), 25) ^ rol((x), 14) ^ ((x) >> 3))
166 #define sigma1_256(x) (rol((x), 15) ^ rol((x), 13) ^ ((x) >> 10))
169 #define blk(i) (block[i] = block[i - 16] + sigma0_256(block[i - 15]) + \
170 sigma1_256(block[i - 2]) + block[i - 7])
172 #define ROUND256(a,b,c,d,e,f,g,h) \
173 T1 += (h) + Sigma1_256(e) + Ch((e), (f), (g)) + K256[i]; \
175 (h) = T1 + Sigma0_256(a) + Maj((a), (b), (c)); \
178 #define ROUND256_0_TO_15(a,b,c,d,e,f,g,h) \
180 ROUND256(a,b,c,d,e,f,g,h)
182 #define ROUND256_16_TO_63(a,b,c,d,e,f,g,h) \
184 ROUND256(a,b,c,d,e,f,g,h)
186 static void sha256_transform(uint32_t *state, const uint8_t buffer[64])
188 unsigned int i, a, b, c, d, e, f, g, h;
201 for (i = 0; i < 64; i++) {
207 T1 += h + Sigma1_256(e) + Ch(e, f, g) + K256[i];
208 T2 = Sigma0_256(a) + Maj(a, b, c);
219 for (i = 0; i < 16;) {
220 ROUND256_0_TO_15(a, b, c, d, e, f, g, h);
221 ROUND256_0_TO_15(h, a, b, c, d, e, f, g);
222 ROUND256_0_TO_15(g, h, a, b, c, d, e, f);
223 ROUND256_0_TO_15(f, g, h, a, b, c, d, e);
224 ROUND256_0_TO_15(e, f, g, h, a, b, c, d);
225 ROUND256_0_TO_15(d, e, f, g, h, a, b, c);
226 ROUND256_0_TO_15(c, d, e, f, g, h, a, b);
227 ROUND256_0_TO_15(b, c, d, e, f, g, h, a);
231 ROUND256_16_TO_63(a, b, c, d, e, f, g, h);
232 ROUND256_16_TO_63(h, a, b, c, d, e, f, g);
233 ROUND256_16_TO_63(g, h, a, b, c, d, e, f);
234 ROUND256_16_TO_63(f, g, h, a, b, c, d, e);
235 ROUND256_16_TO_63(e, f, g, h, a, b, c, d);
236 ROUND256_16_TO_63(d, e, f, g, h, a, b, c);
237 ROUND256_16_TO_63(c, d, e, f, g, h, a, b);
238 ROUND256_16_TO_63(b, c, d, e, f, g, h, a);
252 av_cold int av_sha_init(AVSHA *ctx, int bits)
254 ctx->digest_len = bits >> 5;
257 ctx->state[0] = 0x67452301;
258 ctx->state[1] = 0xEFCDAB89;
259 ctx->state[2] = 0x98BADCFE;
260 ctx->state[3] = 0x10325476;
261 ctx->state[4] = 0xC3D2E1F0;
262 ctx->transform = sha1_transform;
265 ctx->state[0] = 0xC1059ED8;
266 ctx->state[1] = 0x367CD507;
267 ctx->state[2] = 0x3070DD17;
268 ctx->state[3] = 0xF70E5939;
269 ctx->state[4] = 0xFFC00B31;
270 ctx->state[5] = 0x68581511;
271 ctx->state[6] = 0x64F98FA7;
272 ctx->state[7] = 0xBEFA4FA4;
273 ctx->transform = sha256_transform;
276 ctx->state[0] = 0x6A09E667;
277 ctx->state[1] = 0xBB67AE85;
278 ctx->state[2] = 0x3C6EF372;
279 ctx->state[3] = 0xA54FF53A;
280 ctx->state[4] = 0x510E527F;
281 ctx->state[5] = 0x9B05688C;
282 ctx->state[6] = 0x1F83D9AB;
283 ctx->state[7] = 0x5BE0CD19;
284 ctx->transform = sha256_transform;
293 void av_sha_update(AVSHA* ctx, const uint8_t* data, unsigned int len)
300 for (i = 0; i < len; i++) {
301 ctx->buffer[j++] = data[i];
303 ctx->transform(ctx->state, ctx->buffer);
308 if ((j + len) > 63) {
309 memcpy(&ctx->buffer[j], data, (i = 64 - j));
310 ctx->transform(ctx->state, ctx->buffer);
311 for (; i + 63 < len; i += 64)
312 ctx->transform(ctx->state, &data[i]);
316 memcpy(&ctx->buffer[j], &data[i], len - i);
320 void av_sha_final(AVSHA* ctx, uint8_t *digest)
323 uint64_t finalcount = av_be2ne64(ctx->count << 3);
325 av_sha_update(ctx, "\200", 1);
326 while ((ctx->count & 63) != 56)
327 av_sha_update(ctx, "", 1);
328 av_sha_update(ctx, (uint8_t *)&finalcount, 8); /* Should cause a transform() */
329 for (i = 0; i < ctx->digest_len; i++)
330 AV_WB32(digest + i*4, ctx->state[i]);