linux/crypto/sha256_generic.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-or-later
   2/*
   3 * Cryptographic API.
   4 *
   5 * SHA-256, as specified in
   6 * http://csrc.nist.gov/groups/STM/cavp/documents/shs/sha256-384-512.pdf
   7 *
   8 * SHA-256 code by Jean-Luc Cooke <jlcooke@certainkey.com>.
   9 *
  10 * Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com>
  11 * Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk>
  12 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
  13 * SHA224 Support Copyright 2007 Intel Corporation <jonathan.lynch@intel.com>
  14 */
  15#include <crypto/internal/hash.h>
  16#include <linux/init.h>
  17#include <linux/module.h>
  18#include <linux/mm.h>
  19#include <linux/types.h>
  20#include <crypto/sha.h>
  21#include <crypto/sha256_base.h>
  22#include <asm/byteorder.h>
  23#include <asm/unaligned.h>
  24
  25const u8 sha224_zero_message_hash[SHA224_DIGEST_SIZE] = {
  26        0xd1, 0x4a, 0x02, 0x8c, 0x2a, 0x3a, 0x2b, 0xc9, 0x47,
  27        0x61, 0x02, 0xbb, 0x28, 0x82, 0x34, 0xc4, 0x15, 0xa2,
  28        0xb0, 0x1f, 0x82, 0x8e, 0xa6, 0x2a, 0xc5, 0xb3, 0xe4,
  29        0x2f
  30};
  31EXPORT_SYMBOL_GPL(sha224_zero_message_hash);
  32
  33const u8 sha256_zero_message_hash[SHA256_DIGEST_SIZE] = {
  34        0xe3, 0xb0, 0xc4, 0x42, 0x98, 0xfc, 0x1c, 0x14,
  35        0x9a, 0xfb, 0xf4, 0xc8, 0x99, 0x6f, 0xb9, 0x24,
  36        0x27, 0xae, 0x41, 0xe4, 0x64, 0x9b, 0x93, 0x4c,
  37        0xa4, 0x95, 0x99, 0x1b, 0x78, 0x52, 0xb8, 0x55
  38};
  39EXPORT_SYMBOL_GPL(sha256_zero_message_hash);
  40
  41static inline u32 Ch(u32 x, u32 y, u32 z)
  42{
  43        return z ^ (x & (y ^ z));
  44}
  45
  46static inline u32 Maj(u32 x, u32 y, u32 z)
  47{
  48        return (x & y) | (z & (x | y));
  49}
  50
  51#define e0(x)       (ror32(x, 2) ^ ror32(x,13) ^ ror32(x,22))
  52#define e1(x)       (ror32(x, 6) ^ ror32(x,11) ^ ror32(x,25))
  53#define s0(x)       (ror32(x, 7) ^ ror32(x,18) ^ (x >> 3))
  54#define s1(x)       (ror32(x,17) ^ ror32(x,19) ^ (x >> 10))
  55
  56static inline void LOAD_OP(int I, u32 *W, const u8 *input)
  57{
  58        W[I] = get_unaligned_be32((__u32 *)input + I);
  59}
  60
  61static inline void BLEND_OP(int I, u32 *W)
  62{
  63        W[I] = s1(W[I-2]) + W[I-7] + s0(W[I-15]) + W[I-16];
  64}
  65
  66static void sha256_transform(u32 *state, const u8 *input)
  67{
  68        u32 a, b, c, d, e, f, g, h, t1, t2;
  69        u32 W[64];
  70        int i;
  71
  72        /* load the input */
  73        for (i = 0; i < 16; i++)
  74                LOAD_OP(i, W, input);
  75
  76        /* now blend */
  77        for (i = 16; i < 64; i++)
  78                BLEND_OP(i, W);
  79
  80        /* load the state into our registers */
  81        a=state[0];  b=state[1];  c=state[2];  d=state[3];
  82        e=state[4];  f=state[5];  g=state[6];  h=state[7];
  83
  84        /* now iterate */
  85        t1 = h + e1(e) + Ch(e,f,g) + 0x428a2f98 + W[ 0];
  86        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
  87        t1 = g + e1(d) + Ch(d,e,f) + 0x71374491 + W[ 1];
  88        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
  89        t1 = f + e1(c) + Ch(c,d,e) + 0xb5c0fbcf + W[ 2];
  90        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
  91        t1 = e + e1(b) + Ch(b,c,d) + 0xe9b5dba5 + W[ 3];
  92        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
  93        t1 = d + e1(a) + Ch(a,b,c) + 0x3956c25b + W[ 4];
  94        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
  95        t1 = c + e1(h) + Ch(h,a,b) + 0x59f111f1 + W[ 5];
  96        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
  97        t1 = b + e1(g) + Ch(g,h,a) + 0x923f82a4 + W[ 6];
  98        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
  99        t1 = a + e1(f) + Ch(f,g,h) + 0xab1c5ed5 + W[ 7];
 100        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 101
 102        t1 = h + e1(e) + Ch(e,f,g) + 0xd807aa98 + W[ 8];
 103        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
 104        t1 = g + e1(d) + Ch(d,e,f) + 0x12835b01 + W[ 9];
 105        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
 106        t1 = f + e1(c) + Ch(c,d,e) + 0x243185be + W[10];
 107        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
 108        t1 = e + e1(b) + Ch(b,c,d) + 0x550c7dc3 + W[11];
 109        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
 110        t1 = d + e1(a) + Ch(a,b,c) + 0x72be5d74 + W[12];
 111        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
 112        t1 = c + e1(h) + Ch(h,a,b) + 0x80deb1fe + W[13];
 113        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
 114        t1 = b + e1(g) + Ch(g,h,a) + 0x9bdc06a7 + W[14];
 115        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
 116        t1 = a + e1(f) + Ch(f,g,h) + 0xc19bf174 + W[15];
 117        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 118
 119        t1 = h + e1(e) + Ch(e,f,g) + 0xe49b69c1 + W[16];
 120        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
 121        t1 = g + e1(d) + Ch(d,e,f) + 0xefbe4786 + W[17];
 122        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
 123        t1 = f + e1(c) + Ch(c,d,e) + 0x0fc19dc6 + W[18];
 124        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
 125        t1 = e + e1(b) + Ch(b,c,d) + 0x240ca1cc + W[19];
 126        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
 127        t1 = d + e1(a) + Ch(a,b,c) + 0x2de92c6f + W[20];
 128        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
 129        t1 = c + e1(h) + Ch(h,a,b) + 0x4a7484aa + W[21];
 130        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
 131        t1 = b + e1(g) + Ch(g,h,a) + 0x5cb0a9dc + W[22];
 132        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
 133        t1 = a + e1(f) + Ch(f,g,h) + 0x76f988da + W[23];
 134        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 135
 136        t1 = h + e1(e) + Ch(e,f,g) + 0x983e5152 + W[24];
 137        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
 138        t1 = g + e1(d) + Ch(d,e,f) + 0xa831c66d + W[25];
 139        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
 140        t1 = f + e1(c) + Ch(c,d,e) + 0xb00327c8 + W[26];
 141        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
 142        t1 = e + e1(b) + Ch(b,c,d) + 0xbf597fc7 + W[27];
 143        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
 144        t1 = d + e1(a) + Ch(a,b,c) + 0xc6e00bf3 + W[28];
 145        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
 146        t1 = c + e1(h) + Ch(h,a,b) + 0xd5a79147 + W[29];
 147        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
 148        t1 = b + e1(g) + Ch(g,h,a) + 0x06ca6351 + W[30];
 149        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
 150        t1 = a + e1(f) + Ch(f,g,h) + 0x14292967 + W[31];
 151        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 152
 153        t1 = h + e1(e) + Ch(e,f,g) + 0x27b70a85 + W[32];
 154        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
 155        t1 = g + e1(d) + Ch(d,e,f) + 0x2e1b2138 + W[33];
 156        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
 157        t1 = f + e1(c) + Ch(c,d,e) + 0x4d2c6dfc + W[34];
 158        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
 159        t1 = e + e1(b) + Ch(b,c,d) + 0x53380d13 + W[35];
 160        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
 161        t1 = d + e1(a) + Ch(a,b,c) + 0x650a7354 + W[36];
 162        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
 163        t1 = c + e1(h) + Ch(h,a,b) + 0x766a0abb + W[37];
 164        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
 165        t1 = b + e1(g) + Ch(g,h,a) + 0x81c2c92e + W[38];
 166        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
 167        t1 = a + e1(f) + Ch(f,g,h) + 0x92722c85 + W[39];
 168        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 169
 170        t1 = h + e1(e) + Ch(e,f,g) + 0xa2bfe8a1 + W[40];
 171        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
 172        t1 = g + e1(d) + Ch(d,e,f) + 0xa81a664b + W[41];
 173        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
 174        t1 = f + e1(c) + Ch(c,d,e) + 0xc24b8b70 + W[42];
 175        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
 176        t1 = e + e1(b) + Ch(b,c,d) + 0xc76c51a3 + W[43];
 177        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
 178        t1 = d + e1(a) + Ch(a,b,c) + 0xd192e819 + W[44];
 179        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
 180        t1 = c + e1(h) + Ch(h,a,b) + 0xd6990624 + W[45];
 181        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
 182        t1 = b + e1(g) + Ch(g,h,a) + 0xf40e3585 + W[46];
 183        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
 184        t1 = a + e1(f) + Ch(f,g,h) + 0x106aa070 + W[47];
 185        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 186
 187        t1 = h + e1(e) + Ch(e,f,g) + 0x19a4c116 + W[48];
 188        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
 189        t1 = g + e1(d) + Ch(d,e,f) + 0x1e376c08 + W[49];
 190        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
 191        t1 = f + e1(c) + Ch(c,d,e) + 0x2748774c + W[50];
 192        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
 193        t1 = e + e1(b) + Ch(b,c,d) + 0x34b0bcb5 + W[51];
 194        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
 195        t1 = d + e1(a) + Ch(a,b,c) + 0x391c0cb3 + W[52];
 196        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
 197        t1 = c + e1(h) + Ch(h,a,b) + 0x4ed8aa4a + W[53];
 198        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
 199        t1 = b + e1(g) + Ch(g,h,a) + 0x5b9cca4f + W[54];
 200        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
 201        t1 = a + e1(f) + Ch(f,g,h) + 0x682e6ff3 + W[55];
 202        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 203
 204        t1 = h + e1(e) + Ch(e,f,g) + 0x748f82ee + W[56];
 205        t2 = e0(a) + Maj(a,b,c);    d+=t1;    h=t1+t2;
 206        t1 = g + e1(d) + Ch(d,e,f) + 0x78a5636f + W[57];
 207        t2 = e0(h) + Maj(h,a,b);    c+=t1;    g=t1+t2;
 208        t1 = f + e1(c) + Ch(c,d,e) + 0x84c87814 + W[58];
 209        t2 = e0(g) + Maj(g,h,a);    b+=t1;    f=t1+t2;
 210        t1 = e + e1(b) + Ch(b,c,d) + 0x8cc70208 + W[59];
 211        t2 = e0(f) + Maj(f,g,h);    a+=t1;    e=t1+t2;
 212        t1 = d + e1(a) + Ch(a,b,c) + 0x90befffa + W[60];
 213        t2 = e0(e) + Maj(e,f,g);    h+=t1;    d=t1+t2;
 214        t1 = c + e1(h) + Ch(h,a,b) + 0xa4506ceb + W[61];
 215        t2 = e0(d) + Maj(d,e,f);    g+=t1;    c=t1+t2;
 216        t1 = b + e1(g) + Ch(g,h,a) + 0xbef9a3f7 + W[62];
 217        t2 = e0(c) + Maj(c,d,e);    f+=t1;    b=t1+t2;
 218        t1 = a + e1(f) + Ch(f,g,h) + 0xc67178f2 + W[63];
 219        t2 = e0(b) + Maj(b,c,d);    e+=t1;    a=t1+t2;
 220
 221        state[0] += a; state[1] += b; state[2] += c; state[3] += d;
 222        state[4] += e; state[5] += f; state[6] += g; state[7] += h;
 223
 224        /* clear any sensitive info... */
 225        a = b = c = d = e = f = g = h = t1 = t2 = 0;
 226        memzero_explicit(W, 64 * sizeof(u32));
 227}
 228
 229static void sha256_generic_block_fn(struct sha256_state *sst, u8 const *src,
 230                                    int blocks)
 231{
 232        while (blocks--) {
 233                sha256_transform(sst->state, src);
 234                src += SHA256_BLOCK_SIZE;
 235        }
 236}
 237
 238int crypto_sha256_update(struct shash_desc *desc, const u8 *data,
 239                          unsigned int len)
 240{
 241        return sha256_base_do_update(desc, data, len, sha256_generic_block_fn);
 242}
 243EXPORT_SYMBOL(crypto_sha256_update);
 244
 245static int sha256_final(struct shash_desc *desc, u8 *out)
 246{
 247        sha256_base_do_finalize(desc, sha256_generic_block_fn);
 248        return sha256_base_finish(desc, out);
 249}
 250
 251int crypto_sha256_finup(struct shash_desc *desc, const u8 *data,
 252                        unsigned int len, u8 *hash)
 253{
 254        sha256_base_do_update(desc, data, len, sha256_generic_block_fn);
 255        return sha256_final(desc, hash);
 256}
 257EXPORT_SYMBOL(crypto_sha256_finup);
 258
 259static struct shash_alg sha256_algs[2] = { {
 260        .digestsize     =       SHA256_DIGEST_SIZE,
 261        .init           =       sha256_base_init,
 262        .update         =       crypto_sha256_update,
 263        .final          =       sha256_final,
 264        .finup          =       crypto_sha256_finup,
 265        .descsize       =       sizeof(struct sha256_state),
 266        .base           =       {
 267                .cra_name       =       "sha256",
 268                .cra_driver_name=       "sha256-generic",
 269                .cra_priority   =       100,
 270                .cra_blocksize  =       SHA256_BLOCK_SIZE,
 271                .cra_module     =       THIS_MODULE,
 272        }
 273}, {
 274        .digestsize     =       SHA224_DIGEST_SIZE,
 275        .init           =       sha224_base_init,
 276        .update         =       crypto_sha256_update,
 277        .final          =       sha256_final,
 278        .finup          =       crypto_sha256_finup,
 279        .descsize       =       sizeof(struct sha256_state),
 280        .base           =       {
 281                .cra_name       =       "sha224",
 282                .cra_driver_name=       "sha224-generic",
 283                .cra_priority   =       100,
 284                .cra_blocksize  =       SHA224_BLOCK_SIZE,
 285                .cra_module     =       THIS_MODULE,
 286        }
 287} };
 288
 289static int __init sha256_generic_mod_init(void)
 290{
 291        return crypto_register_shashes(sha256_algs, ARRAY_SIZE(sha256_algs));
 292}
 293
 294static void __exit sha256_generic_mod_fini(void)
 295{
 296        crypto_unregister_shashes(sha256_algs, ARRAY_SIZE(sha256_algs));
 297}
 298
 299subsys_initcall(sha256_generic_mod_init);
 300module_exit(sha256_generic_mod_fini);
 301
 302MODULE_LICENSE("GPL");
 303MODULE_DESCRIPTION("SHA-224 and SHA-256 Secure Hash Algorithm");
 304
 305MODULE_ALIAS_CRYPTO("sha224");
 306MODULE_ALIAS_CRYPTO("sha224-generic");
 307MODULE_ALIAS_CRYPTO("sha256");
 308MODULE_ALIAS_CRYPTO("sha256-generic");
 309