linux/arch/x86/crypto/twofish-avx-x86_64-asm_64.S
<<
>>
Prefs
   1/* SPDX-License-Identifier: GPL-2.0-or-later */
   2/*
   3 * Twofish Cipher 8-way parallel algorithm (AVX/x86_64)
   4 *
   5 * Copyright (C) 2012 Johannes Goetzfried
   6 *     <Johannes.Goetzfried@informatik.stud.uni-erlangen.de>
   7 *
   8 * Copyright © 2012-2013 Jussi Kivilinna <jussi.kivilinna@iki.fi>
   9 */
  10
  11#include <linux/linkage.h>
  12#include <asm/frame.h>
  13#include "glue_helper-asm-avx.S"
  14
  15.file "twofish-avx-x86_64-asm_64.S"
  16
  17.section        .rodata.cst16.bswap128_mask, "aM", @progbits, 16
  18.align 16
  19.Lbswap128_mask:
  20        .byte 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0
  21
  22.section        .rodata.cst16.xts_gf128mul_and_shl1_mask, "aM", @progbits, 16
  23.align 16
  24.Lxts_gf128mul_and_shl1_mask:
  25        .byte 0x87, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0
  26
  27.text
  28
  29/* structure of crypto context */
  30#define s0      0
  31#define s1      1024
  32#define s2      2048
  33#define s3      3072
  34#define w       4096
  35#define k       4128
  36
  37/**********************************************************************
  38  8-way AVX twofish
  39 **********************************************************************/
  40#define CTX %rdi
  41
  42#define RA1 %xmm0
  43#define RB1 %xmm1
  44#define RC1 %xmm2
  45#define RD1 %xmm3
  46
  47#define RA2 %xmm4
  48#define RB2 %xmm5
  49#define RC2 %xmm6
  50#define RD2 %xmm7
  51
  52#define RX0 %xmm8
  53#define RY0 %xmm9
  54
  55#define RX1 %xmm10
  56#define RY1 %xmm11
  57
  58#define RK1 %xmm12
  59#define RK2 %xmm13
  60
  61#define RT %xmm14
  62#define RR %xmm15
  63
  64#define RID1  %r13
  65#define RID1d %r13d
  66#define RID2  %rsi
  67#define RID2d %esi
  68
  69#define RGI1   %rdx
  70#define RGI1bl %dl
  71#define RGI1bh %dh
  72#define RGI2   %rcx
  73#define RGI2bl %cl
  74#define RGI2bh %ch
  75
  76#define RGI3   %rax
  77#define RGI3bl %al
  78#define RGI3bh %ah
  79#define RGI4   %rbx
  80#define RGI4bl %bl
  81#define RGI4bh %bh
  82
  83#define RGS1  %r8
  84#define RGS1d %r8d
  85#define RGS2  %r9
  86#define RGS2d %r9d
  87#define RGS3  %r10
  88#define RGS3d %r10d
  89
  90
  91#define lookup_32bit(t0, t1, t2, t3, src, dst, interleave_op, il_reg) \
  92        movzbl          src ## bl,        RID1d;     \
  93        movzbl          src ## bh,        RID2d;     \
  94        shrq $16,       src;                         \
  95        movl            t0(CTX, RID1, 4), dst ## d;  \
  96        movl            t1(CTX, RID2, 4), RID2d;     \
  97        movzbl          src ## bl,        RID1d;     \
  98        xorl            RID2d,            dst ## d;  \
  99        movzbl          src ## bh,        RID2d;     \
 100        interleave_op(il_reg);                       \
 101        xorl            t2(CTX, RID1, 4), dst ## d;  \
 102        xorl            t3(CTX, RID2, 4), dst ## d;
 103
 104#define dummy(d) /* do nothing */
 105
 106#define shr_next(reg) \
 107        shrq $16,       reg;
 108
 109#define G(gi1, gi2, x, t0, t1, t2, t3) \
 110        lookup_32bit(t0, t1, t2, t3, ##gi1, RGS1, shr_next, ##gi1);  \
 111        lookup_32bit(t0, t1, t2, t3, ##gi2, RGS3, shr_next, ##gi2);  \
 112        \
 113        lookup_32bit(t0, t1, t2, t3, ##gi1, RGS2, dummy, none);      \
 114        shlq $32,       RGS2;                                        \
 115        orq             RGS1, RGS2;                                  \
 116        lookup_32bit(t0, t1, t2, t3, ##gi2, RGS1, dummy, none);      \
 117        shlq $32,       RGS1;                                        \
 118        orq             RGS1, RGS3;
 119
 120#define round_head_2(a, b, x1, y1, x2, y2) \
 121        vmovq           b ## 1, RGI3;           \
 122        vpextrq $1,     b ## 1, RGI4;           \
 123        \
 124        G(RGI1, RGI2, x1, s0, s1, s2, s3);      \
 125        vmovq           a ## 2, RGI1;           \
 126        vpextrq $1,     a ## 2, RGI2;           \
 127        vmovq           RGS2, x1;               \
 128        vpinsrq $1,     RGS3, x1, x1;           \
 129        \
 130        G(RGI3, RGI4, y1, s1, s2, s3, s0);      \
 131        vmovq           b ## 2, RGI3;           \
 132        vpextrq $1,     b ## 2, RGI4;           \
 133        vmovq           RGS2, y1;               \
 134        vpinsrq $1,     RGS3, y1, y1;           \
 135        \
 136        G(RGI1, RGI2, x2, s0, s1, s2, s3);      \
 137        vmovq           RGS2, x2;               \
 138        vpinsrq $1,     RGS3, x2, x2;           \
 139        \
 140        G(RGI3, RGI4, y2, s1, s2, s3, s0);      \
 141        vmovq           RGS2, y2;               \
 142        vpinsrq $1,     RGS3, y2, y2;
 143
 144#define encround_tail(a, b, c, d, x, y, prerotate) \
 145        vpaddd                  x, y,   x; \
 146        vpaddd                  x, RK1, RT;\
 147        prerotate(b);                      \
 148        vpxor                   RT, c,  c; \
 149        vpaddd                  y, x,   y; \
 150        vpaddd                  y, RK2, y; \
 151        vpsrld $1,              c, RT;     \
 152        vpslld $(32 - 1),       c, c;      \
 153        vpor                    c, RT,  c; \
 154        vpxor                   d, y,   d; \
 155
 156#define decround_tail(a, b, c, d, x, y, prerotate) \
 157        vpaddd                  x, y,   x; \
 158        vpaddd                  x, RK1, RT;\
 159        prerotate(a);                      \
 160        vpxor                   RT, c,  c; \
 161        vpaddd                  y, x,   y; \
 162        vpaddd                  y, RK2, y; \
 163        vpxor                   d, y,   d; \
 164        vpsrld $1,              d, y;      \
 165        vpslld $(32 - 1),       d, d;      \
 166        vpor                    d, y,   d; \
 167
 168#define rotate_1l(x) \
 169        vpslld $1,              x, RR;     \
 170        vpsrld $(32 - 1),       x, x;      \
 171        vpor                    x, RR,  x;
 172
 173#define preload_rgi(c) \
 174        vmovq                   c, RGI1; \
 175        vpextrq $1,             c, RGI2;
 176
 177#define encrypt_round(n, a, b, c, d, preload, prerotate) \
 178        vbroadcastss (k+4*(2*(n)))(CTX),   RK1;                  \
 179        vbroadcastss (k+4*(2*(n)+1))(CTX), RK2;                  \
 180        round_head_2(a, b, RX0, RY0, RX1, RY1);                  \
 181        encround_tail(a ## 1, b ## 1, c ## 1, d ## 1, RX0, RY0, prerotate); \
 182        preload(c ## 1);                                         \
 183        encround_tail(a ## 2, b ## 2, c ## 2, d ## 2, RX1, RY1, prerotate);
 184
 185#define decrypt_round(n, a, b, c, d, preload, prerotate) \
 186        vbroadcastss (k+4*(2*(n)))(CTX),   RK1;                  \
 187        vbroadcastss (k+4*(2*(n)+1))(CTX), RK2;                  \
 188        round_head_2(a, b, RX0, RY0, RX1, RY1);                  \
 189        decround_tail(a ## 1, b ## 1, c ## 1, d ## 1, RX0, RY0, prerotate); \
 190        preload(c ## 1);                                         \
 191        decround_tail(a ## 2, b ## 2, c ## 2, d ## 2, RX1, RY1, prerotate);
 192
 193#define encrypt_cycle(n) \
 194        encrypt_round((2*n), RA, RB, RC, RD, preload_rgi, rotate_1l); \
 195        encrypt_round(((2*n) + 1), RC, RD, RA, RB, preload_rgi, rotate_1l);
 196
 197#define encrypt_cycle_last(n) \
 198        encrypt_round((2*n), RA, RB, RC, RD, preload_rgi, rotate_1l); \
 199        encrypt_round(((2*n) + 1), RC, RD, RA, RB, dummy, dummy);
 200
 201#define decrypt_cycle(n) \
 202        decrypt_round(((2*n) + 1), RC, RD, RA, RB, preload_rgi, rotate_1l); \
 203        decrypt_round((2*n), RA, RB, RC, RD, preload_rgi, rotate_1l);
 204
 205#define decrypt_cycle_last(n) \
 206        decrypt_round(((2*n) + 1), RC, RD, RA, RB, preload_rgi, rotate_1l); \
 207        decrypt_round((2*n), RA, RB, RC, RD, dummy, dummy);
 208
 209#define transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \
 210        vpunpckldq              x1, x0, t0; \
 211        vpunpckhdq              x1, x0, t2; \
 212        vpunpckldq              x3, x2, t1; \
 213        vpunpckhdq              x3, x2, x3; \
 214        \
 215        vpunpcklqdq             t1, t0, x0; \
 216        vpunpckhqdq             t1, t0, x1; \
 217        vpunpcklqdq             x3, t2, x2; \
 218        vpunpckhqdq             x3, t2, x3;
 219
 220#define inpack_blocks(x0, x1, x2, x3, wkey, t0, t1, t2) \
 221        vpxor           x0, wkey, x0; \
 222        vpxor           x1, wkey, x1; \
 223        vpxor           x2, wkey, x2; \
 224        vpxor           x3, wkey, x3; \
 225        \
 226        transpose_4x4(x0, x1, x2, x3, t0, t1, t2)
 227
 228#define outunpack_blocks(x0, x1, x2, x3, wkey, t0, t1, t2) \
 229        transpose_4x4(x0, x1, x2, x3, t0, t1, t2) \
 230        \
 231        vpxor           x0, wkey, x0; \
 232        vpxor           x1, wkey, x1; \
 233        vpxor           x2, wkey, x2; \
 234        vpxor           x3, wkey, x3;
 235
 236.align 8
 237SYM_FUNC_START_LOCAL(__twofish_enc_blk8)
 238        /* input:
 239         *      %rdi: ctx, CTX
 240         *      RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: blocks
 241         * output:
 242         *      RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks
 243         */
 244
 245        vmovdqu w(CTX), RK1;
 246
 247        pushq %r13;
 248        pushq %rbx;
 249        pushq %rcx;
 250
 251        inpack_blocks(RA1, RB1, RC1, RD1, RK1, RX0, RY0, RK2);
 252        preload_rgi(RA1);
 253        rotate_1l(RD1);
 254        inpack_blocks(RA2, RB2, RC2, RD2, RK1, RX0, RY0, RK2);
 255        rotate_1l(RD2);
 256
 257        encrypt_cycle(0);
 258        encrypt_cycle(1);
 259        encrypt_cycle(2);
 260        encrypt_cycle(3);
 261        encrypt_cycle(4);
 262        encrypt_cycle(5);
 263        encrypt_cycle(6);
 264        encrypt_cycle_last(7);
 265
 266        vmovdqu (w+4*4)(CTX), RK1;
 267
 268        popq %rcx;
 269        popq %rbx;
 270        popq %r13;
 271
 272        outunpack_blocks(RC1, RD1, RA1, RB1, RK1, RX0, RY0, RK2);
 273        outunpack_blocks(RC2, RD2, RA2, RB2, RK1, RX0, RY0, RK2);
 274
 275        ret;
 276SYM_FUNC_END(__twofish_enc_blk8)
 277
 278.align 8
 279SYM_FUNC_START_LOCAL(__twofish_dec_blk8)
 280        /* input:
 281         *      %rdi: ctx, CTX
 282         *      RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2: encrypted blocks
 283         * output:
 284         *      RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2: decrypted blocks
 285         */
 286
 287        vmovdqu (w+4*4)(CTX), RK1;
 288
 289        pushq %r13;
 290        pushq %rbx;
 291
 292        inpack_blocks(RC1, RD1, RA1, RB1, RK1, RX0, RY0, RK2);
 293        preload_rgi(RC1);
 294        rotate_1l(RA1);
 295        inpack_blocks(RC2, RD2, RA2, RB2, RK1, RX0, RY0, RK2);
 296        rotate_1l(RA2);
 297
 298        decrypt_cycle(7);
 299        decrypt_cycle(6);
 300        decrypt_cycle(5);
 301        decrypt_cycle(4);
 302        decrypt_cycle(3);
 303        decrypt_cycle(2);
 304        decrypt_cycle(1);
 305        decrypt_cycle_last(0);
 306
 307        vmovdqu (w)(CTX), RK1;
 308
 309        popq %rbx;
 310        popq %r13;
 311
 312        outunpack_blocks(RA1, RB1, RC1, RD1, RK1, RX0, RY0, RK2);
 313        outunpack_blocks(RA2, RB2, RC2, RD2, RK1, RX0, RY0, RK2);
 314
 315        ret;
 316SYM_FUNC_END(__twofish_dec_blk8)
 317
 318SYM_FUNC_START(twofish_ecb_enc_8way)
 319        /* input:
 320         *      %rdi: ctx, CTX
 321         *      %rsi: dst
 322         *      %rdx: src
 323         */
 324        FRAME_BEGIN
 325
 326        movq %rsi, %r11;
 327
 328        load_8way(%rdx, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2);
 329
 330        call __twofish_enc_blk8;
 331
 332        store_8way(%r11, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2);
 333
 334        FRAME_END
 335        ret;
 336SYM_FUNC_END(twofish_ecb_enc_8way)
 337
 338SYM_FUNC_START(twofish_ecb_dec_8way)
 339        /* input:
 340         *      %rdi: ctx, CTX
 341         *      %rsi: dst
 342         *      %rdx: src
 343         */
 344        FRAME_BEGIN
 345
 346        movq %rsi, %r11;
 347
 348        load_8way(%rdx, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2);
 349
 350        call __twofish_dec_blk8;
 351
 352        store_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2);
 353
 354        FRAME_END
 355        ret;
 356SYM_FUNC_END(twofish_ecb_dec_8way)
 357
 358SYM_FUNC_START(twofish_cbc_dec_8way)
 359        /* input:
 360         *      %rdi: ctx, CTX
 361         *      %rsi: dst
 362         *      %rdx: src
 363         */
 364        FRAME_BEGIN
 365
 366        pushq %r12;
 367
 368        movq %rsi, %r11;
 369        movq %rdx, %r12;
 370
 371        load_8way(%rdx, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2);
 372
 373        call __twofish_dec_blk8;
 374
 375        store_cbc_8way(%r12, %r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2);
 376
 377        popq %r12;
 378
 379        FRAME_END
 380        ret;
 381SYM_FUNC_END(twofish_cbc_dec_8way)
 382
 383SYM_FUNC_START(twofish_ctr_8way)
 384        /* input:
 385         *      %rdi: ctx, CTX
 386         *      %rsi: dst
 387         *      %rdx: src
 388         *      %rcx: iv (little endian, 128bit)
 389         */
 390        FRAME_BEGIN
 391
 392        pushq %r12;
 393
 394        movq %rsi, %r11;
 395        movq %rdx, %r12;
 396
 397        load_ctr_8way(%rcx, .Lbswap128_mask, RA1, RB1, RC1, RD1, RA2, RB2, RC2,
 398                      RD2, RX0, RX1, RY0);
 399
 400        call __twofish_enc_blk8;
 401
 402        store_ctr_8way(%r12, %r11, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2);
 403
 404        popq %r12;
 405
 406        FRAME_END
 407        ret;
 408SYM_FUNC_END(twofish_ctr_8way)
 409
 410SYM_FUNC_START(twofish_xts_enc_8way)
 411        /* input:
 412         *      %rdi: ctx, CTX
 413         *      %rsi: dst
 414         *      %rdx: src
 415         *      %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸))
 416         */
 417        FRAME_BEGIN
 418
 419        movq %rsi, %r11;
 420
 421        /* regs <= src, dst <= IVs, regs <= regs xor IVs */
 422        load_xts_8way(%rcx, %rdx, %rsi, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2,
 423                      RX0, RX1, RY0, .Lxts_gf128mul_and_shl1_mask);
 424
 425        call __twofish_enc_blk8;
 426
 427        /* dst <= regs xor IVs(in dst) */
 428        store_xts_8way(%r11, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2);
 429
 430        FRAME_END
 431        ret;
 432SYM_FUNC_END(twofish_xts_enc_8way)
 433
 434SYM_FUNC_START(twofish_xts_dec_8way)
 435        /* input:
 436         *      %rdi: ctx, CTX
 437         *      %rsi: dst
 438         *      %rdx: src
 439         *      %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸))
 440         */
 441        FRAME_BEGIN
 442
 443        movq %rsi, %r11;
 444
 445        /* regs <= src, dst <= IVs, regs <= regs xor IVs */
 446        load_xts_8way(%rcx, %rdx, %rsi, RC1, RD1, RA1, RB1, RC2, RD2, RA2, RB2,
 447                      RX0, RX1, RY0, .Lxts_gf128mul_and_shl1_mask);
 448
 449        call __twofish_dec_blk8;
 450
 451        /* dst <= regs xor IVs(in dst) */
 452        store_xts_8way(%r11, RA1, RB1, RC1, RD1, RA2, RB2, RC2, RD2);
 453
 454        FRAME_END
 455        ret;
 456SYM_FUNC_END(twofish_xts_dec_8way)
 457