diff options
author | Jerry Shih <bignose1007@gmail.com> | 2024-03-09 15:03:56 +0800 |
---|---|---|
committer | Tomas Mraz <tomas@openssl.org> | 2024-04-17 16:55:03 +0200 |
commit | da8b6308bd7ad5b7c779aa2d9123bf5faacaec7f (patch) | |
tree | 988c7558b2b03f0b6d01cb69c72962fb71526f0f | |
parent | 96939f1e2c6ac1264142ef62c9925786f5723649 (diff) |
Use scalar ALU and vector ALU together for chacha20 stream cipher
Fixes #24070
Use scalar ALU for 1 chacha block with rvv ALU simultaneously.
The tail elements(non-multiple of block length) will be handled by
the scalar logic.
Use rvv path if the input length > chacha_block_size.
And we have about 1.2x improvement comparing with the original code.
Reviewed-by: Hongren Zheng <i@zenithal.me>
Reviewed-by: Paul Dale <ppzgs1@gmail.com>
Reviewed-by: Tomas Mraz <tomas@openssl.org>
(Merged from https://github.com/openssl/openssl/pull/24097)
-rw-r--r-- | crypto/chacha/asm/chacha-riscv64-zbb-zvkb.pl (renamed from crypto/chacha/asm/chacha-riscv64-zvkb.pl) | 306 | ||||
-rw-r--r-- | crypto/chacha/build.info | 4 | ||||
-rw-r--r-- | crypto/chacha/chacha_riscv.c | 11 | ||||
-rw-r--r-- | crypto/perlasm/riscv.pm | 30 |
4 files changed, 269 insertions, 82 deletions
diff --git a/crypto/chacha/asm/chacha-riscv64-zvkb.pl b/crypto/chacha/asm/chacha-riscv64-zbb-zvkb.pl index caa7a1a303..6508af4765 100644 --- a/crypto/chacha/asm/chacha-riscv64-zvkb.pl +++ b/crypto/chacha/asm/chacha-riscv64-zbb-zvkb.pl @@ -38,6 +38,7 @@ # - RV64I # - RISC-V Vector ('V') with VLEN >= 128 # - RISC-V Vector Cryptography Bit-manipulation extension ('Zvkb') +# - RISC-V Basic Bit-manipulation extension ('Zbb') # - RISC-V Zicclsm(Main memory supports misaligned loads/stores) use strict; @@ -59,19 +60,30 @@ my $code = <<___; .text ___ -# void ChaCha20_ctr32_zvkb(unsigned char *out, const unsigned char *inp, -# size_t len, const unsigned int key[8], -# const unsigned int counter[4]); +# void ChaCha20_ctr32_zbb_zvkb(unsigned char *out, const unsigned char *inp, +# size_t len, const unsigned int key[8], +# const unsigned int counter[4]); ################################################################################ my ( $OUTPUT, $INPUT, $LEN, $KEY, $COUNTER ) = ( "a0", "a1", "a2", "a3", "a4" ); -my ( $T0 ) = ( "t0" ); -my ( $CONST_DATA0, $CONST_DATA1, $CONST_DATA2, $CONST_DATA3 ) = - ( "a5", "a6", "a7", "t1" ); -my ( $KEY0, $KEY1, $KEY2,$KEY3, $KEY4, $KEY5, $KEY6, $KEY7, - $COUNTER0, $COUNTER1, $NONCE0, $NONCE1 -) = ( "s0", "s1", "s2", "s3", "s4", "s5", "s6", - "s7", "s8", "s9", "s10", "s11" ); -my ( $VL, $STRIDE, $CHACHA_LOOP_COUNT ) = ( "t2", "t3", "t4" ); +my ( $CONST_DATA0, $CONST_DATA1, $CONST_DATA2, $CONST_DATA3 ) = ( "a5", "a6", + "a7", "s0" ); +my ( $KEY0, $KEY1, $KEY2, $KEY3, $KEY4, $KEY5, $KEY6, $KEY7, $COUNTER0, + $COUNTER1, $NONCE0, $NONCE1) = ( "s1", "s2", "s3", "s4", "s5", "s6", "s7", + "s8", "s9", "s10", "s11", "t0" ); +my ( $STATE0, $STATE1, $STATE2, $STATE3, + $STATE4, $STATE5, $STATE6, $STATE7, + $STATE8, $STATE9, $STATE10, $STATE11, + $STATE12, $STATE13, $STATE14, $STATE15) = ( + $CONST_DATA0, $CONST_DATA1, $CONST_DATA2, $CONST_DATA3, + $KEY0, $KEY1, $KEY2, $KEY3, + $KEY4, $KEY5, $KEY6, $KEY7, + $COUNTER0, $COUNTER1, $NONCE0, $NONCE1 ); +my ( $VL ) = ( "t1" ); +my ( $CURRENT_COUNTER ) = ( "t2" ); +my ( $T0 ) = ( "t3" ); +my ( $T1 ) = ( "t4" ); +my ( $T2 ) = ( "t5" ); +my ( $T3 ) = ( "t6" ); my ( $V0, $V1, $V2, $V3, $V4, $V5, $V6, $V7, $V8, $V9, $V10, $V11, $V12, $V13, $V14, $V15, $V16, $V17, $V18, $V19, $V20, $V21, @@ -80,63 +92,118 @@ my ( sub chacha_quad_round_group { my ( - $A0, $B0, $C0, $D0, $A1, $B1, $C1, $D1, - $A2, $B2, $C2, $D2, $A3, $B3, $C3, $D3 + $A0, $B0, $C0, $D0, + $A1, $B1, $C1, $D1, + $A2, $B2, $C2, $D2, + $A3, $B3, $C3, $D3, + + $S_A0, $S_B0, $S_C0, $S_D0, + $S_A1, $S_B1, $S_C1, $S_D1, + $S_A2, $S_B2, $S_C2, $S_D2, + $S_A3, $S_B3, $S_C3, $S_D3, ) = @_; my $code = <<___; # a += b; d ^= a; d <<<= 16; @{[vadd_vv $A0, $A0, $B0]} + add $S_A0, $S_A0, $S_B0 @{[vadd_vv $A1, $A1, $B1]} + add $S_A1, $S_A1, $S_B1 @{[vadd_vv $A2, $A2, $B2]} + add $S_A2, $S_A2, $S_B2 @{[vadd_vv $A3, $A3, $B3]} + add $S_A3, $S_A3, $S_B3 @{[vxor_vv $D0, $D0, $A0]} + xor $S_D0, $S_D0, $S_A0 @{[vxor_vv $D1, $D1, $A1]} + xor $S_D1, $S_D1, $S_A1 @{[vxor_vv $D2, $D2, $A2]} + xor $S_D2, $S_D2, $S_A2 @{[vxor_vv $D3, $D3, $A3]} + xor $S_D3, $S_D3, $S_A3 @{[vror_vi $D0, $D0, 32 - 16]} + @{[roriw $S_D0, $S_D0, 32 - 16]} @{[vror_vi $D1, $D1, 32 - 16]} + @{[roriw $S_D1, $S_D1, 32 - 16]} @{[vror_vi $D2, $D2, 32 - 16]} + @{[roriw $S_D2, $S_D2, 32 - 16]} @{[vror_vi $D3, $D3, 32 - 16]} + @{[roriw $S_D3, $S_D3, 32 - 16]} # c += d; b ^= c; b <<<= 12; @{[vadd_vv $C0, $C0, $D0]} + add $S_C0, $S_C0, $S_D0 @{[vadd_vv $C1, $C1, $D1]} + add $S_C1, $S_C1, $S_D1 @{[vadd_vv $C2, $C2, $D2]} + add $S_C2, $S_C2, $S_D2 @{[vadd_vv $C3, $C3, $D3]} + add $S_C3, $S_C3, $S_D3 @{[vxor_vv $B0, $B0, $C0]} + xor $S_B0, $S_B0, $S_C0 @{[vxor_vv $B1, $B1, $C1]} + xor $S_B1, $S_B1, $S_C1 @{[vxor_vv $B2, $B2, $C2]} + xor $S_B2, $S_B2, $S_C2 @{[vxor_vv $B3, $B3, $C3]} + xor $S_B3, $S_B3, $S_C3 @{[vror_vi $B0, $B0, 32 - 12]} + @{[roriw $S_B0, $S_B0, 32 - 12]} @{[vror_vi $B1, $B1, 32 - 12]} + @{[roriw $S_B1, $S_B1, 32 - 12]} @{[vror_vi $B2, $B2, 32 - 12]} + @{[roriw $S_B2, $S_B2, 32 - 12]} @{[vror_vi $B3, $B3, 32 - 12]} + @{[roriw $S_B3, $S_B3, 32 - 12]} # a += b; d ^= a; d <<<= 8; @{[vadd_vv $A0, $A0, $B0]} + add $S_A0, $S_A0, $S_B0 @{[vadd_vv $A1, $A1, $B1]} + add $S_A1, $S_A1, $S_B1 @{[vadd_vv $A2, $A2, $B2]} + add $S_A2, $S_A2, $S_B2 @{[vadd_vv $A3, $A3, $B3]} + add $S_A3, $S_A3, $S_B3 @{[vxor_vv $D0, $D0, $A0]} + xor $S_D0, $S_D0, $S_A0 @{[vxor_vv $D1, $D1, $A1]} + xor $S_D1, $S_D1, $S_A1 @{[vxor_vv $D2, $D2, $A2]} + xor $S_D2, $S_D2, $S_A2 @{[vxor_vv $D3, $D3, $A3]} + xor $S_D3, $S_D3, $S_A3 @{[vror_vi $D0, $D0, 32 - 8]} + @{[roriw $S_D0, $S_D0, 32 - 8]} @{[vror_vi $D1, $D1, 32 - 8]} + @{[roriw $S_D1, $S_D1, 32 - 8]} @{[vror_vi $D2, $D2, 32 - 8]} + @{[roriw $S_D2, $S_D2, 32 - 8]} @{[vror_vi $D3, $D3, 32 - 8]} + @{[roriw $S_D3, $S_D3, 32 - 8]} # c += d; b ^= c; b <<<= 7; @{[vadd_vv $C0, $C0, $D0]} + add $S_C0, $S_C0, $S_D0 @{[vadd_vv $C1, $C1, $D1]} + add $S_C1, $S_C1, $S_D1 @{[vadd_vv $C2, $C2, $D2]} + add $S_C2, $S_C2, $S_D2 @{[vadd_vv $C3, $C3, $D3]} + add $S_C3, $S_C3, $S_D3 @{[vxor_vv $B0, $B0, $C0]} + xor $S_B0, $S_B0, $S_C0 @{[vxor_vv $B1, $B1, $C1]} + xor $S_B1, $S_B1, $S_C1 @{[vxor_vv $B2, $B2, $C2]} + xor $S_B2, $S_B2, $S_C2 @{[vxor_vv $B3, $B3, $C3]} + xor $S_B3, $S_B3, $S_C3 @{[vror_vi $B0, $B0, 32 - 7]} + @{[roriw $S_B0, $S_B0, 32 - 7]} @{[vror_vi $B1, $B1, 32 - 7]} + @{[roriw $S_B1, $S_B1, 32 - 7]} @{[vror_vi $B2, $B2, 32 - 7]} + @{[roriw $S_B2, $S_B2, 32 - 7]} @{[vror_vi $B3, $B3, 32 - 7]} + @{[roriw $S_B3, $S_B3, 32 - 7]} ___ return $code; @@ -144,12 +211,9 @@ ___ $code .= <<___; .p2align 3 -.globl ChaCha20_ctr32_zvkb -.type ChaCha20_ctr32_zvkb,\@function -ChaCha20_ctr32_zvkb: - srli $LEN, $LEN, 6 - beqz $LEN, .Lend - +.globl ChaCha20_ctr32_zbb_zvkb +.type ChaCha20_ctr32_zbb_zvkb,\@function +ChaCha20_ctr32_zbb_zvkb: addi sp, sp, -96 sd s0, 0(sp) sd s1, 8(sp) @@ -163,139 +227,232 @@ ChaCha20_ctr32_zvkb: sd s9, 72(sp) sd s10, 80(sp) sd s11, 88(sp) + addi sp, sp, -64 + + lw $CURRENT_COUNTER, 0($COUNTER) - li $STRIDE, 64 +.Lblock_loop: + # We will use the scalar ALU for 1 chacha block. + srli $T0, $LEN, 6 + @{[vsetvli $VL, $T0, "e32", "m1", "ta", "ma"]} + slli $T1, $VL, 6 + bltu $T1, $LEN, 1f + # Since there is no more chacha block existed, we need to split 1 block + # from vector ALU. + addi $T1, $VL, -1 + @{[vsetvli $VL, $T1, "e32", "m1", "ta", "ma"]} +1: #### chacha block data + # init chacha const states # "expa" little endian li $CONST_DATA0, 0x61707865 + @{[vmv_v_x $V0, $CONST_DATA0]} # "nd 3" little endian li $CONST_DATA1, 0x3320646e + @{[vmv_v_x $V1, $CONST_DATA1]} # "2-by" little endian li $CONST_DATA2, 0x79622d32 + @{[vmv_v_x $V2, $CONST_DATA2]} # "te k" little endian li $CONST_DATA3, 0x6b206574 - lw $KEY0, 0($KEY) - lw $KEY1, 4($KEY) - lw $KEY2, 8($KEY) - lw $KEY3, 12($KEY) - lw $KEY4, 16($KEY) - lw $KEY5, 20($KEY) - lw $KEY6, 24($KEY) - lw $KEY7, 28($KEY) - - lw $COUNTER0, 0($COUNTER) - lw $COUNTER1, 4($COUNTER) - lw $NONCE0, 8($COUNTER) - lw $NONCE1, 12($COUNTER) - -.Lblock_loop: - @{[vsetvli $VL, $LEN, "e32", "m1", "ta", "ma"]} - - # init chacha const states - @{[vmv_v_x $V0, $CONST_DATA0]} - @{[vmv_v_x $V1, $CONST_DATA1]} - @{[vmv_v_x $V2, $CONST_DATA2]} @{[vmv_v_x $V3, $CONST_DATA3]} # init chacha key states + lw $KEY1, 4($KEY) @{[vmv_v_x $V4, $KEY0]} + lw $KEY2, 8($KEY) @{[vmv_v_x $V5, $KEY1]} + lw $KEY3, 12($KEY) @{[vmv_v_x $V6, $KEY2]} + lw $KEY4, 16($KEY) @{[vmv_v_x $V7, $KEY3]} + lw $KEY5, 20($KEY) @{[vmv_v_x $V8, $KEY4]} + lw $KEY6, 24($KEY) @{[vmv_v_x $V9, $KEY5]} + lw $KEY7, 28($KEY) @{[vmv_v_x $V10, $KEY6]} @{[vmv_v_x $V11, $KEY7]} # init chacha key states + lw $COUNTER1, 4($COUNTER) @{[vid_v $V12]} - @{[vadd_vx $V12, $V12, $COUNTER0]} + lw $NONCE0, 8($COUNTER) + @{[vadd_vx $V12, $V12, $CURRENT_COUNTER]} + lw $NONCE1, 12($COUNTER) @{[vmv_v_x $V13, $COUNTER1]} + add $COUNTER0, $CURRENT_COUNTER, $VL # init chacha nonce states @{[vmv_v_x $V14, $NONCE0]} @{[vmv_v_x $V15, $NONCE1]} + li $T0, 64 # load the top-half of input data - @{[vlsseg_nf_e32_v 8, $V16, $INPUT, $STRIDE]} + @{[vlsseg_nf_e32_v 8, $V16, $INPUT, $T0]} - li $CHACHA_LOOP_COUNT, 10 + # 20 round groups + li $T0, 10 .Lround_loop: - addi $CHACHA_LOOP_COUNT, $CHACHA_LOOP_COUNT, -1 + addi $T0, $T0, -1 @{[chacha_quad_round_group $V0, $V4, $V8, $V12, $V1, $V5, $V9, $V13, $V2, $V6, $V10, $V14, - $V3, $V7, $V11, $V15]} + $V3, $V7, $V11, $V15, + $STATE0, $STATE4, $STATE8, $STATE12, + $STATE1, $STATE5, $STATE9, $STATE13, + $STATE2, $STATE6, $STATE10, $STATE14, + $STATE3, $STATE7, $STATE11, $STATE15]} @{[chacha_quad_round_group + $V3, $V4, $V9, $V14, $V0, $V5, $V10, $V15, $V1, $V6, $V11, $V12, $V2, $V7, $V8, $V13, - $V3, $V4, $V9, $V14]} - bnez $CHACHA_LOOP_COUNT, .Lround_loop + $STATE3, $STATE4, $STATE9, $STATE14, + $STATE0, $STATE5, $STATE10, $STATE15, + $STATE1, $STATE6, $STATE11, $STATE12, + $STATE2, $STATE7, $STATE8, $STATE13]} + bnez $T0, .Lround_loop + li $T0, 64 # load the bottom-half of input data - addi $T0, $INPUT, 32 - @{[vlsseg_nf_e32_v 8, $V24, $T0, $STRIDE]} + addi $T1, $INPUT, 32 + @{[vlsseg_nf_e32_v 8, $V24, $T1, $T0]} # add chacha top-half initial block states - @{[vadd_vx $V0, $V0, $CONST_DATA0]} - @{[vadd_vx $V1, $V1, $CONST_DATA1]} - @{[vadd_vx $V2, $V2, $CONST_DATA2]} - @{[vadd_vx $V3, $V3, $CONST_DATA3]} - @{[vadd_vx $V4, $V4, $KEY0]} - @{[vadd_vx $V5, $V5, $KEY1]} - @{[vadd_vx $V6, $V6, $KEY2]} - @{[vadd_vx $V7, $V7, $KEY3]} + # "expa" little endian + li $T0, 0x61707865 + @{[vadd_vx $V0, $V0, $T0]} + add $STATE0, $STATE0, $T0 + # "nd 3" little endian + li $T1, 0x3320646e + @{[vadd_vx $V1, $V1, $T1]} + add $STATE1, $STATE1, $T1 + lw $T0, 0($KEY) + # "2-by" little endian + li $T2, 0x79622d32 + @{[vadd_vx $V2, $V2, $T2]} + add $STATE2, $STATE2, $T2 + lw $T1, 4($KEY) + # "te k" little endian + li $T3, 0x6b206574 + @{[vadd_vx $V3, $V3, $T3]} + add $STATE3, $STATE3, $T3 + lw $T2, 8($KEY) + @{[vadd_vx $V4, $V4, $T0]} + add $STATE4, $STATE4, $T0 + lw $T3, 12($KEY) + @{[vadd_vx $V5, $V5, $T1]} + add $STATE5, $STATE5, $T1 + @{[vadd_vx $V6, $V6, $T2]} + add $STATE6, $STATE6, $T2 + @{[vadd_vx $V7, $V7, $T3]} + add $STATE7, $STATE7, $T3 + # xor with the top-half input @{[vxor_vv $V16, $V16, $V0]} + sw $STATE0, 0(sp) + sw $STATE1, 4(sp) @{[vxor_vv $V17, $V17, $V1]} + sw $STATE2, 8(sp) + sw $STATE3, 12(sp) @{[vxor_vv $V18, $V18, $V2]} + sw $STATE4, 16(sp) + sw $STATE5, 20(sp) @{[vxor_vv $V19, $V19, $V3]} + sw $STATE6, 24(sp) + sw $STATE7, 28(sp) @{[vxor_vv $V20, $V20, $V4]} + lw $T0, 16($KEY) @{[vxor_vv $V21, $V21, $V5]} + lw $T1, 20($KEY) @{[vxor_vv $V22, $V22, $V6]} + lw $T2, 24($KEY) @{[vxor_vv $V23, $V23, $V7]} # save the top-half of output - @{[vssseg_nf_e32_v 8, $V16, $OUTPUT, $STRIDE]} + li $T3, 64 + @{[vssseg_nf_e32_v 8, $V16, $OUTPUT, $T3]} # add chacha bottom-half initial block states - @{[vadd_vx $V8, $V8, $KEY4]} - @{[vadd_vx $V9, $V9, $KEY5]} - @{[vadd_vx $V10, $V10, $KEY6]} - @{[vadd_vx $V11, $V11, $KEY7]} + @{[vadd_vx $V8, $V8, $T0]} + add $STATE8, $STATE8, $T0 + lw $T3, 28($KEY) + @{[vadd_vx $V9, $V9, $T1]} + add $STATE9, $STATE9, $T1 + lw $T0, 4($COUNTER) + @{[vadd_vx $V10, $V10, $T2]} + add $STATE10, $STATE10, $T2 + lw $T1, 8($COUNTER) + @{[vadd_vx $V11, $V11, $T3]} + add $STATE11, $STATE11, $T3 + lw $T2, 12($COUNTER) @{[vid_v $V0]} - @{[vadd_vx $V12, $V12, $COUNTER0]} - @{[vadd_vx $V13, $V13, $COUNTER1]} - @{[vadd_vx $V14, $V14, $NONCE0]} - @{[vadd_vx $V15, $V15, $NONCE1]} + add $STATE12, $STATE12, $CURRENT_COUNTER + @{[vadd_vx $V12, $V12, $CURRENT_COUNTER]} + add $STATE12, $STATE12, $VL + @{[vadd_vx $V13, $V13, $T0]} + add $STATE13, $STATE13, $T0 + @{[vadd_vx $V14, $V14, $T1]} + add $STATE14, $STATE14, $T1 + @{[vadd_vx $V15, $V15, $T2]} + add $STATE15, $STATE15, $T2 @{[vadd_vv $V12, $V12, $V0]} # xor with the bottom-half input @{[vxor_vv $V24, $V24, $V8]} + sw $STATE8, 32(sp) @{[vxor_vv $V25, $V25, $V9]} + sw $STATE9, 36(sp) @{[vxor_vv $V26, $V26, $V10]} + sw $STATE10, 40(sp) @{[vxor_vv $V27, $V27, $V11]} + sw $STATE11, 44(sp) @{[vxor_vv $V29, $V29, $V13]} + sw $STATE12, 48(sp) @{[vxor_vv $V28, $V28, $V12]} + sw $STATE13, 52(sp) @{[vxor_vv $V30, $V30, $V14]} + sw $STATE14, 56(sp) @{[vxor_vv $V31, $V31, $V15]} + sw $STATE15, 60(sp) # save the bottom-half of output - addi $T0, $OUTPUT, 32 - @{[vssseg_nf_e32_v 8, $V24, $T0, $STRIDE]} + li $T0, 64 + addi $T1, $OUTPUT, 32 + @{[vssseg_nf_e32_v 8, $V24, $T1, $T0]} - # update counter - add $COUNTER0, $COUNTER0, $VL - sub $LEN, $LEN, $VL - # increase offset for `4 * 16 * VL = 64 * VL` + # the computed vector parts: `64 * VL` slli $T0, $VL, 6 + add $INPUT, $INPUT, $T0 add $OUTPUT, $OUTPUT, $T0 + sub $LEN, $LEN, $T0 + add $CURRENT_COUNTER, $CURRENT_COUNTER, $VL + + # process the scalar data block + addi $CURRENT_COUNTER, $CURRENT_COUNTER, 1 + li $T0, 64 + @{[minu $T1, $LEN, $T0]} + sub $LEN, $LEN, $T1 + mv $T2, sp +.Lscalar_data_loop: + @{[vsetvli $VL, $T1, "e8", "m8", "ta", "ma"]} + @{[vle8_v $V8, $INPUT]} + @{[vle8_v $V16, $T2]} + @{[vxor_vv $V8, $V8, $V16]} + @{[vse8_v $V8, $OUTPUT]} + add $INPUT, $INPUT, $VL + add $OUTPUT, $OUTPUT, $VL + add $T2, $T2, $VL + sub $T1, $T1, $VL + bnez $T1, .Lscalar_data_loop + bnez $LEN, .Lblock_loop + addi sp, sp, 64 ld s0, 0(sp) ld s1, 8(sp) ld s2, 16(sp) @@ -310,9 +467,8 @@ ChaCha20_ctr32_zvkb: ld s11, 88(sp) addi sp, sp, 96 -.Lend: ret -.size ChaCha20_ctr32_zvkb,.-ChaCha20_ctr32_zvkb +.size ChaCha20_ctr32_zbb_zvkb,.-ChaCha20_ctr32_zbb_zvkb ___ print $code; diff --git a/crypto/chacha/build.info b/crypto/chacha/build.info index fb30b7f74f..c1352c9b62 100644 --- a/crypto/chacha/build.info +++ b/crypto/chacha/build.info @@ -22,7 +22,7 @@ IF[{- !$disabled{asm} -}] $CHACHAASM_c64xplus=chacha-c64xplus.s - $CHACHAASM_riscv64=chacha_riscv.c chacha_enc.c chacha-riscv64-zvkb.s + $CHACHAASM_riscv64=chacha_riscv.c chacha_enc.c chacha-riscv64-zbb-zvkb.s $CHACHADEF_riscv64=INCLUDE_C_CHACHA20 # Now that we have defined all the arch specific variables, use the @@ -53,4 +53,4 @@ GENERATE[chacha-s390x.S]=asm/chacha-s390x.pl GENERATE[chacha-ia64.S]=asm/chacha-ia64.pl GENERATE[chacha-ia64.s]=chacha-ia64.S GENERATE[chacha-loongarch64.S]=asm/chacha-loongarch64.pl -GENERATE[chacha-riscv64-zvkb.s]=asm/chacha-riscv64-zvkb.pl +GENERATE[chacha-riscv64-zbb-zvkb.s]=asm/chacha-riscv64-zbb-zvkb.pl diff --git a/crypto/chacha/chacha_riscv.c b/crypto/chacha/chacha_riscv.c index 729a6282d2..8873460e5e 100644 --- a/crypto/chacha/chacha_riscv.c +++ b/crypto/chacha/chacha_riscv.c @@ -40,15 +40,16 @@ #include "crypto/chacha.h" #include "crypto/riscv_arch.h" -void ChaCha20_ctr32_zvkb(unsigned char *out, const unsigned char *inp, - size_t len, const unsigned int key[8], - const unsigned int counter[4]); +void ChaCha20_ctr32_zbb_zvkb(unsigned char *out, const unsigned char *inp, + size_t len, const unsigned int key[8], + const unsigned int counter[4]); void ChaCha20_ctr32(unsigned char *out, const unsigned char *inp, size_t len, const unsigned int key[8], const unsigned int counter[4]) { - if (RISCV_HAS_ZVKB() && riscv_vlen() >= 128) { - ChaCha20_ctr32_zvkb(out, inp, len, key, counter); + if (len > CHACHA_BLK_SIZE && RISCV_HAS_ZVKB() && RISCV_HAS_ZBB() && + riscv_vlen() >= 128) { + ChaCha20_ctr32_zbb_zvkb(out, inp, len, key, counter); } else { ChaCha20_ctr32_c(out, inp, len, key, counter); } diff --git a/crypto/perlasm/riscv.pm b/crypto/perlasm/riscv.pm index 9fdee77793..7e45bec812 100644 --- a/crypto/perlasm/riscv.pm +++ b/crypto/perlasm/riscv.pm @@ -384,6 +384,36 @@ sub rev8 { return ".word ".($template | ($rs << 15) | ($rd << 7)); } +sub roriw { + # Encoding for roriw rd, rs1, shamt instruction on RV64 + # XXXXXXX_ shamt _ rs1 _XXX_ rd _XXXXXXX + my $template = 0b0110000_00000_00000_101_00000_0011011; + my $rd = read_reg shift; + my $rs1 = read_reg shift; + my $shamt = shift; + return ".word ".($template | ($shamt << 20) | ($rs1 << 15) | ($rd << 7)); +} + +sub maxu { + # Encoding for maxu rd, rs1, rs2 instruction on RV64 + # XXXXXXX_ rs2 _ rs1 _XXX_ rd _XXXXXXX + my $template = 0b0000101_00000_00000_111_00000_0110011; + my $rd = read_reg shift; + my $rs1 = read_reg shift; + my $rs2 = read_reg shift; + return ".word ".($template | ($rs2 << 20) | ($rs1 << 15) | ($rd << 7)); +} + +sub minu { + # Encoding for minu rd, rs1, rs2 instruction on RV64 + # XXXXXXX_ rs2 _ rs1 _XXX_ rd _XXXXXXX + my $template = 0b0000101_00000_00000_101_00000_0110011; + my $rd = read_reg shift; + my $rs1 = read_reg shift; + my $rs2 = read_reg shift; + return ".word ".($template | ($rs2 << 20) | ($rs1 << 15) | ($rd << 7)); +} + # Vector instructions sub vadd_vv { |