From a163e60d950f5cbfa56778a10cc34c95681861f1 Mon Sep 17 00:00:00 2001 From: Andy Polyakov Date: Wed, 21 Jun 2017 15:29:53 +0200 Subject: sha/asm/keccak1600-mmx.pl: optimize for Atom and add comparison data. Curiously enough out-of-order Silvermont benefited most from optimization, 33%. [Originally mentioned "anomaly" turned to be misreported frequency scaling problem. Correct results were collected under older kernel.] Reviewed-by: Rich Salz Reviewed-by: Bernd Edlinger (Merged from https://github.com/openssl/openssl/pull/3739) --- crypto/sha/asm/keccak1600-mmx.pl | 241 ++++++++++++++++++++------------------- 1 file changed, 126 insertions(+), 115 deletions(-) (limited to 'crypto/sha') diff --git a/crypto/sha/asm/keccak1600-mmx.pl b/crypto/sha/asm/keccak1600-mmx.pl index 5dccedf2e6..c7685add79 100755 --- a/crypto/sha/asm/keccak1600-mmx.pl +++ b/crypto/sha/asm/keccak1600-mmx.pl @@ -21,29 +21,45 @@ # C[5] held in register bank and D[5] offloaded to memory. Though # instead of actually unrolling the loop pair-wise I simply flip # pointers to T[][] and A[][] and the end of round. Since number of -# rounds is even last round writes to A[][] and everything works out. +# rounds is even, last round writes to A[][] and everything works out. +# It's argued that MMX is the only code path meaningful to implement +# for x86. This is because non-MMX-capable processors is an extinct +# breed, and they as well can lurk executing compiler-generated code. +# For reference gcc-5.x-generated KECCAK_2X code takes 89 cycles per +# processed byte on Pentium. Which is fair result. But older compilers +# produce worse code. On the other hand one can wonder why not 128-bit +# SSE2? Well, SSE2 won't provide double improvement, rather far from +# that, if any at all on some processors, because it will take extra +# permutations and inter-bank data trasfers. Besides, contemporary +# CPUs are better off executing 64-bit code, and it makes lesser sense +# to invest into fancy 32-bit code. And the decision doesn't seem to +# be inadequate, if one compares below results to "64-bit platforms in +# 32-bit mode" SIMD data points available at +# http://keccak.noekeon.org/sw_performance.html. # ######################################################################## # Numbers are cycles per processed byte out of large message. # # r=1088(i) # -# PIII 31 -# Pentium M 27 -# P4 42 -# Core 2 20 -# Sandy Bridge(ii) 18 -# Atom 37 -# Silvermont(ii) 80(iv) -# VIA Nano(ii) 44 -# Sledgehammer(ii)(iii) 25 +# PIII 30/+150% +# Pentium M 27/+150% +# P4 40/+85% +# Core 2 19/+170% +# Sandy Bridge(ii) 18/+140% +# Atom 33/+180% +# Silvermont(ii) 30/+180% +# VIA Nano(ii) 43/+60% +# Sledgehammer(ii)(iii) 24/+130% # -# (i) Corresponds to SHA3-256. +# (i) Corresponds to SHA3-256. Numbers after slash are improvement +# coefficients over KECCAK_2X [with bit interleave and lane +# complementing] position-independent *scalar* code generated +# by gcc-5.x. It's not exactly fair comparison, but it's a +# datapoint... # (ii) 64-bit processor executing 32-bit code. # (iii) Result is considered to be representative even for older AMD # processors. -# (iv) This seems to be some processor anomaly. Successor doesn't -# have this problem... $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1; push(@INC,"${dir}","${dir}../../perlasm"); @@ -97,184 +113,179 @@ my @rhotates = ([ 0, 1, 62, 28, 27 ], &pxor (@C[3],&QWP($A[2][3],"esi")); &pxor (@C[4],&QWP($A[2][4],"esi")); + &pxor (@C[2],&QWP($A[3][2],"esi")); &pxor (@C[0],&QWP($A[3][0],"esi")); &pxor (@C[1],&QWP($A[3][1],"esi")); - &pxor (@C[2],&QWP($A[3][2],"esi")); &pxor (@C[3],&QWP($A[3][3],"esi")); + &movq (@T[0],@C[2]); &pxor (@C[4],&QWP($A[3][4],"esi")); - &movq (@T[0],@C[2]); - &movq (@T[2],@C[2]); - &psrlq (@T[0],63); - &psllq (@T[2],1); - &pxor (@T[0],@C[0]); - &pxor (@T[0],@T[2]); - &movq (&QWP(@D[1],"esp"),@T[0]); # D[1] = E[0] = ROL64(C[2], 1) ^ C[0]; - + &movq (@T[2],@C[2]); + &psrlq (@T[0],63); &movq (@T[1],@C[0]); + &psllq (@T[2],1); + &pxor (@T[0],@C[0]); &psrlq (@C[0],63); + &pxor (@T[0],@T[2]); &psllq (@T[1],1); + &movq (@T[2],@C[1]); + &movq (&QWP(@D[1],"esp"),@T[0]); # D[1] = E[0] = ROL64(C[2], 1) ^ C[0]; + &pxor (@T[1],@C[0]); + &psrlq (@T[2],63); &pxor (@T[1],@C[3]); + &movq (@C[0],@C[1]); &movq (&QWP(@D[4],"esp"),@T[1]); # D[4] = E[1] = ROL64(C[0], 1) ^ C[3]; - &movq (@C[0],@C[1]); - &movq (@T[2],@C[1]); - &psrlq (@C[0],63); - &psllq (@T[2],1); - &pxor (@C[0],@C[4]); - &pxor (@C[0],@T[2]); - &movq (&QWP(@D[0],"esp"),@C[0]); # D[0] = C[0] = ROL64(C[1], 1) ^ C[4]; + &psllq (@C[0],1); + &pxor (@T[2],@C[4]); + &pxor (@C[0],@T[2]); &movq (@T[2],@C[3]); &psrlq (@C[3],63); + &movq (&QWP(@D[0],"esp"),@C[0]); # D[0] = C[0] = ROL64(C[1], 1) ^ C[4]; &psllq (@T[2],1); + &movq (@T[0],@C[4]); + &psrlq (@C[4],63); &pxor (@C[1],@C[3]); + &psllq (@T[0],1); &pxor (@C[1],@T[2]); + &pxor (@C[2],@C[4]); &movq (&QWP(@D[2],"esp"),@C[1]); # D[2] = C[1] = ROL64(C[3], 1) ^ C[1]; + &pxor (@C[2],@T[0]); - &movq (@T[2],@C[4]); - &psrlq (@C[4],63); - &psllq (@T[2],1); - &pxor (@C[2],@C[4]); - &pxor (@C[2],@T[2]); - &movq (&QWP(@D[3],"esp"),@C[2]); # D[3] = C[2] = ROL64(C[4], 1) ^ C[2]; - - ######################################### first Rho step is special + ######################################### first Rho(0) is special &movq (@C[3],&QWP($A[3][3],"esi")); + &movq (&QWP(@D[3],"esp"),@C[2]); # D[3] = C[2] = ROL64(C[4], 1) ^ C[2]; &pxor (@C[3],@C[2]); + &movq (@C[4],&QWP($A[4][4],"esi")); &movq (@T[2],@C[3]); &psrlq (@C[3],64-$rhotates[3][3]); + &pxor (@C[4],@T[1]); &psllq (@T[2],$rhotates[3][3]); + &movq (@T[1],@C[4]); + &psrlq (@C[4],64-$rhotates[4][4]); &por (@C[3],@T[2]); # C[3] = ROL64(A[3][3] ^ C[2], rhotates[3][3]); /* D[3] */ - - &movq (@C[4],&QWP($A[4][4],"esi")); - &pxor (@C[4],@T[1]); - &movq (@T[2],@C[4]); - &psrlq (@C[4],64-$rhotates[4][4]); - &psllq (@T[2],$rhotates[4][4]); - &por (@C[4],@T[2]); # C[4] = ROL64(A[4][4] ^ E[1], rhotates[4][4]); /* D[4] */ - - &pxor (@C[0],&QWP($A[0][0],"esi")); # /* rotate by 0 */ /* D[0] */ + &psllq (@T[1],$rhotates[4][4]); &movq (@C[2],&QWP($A[2][2],"esi")); + &por (@C[4],@T[1]); # C[4] = ROL64(A[4][4] ^ E[1], rhotates[4][4]); /* D[4] */ &pxor (@C[2],@C[1]); + &movq (@C[1],&QWP($A[1][1],"esi")); &movq (@T[1],@C[2]); &psrlq (@C[2],64-$rhotates[2][2]); + &pxor (@C[1],&QWP(@D[1],"esp")); &psllq (@T[1],$rhotates[2][2]); - &por (@C[2],@T[1]); # C[2] = ROL64(A[2][2] ^ C[1], rhotates[2][2]); /* D[2] */ - &movq (@C[1],&QWP($A[1][1],"esi")); - &pxor (@C[1],@T[0]); - &movq (@T[2],@C[1]); - &psrlq (@C[1],64-$rhotates[1][1]); - &psllq (@T[2],$rhotates[1][1]); - &por (@C[1],@T[2]); # C[1] = ROL64(A[1][1] ^ E[0], rhotates[1][1]); /* D[1] */ + &movq (@T[2],@C[1]); + &psrlq (@C[1],64-$rhotates[1][1]); + &por (@C[2],@T[1]); # C[2] = ROL64(A[2][2] ^ C[1], rhotates[2][2]); /* D[2] */ + &psllq (@T[2],$rhotates[1][1]); + &pxor (@C[0],&QWP($A[0][0],"esi")); # /* rotate by 0 */ /* D[0] */ + &por (@C[1],@T[2]); # C[1] = ROL64(A[1][1] ^ D[1], rhotates[1][1]); sub Chi() { ######### regular Chi step - my $y = shift; + my ($y,$xrho) = @_; &movq (@T[0],@C[1]); + &movq (@T[1],@C[2]); &pandn (@T[0],@C[2]); + &pandn (@C[2],@C[3]); &pxor (@T[0],@C[0]); + &pxor (@C[2],@C[1]); &pxor (@T[0],&QWP(0,"ebx")) if ($y == 0); &lea ("ebx",&DWP(8,"ebx")) if ($y == 0); - &movq (&QWP($A[$y][0],"edi"),@T[0]); # R[0][0] = C[0] ^ (~C[1] & C[2]) ^ iotas[i]; - - &movq (@T[1],@C[2]); - &pandn (@T[1],@C[3]); - &pxor (@T[1],@C[1]); - &movq (&QWP($A[$y][1],"edi"),@T[1]); # R[0][1] = C[1] ^ (~C[2] & C[3]); &movq (@T[2],@C[3]); - &pandn (@T[2],@C[4]); - &pxor (@T[2],@C[2]); - &movq (&QWP($A[$y][2],"edi"),@T[2]); # R[0][2] = C[2] ^ (~C[3] & C[4]); - - &movq (@T[0],@C[4]); - &pandn (@T[0],@C[0]); - &pxor (@T[0],@C[3]); - &movq (&QWP($A[$y][3],"edi"),@T[0]); # R[0][3] = C[3] ^ (~C[4] & C[0]); - - &movq (@T[1],@C[0]); - &pandn (@T[1],@C[1]); - &pxor (@T[1],@C[4]); - &movq (&QWP($A[$y][4],"edi"),@T[1]); # R[0][4] = C[4] ^ (~C[0] & C[1]); + &movq (&QWP($A[$y][0],"edi"),@T[0]); # R[0][0] = C[0] ^ (~C[1] & C[2]) ^ iotas[i]; + &movq (@T[0],@C[4]); + &pandn (@C[3],@C[4]); + &pandn (@C[4],@C[0]); + &pxor (@C[3],@T[1]); + &movq (&QWP($A[$y][1],"edi"),@C[2]); # R[0][1] = C[1] ^ (~C[2] & C[3]); + &pxor (@C[4],@T[2]); + &movq (@T[2],&QWP($A[0][$xrho],"esi")) if (defined($xrho)); + + &movq (&QWP($A[$y][2],"edi"),@C[3]); # R[0][2] = C[2] ^ (~C[3] & C[4]); + &pandn (@C[0],@C[1]); + &movq (&QWP($A[$y][3],"edi"),@C[4]); # R[0][3] = C[3] ^ (~C[4] & C[0]); + &pxor (@C[0],@T[0]); + &pxor (@T[2],&QWP(@D[$xrho],"esp")) if (defined($xrho)); + &movq (&QWP($A[$y][4],"edi"),@C[0]); # R[0][4] = C[4] ^ (~C[0] & C[1]); } - &Chi (0); + &Chi (0, 3); sub Rho() { ######### regular Rho step my $x = shift; - &movq (@C[0],&QWP($A[0][$x],"esi")); - &pxor (@C[0],&QWP(@D[$x],"esp")); - &movq (@T[0],@C[0]); - &psrlq (@C[0],64-$rhotates[0][$x]); - &psllq (@T[0],$rhotates[0][$x]); - &por (@C[0],@T[0]); # C[0] = ROL64(A[0][3] ^ D[3], rhotates[0][3]); - - &movq (@C[1],&QWP($A[1][($x+1)%5],"esi")); - &pxor (@C[1],&QWP(@D[($x+1)%5],"esp")); - &movq (@T[1],@C[1]); - &psrlq (@C[1],64-$rhotates[1][($x+1)%5]); - &psllq (@T[1],$rhotates[1][($x+1)%5]); - &por (@C[1],@T[1]); # C[1] = ROL64(A[1][4] ^ D[4], rhotates[1][4]); - + #&movq (@T[2],&QWP($A[0][$x],"esi")); # moved to Chi + #&pxor (@T[2],&QWP(@D[$x],"esp")); # moved to Chi + &movq (@C[0],@T[2]); + &psrlq (@T[2],64-$rhotates[0][$x]); + &movq (@C[1],&QWP($A[1][($x+1)%5],"esi")); + &psllq (@C[0],$rhotates[0][$x]); + &pxor (@C[1],&QWP(@D[($x+1)%5],"esp")); + &por (@C[0],@T[2]); # C[0] = ROL64(A[0][3] ^ D[3], rhotates[0][3]); + + &movq (@T[1],@C[1]); + &psrlq (@C[1],64-$rhotates[1][($x+1)%5]); &movq (@C[2],&QWP($A[2][($x+2)%5],"esi")); + &psllq (@T[1],$rhotates[1][($x+1)%5]); &pxor (@C[2],&QWP(@D[($x+2)%5],"esp")); + &por (@C[1],@T[1]); # C[1] = ROL64(A[1][4] ^ D[4], rhotates[1][4]); + &movq (@T[2],@C[2]); &psrlq (@C[2],64-$rhotates[2][($x+2)%5]); + &movq (@C[3],&QWP($A[3][($x+3)%5],"esi")); &psllq (@T[2],$rhotates[2][($x+2)%5]); + &pxor (@C[3],&QWP(@D[($x+3)%5],"esp")); &por (@C[2],@T[2]); # C[2] = ROL64(A[2][0] ^ D[0], rhotates[2][0]); - &movq (@C[3],&QWP($A[3][($x+3)%5],"esi")); - &pxor (@C[3],&QWP(@D[($x+3)%5],"esp")); - &movq (@T[0],@C[3]); - &psrlq (@C[3],64-$rhotates[3][($x+3)%5]); - &psllq (@T[0],$rhotates[3][($x+3)%5]); - &por (@C[3],@T[0]); # C[3] = ROL64(A[3][1] ^ D[1], rhotates[3][1]); - + &movq (@T[0],@C[3]); + &psrlq (@C[3],64-$rhotates[3][($x+3)%5]); &movq (@C[4],&QWP($A[4][($x+4)%5],"esi")); + &psllq (@T[0],$rhotates[3][($x+3)%5]); &pxor (@C[4],&QWP(@D[($x+4)%5],"esp")); + &por (@C[3],@T[0]); # C[3] = ROL64(A[3][1] ^ D[1], rhotates[3][1]); + &movq (@T[1],@C[4]); &psrlq (@C[4],64-$rhotates[4][($x+4)%5]); &psllq (@T[1],$rhotates[4][($x+4)%5]); &por (@C[4],@T[1]); # C[4] = ROL64(A[4][2] ^ D[2], rhotates[4][2]); } - &Rho (3); &Chi (1); - &Rho (1); &Chi (2); - &Rho (4); &Chi (3); - &Rho (2); #&Chi (4); + &Rho (3); &Chi (1, 1); + &Rho (1); &Chi (2, 4); + &Rho (4); &Chi (3, 2); + &Rho (2); ###&Chi (4); &movq (@T[0],@C[0]); ######### last Chi(4) is special + &xor ("edi","esi"); # &xchg ("esi","edi"); &movq (&QWP(@D[1],"esp"),@C[1]); + &xor ("esi","edi"); + &xor ("edi","esi"); &movq (@T[1],@C[1]); + &movq (@T[2],@C[2]); &pandn (@T[1],@C[2]); + &pandn (@T[2],@C[3]); &pxor (@C[0],@T[1]); - &movq (&QWP($A[4][0],"edi"),@C[0]); # R[4][0] = C[0] ^= (~C[1] & C[2]); - - &movq (@T[2],@C[2]); - &pandn (@T[2],@C[3]); - &pxor (@C[1],@T[2]); - &movq (&QWP($A[4][1],"edi"),@C[1]); # R[4][1] = C[1] ^= (~C[2] & C[3]); + &pxor (@C[1],@T[2]); &movq (@T[1],@C[3]); + &movq (&QWP($A[4][0],"esi"),@C[0]); # R[4][0] = C[0] ^= (~C[1] & C[2]); &pandn (@T[1],@C[4]); + &movq (&QWP($A[4][1],"esi"),@C[1]); # R[4][1] = C[1] ^= (~C[2] & C[3]); &pxor (@C[2],@T[1]); - &movq (&QWP($A[4][2],"edi"),@C[2]); # R[4][2] = C[2] ^= (~C[3] & C[4]); + &movq (@T[2],@C[4]); + &movq (&QWP($A[4][2],"esi"),@C[2]); # R[4][2] = C[2] ^= (~C[3] & C[4]); - &movq (@T[2],@C[4]); &pandn (@T[2],@T[0]); + &pandn (@T[0],&QWP(@D[1],"esp")); &pxor (@C[3],@T[2]); - &movq (&QWP($A[4][3],"edi"),@C[3]); # R[4][3] = C[3] ^= (~C[4] & D[0]); - - &pandn (@T[0],&QWP(@D[1],"esp")); - &pxor (@C[4],@T[0]); - &movq (&QWP($A[4][4],"edi"),@C[4]); # R[4][4] = C[4] ^= (~D[0] & D[1]); - - &xchg ("esi","edi"); - &dec ("ecx"); + &pxor (@C[4],@T[0]); + &movq (&QWP($A[4][3],"esi"),@C[3]); # R[4][3] = C[3] ^= (~C[4] & D[0]); + &sub ("ecx",1); + &movq (&QWP($A[4][4],"esi"),@C[4]); # R[4][4] = C[4] ^= (~D[0] & D[1]); &jnz (&label("loop")); &lea ("ebx",&DWP(-192,"ebx")); # rewind iotas -- cgit v1.2.3