mirror of
https://github.com/openssl/openssl.git
synced 2024-12-09 05:51:54 +08:00
32be631ca1
If one of the perlasm xlate drivers crashes, OpenSSL's build will currently swallow the error and silently truncate the output to however far the driver got. This will hopefully fail to build, but better to check such things. Handle this by checking for errors when closing STDOUT (which is a pipe to the xlate driver). Reviewed-by: Richard Levitte <levitte@openssl.org> Reviewed-by: Tim Hudson <tjh@openssl.org> Reviewed-by: Tomas Mraz <tmraz@fedoraproject.org> (Merged from https://github.com/openssl/openssl/pull/10883)
384 lines
9.1 KiB
Raku
384 lines
9.1 KiB
Raku
#! /usr/bin/env perl
|
||
# Copyright 2012-2018 The OpenSSL Project Authors. All Rights Reserved.
|
||
#
|
||
# Licensed under the Apache License 2.0 (the "License"). You may not use
|
||
# this file except in compliance with the License. You can obtain a copy
|
||
# in the file LICENSE in the source distribution or at
|
||
# https://www.openssl.org/source/license.html
|
||
|
||
|
||
# ====================================================================
|
||
# Written by Andy Polyakov <appro@openssl.org> for the OpenSSL
|
||
# project. The module is, however, dual licensed under OpenSSL and
|
||
# CRYPTOGAMS licenses depending on where you obtain it. For further
|
||
# details see http://www.openssl.org/~appro/cryptogams/.
|
||
# ====================================================================
|
||
|
||
# October 2012.
|
||
#
|
||
# SPARCv9 VIS3 Montgomery multiplication procedure suitable for T3 and
|
||
# onward. There are three new instructions used here: umulxhi,
|
||
# addxc[cc] and initializing store. On T3 RSA private key operations
|
||
# are 1.54/1.87/2.11/2.26 times faster for 512/1024/2048/4096-bit key
|
||
# lengths. This is without dedicated squaring procedure. On T4
|
||
# corresponding coefficients are 1.47/2.10/2.80/2.90x, which is mostly
|
||
# for reference purposes, because T4 has dedicated Montgomery
|
||
# multiplication and squaring *instructions* that deliver even more.
|
||
|
||
$output = pop and open STDOUT,">$output";
|
||
|
||
$frame = "STACK_FRAME";
|
||
$bias = "STACK_BIAS";
|
||
|
||
$code.=<<___;
|
||
#include "sparc_arch.h"
|
||
|
||
#ifdef __arch64__
|
||
.register %g2,#scratch
|
||
.register %g3,#scratch
|
||
#endif
|
||
|
||
.section ".text",#alloc,#execinstr
|
||
___
|
||
|
||
($n0,$m0,$m1,$lo0,$hi0, $lo1,$hi1,$aj,$alo,$nj,$nlo,$tj)=
|
||
(map("%g$_",(1..5)),map("%o$_",(0..5,7)));
|
||
|
||
# int bn_mul_mont(
|
||
$rp="%o0"; # BN_ULONG *rp,
|
||
$ap="%o1"; # const BN_ULONG *ap,
|
||
$bp="%o2"; # const BN_ULONG *bp,
|
||
$np="%o3"; # const BN_ULONG *np,
|
||
$n0p="%o4"; # const BN_ULONG *n0,
|
||
$num="%o5"; # int num); # caller ensures that num is even
|
||
# and >=6
|
||
$code.=<<___;
|
||
.globl bn_mul_mont_vis3
|
||
.align 32
|
||
bn_mul_mont_vis3:
|
||
add %sp, $bias, %g4 ! real top of stack
|
||
sll $num, 2, $num ! size in bytes
|
||
add $num, 63, %g5
|
||
andn %g5, 63, %g5 ! buffer size rounded up to 64 bytes
|
||
add %g5, %g5, %g1
|
||
add %g5, %g1, %g1 ! 3*buffer size
|
||
sub %g4, %g1, %g1
|
||
andn %g1, 63, %g1 ! align at 64 byte
|
||
sub %g1, $frame, %g1 ! new top of stack
|
||
sub %g1, %g4, %g1
|
||
|
||
save %sp, %g1, %sp
|
||
___
|
||
|
||
# +-------------------------------+<----- %sp
|
||
# . .
|
||
# +-------------------------------+<----- aligned at 64 bytes
|
||
# | __int64 tmp[0] |
|
||
# +-------------------------------+
|
||
# . .
|
||
# . .
|
||
# +-------------------------------+<----- aligned at 64 bytes
|
||
# | __int64 ap[1..0] | converted ap[]
|
||
# +-------------------------------+
|
||
# | __int64 np[1..0] | converted np[]
|
||
# +-------------------------------+
|
||
# | __int64 ap[3..2] |
|
||
# . .
|
||
# . .
|
||
# +-------------------------------+
|
||
($rp,$ap,$bp,$np,$n0p,$num)=map("%i$_",(0..5));
|
||
($t0,$t1,$t2,$t3,$cnt,$tp,$bufsz,$anp)=map("%l$_",(0..7));
|
||
($ovf,$i)=($t0,$t1);
|
||
$code.=<<___;
|
||
ld [$n0p+0], $t0 ! pull n0[0..1] value
|
||
add %sp, $bias+$frame, $tp
|
||
ld [$n0p+4], $t1
|
||
add $tp, %g5, $anp
|
||
ld [$bp+0], $t2 ! m0=bp[0]
|
||
sllx $t1, 32, $n0
|
||
ld [$bp+4], $t3
|
||
or $t0, $n0, $n0
|
||
add $bp, 8, $bp
|
||
|
||
ld [$ap+0], $t0 ! ap[0]
|
||
sllx $t3, 32, $m0
|
||
ld [$ap+4], $t1
|
||
or $t2, $m0, $m0
|
||
|
||
ld [$ap+8], $t2 ! ap[1]
|
||
sllx $t1, 32, $aj
|
||
ld [$ap+12], $t3
|
||
or $t0, $aj, $aj
|
||
add $ap, 16, $ap
|
||
stx $aj, [$anp] ! converted ap[0]
|
||
|
||
mulx $aj, $m0, $lo0 ! ap[0]*bp[0]
|
||
umulxhi $aj, $m0, $hi0
|
||
|
||
ld [$np+0], $t0 ! np[0]
|
||
sllx $t3, 32, $aj
|
||
ld [$np+4], $t1
|
||
or $t2, $aj, $aj
|
||
|
||
ld [$np+8], $t2 ! np[1]
|
||
sllx $t1, 32, $nj
|
||
ld [$np+12], $t3
|
||
or $t0, $nj, $nj
|
||
add $np, 16, $np
|
||
stx $nj, [$anp+8] ! converted np[0]
|
||
|
||
mulx $lo0, $n0, $m1 ! "tp[0]"*n0
|
||
stx $aj, [$anp+16] ! converted ap[1]
|
||
|
||
mulx $aj, $m0, $alo ! ap[1]*bp[0]
|
||
umulxhi $aj, $m0, $aj ! ahi=aj
|
||
|
||
mulx $nj, $m1, $lo1 ! np[0]*m1
|
||
umulxhi $nj, $m1, $hi1
|
||
|
||
sllx $t3, 32, $nj
|
||
or $t2, $nj, $nj
|
||
stx $nj, [$anp+24] ! converted np[1]
|
||
add $anp, 32, $anp
|
||
|
||
addcc $lo0, $lo1, $lo1
|
||
addxc %g0, $hi1, $hi1
|
||
|
||
mulx $nj, $m1, $nlo ! np[1]*m1
|
||
umulxhi $nj, $m1, $nj ! nhi=nj
|
||
|
||
ba .L1st
|
||
sub $num, 24, $cnt ! cnt=num-3
|
||
|
||
.align 16
|
||
.L1st:
|
||
ld [$ap+0], $t0 ! ap[j]
|
||
addcc $alo, $hi0, $lo0
|
||
ld [$ap+4], $t1
|
||
addxc $aj, %g0, $hi0
|
||
|
||
sllx $t1, 32, $aj
|
||
add $ap, 8, $ap
|
||
or $t0, $aj, $aj
|
||
stx $aj, [$anp] ! converted ap[j]
|
||
|
||
ld [$np+0], $t2 ! np[j]
|
||
addcc $nlo, $hi1, $lo1
|
||
ld [$np+4], $t3
|
||
addxc $nj, %g0, $hi1 ! nhi=nj
|
||
|
||
sllx $t3, 32, $nj
|
||
add $np, 8, $np
|
||
mulx $aj, $m0, $alo ! ap[j]*bp[0]
|
||
or $t2, $nj, $nj
|
||
umulxhi $aj, $m0, $aj ! ahi=aj
|
||
stx $nj, [$anp+8] ! converted np[j]
|
||
add $anp, 16, $anp ! anp++
|
||
|
||
mulx $nj, $m1, $nlo ! np[j]*m1
|
||
addcc $lo0, $lo1, $lo1 ! np[j]*m1+ap[j]*bp[0]
|
||
umulxhi $nj, $m1, $nj ! nhi=nj
|
||
addxc %g0, $hi1, $hi1
|
||
stx $lo1, [$tp] ! tp[j-1]
|
||
add $tp, 8, $tp ! tp++
|
||
|
||
brnz,pt $cnt, .L1st
|
||
sub $cnt, 8, $cnt ! j--
|
||
!.L1st
|
||
addcc $alo, $hi0, $lo0
|
||
addxc $aj, %g0, $hi0 ! ahi=aj
|
||
|
||
addcc $nlo, $hi1, $lo1
|
||
addxc $nj, %g0, $hi1
|
||
addcc $lo0, $lo1, $lo1 ! np[j]*m1+ap[j]*bp[0]
|
||
addxc %g0, $hi1, $hi1
|
||
stx $lo1, [$tp] ! tp[j-1]
|
||
add $tp, 8, $tp
|
||
|
||
addcc $hi0, $hi1, $hi1
|
||
addxc %g0, %g0, $ovf ! upmost overflow bit
|
||
stx $hi1, [$tp]
|
||
add $tp, 8, $tp
|
||
|
||
ba .Louter
|
||
sub $num, 16, $i ! i=num-2
|
||
|
||
.align 16
|
||
.Louter:
|
||
ld [$bp+0], $t2 ! m0=bp[i]
|
||
ld [$bp+4], $t3
|
||
|
||
sub $anp, $num, $anp ! rewind
|
||
sub $tp, $num, $tp
|
||
sub $anp, $num, $anp
|
||
|
||
add $bp, 8, $bp
|
||
sllx $t3, 32, $m0
|
||
ldx [$anp+0], $aj ! ap[0]
|
||
or $t2, $m0, $m0
|
||
ldx [$anp+8], $nj ! np[0]
|
||
|
||
mulx $aj, $m0, $lo0 ! ap[0]*bp[i]
|
||
ldx [$tp], $tj ! tp[0]
|
||
umulxhi $aj, $m0, $hi0
|
||
ldx [$anp+16], $aj ! ap[1]
|
||
addcc $lo0, $tj, $lo0 ! ap[0]*bp[i]+tp[0]
|
||
mulx $aj, $m0, $alo ! ap[1]*bp[i]
|
||
addxc %g0, $hi0, $hi0
|
||
mulx $lo0, $n0, $m1 ! tp[0]*n0
|
||
umulxhi $aj, $m0, $aj ! ahi=aj
|
||
mulx $nj, $m1, $lo1 ! np[0]*m1
|
||
umulxhi $nj, $m1, $hi1
|
||
ldx [$anp+24], $nj ! np[1]
|
||
add $anp, 32, $anp
|
||
addcc $lo1, $lo0, $lo1
|
||
mulx $nj, $m1, $nlo ! np[1]*m1
|
||
addxc %g0, $hi1, $hi1
|
||
umulxhi $nj, $m1, $nj ! nhi=nj
|
||
|
||
ba .Linner
|
||
sub $num, 24, $cnt ! cnt=num-3
|
||
.align 16
|
||
.Linner:
|
||
addcc $alo, $hi0, $lo0
|
||
ldx [$tp+8], $tj ! tp[j]
|
||
addxc $aj, %g0, $hi0 ! ahi=aj
|
||
ldx [$anp+0], $aj ! ap[j]
|
||
addcc $nlo, $hi1, $lo1
|
||
mulx $aj, $m0, $alo ! ap[j]*bp[i]
|
||
addxc $nj, %g0, $hi1 ! nhi=nj
|
||
ldx [$anp+8], $nj ! np[j]
|
||
add $anp, 16, $anp
|
||
umulxhi $aj, $m0, $aj ! ahi=aj
|
||
addcc $lo0, $tj, $lo0 ! ap[j]*bp[i]+tp[j]
|
||
mulx $nj, $m1, $nlo ! np[j]*m1
|
||
addxc %g0, $hi0, $hi0
|
||
umulxhi $nj, $m1, $nj ! nhi=nj
|
||
addcc $lo1, $lo0, $lo1 ! np[j]*m1+ap[j]*bp[i]+tp[j]
|
||
addxc %g0, $hi1, $hi1
|
||
stx $lo1, [$tp] ! tp[j-1]
|
||
add $tp, 8, $tp
|
||
brnz,pt $cnt, .Linner
|
||
sub $cnt, 8, $cnt
|
||
!.Linner
|
||
ldx [$tp+8], $tj ! tp[j]
|
||
addcc $alo, $hi0, $lo0
|
||
addxc $aj, %g0, $hi0 ! ahi=aj
|
||
addcc $lo0, $tj, $lo0 ! ap[j]*bp[i]+tp[j]
|
||
addxc %g0, $hi0, $hi0
|
||
|
||
addcc $nlo, $hi1, $lo1
|
||
addxc $nj, %g0, $hi1 ! nhi=nj
|
||
addcc $lo1, $lo0, $lo1 ! np[j]*m1+ap[j]*bp[i]+tp[j]
|
||
addxc %g0, $hi1, $hi1
|
||
stx $lo1, [$tp] ! tp[j-1]
|
||
|
||
subcc %g0, $ovf, %g0 ! move upmost overflow to CCR.xcc
|
||
addxccc $hi1, $hi0, $hi1
|
||
addxc %g0, %g0, $ovf
|
||
stx $hi1, [$tp+8]
|
||
add $tp, 16, $tp
|
||
|
||
brnz,pt $i, .Louter
|
||
sub $i, 8, $i
|
||
|
||
sub $anp, $num, $anp ! rewind
|
||
sub $tp, $num, $tp
|
||
sub $anp, $num, $anp
|
||
ba .Lsub
|
||
subcc $num, 8, $cnt ! cnt=num-1 and clear CCR.xcc
|
||
|
||
.align 16
|
||
.Lsub:
|
||
ldx [$tp], $tj
|
||
add $tp, 8, $tp
|
||
ldx [$anp+8], $nj
|
||
add $anp, 16, $anp
|
||
subccc $tj, $nj, $t2 ! tp[j]-np[j]
|
||
srlx $tj, 32, $tj
|
||
srlx $nj, 32, $nj
|
||
subccc $tj, $nj, $t3
|
||
add $rp, 8, $rp
|
||
st $t2, [$rp-4] ! reverse order
|
||
st $t3, [$rp-8]
|
||
brnz,pt $cnt, .Lsub
|
||
sub $cnt, 8, $cnt
|
||
|
||
sub $anp, $num, $anp ! rewind
|
||
sub $tp, $num, $tp
|
||
sub $anp, $num, $anp
|
||
sub $rp, $num, $rp
|
||
|
||
subccc $ovf, %g0, $ovf ! handle upmost overflow bit
|
||
ba .Lcopy
|
||
sub $num, 8, $cnt
|
||
|
||
.align 16
|
||
.Lcopy: ! conditional copy
|
||
ld [$tp+0], $t0
|
||
ld [$tp+4], $t1
|
||
ld [$rp+0], $t2
|
||
ld [$rp+4], $t3
|
||
stx %g0, [$tp] ! zap
|
||
add $tp, 8, $tp
|
||
stx %g0, [$anp] ! zap
|
||
stx %g0, [$anp+8]
|
||
add $anp, 16, $anp
|
||
movcs %icc, $t0, $t2
|
||
movcs %icc, $t1, $t3
|
||
st $t3, [$rp+0] ! flip order
|
||
st $t2, [$rp+4]
|
||
add $rp, 8, $rp
|
||
brnz $cnt, .Lcopy
|
||
sub $cnt, 8, $cnt
|
||
|
||
mov 1, %o0
|
||
ret
|
||
restore
|
||
.type bn_mul_mont_vis3, #function
|
||
.size bn_mul_mont_vis3, .-bn_mul_mont_vis3
|
||
.asciz "Montgomery Multiplication for SPARCv9 VIS3, CRYPTOGAMS by <appro\@openssl.org>"
|
||
.align 4
|
||
___
|
||
|
||
# Purpose of these subroutines is to explicitly encode VIS instructions,
|
||
# so that one can compile the module without having to specify VIS
|
||
# extensions on compiler command line, e.g. -xarch=v9 vs. -xarch=v9a.
|
||
# Idea is to reserve for option to produce "universal" binary and let
|
||
# programmer detect if current CPU is VIS capable at run-time.
|
||
sub unvis3 {
|
||
my ($mnemonic,$rs1,$rs2,$rd)=@_;
|
||
my %bias = ( "g" => 0, "o" => 8, "l" => 16, "i" => 24 );
|
||
my ($ref,$opf);
|
||
my %visopf = ( "addxc" => 0x011,
|
||
"addxccc" => 0x013,
|
||
"umulxhi" => 0x016 );
|
||
|
||
$ref = "$mnemonic\t$rs1,$rs2,$rd";
|
||
|
||
if ($opf=$visopf{$mnemonic}) {
|
||
foreach ($rs1,$rs2,$rd) {
|
||
return $ref if (!/%([goli])([0-9])/);
|
||
$_=$bias{$1}+$2;
|
||
}
|
||
|
||
return sprintf ".word\t0x%08x !%s",
|
||
0x81b00000|$rd<<25|$rs1<<14|$opf<<5|$rs2,
|
||
$ref;
|
||
} else {
|
||
return $ref;
|
||
}
|
||
}
|
||
|
||
foreach (split("\n",$code)) {
|
||
s/\`([^\`]*)\`/eval $1/ge;
|
||
|
||
s/\b(umulxhi|addxc[c]{0,2})\s+(%[goli][0-7]),\s*(%[goli][0-7]),\s*(%[goli][0-7])/
|
||
&unvis3($1,$2,$3,$4)
|
||
/ge;
|
||
|
||
print $_,"\n";
|
||
}
|
||
|
||
close STDOUT or die "error closing STDOUT";
|