#!/usr/bin/env perl # ==================================================================== # Written by Andy Polyakov <appro@fy.chalmers.se> for the OpenSSL # project. The module is, however, dual licensed under OpenSSL and # CRYPTOGAMS licenses depending on where you obtain it. For further # details see http://www.openssl.org/~appro/cryptogams/. # ==================================================================== # October 2005 # # "Teaser" Montgomery multiplication module for UltraSPARC. Why FPU? # Because unlike integer multiplier, which simply stalls whole CPU, # FPU is fully pipelined and can effectively emit 48 bit partial # product every cycle. Why not blended SPARC v9? One can argue that # making this module dependent on UltraSPARC VIS extension limits its # binary compatibility. Well yes, it does exclude SPARC64 prior-V(!) # implementations from compatibility matrix. But the rest, whole Sun # UltraSPARC family and brand new Fujitsu's SPARC64 V, all support # VIS extension instructions used in this module. This is considered # good enough to not care about HAL SPARC64 users [if any] who have # integer-only pure SPARCv9 module to "fall down" to. # USI&II cores currently exhibit uniform 2x improvement [over pre- # bn_mul_mont codebase] for all key lengths and benchmarks. On USIII # performance improves few percents for shorter keys and worsens few # percents for longer keys. This is because USIII integer multiplier # is >3x faster than USI&II one, which is harder to match [but see # TODO list below]. It should also be noted that SPARC64 V features # out-of-order execution, which *might* mean that integer multiplier # is pipelined, which in turn *might* be impossible to match... On # additional note, SPARC64 V implements FP Multiply-Add instruction, # which is perfectly usable in this context... In other words, as far # as Fujitsu SPARC64 V goes, talk to the author:-) # The implementation implies following "non-natural" limitations on # input arguments: # - num may not be less than 4; # - num has to be even; # Failure to meet either condition has no fatal effects, simply # doesn't give any performance gain. # TODO: # - modulo-schedule inner loop for better performance (on in-order # execution core such as UltraSPARC this shall result in further # noticeable(!) improvement); # - dedicated squaring procedure[?]; ###################################################################### # November 2006 # # Modulo-scheduled inner loops allow to interleave floating point and # integer instructions and minimize Read-After-Write penalties. This # results in *further* 20-50% perfromance improvement [depending on # key length, more for longer keys] on USI&II cores and 30-80% - on # USIII&IV. $fname="bn_mul_mont_fpu"; $bits=32; for (@ARGV) { $bits=64 if (/\-m64/ || /\-xarch\=v9/); } if ($bits==64) { $bias=2047; $frame=192; } else { $bias=0; $frame=128; # 96 rounded up to largest known cache-line } $locals=64; # In order to provide for 32-/64-bit ABI duality, I keep integers wider # than 32 bit in %g1-%g4 and %o0-%o5. %l0-%l7 and %i0-%i5 are used # exclusively for pointers, indexes and other small values... # int bn_mul_mont( $rp="%i0"; # BN_ULONG *rp, $ap="%i1"; # const BN_ULONG *ap, $bp="%i2"; # const BN_ULONG *bp, $np="%i3"; # const BN_ULONG *np, $n0="%i4"; # const BN_ULONG *n0, $num="%i5"; # int num); $tp="%l0"; # t[num] $ap_l="%l1"; # a[num],n[num] are smashed to 32-bit words and saved $ap_h="%l2"; # to these four vectors as double-precision FP values. $np_l="%l3"; # This way a bunch of fxtods are eliminated in second $np_h="%l4"; # loop and L1-cache aliasing is minimized... $i="%l5"; $j="%l6"; $mask="%l7"; # 16-bit mask, 0xffff $n0="%g4"; # reassigned(!) to "64-bit" register $carry="%i4"; # %i4 reused(!) for a carry bit # FP register naming chart # # ..HILO # dcba # -------- # LOa # LOb # LOc # LOd # HIa # HIb # HIc # HId # ..a # ..b $ba="%f0"; $bb="%f2"; $bc="%f4"; $bd="%f6"; $na="%f8"; $nb="%f10"; $nc="%f12"; $nd="%f14"; $alo="%f16"; $alo_="%f17"; $ahi="%f18"; $ahi_="%f19"; $nlo="%f20"; $nlo_="%f21"; $nhi="%f22"; $nhi_="%f23"; $dota="%f24"; $dotb="%f26"; $aloa="%f32"; $alob="%f34"; $aloc="%f36"; $alod="%f38"; $ahia="%f40"; $ahib="%f42"; $ahic="%f44"; $ahid="%f46"; $nloa="%f48"; $nlob="%f50"; $nloc="%f52"; $nlod="%f54"; $nhia="%f56"; $nhib="%f58"; $nhic="%f60"; $nhid="%f62"; $ASI_FL16_P=0xD2; # magic ASI value to engage 16-bit FP load $code=<<___; .section ".text",#alloc,#execinstr .global $fname .align 32 $fname: save %sp,-$frame-$locals,%sp cmp $num,4 bl,a,pn %icc,.Lret clr %i0 andcc $num,1,%g0 ! $num has to be even... bnz,a,pn %icc,.Lret clr %i0 ! signal "unsupported input value" srl $num,1,$num sethi %hi(0xffff),$mask ld [%i4+0],$n0 ! $n0 reassigned, remember? or $mask,%lo(0xffff),$mask ld [%i4+4],%o0 sllx %o0,32,%o0 or %o0,$n0,$n0 ! $n0=n0[1].n0[0] sll $num,3,$num ! num*=8 add %sp,$bias,%o0 ! real top of stack sll $num,2,%o1 add %o1,$num,%o1 ! %o1=num*5 sub %o0,%o1,%o0 and %o0,-2048,%o0 ! optimize TLB utilization sub %o0,$bias,%sp ! alloca(5*num*8) rd %asi,%o7 ! save %asi add %sp,$bias+$frame+$locals,$tp add $tp,$num,$ap_l add $ap_l,$num,$ap_l ! [an]p_[lh] point at the vectors' ends ! add $ap_l,$num,$ap_h add $ap_h,$num,$np_l add $np_l,$num,$np_h wr %g0,$ASI_FL16_P,%asi ! setup %asi for 16-bit FP loads add $rp,$num,$rp ! readjust input pointers to point add $ap,$num,$ap ! at the ends too... add $bp,$num,$bp add $np,$num,$np stx %o7,[%sp+$bias+$frame+48] ! save %asi sub %g0,$num,$i ! i=-num sub %g0,$num,$j ! j=-num add $ap,$j,%o3 add $bp,$i,%o4 ld [%o3+4],%g1 ! bp[0] ld [%o3+0],%o0 ld [%o4+4],%g5 ! ap[0] sllx %g1,32,%g1 ld [%o4+0],%o1 sllx %g5,32,%g5 or %g1,%o0,%o0 or %g5,%o1,%o1 add $np,$j,%o5 mulx %o1,%o0,%o0 ! ap[0]*bp[0] mulx $n0,%o0,%o0 ! ap[0]*bp[0]*n0 stx %o0,[%sp+$bias+$frame+0] ld [%o3+0],$alo_ ! load a[j] as pair of 32-bit words fzeros $alo ld [%o3+4],$ahi_ fzeros $ahi ld [%o5+0],$nlo_ ! load n[j] as pair of 32-bit words fzeros $nlo ld [%o5+4],$nhi_ fzeros $nhi ! transfer b[i] to FPU as 4x16-bit values ldda [%o4+2]%asi,$ba fxtod $alo,$alo ldda [%o4+0]%asi,$bb fxtod $ahi,$ahi ldda [%o4+6]%asi,$bc fxtod $nlo,$nlo ldda [%o4+4]%asi,$bd fxtod $nhi,$nhi ! transfer ap[0]*b[0]*n0 to FPU as 4x16-bit values ldda [%sp+$bias+$frame+6]%asi,$na fxtod $ba,$ba ldda [%sp+$bias+$frame+4]%asi,$nb fxtod $bb,$bb ldda [%sp+$bias+$frame+2]%asi,$nc fxtod $bc,$bc ldda [%sp+$bias+$frame+0]%asi,$nd fxtod $bd,$bd std $alo,[$ap_l+$j] ! save smashed ap[j] in double format fxtod $na,$na std $ahi,[$ap_h+$j] fxtod $nb,$nb std $nlo,[$np_l+$j] ! save smashed np[j] in double format fxtod $nc,$nc std $nhi,[$np_h+$j] fxtod $nd,$nd fmuld $alo,$ba,$aloa fmuld $nlo,$na,$nloa fmuld $alo,$bb,$alob fmuld $nlo,$nb,$nlob fmuld $alo,$bc,$aloc faddd $aloa,$nloa,$nloa fmuld $nlo,$nc,$nloc fmuld $alo,$bd,$alod faddd $alob,$nlob,$nlob fmuld $nlo,$nd,$nlod fmuld $ahi,$ba,$ahia faddd $aloc,$nloc,$nloc fmuld $nhi,$na,$nhia fmuld $ahi,$bb,$ahib faddd $alod,$nlod,$nlod fmuld $nhi,$nb,$nhib fmuld $ahi,$bc,$ahic faddd $ahia,$nhia,$nhia fmuld $nhi,$nc,$nhic fmuld $ahi,$bd,$ahid faddd $ahib,$nhib,$nhib fmuld $nhi,$nd,$nhid faddd $ahic,$nhic,$dota ! $nhic faddd $ahid,$nhid,$dotb ! $nhid faddd $nloc,$nhia,$nloc faddd $nlod,$nhib,$nlod fdtox $nloa,$nloa fdtox $nlob,$nlob fdtox $nloc,$nloc fdtox $nlod,$nlod std $nloa,[%sp+$bias+$frame+0] add $j,8,$j std $nlob,[%sp+$bias+$frame+8] add $ap,$j,%o4 std $nloc,[%sp+$bias+$frame+16] add $np,$j,%o5 std $nlod,[%sp+$bias+$frame+24] ld [%o4+0],$alo_ ! load a[j] as pair of 32-bit words fzeros $alo ld [%o4+4],$ahi_ fzeros $ahi ld [%o5+0],$nlo_ ! load n[j] as pair of 32-bit words fzeros $nlo ld [%o5+4],$nhi_ fzeros $nhi fxtod $alo,$alo fxtod $ahi,$ahi fxtod $nlo,$nlo fxtod $nhi,$nhi ldx [%sp+$bias+$frame+0],%o0 fmuld $alo,$ba,$aloa ldx [%sp+$bias+$frame+8],%o1 fmuld $nlo,$na,$nloa ldx [%sp+$bias+$frame+16],%o2 fmuld $alo,$bb,$alob ldx [%sp+$bias+$frame+24],%o3 fmuld $nlo,$nb,$nlob srlx %o0,16,%o7 std $alo,[$ap_l+$j] ! save smashed ap[j] in double format fmuld $alo,$bc,$aloc add %o7,%o1,%o1 std $ahi,[$ap_h+$j] faddd $aloa,$nloa,$nloa fmuld $nlo,$nc,$nloc srlx %o1,16,%o7 std $nlo,[$np_l+$j] ! save smashed np[j] in double format fmuld $alo,$bd,$alod add %o7,%o2,%o2 std $nhi,[$np_h+$j] faddd $alob,$nlob,$nlob fmuld $nlo,$nd,$nlod srlx %o2,16,%o7 fmuld $ahi,$ba,$ahia add %o7,%o3,%o3 ! %o3.%o2[0..15].%o1[0..15].%o0[0..15] faddd $aloc,$nloc,$nloc fmuld $nhi,$na,$nhia !and %o0,$mask,%o0 !and %o1,$mask,%o1 !and %o2,$mask,%o2 !sllx %o1,16,%o1 !sllx %o2,32,%o2 !sllx %o3,48,%o7 !or %o1,%o0,%o0 !or %o2,%o0,%o0 !or %o7,%o0,%o0 ! 64-bit result srlx %o3,16,%g1 ! 34-bit carry fmuld $ahi,$bb,$ahib faddd $alod,$nlod,$nlod fmuld $nhi,$nb,$nhib fmuld $ahi,$bc,$ahic faddd $ahia,$nhia,$nhia fmuld $nhi,$nc,$nhic fmuld $ahi,$bd,$ahid faddd $ahib,$nhib,$nhib fmuld $nhi,$nd,$nhid faddd $dota,$nloa,$nloa faddd $dotb,$nlob,$nlob faddd $ahic,$nhic,$dota ! $nhic faddd $ahid,$nhid,$dotb ! $nhid faddd $nloc,$nhia,$nloc faddd $nlod,$nhib,$nlod fdtox $nloa,$nloa fdtox $nlob,$nlob fdtox $nloc,$nloc fdtox $nlod,$nlod std $nloa,[%sp+$bias+$frame+0] std $nlob,[%sp+$bias+$frame+8] addcc $j,8,$j std $nloc,[%sp+$bias+$frame+16] bz,pn %icc,.L1stskip std $nlod,[%sp+$bias+$frame+24] .align 32 ! incidentally already aligned ! .L1st: add $ap,$j,%o4 add $np,$j,%o5 ld [%o4+0],$alo_ ! load a[j] as pair of 32-bit words fzeros $alo ld [%o4+4],$ahi_ fzeros $ahi ld [%o5+0],$nlo_ ! load n[j] as pair of 32-bit words fzeros $nlo ld [%o5+4],$nhi_ fzeros $nhi fxtod $alo,$alo fxtod $ahi,$ahi fxtod $nlo,$nlo fxtod $nhi,$nhi ldx [%sp+$bias+$frame+0],%o0 fmuld $alo,$ba,$aloa ldx [%sp+$bias+$frame+8],%o1 fmuld $nlo,$na,$nloa ldx [%sp+$bias+$frame+16],%o2 fmuld $alo,$bb,$alob ldx [%sp+$bias+$frame+24],%o3 fmuld $nlo,$nb,$nlob srlx %o0,16,%o7 std $alo,[$ap_l+$j] ! save smashed ap[j] in double format fmuld $alo,$bc,$aloc add %o7,%o1,%o1 std $ahi,[$ap_h+$j] faddd $aloa,$nloa,$nloa fmuld $nlo,$nc,$nloc srlx %o1,16,%o7 std $nlo,[$np_l+$j] ! save smashed np[j] in double format fmuld $alo,$bd,$alod add %o7,%o2,%o2 std $nhi,[$np_h+$j] faddd $alob,$nlob,$nlob fmuld $nlo,$nd,$nlod srlx %o2,16,%o7 fmuld $ahi,$ba,$ahia add %o7,%o3,%o3 ! %o3.%o2[0..15].%o1[0..15].%o0[0..15] and %o0,$mask,%o0 faddd $aloc,$nloc,$nloc fmuld $nhi,$na,$nhia and %o1,$mask,%o1 and %o2,$mask,%o2 fmuld $ahi,$bb,$ahib sllx %o1,16,%o1 faddd $alod,$nlod,$nlod fmuld $nhi,$nb,$nhib sllx %o2,32,%o2 fmuld $ahi,$bc,$ahic sllx %o3,48,%o7 or %o1,%o0,%o0 faddd $ahia,$nhia,$nhia fmuld $nhi,$nc,$nhic or %o2,%o0,%o0 fmuld $ahi,$bd,$ahid or %o7,%o0,%o0 ! 64-bit result faddd $ahib,$nhib,$nhib fmuld $nhi,$nd,$nhid addcc %g1,%o0,%o0 faddd $dota,$nloa,$nloa srlx %o3,16,%g1 ! 34-bit carry faddd $dotb,$nlob,$nlob bcs,a %xcc,.+8 add %g1,1,%g1 stx %o0,[$tp] ! tp[j-1]= faddd $ahic,$nhic,$dota ! $nhic faddd $ahid,$nhid,$dotb ! $nhid faddd $nloc,$nhia,$nloc faddd $nlod,$nhib,$nlod fdtox $nloa,$nloa fdtox $nlob,$nlob fdtox $nloc,$nloc fdtox $nlod,$nlod std $nloa,[%sp+$bias+$frame+0] std $nlob,[%sp+$bias+$frame+8] std $nloc,[%sp+$bias+$frame+16] std $nlod,[%sp+$bias+$frame+24] addcc $j,8,$j bnz,pt %icc,.L1st add $tp,8,$tp .L1stskip: fdtox $dota,$dota fdtox $dotb,$dotb ldx [%sp+$bias+$frame+0],%o0 ldx [%sp+$bias+$frame+8],%o1 ldx [%sp+$bias+$frame+16],%o2 ldx [%sp+$bias+$frame+24],%o3 srlx %o0,16,%o7 std $dota,[%sp+$bias+$frame+32] add %o7,%o1,%o1 std $dotb,[%sp+$bias+$frame+40] srlx %o1,16,%o7 add %o7,%o2,%o2 srlx %o2,16,%o7 add %o7,%o3,%o3 ! %o3.%o2[0..15].%o1[0..15].%o0[0..15] and %o0,$mask,%o0 and %o1,$mask,%o1 and %o2,$mask,%o2 sllx %o1,16,%o1 sllx %o2,32,%o2 sllx %o3,48,%o7 or %o1,%o0,%o0 or %o2,%o0,%o0 or %o7,%o0,%o0 ! 64-bit result ldx [%sp+$bias+$frame+32],%o4 addcc %g1,%o0,%o0 ldx [%sp+$bias+$frame+40],%o5 srlx %o3,16,%g1 ! 34-bit carry bcs,a %xcc,.+8 add %g1,1,%g1 stx %o0,[$tp] ! tp[j-1]= add $tp,8,$tp srlx %o4,16,%o7 add %o7,%o5,%o5 and %o4,$mask,%o4 sllx %o5,16,%o7 or %o7,%o4,%o4 addcc %g1,%o4,%o4 srlx %o5,48,%g1 bcs,a %xcc,.+8 add %g1,1,%g1 mov %g1,$carry stx %o4,[$tp] ! tp[num-1]= ba .Louter add $i,8,$i .align 32 .Louter: sub %g0,$num,$j ! j=-num add %sp,$bias+$frame+$locals,$tp add $ap,$j,%o3 add $bp,$i,%o4 ld [%o3+4],%g1 ! bp[i] ld [%o3+0],%o0 ld [%o4+4],%g5 ! ap[0] sllx %g1,32,%g1 ld [%o4+0],%o1 sllx %g5,32,%g5 or %g1,%o0,%o0 or %g5,%o1,%o1 ldx [$tp],%o2 ! tp[0] mulx %o1,%o0,%o0 addcc %o2,%o0,%o0 mulx $n0,%o0,%o0 ! (ap[0]*bp[i]+t[0])*n0 stx %o0,[%sp+$bias+$frame+0] ! transfer b[i] to FPU as 4x16-bit values ldda [%o4+2]%asi,$ba ldda [%o4+0]%asi,$bb ldda [%o4+6]%asi,$bc ldda [%o4+4]%asi,$bd ! transfer (ap[0]*b[i]+t[0])*n0 to FPU as 4x16-bit values ldda [%sp+$bias+$frame+6]%asi,$na fxtod $ba,$ba ldda [%sp+$bias+$frame+4]%asi,$nb fxtod $bb,$bb ldda [%sp+$bias+$frame+2]%asi,$nc fxtod $bc,$bc ldda [%sp+$bias+$frame+0]%asi,$nd fxtod $bd,$bd ldd [$ap_l+$j],$alo ! load a[j] in double format fxtod $na,$na ldd [$ap_h+$j],$ahi fxtod $nb,$nb ldd [$np_l+$j],$nlo ! load n[j] in double format fxtod $nc,$nc ldd [$np_h+$j],$nhi fxtod $nd,$nd fmuld $alo,$ba,$aloa fmuld $nlo,$na,$nloa fmuld $alo,$bb,$alob fmuld $nlo,$nb,$nlob fmuld $alo,$bc,$aloc faddd $aloa,$nloa,$nloa fmuld $nlo,$nc,$nloc fmuld $alo,$bd,$alod faddd $alob,$nlob,$nlob fmuld $nlo,$nd,$nlod fmuld $ahi,$ba,$ahia faddd $aloc,$nloc,$nloc fmuld $nhi,$na,$nhia fmuld $ahi,$bb,$ahib faddd $alod,$nlod,$nlod fmuld $nhi,$nb,$nhib fmuld $ahi,$bc,$ahic faddd $ahia,$nhia,$nhia fmuld $nhi,$nc,$nhic fmuld $ahi,$bd,$ahid faddd $ahib,$nhib,$nhib fmuld $nhi,$nd,$nhid faddd $ahic,$nhic,$dota ! $nhic faddd $ahid,$nhid,$dotb ! $nhid faddd $nloc,$nhia,$nloc faddd $nlod,$nhib,$nlod fdtox $nloa,$nloa fdtox $nlob,$nlob fdtox $nloc,$nloc fdtox $nlod,$nlod std $nloa,[%sp+$bias+$frame+0] std $nlob,[%sp+$bias+$frame+8] std $nloc,[%sp+$bias+$frame+16] add $j,8,$j std $nlod,[%sp+$bias+$frame+24] ldd [$ap_l+$j],$alo ! load a[j] in double format ldd [$ap_h+$j],$ahi ldd [$np_l+$j],$nlo ! load n[j] in double format ldd [$np_h+$j],$nhi fmuld $alo,$ba,$aloa fmuld $nlo,$na,$nloa fmuld $alo,$bb,$alob fmuld $nlo,$nb,$nlob fmuld $alo,$bc,$aloc ldx [%sp+$bias+$frame+0],%o0 faddd $aloa,$nloa,$nloa fmuld $nlo,$nc,$nloc ldx [%sp+$bias+$frame+8],%o1 fmuld $alo,$bd,$alod ldx [%sp+$bias+$frame+16],%o2 faddd $alob,$nlob,$nlob fmuld $nlo,$nd,$nlod ldx [%sp+$bias+$frame+24],%o3 fmuld $ahi,$ba,$ahia srlx %o0,16,%o7 faddd $aloc,$nloc,$nloc fmuld $nhi,$na,$nhia add %o7,%o1,%o1 fmuld $ahi,$bb,$ahib srlx %o1,16,%o7 faddd $alod,$nlod,$nlod fmuld $nhi,$nb,$nhib add %o7,%o2,%o2 fmuld $ahi,$bc,$ahic srlx %o2,16,%o7 faddd $ahia,$nhia,$nhia fmuld $nhi,$nc,$nhic add %o7,%o3,%o3 ! %o3.%o2[0..15].%o1[0..15].%o0[0..15] ! why? and %o0,$mask,%o0 fmuld $ahi,$bd,$ahid and %o1,$mask,%o1 and %o2,$mask,%o2 faddd $ahib,$nhib,$nhib fmuld $nhi,$nd,$nhid sllx %o1,16,%o1 faddd $dota,$nloa,$nloa sllx %o2,32,%o2 faddd $dotb,$nlob,$nlob sllx %o3,48,%o7 or %o1,%o0,%o0 faddd $ahic,$nhic,$dota ! $nhic or %o2,%o0,%o0 faddd $ahid,$nhid,$dotb ! $nhid or %o7,%o0,%o0 ! 64-bit result ldx [$tp],%o7 faddd $nloc,$nhia,$nloc addcc %o7,%o0,%o0 ! end-of-why? faddd $nlod,$nhib,$nlod srlx %o3,16,%g1 ! 34-bit carry fdtox $nloa,$nloa bcs,a %xcc,.+8 add %g1,1,%g1 fdtox $nlob,$nlob fdtox $nloc,$nloc fdtox $nlod,$nlod std $nloa,[%sp+$bias+$frame+0] std $nlob,[%sp+$bias+$frame+8] addcc $j,8,$j std $nloc,[%sp+$bias+$frame+16] bz,pn %icc,.Linnerskip std $nlod,[%sp+$bias+$frame+24] ba .Linner nop .align 32 .Linner: ldd [$ap_l+$j],$alo ! load a[j] in double format ldd [$ap_h+$j],$ahi ldd [$np_l+$j],$nlo ! load n[j] in double format ldd [$np_h+$j],$nhi fmuld $alo,$ba,$aloa fmuld $nlo,$na,$nloa fmuld $alo,$bb,$alob fmuld $nlo,$nb,$nlob fmuld $alo,$bc,$aloc ldx [%sp+$bias+$frame+0],%o0 faddd $aloa,$nloa,$nloa fmuld $nlo,$nc,$nloc ldx [%sp+$bias+$frame+8],%o1 fmuld $alo,$bd,$alod ldx [%sp+$bias+$frame+16],%o2 faddd $alob,$nlob,$nlob fmuld $nlo,$nd,$nlod ldx [%sp+$bias+$frame+24],%o3 fmuld $ahi,$ba,$ahia srlx %o0,16,%o7 faddd $aloc,$nloc,$nloc fmuld $nhi,$na,$nhia add %o7,%o1,%o1 fmuld $ahi,$bb,$ahib srlx %o1,16,%o7 faddd $alod,$nlod,$nlod fmuld $nhi,$nb,$nhib add %o7,%o2,%o2 fmuld $ahi,$bc,$ahic srlx %o2,16,%o7 faddd $ahia,$nhia,$nhia fmuld $nhi,$nc,$nhic add %o7,%o3,%o3 ! %o3.%o2[0..15].%o1[0..15].%o0[0..15] and %o0,$mask,%o0 fmuld $ahi,$bd,$ahid and %o1,$mask,%o1 and %o2,$mask,%o2 faddd $ahib,$nhib,$nhib fmuld $nhi,$nd,$nhid sllx %o1,16,%o1 faddd $dota,$nloa,$nloa sllx %o2,32,%o2 faddd $dotb,$nlob,$nlob sllx %o3,48,%o7 or %o1,%o0,%o0 faddd $ahic,$nhic,$dota ! $nhic or %o2,%o0,%o0 faddd $ahid,$nhid,$dotb ! $nhid or %o7,%o0,%o0 ! 64-bit result faddd $nloc,$nhia,$nloc addcc %g1,%o0,%o0 ldx [$tp+8],%o7 ! tp[j] faddd $nlod,$nhib,$nlod srlx %o3,16,%g1 ! 34-bit carry fdtox $nloa,$nloa bcs,a %xcc,.+8 add %g1,1,%g1 fdtox $nlob,$nlob addcc %o7,%o0,%o0 fdtox $nloc,$nloc bcs,a %xcc,.+8 add %g1,1,%g1 stx %o0,[$tp] ! tp[j-1] fdtox $nlod,$nlod std $nloa,[%sp+$bias+$frame+0] std $nlob,[%sp+$bias+$frame+8] std $nloc,[%sp+$bias+$frame+16] addcc $j,8,$j std $nlod,[%sp+$bias+$frame+24] bnz,pt %icc,.Linner add $tp,8,$tp .Linnerskip: fdtox $dota,$dota fdtox $dotb,$dotb ldx [%sp+$bias+$frame+0],%o0 ldx [%sp+$bias+$frame+8],%o1 ldx [%sp+$bias+$frame+16],%o2 ldx [%sp+$bias+$frame+24],%o3 srlx %o0,16,%o7 std $dota,[%sp+$bias+$frame+32] add %o7,%o1,%o1 std $dotb,[%sp+$bias+$frame+40] srlx %o1,16,%o7 add %o7,%o2,%o2 srlx %o2,16,%o7 add %o7,%o3,%o3 ! %o3.%o2[0..15].%o1[0..15].%o0[0..15] and %o0,$mask,%o0 and %o1,$mask,%o1 and %o2,$mask,%o2 sllx %o1,16,%o1 sllx %o2,32,%o2 sllx %o3,48,%o7 or %o1,%o0,%o0 or %o2,%o0,%o0 ldx [%sp+$bias+$frame+32],%o4 or %o7,%o0,%o0 ! 64-bit result ldx [%sp+$bias+$frame+40],%o5 addcc %g1,%o0,%o0 ldx [$tp+8],%o7 ! tp[j] srlx %o3,16,%g1 ! 34-bit carry bcs,a %xcc,.+8 add %g1,1,%g1 addcc %o7,%o0,%o0 bcs,a %xcc,.+8 add %g1,1,%g1 stx %o0,[$tp] ! tp[j-1] add $tp,8,$tp srlx %o4,16,%o7 add %o7,%o5,%o5 and %o4,$mask,%o4 sllx %o5,16,%o7 or %o7,%o4,%o4 addcc %g1,%o4,%o4 srlx %o5,48,%g1 bcs,a %xcc,.+8 add %g1,1,%g1 addcc $carry,%o4,%o4 stx %o4,[$tp] ! tp[num-1] mov %g1,$carry bcs,a %xcc,.+8 add $carry,1,$carry addcc $i,8,$i bnz %icc,.Louter nop add $tp,8,$tp ! adjust tp to point at the end orn %g0,%g0,%g4 sub %g0,$num,%o7 ! n=-num ba .Lsub subcc %g0,%g0,%g0 ! clear %icc.c .align 32 .Lsub: ldx [$tp+%o7],%o0 add $np,%o7,%g1 ld [%g1+0],%o2 ld [%g1+4],%o3 srlx %o0,32,%o1 subccc %o0,%o2,%o2 add $rp,%o7,%g1 subccc %o1,%o3,%o3 st %o2,[%g1+0] add %o7,8,%o7 brnz,pt %o7,.Lsub st %o3,[%g1+4] subc $carry,0,%g4 sub %g0,$num,%o7 ! n=-num ba .Lcopy nop .align 32 .Lcopy: ldx [$tp+%o7],%o0 add $rp,%o7,%g1 ld [%g1+0],%o2 ld [%g1+4],%o3 stx %g0,[$tp+%o7] and %o0,%g4,%o0 srlx %o0,32,%o1 andn %o2,%g4,%o2 andn %o3,%g4,%o3 or %o2,%o0,%o0 or %o3,%o1,%o1 st %o0,[%g1+0] add %o7,8,%o7 brnz,pt %o7,.Lcopy st %o1,[%g1+4] sub %g0,$num,%o7 ! n=-num .Lzap: stx %g0,[$ap_l+%o7] stx %g0,[$ap_h+%o7] stx %g0,[$np_l+%o7] stx %g0,[$np_h+%o7] add %o7,8,%o7 brnz,pt %o7,.Lzap nop ldx [%sp+$bias+$frame+48],%o7 wr %g0,%o7,%asi ! restore %asi mov 1,%i0 .Lret: ret restore .type $fname,#function .size $fname,(.-$fname) .asciz "Montgomery Multipltication for UltraSPARC, CRYPTOGAMS by <appro\@openssl.org>" .align 32 ___ $code =~ s/\`([^\`]*)\`/eval($1)/gem; # Below substitution makes it possible to compile without demanding # VIS extentions on command line, e.g. -xarch=v9 vs. -xarch=v9a. I # dare to do this, because VIS capability is detected at run-time now # and this routine is not called on CPU not capable to execute it. Do # note that fzeros is not the only VIS dependency! Another dependency # is implicit and is just _a_ numerical value loaded to %asi register, # which assembler can't recognize as VIS specific... $code =~ s/fzeros\s+%f([0-9]+)/ sprintf(".word\t0x%x\t! fzeros %%f%d",0x81b00c20|($1<<25),$1) /gem; print $code; # flush close STDOUT;