Alexandre Lision | f26d3e5 | 2014-04-14 16:22:31 -0400 | [diff] [blame] | 1 | #!/usr/bin/env perl |
| 2 | |
| 3 | # ==================================================================== |
| 4 | # Written by Andy Polyakov <appro@openssl.org> for the OpenSSL |
| 5 | # project. The module is, however, dual licensed under OpenSSL and |
| 6 | # CRYPTOGAMS licenses depending on where you obtain it. For further |
| 7 | # details see http://www.openssl.org/~appro/cryptogams/. |
| 8 | # ==================================================================== |
| 9 | |
| 10 | # March 2010 |
| 11 | # |
| 12 | # The module implements "4-bit" GCM GHASH function and underlying |
| 13 | # single multiplication operation in GF(2^128). "4-bit" means that it |
| 14 | # uses 256 bytes per-key table [+128 bytes shared table]. Performance |
| 15 | # results are for streamed GHASH subroutine on UltraSPARC pre-Tx CPU |
| 16 | # and are expressed in cycles per processed byte, less is better: |
| 17 | # |
| 18 | # gcc 3.3.x cc 5.2 this assembler |
| 19 | # |
| 20 | # 32-bit build 81.4 43.3 12.6 (+546%/+244%) |
| 21 | # 64-bit build 20.2 21.2 12.6 (+60%/+68%) |
| 22 | # |
| 23 | # Here is data collected on UltraSPARC T1 system running Linux: |
| 24 | # |
| 25 | # gcc 4.4.1 this assembler |
| 26 | # |
| 27 | # 32-bit build 566 50 (+1000%) |
| 28 | # 64-bit build 56 50 (+12%) |
| 29 | # |
| 30 | # I don't quite understand why difference between 32-bit and 64-bit |
| 31 | # compiler-generated code is so big. Compilers *were* instructed to |
| 32 | # generate code for UltraSPARC and should have used 64-bit registers |
| 33 | # for Z vector (see C code) even in 32-bit build... Oh well, it only |
| 34 | # means more impressive improvement coefficients for this assembler |
| 35 | # module;-) Loops are aggressively modulo-scheduled in respect to |
| 36 | # references to input data and Z.hi updates to achieve 12 cycles |
| 37 | # timing. To anchor to something else, sha1-sparcv9.pl spends 11.6 |
| 38 | # cycles to process one byte on UltraSPARC pre-Tx CPU and ~24 on T1. |
| 39 | |
| 40 | $bits=32; |
| 41 | for (@ARGV) { $bits=64 if (/\-m64/ || /\-xarch\=v9/); } |
| 42 | if ($bits==64) { $bias=2047; $frame=192; } |
| 43 | else { $bias=0; $frame=112; } |
| 44 | |
| 45 | $output=shift; |
| 46 | open STDOUT,">$output"; |
| 47 | |
| 48 | $Zhi="%o0"; # 64-bit values |
| 49 | $Zlo="%o1"; |
| 50 | $Thi="%o2"; |
| 51 | $Tlo="%o3"; |
| 52 | $rem="%o4"; |
| 53 | $tmp="%o5"; |
| 54 | |
| 55 | $nhi="%l0"; # small values and pointers |
| 56 | $nlo="%l1"; |
| 57 | $xi0="%l2"; |
| 58 | $xi1="%l3"; |
| 59 | $rem_4bit="%l4"; |
| 60 | $remi="%l5"; |
| 61 | $Htblo="%l6"; |
| 62 | $cnt="%l7"; |
| 63 | |
| 64 | $Xi="%i0"; # input argument block |
| 65 | $Htbl="%i1"; |
| 66 | $inp="%i2"; |
| 67 | $len="%i3"; |
| 68 | |
| 69 | $code.=<<___; |
| 70 | .section ".text",#alloc,#execinstr |
| 71 | |
| 72 | .align 64 |
| 73 | rem_4bit: |
| 74 | .long `0x0000<<16`,0,`0x1C20<<16`,0,`0x3840<<16`,0,`0x2460<<16`,0 |
| 75 | .long `0x7080<<16`,0,`0x6CA0<<16`,0,`0x48C0<<16`,0,`0x54E0<<16`,0 |
| 76 | .long `0xE100<<16`,0,`0xFD20<<16`,0,`0xD940<<16`,0,`0xC560<<16`,0 |
| 77 | .long `0x9180<<16`,0,`0x8DA0<<16`,0,`0xA9C0<<16`,0,`0xB5E0<<16`,0 |
| 78 | .type rem_4bit,#object |
| 79 | .size rem_4bit,(.-rem_4bit) |
| 80 | |
| 81 | .globl gcm_ghash_4bit |
| 82 | .align 32 |
| 83 | gcm_ghash_4bit: |
| 84 | save %sp,-$frame,%sp |
| 85 | ldub [$inp+15],$nlo |
| 86 | ldub [$Xi+15],$xi0 |
| 87 | ldub [$Xi+14],$xi1 |
| 88 | add $len,$inp,$len |
| 89 | add $Htbl,8,$Htblo |
| 90 | |
| 91 | 1: call .+8 |
| 92 | add %o7,rem_4bit-1b,$rem_4bit |
| 93 | |
| 94 | .Louter: |
| 95 | xor $xi0,$nlo,$nlo |
| 96 | and $nlo,0xf0,$nhi |
| 97 | and $nlo,0x0f,$nlo |
| 98 | sll $nlo,4,$nlo |
| 99 | ldx [$Htblo+$nlo],$Zlo |
| 100 | ldx [$Htbl+$nlo],$Zhi |
| 101 | |
| 102 | ldub [$inp+14],$nlo |
| 103 | |
| 104 | ldx [$Htblo+$nhi],$Tlo |
| 105 | and $Zlo,0xf,$remi |
| 106 | ldx [$Htbl+$nhi],$Thi |
| 107 | sll $remi,3,$remi |
| 108 | ldx [$rem_4bit+$remi],$rem |
| 109 | srlx $Zlo,4,$Zlo |
| 110 | mov 13,$cnt |
| 111 | sllx $Zhi,60,$tmp |
| 112 | xor $Tlo,$Zlo,$Zlo |
| 113 | srlx $Zhi,4,$Zhi |
| 114 | xor $Zlo,$tmp,$Zlo |
| 115 | |
| 116 | xor $xi1,$nlo,$nlo |
| 117 | and $Zlo,0xf,$remi |
| 118 | and $nlo,0xf0,$nhi |
| 119 | and $nlo,0x0f,$nlo |
| 120 | ba .Lghash_inner |
| 121 | sll $nlo,4,$nlo |
| 122 | .align 32 |
| 123 | .Lghash_inner: |
| 124 | ldx [$Htblo+$nlo],$Tlo |
| 125 | sll $remi,3,$remi |
| 126 | xor $Thi,$Zhi,$Zhi |
| 127 | ldx [$Htbl+$nlo],$Thi |
| 128 | srlx $Zlo,4,$Zlo |
| 129 | xor $rem,$Zhi,$Zhi |
| 130 | ldx [$rem_4bit+$remi],$rem |
| 131 | sllx $Zhi,60,$tmp |
| 132 | xor $Tlo,$Zlo,$Zlo |
| 133 | ldub [$inp+$cnt],$nlo |
| 134 | srlx $Zhi,4,$Zhi |
| 135 | xor $Zlo,$tmp,$Zlo |
| 136 | ldub [$Xi+$cnt],$xi1 |
| 137 | xor $Thi,$Zhi,$Zhi |
| 138 | and $Zlo,0xf,$remi |
| 139 | |
| 140 | ldx [$Htblo+$nhi],$Tlo |
| 141 | sll $remi,3,$remi |
| 142 | xor $rem,$Zhi,$Zhi |
| 143 | ldx [$Htbl+$nhi],$Thi |
| 144 | srlx $Zlo,4,$Zlo |
| 145 | ldx [$rem_4bit+$remi],$rem |
| 146 | sllx $Zhi,60,$tmp |
| 147 | xor $xi1,$nlo,$nlo |
| 148 | srlx $Zhi,4,$Zhi |
| 149 | and $nlo,0xf0,$nhi |
| 150 | addcc $cnt,-1,$cnt |
| 151 | xor $Zlo,$tmp,$Zlo |
| 152 | and $nlo,0x0f,$nlo |
| 153 | xor $Tlo,$Zlo,$Zlo |
| 154 | sll $nlo,4,$nlo |
| 155 | blu .Lghash_inner |
| 156 | and $Zlo,0xf,$remi |
| 157 | |
| 158 | ldx [$Htblo+$nlo],$Tlo |
| 159 | sll $remi,3,$remi |
| 160 | xor $Thi,$Zhi,$Zhi |
| 161 | ldx [$Htbl+$nlo],$Thi |
| 162 | srlx $Zlo,4,$Zlo |
| 163 | xor $rem,$Zhi,$Zhi |
| 164 | ldx [$rem_4bit+$remi],$rem |
| 165 | sllx $Zhi,60,$tmp |
| 166 | xor $Tlo,$Zlo,$Zlo |
| 167 | srlx $Zhi,4,$Zhi |
| 168 | xor $Zlo,$tmp,$Zlo |
| 169 | xor $Thi,$Zhi,$Zhi |
| 170 | |
| 171 | add $inp,16,$inp |
| 172 | cmp $inp,$len |
| 173 | be,pn `$bits==64?"%xcc":"%icc"`,.Ldone |
| 174 | and $Zlo,0xf,$remi |
| 175 | |
| 176 | ldx [$Htblo+$nhi],$Tlo |
| 177 | sll $remi,3,$remi |
| 178 | xor $rem,$Zhi,$Zhi |
| 179 | ldx [$Htbl+$nhi],$Thi |
| 180 | srlx $Zlo,4,$Zlo |
| 181 | ldx [$rem_4bit+$remi],$rem |
| 182 | sllx $Zhi,60,$tmp |
| 183 | xor $Tlo,$Zlo,$Zlo |
| 184 | ldub [$inp+15],$nlo |
| 185 | srlx $Zhi,4,$Zhi |
| 186 | xor $Zlo,$tmp,$Zlo |
| 187 | xor $Thi,$Zhi,$Zhi |
| 188 | stx $Zlo,[$Xi+8] |
| 189 | xor $rem,$Zhi,$Zhi |
| 190 | stx $Zhi,[$Xi] |
| 191 | srl $Zlo,8,$xi1 |
| 192 | and $Zlo,0xff,$xi0 |
| 193 | ba .Louter |
| 194 | and $xi1,0xff,$xi1 |
| 195 | .align 32 |
| 196 | .Ldone: |
| 197 | ldx [$Htblo+$nhi],$Tlo |
| 198 | sll $remi,3,$remi |
| 199 | xor $rem,$Zhi,$Zhi |
| 200 | ldx [$Htbl+$nhi],$Thi |
| 201 | srlx $Zlo,4,$Zlo |
| 202 | ldx [$rem_4bit+$remi],$rem |
| 203 | sllx $Zhi,60,$tmp |
| 204 | xor $Tlo,$Zlo,$Zlo |
| 205 | srlx $Zhi,4,$Zhi |
| 206 | xor $Zlo,$tmp,$Zlo |
| 207 | xor $Thi,$Zhi,$Zhi |
| 208 | stx $Zlo,[$Xi+8] |
| 209 | xor $rem,$Zhi,$Zhi |
| 210 | stx $Zhi,[$Xi] |
| 211 | |
| 212 | ret |
| 213 | restore |
| 214 | .type gcm_ghash_4bit,#function |
| 215 | .size gcm_ghash_4bit,(.-gcm_ghash_4bit) |
| 216 | ___ |
| 217 | |
| 218 | undef $inp; |
| 219 | undef $len; |
| 220 | |
| 221 | $code.=<<___; |
| 222 | .globl gcm_gmult_4bit |
| 223 | .align 32 |
| 224 | gcm_gmult_4bit: |
| 225 | save %sp,-$frame,%sp |
| 226 | ldub [$Xi+15],$nlo |
| 227 | add $Htbl,8,$Htblo |
| 228 | |
| 229 | 1: call .+8 |
| 230 | add %o7,rem_4bit-1b,$rem_4bit |
| 231 | |
| 232 | and $nlo,0xf0,$nhi |
| 233 | and $nlo,0x0f,$nlo |
| 234 | sll $nlo,4,$nlo |
| 235 | ldx [$Htblo+$nlo],$Zlo |
| 236 | ldx [$Htbl+$nlo],$Zhi |
| 237 | |
| 238 | ldub [$Xi+14],$nlo |
| 239 | |
| 240 | ldx [$Htblo+$nhi],$Tlo |
| 241 | and $Zlo,0xf,$remi |
| 242 | ldx [$Htbl+$nhi],$Thi |
| 243 | sll $remi,3,$remi |
| 244 | ldx [$rem_4bit+$remi],$rem |
| 245 | srlx $Zlo,4,$Zlo |
| 246 | mov 13,$cnt |
| 247 | sllx $Zhi,60,$tmp |
| 248 | xor $Tlo,$Zlo,$Zlo |
| 249 | srlx $Zhi,4,$Zhi |
| 250 | xor $Zlo,$tmp,$Zlo |
| 251 | |
| 252 | and $Zlo,0xf,$remi |
| 253 | and $nlo,0xf0,$nhi |
| 254 | and $nlo,0x0f,$nlo |
| 255 | ba .Lgmult_inner |
| 256 | sll $nlo,4,$nlo |
| 257 | .align 32 |
| 258 | .Lgmult_inner: |
| 259 | ldx [$Htblo+$nlo],$Tlo |
| 260 | sll $remi,3,$remi |
| 261 | xor $Thi,$Zhi,$Zhi |
| 262 | ldx [$Htbl+$nlo],$Thi |
| 263 | srlx $Zlo,4,$Zlo |
| 264 | xor $rem,$Zhi,$Zhi |
| 265 | ldx [$rem_4bit+$remi],$rem |
| 266 | sllx $Zhi,60,$tmp |
| 267 | xor $Tlo,$Zlo,$Zlo |
| 268 | ldub [$Xi+$cnt],$nlo |
| 269 | srlx $Zhi,4,$Zhi |
| 270 | xor $Zlo,$tmp,$Zlo |
| 271 | xor $Thi,$Zhi,$Zhi |
| 272 | and $Zlo,0xf,$remi |
| 273 | |
| 274 | ldx [$Htblo+$nhi],$Tlo |
| 275 | sll $remi,3,$remi |
| 276 | xor $rem,$Zhi,$Zhi |
| 277 | ldx [$Htbl+$nhi],$Thi |
| 278 | srlx $Zlo,4,$Zlo |
| 279 | ldx [$rem_4bit+$remi],$rem |
| 280 | sllx $Zhi,60,$tmp |
| 281 | srlx $Zhi,4,$Zhi |
| 282 | and $nlo,0xf0,$nhi |
| 283 | addcc $cnt,-1,$cnt |
| 284 | xor $Zlo,$tmp,$Zlo |
| 285 | and $nlo,0x0f,$nlo |
| 286 | xor $Tlo,$Zlo,$Zlo |
| 287 | sll $nlo,4,$nlo |
| 288 | blu .Lgmult_inner |
| 289 | and $Zlo,0xf,$remi |
| 290 | |
| 291 | ldx [$Htblo+$nlo],$Tlo |
| 292 | sll $remi,3,$remi |
| 293 | xor $Thi,$Zhi,$Zhi |
| 294 | ldx [$Htbl+$nlo],$Thi |
| 295 | srlx $Zlo,4,$Zlo |
| 296 | xor $rem,$Zhi,$Zhi |
| 297 | ldx [$rem_4bit+$remi],$rem |
| 298 | sllx $Zhi,60,$tmp |
| 299 | xor $Tlo,$Zlo,$Zlo |
| 300 | srlx $Zhi,4,$Zhi |
| 301 | xor $Zlo,$tmp,$Zlo |
| 302 | xor $Thi,$Zhi,$Zhi |
| 303 | and $Zlo,0xf,$remi |
| 304 | |
| 305 | ldx [$Htblo+$nhi],$Tlo |
| 306 | sll $remi,3,$remi |
| 307 | xor $rem,$Zhi,$Zhi |
| 308 | ldx [$Htbl+$nhi],$Thi |
| 309 | srlx $Zlo,4,$Zlo |
| 310 | ldx [$rem_4bit+$remi],$rem |
| 311 | sllx $Zhi,60,$tmp |
| 312 | xor $Tlo,$Zlo,$Zlo |
| 313 | srlx $Zhi,4,$Zhi |
| 314 | xor $Zlo,$tmp,$Zlo |
| 315 | xor $Thi,$Zhi,$Zhi |
| 316 | stx $Zlo,[$Xi+8] |
| 317 | xor $rem,$Zhi,$Zhi |
| 318 | stx $Zhi,[$Xi] |
| 319 | |
| 320 | ret |
| 321 | restore |
| 322 | .type gcm_gmult_4bit,#function |
| 323 | .size gcm_gmult_4bit,(.-gcm_gmult_4bit) |
| 324 | .asciz "GHASH for SPARCv9, CRYPTOGAMS by <appro\@openssl.org>" |
| 325 | .align 4 |
| 326 | ___ |
| 327 | |
| 328 | $code =~ s/\`([^\`]*)\`/eval $1/gem; |
| 329 | print $code; |
| 330 | close STDOUT; |