VirtualBox

source: vbox/trunk/src/libs/openssl-1.1.1f/crypto/modes/asm/aesni-gcm-x86_64.pl@ 83531

Last change on this file since 83531 was 83531, checked in by vboxsync, 5 years ago

setting svn:sync-process=export for openssl-1.1.1f, all files except tests

File size: 28.9 KB
Line 
1#! /usr/bin/env perl
2# Copyright 2013-2020 The OpenSSL Project Authors. All Rights Reserved.
3#
4# Licensed under the OpenSSL license (the "License"). You may not use
5# this file except in compliance with the License. You can obtain a copy
6# in the file LICENSE in the source distribution or at
7# https://www.openssl.org/source/license.html
8
9#
10# ====================================================================
11# Written by Andy Polyakov <[email protected]> for the OpenSSL
12# project. The module is, however, dual licensed under OpenSSL and
13# CRYPTOGAMS licenses depending on where you obtain it. For further
14# details see http://www.openssl.org/~appro/cryptogams/.
15# ====================================================================
16#
17#
18# AES-NI-CTR+GHASH stitch.
19#
20# February 2013
21#
22# OpenSSL GCM implementation is organized in such way that its
23# performance is rather close to the sum of its streamed components,
24# in the context parallelized AES-NI CTR and modulo-scheduled
25# PCLMULQDQ-enabled GHASH. Unfortunately, as no stitch implementation
26# was observed to perform significantly better than the sum of the
27# components on contemporary CPUs, the effort was deemed impossible to
28# justify. This module is based on combination of Intel submissions,
29# [1] and [2], with MOVBE twist suggested by Ilya Albrekht and Max
30# Locktyukhin of Intel Corp. who verified that it reduces shuffles
31# pressure with notable relative improvement, achieving 1.0 cycle per
32# byte processed with 128-bit key on Haswell processor, 0.74 - on
33# Broadwell, 0.63 - on Skylake... [Mentioned results are raw profiled
34# measurements for favourable packet size, one divisible by 96.
35# Applications using the EVP interface will observe a few percent
36# worse performance.]
37#
38# Knights Landing processes 1 byte in 1.25 cycles (measured with EVP).
39#
40# [1] http://rt.openssl.org/Ticket/Display.html?id=2900&user=guest&pass=guest
41# [2] http://www.intel.com/content/dam/www/public/us/en/documents/software-support/enabling-high-performance-gcm.pdf
42
43$flavour = shift;
44$output = shift;
45if ($flavour =~ /\./) { $output = $flavour; undef $flavour; }
46
47$win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/);
48
49$0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
50( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or
51( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or
52die "can't locate x86_64-xlate.pl";
53
54if (`$ENV{CC} -Wa,-v -c -o /dev/null -x assembler /dev/null 2>&1`
55 =~ /GNU assembler version ([2-9]\.[0-9]+)/) {
56 $avx = ($1>=2.20) + ($1>=2.22);
57}
58
59if (!$avx && $win64 && ($flavour =~ /nasm/ || $ENV{ASM} =~ /nasm/) &&
60 `nasm -v 2>&1` =~ /NASM version ([2-9]\.[0-9]+)/) {
61 $avx = ($1>=2.09) + ($1>=2.10);
62}
63
64if (!$avx && $win64 && ($flavour =~ /masm/ || $ENV{ASM} =~ /ml64/) &&
65 `ml64 2>&1` =~ /Version ([0-9]+)\./) {
66 $avx = ($1>=10) + ($1>=11);
67}
68
69if (!$avx && `$ENV{CC} -v 2>&1` =~ /((?:^clang|LLVM) version|.*based on LLVM) ([0-9]+\.[0-9]+)/) {
70 $avx = ($2>=3.0) + ($2>3.0);
71}
72
73open OUT,"| \"$^X\" \"$xlate\" $flavour \"$output\"";
74*STDOUT=*OUT;
75
76if ($avx>1) {{{
77
78($inp,$out,$len,$key,$ivp,$Xip)=("%rdi","%rsi","%rdx","%rcx","%r8","%r9");
79
80($Ii,$T1,$T2,$Hkey,
81 $Z0,$Z1,$Z2,$Z3,$Xi) = map("%xmm$_",(0..8));
82
83($inout0,$inout1,$inout2,$inout3,$inout4,$inout5,$rndkey) = map("%xmm$_",(9..15));
84
85($counter,$rounds,$ret,$const,$in0,$end0)=("%ebx","%ebp","%r10","%r11","%r14","%r15");
86
87$code=<<___;
88.text
89
90.type _aesni_ctr32_ghash_6x,\@abi-omnipotent
91.align 32
92_aesni_ctr32_ghash_6x:
93.cfi_startproc
94 vmovdqu 0x20($const),$T2 # borrow $T2, .Lone_msb
95 sub \$6,$len
96 vpxor $Z0,$Z0,$Z0 # $Z0 = 0
97 vmovdqu 0x00-0x80($key),$rndkey
98 vpaddb $T2,$T1,$inout1
99 vpaddb $T2,$inout1,$inout2
100 vpaddb $T2,$inout2,$inout3
101 vpaddb $T2,$inout3,$inout4
102 vpaddb $T2,$inout4,$inout5
103 vpxor $rndkey,$T1,$inout0
104 vmovdqu $Z0,16+8(%rsp) # "$Z3" = 0
105 jmp .Loop6x
106
107.align 32
108.Loop6x:
109 add \$`6<<24`,$counter
110 jc .Lhandle_ctr32 # discard $inout[1-5]?
111 vmovdqu 0x00-0x20($Xip),$Hkey # $Hkey^1
112 vpaddb $T2,$inout5,$T1 # next counter value
113 vpxor $rndkey,$inout1,$inout1
114 vpxor $rndkey,$inout2,$inout2
115
116.Lresume_ctr32:
117 vmovdqu $T1,($ivp) # save next counter value
118 vpclmulqdq \$0x10,$Hkey,$Z3,$Z1
119 vpxor $rndkey,$inout3,$inout3
120 vmovups 0x10-0x80($key),$T2 # borrow $T2 for $rndkey
121 vpclmulqdq \$0x01,$Hkey,$Z3,$Z2
122 xor %r12,%r12
123 cmp $in0,$end0
124
125 vaesenc $T2,$inout0,$inout0
126 vmovdqu 0x30+8(%rsp),$Ii # I[4]
127 vpxor $rndkey,$inout4,$inout4
128 vpclmulqdq \$0x00,$Hkey,$Z3,$T1
129 vaesenc $T2,$inout1,$inout1
130 vpxor $rndkey,$inout5,$inout5
131 setnc %r12b
132 vpclmulqdq \$0x11,$Hkey,$Z3,$Z3
133 vaesenc $T2,$inout2,$inout2
134 vmovdqu 0x10-0x20($Xip),$Hkey # $Hkey^2
135 neg %r12
136 vaesenc $T2,$inout3,$inout3
137 vpxor $Z1,$Z2,$Z2
138 vpclmulqdq \$0x00,$Hkey,$Ii,$Z1
139 vpxor $Z0,$Xi,$Xi # modulo-scheduled
140 vaesenc $T2,$inout4,$inout4
141 vpxor $Z1,$T1,$Z0
142 and \$0x60,%r12
143 vmovups 0x20-0x80($key),$rndkey
144 vpclmulqdq \$0x10,$Hkey,$Ii,$T1
145 vaesenc $T2,$inout5,$inout5
146
147 vpclmulqdq \$0x01,$Hkey,$Ii,$T2
148 lea ($in0,%r12),$in0
149 vaesenc $rndkey,$inout0,$inout0
150 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled [vpxor $Z3,$Xi,$Xi]
151 vpclmulqdq \$0x11,$Hkey,$Ii,$Hkey
152 vmovdqu 0x40+8(%rsp),$Ii # I[3]
153 vaesenc $rndkey,$inout1,$inout1
154 movbe 0x58($in0),%r13
155 vaesenc $rndkey,$inout2,$inout2
156 movbe 0x50($in0),%r12
157 vaesenc $rndkey,$inout3,$inout3
158 mov %r13,0x20+8(%rsp)
159 vaesenc $rndkey,$inout4,$inout4
160 mov %r12,0x28+8(%rsp)
161 vmovdqu 0x30-0x20($Xip),$Z1 # borrow $Z1 for $Hkey^3
162 vaesenc $rndkey,$inout5,$inout5
163
164 vmovups 0x30-0x80($key),$rndkey
165 vpxor $T1,$Z2,$Z2
166 vpclmulqdq \$0x00,$Z1,$Ii,$T1
167 vaesenc $rndkey,$inout0,$inout0
168 vpxor $T2,$Z2,$Z2
169 vpclmulqdq \$0x10,$Z1,$Ii,$T2
170 vaesenc $rndkey,$inout1,$inout1
171 vpxor $Hkey,$Z3,$Z3
172 vpclmulqdq \$0x01,$Z1,$Ii,$Hkey
173 vaesenc $rndkey,$inout2,$inout2
174 vpclmulqdq \$0x11,$Z1,$Ii,$Z1
175 vmovdqu 0x50+8(%rsp),$Ii # I[2]
176 vaesenc $rndkey,$inout3,$inout3
177 vaesenc $rndkey,$inout4,$inout4
178 vpxor $T1,$Z0,$Z0
179 vmovdqu 0x40-0x20($Xip),$T1 # borrow $T1 for $Hkey^4
180 vaesenc $rndkey,$inout5,$inout5
181
182 vmovups 0x40-0x80($key),$rndkey
183 vpxor $T2,$Z2,$Z2
184 vpclmulqdq \$0x00,$T1,$Ii,$T2
185 vaesenc $rndkey,$inout0,$inout0
186 vpxor $Hkey,$Z2,$Z2
187 vpclmulqdq \$0x10,$T1,$Ii,$Hkey
188 vaesenc $rndkey,$inout1,$inout1
189 movbe 0x48($in0),%r13
190 vpxor $Z1,$Z3,$Z3
191 vpclmulqdq \$0x01,$T1,$Ii,$Z1
192 vaesenc $rndkey,$inout2,$inout2
193 movbe 0x40($in0),%r12
194 vpclmulqdq \$0x11,$T1,$Ii,$T1
195 vmovdqu 0x60+8(%rsp),$Ii # I[1]
196 vaesenc $rndkey,$inout3,$inout3
197 mov %r13,0x30+8(%rsp)
198 vaesenc $rndkey,$inout4,$inout4
199 mov %r12,0x38+8(%rsp)
200 vpxor $T2,$Z0,$Z0
201 vmovdqu 0x60-0x20($Xip),$T2 # borrow $T2 for $Hkey^5
202 vaesenc $rndkey,$inout5,$inout5
203
204 vmovups 0x50-0x80($key),$rndkey
205 vpxor $Hkey,$Z2,$Z2
206 vpclmulqdq \$0x00,$T2,$Ii,$Hkey
207 vaesenc $rndkey,$inout0,$inout0
208 vpxor $Z1,$Z2,$Z2
209 vpclmulqdq \$0x10,$T2,$Ii,$Z1
210 vaesenc $rndkey,$inout1,$inout1
211 movbe 0x38($in0),%r13
212 vpxor $T1,$Z3,$Z3
213 vpclmulqdq \$0x01,$T2,$Ii,$T1
214 vpxor 0x70+8(%rsp),$Xi,$Xi # accumulate I[0]
215 vaesenc $rndkey,$inout2,$inout2
216 movbe 0x30($in0),%r12
217 vpclmulqdq \$0x11,$T2,$Ii,$T2
218 vaesenc $rndkey,$inout3,$inout3
219 mov %r13,0x40+8(%rsp)
220 vaesenc $rndkey,$inout4,$inout4
221 mov %r12,0x48+8(%rsp)
222 vpxor $Hkey,$Z0,$Z0
223 vmovdqu 0x70-0x20($Xip),$Hkey # $Hkey^6
224 vaesenc $rndkey,$inout5,$inout5
225
226 vmovups 0x60-0x80($key),$rndkey
227 vpxor $Z1,$Z2,$Z2
228 vpclmulqdq \$0x10,$Hkey,$Xi,$Z1
229 vaesenc $rndkey,$inout0,$inout0
230 vpxor $T1,$Z2,$Z2
231 vpclmulqdq \$0x01,$Hkey,$Xi,$T1
232 vaesenc $rndkey,$inout1,$inout1
233 movbe 0x28($in0),%r13
234 vpxor $T2,$Z3,$Z3
235 vpclmulqdq \$0x00,$Hkey,$Xi,$T2
236 vaesenc $rndkey,$inout2,$inout2
237 movbe 0x20($in0),%r12
238 vpclmulqdq \$0x11,$Hkey,$Xi,$Xi
239 vaesenc $rndkey,$inout3,$inout3
240 mov %r13,0x50+8(%rsp)
241 vaesenc $rndkey,$inout4,$inout4
242 mov %r12,0x58+8(%rsp)
243 vpxor $Z1,$Z2,$Z2
244 vaesenc $rndkey,$inout5,$inout5
245 vpxor $T1,$Z2,$Z2
246
247 vmovups 0x70-0x80($key),$rndkey
248 vpslldq \$8,$Z2,$Z1
249 vpxor $T2,$Z0,$Z0
250 vmovdqu 0x10($const),$Hkey # .Lpoly
251
252 vaesenc $rndkey,$inout0,$inout0
253 vpxor $Xi,$Z3,$Z3
254 vaesenc $rndkey,$inout1,$inout1
255 vpxor $Z1,$Z0,$Z0
256 movbe 0x18($in0),%r13
257 vaesenc $rndkey,$inout2,$inout2
258 movbe 0x10($in0),%r12
259 vpalignr \$8,$Z0,$Z0,$Ii # 1st phase
260 vpclmulqdq \$0x10,$Hkey,$Z0,$Z0
261 mov %r13,0x60+8(%rsp)
262 vaesenc $rndkey,$inout3,$inout3
263 mov %r12,0x68+8(%rsp)
264 vaesenc $rndkey,$inout4,$inout4
265 vmovups 0x80-0x80($key),$T1 # borrow $T1 for $rndkey
266 vaesenc $rndkey,$inout5,$inout5
267
268 vaesenc $T1,$inout0,$inout0
269 vmovups 0x90-0x80($key),$rndkey
270 vaesenc $T1,$inout1,$inout1
271 vpsrldq \$8,$Z2,$Z2
272 vaesenc $T1,$inout2,$inout2
273 vpxor $Z2,$Z3,$Z3
274 vaesenc $T1,$inout3,$inout3
275 vpxor $Ii,$Z0,$Z0
276 movbe 0x08($in0),%r13
277 vaesenc $T1,$inout4,$inout4
278 movbe 0x00($in0),%r12
279 vaesenc $T1,$inout5,$inout5
280 vmovups 0xa0-0x80($key),$T1
281 cmp \$11,$rounds
282 jb .Lenc_tail # 128-bit key
283
284 vaesenc $rndkey,$inout0,$inout0
285 vaesenc $rndkey,$inout1,$inout1
286 vaesenc $rndkey,$inout2,$inout2
287 vaesenc $rndkey,$inout3,$inout3
288 vaesenc $rndkey,$inout4,$inout4
289 vaesenc $rndkey,$inout5,$inout5
290
291 vaesenc $T1,$inout0,$inout0
292 vaesenc $T1,$inout1,$inout1
293 vaesenc $T1,$inout2,$inout2
294 vaesenc $T1,$inout3,$inout3
295 vaesenc $T1,$inout4,$inout4
296 vmovups 0xb0-0x80($key),$rndkey
297 vaesenc $T1,$inout5,$inout5
298 vmovups 0xc0-0x80($key),$T1
299 je .Lenc_tail # 192-bit key
300
301 vaesenc $rndkey,$inout0,$inout0
302 vaesenc $rndkey,$inout1,$inout1
303 vaesenc $rndkey,$inout2,$inout2
304 vaesenc $rndkey,$inout3,$inout3
305 vaesenc $rndkey,$inout4,$inout4
306 vaesenc $rndkey,$inout5,$inout5
307
308 vaesenc $T1,$inout0,$inout0
309 vaesenc $T1,$inout1,$inout1
310 vaesenc $T1,$inout2,$inout2
311 vaesenc $T1,$inout3,$inout3
312 vaesenc $T1,$inout4,$inout4
313 vmovups 0xd0-0x80($key),$rndkey
314 vaesenc $T1,$inout5,$inout5
315 vmovups 0xe0-0x80($key),$T1
316 jmp .Lenc_tail # 256-bit key
317
318.align 32
319.Lhandle_ctr32:
320 vmovdqu ($const),$Ii # borrow $Ii for .Lbswap_mask
321 vpshufb $Ii,$T1,$Z2 # byte-swap counter
322 vmovdqu 0x30($const),$Z1 # borrow $Z1, .Ltwo_lsb
323 vpaddd 0x40($const),$Z2,$inout1 # .Lone_lsb
324 vpaddd $Z1,$Z2,$inout2
325 vmovdqu 0x00-0x20($Xip),$Hkey # $Hkey^1
326 vpaddd $Z1,$inout1,$inout3
327 vpshufb $Ii,$inout1,$inout1
328 vpaddd $Z1,$inout2,$inout4
329 vpshufb $Ii,$inout2,$inout2
330 vpxor $rndkey,$inout1,$inout1
331 vpaddd $Z1,$inout3,$inout5
332 vpshufb $Ii,$inout3,$inout3
333 vpxor $rndkey,$inout2,$inout2
334 vpaddd $Z1,$inout4,$T1 # byte-swapped next counter value
335 vpshufb $Ii,$inout4,$inout4
336 vpshufb $Ii,$inout5,$inout5
337 vpshufb $Ii,$T1,$T1 # next counter value
338 jmp .Lresume_ctr32
339
340.align 32
341.Lenc_tail:
342 vaesenc $rndkey,$inout0,$inout0
343 vmovdqu $Z3,16+8(%rsp) # postpone vpxor $Z3,$Xi,$Xi
344 vpalignr \$8,$Z0,$Z0,$Xi # 2nd phase
345 vaesenc $rndkey,$inout1,$inout1
346 vpclmulqdq \$0x10,$Hkey,$Z0,$Z0
347 vpxor 0x00($inp),$T1,$T2
348 vaesenc $rndkey,$inout2,$inout2
349 vpxor 0x10($inp),$T1,$Ii
350 vaesenc $rndkey,$inout3,$inout3
351 vpxor 0x20($inp),$T1,$Z1
352 vaesenc $rndkey,$inout4,$inout4
353 vpxor 0x30($inp),$T1,$Z2
354 vaesenc $rndkey,$inout5,$inout5
355 vpxor 0x40($inp),$T1,$Z3
356 vpxor 0x50($inp),$T1,$Hkey
357 vmovdqu ($ivp),$T1 # load next counter value
358
359 vaesenclast $T2,$inout0,$inout0
360 vmovdqu 0x20($const),$T2 # borrow $T2, .Lone_msb
361 vaesenclast $Ii,$inout1,$inout1
362 vpaddb $T2,$T1,$Ii
363 mov %r13,0x70+8(%rsp)
364 lea 0x60($inp),$inp
365 vaesenclast $Z1,$inout2,$inout2
366 vpaddb $T2,$Ii,$Z1
367 mov %r12,0x78+8(%rsp)
368 lea 0x60($out),$out
369 vmovdqu 0x00-0x80($key),$rndkey
370 vaesenclast $Z2,$inout3,$inout3
371 vpaddb $T2,$Z1,$Z2
372 vaesenclast $Z3, $inout4,$inout4
373 vpaddb $T2,$Z2,$Z3
374 vaesenclast $Hkey,$inout5,$inout5
375 vpaddb $T2,$Z3,$Hkey
376
377 add \$0x60,$ret
378 sub \$0x6,$len
379 jc .L6x_done
380
381 vmovups $inout0,-0x60($out) # save output
382 vpxor $rndkey,$T1,$inout0
383 vmovups $inout1,-0x50($out)
384 vmovdqa $Ii,$inout1 # 0 latency
385 vmovups $inout2,-0x40($out)
386 vmovdqa $Z1,$inout2 # 0 latency
387 vmovups $inout3,-0x30($out)
388 vmovdqa $Z2,$inout3 # 0 latency
389 vmovups $inout4,-0x20($out)
390 vmovdqa $Z3,$inout4 # 0 latency
391 vmovups $inout5,-0x10($out)
392 vmovdqa $Hkey,$inout5 # 0 latency
393 vmovdqu 0x20+8(%rsp),$Z3 # I[5]
394 jmp .Loop6x
395
396.L6x_done:
397 vpxor 16+8(%rsp),$Xi,$Xi # modulo-scheduled
398 vpxor $Z0,$Xi,$Xi # modulo-scheduled
399
400 ret
401.cfi_endproc
402.size _aesni_ctr32_ghash_6x,.-_aesni_ctr32_ghash_6x
403___
404######################################################################
405#
406# size_t aesni_gcm_[en|de]crypt(const void *inp, void *out, size_t len,
407# const AES_KEY *key, unsigned char iv[16],
408# struct { u128 Xi,H,Htbl[9]; } *Xip);
409$code.=<<___;
410.globl aesni_gcm_decrypt
411.type aesni_gcm_decrypt,\@function,6
412.align 32
413aesni_gcm_decrypt:
414.cfi_startproc
415 xor $ret,$ret
416 cmp \$0x60,$len # minimal accepted length
417 jb .Lgcm_dec_abort
418
419 lea (%rsp),%rax # save stack pointer
420.cfi_def_cfa_register %rax
421 push %rbx
422.cfi_push %rbx
423 push %rbp
424.cfi_push %rbp
425 push %r12
426.cfi_push %r12
427 push %r13
428.cfi_push %r13
429 push %r14
430.cfi_push %r14
431 push %r15
432.cfi_push %r15
433___
434$code.=<<___ if ($win64);
435 lea -0xa8(%rsp),%rsp
436 movaps %xmm6,-0xd8(%rax)
437 movaps %xmm7,-0xc8(%rax)
438 movaps %xmm8,-0xb8(%rax)
439 movaps %xmm9,-0xa8(%rax)
440 movaps %xmm10,-0x98(%rax)
441 movaps %xmm11,-0x88(%rax)
442 movaps %xmm12,-0x78(%rax)
443 movaps %xmm13,-0x68(%rax)
444 movaps %xmm14,-0x58(%rax)
445 movaps %xmm15,-0x48(%rax)
446.Lgcm_dec_body:
447___
448$code.=<<___;
449 vzeroupper
450
451 vmovdqu ($ivp),$T1 # input counter value
452 add \$-128,%rsp
453 mov 12($ivp),$counter
454 lea .Lbswap_mask(%rip),$const
455 lea -0x80($key),$in0 # borrow $in0
456 mov \$0xf80,$end0 # borrow $end0
457 vmovdqu ($Xip),$Xi # load Xi
458 and \$-128,%rsp # ensure stack alignment
459 vmovdqu ($const),$Ii # borrow $Ii for .Lbswap_mask
460 lea 0x80($key),$key # size optimization
461 lea 0x20+0x20($Xip),$Xip # size optimization
462 mov 0xf0-0x80($key),$rounds
463 vpshufb $Ii,$Xi,$Xi
464
465 and $end0,$in0
466 and %rsp,$end0
467 sub $in0,$end0
468 jc .Ldec_no_key_aliasing
469 cmp \$768,$end0
470 jnc .Ldec_no_key_aliasing
471 sub $end0,%rsp # avoid aliasing with key
472.Ldec_no_key_aliasing:
473
474 vmovdqu 0x50($inp),$Z3 # I[5]
475 lea ($inp),$in0
476 vmovdqu 0x40($inp),$Z0
477 lea -0xc0($inp,$len),$end0
478 vmovdqu 0x30($inp),$Z1
479 shr \$4,$len
480 xor $ret,$ret
481 vmovdqu 0x20($inp),$Z2
482 vpshufb $Ii,$Z3,$Z3 # passed to _aesni_ctr32_ghash_6x
483 vmovdqu 0x10($inp),$T2
484 vpshufb $Ii,$Z0,$Z0
485 vmovdqu ($inp),$Hkey
486 vpshufb $Ii,$Z1,$Z1
487 vmovdqu $Z0,0x30(%rsp)
488 vpshufb $Ii,$Z2,$Z2
489 vmovdqu $Z1,0x40(%rsp)
490 vpshufb $Ii,$T2,$T2
491 vmovdqu $Z2,0x50(%rsp)
492 vpshufb $Ii,$Hkey,$Hkey
493 vmovdqu $T2,0x60(%rsp)
494 vmovdqu $Hkey,0x70(%rsp)
495
496 call _aesni_ctr32_ghash_6x
497
498 vmovups $inout0,-0x60($out) # save output
499 vmovups $inout1,-0x50($out)
500 vmovups $inout2,-0x40($out)
501 vmovups $inout3,-0x30($out)
502 vmovups $inout4,-0x20($out)
503 vmovups $inout5,-0x10($out)
504
505 vpshufb ($const),$Xi,$Xi # .Lbswap_mask
506 vmovdqu $Xi,-0x40($Xip) # output Xi
507
508 vzeroupper
509___
510$code.=<<___ if ($win64);
511 movaps -0xd8(%rax),%xmm6
512 movaps -0xc8(%rax),%xmm7
513 movaps -0xb8(%rax),%xmm8
514 movaps -0xa8(%rax),%xmm9
515 movaps -0x98(%rax),%xmm10
516 movaps -0x88(%rax),%xmm11
517 movaps -0x78(%rax),%xmm12
518 movaps -0x68(%rax),%xmm13
519 movaps -0x58(%rax),%xmm14
520 movaps -0x48(%rax),%xmm15
521___
522$code.=<<___;
523 mov -48(%rax),%r15
524.cfi_restore %r15
525 mov -40(%rax),%r14
526.cfi_restore %r14
527 mov -32(%rax),%r13
528.cfi_restore %r13
529 mov -24(%rax),%r12
530.cfi_restore %r12
531 mov -16(%rax),%rbp
532.cfi_restore %rbp
533 mov -8(%rax),%rbx
534.cfi_restore %rbx
535 lea (%rax),%rsp # restore %rsp
536.cfi_def_cfa_register %rsp
537.Lgcm_dec_abort:
538 mov $ret,%rax # return value
539 ret
540.cfi_endproc
541.size aesni_gcm_decrypt,.-aesni_gcm_decrypt
542___
543
544$code.=<<___;
545.type _aesni_ctr32_6x,\@abi-omnipotent
546.align 32
547_aesni_ctr32_6x:
548.cfi_startproc
549 vmovdqu 0x00-0x80($key),$Z0 # borrow $Z0 for $rndkey
550 vmovdqu 0x20($const),$T2 # borrow $T2, .Lone_msb
551 lea -1($rounds),%r13
552 vmovups 0x10-0x80($key),$rndkey
553 lea 0x20-0x80($key),%r12
554 vpxor $Z0,$T1,$inout0
555 add \$`6<<24`,$counter
556 jc .Lhandle_ctr32_2
557 vpaddb $T2,$T1,$inout1
558 vpaddb $T2,$inout1,$inout2
559 vpxor $Z0,$inout1,$inout1
560 vpaddb $T2,$inout2,$inout3
561 vpxor $Z0,$inout2,$inout2
562 vpaddb $T2,$inout3,$inout4
563 vpxor $Z0,$inout3,$inout3
564 vpaddb $T2,$inout4,$inout5
565 vpxor $Z0,$inout4,$inout4
566 vpaddb $T2,$inout5,$T1
567 vpxor $Z0,$inout5,$inout5
568 jmp .Loop_ctr32
569
570.align 16
571.Loop_ctr32:
572 vaesenc $rndkey,$inout0,$inout0
573 vaesenc $rndkey,$inout1,$inout1
574 vaesenc $rndkey,$inout2,$inout2
575 vaesenc $rndkey,$inout3,$inout3
576 vaesenc $rndkey,$inout4,$inout4
577 vaesenc $rndkey,$inout5,$inout5
578 vmovups (%r12),$rndkey
579 lea 0x10(%r12),%r12
580 dec %r13d
581 jnz .Loop_ctr32
582
583 vmovdqu (%r12),$Hkey # last round key
584 vaesenc $rndkey,$inout0,$inout0
585 vpxor 0x00($inp),$Hkey,$Z0
586 vaesenc $rndkey,$inout1,$inout1
587 vpxor 0x10($inp),$Hkey,$Z1
588 vaesenc $rndkey,$inout2,$inout2
589 vpxor 0x20($inp),$Hkey,$Z2
590 vaesenc $rndkey,$inout3,$inout3
591 vpxor 0x30($inp),$Hkey,$Xi
592 vaesenc $rndkey,$inout4,$inout4
593 vpxor 0x40($inp),$Hkey,$T2
594 vaesenc $rndkey,$inout5,$inout5
595 vpxor 0x50($inp),$Hkey,$Hkey
596 lea 0x60($inp),$inp
597
598 vaesenclast $Z0,$inout0,$inout0
599 vaesenclast $Z1,$inout1,$inout1
600 vaesenclast $Z2,$inout2,$inout2
601 vaesenclast $Xi,$inout3,$inout3
602 vaesenclast $T2,$inout4,$inout4
603 vaesenclast $Hkey,$inout5,$inout5
604 vmovups $inout0,0x00($out)
605 vmovups $inout1,0x10($out)
606 vmovups $inout2,0x20($out)
607 vmovups $inout3,0x30($out)
608 vmovups $inout4,0x40($out)
609 vmovups $inout5,0x50($out)
610 lea 0x60($out),$out
611
612 ret
613.align 32
614.Lhandle_ctr32_2:
615 vpshufb $Ii,$T1,$Z2 # byte-swap counter
616 vmovdqu 0x30($const),$Z1 # borrow $Z1, .Ltwo_lsb
617 vpaddd 0x40($const),$Z2,$inout1 # .Lone_lsb
618 vpaddd $Z1,$Z2,$inout2
619 vpaddd $Z1,$inout1,$inout3
620 vpshufb $Ii,$inout1,$inout1
621 vpaddd $Z1,$inout2,$inout4
622 vpshufb $Ii,$inout2,$inout2
623 vpxor $Z0,$inout1,$inout1
624 vpaddd $Z1,$inout3,$inout5
625 vpshufb $Ii,$inout3,$inout3
626 vpxor $Z0,$inout2,$inout2
627 vpaddd $Z1,$inout4,$T1 # byte-swapped next counter value
628 vpshufb $Ii,$inout4,$inout4
629 vpxor $Z0,$inout3,$inout3
630 vpshufb $Ii,$inout5,$inout5
631 vpxor $Z0,$inout4,$inout4
632 vpshufb $Ii,$T1,$T1 # next counter value
633 vpxor $Z0,$inout5,$inout5
634 jmp .Loop_ctr32
635.cfi_endproc
636.size _aesni_ctr32_6x,.-_aesni_ctr32_6x
637
638.globl aesni_gcm_encrypt
639.type aesni_gcm_encrypt,\@function,6
640.align 32
641aesni_gcm_encrypt:
642.cfi_startproc
643 xor $ret,$ret
644 cmp \$0x60*3,$len # minimal accepted length
645 jb .Lgcm_enc_abort
646
647 lea (%rsp),%rax # save stack pointer
648.cfi_def_cfa_register %rax
649 push %rbx
650.cfi_push %rbx
651 push %rbp
652.cfi_push %rbp
653 push %r12
654.cfi_push %r12
655 push %r13
656.cfi_push %r13
657 push %r14
658.cfi_push %r14
659 push %r15
660.cfi_push %r15
661___
662$code.=<<___ if ($win64);
663 lea -0xa8(%rsp),%rsp
664 movaps %xmm6,-0xd8(%rax)
665 movaps %xmm7,-0xc8(%rax)
666 movaps %xmm8,-0xb8(%rax)
667 movaps %xmm9,-0xa8(%rax)
668 movaps %xmm10,-0x98(%rax)
669 movaps %xmm11,-0x88(%rax)
670 movaps %xmm12,-0x78(%rax)
671 movaps %xmm13,-0x68(%rax)
672 movaps %xmm14,-0x58(%rax)
673 movaps %xmm15,-0x48(%rax)
674.Lgcm_enc_body:
675___
676$code.=<<___;
677 vzeroupper
678
679 vmovdqu ($ivp),$T1 # input counter value
680 add \$-128,%rsp
681 mov 12($ivp),$counter
682 lea .Lbswap_mask(%rip),$const
683 lea -0x80($key),$in0 # borrow $in0
684 mov \$0xf80,$end0 # borrow $end0
685 lea 0x80($key),$key # size optimization
686 vmovdqu ($const),$Ii # borrow $Ii for .Lbswap_mask
687 and \$-128,%rsp # ensure stack alignment
688 mov 0xf0-0x80($key),$rounds
689
690 and $end0,$in0
691 and %rsp,$end0
692 sub $in0,$end0
693 jc .Lenc_no_key_aliasing
694 cmp \$768,$end0
695 jnc .Lenc_no_key_aliasing
696 sub $end0,%rsp # avoid aliasing with key
697.Lenc_no_key_aliasing:
698
699 lea ($out),$in0
700 lea -0xc0($out,$len),$end0
701 shr \$4,$len
702
703 call _aesni_ctr32_6x
704 vpshufb $Ii,$inout0,$Xi # save bswapped output on stack
705 vpshufb $Ii,$inout1,$T2
706 vmovdqu $Xi,0x70(%rsp)
707 vpshufb $Ii,$inout2,$Z0
708 vmovdqu $T2,0x60(%rsp)
709 vpshufb $Ii,$inout3,$Z1
710 vmovdqu $Z0,0x50(%rsp)
711 vpshufb $Ii,$inout4,$Z2
712 vmovdqu $Z1,0x40(%rsp)
713 vpshufb $Ii,$inout5,$Z3 # passed to _aesni_ctr32_ghash_6x
714 vmovdqu $Z2,0x30(%rsp)
715
716 call _aesni_ctr32_6x
717
718 vmovdqu ($Xip),$Xi # load Xi
719 lea 0x20+0x20($Xip),$Xip # size optimization
720 sub \$12,$len
721 mov \$0x60*2,$ret
722 vpshufb $Ii,$Xi,$Xi
723
724 call _aesni_ctr32_ghash_6x
725 vmovdqu 0x20(%rsp),$Z3 # I[5]
726 vmovdqu ($const),$Ii # borrow $Ii for .Lbswap_mask
727 vmovdqu 0x00-0x20($Xip),$Hkey # $Hkey^1
728 vpunpckhqdq $Z3,$Z3,$T1
729 vmovdqu 0x20-0x20($Xip),$rndkey # borrow $rndkey for $HK
730 vmovups $inout0,-0x60($out) # save output
731 vpshufb $Ii,$inout0,$inout0 # but keep bswapped copy
732 vpxor $Z3,$T1,$T1
733 vmovups $inout1,-0x50($out)
734 vpshufb $Ii,$inout1,$inout1
735 vmovups $inout2,-0x40($out)
736 vpshufb $Ii,$inout2,$inout2
737 vmovups $inout3,-0x30($out)
738 vpshufb $Ii,$inout3,$inout3
739 vmovups $inout4,-0x20($out)
740 vpshufb $Ii,$inout4,$inout4
741 vmovups $inout5,-0x10($out)
742 vpshufb $Ii,$inout5,$inout5
743 vmovdqu $inout0,0x10(%rsp) # free $inout0
744___
745{ my ($HK,$T3)=($rndkey,$inout0);
746
747$code.=<<___;
748 vmovdqu 0x30(%rsp),$Z2 # I[4]
749 vmovdqu 0x10-0x20($Xip),$Ii # borrow $Ii for $Hkey^2
750 vpunpckhqdq $Z2,$Z2,$T2
751 vpclmulqdq \$0x00,$Hkey,$Z3,$Z1
752 vpxor $Z2,$T2,$T2
753 vpclmulqdq \$0x11,$Hkey,$Z3,$Z3
754 vpclmulqdq \$0x00,$HK,$T1,$T1
755
756 vmovdqu 0x40(%rsp),$T3 # I[3]
757 vpclmulqdq \$0x00,$Ii,$Z2,$Z0
758 vmovdqu 0x30-0x20($Xip),$Hkey # $Hkey^3
759 vpxor $Z1,$Z0,$Z0
760 vpunpckhqdq $T3,$T3,$Z1
761 vpclmulqdq \$0x11,$Ii,$Z2,$Z2
762 vpxor $T3,$Z1,$Z1
763 vpxor $Z3,$Z2,$Z2
764 vpclmulqdq \$0x10,$HK,$T2,$T2
765 vmovdqu 0x50-0x20($Xip),$HK
766 vpxor $T1,$T2,$T2
767
768 vmovdqu 0x50(%rsp),$T1 # I[2]
769 vpclmulqdq \$0x00,$Hkey,$T3,$Z3
770 vmovdqu 0x40-0x20($Xip),$Ii # borrow $Ii for $Hkey^4
771 vpxor $Z0,$Z3,$Z3
772 vpunpckhqdq $T1,$T1,$Z0
773 vpclmulqdq \$0x11,$Hkey,$T3,$T3
774 vpxor $T1,$Z0,$Z0
775 vpxor $Z2,$T3,$T3
776 vpclmulqdq \$0x00,$HK,$Z1,$Z1
777 vpxor $T2,$Z1,$Z1
778
779 vmovdqu 0x60(%rsp),$T2 # I[1]
780 vpclmulqdq \$0x00,$Ii,$T1,$Z2
781 vmovdqu 0x60-0x20($Xip),$Hkey # $Hkey^5
782 vpxor $Z3,$Z2,$Z2
783 vpunpckhqdq $T2,$T2,$Z3
784 vpclmulqdq \$0x11,$Ii,$T1,$T1
785 vpxor $T2,$Z3,$Z3
786 vpxor $T3,$T1,$T1
787 vpclmulqdq \$0x10,$HK,$Z0,$Z0
788 vmovdqu 0x80-0x20($Xip),$HK
789 vpxor $Z1,$Z0,$Z0
790
791 vpxor 0x70(%rsp),$Xi,$Xi # accumulate I[0]
792 vpclmulqdq \$0x00,$Hkey,$T2,$Z1
793 vmovdqu 0x70-0x20($Xip),$Ii # borrow $Ii for $Hkey^6
794 vpunpckhqdq $Xi,$Xi,$T3
795 vpxor $Z2,$Z1,$Z1
796 vpclmulqdq \$0x11,$Hkey,$T2,$T2
797 vpxor $Xi,$T3,$T3
798 vpxor $T1,$T2,$T2
799 vpclmulqdq \$0x00,$HK,$Z3,$Z3
800 vpxor $Z0,$Z3,$Z0
801
802 vpclmulqdq \$0x00,$Ii,$Xi,$Z2
803 vmovdqu 0x00-0x20($Xip),$Hkey # $Hkey^1
804 vpunpckhqdq $inout5,$inout5,$T1
805 vpclmulqdq \$0x11,$Ii,$Xi,$Xi
806 vpxor $inout5,$T1,$T1
807 vpxor $Z1,$Z2,$Z1
808 vpclmulqdq \$0x10,$HK,$T3,$T3
809 vmovdqu 0x20-0x20($Xip),$HK
810 vpxor $T2,$Xi,$Z3
811 vpxor $Z0,$T3,$Z2
812
813 vmovdqu 0x10-0x20($Xip),$Ii # borrow $Ii for $Hkey^2
814 vpxor $Z1,$Z3,$T3 # aggregated Karatsuba post-processing
815 vpclmulqdq \$0x00,$Hkey,$inout5,$Z0
816 vpxor $T3,$Z2,$Z2
817 vpunpckhqdq $inout4,$inout4,$T2
818 vpclmulqdq \$0x11,$Hkey,$inout5,$inout5
819 vpxor $inout4,$T2,$T2
820 vpslldq \$8,$Z2,$T3
821 vpclmulqdq \$0x00,$HK,$T1,$T1
822 vpxor $T3,$Z1,$Xi
823 vpsrldq \$8,$Z2,$Z2
824 vpxor $Z2,$Z3,$Z3
825
826 vpclmulqdq \$0x00,$Ii,$inout4,$Z1
827 vmovdqu 0x30-0x20($Xip),$Hkey # $Hkey^3
828 vpxor $Z0,$Z1,$Z1
829 vpunpckhqdq $inout3,$inout3,$T3
830 vpclmulqdq \$0x11,$Ii,$inout4,$inout4
831 vpxor $inout3,$T3,$T3
832 vpxor $inout5,$inout4,$inout4
833 vpalignr \$8,$Xi,$Xi,$inout5 # 1st phase
834 vpclmulqdq \$0x10,$HK,$T2,$T2
835 vmovdqu 0x50-0x20($Xip),$HK
836 vpxor $T1,$T2,$T2
837
838 vpclmulqdq \$0x00,$Hkey,$inout3,$Z0
839 vmovdqu 0x40-0x20($Xip),$Ii # borrow $Ii for $Hkey^4
840 vpxor $Z1,$Z0,$Z0
841 vpunpckhqdq $inout2,$inout2,$T1
842 vpclmulqdq \$0x11,$Hkey,$inout3,$inout3
843 vpxor $inout2,$T1,$T1
844 vpxor $inout4,$inout3,$inout3
845 vxorps 0x10(%rsp),$Z3,$Z3 # accumulate $inout0
846 vpclmulqdq \$0x00,$HK,$T3,$T3
847 vpxor $T2,$T3,$T3
848
849 vpclmulqdq \$0x10,0x10($const),$Xi,$Xi
850 vxorps $inout5,$Xi,$Xi
851
852 vpclmulqdq \$0x00,$Ii,$inout2,$Z1
853 vmovdqu 0x60-0x20($Xip),$Hkey # $Hkey^5
854 vpxor $Z0,$Z1,$Z1
855 vpunpckhqdq $inout1,$inout1,$T2
856 vpclmulqdq \$0x11,$Ii,$inout2,$inout2
857 vpxor $inout1,$T2,$T2
858 vpalignr \$8,$Xi,$Xi,$inout5 # 2nd phase
859 vpxor $inout3,$inout2,$inout2
860 vpclmulqdq \$0x10,$HK,$T1,$T1
861 vmovdqu 0x80-0x20($Xip),$HK
862 vpxor $T3,$T1,$T1
863
864 vxorps $Z3,$inout5,$inout5
865 vpclmulqdq \$0x10,0x10($const),$Xi,$Xi
866 vxorps $inout5,$Xi,$Xi
867
868 vpclmulqdq \$0x00,$Hkey,$inout1,$Z0
869 vmovdqu 0x70-0x20($Xip),$Ii # borrow $Ii for $Hkey^6
870 vpxor $Z1,$Z0,$Z0
871 vpunpckhqdq $Xi,$Xi,$T3
872 vpclmulqdq \$0x11,$Hkey,$inout1,$inout1
873 vpxor $Xi,$T3,$T3
874 vpxor $inout2,$inout1,$inout1
875 vpclmulqdq \$0x00,$HK,$T2,$T2
876 vpxor $T1,$T2,$T2
877
878 vpclmulqdq \$0x00,$Ii,$Xi,$Z1
879 vpclmulqdq \$0x11,$Ii,$Xi,$Z3
880 vpxor $Z0,$Z1,$Z1
881 vpclmulqdq \$0x10,$HK,$T3,$Z2
882 vpxor $inout1,$Z3,$Z3
883 vpxor $T2,$Z2,$Z2
884
885 vpxor $Z1,$Z3,$Z0 # aggregated Karatsuba post-processing
886 vpxor $Z0,$Z2,$Z2
887 vpslldq \$8,$Z2,$T1
888 vmovdqu 0x10($const),$Hkey # .Lpoly
889 vpsrldq \$8,$Z2,$Z2
890 vpxor $T1,$Z1,$Xi
891 vpxor $Z2,$Z3,$Z3
892
893 vpalignr \$8,$Xi,$Xi,$T2 # 1st phase
894 vpclmulqdq \$0x10,$Hkey,$Xi,$Xi
895 vpxor $T2,$Xi,$Xi
896
897 vpalignr \$8,$Xi,$Xi,$T2 # 2nd phase
898 vpclmulqdq \$0x10,$Hkey,$Xi,$Xi
899 vpxor $Z3,$T2,$T2
900 vpxor $T2,$Xi,$Xi
901___
902}
903$code.=<<___;
904 vpshufb ($const),$Xi,$Xi # .Lbswap_mask
905 vmovdqu $Xi,-0x40($Xip) # output Xi
906
907 vzeroupper
908___
909$code.=<<___ if ($win64);
910 movaps -0xd8(%rax),%xmm6
911 movaps -0xc8(%rax),%xmm7
912 movaps -0xb8(%rax),%xmm8
913 movaps -0xa8(%rax),%xmm9
914 movaps -0x98(%rax),%xmm10
915 movaps -0x88(%rax),%xmm11
916 movaps -0x78(%rax),%xmm12
917 movaps -0x68(%rax),%xmm13
918 movaps -0x58(%rax),%xmm14
919 movaps -0x48(%rax),%xmm15
920___
921$code.=<<___;
922 mov -48(%rax),%r15
923.cfi_restore %r15
924 mov -40(%rax),%r14
925.cfi_restore %r14
926 mov -32(%rax),%r13
927.cfi_restore %r13
928 mov -24(%rax),%r12
929.cfi_restore %r12
930 mov -16(%rax),%rbp
931.cfi_restore %rbp
932 mov -8(%rax),%rbx
933.cfi_restore %rbx
934 lea (%rax),%rsp # restore %rsp
935.cfi_def_cfa_register %rsp
936.Lgcm_enc_abort:
937 mov $ret,%rax # return value
938 ret
939.cfi_endproc
940.size aesni_gcm_encrypt,.-aesni_gcm_encrypt
941___
942
943$code.=<<___;
944.align 64
945.Lbswap_mask:
946 .byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
947.Lpoly:
948 .byte 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2
949.Lone_msb:
950 .byte 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1
951.Ltwo_lsb:
952 .byte 2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
953.Lone_lsb:
954 .byte 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
955.asciz "AES-NI GCM module for x86_64, CRYPTOGAMS by <appro\@openssl.org>"
956.align 64
957___
958if ($win64) {
959$rec="%rcx";
960$frame="%rdx";
961$context="%r8";
962$disp="%r9";
963
964$code.=<<___
965.extern __imp_RtlVirtualUnwind
966.type gcm_se_handler,\@abi-omnipotent
967.align 16
968gcm_se_handler:
969 push %rsi
970 push %rdi
971 push %rbx
972 push %rbp
973 push %r12
974 push %r13
975 push %r14
976 push %r15
977 pushfq
978 sub \$64,%rsp
979
980 mov 120($context),%rax # pull context->Rax
981 mov 248($context),%rbx # pull context->Rip
982
983 mov 8($disp),%rsi # disp->ImageBase
984 mov 56($disp),%r11 # disp->HandlerData
985
986 mov 0(%r11),%r10d # HandlerData[0]
987 lea (%rsi,%r10),%r10 # prologue label
988 cmp %r10,%rbx # context->Rip<prologue label
989 jb .Lcommon_seh_tail
990
991 mov 152($context),%rax # pull context->Rsp
992
993 mov 4(%r11),%r10d # HandlerData[1]
994 lea (%rsi,%r10),%r10 # epilogue label
995 cmp %r10,%rbx # context->Rip>=epilogue label
996 jae .Lcommon_seh_tail
997
998 mov 120($context),%rax # pull context->Rax
999
1000 mov -48(%rax),%r15
1001 mov -40(%rax),%r14
1002 mov -32(%rax),%r13
1003 mov -24(%rax),%r12
1004 mov -16(%rax),%rbp
1005 mov -8(%rax),%rbx
1006 mov %r15,240($context)
1007 mov %r14,232($context)
1008 mov %r13,224($context)
1009 mov %r12,216($context)
1010 mov %rbp,160($context)
1011 mov %rbx,144($context)
1012
1013 lea -0xd8(%rax),%rsi # %xmm save area
1014 lea 512($context),%rdi # & context.Xmm6
1015 mov \$20,%ecx # 10*sizeof(%xmm0)/sizeof(%rax)
1016 .long 0xa548f3fc # cld; rep movsq
1017
1018.Lcommon_seh_tail:
1019 mov 8(%rax),%rdi
1020 mov 16(%rax),%rsi
1021 mov %rax,152($context) # restore context->Rsp
1022 mov %rsi,168($context) # restore context->Rsi
1023 mov %rdi,176($context) # restore context->Rdi
1024
1025 mov 40($disp),%rdi # disp->ContextRecord
1026 mov $context,%rsi # context
1027 mov \$154,%ecx # sizeof(CONTEXT)
1028 .long 0xa548f3fc # cld; rep movsq
1029
1030 mov $disp,%rsi
1031 xor %rcx,%rcx # arg1, UNW_FLAG_NHANDLER
1032 mov 8(%rsi),%rdx # arg2, disp->ImageBase
1033 mov 0(%rsi),%r8 # arg3, disp->ControlPc
1034 mov 16(%rsi),%r9 # arg4, disp->FunctionEntry
1035 mov 40(%rsi),%r10 # disp->ContextRecord
1036 lea 56(%rsi),%r11 # &disp->HandlerData
1037 lea 24(%rsi),%r12 # &disp->EstablisherFrame
1038 mov %r10,32(%rsp) # arg5
1039 mov %r11,40(%rsp) # arg6
1040 mov %r12,48(%rsp) # arg7
1041 mov %rcx,56(%rsp) # arg8, (NULL)
1042 call *__imp_RtlVirtualUnwind(%rip)
1043
1044 mov \$1,%eax # ExceptionContinueSearch
1045 add \$64,%rsp
1046 popfq
1047 pop %r15
1048 pop %r14
1049 pop %r13
1050 pop %r12
1051 pop %rbp
1052 pop %rbx
1053 pop %rdi
1054 pop %rsi
1055 ret
1056.size gcm_se_handler,.-gcm_se_handler
1057
1058.section .pdata
1059.align 4
1060 .rva .LSEH_begin_aesni_gcm_decrypt
1061 .rva .LSEH_end_aesni_gcm_decrypt
1062 .rva .LSEH_gcm_dec_info
1063
1064 .rva .LSEH_begin_aesni_gcm_encrypt
1065 .rva .LSEH_end_aesni_gcm_encrypt
1066 .rva .LSEH_gcm_enc_info
1067.section .xdata
1068.align 8
1069.LSEH_gcm_dec_info:
1070 .byte 9,0,0,0
1071 .rva gcm_se_handler
1072 .rva .Lgcm_dec_body,.Lgcm_dec_abort
1073.LSEH_gcm_enc_info:
1074 .byte 9,0,0,0
1075 .rva gcm_se_handler
1076 .rva .Lgcm_enc_body,.Lgcm_enc_abort
1077___
1078}
1079}}} else {{{
1080$code=<<___; # assembler is too old
1081.text
1082
1083.globl aesni_gcm_encrypt
1084.type aesni_gcm_encrypt,\@abi-omnipotent
1085aesni_gcm_encrypt:
1086.cfi_startproc
1087 xor %eax,%eax
1088 ret
1089.cfi_endproc
1090.size aesni_gcm_encrypt,.-aesni_gcm_encrypt
1091
1092.globl aesni_gcm_decrypt
1093.type aesni_gcm_decrypt,\@abi-omnipotent
1094aesni_gcm_decrypt:
1095.cfi_startproc
1096 xor %eax,%eax
1097 ret
1098.cfi_endproc
1099.size aesni_gcm_decrypt,.-aesni_gcm_decrypt
1100___
1101}}}
1102
1103$code =~ s/\`([^\`]*)\`/eval($1)/gem;
1104
1105print $code;
1106
1107close STDOUT or die "error closing STDOUT: $!";
Note: See TracBrowser for help on using the repository browser.

© 2024 Oracle Support Privacy / Do Not Sell My Info Terms of Use Trademark Policy Automated Access Etiquette