summaryrefslogtreecommitdiffstats
path: root/win-x86/crypto/aes
diff options
context:
space:
mode:
authorAdam Langley <agl@google.com>2015-05-11 17:20:37 -0700
committerKenny Root <kroot@google.com>2015-05-12 23:06:14 +0000
commite9ada863a7b3e81f5d2b1e3bdd2305da902a87f5 (patch)
tree6e43e34595ecf887c26c32b86d8ab097fe8cac64 /win-x86/crypto/aes
parentb3106a0cc1493bbe0505c0ec0ce3da4ca90a29ae (diff)
downloadexternal_boringssl-e9ada863a7b3e81f5d2b1e3bdd2305da902a87f5.zip
external_boringssl-e9ada863a7b3e81f5d2b1e3bdd2305da902a87f5.tar.gz
external_boringssl-e9ada863a7b3e81f5d2b1e3bdd2305da902a87f5.tar.bz2
external/boringssl: bump revision.
This change bumps the BoringSSL revision to the current tip-of-tree. Change-Id: I91d5bf467e16e8d86cb19a4de873985f524e5faa
Diffstat (limited to 'win-x86/crypto/aes')
-rw-r--r--win-x86/crypto/aes/aes-586.asm3219
-rw-r--r--win-x86/crypto/aes/aesni-x86.asm2424
-rw-r--r--win-x86/crypto/aes/vpaes-x86.asm649
3 files changed, 6292 insertions, 0 deletions
diff --git a/win-x86/crypto/aes/aes-586.asm b/win-x86/crypto/aes/aes-586.asm
new file mode 100644
index 0000000..42ca026
--- /dev/null
+++ b/win-x86/crypto/aes/aes-586.asm
@@ -0,0 +1,3219 @@
+%ifidn __OUTPUT_FORMAT__,obj
+section code use32 class=code align=64
+%elifidn __OUTPUT_FORMAT__,win32
+%ifdef __YASM_VERSION_ID__
+%if __YASM_VERSION_ID__ < 01010000h
+%error yasm version 1.1.0 or later needed.
+%endif
+; Yasm automatically includes .00 and complains about redefining it.
+; https://www.tortall.net/projects/yasm/manual/html/objfmt-win32-safeseh.html
+%else
+$@feat.00 equ 1
+%endif
+section .text code align=64
+%else
+section .text code
+%endif
+align 16
+__x86_AES_encrypt_compact:
+ mov DWORD [20+esp],edi
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ mov esi,DWORD [240+edi]
+ lea esi,[esi*1+esi-2]
+ lea esi,[esi*8+edi]
+ mov DWORD [24+esp],esi
+ mov edi,DWORD [ebp-128]
+ mov esi,DWORD [ebp-96]
+ mov edi,DWORD [ebp-64]
+ mov esi,DWORD [ebp-32]
+ mov edi,DWORD [ebp]
+ mov esi,DWORD [32+ebp]
+ mov edi,DWORD [64+ebp]
+ mov esi,DWORD [96+ebp]
+align 16
+L$000loop:
+ mov esi,eax
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,bh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,edx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ shr ebx,16
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,ch
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,eax
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ shr ecx,24
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,dh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,eax
+ shr edi,16
+ and edx,255
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ movzx edi,bh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ and edx,255
+ movzx edx,BYTE [edx*1+ebp-128]
+ movzx eax,ah
+ movzx eax,BYTE [eax*1+ebp-128]
+ shl eax,8
+ xor edx,eax
+ mov eax,DWORD [4+esp]
+ and ebx,255
+ movzx ebx,BYTE [ebx*1+ebp-128]
+ shl ebx,16
+ xor edx,ebx
+ mov ebx,DWORD [8+esp]
+ movzx ecx,BYTE [ecx*1+ebp-128]
+ shl ecx,24
+ xor edx,ecx
+ mov ecx,esi
+ mov ebp,2155905152
+ and ebp,ecx
+ lea edi,[ecx*1+ecx]
+ mov esi,ebp
+ shr ebp,7
+ and edi,4278124286
+ sub esi,ebp
+ mov ebp,ecx
+ and esi,454761243
+ ror ebp,16
+ xor esi,edi
+ mov edi,ecx
+ xor ecx,esi
+ ror edi,24
+ xor esi,ebp
+ rol ecx,24
+ xor esi,edi
+ mov ebp,2155905152
+ xor ecx,esi
+ and ebp,edx
+ lea edi,[edx*1+edx]
+ mov esi,ebp
+ shr ebp,7
+ and edi,4278124286
+ sub esi,ebp
+ mov ebp,edx
+ and esi,454761243
+ ror ebp,16
+ xor esi,edi
+ mov edi,edx
+ xor edx,esi
+ ror edi,24
+ xor esi,ebp
+ rol edx,24
+ xor esi,edi
+ mov ebp,2155905152
+ xor edx,esi
+ and ebp,eax
+ lea edi,[eax*1+eax]
+ mov esi,ebp
+ shr ebp,7
+ and edi,4278124286
+ sub esi,ebp
+ mov ebp,eax
+ and esi,454761243
+ ror ebp,16
+ xor esi,edi
+ mov edi,eax
+ xor eax,esi
+ ror edi,24
+ xor esi,ebp
+ rol eax,24
+ xor esi,edi
+ mov ebp,2155905152
+ xor eax,esi
+ and ebp,ebx
+ lea edi,[ebx*1+ebx]
+ mov esi,ebp
+ shr ebp,7
+ and edi,4278124286
+ sub esi,ebp
+ mov ebp,ebx
+ and esi,454761243
+ ror ebp,16
+ xor esi,edi
+ mov edi,ebx
+ xor ebx,esi
+ ror edi,24
+ xor esi,ebp
+ rol ebx,24
+ xor esi,edi
+ xor ebx,esi
+ mov edi,DWORD [20+esp]
+ mov ebp,DWORD [28+esp]
+ add edi,16
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ cmp edi,DWORD [24+esp]
+ mov DWORD [20+esp],edi
+ jb NEAR L$000loop
+ mov esi,eax
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,bh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,edx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ shr ebx,16
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,ch
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,eax
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ shr ecx,24
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,dh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,eax
+ shr edi,16
+ and edx,255
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ movzx edi,bh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov edi,DWORD [20+esp]
+ and edx,255
+ movzx edx,BYTE [edx*1+ebp-128]
+ movzx eax,ah
+ movzx eax,BYTE [eax*1+ebp-128]
+ shl eax,8
+ xor edx,eax
+ mov eax,DWORD [4+esp]
+ and ebx,255
+ movzx ebx,BYTE [ebx*1+ebp-128]
+ shl ebx,16
+ xor edx,ebx
+ mov ebx,DWORD [8+esp]
+ movzx ecx,BYTE [ecx*1+ebp-128]
+ shl ecx,24
+ xor edx,ecx
+ mov ecx,esi
+ xor eax,DWORD [16+edi]
+ xor ebx,DWORD [20+edi]
+ xor ecx,DWORD [24+edi]
+ xor edx,DWORD [28+edi]
+ ret
+align 16
+__sse_AES_encrypt_compact:
+ pxor mm0,[edi]
+ pxor mm4,[8+edi]
+ mov esi,DWORD [240+edi]
+ lea esi,[esi*1+esi-2]
+ lea esi,[esi*8+edi]
+ mov DWORD [24+esp],esi
+ mov eax,454761243
+ mov DWORD [8+esp],eax
+ mov DWORD [12+esp],eax
+ mov eax,DWORD [ebp-128]
+ mov ebx,DWORD [ebp-96]
+ mov ecx,DWORD [ebp-64]
+ mov edx,DWORD [ebp-32]
+ mov eax,DWORD [ebp]
+ mov ebx,DWORD [32+ebp]
+ mov ecx,DWORD [64+ebp]
+ mov edx,DWORD [96+ebp]
+align 16
+L$001loop:
+ pshufw mm1,mm0,8
+ pshufw mm5,mm4,13
+ movd eax,mm1
+ movd ebx,mm5
+ mov DWORD [20+esp],edi
+ movzx esi,al
+ movzx edx,ah
+ pshufw mm2,mm0,13
+ movzx ecx,BYTE [esi*1+ebp-128]
+ movzx edi,bl
+ movzx edx,BYTE [edx*1+ebp-128]
+ shr eax,16
+ shl edx,8
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shl esi,16
+ pshufw mm6,mm4,8
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,ah
+ shl esi,24
+ shr ebx,16
+ or edx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shl esi,8
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,al
+ shl esi,24
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bl
+ movd eax,mm2
+ movd mm0,ecx
+ movzx ecx,BYTE [edi*1+ebp-128]
+ movzx edi,ah
+ shl ecx,16
+ movd ebx,mm6
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shl esi,24
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bl
+ shl esi,8
+ shr ebx,16
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,al
+ shr eax,16
+ movd mm1,ecx
+ movzx ecx,BYTE [edi*1+ebp-128]
+ movzx edi,ah
+ shl ecx,16
+ and eax,255
+ or ecx,esi
+ punpckldq mm0,mm1
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shl esi,24
+ and ebx,255
+ movzx eax,BYTE [eax*1+ebp-128]
+ or ecx,esi
+ shl eax,16
+ movzx esi,BYTE [edi*1+ebp-128]
+ or edx,eax
+ shl esi,8
+ movzx ebx,BYTE [ebx*1+ebp-128]
+ or ecx,esi
+ or edx,ebx
+ mov edi,DWORD [20+esp]
+ movd mm4,ecx
+ movd mm5,edx
+ punpckldq mm4,mm5
+ add edi,16
+ cmp edi,DWORD [24+esp]
+ ja NEAR L$002out
+ movq mm2,[8+esp]
+ pxor mm3,mm3
+ pxor mm7,mm7
+ movq mm1,mm0
+ movq mm5,mm4
+ pcmpgtb mm3,mm0
+ pcmpgtb mm7,mm4
+ pand mm3,mm2
+ pand mm7,mm2
+ pshufw mm2,mm0,177
+ pshufw mm6,mm4,177
+ paddb mm0,mm0
+ paddb mm4,mm4
+ pxor mm0,mm3
+ pxor mm4,mm7
+ pshufw mm3,mm2,177
+ pshufw mm7,mm6,177
+ pxor mm1,mm0
+ pxor mm5,mm4
+ pxor mm0,mm2
+ pxor mm4,mm6
+ movq mm2,mm3
+ movq mm6,mm7
+ pslld mm3,8
+ pslld mm7,8
+ psrld mm2,24
+ psrld mm6,24
+ pxor mm0,mm3
+ pxor mm4,mm7
+ pxor mm0,mm2
+ pxor mm4,mm6
+ movq mm3,mm1
+ movq mm7,mm5
+ movq mm2,[edi]
+ movq mm6,[8+edi]
+ psrld mm1,8
+ psrld mm5,8
+ mov eax,DWORD [ebp-128]
+ pslld mm3,24
+ pslld mm7,24
+ mov ebx,DWORD [ebp-64]
+ pxor mm0,mm1
+ pxor mm4,mm5
+ mov ecx,DWORD [ebp]
+ pxor mm0,mm3
+ pxor mm4,mm7
+ mov edx,DWORD [64+ebp]
+ pxor mm0,mm2
+ pxor mm4,mm6
+ jmp NEAR L$001loop
+align 16
+L$002out:
+ pxor mm0,[edi]
+ pxor mm4,[8+edi]
+ ret
+align 16
+__x86_AES_encrypt:
+ mov DWORD [20+esp],edi
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ mov esi,DWORD [240+edi]
+ lea esi,[esi*1+esi-2]
+ lea esi,[esi*8+edi]
+ mov DWORD [24+esp],esi
+align 16
+L$003loop:
+ mov esi,eax
+ and esi,255
+ mov esi,DWORD [esi*8+ebp]
+ movzx edi,bh
+ xor esi,DWORD [3+edi*8+ebp]
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ xor esi,DWORD [2+edi*8+ebp]
+ mov edi,edx
+ shr edi,24
+ xor esi,DWORD [1+edi*8+ebp]
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ shr ebx,16
+ mov esi,DWORD [esi*8+ebp]
+ movzx edi,ch
+ xor esi,DWORD [3+edi*8+ebp]
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ xor esi,DWORD [2+edi*8+ebp]
+ mov edi,eax
+ shr edi,24
+ xor esi,DWORD [1+edi*8+ebp]
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ shr ecx,24
+ mov esi,DWORD [esi*8+ebp]
+ movzx edi,dh
+ xor esi,DWORD [3+edi*8+ebp]
+ mov edi,eax
+ shr edi,16
+ and edx,255
+ and edi,255
+ xor esi,DWORD [2+edi*8+ebp]
+ movzx edi,bh
+ xor esi,DWORD [1+edi*8+ebp]
+ mov edi,DWORD [20+esp]
+ mov edx,DWORD [edx*8+ebp]
+ movzx eax,ah
+ xor edx,DWORD [3+eax*8+ebp]
+ mov eax,DWORD [4+esp]
+ and ebx,255
+ xor edx,DWORD [2+ebx*8+ebp]
+ mov ebx,DWORD [8+esp]
+ xor edx,DWORD [1+ecx*8+ebp]
+ mov ecx,esi
+ add edi,16
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ cmp edi,DWORD [24+esp]
+ mov DWORD [20+esp],edi
+ jb NEAR L$003loop
+ mov esi,eax
+ and esi,255
+ mov esi,DWORD [2+esi*8+ebp]
+ and esi,255
+ movzx edi,bh
+ mov edi,DWORD [edi*8+ebp]
+ and edi,65280
+ xor esi,edi
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ mov edi,DWORD [edi*8+ebp]
+ and edi,16711680
+ xor esi,edi
+ mov edi,edx
+ shr edi,24
+ mov edi,DWORD [2+edi*8+ebp]
+ and edi,4278190080
+ xor esi,edi
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ shr ebx,16
+ mov esi,DWORD [2+esi*8+ebp]
+ and esi,255
+ movzx edi,ch
+ mov edi,DWORD [edi*8+ebp]
+ and edi,65280
+ xor esi,edi
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ mov edi,DWORD [edi*8+ebp]
+ and edi,16711680
+ xor esi,edi
+ mov edi,eax
+ shr edi,24
+ mov edi,DWORD [2+edi*8+ebp]
+ and edi,4278190080
+ xor esi,edi
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ shr ecx,24
+ mov esi,DWORD [2+esi*8+ebp]
+ and esi,255
+ movzx edi,dh
+ mov edi,DWORD [edi*8+ebp]
+ and edi,65280
+ xor esi,edi
+ mov edi,eax
+ shr edi,16
+ and edx,255
+ and edi,255
+ mov edi,DWORD [edi*8+ebp]
+ and edi,16711680
+ xor esi,edi
+ movzx edi,bh
+ mov edi,DWORD [2+edi*8+ebp]
+ and edi,4278190080
+ xor esi,edi
+ mov edi,DWORD [20+esp]
+ and edx,255
+ mov edx,DWORD [2+edx*8+ebp]
+ and edx,255
+ movzx eax,ah
+ mov eax,DWORD [eax*8+ebp]
+ and eax,65280
+ xor edx,eax
+ mov eax,DWORD [4+esp]
+ and ebx,255
+ mov ebx,DWORD [ebx*8+ebp]
+ and ebx,16711680
+ xor edx,ebx
+ mov ebx,DWORD [8+esp]
+ mov ecx,DWORD [2+ecx*8+ebp]
+ and ecx,4278190080
+ xor edx,ecx
+ mov ecx,esi
+ add edi,16
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ ret
+align 64
+L$AES_Te:
+dd 2774754246,2774754246
+dd 2222750968,2222750968
+dd 2574743534,2574743534
+dd 2373680118,2373680118
+dd 234025727,234025727
+dd 3177933782,3177933782
+dd 2976870366,2976870366
+dd 1422247313,1422247313
+dd 1345335392,1345335392
+dd 50397442,50397442
+dd 2842126286,2842126286
+dd 2099981142,2099981142
+dd 436141799,436141799
+dd 1658312629,1658312629
+dd 3870010189,3870010189
+dd 2591454956,2591454956
+dd 1170918031,1170918031
+dd 2642575903,2642575903
+dd 1086966153,1086966153
+dd 2273148410,2273148410
+dd 368769775,368769775
+dd 3948501426,3948501426
+dd 3376891790,3376891790
+dd 200339707,200339707
+dd 3970805057,3970805057
+dd 1742001331,1742001331
+dd 4255294047,4255294047
+dd 3937382213,3937382213
+dd 3214711843,3214711843
+dd 4154762323,4154762323
+dd 2524082916,2524082916
+dd 1539358875,1539358875
+dd 3266819957,3266819957
+dd 486407649,486407649
+dd 2928907069,2928907069
+dd 1780885068,1780885068
+dd 1513502316,1513502316
+dd 1094664062,1094664062
+dd 49805301,49805301
+dd 1338821763,1338821763
+dd 1546925160,1546925160
+dd 4104496465,4104496465
+dd 887481809,887481809
+dd 150073849,150073849
+dd 2473685474,2473685474
+dd 1943591083,1943591083
+dd 1395732834,1395732834
+dd 1058346282,1058346282
+dd 201589768,201589768
+dd 1388824469,1388824469
+dd 1696801606,1696801606
+dd 1589887901,1589887901
+dd 672667696,672667696
+dd 2711000631,2711000631
+dd 251987210,251987210
+dd 3046808111,3046808111
+dd 151455502,151455502
+dd 907153956,907153956
+dd 2608889883,2608889883
+dd 1038279391,1038279391
+dd 652995533,652995533
+dd 1764173646,1764173646
+dd 3451040383,3451040383
+dd 2675275242,2675275242
+dd 453576978,453576978
+dd 2659418909,2659418909
+dd 1949051992,1949051992
+dd 773462580,773462580
+dd 756751158,756751158
+dd 2993581788,2993581788
+dd 3998898868,3998898868
+dd 4221608027,4221608027
+dd 4132590244,4132590244
+dd 1295727478,1295727478
+dd 1641469623,1641469623
+dd 3467883389,3467883389
+dd 2066295122,2066295122
+dd 1055122397,1055122397
+dd 1898917726,1898917726
+dd 2542044179,2542044179
+dd 4115878822,4115878822
+dd 1758581177,1758581177
+dd 0,0
+dd 753790401,753790401
+dd 1612718144,1612718144
+dd 536673507,536673507
+dd 3367088505,3367088505
+dd 3982187446,3982187446
+dd 3194645204,3194645204
+dd 1187761037,1187761037
+dd 3653156455,3653156455
+dd 1262041458,1262041458
+dd 3729410708,3729410708
+dd 3561770136,3561770136
+dd 3898103984,3898103984
+dd 1255133061,1255133061
+dd 1808847035,1808847035
+dd 720367557,720367557
+dd 3853167183,3853167183
+dd 385612781,385612781
+dd 3309519750,3309519750
+dd 3612167578,3612167578
+dd 1429418854,1429418854
+dd 2491778321,2491778321
+dd 3477423498,3477423498
+dd 284817897,284817897
+dd 100794884,100794884
+dd 2172616702,2172616702
+dd 4031795360,4031795360
+dd 1144798328,1144798328
+dd 3131023141,3131023141
+dd 3819481163,3819481163
+dd 4082192802,4082192802
+dd 4272137053,4272137053
+dd 3225436288,3225436288
+dd 2324664069,2324664069
+dd 2912064063,2912064063
+dd 3164445985,3164445985
+dd 1211644016,1211644016
+dd 83228145,83228145
+dd 3753688163,3753688163
+dd 3249976951,3249976951
+dd 1977277103,1977277103
+dd 1663115586,1663115586
+dd 806359072,806359072
+dd 452984805,452984805
+dd 250868733,250868733
+dd 1842533055,1842533055
+dd 1288555905,1288555905
+dd 336333848,336333848
+dd 890442534,890442534
+dd 804056259,804056259
+dd 3781124030,3781124030
+dd 2727843637,2727843637
+dd 3427026056,3427026056
+dd 957814574,957814574
+dd 1472513171,1472513171
+dd 4071073621,4071073621
+dd 2189328124,2189328124
+dd 1195195770,1195195770
+dd 2892260552,2892260552
+dd 3881655738,3881655738
+dd 723065138,723065138
+dd 2507371494,2507371494
+dd 2690670784,2690670784
+dd 2558624025,2558624025
+dd 3511635870,3511635870
+dd 2145180835,2145180835
+dd 1713513028,1713513028
+dd 2116692564,2116692564
+dd 2878378043,2878378043
+dd 2206763019,2206763019
+dd 3393603212,3393603212
+dd 703524551,703524551
+dd 3552098411,3552098411
+dd 1007948840,1007948840
+dd 2044649127,2044649127
+dd 3797835452,3797835452
+dd 487262998,487262998
+dd 1994120109,1994120109
+dd 1004593371,1004593371
+dd 1446130276,1446130276
+dd 1312438900,1312438900
+dd 503974420,503974420
+dd 3679013266,3679013266
+dd 168166924,168166924
+dd 1814307912,1814307912
+dd 3831258296,3831258296
+dd 1573044895,1573044895
+dd 1859376061,1859376061
+dd 4021070915,4021070915
+dd 2791465668,2791465668
+dd 2828112185,2828112185
+dd 2761266481,2761266481
+dd 937747667,937747667
+dd 2339994098,2339994098
+dd 854058965,854058965
+dd 1137232011,1137232011
+dd 1496790894,1496790894
+dd 3077402074,3077402074
+dd 2358086913,2358086913
+dd 1691735473,1691735473
+dd 3528347292,3528347292
+dd 3769215305,3769215305
+dd 3027004632,3027004632
+dd 4199962284,4199962284
+dd 133494003,133494003
+dd 636152527,636152527
+dd 2942657994,2942657994
+dd 2390391540,2390391540
+dd 3920539207,3920539207
+dd 403179536,403179536
+dd 3585784431,3585784431
+dd 2289596656,2289596656
+dd 1864705354,1864705354
+dd 1915629148,1915629148
+dd 605822008,605822008
+dd 4054230615,4054230615
+dd 3350508659,3350508659
+dd 1371981463,1371981463
+dd 602466507,602466507
+dd 2094914977,2094914977
+dd 2624877800,2624877800
+dd 555687742,555687742
+dd 3712699286,3712699286
+dd 3703422305,3703422305
+dd 2257292045,2257292045
+dd 2240449039,2240449039
+dd 2423288032,2423288032
+dd 1111375484,1111375484
+dd 3300242801,3300242801
+dd 2858837708,2858837708
+dd 3628615824,3628615824
+dd 84083462,84083462
+dd 32962295,32962295
+dd 302911004,302911004
+dd 2741068226,2741068226
+dd 1597322602,1597322602
+dd 4183250862,4183250862
+dd 3501832553,3501832553
+dd 2441512471,2441512471
+dd 1489093017,1489093017
+dd 656219450,656219450
+dd 3114180135,3114180135
+dd 954327513,954327513
+dd 335083755,335083755
+dd 3013122091,3013122091
+dd 856756514,856756514
+dd 3144247762,3144247762
+dd 1893325225,1893325225
+dd 2307821063,2307821063
+dd 2811532339,2811532339
+dd 3063651117,3063651117
+dd 572399164,572399164
+dd 2458355477,2458355477
+dd 552200649,552200649
+dd 1238290055,1238290055
+dd 4283782570,4283782570
+dd 2015897680,2015897680
+dd 2061492133,2061492133
+dd 2408352771,2408352771
+dd 4171342169,4171342169
+dd 2156497161,2156497161
+dd 386731290,386731290
+dd 3669999461,3669999461
+dd 837215959,837215959
+dd 3326231172,3326231172
+dd 3093850320,3093850320
+dd 3275833730,3275833730
+dd 2962856233,2962856233
+dd 1999449434,1999449434
+dd 286199582,286199582
+dd 3417354363,3417354363
+dd 4233385128,4233385128
+dd 3602627437,3602627437
+dd 974525996,974525996
+db 99,124,119,123,242,107,111,197
+db 48,1,103,43,254,215,171,118
+db 202,130,201,125,250,89,71,240
+db 173,212,162,175,156,164,114,192
+db 183,253,147,38,54,63,247,204
+db 52,165,229,241,113,216,49,21
+db 4,199,35,195,24,150,5,154
+db 7,18,128,226,235,39,178,117
+db 9,131,44,26,27,110,90,160
+db 82,59,214,179,41,227,47,132
+db 83,209,0,237,32,252,177,91
+db 106,203,190,57,74,76,88,207
+db 208,239,170,251,67,77,51,133
+db 69,249,2,127,80,60,159,168
+db 81,163,64,143,146,157,56,245
+db 188,182,218,33,16,255,243,210
+db 205,12,19,236,95,151,68,23
+db 196,167,126,61,100,93,25,115
+db 96,129,79,220,34,42,144,136
+db 70,238,184,20,222,94,11,219
+db 224,50,58,10,73,6,36,92
+db 194,211,172,98,145,149,228,121
+db 231,200,55,109,141,213,78,169
+db 108,86,244,234,101,122,174,8
+db 186,120,37,46,28,166,180,198
+db 232,221,116,31,75,189,139,138
+db 112,62,181,102,72,3,246,14
+db 97,53,87,185,134,193,29,158
+db 225,248,152,17,105,217,142,148
+db 155,30,135,233,206,85,40,223
+db 140,161,137,13,191,230,66,104
+db 65,153,45,15,176,84,187,22
+db 99,124,119,123,242,107,111,197
+db 48,1,103,43,254,215,171,118
+db 202,130,201,125,250,89,71,240
+db 173,212,162,175,156,164,114,192
+db 183,253,147,38,54,63,247,204
+db 52,165,229,241,113,216,49,21
+db 4,199,35,195,24,150,5,154
+db 7,18,128,226,235,39,178,117
+db 9,131,44,26,27,110,90,160
+db 82,59,214,179,41,227,47,132
+db 83,209,0,237,32,252,177,91
+db 106,203,190,57,74,76,88,207
+db 208,239,170,251,67,77,51,133
+db 69,249,2,127,80,60,159,168
+db 81,163,64,143,146,157,56,245
+db 188,182,218,33,16,255,243,210
+db 205,12,19,236,95,151,68,23
+db 196,167,126,61,100,93,25,115
+db 96,129,79,220,34,42,144,136
+db 70,238,184,20,222,94,11,219
+db 224,50,58,10,73,6,36,92
+db 194,211,172,98,145,149,228,121
+db 231,200,55,109,141,213,78,169
+db 108,86,244,234,101,122,174,8
+db 186,120,37,46,28,166,180,198
+db 232,221,116,31,75,189,139,138
+db 112,62,181,102,72,3,246,14
+db 97,53,87,185,134,193,29,158
+db 225,248,152,17,105,217,142,148
+db 155,30,135,233,206,85,40,223
+db 140,161,137,13,191,230,66,104
+db 65,153,45,15,176,84,187,22
+db 99,124,119,123,242,107,111,197
+db 48,1,103,43,254,215,171,118
+db 202,130,201,125,250,89,71,240
+db 173,212,162,175,156,164,114,192
+db 183,253,147,38,54,63,247,204
+db 52,165,229,241,113,216,49,21
+db 4,199,35,195,24,150,5,154
+db 7,18,128,226,235,39,178,117
+db 9,131,44,26,27,110,90,160
+db 82,59,214,179,41,227,47,132
+db 83,209,0,237,32,252,177,91
+db 106,203,190,57,74,76,88,207
+db 208,239,170,251,67,77,51,133
+db 69,249,2,127,80,60,159,168
+db 81,163,64,143,146,157,56,245
+db 188,182,218,33,16,255,243,210
+db 205,12,19,236,95,151,68,23
+db 196,167,126,61,100,93,25,115
+db 96,129,79,220,34,42,144,136
+db 70,238,184,20,222,94,11,219
+db 224,50,58,10,73,6,36,92
+db 194,211,172,98,145,149,228,121
+db 231,200,55,109,141,213,78,169
+db 108,86,244,234,101,122,174,8
+db 186,120,37,46,28,166,180,198
+db 232,221,116,31,75,189,139,138
+db 112,62,181,102,72,3,246,14
+db 97,53,87,185,134,193,29,158
+db 225,248,152,17,105,217,142,148
+db 155,30,135,233,206,85,40,223
+db 140,161,137,13,191,230,66,104
+db 65,153,45,15,176,84,187,22
+db 99,124,119,123,242,107,111,197
+db 48,1,103,43,254,215,171,118
+db 202,130,201,125,250,89,71,240
+db 173,212,162,175,156,164,114,192
+db 183,253,147,38,54,63,247,204
+db 52,165,229,241,113,216,49,21
+db 4,199,35,195,24,150,5,154
+db 7,18,128,226,235,39,178,117
+db 9,131,44,26,27,110,90,160
+db 82,59,214,179,41,227,47,132
+db 83,209,0,237,32,252,177,91
+db 106,203,190,57,74,76,88,207
+db 208,239,170,251,67,77,51,133
+db 69,249,2,127,80,60,159,168
+db 81,163,64,143,146,157,56,245
+db 188,182,218,33,16,255,243,210
+db 205,12,19,236,95,151,68,23
+db 196,167,126,61,100,93,25,115
+db 96,129,79,220,34,42,144,136
+db 70,238,184,20,222,94,11,219
+db 224,50,58,10,73,6,36,92
+db 194,211,172,98,145,149,228,121
+db 231,200,55,109,141,213,78,169
+db 108,86,244,234,101,122,174,8
+db 186,120,37,46,28,166,180,198
+db 232,221,116,31,75,189,139,138
+db 112,62,181,102,72,3,246,14
+db 97,53,87,185,134,193,29,158
+db 225,248,152,17,105,217,142,148
+db 155,30,135,233,206,85,40,223
+db 140,161,137,13,191,230,66,104
+db 65,153,45,15,176,84,187,22
+dd 1,2,4,8
+dd 16,32,64,128
+dd 27,54,0,0
+dd 0,0,0,0
+global _asm_AES_encrypt
+align 16
+_asm_AES_encrypt:
+L$_asm_AES_encrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [28+esp]
+ mov eax,esp
+ sub esp,36
+ and esp,-64
+ lea ebx,[edi-127]
+ sub ebx,esp
+ neg ebx
+ and ebx,960
+ sub esp,ebx
+ add esp,4
+ mov DWORD [28+esp],eax
+ call L$004pic_point
+L$004pic_point:
+ pop ebp
+ lea eax,[_OPENSSL_ia32cap_P]
+ lea ebp,[(L$AES_Te-L$004pic_point)+ebp]
+ lea ebx,[764+esp]
+ sub ebx,ebp
+ and ebx,768
+ lea ebp,[2176+ebx*1+ebp]
+ bt DWORD [eax],25
+ jnc NEAR L$005x86
+ movq mm0,[esi]
+ movq mm4,[8+esi]
+ call __sse_AES_encrypt_compact
+ mov esp,DWORD [28+esp]
+ mov esi,DWORD [24+esp]
+ movq [esi],mm0
+ movq [8+esi],mm4
+ emms
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+align 16
+L$005x86:
+ mov DWORD [24+esp],ebp
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ call __x86_AES_encrypt_compact
+ mov esp,DWORD [28+esp]
+ mov esi,DWORD [24+esp]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+align 16
+__x86_AES_decrypt_compact:
+ mov DWORD [20+esp],edi
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ mov esi,DWORD [240+edi]
+ lea esi,[esi*1+esi-2]
+ lea esi,[esi*8+edi]
+ mov DWORD [24+esp],esi
+ mov edi,DWORD [ebp-128]
+ mov esi,DWORD [ebp-96]
+ mov edi,DWORD [ebp-64]
+ mov esi,DWORD [ebp-32]
+ mov edi,DWORD [ebp]
+ mov esi,DWORD [32+ebp]
+ mov edi,DWORD [64+ebp]
+ mov esi,DWORD [96+ebp]
+align 16
+L$006loop:
+ mov esi,eax
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,dh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,ebx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,ah
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,ecx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,bh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,eax
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,edx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ and edx,255
+ movzx edx,BYTE [edx*1+ebp-128]
+ movzx ecx,ch
+ movzx ecx,BYTE [ecx*1+ebp-128]
+ shl ecx,8
+ xor edx,ecx
+ mov ecx,esi
+ shr ebx,16
+ and ebx,255
+ movzx ebx,BYTE [ebx*1+ebp-128]
+ shl ebx,16
+ xor edx,ebx
+ shr eax,24
+ movzx eax,BYTE [eax*1+ebp-128]
+ shl eax,24
+ xor edx,eax
+ mov edi,2155905152
+ and edi,ecx
+ mov esi,edi
+ shr edi,7
+ lea eax,[ecx*1+ecx]
+ sub esi,edi
+ and eax,4278124286
+ and esi,454761243
+ xor eax,esi
+ mov edi,2155905152
+ and edi,eax
+ mov esi,edi
+ shr edi,7
+ lea ebx,[eax*1+eax]
+ sub esi,edi
+ and ebx,4278124286
+ and esi,454761243
+ xor eax,ecx
+ xor ebx,esi
+ mov edi,2155905152
+ and edi,ebx
+ mov esi,edi
+ shr edi,7
+ lea ebp,[ebx*1+ebx]
+ sub esi,edi
+ and ebp,4278124286
+ and esi,454761243
+ xor ebx,ecx
+ rol ecx,8
+ xor ebp,esi
+ xor ecx,eax
+ xor eax,ebp
+ xor ecx,ebx
+ xor ebx,ebp
+ rol eax,24
+ xor ecx,ebp
+ rol ebx,16
+ xor ecx,eax
+ rol ebp,8
+ xor ecx,ebx
+ mov eax,DWORD [4+esp]
+ xor ecx,ebp
+ mov DWORD [12+esp],ecx
+ mov edi,2155905152
+ and edi,edx
+ mov esi,edi
+ shr edi,7
+ lea ebx,[edx*1+edx]
+ sub esi,edi
+ and ebx,4278124286
+ and esi,454761243
+ xor ebx,esi
+ mov edi,2155905152
+ and edi,ebx
+ mov esi,edi
+ shr edi,7
+ lea ecx,[ebx*1+ebx]
+ sub esi,edi
+ and ecx,4278124286
+ and esi,454761243
+ xor ebx,edx
+ xor ecx,esi
+ mov edi,2155905152
+ and edi,ecx
+ mov esi,edi
+ shr edi,7
+ lea ebp,[ecx*1+ecx]
+ sub esi,edi
+ and ebp,4278124286
+ and esi,454761243
+ xor ecx,edx
+ rol edx,8
+ xor ebp,esi
+ xor edx,ebx
+ xor ebx,ebp
+ xor edx,ecx
+ xor ecx,ebp
+ rol ebx,24
+ xor edx,ebp
+ rol ecx,16
+ xor edx,ebx
+ rol ebp,8
+ xor edx,ecx
+ mov ebx,DWORD [8+esp]
+ xor edx,ebp
+ mov DWORD [16+esp],edx
+ mov edi,2155905152
+ and edi,eax
+ mov esi,edi
+ shr edi,7
+ lea ecx,[eax*1+eax]
+ sub esi,edi
+ and ecx,4278124286
+ and esi,454761243
+ xor ecx,esi
+ mov edi,2155905152
+ and edi,ecx
+ mov esi,edi
+ shr edi,7
+ lea edx,[ecx*1+ecx]
+ sub esi,edi
+ and edx,4278124286
+ and esi,454761243
+ xor ecx,eax
+ xor edx,esi
+ mov edi,2155905152
+ and edi,edx
+ mov esi,edi
+ shr edi,7
+ lea ebp,[edx*1+edx]
+ sub esi,edi
+ and ebp,4278124286
+ and esi,454761243
+ xor edx,eax
+ rol eax,8
+ xor ebp,esi
+ xor eax,ecx
+ xor ecx,ebp
+ xor eax,edx
+ xor edx,ebp
+ rol ecx,24
+ xor eax,ebp
+ rol edx,16
+ xor eax,ecx
+ rol ebp,8
+ xor eax,edx
+ xor eax,ebp
+ mov edi,2155905152
+ and edi,ebx
+ mov esi,edi
+ shr edi,7
+ lea ecx,[ebx*1+ebx]
+ sub esi,edi
+ and ecx,4278124286
+ and esi,454761243
+ xor ecx,esi
+ mov edi,2155905152
+ and edi,ecx
+ mov esi,edi
+ shr edi,7
+ lea edx,[ecx*1+ecx]
+ sub esi,edi
+ and edx,4278124286
+ and esi,454761243
+ xor ecx,ebx
+ xor edx,esi
+ mov edi,2155905152
+ and edi,edx
+ mov esi,edi
+ shr edi,7
+ lea ebp,[edx*1+edx]
+ sub esi,edi
+ and ebp,4278124286
+ and esi,454761243
+ xor edx,ebx
+ rol ebx,8
+ xor ebp,esi
+ xor ebx,ecx
+ xor ecx,ebp
+ xor ebx,edx
+ xor edx,ebp
+ rol ecx,24
+ xor ebx,ebp
+ rol edx,16
+ xor ebx,ecx
+ rol ebp,8
+ xor ebx,edx
+ mov ecx,DWORD [12+esp]
+ xor ebx,ebp
+ mov edx,DWORD [16+esp]
+ mov edi,DWORD [20+esp]
+ mov ebp,DWORD [28+esp]
+ add edi,16
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ cmp edi,DWORD [24+esp]
+ mov DWORD [20+esp],edi
+ jb NEAR L$006loop
+ mov esi,eax
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,dh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,ebx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,ah
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,ecx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp-128]
+ movzx edi,bh
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,8
+ xor esi,edi
+ mov edi,eax
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,16
+ xor esi,edi
+ mov edi,edx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp-128]
+ shl edi,24
+ xor esi,edi
+ mov edi,DWORD [20+esp]
+ and edx,255
+ movzx edx,BYTE [edx*1+ebp-128]
+ movzx ecx,ch
+ movzx ecx,BYTE [ecx*1+ebp-128]
+ shl ecx,8
+ xor edx,ecx
+ mov ecx,esi
+ shr ebx,16
+ and ebx,255
+ movzx ebx,BYTE [ebx*1+ebp-128]
+ shl ebx,16
+ xor edx,ebx
+ mov ebx,DWORD [8+esp]
+ shr eax,24
+ movzx eax,BYTE [eax*1+ebp-128]
+ shl eax,24
+ xor edx,eax
+ mov eax,DWORD [4+esp]
+ xor eax,DWORD [16+edi]
+ xor ebx,DWORD [20+edi]
+ xor ecx,DWORD [24+edi]
+ xor edx,DWORD [28+edi]
+ ret
+align 16
+__sse_AES_decrypt_compact:
+ pxor mm0,[edi]
+ pxor mm4,[8+edi]
+ mov esi,DWORD [240+edi]
+ lea esi,[esi*1+esi-2]
+ lea esi,[esi*8+edi]
+ mov DWORD [24+esp],esi
+ mov eax,454761243
+ mov DWORD [8+esp],eax
+ mov DWORD [12+esp],eax
+ mov eax,DWORD [ebp-128]
+ mov ebx,DWORD [ebp-96]
+ mov ecx,DWORD [ebp-64]
+ mov edx,DWORD [ebp-32]
+ mov eax,DWORD [ebp]
+ mov ebx,DWORD [32+ebp]
+ mov ecx,DWORD [64+ebp]
+ mov edx,DWORD [96+ebp]
+align 16
+L$007loop:
+ pshufw mm1,mm0,12
+ pshufw mm5,mm4,9
+ movd eax,mm1
+ movd ebx,mm5
+ mov DWORD [20+esp],edi
+ movzx esi,al
+ movzx edx,ah
+ pshufw mm2,mm0,6
+ movzx ecx,BYTE [esi*1+ebp-128]
+ movzx edi,bl
+ movzx edx,BYTE [edx*1+ebp-128]
+ shr eax,16
+ shl edx,8
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shl esi,16
+ pshufw mm6,mm4,3
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,ah
+ shl esi,24
+ shr ebx,16
+ or edx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shl esi,24
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,al
+ shl esi,8
+ movd eax,mm2
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bl
+ shl esi,16
+ movd ebx,mm6
+ movd mm0,ecx
+ movzx ecx,BYTE [edi*1+ebp-128]
+ movzx edi,al
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bl
+ or edx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,ah
+ shl esi,16
+ shr eax,16
+ or edx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shr ebx,16
+ shl esi,8
+ movd mm1,edx
+ movzx edx,BYTE [edi*1+ebp-128]
+ movzx edi,bh
+ shl edx,24
+ and ebx,255
+ or edx,esi
+ punpckldq mm0,mm1
+ movzx esi,BYTE [edi*1+ebp-128]
+ movzx edi,al
+ shl esi,8
+ movzx eax,ah
+ movzx ebx,BYTE [ebx*1+ebp-128]
+ or ecx,esi
+ movzx esi,BYTE [edi*1+ebp-128]
+ or edx,ebx
+ shl esi,16
+ movzx eax,BYTE [eax*1+ebp-128]
+ or edx,esi
+ shl eax,24
+ or ecx,eax
+ mov edi,DWORD [20+esp]
+ movd mm4,edx
+ movd mm5,ecx
+ punpckldq mm4,mm5
+ add edi,16
+ cmp edi,DWORD [24+esp]
+ ja NEAR L$008out
+ movq mm3,mm0
+ movq mm7,mm4
+ pshufw mm2,mm0,228
+ pshufw mm6,mm4,228
+ movq mm1,mm0
+ movq mm5,mm4
+ pshufw mm0,mm0,177
+ pshufw mm4,mm4,177
+ pslld mm2,8
+ pslld mm6,8
+ psrld mm3,8
+ psrld mm7,8
+ pxor mm0,mm2
+ pxor mm4,mm6
+ pxor mm0,mm3
+ pxor mm4,mm7
+ pslld mm2,16
+ pslld mm6,16
+ psrld mm3,16
+ psrld mm7,16
+ pxor mm0,mm2
+ pxor mm4,mm6
+ pxor mm0,mm3
+ pxor mm4,mm7
+ movq mm3,[8+esp]
+ pxor mm2,mm2
+ pxor mm6,mm6
+ pcmpgtb mm2,mm1
+ pcmpgtb mm6,mm5
+ pand mm2,mm3
+ pand mm6,mm3
+ paddb mm1,mm1
+ paddb mm5,mm5
+ pxor mm1,mm2
+ pxor mm5,mm6
+ movq mm3,mm1
+ movq mm7,mm5
+ movq mm2,mm1
+ movq mm6,mm5
+ pxor mm0,mm1
+ pxor mm4,mm5
+ pslld mm3,24
+ pslld mm7,24
+ psrld mm2,8
+ psrld mm6,8
+ pxor mm0,mm3
+ pxor mm4,mm7
+ pxor mm0,mm2
+ pxor mm4,mm6
+ movq mm2,[8+esp]
+ pxor mm3,mm3
+ pxor mm7,mm7
+ pcmpgtb mm3,mm1
+ pcmpgtb mm7,mm5
+ pand mm3,mm2
+ pand mm7,mm2
+ paddb mm1,mm1
+ paddb mm5,mm5
+ pxor mm1,mm3
+ pxor mm5,mm7
+ pshufw mm3,mm1,177
+ pshufw mm7,mm5,177
+ pxor mm0,mm1
+ pxor mm4,mm5
+ pxor mm0,mm3
+ pxor mm4,mm7
+ pxor mm3,mm3
+ pxor mm7,mm7
+ pcmpgtb mm3,mm1
+ pcmpgtb mm7,mm5
+ pand mm3,mm2
+ pand mm7,mm2
+ paddb mm1,mm1
+ paddb mm5,mm5
+ pxor mm1,mm3
+ pxor mm5,mm7
+ pxor mm0,mm1
+ pxor mm4,mm5
+ movq mm3,mm1
+ movq mm7,mm5
+ pshufw mm2,mm1,177
+ pshufw mm6,mm5,177
+ pxor mm0,mm2
+ pxor mm4,mm6
+ pslld mm1,8
+ pslld mm5,8
+ psrld mm3,8
+ psrld mm7,8
+ movq mm2,[edi]
+ movq mm6,[8+edi]
+ pxor mm0,mm1
+ pxor mm4,mm5
+ pxor mm0,mm3
+ pxor mm4,mm7
+ mov eax,DWORD [ebp-128]
+ pslld mm1,16
+ pslld mm5,16
+ mov ebx,DWORD [ebp-64]
+ psrld mm3,16
+ psrld mm7,16
+ mov ecx,DWORD [ebp]
+ pxor mm0,mm1
+ pxor mm4,mm5
+ mov edx,DWORD [64+ebp]
+ pxor mm0,mm3
+ pxor mm4,mm7
+ pxor mm0,mm2
+ pxor mm4,mm6
+ jmp NEAR L$007loop
+align 16
+L$008out:
+ pxor mm0,[edi]
+ pxor mm4,[8+edi]
+ ret
+align 16
+__x86_AES_decrypt:
+ mov DWORD [20+esp],edi
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ mov esi,DWORD [240+edi]
+ lea esi,[esi*1+esi-2]
+ lea esi,[esi*8+edi]
+ mov DWORD [24+esp],esi
+align 16
+L$009loop:
+ mov esi,eax
+ and esi,255
+ mov esi,DWORD [esi*8+ebp]
+ movzx edi,dh
+ xor esi,DWORD [3+edi*8+ebp]
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ xor esi,DWORD [2+edi*8+ebp]
+ mov edi,ebx
+ shr edi,24
+ xor esi,DWORD [1+edi*8+ebp]
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ mov esi,DWORD [esi*8+ebp]
+ movzx edi,ah
+ xor esi,DWORD [3+edi*8+ebp]
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ xor esi,DWORD [2+edi*8+ebp]
+ mov edi,ecx
+ shr edi,24
+ xor esi,DWORD [1+edi*8+ebp]
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ mov esi,DWORD [esi*8+ebp]
+ movzx edi,bh
+ xor esi,DWORD [3+edi*8+ebp]
+ mov edi,eax
+ shr edi,16
+ and edi,255
+ xor esi,DWORD [2+edi*8+ebp]
+ mov edi,edx
+ shr edi,24
+ xor esi,DWORD [1+edi*8+ebp]
+ mov edi,DWORD [20+esp]
+ and edx,255
+ mov edx,DWORD [edx*8+ebp]
+ movzx ecx,ch
+ xor edx,DWORD [3+ecx*8+ebp]
+ mov ecx,esi
+ shr ebx,16
+ and ebx,255
+ xor edx,DWORD [2+ebx*8+ebp]
+ mov ebx,DWORD [8+esp]
+ shr eax,24
+ xor edx,DWORD [1+eax*8+ebp]
+ mov eax,DWORD [4+esp]
+ add edi,16
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ cmp edi,DWORD [24+esp]
+ mov DWORD [20+esp],edi
+ jb NEAR L$009loop
+ lea ebp,[2176+ebp]
+ mov edi,DWORD [ebp-128]
+ mov esi,DWORD [ebp-96]
+ mov edi,DWORD [ebp-64]
+ mov esi,DWORD [ebp-32]
+ mov edi,DWORD [ebp]
+ mov esi,DWORD [32+ebp]
+ mov edi,DWORD [64+ebp]
+ mov esi,DWORD [96+ebp]
+ lea ebp,[ebp-128]
+ mov esi,eax
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp]
+ movzx edi,dh
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,8
+ xor esi,edi
+ mov edi,ecx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,16
+ xor esi,edi
+ mov edi,ebx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [4+esp],esi
+ mov esi,ebx
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp]
+ movzx edi,ah
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,8
+ xor esi,edi
+ mov edi,edx
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,16
+ xor esi,edi
+ mov edi,ecx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,24
+ xor esi,edi
+ mov DWORD [8+esp],esi
+ mov esi,ecx
+ and esi,255
+ movzx esi,BYTE [esi*1+ebp]
+ movzx edi,bh
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,8
+ xor esi,edi
+ mov edi,eax
+ shr edi,16
+ and edi,255
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,16
+ xor esi,edi
+ mov edi,edx
+ shr edi,24
+ movzx edi,BYTE [edi*1+ebp]
+ shl edi,24
+ xor esi,edi
+ mov edi,DWORD [20+esp]
+ and edx,255
+ movzx edx,BYTE [edx*1+ebp]
+ movzx ecx,ch
+ movzx ecx,BYTE [ecx*1+ebp]
+ shl ecx,8
+ xor edx,ecx
+ mov ecx,esi
+ shr ebx,16
+ and ebx,255
+ movzx ebx,BYTE [ebx*1+ebp]
+ shl ebx,16
+ xor edx,ebx
+ mov ebx,DWORD [8+esp]
+ shr eax,24
+ movzx eax,BYTE [eax*1+ebp]
+ shl eax,24
+ xor edx,eax
+ mov eax,DWORD [4+esp]
+ lea ebp,[ebp-2048]
+ add edi,16
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ ret
+align 64
+L$AES_Td:
+dd 1353184337,1353184337
+dd 1399144830,1399144830
+dd 3282310938,3282310938
+dd 2522752826,2522752826
+dd 3412831035,3412831035
+dd 4047871263,4047871263
+dd 2874735276,2874735276
+dd 2466505547,2466505547
+dd 1442459680,1442459680
+dd 4134368941,4134368941
+dd 2440481928,2440481928
+dd 625738485,625738485
+dd 4242007375,4242007375
+dd 3620416197,3620416197
+dd 2151953702,2151953702
+dd 2409849525,2409849525
+dd 1230680542,1230680542
+dd 1729870373,1729870373
+dd 2551114309,2551114309
+dd 3787521629,3787521629
+dd 41234371,41234371
+dd 317738113,317738113
+dd 2744600205,2744600205
+dd 3338261355,3338261355
+dd 3881799427,3881799427
+dd 2510066197,2510066197
+dd 3950669247,3950669247
+dd 3663286933,3663286933
+dd 763608788,763608788
+dd 3542185048,3542185048
+dd 694804553,694804553
+dd 1154009486,1154009486
+dd 1787413109,1787413109
+dd 2021232372,2021232372
+dd 1799248025,1799248025
+dd 3715217703,3715217703
+dd 3058688446,3058688446
+dd 397248752,397248752
+dd 1722556617,1722556617
+dd 3023752829,3023752829
+dd 407560035,407560035
+dd 2184256229,2184256229
+dd 1613975959,1613975959
+dd 1165972322,1165972322
+dd 3765920945,3765920945
+dd 2226023355,2226023355
+dd 480281086,480281086
+dd 2485848313,2485848313
+dd 1483229296,1483229296
+dd 436028815,436028815
+dd 2272059028,2272059028
+dd 3086515026,3086515026
+dd 601060267,601060267
+dd 3791801202,3791801202
+dd 1468997603,1468997603
+dd 715871590,715871590
+dd 120122290,120122290
+dd 63092015,63092015
+dd 2591802758,2591802758
+dd 2768779219,2768779219
+dd 4068943920,4068943920
+dd 2997206819,2997206819
+dd 3127509762,3127509762
+dd 1552029421,1552029421
+dd 723308426,723308426
+dd 2461301159,2461301159
+dd 4042393587,4042393587
+dd 2715969870,2715969870
+dd 3455375973,3455375973
+dd 3586000134,3586000134
+dd 526529745,526529745
+dd 2331944644,2331944644
+dd 2639474228,2639474228
+dd 2689987490,2689987490
+dd 853641733,853641733
+dd 1978398372,1978398372
+dd 971801355,971801355
+dd 2867814464,2867814464
+dd 111112542,111112542
+dd 1360031421,1360031421
+dd 4186579262,4186579262
+dd 1023860118,1023860118
+dd 2919579357,2919579357
+dd 1186850381,1186850381
+dd 3045938321,3045938321
+dd 90031217,90031217
+dd 1876166148,1876166148
+dd 4279586912,4279586912
+dd 620468249,620468249
+dd 2548678102,2548678102
+dd 3426959497,3426959497
+dd 2006899047,2006899047
+dd 3175278768,3175278768
+dd 2290845959,2290845959
+dd 945494503,945494503
+dd 3689859193,3689859193
+dd 1191869601,1191869601
+dd 3910091388,3910091388
+dd 3374220536,3374220536
+dd 0,0
+dd 2206629897,2206629897
+dd 1223502642,1223502642
+dd 2893025566,2893025566
+dd 1316117100,1316117100
+dd 4227796733,4227796733
+dd 1446544655,1446544655
+dd 517320253,517320253
+dd 658058550,658058550
+dd 1691946762,1691946762
+dd 564550760,564550760
+dd 3511966619,3511966619
+dd 976107044,976107044
+dd 2976320012,2976320012
+dd 266819475,266819475
+dd 3533106868,3533106868
+dd 2660342555,2660342555
+dd 1338359936,1338359936
+dd 2720062561,2720062561
+dd 1766553434,1766553434
+dd 370807324,370807324
+dd 179999714,179999714
+dd 3844776128,3844776128
+dd 1138762300,1138762300
+dd 488053522,488053522
+dd 185403662,185403662
+dd 2915535858,2915535858
+dd 3114841645,3114841645
+dd 3366526484,3366526484
+dd 2233069911,2233069911
+dd 1275557295,1275557295
+dd 3151862254,3151862254
+dd 4250959779,4250959779
+dd 2670068215,2670068215
+dd 3170202204,3170202204
+dd 3309004356,3309004356
+dd 880737115,880737115
+dd 1982415755,1982415755
+dd 3703972811,3703972811
+dd 1761406390,1761406390
+dd 1676797112,1676797112
+dd 3403428311,3403428311
+dd 277177154,277177154
+dd 1076008723,1076008723
+dd 538035844,538035844
+dd 2099530373,2099530373
+dd 4164795346,4164795346
+dd 288553390,288553390
+dd 1839278535,1839278535
+dd 1261411869,1261411869
+dd 4080055004,4080055004
+dd 3964831245,3964831245
+dd 3504587127,3504587127
+dd 1813426987,1813426987
+dd 2579067049,2579067049
+dd 4199060497,4199060497
+dd 577038663,577038663
+dd 3297574056,3297574056
+dd 440397984,440397984
+dd 3626794326,3626794326
+dd 4019204898,4019204898
+dd 3343796615,3343796615
+dd 3251714265,3251714265
+dd 4272081548,4272081548
+dd 906744984,906744984
+dd 3481400742,3481400742
+dd 685669029,685669029
+dd 646887386,646887386
+dd 2764025151,2764025151
+dd 3835509292,3835509292
+dd 227702864,227702864
+dd 2613862250,2613862250
+dd 1648787028,1648787028
+dd 3256061430,3256061430
+dd 3904428176,3904428176
+dd 1593260334,1593260334
+dd 4121936770,4121936770
+dd 3196083615,3196083615
+dd 2090061929,2090061929
+dd 2838353263,2838353263
+dd 3004310991,3004310991
+dd 999926984,999926984
+dd 2809993232,2809993232
+dd 1852021992,1852021992
+dd 2075868123,2075868123
+dd 158869197,158869197
+dd 4095236462,4095236462
+dd 28809964,28809964
+dd 2828685187,2828685187
+dd 1701746150,1701746150
+dd 2129067946,2129067946
+dd 147831841,147831841
+dd 3873969647,3873969647
+dd 3650873274,3650873274
+dd 3459673930,3459673930
+dd 3557400554,3557400554
+dd 3598495785,3598495785
+dd 2947720241,2947720241
+dd 824393514,824393514
+dd 815048134,815048134
+dd 3227951669,3227951669
+dd 935087732,935087732
+dd 2798289660,2798289660
+dd 2966458592,2966458592
+dd 366520115,366520115
+dd 1251476721,1251476721
+dd 4158319681,4158319681
+dd 240176511,240176511
+dd 804688151,804688151
+dd 2379631990,2379631990
+dd 1303441219,1303441219
+dd 1414376140,1414376140
+dd 3741619940,3741619940
+dd 3820343710,3820343710
+dd 461924940,461924940
+dd 3089050817,3089050817
+dd 2136040774,2136040774
+dd 82468509,82468509
+dd 1563790337,1563790337
+dd 1937016826,1937016826
+dd 776014843,776014843
+dd 1511876531,1511876531
+dd 1389550482,1389550482
+dd 861278441,861278441
+dd 323475053,323475053
+dd 2355222426,2355222426
+dd 2047648055,2047648055
+dd 2383738969,2383738969
+dd 2302415851,2302415851
+dd 3995576782,3995576782
+dd 902390199,902390199
+dd 3991215329,3991215329
+dd 1018251130,1018251130
+dd 1507840668,1507840668
+dd 1064563285,1064563285
+dd 2043548696,2043548696
+dd 3208103795,3208103795
+dd 3939366739,3939366739
+dd 1537932639,1537932639
+dd 342834655,342834655
+dd 2262516856,2262516856
+dd 2180231114,2180231114
+dd 1053059257,1053059257
+dd 741614648,741614648
+dd 1598071746,1598071746
+dd 1925389590,1925389590
+dd 203809468,203809468
+dd 2336832552,2336832552
+dd 1100287487,1100287487
+dd 1895934009,1895934009
+dd 3736275976,3736275976
+dd 2632234200,2632234200
+dd 2428589668,2428589668
+dd 1636092795,1636092795
+dd 1890988757,1890988757
+dd 1952214088,1952214088
+dd 1113045200,1113045200
+db 82,9,106,213,48,54,165,56
+db 191,64,163,158,129,243,215,251
+db 124,227,57,130,155,47,255,135
+db 52,142,67,68,196,222,233,203
+db 84,123,148,50,166,194,35,61
+db 238,76,149,11,66,250,195,78
+db 8,46,161,102,40,217,36,178
+db 118,91,162,73,109,139,209,37
+db 114,248,246,100,134,104,152,22
+db 212,164,92,204,93,101,182,146
+db 108,112,72,80,253,237,185,218
+db 94,21,70,87,167,141,157,132
+db 144,216,171,0,140,188,211,10
+db 247,228,88,5,184,179,69,6
+db 208,44,30,143,202,63,15,2
+db 193,175,189,3,1,19,138,107
+db 58,145,17,65,79,103,220,234
+db 151,242,207,206,240,180,230,115
+db 150,172,116,34,231,173,53,133
+db 226,249,55,232,28,117,223,110
+db 71,241,26,113,29,41,197,137
+db 111,183,98,14,170,24,190,27
+db 252,86,62,75,198,210,121,32
+db 154,219,192,254,120,205,90,244
+db 31,221,168,51,136,7,199,49
+db 177,18,16,89,39,128,236,95
+db 96,81,127,169,25,181,74,13
+db 45,229,122,159,147,201,156,239
+db 160,224,59,77,174,42,245,176
+db 200,235,187,60,131,83,153,97
+db 23,43,4,126,186,119,214,38
+db 225,105,20,99,85,33,12,125
+db 82,9,106,213,48,54,165,56
+db 191,64,163,158,129,243,215,251
+db 124,227,57,130,155,47,255,135
+db 52,142,67,68,196,222,233,203
+db 84,123,148,50,166,194,35,61
+db 238,76,149,11,66,250,195,78
+db 8,46,161,102,40,217,36,178
+db 118,91,162,73,109,139,209,37
+db 114,248,246,100,134,104,152,22
+db 212,164,92,204,93,101,182,146
+db 108,112,72,80,253,237,185,218
+db 94,21,70,87,167,141,157,132
+db 144,216,171,0,140,188,211,10
+db 247,228,88,5,184,179,69,6
+db 208,44,30,143,202,63,15,2
+db 193,175,189,3,1,19,138,107
+db 58,145,17,65,79,103,220,234
+db 151,242,207,206,240,180,230,115
+db 150,172,116,34,231,173,53,133
+db 226,249,55,232,28,117,223,110
+db 71,241,26,113,29,41,197,137
+db 111,183,98,14,170,24,190,27
+db 252,86,62,75,198,210,121,32
+db 154,219,192,254,120,205,90,244
+db 31,221,168,51,136,7,199,49
+db 177,18,16,89,39,128,236,95
+db 96,81,127,169,25,181,74,13
+db 45,229,122,159,147,201,156,239
+db 160,224,59,77,174,42,245,176
+db 200,235,187,60,131,83,153,97
+db 23,43,4,126,186,119,214,38
+db 225,105,20,99,85,33,12,125
+db 82,9,106,213,48,54,165,56
+db 191,64,163,158,129,243,215,251
+db 124,227,57,130,155,47,255,135
+db 52,142,67,68,196,222,233,203
+db 84,123,148,50,166,194,35,61
+db 238,76,149,11,66,250,195,78
+db 8,46,161,102,40,217,36,178
+db 118,91,162,73,109,139,209,37
+db 114,248,246,100,134,104,152,22
+db 212,164,92,204,93,101,182,146
+db 108,112,72,80,253,237,185,218
+db 94,21,70,87,167,141,157,132
+db 144,216,171,0,140,188,211,10
+db 247,228,88,5,184,179,69,6
+db 208,44,30,143,202,63,15,2
+db 193,175,189,3,1,19,138,107
+db 58,145,17,65,79,103,220,234
+db 151,242,207,206,240,180,230,115
+db 150,172,116,34,231,173,53,133
+db 226,249,55,232,28,117,223,110
+db 71,241,26,113,29,41,197,137
+db 111,183,98,14,170,24,190,27
+db 252,86,62,75,198,210,121,32
+db 154,219,192,254,120,205,90,244
+db 31,221,168,51,136,7,199,49
+db 177,18,16,89,39,128,236,95
+db 96,81,127,169,25,181,74,13
+db 45,229,122,159,147,201,156,239
+db 160,224,59,77,174,42,245,176
+db 200,235,187,60,131,83,153,97
+db 23,43,4,126,186,119,214,38
+db 225,105,20,99,85,33,12,125
+db 82,9,106,213,48,54,165,56
+db 191,64,163,158,129,243,215,251
+db 124,227,57,130,155,47,255,135
+db 52,142,67,68,196,222,233,203
+db 84,123,148,50,166,194,35,61
+db 238,76,149,11,66,250,195,78
+db 8,46,161,102,40,217,36,178
+db 118,91,162,73,109,139,209,37
+db 114,248,246,100,134,104,152,22
+db 212,164,92,204,93,101,182,146
+db 108,112,72,80,253,237,185,218
+db 94,21,70,87,167,141,157,132
+db 144,216,171,0,140,188,211,10
+db 247,228,88,5,184,179,69,6
+db 208,44,30,143,202,63,15,2
+db 193,175,189,3,1,19,138,107
+db 58,145,17,65,79,103,220,234
+db 151,242,207,206,240,180,230,115
+db 150,172,116,34,231,173,53,133
+db 226,249,55,232,28,117,223,110
+db 71,241,26,113,29,41,197,137
+db 111,183,98,14,170,24,190,27
+db 252,86,62,75,198,210,121,32
+db 154,219,192,254,120,205,90,244
+db 31,221,168,51,136,7,199,49
+db 177,18,16,89,39,128,236,95
+db 96,81,127,169,25,181,74,13
+db 45,229,122,159,147,201,156,239
+db 160,224,59,77,174,42,245,176
+db 200,235,187,60,131,83,153,97
+db 23,43,4,126,186,119,214,38
+db 225,105,20,99,85,33,12,125
+global _asm_AES_decrypt
+align 16
+_asm_AES_decrypt:
+L$_asm_AES_decrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [28+esp]
+ mov eax,esp
+ sub esp,36
+ and esp,-64
+ lea ebx,[edi-127]
+ sub ebx,esp
+ neg ebx
+ and ebx,960
+ sub esp,ebx
+ add esp,4
+ mov DWORD [28+esp],eax
+ call L$010pic_point
+L$010pic_point:
+ pop ebp
+ lea eax,[_OPENSSL_ia32cap_P]
+ lea ebp,[(L$AES_Td-L$010pic_point)+ebp]
+ lea ebx,[764+esp]
+ sub ebx,ebp
+ and ebx,768
+ lea ebp,[2176+ebx*1+ebp]
+ bt DWORD [eax],25
+ jnc NEAR L$011x86
+ movq mm0,[esi]
+ movq mm4,[8+esi]
+ call __sse_AES_decrypt_compact
+ mov esp,DWORD [28+esp]
+ mov esi,DWORD [24+esp]
+ movq [esi],mm0
+ movq [8+esi],mm4
+ emms
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+align 16
+L$011x86:
+ mov DWORD [24+esp],ebp
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ call __x86_AES_decrypt_compact
+ mov esp,DWORD [28+esp]
+ mov esi,DWORD [24+esp]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _asm_AES_cbc_encrypt
+align 16
+_asm_AES_cbc_encrypt:
+L$_asm_AES_cbc_encrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov ecx,DWORD [28+esp]
+ cmp ecx,0
+ je NEAR L$012drop_out
+ call L$013pic_point
+L$013pic_point:
+ pop ebp
+ lea eax,[_OPENSSL_ia32cap_P]
+ cmp DWORD [40+esp],0
+ lea ebp,[(L$AES_Te-L$013pic_point)+ebp]
+ jne NEAR L$014picked_te
+ lea ebp,[(L$AES_Td-L$AES_Te)+ebp]
+L$014picked_te:
+ pushfd
+ cld
+ cmp ecx,512
+ jb NEAR L$015slow_way
+ test ecx,15
+ jnz NEAR L$015slow_way
+ bt DWORD [eax],28
+ jc NEAR L$015slow_way
+ lea esi,[esp-324]
+ and esi,-64
+ mov eax,ebp
+ lea ebx,[2304+ebp]
+ mov edx,esi
+ and eax,4095
+ and ebx,4095
+ and edx,4095
+ cmp edx,ebx
+ jb NEAR L$016tbl_break_out
+ sub edx,ebx
+ sub esi,edx
+ jmp NEAR L$017tbl_ok
+align 4
+L$016tbl_break_out:
+ sub edx,eax
+ and edx,4095
+ add edx,384
+ sub esi,edx
+align 4
+L$017tbl_ok:
+ lea edx,[24+esp]
+ xchg esp,esi
+ add esp,4
+ mov DWORD [24+esp],ebp
+ mov DWORD [28+esp],esi
+ mov eax,DWORD [edx]
+ mov ebx,DWORD [4+edx]
+ mov edi,DWORD [12+edx]
+ mov esi,DWORD [16+edx]
+ mov edx,DWORD [20+edx]
+ mov DWORD [32+esp],eax
+ mov DWORD [36+esp],ebx
+ mov DWORD [40+esp],ecx
+ mov DWORD [44+esp],edi
+ mov DWORD [48+esp],esi
+ mov DWORD [316+esp],0
+ mov ebx,edi
+ mov ecx,61
+ sub ebx,ebp
+ mov esi,edi
+ and ebx,4095
+ lea edi,[76+esp]
+ cmp ebx,2304
+ jb NEAR L$018do_copy
+ cmp ebx,3852
+ jb NEAR L$019skip_copy
+align 4
+L$018do_copy:
+ mov DWORD [44+esp],edi
+dd 2784229001
+L$019skip_copy:
+ mov edi,16
+align 4
+L$020prefetch_tbl:
+ mov eax,DWORD [ebp]
+ mov ebx,DWORD [32+ebp]
+ mov ecx,DWORD [64+ebp]
+ mov esi,DWORD [96+ebp]
+ lea ebp,[128+ebp]
+ sub edi,1
+ jnz NEAR L$020prefetch_tbl
+ sub ebp,2048
+ mov esi,DWORD [32+esp]
+ mov edi,DWORD [48+esp]
+ cmp edx,0
+ je NEAR L$021fast_decrypt
+ mov eax,DWORD [edi]
+ mov ebx,DWORD [4+edi]
+align 16
+L$022fast_enc_loop:
+ mov ecx,DWORD [8+edi]
+ mov edx,DWORD [12+edi]
+ xor eax,DWORD [esi]
+ xor ebx,DWORD [4+esi]
+ xor ecx,DWORD [8+esi]
+ xor edx,DWORD [12+esi]
+ mov edi,DWORD [44+esp]
+ call __x86_AES_encrypt
+ mov esi,DWORD [32+esp]
+ mov edi,DWORD [36+esp]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ lea esi,[16+esi]
+ mov ecx,DWORD [40+esp]
+ mov DWORD [32+esp],esi
+ lea edx,[16+edi]
+ mov DWORD [36+esp],edx
+ sub ecx,16
+ mov DWORD [40+esp],ecx
+ jnz NEAR L$022fast_enc_loop
+ mov esi,DWORD [48+esp]
+ mov ecx,DWORD [8+edi]
+ mov edx,DWORD [12+edi]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ cmp DWORD [316+esp],0
+ mov edi,DWORD [44+esp]
+ je NEAR L$023skip_ezero
+ mov ecx,60
+ xor eax,eax
+align 4
+dd 2884892297
+L$023skip_ezero:
+ mov esp,DWORD [28+esp]
+ popfd
+L$012drop_out:
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+ pushfd
+align 16
+L$021fast_decrypt:
+ cmp esi,DWORD [36+esp]
+ je NEAR L$024fast_dec_in_place
+ mov DWORD [52+esp],edi
+align 4
+align 16
+L$025fast_dec_loop:
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ mov edi,DWORD [44+esp]
+ call __x86_AES_decrypt
+ mov edi,DWORD [52+esp]
+ mov esi,DWORD [40+esp]
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ mov edi,DWORD [36+esp]
+ mov esi,DWORD [32+esp]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov ecx,DWORD [40+esp]
+ mov DWORD [52+esp],esi
+ lea esi,[16+esi]
+ mov DWORD [32+esp],esi
+ lea edi,[16+edi]
+ mov DWORD [36+esp],edi
+ sub ecx,16
+ mov DWORD [40+esp],ecx
+ jnz NEAR L$025fast_dec_loop
+ mov edi,DWORD [52+esp]
+ mov esi,DWORD [48+esp]
+ mov eax,DWORD [edi]
+ mov ebx,DWORD [4+edi]
+ mov ecx,DWORD [8+edi]
+ mov edx,DWORD [12+edi]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ jmp NEAR L$026fast_dec_out
+align 16
+L$024fast_dec_in_place:
+L$027fast_dec_in_place_loop:
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ lea edi,[60+esp]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov edi,DWORD [44+esp]
+ call __x86_AES_decrypt
+ mov edi,DWORD [48+esp]
+ mov esi,DWORD [36+esp]
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ lea esi,[16+esi]
+ mov DWORD [36+esp],esi
+ lea esi,[60+esp]
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov esi,DWORD [32+esp]
+ mov ecx,DWORD [40+esp]
+ lea esi,[16+esi]
+ mov DWORD [32+esp],esi
+ sub ecx,16
+ mov DWORD [40+esp],ecx
+ jnz NEAR L$027fast_dec_in_place_loop
+align 4
+L$026fast_dec_out:
+ cmp DWORD [316+esp],0
+ mov edi,DWORD [44+esp]
+ je NEAR L$028skip_dzero
+ mov ecx,60
+ xor eax,eax
+align 4
+dd 2884892297
+L$028skip_dzero:
+ mov esp,DWORD [28+esp]
+ popfd
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+ pushfd
+align 16
+L$015slow_way:
+ mov eax,DWORD [eax]
+ mov edi,DWORD [36+esp]
+ lea esi,[esp-80]
+ and esi,-64
+ lea ebx,[edi-143]
+ sub ebx,esi
+ neg ebx
+ and ebx,960
+ sub esi,ebx
+ lea ebx,[768+esi]
+ sub ebx,ebp
+ and ebx,768
+ lea ebp,[2176+ebx*1+ebp]
+ lea edx,[24+esp]
+ xchg esp,esi
+ add esp,4
+ mov DWORD [24+esp],ebp
+ mov DWORD [28+esp],esi
+ mov DWORD [52+esp],eax
+ mov eax,DWORD [edx]
+ mov ebx,DWORD [4+edx]
+ mov esi,DWORD [16+edx]
+ mov edx,DWORD [20+edx]
+ mov DWORD [32+esp],eax
+ mov DWORD [36+esp],ebx
+ mov DWORD [40+esp],ecx
+ mov DWORD [44+esp],edi
+ mov DWORD [48+esp],esi
+ mov edi,esi
+ mov esi,eax
+ cmp edx,0
+ je NEAR L$029slow_decrypt
+ cmp ecx,16
+ mov edx,ebx
+ jb NEAR L$030slow_enc_tail
+ bt DWORD [52+esp],25
+ jnc NEAR L$031slow_enc_x86
+ movq mm0,[edi]
+ movq mm4,[8+edi]
+align 16
+L$032slow_enc_loop_sse:
+ pxor mm0,[esi]
+ pxor mm4,[8+esi]
+ mov edi,DWORD [44+esp]
+ call __sse_AES_encrypt_compact
+ mov esi,DWORD [32+esp]
+ mov edi,DWORD [36+esp]
+ mov ecx,DWORD [40+esp]
+ movq [edi],mm0
+ movq [8+edi],mm4
+ lea esi,[16+esi]
+ mov DWORD [32+esp],esi
+ lea edx,[16+edi]
+ mov DWORD [36+esp],edx
+ sub ecx,16
+ cmp ecx,16
+ mov DWORD [40+esp],ecx
+ jae NEAR L$032slow_enc_loop_sse
+ test ecx,15
+ jnz NEAR L$030slow_enc_tail
+ mov esi,DWORD [48+esp]
+ movq [esi],mm0
+ movq [8+esi],mm4
+ emms
+ mov esp,DWORD [28+esp]
+ popfd
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+ pushfd
+align 16
+L$031slow_enc_x86:
+ mov eax,DWORD [edi]
+ mov ebx,DWORD [4+edi]
+align 4
+L$033slow_enc_loop_x86:
+ mov ecx,DWORD [8+edi]
+ mov edx,DWORD [12+edi]
+ xor eax,DWORD [esi]
+ xor ebx,DWORD [4+esi]
+ xor ecx,DWORD [8+esi]
+ xor edx,DWORD [12+esi]
+ mov edi,DWORD [44+esp]
+ call __x86_AES_encrypt_compact
+ mov esi,DWORD [32+esp]
+ mov edi,DWORD [36+esp]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov ecx,DWORD [40+esp]
+ lea esi,[16+esi]
+ mov DWORD [32+esp],esi
+ lea edx,[16+edi]
+ mov DWORD [36+esp],edx
+ sub ecx,16
+ cmp ecx,16
+ mov DWORD [40+esp],ecx
+ jae NEAR L$033slow_enc_loop_x86
+ test ecx,15
+ jnz NEAR L$030slow_enc_tail
+ mov esi,DWORD [48+esp]
+ mov ecx,DWORD [8+edi]
+ mov edx,DWORD [12+edi]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ mov esp,DWORD [28+esp]
+ popfd
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+ pushfd
+align 16
+L$030slow_enc_tail:
+ emms
+ mov edi,edx
+ mov ebx,16
+ sub ebx,ecx
+ cmp edi,esi
+ je NEAR L$034enc_in_place
+align 4
+dd 2767451785
+ jmp NEAR L$035enc_skip_in_place
+L$034enc_in_place:
+ lea edi,[ecx*1+edi]
+L$035enc_skip_in_place:
+ mov ecx,ebx
+ xor eax,eax
+align 4
+dd 2868115081
+ mov edi,DWORD [48+esp]
+ mov esi,edx
+ mov eax,DWORD [edi]
+ mov ebx,DWORD [4+edi]
+ mov DWORD [40+esp],16
+ jmp NEAR L$033slow_enc_loop_x86
+align 16
+L$029slow_decrypt:
+ bt DWORD [52+esp],25
+ jnc NEAR L$036slow_dec_loop_x86
+align 4
+L$037slow_dec_loop_sse:
+ movq mm0,[esi]
+ movq mm4,[8+esi]
+ mov edi,DWORD [44+esp]
+ call __sse_AES_decrypt_compact
+ mov esi,DWORD [32+esp]
+ lea eax,[60+esp]
+ mov ebx,DWORD [36+esp]
+ mov ecx,DWORD [40+esp]
+ mov edi,DWORD [48+esp]
+ movq mm1,[esi]
+ movq mm5,[8+esi]
+ pxor mm0,[edi]
+ pxor mm4,[8+edi]
+ movq [edi],mm1
+ movq [8+edi],mm5
+ sub ecx,16
+ jc NEAR L$038slow_dec_partial_sse
+ movq [ebx],mm0
+ movq [8+ebx],mm4
+ lea ebx,[16+ebx]
+ mov DWORD [36+esp],ebx
+ lea esi,[16+esi]
+ mov DWORD [32+esp],esi
+ mov DWORD [40+esp],ecx
+ jnz NEAR L$037slow_dec_loop_sse
+ emms
+ mov esp,DWORD [28+esp]
+ popfd
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+ pushfd
+align 16
+L$038slow_dec_partial_sse:
+ movq [eax],mm0
+ movq [8+eax],mm4
+ emms
+ add ecx,16
+ mov edi,ebx
+ mov esi,eax
+align 4
+dd 2767451785
+ mov esp,DWORD [28+esp]
+ popfd
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+ pushfd
+align 16
+L$036slow_dec_loop_x86:
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ lea edi,[60+esp]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov edi,DWORD [44+esp]
+ call __x86_AES_decrypt_compact
+ mov edi,DWORD [48+esp]
+ mov esi,DWORD [40+esp]
+ xor eax,DWORD [edi]
+ xor ebx,DWORD [4+edi]
+ xor ecx,DWORD [8+edi]
+ xor edx,DWORD [12+edi]
+ sub esi,16
+ jc NEAR L$039slow_dec_partial_x86
+ mov DWORD [40+esp],esi
+ mov esi,DWORD [36+esp]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ lea esi,[16+esi]
+ mov DWORD [36+esp],esi
+ lea esi,[60+esp]
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov esi,DWORD [32+esp]
+ lea esi,[16+esi]
+ mov DWORD [32+esp],esi
+ jnz NEAR L$036slow_dec_loop_x86
+ mov esp,DWORD [28+esp]
+ popfd
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+ pushfd
+align 16
+L$039slow_dec_partial_x86:
+ lea esi,[60+esp]
+ mov DWORD [esi],eax
+ mov DWORD [4+esi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ mov esi,DWORD [32+esp]
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov ecx,DWORD [40+esp]
+ mov edi,DWORD [36+esp]
+ lea esi,[60+esp]
+align 4
+dd 2767451785
+ mov esp,DWORD [28+esp]
+ popfd
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+align 16
+__x86_AES_set_encrypt_key:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [24+esp]
+ mov edi,DWORD [32+esp]
+ test esi,-1
+ jz NEAR L$040badpointer
+ test edi,-1
+ jz NEAR L$040badpointer
+ call L$041pic_point
+L$041pic_point:
+ pop ebp
+ lea ebp,[(L$AES_Te-L$041pic_point)+ebp]
+ lea ebp,[2176+ebp]
+ mov eax,DWORD [ebp-128]
+ mov ebx,DWORD [ebp-96]
+ mov ecx,DWORD [ebp-64]
+ mov edx,DWORD [ebp-32]
+ mov eax,DWORD [ebp]
+ mov ebx,DWORD [32+ebp]
+ mov ecx,DWORD [64+ebp]
+ mov edx,DWORD [96+ebp]
+ mov ecx,DWORD [28+esp]
+ cmp ecx,128
+ je NEAR L$04210rounds
+ cmp ecx,192
+ je NEAR L$04312rounds
+ cmp ecx,256
+ je NEAR L$04414rounds
+ mov eax,-2
+ jmp NEAR L$045exit
+L$04210rounds:
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ xor ecx,ecx
+ jmp NEAR L$04610shortcut
+align 4
+L$04710loop:
+ mov eax,DWORD [edi]
+ mov edx,DWORD [12+edi]
+L$04610shortcut:
+ movzx esi,dl
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ shl ebx,24
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shr edx,16
+ movzx esi,dl
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ shl ebx,8
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shl ebx,16
+ xor eax,ebx
+ xor eax,DWORD [896+ecx*4+ebp]
+ mov DWORD [16+edi],eax
+ xor eax,DWORD [4+edi]
+ mov DWORD [20+edi],eax
+ xor eax,DWORD [8+edi]
+ mov DWORD [24+edi],eax
+ xor eax,DWORD [12+edi]
+ mov DWORD [28+edi],eax
+ inc ecx
+ add edi,16
+ cmp ecx,10
+ jl NEAR L$04710loop
+ mov DWORD [80+edi],10
+ xor eax,eax
+ jmp NEAR L$045exit
+L$04312rounds:
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov ecx,DWORD [16+esi]
+ mov edx,DWORD [20+esi]
+ mov DWORD [16+edi],ecx
+ mov DWORD [20+edi],edx
+ xor ecx,ecx
+ jmp NEAR L$04812shortcut
+align 4
+L$04912loop:
+ mov eax,DWORD [edi]
+ mov edx,DWORD [20+edi]
+L$04812shortcut:
+ movzx esi,dl
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ shl ebx,24
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shr edx,16
+ movzx esi,dl
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ shl ebx,8
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shl ebx,16
+ xor eax,ebx
+ xor eax,DWORD [896+ecx*4+ebp]
+ mov DWORD [24+edi],eax
+ xor eax,DWORD [4+edi]
+ mov DWORD [28+edi],eax
+ xor eax,DWORD [8+edi]
+ mov DWORD [32+edi],eax
+ xor eax,DWORD [12+edi]
+ mov DWORD [36+edi],eax
+ cmp ecx,7
+ je NEAR L$05012break
+ inc ecx
+ xor eax,DWORD [16+edi]
+ mov DWORD [40+edi],eax
+ xor eax,DWORD [20+edi]
+ mov DWORD [44+edi],eax
+ add edi,24
+ jmp NEAR L$04912loop
+L$05012break:
+ mov DWORD [72+edi],12
+ xor eax,eax
+ jmp NEAR L$045exit
+L$04414rounds:
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [8+esi]
+ mov edx,DWORD [12+esi]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [8+edi],ecx
+ mov DWORD [12+edi],edx
+ mov eax,DWORD [16+esi]
+ mov ebx,DWORD [20+esi]
+ mov ecx,DWORD [24+esi]
+ mov edx,DWORD [28+esi]
+ mov DWORD [16+edi],eax
+ mov DWORD [20+edi],ebx
+ mov DWORD [24+edi],ecx
+ mov DWORD [28+edi],edx
+ xor ecx,ecx
+ jmp NEAR L$05114shortcut
+align 4
+L$05214loop:
+ mov edx,DWORD [28+edi]
+L$05114shortcut:
+ mov eax,DWORD [edi]
+ movzx esi,dl
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ shl ebx,24
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shr edx,16
+ movzx esi,dl
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ shl ebx,8
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shl ebx,16
+ xor eax,ebx
+ xor eax,DWORD [896+ecx*4+ebp]
+ mov DWORD [32+edi],eax
+ xor eax,DWORD [4+edi]
+ mov DWORD [36+edi],eax
+ xor eax,DWORD [8+edi]
+ mov DWORD [40+edi],eax
+ xor eax,DWORD [12+edi]
+ mov DWORD [44+edi],eax
+ cmp ecx,6
+ je NEAR L$05314break
+ inc ecx
+ mov edx,eax
+ mov eax,DWORD [16+edi]
+ movzx esi,dl
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shr edx,16
+ shl ebx,8
+ movzx esi,dl
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ movzx esi,dh
+ shl ebx,16
+ xor eax,ebx
+ movzx ebx,BYTE [esi*1+ebp-128]
+ shl ebx,24
+ xor eax,ebx
+ mov DWORD [48+edi],eax
+ xor eax,DWORD [20+edi]
+ mov DWORD [52+edi],eax
+ xor eax,DWORD [24+edi]
+ mov DWORD [56+edi],eax
+ xor eax,DWORD [28+edi]
+ mov DWORD [60+edi],eax
+ add edi,32
+ jmp NEAR L$05214loop
+L$05314break:
+ mov DWORD [48+edi],14
+ xor eax,eax
+ jmp NEAR L$045exit
+L$040badpointer:
+ mov eax,-1
+L$045exit:
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _asm_AES_set_encrypt_key
+align 16
+_asm_AES_set_encrypt_key:
+L$_asm_AES_set_encrypt_key_begin:
+ call __x86_AES_set_encrypt_key
+ ret
+global _asm_AES_set_decrypt_key
+align 16
+_asm_AES_set_decrypt_key:
+L$_asm_AES_set_decrypt_key_begin:
+ call __x86_AES_set_encrypt_key
+ cmp eax,0
+ je NEAR L$054proceed
+ ret
+L$054proceed:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [28+esp]
+ mov ecx,DWORD [240+esi]
+ lea ecx,[ecx*4]
+ lea edi,[ecx*4+esi]
+align 4
+L$055invert:
+ mov eax,DWORD [esi]
+ mov ebx,DWORD [4+esi]
+ mov ecx,DWORD [edi]
+ mov edx,DWORD [4+edi]
+ mov DWORD [edi],eax
+ mov DWORD [4+edi],ebx
+ mov DWORD [esi],ecx
+ mov DWORD [4+esi],edx
+ mov eax,DWORD [8+esi]
+ mov ebx,DWORD [12+esi]
+ mov ecx,DWORD [8+edi]
+ mov edx,DWORD [12+edi]
+ mov DWORD [8+edi],eax
+ mov DWORD [12+edi],ebx
+ mov DWORD [8+esi],ecx
+ mov DWORD [12+esi],edx
+ add esi,16
+ sub edi,16
+ cmp esi,edi
+ jne NEAR L$055invert
+ mov edi,DWORD [28+esp]
+ mov esi,DWORD [240+edi]
+ lea esi,[esi*1+esi-2]
+ lea esi,[esi*8+edi]
+ mov DWORD [28+esp],esi
+ mov eax,DWORD [16+edi]
+align 4
+L$056permute:
+ add edi,16
+ mov ebp,2155905152
+ and ebp,eax
+ lea ebx,[eax*1+eax]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and ebx,4278124286
+ and esi,454761243
+ xor ebx,esi
+ mov ebp,2155905152
+ and ebp,ebx
+ lea ecx,[ebx*1+ebx]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and ecx,4278124286
+ and esi,454761243
+ xor ebx,eax
+ xor ecx,esi
+ mov ebp,2155905152
+ and ebp,ecx
+ lea edx,[ecx*1+ecx]
+ mov esi,ebp
+ shr ebp,7
+ xor ecx,eax
+ sub esi,ebp
+ and edx,4278124286
+ and esi,454761243
+ rol eax,8
+ xor edx,esi
+ mov ebp,DWORD [4+edi]
+ xor eax,ebx
+ xor ebx,edx
+ xor eax,ecx
+ rol ebx,24
+ xor ecx,edx
+ xor eax,edx
+ rol ecx,16
+ xor eax,ebx
+ rol edx,8
+ xor eax,ecx
+ mov ebx,ebp
+ xor eax,edx
+ mov DWORD [edi],eax
+ mov ebp,2155905152
+ and ebp,ebx
+ lea ecx,[ebx*1+ebx]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and ecx,4278124286
+ and esi,454761243
+ xor ecx,esi
+ mov ebp,2155905152
+ and ebp,ecx
+ lea edx,[ecx*1+ecx]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and edx,4278124286
+ and esi,454761243
+ xor ecx,ebx
+ xor edx,esi
+ mov ebp,2155905152
+ and ebp,edx
+ lea eax,[edx*1+edx]
+ mov esi,ebp
+ shr ebp,7
+ xor edx,ebx
+ sub esi,ebp
+ and eax,4278124286
+ and esi,454761243
+ rol ebx,8
+ xor eax,esi
+ mov ebp,DWORD [8+edi]
+ xor ebx,ecx
+ xor ecx,eax
+ xor ebx,edx
+ rol ecx,24
+ xor edx,eax
+ xor ebx,eax
+ rol edx,16
+ xor ebx,ecx
+ rol eax,8
+ xor ebx,edx
+ mov ecx,ebp
+ xor ebx,eax
+ mov DWORD [4+edi],ebx
+ mov ebp,2155905152
+ and ebp,ecx
+ lea edx,[ecx*1+ecx]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and edx,4278124286
+ and esi,454761243
+ xor edx,esi
+ mov ebp,2155905152
+ and ebp,edx
+ lea eax,[edx*1+edx]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and eax,4278124286
+ and esi,454761243
+ xor edx,ecx
+ xor eax,esi
+ mov ebp,2155905152
+ and ebp,eax
+ lea ebx,[eax*1+eax]
+ mov esi,ebp
+ shr ebp,7
+ xor eax,ecx
+ sub esi,ebp
+ and ebx,4278124286
+ and esi,454761243
+ rol ecx,8
+ xor ebx,esi
+ mov ebp,DWORD [12+edi]
+ xor ecx,edx
+ xor edx,ebx
+ xor ecx,eax
+ rol edx,24
+ xor eax,ebx
+ xor ecx,ebx
+ rol eax,16
+ xor ecx,edx
+ rol ebx,8
+ xor ecx,eax
+ mov edx,ebp
+ xor ecx,ebx
+ mov DWORD [8+edi],ecx
+ mov ebp,2155905152
+ and ebp,edx
+ lea eax,[edx*1+edx]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and eax,4278124286
+ and esi,454761243
+ xor eax,esi
+ mov ebp,2155905152
+ and ebp,eax
+ lea ebx,[eax*1+eax]
+ mov esi,ebp
+ shr ebp,7
+ sub esi,ebp
+ and ebx,4278124286
+ and esi,454761243
+ xor eax,edx
+ xor ebx,esi
+ mov ebp,2155905152
+ and ebp,ebx
+ lea ecx,[ebx*1+ebx]
+ mov esi,ebp
+ shr ebp,7
+ xor ebx,edx
+ sub esi,ebp
+ and ecx,4278124286
+ and esi,454761243
+ rol edx,8
+ xor ecx,esi
+ mov ebp,DWORD [16+edi]
+ xor edx,eax
+ xor eax,ecx
+ xor edx,ebx
+ rol eax,24
+ xor ebx,ecx
+ xor edx,ecx
+ rol ebx,16
+ xor edx,eax
+ rol ecx,8
+ xor edx,ebx
+ mov eax,ebp
+ xor edx,ecx
+ mov DWORD [12+edi],edx
+ cmp edi,DWORD [28+esp]
+ jb NEAR L$056permute
+ xor eax,eax
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+db 65,69,83,32,102,111,114,32,120,56,54,44,32,67,82,89
+db 80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114
+db 111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
+segment .bss
+common _OPENSSL_ia32cap_P 16
diff --git a/win-x86/crypto/aes/aesni-x86.asm b/win-x86/crypto/aes/aesni-x86.asm
new file mode 100644
index 0000000..a9a5956
--- /dev/null
+++ b/win-x86/crypto/aes/aesni-x86.asm
@@ -0,0 +1,2424 @@
+%ifidn __OUTPUT_FORMAT__,obj
+section code use32 class=code align=64
+%elifidn __OUTPUT_FORMAT__,win32
+%ifdef __YASM_VERSION_ID__
+%if __YASM_VERSION_ID__ < 01010000h
+%error yasm version 1.1.0 or later needed.
+%endif
+; Yasm automatically includes .00 and complains about redefining it.
+; https://www.tortall.net/projects/yasm/manual/html/objfmt-win32-safeseh.html
+%else
+$@feat.00 equ 1
+%endif
+section .text code align=64
+%else
+section .text code
+%endif
+;extern _OPENSSL_ia32cap_P
+global _aesni_encrypt
+align 16
+_aesni_encrypt:
+L$_aesni_encrypt_begin:
+ mov eax,DWORD [4+esp]
+ mov edx,DWORD [12+esp]
+ movups xmm2,[eax]
+ mov ecx,DWORD [240+edx]
+ mov eax,DWORD [8+esp]
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$000enc1_loop_1:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$000enc1_loop_1
+db 102,15,56,221,209
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ movups [eax],xmm2
+ pxor xmm2,xmm2
+ ret
+global _aesni_decrypt
+align 16
+_aesni_decrypt:
+L$_aesni_decrypt_begin:
+ mov eax,DWORD [4+esp]
+ mov edx,DWORD [12+esp]
+ movups xmm2,[eax]
+ mov ecx,DWORD [240+edx]
+ mov eax,DWORD [8+esp]
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$001dec1_loop_2:
+db 102,15,56,222,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$001dec1_loop_2
+db 102,15,56,223,209
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ movups [eax],xmm2
+ pxor xmm2,xmm2
+ ret
+align 16
+__aesni_encrypt2:
+ movups xmm0,[edx]
+ shl ecx,4
+ movups xmm1,[16+edx]
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ movups xmm0,[32+edx]
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+ add ecx,16
+L$002enc2_loop:
+db 102,15,56,220,209
+db 102,15,56,220,217
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,220,208
+db 102,15,56,220,216
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$002enc2_loop
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,221,208
+db 102,15,56,221,216
+ ret
+align 16
+__aesni_decrypt2:
+ movups xmm0,[edx]
+ shl ecx,4
+ movups xmm1,[16+edx]
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ movups xmm0,[32+edx]
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+ add ecx,16
+L$003dec2_loop:
+db 102,15,56,222,209
+db 102,15,56,222,217
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,222,208
+db 102,15,56,222,216
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$003dec2_loop
+db 102,15,56,222,209
+db 102,15,56,222,217
+db 102,15,56,223,208
+db 102,15,56,223,216
+ ret
+align 16
+__aesni_encrypt3:
+ movups xmm0,[edx]
+ shl ecx,4
+ movups xmm1,[16+edx]
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ pxor xmm4,xmm0
+ movups xmm0,[32+edx]
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+ add ecx,16
+L$004enc3_loop:
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,220,225
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,220,208
+db 102,15,56,220,216
+db 102,15,56,220,224
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$004enc3_loop
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,220,225
+db 102,15,56,221,208
+db 102,15,56,221,216
+db 102,15,56,221,224
+ ret
+align 16
+__aesni_decrypt3:
+ movups xmm0,[edx]
+ shl ecx,4
+ movups xmm1,[16+edx]
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ pxor xmm4,xmm0
+ movups xmm0,[32+edx]
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+ add ecx,16
+L$005dec3_loop:
+db 102,15,56,222,209
+db 102,15,56,222,217
+db 102,15,56,222,225
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,222,208
+db 102,15,56,222,216
+db 102,15,56,222,224
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$005dec3_loop
+db 102,15,56,222,209
+db 102,15,56,222,217
+db 102,15,56,222,225
+db 102,15,56,223,208
+db 102,15,56,223,216
+db 102,15,56,223,224
+ ret
+align 16
+__aesni_encrypt4:
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ shl ecx,4
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ pxor xmm4,xmm0
+ pxor xmm5,xmm0
+ movups xmm0,[32+edx]
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+db 15,31,64,0
+ add ecx,16
+L$006enc4_loop:
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,220,225
+db 102,15,56,220,233
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,220,208
+db 102,15,56,220,216
+db 102,15,56,220,224
+db 102,15,56,220,232
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$006enc4_loop
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,220,225
+db 102,15,56,220,233
+db 102,15,56,221,208
+db 102,15,56,221,216
+db 102,15,56,221,224
+db 102,15,56,221,232
+ ret
+align 16
+__aesni_decrypt4:
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ shl ecx,4
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ pxor xmm4,xmm0
+ pxor xmm5,xmm0
+ movups xmm0,[32+edx]
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+db 15,31,64,0
+ add ecx,16
+L$007dec4_loop:
+db 102,15,56,222,209
+db 102,15,56,222,217
+db 102,15,56,222,225
+db 102,15,56,222,233
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,222,208
+db 102,15,56,222,216
+db 102,15,56,222,224
+db 102,15,56,222,232
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$007dec4_loop
+db 102,15,56,222,209
+db 102,15,56,222,217
+db 102,15,56,222,225
+db 102,15,56,222,233
+db 102,15,56,223,208
+db 102,15,56,223,216
+db 102,15,56,223,224
+db 102,15,56,223,232
+ ret
+align 16
+__aesni_encrypt6:
+ movups xmm0,[edx]
+ shl ecx,4
+ movups xmm1,[16+edx]
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ pxor xmm4,xmm0
+db 102,15,56,220,209
+ pxor xmm5,xmm0
+ pxor xmm6,xmm0
+db 102,15,56,220,217
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+db 102,15,56,220,225
+ pxor xmm7,xmm0
+ movups xmm0,[ecx*1+edx]
+ add ecx,16
+ jmp NEAR L$008_aesni_encrypt6_inner
+align 16
+L$009enc6_loop:
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,220,225
+L$008_aesni_encrypt6_inner:
+db 102,15,56,220,233
+db 102,15,56,220,241
+db 102,15,56,220,249
+L$_aesni_encrypt6_enter:
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,220,208
+db 102,15,56,220,216
+db 102,15,56,220,224
+db 102,15,56,220,232
+db 102,15,56,220,240
+db 102,15,56,220,248
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$009enc6_loop
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,220,225
+db 102,15,56,220,233
+db 102,15,56,220,241
+db 102,15,56,220,249
+db 102,15,56,221,208
+db 102,15,56,221,216
+db 102,15,56,221,224
+db 102,15,56,221,232
+db 102,15,56,221,240
+db 102,15,56,221,248
+ ret
+align 16
+__aesni_decrypt6:
+ movups xmm0,[edx]
+ shl ecx,4
+ movups xmm1,[16+edx]
+ xorps xmm2,xmm0
+ pxor xmm3,xmm0
+ pxor xmm4,xmm0
+db 102,15,56,222,209
+ pxor xmm5,xmm0
+ pxor xmm6,xmm0
+db 102,15,56,222,217
+ lea edx,[32+ecx*1+edx]
+ neg ecx
+db 102,15,56,222,225
+ pxor xmm7,xmm0
+ movups xmm0,[ecx*1+edx]
+ add ecx,16
+ jmp NEAR L$010_aesni_decrypt6_inner
+align 16
+L$011dec6_loop:
+db 102,15,56,222,209
+db 102,15,56,222,217
+db 102,15,56,222,225
+L$010_aesni_decrypt6_inner:
+db 102,15,56,222,233
+db 102,15,56,222,241
+db 102,15,56,222,249
+L$_aesni_decrypt6_enter:
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,222,208
+db 102,15,56,222,216
+db 102,15,56,222,224
+db 102,15,56,222,232
+db 102,15,56,222,240
+db 102,15,56,222,248
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$011dec6_loop
+db 102,15,56,222,209
+db 102,15,56,222,217
+db 102,15,56,222,225
+db 102,15,56,222,233
+db 102,15,56,222,241
+db 102,15,56,222,249
+db 102,15,56,223,208
+db 102,15,56,223,216
+db 102,15,56,223,224
+db 102,15,56,223,232
+db 102,15,56,223,240
+db 102,15,56,223,248
+ ret
+global _aesni_ecb_encrypt
+align 16
+_aesni_ecb_encrypt:
+L$_aesni_ecb_encrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [24+esp]
+ mov eax,DWORD [28+esp]
+ mov edx,DWORD [32+esp]
+ mov ebx,DWORD [36+esp]
+ and eax,-16
+ jz NEAR L$012ecb_ret
+ mov ecx,DWORD [240+edx]
+ test ebx,ebx
+ jz NEAR L$013ecb_decrypt
+ mov ebp,edx
+ mov ebx,ecx
+ cmp eax,96
+ jb NEAR L$014ecb_enc_tail
+ movdqu xmm2,[esi]
+ movdqu xmm3,[16+esi]
+ movdqu xmm4,[32+esi]
+ movdqu xmm5,[48+esi]
+ movdqu xmm6,[64+esi]
+ movdqu xmm7,[80+esi]
+ lea esi,[96+esi]
+ sub eax,96
+ jmp NEAR L$015ecb_enc_loop6_enter
+align 16
+L$016ecb_enc_loop6:
+ movups [edi],xmm2
+ movdqu xmm2,[esi]
+ movups [16+edi],xmm3
+ movdqu xmm3,[16+esi]
+ movups [32+edi],xmm4
+ movdqu xmm4,[32+esi]
+ movups [48+edi],xmm5
+ movdqu xmm5,[48+esi]
+ movups [64+edi],xmm6
+ movdqu xmm6,[64+esi]
+ movups [80+edi],xmm7
+ lea edi,[96+edi]
+ movdqu xmm7,[80+esi]
+ lea esi,[96+esi]
+L$015ecb_enc_loop6_enter:
+ call __aesni_encrypt6
+ mov edx,ebp
+ mov ecx,ebx
+ sub eax,96
+ jnc NEAR L$016ecb_enc_loop6
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ movups [64+edi],xmm6
+ movups [80+edi],xmm7
+ lea edi,[96+edi]
+ add eax,96
+ jz NEAR L$012ecb_ret
+L$014ecb_enc_tail:
+ movups xmm2,[esi]
+ cmp eax,32
+ jb NEAR L$017ecb_enc_one
+ movups xmm3,[16+esi]
+ je NEAR L$018ecb_enc_two
+ movups xmm4,[32+esi]
+ cmp eax,64
+ jb NEAR L$019ecb_enc_three
+ movups xmm5,[48+esi]
+ je NEAR L$020ecb_enc_four
+ movups xmm6,[64+esi]
+ xorps xmm7,xmm7
+ call __aesni_encrypt6
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ movups [64+edi],xmm6
+ jmp NEAR L$012ecb_ret
+align 16
+L$017ecb_enc_one:
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$021enc1_loop_3:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$021enc1_loop_3
+db 102,15,56,221,209
+ movups [edi],xmm2
+ jmp NEAR L$012ecb_ret
+align 16
+L$018ecb_enc_two:
+ call __aesni_encrypt2
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ jmp NEAR L$012ecb_ret
+align 16
+L$019ecb_enc_three:
+ call __aesni_encrypt3
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ jmp NEAR L$012ecb_ret
+align 16
+L$020ecb_enc_four:
+ call __aesni_encrypt4
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ jmp NEAR L$012ecb_ret
+align 16
+L$013ecb_decrypt:
+ mov ebp,edx
+ mov ebx,ecx
+ cmp eax,96
+ jb NEAR L$022ecb_dec_tail
+ movdqu xmm2,[esi]
+ movdqu xmm3,[16+esi]
+ movdqu xmm4,[32+esi]
+ movdqu xmm5,[48+esi]
+ movdqu xmm6,[64+esi]
+ movdqu xmm7,[80+esi]
+ lea esi,[96+esi]
+ sub eax,96
+ jmp NEAR L$023ecb_dec_loop6_enter
+align 16
+L$024ecb_dec_loop6:
+ movups [edi],xmm2
+ movdqu xmm2,[esi]
+ movups [16+edi],xmm3
+ movdqu xmm3,[16+esi]
+ movups [32+edi],xmm4
+ movdqu xmm4,[32+esi]
+ movups [48+edi],xmm5
+ movdqu xmm5,[48+esi]
+ movups [64+edi],xmm6
+ movdqu xmm6,[64+esi]
+ movups [80+edi],xmm7
+ lea edi,[96+edi]
+ movdqu xmm7,[80+esi]
+ lea esi,[96+esi]
+L$023ecb_dec_loop6_enter:
+ call __aesni_decrypt6
+ mov edx,ebp
+ mov ecx,ebx
+ sub eax,96
+ jnc NEAR L$024ecb_dec_loop6
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ movups [64+edi],xmm6
+ movups [80+edi],xmm7
+ lea edi,[96+edi]
+ add eax,96
+ jz NEAR L$012ecb_ret
+L$022ecb_dec_tail:
+ movups xmm2,[esi]
+ cmp eax,32
+ jb NEAR L$025ecb_dec_one
+ movups xmm3,[16+esi]
+ je NEAR L$026ecb_dec_two
+ movups xmm4,[32+esi]
+ cmp eax,64
+ jb NEAR L$027ecb_dec_three
+ movups xmm5,[48+esi]
+ je NEAR L$028ecb_dec_four
+ movups xmm6,[64+esi]
+ xorps xmm7,xmm7
+ call __aesni_decrypt6
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ movups [64+edi],xmm6
+ jmp NEAR L$012ecb_ret
+align 16
+L$025ecb_dec_one:
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$029dec1_loop_4:
+db 102,15,56,222,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$029dec1_loop_4
+db 102,15,56,223,209
+ movups [edi],xmm2
+ jmp NEAR L$012ecb_ret
+align 16
+L$026ecb_dec_two:
+ call __aesni_decrypt2
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ jmp NEAR L$012ecb_ret
+align 16
+L$027ecb_dec_three:
+ call __aesni_decrypt3
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ jmp NEAR L$012ecb_ret
+align 16
+L$028ecb_dec_four:
+ call __aesni_decrypt4
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+L$012ecb_ret:
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ pxor xmm3,xmm3
+ pxor xmm4,xmm4
+ pxor xmm5,xmm5
+ pxor xmm6,xmm6
+ pxor xmm7,xmm7
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _aesni_ccm64_encrypt_blocks
+align 16
+_aesni_ccm64_encrypt_blocks:
+L$_aesni_ccm64_encrypt_blocks_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [24+esp]
+ mov eax,DWORD [28+esp]
+ mov edx,DWORD [32+esp]
+ mov ebx,DWORD [36+esp]
+ mov ecx,DWORD [40+esp]
+ mov ebp,esp
+ sub esp,60
+ and esp,-16
+ mov DWORD [48+esp],ebp
+ movdqu xmm7,[ebx]
+ movdqu xmm3,[ecx]
+ mov ecx,DWORD [240+edx]
+ mov DWORD [esp],202182159
+ mov DWORD [4+esp],134810123
+ mov DWORD [8+esp],67438087
+ mov DWORD [12+esp],66051
+ mov ebx,1
+ xor ebp,ebp
+ mov DWORD [16+esp],ebx
+ mov DWORD [20+esp],ebp
+ mov DWORD [24+esp],ebp
+ mov DWORD [28+esp],ebp
+ shl ecx,4
+ mov ebx,16
+ lea ebp,[edx]
+ movdqa xmm5,[esp]
+ movdqa xmm2,xmm7
+ lea edx,[32+ecx*1+edx]
+ sub ebx,ecx
+db 102,15,56,0,253
+L$030ccm64_enc_outer:
+ movups xmm0,[ebp]
+ mov ecx,ebx
+ movups xmm6,[esi]
+ xorps xmm2,xmm0
+ movups xmm1,[16+ebp]
+ xorps xmm0,xmm6
+ xorps xmm3,xmm0
+ movups xmm0,[32+ebp]
+L$031ccm64_enc2_loop:
+db 102,15,56,220,209
+db 102,15,56,220,217
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,220,208
+db 102,15,56,220,216
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$031ccm64_enc2_loop
+db 102,15,56,220,209
+db 102,15,56,220,217
+ paddq xmm7,[16+esp]
+ dec eax
+db 102,15,56,221,208
+db 102,15,56,221,216
+ lea esi,[16+esi]
+ xorps xmm6,xmm2
+ movdqa xmm2,xmm7
+ movups [edi],xmm6
+db 102,15,56,0,213
+ lea edi,[16+edi]
+ jnz NEAR L$030ccm64_enc_outer
+ mov esp,DWORD [48+esp]
+ mov edi,DWORD [40+esp]
+ movups [edi],xmm3
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ pxor xmm3,xmm3
+ pxor xmm4,xmm4
+ pxor xmm5,xmm5
+ pxor xmm6,xmm6
+ pxor xmm7,xmm7
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _aesni_ccm64_decrypt_blocks
+align 16
+_aesni_ccm64_decrypt_blocks:
+L$_aesni_ccm64_decrypt_blocks_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [24+esp]
+ mov eax,DWORD [28+esp]
+ mov edx,DWORD [32+esp]
+ mov ebx,DWORD [36+esp]
+ mov ecx,DWORD [40+esp]
+ mov ebp,esp
+ sub esp,60
+ and esp,-16
+ mov DWORD [48+esp],ebp
+ movdqu xmm7,[ebx]
+ movdqu xmm3,[ecx]
+ mov ecx,DWORD [240+edx]
+ mov DWORD [esp],202182159
+ mov DWORD [4+esp],134810123
+ mov DWORD [8+esp],67438087
+ mov DWORD [12+esp],66051
+ mov ebx,1
+ xor ebp,ebp
+ mov DWORD [16+esp],ebx
+ mov DWORD [20+esp],ebp
+ mov DWORD [24+esp],ebp
+ mov DWORD [28+esp],ebp
+ movdqa xmm5,[esp]
+ movdqa xmm2,xmm7
+ mov ebp,edx
+ mov ebx,ecx
+db 102,15,56,0,253
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$032enc1_loop_5:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$032enc1_loop_5
+db 102,15,56,221,209
+ shl ebx,4
+ mov ecx,16
+ movups xmm6,[esi]
+ paddq xmm7,[16+esp]
+ lea esi,[16+esi]
+ sub ecx,ebx
+ lea edx,[32+ebx*1+ebp]
+ mov ebx,ecx
+ jmp NEAR L$033ccm64_dec_outer
+align 16
+L$033ccm64_dec_outer:
+ xorps xmm6,xmm2
+ movdqa xmm2,xmm7
+ movups [edi],xmm6
+ lea edi,[16+edi]
+db 102,15,56,0,213
+ sub eax,1
+ jz NEAR L$034ccm64_dec_break
+ movups xmm0,[ebp]
+ mov ecx,ebx
+ movups xmm1,[16+ebp]
+ xorps xmm6,xmm0
+ xorps xmm2,xmm0
+ xorps xmm3,xmm6
+ movups xmm0,[32+ebp]
+L$035ccm64_dec2_loop:
+db 102,15,56,220,209
+db 102,15,56,220,217
+ movups xmm1,[ecx*1+edx]
+ add ecx,32
+db 102,15,56,220,208
+db 102,15,56,220,216
+ movups xmm0,[ecx*1+edx-16]
+ jnz NEAR L$035ccm64_dec2_loop
+ movups xmm6,[esi]
+ paddq xmm7,[16+esp]
+db 102,15,56,220,209
+db 102,15,56,220,217
+db 102,15,56,221,208
+db 102,15,56,221,216
+ lea esi,[16+esi]
+ jmp NEAR L$033ccm64_dec_outer
+align 16
+L$034ccm64_dec_break:
+ mov ecx,DWORD [240+ebp]
+ mov edx,ebp
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ xorps xmm6,xmm0
+ lea edx,[32+edx]
+ xorps xmm3,xmm6
+L$036enc1_loop_6:
+db 102,15,56,220,217
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$036enc1_loop_6
+db 102,15,56,221,217
+ mov esp,DWORD [48+esp]
+ mov edi,DWORD [40+esp]
+ movups [edi],xmm3
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ pxor xmm3,xmm3
+ pxor xmm4,xmm4
+ pxor xmm5,xmm5
+ pxor xmm6,xmm6
+ pxor xmm7,xmm7
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _aesni_ctr32_encrypt_blocks
+align 16
+_aesni_ctr32_encrypt_blocks:
+L$_aesni_ctr32_encrypt_blocks_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [24+esp]
+ mov eax,DWORD [28+esp]
+ mov edx,DWORD [32+esp]
+ mov ebx,DWORD [36+esp]
+ mov ebp,esp
+ sub esp,88
+ and esp,-16
+ mov DWORD [80+esp],ebp
+ cmp eax,1
+ je NEAR L$037ctr32_one_shortcut
+ movdqu xmm7,[ebx]
+ mov DWORD [esp],202182159
+ mov DWORD [4+esp],134810123
+ mov DWORD [8+esp],67438087
+ mov DWORD [12+esp],66051
+ mov ecx,6
+ xor ebp,ebp
+ mov DWORD [16+esp],ecx
+ mov DWORD [20+esp],ecx
+ mov DWORD [24+esp],ecx
+ mov DWORD [28+esp],ebp
+db 102,15,58,22,251,3
+db 102,15,58,34,253,3
+ mov ecx,DWORD [240+edx]
+ bswap ebx
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ movdqa xmm2,[esp]
+db 102,15,58,34,195,0
+ lea ebp,[3+ebx]
+db 102,15,58,34,205,0
+ inc ebx
+db 102,15,58,34,195,1
+ inc ebp
+db 102,15,58,34,205,1
+ inc ebx
+db 102,15,58,34,195,2
+ inc ebp
+db 102,15,58,34,205,2
+ movdqa [48+esp],xmm0
+db 102,15,56,0,194
+ movdqu xmm6,[edx]
+ movdqa [64+esp],xmm1
+db 102,15,56,0,202
+ pshufd xmm2,xmm0,192
+ pshufd xmm3,xmm0,128
+ cmp eax,6
+ jb NEAR L$038ctr32_tail
+ pxor xmm7,xmm6
+ shl ecx,4
+ mov ebx,16
+ movdqa [32+esp],xmm7
+ mov ebp,edx
+ sub ebx,ecx
+ lea edx,[32+ecx*1+edx]
+ sub eax,6
+ jmp NEAR L$039ctr32_loop6
+align 16
+L$039ctr32_loop6:
+ pshufd xmm4,xmm0,64
+ movdqa xmm0,[32+esp]
+ pshufd xmm5,xmm1,192
+ pxor xmm2,xmm0
+ pshufd xmm6,xmm1,128
+ pxor xmm3,xmm0
+ pshufd xmm7,xmm1,64
+ movups xmm1,[16+ebp]
+ pxor xmm4,xmm0
+ pxor xmm5,xmm0
+db 102,15,56,220,209
+ pxor xmm6,xmm0
+ pxor xmm7,xmm0
+db 102,15,56,220,217
+ movups xmm0,[32+ebp]
+ mov ecx,ebx
+db 102,15,56,220,225
+db 102,15,56,220,233
+db 102,15,56,220,241
+db 102,15,56,220,249
+ call L$_aesni_encrypt6_enter
+ movups xmm1,[esi]
+ movups xmm0,[16+esi]
+ xorps xmm2,xmm1
+ movups xmm1,[32+esi]
+ xorps xmm3,xmm0
+ movups [edi],xmm2
+ movdqa xmm0,[16+esp]
+ xorps xmm4,xmm1
+ movdqa xmm1,[64+esp]
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ paddd xmm1,xmm0
+ paddd xmm0,[48+esp]
+ movdqa xmm2,[esp]
+ movups xmm3,[48+esi]
+ movups xmm4,[64+esi]
+ xorps xmm5,xmm3
+ movups xmm3,[80+esi]
+ lea esi,[96+esi]
+ movdqa [48+esp],xmm0
+db 102,15,56,0,194
+ xorps xmm6,xmm4
+ movups [48+edi],xmm5
+ xorps xmm7,xmm3
+ movdqa [64+esp],xmm1
+db 102,15,56,0,202
+ movups [64+edi],xmm6
+ pshufd xmm2,xmm0,192
+ movups [80+edi],xmm7
+ lea edi,[96+edi]
+ pshufd xmm3,xmm0,128
+ sub eax,6
+ jnc NEAR L$039ctr32_loop6
+ add eax,6
+ jz NEAR L$040ctr32_ret
+ movdqu xmm7,[ebp]
+ mov edx,ebp
+ pxor xmm7,[32+esp]
+ mov ecx,DWORD [240+ebp]
+L$038ctr32_tail:
+ por xmm2,xmm7
+ cmp eax,2
+ jb NEAR L$041ctr32_one
+ pshufd xmm4,xmm0,64
+ por xmm3,xmm7
+ je NEAR L$042ctr32_two
+ pshufd xmm5,xmm1,192
+ por xmm4,xmm7
+ cmp eax,4
+ jb NEAR L$043ctr32_three
+ pshufd xmm6,xmm1,128
+ por xmm5,xmm7
+ je NEAR L$044ctr32_four
+ por xmm6,xmm7
+ call __aesni_encrypt6
+ movups xmm1,[esi]
+ movups xmm0,[16+esi]
+ xorps xmm2,xmm1
+ movups xmm1,[32+esi]
+ xorps xmm3,xmm0
+ movups xmm0,[48+esi]
+ xorps xmm4,xmm1
+ movups xmm1,[64+esi]
+ xorps xmm5,xmm0
+ movups [edi],xmm2
+ xorps xmm6,xmm1
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ movups [64+edi],xmm6
+ jmp NEAR L$040ctr32_ret
+align 16
+L$037ctr32_one_shortcut:
+ movups xmm2,[ebx]
+ mov ecx,DWORD [240+edx]
+L$041ctr32_one:
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$045enc1_loop_7:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$045enc1_loop_7
+db 102,15,56,221,209
+ movups xmm6,[esi]
+ xorps xmm6,xmm2
+ movups [edi],xmm6
+ jmp NEAR L$040ctr32_ret
+align 16
+L$042ctr32_two:
+ call __aesni_encrypt2
+ movups xmm5,[esi]
+ movups xmm6,[16+esi]
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ jmp NEAR L$040ctr32_ret
+align 16
+L$043ctr32_three:
+ call __aesni_encrypt3
+ movups xmm5,[esi]
+ movups xmm6,[16+esi]
+ xorps xmm2,xmm5
+ movups xmm7,[32+esi]
+ xorps xmm3,xmm6
+ movups [edi],xmm2
+ xorps xmm4,xmm7
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ jmp NEAR L$040ctr32_ret
+align 16
+L$044ctr32_four:
+ call __aesni_encrypt4
+ movups xmm6,[esi]
+ movups xmm7,[16+esi]
+ movups xmm1,[32+esi]
+ xorps xmm2,xmm6
+ movups xmm0,[48+esi]
+ xorps xmm3,xmm7
+ movups [edi],xmm2
+ xorps xmm4,xmm1
+ movups [16+edi],xmm3
+ xorps xmm5,xmm0
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+L$040ctr32_ret:
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ pxor xmm3,xmm3
+ pxor xmm4,xmm4
+ movdqa [32+esp],xmm0
+ pxor xmm5,xmm5
+ movdqa [48+esp],xmm0
+ pxor xmm6,xmm6
+ movdqa [64+esp],xmm0
+ pxor xmm7,xmm7
+ mov esp,DWORD [80+esp]
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _aesni_xts_encrypt
+align 16
+_aesni_xts_encrypt:
+L$_aesni_xts_encrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov edx,DWORD [36+esp]
+ mov esi,DWORD [40+esp]
+ mov ecx,DWORD [240+edx]
+ movups xmm2,[esi]
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$046enc1_loop_8:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$046enc1_loop_8
+db 102,15,56,221,209
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [24+esp]
+ mov eax,DWORD [28+esp]
+ mov edx,DWORD [32+esp]
+ mov ebp,esp
+ sub esp,120
+ mov ecx,DWORD [240+edx]
+ and esp,-16
+ mov DWORD [96+esp],135
+ mov DWORD [100+esp],0
+ mov DWORD [104+esp],1
+ mov DWORD [108+esp],0
+ mov DWORD [112+esp],eax
+ mov DWORD [116+esp],ebp
+ movdqa xmm1,xmm2
+ pxor xmm0,xmm0
+ movdqa xmm3,[96+esp]
+ pcmpgtd xmm0,xmm1
+ and eax,-16
+ mov ebp,edx
+ mov ebx,ecx
+ sub eax,96
+ jc NEAR L$047xts_enc_short
+ shl ecx,4
+ mov ebx,16
+ sub ebx,ecx
+ lea edx,[32+ecx*1+edx]
+ jmp NEAR L$048xts_enc_loop6
+align 16
+L$048xts_enc_loop6:
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [16+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [32+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [48+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm7,xmm0,19
+ movdqa [64+esp],xmm1
+ paddq xmm1,xmm1
+ movups xmm0,[ebp]
+ pand xmm7,xmm3
+ movups xmm2,[esi]
+ pxor xmm7,xmm1
+ mov ecx,ebx
+ movdqu xmm3,[16+esi]
+ xorps xmm2,xmm0
+ movdqu xmm4,[32+esi]
+ pxor xmm3,xmm0
+ movdqu xmm5,[48+esi]
+ pxor xmm4,xmm0
+ movdqu xmm6,[64+esi]
+ pxor xmm5,xmm0
+ movdqu xmm1,[80+esi]
+ pxor xmm6,xmm0
+ lea esi,[96+esi]
+ pxor xmm2,[esp]
+ movdqa [80+esp],xmm7
+ pxor xmm7,xmm1
+ movups xmm1,[16+ebp]
+ pxor xmm3,[16+esp]
+ pxor xmm4,[32+esp]
+db 102,15,56,220,209
+ pxor xmm5,[48+esp]
+ pxor xmm6,[64+esp]
+db 102,15,56,220,217
+ pxor xmm7,xmm0
+ movups xmm0,[32+ebp]
+db 102,15,56,220,225
+db 102,15,56,220,233
+db 102,15,56,220,241
+db 102,15,56,220,249
+ call L$_aesni_encrypt6_enter
+ movdqa xmm1,[80+esp]
+ pxor xmm0,xmm0
+ xorps xmm2,[esp]
+ pcmpgtd xmm0,xmm1
+ xorps xmm3,[16+esp]
+ movups [edi],xmm2
+ xorps xmm4,[32+esp]
+ movups [16+edi],xmm3
+ xorps xmm5,[48+esp]
+ movups [32+edi],xmm4
+ xorps xmm6,[64+esp]
+ movups [48+edi],xmm5
+ xorps xmm7,xmm1
+ movups [64+edi],xmm6
+ pshufd xmm2,xmm0,19
+ movups [80+edi],xmm7
+ lea edi,[96+edi]
+ movdqa xmm3,[96+esp]
+ pxor xmm0,xmm0
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ sub eax,96
+ jnc NEAR L$048xts_enc_loop6
+ mov ecx,DWORD [240+ebp]
+ mov edx,ebp
+ mov ebx,ecx
+L$047xts_enc_short:
+ add eax,96
+ jz NEAR L$049xts_enc_done6x
+ movdqa xmm5,xmm1
+ cmp eax,32
+ jb NEAR L$050xts_enc_one
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ je NEAR L$051xts_enc_two
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa xmm6,xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ cmp eax,64
+ jb NEAR L$052xts_enc_three
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa xmm7,xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ movdqa [esp],xmm5
+ movdqa [16+esp],xmm6
+ je NEAR L$053xts_enc_four
+ movdqa [32+esp],xmm7
+ pshufd xmm7,xmm0,19
+ movdqa [48+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm7,xmm3
+ pxor xmm7,xmm1
+ movdqu xmm2,[esi]
+ movdqu xmm3,[16+esi]
+ movdqu xmm4,[32+esi]
+ pxor xmm2,[esp]
+ movdqu xmm5,[48+esi]
+ pxor xmm3,[16+esp]
+ movdqu xmm6,[64+esi]
+ pxor xmm4,[32+esp]
+ lea esi,[80+esi]
+ pxor xmm5,[48+esp]
+ movdqa [64+esp],xmm7
+ pxor xmm6,xmm7
+ call __aesni_encrypt6
+ movaps xmm1,[64+esp]
+ xorps xmm2,[esp]
+ xorps xmm3,[16+esp]
+ xorps xmm4,[32+esp]
+ movups [edi],xmm2
+ xorps xmm5,[48+esp]
+ movups [16+edi],xmm3
+ xorps xmm6,xmm1
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ movups [64+edi],xmm6
+ lea edi,[80+edi]
+ jmp NEAR L$054xts_enc_done
+align 16
+L$050xts_enc_one:
+ movups xmm2,[esi]
+ lea esi,[16+esi]
+ xorps xmm2,xmm5
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$055enc1_loop_9:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$055enc1_loop_9
+db 102,15,56,221,209
+ xorps xmm2,xmm5
+ movups [edi],xmm2
+ lea edi,[16+edi]
+ movdqa xmm1,xmm5
+ jmp NEAR L$054xts_enc_done
+align 16
+L$051xts_enc_two:
+ movaps xmm6,xmm1
+ movups xmm2,[esi]
+ movups xmm3,[16+esi]
+ lea esi,[32+esi]
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ call __aesni_encrypt2
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ lea edi,[32+edi]
+ movdqa xmm1,xmm6
+ jmp NEAR L$054xts_enc_done
+align 16
+L$052xts_enc_three:
+ movaps xmm7,xmm1
+ movups xmm2,[esi]
+ movups xmm3,[16+esi]
+ movups xmm4,[32+esi]
+ lea esi,[48+esi]
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ xorps xmm4,xmm7
+ call __aesni_encrypt3
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ xorps xmm4,xmm7
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ lea edi,[48+edi]
+ movdqa xmm1,xmm7
+ jmp NEAR L$054xts_enc_done
+align 16
+L$053xts_enc_four:
+ movaps xmm6,xmm1
+ movups xmm2,[esi]
+ movups xmm3,[16+esi]
+ movups xmm4,[32+esi]
+ xorps xmm2,[esp]
+ movups xmm5,[48+esi]
+ lea esi,[64+esi]
+ xorps xmm3,[16+esp]
+ xorps xmm4,xmm7
+ xorps xmm5,xmm6
+ call __aesni_encrypt4
+ xorps xmm2,[esp]
+ xorps xmm3,[16+esp]
+ xorps xmm4,xmm7
+ movups [edi],xmm2
+ xorps xmm5,xmm6
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ lea edi,[64+edi]
+ movdqa xmm1,xmm6
+ jmp NEAR L$054xts_enc_done
+align 16
+L$049xts_enc_done6x:
+ mov eax,DWORD [112+esp]
+ and eax,15
+ jz NEAR L$056xts_enc_ret
+ movdqa xmm5,xmm1
+ mov DWORD [112+esp],eax
+ jmp NEAR L$057xts_enc_steal
+align 16
+L$054xts_enc_done:
+ mov eax,DWORD [112+esp]
+ pxor xmm0,xmm0
+ and eax,15
+ jz NEAR L$056xts_enc_ret
+ pcmpgtd xmm0,xmm1
+ mov DWORD [112+esp],eax
+ pshufd xmm5,xmm0,19
+ paddq xmm1,xmm1
+ pand xmm5,[96+esp]
+ pxor xmm5,xmm1
+L$057xts_enc_steal:
+ movzx ecx,BYTE [esi]
+ movzx edx,BYTE [edi-16]
+ lea esi,[1+esi]
+ mov BYTE [edi-16],cl
+ mov BYTE [edi],dl
+ lea edi,[1+edi]
+ sub eax,1
+ jnz NEAR L$057xts_enc_steal
+ sub edi,DWORD [112+esp]
+ mov edx,ebp
+ mov ecx,ebx
+ movups xmm2,[edi-16]
+ xorps xmm2,xmm5
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$058enc1_loop_10:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$058enc1_loop_10
+db 102,15,56,221,209
+ xorps xmm2,xmm5
+ movups [edi-16],xmm2
+L$056xts_enc_ret:
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ movdqa [esp],xmm0
+ pxor xmm3,xmm3
+ movdqa [16+esp],xmm0
+ pxor xmm4,xmm4
+ movdqa [32+esp],xmm0
+ pxor xmm5,xmm5
+ movdqa [48+esp],xmm0
+ pxor xmm6,xmm6
+ movdqa [64+esp],xmm0
+ pxor xmm7,xmm7
+ movdqa [80+esp],xmm0
+ mov esp,DWORD [116+esp]
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _aesni_xts_decrypt
+align 16
+_aesni_xts_decrypt:
+L$_aesni_xts_decrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov edx,DWORD [36+esp]
+ mov esi,DWORD [40+esp]
+ mov ecx,DWORD [240+edx]
+ movups xmm2,[esi]
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$059enc1_loop_11:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$059enc1_loop_11
+db 102,15,56,221,209
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [24+esp]
+ mov eax,DWORD [28+esp]
+ mov edx,DWORD [32+esp]
+ mov ebp,esp
+ sub esp,120
+ and esp,-16
+ xor ebx,ebx
+ test eax,15
+ setnz bl
+ shl ebx,4
+ sub eax,ebx
+ mov DWORD [96+esp],135
+ mov DWORD [100+esp],0
+ mov DWORD [104+esp],1
+ mov DWORD [108+esp],0
+ mov DWORD [112+esp],eax
+ mov DWORD [116+esp],ebp
+ mov ecx,DWORD [240+edx]
+ mov ebp,edx
+ mov ebx,ecx
+ movdqa xmm1,xmm2
+ pxor xmm0,xmm0
+ movdqa xmm3,[96+esp]
+ pcmpgtd xmm0,xmm1
+ and eax,-16
+ sub eax,96
+ jc NEAR L$060xts_dec_short
+ shl ecx,4
+ mov ebx,16
+ sub ebx,ecx
+ lea edx,[32+ecx*1+edx]
+ jmp NEAR L$061xts_dec_loop6
+align 16
+L$061xts_dec_loop6:
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [16+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [32+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa [48+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ pshufd xmm7,xmm0,19
+ movdqa [64+esp],xmm1
+ paddq xmm1,xmm1
+ movups xmm0,[ebp]
+ pand xmm7,xmm3
+ movups xmm2,[esi]
+ pxor xmm7,xmm1
+ mov ecx,ebx
+ movdqu xmm3,[16+esi]
+ xorps xmm2,xmm0
+ movdqu xmm4,[32+esi]
+ pxor xmm3,xmm0
+ movdqu xmm5,[48+esi]
+ pxor xmm4,xmm0
+ movdqu xmm6,[64+esi]
+ pxor xmm5,xmm0
+ movdqu xmm1,[80+esi]
+ pxor xmm6,xmm0
+ lea esi,[96+esi]
+ pxor xmm2,[esp]
+ movdqa [80+esp],xmm7
+ pxor xmm7,xmm1
+ movups xmm1,[16+ebp]
+ pxor xmm3,[16+esp]
+ pxor xmm4,[32+esp]
+db 102,15,56,222,209
+ pxor xmm5,[48+esp]
+ pxor xmm6,[64+esp]
+db 102,15,56,222,217
+ pxor xmm7,xmm0
+ movups xmm0,[32+ebp]
+db 102,15,56,222,225
+db 102,15,56,222,233
+db 102,15,56,222,241
+db 102,15,56,222,249
+ call L$_aesni_decrypt6_enter
+ movdqa xmm1,[80+esp]
+ pxor xmm0,xmm0
+ xorps xmm2,[esp]
+ pcmpgtd xmm0,xmm1
+ xorps xmm3,[16+esp]
+ movups [edi],xmm2
+ xorps xmm4,[32+esp]
+ movups [16+edi],xmm3
+ xorps xmm5,[48+esp]
+ movups [32+edi],xmm4
+ xorps xmm6,[64+esp]
+ movups [48+edi],xmm5
+ xorps xmm7,xmm1
+ movups [64+edi],xmm6
+ pshufd xmm2,xmm0,19
+ movups [80+edi],xmm7
+ lea edi,[96+edi]
+ movdqa xmm3,[96+esp]
+ pxor xmm0,xmm0
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ sub eax,96
+ jnc NEAR L$061xts_dec_loop6
+ mov ecx,DWORD [240+ebp]
+ mov edx,ebp
+ mov ebx,ecx
+L$060xts_dec_short:
+ add eax,96
+ jz NEAR L$062xts_dec_done6x
+ movdqa xmm5,xmm1
+ cmp eax,32
+ jb NEAR L$063xts_dec_one
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ je NEAR L$064xts_dec_two
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa xmm6,xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ cmp eax,64
+ jb NEAR L$065xts_dec_three
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa xmm7,xmm1
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+ movdqa [esp],xmm5
+ movdqa [16+esp],xmm6
+ je NEAR L$066xts_dec_four
+ movdqa [32+esp],xmm7
+ pshufd xmm7,xmm0,19
+ movdqa [48+esp],xmm1
+ paddq xmm1,xmm1
+ pand xmm7,xmm3
+ pxor xmm7,xmm1
+ movdqu xmm2,[esi]
+ movdqu xmm3,[16+esi]
+ movdqu xmm4,[32+esi]
+ pxor xmm2,[esp]
+ movdqu xmm5,[48+esi]
+ pxor xmm3,[16+esp]
+ movdqu xmm6,[64+esi]
+ pxor xmm4,[32+esp]
+ lea esi,[80+esi]
+ pxor xmm5,[48+esp]
+ movdqa [64+esp],xmm7
+ pxor xmm6,xmm7
+ call __aesni_decrypt6
+ movaps xmm1,[64+esp]
+ xorps xmm2,[esp]
+ xorps xmm3,[16+esp]
+ xorps xmm4,[32+esp]
+ movups [edi],xmm2
+ xorps xmm5,[48+esp]
+ movups [16+edi],xmm3
+ xorps xmm6,xmm1
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ movups [64+edi],xmm6
+ lea edi,[80+edi]
+ jmp NEAR L$067xts_dec_done
+align 16
+L$063xts_dec_one:
+ movups xmm2,[esi]
+ lea esi,[16+esi]
+ xorps xmm2,xmm5
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$068dec1_loop_12:
+db 102,15,56,222,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$068dec1_loop_12
+db 102,15,56,223,209
+ xorps xmm2,xmm5
+ movups [edi],xmm2
+ lea edi,[16+edi]
+ movdqa xmm1,xmm5
+ jmp NEAR L$067xts_dec_done
+align 16
+L$064xts_dec_two:
+ movaps xmm6,xmm1
+ movups xmm2,[esi]
+ movups xmm3,[16+esi]
+ lea esi,[32+esi]
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ call __aesni_decrypt2
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ lea edi,[32+edi]
+ movdqa xmm1,xmm6
+ jmp NEAR L$067xts_dec_done
+align 16
+L$065xts_dec_three:
+ movaps xmm7,xmm1
+ movups xmm2,[esi]
+ movups xmm3,[16+esi]
+ movups xmm4,[32+esi]
+ lea esi,[48+esi]
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ xorps xmm4,xmm7
+ call __aesni_decrypt3
+ xorps xmm2,xmm5
+ xorps xmm3,xmm6
+ xorps xmm4,xmm7
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ lea edi,[48+edi]
+ movdqa xmm1,xmm7
+ jmp NEAR L$067xts_dec_done
+align 16
+L$066xts_dec_four:
+ movaps xmm6,xmm1
+ movups xmm2,[esi]
+ movups xmm3,[16+esi]
+ movups xmm4,[32+esi]
+ xorps xmm2,[esp]
+ movups xmm5,[48+esi]
+ lea esi,[64+esi]
+ xorps xmm3,[16+esp]
+ xorps xmm4,xmm7
+ xorps xmm5,xmm6
+ call __aesni_decrypt4
+ xorps xmm2,[esp]
+ xorps xmm3,[16+esp]
+ xorps xmm4,xmm7
+ movups [edi],xmm2
+ xorps xmm5,xmm6
+ movups [16+edi],xmm3
+ movups [32+edi],xmm4
+ movups [48+edi],xmm5
+ lea edi,[64+edi]
+ movdqa xmm1,xmm6
+ jmp NEAR L$067xts_dec_done
+align 16
+L$062xts_dec_done6x:
+ mov eax,DWORD [112+esp]
+ and eax,15
+ jz NEAR L$069xts_dec_ret
+ mov DWORD [112+esp],eax
+ jmp NEAR L$070xts_dec_only_one_more
+align 16
+L$067xts_dec_done:
+ mov eax,DWORD [112+esp]
+ pxor xmm0,xmm0
+ and eax,15
+ jz NEAR L$069xts_dec_ret
+ pcmpgtd xmm0,xmm1
+ mov DWORD [112+esp],eax
+ pshufd xmm2,xmm0,19
+ pxor xmm0,xmm0
+ movdqa xmm3,[96+esp]
+ paddq xmm1,xmm1
+ pand xmm2,xmm3
+ pcmpgtd xmm0,xmm1
+ pxor xmm1,xmm2
+L$070xts_dec_only_one_more:
+ pshufd xmm5,xmm0,19
+ movdqa xmm6,xmm1
+ paddq xmm1,xmm1
+ pand xmm5,xmm3
+ pxor xmm5,xmm1
+ mov edx,ebp
+ mov ecx,ebx
+ movups xmm2,[esi]
+ xorps xmm2,xmm5
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$071dec1_loop_13:
+db 102,15,56,222,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$071dec1_loop_13
+db 102,15,56,223,209
+ xorps xmm2,xmm5
+ movups [edi],xmm2
+L$072xts_dec_steal:
+ movzx ecx,BYTE [16+esi]
+ movzx edx,BYTE [edi]
+ lea esi,[1+esi]
+ mov BYTE [edi],cl
+ mov BYTE [16+edi],dl
+ lea edi,[1+edi]
+ sub eax,1
+ jnz NEAR L$072xts_dec_steal
+ sub edi,DWORD [112+esp]
+ mov edx,ebp
+ mov ecx,ebx
+ movups xmm2,[edi]
+ xorps xmm2,xmm6
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$073dec1_loop_14:
+db 102,15,56,222,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$073dec1_loop_14
+db 102,15,56,223,209
+ xorps xmm2,xmm6
+ movups [edi],xmm2
+L$069xts_dec_ret:
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ movdqa [esp],xmm0
+ pxor xmm3,xmm3
+ movdqa [16+esp],xmm0
+ pxor xmm4,xmm4
+ movdqa [32+esp],xmm0
+ pxor xmm5,xmm5
+ movdqa [48+esp],xmm0
+ pxor xmm6,xmm6
+ movdqa [64+esp],xmm0
+ pxor xmm7,xmm7
+ movdqa [80+esp],xmm0
+ mov esp,DWORD [116+esp]
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _aesni_cbc_encrypt
+align 16
+_aesni_cbc_encrypt:
+L$_aesni_cbc_encrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov ebx,esp
+ mov edi,DWORD [24+esp]
+ sub ebx,24
+ mov eax,DWORD [28+esp]
+ and ebx,-16
+ mov edx,DWORD [32+esp]
+ mov ebp,DWORD [36+esp]
+ test eax,eax
+ jz NEAR L$074cbc_abort
+ cmp DWORD [40+esp],0
+ xchg ebx,esp
+ movups xmm7,[ebp]
+ mov ecx,DWORD [240+edx]
+ mov ebp,edx
+ mov DWORD [16+esp],ebx
+ mov ebx,ecx
+ je NEAR L$075cbc_decrypt
+ movaps xmm2,xmm7
+ cmp eax,16
+ jb NEAR L$076cbc_enc_tail
+ sub eax,16
+ jmp NEAR L$077cbc_enc_loop
+align 16
+L$077cbc_enc_loop:
+ movups xmm7,[esi]
+ lea esi,[16+esi]
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ xorps xmm7,xmm0
+ lea edx,[32+edx]
+ xorps xmm2,xmm7
+L$078enc1_loop_15:
+db 102,15,56,220,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$078enc1_loop_15
+db 102,15,56,221,209
+ mov ecx,ebx
+ mov edx,ebp
+ movups [edi],xmm2
+ lea edi,[16+edi]
+ sub eax,16
+ jnc NEAR L$077cbc_enc_loop
+ add eax,16
+ jnz NEAR L$076cbc_enc_tail
+ movaps xmm7,xmm2
+ pxor xmm2,xmm2
+ jmp NEAR L$079cbc_ret
+L$076cbc_enc_tail:
+ mov ecx,eax
+dd 2767451785
+ mov ecx,16
+ sub ecx,eax
+ xor eax,eax
+dd 2868115081
+ lea edi,[edi-16]
+ mov ecx,ebx
+ mov esi,edi
+ mov edx,ebp
+ jmp NEAR L$077cbc_enc_loop
+align 16
+L$075cbc_decrypt:
+ cmp eax,80
+ jbe NEAR L$080cbc_dec_tail
+ movaps [esp],xmm7
+ sub eax,80
+ jmp NEAR L$081cbc_dec_loop6_enter
+align 16
+L$082cbc_dec_loop6:
+ movaps [esp],xmm0
+ movups [edi],xmm7
+ lea edi,[16+edi]
+L$081cbc_dec_loop6_enter:
+ movdqu xmm2,[esi]
+ movdqu xmm3,[16+esi]
+ movdqu xmm4,[32+esi]
+ movdqu xmm5,[48+esi]
+ movdqu xmm6,[64+esi]
+ movdqu xmm7,[80+esi]
+ call __aesni_decrypt6
+ movups xmm1,[esi]
+ movups xmm0,[16+esi]
+ xorps xmm2,[esp]
+ xorps xmm3,xmm1
+ movups xmm1,[32+esi]
+ xorps xmm4,xmm0
+ movups xmm0,[48+esi]
+ xorps xmm5,xmm1
+ movups xmm1,[64+esi]
+ xorps xmm6,xmm0
+ movups xmm0,[80+esi]
+ xorps xmm7,xmm1
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ lea esi,[96+esi]
+ movups [32+edi],xmm4
+ mov ecx,ebx
+ movups [48+edi],xmm5
+ mov edx,ebp
+ movups [64+edi],xmm6
+ lea edi,[80+edi]
+ sub eax,96
+ ja NEAR L$082cbc_dec_loop6
+ movaps xmm2,xmm7
+ movaps xmm7,xmm0
+ add eax,80
+ jle NEAR L$083cbc_dec_clear_tail_collected
+ movups [edi],xmm2
+ lea edi,[16+edi]
+L$080cbc_dec_tail:
+ movups xmm2,[esi]
+ movaps xmm6,xmm2
+ cmp eax,16
+ jbe NEAR L$084cbc_dec_one
+ movups xmm3,[16+esi]
+ movaps xmm5,xmm3
+ cmp eax,32
+ jbe NEAR L$085cbc_dec_two
+ movups xmm4,[32+esi]
+ cmp eax,48
+ jbe NEAR L$086cbc_dec_three
+ movups xmm5,[48+esi]
+ cmp eax,64
+ jbe NEAR L$087cbc_dec_four
+ movups xmm6,[64+esi]
+ movaps [esp],xmm7
+ movups xmm2,[esi]
+ xorps xmm7,xmm7
+ call __aesni_decrypt6
+ movups xmm1,[esi]
+ movups xmm0,[16+esi]
+ xorps xmm2,[esp]
+ xorps xmm3,xmm1
+ movups xmm1,[32+esi]
+ xorps xmm4,xmm0
+ movups xmm0,[48+esi]
+ xorps xmm5,xmm1
+ movups xmm7,[64+esi]
+ xorps xmm6,xmm0
+ movups [edi],xmm2
+ movups [16+edi],xmm3
+ pxor xmm3,xmm3
+ movups [32+edi],xmm4
+ pxor xmm4,xmm4
+ movups [48+edi],xmm5
+ pxor xmm5,xmm5
+ lea edi,[64+edi]
+ movaps xmm2,xmm6
+ pxor xmm6,xmm6
+ sub eax,80
+ jmp NEAR L$088cbc_dec_tail_collected
+align 16
+L$084cbc_dec_one:
+ movups xmm0,[edx]
+ movups xmm1,[16+edx]
+ lea edx,[32+edx]
+ xorps xmm2,xmm0
+L$089dec1_loop_16:
+db 102,15,56,222,209
+ dec ecx
+ movups xmm1,[edx]
+ lea edx,[16+edx]
+ jnz NEAR L$089dec1_loop_16
+db 102,15,56,223,209
+ xorps xmm2,xmm7
+ movaps xmm7,xmm6
+ sub eax,16
+ jmp NEAR L$088cbc_dec_tail_collected
+align 16
+L$085cbc_dec_two:
+ call __aesni_decrypt2
+ xorps xmm2,xmm7
+ xorps xmm3,xmm6
+ movups [edi],xmm2
+ movaps xmm2,xmm3
+ pxor xmm3,xmm3
+ lea edi,[16+edi]
+ movaps xmm7,xmm5
+ sub eax,32
+ jmp NEAR L$088cbc_dec_tail_collected
+align 16
+L$086cbc_dec_three:
+ call __aesni_decrypt3
+ xorps xmm2,xmm7
+ xorps xmm3,xmm6
+ xorps xmm4,xmm5
+ movups [edi],xmm2
+ movaps xmm2,xmm4
+ pxor xmm4,xmm4
+ movups [16+edi],xmm3
+ pxor xmm3,xmm3
+ lea edi,[32+edi]
+ movups xmm7,[32+esi]
+ sub eax,48
+ jmp NEAR L$088cbc_dec_tail_collected
+align 16
+L$087cbc_dec_four:
+ call __aesni_decrypt4
+ movups xmm1,[16+esi]
+ movups xmm0,[32+esi]
+ xorps xmm2,xmm7
+ movups xmm7,[48+esi]
+ xorps xmm3,xmm6
+ movups [edi],xmm2
+ xorps xmm4,xmm1
+ movups [16+edi],xmm3
+ pxor xmm3,xmm3
+ xorps xmm5,xmm0
+ movups [32+edi],xmm4
+ pxor xmm4,xmm4
+ lea edi,[48+edi]
+ movaps xmm2,xmm5
+ pxor xmm5,xmm5
+ sub eax,64
+ jmp NEAR L$088cbc_dec_tail_collected
+align 16
+L$083cbc_dec_clear_tail_collected:
+ pxor xmm3,xmm3
+ pxor xmm4,xmm4
+ pxor xmm5,xmm5
+ pxor xmm6,xmm6
+L$088cbc_dec_tail_collected:
+ and eax,15
+ jnz NEAR L$090cbc_dec_tail_partial
+ movups [edi],xmm2
+ pxor xmm0,xmm0
+ jmp NEAR L$079cbc_ret
+align 16
+L$090cbc_dec_tail_partial:
+ movaps [esp],xmm2
+ pxor xmm0,xmm0
+ mov ecx,16
+ mov esi,esp
+ sub ecx,eax
+dd 2767451785
+ movdqa [esp],xmm2
+L$079cbc_ret:
+ mov esp,DWORD [16+esp]
+ mov ebp,DWORD [36+esp]
+ pxor xmm2,xmm2
+ pxor xmm1,xmm1
+ movups [ebp],xmm7
+ pxor xmm7,xmm7
+L$074cbc_abort:
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+align 16
+__aesni_set_encrypt_key:
+ push ebp
+ push ebx
+ test eax,eax
+ jz NEAR L$091bad_pointer
+ test edx,edx
+ jz NEAR L$091bad_pointer
+ call L$092pic
+L$092pic:
+ pop ebx
+ lea ebx,[(L$key_const-L$092pic)+ebx]
+ lea ebp,[_OPENSSL_ia32cap_P]
+ movups xmm0,[eax]
+ xorps xmm4,xmm4
+ mov ebp,DWORD [4+ebp]
+ lea edx,[16+edx]
+ and ebp,268437504
+ cmp ecx,256
+ je NEAR L$09314rounds
+ cmp ecx,192
+ je NEAR L$09412rounds
+ cmp ecx,128
+ jne NEAR L$095bad_keybits
+align 16
+L$09610rounds:
+ cmp ebp,268435456
+ je NEAR L$09710rounds_alt
+ mov ecx,9
+ movups [edx-16],xmm0
+db 102,15,58,223,200,1
+ call L$098key_128_cold
+db 102,15,58,223,200,2
+ call L$099key_128
+db 102,15,58,223,200,4
+ call L$099key_128
+db 102,15,58,223,200,8
+ call L$099key_128
+db 102,15,58,223,200,16
+ call L$099key_128
+db 102,15,58,223,200,32
+ call L$099key_128
+db 102,15,58,223,200,64
+ call L$099key_128
+db 102,15,58,223,200,128
+ call L$099key_128
+db 102,15,58,223,200,27
+ call L$099key_128
+db 102,15,58,223,200,54
+ call L$099key_128
+ movups [edx],xmm0
+ mov DWORD [80+edx],ecx
+ jmp NEAR L$100good_key
+align 16
+L$099key_128:
+ movups [edx],xmm0
+ lea edx,[16+edx]
+L$098key_128_cold:
+ shufps xmm4,xmm0,16
+ xorps xmm0,xmm4
+ shufps xmm4,xmm0,140
+ xorps xmm0,xmm4
+ shufps xmm1,xmm1,255
+ xorps xmm0,xmm1
+ ret
+align 16
+L$09710rounds_alt:
+ movdqa xmm5,[ebx]
+ mov ecx,8
+ movdqa xmm4,[32+ebx]
+ movdqa xmm2,xmm0
+ movdqu [edx-16],xmm0
+L$101loop_key128:
+db 102,15,56,0,197
+db 102,15,56,221,196
+ pslld xmm4,1
+ lea edx,[16+edx]
+ movdqa xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm2,xmm3
+ pxor xmm0,xmm2
+ movdqu [edx-16],xmm0
+ movdqa xmm2,xmm0
+ dec ecx
+ jnz NEAR L$101loop_key128
+ movdqa xmm4,[48+ebx]
+db 102,15,56,0,197
+db 102,15,56,221,196
+ pslld xmm4,1
+ movdqa xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm2,xmm3
+ pxor xmm0,xmm2
+ movdqu [edx],xmm0
+ movdqa xmm2,xmm0
+db 102,15,56,0,197
+db 102,15,56,221,196
+ movdqa xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm3,xmm2
+ pslldq xmm2,4
+ pxor xmm2,xmm3
+ pxor xmm0,xmm2
+ movdqu [16+edx],xmm0
+ mov ecx,9
+ mov DWORD [96+edx],ecx
+ jmp NEAR L$100good_key
+align 16
+L$09412rounds:
+ movq xmm2,[16+eax]
+ cmp ebp,268435456
+ je NEAR L$10212rounds_alt
+ mov ecx,11
+ movups [edx-16],xmm0
+db 102,15,58,223,202,1
+ call L$103key_192a_cold
+db 102,15,58,223,202,2
+ call L$104key_192b
+db 102,15,58,223,202,4
+ call L$105key_192a
+db 102,15,58,223,202,8
+ call L$104key_192b
+db 102,15,58,223,202,16
+ call L$105key_192a
+db 102,15,58,223,202,32
+ call L$104key_192b
+db 102,15,58,223,202,64
+ call L$105key_192a
+db 102,15,58,223,202,128
+ call L$104key_192b
+ movups [edx],xmm0
+ mov DWORD [48+edx],ecx
+ jmp NEAR L$100good_key
+align 16
+L$105key_192a:
+ movups [edx],xmm0
+ lea edx,[16+edx]
+align 16
+L$103key_192a_cold:
+ movaps xmm5,xmm2
+L$106key_192b_warm:
+ shufps xmm4,xmm0,16
+ movdqa xmm3,xmm2
+ xorps xmm0,xmm4
+ shufps xmm4,xmm0,140
+ pslldq xmm3,4
+ xorps xmm0,xmm4
+ pshufd xmm1,xmm1,85
+ pxor xmm2,xmm3
+ pxor xmm0,xmm1
+ pshufd xmm3,xmm0,255
+ pxor xmm2,xmm3
+ ret
+align 16
+L$104key_192b:
+ movaps xmm3,xmm0
+ shufps xmm5,xmm0,68
+ movups [edx],xmm5
+ shufps xmm3,xmm2,78
+ movups [16+edx],xmm3
+ lea edx,[32+edx]
+ jmp NEAR L$106key_192b_warm
+align 16
+L$10212rounds_alt:
+ movdqa xmm5,[16+ebx]
+ movdqa xmm4,[32+ebx]
+ mov ecx,8
+ movdqu [edx-16],xmm0
+L$107loop_key192:
+ movq [edx],xmm2
+ movdqa xmm1,xmm2
+db 102,15,56,0,213
+db 102,15,56,221,212
+ pslld xmm4,1
+ lea edx,[24+edx]
+ movdqa xmm3,xmm0
+ pslldq xmm0,4
+ pxor xmm3,xmm0
+ pslldq xmm0,4
+ pxor xmm3,xmm0
+ pslldq xmm0,4
+ pxor xmm0,xmm3
+ pshufd xmm3,xmm0,255
+ pxor xmm3,xmm1
+ pslldq xmm1,4
+ pxor xmm3,xmm1
+ pxor xmm0,xmm2
+ pxor xmm2,xmm3
+ movdqu [edx-16],xmm0
+ dec ecx
+ jnz NEAR L$107loop_key192
+ mov ecx,11
+ mov DWORD [32+edx],ecx
+ jmp NEAR L$100good_key
+align 16
+L$09314rounds:
+ movups xmm2,[16+eax]
+ lea edx,[16+edx]
+ cmp ebp,268435456
+ je NEAR L$10814rounds_alt
+ mov ecx,13
+ movups [edx-32],xmm0
+ movups [edx-16],xmm2
+db 102,15,58,223,202,1
+ call L$109key_256a_cold
+db 102,15,58,223,200,1
+ call L$110key_256b
+db 102,15,58,223,202,2
+ call L$111key_256a
+db 102,15,58,223,200,2
+ call L$110key_256b
+db 102,15,58,223,202,4
+ call L$111key_256a
+db 102,15,58,223,200,4
+ call L$110key_256b
+db 102,15,58,223,202,8
+ call L$111key_256a
+db 102,15,58,223,200,8
+ call L$110key_256b
+db 102,15,58,223,202,16
+ call L$111key_256a
+db 102,15,58,223,200,16
+ call L$110key_256b
+db 102,15,58,223,202,32
+ call L$111key_256a
+db 102,15,58,223,200,32
+ call L$110key_256b
+db 102,15,58,223,202,64
+ call L$111key_256a
+ movups [edx],xmm0
+ mov DWORD [16+edx],ecx
+ xor eax,eax
+ jmp NEAR L$100good_key
+align 16
+L$111key_256a:
+ movups [edx],xmm2
+ lea edx,[16+edx]
+L$109key_256a_cold:
+ shufps xmm4,xmm0,16
+ xorps xmm0,xmm4
+ shufps xmm4,xmm0,140
+ xorps xmm0,xmm4
+ shufps xmm1,xmm1,255
+ xorps xmm0,xmm1
+ ret
+align 16
+L$110key_256b:
+ movups [edx],xmm0
+ lea edx,[16+edx]
+ shufps xmm4,xmm2,16
+ xorps xmm2,xmm4
+ shufps xmm4,xmm2,140
+ xorps xmm2,xmm4
+ shufps xmm1,xmm1,170
+ xorps xmm2,xmm1
+ ret
+align 16
+L$10814rounds_alt:
+ movdqa xmm5,[ebx]
+ movdqa xmm4,[32+ebx]
+ mov ecx,7
+ movdqu [edx-32],xmm0
+ movdqa xmm1,xmm2
+ movdqu [edx-16],xmm2
+L$112loop_key256:
+db 102,15,56,0,213
+db 102,15,56,221,212
+ movdqa xmm3,xmm0
+ pslldq xmm0,4
+ pxor xmm3,xmm0
+ pslldq xmm0,4
+ pxor xmm3,xmm0
+ pslldq xmm0,4
+ pxor xmm0,xmm3
+ pslld xmm4,1
+ pxor xmm0,xmm2
+ movdqu [edx],xmm0
+ dec ecx
+ jz NEAR L$113done_key256
+ pshufd xmm2,xmm0,255
+ pxor xmm3,xmm3
+db 102,15,56,221,211
+ movdqa xmm3,xmm1
+ pslldq xmm1,4
+ pxor xmm3,xmm1
+ pslldq xmm1,4
+ pxor xmm3,xmm1
+ pslldq xmm1,4
+ pxor xmm1,xmm3
+ pxor xmm2,xmm1
+ movdqu [16+edx],xmm2
+ lea edx,[32+edx]
+ movdqa xmm1,xmm2
+ jmp NEAR L$112loop_key256
+L$113done_key256:
+ mov ecx,13
+ mov DWORD [16+edx],ecx
+L$100good_key:
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ pxor xmm3,xmm3
+ pxor xmm4,xmm4
+ pxor xmm5,xmm5
+ xor eax,eax
+ pop ebx
+ pop ebp
+ ret
+align 4
+L$091bad_pointer:
+ mov eax,-1
+ pop ebx
+ pop ebp
+ ret
+align 4
+L$095bad_keybits:
+ pxor xmm0,xmm0
+ mov eax,-2
+ pop ebx
+ pop ebp
+ ret
+global _aesni_set_encrypt_key
+align 16
+_aesni_set_encrypt_key:
+L$_aesni_set_encrypt_key_begin:
+ mov eax,DWORD [4+esp]
+ mov ecx,DWORD [8+esp]
+ mov edx,DWORD [12+esp]
+ call __aesni_set_encrypt_key
+ ret
+global _aesni_set_decrypt_key
+align 16
+_aesni_set_decrypt_key:
+L$_aesni_set_decrypt_key_begin:
+ mov eax,DWORD [4+esp]
+ mov ecx,DWORD [8+esp]
+ mov edx,DWORD [12+esp]
+ call __aesni_set_encrypt_key
+ mov edx,DWORD [12+esp]
+ shl ecx,4
+ test eax,eax
+ jnz NEAR L$114dec_key_ret
+ lea eax,[16+ecx*1+edx]
+ movups xmm0,[edx]
+ movups xmm1,[eax]
+ movups [eax],xmm0
+ movups [edx],xmm1
+ lea edx,[16+edx]
+ lea eax,[eax-16]
+L$115dec_key_inverse:
+ movups xmm0,[edx]
+ movups xmm1,[eax]
+db 102,15,56,219,192
+db 102,15,56,219,201
+ lea edx,[16+edx]
+ lea eax,[eax-16]
+ movups [16+eax],xmm0
+ movups [edx-16],xmm1
+ cmp eax,edx
+ ja NEAR L$115dec_key_inverse
+ movups xmm0,[edx]
+db 102,15,56,219,192
+ movups [edx],xmm0
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ xor eax,eax
+L$114dec_key_ret:
+ ret
+align 64
+L$key_const:
+dd 202313229,202313229,202313229,202313229
+dd 67569157,67569157,67569157,67569157
+dd 1,1,1,1
+dd 27,27,27,27
+db 65,69,83,32,102,111,114,32,73,110,116,101,108,32,65,69
+db 83,45,78,73,44,32,67,82,89,80,84,79,71,65,77,83
+db 32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115
+db 115,108,46,111,114,103,62,0
+segment .bss
+common _OPENSSL_ia32cap_P 16
diff --git a/win-x86/crypto/aes/vpaes-x86.asm b/win-x86/crypto/aes/vpaes-x86.asm
new file mode 100644
index 0000000..b08b056
--- /dev/null
+++ b/win-x86/crypto/aes/vpaes-x86.asm
@@ -0,0 +1,649 @@
+%ifidn __OUTPUT_FORMAT__,obj
+section code use32 class=code align=64
+%elifidn __OUTPUT_FORMAT__,win32
+%ifdef __YASM_VERSION_ID__
+%if __YASM_VERSION_ID__ < 01010000h
+%error yasm version 1.1.0 or later needed.
+%endif
+; Yasm automatically includes .00 and complains about redefining it.
+; https://www.tortall.net/projects/yasm/manual/html/objfmt-win32-safeseh.html
+%else
+$@feat.00 equ 1
+%endif
+section .text code align=64
+%else
+section .text code
+%endif
+align 64
+L$_vpaes_consts:
+dd 218628480,235210255,168496130,67568393
+dd 252381056,17041926,33884169,51187212
+dd 252645135,252645135,252645135,252645135
+dd 1512730624,3266504856,1377990664,3401244816
+dd 830229760,1275146365,2969422977,3447763452
+dd 3411033600,2979783055,338359620,2782886510
+dd 4209124096,907596821,221174255,1006095553
+dd 191964160,3799684038,3164090317,1589111125
+dd 182528256,1777043520,2877432650,3265356744
+dd 1874708224,3503451415,3305285752,363511674
+dd 1606117888,3487855781,1093350906,2384367825
+dd 197121,67569157,134941193,202313229
+dd 67569157,134941193,202313229,197121
+dd 134941193,202313229,197121,67569157
+dd 202313229,197121,67569157,134941193
+dd 33619971,100992007,168364043,235736079
+dd 235736079,33619971,100992007,168364043
+dd 168364043,235736079,33619971,100992007
+dd 100992007,168364043,235736079,33619971
+dd 50462976,117835012,185207048,252579084
+dd 252314880,51251460,117574920,184942860
+dd 184682752,252054788,50987272,118359308
+dd 118099200,185467140,251790600,50727180
+dd 2946363062,528716217,1300004225,1881839624
+dd 1532713819,1532713819,1532713819,1532713819
+dd 3602276352,4288629033,3737020424,4153884961
+dd 1354558464,32357713,2958822624,3775749553
+dd 1201988352,132424512,1572796698,503232858
+dd 2213177600,1597421020,4103937655,675398315
+dd 2749646592,4273543773,1511898873,121693092
+dd 3040248576,1103263732,2871565598,1608280554
+dd 2236667136,2588920351,482954393,64377734
+dd 3069987328,291237287,2117370568,3650299247
+dd 533321216,3573750986,2572112006,1401264716
+dd 1339849704,2721158661,548607111,3445553514
+dd 2128193280,3054596040,2183486460,1257083700
+dd 655635200,1165381986,3923443150,2344132524
+dd 190078720,256924420,290342170,357187870
+dd 1610966272,2263057382,4103205268,309794674
+dd 2592527872,2233205587,1335446729,3402964816
+dd 3973531904,3225098121,3002836325,1918774430
+dd 3870401024,2102906079,2284471353,4117666579
+dd 617007872,1021508343,366931923,691083277
+dd 2528395776,3491914898,2968704004,1613121270
+dd 3445188352,3247741094,844474987,4093578302
+dd 651481088,1190302358,1689581232,574775300
+dd 4289380608,206939853,2555985458,2489840491
+dd 2130264064,327674451,3566485037,3349835193
+dd 2470714624,316102159,3636825756,3393945945
+db 86,101,99,116,111,114,32,80,101,114,109,117,116,97,116,105
+db 111,110,32,65,69,83,32,102,111,114,32,120,56,54,47,83
+db 83,83,69,51,44,32,77,105,107,101,32,72,97,109,98,117
+db 114,103,32,40,83,116,97,110,102,111,114,100,32,85,110,105
+db 118,101,114,115,105,116,121,41,0
+align 64
+align 16
+__vpaes_preheat:
+ add ebp,DWORD [esp]
+ movdqa xmm7,[ebp-48]
+ movdqa xmm6,[ebp-16]
+ ret
+align 16
+__vpaes_encrypt_core:
+ mov ecx,16
+ mov eax,DWORD [240+edx]
+ movdqa xmm1,xmm6
+ movdqa xmm2,[ebp]
+ pandn xmm1,xmm0
+ pand xmm0,xmm6
+ movdqu xmm5,[edx]
+db 102,15,56,0,208
+ movdqa xmm0,[16+ebp]
+ pxor xmm2,xmm5
+ psrld xmm1,4
+ add edx,16
+db 102,15,56,0,193
+ lea ebx,[192+ebp]
+ pxor xmm0,xmm2
+ jmp NEAR L$000enc_entry
+align 16
+L$001enc_loop:
+ movdqa xmm4,[32+ebp]
+ movdqa xmm0,[48+ebp]
+db 102,15,56,0,226
+db 102,15,56,0,195
+ pxor xmm4,xmm5
+ movdqa xmm5,[64+ebp]
+ pxor xmm0,xmm4
+ movdqa xmm1,[ecx*1+ebx-64]
+db 102,15,56,0,234
+ movdqa xmm2,[80+ebp]
+ movdqa xmm4,[ecx*1+ebx]
+db 102,15,56,0,211
+ movdqa xmm3,xmm0
+ pxor xmm2,xmm5
+db 102,15,56,0,193
+ add edx,16
+ pxor xmm0,xmm2
+db 102,15,56,0,220
+ add ecx,16
+ pxor xmm3,xmm0
+db 102,15,56,0,193
+ and ecx,48
+ sub eax,1
+ pxor xmm0,xmm3
+L$000enc_entry:
+ movdqa xmm1,xmm6
+ movdqa xmm5,[ebp-32]
+ pandn xmm1,xmm0
+ psrld xmm1,4
+ pand xmm0,xmm6
+db 102,15,56,0,232
+ movdqa xmm3,xmm7
+ pxor xmm0,xmm1
+db 102,15,56,0,217
+ movdqa xmm4,xmm7
+ pxor xmm3,xmm5
+db 102,15,56,0,224
+ movdqa xmm2,xmm7
+ pxor xmm4,xmm5
+db 102,15,56,0,211
+ movdqa xmm3,xmm7
+ pxor xmm2,xmm0
+db 102,15,56,0,220
+ movdqu xmm5,[edx]
+ pxor xmm3,xmm1
+ jnz NEAR L$001enc_loop
+ movdqa xmm4,[96+ebp]
+ movdqa xmm0,[112+ebp]
+db 102,15,56,0,226
+ pxor xmm4,xmm5
+db 102,15,56,0,195
+ movdqa xmm1,[64+ecx*1+ebx]
+ pxor xmm0,xmm4
+db 102,15,56,0,193
+ ret
+align 16
+__vpaes_decrypt_core:
+ lea ebx,[608+ebp]
+ mov eax,DWORD [240+edx]
+ movdqa xmm1,xmm6
+ movdqa xmm2,[ebx-64]
+ pandn xmm1,xmm0
+ mov ecx,eax
+ psrld xmm1,4
+ movdqu xmm5,[edx]
+ shl ecx,4
+ pand xmm0,xmm6
+db 102,15,56,0,208
+ movdqa xmm0,[ebx-48]
+ xor ecx,48
+db 102,15,56,0,193
+ and ecx,48
+ pxor xmm2,xmm5
+ movdqa xmm5,[176+ebp]
+ pxor xmm0,xmm2
+ add edx,16
+ lea ecx,[ecx*1+ebx-352]
+ jmp NEAR L$002dec_entry
+align 16
+L$003dec_loop:
+ movdqa xmm4,[ebx-32]
+ movdqa xmm1,[ebx-16]
+db 102,15,56,0,226
+db 102,15,56,0,203
+ pxor xmm0,xmm4
+ movdqa xmm4,[ebx]
+ pxor xmm0,xmm1
+ movdqa xmm1,[16+ebx]
+db 102,15,56,0,226
+db 102,15,56,0,197
+db 102,15,56,0,203
+ pxor xmm0,xmm4
+ movdqa xmm4,[32+ebx]
+ pxor xmm0,xmm1
+ movdqa xmm1,[48+ebx]
+db 102,15,56,0,226
+db 102,15,56,0,197
+db 102,15,56,0,203
+ pxor xmm0,xmm4
+ movdqa xmm4,[64+ebx]
+ pxor xmm0,xmm1
+ movdqa xmm1,[80+ebx]
+db 102,15,56,0,226
+db 102,15,56,0,197
+db 102,15,56,0,203
+ pxor xmm0,xmm4
+ add edx,16
+db 102,15,58,15,237,12
+ pxor xmm0,xmm1
+ sub eax,1
+L$002dec_entry:
+ movdqa xmm1,xmm6
+ movdqa xmm2,[ebp-32]
+ pandn xmm1,xmm0
+ pand xmm0,xmm6
+ psrld xmm1,4
+db 102,15,56,0,208
+ movdqa xmm3,xmm7
+ pxor xmm0,xmm1
+db 102,15,56,0,217
+ movdqa xmm4,xmm7
+ pxor xmm3,xmm2
+db 102,15,56,0,224
+ pxor xmm4,xmm2
+ movdqa xmm2,xmm7
+db 102,15,56,0,211
+ movdqa xmm3,xmm7
+ pxor xmm2,xmm0
+db 102,15,56,0,220
+ movdqu xmm0,[edx]
+ pxor xmm3,xmm1
+ jnz NEAR L$003dec_loop
+ movdqa xmm4,[96+ebx]
+db 102,15,56,0,226
+ pxor xmm4,xmm0
+ movdqa xmm0,[112+ebx]
+ movdqa xmm2,[ecx]
+db 102,15,56,0,195
+ pxor xmm0,xmm4
+db 102,15,56,0,194
+ ret
+align 16
+__vpaes_schedule_core:
+ add ebp,DWORD [esp]
+ movdqu xmm0,[esi]
+ movdqa xmm2,[320+ebp]
+ movdqa xmm3,xmm0
+ lea ebx,[ebp]
+ movdqa [4+esp],xmm2
+ call __vpaes_schedule_transform
+ movdqa xmm7,xmm0
+ test edi,edi
+ jnz NEAR L$004schedule_am_decrypting
+ movdqu [edx],xmm0
+ jmp NEAR L$005schedule_go
+L$004schedule_am_decrypting:
+ movdqa xmm1,[256+ecx*1+ebp]
+db 102,15,56,0,217
+ movdqu [edx],xmm3
+ xor ecx,48
+L$005schedule_go:
+ cmp eax,192
+ ja NEAR L$006schedule_256
+ je NEAR L$007schedule_192
+L$008schedule_128:
+ mov eax,10
+L$009loop_schedule_128:
+ call __vpaes_schedule_round
+ dec eax
+ jz NEAR L$010schedule_mangle_last
+ call __vpaes_schedule_mangle
+ jmp NEAR L$009loop_schedule_128
+align 16
+L$007schedule_192:
+ movdqu xmm0,[8+esi]
+ call __vpaes_schedule_transform
+ movdqa xmm6,xmm0
+ pxor xmm4,xmm4
+ movhlps xmm6,xmm4
+ mov eax,4
+L$011loop_schedule_192:
+ call __vpaes_schedule_round
+db 102,15,58,15,198,8
+ call __vpaes_schedule_mangle
+ call __vpaes_schedule_192_smear
+ call __vpaes_schedule_mangle
+ call __vpaes_schedule_round
+ dec eax
+ jz NEAR L$010schedule_mangle_last
+ call __vpaes_schedule_mangle
+ call __vpaes_schedule_192_smear
+ jmp NEAR L$011loop_schedule_192
+align 16
+L$006schedule_256:
+ movdqu xmm0,[16+esi]
+ call __vpaes_schedule_transform
+ mov eax,7
+L$012loop_schedule_256:
+ call __vpaes_schedule_mangle
+ movdqa xmm6,xmm0
+ call __vpaes_schedule_round
+ dec eax
+ jz NEAR L$010schedule_mangle_last
+ call __vpaes_schedule_mangle
+ pshufd xmm0,xmm0,255
+ movdqa [20+esp],xmm7
+ movdqa xmm7,xmm6
+ call L$_vpaes_schedule_low_round
+ movdqa xmm7,[20+esp]
+ jmp NEAR L$012loop_schedule_256
+align 16
+L$010schedule_mangle_last:
+ lea ebx,[384+ebp]
+ test edi,edi
+ jnz NEAR L$013schedule_mangle_last_dec
+ movdqa xmm1,[256+ecx*1+ebp]
+db 102,15,56,0,193
+ lea ebx,[352+ebp]
+ add edx,32
+L$013schedule_mangle_last_dec:
+ add edx,-16
+ pxor xmm0,[336+ebp]
+ call __vpaes_schedule_transform
+ movdqu [edx],xmm0
+ pxor xmm0,xmm0
+ pxor xmm1,xmm1
+ pxor xmm2,xmm2
+ pxor xmm3,xmm3
+ pxor xmm4,xmm4
+ pxor xmm5,xmm5
+ pxor xmm6,xmm6
+ pxor xmm7,xmm7
+ ret
+align 16
+__vpaes_schedule_192_smear:
+ pshufd xmm1,xmm6,128
+ pshufd xmm0,xmm7,254
+ pxor xmm6,xmm1
+ pxor xmm1,xmm1
+ pxor xmm6,xmm0
+ movdqa xmm0,xmm6
+ movhlps xmm6,xmm1
+ ret
+align 16
+__vpaes_schedule_round:
+ movdqa xmm2,[8+esp]
+ pxor xmm1,xmm1
+db 102,15,58,15,202,15
+db 102,15,58,15,210,15
+ pxor xmm7,xmm1
+ pshufd xmm0,xmm0,255
+db 102,15,58,15,192,1
+ movdqa [8+esp],xmm2
+L$_vpaes_schedule_low_round:
+ movdqa xmm1,xmm7
+ pslldq xmm7,4
+ pxor xmm7,xmm1
+ movdqa xmm1,xmm7
+ pslldq xmm7,8
+ pxor xmm7,xmm1
+ pxor xmm7,[336+ebp]
+ movdqa xmm4,[ebp-16]
+ movdqa xmm5,[ebp-48]
+ movdqa xmm1,xmm4
+ pandn xmm1,xmm0
+ psrld xmm1,4
+ pand xmm0,xmm4
+ movdqa xmm2,[ebp-32]
+db 102,15,56,0,208
+ pxor xmm0,xmm1
+ movdqa xmm3,xmm5
+db 102,15,56,0,217
+ pxor xmm3,xmm2
+ movdqa xmm4,xmm5
+db 102,15,56,0,224
+ pxor xmm4,xmm2
+ movdqa xmm2,xmm5
+db 102,15,56,0,211
+ pxor xmm2,xmm0
+ movdqa xmm3,xmm5
+db 102,15,56,0,220
+ pxor xmm3,xmm1
+ movdqa xmm4,[32+ebp]
+db 102,15,56,0,226
+ movdqa xmm0,[48+ebp]
+db 102,15,56,0,195
+ pxor xmm0,xmm4
+ pxor xmm0,xmm7
+ movdqa xmm7,xmm0
+ ret
+align 16
+__vpaes_schedule_transform:
+ movdqa xmm2,[ebp-16]
+ movdqa xmm1,xmm2
+ pandn xmm1,xmm0
+ psrld xmm1,4
+ pand xmm0,xmm2
+ movdqa xmm2,[ebx]
+db 102,15,56,0,208
+ movdqa xmm0,[16+ebx]
+db 102,15,56,0,193
+ pxor xmm0,xmm2
+ ret
+align 16
+__vpaes_schedule_mangle:
+ movdqa xmm4,xmm0
+ movdqa xmm5,[128+ebp]
+ test edi,edi
+ jnz NEAR L$014schedule_mangle_dec
+ add edx,16
+ pxor xmm4,[336+ebp]
+db 102,15,56,0,229
+ movdqa xmm3,xmm4
+db 102,15,56,0,229
+ pxor xmm3,xmm4
+db 102,15,56,0,229
+ pxor xmm3,xmm4
+ jmp NEAR L$015schedule_mangle_both
+align 16
+L$014schedule_mangle_dec:
+ movdqa xmm2,[ebp-16]
+ lea esi,[416+ebp]
+ movdqa xmm1,xmm2
+ pandn xmm1,xmm4
+ psrld xmm1,4
+ pand xmm4,xmm2
+ movdqa xmm2,[esi]
+db 102,15,56,0,212
+ movdqa xmm3,[16+esi]
+db 102,15,56,0,217
+ pxor xmm3,xmm2
+db 102,15,56,0,221
+ movdqa xmm2,[32+esi]
+db 102,15,56,0,212
+ pxor xmm2,xmm3
+ movdqa xmm3,[48+esi]
+db 102,15,56,0,217
+ pxor xmm3,xmm2
+db 102,15,56,0,221
+ movdqa xmm2,[64+esi]
+db 102,15,56,0,212
+ pxor xmm2,xmm3
+ movdqa xmm3,[80+esi]
+db 102,15,56,0,217
+ pxor xmm3,xmm2
+db 102,15,56,0,221
+ movdqa xmm2,[96+esi]
+db 102,15,56,0,212
+ pxor xmm2,xmm3
+ movdqa xmm3,[112+esi]
+db 102,15,56,0,217
+ pxor xmm3,xmm2
+ add edx,-16
+L$015schedule_mangle_both:
+ movdqa xmm1,[256+ecx*1+ebp]
+db 102,15,56,0,217
+ add ecx,-16
+ and ecx,48
+ movdqu [edx],xmm3
+ ret
+global _vpaes_set_encrypt_key
+align 16
+_vpaes_set_encrypt_key:
+L$_vpaes_set_encrypt_key_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ lea ebx,[esp-56]
+ mov eax,DWORD [24+esp]
+ and ebx,-16
+ mov edx,DWORD [28+esp]
+ xchg ebx,esp
+ mov DWORD [48+esp],ebx
+ mov ebx,eax
+ shr ebx,5
+ add ebx,5
+ mov DWORD [240+edx],ebx
+ mov ecx,48
+ mov edi,0
+ lea ebp,[(L$_vpaes_consts+0x30-L$016pic_point)]
+ call __vpaes_schedule_core
+L$016pic_point:
+ mov esp,DWORD [48+esp]
+ xor eax,eax
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _vpaes_set_decrypt_key
+align 16
+_vpaes_set_decrypt_key:
+L$_vpaes_set_decrypt_key_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ lea ebx,[esp-56]
+ mov eax,DWORD [24+esp]
+ and ebx,-16
+ mov edx,DWORD [28+esp]
+ xchg ebx,esp
+ mov DWORD [48+esp],ebx
+ mov ebx,eax
+ shr ebx,5
+ add ebx,5
+ mov DWORD [240+edx],ebx
+ shl ebx,4
+ lea edx,[16+ebx*1+edx]
+ mov edi,1
+ mov ecx,eax
+ shr ecx,1
+ and ecx,32
+ xor ecx,32
+ lea ebp,[(L$_vpaes_consts+0x30-L$017pic_point)]
+ call __vpaes_schedule_core
+L$017pic_point:
+ mov esp,DWORD [48+esp]
+ xor eax,eax
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _vpaes_encrypt
+align 16
+_vpaes_encrypt:
+L$_vpaes_encrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ lea ebp,[(L$_vpaes_consts+0x30-L$018pic_point)]
+ call __vpaes_preheat
+L$018pic_point:
+ mov esi,DWORD [20+esp]
+ lea ebx,[esp-56]
+ mov edi,DWORD [24+esp]
+ and ebx,-16
+ mov edx,DWORD [28+esp]
+ xchg ebx,esp
+ mov DWORD [48+esp],ebx
+ movdqu xmm0,[esi]
+ call __vpaes_encrypt_core
+ movdqu [edi],xmm0
+ mov esp,DWORD [48+esp]
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _vpaes_decrypt
+align 16
+_vpaes_decrypt:
+L$_vpaes_decrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ lea ebp,[(L$_vpaes_consts+0x30-L$019pic_point)]
+ call __vpaes_preheat
+L$019pic_point:
+ mov esi,DWORD [20+esp]
+ lea ebx,[esp-56]
+ mov edi,DWORD [24+esp]
+ and ebx,-16
+ mov edx,DWORD [28+esp]
+ xchg ebx,esp
+ mov DWORD [48+esp],ebx
+ movdqu xmm0,[esi]
+ call __vpaes_decrypt_core
+ movdqu [edi],xmm0
+ mov esp,DWORD [48+esp]
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret
+global _vpaes_cbc_encrypt
+align 16
+_vpaes_cbc_encrypt:
+L$_vpaes_cbc_encrypt_begin:
+ push ebp
+ push ebx
+ push esi
+ push edi
+ mov esi,DWORD [20+esp]
+ mov edi,DWORD [24+esp]
+ mov eax,DWORD [28+esp]
+ mov edx,DWORD [32+esp]
+ sub eax,16
+ jc NEAR L$020cbc_abort
+ lea ebx,[esp-56]
+ mov ebp,DWORD [36+esp]
+ and ebx,-16
+ mov ecx,DWORD [40+esp]
+ xchg ebx,esp
+ movdqu xmm1,[ebp]
+ sub edi,esi
+ mov DWORD [48+esp],ebx
+ mov DWORD [esp],edi
+ mov DWORD [4+esp],edx
+ mov DWORD [8+esp],ebp
+ mov edi,eax
+ lea ebp,[(L$_vpaes_consts+0x30-L$021pic_point)]
+ call __vpaes_preheat
+L$021pic_point:
+ cmp ecx,0
+ je NEAR L$022cbc_dec_loop
+ jmp NEAR L$023cbc_enc_loop
+align 16
+L$023cbc_enc_loop:
+ movdqu xmm0,[esi]
+ pxor xmm0,xmm1
+ call __vpaes_encrypt_core
+ mov ebx,DWORD [esp]
+ mov edx,DWORD [4+esp]
+ movdqa xmm1,xmm0
+ movdqu [esi*1+ebx],xmm0
+ lea esi,[16+esi]
+ sub edi,16
+ jnc NEAR L$023cbc_enc_loop
+ jmp NEAR L$024cbc_done
+align 16
+L$022cbc_dec_loop:
+ movdqu xmm0,[esi]
+ movdqa [16+esp],xmm1
+ movdqa [32+esp],xmm0
+ call __vpaes_decrypt_core
+ mov ebx,DWORD [esp]
+ mov edx,DWORD [4+esp]
+ pxor xmm0,[16+esp]
+ movdqa xmm1,[32+esp]
+ movdqu [esi*1+ebx],xmm0
+ lea esi,[16+esi]
+ sub edi,16
+ jnc NEAR L$022cbc_dec_loop
+L$024cbc_done:
+ mov ebx,DWORD [8+esp]
+ mov esp,DWORD [48+esp]
+ movdqu [ebx],xmm1
+L$020cbc_abort:
+ pop edi
+ pop esi
+ pop ebx
+ pop ebp
+ ret