#if defined(__i386__)
.file	"aes-586.S"
.text
.hidden	_x86_AES_encrypt_compact
.type	_x86_AES_encrypt_compact,@function
.align	16
_x86_AES_encrypt_compact:
	movl	%edi,20(%esp)
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	movl	240(%edi),%esi
	leal	-2(%esi,%esi,1),%esi
	leal	(%edi,%esi,8),%esi
	movl	%esi,24(%esp)
	movl	-128(%ebp),%edi
	movl	-96(%ebp),%esi
	movl	-64(%ebp),%edi
	movl	-32(%ebp),%esi
	movl	(%ebp),%edi
	movl	32(%ebp),%esi
	movl	64(%ebp),%edi
	movl	96(%ebp),%esi
.align	16
.L000loop:
	movl	%eax,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%bh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,4(%esp)

	movl	%ebx,%esi
	andl	$255,%esi
	shrl	$16,%ebx
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%ch,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,8(%esp)

	movl	%ecx,%esi
	andl	$255,%esi
	shrl	$24,%ecx
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%dh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edx
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movzbl	%bh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi

	andl	$255,%edx
	movzbl	-128(%ebp,%edx,1),%edx
	movzbl	%ah,%eax
	movzbl	-128(%ebp,%eax,1),%eax
	shll	$8,%eax
	xorl	%eax,%edx
	movl	4(%esp),%eax
	andl	$255,%ebx
	movzbl	-128(%ebp,%ebx,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%edx
	movl	8(%esp),%ebx
	movzbl	-128(%ebp,%ecx,1),%ecx
	shll	$24,%ecx
	xorl	%ecx,%edx
	movl	%esi,%ecx

	movl	$2155905152,%ebp
	andl	%ecx,%ebp
	leal	(%ecx,%ecx,1),%edi
	movl	%ebp,%esi
	shrl	$7,%ebp
	andl	$4278124286,%edi
	subl	%ebp,%esi
	movl	%ecx,%ebp
	andl	$454761243,%esi
	rorl	$16,%ebp
	xorl	%edi,%esi
	movl	%ecx,%edi
	xorl	%esi,%ecx
	rorl	$24,%edi
	xorl	%ebp,%esi
	roll	$24,%ecx
	xorl	%edi,%esi
	movl	$2155905152,%ebp
	xorl	%esi,%ecx
	andl	%edx,%ebp
	leal	(%edx,%edx,1),%edi
	movl	%ebp,%esi
	shrl	$7,%ebp
	andl	$4278124286,%edi
	subl	%ebp,%esi
	movl	%edx,%ebp
	andl	$454761243,%esi
	rorl	$16,%ebp
	xorl	%edi,%esi
	movl	%edx,%edi
	xorl	%esi,%edx
	rorl	$24,%edi
	xorl	%ebp,%esi
	roll	$24,%edx
	xorl	%edi,%esi
	movl	$2155905152,%ebp
	xorl	%esi,%edx
	andl	%eax,%ebp
	leal	(%eax,%eax,1),%edi
	movl	%ebp,%esi
	shrl	$7,%ebp
	andl	$4278124286,%edi
	subl	%ebp,%esi
	movl	%eax,%ebp
	andl	$454761243,%esi
	rorl	$16,%ebp
	xorl	%edi,%esi
	movl	%eax,%edi
	xorl	%esi,%eax
	rorl	$24,%edi
	xorl	%ebp,%esi
	roll	$24,%eax
	xorl	%edi,%esi
	movl	$2155905152,%ebp
	xorl	%esi,%eax
	andl	%ebx,%ebp
	leal	(%ebx,%ebx,1),%edi
	movl	%ebp,%esi
	shrl	$7,%ebp
	andl	$4278124286,%edi
	subl	%ebp,%esi
	movl	%ebx,%ebp
	andl	$454761243,%esi
	rorl	$16,%ebp
	xorl	%edi,%esi
	movl	%ebx,%edi
	xorl	%esi,%ebx
	rorl	$24,%edi
	xorl	%ebp,%esi
	roll	$24,%ebx
	xorl	%edi,%esi
	xorl	%esi,%ebx
	movl	20(%esp),%edi
	movl	28(%esp),%ebp
	addl	$16,%edi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	cmpl	24(%esp),%edi
	movl	%edi,20(%esp)
	jb	.L000loop
	movl	%eax,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%bh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,4(%esp)

	movl	%ebx,%esi
	andl	$255,%esi
	shrl	$16,%ebx
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%ch,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,8(%esp)

	movl	%ecx,%esi
	andl	$255,%esi
	shrl	$24,%ecx
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%dh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edx
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movzbl	%bh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi

	movl	20(%esp),%edi
	andl	$255,%edx
	movzbl	-128(%ebp,%edx,1),%edx
	movzbl	%ah,%eax
	movzbl	-128(%ebp,%eax,1),%eax
	shll	$8,%eax
	xorl	%eax,%edx
	movl	4(%esp),%eax
	andl	$255,%ebx
	movzbl	-128(%ebp,%ebx,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%edx
	movl	8(%esp),%ebx
	movzbl	-128(%ebp,%ecx,1),%ecx
	shll	$24,%ecx
	xorl	%ecx,%edx
	movl	%esi,%ecx

	xorl	16(%edi),%eax
	xorl	20(%edi),%ebx
	xorl	24(%edi),%ecx
	xorl	28(%edi),%edx
	ret
.size	_x86_AES_encrypt_compact,.-_x86_AES_encrypt_compact
.hidden	_sse_AES_encrypt_compact
.type	_sse_AES_encrypt_compact,@function
.align	16
_sse_AES_encrypt_compact:
	pxor	(%edi),%mm0
	pxor	8(%edi),%mm4
	movl	240(%edi),%esi
	leal	-2(%esi,%esi,1),%esi
	leal	(%edi,%esi,8),%esi
	movl	%esi,24(%esp)
	movl	$454761243,%eax
	movl	%eax,8(%esp)
	movl	%eax,12(%esp)
	movl	-128(%ebp),%eax
	movl	-96(%ebp),%ebx
	movl	-64(%ebp),%ecx
	movl	-32(%ebp),%edx
	movl	(%ebp),%eax
	movl	32(%ebp),%ebx
	movl	64(%ebp),%ecx
	movl	96(%ebp),%edx
.align	16
.L001loop:
	pshufw	$8,%mm0,%mm1
	pshufw	$13,%mm4,%mm5
	movd	%mm1,%eax
	movd	%mm5,%ebx
	movl	%edi,20(%esp)
	movzbl	%al,%esi
	movzbl	%ah,%edx
	pshufw	$13,%mm0,%mm2
	movzbl	-128(%ebp,%esi,1),%ecx
	movzbl	%bl,%edi
	movzbl	-128(%ebp,%edx,1),%edx
	shrl	$16,%eax
	shll	$8,%edx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bh,%edi
	shll	$16,%esi
	pshufw	$8,%mm4,%mm6
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%ah,%edi
	shll	$24,%esi
	shrl	$16,%ebx
	orl	%esi,%edx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bh,%edi
	shll	$8,%esi
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%al,%edi
	shll	$24,%esi
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bl,%edi
	movd	%mm2,%eax
	movd	%ecx,%mm0
	movzbl	-128(%ebp,%edi,1),%ecx
	movzbl	%ah,%edi
	shll	$16,%ecx
	movd	%mm6,%ebx
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bh,%edi
	shll	$24,%esi
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bl,%edi
	shll	$8,%esi
	shrl	$16,%ebx
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%al,%edi
	shrl	$16,%eax
	movd	%ecx,%mm1
	movzbl	-128(%ebp,%edi,1),%ecx
	movzbl	%ah,%edi
	shll	$16,%ecx
	andl	$255,%eax
	orl	%esi,%ecx
	punpckldq	%mm1,%mm0
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bh,%edi
	shll	$24,%esi
	andl	$255,%ebx
	movzbl	-128(%ebp,%eax,1),%eax
	orl	%esi,%ecx
	shll	$16,%eax
	movzbl	-128(%ebp,%edi,1),%esi
	orl	%eax,%edx
	shll	$8,%esi
	movzbl	-128(%ebp,%ebx,1),%ebx
	orl	%esi,%ecx
	orl	%ebx,%edx
	movl	20(%esp),%edi
	movd	%ecx,%mm4
	movd	%edx,%mm5
	punpckldq	%mm5,%mm4
	addl	$16,%edi
	cmpl	24(%esp),%edi
	ja	.L002out
	movq	8(%esp),%mm2
	pxor	%mm3,%mm3
	pxor	%mm7,%mm7
	movq	%mm0,%mm1
	movq	%mm4,%mm5
	pcmpgtb	%mm0,%mm3
	pcmpgtb	%mm4,%mm7
	pand	%mm2,%mm3
	pand	%mm2,%mm7
	pshufw	$177,%mm0,%mm2
	pshufw	$177,%mm4,%mm6
	paddb	%mm0,%mm0
	paddb	%mm4,%mm4
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	pshufw	$177,%mm2,%mm3
	pshufw	$177,%mm6,%mm7
	pxor	%mm0,%mm1
	pxor	%mm4,%mm5
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	movq	%mm3,%mm2
	movq	%mm7,%mm6
	pslld	$8,%mm3
	pslld	$8,%mm7
	psrld	$24,%mm2
	psrld	$24,%mm6
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	movq	%mm1,%mm3
	movq	%mm5,%mm7
	movq	(%edi),%mm2
	movq	8(%edi),%mm6
	psrld	$8,%mm1
	psrld	$8,%mm5
	movl	-128(%ebp),%eax
	pslld	$24,%mm3
	pslld	$24,%mm7
	movl	-64(%ebp),%ebx
	pxor	%mm1,%mm0
	pxor	%mm5,%mm4
	movl	(%ebp),%ecx
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	movl	64(%ebp),%edx
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	jmp	.L001loop
.align	16
.L002out:
	pxor	(%edi),%mm0
	pxor	8(%edi),%mm4
	ret
.size	_sse_AES_encrypt_compact,.-_sse_AES_encrypt_compact
.hidden	_x86_AES_encrypt
.type	_x86_AES_encrypt,@function
.align	16
_x86_AES_encrypt:
	movl	%edi,20(%esp)
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	movl	240(%edi),%esi
	leal	-2(%esi,%esi,1),%esi
	leal	(%edi,%esi,8),%esi
	movl	%esi,24(%esp)
.align	16
.L003loop:
	movl	%eax,%esi
	andl	$255,%esi
	movl	(%ebp,%esi,8),%esi
	movzbl	%bh,%edi
	xorl	3(%ebp,%edi,8),%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	xorl	2(%ebp,%edi,8),%esi
	movl	%edx,%edi
	shrl	$24,%edi
	xorl	1(%ebp,%edi,8),%esi
	movl	%esi,4(%esp)

	movl	%ebx,%esi
	andl	$255,%esi
	shrl	$16,%ebx
	movl	(%ebp,%esi,8),%esi
	movzbl	%ch,%edi
	xorl	3(%ebp,%edi,8),%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	xorl	2(%ebp,%edi,8),%esi
	movl	%eax,%edi
	shrl	$24,%edi
	xorl	1(%ebp,%edi,8),%esi
	movl	%esi,8(%esp)

	movl	%ecx,%esi
	andl	$255,%esi
	shrl	$24,%ecx
	movl	(%ebp,%esi,8),%esi
	movzbl	%dh,%edi
	xorl	3(%ebp,%edi,8),%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edx
	andl	$255,%edi
	xorl	2(%ebp,%edi,8),%esi
	movzbl	%bh,%edi
	xorl	1(%ebp,%edi,8),%esi

	movl	20(%esp),%edi
	movl	(%ebp,%edx,8),%edx
	movzbl	%ah,%eax
	xorl	3(%ebp,%eax,8),%edx
	movl	4(%esp),%eax
	andl	$255,%ebx
	xorl	2(%ebp,%ebx,8),%edx
	movl	8(%esp),%ebx
	xorl	1(%ebp,%ecx,8),%edx
	movl	%esi,%ecx

	addl	$16,%edi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	cmpl	24(%esp),%edi
	movl	%edi,20(%esp)
	jb	.L003loop
	movl	%eax,%esi
	andl	$255,%esi
	movl	2(%ebp,%esi,8),%esi
	andl	$255,%esi
	movzbl	%bh,%edi
	movl	(%ebp,%edi,8),%edi
	andl	$65280,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movl	(%ebp,%edi,8),%edi
	andl	$16711680,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$24,%edi
	movl	2(%ebp,%edi,8),%edi
	andl	$4278190080,%edi
	xorl	%edi,%esi
	movl	%esi,4(%esp)
	movl	%ebx,%esi
	andl	$255,%esi
	shrl	$16,%ebx
	movl	2(%ebp,%esi,8),%esi
	andl	$255,%esi
	movzbl	%ch,%edi
	movl	(%ebp,%edi,8),%edi
	andl	$65280,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movl	(%ebp,%edi,8),%edi
	andl	$16711680,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$24,%edi
	movl	2(%ebp,%edi,8),%edi
	andl	$4278190080,%edi
	xorl	%edi,%esi
	movl	%esi,8(%esp)
	movl	%ecx,%esi
	andl	$255,%esi
	shrl	$24,%ecx
	movl	2(%ebp,%esi,8),%esi
	andl	$255,%esi
	movzbl	%dh,%edi
	movl	(%ebp,%edi,8),%edi
	andl	$65280,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edx
	andl	$255,%edi
	movl	(%ebp,%edi,8),%edi
	andl	$16711680,%edi
	xorl	%edi,%esi
	movzbl	%bh,%edi
	movl	2(%ebp,%edi,8),%edi
	andl	$4278190080,%edi
	xorl	%edi,%esi
	movl	20(%esp),%edi
	andl	$255,%edx
	movl	2(%ebp,%edx,8),%edx
	andl	$255,%edx
	movzbl	%ah,%eax
	movl	(%ebp,%eax,8),%eax
	andl	$65280,%eax
	xorl	%eax,%edx
	movl	4(%esp),%eax
	andl	$255,%ebx
	movl	(%ebp,%ebx,8),%ebx
	andl	$16711680,%ebx
	xorl	%ebx,%edx
	movl	8(%esp),%ebx
	movl	2(%ebp,%ecx,8),%ecx
	andl	$4278190080,%ecx
	xorl	%ecx,%edx
	movl	%esi,%ecx
	addl	$16,%edi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	ret
.align	64
.LAES_Te:
.long	2774754246,2774754246
.long	2222750968,2222750968
.long	2574743534,2574743534
.long	2373680118,2373680118
.long	234025727,234025727
.long	3177933782,3177933782
.long	2976870366,2976870366
.long	1422247313,1422247313
.long	1345335392,1345335392
.long	50397442,50397442
.long	2842126286,2842126286
.long	2099981142,2099981142
.long	436141799,436141799
.long	1658312629,1658312629
.long	3870010189,3870010189
.long	2591454956,2591454956
.long	1170918031,1170918031
.long	2642575903,2642575903
.long	1086966153,1086966153
.long	2273148410,2273148410
.long	368769775,368769775
.long	3948501426,3948501426
.long	3376891790,3376891790
.long	200339707,200339707
.long	3970805057,3970805057
.long	1742001331,1742001331
.long	4255294047,4255294047
.long	3937382213,3937382213
.long	3214711843,3214711843
.long	4154762323,4154762323
.long	2524082916,2524082916
.long	1539358875,1539358875
.long	3266819957,3266819957
.long	486407649,486407649
.long	2928907069,2928907069
.long	1780885068,1780885068
.long	1513502316,1513502316
.long	1094664062,1094664062
.long	49805301,49805301
.long	1338821763,1338821763
.long	1546925160,1546925160
.long	4104496465,4104496465
.long	887481809,887481809
.long	150073849,150073849
.long	2473685474,2473685474
.long	1943591083,1943591083
.long	1395732834,1395732834
.long	1058346282,1058346282
.long	201589768,201589768
.long	1388824469,1388824469
.long	1696801606,1696801606
.long	1589887901,1589887901
.long	672667696,672667696
.long	2711000631,2711000631
.long	251987210,251987210
.long	3046808111,3046808111
.long	151455502,151455502
.long	907153956,907153956
.long	2608889883,2608889883
.long	1038279391,1038279391
.long	652995533,652995533
.long	1764173646,1764173646
.long	3451040383,3451040383
.long	2675275242,2675275242
.long	453576978,453576978
.long	2659418909,2659418909
.long	1949051992,1949051992
.long	773462580,773462580
.long	756751158,756751158
.long	2993581788,2993581788
.long	3998898868,3998898868
.long	4221608027,4221608027
.long	4132590244,4132590244
.long	1295727478,1295727478
.long	1641469623,1641469623
.long	3467883389,3467883389
.long	2066295122,2066295122
.long	1055122397,1055122397
.long	1898917726,1898917726
.long	2542044179,2542044179
.long	4115878822,4115878822
.long	1758581177,1758581177
.long	0,0
.long	753790401,753790401
.long	1612718144,1612718144
.long	536673507,536673507
.long	3367088505,3367088505
.long	3982187446,3982187446
.long	3194645204,3194645204
.long	1187761037,1187761037
.long	3653156455,3653156455
.long	1262041458,1262041458
.long	3729410708,3729410708
.long	3561770136,3561770136
.long	3898103984,3898103984
.long	1255133061,1255133061
.long	1808847035,1808847035
.long	720367557,720367557
.long	3853167183,3853167183
.long	385612781,385612781
.long	3309519750,3309519750
.long	3612167578,3612167578
.long	1429418854,1429418854
.long	2491778321,2491778321
.long	3477423498,3477423498
.long	284817897,284817897
.long	100794884,100794884
.long	2172616702,2172616702
.long	4031795360,4031795360
.long	1144798328,1144798328
.long	3131023141,3131023141
.long	3819481163,3819481163
.long	4082192802,4082192802
.long	4272137053,4272137053
.long	3225436288,3225436288
.long	2324664069,2324664069
.long	2912064063,2912064063
.long	3164445985,3164445985
.long	1211644016,1211644016
.long	83228145,83228145
.long	3753688163,3753688163
.long	3249976951,3249976951
.long	1977277103,1977277103
.long	1663115586,1663115586
.long	806359072,806359072
.long	452984805,452984805
.long	250868733,250868733
.long	1842533055,1842533055
.long	1288555905,1288555905
.long	336333848,336333848
.long	890442534,890442534
.long	804056259,804056259
.long	3781124030,3781124030
.long	2727843637,2727843637
.long	3427026056,3427026056
.long	957814574,957814574
.long	1472513171,1472513171
.long	4071073621,4071073621
.long	2189328124,2189328124
.long	1195195770,1195195770
.long	2892260552,2892260552
.long	3881655738,3881655738
.long	723065138,723065138
.long	2507371494,2507371494
.long	2690670784,2690670784
.long	2558624025,2558624025
.long	3511635870,3511635870
.long	2145180835,2145180835
.long	1713513028,1713513028
.long	2116692564,2116692564
.long	2878378043,2878378043
.long	2206763019,2206763019
.long	3393603212,3393603212
.long	703524551,703524551
.long	3552098411,3552098411
.long	1007948840,1007948840
.long	2044649127,2044649127
.long	3797835452,3797835452
.long	487262998,487262998
.long	1994120109,1994120109
.long	1004593371,1004593371
.long	1446130276,1446130276
.long	1312438900,1312438900
.long	503974420,503974420
.long	3679013266,3679013266
.long	168166924,168166924
.long	1814307912,1814307912
.long	3831258296,3831258296
.long	1573044895,1573044895
.long	1859376061,1859376061
.long	4021070915,4021070915
.long	2791465668,2791465668
.long	2828112185,2828112185
.long	2761266481,2761266481
.long	937747667,937747667
.long	2339994098,2339994098
.long	854058965,854058965
.long	1137232011,1137232011
.long	1496790894,1496790894
.long	3077402074,3077402074
.long	2358086913,2358086913
.long	1691735473,1691735473
.long	3528347292,3528347292
.long	3769215305,3769215305
.long	3027004632,3027004632
.long	4199962284,4199962284
.long	133494003,133494003
.long	636152527,636152527
.long	2942657994,2942657994
.long	2390391540,2390391540
.long	3920539207,3920539207
.long	403179536,403179536
.long	3585784431,3585784431
.long	2289596656,2289596656
.long	1864705354,1864705354
.long	1915629148,1915629148
.long	605822008,605822008
.long	4054230615,4054230615
.long	3350508659,3350508659
.long	1371981463,1371981463
.long	602466507,602466507
.long	2094914977,2094914977
.long	2624877800,2624877800
.long	555687742,555687742
.long	3712699286,3712699286
.long	3703422305,3703422305
.long	2257292045,2257292045
.long	2240449039,2240449039
.long	2423288032,2423288032
.long	1111375484,1111375484
.long	3300242801,3300242801
.long	2858837708,2858837708
.long	3628615824,3628615824
.long	84083462,84083462
.long	32962295,32962295
.long	302911004,302911004
.long	2741068226,2741068226
.long	1597322602,1597322602
.long	4183250862,4183250862
.long	3501832553,3501832553
.long	2441512471,2441512471
.long	1489093017,1489093017
.long	656219450,656219450
.long	3114180135,3114180135
.long	954327513,954327513
.long	335083755,335083755
.long	3013122091,3013122091
.long	856756514,856756514
.long	3144247762,3144247762
.long	1893325225,1893325225
.long	2307821063,2307821063
.long	2811532339,2811532339
.long	3063651117,3063651117
.long	572399164,572399164
.long	2458355477,2458355477
.long	552200649,552200649
.long	1238290055,1238290055
.long	4283782570,4283782570
.long	2015897680,2015897680
.long	2061492133,2061492133
.long	2408352771,2408352771
.long	4171342169,4171342169
.long	2156497161,2156497161
.long	386731290,386731290
.long	3669999461,3669999461
.long	837215959,837215959
.long	3326231172,3326231172
.long	3093850320,3093850320
.long	3275833730,3275833730
.long	2962856233,2962856233
.long	1999449434,1999449434
.long	286199582,286199582
.long	3417354363,3417354363
.long	4233385128,4233385128
.long	3602627437,3602627437
.long	974525996,974525996
.byte	99,124,119,123,242,107,111,197
.byte	48,1,103,43,254,215,171,118
.byte	202,130,201,125,250,89,71,240
.byte	173,212,162,175,156,164,114,192
.byte	183,253,147,38,54,63,247,204
.byte	52,165,229,241,113,216,49,21
.byte	4,199,35,195,24,150,5,154
.byte	7,18,128,226,235,39,178,117
.byte	9,131,44,26,27,110,90,160
.byte	82,59,214,179,41,227,47,132
.byte	83,209,0,237,32,252,177,91
.byte	106,203,190,57,74,76,88,207
.byte	208,239,170,251,67,77,51,133
.byte	69,249,2,127,80,60,159,168
.byte	81,163,64,143,146,157,56,245
.byte	188,182,218,33,16,255,243,210
.byte	205,12,19,236,95,151,68,23
.byte	196,167,126,61,100,93,25,115
.byte	96,129,79,220,34,42,144,136
.byte	70,238,184,20,222,94,11,219
.byte	224,50,58,10,73,6,36,92
.byte	194,211,172,98,145,149,228,121
.byte	231,200,55,109,141,213,78,169
.byte	108,86,244,234,101,122,174,8
.byte	186,120,37,46,28,166,180,198
.byte	232,221,116,31,75,189,139,138
.byte	112,62,181,102,72,3,246,14
.byte	97,53,87,185,134,193,29,158
.byte	225,248,152,17,105,217,142,148
.byte	155,30,135,233,206,85,40,223
.byte	140,161,137,13,191,230,66,104
.byte	65,153,45,15,176,84,187,22
.byte	99,124,119,123,242,107,111,197
.byte	48,1,103,43,254,215,171,118
.byte	202,130,201,125,250,89,71,240
.byte	173,212,162,175,156,164,114,192
.byte	183,253,147,38,54,63,247,204
.byte	52,165,229,241,113,216,49,21
.byte	4,199,35,195,24,150,5,154
.byte	7,18,128,226,235,39,178,117
.byte	9,131,44,26,27,110,90,160
.byte	82,59,214,179,41,227,47,132
.byte	83,209,0,237,32,252,177,91
.byte	106,203,190,57,74,76,88,207
.byte	208,239,170,251,67,77,51,133
.byte	69,249,2,127,80,60,159,168
.byte	81,163,64,143,146,157,56,245
.byte	188,182,218,33,16,255,243,210
.byte	205,12,19,236,95,151,68,23
.byte	196,167,126,61,100,93,25,115
.byte	96,129,79,220,34,42,144,136
.byte	70,238,184,20,222,94,11,219
.byte	224,50,58,10,73,6,36,92
.byte	194,211,172,98,145,149,228,121
.byte	231,200,55,109,141,213,78,169
.byte	108,86,244,234,101,122,174,8
.byte	186,120,37,46,28,166,180,198
.byte	232,221,116,31,75,189,139,138
.byte	112,62,181,102,72,3,246,14
.byte	97,53,87,185,134,193,29,158
.byte	225,248,152,17,105,217,142,148
.byte	155,30,135,233,206,85,40,223
.byte	140,161,137,13,191,230,66,104
.byte	65,153,45,15,176,84,187,22
.byte	99,124,119,123,242,107,111,197
.byte	48,1,103,43,254,215,171,118
.byte	202,130,201,125,250,89,71,240
.byte	173,212,162,175,156,164,114,192
.byte	183,253,147,38,54,63,247,204
.byte	52,165,229,241,113,216,49,21
.byte	4,199,35,195,24,150,5,154
.byte	7,18,128,226,235,39,178,117
.byte	9,131,44,26,27,110,90,160
.byte	82,59,214,179,41,227,47,132
.byte	83,209,0,237,32,252,177,91
.byte	106,203,190,57,74,76,88,207
.byte	208,239,170,251,67,77,51,133
.byte	69,249,2,127,80,60,159,168
.byte	81,163,64,143,146,157,56,245
.byte	188,182,218,33,16,255,243,210
.byte	205,12,19,236,95,151,68,23
.byte	196,167,126,61,100,93,25,115
.byte	96,129,79,220,34,42,144,136
.byte	70,238,184,20,222,94,11,219
.byte	224,50,58,10,73,6,36,92
.byte	194,211,172,98,145,149,228,121
.byte	231,200,55,109,141,213,78,169
.byte	108,86,244,234,101,122,174,8
.byte	186,120,37,46,28,166,180,198
.byte	232,221,116,31,75,189,139,138
.byte	112,62,181,102,72,3,246,14
.byte	97,53,87,185,134,193,29,158
.byte	225,248,152,17,105,217,142,148
.byte	155,30,135,233,206,85,40,223
.byte	140,161,137,13,191,230,66,104
.byte	65,153,45,15,176,84,187,22
.byte	99,124,119,123,242,107,111,197
.byte	48,1,103,43,254,215,171,118
.byte	202,130,201,125,250,89,71,240
.byte	173,212,162,175,156,164,114,192
.byte	183,253,147,38,54,63,247,204
.byte	52,165,229,241,113,216,49,21
.byte	4,199,35,195,24,150,5,154
.byte	7,18,128,226,235,39,178,117
.byte	9,131,44,26,27,110,90,160
.byte	82,59,214,179,41,227,47,132
.byte	83,209,0,237,32,252,177,91
.byte	106,203,190,57,74,76,88,207
.byte	208,239,170,251,67,77,51,133
.byte	69,249,2,127,80,60,159,168
.byte	81,163,64,143,146,157,56,245
.byte	188,182,218,33,16,255,243,210
.byte	205,12,19,236,95,151,68,23
.byte	196,167,126,61,100,93,25,115
.byte	96,129,79,220,34,42,144,136
.byte	70,238,184,20,222,94,11,219
.byte	224,50,58,10,73,6,36,92
.byte	194,211,172,98,145,149,228,121
.byte	231,200,55,109,141,213,78,169
.byte	108,86,244,234,101,122,174,8
.byte	186,120,37,46,28,166,180,198
.byte	232,221,116,31,75,189,139,138
.byte	112,62,181,102,72,3,246,14
.byte	97,53,87,185,134,193,29,158
.byte	225,248,152,17,105,217,142,148
.byte	155,30,135,233,206,85,40,223
.byte	140,161,137,13,191,230,66,104
.byte	65,153,45,15,176,84,187,22
.long	1,2,4,8
.long	16,32,64,128
.long	27,54,0,0
.long	0,0,0,0
.size	_x86_AES_encrypt,.-_x86_AES_encrypt
.globl	asm_AES_encrypt
.hidden	asm_AES_encrypt
.type	asm_AES_encrypt,@function
.align	16
asm_AES_encrypt:
.L_asm_AES_encrypt_begin:
	pushl	%ebp
	pushl	%ebx
	pushl	%esi
	pushl	%edi
	movl	20(%esp),%esi
	movl	28(%esp),%edi
	movl	%esp,%eax
	subl	$36,%esp
	andl	$-64,%esp
	leal	-127(%edi),%ebx
	subl	%esp,%ebx
	negl	%ebx
	andl	$960,%ebx
	subl	%ebx,%esp
	addl	$4,%esp
	movl	%eax,28(%esp)
	call	.L004pic_point
.L004pic_point:
	popl	%ebp
	leal	OPENSSL_ia32cap_P-.L004pic_point(%ebp),%eax
	leal	.LAES_Te-.L004pic_point(%ebp),%ebp
	leal	764(%esp),%ebx
	subl	%ebp,%ebx
	andl	$768,%ebx
	leal	2176(%ebp,%ebx,1),%ebp
	btl	$25,(%eax)
	jnc	.L005x86
	movq	(%esi),%mm0
	movq	8(%esi),%mm4
	call	_sse_AES_encrypt_compact
	movl	28(%esp),%esp
	movl	24(%esp),%esi
	movq	%mm0,(%esi)
	movq	%mm4,8(%esi)
	emms
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
.align	16
.L005x86:
	movl	%ebp,24(%esp)
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	call	_x86_AES_encrypt_compact
	movl	28(%esp),%esp
	movl	24(%esp),%esi
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
.size	asm_AES_encrypt,.-.L_asm_AES_encrypt_begin
.hidden	_x86_AES_decrypt_compact
.type	_x86_AES_decrypt_compact,@function
.align	16
_x86_AES_decrypt_compact:
	movl	%edi,20(%esp)
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	movl	240(%edi),%esi
	leal	-2(%esi,%esi,1),%esi
	leal	(%edi,%esi,8),%esi
	movl	%esi,24(%esp)
	movl	-128(%ebp),%edi
	movl	-96(%ebp),%esi
	movl	-64(%ebp),%edi
	movl	-32(%ebp),%esi
	movl	(%ebp),%edi
	movl	32(%ebp),%esi
	movl	64(%ebp),%edi
	movl	96(%ebp),%esi
.align	16
.L006loop:
	movl	%eax,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%dh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%ebx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,4(%esp)
	movl	%ebx,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%ah,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,8(%esp)
	movl	%ecx,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%bh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	andl	$255,%edx
	movzbl	-128(%ebp,%edx,1),%edx
	movzbl	%ch,%ecx
	movzbl	-128(%ebp,%ecx,1),%ecx
	shll	$8,%ecx
	xorl	%ecx,%edx
	movl	%esi,%ecx
	shrl	$16,%ebx
	andl	$255,%ebx
	movzbl	-128(%ebp,%ebx,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%edx
	shrl	$24,%eax
	movzbl	-128(%ebp,%eax,1),%eax
	shll	$24,%eax
	xorl	%eax,%edx
	movl	$2155905152,%edi
	andl	%ecx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%ecx,%ecx,1),%eax
	subl	%edi,%esi
	andl	$4278124286,%eax
	andl	$454761243,%esi
	xorl	%esi,%eax
	movl	$2155905152,%edi
	andl	%eax,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%eax,%eax,1),%ebx
	subl	%edi,%esi
	andl	$4278124286,%ebx
	andl	$454761243,%esi
	xorl	%ecx,%eax
	xorl	%esi,%ebx
	movl	$2155905152,%edi
	andl	%ebx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%ebx,%ebx,1),%ebp
	subl	%edi,%esi
	andl	$4278124286,%ebp
	andl	$454761243,%esi
	xorl	%ecx,%ebx
	roll	$8,%ecx
	xorl	%esi,%ebp
	xorl	%eax,%ecx
	xorl	%ebp,%eax
	xorl	%ebx,%ecx
	xorl	%ebp,%ebx
	roll	$24,%eax
	xorl	%ebp,%ecx
	roll	$16,%ebx
	xorl	%eax,%ecx
	roll	$8,%ebp
	xorl	%ebx,%ecx
	movl	4(%esp),%eax
	xorl	%ebp,%ecx
	movl	%ecx,12(%esp)
	movl	$2155905152,%edi
	andl	%edx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%edx,%edx,1),%ebx
	subl	%edi,%esi
	andl	$4278124286,%ebx
	andl	$454761243,%esi
	xorl	%esi,%ebx
	movl	$2155905152,%edi
	andl	%ebx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%ebx,%ebx,1),%ecx
	subl	%edi,%esi
	andl	$4278124286,%ecx
	andl	$454761243,%esi
	xorl	%edx,%ebx
	xorl	%esi,%ecx
	movl	$2155905152,%edi
	andl	%ecx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%ecx,%ecx,1),%ebp
	subl	%edi,%esi
	andl	$4278124286,%ebp
	andl	$454761243,%esi
	xorl	%edx,%ecx
	roll	$8,%edx
	xorl	%esi,%ebp
	xorl	%ebx,%edx
	xorl	%ebp,%ebx
	xorl	%ecx,%edx
	xorl	%ebp,%ecx
	roll	$24,%ebx
	xorl	%ebp,%edx
	roll	$16,%ecx
	xorl	%ebx,%edx
	roll	$8,%ebp
	xorl	%ecx,%edx
	movl	8(%esp),%ebx
	xorl	%ebp,%edx
	movl	%edx,16(%esp)
	movl	$2155905152,%edi
	andl	%eax,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%eax,%eax,1),%ecx
	subl	%edi,%esi
	andl	$4278124286,%ecx
	andl	$454761243,%esi
	xorl	%esi,%ecx
	movl	$2155905152,%edi
	andl	%ecx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%ecx,%ecx,1),%edx
	subl	%edi,%esi
	andl	$4278124286,%edx
	andl	$454761243,%esi
	xorl	%eax,%ecx
	xorl	%esi,%edx
	movl	$2155905152,%edi
	andl	%edx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%edx,%edx,1),%ebp
	subl	%edi,%esi
	andl	$4278124286,%ebp
	andl	$454761243,%esi
	xorl	%eax,%edx
	roll	$8,%eax
	xorl	%esi,%ebp
	xorl	%ecx,%eax
	xorl	%ebp,%ecx
	xorl	%edx,%eax
	xorl	%ebp,%edx
	roll	$24,%ecx
	xorl	%ebp,%eax
	roll	$16,%edx
	xorl	%ecx,%eax
	roll	$8,%ebp
	xorl	%edx,%eax
	xorl	%ebp,%eax
	movl	$2155905152,%edi
	andl	%ebx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%ebx,%ebx,1),%ecx
	subl	%edi,%esi
	andl	$4278124286,%ecx
	andl	$454761243,%esi
	xorl	%esi,%ecx
	movl	$2155905152,%edi
	andl	%ecx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%ecx,%ecx,1),%edx
	subl	%edi,%esi
	andl	$4278124286,%edx
	andl	$454761243,%esi
	xorl	%ebx,%ecx
	xorl	%esi,%edx
	movl	$2155905152,%edi
	andl	%edx,%edi
	movl	%edi,%esi
	shrl	$7,%edi
	leal	(%edx,%edx,1),%ebp
	subl	%edi,%esi
	andl	$4278124286,%ebp
	andl	$454761243,%esi
	xorl	%ebx,%edx
	roll	$8,%ebx
	xorl	%esi,%ebp
	xorl	%ecx,%ebx
	xorl	%ebp,%ecx
	xorl	%edx,%ebx
	xorl	%ebp,%edx
	roll	$24,%ecx
	xorl	%ebp,%ebx
	roll	$16,%edx
	xorl	%ecx,%ebx
	roll	$8,%ebp
	xorl	%edx,%ebx
	movl	12(%esp),%ecx
	xorl	%ebp,%ebx
	movl	16(%esp),%edx
	movl	20(%esp),%edi
	movl	28(%esp),%ebp
	addl	$16,%edi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	cmpl	24(%esp),%edi
	movl	%edi,20(%esp)
	jb	.L006loop
	movl	%eax,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%dh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%ebx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,4(%esp)
	movl	%ebx,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%ah,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,8(%esp)
	movl	%ecx,%esi
	andl	$255,%esi
	movzbl	-128(%ebp,%esi,1),%esi
	movzbl	%bh,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$24,%edi
	movzbl	-128(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	20(%esp),%edi
	andl	$255,%edx
	movzbl	-128(%ebp,%edx,1),%edx
	movzbl	%ch,%ecx
	movzbl	-128(%ebp,%ecx,1),%ecx
	shll	$8,%ecx
	xorl	%ecx,%edx
	movl	%esi,%ecx
	shrl	$16,%ebx
	andl	$255,%ebx
	movzbl	-128(%ebp,%ebx,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%edx
	movl	8(%esp),%ebx
	shrl	$24,%eax
	movzbl	-128(%ebp,%eax,1),%eax
	shll	$24,%eax
	xorl	%eax,%edx
	movl	4(%esp),%eax
	xorl	16(%edi),%eax
	xorl	20(%edi),%ebx
	xorl	24(%edi),%ecx
	xorl	28(%edi),%edx
	ret
.size	_x86_AES_decrypt_compact,.-_x86_AES_decrypt_compact
.hidden	_sse_AES_decrypt_compact
.type	_sse_AES_decrypt_compact,@function
.align	16
_sse_AES_decrypt_compact:
	pxor	(%edi),%mm0
	pxor	8(%edi),%mm4
	movl	240(%edi),%esi
	leal	-2(%esi,%esi,1),%esi
	leal	(%edi,%esi,8),%esi
	movl	%esi,24(%esp)
	movl	$454761243,%eax
	movl	%eax,8(%esp)
	movl	%eax,12(%esp)
	movl	-128(%ebp),%eax
	movl	-96(%ebp),%ebx
	movl	-64(%ebp),%ecx
	movl	-32(%ebp),%edx
	movl	(%ebp),%eax
	movl	32(%ebp),%ebx
	movl	64(%ebp),%ecx
	movl	96(%ebp),%edx
.align	16
.L007loop:
	pshufw	$12,%mm0,%mm1
	pshufw	$9,%mm4,%mm5
	movd	%mm1,%eax
	movd	%mm5,%ebx
	movl	%edi,20(%esp)
	movzbl	%al,%esi
	movzbl	%ah,%edx
	pshufw	$6,%mm0,%mm2
	movzbl	-128(%ebp,%esi,1),%ecx
	movzbl	%bl,%edi
	movzbl	-128(%ebp,%edx,1),%edx
	shrl	$16,%eax
	shll	$8,%edx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bh,%edi
	shll	$16,%esi
	pshufw	$3,%mm4,%mm6
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%ah,%edi
	shll	$24,%esi
	shrl	$16,%ebx
	orl	%esi,%edx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bh,%edi
	shll	$24,%esi
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%al,%edi
	shll	$8,%esi
	movd	%mm2,%eax
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bl,%edi
	shll	$16,%esi
	movd	%mm6,%ebx
	movd	%ecx,%mm0
	movzbl	-128(%ebp,%edi,1),%ecx
	movzbl	%al,%edi
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bl,%edi
	orl	%esi,%edx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%ah,%edi
	shll	$16,%esi
	shrl	$16,%eax
	orl	%esi,%edx
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%bh,%edi
	shrl	$16,%ebx
	shll	$8,%esi
	movd	%edx,%mm1
	movzbl	-128(%ebp,%edi,1),%edx
	movzbl	%bh,%edi
	shll	$24,%edx
	andl	$255,%ebx
	orl	%esi,%edx
	punpckldq	%mm1,%mm0
	movzbl	-128(%ebp,%edi,1),%esi
	movzbl	%al,%edi
	shll	$8,%esi
	movzbl	%ah,%eax
	movzbl	-128(%ebp,%ebx,1),%ebx
	orl	%esi,%ecx
	movzbl	-128(%ebp,%edi,1),%esi
	orl	%ebx,%edx
	shll	$16,%esi
	movzbl	-128(%ebp,%eax,1),%eax
	orl	%esi,%edx
	shll	$24,%eax
	orl	%eax,%ecx
	movl	20(%esp),%edi
	movd	%edx,%mm4
	movd	%ecx,%mm5
	punpckldq	%mm5,%mm4
	addl	$16,%edi
	cmpl	24(%esp),%edi
	ja	.L008out
	movq	%mm0,%mm3
	movq	%mm4,%mm7
	pshufw	$228,%mm0,%mm2
	pshufw	$228,%mm4,%mm6
	movq	%mm0,%mm1
	movq	%mm4,%mm5
	pshufw	$177,%mm0,%mm0
	pshufw	$177,%mm4,%mm4
	pslld	$8,%mm2
	pslld	$8,%mm6
	psrld	$8,%mm3
	psrld	$8,%mm7
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	pslld	$16,%mm2
	pslld	$16,%mm6
	psrld	$16,%mm3
	psrld	$16,%mm7
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	movq	8(%esp),%mm3
	pxor	%mm2,%mm2
	pxor	%mm6,%mm6
	pcmpgtb	%mm1,%mm2
	pcmpgtb	%mm5,%mm6
	pand	%mm3,%mm2
	pand	%mm3,%mm6
	paddb	%mm1,%mm1
	paddb	%mm5,%mm5
	pxor	%mm2,%mm1
	pxor	%mm6,%mm5
	movq	%mm1,%mm3
	movq	%mm5,%mm7
	movq	%mm1,%mm2
	movq	%mm5,%mm6
	pxor	%mm1,%mm0
	pxor	%mm5,%mm4
	pslld	$24,%mm3
	pslld	$24,%mm7
	psrld	$8,%mm2
	psrld	$8,%mm6
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	movq	8(%esp),%mm2
	pxor	%mm3,%mm3
	pxor	%mm7,%mm7
	pcmpgtb	%mm1,%mm3
	pcmpgtb	%mm5,%mm7
	pand	%mm2,%mm3
	pand	%mm2,%mm7
	paddb	%mm1,%mm1
	paddb	%mm5,%mm5
	pxor	%mm3,%mm1
	pxor	%mm7,%mm5
	pshufw	$177,%mm1,%mm3
	pshufw	$177,%mm5,%mm7
	pxor	%mm1,%mm0
	pxor	%mm5,%mm4
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	pxor	%mm3,%mm3
	pxor	%mm7,%mm7
	pcmpgtb	%mm1,%mm3
	pcmpgtb	%mm5,%mm7
	pand	%mm2,%mm3
	pand	%mm2,%mm7
	paddb	%mm1,%mm1
	paddb	%mm5,%mm5
	pxor	%mm3,%mm1
	pxor	%mm7,%mm5
	pxor	%mm1,%mm0
	pxor	%mm5,%mm4
	movq	%mm1,%mm3
	movq	%mm5,%mm7
	pshufw	$177,%mm1,%mm2
	pshufw	$177,%mm5,%mm6
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	pslld	$8,%mm1
	pslld	$8,%mm5
	psrld	$8,%mm3
	psrld	$8,%mm7
	movq	(%edi),%mm2
	movq	8(%edi),%mm6
	pxor	%mm1,%mm0
	pxor	%mm5,%mm4
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	movl	-128(%ebp),%eax
	pslld	$16,%mm1
	pslld	$16,%mm5
	movl	-64(%ebp),%ebx
	psrld	$16,%mm3
	psrld	$16,%mm7
	movl	(%ebp),%ecx
	pxor	%mm1,%mm0
	pxor	%mm5,%mm4
	movl	64(%ebp),%edx
	pxor	%mm3,%mm0
	pxor	%mm7,%mm4
	pxor	%mm2,%mm0
	pxor	%mm6,%mm4
	jmp	.L007loop
.align	16
.L008out:
	pxor	(%edi),%mm0
	pxor	8(%edi),%mm4
	ret
.size	_sse_AES_decrypt_compact,.-_sse_AES_decrypt_compact
.hidden	_x86_AES_decrypt
.type	_x86_AES_decrypt,@function
.align	16
_x86_AES_decrypt:
	movl	%edi,20(%esp)
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	movl	240(%edi),%esi
	leal	-2(%esi,%esi,1),%esi
	leal	(%edi,%esi,8),%esi
	movl	%esi,24(%esp)
.align	16
.L009loop:
	movl	%eax,%esi
	andl	$255,%esi
	movl	(%ebp,%esi,8),%esi
	movzbl	%dh,%edi
	xorl	3(%ebp,%edi,8),%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	xorl	2(%ebp,%edi,8),%esi
	movl	%ebx,%edi
	shrl	$24,%edi
	xorl	1(%ebp,%edi,8),%esi
	movl	%esi,4(%esp)

	movl	%ebx,%esi
	andl	$255,%esi
	movl	(%ebp,%esi,8),%esi
	movzbl	%ah,%edi
	xorl	3(%ebp,%edi,8),%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	xorl	2(%ebp,%edi,8),%esi
	movl	%ecx,%edi
	shrl	$24,%edi
	xorl	1(%ebp,%edi,8),%esi
	movl	%esi,8(%esp)

	movl	%ecx,%esi
	andl	$255,%esi
	movl	(%ebp,%esi,8),%esi
	movzbl	%bh,%edi
	xorl	3(%ebp,%edi,8),%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edi
	xorl	2(%ebp,%edi,8),%esi
	movl	%edx,%edi
	shrl	$24,%edi
	xorl	1(%ebp,%edi,8),%esi

	movl	20(%esp),%edi
	andl	$255,%edx
	movl	(%ebp,%edx,8),%edx
	movzbl	%ch,%ecx
	xorl	3(%ebp,%ecx,8),%edx
	movl	%esi,%ecx
	shrl	$16,%ebx
	andl	$255,%ebx
	xorl	2(%ebp,%ebx,8),%edx
	movl	8(%esp),%ebx
	shrl	$24,%eax
	xorl	1(%ebp,%eax,8),%edx
	movl	4(%esp),%eax

	addl	$16,%edi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	cmpl	24(%esp),%edi
	movl	%edi,20(%esp)
	jb	.L009loop
	leal	2176(%ebp),%ebp
	movl	-128(%ebp),%edi
	movl	-96(%ebp),%esi
	movl	-64(%ebp),%edi
	movl	-32(%ebp),%esi
	movl	(%ebp),%edi
	movl	32(%ebp),%esi
	movl	64(%ebp),%edi
	movl	96(%ebp),%esi
	leal	-128(%ebp),%ebp
	movl	%eax,%esi
	andl	$255,%esi
	movzbl	(%ebp,%esi,1),%esi
	movzbl	%dh,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%ebx,%edi
	shrl	$24,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,4(%esp)
	movl	%ebx,%esi
	andl	$255,%esi
	movzbl	(%ebp,%esi,1),%esi
	movzbl	%ah,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%ecx,%edi
	shrl	$24,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	%esi,8(%esp)
	movl	%ecx,%esi
	andl	$255,%esi
	movzbl	(%ebp,%esi,1),%esi
	movzbl	%bh,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$8,%edi
	xorl	%edi,%esi
	movl	%eax,%edi
	shrl	$16,%edi
	andl	$255,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$16,%edi
	xorl	%edi,%esi
	movl	%edx,%edi
	shrl	$24,%edi
	movzbl	(%ebp,%edi,1),%edi
	shll	$24,%edi
	xorl	%edi,%esi
	movl	20(%esp),%edi
	andl	$255,%edx
	movzbl	(%ebp,%edx,1),%edx
	movzbl	%ch,%ecx
	movzbl	(%ebp,%ecx,1),%ecx
	shll	$8,%ecx
	xorl	%ecx,%edx
	movl	%esi,%ecx
	shrl	$16,%ebx
	andl	$255,%ebx
	movzbl	(%ebp,%ebx,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%edx
	movl	8(%esp),%ebx
	shrl	$24,%eax
	movzbl	(%ebp,%eax,1),%eax
	shll	$24,%eax
	xorl	%eax,%edx
	movl	4(%esp),%eax
	leal	-2048(%ebp),%ebp
	addl	$16,%edi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	ret
.align	64
.LAES_Td:
.long	1353184337,1353184337
.long	1399144830,1399144830
.long	3282310938,3282310938
.long	2522752826,2522752826
.long	3412831035,3412831035
.long	4047871263,4047871263
.long	2874735276,2874735276
.long	2466505547,2466505547
.long	1442459680,1442459680
.long	4134368941,4134368941
.long	2440481928,2440481928
.long	625738485,625738485
.long	4242007375,4242007375
.long	3620416197,3620416197
.long	2151953702,2151953702
.long	2409849525,2409849525
.long	1230680542,1230680542
.long	1729870373,1729870373
.long	2551114309,2551114309
.long	3787521629,3787521629
.long	41234371,41234371
.long	317738113,317738113
.long	2744600205,2744600205
.long	3338261355,3338261355
.long	3881799427,3881799427
.long	2510066197,2510066197
.long	3950669247,3950669247
.long	3663286933,3663286933
.long	763608788,763608788
.long	3542185048,3542185048
.long	694804553,694804553
.long	1154009486,1154009486
.long	1787413109,1787413109
.long	2021232372,2021232372
.long	1799248025,1799248025
.long	3715217703,3715217703
.long	3058688446,3058688446
.long	397248752,397248752
.long	1722556617,1722556617
.long	3023752829,3023752829
.long	407560035,407560035
.long	2184256229,2184256229
.long	1613975959,1613975959
.long	1165972322,1165972322
.long	3765920945,3765920945
.long	2226023355,2226023355
.long	480281086,480281086
.long	2485848313,2485848313
.long	1483229296,1483229296
.long	436028815,436028815
.long	2272059028,2272059028
.long	3086515026,3086515026
.long	601060267,601060267
.long	3791801202,3791801202
.long	1468997603,1468997603
.long	715871590,715871590
.long	120122290,120122290
.long	63092015,63092015
.long	2591802758,2591802758
.long	2768779219,2768779219
.long	4068943920,4068943920
.long	2997206819,2997206819
.long	3127509762,3127509762
.long	1552029421,1552029421
.long	723308426,723308426
.long	2461301159,2461301159
.long	4042393587,4042393587
.long	2715969870,2715969870
.long	3455375973,3455375973
.long	3586000134,3586000134
.long	526529745,526529745
.long	2331944644,2331944644
.long	2639474228,2639474228
.long	2689987490,2689987490
.long	853641733,853641733
.long	1978398372,1978398372
.long	971801355,971801355
.long	2867814464,2867814464
.long	111112542,111112542
.long	1360031421,1360031421
.long	4186579262,4186579262
.long	1023860118,1023860118
.long	2919579357,2919579357
.long	1186850381,1186850381
.long	3045938321,3045938321
.long	90031217,90031217
.long	1876166148,1876166148
.long	4279586912,4279586912
.long	620468249,620468249
.long	2548678102,2548678102
.long	3426959497,3426959497
.long	2006899047,2006899047
.long	3175278768,3175278768
.long	2290845959,2290845959
.long	945494503,945494503
.long	3689859193,3689859193
.long	1191869601,1191869601
.long	3910091388,3910091388
.long	3374220536,3374220536
.long	0,0
.long	2206629897,2206629897
.long	1223502642,1223502642
.long	2893025566,2893025566
.long	1316117100,1316117100
.long	4227796733,4227796733
.long	1446544655,1446544655
.long	517320253,517320253
.long	658058550,658058550
.long	1691946762,1691946762
.long	564550760,564550760
.long	3511966619,3511966619
.long	976107044,976107044
.long	2976320012,2976320012
.long	266819475,266819475
.long	3533106868,3533106868
.long	2660342555,2660342555
.long	1338359936,1338359936
.long	2720062561,2720062561
.long	1766553434,1766553434
.long	370807324,370807324
.long	179999714,179999714
.long	3844776128,3844776128
.long	1138762300,1138762300
.long	488053522,488053522
.long	185403662,185403662
.long	2915535858,2915535858
.long	3114841645,3114841645
.long	3366526484,3366526484
.long	2233069911,2233069911
.long	1275557295,1275557295
.long	3151862254,3151862254
.long	4250959779,4250959779
.long	2670068215,2670068215
.long	3170202204,3170202204
.long	3309004356,3309004356
.long	880737115,880737115
.long	1982415755,1982415755
.long	3703972811,3703972811
.long	1761406390,1761406390
.long	1676797112,1676797112
.long	3403428311,3403428311
.long	277177154,277177154
.long	1076008723,1076008723
.long	538035844,538035844
.long	2099530373,2099530373
.long	4164795346,4164795346
.long	288553390,288553390
.long	1839278535,1839278535
.long	1261411869,1261411869
.long	4080055004,4080055004
.long	3964831245,3964831245
.long	3504587127,3504587127
.long	1813426987,1813426987
.long	2579067049,2579067049
.long	4199060497,4199060497
.long	577038663,577038663
.long	3297574056,3297574056
.long	440397984,440397984
.long	3626794326,3626794326
.long	4019204898,4019204898
.long	3343796615,3343796615
.long	3251714265,3251714265
.long	4272081548,4272081548
.long	906744984,906744984
.long	3481400742,3481400742
.long	685669029,685669029
.long	646887386,646887386
.long	2764025151,2764025151
.long	3835509292,3835509292
.long	227702864,227702864
.long	2613862250,2613862250
.long	1648787028,1648787028
.long	3256061430,3256061430
.long	3904428176,3904428176
.long	1593260334,1593260334
.long	4121936770,4121936770
.long	3196083615,3196083615
.long	2090061929,2090061929
.long	2838353263,2838353263
.long	3004310991,3004310991
.long	999926984,999926984
.long	2809993232,2809993232
.long	1852021992,1852021992
.long	2075868123,2075868123
.long	158869197,158869197
.long	4095236462,4095236462
.long	28809964,28809964
.long	2828685187,2828685187
.long	1701746150,1701746150
.long	2129067946,2129067946
.long	147831841,147831841
.long	3873969647,3873969647
.long	3650873274,3650873274
.long	3459673930,3459673930
.long	3557400554,3557400554
.long	3598495785,3598495785
.long	2947720241,2947720241
.long	824393514,824393514
.long	815048134,815048134
.long	3227951669,3227951669
.long	935087732,935087732
.long	2798289660,2798289660
.long	2966458592,2966458592
.long	366520115,366520115
.long	1251476721,1251476721
.long	4158319681,4158319681
.long	240176511,240176511
.long	804688151,804688151
.long	2379631990,2379631990
.long	1303441219,1303441219
.long	1414376140,1414376140
.long	3741619940,3741619940
.long	3820343710,3820343710
.long	461924940,461924940
.long	3089050817,3089050817
.long	2136040774,2136040774
.long	82468509,82468509
.long	1563790337,1563790337
.long	1937016826,1937016826
.long	776014843,776014843
.long	1511876531,1511876531
.long	1389550482,1389550482
.long	861278441,861278441
.long	323475053,323475053
.long	2355222426,2355222426
.long	2047648055,2047648055
.long	2383738969,2383738969
.long	2302415851,2302415851
.long	3995576782,3995576782
.long	902390199,902390199
.long	3991215329,3991215329
.long	1018251130,1018251130
.long	1507840668,1507840668
.long	1064563285,1064563285
.long	2043548696,2043548696
.long	3208103795,3208103795
.long	3939366739,3939366739
.long	1537932639,1537932639
.long	342834655,342834655
.long	2262516856,2262516856
.long	2180231114,2180231114
.long	1053059257,1053059257
.long	741614648,741614648
.long	1598071746,1598071746
.long	1925389590,1925389590
.long	203809468,203809468
.long	2336832552,2336832552
.long	1100287487,1100287487
.long	1895934009,1895934009
.long	3736275976,3736275976
.long	2632234200,2632234200
.long	2428589668,2428589668
.long	1636092795,1636092795
.long	1890988757,1890988757
.long	1952214088,1952214088
.long	1113045200,1113045200
.byte	82,9,106,213,48,54,165,56
.byte	191,64,163,158,129,243,215,251
.byte	124,227,57,130,155,47,255,135
.byte	52,142,67,68,196,222,233,203
.byte	84,123,148,50,166,194,35,61
.byte	238,76,149,11,66,250,195,78
.byte	8,46,161,102,40,217,36,178
.byte	118,91,162,73,109,139,209,37
.byte	114,248,246,100,134,104,152,22
.byte	212,164,92,204,93,101,182,146
.byte	108,112,72,80,253,237,185,218
.byte	94,21,70,87,167,141,157,132
.byte	144,216,171,0,140,188,211,10
.byte	247,228,88,5,184,179,69,6
.byte	208,44,30,143,202,63,15,2
.byte	193,175,189,3,1,19,138,107
.byte	58,145,17,65,79,103,220,234
.byte	151,242,207,206,240,180,230,115
.byte	150,172,116,34,231,173,53,133
.byte	226,249,55,232,28,117,223,110
.byte	71,241,26,113,29,41,197,137
.byte	111,183,98,14,170,24,190,27
.byte	252,86,62,75,198,210,121,32
.byte	154,219,192,254,120,205,90,244
.byte	31,221,168,51,136,7,199,49
.byte	177,18,16,89,39,128,236,95
.byte	96,81,127,169,25,181,74,13
.byte	45,229,122,159,147,201,156,239
.byte	160,224,59,77,174,42,245,176
.byte	200,235,187,60,131,83,153,97
.byte	23,43,4,126,186,119,214,38
.byte	225,105,20,99,85,33,12,125
.byte	82,9,106,213,48,54,165,56
.byte	191,64,163,158,129,243,215,251
.byte	124,227,57,130,155,47,255,135
.byte	52,142,67,68,196,222,233,203
.byte	84,123,148,50,166,194,35,61
.byte	238,76,149,11,66,250,195,78
.byte	8,46,161,102,40,217,36,178
.byte	118,91,162,73,109,139,209,37
.byte	114,248,246,100,134,104,152,22
.byte	212,164,92,204,93,101,182,146
.byte	108,112,72,80,253,237,185,218
.byte	94,21,70,87,167,141,157,132
.byte	144,216,171,0,140,188,211,10
.byte	247,228,88,5,184,179,69,6
.byte	208,44,30,143,202,63,15,2
.byte	193,175,189,3,1,19,138,107
.byte	58,145,17,65,79,103,220,234
.byte	151,242,207,206,240,180,230,115
.byte	150,172,116,34,231,173,53,133
.byte	226,249,55,232,28,117,223,110
.byte	71,241,26,113,29,41,197,137
.byte	111,183,98,14,170,24,190,27
.byte	252,86,62,75,198,210,121,32
.byte	154,219,192,254,120,205,90,244
.byte	31,221,168,51,136,7,199,49
.byte	177,18,16,89,39,128,236,95
.byte	96,81,127,169,25,181,74,13
.byte	45,229,122,159,147,201,156,239
.byte	160,224,59,77,174,42,245,176
.byte	200,235,187,60,131,83,153,97
.byte	23,43,4,126,186,119,214,38
.byte	225,105,20,99,85,33,12,125
.byte	82,9,106,213,48,54,165,56
.byte	191,64,163,158,129,243,215,251
.byte	124,227,57,130,155,47,255,135
.byte	52,142,67,68,196,222,233,203
.byte	84,123,148,50,166,194,35,61
.byte	238,76,149,11,66,250,195,78
.byte	8,46,161,102,40,217,36,178
.byte	118,91,162,73,109,139,209,37
.byte	114,248,246,100,134,104,152,22
.byte	212,164,92,204,93,101,182,146
.byte	108,112,72,80,253,237,185,218
.byte	94,21,70,87,167,141,157,132
.byte	144,216,171,0,140,188,211,10
.byte	247,228,88,5,184,179,69,6
.byte	208,44,30,143,202,63,15,2
.byte	193,175,189,3,1,19,138,107
.byte	58,145,17,65,79,103,220,234
.byte	151,242,207,206,240,180,230,115
.byte	150,172,116,34,231,173,53,133
.byte	226,249,55,232,28,117,223,110
.byte	71,241,26,113,29,41,197,137
.byte	111,183,98,14,170,24,190,27
.byte	252,86,62,75,198,210,121,32
.byte	154,219,192,254,120,205,90,244
.byte	31,221,168,51,136,7,199,49
.byte	177,18,16,89,39,128,236,95
.byte	96,81,127,169,25,181,74,13
.byte	45,229,122,159,147,201,156,239
.byte	160,224,59,77,174,42,245,176
.byte	200,235,187,60,131,83,153,97
.byte	23,43,4,126,186,119,214,38
.byte	225,105,20,99,85,33,12,125
.byte	82,9,106,213,48,54,165,56
.byte	191,64,163,158,129,243,215,251
.byte	124,227,57,130,155,47,255,135
.byte	52,142,67,68,196,222,233,203
.byte	84,123,148,50,166,194,35,61
.byte	238,76,149,11,66,250,195,78
.byte	8,46,161,102,40,217,36,178
.byte	118,91,162,73,109,139,209,37
.byte	114,248,246,100,134,104,152,22
.byte	212,164,92,204,93,101,182,146
.byte	108,112,72,80,253,237,185,218
.byte	94,21,70,87,167,141,157,132
.byte	144,216,171,0,140,188,211,10
.byte	247,228,88,5,184,179,69,6
.byte	208,44,30,143,202,63,15,2
.byte	193,175,189,3,1,19,138,107
.byte	58,145,17,65,79,103,220,234
.byte	151,242,207,206,240,180,230,115
.byte	150,172,116,34,231,173,53,133
.byte	226,249,55,232,28,117,223,110
.byte	71,241,26,113,29,41,197,137
.byte	111,183,98,14,170,24,190,27
.byte	252,86,62,75,198,210,121,32
.byte	154,219,192,254,120,205,90,244
.byte	31,221,168,51,136,7,199,49
.byte	177,18,16,89,39,128,236,95
.byte	96,81,127,169,25,181,74,13
.byte	45,229,122,159,147,201,156,239
.byte	160,224,59,77,174,42,245,176
.byte	200,235,187,60,131,83,153,97
.byte	23,43,4,126,186,119,214,38
.byte	225,105,20,99,85,33,12,125
.size	_x86_AES_decrypt,.-_x86_AES_decrypt
.globl	asm_AES_decrypt
.hidden	asm_AES_decrypt
.type	asm_AES_decrypt,@function
.align	16
asm_AES_decrypt:
.L_asm_AES_decrypt_begin:
	pushl	%ebp
	pushl	%ebx
	pushl	%esi
	pushl	%edi
	movl	20(%esp),%esi
	movl	28(%esp),%edi
	movl	%esp,%eax
	subl	$36,%esp
	andl	$-64,%esp
	leal	-127(%edi),%ebx
	subl	%esp,%ebx
	negl	%ebx
	andl	$960,%ebx
	subl	%ebx,%esp
	addl	$4,%esp
	movl	%eax,28(%esp)
	call	.L010pic_point
.L010pic_point:
	popl	%ebp
	leal	OPENSSL_ia32cap_P-.L010pic_point(%ebp),%eax
	leal	.LAES_Td-.L010pic_point(%ebp),%ebp
	leal	764(%esp),%ebx
	subl	%ebp,%ebx
	andl	$768,%ebx
	leal	2176(%ebp,%ebx,1),%ebp
	btl	$25,(%eax)
	jnc	.L011x86
	movq	(%esi),%mm0
	movq	8(%esi),%mm4
	call	_sse_AES_decrypt_compact
	movl	28(%esp),%esp
	movl	24(%esp),%esi
	movq	%mm0,(%esi)
	movq	%mm4,8(%esi)
	emms
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
.align	16
.L011x86:
	movl	%ebp,24(%esp)
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	call	_x86_AES_decrypt_compact
	movl	28(%esp),%esp
	movl	24(%esp),%esi
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
.size	asm_AES_decrypt,.-.L_asm_AES_decrypt_begin
.globl	asm_AES_cbc_encrypt
.hidden	asm_AES_cbc_encrypt
.type	asm_AES_cbc_encrypt,@function
.align	16
asm_AES_cbc_encrypt:
.L_asm_AES_cbc_encrypt_begin:
	pushl	%ebp
	pushl	%ebx
	pushl	%esi
	pushl	%edi
	movl	28(%esp),%ecx
	cmpl	$0,%ecx
	je	.L012drop_out
	call	.L013pic_point
.L013pic_point:
	popl	%ebp
	leal	OPENSSL_ia32cap_P-.L013pic_point(%ebp),%eax
	cmpl	$0,40(%esp)
	leal	.LAES_Te-.L013pic_point(%ebp),%ebp
	jne	.L014picked_te
	leal	.LAES_Td-.LAES_Te(%ebp),%ebp
.L014picked_te:
	pushfl
	cld
	cmpl	$512,%ecx
	jb	.L015slow_way
	testl	$15,%ecx
	jnz	.L015slow_way
	btl	$28,(%eax)
	jc	.L015slow_way
	leal	-324(%esp),%esi
	andl	$-64,%esi
	movl	%ebp,%eax
	leal	2304(%ebp),%ebx
	movl	%esi,%edx
	andl	$4095,%eax
	andl	$4095,%ebx
	andl	$4095,%edx
	cmpl	%ebx,%edx
	jb	.L016tbl_break_out
	subl	%ebx,%edx
	subl	%edx,%esi
	jmp	.L017tbl_ok
.align	4
.L016tbl_break_out:
	subl	%eax,%edx
	andl	$4095,%edx
	addl	$384,%edx
	subl	%edx,%esi
.align	4
.L017tbl_ok:
	leal	24(%esp),%edx
	xchgl	%esi,%esp
	addl	$4,%esp
	movl	%ebp,24(%esp)
	movl	%esi,28(%esp)
	movl	(%edx),%eax
	movl	4(%edx),%ebx
	movl	12(%edx),%edi
	movl	16(%edx),%esi
	movl	20(%edx),%edx
	movl	%eax,32(%esp)
	movl	%ebx,36(%esp)
	movl	%ecx,40(%esp)
	movl	%edi,44(%esp)
	movl	%esi,48(%esp)
	movl	$0,316(%esp)
	movl	%edi,%ebx
	movl	$61,%ecx
	subl	%ebp,%ebx
	movl	%edi,%esi
	andl	$4095,%ebx
	leal	76(%esp),%edi
	cmpl	$2304,%ebx
	jb	.L018do_copy
	cmpl	$3852,%ebx
	jb	.L019skip_copy
.align	4
.L018do_copy:
	movl	%edi,44(%esp)
.long	2784229001
.L019skip_copy:
	movl	$16,%edi
.align	4
.L020prefetch_tbl:
	movl	(%ebp),%eax
	movl	32(%ebp),%ebx
	movl	64(%ebp),%ecx
	movl	96(%ebp),%esi
	leal	128(%ebp),%ebp
	subl	$1,%edi
	jnz	.L020prefetch_tbl
	subl	$2048,%ebp
	movl	32(%esp),%esi
	movl	48(%esp),%edi
	cmpl	$0,%edx
	je	.L021fast_decrypt
	movl	(%edi),%eax
	movl	4(%edi),%ebx
.align	16
.L022fast_enc_loop:
	movl	8(%edi),%ecx
	movl	12(%edi),%edx
	xorl	(%esi),%eax
	xorl	4(%esi),%ebx
	xorl	8(%esi),%ecx
	xorl	12(%esi),%edx
	movl	44(%esp),%edi
	call	_x86_AES_encrypt
	movl	32(%esp),%esi
	movl	36(%esp),%edi
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	leal	16(%esi),%esi
	movl	40(%esp),%ecx
	movl	%esi,32(%esp)
	leal	16(%edi),%edx
	movl	%edx,36(%esp)
	subl	$16,%ecx
	movl	%ecx,40(%esp)
	jnz	.L022fast_enc_loop
	movl	48(%esp),%esi
	movl	8(%edi),%ecx
	movl	12(%edi),%edx
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	cmpl	$0,316(%esp)
	movl	44(%esp),%edi
	je	.L023skip_ezero
	movl	$60,%ecx
	xorl	%eax,%eax
.align	4
.long	2884892297
.L023skip_ezero:
	movl	28(%esp),%esp
	popfl
.L012drop_out:
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
	pushfl
.align	16
.L021fast_decrypt:
	cmpl	36(%esp),%esi
	je	.L024fast_dec_in_place
	movl	%edi,52(%esp)
.align	4
.align	16
.L025fast_dec_loop:
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	movl	44(%esp),%edi
	call	_x86_AES_decrypt
	movl	52(%esp),%edi
	movl	40(%esp),%esi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	movl	36(%esp),%edi
	movl	32(%esp),%esi
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	40(%esp),%ecx
	movl	%esi,52(%esp)
	leal	16(%esi),%esi
	movl	%esi,32(%esp)
	leal	16(%edi),%edi
	movl	%edi,36(%esp)
	subl	$16,%ecx
	movl	%ecx,40(%esp)
	jnz	.L025fast_dec_loop
	movl	52(%esp),%edi
	movl	48(%esp),%esi
	movl	(%edi),%eax
	movl	4(%edi),%ebx
	movl	8(%edi),%ecx
	movl	12(%edi),%edx
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	jmp	.L026fast_dec_out
.align	16
.L024fast_dec_in_place:
.L027fast_dec_in_place_loop:
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	leal	60(%esp),%edi
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	44(%esp),%edi
	call	_x86_AES_decrypt
	movl	48(%esp),%edi
	movl	36(%esp),%esi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	leal	16(%esi),%esi
	movl	%esi,36(%esp)
	leal	60(%esp),%esi
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	32(%esp),%esi
	movl	40(%esp),%ecx
	leal	16(%esi),%esi
	movl	%esi,32(%esp)
	subl	$16,%ecx
	movl	%ecx,40(%esp)
	jnz	.L027fast_dec_in_place_loop
.align	4
.L026fast_dec_out:
	cmpl	$0,316(%esp)
	movl	44(%esp),%edi
	je	.L028skip_dzero
	movl	$60,%ecx
	xorl	%eax,%eax
.align	4
.long	2884892297
.L028skip_dzero:
	movl	28(%esp),%esp
	popfl
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
	pushfl
.align	16
.L015slow_way:
	movl	(%eax),%eax
	movl	36(%esp),%edi
	leal	-80(%esp),%esi
	andl	$-64,%esi
	leal	-143(%edi),%ebx
	subl	%esi,%ebx
	negl	%ebx
	andl	$960,%ebx
	subl	%ebx,%esi
	leal	768(%esi),%ebx
	subl	%ebp,%ebx
	andl	$768,%ebx
	leal	2176(%ebp,%ebx,1),%ebp
	leal	24(%esp),%edx
	xchgl	%esi,%esp
	addl	$4,%esp
	movl	%ebp,24(%esp)
	movl	%esi,28(%esp)
	movl	%eax,52(%esp)
	movl	(%edx),%eax
	movl	4(%edx),%ebx
	movl	16(%edx),%esi
	movl	20(%edx),%edx
	movl	%eax,32(%esp)
	movl	%ebx,36(%esp)
	movl	%ecx,40(%esp)
	movl	%edi,44(%esp)
	movl	%esi,48(%esp)
	movl	%esi,%edi
	movl	%eax,%esi
	cmpl	$0,%edx
	je	.L029slow_decrypt
	cmpl	$16,%ecx
	movl	%ebx,%edx
	jb	.L030slow_enc_tail
	btl	$25,52(%esp)
	jnc	.L031slow_enc_x86
	movq	(%edi),%mm0
	movq	8(%edi),%mm4
.align	16
.L032slow_enc_loop_sse:
	pxor	(%esi),%mm0
	pxor	8(%esi),%mm4
	movl	44(%esp),%edi
	call	_sse_AES_encrypt_compact
	movl	32(%esp),%esi
	movl	36(%esp),%edi
	movl	40(%esp),%ecx
	movq	%mm0,(%edi)
	movq	%mm4,8(%edi)
	leal	16(%esi),%esi
	movl	%esi,32(%esp)
	leal	16(%edi),%edx
	movl	%edx,36(%esp)
	subl	$16,%ecx
	cmpl	$16,%ecx
	movl	%ecx,40(%esp)
	jae	.L032slow_enc_loop_sse
	testl	$15,%ecx
	jnz	.L030slow_enc_tail
	movl	48(%esp),%esi
	movq	%mm0,(%esi)
	movq	%mm4,8(%esi)
	emms
	movl	28(%esp),%esp
	popfl
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
	pushfl
.align	16
.L031slow_enc_x86:
	movl	(%edi),%eax
	movl	4(%edi),%ebx
.align	4
.L033slow_enc_loop_x86:
	movl	8(%edi),%ecx
	movl	12(%edi),%edx
	xorl	(%esi),%eax
	xorl	4(%esi),%ebx
	xorl	8(%esi),%ecx
	xorl	12(%esi),%edx
	movl	44(%esp),%edi
	call	_x86_AES_encrypt_compact
	movl	32(%esp),%esi
	movl	36(%esp),%edi
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	40(%esp),%ecx
	leal	16(%esi),%esi
	movl	%esi,32(%esp)
	leal	16(%edi),%edx
	movl	%edx,36(%esp)
	subl	$16,%ecx
	cmpl	$16,%ecx
	movl	%ecx,40(%esp)
	jae	.L033slow_enc_loop_x86
	testl	$15,%ecx
	jnz	.L030slow_enc_tail
	movl	48(%esp),%esi
	movl	8(%edi),%ecx
	movl	12(%edi),%edx
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	movl	28(%esp),%esp
	popfl
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
	pushfl
.align	16
.L030slow_enc_tail:
	emms
	movl	%edx,%edi
	movl	$16,%ebx
	subl	%ecx,%ebx
	cmpl	%esi,%edi
	je	.L034enc_in_place
.align	4
.long	2767451785
	jmp	.L035enc_skip_in_place
.L034enc_in_place:
	leal	(%edi,%ecx,1),%edi
.L035enc_skip_in_place:
	movl	%ebx,%ecx
	xorl	%eax,%eax
.align	4
.long	2868115081
	movl	48(%esp),%edi
	movl	%edx,%esi
	movl	(%edi),%eax
	movl	4(%edi),%ebx
	movl	$16,40(%esp)
	jmp	.L033slow_enc_loop_x86
.align	16
.L029slow_decrypt:
	btl	$25,52(%esp)
	jnc	.L036slow_dec_loop_x86
.align	4
.L037slow_dec_loop_sse:
	movq	(%esi),%mm0
	movq	8(%esi),%mm4
	movl	44(%esp),%edi
	call	_sse_AES_decrypt_compact
	movl	32(%esp),%esi
	leal	60(%esp),%eax
	movl	36(%esp),%ebx
	movl	40(%esp),%ecx
	movl	48(%esp),%edi
	movq	(%esi),%mm1
	movq	8(%esi),%mm5
	pxor	(%edi),%mm0
	pxor	8(%edi),%mm4
	movq	%mm1,(%edi)
	movq	%mm5,8(%edi)
	subl	$16,%ecx
	jc	.L038slow_dec_partial_sse
	movq	%mm0,(%ebx)
	movq	%mm4,8(%ebx)
	leal	16(%ebx),%ebx
	movl	%ebx,36(%esp)
	leal	16(%esi),%esi
	movl	%esi,32(%esp)
	movl	%ecx,40(%esp)
	jnz	.L037slow_dec_loop_sse
	emms
	movl	28(%esp),%esp
	popfl
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
	pushfl
.align	16
.L038slow_dec_partial_sse:
	movq	%mm0,(%eax)
	movq	%mm4,8(%eax)
	emms
	addl	$16,%ecx
	movl	%ebx,%edi
	movl	%eax,%esi
.align	4
.long	2767451785
	movl	28(%esp),%esp
	popfl
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
	pushfl
.align	16
.L036slow_dec_loop_x86:
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	leal	60(%esp),%edi
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	44(%esp),%edi
	call	_x86_AES_decrypt_compact
	movl	48(%esp),%edi
	movl	40(%esp),%esi
	xorl	(%edi),%eax
	xorl	4(%edi),%ebx
	xorl	8(%edi),%ecx
	xorl	12(%edi),%edx
	subl	$16,%esi
	jc	.L039slow_dec_partial_x86
	movl	%esi,40(%esp)
	movl	36(%esp),%esi
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	leal	16(%esi),%esi
	movl	%esi,36(%esp)
	leal	60(%esp),%esi
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	32(%esp),%esi
	leal	16(%esi),%esi
	movl	%esi,32(%esp)
	jnz	.L036slow_dec_loop_x86
	movl	28(%esp),%esp
	popfl
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
	pushfl
.align	16
.L039slow_dec_partial_x86:
	leal	60(%esp),%esi
	movl	%eax,(%esi)
	movl	%ebx,4(%esi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	movl	32(%esp),%esi
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	40(%esp),%ecx
	movl	36(%esp),%edi
	leal	60(%esp),%esi
.align	4
.long	2767451785
	movl	28(%esp),%esp
	popfl
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
.size	asm_AES_cbc_encrypt,.-.L_asm_AES_cbc_encrypt_begin
.hidden	_x86_AES_set_encrypt_key
.type	_x86_AES_set_encrypt_key,@function
.align	16
_x86_AES_set_encrypt_key:
	pushl	%ebp
	pushl	%ebx
	pushl	%esi
	pushl	%edi
	movl	24(%esp),%esi
	movl	32(%esp),%edi
	testl	$-1,%esi
	jz	.L040badpointer
	testl	$-1,%edi
	jz	.L040badpointer
	call	.L041pic_point
.L041pic_point:
	popl	%ebp
	leal	.LAES_Te-.L041pic_point(%ebp),%ebp
	leal	2176(%ebp),%ebp
	movl	-128(%ebp),%eax
	movl	-96(%ebp),%ebx
	movl	-64(%ebp),%ecx
	movl	-32(%ebp),%edx
	movl	(%ebp),%eax
	movl	32(%ebp),%ebx
	movl	64(%ebp),%ecx
	movl	96(%ebp),%edx
	movl	28(%esp),%ecx
	cmpl	$128,%ecx
	je	.L04210rounds
	cmpl	$192,%ecx
	je	.L04312rounds
	cmpl	$256,%ecx
	je	.L04414rounds
	movl	$-2,%eax
	jmp	.L045exit
.L04210rounds:
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	xorl	%ecx,%ecx
	jmp	.L04610shortcut
.align	4
.L04710loop:
	movl	(%edi),%eax
	movl	12(%edi),%edx
.L04610shortcut:
	movzbl	%dl,%esi
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	shll	$24,%ebx
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shrl	$16,%edx
	movzbl	%dl,%esi
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	shll	$8,%ebx
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%eax
	xorl	896(%ebp,%ecx,4),%eax
	movl	%eax,16(%edi)
	xorl	4(%edi),%eax
	movl	%eax,20(%edi)
	xorl	8(%edi),%eax
	movl	%eax,24(%edi)
	xorl	12(%edi),%eax
	movl	%eax,28(%edi)
	incl	%ecx
	addl	$16,%edi
	cmpl	$10,%ecx
	jl	.L04710loop
	movl	$10,80(%edi)
	xorl	%eax,%eax
	jmp	.L045exit
.L04312rounds:
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	16(%esi),%ecx
	movl	20(%esi),%edx
	movl	%ecx,16(%edi)
	movl	%edx,20(%edi)
	xorl	%ecx,%ecx
	jmp	.L04812shortcut
.align	4
.L04912loop:
	movl	(%edi),%eax
	movl	20(%edi),%edx
.L04812shortcut:
	movzbl	%dl,%esi
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	shll	$24,%ebx
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shrl	$16,%edx
	movzbl	%dl,%esi
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	shll	$8,%ebx
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%eax
	xorl	896(%ebp,%ecx,4),%eax
	movl	%eax,24(%edi)
	xorl	4(%edi),%eax
	movl	%eax,28(%edi)
	xorl	8(%edi),%eax
	movl	%eax,32(%edi)
	xorl	12(%edi),%eax
	movl	%eax,36(%edi)
	cmpl	$7,%ecx
	je	.L05012break
	incl	%ecx
	xorl	16(%edi),%eax
	movl	%eax,40(%edi)
	xorl	20(%edi),%eax
	movl	%eax,44(%edi)
	addl	$24,%edi
	jmp	.L04912loop
.L05012break:
	movl	$12,72(%edi)
	xorl	%eax,%eax
	jmp	.L045exit
.L04414rounds:
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	8(%esi),%ecx
	movl	12(%esi),%edx
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,8(%edi)
	movl	%edx,12(%edi)
	movl	16(%esi),%eax
	movl	20(%esi),%ebx
	movl	24(%esi),%ecx
	movl	28(%esi),%edx
	movl	%eax,16(%edi)
	movl	%ebx,20(%edi)
	movl	%ecx,24(%edi)
	movl	%edx,28(%edi)
	xorl	%ecx,%ecx
	jmp	.L05114shortcut
.align	4
.L05214loop:
	movl	28(%edi),%edx
.L05114shortcut:
	movl	(%edi),%eax
	movzbl	%dl,%esi
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	shll	$24,%ebx
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shrl	$16,%edx
	movzbl	%dl,%esi
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	shll	$8,%ebx
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shll	$16,%ebx
	xorl	%ebx,%eax
	xorl	896(%ebp,%ecx,4),%eax
	movl	%eax,32(%edi)
	xorl	4(%edi),%eax
	movl	%eax,36(%edi)
	xorl	8(%edi),%eax
	movl	%eax,40(%edi)
	xorl	12(%edi),%eax
	movl	%eax,44(%edi)
	cmpl	$6,%ecx
	je	.L05314break
	incl	%ecx
	movl	%eax,%edx
	movl	16(%edi),%eax
	movzbl	%dl,%esi
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shrl	$16,%edx
	shll	$8,%ebx
	movzbl	%dl,%esi
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	movzbl	%dh,%esi
	shll	$16,%ebx
	xorl	%ebx,%eax
	movzbl	-128(%ebp,%esi,1),%ebx
	shll	$24,%ebx
	xorl	%ebx,%eax
	movl	%eax,48(%edi)
	xorl	20(%edi),%eax
	movl	%eax,52(%edi)
	xorl	24(%edi),%eax
	movl	%eax,56(%edi)
	xorl	28(%edi),%eax
	movl	%eax,60(%edi)
	addl	$32,%edi
	jmp	.L05214loop
.L05314break:
	movl	$14,48(%edi)
	xorl	%eax,%eax
	jmp	.L045exit
.L040badpointer:
	movl	$-1,%eax
.L045exit:
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
.size	_x86_AES_set_encrypt_key,.-_x86_AES_set_encrypt_key
.globl	asm_AES_set_encrypt_key
.hidden	asm_AES_set_encrypt_key
.type	asm_AES_set_encrypt_key,@function
.align	16
asm_AES_set_encrypt_key:
.L_asm_AES_set_encrypt_key_begin:
	call	_x86_AES_set_encrypt_key
	ret
.size	asm_AES_set_encrypt_key,.-.L_asm_AES_set_encrypt_key_begin
.globl	asm_AES_set_decrypt_key
.hidden	asm_AES_set_decrypt_key
.type	asm_AES_set_decrypt_key,@function
.align	16
asm_AES_set_decrypt_key:
.L_asm_AES_set_decrypt_key_begin:
	call	_x86_AES_set_encrypt_key
	cmpl	$0,%eax
	je	.L054proceed
	ret
.L054proceed:
	pushl	%ebp
	pushl	%ebx
	pushl	%esi
	pushl	%edi
	movl	28(%esp),%esi
	movl	240(%esi),%ecx
	leal	(,%ecx,4),%ecx
	leal	(%esi,%ecx,4),%edi
.align	4
.L055invert:
	movl	(%esi),%eax
	movl	4(%esi),%ebx
	movl	(%edi),%ecx
	movl	4(%edi),%edx
	movl	%eax,(%edi)
	movl	%ebx,4(%edi)
	movl	%ecx,(%esi)
	movl	%edx,4(%esi)
	movl	8(%esi),%eax
	movl	12(%esi),%ebx
	movl	8(%edi),%ecx
	movl	12(%edi),%edx
	movl	%eax,8(%edi)
	movl	%ebx,12(%edi)
	movl	%ecx,8(%esi)
	movl	%edx,12(%esi)
	addl	$16,%esi
	subl	$16,%edi
	cmpl	%edi,%esi
	jne	.L055invert
	movl	28(%esp),%edi
	movl	240(%edi),%esi
	leal	-2(%esi,%esi,1),%esi
	leal	(%edi,%esi,8),%esi
	movl	%esi,28(%esp)
	movl	16(%edi),%eax
.align	4
.L056permute:
	addl	$16,%edi
	movl	$2155905152,%ebp
	andl	%eax,%ebp
	leal	(%eax,%eax,1),%ebx
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%ebx
	andl	$454761243,%esi
	xorl	%esi,%ebx
	movl	$2155905152,%ebp
	andl	%ebx,%ebp
	leal	(%ebx,%ebx,1),%ecx
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%ecx
	andl	$454761243,%esi
	xorl	%eax,%ebx
	xorl	%esi,%ecx
	movl	$2155905152,%ebp
	andl	%ecx,%ebp
	leal	(%ecx,%ecx,1),%edx
	movl	%ebp,%esi
	shrl	$7,%ebp
	xorl	%eax,%ecx
	subl	%ebp,%esi
	andl	$4278124286,%edx
	andl	$454761243,%esi
	roll	$8,%eax
	xorl	%esi,%edx
	movl	4(%edi),%ebp
	xorl	%ebx,%eax
	xorl	%edx,%ebx
	xorl	%ecx,%eax
	roll	$24,%ebx
	xorl	%edx,%ecx
	xorl	%edx,%eax
	roll	$16,%ecx
	xorl	%ebx,%eax
	roll	$8,%edx
	xorl	%ecx,%eax
	movl	%ebp,%ebx
	xorl	%edx,%eax
	movl	%eax,(%edi)
	movl	$2155905152,%ebp
	andl	%ebx,%ebp
	leal	(%ebx,%ebx,1),%ecx
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%ecx
	andl	$454761243,%esi
	xorl	%esi,%ecx
	movl	$2155905152,%ebp
	andl	%ecx,%ebp
	leal	(%ecx,%ecx,1),%edx
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%edx
	andl	$454761243,%esi
	xorl	%ebx,%ecx
	xorl	%esi,%edx
	movl	$2155905152,%ebp
	andl	%edx,%ebp
	leal	(%edx,%edx,1),%eax
	movl	%ebp,%esi
	shrl	$7,%ebp
	xorl	%ebx,%edx
	subl	%ebp,%esi
	andl	$4278124286,%eax
	andl	$454761243,%esi
	roll	$8,%ebx
	xorl	%esi,%eax
	movl	8(%edi),%ebp
	xorl	%ecx,%ebx
	xorl	%eax,%ecx
	xorl	%edx,%ebx
	roll	$24,%ecx
	xorl	%eax,%edx
	xorl	%eax,%ebx
	roll	$16,%edx
	xorl	%ecx,%ebx
	roll	$8,%eax
	xorl	%edx,%ebx
	movl	%ebp,%ecx
	xorl	%eax,%ebx
	movl	%ebx,4(%edi)
	movl	$2155905152,%ebp
	andl	%ecx,%ebp
	leal	(%ecx,%ecx,1),%edx
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%edx
	andl	$454761243,%esi
	xorl	%esi,%edx
	movl	$2155905152,%ebp
	andl	%edx,%ebp
	leal	(%edx,%edx,1),%eax
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%eax
	andl	$454761243,%esi
	xorl	%ecx,%edx
	xorl	%esi,%eax
	movl	$2155905152,%ebp
	andl	%eax,%ebp
	leal	(%eax,%eax,1),%ebx
	movl	%ebp,%esi
	shrl	$7,%ebp
	xorl	%ecx,%eax
	subl	%ebp,%esi
	andl	$4278124286,%ebx
	andl	$454761243,%esi
	roll	$8,%ecx
	xorl	%esi,%ebx
	movl	12(%edi),%ebp
	xorl	%edx,%ecx
	xorl	%ebx,%edx
	xorl	%eax,%ecx
	roll	$24,%edx
	xorl	%ebx,%eax
	xorl	%ebx,%ecx
	roll	$16,%eax
	xorl	%edx,%ecx
	roll	$8,%ebx
	xorl	%eax,%ecx
	movl	%ebp,%edx
	xorl	%ebx,%ecx
	movl	%ecx,8(%edi)
	movl	$2155905152,%ebp
	andl	%edx,%ebp
	leal	(%edx,%edx,1),%eax
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%eax
	andl	$454761243,%esi
	xorl	%esi,%eax
	movl	$2155905152,%ebp
	andl	%eax,%ebp
	leal	(%eax,%eax,1),%ebx
	movl	%ebp,%esi
	shrl	$7,%ebp
	subl	%ebp,%esi
	andl	$4278124286,%ebx
	andl	$454761243,%esi
	xorl	%edx,%eax
	xorl	%esi,%ebx
	movl	$2155905152,%ebp
	andl	%ebx,%ebp
	leal	(%ebx,%ebx,1),%ecx
	movl	%ebp,%esi
	shrl	$7,%ebp
	xorl	%edx,%ebx
	subl	%ebp,%esi
	andl	$4278124286,%ecx
	andl	$454761243,%esi
	roll	$8,%edx
	xorl	%esi,%ecx
	movl	16(%edi),%ebp
	xorl	%eax,%edx
	xorl	%ecx,%eax
	xorl	%ebx,%edx
	roll	$24,%eax
	xorl	%ecx,%ebx
	xorl	%ecx,%edx
	roll	$16,%ebx
	xorl	%eax,%edx
	roll	$8,%ecx
	xorl	%ebx,%edx
	movl	%ebp,%eax
	xorl	%ecx,%edx
	movl	%edx,12(%edi)
	cmpl	28(%esp),%edi
	jb	.L056permute
	xorl	%eax,%eax
	popl	%edi
	popl	%esi
	popl	%ebx
	popl	%ebp
	ret
.size	asm_AES_set_decrypt_key,.-.L_asm_AES_set_decrypt_key_begin
.byte	65,69,83,32,102,111,114,32,120,56,54,44,32,67,82,89
.byte	80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114
.byte	111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
#endif