mirror of
https://github.com/pineappleEA/pineapple-src.git
synced 2024-11-26 05:18:24 -05:00
1176 lines
22 KiB
ArmAsm
Executable File
1176 lines
22 KiB
ArmAsm
Executable File
#include "x86_arch.h"
|
|
.text
|
|
|
|
.globl gcm_gmult_4bit
|
|
.def gcm_gmult_4bit; .scl 2; .type 32; .endef
|
|
.p2align 4
|
|
gcm_gmult_4bit:
|
|
movq %rdi,8(%rsp)
|
|
movq %rsi,16(%rsp)
|
|
movq %rsp,%rax
|
|
.LSEH_begin_gcm_gmult_4bit:
|
|
movq %rcx,%rdi
|
|
movq %rdx,%rsi
|
|
|
|
pushq %rbx
|
|
pushq %rbp
|
|
pushq %r12
|
|
.Lgmult_prologue:
|
|
|
|
movzbq 15(%rdi),%r8
|
|
leaq .Lrem_4bit(%rip),%r11
|
|
xorq %rax,%rax
|
|
xorq %rbx,%rbx
|
|
movb %r8b,%al
|
|
movb %r8b,%bl
|
|
shlb $4,%al
|
|
movq $14,%rcx
|
|
movq 8(%rsi,%rax,1),%r8
|
|
movq (%rsi,%rax,1),%r9
|
|
andb $240,%bl
|
|
movq %r8,%rdx
|
|
jmp .Loop1
|
|
|
|
.p2align 4
|
|
.Loop1:
|
|
shrq $4,%r8
|
|
andq $15,%rdx
|
|
movq %r9,%r10
|
|
movb (%rdi,%rcx,1),%al
|
|
shrq $4,%r9
|
|
xorq 8(%rsi,%rbx,1),%r8
|
|
shlq $60,%r10
|
|
xorq (%rsi,%rbx,1),%r9
|
|
movb %al,%bl
|
|
xorq (%r11,%rdx,8),%r9
|
|
movq %r8,%rdx
|
|
shlb $4,%al
|
|
xorq %r10,%r8
|
|
decq %rcx
|
|
js .Lbreak1
|
|
|
|
shrq $4,%r8
|
|
andq $15,%rdx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
shlq $60,%r10
|
|
xorq (%rsi,%rax,1),%r9
|
|
andb $240,%bl
|
|
xorq (%r11,%rdx,8),%r9
|
|
movq %r8,%rdx
|
|
xorq %r10,%r8
|
|
jmp .Loop1
|
|
|
|
.p2align 4
|
|
.Lbreak1:
|
|
shrq $4,%r8
|
|
andq $15,%rdx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
shlq $60,%r10
|
|
xorq (%rsi,%rax,1),%r9
|
|
andb $240,%bl
|
|
xorq (%r11,%rdx,8),%r9
|
|
movq %r8,%rdx
|
|
xorq %r10,%r8
|
|
|
|
shrq $4,%r8
|
|
andq $15,%rdx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
xorq 8(%rsi,%rbx,1),%r8
|
|
shlq $60,%r10
|
|
xorq (%rsi,%rbx,1),%r9
|
|
xorq %r10,%r8
|
|
xorq (%r11,%rdx,8),%r9
|
|
|
|
bswapq %r8
|
|
bswapq %r9
|
|
movq %r8,8(%rdi)
|
|
movq %r9,(%rdi)
|
|
|
|
movq 16(%rsp),%rbx
|
|
leaq 24(%rsp),%rsp
|
|
.Lgmult_epilogue:
|
|
movq 8(%rsp),%rdi
|
|
movq 16(%rsp),%rsi
|
|
retq
|
|
.LSEH_end_gcm_gmult_4bit:
|
|
.globl gcm_ghash_4bit
|
|
.def gcm_ghash_4bit; .scl 2; .type 32; .endef
|
|
.p2align 4
|
|
gcm_ghash_4bit:
|
|
movq %rdi,8(%rsp)
|
|
movq %rsi,16(%rsp)
|
|
movq %rsp,%rax
|
|
.LSEH_begin_gcm_ghash_4bit:
|
|
movq %rcx,%rdi
|
|
movq %rdx,%rsi
|
|
movq %r8,%rdx
|
|
movq %r9,%rcx
|
|
|
|
pushq %rbx
|
|
pushq %rbp
|
|
pushq %r12
|
|
pushq %r13
|
|
pushq %r14
|
|
pushq %r15
|
|
subq $280,%rsp
|
|
.Lghash_prologue:
|
|
movq %rdx,%r14
|
|
movq %rcx,%r15
|
|
subq $-128,%rsi
|
|
leaq 16+128(%rsp),%rbp
|
|
xorl %edx,%edx
|
|
movq 0+0-128(%rsi),%r8
|
|
movq 0+8-128(%rsi),%rax
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq 16+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq 16+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,0(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,0(%rbp)
|
|
movq 32+0-128(%rsi),%r8
|
|
shlb $4,%dl
|
|
movq %rax,0-128(%rbp)
|
|
movq 32+8-128(%rsi),%rax
|
|
shlq $60,%r10
|
|
movb %dl,1(%rsp)
|
|
orq %r10,%rbx
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq %r9,8(%rbp)
|
|
movq 48+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq %rbx,8-128(%rbp)
|
|
movq 48+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,2(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,16(%rbp)
|
|
movq 64+0-128(%rsi),%r8
|
|
shlb $4,%dl
|
|
movq %rax,16-128(%rbp)
|
|
movq 64+8-128(%rsi),%rax
|
|
shlq $60,%r10
|
|
movb %dl,3(%rsp)
|
|
orq %r10,%rbx
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq %r9,24(%rbp)
|
|
movq 80+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq %rbx,24-128(%rbp)
|
|
movq 80+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,4(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,32(%rbp)
|
|
movq 96+0-128(%rsi),%r8
|
|
shlb $4,%dl
|
|
movq %rax,32-128(%rbp)
|
|
movq 96+8-128(%rsi),%rax
|
|
shlq $60,%r10
|
|
movb %dl,5(%rsp)
|
|
orq %r10,%rbx
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq %r9,40(%rbp)
|
|
movq 112+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq %rbx,40-128(%rbp)
|
|
movq 112+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,6(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,48(%rbp)
|
|
movq 128+0-128(%rsi),%r8
|
|
shlb $4,%dl
|
|
movq %rax,48-128(%rbp)
|
|
movq 128+8-128(%rsi),%rax
|
|
shlq $60,%r10
|
|
movb %dl,7(%rsp)
|
|
orq %r10,%rbx
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq %r9,56(%rbp)
|
|
movq 144+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq %rbx,56-128(%rbp)
|
|
movq 144+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,8(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,64(%rbp)
|
|
movq 160+0-128(%rsi),%r8
|
|
shlb $4,%dl
|
|
movq %rax,64-128(%rbp)
|
|
movq 160+8-128(%rsi),%rax
|
|
shlq $60,%r10
|
|
movb %dl,9(%rsp)
|
|
orq %r10,%rbx
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq %r9,72(%rbp)
|
|
movq 176+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq %rbx,72-128(%rbp)
|
|
movq 176+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,10(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,80(%rbp)
|
|
movq 192+0-128(%rsi),%r8
|
|
shlb $4,%dl
|
|
movq %rax,80-128(%rbp)
|
|
movq 192+8-128(%rsi),%rax
|
|
shlq $60,%r10
|
|
movb %dl,11(%rsp)
|
|
orq %r10,%rbx
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq %r9,88(%rbp)
|
|
movq 208+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq %rbx,88-128(%rbp)
|
|
movq 208+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,12(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,96(%rbp)
|
|
movq 224+0-128(%rsi),%r8
|
|
shlb $4,%dl
|
|
movq %rax,96-128(%rbp)
|
|
movq 224+8-128(%rsi),%rax
|
|
shlq $60,%r10
|
|
movb %dl,13(%rsp)
|
|
orq %r10,%rbx
|
|
movb %al,%dl
|
|
shrq $4,%rax
|
|
movq %r8,%r10
|
|
shrq $4,%r8
|
|
movq %r9,104(%rbp)
|
|
movq 240+0-128(%rsi),%r9
|
|
shlb $4,%dl
|
|
movq %rbx,104-128(%rbp)
|
|
movq 240+8-128(%rsi),%rbx
|
|
shlq $60,%r10
|
|
movb %dl,14(%rsp)
|
|
orq %r10,%rax
|
|
movb %bl,%dl
|
|
shrq $4,%rbx
|
|
movq %r9,%r10
|
|
shrq $4,%r9
|
|
movq %r8,112(%rbp)
|
|
shlb $4,%dl
|
|
movq %rax,112-128(%rbp)
|
|
shlq $60,%r10
|
|
movb %dl,15(%rsp)
|
|
orq %r10,%rbx
|
|
movq %r9,120(%rbp)
|
|
movq %rbx,120-128(%rbp)
|
|
addq $-128,%rsi
|
|
movq 8(%rdi),%r8
|
|
movq 0(%rdi),%r9
|
|
addq %r14,%r15
|
|
leaq .Lrem_8bit(%rip),%r11
|
|
jmp .Louter_loop
|
|
.p2align 4
|
|
.Louter_loop:
|
|
xorq (%r14),%r9
|
|
movq 8(%r14),%rdx
|
|
leaq 16(%r14),%r14
|
|
xorq %r8,%rdx
|
|
movq %r9,(%rdi)
|
|
movq %rdx,8(%rdi)
|
|
shrq $32,%rdx
|
|
xorq %rax,%rax
|
|
roll $8,%edx
|
|
movb %dl,%al
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
shrl $4,%ebx
|
|
roll $8,%edx
|
|
movq 8(%rsi,%rax,1),%r8
|
|
movq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
shrl $4,%ecx
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r12,2),%r12
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rcx,1),%r13
|
|
shrl $4,%ebx
|
|
shlq $48,%r12
|
|
xorq %r8,%r13
|
|
movq %r9,%r10
|
|
xorq %r12,%r9
|
|
shrq $8,%r8
|
|
movzbq %r13b,%r13
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rcx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rcx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
shrl $4,%ecx
|
|
shlq $48,%r13
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
xorq %r13,%r9
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
movl 8(%rdi),%edx
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r12,2),%r12
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rcx,1),%r13
|
|
shrl $4,%ebx
|
|
shlq $48,%r12
|
|
xorq %r8,%r13
|
|
movq %r9,%r10
|
|
xorq %r12,%r9
|
|
shrq $8,%r8
|
|
movzbq %r13b,%r13
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rcx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rcx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
shrl $4,%ecx
|
|
shlq $48,%r13
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
xorq %r13,%r9
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r12,2),%r12
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rcx,1),%r13
|
|
shrl $4,%ebx
|
|
shlq $48,%r12
|
|
xorq %r8,%r13
|
|
movq %r9,%r10
|
|
xorq %r12,%r9
|
|
shrq $8,%r8
|
|
movzbq %r13b,%r13
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rcx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rcx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
shrl $4,%ecx
|
|
shlq $48,%r13
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
xorq %r13,%r9
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
movl 4(%rdi),%edx
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r12,2),%r12
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rcx,1),%r13
|
|
shrl $4,%ebx
|
|
shlq $48,%r12
|
|
xorq %r8,%r13
|
|
movq %r9,%r10
|
|
xorq %r12,%r9
|
|
shrq $8,%r8
|
|
movzbq %r13b,%r13
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rcx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rcx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
shrl $4,%ecx
|
|
shlq $48,%r13
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
xorq %r13,%r9
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r12,2),%r12
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rcx,1),%r13
|
|
shrl $4,%ebx
|
|
shlq $48,%r12
|
|
xorq %r8,%r13
|
|
movq %r9,%r10
|
|
xorq %r12,%r9
|
|
shrq $8,%r8
|
|
movzbq %r13b,%r13
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rcx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rcx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
shrl $4,%ecx
|
|
shlq $48,%r13
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
xorq %r13,%r9
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
movl 0(%rdi),%edx
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r12,2),%r12
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rcx,1),%r13
|
|
shrl $4,%ebx
|
|
shlq $48,%r12
|
|
xorq %r8,%r13
|
|
movq %r9,%r10
|
|
xorq %r12,%r9
|
|
shrq $8,%r8
|
|
movzbq %r13b,%r13
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rcx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rcx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
shrl $4,%ecx
|
|
shlq $48,%r13
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
xorq %r13,%r9
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r12,2),%r12
|
|
movzbl %dl,%ebx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rcx,1),%r13
|
|
shrl $4,%ebx
|
|
shlq $48,%r12
|
|
xorq %r8,%r13
|
|
movq %r9,%r10
|
|
xorq %r12,%r9
|
|
shrq $8,%r8
|
|
movzbq %r13b,%r13
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rcx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rcx,8),%r9
|
|
roll $8,%edx
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
movb %dl,%al
|
|
xorq %r10,%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
movzbl %dl,%ecx
|
|
shlb $4,%al
|
|
movzbq (%rsp,%rbx,1),%r12
|
|
andl $240,%ecx
|
|
shlq $48,%r13
|
|
xorq %r8,%r12
|
|
movq %r9,%r10
|
|
xorq %r13,%r9
|
|
shrq $8,%r8
|
|
movzbq %r12b,%r12
|
|
movl -4(%rdi),%edx
|
|
shrq $8,%r9
|
|
xorq -128(%rbp,%rbx,8),%r8
|
|
shlq $56,%r10
|
|
xorq (%rbp,%rbx,8),%r9
|
|
movzwq (%r11,%r12,2),%r12
|
|
xorq 8(%rsi,%rax,1),%r8
|
|
xorq (%rsi,%rax,1),%r9
|
|
shlq $48,%r12
|
|
xorq %r10,%r8
|
|
xorq %r12,%r9
|
|
movzbq %r8b,%r13
|
|
shrq $4,%r8
|
|
movq %r9,%r10
|
|
shlb $4,%r13b
|
|
shrq $4,%r9
|
|
xorq 8(%rsi,%rcx,1),%r8
|
|
movzwq (%r11,%r13,2),%r13
|
|
shlq $60,%r10
|
|
xorq (%rsi,%rcx,1),%r9
|
|
xorq %r10,%r8
|
|
shlq $48,%r13
|
|
bswapq %r8
|
|
xorq %r13,%r9
|
|
bswapq %r9
|
|
cmpq %r15,%r14
|
|
jb .Louter_loop
|
|
movq %r8,8(%rdi)
|
|
movq %r9,(%rdi)
|
|
|
|
leaq 280(%rsp),%rsi
|
|
movq 0(%rsi),%r15
|
|
movq 8(%rsi),%r14
|
|
movq 16(%rsi),%r13
|
|
movq 24(%rsi),%r12
|
|
movq 32(%rsi),%rbp
|
|
movq 40(%rsi),%rbx
|
|
leaq 48(%rsi),%rsp
|
|
.Lghash_epilogue:
|
|
movq 8(%rsp),%rdi
|
|
movq 16(%rsp),%rsi
|
|
retq
|
|
.LSEH_end_gcm_ghash_4bit:
|
|
.globl gcm_init_clmul
|
|
.def gcm_init_clmul; .scl 2; .type 32; .endef
|
|
.p2align 4
|
|
gcm_init_clmul:
|
|
movdqu (%rdx),%xmm2
|
|
pshufd $78,%xmm2,%xmm2
|
|
|
|
|
|
pshufd $255,%xmm2,%xmm4
|
|
movdqa %xmm2,%xmm3
|
|
psllq $1,%xmm2
|
|
pxor %xmm5,%xmm5
|
|
psrlq $63,%xmm3
|
|
pcmpgtd %xmm4,%xmm5
|
|
pslldq $8,%xmm3
|
|
por %xmm3,%xmm2
|
|
|
|
|
|
pand .L0x1c2_polynomial(%rip),%xmm5
|
|
pxor %xmm5,%xmm2
|
|
|
|
|
|
movdqa %xmm2,%xmm0
|
|
movdqa %xmm0,%xmm1
|
|
pshufd $78,%xmm0,%xmm3
|
|
pshufd $78,%xmm2,%xmm4
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm2,%xmm4
|
|
.byte 102,15,58,68,194,0
|
|
.byte 102,15,58,68,202,17
|
|
.byte 102,15,58,68,220,0
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm1,%xmm3
|
|
|
|
movdqa %xmm3,%xmm4
|
|
psrldq $8,%xmm3
|
|
pslldq $8,%xmm4
|
|
pxor %xmm3,%xmm1
|
|
pxor %xmm4,%xmm0
|
|
|
|
movdqa %xmm0,%xmm3
|
|
psllq $1,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $5,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $57,%xmm0
|
|
movdqa %xmm0,%xmm4
|
|
pslldq $8,%xmm0
|
|
psrldq $8,%xmm4
|
|
pxor %xmm3,%xmm0
|
|
pxor %xmm4,%xmm1
|
|
|
|
|
|
movdqa %xmm0,%xmm4
|
|
psrlq $5,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
pxor %xmm1,%xmm4
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
movdqu %xmm2,(%rcx)
|
|
movdqu %xmm0,16(%rcx)
|
|
retq
|
|
|
|
.globl gcm_gmult_clmul
|
|
.def gcm_gmult_clmul; .scl 2; .type 32; .endef
|
|
.p2align 4
|
|
gcm_gmult_clmul:
|
|
movdqu (%rcx),%xmm0
|
|
movdqa .Lbswap_mask(%rip),%xmm5
|
|
movdqu (%rdx),%xmm2
|
|
.byte 102,15,56,0,197
|
|
movdqa %xmm0,%xmm1
|
|
pshufd $78,%xmm0,%xmm3
|
|
pshufd $78,%xmm2,%xmm4
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm2,%xmm4
|
|
.byte 102,15,58,68,194,0
|
|
.byte 102,15,58,68,202,17
|
|
.byte 102,15,58,68,220,0
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm1,%xmm3
|
|
|
|
movdqa %xmm3,%xmm4
|
|
psrldq $8,%xmm3
|
|
pslldq $8,%xmm4
|
|
pxor %xmm3,%xmm1
|
|
pxor %xmm4,%xmm0
|
|
|
|
movdqa %xmm0,%xmm3
|
|
psllq $1,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $5,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $57,%xmm0
|
|
movdqa %xmm0,%xmm4
|
|
pslldq $8,%xmm0
|
|
psrldq $8,%xmm4
|
|
pxor %xmm3,%xmm0
|
|
pxor %xmm4,%xmm1
|
|
|
|
|
|
movdqa %xmm0,%xmm4
|
|
psrlq $5,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
pxor %xmm1,%xmm4
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
.byte 102,15,56,0,197
|
|
movdqu %xmm0,(%rcx)
|
|
retq
|
|
|
|
.globl gcm_ghash_clmul
|
|
.def gcm_ghash_clmul; .scl 2; .type 32; .endef
|
|
.p2align 4
|
|
gcm_ghash_clmul:
|
|
.LSEH_begin_gcm_ghash_clmul:
|
|
|
|
.byte 0x48,0x83,0xec,0x58
|
|
.byte 0x0f,0x29,0x34,0x24
|
|
.byte 0x0f,0x29,0x7c,0x24,0x10
|
|
.byte 0x44,0x0f,0x29,0x44,0x24,0x20
|
|
.byte 0x44,0x0f,0x29,0x4c,0x24,0x30
|
|
.byte 0x44,0x0f,0x29,0x54,0x24,0x40
|
|
movdqa .Lbswap_mask(%rip),%xmm5
|
|
|
|
movdqu (%rcx),%xmm0
|
|
movdqu (%rdx),%xmm2
|
|
.byte 102,15,56,0,197
|
|
|
|
subq $16,%r9
|
|
jz .Lodd_tail
|
|
|
|
movdqu 16(%rdx),%xmm8
|
|
|
|
|
|
|
|
|
|
|
|
movdqu (%r8),%xmm3
|
|
movdqu 16(%r8),%xmm6
|
|
.byte 102,15,56,0,221
|
|
.byte 102,15,56,0,245
|
|
pxor %xmm3,%xmm0
|
|
movdqa %xmm6,%xmm7
|
|
pshufd $78,%xmm6,%xmm3
|
|
pshufd $78,%xmm2,%xmm4
|
|
pxor %xmm6,%xmm3
|
|
pxor %xmm2,%xmm4
|
|
.byte 102,15,58,68,242,0
|
|
.byte 102,15,58,68,250,17
|
|
.byte 102,15,58,68,220,0
|
|
pxor %xmm6,%xmm3
|
|
pxor %xmm7,%xmm3
|
|
|
|
movdqa %xmm3,%xmm4
|
|
psrldq $8,%xmm3
|
|
pslldq $8,%xmm4
|
|
pxor %xmm3,%xmm7
|
|
pxor %xmm4,%xmm6
|
|
movdqa %xmm0,%xmm1
|
|
pshufd $78,%xmm0,%xmm3
|
|
pshufd $78,%xmm8,%xmm4
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm8,%xmm4
|
|
|
|
leaq 32(%r8),%r8
|
|
subq $32,%r9
|
|
jbe .Leven_tail
|
|
|
|
.Lmod_loop:
|
|
.byte 102,65,15,58,68,192,0
|
|
.byte 102,65,15,58,68,200,17
|
|
.byte 102,15,58,68,220,0
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm1,%xmm3
|
|
|
|
movdqa %xmm3,%xmm4
|
|
psrldq $8,%xmm3
|
|
pslldq $8,%xmm4
|
|
pxor %xmm3,%xmm1
|
|
pxor %xmm4,%xmm0
|
|
movdqu (%r8),%xmm3
|
|
pxor %xmm6,%xmm0
|
|
pxor %xmm7,%xmm1
|
|
|
|
movdqu 16(%r8),%xmm6
|
|
.byte 102,15,56,0,221
|
|
.byte 102,15,56,0,245
|
|
|
|
movdqa %xmm6,%xmm7
|
|
pshufd $78,%xmm6,%xmm9
|
|
pshufd $78,%xmm2,%xmm10
|
|
pxor %xmm6,%xmm9
|
|
pxor %xmm2,%xmm10
|
|
pxor %xmm3,%xmm1
|
|
|
|
movdqa %xmm0,%xmm3
|
|
psllq $1,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $5,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
.byte 102,15,58,68,242,0
|
|
psllq $57,%xmm0
|
|
movdqa %xmm0,%xmm4
|
|
pslldq $8,%xmm0
|
|
psrldq $8,%xmm4
|
|
pxor %xmm3,%xmm0
|
|
pxor %xmm4,%xmm1
|
|
|
|
.byte 102,15,58,68,250,17
|
|
movdqa %xmm0,%xmm4
|
|
psrlq $5,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
pxor %xmm1,%xmm4
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
|
|
.byte 102,69,15,58,68,202,0
|
|
movdqa %xmm0,%xmm1
|
|
pshufd $78,%xmm0,%xmm3
|
|
pshufd $78,%xmm8,%xmm4
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm8,%xmm4
|
|
|
|
pxor %xmm6,%xmm9
|
|
pxor %xmm7,%xmm9
|
|
movdqa %xmm9,%xmm10
|
|
psrldq $8,%xmm9
|
|
pslldq $8,%xmm10
|
|
pxor %xmm9,%xmm7
|
|
pxor %xmm10,%xmm6
|
|
|
|
leaq 32(%r8),%r8
|
|
subq $32,%r9
|
|
ja .Lmod_loop
|
|
|
|
.Leven_tail:
|
|
.byte 102,65,15,58,68,192,0
|
|
.byte 102,65,15,58,68,200,17
|
|
.byte 102,15,58,68,220,0
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm1,%xmm3
|
|
|
|
movdqa %xmm3,%xmm4
|
|
psrldq $8,%xmm3
|
|
pslldq $8,%xmm4
|
|
pxor %xmm3,%xmm1
|
|
pxor %xmm4,%xmm0
|
|
pxor %xmm6,%xmm0
|
|
pxor %xmm7,%xmm1
|
|
|
|
movdqa %xmm0,%xmm3
|
|
psllq $1,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $5,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $57,%xmm0
|
|
movdqa %xmm0,%xmm4
|
|
pslldq $8,%xmm0
|
|
psrldq $8,%xmm4
|
|
pxor %xmm3,%xmm0
|
|
pxor %xmm4,%xmm1
|
|
|
|
|
|
movdqa %xmm0,%xmm4
|
|
psrlq $5,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
pxor %xmm1,%xmm4
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
testq %r9,%r9
|
|
jnz .Ldone
|
|
|
|
.Lodd_tail:
|
|
movdqu (%r8),%xmm3
|
|
.byte 102,15,56,0,221
|
|
pxor %xmm3,%xmm0
|
|
movdqa %xmm0,%xmm1
|
|
pshufd $78,%xmm0,%xmm3
|
|
pshufd $78,%xmm2,%xmm4
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm2,%xmm4
|
|
.byte 102,15,58,68,194,0
|
|
.byte 102,15,58,68,202,17
|
|
.byte 102,15,58,68,220,0
|
|
pxor %xmm0,%xmm3
|
|
pxor %xmm1,%xmm3
|
|
|
|
movdqa %xmm3,%xmm4
|
|
psrldq $8,%xmm3
|
|
pslldq $8,%xmm4
|
|
pxor %xmm3,%xmm1
|
|
pxor %xmm4,%xmm0
|
|
|
|
movdqa %xmm0,%xmm3
|
|
psllq $1,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $5,%xmm0
|
|
pxor %xmm3,%xmm0
|
|
psllq $57,%xmm0
|
|
movdqa %xmm0,%xmm4
|
|
pslldq $8,%xmm0
|
|
psrldq $8,%xmm4
|
|
pxor %xmm3,%xmm0
|
|
pxor %xmm4,%xmm1
|
|
|
|
|
|
movdqa %xmm0,%xmm4
|
|
psrlq $5,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
pxor %xmm1,%xmm4
|
|
psrlq $1,%xmm0
|
|
pxor %xmm4,%xmm0
|
|
.Ldone:
|
|
.byte 102,15,56,0,197
|
|
movdqu %xmm0,(%rcx)
|
|
movaps (%rsp),%xmm6
|
|
movaps 16(%rsp),%xmm7
|
|
movaps 32(%rsp),%xmm8
|
|
movaps 48(%rsp),%xmm9
|
|
movaps 64(%rsp),%xmm10
|
|
addq $88,%rsp
|
|
retq
|
|
.LSEH_end_gcm_ghash_clmul:
|
|
|
|
.p2align 6
|
|
.Lbswap_mask:
|
|
.byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
|
|
.L0x1c2_polynomial:
|
|
.byte 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2
|
|
.p2align 6
|
|
|
|
.Lrem_4bit:
|
|
.long 0,0,0,471859200,0,943718400,0,610271232
|
|
.long 0,1887436800,0,1822425088,0,1220542464,0,1423966208
|
|
.long 0,3774873600,0,4246732800,0,3644850176,0,3311403008
|
|
.long 0,2441084928,0,2376073216,0,2847932416,0,3051356160
|
|
|
|
.Lrem_8bit:
|
|
.value 0x0000,0x01C2,0x0384,0x0246,0x0708,0x06CA,0x048C,0x054E
|
|
.value 0x0E10,0x0FD2,0x0D94,0x0C56,0x0918,0x08DA,0x0A9C,0x0B5E
|
|
.value 0x1C20,0x1DE2,0x1FA4,0x1E66,0x1B28,0x1AEA,0x18AC,0x196E
|
|
.value 0x1230,0x13F2,0x11B4,0x1076,0x1538,0x14FA,0x16BC,0x177E
|
|
.value 0x3840,0x3982,0x3BC4,0x3A06,0x3F48,0x3E8A,0x3CCC,0x3D0E
|
|
.value 0x3650,0x3792,0x35D4,0x3416,0x3158,0x309A,0x32DC,0x331E
|
|
.value 0x2460,0x25A2,0x27E4,0x2626,0x2368,0x22AA,0x20EC,0x212E
|
|
.value 0x2A70,0x2BB2,0x29F4,0x2836,0x2D78,0x2CBA,0x2EFC,0x2F3E
|
|
.value 0x7080,0x7142,0x7304,0x72C6,0x7788,0x764A,0x740C,0x75CE
|
|
.value 0x7E90,0x7F52,0x7D14,0x7CD6,0x7998,0x785A,0x7A1C,0x7BDE
|
|
.value 0x6CA0,0x6D62,0x6F24,0x6EE6,0x6BA8,0x6A6A,0x682C,0x69EE
|
|
.value 0x62B0,0x6372,0x6134,0x60F6,0x65B8,0x647A,0x663C,0x67FE
|
|
.value 0x48C0,0x4902,0x4B44,0x4A86,0x4FC8,0x4E0A,0x4C4C,0x4D8E
|
|
.value 0x46D0,0x4712,0x4554,0x4496,0x41D8,0x401A,0x425C,0x439E
|
|
.value 0x54E0,0x5522,0x5764,0x56A6,0x53E8,0x522A,0x506C,0x51AE
|
|
.value 0x5AF0,0x5B32,0x5974,0x58B6,0x5DF8,0x5C3A,0x5E7C,0x5FBE
|
|
.value 0xE100,0xE0C2,0xE284,0xE346,0xE608,0xE7CA,0xE58C,0xE44E
|
|
.value 0xEF10,0xEED2,0xEC94,0xED56,0xE818,0xE9DA,0xEB9C,0xEA5E
|
|
.value 0xFD20,0xFCE2,0xFEA4,0xFF66,0xFA28,0xFBEA,0xF9AC,0xF86E
|
|
.value 0xF330,0xF2F2,0xF0B4,0xF176,0xF438,0xF5FA,0xF7BC,0xF67E
|
|
.value 0xD940,0xD882,0xDAC4,0xDB06,0xDE48,0xDF8A,0xDDCC,0xDC0E
|
|
.value 0xD750,0xD692,0xD4D4,0xD516,0xD058,0xD19A,0xD3DC,0xD21E
|
|
.value 0xC560,0xC4A2,0xC6E4,0xC726,0xC268,0xC3AA,0xC1EC,0xC02E
|
|
.value 0xCB70,0xCAB2,0xC8F4,0xC936,0xCC78,0xCDBA,0xCFFC,0xCE3E
|
|
.value 0x9180,0x9042,0x9204,0x93C6,0x9688,0x974A,0x950C,0x94CE
|
|
.value 0x9F90,0x9E52,0x9C14,0x9DD6,0x9898,0x995A,0x9B1C,0x9ADE
|
|
.value 0x8DA0,0x8C62,0x8E24,0x8FE6,0x8AA8,0x8B6A,0x892C,0x88EE
|
|
.value 0x83B0,0x8272,0x8034,0x81F6,0x84B8,0x857A,0x873C,0x86FE
|
|
.value 0xA9C0,0xA802,0xAA44,0xAB86,0xAEC8,0xAF0A,0xAD4C,0xAC8E
|
|
.value 0xA7D0,0xA612,0xA454,0xA596,0xA0D8,0xA11A,0xA35C,0xA29E
|
|
.value 0xB5E0,0xB422,0xB664,0xB7A6,0xB2E8,0xB32A,0xB16C,0xB0AE
|
|
.value 0xBBF0,0xBA32,0xB874,0xB9B6,0xBCF8,0xBD3A,0xBF7C,0xBEBE
|
|
|
|
.byte 71,72,65,83,72,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
|
|
.p2align 6
|
|
|
|
.def se_handler; .scl 3; .type 32; .endef
|
|
.p2align 4
|
|
se_handler:
|
|
pushq %rsi
|
|
pushq %rdi
|
|
pushq %rbx
|
|
pushq %rbp
|
|
pushq %r12
|
|
pushq %r13
|
|
pushq %r14
|
|
pushq %r15
|
|
pushfq
|
|
subq $64,%rsp
|
|
|
|
movq 120(%r8),%rax
|
|
movq 248(%r8),%rbx
|
|
|
|
movq 8(%r9),%rsi
|
|
movq 56(%r9),%r11
|
|
|
|
movl 0(%r11),%r10d
|
|
leaq (%rsi,%r10,1),%r10
|
|
cmpq %r10,%rbx
|
|
jb .Lin_prologue
|
|
|
|
movq 152(%r8),%rax
|
|
|
|
movl 4(%r11),%r10d
|
|
leaq (%rsi,%r10,1),%r10
|
|
cmpq %r10,%rbx
|
|
jae .Lin_prologue
|
|
|
|
leaq 24(%rax),%rax
|
|
|
|
movq -8(%rax),%rbx
|
|
movq -16(%rax),%rbp
|
|
movq -24(%rax),%r12
|
|
movq %rbx,144(%r8)
|
|
movq %rbp,160(%r8)
|
|
movq %r12,216(%r8)
|
|
|
|
.Lin_prologue:
|
|
movq 8(%rax),%rdi
|
|
movq 16(%rax),%rsi
|
|
movq %rax,152(%r8)
|
|
movq %rsi,168(%r8)
|
|
movq %rdi,176(%r8)
|
|
|
|
movq 40(%r9),%rdi
|
|
movq %r8,%rsi
|
|
movl $154,%ecx
|
|
.long 0xa548f3fc
|
|
|
|
movq %r9,%rsi
|
|
xorq %rcx,%rcx
|
|
movq 8(%rsi),%rdx
|
|
movq 0(%rsi),%r8
|
|
movq 16(%rsi),%r9
|
|
movq 40(%rsi),%r10
|
|
leaq 56(%rsi),%r11
|
|
leaq 24(%rsi),%r12
|
|
movq %r10,32(%rsp)
|
|
movq %r11,40(%rsp)
|
|
movq %r12,48(%rsp)
|
|
movq %rcx,56(%rsp)
|
|
call *__imp_RtlVirtualUnwind(%rip)
|
|
|
|
movl $1,%eax
|
|
addq $64,%rsp
|
|
popfq
|
|
popq %r15
|
|
popq %r14
|
|
popq %r13
|
|
popq %r12
|
|
popq %rbp
|
|
popq %rbx
|
|
popq %rdi
|
|
popq %rsi
|
|
retq
|
|
|
|
|
|
.section .pdata
|
|
.p2align 2
|
|
.rva .LSEH_begin_gcm_gmult_4bit
|
|
.rva .LSEH_end_gcm_gmult_4bit
|
|
.rva .LSEH_info_gcm_gmult_4bit
|
|
|
|
.rva .LSEH_begin_gcm_ghash_4bit
|
|
.rva .LSEH_end_gcm_ghash_4bit
|
|
.rva .LSEH_info_gcm_ghash_4bit
|
|
|
|
.rva .LSEH_begin_gcm_ghash_clmul
|
|
.rva .LSEH_end_gcm_ghash_clmul
|
|
.rva .LSEH_info_gcm_ghash_clmul
|
|
|
|
.section .xdata
|
|
.p2align 3
|
|
.LSEH_info_gcm_gmult_4bit:
|
|
.byte 9,0,0,0
|
|
.rva se_handler
|
|
.rva .Lgmult_prologue,.Lgmult_epilogue
|
|
.LSEH_info_gcm_ghash_4bit:
|
|
.byte 9,0,0,0
|
|
.rva se_handler
|
|
.rva .Lghash_prologue,.Lghash_epilogue
|
|
.LSEH_info_gcm_ghash_clmul:
|
|
.byte 0x01,0x1f,0x0b,0x00
|
|
.byte 0x1f,0xa8,0x04,0x00
|
|
.byte 0x19,0x98,0x03,0x00
|
|
.byte 0x13,0x88,0x02,0x00
|
|
.byte 0x0d,0x78,0x01,0x00
|
|
.byte 0x08,0x68,0x00,0x00
|
|
.byte 0x04,0xa2,0x00,0x00
|