| // This file is generated from a similarly-named Perl script in the BoringSSL |
| // source tree. Do not edit by hand. |
| |
| #if defined(__has_feature) |
| #if __has_feature(memory_sanitizer) && !defined(OPENSSL_NO_ASM) |
| #define OPENSSL_NO_ASM |
| #endif |
| #endif |
| |
| #if defined(__x86_64__) && !defined(OPENSSL_NO_ASM) |
| #if defined(BORINGSSL_PREFIX) |
| #include <boringssl_prefix_symbols_asm.h> |
| #endif |
| .text |
| |
| .extern OPENSSL_ia32cap_P |
| .hidden OPENSSL_ia32cap_P |
| .globl sha256_block_data_order |
| .hidden sha256_block_data_order |
| .type sha256_block_data_order,@function |
| .align 16 |
| sha256_block_data_order: |
| .cfi_startproc |
| leaq OPENSSL_ia32cap_P(%rip),%r11 |
| movl 0(%r11),%r9d |
| movl 4(%r11),%r10d |
| movl 8(%r11),%r11d |
| testl $536870912,%r11d |
| jnz .Lshaext_shortcut |
| andl $1073741824,%r9d |
| andl $268435968,%r10d |
| orl %r9d,%r10d |
| cmpl $1342177792,%r10d |
| je .Lavx_shortcut |
| testl $512,%r10d |
| jnz .Lssse3_shortcut |
| movq %rsp,%rax |
| .cfi_def_cfa_register %rax |
| pushq %rbx |
| .cfi_offset %rbx,-16 |
| pushq %rbp |
| .cfi_offset %rbp,-24 |
| pushq %r12 |
| .cfi_offset %r12,-32 |
| pushq %r13 |
| .cfi_offset %r13,-40 |
| pushq %r14 |
| .cfi_offset %r14,-48 |
| pushq %r15 |
| .cfi_offset %r15,-56 |
| shlq $4,%rdx |
| subq $64+32,%rsp |
| leaq (%rsi,%rdx,4),%rdx |
| andq $-64,%rsp |
| movq %rdi,64+0(%rsp) |
| movq %rsi,64+8(%rsp) |
| movq %rdx,64+16(%rsp) |
| movq %rax,88(%rsp) |
| .cfi_escape 0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08 |
| .Lprologue: |
| |
| movl 0(%rdi),%eax |
| movl 4(%rdi),%ebx |
| movl 8(%rdi),%ecx |
| movl 12(%rdi),%edx |
| movl 16(%rdi),%r8d |
| movl 20(%rdi),%r9d |
| movl 24(%rdi),%r10d |
| movl 28(%rdi),%r11d |
| jmp .Lloop |
| |
| .align 16 |
| .Lloop: |
| movl %ebx,%edi |
| leaq K256(%rip),%rbp |
| xorl %ecx,%edi |
| movl 0(%rsi),%r12d |
| movl %r8d,%r13d |
| movl %eax,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r9d,%r15d |
| |
| xorl %r8d,%r13d |
| rorl $9,%r14d |
| xorl %r10d,%r15d |
| |
| movl %r12d,0(%rsp) |
| xorl %eax,%r14d |
| andl %r8d,%r15d |
| |
| rorl $5,%r13d |
| addl %r11d,%r12d |
| xorl %r10d,%r15d |
| |
| rorl $11,%r14d |
| xorl %r8d,%r13d |
| addl %r15d,%r12d |
| |
| movl %eax,%r15d |
| addl (%rbp),%r12d |
| xorl %eax,%r14d |
| |
| xorl %ebx,%r15d |
| rorl $6,%r13d |
| movl %ebx,%r11d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r11d |
| addl %r12d,%edx |
| addl %r12d,%r11d |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%r11d |
| movl 4(%rsi),%r12d |
| movl %edx,%r13d |
| movl %r11d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r8d,%edi |
| |
| xorl %edx,%r13d |
| rorl $9,%r14d |
| xorl %r9d,%edi |
| |
| movl %r12d,4(%rsp) |
| xorl %r11d,%r14d |
| andl %edx,%edi |
| |
| rorl $5,%r13d |
| addl %r10d,%r12d |
| xorl %r9d,%edi |
| |
| rorl $11,%r14d |
| xorl %edx,%r13d |
| addl %edi,%r12d |
| |
| movl %r11d,%edi |
| addl (%rbp),%r12d |
| xorl %r11d,%r14d |
| |
| xorl %eax,%edi |
| rorl $6,%r13d |
| movl %eax,%r10d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r10d |
| addl %r12d,%ecx |
| addl %r12d,%r10d |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%r10d |
| movl 8(%rsi),%r12d |
| movl %ecx,%r13d |
| movl %r10d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %edx,%r15d |
| |
| xorl %ecx,%r13d |
| rorl $9,%r14d |
| xorl %r8d,%r15d |
| |
| movl %r12d,8(%rsp) |
| xorl %r10d,%r14d |
| andl %ecx,%r15d |
| |
| rorl $5,%r13d |
| addl %r9d,%r12d |
| xorl %r8d,%r15d |
| |
| rorl $11,%r14d |
| xorl %ecx,%r13d |
| addl %r15d,%r12d |
| |
| movl %r10d,%r15d |
| addl (%rbp),%r12d |
| xorl %r10d,%r14d |
| |
| xorl %r11d,%r15d |
| rorl $6,%r13d |
| movl %r11d,%r9d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r9d |
| addl %r12d,%ebx |
| addl %r12d,%r9d |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%r9d |
| movl 12(%rsi),%r12d |
| movl %ebx,%r13d |
| movl %r9d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %ecx,%edi |
| |
| xorl %ebx,%r13d |
| rorl $9,%r14d |
| xorl %edx,%edi |
| |
| movl %r12d,12(%rsp) |
| xorl %r9d,%r14d |
| andl %ebx,%edi |
| |
| rorl $5,%r13d |
| addl %r8d,%r12d |
| xorl %edx,%edi |
| |
| rorl $11,%r14d |
| xorl %ebx,%r13d |
| addl %edi,%r12d |
| |
| movl %r9d,%edi |
| addl (%rbp),%r12d |
| xorl %r9d,%r14d |
| |
| xorl %r10d,%edi |
| rorl $6,%r13d |
| movl %r10d,%r8d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r8d |
| addl %r12d,%eax |
| addl %r12d,%r8d |
| |
| leaq 20(%rbp),%rbp |
| addl %r14d,%r8d |
| movl 16(%rsi),%r12d |
| movl %eax,%r13d |
| movl %r8d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %ebx,%r15d |
| |
| xorl %eax,%r13d |
| rorl $9,%r14d |
| xorl %ecx,%r15d |
| |
| movl %r12d,16(%rsp) |
| xorl %r8d,%r14d |
| andl %eax,%r15d |
| |
| rorl $5,%r13d |
| addl %edx,%r12d |
| xorl %ecx,%r15d |
| |
| rorl $11,%r14d |
| xorl %eax,%r13d |
| addl %r15d,%r12d |
| |
| movl %r8d,%r15d |
| addl (%rbp),%r12d |
| xorl %r8d,%r14d |
| |
| xorl %r9d,%r15d |
| rorl $6,%r13d |
| movl %r9d,%edx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%edx |
| addl %r12d,%r11d |
| addl %r12d,%edx |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%edx |
| movl 20(%rsi),%r12d |
| movl %r11d,%r13d |
| movl %edx,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %eax,%edi |
| |
| xorl %r11d,%r13d |
| rorl $9,%r14d |
| xorl %ebx,%edi |
| |
| movl %r12d,20(%rsp) |
| xorl %edx,%r14d |
| andl %r11d,%edi |
| |
| rorl $5,%r13d |
| addl %ecx,%r12d |
| xorl %ebx,%edi |
| |
| rorl $11,%r14d |
| xorl %r11d,%r13d |
| addl %edi,%r12d |
| |
| movl %edx,%edi |
| addl (%rbp),%r12d |
| xorl %edx,%r14d |
| |
| xorl %r8d,%edi |
| rorl $6,%r13d |
| movl %r8d,%ecx |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%ecx |
| addl %r12d,%r10d |
| addl %r12d,%ecx |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%ecx |
| movl 24(%rsi),%r12d |
| movl %r10d,%r13d |
| movl %ecx,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r11d,%r15d |
| |
| xorl %r10d,%r13d |
| rorl $9,%r14d |
| xorl %eax,%r15d |
| |
| movl %r12d,24(%rsp) |
| xorl %ecx,%r14d |
| andl %r10d,%r15d |
| |
| rorl $5,%r13d |
| addl %ebx,%r12d |
| xorl %eax,%r15d |
| |
| rorl $11,%r14d |
| xorl %r10d,%r13d |
| addl %r15d,%r12d |
| |
| movl %ecx,%r15d |
| addl (%rbp),%r12d |
| xorl %ecx,%r14d |
| |
| xorl %edx,%r15d |
| rorl $6,%r13d |
| movl %edx,%ebx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%ebx |
| addl %r12d,%r9d |
| addl %r12d,%ebx |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%ebx |
| movl 28(%rsi),%r12d |
| movl %r9d,%r13d |
| movl %ebx,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r10d,%edi |
| |
| xorl %r9d,%r13d |
| rorl $9,%r14d |
| xorl %r11d,%edi |
| |
| movl %r12d,28(%rsp) |
| xorl %ebx,%r14d |
| andl %r9d,%edi |
| |
| rorl $5,%r13d |
| addl %eax,%r12d |
| xorl %r11d,%edi |
| |
| rorl $11,%r14d |
| xorl %r9d,%r13d |
| addl %edi,%r12d |
| |
| movl %ebx,%edi |
| addl (%rbp),%r12d |
| xorl %ebx,%r14d |
| |
| xorl %ecx,%edi |
| rorl $6,%r13d |
| movl %ecx,%eax |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%eax |
| addl %r12d,%r8d |
| addl %r12d,%eax |
| |
| leaq 20(%rbp),%rbp |
| addl %r14d,%eax |
| movl 32(%rsi),%r12d |
| movl %r8d,%r13d |
| movl %eax,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r9d,%r15d |
| |
| xorl %r8d,%r13d |
| rorl $9,%r14d |
| xorl %r10d,%r15d |
| |
| movl %r12d,32(%rsp) |
| xorl %eax,%r14d |
| andl %r8d,%r15d |
| |
| rorl $5,%r13d |
| addl %r11d,%r12d |
| xorl %r10d,%r15d |
| |
| rorl $11,%r14d |
| xorl %r8d,%r13d |
| addl %r15d,%r12d |
| |
| movl %eax,%r15d |
| addl (%rbp),%r12d |
| xorl %eax,%r14d |
| |
| xorl %ebx,%r15d |
| rorl $6,%r13d |
| movl %ebx,%r11d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r11d |
| addl %r12d,%edx |
| addl %r12d,%r11d |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%r11d |
| movl 36(%rsi),%r12d |
| movl %edx,%r13d |
| movl %r11d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r8d,%edi |
| |
| xorl %edx,%r13d |
| rorl $9,%r14d |
| xorl %r9d,%edi |
| |
| movl %r12d,36(%rsp) |
| xorl %r11d,%r14d |
| andl %edx,%edi |
| |
| rorl $5,%r13d |
| addl %r10d,%r12d |
| xorl %r9d,%edi |
| |
| rorl $11,%r14d |
| xorl %edx,%r13d |
| addl %edi,%r12d |
| |
| movl %r11d,%edi |
| addl (%rbp),%r12d |
| xorl %r11d,%r14d |
| |
| xorl %eax,%edi |
| rorl $6,%r13d |
| movl %eax,%r10d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r10d |
| addl %r12d,%ecx |
| addl %r12d,%r10d |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%r10d |
| movl 40(%rsi),%r12d |
| movl %ecx,%r13d |
| movl %r10d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %edx,%r15d |
| |
| xorl %ecx,%r13d |
| rorl $9,%r14d |
| xorl %r8d,%r15d |
| |
| movl %r12d,40(%rsp) |
| xorl %r10d,%r14d |
| andl %ecx,%r15d |
| |
| rorl $5,%r13d |
| addl %r9d,%r12d |
| xorl %r8d,%r15d |
| |
| rorl $11,%r14d |
| xorl %ecx,%r13d |
| addl %r15d,%r12d |
| |
| movl %r10d,%r15d |
| addl (%rbp),%r12d |
| xorl %r10d,%r14d |
| |
| xorl %r11d,%r15d |
| rorl $6,%r13d |
| movl %r11d,%r9d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r9d |
| addl %r12d,%ebx |
| addl %r12d,%r9d |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%r9d |
| movl 44(%rsi),%r12d |
| movl %ebx,%r13d |
| movl %r9d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %ecx,%edi |
| |
| xorl %ebx,%r13d |
| rorl $9,%r14d |
| xorl %edx,%edi |
| |
| movl %r12d,44(%rsp) |
| xorl %r9d,%r14d |
| andl %ebx,%edi |
| |
| rorl $5,%r13d |
| addl %r8d,%r12d |
| xorl %edx,%edi |
| |
| rorl $11,%r14d |
| xorl %ebx,%r13d |
| addl %edi,%r12d |
| |
| movl %r9d,%edi |
| addl (%rbp),%r12d |
| xorl %r9d,%r14d |
| |
| xorl %r10d,%edi |
| rorl $6,%r13d |
| movl %r10d,%r8d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r8d |
| addl %r12d,%eax |
| addl %r12d,%r8d |
| |
| leaq 20(%rbp),%rbp |
| addl %r14d,%r8d |
| movl 48(%rsi),%r12d |
| movl %eax,%r13d |
| movl %r8d,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %ebx,%r15d |
| |
| xorl %eax,%r13d |
| rorl $9,%r14d |
| xorl %ecx,%r15d |
| |
| movl %r12d,48(%rsp) |
| xorl %r8d,%r14d |
| andl %eax,%r15d |
| |
| rorl $5,%r13d |
| addl %edx,%r12d |
| xorl %ecx,%r15d |
| |
| rorl $11,%r14d |
| xorl %eax,%r13d |
| addl %r15d,%r12d |
| |
| movl %r8d,%r15d |
| addl (%rbp),%r12d |
| xorl %r8d,%r14d |
| |
| xorl %r9d,%r15d |
| rorl $6,%r13d |
| movl %r9d,%edx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%edx |
| addl %r12d,%r11d |
| addl %r12d,%edx |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%edx |
| movl 52(%rsi),%r12d |
| movl %r11d,%r13d |
| movl %edx,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %eax,%edi |
| |
| xorl %r11d,%r13d |
| rorl $9,%r14d |
| xorl %ebx,%edi |
| |
| movl %r12d,52(%rsp) |
| xorl %edx,%r14d |
| andl %r11d,%edi |
| |
| rorl $5,%r13d |
| addl %ecx,%r12d |
| xorl %ebx,%edi |
| |
| rorl $11,%r14d |
| xorl %r11d,%r13d |
| addl %edi,%r12d |
| |
| movl %edx,%edi |
| addl (%rbp),%r12d |
| xorl %edx,%r14d |
| |
| xorl %r8d,%edi |
| rorl $6,%r13d |
| movl %r8d,%ecx |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%ecx |
| addl %r12d,%r10d |
| addl %r12d,%ecx |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%ecx |
| movl 56(%rsi),%r12d |
| movl %r10d,%r13d |
| movl %ecx,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r11d,%r15d |
| |
| xorl %r10d,%r13d |
| rorl $9,%r14d |
| xorl %eax,%r15d |
| |
| movl %r12d,56(%rsp) |
| xorl %ecx,%r14d |
| andl %r10d,%r15d |
| |
| rorl $5,%r13d |
| addl %ebx,%r12d |
| xorl %eax,%r15d |
| |
| rorl $11,%r14d |
| xorl %r10d,%r13d |
| addl %r15d,%r12d |
| |
| movl %ecx,%r15d |
| addl (%rbp),%r12d |
| xorl %ecx,%r14d |
| |
| xorl %edx,%r15d |
| rorl $6,%r13d |
| movl %edx,%ebx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%ebx |
| addl %r12d,%r9d |
| addl %r12d,%ebx |
| |
| leaq 4(%rbp),%rbp |
| addl %r14d,%ebx |
| movl 60(%rsi),%r12d |
| movl %r9d,%r13d |
| movl %ebx,%r14d |
| bswapl %r12d |
| rorl $14,%r13d |
| movl %r10d,%edi |
| |
| xorl %r9d,%r13d |
| rorl $9,%r14d |
| xorl %r11d,%edi |
| |
| movl %r12d,60(%rsp) |
| xorl %ebx,%r14d |
| andl %r9d,%edi |
| |
| rorl $5,%r13d |
| addl %eax,%r12d |
| xorl %r11d,%edi |
| |
| rorl $11,%r14d |
| xorl %r9d,%r13d |
| addl %edi,%r12d |
| |
| movl %ebx,%edi |
| addl (%rbp),%r12d |
| xorl %ebx,%r14d |
| |
| xorl %ecx,%edi |
| rorl $6,%r13d |
| movl %ecx,%eax |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%eax |
| addl %r12d,%r8d |
| addl %r12d,%eax |
| |
| leaq 20(%rbp),%rbp |
| jmp .Lrounds_16_xx |
| .align 16 |
| .Lrounds_16_xx: |
| movl 4(%rsp),%r13d |
| movl 56(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%eax |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 36(%rsp),%r12d |
| |
| addl 0(%rsp),%r12d |
| movl %r8d,%r13d |
| addl %r15d,%r12d |
| movl %eax,%r14d |
| rorl $14,%r13d |
| movl %r9d,%r15d |
| |
| xorl %r8d,%r13d |
| rorl $9,%r14d |
| xorl %r10d,%r15d |
| |
| movl %r12d,0(%rsp) |
| xorl %eax,%r14d |
| andl %r8d,%r15d |
| |
| rorl $5,%r13d |
| addl %r11d,%r12d |
| xorl %r10d,%r15d |
| |
| rorl $11,%r14d |
| xorl %r8d,%r13d |
| addl %r15d,%r12d |
| |
| movl %eax,%r15d |
| addl (%rbp),%r12d |
| xorl %eax,%r14d |
| |
| xorl %ebx,%r15d |
| rorl $6,%r13d |
| movl %ebx,%r11d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r11d |
| addl %r12d,%edx |
| addl %r12d,%r11d |
| |
| leaq 4(%rbp),%rbp |
| movl 8(%rsp),%r13d |
| movl 60(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r11d |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 40(%rsp),%r12d |
| |
| addl 4(%rsp),%r12d |
| movl %edx,%r13d |
| addl %edi,%r12d |
| movl %r11d,%r14d |
| rorl $14,%r13d |
| movl %r8d,%edi |
| |
| xorl %edx,%r13d |
| rorl $9,%r14d |
| xorl %r9d,%edi |
| |
| movl %r12d,4(%rsp) |
| xorl %r11d,%r14d |
| andl %edx,%edi |
| |
| rorl $5,%r13d |
| addl %r10d,%r12d |
| xorl %r9d,%edi |
| |
| rorl $11,%r14d |
| xorl %edx,%r13d |
| addl %edi,%r12d |
| |
| movl %r11d,%edi |
| addl (%rbp),%r12d |
| xorl %r11d,%r14d |
| |
| xorl %eax,%edi |
| rorl $6,%r13d |
| movl %eax,%r10d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r10d |
| addl %r12d,%ecx |
| addl %r12d,%r10d |
| |
| leaq 4(%rbp),%rbp |
| movl 12(%rsp),%r13d |
| movl 0(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r10d |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 44(%rsp),%r12d |
| |
| addl 8(%rsp),%r12d |
| movl %ecx,%r13d |
| addl %r15d,%r12d |
| movl %r10d,%r14d |
| rorl $14,%r13d |
| movl %edx,%r15d |
| |
| xorl %ecx,%r13d |
| rorl $9,%r14d |
| xorl %r8d,%r15d |
| |
| movl %r12d,8(%rsp) |
| xorl %r10d,%r14d |
| andl %ecx,%r15d |
| |
| rorl $5,%r13d |
| addl %r9d,%r12d |
| xorl %r8d,%r15d |
| |
| rorl $11,%r14d |
| xorl %ecx,%r13d |
| addl %r15d,%r12d |
| |
| movl %r10d,%r15d |
| addl (%rbp),%r12d |
| xorl %r10d,%r14d |
| |
| xorl %r11d,%r15d |
| rorl $6,%r13d |
| movl %r11d,%r9d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r9d |
| addl %r12d,%ebx |
| addl %r12d,%r9d |
| |
| leaq 4(%rbp),%rbp |
| movl 16(%rsp),%r13d |
| movl 4(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r9d |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 48(%rsp),%r12d |
| |
| addl 12(%rsp),%r12d |
| movl %ebx,%r13d |
| addl %edi,%r12d |
| movl %r9d,%r14d |
| rorl $14,%r13d |
| movl %ecx,%edi |
| |
| xorl %ebx,%r13d |
| rorl $9,%r14d |
| xorl %edx,%edi |
| |
| movl %r12d,12(%rsp) |
| xorl %r9d,%r14d |
| andl %ebx,%edi |
| |
| rorl $5,%r13d |
| addl %r8d,%r12d |
| xorl %edx,%edi |
| |
| rorl $11,%r14d |
| xorl %ebx,%r13d |
| addl %edi,%r12d |
| |
| movl %r9d,%edi |
| addl (%rbp),%r12d |
| xorl %r9d,%r14d |
| |
| xorl %r10d,%edi |
| rorl $6,%r13d |
| movl %r10d,%r8d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r8d |
| addl %r12d,%eax |
| addl %r12d,%r8d |
| |
| leaq 20(%rbp),%rbp |
| movl 20(%rsp),%r13d |
| movl 8(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r8d |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 52(%rsp),%r12d |
| |
| addl 16(%rsp),%r12d |
| movl %eax,%r13d |
| addl %r15d,%r12d |
| movl %r8d,%r14d |
| rorl $14,%r13d |
| movl %ebx,%r15d |
| |
| xorl %eax,%r13d |
| rorl $9,%r14d |
| xorl %ecx,%r15d |
| |
| movl %r12d,16(%rsp) |
| xorl %r8d,%r14d |
| andl %eax,%r15d |
| |
| rorl $5,%r13d |
| addl %edx,%r12d |
| xorl %ecx,%r15d |
| |
| rorl $11,%r14d |
| xorl %eax,%r13d |
| addl %r15d,%r12d |
| |
| movl %r8d,%r15d |
| addl (%rbp),%r12d |
| xorl %r8d,%r14d |
| |
| xorl %r9d,%r15d |
| rorl $6,%r13d |
| movl %r9d,%edx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%edx |
| addl %r12d,%r11d |
| addl %r12d,%edx |
| |
| leaq 4(%rbp),%rbp |
| movl 24(%rsp),%r13d |
| movl 12(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%edx |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 56(%rsp),%r12d |
| |
| addl 20(%rsp),%r12d |
| movl %r11d,%r13d |
| addl %edi,%r12d |
| movl %edx,%r14d |
| rorl $14,%r13d |
| movl %eax,%edi |
| |
| xorl %r11d,%r13d |
| rorl $9,%r14d |
| xorl %ebx,%edi |
| |
| movl %r12d,20(%rsp) |
| xorl %edx,%r14d |
| andl %r11d,%edi |
| |
| rorl $5,%r13d |
| addl %ecx,%r12d |
| xorl %ebx,%edi |
| |
| rorl $11,%r14d |
| xorl %r11d,%r13d |
| addl %edi,%r12d |
| |
| movl %edx,%edi |
| addl (%rbp),%r12d |
| xorl %edx,%r14d |
| |
| xorl %r8d,%edi |
| rorl $6,%r13d |
| movl %r8d,%ecx |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%ecx |
| addl %r12d,%r10d |
| addl %r12d,%ecx |
| |
| leaq 4(%rbp),%rbp |
| movl 28(%rsp),%r13d |
| movl 16(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%ecx |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 60(%rsp),%r12d |
| |
| addl 24(%rsp),%r12d |
| movl %r10d,%r13d |
| addl %r15d,%r12d |
| movl %ecx,%r14d |
| rorl $14,%r13d |
| movl %r11d,%r15d |
| |
| xorl %r10d,%r13d |
| rorl $9,%r14d |
| xorl %eax,%r15d |
| |
| movl %r12d,24(%rsp) |
| xorl %ecx,%r14d |
| andl %r10d,%r15d |
| |
| rorl $5,%r13d |
| addl %ebx,%r12d |
| xorl %eax,%r15d |
| |
| rorl $11,%r14d |
| xorl %r10d,%r13d |
| addl %r15d,%r12d |
| |
| movl %ecx,%r15d |
| addl (%rbp),%r12d |
| xorl %ecx,%r14d |
| |
| xorl %edx,%r15d |
| rorl $6,%r13d |
| movl %edx,%ebx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%ebx |
| addl %r12d,%r9d |
| addl %r12d,%ebx |
| |
| leaq 4(%rbp),%rbp |
| movl 32(%rsp),%r13d |
| movl 20(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%ebx |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 0(%rsp),%r12d |
| |
| addl 28(%rsp),%r12d |
| movl %r9d,%r13d |
| addl %edi,%r12d |
| movl %ebx,%r14d |
| rorl $14,%r13d |
| movl %r10d,%edi |
| |
| xorl %r9d,%r13d |
| rorl $9,%r14d |
| xorl %r11d,%edi |
| |
| movl %r12d,28(%rsp) |
| xorl %ebx,%r14d |
| andl %r9d,%edi |
| |
| rorl $5,%r13d |
| addl %eax,%r12d |
| xorl %r11d,%edi |
| |
| rorl $11,%r14d |
| xorl %r9d,%r13d |
| addl %edi,%r12d |
| |
| movl %ebx,%edi |
| addl (%rbp),%r12d |
| xorl %ebx,%r14d |
| |
| xorl %ecx,%edi |
| rorl $6,%r13d |
| movl %ecx,%eax |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%eax |
| addl %r12d,%r8d |
| addl %r12d,%eax |
| |
| leaq 20(%rbp),%rbp |
| movl 36(%rsp),%r13d |
| movl 24(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%eax |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 4(%rsp),%r12d |
| |
| addl 32(%rsp),%r12d |
| movl %r8d,%r13d |
| addl %r15d,%r12d |
| movl %eax,%r14d |
| rorl $14,%r13d |
| movl %r9d,%r15d |
| |
| xorl %r8d,%r13d |
| rorl $9,%r14d |
| xorl %r10d,%r15d |
| |
| movl %r12d,32(%rsp) |
| xorl %eax,%r14d |
| andl %r8d,%r15d |
| |
| rorl $5,%r13d |
| addl %r11d,%r12d |
| xorl %r10d,%r15d |
| |
| rorl $11,%r14d |
| xorl %r8d,%r13d |
| addl %r15d,%r12d |
| |
| movl %eax,%r15d |
| addl (%rbp),%r12d |
| xorl %eax,%r14d |
| |
| xorl %ebx,%r15d |
| rorl $6,%r13d |
| movl %ebx,%r11d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r11d |
| addl %r12d,%edx |
| addl %r12d,%r11d |
| |
| leaq 4(%rbp),%rbp |
| movl 40(%rsp),%r13d |
| movl 28(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r11d |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 8(%rsp),%r12d |
| |
| addl 36(%rsp),%r12d |
| movl %edx,%r13d |
| addl %edi,%r12d |
| movl %r11d,%r14d |
| rorl $14,%r13d |
| movl %r8d,%edi |
| |
| xorl %edx,%r13d |
| rorl $9,%r14d |
| xorl %r9d,%edi |
| |
| movl %r12d,36(%rsp) |
| xorl %r11d,%r14d |
| andl %edx,%edi |
| |
| rorl $5,%r13d |
| addl %r10d,%r12d |
| xorl %r9d,%edi |
| |
| rorl $11,%r14d |
| xorl %edx,%r13d |
| addl %edi,%r12d |
| |
| movl %r11d,%edi |
| addl (%rbp),%r12d |
| xorl %r11d,%r14d |
| |
| xorl %eax,%edi |
| rorl $6,%r13d |
| movl %eax,%r10d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r10d |
| addl %r12d,%ecx |
| addl %r12d,%r10d |
| |
| leaq 4(%rbp),%rbp |
| movl 44(%rsp),%r13d |
| movl 32(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r10d |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 12(%rsp),%r12d |
| |
| addl 40(%rsp),%r12d |
| movl %ecx,%r13d |
| addl %r15d,%r12d |
| movl %r10d,%r14d |
| rorl $14,%r13d |
| movl %edx,%r15d |
| |
| xorl %ecx,%r13d |
| rorl $9,%r14d |
| xorl %r8d,%r15d |
| |
| movl %r12d,40(%rsp) |
| xorl %r10d,%r14d |
| andl %ecx,%r15d |
| |
| rorl $5,%r13d |
| addl %r9d,%r12d |
| xorl %r8d,%r15d |
| |
| rorl $11,%r14d |
| xorl %ecx,%r13d |
| addl %r15d,%r12d |
| |
| movl %r10d,%r15d |
| addl (%rbp),%r12d |
| xorl %r10d,%r14d |
| |
| xorl %r11d,%r15d |
| rorl $6,%r13d |
| movl %r11d,%r9d |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%r9d |
| addl %r12d,%ebx |
| addl %r12d,%r9d |
| |
| leaq 4(%rbp),%rbp |
| movl 48(%rsp),%r13d |
| movl 36(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r9d |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 16(%rsp),%r12d |
| |
| addl 44(%rsp),%r12d |
| movl %ebx,%r13d |
| addl %edi,%r12d |
| movl %r9d,%r14d |
| rorl $14,%r13d |
| movl %ecx,%edi |
| |
| xorl %ebx,%r13d |
| rorl $9,%r14d |
| xorl %edx,%edi |
| |
| movl %r12d,44(%rsp) |
| xorl %r9d,%r14d |
| andl %ebx,%edi |
| |
| rorl $5,%r13d |
| addl %r8d,%r12d |
| xorl %edx,%edi |
| |
| rorl $11,%r14d |
| xorl %ebx,%r13d |
| addl %edi,%r12d |
| |
| movl %r9d,%edi |
| addl (%rbp),%r12d |
| xorl %r9d,%r14d |
| |
| xorl %r10d,%edi |
| rorl $6,%r13d |
| movl %r10d,%r8d |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%r8d |
| addl %r12d,%eax |
| addl %r12d,%r8d |
| |
| leaq 20(%rbp),%rbp |
| movl 52(%rsp),%r13d |
| movl 40(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%r8d |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 20(%rsp),%r12d |
| |
| addl 48(%rsp),%r12d |
| movl %eax,%r13d |
| addl %r15d,%r12d |
| movl %r8d,%r14d |
| rorl $14,%r13d |
| movl %ebx,%r15d |
| |
| xorl %eax,%r13d |
| rorl $9,%r14d |
| xorl %ecx,%r15d |
| |
| movl %r12d,48(%rsp) |
| xorl %r8d,%r14d |
| andl %eax,%r15d |
| |
| rorl $5,%r13d |
| addl %edx,%r12d |
| xorl %ecx,%r15d |
| |
| rorl $11,%r14d |
| xorl %eax,%r13d |
| addl %r15d,%r12d |
| |
| movl %r8d,%r15d |
| addl (%rbp),%r12d |
| xorl %r8d,%r14d |
| |
| xorl %r9d,%r15d |
| rorl $6,%r13d |
| movl %r9d,%edx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%edx |
| addl %r12d,%r11d |
| addl %r12d,%edx |
| |
| leaq 4(%rbp),%rbp |
| movl 56(%rsp),%r13d |
| movl 44(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%edx |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 24(%rsp),%r12d |
| |
| addl 52(%rsp),%r12d |
| movl %r11d,%r13d |
| addl %edi,%r12d |
| movl %edx,%r14d |
| rorl $14,%r13d |
| movl %eax,%edi |
| |
| xorl %r11d,%r13d |
| rorl $9,%r14d |
| xorl %ebx,%edi |
| |
| movl %r12d,52(%rsp) |
| xorl %edx,%r14d |
| andl %r11d,%edi |
| |
| rorl $5,%r13d |
| addl %ecx,%r12d |
| xorl %ebx,%edi |
| |
| rorl $11,%r14d |
| xorl %r11d,%r13d |
| addl %edi,%r12d |
| |
| movl %edx,%edi |
| addl (%rbp),%r12d |
| xorl %edx,%r14d |
| |
| xorl %r8d,%edi |
| rorl $6,%r13d |
| movl %r8d,%ecx |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%ecx |
| addl %r12d,%r10d |
| addl %r12d,%ecx |
| |
| leaq 4(%rbp),%rbp |
| movl 60(%rsp),%r13d |
| movl 48(%rsp),%r15d |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%ecx |
| movl %r15d,%r14d |
| rorl $2,%r15d |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%r15d |
| shrl $10,%r14d |
| |
| rorl $17,%r15d |
| xorl %r13d,%r12d |
| xorl %r14d,%r15d |
| addl 28(%rsp),%r12d |
| |
| addl 56(%rsp),%r12d |
| movl %r10d,%r13d |
| addl %r15d,%r12d |
| movl %ecx,%r14d |
| rorl $14,%r13d |
| movl %r11d,%r15d |
| |
| xorl %r10d,%r13d |
| rorl $9,%r14d |
| xorl %eax,%r15d |
| |
| movl %r12d,56(%rsp) |
| xorl %ecx,%r14d |
| andl %r10d,%r15d |
| |
| rorl $5,%r13d |
| addl %ebx,%r12d |
| xorl %eax,%r15d |
| |
| rorl $11,%r14d |
| xorl %r10d,%r13d |
| addl %r15d,%r12d |
| |
| movl %ecx,%r15d |
| addl (%rbp),%r12d |
| xorl %ecx,%r14d |
| |
| xorl %edx,%r15d |
| rorl $6,%r13d |
| movl %edx,%ebx |
| |
| andl %r15d,%edi |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %edi,%ebx |
| addl %r12d,%r9d |
| addl %r12d,%ebx |
| |
| leaq 4(%rbp),%rbp |
| movl 0(%rsp),%r13d |
| movl 52(%rsp),%edi |
| |
| movl %r13d,%r12d |
| rorl $11,%r13d |
| addl %r14d,%ebx |
| movl %edi,%r14d |
| rorl $2,%edi |
| |
| xorl %r12d,%r13d |
| shrl $3,%r12d |
| rorl $7,%r13d |
| xorl %r14d,%edi |
| shrl $10,%r14d |
| |
| rorl $17,%edi |
| xorl %r13d,%r12d |
| xorl %r14d,%edi |
| addl 32(%rsp),%r12d |
| |
| addl 60(%rsp),%r12d |
| movl %r9d,%r13d |
| addl %edi,%r12d |
| movl %ebx,%r14d |
| rorl $14,%r13d |
| movl %r10d,%edi |
| |
| xorl %r9d,%r13d |
| rorl $9,%r14d |
| xorl %r11d,%edi |
| |
| movl %r12d,60(%rsp) |
| xorl %ebx,%r14d |
| andl %r9d,%edi |
| |
| rorl $5,%r13d |
| addl %eax,%r12d |
| xorl %r11d,%edi |
| |
| rorl $11,%r14d |
| xorl %r9d,%r13d |
| addl %edi,%r12d |
| |
| movl %ebx,%edi |
| addl (%rbp),%r12d |
| xorl %ebx,%r14d |
| |
| xorl %ecx,%edi |
| rorl $6,%r13d |
| movl %ecx,%eax |
| |
| andl %edi,%r15d |
| rorl $2,%r14d |
| addl %r13d,%r12d |
| |
| xorl %r15d,%eax |
| addl %r12d,%r8d |
| addl %r12d,%eax |
| |
| leaq 20(%rbp),%rbp |
| cmpb $0,3(%rbp) |
| jnz .Lrounds_16_xx |
| |
| movq 64+0(%rsp),%rdi |
| addl %r14d,%eax |
| leaq 64(%rsi),%rsi |
| |
| addl 0(%rdi),%eax |
| addl 4(%rdi),%ebx |
| addl 8(%rdi),%ecx |
| addl 12(%rdi),%edx |
| addl 16(%rdi),%r8d |
| addl 20(%rdi),%r9d |
| addl 24(%rdi),%r10d |
| addl 28(%rdi),%r11d |
| |
| cmpq 64+16(%rsp),%rsi |
| |
| movl %eax,0(%rdi) |
| movl %ebx,4(%rdi) |
| movl %ecx,8(%rdi) |
| movl %edx,12(%rdi) |
| movl %r8d,16(%rdi) |
| movl %r9d,20(%rdi) |
| movl %r10d,24(%rdi) |
| movl %r11d,28(%rdi) |
| jb .Lloop |
| |
| movq 88(%rsp),%rsi |
| .cfi_def_cfa %rsi,8 |
| movq -48(%rsi),%r15 |
| .cfi_restore %r15 |
| movq -40(%rsi),%r14 |
| .cfi_restore %r14 |
| movq -32(%rsi),%r13 |
| .cfi_restore %r13 |
| movq -24(%rsi),%r12 |
| .cfi_restore %r12 |
| movq -16(%rsi),%rbp |
| .cfi_restore %rbp |
| movq -8(%rsi),%rbx |
| .cfi_restore %rbx |
| leaq (%rsi),%rsp |
| .cfi_def_cfa_register %rsp |
| .Lepilogue: |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size sha256_block_data_order,.-sha256_block_data_order |
| .align 64 |
| .type K256,@object |
| K256: |
| .long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5 |
| .long 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5 |
| .long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5 |
| .long 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5 |
| .long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3 |
| .long 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3 |
| .long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174 |
| .long 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174 |
| .long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc |
| .long 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc |
| .long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da |
| .long 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da |
| .long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7 |
| .long 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7 |
| .long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967 |
| .long 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967 |
| .long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13 |
| .long 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13 |
| .long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85 |
| .long 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85 |
| .long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3 |
| .long 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3 |
| .long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070 |
| .long 0xd192e819,0xd6990624,0xf40e3585,0x106aa070 |
| .long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5 |
| .long 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5 |
| .long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3 |
| .long 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3 |
| .long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208 |
| .long 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208 |
| .long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2 |
| .long 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2 |
| |
| .long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f |
| .long 0x00010203,0x04050607,0x08090a0b,0x0c0d0e0f |
| .long 0x03020100,0x0b0a0908,0xffffffff,0xffffffff |
| .long 0x03020100,0x0b0a0908,0xffffffff,0xffffffff |
| .long 0xffffffff,0xffffffff,0x03020100,0x0b0a0908 |
| .long 0xffffffff,0xffffffff,0x03020100,0x0b0a0908 |
| .byte 83,72,65,50,53,54,32,98,108,111,99,107,32,116,114,97,110,115,102,111,114,109,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |
| .type sha256_block_data_order_shaext,@function |
| .align 64 |
| sha256_block_data_order_shaext: |
| .cfi_startproc |
| .Lshaext_shortcut: |
| leaq K256+128(%rip),%rcx |
| movdqu (%rdi),%xmm1 |
| movdqu 16(%rdi),%xmm2 |
| movdqa 512-128(%rcx),%xmm7 |
| |
| pshufd $0x1b,%xmm1,%xmm0 |
| pshufd $0xb1,%xmm1,%xmm1 |
| pshufd $0x1b,%xmm2,%xmm2 |
| movdqa %xmm7,%xmm8 |
| .byte 102,15,58,15,202,8 |
| punpcklqdq %xmm0,%xmm2 |
| jmp .Loop_shaext |
| |
| .align 16 |
| .Loop_shaext: |
| movdqu (%rsi),%xmm3 |
| movdqu 16(%rsi),%xmm4 |
| movdqu 32(%rsi),%xmm5 |
| .byte 102,15,56,0,223 |
| movdqu 48(%rsi),%xmm6 |
| |
| movdqa 0-128(%rcx),%xmm0 |
| paddd %xmm3,%xmm0 |
| .byte 102,15,56,0,231 |
| movdqa %xmm2,%xmm10 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| nop |
| movdqa %xmm1,%xmm9 |
| .byte 15,56,203,202 |
| |
| movdqa 32-128(%rcx),%xmm0 |
| paddd %xmm4,%xmm0 |
| .byte 102,15,56,0,239 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| leaq 64(%rsi),%rsi |
| .byte 15,56,204,220 |
| .byte 15,56,203,202 |
| |
| movdqa 64-128(%rcx),%xmm0 |
| paddd %xmm5,%xmm0 |
| .byte 102,15,56,0,247 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm6,%xmm7 |
| .byte 102,15,58,15,253,4 |
| nop |
| paddd %xmm7,%xmm3 |
| .byte 15,56,204,229 |
| .byte 15,56,203,202 |
| |
| movdqa 96-128(%rcx),%xmm0 |
| paddd %xmm6,%xmm0 |
| .byte 15,56,205,222 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm3,%xmm7 |
| .byte 102,15,58,15,254,4 |
| nop |
| paddd %xmm7,%xmm4 |
| .byte 15,56,204,238 |
| .byte 15,56,203,202 |
| movdqa 128-128(%rcx),%xmm0 |
| paddd %xmm3,%xmm0 |
| .byte 15,56,205,227 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm4,%xmm7 |
| .byte 102,15,58,15,251,4 |
| nop |
| paddd %xmm7,%xmm5 |
| .byte 15,56,204,243 |
| .byte 15,56,203,202 |
| movdqa 160-128(%rcx),%xmm0 |
| paddd %xmm4,%xmm0 |
| .byte 15,56,205,236 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm5,%xmm7 |
| .byte 102,15,58,15,252,4 |
| nop |
| paddd %xmm7,%xmm6 |
| .byte 15,56,204,220 |
| .byte 15,56,203,202 |
| movdqa 192-128(%rcx),%xmm0 |
| paddd %xmm5,%xmm0 |
| .byte 15,56,205,245 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm6,%xmm7 |
| .byte 102,15,58,15,253,4 |
| nop |
| paddd %xmm7,%xmm3 |
| .byte 15,56,204,229 |
| .byte 15,56,203,202 |
| movdqa 224-128(%rcx),%xmm0 |
| paddd %xmm6,%xmm0 |
| .byte 15,56,205,222 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm3,%xmm7 |
| .byte 102,15,58,15,254,4 |
| nop |
| paddd %xmm7,%xmm4 |
| .byte 15,56,204,238 |
| .byte 15,56,203,202 |
| movdqa 256-128(%rcx),%xmm0 |
| paddd %xmm3,%xmm0 |
| .byte 15,56,205,227 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm4,%xmm7 |
| .byte 102,15,58,15,251,4 |
| nop |
| paddd %xmm7,%xmm5 |
| .byte 15,56,204,243 |
| .byte 15,56,203,202 |
| movdqa 288-128(%rcx),%xmm0 |
| paddd %xmm4,%xmm0 |
| .byte 15,56,205,236 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm5,%xmm7 |
| .byte 102,15,58,15,252,4 |
| nop |
| paddd %xmm7,%xmm6 |
| .byte 15,56,204,220 |
| .byte 15,56,203,202 |
| movdqa 320-128(%rcx),%xmm0 |
| paddd %xmm5,%xmm0 |
| .byte 15,56,205,245 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm6,%xmm7 |
| .byte 102,15,58,15,253,4 |
| nop |
| paddd %xmm7,%xmm3 |
| .byte 15,56,204,229 |
| .byte 15,56,203,202 |
| movdqa 352-128(%rcx),%xmm0 |
| paddd %xmm6,%xmm0 |
| .byte 15,56,205,222 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm3,%xmm7 |
| .byte 102,15,58,15,254,4 |
| nop |
| paddd %xmm7,%xmm4 |
| .byte 15,56,204,238 |
| .byte 15,56,203,202 |
| movdqa 384-128(%rcx),%xmm0 |
| paddd %xmm3,%xmm0 |
| .byte 15,56,205,227 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm4,%xmm7 |
| .byte 102,15,58,15,251,4 |
| nop |
| paddd %xmm7,%xmm5 |
| .byte 15,56,204,243 |
| .byte 15,56,203,202 |
| movdqa 416-128(%rcx),%xmm0 |
| paddd %xmm4,%xmm0 |
| .byte 15,56,205,236 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| movdqa %xmm5,%xmm7 |
| .byte 102,15,58,15,252,4 |
| .byte 15,56,203,202 |
| paddd %xmm7,%xmm6 |
| |
| movdqa 448-128(%rcx),%xmm0 |
| paddd %xmm5,%xmm0 |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| .byte 15,56,205,245 |
| movdqa %xmm8,%xmm7 |
| .byte 15,56,203,202 |
| |
| movdqa 480-128(%rcx),%xmm0 |
| paddd %xmm6,%xmm0 |
| nop |
| .byte 15,56,203,209 |
| pshufd $0x0e,%xmm0,%xmm0 |
| decq %rdx |
| nop |
| .byte 15,56,203,202 |
| |
| paddd %xmm10,%xmm2 |
| paddd %xmm9,%xmm1 |
| jnz .Loop_shaext |
| |
| pshufd $0xb1,%xmm2,%xmm2 |
| pshufd $0x1b,%xmm1,%xmm7 |
| pshufd $0xb1,%xmm1,%xmm1 |
| punpckhqdq %xmm2,%xmm1 |
| .byte 102,15,58,15,215,8 |
| |
| movdqu %xmm1,(%rdi) |
| movdqu %xmm2,16(%rdi) |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size sha256_block_data_order_shaext,.-sha256_block_data_order_shaext |
| .type sha256_block_data_order_ssse3,@function |
| .align 64 |
| sha256_block_data_order_ssse3: |
| .cfi_startproc |
| .Lssse3_shortcut: |
| movq %rsp,%rax |
| .cfi_def_cfa_register %rax |
| pushq %rbx |
| .cfi_offset %rbx,-16 |
| pushq %rbp |
| .cfi_offset %rbp,-24 |
| pushq %r12 |
| .cfi_offset %r12,-32 |
| pushq %r13 |
| .cfi_offset %r13,-40 |
| pushq %r14 |
| .cfi_offset %r14,-48 |
| pushq %r15 |
| .cfi_offset %r15,-56 |
| shlq $4,%rdx |
| subq $96,%rsp |
| leaq (%rsi,%rdx,4),%rdx |
| andq $-64,%rsp |
| movq %rdi,64+0(%rsp) |
| movq %rsi,64+8(%rsp) |
| movq %rdx,64+16(%rsp) |
| movq %rax,88(%rsp) |
| .cfi_escape 0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08 |
| .Lprologue_ssse3: |
| |
| movl 0(%rdi),%eax |
| movl 4(%rdi),%ebx |
| movl 8(%rdi),%ecx |
| movl 12(%rdi),%edx |
| movl 16(%rdi),%r8d |
| movl 20(%rdi),%r9d |
| movl 24(%rdi),%r10d |
| movl 28(%rdi),%r11d |
| |
| |
| jmp .Lloop_ssse3 |
| .align 16 |
| .Lloop_ssse3: |
| movdqa K256+512(%rip),%xmm7 |
| movdqu 0(%rsi),%xmm0 |
| movdqu 16(%rsi),%xmm1 |
| movdqu 32(%rsi),%xmm2 |
| .byte 102,15,56,0,199 |
| movdqu 48(%rsi),%xmm3 |
| leaq K256(%rip),%rbp |
| .byte 102,15,56,0,207 |
| movdqa 0(%rbp),%xmm4 |
| movdqa 32(%rbp),%xmm5 |
| .byte 102,15,56,0,215 |
| paddd %xmm0,%xmm4 |
| movdqa 64(%rbp),%xmm6 |
| .byte 102,15,56,0,223 |
| movdqa 96(%rbp),%xmm7 |
| paddd %xmm1,%xmm5 |
| paddd %xmm2,%xmm6 |
| paddd %xmm3,%xmm7 |
| movdqa %xmm4,0(%rsp) |
| movl %eax,%r14d |
| movdqa %xmm5,16(%rsp) |
| movl %ebx,%edi |
| movdqa %xmm6,32(%rsp) |
| xorl %ecx,%edi |
| movdqa %xmm7,48(%rsp) |
| movl %r8d,%r13d |
| jmp .Lssse3_00_47 |
| |
| .align 16 |
| .Lssse3_00_47: |
| subq $-128,%rbp |
| rorl $14,%r13d |
| movdqa %xmm1,%xmm4 |
| movl %r14d,%eax |
| movl %r9d,%r12d |
| movdqa %xmm3,%xmm7 |
| rorl $9,%r14d |
| xorl %r8d,%r13d |
| xorl %r10d,%r12d |
| rorl $5,%r13d |
| xorl %eax,%r14d |
| .byte 102,15,58,15,224,4 |
| andl %r8d,%r12d |
| xorl %r8d,%r13d |
| .byte 102,15,58,15,250,4 |
| addl 0(%rsp),%r11d |
| movl %eax,%r15d |
| xorl %r10d,%r12d |
| rorl $11,%r14d |
| movdqa %xmm4,%xmm5 |
| xorl %ebx,%r15d |
| addl %r12d,%r11d |
| movdqa %xmm4,%xmm6 |
| rorl $6,%r13d |
| andl %r15d,%edi |
| psrld $3,%xmm4 |
| xorl %eax,%r14d |
| addl %r13d,%r11d |
| xorl %ebx,%edi |
| paddd %xmm7,%xmm0 |
| rorl $2,%r14d |
| addl %r11d,%edx |
| psrld $7,%xmm6 |
| addl %edi,%r11d |
| movl %edx,%r13d |
| pshufd $250,%xmm3,%xmm7 |
| addl %r11d,%r14d |
| rorl $14,%r13d |
| pslld $14,%xmm5 |
| movl %r14d,%r11d |
| movl %r8d,%r12d |
| pxor %xmm6,%xmm4 |
| rorl $9,%r14d |
| xorl %edx,%r13d |
| xorl %r9d,%r12d |
| rorl $5,%r13d |
| psrld $11,%xmm6 |
| xorl %r11d,%r14d |
| pxor %xmm5,%xmm4 |
| andl %edx,%r12d |
| xorl %edx,%r13d |
| pslld $11,%xmm5 |
| addl 4(%rsp),%r10d |
| movl %r11d,%edi |
| pxor %xmm6,%xmm4 |
| xorl %r9d,%r12d |
| rorl $11,%r14d |
| movdqa %xmm7,%xmm6 |
| xorl %eax,%edi |
| addl %r12d,%r10d |
| pxor %xmm5,%xmm4 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %r11d,%r14d |
| psrld $10,%xmm7 |
| addl %r13d,%r10d |
| xorl %eax,%r15d |
| paddd %xmm4,%xmm0 |
| rorl $2,%r14d |
| addl %r10d,%ecx |
| psrlq $17,%xmm6 |
| addl %r15d,%r10d |
| movl %ecx,%r13d |
| addl %r10d,%r14d |
| pxor %xmm6,%xmm7 |
| rorl $14,%r13d |
| movl %r14d,%r10d |
| movl %edx,%r12d |
| rorl $9,%r14d |
| psrlq $2,%xmm6 |
| xorl %ecx,%r13d |
| xorl %r8d,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $5,%r13d |
| xorl %r10d,%r14d |
| andl %ecx,%r12d |
| pshufd $128,%xmm7,%xmm7 |
| xorl %ecx,%r13d |
| addl 8(%rsp),%r9d |
| movl %r10d,%r15d |
| psrldq $8,%xmm7 |
| xorl %r8d,%r12d |
| rorl $11,%r14d |
| xorl %r11d,%r15d |
| addl %r12d,%r9d |
| rorl $6,%r13d |
| paddd %xmm7,%xmm0 |
| andl %r15d,%edi |
| xorl %r10d,%r14d |
| addl %r13d,%r9d |
| pshufd $80,%xmm0,%xmm7 |
| xorl %r11d,%edi |
| rorl $2,%r14d |
| addl %r9d,%ebx |
| movdqa %xmm7,%xmm6 |
| addl %edi,%r9d |
| movl %ebx,%r13d |
| psrld $10,%xmm7 |
| addl %r9d,%r14d |
| rorl $14,%r13d |
| psrlq $17,%xmm6 |
| movl %r14d,%r9d |
| movl %ecx,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $9,%r14d |
| xorl %ebx,%r13d |
| xorl %edx,%r12d |
| rorl $5,%r13d |
| xorl %r9d,%r14d |
| psrlq $2,%xmm6 |
| andl %ebx,%r12d |
| xorl %ebx,%r13d |
| addl 12(%rsp),%r8d |
| pxor %xmm6,%xmm7 |
| movl %r9d,%edi |
| xorl %edx,%r12d |
| rorl $11,%r14d |
| pshufd $8,%xmm7,%xmm7 |
| xorl %r10d,%edi |
| addl %r12d,%r8d |
| movdqa 0(%rbp),%xmm6 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| pslldq $8,%xmm7 |
| xorl %r9d,%r14d |
| addl %r13d,%r8d |
| xorl %r10d,%r15d |
| paddd %xmm7,%xmm0 |
| rorl $2,%r14d |
| addl %r8d,%eax |
| addl %r15d,%r8d |
| paddd %xmm0,%xmm6 |
| movl %eax,%r13d |
| addl %r8d,%r14d |
| movdqa %xmm6,0(%rsp) |
| rorl $14,%r13d |
| movdqa %xmm2,%xmm4 |
| movl %r14d,%r8d |
| movl %ebx,%r12d |
| movdqa %xmm0,%xmm7 |
| rorl $9,%r14d |
| xorl %eax,%r13d |
| xorl %ecx,%r12d |
| rorl $5,%r13d |
| xorl %r8d,%r14d |
| .byte 102,15,58,15,225,4 |
| andl %eax,%r12d |
| xorl %eax,%r13d |
| .byte 102,15,58,15,251,4 |
| addl 16(%rsp),%edx |
| movl %r8d,%r15d |
| xorl %ecx,%r12d |
| rorl $11,%r14d |
| movdqa %xmm4,%xmm5 |
| xorl %r9d,%r15d |
| addl %r12d,%edx |
| movdqa %xmm4,%xmm6 |
| rorl $6,%r13d |
| andl %r15d,%edi |
| psrld $3,%xmm4 |
| xorl %r8d,%r14d |
| addl %r13d,%edx |
| xorl %r9d,%edi |
| paddd %xmm7,%xmm1 |
| rorl $2,%r14d |
| addl %edx,%r11d |
| psrld $7,%xmm6 |
| addl %edi,%edx |
| movl %r11d,%r13d |
| pshufd $250,%xmm0,%xmm7 |
| addl %edx,%r14d |
| rorl $14,%r13d |
| pslld $14,%xmm5 |
| movl %r14d,%edx |
| movl %eax,%r12d |
| pxor %xmm6,%xmm4 |
| rorl $9,%r14d |
| xorl %r11d,%r13d |
| xorl %ebx,%r12d |
| rorl $5,%r13d |
| psrld $11,%xmm6 |
| xorl %edx,%r14d |
| pxor %xmm5,%xmm4 |
| andl %r11d,%r12d |
| xorl %r11d,%r13d |
| pslld $11,%xmm5 |
| addl 20(%rsp),%ecx |
| movl %edx,%edi |
| pxor %xmm6,%xmm4 |
| xorl %ebx,%r12d |
| rorl $11,%r14d |
| movdqa %xmm7,%xmm6 |
| xorl %r8d,%edi |
| addl %r12d,%ecx |
| pxor %xmm5,%xmm4 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %edx,%r14d |
| psrld $10,%xmm7 |
| addl %r13d,%ecx |
| xorl %r8d,%r15d |
| paddd %xmm4,%xmm1 |
| rorl $2,%r14d |
| addl %ecx,%r10d |
| psrlq $17,%xmm6 |
| addl %r15d,%ecx |
| movl %r10d,%r13d |
| addl %ecx,%r14d |
| pxor %xmm6,%xmm7 |
| rorl $14,%r13d |
| movl %r14d,%ecx |
| movl %r11d,%r12d |
| rorl $9,%r14d |
| psrlq $2,%xmm6 |
| xorl %r10d,%r13d |
| xorl %eax,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $5,%r13d |
| xorl %ecx,%r14d |
| andl %r10d,%r12d |
| pshufd $128,%xmm7,%xmm7 |
| xorl %r10d,%r13d |
| addl 24(%rsp),%ebx |
| movl %ecx,%r15d |
| psrldq $8,%xmm7 |
| xorl %eax,%r12d |
| rorl $11,%r14d |
| xorl %edx,%r15d |
| addl %r12d,%ebx |
| rorl $6,%r13d |
| paddd %xmm7,%xmm1 |
| andl %r15d,%edi |
| xorl %ecx,%r14d |
| addl %r13d,%ebx |
| pshufd $80,%xmm1,%xmm7 |
| xorl %edx,%edi |
| rorl $2,%r14d |
| addl %ebx,%r9d |
| movdqa %xmm7,%xmm6 |
| addl %edi,%ebx |
| movl %r9d,%r13d |
| psrld $10,%xmm7 |
| addl %ebx,%r14d |
| rorl $14,%r13d |
| psrlq $17,%xmm6 |
| movl %r14d,%ebx |
| movl %r10d,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $9,%r14d |
| xorl %r9d,%r13d |
| xorl %r11d,%r12d |
| rorl $5,%r13d |
| xorl %ebx,%r14d |
| psrlq $2,%xmm6 |
| andl %r9d,%r12d |
| xorl %r9d,%r13d |
| addl 28(%rsp),%eax |
| pxor %xmm6,%xmm7 |
| movl %ebx,%edi |
| xorl %r11d,%r12d |
| rorl $11,%r14d |
| pshufd $8,%xmm7,%xmm7 |
| xorl %ecx,%edi |
| addl %r12d,%eax |
| movdqa 32(%rbp),%xmm6 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| pslldq $8,%xmm7 |
| xorl %ebx,%r14d |
| addl %r13d,%eax |
| xorl %ecx,%r15d |
| paddd %xmm7,%xmm1 |
| rorl $2,%r14d |
| addl %eax,%r8d |
| addl %r15d,%eax |
| paddd %xmm1,%xmm6 |
| movl %r8d,%r13d |
| addl %eax,%r14d |
| movdqa %xmm6,16(%rsp) |
| rorl $14,%r13d |
| movdqa %xmm3,%xmm4 |
| movl %r14d,%eax |
| movl %r9d,%r12d |
| movdqa %xmm1,%xmm7 |
| rorl $9,%r14d |
| xorl %r8d,%r13d |
| xorl %r10d,%r12d |
| rorl $5,%r13d |
| xorl %eax,%r14d |
| .byte 102,15,58,15,226,4 |
| andl %r8d,%r12d |
| xorl %r8d,%r13d |
| .byte 102,15,58,15,248,4 |
| addl 32(%rsp),%r11d |
| movl %eax,%r15d |
| xorl %r10d,%r12d |
| rorl $11,%r14d |
| movdqa %xmm4,%xmm5 |
| xorl %ebx,%r15d |
| addl %r12d,%r11d |
| movdqa %xmm4,%xmm6 |
| rorl $6,%r13d |
| andl %r15d,%edi |
| psrld $3,%xmm4 |
| xorl %eax,%r14d |
| addl %r13d,%r11d |
| xorl %ebx,%edi |
| paddd %xmm7,%xmm2 |
| rorl $2,%r14d |
| addl %r11d,%edx |
| psrld $7,%xmm6 |
| addl %edi,%r11d |
| movl %edx,%r13d |
| pshufd $250,%xmm1,%xmm7 |
| addl %r11d,%r14d |
| rorl $14,%r13d |
| pslld $14,%xmm5 |
| movl %r14d,%r11d |
| movl %r8d,%r12d |
| pxor %xmm6,%xmm4 |
| rorl $9,%r14d |
| xorl %edx,%r13d |
| xorl %r9d,%r12d |
| rorl $5,%r13d |
| psrld $11,%xmm6 |
| xorl %r11d,%r14d |
| pxor %xmm5,%xmm4 |
| andl %edx,%r12d |
| xorl %edx,%r13d |
| pslld $11,%xmm5 |
| addl 36(%rsp),%r10d |
| movl %r11d,%edi |
| pxor %xmm6,%xmm4 |
| xorl %r9d,%r12d |
| rorl $11,%r14d |
| movdqa %xmm7,%xmm6 |
| xorl %eax,%edi |
| addl %r12d,%r10d |
| pxor %xmm5,%xmm4 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %r11d,%r14d |
| psrld $10,%xmm7 |
| addl %r13d,%r10d |
| xorl %eax,%r15d |
| paddd %xmm4,%xmm2 |
| rorl $2,%r14d |
| addl %r10d,%ecx |
| psrlq $17,%xmm6 |
| addl %r15d,%r10d |
| movl %ecx,%r13d |
| addl %r10d,%r14d |
| pxor %xmm6,%xmm7 |
| rorl $14,%r13d |
| movl %r14d,%r10d |
| movl %edx,%r12d |
| rorl $9,%r14d |
| psrlq $2,%xmm6 |
| xorl %ecx,%r13d |
| xorl %r8d,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $5,%r13d |
| xorl %r10d,%r14d |
| andl %ecx,%r12d |
| pshufd $128,%xmm7,%xmm7 |
| xorl %ecx,%r13d |
| addl 40(%rsp),%r9d |
| movl %r10d,%r15d |
| psrldq $8,%xmm7 |
| xorl %r8d,%r12d |
| rorl $11,%r14d |
| xorl %r11d,%r15d |
| addl %r12d,%r9d |
| rorl $6,%r13d |
| paddd %xmm7,%xmm2 |
| andl %r15d,%edi |
| xorl %r10d,%r14d |
| addl %r13d,%r9d |
| pshufd $80,%xmm2,%xmm7 |
| xorl %r11d,%edi |
| rorl $2,%r14d |
| addl %r9d,%ebx |
| movdqa %xmm7,%xmm6 |
| addl %edi,%r9d |
| movl %ebx,%r13d |
| psrld $10,%xmm7 |
| addl %r9d,%r14d |
| rorl $14,%r13d |
| psrlq $17,%xmm6 |
| movl %r14d,%r9d |
| movl %ecx,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $9,%r14d |
| xorl %ebx,%r13d |
| xorl %edx,%r12d |
| rorl $5,%r13d |
| xorl %r9d,%r14d |
| psrlq $2,%xmm6 |
| andl %ebx,%r12d |
| xorl %ebx,%r13d |
| addl 44(%rsp),%r8d |
| pxor %xmm6,%xmm7 |
| movl %r9d,%edi |
| xorl %edx,%r12d |
| rorl $11,%r14d |
| pshufd $8,%xmm7,%xmm7 |
| xorl %r10d,%edi |
| addl %r12d,%r8d |
| movdqa 64(%rbp),%xmm6 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| pslldq $8,%xmm7 |
| xorl %r9d,%r14d |
| addl %r13d,%r8d |
| xorl %r10d,%r15d |
| paddd %xmm7,%xmm2 |
| rorl $2,%r14d |
| addl %r8d,%eax |
| addl %r15d,%r8d |
| paddd %xmm2,%xmm6 |
| movl %eax,%r13d |
| addl %r8d,%r14d |
| movdqa %xmm6,32(%rsp) |
| rorl $14,%r13d |
| movdqa %xmm0,%xmm4 |
| movl %r14d,%r8d |
| movl %ebx,%r12d |
| movdqa %xmm2,%xmm7 |
| rorl $9,%r14d |
| xorl %eax,%r13d |
| xorl %ecx,%r12d |
| rorl $5,%r13d |
| xorl %r8d,%r14d |
| .byte 102,15,58,15,227,4 |
| andl %eax,%r12d |
| xorl %eax,%r13d |
| .byte 102,15,58,15,249,4 |
| addl 48(%rsp),%edx |
| movl %r8d,%r15d |
| xorl %ecx,%r12d |
| rorl $11,%r14d |
| movdqa %xmm4,%xmm5 |
| xorl %r9d,%r15d |
| addl %r12d,%edx |
| movdqa %xmm4,%xmm6 |
| rorl $6,%r13d |
| andl %r15d,%edi |
| psrld $3,%xmm4 |
| xorl %r8d,%r14d |
| addl %r13d,%edx |
| xorl %r9d,%edi |
| paddd %xmm7,%xmm3 |
| rorl $2,%r14d |
| addl %edx,%r11d |
| psrld $7,%xmm6 |
| addl %edi,%edx |
| movl %r11d,%r13d |
| pshufd $250,%xmm2,%xmm7 |
| addl %edx,%r14d |
| rorl $14,%r13d |
| pslld $14,%xmm5 |
| movl %r14d,%edx |
| movl %eax,%r12d |
| pxor %xmm6,%xmm4 |
| rorl $9,%r14d |
| xorl %r11d,%r13d |
| xorl %ebx,%r12d |
| rorl $5,%r13d |
| psrld $11,%xmm6 |
| xorl %edx,%r14d |
| pxor %xmm5,%xmm4 |
| andl %r11d,%r12d |
| xorl %r11d,%r13d |
| pslld $11,%xmm5 |
| addl 52(%rsp),%ecx |
| movl %edx,%edi |
| pxor %xmm6,%xmm4 |
| xorl %ebx,%r12d |
| rorl $11,%r14d |
| movdqa %xmm7,%xmm6 |
| xorl %r8d,%edi |
| addl %r12d,%ecx |
| pxor %xmm5,%xmm4 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %edx,%r14d |
| psrld $10,%xmm7 |
| addl %r13d,%ecx |
| xorl %r8d,%r15d |
| paddd %xmm4,%xmm3 |
| rorl $2,%r14d |
| addl %ecx,%r10d |
| psrlq $17,%xmm6 |
| addl %r15d,%ecx |
| movl %r10d,%r13d |
| addl %ecx,%r14d |
| pxor %xmm6,%xmm7 |
| rorl $14,%r13d |
| movl %r14d,%ecx |
| movl %r11d,%r12d |
| rorl $9,%r14d |
| psrlq $2,%xmm6 |
| xorl %r10d,%r13d |
| xorl %eax,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $5,%r13d |
| xorl %ecx,%r14d |
| andl %r10d,%r12d |
| pshufd $128,%xmm7,%xmm7 |
| xorl %r10d,%r13d |
| addl 56(%rsp),%ebx |
| movl %ecx,%r15d |
| psrldq $8,%xmm7 |
| xorl %eax,%r12d |
| rorl $11,%r14d |
| xorl %edx,%r15d |
| addl %r12d,%ebx |
| rorl $6,%r13d |
| paddd %xmm7,%xmm3 |
| andl %r15d,%edi |
| xorl %ecx,%r14d |
| addl %r13d,%ebx |
| pshufd $80,%xmm3,%xmm7 |
| xorl %edx,%edi |
| rorl $2,%r14d |
| addl %ebx,%r9d |
| movdqa %xmm7,%xmm6 |
| addl %edi,%ebx |
| movl %r9d,%r13d |
| psrld $10,%xmm7 |
| addl %ebx,%r14d |
| rorl $14,%r13d |
| psrlq $17,%xmm6 |
| movl %r14d,%ebx |
| movl %r10d,%r12d |
| pxor %xmm6,%xmm7 |
| rorl $9,%r14d |
| xorl %r9d,%r13d |
| xorl %r11d,%r12d |
| rorl $5,%r13d |
| xorl %ebx,%r14d |
| psrlq $2,%xmm6 |
| andl %r9d,%r12d |
| xorl %r9d,%r13d |
| addl 60(%rsp),%eax |
| pxor %xmm6,%xmm7 |
| movl %ebx,%edi |
| xorl %r11d,%r12d |
| rorl $11,%r14d |
| pshufd $8,%xmm7,%xmm7 |
| xorl %ecx,%edi |
| addl %r12d,%eax |
| movdqa 96(%rbp),%xmm6 |
| rorl $6,%r13d |
| andl %edi,%r15d |
| pslldq $8,%xmm7 |
| xorl %ebx,%r14d |
| addl %r13d,%eax |
| xorl %ecx,%r15d |
| paddd %xmm7,%xmm3 |
| rorl $2,%r14d |
| addl %eax,%r8d |
| addl %r15d,%eax |
| paddd %xmm3,%xmm6 |
| movl %r8d,%r13d |
| addl %eax,%r14d |
| movdqa %xmm6,48(%rsp) |
| cmpb $0,131(%rbp) |
| jne .Lssse3_00_47 |
| rorl $14,%r13d |
| movl %r14d,%eax |
| movl %r9d,%r12d |
| rorl $9,%r14d |
| xorl %r8d,%r13d |
| xorl %r10d,%r12d |
| rorl $5,%r13d |
| xorl %eax,%r14d |
| andl %r8d,%r12d |
| xorl %r8d,%r13d |
| addl 0(%rsp),%r11d |
| movl %eax,%r15d |
| xorl %r10d,%r12d |
| rorl $11,%r14d |
| xorl %ebx,%r15d |
| addl %r12d,%r11d |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %eax,%r14d |
| addl %r13d,%r11d |
| xorl %ebx,%edi |
| rorl $2,%r14d |
| addl %r11d,%edx |
| addl %edi,%r11d |
| movl %edx,%r13d |
| addl %r11d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r11d |
| movl %r8d,%r12d |
| rorl $9,%r14d |
| xorl %edx,%r13d |
| xorl %r9d,%r12d |
| rorl $5,%r13d |
| xorl %r11d,%r14d |
| andl %edx,%r12d |
| xorl %edx,%r13d |
| addl 4(%rsp),%r10d |
| movl %r11d,%edi |
| xorl %r9d,%r12d |
| rorl $11,%r14d |
| xorl %eax,%edi |
| addl %r12d,%r10d |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %r11d,%r14d |
| addl %r13d,%r10d |
| xorl %eax,%r15d |
| rorl $2,%r14d |
| addl %r10d,%ecx |
| addl %r15d,%r10d |
| movl %ecx,%r13d |
| addl %r10d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r10d |
| movl %edx,%r12d |
| rorl $9,%r14d |
| xorl %ecx,%r13d |
| xorl %r8d,%r12d |
| rorl $5,%r13d |
| xorl %r10d,%r14d |
| andl %ecx,%r12d |
| xorl %ecx,%r13d |
| addl 8(%rsp),%r9d |
| movl %r10d,%r15d |
| xorl %r8d,%r12d |
| rorl $11,%r14d |
| xorl %r11d,%r15d |
| addl %r12d,%r9d |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %r10d,%r14d |
| addl %r13d,%r9d |
| xorl %r11d,%edi |
| rorl $2,%r14d |
| addl %r9d,%ebx |
| addl %edi,%r9d |
| movl %ebx,%r13d |
| addl %r9d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r9d |
| movl %ecx,%r12d |
| rorl $9,%r14d |
| xorl %ebx,%r13d |
| xorl %edx,%r12d |
| rorl $5,%r13d |
| xorl %r9d,%r14d |
| andl %ebx,%r12d |
| xorl %ebx,%r13d |
| addl 12(%rsp),%r8d |
| movl %r9d,%edi |
| xorl %edx,%r12d |
| rorl $11,%r14d |
| xorl %r10d,%edi |
| addl %r12d,%r8d |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %r9d,%r14d |
| addl %r13d,%r8d |
| xorl %r10d,%r15d |
| rorl $2,%r14d |
| addl %r8d,%eax |
| addl %r15d,%r8d |
| movl %eax,%r13d |
| addl %r8d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r8d |
| movl %ebx,%r12d |
| rorl $9,%r14d |
| xorl %eax,%r13d |
| xorl %ecx,%r12d |
| rorl $5,%r13d |
| xorl %r8d,%r14d |
| andl %eax,%r12d |
| xorl %eax,%r13d |
| addl 16(%rsp),%edx |
| movl %r8d,%r15d |
| xorl %ecx,%r12d |
| rorl $11,%r14d |
| xorl %r9d,%r15d |
| addl %r12d,%edx |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %r8d,%r14d |
| addl %r13d,%edx |
| xorl %r9d,%edi |
| rorl $2,%r14d |
| addl %edx,%r11d |
| addl %edi,%edx |
| movl %r11d,%r13d |
| addl %edx,%r14d |
| rorl $14,%r13d |
| movl %r14d,%edx |
| movl %eax,%r12d |
| rorl $9,%r14d |
| xorl %r11d,%r13d |
| xorl %ebx,%r12d |
| rorl $5,%r13d |
| xorl %edx,%r14d |
| andl %r11d,%r12d |
| xorl %r11d,%r13d |
| addl 20(%rsp),%ecx |
| movl %edx,%edi |
| xorl %ebx,%r12d |
| rorl $11,%r14d |
| xorl %r8d,%edi |
| addl %r12d,%ecx |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %edx,%r14d |
| addl %r13d,%ecx |
| xorl %r8d,%r15d |
| rorl $2,%r14d |
| addl %ecx,%r10d |
| addl %r15d,%ecx |
| movl %r10d,%r13d |
| addl %ecx,%r14d |
| rorl $14,%r13d |
| movl %r14d,%ecx |
| movl %r11d,%r12d |
| rorl $9,%r14d |
| xorl %r10d,%r13d |
| xorl %eax,%r12d |
| rorl $5,%r13d |
| xorl %ecx,%r14d |
| andl %r10d,%r12d |
| xorl %r10d,%r13d |
| addl 24(%rsp),%ebx |
| movl %ecx,%r15d |
| xorl %eax,%r12d |
| rorl $11,%r14d |
| xorl %edx,%r15d |
| addl %r12d,%ebx |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %ecx,%r14d |
| addl %r13d,%ebx |
| xorl %edx,%edi |
| rorl $2,%r14d |
| addl %ebx,%r9d |
| addl %edi,%ebx |
| movl %r9d,%r13d |
| addl %ebx,%r14d |
| rorl $14,%r13d |
| movl %r14d,%ebx |
| movl %r10d,%r12d |
| rorl $9,%r14d |
| xorl %r9d,%r13d |
| xorl %r11d,%r12d |
| rorl $5,%r13d |
| xorl %ebx,%r14d |
| andl %r9d,%r12d |
| xorl %r9d,%r13d |
| addl 28(%rsp),%eax |
| movl %ebx,%edi |
| xorl %r11d,%r12d |
| rorl $11,%r14d |
| xorl %ecx,%edi |
| addl %r12d,%eax |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %ebx,%r14d |
| addl %r13d,%eax |
| xorl %ecx,%r15d |
| rorl $2,%r14d |
| addl %eax,%r8d |
| addl %r15d,%eax |
| movl %r8d,%r13d |
| addl %eax,%r14d |
| rorl $14,%r13d |
| movl %r14d,%eax |
| movl %r9d,%r12d |
| rorl $9,%r14d |
| xorl %r8d,%r13d |
| xorl %r10d,%r12d |
| rorl $5,%r13d |
| xorl %eax,%r14d |
| andl %r8d,%r12d |
| xorl %r8d,%r13d |
| addl 32(%rsp),%r11d |
| movl %eax,%r15d |
| xorl %r10d,%r12d |
| rorl $11,%r14d |
| xorl %ebx,%r15d |
| addl %r12d,%r11d |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %eax,%r14d |
| addl %r13d,%r11d |
| xorl %ebx,%edi |
| rorl $2,%r14d |
| addl %r11d,%edx |
| addl %edi,%r11d |
| movl %edx,%r13d |
| addl %r11d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r11d |
| movl %r8d,%r12d |
| rorl $9,%r14d |
| xorl %edx,%r13d |
| xorl %r9d,%r12d |
| rorl $5,%r13d |
| xorl %r11d,%r14d |
| andl %edx,%r12d |
| xorl %edx,%r13d |
| addl 36(%rsp),%r10d |
| movl %r11d,%edi |
| xorl %r9d,%r12d |
| rorl $11,%r14d |
| xorl %eax,%edi |
| addl %r12d,%r10d |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %r11d,%r14d |
| addl %r13d,%r10d |
| xorl %eax,%r15d |
| rorl $2,%r14d |
| addl %r10d,%ecx |
| addl %r15d,%r10d |
| movl %ecx,%r13d |
| addl %r10d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r10d |
| movl %edx,%r12d |
| rorl $9,%r14d |
| xorl %ecx,%r13d |
| xorl %r8d,%r12d |
| rorl $5,%r13d |
| xorl %r10d,%r14d |
| andl %ecx,%r12d |
| xorl %ecx,%r13d |
| addl 40(%rsp),%r9d |
| movl %r10d,%r15d |
| xorl %r8d,%r12d |
| rorl $11,%r14d |
| xorl %r11d,%r15d |
| addl %r12d,%r9d |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %r10d,%r14d |
| addl %r13d,%r9d |
| xorl %r11d,%edi |
| rorl $2,%r14d |
| addl %r9d,%ebx |
| addl %edi,%r9d |
| movl %ebx,%r13d |
| addl %r9d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r9d |
| movl %ecx,%r12d |
| rorl $9,%r14d |
| xorl %ebx,%r13d |
| xorl %edx,%r12d |
| rorl $5,%r13d |
| xorl %r9d,%r14d |
| andl %ebx,%r12d |
| xorl %ebx,%r13d |
| addl 44(%rsp),%r8d |
| movl %r9d,%edi |
| xorl %edx,%r12d |
| rorl $11,%r14d |
| xorl %r10d,%edi |
| addl %r12d,%r8d |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %r9d,%r14d |
| addl %r13d,%r8d |
| xorl %r10d,%r15d |
| rorl $2,%r14d |
| addl %r8d,%eax |
| addl %r15d,%r8d |
| movl %eax,%r13d |
| addl %r8d,%r14d |
| rorl $14,%r13d |
| movl %r14d,%r8d |
| movl %ebx,%r12d |
| rorl $9,%r14d |
| xorl %eax,%r13d |
| xorl %ecx,%r12d |
| rorl $5,%r13d |
| xorl %r8d,%r14d |
| andl %eax,%r12d |
| xorl %eax,%r13d |
| addl 48(%rsp),%edx |
| movl %r8d,%r15d |
| xorl %ecx,%r12d |
| rorl $11,%r14d |
| xorl %r9d,%r15d |
| addl %r12d,%edx |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %r8d,%r14d |
| addl %r13d,%edx |
| xorl %r9d,%edi |
| rorl $2,%r14d |
| addl %edx,%r11d |
| addl %edi,%edx |
| movl %r11d,%r13d |
| addl %edx,%r14d |
| rorl $14,%r13d |
| movl %r14d,%edx |
| movl %eax,%r12d |
| rorl $9,%r14d |
| xorl %r11d,%r13d |
| xorl %ebx,%r12d |
| rorl $5,%r13d |
| xorl %edx,%r14d |
| andl %r11d,%r12d |
| xorl %r11d,%r13d |
| addl 52(%rsp),%ecx |
| movl %edx,%edi |
| xorl %ebx,%r12d |
| rorl $11,%r14d |
| xorl %r8d,%edi |
| addl %r12d,%ecx |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %edx,%r14d |
| addl %r13d,%ecx |
| xorl %r8d,%r15d |
| rorl $2,%r14d |
| addl %ecx,%r10d |
| addl %r15d,%ecx |
| movl %r10d,%r13d |
| addl %ecx,%r14d |
| rorl $14,%r13d |
| movl %r14d,%ecx |
| movl %r11d,%r12d |
| rorl $9,%r14d |
| xorl %r10d,%r13d |
| xorl %eax,%r12d |
| rorl $5,%r13d |
| xorl %ecx,%r14d |
| andl %r10d,%r12d |
| xorl %r10d,%r13d |
| addl 56(%rsp),%ebx |
| movl %ecx,%r15d |
| xorl %eax,%r12d |
| rorl $11,%r14d |
| xorl %edx,%r15d |
| addl %r12d,%ebx |
| rorl $6,%r13d |
| andl %r15d,%edi |
| xorl %ecx,%r14d |
| addl %r13d,%ebx |
| xorl %edx,%edi |
| rorl $2,%r14d |
| addl %ebx,%r9d |
| addl %edi,%ebx |
| movl %r9d,%r13d |
| addl %ebx,%r14d |
| rorl $14,%r13d |
| movl %r14d,%ebx |
| movl %r10d,%r12d |
| rorl $9,%r14d |
| xorl %r9d,%r13d |
| xorl %r11d,%r12d |
| rorl $5,%r13d |
| xorl %ebx,%r14d |
| andl %r9d,%r12d |
| xorl %r9d,%r13d |
| addl 60(%rsp),%eax |
| movl %ebx,%edi |
| xorl %r11d,%r12d |
| rorl $11,%r14d |
| xorl %ecx,%edi |
| addl %r12d,%eax |
| rorl $6,%r13d |
| andl %edi,%r15d |
| xorl %ebx,%r14d |
| addl %r13d,%eax |
| xorl %ecx,%r15d |
| rorl $2,%r14d |
| addl %eax,%r8d |
| addl %r15d,%eax |
| movl %r8d,%r13d |
| addl %eax,%r14d |
| movq 64+0(%rsp),%rdi |
| movl %r14d,%eax |
| |
| addl 0(%rdi),%eax |
| leaq 64(%rsi),%rsi |
| addl 4(%rdi),%ebx |
| addl 8(%rdi),%ecx |
| addl 12(%rdi),%edx |
| addl 16(%rdi),%r8d |
| addl 20(%rdi),%r9d |
| addl 24(%rdi),%r10d |
| addl 28(%rdi),%r11d |
| |
| cmpq 64+16(%rsp),%rsi |
| |
| movl %eax,0(%rdi) |
| movl %ebx,4(%rdi) |
| movl %ecx,8(%rdi) |
| movl %edx,12(%rdi) |
| movl %r8d,16(%rdi) |
| movl %r9d,20(%rdi) |
| movl %r10d,24(%rdi) |
| movl %r11d,28(%rdi) |
| jb .Lloop_ssse3 |
| |
| movq 88(%rsp),%rsi |
| .cfi_def_cfa %rsi,8 |
| movq -48(%rsi),%r15 |
| .cfi_restore %r15 |
| movq -40(%rsi),%r14 |
| .cfi_restore %r14 |
| movq -32(%rsi),%r13 |
| .cfi_restore %r13 |
| movq -24(%rsi),%r12 |
| .cfi_restore %r12 |
| movq -16(%rsi),%rbp |
| .cfi_restore %rbp |
| movq -8(%rsi),%rbx |
| .cfi_restore %rbx |
| leaq (%rsi),%rsp |
| .cfi_def_cfa_register %rsp |
| .Lepilogue_ssse3: |
| .byte 0xf3,0xc3 |
| .cfi_endproc |
| .size sha256_block_data_order_ssse3,.-sha256_block_data_order_ssse3 |
| .type sha256_block_data_order_avx,@function |
| .align 64 |
| sha256_block_data_order_avx: |
| .cfi_startproc |
| .Lavx_shortcut: |
| movq %rsp,%rax |
| .cfi_def_cfa_register %rax |
| pushq %rbx |
| .cfi_offset %rbx,-16 |
| pushq %rbp |
| .cfi_offset %rbp,-24 |
| pushq %r12 |
| .cfi_offset %r12,-32 |
| pushq %r13 |
| .cfi_offset %r13,-40 |
| pushq %r14 |
| .cfi_offset %r14,-48 |
| pushq %r15 |
| .cfi_offset %r15,-56 |
| shlq $4,%rdx |
| subq $96,%rsp |
| leaq (%rsi,%rdx,4),%rdx |
| andq $-64,%rsp |
| movq %rdi,64+0(%rsp) |
| movq %rsi,64+8(%rsp) |
| movq %rdx,64+16(%rsp) |
| movq %rax,88(%rsp) |
| .cfi_escape 0x0f,0x06,0x77,0xd8,0x00,0x06,0x23,0x08 |
| .Lprologue_avx: |
| |
| vzeroupper |
| movl 0(%rdi),%eax |
| movl 4(%rdi),%ebx |
| movl 8(%rdi),%ecx |
| movl 12(%rdi),%edx |
| movl 16(%rdi),%r8d |
| movl 20(%rdi),%r9d |
| movl 24(%rdi),%r10d |
| movl 28(%rdi),%r11d |
| vmovdqa K256+512+32(%rip),%xmm8 |
| vmovdqa K256+512+64(%rip),%xmm9 |
| jmp .Lloop_avx |
| .align 16 |
| .Lloop_avx: |
| vmovdqa K256+512(%rip),%xmm7 |
| vmovdqu 0(%rsi),%xmm0 |
| vmovdqu 16(%rsi),%xmm1 |
| vmovdqu 32(%rsi),%xmm2 |
| vmovdqu 48(%rsi),%xmm3 |
| vpshufb %xmm7,%xmm0,%xmm0 |
| leaq K256(%rip),%rbp |
| vpshufb %xmm7,%xmm1,%xmm1 |
| vpshufb %xmm7,%xmm2,%xmm2 |
| vpaddd 0(%rbp),%xmm0,%xmm4 |
| vpshufb %xmm7,%xmm3,%xmm3 |
| vpaddd 32(%rbp),%xmm1,%xmm5 |
| vpaddd 64(%rbp),%xmm2,%xmm6 |
| vpaddd 96(%rbp),%xmm3,%xmm7 |
| vmovdqa %xmm4,0(%rsp) |
| movl %eax,%r14d |
| vmovdqa %xmm5,16(%rsp) |
| movl %ebx,%edi |
| vmovdqa %xmm6,32(%rsp) |
| xorl %ecx,%edi |
| vmovdqa %xmm7,48(%rsp) |
| movl %r8d,%r13d |
| jmp .Lavx_00_47 |
| |
| .align 16 |
| .Lavx_00_47: |
| subq $-128,%rbp |
| vpalignr $4,%xmm0,%xmm1,%xmm4 |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%eax |
| movl %r9d,%r12d |
| vpalignr $4,%xmm2,%xmm3,%xmm7 |
| shrdl $9,%r14d,%r14d |
| xorl %r8d,%r13d |
| xorl %r10d,%r12d |
| vpsrld $7,%xmm4,%xmm6 |
| shrdl $5,%r13d,%r13d |
| xorl %eax,%r14d |
| andl %r8d,%r12d |
| vpaddd %xmm7,%xmm0,%xmm0 |
| xorl %r8d,%r13d |
| addl 0(%rsp),%r11d |
| movl %eax,%r15d |
| vpsrld $3,%xmm4,%xmm7 |
| xorl %r10d,%r12d |
| shrdl $11,%r14d,%r14d |
| xorl %ebx,%r15d |
| vpslld $14,%xmm4,%xmm5 |
| addl %r12d,%r11d |
| shrdl $6,%r13d,%r13d |
| andl %r15d,%edi |
| vpxor %xmm6,%xmm7,%xmm4 |
| xorl %eax,%r14d |
| addl %r13d,%r11d |
| xorl %ebx,%edi |
| vpshufd $250,%xmm3,%xmm7 |
| shrdl $2,%r14d,%r14d |
| addl %r11d,%edx |
| addl %edi,%r11d |
| vpsrld $11,%xmm6,%xmm6 |
| movl %edx,%r13d |
| addl %r11d,%r14d |
| shrdl $14,%r13d,%r13d |
| vpxor %xmm5,%xmm4,%xmm4 |
| movl %r14d,%r11d |
| movl %r8d,%r12d |
| shrdl $9,%r14d,%r14d |
| vpslld $11,%xmm5,%xmm5 |
| xorl %edx,%r13d |
| xorl %r9d,%r12d |
| shrdl $5,%r13d,%r13d |
| vpxor %xmm6,%xmm4,%xmm4 |
| xorl %r11d,%r14d |
| andl %edx,%r12d |
| xorl %edx,%r13d |
| vpsrld $10,%xmm7,%xmm6 |
| addl 4(%rsp),%r10d |
| movl %r11d,%edi |
| xorl %r9d,%r12d |
| vpxor %xmm5,%xmm4,%xmm4 |
| shrdl $11,%r14d,%r14d |
| xorl %eax,%edi |
| addl %r12d,%r10d |
| vpsrlq $17,%xmm7,%xmm7 |
| shrdl $6,%r13d,%r13d |
| andl %edi,%r15d |
| xorl %r11d,%r14d |
| vpaddd %xmm4,%xmm0,%xmm0 |
| addl %r13d,%r10d |
| xorl %eax,%r15d |
| shrdl $2,%r14d,%r14d |
| vpxor %xmm7,%xmm6,%xmm6 |
| addl %r10d,%ecx |
| addl %r15d,%r10d |
| movl %ecx,%r13d |
| vpsrlq $2,%xmm7,%xmm7 |
| addl %r10d,%r14d |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%r10d |
| vpxor %xmm7,%xmm6,%xmm6 |
| movl %edx,%r12d |
| shrdl $9,%r14d,%r14d |
| xorl %ecx,%r13d |
| vpshufb %xmm8,%xmm6,%xmm6 |
| xorl %r8d,%r12d |
| shrdl $5,%r13d,%r13d |
| xorl %r10d,%r14d |
| vpaddd %xmm6,%xmm0,%xmm0 |
| andl %ecx,%r12d |
| xorl %ecx,%r13d |
| addl 8(%rsp),%r9d |
| vpshufd $80,%xmm0,%xmm7 |
| movl %r10d,%r15d |
| xorl %r8d,%r12d |
| shrdl $11,%r14d,%r14d |
| vpsrld $10,%xmm7,%xmm6 |
| xorl %r11d,%r15d |
| addl %r12d,%r9d |
| shrdl $6,%r13d,%r13d |
| vpsrlq $17,%xmm7,%xmm7 |
| andl %r15d,%edi |
| xorl %r10d,%r14d |
| addl %r13d,%r9d |
| vpxor %xmm7,%xmm6,%xmm6 |
| xorl %r11d,%edi |
| shrdl $2,%r14d,%r14d |
| addl %r9d,%ebx |
| vpsrlq $2,%xmm7,%xmm7 |
| addl %edi,%r9d |
| movl %ebx,%r13d |
| addl %r9d,%r14d |
| vpxor %xmm7,%xmm6,%xmm6 |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%r9d |
| movl %ecx,%r12d |
| vpshufb %xmm9,%xmm6,%xmm6 |
| shrdl $9,%r14d,%r14d |
| xorl %ebx,%r13d |
| xorl %edx,%r12d |
| vpaddd %xmm6,%xmm0,%xmm0 |
| shrdl $5,%r13d,%r13d |
| xorl %r9d,%r14d |
| andl %ebx,%r12d |
| vpaddd 0(%rbp),%xmm0,%xmm6 |
| xorl %ebx,%r13d |
| addl 12(%rsp),%r8d |
| movl %r9d,%edi |
| xorl %edx,%r12d |
| shrdl $11,%r14d,%r14d |
| xorl %r10d,%edi |
| addl %r12d,%r8d |
| shrdl $6,%r13d,%r13d |
| andl %edi,%r15d |
| xorl %r9d,%r14d |
| addl %r13d,%r8d |
| xorl %r10d,%r15d |
| shrdl $2,%r14d,%r14d |
| addl %r8d,%eax |
| addl %r15d,%r8d |
| movl %eax,%r13d |
| addl %r8d,%r14d |
| vmovdqa %xmm6,0(%rsp) |
| vpalignr $4,%xmm1,%xmm2,%xmm4 |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%r8d |
| movl %ebx,%r12d |
| vpalignr $4,%xmm3,%xmm0,%xmm7 |
| shrdl $9,%r14d,%r14d |
| xorl %eax,%r13d |
| xorl %ecx,%r12d |
| vpsrld $7,%xmm4,%xmm6 |
| shrdl $5,%r13d,%r13d |
| xorl %r8d,%r14d |
| andl %eax,%r12d |
| vpaddd %xmm7,%xmm1,%xmm1 |
| xorl %eax,%r13d |
| addl 16(%rsp),%edx |
| movl %r8d,%r15d |
| vpsrld $3,%xmm4,%xmm7 |
| xorl %ecx,%r12d |
| shrdl $11,%r14d,%r14d |
| xorl %r9d,%r15d |
| vpslld $14,%xmm4,%xmm5 |
| addl %r12d,%edx |
| shrdl $6,%r13d,%r13d |
| andl %r15d,%edi |
| vpxor %xmm6,%xmm7,%xmm4 |
| xorl %r8d,%r14d |
| addl %r13d,%edx |
| xorl %r9d,%edi |
| vpshufd $250,%xmm0,%xmm7 |
| shrdl $2,%r14d,%r14d |
| addl %edx,%r11d |
| addl %edi,%edx |
| vpsrld $11,%xmm6,%xmm6 |
| movl %r11d,%r13d |
| addl %edx,%r14d |
| shrdl $14,%r13d,%r13d |
| vpxor %xmm5,%xmm4,%xmm4 |
| movl %r14d,%edx |
| movl %eax,%r12d |
| shrdl $9,%r14d,%r14d |
| vpslld $11,%xmm5,%xmm5 |
| xorl %r11d,%r13d |
| xorl %ebx,%r12d |
| shrdl $5,%r13d,%r13d |
| vpxor %xmm6,%xmm4,%xmm4 |
| xorl %edx,%r14d |
| andl %r11d,%r12d |
| xorl %r11d,%r13d |
| vpsrld $10,%xmm7,%xmm6 |
| addl 20(%rsp),%ecx |
| movl %edx,%edi |
| xorl %ebx,%r12d |
| vpxor %xmm5,%xmm4,%xmm4 |
| shrdl $11,%r14d,%r14d |
| xorl %r8d,%edi |
| addl %r12d,%ecx |
| vpsrlq $17,%xmm7,%xmm7 |
| shrdl $6,%r13d,%r13d |
| andl %edi,%r15d |
| xorl %edx,%r14d |
| vpaddd %xmm4,%xmm1,%xmm1 |
| addl %r13d,%ecx |
| xorl %r8d,%r15d |
| shrdl $2,%r14d,%r14d |
| vpxor %xmm7,%xmm6,%xmm6 |
| addl %ecx,%r10d |
| addl %r15d,%ecx |
| movl %r10d,%r13d |
| vpsrlq $2,%xmm7,%xmm7 |
| addl %ecx,%r14d |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%ecx |
| vpxor %xmm7,%xmm6,%xmm6 |
| movl %r11d,%r12d |
| shrdl $9,%r14d,%r14d |
| xorl %r10d,%r13d |
| vpshufb %xmm8,%xmm6,%xmm6 |
| xorl %eax,%r12d |
| shrdl $5,%r13d,%r13d |
| xorl %ecx,%r14d |
| vpaddd %xmm6,%xmm1,%xmm1 |
| andl %r10d,%r12d |
| xorl %r10d,%r13d |
| addl 24(%rsp),%ebx |
| vpshufd $80,%xmm1,%xmm7 |
| movl %ecx,%r15d |
| xorl %eax,%r12d |
| shrdl $11,%r14d,%r14d |
| vpsrld $10,%xmm7,%xmm6 |
| xorl %edx,%r15d |
| addl %r12d,%ebx |
| shrdl $6,%r13d,%r13d |
| vpsrlq $17,%xmm7,%xmm7 |
| andl %r15d,%edi |
| xorl %ecx,%r14d |
| addl %r13d,%ebx |
| vpxor %xmm7,%xmm6,%xmm6 |
| xorl %edx,%edi |
| shrdl $2,%r14d,%r14d |
| addl %ebx,%r9d |
| vpsrlq $2,%xmm7,%xmm7 |
| addl %edi,%ebx |
| movl %r9d,%r13d |
| addl %ebx,%r14d |
| vpxor %xmm7,%xmm6,%xmm6 |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%ebx |
| movl %r10d,%r12d |
| vpshufb %xmm9,%xmm6,%xmm6 |
| shrdl $9,%r14d,%r14d |
| xorl %r9d,%r13d |
| xorl %r11d,%r12d |
| vpaddd %xmm6,%xmm1,%xmm1 |
| shrdl $5,%r13d,%r13d |
| xorl %ebx,%r14d |
| andl %r9d,%r12d |
| vpaddd 32(%rbp),%xmm1,%xmm6 |
| xorl %r9d,%r13d |
| addl 28(%rsp),%eax |
| movl %ebx,%edi |
| xorl %r11d,%r12d |
| shrdl $11,%r14d,%r14d |
| xorl %ecx,%edi |
| addl %r12d,%eax |
| shrdl $6,%r13d,%r13d |
| andl %edi,%r15d |
| xorl %ebx,%r14d |
| addl %r13d,%eax |
| xorl %ecx,%r15d |
| shrdl $2,%r14d,%r14d |
| addl %eax,%r8d |
| addl %r15d,%eax |
| movl %r8d,%r13d |
| addl %eax,%r14d |
| vmovdqa %xmm6,16(%rsp) |
| vpalignr $4,%xmm2,%xmm3,%xmm4 |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%eax |
| movl %r9d,%r12d |
| vpalignr $4,%xmm0,%xmm1,%xmm7 |
| shrdl $9,%r14d,%r14d |
| xorl %r8d,%r13d |
| xorl %r10d,%r12d |
| vpsrld $7,%xmm4,%xmm6 |
| shrdl $5,%r13d,%r13d |
| xorl %eax,%r14d |
| andl %r8d,%r12d |
| vpaddd %xmm7,%xmm2,%xmm2 |
| xorl %r8d,%r13d |
| addl 32(%rsp),%r11d |
| movl %eax,%r15d |
| vpsrld $3,%xmm4,%xmm7 |
| xorl %r10d,%r12d |
| shrdl $11,%r14d,%r14d |
| xorl %ebx,%r15d |
| vpslld $14,%xmm4,%xmm5 |
| addl %r12d,%r11d |
| shrdl $6,%r13d,%r13d |
| andl %r15d,%edi |
| vpxor %xmm6,%xmm7,%xmm4 |
| xorl %eax,%r14d |
| addl %r13d,%r11d |
| xorl %ebx,%edi |
| vpshufd $250,%xmm1,%xmm7 |
| shrdl $2,%r14d,%r14d |
| addl %r11d,%edx |
| addl %edi,%r11d |
| vpsrld $11,%xmm6,%xmm6 |
| movl %edx,%r13d |
| addl %r11d,%r14d |
| shrdl $14,%r13d,%r13d |
| vpxor %xmm5,%xmm4,%xmm4 |
| movl %r14d,%r11d |
| movl %r8d,%r12d |
| shrdl $9,%r14d,%r14d |
| vpslld $11,%xmm5,%xmm5 |
| xorl %edx,%r13d |
| xorl %r9d,%r12d |
| shrdl $5,%r13d,%r13d |
| vpxor %xmm6,%xmm4,%xmm4 |
| xorl %r11d,%r14d |
| andl %edx,%r12d |
| xorl %edx,%r13d |
| vpsrld $10,%xmm7,%xmm6 |
| addl 36(%rsp),%r10d |
| movl %r11d,%edi |
| xorl %r9d,%r12d |
| vpxor %xmm5,%xmm4,%xmm4 |
| shrdl $11,%r14d,%r14d |
| xorl %eax,%edi |
| addl %r12d,%r10d |
| vpsrlq $17,%xmm7,%xmm7 |
| shrdl $6,%r13d,%r13d |
| andl %edi,%r15d |
| xorl %r11d,%r14d |
| vpaddd %xmm4,%xmm2,%xmm2 |
| addl %r13d,%r10d |
| xorl %eax,%r15d |
| shrdl $2,%r14d,%r14d |
| vpxor %xmm7,%xmm6,%xmm6 |
| addl %r10d,%ecx |
| addl %r15d,%r10d |
| movl %ecx,%r13d |
| vpsrlq $2,%xmm7,%xmm7 |
| addl %r10d,%r14d |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%r10d |
| vpxor %xmm7,%xmm6,%xmm6 |
| movl %edx,%r12d |
| shrdl $9,%r14d,%r14d |
| xorl %ecx,%r13d |
| vpshufb %xmm8,%xmm6,%xmm6 |
| xorl %r8d,%r12d |
| shrdl $5,%r13d,%r13d |
| xorl %r10d,%r14d |
| vpaddd %xmm6,%xmm2,%xmm2 |
| andl %ecx,%r12d |
| xorl %ecx,%r13d |
| addl 40(%rsp),%r9d |
| vpshufd $80,%xmm2,%xmm7 |
| movl %r10d,%r15d |
| xorl %r8d,%r12d |
| shrdl $11,%r14d,%r14d |
| vpsrld $10,%xmm7,%xmm6 |
| xorl %r11d,%r15d |
| addl %r12d,%r9d |
| shrdl $6,%r13d,%r13d |
| vpsrlq $17,%xmm7,%xmm7 |
| andl %r15d,%edi |
| xorl %r10d,%r14d |
| addl %r13d,%r9d |
| vpxor %xmm7,%xmm6,%xmm6 |
| xorl %r11d,%edi |
| shrdl $2,%r14d,%r14d |
| addl %r9d,%ebx |
| vpsrlq $2,%xmm7,%xmm7 |
| addl %edi,%r9d |
| movl %ebx,%r13d |
| addl %r9d,%r14d |
| vpxor %xmm7,%xmm6,%xmm6 |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%r9d |
| movl %ecx,%r12d |
| vpshufb %xmm9,%xmm6,%xmm6 |
| shrdl $9,%r14d,%r14d |
| xorl %ebx,%r13d |
| xorl %edx,%r12d |
| vpaddd %xmm6,%xmm2,%xmm2 |
| shrdl $5,%r13d,%r13d |
| xorl %r9d,%r14d |
| andl %ebx,%r12d |
| vpaddd 64(%rbp),%xmm2,%xmm6 |
| xorl %ebx,%r13d |
| addl 44(%rsp),%r8d |
| movl %r9d,%edi |
| xorl %edx,%r12d |
| shrdl $11,%r14d,%r14d |
| xorl %r10d,%edi |
| addl %r12d,%r8d |
| shrdl $6,%r13d,%r13d |
| andl %edi,%r15d |
| xorl %r9d,%r14d |
| addl %r13d,%r8d |
| xorl %r10d,%r15d |
| shrdl $2,%r14d,%r14d |
| addl %r8d,%eax |
| addl %r15d,%r8d |
| movl %eax,%r13d |
| addl %r8d,%r14d |
| vmovdqa %xmm6,32(%rsp) |
| vpalignr $4,%xmm3,%xmm0,%xmm4 |
| shrdl $14,%r13d,%r13d |
| movl %r14d,%r8d |
| movl %ebx,%r12d |
| vpalignr $4,%xmm1,%xmm2,%xmm7 |
| shrdl $9,%r14d,%r14d |
| xorl %eax,%r13d |
| xorl %ecx,%r12d |
| vpsrld $7,%xmm4,%xmm6 |
| shrdl $5,%r13d,%r13d |
| xorl %r8d,%r14d |
| andl %eax,%r12d |
| vpaddd %xmm7,%xmm3,%xmm3 |
| xorl %eax,%r13d |
| addl 48(%rsp),%edx |
| movl %r8d,%r15d |
| vpsrld $3,%xmm4,%xmm7 |
| xorl %ecx,%r12d |
| shrdl $11,%r14d,%r14d |
| xorl %r9d,%r15d |
| vpslld $14,%xmm4,%xmm5 |
| addl %r12d,%edx |
| shrdl $6,%r13d,%r13d |
| andl %r15d,%edi |
| vpxor %xmm6,%xmm7,%xmm4 |
| xorl %r8d,%r14d |
| addl %r13d,%edx |
| xorl %r9d,%edi |
| vpshufd $250,%xmm2,%xmm7 |
| shrdl $2,%r14d,%r14d |
| addl %edx,%r11d |
| addl %edi,%edx |
| vpsrld $11,%xmm6,%xmm6 |
| movl %r11d,%r13d |
| addl |