#include "ladder_base_namespace.h" #include "consts_namespace.h" # qhasm: int64 input_0 # qhasm: int64 input_1 # qhasm: int64 input_2 # qhasm: int64 input_3 # qhasm: int64 input_4 # qhasm: int64 input_5 # qhasm: stack64 input_6 # qhasm: stack64 input_7 # qhasm: int64 caller_r11 # qhasm: int64 caller_r12 # qhasm: int64 caller_r13 # qhasm: int64 caller_r14 # qhasm: int64 caller_r15 # qhasm: int64 caller_rbx # qhasm: int64 caller_rbp # qhasm: stack128 x3_0 # qhasm: stack128 x3_2 # qhasm: stack128 x3_4 # qhasm: stack128 x3_6 # qhasm: stack128 x3_8 # qhasm: stack128 z3_0 # qhasm: stack128 z3_2 # qhasm: stack128 z3_4 # qhasm: stack128 z3_6 # qhasm: stack128 z3_8 # qhasm: stack128 t0_0 # qhasm: stack128 t0_2 # qhasm: stack128 t0_4 # qhasm: stack128 t0_6 # qhasm: stack128 t0_8 # qhasm: reg128 init0 # qhasm: reg128 init1 # qhasm: reg128 init9 # qhasm: int64 b # qhasm: int64 b0 # qhasm: int64 tmp0 # qhasm: int64 b1 # qhasm: int64 tmp1 # qhasm: int64 b2 # qhasm: int64 tmp2 # qhasm: int64 b3 # qhasm: int64 tmp3 # qhasm: int64 byte # qhasm: int64 ptr # qhasm: int64 pos # qhasm: reg128 r # qhasm: reg128 r0 # qhasm: reg128 r1 # qhasm: reg128 r2 # qhasm: reg128 r3 # qhasm: reg128 r4 # qhasm: reg128 r5 # qhasm: reg128 r6 # qhasm: reg128 r7 # qhasm: reg128 r8 # qhasm: reg128 r9 # qhasm: reg128 f0 # qhasm: reg128 f1 # qhasm: reg128 f2 # qhasm: reg128 f3 # qhasm: reg128 f4 # qhasm: reg128 f5 # qhasm: reg128 f6 # qhasm: reg128 f7 # qhasm: reg128 f8 # qhasm: reg128 f9 # qhasm: stack128 f0_stack # qhasm: stack128 f1_stack # qhasm: stack128 f2_stack # qhasm: stack128 f3_stack # qhasm: stack128 f4_stack # qhasm: stack128 f5_stack # qhasm: stack128 f6_stack # qhasm: stack128 f7_stack # qhasm: stack128 f8_stack # qhasm: stack128 f9_stack # qhasm: stack128 f0_2_stack # qhasm: stack128 f1_2_stack # qhasm: stack128 f2_2_stack # qhasm: stack128 f3_2_stack # qhasm: stack128 f4_2_stack # qhasm: stack128 f5_2_stack # qhasm: stack128 f6_2_stack # qhasm: stack128 f7_2_stack # qhasm: stack128 f8_2_stack # qhasm: stack128 f9_2_stack # qhasm: reg128 f1_2 # qhasm: reg128 f2_2 # qhasm: reg128 f3_2 # qhasm: reg128 f4_2 # qhasm: reg128 f9_38 # qhasm: stack128 f5_38_stack # qhasm: stack128 f6_19_stack # qhasm: stack128 f7_38_stack # qhasm: stack128 f8_19_stack # qhasm: reg128 g0 # qhasm: reg128 g1 # qhasm: reg128 g2 # qhasm: reg128 g3 # qhasm: reg128 g4 # qhasm: reg128 g5 # qhasm: reg128 g6 # qhasm: reg128 g7 # qhasm: reg128 g8 # qhasm: reg128 g9 # qhasm: reg128 h0 # qhasm: reg128 h1 # qhasm: reg128 h2 # qhasm: reg128 h3 # qhasm: reg128 h4 # qhasm: reg128 h5 # qhasm: reg128 h6 # qhasm: reg128 h7 # qhasm: reg128 h8 # qhasm: reg128 h9 # qhasm: reg128 h1_2 # qhasm: reg128 h2_2 # qhasm: reg128 h3_2 # qhasm: reg128 h4_2 # qhasm: reg128 h9_38 # qhasm: reg128 carry0 # qhasm: reg128 carry1 # qhasm: reg128 carry2 # qhasm: reg128 carry3 # qhasm: reg128 carry4 # qhasm: reg128 carry5 # qhasm: reg128 carry6 # qhasm: reg128 carry7 # qhasm: reg128 carry8 # qhasm: reg128 carry9 # qhasm: stack128 s # qhasm: stack8192 masks # qhasm: reg128 mask # qhasm: reg128 diff # qhasm: stack128 h5_stack # qhasm: stack128 h6_stack # qhasm: stack128 h7_stack # qhasm: stack128 h8_stack # qhasm: stack128 h5_2_stack # qhasm: stack128 h6_2_stack # qhasm: stack128 h7_2_stack # qhasm: stack128 h8_2_stack # qhasm: stack128 h5_38_stack # qhasm: stack128 h6_19_stack # qhasm: stack128 h7_38_stack # qhasm: stack128 h8_19_stack # qhasm: reg128 m0 # qhasm: reg128 m1 # qhasm: reg128 m2 # qhasm: reg128 m3 # qhasm: reg128 m4 # qhasm: reg128 m5 # qhasm: reg128 m6 # qhasm: reg128 m7 # qhasm: reg128 m8 # qhasm: reg128 m9 # qhasm: reg128 h_01 # qhasm: reg128 h_23 # qhasm: reg128 h_45 # qhasm: reg128 h_67 # qhasm: reg128 h_89 # qhasm: int64 h_0 # qhasm: int64 h_1 # qhasm: int64 h_2 # qhasm: int64 h_3 # qhasm: int64 h_4 # qhasm: int64 h_5 # qhasm: int64 h_6 # qhasm: int64 h_7 # qhasm: int64 h_8 # qhasm: int64 h_9 # qhasm: int64 carry_0 # qhasm: int64 carry_1 # qhasm: int64 carry_2 # qhasm: int64 carry_3 # qhasm: int64 carry_4 # qhasm: int64 carry_5 # qhasm: int64 carry_6 # qhasm: int64 carry_7 # qhasm: int64 carry_8 # qhasm: int64 carry_9 # qhasm: stack128 buf0 # qhasm: stack128 buf1 # qhasm: stack128 buf2 # qhasm: stack128 buf3 # qhasm: stack128 buf4 # qhasm: stack128 buf5 # qhasm: stack128 buf6 # qhasm: stack128 buf7 # qhasm: stack128 buf8 # qhasm: stack128 buf9 # qhasm: stack64 r11_stack # qhasm: stack64 r12_stack # qhasm: stack64 r13_stack # qhasm: stack64 r14_stack # qhasm: stack64 r15_stack # qhasm: stack64 rbx_stack # qhasm: stack64 rbp_stack # qhasm: enter ladder_base .p2align 5 .global _ladder_base .global ladder_base _ladder_base: ladder_base: mov %rsp,%r11 and $31,%r11 add $1568,%r11 sub %r11,%rsp # qhasm: r11_stack = caller_r11 # asm 1: movq r11_stack=stack64#1 # asm 2: movq r11_stack=1536(%rsp) movq %r11,1536(%rsp) # qhasm: r12_stack = caller_r12 # asm 1: movq r12_stack=stack64#2 # asm 2: movq r12_stack=1544(%rsp) movq %r12,1544(%rsp) # qhasm: r13_stack = caller_r13 # asm 1: movq r13_stack=stack64#3 # asm 2: movq r13_stack=1552(%rsp) movq %r13,1552(%rsp) # qhasm: init0 = v0_0 # asm 1: movdqa v0_0,>init0=reg128#1 # asm 2: movdqa v0_0,>init0=%xmm0 movdqa v0_0(%rip),%xmm0 # qhasm: init1 = v1_0 # asm 1: movdqa v1_0,>init1=reg128#2 # asm 2: movdqa v1_0,>init1=%xmm1 movdqa v1_0(%rip),%xmm1 # qhasm: init9 = v9_0 # asm 1: movdqa v9_0,>init9=reg128#3 # asm 2: movdqa v9_0,>init9=%xmm2 movdqa v9_0(%rip),%xmm2 # qhasm: x3_0 = init9 # asm 1: movdqa x3_0=stack128#1 # asm 2: movdqa x3_0=0(%rsp) movdqa %xmm2,0(%rsp) # qhasm: x3_2 = init0 # asm 1: movdqa x3_2=stack128#2 # asm 2: movdqa x3_2=16(%rsp) movdqa %xmm0,16(%rsp) # qhasm: x3_4 = init0 # asm 1: movdqa x3_4=stack128#3 # asm 2: movdqa x3_4=32(%rsp) movdqa %xmm0,32(%rsp) # qhasm: x3_6 = init0 # asm 1: movdqa x3_6=stack128#4 # asm 2: movdqa x3_6=48(%rsp) movdqa %xmm0,48(%rsp) # qhasm: x3_8 = init0 # asm 1: movdqa x3_8=stack128#5 # asm 2: movdqa x3_8=64(%rsp) movdqa %xmm0,64(%rsp) # qhasm: z3_0 = init1 # asm 1: movdqa z3_0=stack128#6 # asm 2: movdqa z3_0=80(%rsp) movdqa %xmm1,80(%rsp) # qhasm: z3_2 = init0 # asm 1: movdqa z3_2=stack128#7 # asm 2: movdqa z3_2=96(%rsp) movdqa %xmm0,96(%rsp) # qhasm: z3_4 = init0 # asm 1: movdqa z3_4=stack128#8 # asm 2: movdqa z3_4=112(%rsp) movdqa %xmm0,112(%rsp) # qhasm: z3_6 = init0 # asm 1: movdqa z3_6=stack128#9 # asm 2: movdqa z3_6=128(%rsp) movdqa %xmm0,128(%rsp) # qhasm: z3_8 = init0 # asm 1: movdqa z3_8=stack128#10 # asm 2: movdqa z3_8=144(%rsp) movdqa %xmm0,144(%rsp) # qhasm: f0 = init1 # asm 1: movdqa f0=reg128#1 # asm 2: movdqa f0=%xmm0 movdqa %xmm1,%xmm0 # qhasm: f1 ^= f1 # asm 1: pxor >f1=reg128#2,>f1=reg128#2 # asm 2: pxor >f1=%xmm1,>f1=%xmm1 pxor %xmm1,%xmm1 # qhasm: f2 ^= f2 # asm 1: pxor >f2=reg128#3,>f2=reg128#3 # asm 2: pxor >f2=%xmm2,>f2=%xmm2 pxor %xmm2,%xmm2 # qhasm: f3 ^= f3 # asm 1: pxor >f3=reg128#4,>f3=reg128#4 # asm 2: pxor >f3=%xmm3,>f3=%xmm3 pxor %xmm3,%xmm3 # qhasm: f4 ^= f4 # asm 1: pxor >f4=reg128#5,>f4=reg128#5 # asm 2: pxor >f4=%xmm4,>f4=%xmm4 pxor %xmm4,%xmm4 # qhasm: f5 ^= f5 # asm 1: pxor >f5=reg128#6,>f5=reg128#6 # asm 2: pxor >f5=%xmm5,>f5=%xmm5 pxor %xmm5,%xmm5 # qhasm: f6 ^= f6 # asm 1: pxor >f6=reg128#7,>f6=reg128#7 # asm 2: pxor >f6=%xmm6,>f6=%xmm6 pxor %xmm6,%xmm6 # qhasm: f7 ^= f7 # asm 1: pxor >f7=reg128#8,>f7=reg128#8 # asm 2: pxor >f7=%xmm7,>f7=%xmm7 pxor %xmm7,%xmm7 # qhasm: f8 ^= f8 # asm 1: pxor >f8=reg128#9,>f8=reg128#9 # asm 2: pxor >f8=%xmm8,>f8=%xmm8 pxor %xmm8,%xmm8 # qhasm: f9 ^= f9 # asm 1: pxor >f9=reg128#10,>f9=reg128#10 # asm 2: pxor >f9=%xmm9,>f9=%xmm9 pxor %xmm9,%xmm9 # qhasm: b0 = mem64[input_1 + 0] # asm 1: movq 0(b0=int64#3 # asm 2: movq 0(b0=%rdx movq 0(%rsi),%rdx # qhasm: b1 = mem64[input_1 + 8] # asm 1: movq 8(b1=int64#4 # asm 2: movq 8(b1=%rcx movq 8(%rsi),%rcx # qhasm: b2 = mem64[input_1 + 16] # asm 1: movq 16(b2=int64#5 # asm 2: movq 16(b2=%r8 movq 16(%rsi),%r8 # qhasm: b3 = mem64[input_1 + 24] # asm 1: movq 24(b3=int64#6 # asm 2: movq 24(b3=%r9 movq 24(%rsi),%r9 # qhasm: b0 = (b1 b0) >> 1 # asm 1: shrd $1,> 1 # asm 1: shrd $1,> 1 # asm 1: shrd $1,>= 1 # asm 1: shr $1,ptr=int64#2 # asm 2: leaq ptr=%rsi leaq 512(%rsp),%rsi # qhasm: pos = 64 # asm 1: mov $64,>pos=int64#7 # asm 2: mov $64,>pos=%rax mov $64,%rax # qhasm: small_loop: ._small_loop: # qhasm: tmp0 = b0 # asm 1: mov tmp0=int64#8 # asm 2: mov tmp0=%r10 mov %rdx,%r10 # qhasm: tmp1 = b1 # asm 1: mov tmp1=int64#9 # asm 2: mov tmp1=%r11 mov %rcx,%r11 # qhasm: tmp2 = b2 # asm 1: mov tmp2=int64#10 # asm 2: mov tmp2=%r12 mov %r8,%r12 # qhasm: tmp3 = b3 # asm 1: mov tmp3=int64#11 # asm 2: mov tmp3=%r13 mov %r9,%r13 # qhasm: (uint64) b0 >>= 1 # asm 1: shr $1,>= 1 # asm 1: shr $1,>= 1 # asm 1: shr $1,>= 1 # asm 1: shr $1,pos=int64#3 # asm 2: mov $255,>pos=%rdx mov $255,%rdx # qhasm: ptr += 760 # asm 1: add $760,mask=reg128#11 # asm 2: vbroadcastss 0(mask=%xmm10 vbroadcastss 0(%rsi),%xmm10 # qhasm: ptr -= 4 # asm 1: sub $4,g0=reg128#12 # asm 2: movdqa g0=%xmm11 movdqa 0(%rsp),%xmm11 # qhasm: g1 = z3_0 # asm 1: movdqa g1=reg128#13 # asm 2: movdqa g1=%xmm12 movdqa 80(%rsp),%xmm12 # qhasm: diff = f0 ^ g0 # asm 1: vpxor diff=reg128#14 # asm 2: vpxor diff=%xmm13 vpxor %xmm11,%xmm0,%xmm13 # qhasm: diff &= mask # asm 1: pand diff=reg128#14 # asm 2: vpxor diff=%xmm13 vpxor %xmm12,%xmm1,%xmm13 # qhasm: diff &= mask # asm 1: pand g2=reg128#14 # asm 2: movdqa g2=%xmm13 movdqa 16(%rsp),%xmm13 # qhasm: g3 = z3_2 # asm 1: movdqa g3=reg128#15 # asm 2: movdqa g3=%xmm14 movdqa 96(%rsp),%xmm14 # qhasm: diff = f2 ^ g2 # asm 1: vpxor diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm13,%xmm2,%xmm15 # qhasm: diff &= mask # asm 1: pand diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm14,%xmm3,%xmm15 # qhasm: diff &= mask # asm 1: pand x3_2=stack128#1 # asm 2: movdqa x3_2=0(%rsp) movdqa %xmm13,0(%rsp) # qhasm: z3_2 = g3 # asm 1: movdqa z3_2=stack128#2 # asm 2: movdqa z3_2=16(%rsp) movdqa %xmm14,16(%rsp) # qhasm: g4 = x3_4 # asm 1: movdqa g4=reg128#14 # asm 2: movdqa g4=%xmm13 movdqa 32(%rsp),%xmm13 # qhasm: g5 = z3_4 # asm 1: movdqa g5=reg128#15 # asm 2: movdqa g5=%xmm14 movdqa 112(%rsp),%xmm14 # qhasm: diff = f4 ^ g4 # asm 1: vpxor diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm13,%xmm4,%xmm15 # qhasm: diff &= mask # asm 1: pand diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm14,%xmm5,%xmm15 # qhasm: diff &= mask # asm 1: pand x3_4=stack128#3 # asm 2: movdqa x3_4=32(%rsp) movdqa %xmm13,32(%rsp) # qhasm: z3_4 = g5 # asm 1: movdqa z3_4=stack128#6 # asm 2: movdqa z3_4=80(%rsp) movdqa %xmm14,80(%rsp) # qhasm: g6 = x3_6 # asm 1: movdqa g6=reg128#14 # asm 2: movdqa g6=%xmm13 movdqa 48(%rsp),%xmm13 # qhasm: g7 = z3_6 # asm 1: movdqa g7=reg128#15 # asm 2: movdqa g7=%xmm14 movdqa 128(%rsp),%xmm14 # qhasm: diff = f6 ^ g6 # asm 1: vpxor diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm13,%xmm6,%xmm15 # qhasm: diff &= mask # asm 1: pand diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm14,%xmm7,%xmm15 # qhasm: diff &= mask # asm 1: pand x3_6=stack128#4 # asm 2: movdqa x3_6=48(%rsp) movdqa %xmm13,48(%rsp) # qhasm: z3_6 = g7 # asm 1: movdqa z3_6=stack128#7 # asm 2: movdqa z3_6=96(%rsp) movdqa %xmm14,96(%rsp) # qhasm: g8 = x3_8 # asm 1: movdqa g8=reg128#14 # asm 2: movdqa g8=%xmm13 movdqa 64(%rsp),%xmm13 # qhasm: g9 = z3_8 # asm 1: movdqa g9=reg128#15 # asm 2: movdqa g9=%xmm14 movdqa 144(%rsp),%xmm14 # qhasm: diff = f8 ^ g8 # asm 1: vpxor diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm13,%xmm8,%xmm15 # qhasm: diff &= mask # asm 1: pand diff=reg128#16 # asm 2: vpxor diff=%xmm15 vpxor %xmm14,%xmm9,%xmm15 # qhasm: diff &= mask # asm 1: pand x3_8=stack128#5 # asm 2: movdqa x3_8=64(%rsp) movdqa %xmm13,64(%rsp) # qhasm: z3_8 = g9 # asm 1: movdqa z3_8=stack128#8 # asm 2: movdqa z3_8=112(%rsp) movdqa %xmm14,112(%rsp) # qhasm: 2x r = g0 + mem128[ subc0 ] # asm 1: vpaddq subc0,r=reg128#11 # asm 2: vpaddq subc0,r=%xmm10 vpaddq subc0(%rip),%xmm11,%xmm10 # qhasm: 2x r -= g1 # asm 1: psubq g1=reg128#13 # asm 2: vpunpckhqdq g1=%xmm12 vpunpckhqdq %xmm10,%xmm11,%xmm12 # qhasm: g0 = unpack_low(g0, r) # asm 1: vpunpcklqdq g0=reg128#11 # asm 2: vpunpcklqdq g0=%xmm10 vpunpcklqdq %xmm10,%xmm11,%xmm10 # qhasm: 2x r = f0 + f1 # asm 1: vpaddq r=reg128#12 # asm 2: vpaddq r=%xmm11 vpaddq %xmm1,%xmm0,%xmm11 # qhasm: 2x f0 += mem128[ subc0 ] # asm 1: paddq subc0,f1=reg128#2 # asm 2: vpunpckhqdq f1=%xmm1 vpunpckhqdq %xmm11,%xmm0,%xmm1 # qhasm: f0 = unpack_low(f0, r) # asm 1: vpunpcklqdq f0=reg128#1 # asm 2: vpunpcklqdq f0=%xmm0 vpunpcklqdq %xmm11,%xmm0,%xmm0 # qhasm: 2x h0 = g0 * f0 # asm 1: vpmuludq h0=reg128#12 # asm 2: vpmuludq h0=%xmm11 vpmuludq %xmm0,%xmm10,%xmm11 # qhasm: 2x h1 = g0 * f1 # asm 1: vpmuludq h1=reg128#14 # asm 2: vpmuludq h1=%xmm13 vpmuludq %xmm1,%xmm10,%xmm13 # qhasm: f1_stack = f1 # asm 1: movdqa f1_stack=stack128#9 # asm 2: movdqa f1_stack=128(%rsp) movdqa %xmm1,128(%rsp) # qhasm: 2x f1 += f1 # asm 1: paddq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm0,%xmm12,%xmm14 # qhasm: f0_stack = f0 # asm 1: movdqa f0_stack=stack128#10 # asm 2: movdqa f0_stack=144(%rsp) movdqa %xmm0,144(%rsp) # qhasm: 2x h1 += r # asm 1: paddq h2=reg128#1 # asm 2: vpmuludq h2=%xmm0 vpmuludq %xmm1,%xmm12,%xmm0 # qhasm: f1_2_stack = f1 # asm 1: movdqa f1_2_stack=stack128#11 # asm 2: movdqa f1_2_stack=160(%rsp) movdqa %xmm1,160(%rsp) # qhasm: 2x r = f2 + f3 # asm 1: vpaddq r=reg128#2 # asm 2: vpaddq r=%xmm1 vpaddq %xmm3,%xmm2,%xmm1 # qhasm: 2x f2 += mem128[ subc2 ] # asm 1: paddq subc2,f3=reg128#4 # asm 2: vpunpckhqdq f3=%xmm3 vpunpckhqdq %xmm1,%xmm2,%xmm3 # qhasm: f2 = unpack_low(f2, r) # asm 1: vpunpcklqdq f2=reg128#2 # asm 2: vpunpcklqdq f2=%xmm1 vpunpcklqdq %xmm1,%xmm2,%xmm1 # qhasm: 2x r = g0 * f2 # asm 1: vpmuludq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm1,%xmm10,%xmm2 # qhasm: 2x h2 += r # asm 1: paddq h3=reg128#3 # asm 2: vpmuludq h3=%xmm2 vpmuludq %xmm3,%xmm10,%xmm2 # qhasm: f3_stack = f3 # asm 1: movdqa f3_stack=stack128#12 # asm 2: movdqa f3_stack=176(%rsp) movdqa %xmm3,176(%rsp) # qhasm: 2x f3 += f3 # asm 1: paddq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm1,%xmm12,%xmm14 # qhasm: f2_stack = f2 # asm 1: movdqa f2_stack=stack128#13 # asm 2: movdqa f2_stack=192(%rsp) movdqa %xmm1,192(%rsp) # qhasm: 2x h3 += r # asm 1: paddq h4=reg128#2 # asm 2: vpmuludq h4=%xmm1 vpmuludq %xmm3,%xmm12,%xmm1 # qhasm: f3_2_stack = f3 # asm 1: movdqa f3_2_stack=stack128#14 # asm 2: movdqa f3_2_stack=208(%rsp) movdqa %xmm3,208(%rsp) # qhasm: 2x r = f4 + f5 # asm 1: vpaddq r=reg128#4 # asm 2: vpaddq r=%xmm3 vpaddq %xmm5,%xmm4,%xmm3 # qhasm: 2x f4 += mem128[ subc2 ] # asm 1: paddq subc2,f5=reg128#6 # asm 2: vpunpckhqdq f5=%xmm5 vpunpckhqdq %xmm3,%xmm4,%xmm5 # qhasm: f4 = unpack_low(f4, r) # asm 1: vpunpcklqdq f4=reg128#4 # asm 2: vpunpcklqdq f4=%xmm3 vpunpcklqdq %xmm3,%xmm4,%xmm3 # qhasm: 2x r = g0 * f4 # asm 1: vpmuludq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq %xmm3,%xmm10,%xmm4 # qhasm: 2x h4 += r # asm 1: paddq h5=reg128#5 # asm 2: vpmuludq h5=%xmm4 vpmuludq %xmm5,%xmm10,%xmm4 # qhasm: f5_stack = f5 # asm 1: movdqa f5_stack=stack128#15 # asm 2: movdqa f5_stack=224(%rsp) movdqa %xmm5,224(%rsp) # qhasm: 2x f5 += f5 # asm 1: paddq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm3,%xmm12,%xmm14 # qhasm: f4_stack = f4 # asm 1: movdqa f4_stack=stack128#16 # asm 2: movdqa f4_stack=240(%rsp) movdqa %xmm3,240(%rsp) # qhasm: 2x h5 += r # asm 1: paddq r=reg128#4 # asm 2: vpaddq r=%xmm3 vpaddq %xmm7,%xmm6,%xmm3 # qhasm: 2x f6 += mem128[ subc2 ] # asm 1: paddq subc2,f7=reg128#8 # asm 2: vpunpckhqdq f7=%xmm7 vpunpckhqdq %xmm3,%xmm6,%xmm7 # qhasm: f6 = unpack_low(f6, r) # asm 1: vpunpcklqdq f6=reg128#4 # asm 2: vpunpcklqdq f6=%xmm3 vpunpcklqdq %xmm3,%xmm6,%xmm3 # qhasm: 2x h6 = g0 * f6 # asm 1: vpmuludq h6=reg128#7 # asm 2: vpmuludq h6=%xmm6 vpmuludq %xmm3,%xmm10,%xmm6 # qhasm: 2x r = g1 * f5 # asm 1: vpmuludq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm5,%xmm12,%xmm14 # qhasm: f5_2_stack = f5 # asm 1: movdqa f5_2_stack=stack128#17 # asm 2: movdqa f5_2_stack=256(%rsp) movdqa %xmm5,256(%rsp) # qhasm: 2x f5 *= mem128[ v19_19 ] # asm 1: pmuludq v19_19,f5_38_stack=stack128#18 # asm 2: movdqa f5_38_stack=272(%rsp) movdqa %xmm5,272(%rsp) # qhasm: 2x h6 += r # asm 1: paddq h7=reg128#6 # asm 2: vpmuludq h7=%xmm5 vpmuludq %xmm7,%xmm10,%xmm5 # qhasm: f7_stack = f7 # asm 1: movdqa f7_stack=stack128#19 # asm 2: movdqa f7_stack=288(%rsp) movdqa %xmm7,288(%rsp) # qhasm: 2x f7 += f7 # asm 1: paddq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm3,%xmm12,%xmm14 # qhasm: f6_stack = f6 # asm 1: movdqa f6_stack=stack128#20 # asm 2: movdqa f6_stack=304(%rsp) movdqa %xmm3,304(%rsp) # qhasm: 2x h7 += r # asm 1: paddq f6_19_stack=stack128#21 # asm 2: movdqa f6_19_stack=320(%rsp) movdqa %xmm3,320(%rsp) # qhasm: 2x r = f8 + f9 # asm 1: vpaddq r=reg128#4 # asm 2: vpaddq r=%xmm3 vpaddq %xmm9,%xmm8,%xmm3 # qhasm: 2x f8 += mem128[ subc2 ] # asm 1: paddq subc2,f9=reg128#10 # asm 2: vpunpckhqdq f9=%xmm9 vpunpckhqdq %xmm3,%xmm8,%xmm9 # qhasm: f8 = unpack_low(f8, r) # asm 1: vpunpcklqdq f8=reg128#4 # asm 2: vpunpcklqdq f8=%xmm3 vpunpcklqdq %xmm3,%xmm8,%xmm3 # qhasm: f8_stack = f8 # asm 1: movdqa f8_stack=stack128#22 # asm 2: movdqa f8_stack=336(%rsp) movdqa %xmm3,336(%rsp) # qhasm: 2x h8 = g1 * f7 # asm 1: vpmuludq h8=reg128#9 # asm 2: vpmuludq h8=%xmm8 vpmuludq %xmm7,%xmm12,%xmm8 # qhasm: f7_2_stack = f7 # asm 1: movdqa f7_2_stack=stack128#23 # asm 2: movdqa f7_2_stack=352(%rsp) movdqa %xmm7,352(%rsp) # qhasm: 2x f7 *= mem128[ v19_19 ] # asm 1: pmuludq v19_19,f7_38_stack=stack128#24 # asm 2: movdqa f7_38_stack=368(%rsp) movdqa %xmm7,368(%rsp) # qhasm: 2x r = g0 * f8 # asm 1: vpmuludq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq %xmm3,%xmm10,%xmm7 # qhasm: 2x h8 += r # asm 1: paddq h9=reg128#8 # asm 2: vpmuludq h9=%xmm7 vpmuludq %xmm9,%xmm10,%xmm7 # qhasm: f9_stack = f9 # asm 1: movdqa f9_stack=stack128#25 # asm 2: movdqa f9_stack=384(%rsp) movdqa %xmm9,384(%rsp) # qhasm: 2x f9 += f9 # asm 1: paddq r=reg128#11 # asm 2: vpmuludq r=%xmm10 vpmuludq %xmm3,%xmm12,%xmm10 # qhasm: 2x h9 += r # asm 1: paddq f8_19_stack=stack128#26 # asm 2: movdqa f8_19_stack=400(%rsp) movdqa %xmm3,400(%rsp) # qhasm: 2x g1 *= mem128[ v19_19 ] # asm 1: pmuludq v19_19,r=reg128#4 # asm 2: vpmuludq r=%xmm3 vpmuludq %xmm9,%xmm12,%xmm3 # qhasm: f9_2_stack = f9 # asm 1: movdqa f9_2_stack=stack128#27 # asm 2: movdqa f9_2_stack=416(%rsp) movdqa %xmm9,416(%rsp) # qhasm: 2x h0 += r # asm 1: paddq g2=reg128#4 # asm 2: movdqa g2=%xmm3 movdqa 0(%rsp),%xmm3 # qhasm: g3 = z3_2 # asm 1: movdqa g3=reg128#10 # asm 2: movdqa g3=%xmm9 movdqa 16(%rsp),%xmm9 # qhasm: 2x r = g2 + mem128[ subc2 ] # asm 1: vpaddq subc2,r=reg128#11 # asm 2: vpaddq subc2,r=%xmm10 vpaddq subc2(%rip),%xmm3,%xmm10 # qhasm: 2x r -= g3 # asm 1: psubq g3=reg128#10 # asm 2: vpunpckhqdq g3=%xmm9 vpunpckhqdq %xmm10,%xmm3,%xmm9 # qhasm: g2 = unpack_low(g2, r) # asm 1: vpunpcklqdq g2=reg128#4 # asm 2: vpunpcklqdq g2=%xmm3 vpunpcklqdq %xmm10,%xmm3,%xmm3 # qhasm: 2x r2 = g2 * f0_stack # asm 1: vpmuludq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 144(%rsp),%xmm3,%xmm10 # qhasm: 2x h2 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 128(%rsp),%xmm3,%xmm10 # qhasm: 2x h3 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 192(%rsp),%xmm3,%xmm10 # qhasm: 2x h4 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 176(%rsp),%xmm3,%xmm10 # qhasm: 2x h5 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 240(%rsp),%xmm3,%xmm10 # qhasm: 2x h6 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 224(%rsp),%xmm3,%xmm10 # qhasm: 2x h7 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 304(%rsp),%xmm3,%xmm10 # qhasm: 2x h8 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 288(%rsp),%xmm3,%xmm10 # qhasm: 2x h9 += r2 # asm 1: paddq r2=reg128#11 # asm 2: vpmuludq r2=%xmm10 vpmuludq 336(%rsp),%xmm3,%xmm10 # qhasm: 2x h0 += r2 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 144(%rsp),%xmm9,%xmm3 # qhasm: 2x h3 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 160(%rsp),%xmm9,%xmm3 # qhasm: 2x h4 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 192(%rsp),%xmm9,%xmm3 # qhasm: 2x h5 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 208(%rsp),%xmm9,%xmm3 # qhasm: 2x h6 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 240(%rsp),%xmm9,%xmm3 # qhasm: 2x h7 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 256(%rsp),%xmm9,%xmm3 # qhasm: 2x h8 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 304(%rsp),%xmm9,%xmm3 # qhasm: 2x h9 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 352(%rsp),%xmm9,%xmm3 # qhasm: 2x h0 += r3 # asm 1: paddq r3=reg128#4 # asm 2: vpmuludq r3=%xmm3 vpmuludq 336(%rsp),%xmm9,%xmm3 # qhasm: 2x h1 += r3 # asm 1: paddq g4=reg128#4 # asm 2: movdqa g4=%xmm3 movdqa 32(%rsp),%xmm3 # qhasm: g5 = z3_4 # asm 1: movdqa g5=reg128#10 # asm 2: movdqa g5=%xmm9 movdqa 80(%rsp),%xmm9 # qhasm: 2x r = g4 + mem128[ subc2 ] # asm 1: vpaddq subc2,r=reg128#11 # asm 2: vpaddq subc2,r=%xmm10 vpaddq subc2(%rip),%xmm3,%xmm10 # qhasm: 2x r -= g5 # asm 1: psubq g5=reg128#10 # asm 2: vpunpckhqdq g5=%xmm9 vpunpckhqdq %xmm10,%xmm3,%xmm9 # qhasm: g4 = unpack_low(g4, r) # asm 1: vpunpcklqdq g4=reg128#4 # asm 2: vpunpcklqdq g4=%xmm3 vpunpcklqdq %xmm10,%xmm3,%xmm3 # qhasm: 2x r4 = g4 * f0_stack # asm 1: vpmuludq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 144(%rsp),%xmm3,%xmm10 # qhasm: 2x h4 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 128(%rsp),%xmm3,%xmm10 # qhasm: 2x h5 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 192(%rsp),%xmm3,%xmm10 # qhasm: 2x h6 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 176(%rsp),%xmm3,%xmm10 # qhasm: 2x h7 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 240(%rsp),%xmm3,%xmm10 # qhasm: 2x h8 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 224(%rsp),%xmm3,%xmm10 # qhasm: 2x h9 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 304(%rsp),%xmm3,%xmm10 # qhasm: 2x h0 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 288(%rsp),%xmm3,%xmm10 # qhasm: 2x h1 += r4 # asm 1: paddq r4=reg128#11 # asm 2: vpmuludq r4=%xmm10 vpmuludq 336(%rsp),%xmm3,%xmm10 # qhasm: 2x h2 += r4 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 144(%rsp),%xmm9,%xmm3 # qhasm: 2x h5 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 160(%rsp),%xmm9,%xmm3 # qhasm: 2x h6 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 192(%rsp),%xmm9,%xmm3 # qhasm: 2x h7 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 208(%rsp),%xmm9,%xmm3 # qhasm: 2x h8 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 240(%rsp),%xmm9,%xmm3 # qhasm: 2x h9 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 256(%rsp),%xmm9,%xmm3 # qhasm: 2x h0 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 304(%rsp),%xmm9,%xmm3 # qhasm: 2x h1 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 352(%rsp),%xmm9,%xmm3 # qhasm: 2x h2 += r5 # asm 1: paddq r5=reg128#4 # asm 2: vpmuludq r5=%xmm3 vpmuludq 336(%rsp),%xmm9,%xmm3 # qhasm: 2x h3 += r5 # asm 1: paddq g6=reg128#4 # asm 2: movdqa g6=%xmm3 movdqa 48(%rsp),%xmm3 # qhasm: g7 = z3_6 # asm 1: movdqa g7=reg128#10 # asm 2: movdqa g7=%xmm9 movdqa 96(%rsp),%xmm9 # qhasm: 2x r = g6 + mem128[ subc2 ] # asm 1: vpaddq subc2,r=reg128#11 # asm 2: vpaddq subc2,r=%xmm10 vpaddq subc2(%rip),%xmm3,%xmm10 # qhasm: 2x r -= g7 # asm 1: psubq g7=reg128#10 # asm 2: vpunpckhqdq g7=%xmm9 vpunpckhqdq %xmm10,%xmm3,%xmm9 # qhasm: g6 = unpack_low(g6, r) # asm 1: vpunpcklqdq g6=reg128#4 # asm 2: vpunpcklqdq g6=%xmm3 vpunpcklqdq %xmm10,%xmm3,%xmm3 # qhasm: 2x r6 = g6 * f0_stack # asm 1: vpmuludq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 144(%rsp),%xmm3,%xmm10 # qhasm: 2x h6 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 128(%rsp),%xmm3,%xmm10 # qhasm: 2x h7 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 192(%rsp),%xmm3,%xmm10 # qhasm: 2x h8 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 176(%rsp),%xmm3,%xmm10 # qhasm: 2x h9 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 240(%rsp),%xmm3,%xmm10 # qhasm: 2x h0 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 224(%rsp),%xmm3,%xmm10 # qhasm: 2x h1 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 304(%rsp),%xmm3,%xmm10 # qhasm: 2x h2 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 288(%rsp),%xmm3,%xmm10 # qhasm: 2x h3 += r6 # asm 1: paddq r6=reg128#11 # asm 2: vpmuludq r6=%xmm10 vpmuludq 336(%rsp),%xmm3,%xmm10 # qhasm: 2x h4 += r6 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 144(%rsp),%xmm9,%xmm3 # qhasm: 2x h7 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 160(%rsp),%xmm9,%xmm3 # qhasm: 2x h8 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 192(%rsp),%xmm9,%xmm3 # qhasm: 2x h9 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 208(%rsp),%xmm9,%xmm3 # qhasm: 2x h0 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 240(%rsp),%xmm9,%xmm3 # qhasm: 2x h1 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 256(%rsp),%xmm9,%xmm3 # qhasm: 2x h2 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 304(%rsp),%xmm9,%xmm3 # qhasm: 2x h3 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 352(%rsp),%xmm9,%xmm3 # qhasm: 2x h4 += r7 # asm 1: paddq r7=reg128#4 # asm 2: vpmuludq r7=%xmm3 vpmuludq 336(%rsp),%xmm9,%xmm3 # qhasm: 2x h5 += r7 # asm 1: paddq g8=reg128#4 # asm 2: movdqa g8=%xmm3 movdqa 64(%rsp),%xmm3 # qhasm: g9 = z3_8 # asm 1: movdqa g9=reg128#10 # asm 2: movdqa g9=%xmm9 movdqa 112(%rsp),%xmm9 # qhasm: 2x r = g8 + mem128[ subc2 ] # asm 1: vpaddq subc2,r=reg128#11 # asm 2: vpaddq subc2,r=%xmm10 vpaddq subc2(%rip),%xmm3,%xmm10 # qhasm: 2x r -= g9 # asm 1: psubq g9=reg128#10 # asm 2: vpunpckhqdq g9=%xmm9 vpunpckhqdq %xmm10,%xmm3,%xmm9 # qhasm: g8 = unpack_low(g8, r) # asm 1: vpunpcklqdq g8=reg128#4 # asm 2: vpunpcklqdq g8=%xmm3 vpunpcklqdq %xmm10,%xmm3,%xmm3 # qhasm: 2x r8 = g8 * f0_stack # asm 1: vpmuludq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 144(%rsp),%xmm3,%xmm10 # qhasm: 2x h8 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 128(%rsp),%xmm3,%xmm10 # qhasm: 2x h9 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 192(%rsp),%xmm3,%xmm10 # qhasm: 2x h0 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 176(%rsp),%xmm3,%xmm10 # qhasm: 2x h1 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 240(%rsp),%xmm3,%xmm10 # qhasm: 2x h2 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 224(%rsp),%xmm3,%xmm10 # qhasm: 2x h3 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 304(%rsp),%xmm3,%xmm10 # qhasm: 2x h4 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 288(%rsp),%xmm3,%xmm10 # qhasm: 2x h5 += r8 # asm 1: paddq r8=reg128#11 # asm 2: vpmuludq r8=%xmm10 vpmuludq 336(%rsp),%xmm3,%xmm10 # qhasm: 2x h6 += r8 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 144(%rsp),%xmm9,%xmm3 # qhasm: 2x h9 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 160(%rsp),%xmm9,%xmm3 # qhasm: 2x h0 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 192(%rsp),%xmm9,%xmm3 # qhasm: 2x h1 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 208(%rsp),%xmm9,%xmm3 # qhasm: 2x h2 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 240(%rsp),%xmm9,%xmm3 # qhasm: 2x h3 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 256(%rsp),%xmm9,%xmm3 # qhasm: 2x h4 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 304(%rsp),%xmm9,%xmm3 # qhasm: 2x h5 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 352(%rsp),%xmm9,%xmm3 # qhasm: 2x h6 += r9 # asm 1: paddq r9=reg128#4 # asm 2: vpmuludq r9=%xmm3 vpmuludq 336(%rsp),%xmm9,%xmm3 # qhasm: 2x h7 += r9 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#4 # asm 2: vpsrlq $25,carry5=%xmm3 vpsrlq $25,%xmm4,%xmm3 # qhasm: 2x h6 += carry5 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry0=reg128#4 # asm 2: vpsrlq $26,carry0=%xmm3 vpsrlq $26,%xmm11,%xmm3 # qhasm: 2x h1 += carry0 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry6=reg128#4 # asm 2: vpsrlq $26,carry6=%xmm3 vpsrlq $26,%xmm6,%xmm3 # qhasm: 2x h7 += carry6 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry1=reg128#4 # asm 2: vpsrlq $25,carry1=%xmm3 vpsrlq $25,%xmm13,%xmm3 # qhasm: 2x h2 += carry1 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry7=reg128#4 # asm 2: vpsrlq $25,carry7=%xmm3 vpsrlq $25,%xmm5,%xmm3 # qhasm: 2x h8 += carry7 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry2=reg128#4 # asm 2: vpsrlq $26,carry2=%xmm3 vpsrlq $26,%xmm0,%xmm3 # qhasm: 2x h3 += carry2 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry8=reg128#4 # asm 2: vpsrlq $26,carry8=%xmm3 vpsrlq $26,%xmm8,%xmm3 # qhasm: 2x h9 += carry8 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry3=reg128#4 # asm 2: vpsrlq $25,carry3=%xmm3 vpsrlq $25,%xmm2,%xmm3 # qhasm: 2x h4 += carry3 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry9=reg128#4 # asm 2: vpsrlq $25,carry9=%xmm3 vpsrlq $25,%xmm7,%xmm3 # qhasm: 2x r0 = carry9 << 4 # asm 1: vpsllq $4,r0=reg128#10 # asm 2: vpsllq $4,r0=%xmm9 vpsllq $4,%xmm3,%xmm9 # qhasm: 2x h0 += carry9 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry4=reg128#4 # asm 2: vpsrlq $26,carry4=%xmm3 vpsrlq $26,%xmm1,%xmm3 # qhasm: 2x h5 += carry4 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry0=reg128#4 # asm 2: vpsrlq $26,carry0=%xmm3 vpsrlq $26,%xmm11,%xmm3 # qhasm: 2x h1 += carry0 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#4 # asm 2: vpsrlq $25,carry5=%xmm3 vpsrlq $25,%xmm4,%xmm3 # qhasm: 2x h6 += carry5 # asm 1: paddq r=reg128#4 # asm 2: vpunpcklqdq r=%xmm3 vpunpcklqdq %xmm13,%xmm11,%xmm3 # qhasm: h1 = unpack_high(h0, h1) # asm 1: vpunpckhqdq h1=reg128#10 # asm 2: vpunpckhqdq h1=%xmm9 vpunpckhqdq %xmm13,%xmm11,%xmm9 # qhasm: 2x h0 = h1 + mem128[ subc0 ] # asm 1: vpaddq subc0,h0=reg128#11 # asm 2: vpaddq subc0,h0=%xmm10 vpaddq subc0(%rip),%xmm9,%xmm10 # qhasm: 2x h0 -= r # asm 1: psubq h1=reg128#10 # asm 2: vpunpckhqdq h1=%xmm9 vpunpckhqdq %xmm3,%xmm10,%xmm9 # qhasm: unpack low qwords of h0 and r # asm 1: punpcklqdq m0=reg128#4 # asm 2: vpmuludq m0=%xmm3 vpmuludq %xmm10,%xmm10,%xmm3 # qhasm: 2x h0 += h0 # asm 1: paddq m1=reg128#12 # asm 2: vpmuludq m1=%xmm11 vpmuludq %xmm9,%xmm10,%xmm11 # qhasm: r = unpack_low(h2, h3) # asm 1: vpunpcklqdq r=reg128#13 # asm 2: vpunpcklqdq r=%xmm12 vpunpcklqdq %xmm2,%xmm0,%xmm12 # qhasm: h3 = unpack_high(h2, h3) # asm 1: vpunpckhqdq h3=reg128#1 # asm 2: vpunpckhqdq h3=%xmm0 vpunpckhqdq %xmm2,%xmm0,%xmm0 # qhasm: 2x h2 = h3 + mem128[ subc2 ] # asm 1: vpaddq subc2,h2=reg128#3 # asm 2: vpaddq subc2,h2=%xmm2 vpaddq subc2(%rip),%xmm0,%xmm2 # qhasm: 2x h2 -= r # asm 1: psubq h3=reg128#1 # asm 2: vpunpckhqdq h3=%xmm0 vpunpckhqdq %xmm12,%xmm2,%xmm0 # qhasm: unpack low qwords of h2 and r # asm 1: punpcklqdq m2=reg128#13 # asm 2: vpmuludq m2=%xmm12 vpmuludq %xmm2,%xmm10,%xmm12 # qhasm: 2x h1_2 = h1 + h1 # asm 1: vpaddq h1_2=reg128#14 # asm 2: vpaddq h1_2=%xmm13 vpaddq %xmm9,%xmm9,%xmm13 # qhasm: 2x r = h1 * h1_2 # asm 1: vpmuludq r=reg128#10 # asm 2: vpmuludq r=%xmm9 vpmuludq %xmm13,%xmm9,%xmm9 # qhasm: 2x m2 += r # asm 1: paddq m3=reg128#10 # asm 2: vpmuludq m3=%xmm9 vpmuludq %xmm0,%xmm10,%xmm9 # qhasm: 2x r = h1_2 * h2 # asm 1: vpmuludq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm2,%xmm13,%xmm14 # qhasm: 2x m3 += r # asm 1: paddq r=reg128#15 # asm 2: vpunpcklqdq r=%xmm14 vpunpcklqdq %xmm4,%xmm1,%xmm14 # qhasm: h5 = unpack_high(h4, h5) # asm 1: vpunpckhqdq h5=reg128#2 # asm 2: vpunpckhqdq h5=%xmm1 vpunpckhqdq %xmm4,%xmm1,%xmm1 # qhasm: 2x h4 = h5 + mem128[ subc2 ] # asm 1: vpaddq subc2,h4=reg128#5 # asm 2: vpaddq subc2,h4=%xmm4 vpaddq subc2(%rip),%xmm1,%xmm4 # qhasm: 2x h4 -= r # asm 1: psubq h5=reg128#2 # asm 2: vpunpckhqdq h5=%xmm1 vpunpckhqdq %xmm14,%xmm4,%xmm1 # qhasm: unpack low qwords of h4 and r # asm 1: punpcklqdq h5_stack=stack128#1 # asm 2: movdqa h5_stack=0(%rsp) movdqa %xmm1,0(%rsp) # qhasm: 2x h5 += h5 # asm 1: paddq h5_2_stack=stack128#2 # asm 2: movdqa h5_2_stack=16(%rsp) movdqa %xmm1,16(%rsp) # qhasm: 2x h5 *= mem128[ v19_19 ] # asm 1: pmuludq v19_19,h5_38_stack=stack128#3 # asm 2: movdqa h5_38_stack=32(%rsp) movdqa %xmm1,32(%rsp) # qhasm: 2x m4 = h0 * h4 # asm 1: vpmuludq m4=reg128#2 # asm 2: vpmuludq m4=%xmm1 vpmuludq %xmm4,%xmm10,%xmm1 # qhasm: 2x r = h2 * h2 # asm 1: vpmuludq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm2,%xmm2,%xmm14 # qhasm: 2x m4 += r # asm 1: paddq m5=reg128#15 # asm 2: vpmuludq m5=%xmm14 vpmuludq 0(%rsp),%xmm10,%xmm14 # qhasm: 2x r = h1_2 * h4 # asm 1: vpmuludq r=reg128#16 # asm 2: vpmuludq r=%xmm15 vpmuludq %xmm4,%xmm13,%xmm15 # qhasm: 2x m5 += r # asm 1: paddq r=reg128#16 # asm 2: vpunpcklqdq r=%xmm15 vpunpcklqdq %xmm5,%xmm6,%xmm15 # qhasm: h7 = unpack_high(h6, h7) # asm 1: vpunpckhqdq h7=reg128#6 # asm 2: vpunpckhqdq h7=%xmm5 vpunpckhqdq %xmm5,%xmm6,%xmm5 # qhasm: 2x h6 = h7 + mem128[ subc2 ] # asm 1: vpaddq subc2,h6=reg128#7 # asm 2: vpaddq subc2,h6=%xmm6 vpaddq subc2(%rip),%xmm5,%xmm6 # qhasm: 2x h6 -= r # asm 1: psubq h7=reg128#6 # asm 2: vpunpckhqdq h7=%xmm5 vpunpckhqdq %xmm15,%xmm6,%xmm5 # qhasm: unpack low qwords of h6 and r # asm 1: punpcklqdq h6_stack=stack128#4 # asm 2: movdqa h6_stack=48(%rsp) movdqa %xmm6,48(%rsp) # qhasm: 2x h6 *= mem128[ v19_19 ] # asm 1: pmuludq v19_19,h6_19_stack=stack128#5 # asm 2: movdqa h6_19_stack=64(%rsp) movdqa %xmm6,64(%rsp) # qhasm: h7_stack = h7 # asm 1: movdqa h7_stack=stack128#6 # asm 2: movdqa h7_stack=80(%rsp) movdqa %xmm5,80(%rsp) # qhasm: 2x h7 *= mem128[ v38_38 ] # asm 1: pmuludq v38_38,h7_38_stack=stack128#7 # asm 2: movdqa h7_38_stack=96(%rsp) movdqa %xmm5,96(%rsp) # qhasm: 2x m6 = h0 * h6_stack # asm 1: vpmuludq m6=reg128#6 # asm 2: vpmuludq m6=%xmm5 vpmuludq 48(%rsp),%xmm10,%xmm5 # qhasm: 2x h3_2 = h3 + h3 # asm 1: vpaddq h3_2=reg128#7 # asm 2: vpaddq h3_2=%xmm6 vpaddq %xmm0,%xmm0,%xmm6 # qhasm: 2x r = h3 * h3_2 # asm 1: vpmuludq r=reg128#1 # asm 2: vpmuludq r=%xmm0 vpmuludq %xmm6,%xmm0,%xmm0 # qhasm: 2x m6 += r # asm 1: paddq m7=reg128#1 # asm 2: vpmuludq m7=%xmm0 vpmuludq 80(%rsp),%xmm10,%xmm0 # qhasm: 2x r = h3_2 * h4 # asm 1: vpmuludq r=reg128#16 # asm 2: vpmuludq r=%xmm15 vpmuludq %xmm4,%xmm6,%xmm15 # qhasm: 2x m7 += r # asm 1: paddq r=reg128#16 # asm 2: vpmuludq r=%xmm15 vpmuludq %xmm6,%xmm13,%xmm15 # qhasm: 2x m4 += r # asm 1: paddq r=reg128#16 # asm 2: vpmuludq r=%xmm15 vpmuludq %xmm6,%xmm2,%xmm15 # qhasm: 2x m5 += r # asm 1: paddq r=reg128#16 # asm 2: vpunpcklqdq r=%xmm15 vpunpcklqdq %xmm7,%xmm8,%xmm15 # qhasm: h9 = unpack_high(h8, h9) # asm 1: vpunpckhqdq h9=reg128#8 # asm 2: vpunpckhqdq h9=%xmm7 vpunpckhqdq %xmm7,%xmm8,%xmm7 # qhasm: 2x h8 = h9 + mem128[ subc2 ] # asm 1: vpaddq subc2,h8=reg128#9 # asm 2: vpaddq subc2,h8=%xmm8 vpaddq subc2(%rip),%xmm7,%xmm8 # qhasm: 2x h8 -= r # asm 1: psubq h9=reg128#8 # asm 2: vpunpckhqdq h9=%xmm7 vpunpckhqdq %xmm15,%xmm8,%xmm7 # qhasm: unpack low qwords of h8 and r # asm 1: punpcklqdq h8_stack=stack128#8 # asm 2: movdqa h8_stack=112(%rsp) movdqa %xmm8,112(%rsp) # qhasm: 2x h8 *= mem128[ v19_19 ] # asm 1: pmuludq v19_19,h8_19_stack=stack128#11 # asm 2: movdqa h8_19_stack=160(%rsp) movdqa %xmm8,160(%rsp) # qhasm: 2x m8 = h0 * h8_stack # asm 1: vpmuludq m8=reg128#9 # asm 2: vpmuludq m8=%xmm8 vpmuludq 112(%rsp),%xmm10,%xmm8 # qhasm: 2x m9 = h0 * h9 # asm 1: vpmuludq m9=reg128#11 # asm 2: vpmuludq m9=%xmm10 vpmuludq %xmm7,%xmm10,%xmm10 # qhasm: 2x h9_38 = h9 * mem128[ v38_38 ] # asm 1: vpmuludq v38_38,h9_38=reg128#16 # asm 2: vpmuludq v38_38,h9_38=%xmm15 vpmuludq v38_38(%rip),%xmm7,%xmm15 # qhasm: 2x r = h9 * h9_38 # asm 1: vpmuludq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq %xmm15,%xmm7,%xmm7 # qhasm: 2x m8 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq %xmm15,%xmm13,%xmm7 # qhasm: 2x m0 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq %xmm15,%xmm2,%xmm7 # qhasm: 2x m1 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 80(%rsp),%xmm13,%xmm7 # qhasm: 2x r += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 16(%rsp),%xmm13,%xmm7 # qhasm: 2x m6 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 48(%rsp),%xmm13,%xmm7 # qhasm: 2x m7 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 112(%rsp),%xmm13,%xmm7 # qhasm: 2x m9 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq %xmm15,%xmm6,%xmm7 # qhasm: 2x m2 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq %xmm15,%xmm4,%xmm7 # qhasm: 2x m3 += r # asm 1: paddq h2_2=reg128#3 # asm 2: vpaddq h2_2=%xmm2 vpaddq %xmm2,%xmm2,%xmm2 # qhasm: 2x r = h2_2 * h4 # asm 1: vpmuludq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq %xmm4,%xmm2,%xmm7 # qhasm: 2x m6 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 160(%rsp),%xmm2,%xmm7 # qhasm: 2x m0 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 160(%rsp),%xmm6,%xmm7 # qhasm: 2x m1 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 0(%rsp),%xmm2,%xmm7 # qhasm: 2x m7 += r # asm 1: paddq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 48(%rsp),%xmm2,%xmm7 # qhasm: 2x m8 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 80(%rsp),%xmm2,%xmm2 # qhasm: 2x m9 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 96(%rsp),%xmm4,%xmm2 # qhasm: 2x m1 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm4,%xmm4,%xmm2 # qhasm: 2x m8 += r # asm 1: paddq h4_2=reg128#3 # asm 2: vpaddq h4_2=%xmm2 vpaddq %xmm4,%xmm4,%xmm2 # qhasm: 2x r = h4_2 * h8_19_stack # asm 1: vpmuludq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 160(%rsp),%xmm2,%xmm4 # qhasm: 2x m2 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 16(%rsp),%xmm15,%xmm4 # qhasm: 2x m4 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 48(%rsp),%xmm15,%xmm4 # qhasm: 2x m5 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 96(%rsp),%xmm6,%xmm4 # qhasm: 2x m0 += r # asm 1: paddq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 16(%rsp),%xmm4 # qhasm: 2x r *= h8_19_stack # asm 1: pmuludq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 16(%rsp),%xmm6,%xmm4 # qhasm: 2x m8 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 48(%rsp),%xmm6,%xmm4 # qhasm: 2x m9 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 80(%rsp),%xmm15,%xmm4 # qhasm: 2x r += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 112(%rsp),%xmm15,%xmm4 # qhasm: 2x m7 += r # asm 1: paddq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 48(%rsp),%xmm4 # qhasm: 2x r += r # asm 1: paddq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 80(%rsp),%xmm4 # qhasm: 2x r += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 64(%rsp),%xmm2,%xmm4 # qhasm: 2x m0 += r # asm 1: paddq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 16(%rsp),%xmm4 # qhasm: 2x r *= h6_19_stack # asm 1: pmuludq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 16(%rsp),%xmm4 # qhasm: 2x r *= h7_38_stack # asm 1: pmuludq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 48(%rsp),%xmm4 # qhasm: 2x r *= h7_38_stack # asm 1: pmuludq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 0(%rsp),%xmm2,%xmm2 # qhasm: 2x m9 += r # asm 1: paddq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 32(%rsp),%xmm2 # qhasm: 2x r *= h5_stack # asm 1: pmuludq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 64(%rsp),%xmm2 # qhasm: 2x r *= h6_stack # asm 1: pmuludq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 96(%rsp),%xmm2 # qhasm: 2x r *= h7_stack # asm 1: pmuludq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 160(%rsp),%xmm2 # qhasm: 2x r *= h8_stack # asm 1: pmuludq >= 26 # asm 1: vpsrlq $26,carry0=reg128#3 # asm 2: vpsrlq $26,carry0=%xmm2 vpsrlq $26,%xmm3,%xmm2 # qhasm: 2x m1 += carry0 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#3 # asm 2: vpsrlq $25,carry5=%xmm2 vpsrlq $25,%xmm14,%xmm2 # qhasm: 2x m6 += carry5 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry1=reg128#3 # asm 2: vpsrlq $25,carry1=%xmm2 vpsrlq $25,%xmm11,%xmm2 # qhasm: 2x m2 += carry1 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry6=reg128#3 # asm 2: vpsrlq $26,carry6=%xmm2 vpsrlq $26,%xmm5,%xmm2 # qhasm: 2x m7 += carry6 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry2=reg128#3 # asm 2: vpsrlq $26,carry2=%xmm2 vpsrlq $26,%xmm12,%xmm2 # qhasm: 2x m3 += carry2 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry7=reg128#3 # asm 2: vpsrlq $25,carry7=%xmm2 vpsrlq $25,%xmm0,%xmm2 # qhasm: 2x m8 += carry7 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry3=reg128#3 # asm 2: vpsrlq $25,carry3=%xmm2 vpsrlq $25,%xmm9,%xmm2 # qhasm: 2x m4 += carry3 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry8=reg128#3 # asm 2: vpsrlq $26,carry8=%xmm2 vpsrlq $26,%xmm8,%xmm2 # qhasm: 2x m9 += carry8 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry4=reg128#3 # asm 2: vpsrlq $26,carry4=%xmm2 vpsrlq $26,%xmm1,%xmm2 # qhasm: 2x m5 += carry4 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry9=reg128#3 # asm 2: vpsrlq $25,carry9=%xmm2 vpsrlq $25,%xmm10,%xmm2 # qhasm: 2x r0 = carry9 << 4 # asm 1: vpsllq $4,r0=reg128#5 # asm 2: vpsllq $4,r0=%xmm4 vpsllq $4,%xmm2,%xmm4 # qhasm: 2x m0 += carry9 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#3 # asm 2: vpsrlq $25,carry5=%xmm2 vpsrlq $25,%xmm14,%xmm2 # qhasm: 2x m6 += carry5 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry0=reg128#3 # asm 2: vpsrlq $26,carry0=%xmm2 vpsrlq $26,%xmm3,%xmm2 # qhasm: 2x m1 += carry0 # asm 1: paddq r=reg128#3 # asm 2: vpunpckhqdq r=%xmm2 vpunpckhqdq %xmm11,%xmm3,%xmm2 # qhasm: x3_0 = r # asm 1: movdqa x3_0=stack128#1 # asm 2: movdqa x3_0=0(%rsp) movdqa %xmm2,0(%rsp) # qhasm: r = unpack_low(m0, m1) # asm 1: vpunpcklqdq r=reg128#3 # asm 2: vpunpcklqdq r=%xmm2 vpunpcklqdq %xmm11,%xmm3,%xmm2 # qhasm: 2x r *= mem128[ v9_9 ] # asm 1: pmuludq v9_9,z3_0=stack128#6 # asm 2: movdqa z3_0=80(%rsp) movdqa %xmm2,80(%rsp) # qhasm: r = unpack_high(m2, m3) # asm 1: vpunpckhqdq r=reg128#3 # asm 2: vpunpckhqdq r=%xmm2 vpunpckhqdq %xmm9,%xmm12,%xmm2 # qhasm: x3_2 = r # asm 1: movdqa x3_2=stack128#2 # asm 2: movdqa x3_2=16(%rsp) movdqa %xmm2,16(%rsp) # qhasm: r = unpack_low(m2, m3) # asm 1: vpunpcklqdq r=reg128#3 # asm 2: vpunpcklqdq r=%xmm2 vpunpcklqdq %xmm9,%xmm12,%xmm2 # qhasm: 2x r *= mem128[ v9_9 ] # asm 1: pmuludq v9_9,z3_2=stack128#7 # asm 2: movdqa z3_2=96(%rsp) movdqa %xmm2,96(%rsp) # qhasm: r = unpack_high(m4, m5) # asm 1: vpunpckhqdq r=reg128#3 # asm 2: vpunpckhqdq r=%xmm2 vpunpckhqdq %xmm14,%xmm1,%xmm2 # qhasm: x3_4 = r # asm 1: movdqa x3_4=stack128#3 # asm 2: movdqa x3_4=32(%rsp) movdqa %xmm2,32(%rsp) # qhasm: r = unpack_low(m4, m5) # asm 1: vpunpcklqdq r=reg128#2 # asm 2: vpunpcklqdq r=%xmm1 vpunpcklqdq %xmm14,%xmm1,%xmm1 # qhasm: 2x r *= mem128[ v9_9 ] # asm 1: pmuludq v9_9,z3_4=stack128#8 # asm 2: movdqa z3_4=112(%rsp) movdqa %xmm1,112(%rsp) # qhasm: r = unpack_high(m6, m7) # asm 1: vpunpckhqdq r=reg128#2 # asm 2: vpunpckhqdq r=%xmm1 vpunpckhqdq %xmm0,%xmm5,%xmm1 # qhasm: x3_6 = r # asm 1: movdqa x3_6=stack128#4 # asm 2: movdqa x3_6=48(%rsp) movdqa %xmm1,48(%rsp) # qhasm: r = unpack_low(m6, m7) # asm 1: vpunpcklqdq r=reg128#1 # asm 2: vpunpcklqdq r=%xmm0 vpunpcklqdq %xmm0,%xmm5,%xmm0 # qhasm: 2x r *= mem128[ v9_9 ] # asm 1: pmuludq v9_9,z3_6=stack128#11 # asm 2: movdqa z3_6=160(%rsp) movdqa %xmm0,160(%rsp) # qhasm: r = unpack_high(m8, m9) # asm 1: vpunpckhqdq r=reg128#1 # asm 2: vpunpckhqdq r=%xmm0 vpunpckhqdq %xmm10,%xmm8,%xmm0 # qhasm: x3_8 = r # asm 1: movdqa x3_8=stack128#5 # asm 2: movdqa x3_8=64(%rsp) movdqa %xmm0,64(%rsp) # qhasm: r = unpack_low(m8, m9) # asm 1: vpunpcklqdq r=reg128#1 # asm 2: vpunpcklqdq r=%xmm0 vpunpcklqdq %xmm10,%xmm8,%xmm0 # qhasm: 2x r *= mem128[ v9_9 ] # asm 1: pmuludq v9_9,z3_8=stack128#14 # asm 2: movdqa z3_8=208(%rsp) movdqa %xmm0,208(%rsp) # qhasm: f0 = f0_stack # asm 1: movdqa f0=reg128#1 # asm 2: movdqa f0=%xmm0 movdqa 144(%rsp),%xmm0 # qhasm: 2x h0 = f0 * f0 # asm 1: vpmuludq h0=reg128#2 # asm 2: vpmuludq h0=%xmm1 vpmuludq %xmm0,%xmm0,%xmm1 # qhasm: 2x f0 += f0 # asm 1: paddq f1=reg128#3 # asm 2: movdqa f1=%xmm2 movdqa 128(%rsp),%xmm2 # qhasm: 2x h1 = f0 * f1 # asm 1: vpmuludq h1=reg128#4 # asm 2: vpmuludq h1=%xmm3 vpmuludq %xmm2,%xmm0,%xmm3 # qhasm: f2 = f2_stack # asm 1: movdqa f2=reg128#5 # asm 2: movdqa f2=%xmm4 movdqa 192(%rsp),%xmm4 # qhasm: 2x h2 = f0 * f2 # asm 1: vpmuludq h2=reg128#6 # asm 2: vpmuludq h2=%xmm5 vpmuludq %xmm4,%xmm0,%xmm5 # qhasm: f3 = f3_stack # asm 1: movdqa f3=reg128#7 # asm 2: movdqa f3=%xmm6 movdqa 176(%rsp),%xmm6 # qhasm: 2x h3 = f0 * f3 # asm 1: vpmuludq h3=reg128#8 # asm 2: vpmuludq h3=%xmm7 vpmuludq %xmm6,%xmm0,%xmm7 # qhasm: f4 = f4_stack # asm 1: movdqa f4=reg128#9 # asm 2: movdqa f4=%xmm8 movdqa 240(%rsp),%xmm8 # qhasm: 2x h4 = f0 * f4 # asm 1: vpmuludq h4=reg128#10 # asm 2: vpmuludq h4=%xmm9 vpmuludq %xmm8,%xmm0,%xmm9 # qhasm: 2x h5 = f0 * f5_stack # asm 1: vpmuludq h5=reg128#11 # asm 2: vpmuludq h5=%xmm10 vpmuludq 224(%rsp),%xmm0,%xmm10 # qhasm: 2x h6 = f0 * f6_stack # asm 1: vpmuludq h6=reg128#12 # asm 2: vpmuludq h6=%xmm11 vpmuludq 304(%rsp),%xmm0,%xmm11 # qhasm: 2x h7 = f0 * f7_stack # asm 1: vpmuludq h7=reg128#13 # asm 2: vpmuludq h7=%xmm12 vpmuludq 288(%rsp),%xmm0,%xmm12 # qhasm: 2x h8 = f0 * f8_stack # asm 1: vpmuludq h8=reg128#14 # asm 2: vpmuludq h8=%xmm13 vpmuludq 336(%rsp),%xmm0,%xmm13 # qhasm: f9 = f9_stack # asm 1: movdqa f9=reg128#15 # asm 2: movdqa f9=%xmm14 movdqa 384(%rsp),%xmm14 # qhasm: 2x h9 = f0 * f9 # asm 1: vpmuludq h9=reg128#1 # asm 2: vpmuludq h9=%xmm0 vpmuludq %xmm14,%xmm0,%xmm0 # qhasm: 2x f9_38 = f9 * mem128[ v38_38 ] # asm 1: vpmuludq v38_38,f9_38=reg128#16 # asm 2: vpmuludq v38_38,f9_38=%xmm15 vpmuludq v38_38(%rip),%xmm14,%xmm15 # qhasm: 2x r = f9 * f9_38 # asm 1: vpmuludq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq %xmm15,%xmm14,%xmm14 # qhasm: 2x h8 += r # asm 1: paddq f3_2=reg128#15 # asm 2: vpaddq f3_2=%xmm14 vpaddq %xmm6,%xmm6,%xmm14 # qhasm: 2x r = f3 * f3_2 # asm 1: vpmuludq r=reg128#7 # asm 2: vpmuludq r=%xmm6 vpmuludq %xmm14,%xmm6,%xmm6 # qhasm: 2x h6 += r # asm 1: paddq f1_2=reg128#7 # asm 2: vpaddq f1_2=%xmm6 vpaddq %xmm2,%xmm2,%xmm6 # qhasm: 2x r = f1 * f1_2 # asm 1: vpmuludq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm6,%xmm2,%xmm2 # qhasm: 2x h2 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm15,%xmm6,%xmm2 # qhasm: 2x h0 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm15,%xmm4,%xmm2 # qhasm: 2x h1 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 256(%rsp),%xmm6,%xmm2 # qhasm: 2x h6 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 304(%rsp),%xmm6,%xmm2 # qhasm: 2x h7 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 352(%rsp),%xmm6,%xmm2 # qhasm: 2x h8 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 336(%rsp),%xmm6,%xmm2 # qhasm: 2x h9 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm4,%xmm6,%xmm2 # qhasm: 2x h3 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm14,%xmm6,%xmm2 # qhasm: 2x h4 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm8,%xmm6,%xmm2 # qhasm: 2x h5 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm15,%xmm14,%xmm2 # qhasm: 2x h2 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm15,%xmm8,%xmm2 # qhasm: 2x h3 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm4,%xmm4,%xmm2 # qhasm: 2x h4 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm14,%xmm4,%xmm2 # qhasm: 2x h5 += r # asm 1: paddq f2_2=reg128#3 # asm 2: vpaddq f2_2=%xmm2 vpaddq %xmm4,%xmm4,%xmm2 # qhasm: 2x r = f2_2 * f4 # asm 1: vpmuludq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq %xmm8,%xmm2,%xmm4 # qhasm: 2x h6 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 400(%rsp),%xmm2,%xmm4 # qhasm: 2x h0 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 400(%rsp),%xmm14,%xmm4 # qhasm: 2x h1 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 224(%rsp),%xmm2,%xmm4 # qhasm: 2x h7 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 304(%rsp),%xmm2,%xmm4 # qhasm: 2x h8 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 288(%rsp),%xmm2,%xmm2 # qhasm: 2x h9 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 368(%rsp),%xmm8,%xmm2 # qhasm: 2x h1 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm8,%xmm14,%xmm2 # qhasm: 2x h7 += r # asm 1: paddq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq %xmm8,%xmm8,%xmm2 # qhasm: 2x h8 += r # asm 1: paddq f4_2=reg128#3 # asm 2: vpaddq f4_2=%xmm2 vpaddq %xmm8,%xmm8,%xmm2 # qhasm: 2x r = f4_2 * f8_19_stack # asm 1: vpmuludq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 400(%rsp),%xmm2,%xmm4 # qhasm: 2x h2 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 256(%rsp),%xmm15,%xmm4 # qhasm: 2x h4 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 304(%rsp),%xmm15,%xmm4 # qhasm: 2x h5 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 368(%rsp),%xmm14,%xmm4 # qhasm: 2x h0 += r # asm 1: paddq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 256(%rsp),%xmm4 # qhasm: 2x r *= f8_19_stack # asm 1: pmuludq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 256(%rsp),%xmm14,%xmm4 # qhasm: 2x h8 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 304(%rsp),%xmm14,%xmm4 # qhasm: 2x h9 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 352(%rsp),%xmm15,%xmm4 # qhasm: 2x h6 += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 336(%rsp),%xmm15,%xmm4 # qhasm: 2x h7 += r # asm 1: paddq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 304(%rsp),%xmm4 # qhasm: 2x r += r # asm 1: paddq r=reg128#5 # asm 2: vpmuludq r=%xmm4 vpmuludq 320(%rsp),%xmm2,%xmm4 # qhasm: 2x h0 += r # asm 1: paddq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 256(%rsp),%xmm4 # qhasm: 2x r *= f6_19_stack # asm 1: pmuludq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 256(%rsp),%xmm4 # qhasm: 2x r *= f7_38_stack # asm 1: pmuludq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 304(%rsp),%xmm4 # qhasm: 2x r *= f7_38_stack # asm 1: pmuludq r=reg128#5 # asm 2: movdqa r=%xmm4 movdqa 352(%rsp),%xmm4 # qhasm: 2x r *= f8_19_stack # asm 1: pmuludq r=reg128#3 # asm 2: vpmuludq r=%xmm2 vpmuludq 224(%rsp),%xmm2,%xmm2 # qhasm: 2x h9 += r # asm 1: paddq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 272(%rsp),%xmm2 # qhasm: 2x r *= f5_stack # asm 1: pmuludq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 320(%rsp),%xmm2 # qhasm: 2x r *= f6_stack # asm 1: pmuludq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 368(%rsp),%xmm2 # qhasm: 2x r *= f7_stack # asm 1: pmuludq r=reg128#3 # asm 2: movdqa r=%xmm2 movdqa 400(%rsp),%xmm2 # qhasm: 2x r *= f8_stack # asm 1: pmuludq >= 26 # asm 1: vpsrlq $26,carry0=reg128#3 # asm 2: vpsrlq $26,carry0=%xmm2 vpsrlq $26,%xmm1,%xmm2 # qhasm: 2x h1 += carry0 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#3 # asm 2: vpsrlq $25,carry5=%xmm2 vpsrlq $25,%xmm10,%xmm2 # qhasm: 2x h6 += carry5 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry1=reg128#3 # asm 2: vpsrlq $25,carry1=%xmm2 vpsrlq $25,%xmm3,%xmm2 # qhasm: 2x h2 += carry1 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry6=reg128#3 # asm 2: vpsrlq $26,carry6=%xmm2 vpsrlq $26,%xmm11,%xmm2 # qhasm: 2x h7 += carry6 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry2=reg128#3 # asm 2: vpsrlq $26,carry2=%xmm2 vpsrlq $26,%xmm5,%xmm2 # qhasm: 2x h3 += carry2 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry7=reg128#3 # asm 2: vpsrlq $25,carry7=%xmm2 vpsrlq $25,%xmm12,%xmm2 # qhasm: 2x h8 += carry7 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry3=reg128#3 # asm 2: vpsrlq $25,carry3=%xmm2 vpsrlq $25,%xmm7,%xmm2 # qhasm: 2x h4 += carry3 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry8=reg128#3 # asm 2: vpsrlq $26,carry8=%xmm2 vpsrlq $26,%xmm13,%xmm2 # qhasm: 2x h9 += carry8 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry4=reg128#3 # asm 2: vpsrlq $26,carry4=%xmm2 vpsrlq $26,%xmm9,%xmm2 # qhasm: 2x h5 += carry4 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry9=reg128#3 # asm 2: vpsrlq $25,carry9=%xmm2 vpsrlq $25,%xmm0,%xmm2 # qhasm: 2x r0 = carry9 << 4 # asm 1: vpsllq $4,r0=reg128#5 # asm 2: vpsllq $4,r0=%xmm4 vpsllq $4,%xmm2,%xmm4 # qhasm: 2x h0 += carry9 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#3 # asm 2: vpsrlq $25,carry5=%xmm2 vpsrlq $25,%xmm10,%xmm2 # qhasm: 2x h6 += carry5 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry0=reg128#3 # asm 2: vpsrlq $26,carry0=%xmm2 vpsrlq $26,%xmm1,%xmm2 # qhasm: 2x h1 += carry0 # asm 1: paddq f0=reg128#3 # asm 2: vpunpckhqdq f0=%xmm2 vpunpckhqdq %xmm3,%xmm1,%xmm2 # qhasm: h0 = unpack_low(h0, h1) # asm 1: vpunpcklqdq h0=reg128#2 # asm 2: vpunpcklqdq h0=%xmm1 vpunpcklqdq %xmm3,%xmm1,%xmm1 # qhasm: t0_0 = h0 # asm 1: movdqa t0_0=stack128#12 # asm 2: movdqa t0_0=176(%rsp) movdqa %xmm1,176(%rsp) # qhasm: 2x h1 = f0 + mem128[ subc0 ] # asm 1: vpaddq subc0,h1=reg128#4 # asm 2: vpaddq subc0,h1=%xmm3 vpaddq subc0(%rip),%xmm2,%xmm3 # qhasm: 2x h1 -= h0 # asm 1: psubq f1=reg128#2 # asm 2: vpunpckhqdq f1=%xmm1 vpunpckhqdq %xmm3,%xmm2,%xmm1 # qhasm: f0 = unpack_low(f0, h1) # asm 1: vpunpcklqdq f0=reg128#3 # asm 2: vpunpcklqdq f0=%xmm2 vpunpcklqdq %xmm3,%xmm2,%xmm2 # qhasm: f0_stack = f0 # asm 1: movdqa f0_stack=stack128#13 # asm 2: movdqa f0_stack=192(%rsp) movdqa %xmm2,192(%rsp) # qhasm: f1_stack = f1 # asm 1: movdqa f1_stack=stack128#15 # asm 2: movdqa f1_stack=224(%rsp) movdqa %xmm1,224(%rsp) # qhasm: 2x f1 <<= 1 # asm 1: psllq $1,f1_2_stack=stack128#16 # asm 2: movdqa f1_2_stack=240(%rsp) movdqa %xmm1,240(%rsp) # qhasm: 2x h1 *= mem128[ v121666_121666 ] # asm 1: pmuludq v121666_121666,r=reg128#2 # asm 2: movdqa r=%xmm1 movdqa 80(%rsp),%xmm1 # qhasm: g0 = unpack_low(h1, r) # asm 1: vpunpcklqdq g0=reg128#3 # asm 2: vpunpcklqdq g0=%xmm2 vpunpcklqdq %xmm1,%xmm3,%xmm2 # qhasm: g1 = unpack_high(h1, r) # asm 1: vpunpckhqdq g1=reg128#2 # asm 2: vpunpckhqdq g1=%xmm1 vpunpckhqdq %xmm1,%xmm3,%xmm1 # qhasm: f2 = unpack_high(h2, h3) # asm 1: vpunpckhqdq f2=reg128#4 # asm 2: vpunpckhqdq f2=%xmm3 vpunpckhqdq %xmm7,%xmm5,%xmm3 # qhasm: h2 = unpack_low(h2, h3) # asm 1: vpunpcklqdq h2=reg128#5 # asm 2: vpunpcklqdq h2=%xmm4 vpunpcklqdq %xmm7,%xmm5,%xmm4 # qhasm: t0_2 = h2 # asm 1: movdqa t0_2=stack128#17 # asm 2: movdqa t0_2=256(%rsp) movdqa %xmm4,256(%rsp) # qhasm: 2x h3 = f2 + mem128[ subc2 ] # asm 1: vpaddq subc2,h3=reg128#6 # asm 2: vpaddq subc2,h3=%xmm5 vpaddq subc2(%rip),%xmm3,%xmm5 # qhasm: 2x h3 -= h2 # asm 1: psubq f3=reg128#5 # asm 2: vpunpckhqdq f3=%xmm4 vpunpckhqdq %xmm5,%xmm3,%xmm4 # qhasm: f2 = unpack_low(f2, h3) # asm 1: vpunpcklqdq f2=reg128#4 # asm 2: vpunpcklqdq f2=%xmm3 vpunpcklqdq %xmm5,%xmm3,%xmm3 # qhasm: f2_stack = f2 # asm 1: movdqa f2_stack=stack128#18 # asm 2: movdqa f2_stack=272(%rsp) movdqa %xmm3,272(%rsp) # qhasm: f3_stack = f3 # asm 1: movdqa f3_stack=stack128#19 # asm 2: movdqa f3_stack=288(%rsp) movdqa %xmm4,288(%rsp) # qhasm: 2x f3 <<= 1 # asm 1: psllq $1,f3_2_stack=stack128#20 # asm 2: movdqa f3_2_stack=304(%rsp) movdqa %xmm4,304(%rsp) # qhasm: 2x h3 *= mem128[ v121666_121666 ] # asm 1: pmuludq v121666_121666,r=reg128#4 # asm 2: movdqa r=%xmm3 movdqa 96(%rsp),%xmm3 # qhasm: g2 = unpack_low(h3, r) # asm 1: vpunpcklqdq g2=reg128#5 # asm 2: vpunpcklqdq g2=%xmm4 vpunpcklqdq %xmm3,%xmm5,%xmm4 # qhasm: g3 = unpack_high(h3, r) # asm 1: vpunpckhqdq g3=reg128#4 # asm 2: vpunpckhqdq g3=%xmm3 vpunpckhqdq %xmm3,%xmm5,%xmm3 # qhasm: f4 = unpack_high(h4, h5) # asm 1: vpunpckhqdq f4=reg128#6 # asm 2: vpunpckhqdq f4=%xmm5 vpunpckhqdq %xmm10,%xmm9,%xmm5 # qhasm: h4 = unpack_low(h4, h5) # asm 1: vpunpcklqdq h4=reg128#7 # asm 2: vpunpcklqdq h4=%xmm6 vpunpcklqdq %xmm10,%xmm9,%xmm6 # qhasm: t0_4 = h4 # asm 1: movdqa t0_4=stack128#21 # asm 2: movdqa t0_4=320(%rsp) movdqa %xmm6,320(%rsp) # qhasm: 2x h5 = f4 + mem128[ subc2 ] # asm 1: vpaddq subc2,h5=reg128#8 # asm 2: vpaddq subc2,h5=%xmm7 vpaddq subc2(%rip),%xmm5,%xmm7 # qhasm: 2x h5 -= h4 # asm 1: psubq f5=reg128#7 # asm 2: vpunpckhqdq f5=%xmm6 vpunpckhqdq %xmm7,%xmm5,%xmm6 # qhasm: f4 = unpack_low(f4, h5) # asm 1: vpunpcklqdq f4=reg128#6 # asm 2: vpunpcklqdq f4=%xmm5 vpunpcklqdq %xmm7,%xmm5,%xmm5 # qhasm: f4_stack = f4 # asm 1: movdqa f4_stack=stack128#22 # asm 2: movdqa f4_stack=336(%rsp) movdqa %xmm5,336(%rsp) # qhasm: f5_stack = f5 # asm 1: movdqa f5_stack=stack128#23 # asm 2: movdqa f5_stack=352(%rsp) movdqa %xmm6,352(%rsp) # qhasm: 2x f5 <<= 1 # asm 1: psllq $1,f5_2_stack=stack128#24 # asm 2: movdqa f5_2_stack=368(%rsp) movdqa %xmm6,368(%rsp) # qhasm: 2x h5 *= mem128[ v121666_121666 ] # asm 1: pmuludq v121666_121666,r=reg128#6 # asm 2: movdqa r=%xmm5 movdqa 112(%rsp),%xmm5 # qhasm: g4 = unpack_low(h5, r) # asm 1: vpunpcklqdq g4=reg128#7 # asm 2: vpunpcklqdq g4=%xmm6 vpunpcklqdq %xmm5,%xmm7,%xmm6 # qhasm: g5 = unpack_high(h5, r) # asm 1: vpunpckhqdq g5=reg128#6 # asm 2: vpunpckhqdq g5=%xmm5 vpunpckhqdq %xmm5,%xmm7,%xmm5 # qhasm: f6 = unpack_high(h6, h7) # asm 1: vpunpckhqdq f6=reg128#8 # asm 2: vpunpckhqdq f6=%xmm7 vpunpckhqdq %xmm12,%xmm11,%xmm7 # qhasm: h6 = unpack_low(h6, h7) # asm 1: vpunpcklqdq h6=reg128#9 # asm 2: vpunpcklqdq h6=%xmm8 vpunpcklqdq %xmm12,%xmm11,%xmm8 # qhasm: t0_6 = h6 # asm 1: movdqa t0_6=stack128#25 # asm 2: movdqa t0_6=384(%rsp) movdqa %xmm8,384(%rsp) # qhasm: 2x h7 = f6 + mem128[ subc2 ] # asm 1: vpaddq subc2,h7=reg128#10 # asm 2: vpaddq subc2,h7=%xmm9 vpaddq subc2(%rip),%xmm7,%xmm9 # qhasm: 2x h7 -= h6 # asm 1: psubq f7=reg128#9 # asm 2: vpunpckhqdq f7=%xmm8 vpunpckhqdq %xmm9,%xmm7,%xmm8 # qhasm: f6 = unpack_low(f6, h7) # asm 1: vpunpcklqdq f6=reg128#8 # asm 2: vpunpcklqdq f6=%xmm7 vpunpcklqdq %xmm9,%xmm7,%xmm7 # qhasm: f6_stack = f6 # asm 1: movdqa f6_stack=stack128#26 # asm 2: movdqa f6_stack=400(%rsp) movdqa %xmm7,400(%rsp) # qhasm: f7_stack = f7 # asm 1: movdqa f7_stack=stack128#27 # asm 2: movdqa f7_stack=416(%rsp) movdqa %xmm8,416(%rsp) # qhasm: 2x f7 <<= 1 # asm 1: psllq $1,f7_2_stack=stack128#28 # asm 2: movdqa f7_2_stack=432(%rsp) movdqa %xmm8,432(%rsp) # qhasm: 2x h7 *= mem128[ v121666_121666 ] # asm 1: pmuludq v121666_121666,r=reg128#8 # asm 2: movdqa r=%xmm7 movdqa 160(%rsp),%xmm7 # qhasm: g6 = unpack_low(h7, r) # asm 1: vpunpcklqdq g6=reg128#9 # asm 2: vpunpcklqdq g6=%xmm8 vpunpcklqdq %xmm7,%xmm9,%xmm8 # qhasm: g7 = unpack_high(h7, r) # asm 1: vpunpckhqdq g7=reg128#8 # asm 2: vpunpckhqdq g7=%xmm7 vpunpckhqdq %xmm7,%xmm9,%xmm7 # qhasm: f8 = unpack_high(h8, h9) # asm 1: vpunpckhqdq f8=reg128#10 # asm 2: vpunpckhqdq f8=%xmm9 vpunpckhqdq %xmm0,%xmm13,%xmm9 # qhasm: h8 = unpack_low(h8, h9) # asm 1: vpunpcklqdq h8=reg128#1 # asm 2: vpunpcklqdq h8=%xmm0 vpunpcklqdq %xmm0,%xmm13,%xmm0 # qhasm: t0_8 = h8 # asm 1: movdqa t0_8=stack128#11 # asm 2: movdqa t0_8=160(%rsp) movdqa %xmm0,160(%rsp) # qhasm: 2x h9 = f8 + mem128[ subc2 ] # asm 1: vpaddq subc2,h9=reg128#11 # asm 2: vpaddq subc2,h9=%xmm10 vpaddq subc2(%rip),%xmm9,%xmm10 # qhasm: 2x h9 -= h8 # asm 1: psubq f9=reg128#1 # asm 2: vpunpckhqdq f9=%xmm0 vpunpckhqdq %xmm10,%xmm9,%xmm0 # qhasm: f8 = unpack_low(f8, h9) # asm 1: vpunpcklqdq f8=reg128#10 # asm 2: vpunpcklqdq f8=%xmm9 vpunpcklqdq %xmm10,%xmm9,%xmm9 # qhasm: f8_stack = f8 # asm 1: movdqa f8_stack=stack128#29 # asm 2: movdqa f8_stack=448(%rsp) movdqa %xmm9,448(%rsp) # qhasm: f9_stack = f9 # asm 1: movdqa f9_stack=stack128#30 # asm 2: movdqa f9_stack=464(%rsp) movdqa %xmm0,464(%rsp) # qhasm: 2x f9 <<= 1 # asm 1: psllq $1,f9_2_stack=stack128#31 # asm 2: movdqa f9_2_stack=480(%rsp) movdqa %xmm0,480(%rsp) # qhasm: 2x h9 *= mem128[ v121666_121666 ] # asm 1: pmuludq v121666_121666,r=reg128#1 # asm 2: movdqa r=%xmm0 movdqa 208(%rsp),%xmm0 # qhasm: g8 = unpack_low(h9, r) # asm 1: vpunpcklqdq g8=reg128#10 # asm 2: vpunpcklqdq g8=%xmm9 vpunpcklqdq %xmm0,%xmm10,%xmm9 # qhasm: g9 = unpack_high(h9, r) # asm 1: vpunpckhqdq g9=reg128#1 # asm 2: vpunpckhqdq g9=%xmm0 vpunpckhqdq %xmm0,%xmm10,%xmm0 # qhasm: 2x carry0 = g0 unsigned>>= 26 # asm 1: vpsrlq $26,carry0=reg128#11 # asm 2: vpsrlq $26,carry0=%xmm10 vpsrlq $26,%xmm2,%xmm10 # qhasm: 2x g1 += carry0 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#11 # asm 2: vpsrlq $25,carry5=%xmm10 vpsrlq $25,%xmm5,%xmm10 # qhasm: 2x g6 += carry5 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry1=reg128#11 # asm 2: vpsrlq $25,carry1=%xmm10 vpsrlq $25,%xmm1,%xmm10 # qhasm: 2x g2 += carry1 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry6=reg128#11 # asm 2: vpsrlq $26,carry6=%xmm10 vpsrlq $26,%xmm8,%xmm10 # qhasm: 2x g7 += carry6 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry2=reg128#11 # asm 2: vpsrlq $26,carry2=%xmm10 vpsrlq $26,%xmm4,%xmm10 # qhasm: 2x g3 += carry2 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry7=reg128#11 # asm 2: vpsrlq $25,carry7=%xmm10 vpsrlq $25,%xmm7,%xmm10 # qhasm: 2x g8 += carry7 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry3=reg128#11 # asm 2: vpsrlq $25,carry3=%xmm10 vpsrlq $25,%xmm3,%xmm10 # qhasm: 2x g4 += carry3 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry8=reg128#11 # asm 2: vpsrlq $26,carry8=%xmm10 vpsrlq $26,%xmm9,%xmm10 # qhasm: 2x g9 += carry8 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry4=reg128#11 # asm 2: vpsrlq $26,carry4=%xmm10 vpsrlq $26,%xmm6,%xmm10 # qhasm: 2x g5 += carry4 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry9=reg128#11 # asm 2: vpsrlq $25,carry9=%xmm10 vpsrlq $25,%xmm0,%xmm10 # qhasm: 2x r0 = carry9 << 4 # asm 1: vpsllq $4,r0=reg128#12 # asm 2: vpsllq $4,r0=%xmm11 vpsllq $4,%xmm10,%xmm11 # qhasm: 2x g0 += carry9 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#11 # asm 2: vpsrlq $25,carry5=%xmm10 vpsrlq $25,%xmm5,%xmm10 # qhasm: 2x g6 += carry5 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry0=reg128#11 # asm 2: vpsrlq $26,carry0=%xmm10 vpsrlq $26,%xmm2,%xmm10 # qhasm: 2x g1 += carry0 # asm 1: paddq r=reg128#11 # asm 2: vpunpckhqdq r=%xmm10 vpunpckhqdq %xmm1,%xmm2,%xmm10 # qhasm: z3_0 = r # asm 1: movdqa z3_0=stack128#6 # asm 2: movdqa z3_0=80(%rsp) movdqa %xmm10,80(%rsp) # qhasm: h1 = unpack_low(g0, g1) # asm 1: vpunpcklqdq h1=reg128#2 # asm 2: vpunpcklqdq h1=%xmm1 vpunpcklqdq %xmm1,%xmm2,%xmm1 # qhasm: r = unpack_high(g2, g3) # asm 1: vpunpckhqdq r=reg128#3 # asm 2: vpunpckhqdq r=%xmm2 vpunpckhqdq %xmm3,%xmm4,%xmm2 # qhasm: z3_2 = r # asm 1: movdqa z3_2=stack128#7 # asm 2: movdqa z3_2=96(%rsp) movdqa %xmm2,96(%rsp) # qhasm: h3 = unpack_low(g2, g3) # asm 1: vpunpcklqdq h3=reg128#3 # asm 2: vpunpcklqdq h3=%xmm2 vpunpcklqdq %xmm3,%xmm4,%xmm2 # qhasm: r = unpack_high(g4, g5) # asm 1: vpunpckhqdq r=reg128#4 # asm 2: vpunpckhqdq r=%xmm3 vpunpckhqdq %xmm5,%xmm6,%xmm3 # qhasm: z3_4 = r # asm 1: movdqa z3_4=stack128#8 # asm 2: movdqa z3_4=112(%rsp) movdqa %xmm3,112(%rsp) # qhasm: h5 = unpack_low(g4, g5) # asm 1: vpunpcklqdq h5=reg128#4 # asm 2: vpunpcklqdq h5=%xmm3 vpunpcklqdq %xmm5,%xmm6,%xmm3 # qhasm: r = unpack_high(g6, g7) # asm 1: vpunpckhqdq r=reg128#5 # asm 2: vpunpckhqdq r=%xmm4 vpunpckhqdq %xmm7,%xmm8,%xmm4 # qhasm: z3_6 = r # asm 1: movdqa z3_6=stack128#9 # asm 2: movdqa z3_6=128(%rsp) movdqa %xmm4,128(%rsp) # qhasm: h7 = unpack_low(g6, g7) # asm 1: vpunpcklqdq h7=reg128#5 # asm 2: vpunpcklqdq h7=%xmm4 vpunpcklqdq %xmm7,%xmm8,%xmm4 # qhasm: r = unpack_high(g8, g9) # asm 1: vpunpckhqdq r=reg128#6 # asm 2: vpunpckhqdq r=%xmm5 vpunpckhqdq %xmm0,%xmm9,%xmm5 # qhasm: z3_8 = r # asm 1: movdqa z3_8=stack128#10 # asm 2: movdqa z3_8=144(%rsp) movdqa %xmm5,144(%rsp) # qhasm: h9 = unpack_low(g8, g9) # asm 1: vpunpcklqdq h9=reg128#1 # asm 2: vpunpcklqdq h9=%xmm0 vpunpcklqdq %xmm0,%xmm9,%xmm0 # qhasm: h0 = t0_0 # asm 1: movdqa h0=reg128#6 # asm 2: movdqa h0=%xmm5 movdqa 176(%rsp),%xmm5 # qhasm: 2x h1 += h0 # asm 1: paddq g0=reg128#7 # asm 2: vpunpcklqdq g0=%xmm6 vpunpcklqdq %xmm1,%xmm5,%xmm6 # qhasm: g1 = unpack_high(h0, h1) # asm 1: vpunpckhqdq g1=reg128#2 # asm 2: vpunpckhqdq g1=%xmm1 vpunpckhqdq %xmm1,%xmm5,%xmm1 # qhasm: 2x m1 = g0 * f1_stack # asm 1: vpmuludq m1=reg128#6 # asm 2: vpmuludq m1=%xmm5 vpmuludq 224(%rsp),%xmm6,%xmm5 # qhasm: 2x r = g1 * f0_stack # asm 1: vpmuludq r=reg128#8 # asm 2: vpmuludq r=%xmm7 vpmuludq 192(%rsp),%xmm1,%xmm7 # qhasm: 2x m1 += r # asm 1: paddq m2=reg128#8 # asm 2: vpmuludq m2=%xmm7 vpmuludq 272(%rsp),%xmm6,%xmm7 # qhasm: 2x r = g1 * f1_2_stack # asm 1: vpmuludq r=reg128#9 # asm 2: vpmuludq r=%xmm8 vpmuludq 240(%rsp),%xmm1,%xmm8 # qhasm: 2x m2 += r # asm 1: paddq m3=reg128#9 # asm 2: vpmuludq m3=%xmm8 vpmuludq 288(%rsp),%xmm6,%xmm8 # qhasm: 2x r = g1 * f2_stack # asm 1: vpmuludq r=reg128#10 # asm 2: vpmuludq r=%xmm9 vpmuludq 272(%rsp),%xmm1,%xmm9 # qhasm: 2x m3 += r # asm 1: paddq m4=reg128#10 # asm 2: vpmuludq m4=%xmm9 vpmuludq 336(%rsp),%xmm6,%xmm9 # qhasm: 2x r = g1 * f3_2_stack # asm 1: vpmuludq r=reg128#11 # asm 2: vpmuludq r=%xmm10 vpmuludq 304(%rsp),%xmm1,%xmm10 # qhasm: 2x m4 += r # asm 1: paddq m5=reg128#11 # asm 2: vpmuludq m5=%xmm10 vpmuludq 352(%rsp),%xmm6,%xmm10 # qhasm: 2x r = g1 * f4_stack # asm 1: vpmuludq r=reg128#12 # asm 2: vpmuludq r=%xmm11 vpmuludq 336(%rsp),%xmm1,%xmm11 # qhasm: 2x m5 += r # asm 1: paddq m6=reg128#12 # asm 2: vpmuludq m6=%xmm11 vpmuludq 400(%rsp),%xmm6,%xmm11 # qhasm: 2x r = g1 * f5_2_stack # asm 1: vpmuludq r=reg128#13 # asm 2: vpmuludq r=%xmm12 vpmuludq 368(%rsp),%xmm1,%xmm12 # qhasm: 2x m6 += r # asm 1: paddq m7=reg128#13 # asm 2: vpmuludq m7=%xmm12 vpmuludq 416(%rsp),%xmm6,%xmm12 # qhasm: 2x r = g1 * f6_stack # asm 1: vpmuludq r=reg128#14 # asm 2: vpmuludq r=%xmm13 vpmuludq 400(%rsp),%xmm1,%xmm13 # qhasm: 2x m7 += r # asm 1: paddq m8=reg128#14 # asm 2: vpmuludq m8=%xmm13 vpmuludq 448(%rsp),%xmm6,%xmm13 # qhasm: 2x r = g1 * f7_2_stack # asm 1: vpmuludq r=reg128#15 # asm 2: vpmuludq r=%xmm14 vpmuludq 432(%rsp),%xmm1,%xmm14 # qhasm: 2x m8 += r # asm 1: paddq m9=reg128#15 # asm 2: vpmuludq m9=%xmm14 vpmuludq 464(%rsp),%xmm6,%xmm14 # qhasm: 2x r = g1 * f8_stack # asm 1: vpmuludq r=reg128#16 # asm 2: vpmuludq r=%xmm15 vpmuludq 448(%rsp),%xmm1,%xmm15 # qhasm: 2x m9 += r # asm 1: paddq m0=reg128#7 # asm 2: vpmuludq m0=%xmm6 vpmuludq 192(%rsp),%xmm6,%xmm6 # qhasm: 2x g1 *= mem128[ v19_19 ] # asm 1: pmuludq v19_19,r=reg128#2 # asm 2: vpmuludq r=%xmm1 vpmuludq 480(%rsp),%xmm1,%xmm1 # qhasm: 2x m0 += r # asm 1: paddq h2=reg128#2 # asm 2: movdqa h2=%xmm1 movdqa 256(%rsp),%xmm1 # qhasm: 2x h3 += h2 # asm 1: paddq g2=reg128#16 # asm 2: vpunpcklqdq g2=%xmm15 vpunpcklqdq %xmm2,%xmm1,%xmm15 # qhasm: g3 = unpack_high(h2, h3) # asm 1: vpunpckhqdq g3=reg128#2 # asm 2: vpunpckhqdq g3=%xmm1 vpunpckhqdq %xmm2,%xmm1,%xmm1 # qhasm: 2x r2 = g2 * f0_stack # asm 1: vpmuludq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 192(%rsp),%xmm15,%xmm2 # qhasm: 2x m2 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 224(%rsp),%xmm15,%xmm2 # qhasm: 2x m3 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 272(%rsp),%xmm15,%xmm2 # qhasm: 2x m4 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 288(%rsp),%xmm15,%xmm2 # qhasm: 2x m5 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 336(%rsp),%xmm15,%xmm2 # qhasm: 2x m6 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 352(%rsp),%xmm15,%xmm2 # qhasm: 2x m7 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 400(%rsp),%xmm15,%xmm2 # qhasm: 2x m8 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 416(%rsp),%xmm15,%xmm2 # qhasm: 2x m9 += r2 # asm 1: paddq r2=reg128#3 # asm 2: vpmuludq r2=%xmm2 vpmuludq 448(%rsp),%xmm15,%xmm2 # qhasm: 2x m0 += r2 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 192(%rsp),%xmm1,%xmm2 # qhasm: 2x m3 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 240(%rsp),%xmm1,%xmm2 # qhasm: 2x m4 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 272(%rsp),%xmm1,%xmm2 # qhasm: 2x m5 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 304(%rsp),%xmm1,%xmm2 # qhasm: 2x m6 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 336(%rsp),%xmm1,%xmm2 # qhasm: 2x m7 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 368(%rsp),%xmm1,%xmm2 # qhasm: 2x m8 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 400(%rsp),%xmm1,%xmm2 # qhasm: 2x m9 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 432(%rsp),%xmm1,%xmm2 # qhasm: 2x m0 += r3 # asm 1: paddq r3=reg128#3 # asm 2: vpmuludq r3=%xmm2 vpmuludq 448(%rsp),%xmm1,%xmm2 # qhasm: 2x m1 += r3 # asm 1: paddq h4=reg128#2 # asm 2: movdqa h4=%xmm1 movdqa 320(%rsp),%xmm1 # qhasm: 2x h5 += h4 # asm 1: paddq g4=reg128#3 # asm 2: vpunpcklqdq g4=%xmm2 vpunpcklqdq %xmm3,%xmm1,%xmm2 # qhasm: g5 = unpack_high(h4, h5) # asm 1: vpunpckhqdq g5=reg128#2 # asm 2: vpunpckhqdq g5=%xmm1 vpunpckhqdq %xmm3,%xmm1,%xmm1 # qhasm: 2x r4 = g4 * f0_stack # asm 1: vpmuludq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 192(%rsp),%xmm2,%xmm3 # qhasm: 2x m4 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 224(%rsp),%xmm2,%xmm3 # qhasm: 2x m5 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 272(%rsp),%xmm2,%xmm3 # qhasm: 2x m6 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 288(%rsp),%xmm2,%xmm3 # qhasm: 2x m7 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 336(%rsp),%xmm2,%xmm3 # qhasm: 2x m8 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 352(%rsp),%xmm2,%xmm3 # qhasm: 2x m9 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 400(%rsp),%xmm2,%xmm3 # qhasm: 2x m0 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 416(%rsp),%xmm2,%xmm3 # qhasm: 2x m1 += r4 # asm 1: paddq r4=reg128#4 # asm 2: vpmuludq r4=%xmm3 vpmuludq 448(%rsp),%xmm2,%xmm3 # qhasm: 2x m2 += r4 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 192(%rsp),%xmm1,%xmm2 # qhasm: 2x m5 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 240(%rsp),%xmm1,%xmm2 # qhasm: 2x m6 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 272(%rsp),%xmm1,%xmm2 # qhasm: 2x m7 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 304(%rsp),%xmm1,%xmm2 # qhasm: 2x m8 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 336(%rsp),%xmm1,%xmm2 # qhasm: 2x m9 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 368(%rsp),%xmm1,%xmm2 # qhasm: 2x m0 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 400(%rsp),%xmm1,%xmm2 # qhasm: 2x m1 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 432(%rsp),%xmm1,%xmm2 # qhasm: 2x m2 += r5 # asm 1: paddq r5=reg128#3 # asm 2: vpmuludq r5=%xmm2 vpmuludq 448(%rsp),%xmm1,%xmm2 # qhasm: 2x m3 += r5 # asm 1: paddq h6=reg128#2 # asm 2: movdqa h6=%xmm1 movdqa 384(%rsp),%xmm1 # qhasm: 2x h7 += h6 # asm 1: paddq g6=reg128#3 # asm 2: vpunpcklqdq g6=%xmm2 vpunpcklqdq %xmm4,%xmm1,%xmm2 # qhasm: g7 = unpack_high(h6, h7) # asm 1: vpunpckhqdq g7=reg128#2 # asm 2: vpunpckhqdq g7=%xmm1 vpunpckhqdq %xmm4,%xmm1,%xmm1 # qhasm: 2x r6 = g6 * f0_stack # asm 1: vpmuludq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 192(%rsp),%xmm2,%xmm3 # qhasm: 2x m6 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 224(%rsp),%xmm2,%xmm3 # qhasm: 2x m7 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 272(%rsp),%xmm2,%xmm3 # qhasm: 2x m8 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 288(%rsp),%xmm2,%xmm3 # qhasm: 2x m9 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 336(%rsp),%xmm2,%xmm3 # qhasm: 2x m0 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 352(%rsp),%xmm2,%xmm3 # qhasm: 2x m1 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 400(%rsp),%xmm2,%xmm3 # qhasm: 2x m2 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 416(%rsp),%xmm2,%xmm3 # qhasm: 2x m3 += r6 # asm 1: paddq r6=reg128#4 # asm 2: vpmuludq r6=%xmm3 vpmuludq 448(%rsp),%xmm2,%xmm3 # qhasm: 2x m4 += r6 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 192(%rsp),%xmm1,%xmm2 # qhasm: 2x m7 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 240(%rsp),%xmm1,%xmm2 # qhasm: 2x m8 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 272(%rsp),%xmm1,%xmm2 # qhasm: 2x m9 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 304(%rsp),%xmm1,%xmm2 # qhasm: 2x m0 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 336(%rsp),%xmm1,%xmm2 # qhasm: 2x m1 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 368(%rsp),%xmm1,%xmm2 # qhasm: 2x m2 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 400(%rsp),%xmm1,%xmm2 # qhasm: 2x m3 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 432(%rsp),%xmm1,%xmm2 # qhasm: 2x m4 += r7 # asm 1: paddq r7=reg128#3 # asm 2: vpmuludq r7=%xmm2 vpmuludq 448(%rsp),%xmm1,%xmm2 # qhasm: 2x m5 += r7 # asm 1: paddq h8=reg128#2 # asm 2: movdqa h8=%xmm1 movdqa 160(%rsp),%xmm1 # qhasm: 2x h9 += h8 # asm 1: paddq g8=reg128#3 # asm 2: vpunpcklqdq g8=%xmm2 vpunpcklqdq %xmm0,%xmm1,%xmm2 # qhasm: g9 = unpack_high(h8, h9) # asm 1: vpunpckhqdq g9=reg128#1 # asm 2: vpunpckhqdq g9=%xmm0 vpunpckhqdq %xmm0,%xmm1,%xmm0 # qhasm: 2x r8 = g8 * f0_stack # asm 1: vpmuludq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 192(%rsp),%xmm2,%xmm1 # qhasm: 2x m8 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 224(%rsp),%xmm2,%xmm1 # qhasm: 2x m9 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 272(%rsp),%xmm2,%xmm1 # qhasm: 2x m0 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 288(%rsp),%xmm2,%xmm1 # qhasm: 2x m1 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 336(%rsp),%xmm2,%xmm1 # qhasm: 2x m2 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 352(%rsp),%xmm2,%xmm1 # qhasm: 2x m3 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 400(%rsp),%xmm2,%xmm1 # qhasm: 2x m4 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 416(%rsp),%xmm2,%xmm1 # qhasm: 2x m5 += r8 # asm 1: paddq r8=reg128#2 # asm 2: vpmuludq r8=%xmm1 vpmuludq 448(%rsp),%xmm2,%xmm1 # qhasm: 2x m6 += r8 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 192(%rsp),%xmm0,%xmm1 # qhasm: 2x m9 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 240(%rsp),%xmm0,%xmm1 # qhasm: 2x m0 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 272(%rsp),%xmm0,%xmm1 # qhasm: 2x m1 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 304(%rsp),%xmm0,%xmm1 # qhasm: 2x m2 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 336(%rsp),%xmm0,%xmm1 # qhasm: 2x m3 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 368(%rsp),%xmm0,%xmm1 # qhasm: 2x m4 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 400(%rsp),%xmm0,%xmm1 # qhasm: 2x m5 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 432(%rsp),%xmm0,%xmm1 # qhasm: 2x m6 += r9 # asm 1: paddq r9=reg128#2 # asm 2: vpmuludq r9=%xmm1 vpmuludq 448(%rsp),%xmm0,%xmm1 # qhasm: 2x m7 += r9 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry0=reg128#1 # asm 2: vpsrlq $26,carry0=%xmm0 vpsrlq $26,%xmm6,%xmm0 # qhasm: 2x m1 += carry0 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#1 # asm 2: vpsrlq $25,carry5=%xmm0 vpsrlq $25,%xmm10,%xmm0 # qhasm: 2x m6 += carry5 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry1=reg128#1 # asm 2: vpsrlq $25,carry1=%xmm0 vpsrlq $25,%xmm5,%xmm0 # qhasm: 2x m2 += carry1 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry6=reg128#1 # asm 2: vpsrlq $26,carry6=%xmm0 vpsrlq $26,%xmm11,%xmm0 # qhasm: 2x m7 += carry6 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry2=reg128#1 # asm 2: vpsrlq $26,carry2=%xmm0 vpsrlq $26,%xmm7,%xmm0 # qhasm: 2x m3 += carry2 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry7=reg128#1 # asm 2: vpsrlq $25,carry7=%xmm0 vpsrlq $25,%xmm12,%xmm0 # qhasm: 2x m8 += carry7 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry3=reg128#1 # asm 2: vpsrlq $25,carry3=%xmm0 vpsrlq $25,%xmm8,%xmm0 # qhasm: 2x m4 += carry3 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry8=reg128#1 # asm 2: vpsrlq $26,carry8=%xmm0 vpsrlq $26,%xmm13,%xmm0 # qhasm: 2x m9 += carry8 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry4=reg128#1 # asm 2: vpsrlq $26,carry4=%xmm0 vpsrlq $26,%xmm9,%xmm0 # qhasm: 2x m5 += carry4 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry9=reg128#1 # asm 2: vpsrlq $25,carry9=%xmm0 vpsrlq $25,%xmm14,%xmm0 # qhasm: 2x r0 = carry9 << 4 # asm 1: vpsllq $4,r0=reg128#2 # asm 2: vpsllq $4,r0=%xmm1 vpsllq $4,%xmm0,%xmm1 # qhasm: 2x m0 += carry9 # asm 1: paddq >= 25 # asm 1: vpsrlq $25,carry5=reg128#1 # asm 2: vpsrlq $25,carry5=%xmm0 vpsrlq $25,%xmm10,%xmm0 # qhasm: 2x m6 += carry5 # asm 1: paddq >= 26 # asm 1: vpsrlq $26,carry0=reg128#1 # asm 2: vpsrlq $26,carry0=%xmm0 vpsrlq $26,%xmm6,%xmm0 # qhasm: 2x m1 += carry0 # asm 1: paddq f1=reg128#2 # asm 2: vpunpckhqdq f1=%xmm1 vpunpckhqdq %xmm5,%xmm6,%xmm1 # qhasm: f0 = unpack_low( m0, m1 ) # asm 1: vpunpcklqdq f0=reg128#1 # asm 2: vpunpcklqdq f0=%xmm0 vpunpcklqdq %xmm5,%xmm6,%xmm0 # qhasm: f3 = unpack_high( m2, m3 ) # asm 1: vpunpckhqdq f3=reg128#4 # asm 2: vpunpckhqdq f3=%xmm3 vpunpckhqdq %xmm8,%xmm7,%xmm3 # qhasm: f2 = unpack_low( m2, m3 ) # asm 1: vpunpcklqdq f2=reg128#3 # asm 2: vpunpcklqdq f2=%xmm2 vpunpcklqdq %xmm8,%xmm7,%xmm2 # qhasm: f5 = unpack_high( m4, m5 ) # asm 1: vpunpckhqdq f5=reg128#6 # asm 2: vpunpckhqdq f5=%xmm5 vpunpckhqdq %xmm10,%xmm9,%xmm5 # qhasm: f4 = unpack_low( m4, m5 ) # asm 1: vpunpcklqdq f4=reg128#5 # asm 2: vpunpcklqdq f4=%xmm4 vpunpcklqdq %xmm10,%xmm9,%xmm4 # qhasm: f7 = unpack_high( m6, m7 ) # asm 1: vpunpckhqdq f7=reg128#8 # asm 2: vpunpckhqdq f7=%xmm7 vpunpckhqdq %xmm12,%xmm11,%xmm7 # qhasm: f6 = unpack_low( m6, m7 ) # asm 1: vpunpcklqdq f6=reg128#7 # asm 2: vpunpcklqdq f6=%xmm6 vpunpcklqdq %xmm12,%xmm11,%xmm6 # qhasm: f9 = unpack_high( m8, m9 ) # asm 1: vpunpckhqdq f9=reg128#10 # asm 2: vpunpckhqdq f9=%xmm9 vpunpckhqdq %xmm14,%xmm13,%xmm9 # qhasm: f8 = unpack_low( m8, m9 ) # asm 1: vpunpcklqdq f8=reg128#9 # asm 2: vpunpcklqdq f8=%xmm8 vpunpcklqdq %xmm14,%xmm13,%xmm8 # qhasm: =? pos - 0 # asm 1: cmp $0,caller_r11=int64#9 # asm 2: movq caller_r11=%r11 movq 1536(%rsp),%r11 # qhasm: caller_r12 = r12_stack # asm 1: movq caller_r12=int64#10 # asm 2: movq caller_r12=%r12 movq 1544(%rsp),%r12 # qhasm: caller_r13 = r13_stack # asm 1: movq caller_r13=int64#11 # asm 2: movq caller_r13=%r13 movq 1552(%rsp),%r13 # qhasm: return add %r11,%rsp ret