# File: dclxvi-20130329/fp2e_square.s # Author: Ruben Niederhagen, Peter Schwabe # Public Domain # qhasm: int64 rop # qhasm: int64 op # qhasm: input rop # qhasm: input op # qhasm: int6464 r0 # qhasm: int6464 r1 # qhasm: int6464 r2 # qhasm: int6464 r3 # qhasm: int6464 r4 # qhasm: int6464 r5 # qhasm: int6464 r6 # qhasm: int6464 r7 # qhasm: int6464 r8 # qhasm: int6464 r9 # qhasm: int6464 r10 # qhasm: int6464 r11 # qhasm: int6464 r12 # qhasm: int6464 r13 # qhasm: int6464 r14 # qhasm: int6464 r15 # qhasm: int6464 r16 # qhasm: int6464 r17 # qhasm: int6464 r18 # qhasm: int6464 r19 # qhasm: int6464 r20 # qhasm: int6464 r21 # qhasm: int6464 r22 # qhasm: int6464 tmp0 # qhasm: int6464 tmp1 # qhasm: int6464 tmp2 # qhasm: int6464 tmp3 # qhasm: int6464 tmp4 # qhasm: int6464 tmp5 # qhasm: int6464 tmp6 # qhasm: int6464 tmp7 # qhasm: int6464 tmp8 # qhasm: int6464 tmp9 # qhasm: int6464 tmp10 # qhasm: int6464 tmp11 # qhasm: int64 t1p # qhasm: int64 t2p # qhasm: int64 rp # qhasm: int6464 0yoff # qhasm: int6464 t0 # qhasm: int6464 t1 # qhasm: int6464 t2 # qhasm: int6464 t3 # qhasm: int6464 t4 # qhasm: int6464 t5 # qhasm: int6464 t6 # qhasm: int6464 t7 # qhasm: int6464 t8 # qhasm: int6464 t9 # qhasm: int6464 t10 # qhasm: int6464 t11 # qhasm: int6464 t12 # qhasm: int6464 t13 # qhasm: int6464 t14 # qhasm: int6464 t15 # qhasm: int6464 t16 # qhasm: int6464 t17 # qhasm: int6464 t18 # qhasm: int6464 t19 # qhasm: int6464 t20 # qhasm: int6464 t21 # qhasm: int6464 t22 # qhasm: int6464 ab0 # qhasm: int6464 ab1 # qhasm: int6464 ab2 # qhasm: int6464 ab3 # qhasm: int6464 ab4 # qhasm: int6464 ab5 # qhasm: int6464 ab6 # qhasm: int6464 ab7 # qhasm: int6464 ab8 # qhasm: int6464 ab9 # qhasm: int6464 ab10 # qhasm: int6464 ab11 # qhasm: int6464 ab0six # qhasm: int6464 ab1six # qhasm: int6464 ab2six # qhasm: int6464 ab3six # qhasm: int6464 ab4six # qhasm: int6464 ab5six # qhasm: int6464 ab6six # qhasm: int6464 ab7six # qhasm: int6464 ab8six # qhasm: int6464 ab9six # qhasm: int6464 ab10six # qhasm: int6464 ab11six # qhasm: int64 myp # qhasm: int6464 round # qhasm: int6464 carry # qhasm: int6464 2t6 # qhasm: stack6144 mystack # qhasm: enter fp2e_square_qhasm .text .p2align 5 .globl _fp2e_square_qhasm .globl fp2e_square_qhasm _fp2e_square_qhasm: fp2e_square_qhasm: mov %rsp,%r11 and $31,%r11 add $768,%r11 sub %r11,%rsp # qhasm: myp = &mystack # asm 1: leaq myp=int64#3 # asm 2: leaq myp=%rdx leaq 0(%rsp),%rdx # qhasm: r0 = *(int128 *)(op + 0) # asm 1: movdqa 0(r0=int6464#1 # asm 2: movdqa 0(r0=%xmm0 movdqa 0(%rsi),%xmm0 # qhasm: tmp0 = r0 # asm 1: movdqa tmp0=int6464#2 # asm 2: movdqa tmp0=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp0 = shuffle float64 of tmp0 and tmp0 by 0x1 # asm 1: shufpd $0x1,r0=int6464#1 # asm 2: movdqa r0=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r0 = unpack high double of r0 and r0 # asm 1: unpckhpd r1=int6464#1 # asm 2: movdqa 16(r1=%xmm0 movdqa 16(%rsi),%xmm0 # qhasm: tmp1 = r1 # asm 1: movdqa tmp1=int6464#2 # asm 2: movdqa tmp1=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp1 = shuffle float64 of tmp1 and tmp1 by 0x1 # asm 1: shufpd $0x1,r1=int6464#1 # asm 2: movdqa r1=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r1 = unpack high double of r1 and r1 # asm 1: unpckhpd r2=int6464#1 # asm 2: movdqa 32(r2=%xmm0 movdqa 32(%rsi),%xmm0 # qhasm: tmp2 = r2 # asm 1: movdqa tmp2=int6464#2 # asm 2: movdqa tmp2=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp2 = shuffle float64 of tmp2 and tmp2 by 0x1 # asm 1: shufpd $0x1,r2=int6464#1 # asm 2: movdqa r2=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r2 = unpack high double of r2 and r2 # asm 1: unpckhpd r3=int6464#1 # asm 2: movdqa 48(r3=%xmm0 movdqa 48(%rsi),%xmm0 # qhasm: tmp3 = r3 # asm 1: movdqa tmp3=int6464#2 # asm 2: movdqa tmp3=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp3 = shuffle float64 of tmp3 and tmp3 by 0x1 # asm 1: shufpd $0x1,r3=int6464#1 # asm 2: movdqa r3=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r3 = unpack high double of r3 and r3 # asm 1: unpckhpd r4=int6464#1 # asm 2: movdqa 64(r4=%xmm0 movdqa 64(%rsi),%xmm0 # qhasm: tmp4 = r4 # asm 1: movdqa tmp4=int6464#2 # asm 2: movdqa tmp4=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp4 = shuffle float64 of tmp4 and tmp4 by 0x1 # asm 1: shufpd $0x1,r4=int6464#1 # asm 2: movdqa r4=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r4 = unpack high double of r4 and r4 # asm 1: unpckhpd r5=int6464#1 # asm 2: movdqa 80(r5=%xmm0 movdqa 80(%rsi),%xmm0 # qhasm: tmp5 = r5 # asm 1: movdqa tmp5=int6464#2 # asm 2: movdqa tmp5=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp5 = shuffle float64 of tmp5 and tmp5 by 0x1 # asm 1: shufpd $0x1,r5=int6464#1 # asm 2: movdqa r5=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r5 = unpack high double of r5 and r5 # asm 1: unpckhpd r6=int6464#1 # asm 2: movdqa 96(r6=%xmm0 movdqa 96(%rsi),%xmm0 # qhasm: tmp6 = r6 # asm 1: movdqa tmp6=int6464#2 # asm 2: movdqa tmp6=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp6 = shuffle float64 of tmp6 and tmp6 by 0x1 # asm 1: shufpd $0x1,r6=int6464#1 # asm 2: movdqa r6=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r6 = unpack high double of r6 and r6 # asm 1: unpckhpd r7=int6464#1 # asm 2: movdqa 112(r7=%xmm0 movdqa 112(%rsi),%xmm0 # qhasm: tmp7 = r7 # asm 1: movdqa tmp7=int6464#2 # asm 2: movdqa tmp7=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp7 = shuffle float64 of tmp7 and tmp7 by 0x1 # asm 1: shufpd $0x1,r7=int6464#1 # asm 2: movdqa r7=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r7 = unpack high double of r7 and r7 # asm 1: unpckhpd r8=int6464#1 # asm 2: movdqa 128(r8=%xmm0 movdqa 128(%rsi),%xmm0 # qhasm: tmp8 = r8 # asm 1: movdqa tmp8=int6464#2 # asm 2: movdqa tmp8=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp8 = shuffle float64 of tmp8 and tmp8 by 0x1 # asm 1: shufpd $0x1,r8=int6464#1 # asm 2: movdqa r8=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r8 = unpack high double of r8 and r8 # asm 1: unpckhpd r9=int6464#1 # asm 2: movdqa 144(r9=%xmm0 movdqa 144(%rsi),%xmm0 # qhasm: tmp9 = r9 # asm 1: movdqa tmp9=int6464#2 # asm 2: movdqa tmp9=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp9 = shuffle float64 of tmp9 and tmp9 by 0x1 # asm 1: shufpd $0x1,r9=int6464#1 # asm 2: movdqa r9=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r9 = unpack high double of r9 and r9 # asm 1: unpckhpd r10=int6464#1 # asm 2: movdqa 160(r10=%xmm0 movdqa 160(%rsi),%xmm0 # qhasm: tmp10 = r10 # asm 1: movdqa tmp10=int6464#2 # asm 2: movdqa tmp10=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp10 = shuffle float64 of tmp10 and tmp10 by 0x1 # asm 1: shufpd $0x1,r10=int6464#1 # asm 2: movdqa r10=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r10 = unpack high double of r10 and r10 # asm 1: unpckhpd r11=int6464#1 # asm 2: movdqa 176(r11=%xmm0 movdqa 176(%rsi),%xmm0 # qhasm: tmp11 = r11 # asm 1: movdqa tmp11=int6464#2 # asm 2: movdqa tmp11=%xmm1 movdqa %xmm0,%xmm1 # qhasm: tmp11 = shuffle float64 of tmp11 and tmp11 by 0x1 # asm 1: shufpd $0x1,r11=int6464#1 # asm 2: movdqa r11=%xmm0 movdqa %xmm1,%xmm0 # qhasm: r11 = unpack high double of r11 and r11 # asm 1: unpckhpd t1p=int64#2 # asm 2: mov t1p=%rsi mov %rdx,%rsi # qhasm: t2p = myp + 192 # asm 1: lea 192(t2p=int64#4 # asm 2: lea 192(t2p=%rcx lea 192(%rdx),%rcx # qhasm: rp = myp + 384 # asm 1: lea 384(rp=int64#3 # asm 2: lea 384(rp=%rdx lea 384(%rdx),%rdx # qhasm: ab0 = *(int128 *)(t1p + 0) # asm 1: movdqa 0(ab0=int6464#1 # asm 2: movdqa 0(ab0=%xmm0 movdqa 0(%rsi),%xmm0 # qhasm: t0 = ab0 # asm 1: movdqa t0=int6464#2 # asm 2: movdqa t0=%xmm1 movdqa %xmm0,%xmm1 # qhasm: float6464 t0 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(r0=int6464#2 # asm 2: movdqa r0=%xmm1 movdqa %xmm1,%xmm1 # qhasm: t1 = ab0 # asm 1: movdqa t1=int6464#3 # asm 2: movdqa t1=%xmm2 movdqa %xmm0,%xmm2 # qhasm: float6464 t1 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(r1=int6464#3 # asm 2: movdqa r1=%xmm2 movdqa %xmm2,%xmm2 # qhasm: t2 = ab0 # asm 1: movdqa t2=int6464#4 # asm 2: movdqa t2=%xmm3 movdqa %xmm0,%xmm3 # qhasm: float6464 t2 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(r2=int6464#4 # asm 2: movdqa r2=%xmm3 movdqa %xmm3,%xmm3 # qhasm: t3 = ab0 # asm 1: movdqa t3=int6464#5 # asm 2: movdqa t3=%xmm4 movdqa %xmm0,%xmm4 # qhasm: float6464 t3 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(r3=int6464#5 # asm 2: movdqa r3=%xmm4 movdqa %xmm4,%xmm4 # qhasm: t4 = ab0 # asm 1: movdqa t4=int6464#6 # asm 2: movdqa t4=%xmm5 movdqa %xmm0,%xmm5 # qhasm: float6464 t4 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(r4=int6464#6 # asm 2: movdqa r4=%xmm5 movdqa %xmm5,%xmm5 # qhasm: t5 = ab0 # asm 1: movdqa t5=int6464#7 # asm 2: movdqa t5=%xmm6 movdqa %xmm0,%xmm6 # qhasm: float6464 t5 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(r5=int6464#7 # asm 2: movdqa r5=%xmm6 movdqa %xmm6,%xmm6 # qhasm: t6 = ab0 # asm 1: movdqa t6=int6464#8 # asm 2: movdqa t6=%xmm7 movdqa %xmm0,%xmm7 # qhasm: float6464 t6 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(r6=int6464#8 # asm 2: movdqa r6=%xmm7 movdqa %xmm7,%xmm7 # qhasm: t7 = ab0 # asm 1: movdqa t7=int6464#9 # asm 2: movdqa t7=%xmm8 movdqa %xmm0,%xmm8 # qhasm: float6464 t7 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(r7=int6464#9 # asm 2: movdqa r7=%xmm8 movdqa %xmm8,%xmm8 # qhasm: t8 = ab0 # asm 1: movdqa t8=int6464#10 # asm 2: movdqa t8=%xmm9 movdqa %xmm0,%xmm9 # qhasm: float6464 t8 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(r8=int6464#10 # asm 2: movdqa r8=%xmm9 movdqa %xmm9,%xmm9 # qhasm: t9 = ab0 # asm 1: movdqa t9=int6464#11 # asm 2: movdqa t9=%xmm10 movdqa %xmm0,%xmm10 # qhasm: float6464 t9 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(r9=int6464#11 # asm 2: movdqa r9=%xmm10 movdqa %xmm10,%xmm10 # qhasm: t10 = ab0 # asm 1: movdqa t10=int6464#12 # asm 2: movdqa t10=%xmm11 movdqa %xmm0,%xmm11 # qhasm: float6464 t10 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r10=int6464#12 # asm 2: movdqa r10=%xmm11 movdqa %xmm11,%xmm11 # qhasm: r11 = ab0 # asm 1: movdqa r11=int6464#1 # asm 2: movdqa r11=%xmm0 movdqa %xmm0,%xmm0 # qhasm: float6464 r11 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(ab1=int6464#2 # asm 2: movdqa 16(ab1=%xmm1 movdqa 16(%rsi),%xmm1 # qhasm: ab1six = ab1 # asm 1: movdqa ab1six=int6464#13 # asm 2: movdqa ab1six=%xmm12 movdqa %xmm1,%xmm12 # qhasm: float6464 ab1six *= SIX_SIX # asm 1: mulpd SIX_SIX,t1=int6464#14 # asm 2: movdqa t1=%xmm13 movdqa %xmm1,%xmm13 # qhasm: float6464 t1 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t7=int6464#2 # asm 2: movdqa t7=%xmm1 movdqa %xmm1,%xmm1 # qhasm: float6464 t7 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t2=int6464#2 # asm 2: movdqa t2=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t2 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t3=int6464#2 # asm 2: movdqa t3=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t3 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t4=int6464#2 # asm 2: movdqa t4=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t4 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t5=int6464#2 # asm 2: movdqa t5=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t5 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t6=int6464#2 # asm 2: movdqa t6=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t6 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t8=int6464#2 # asm 2: movdqa t8=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t8 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t9=int6464#2 # asm 2: movdqa t9=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t9 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t10=int6464#2 # asm 2: movdqa t10=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t10 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t11=int6464#2 # asm 2: movdqa t11=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 t11 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r12=int6464#2 # asm 2: movdqa r12=%xmm1 movdqa %xmm12,%xmm1 # qhasm: float6464 r12 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(ab2=int6464#3 # asm 2: movdqa 32(ab2=%xmm2 movdqa 32(%rsi),%xmm2 # qhasm: ab2six = ab2 # asm 1: movdqa ab2six=int6464#13 # asm 2: movdqa ab2six=%xmm12 movdqa %xmm2,%xmm12 # qhasm: float6464 ab2six *= SIX_SIX # asm 1: mulpd SIX_SIX,t2=int6464#14 # asm 2: movdqa t2=%xmm13 movdqa %xmm2,%xmm13 # qhasm: float6464 t2 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t7=int6464#14 # asm 2: movdqa t7=%xmm13 movdqa %xmm2,%xmm13 # qhasm: float6464 t7 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t8=int6464#14 # asm 2: movdqa t8=%xmm13 movdqa %xmm2,%xmm13 # qhasm: float6464 t8 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(r13=int6464#3 # asm 2: movdqa r13=%xmm2 movdqa %xmm2,%xmm2 # qhasm: float6464 r13 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t3=int6464#14 # asm 2: movdqa t3=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t3 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t4=int6464#14 # asm 2: movdqa t4=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t4 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t5=int6464#14 # asm 2: movdqa t5=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t5 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t6=int6464#14 # asm 2: movdqa t6=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t6 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t9=int6464#14 # asm 2: movdqa t9=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t9 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t10=int6464#14 # asm 2: movdqa t10=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t10 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t12=int6464#13 # asm 2: movdqa t12=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t12 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(ab3=int6464#4 # asm 2: movdqa 48(ab3=%xmm3 movdqa 48(%rsi),%xmm3 # qhasm: ab3six = ab3 # asm 1: movdqa ab3six=int6464#13 # asm 2: movdqa ab3six=%xmm12 movdqa %xmm3,%xmm12 # qhasm: float6464 ab3six *= SIX_SIX # asm 1: mulpd SIX_SIX,t3=int6464#14 # asm 2: movdqa t3=%xmm13 movdqa %xmm3,%xmm13 # qhasm: float6464 t3 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t7=int6464#14 # asm 2: movdqa t7=%xmm13 movdqa %xmm3,%xmm13 # qhasm: float6464 t7 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t8=int6464#14 # asm 2: movdqa t8=%xmm13 movdqa %xmm3,%xmm13 # qhasm: float6464 t8 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t9=int6464#14 # asm 2: movdqa t9=%xmm13 movdqa %xmm3,%xmm13 # qhasm: float6464 t9 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t13=int6464#14 # asm 2: movdqa t13=%xmm13 movdqa %xmm3,%xmm13 # qhasm: float6464 t13 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r14=int6464#4 # asm 2: movdqa r14=%xmm3 movdqa %xmm3,%xmm3 # qhasm: float6464 r14 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t4=int6464#14 # asm 2: movdqa t4=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t4 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t5=int6464#14 # asm 2: movdqa t5=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t5 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t6=int6464#14 # asm 2: movdqa t6=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t6 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t10=int6464#14 # asm 2: movdqa t10=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t10 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t12=int6464#13 # asm 2: movdqa t12=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t12 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(ab4=int6464#5 # asm 2: movdqa 64(ab4=%xmm4 movdqa 64(%rsi),%xmm4 # qhasm: ab4six = ab4 # asm 1: movdqa ab4six=int6464#13 # asm 2: movdqa ab4six=%xmm12 movdqa %xmm4,%xmm12 # qhasm: float6464 ab4six *= SIX_SIX # asm 1: mulpd SIX_SIX,t4=int6464#14 # asm 2: movdqa t4=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t4 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t7=int6464#14 # asm 2: movdqa t7=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t7 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t8=int6464#14 # asm 2: movdqa t8=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t8 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t9=int6464#14 # asm 2: movdqa t9=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t9 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t10=int6464#14 # asm 2: movdqa t10=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t10 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t13=int6464#14 # asm 2: movdqa t13=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t13 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t14=int6464#14 # asm 2: movdqa t14=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t14 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r15=int6464#5 # asm 2: movdqa r15=%xmm4 movdqa %xmm4,%xmm4 # qhasm: float6464 r15 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t5=int6464#14 # asm 2: movdqa t5=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t5 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t6=int6464#14 # asm 2: movdqa t6=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t6 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t12=int6464#13 # asm 2: movdqa t12=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t12 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(ab5=int6464#6 # asm 2: movdqa 80(ab5=%xmm5 movdqa 80(%rsi),%xmm5 # qhasm: ab5six = ab5 # asm 1: movdqa ab5six=int6464#13 # asm 2: movdqa ab5six=%xmm12 movdqa %xmm5,%xmm12 # qhasm: float6464 ab5six *= SIX_SIX # asm 1: mulpd SIX_SIX,t5=int6464#14 # asm 2: movdqa t5=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t5 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t7=int6464#14 # asm 2: movdqa t7=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t7 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t8=int6464#14 # asm 2: movdqa t8=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t8 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t9=int6464#14 # asm 2: movdqa t9=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t9 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t10=int6464#14 # asm 2: movdqa t10=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t10 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t13=int6464#14 # asm 2: movdqa t13=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t13 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t14=int6464#14 # asm 2: movdqa t14=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t14 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t15=int6464#14 # asm 2: movdqa t15=%xmm13 movdqa %xmm5,%xmm13 # qhasm: float6464 t15 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r16=int6464#6 # asm 2: movdqa r16=%xmm5 movdqa %xmm5,%xmm5 # qhasm: float6464 r16 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t6=int6464#14 # asm 2: movdqa t6=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t6 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t12=int6464#13 # asm 2: movdqa t12=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t12 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(ab6=int6464#7 # asm 2: movdqa 96(ab6=%xmm6 movdqa 96(%rsi),%xmm6 # qhasm: t6 = ab6 # asm 1: movdqa t6=int6464#13 # asm 2: movdqa t6=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t6 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t7=int6464#13 # asm 2: movdqa t7=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t7 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t8=int6464#13 # asm 2: movdqa t8=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t8 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t9=int6464#13 # asm 2: movdqa t9=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t9 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t10=int6464#13 # asm 2: movdqa t10=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t10 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t11=int6464#13 # asm 2: movdqa t11=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t11 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t12=int6464#13 # asm 2: movdqa t12=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t12 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t13=int6464#13 # asm 2: movdqa t13=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t13 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t14=int6464#13 # asm 2: movdqa t14=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t14 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t15=int6464#13 # asm 2: movdqa t15=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t15 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t16=int6464#13 # asm 2: movdqa t16=%xmm12 movdqa %xmm6,%xmm12 # qhasm: float6464 t16 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r17=int6464#7 # asm 2: movdqa r17=%xmm6 movdqa %xmm6,%xmm6 # qhasm: float6464 r17 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(ab7=int6464#8 # asm 2: movdqa 112(ab7=%xmm7 movdqa 112(%rsi),%xmm7 # qhasm: ab7six = ab7 # asm 1: movdqa ab7six=int6464#13 # asm 2: movdqa ab7six=%xmm12 movdqa %xmm7,%xmm12 # qhasm: float6464 ab7six *= SIX_SIX # asm 1: mulpd SIX_SIX,t7=int6464#14 # asm 2: movdqa t7=%xmm13 movdqa %xmm7,%xmm13 # qhasm: float6464 t7 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t13=int6464#8 # asm 2: movdqa t13=%xmm7 movdqa %xmm7,%xmm7 # qhasm: float6464 t13 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t8=int6464#8 # asm 2: movdqa t8=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t8 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t9=int6464#8 # asm 2: movdqa t9=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t9 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t10=int6464#8 # asm 2: movdqa t10=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t10 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t11=int6464#8 # asm 2: movdqa t11=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t11 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t12=int6464#8 # asm 2: movdqa t12=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t12 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t14=int6464#8 # asm 2: movdqa t14=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t14 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t15=int6464#8 # asm 2: movdqa t15=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t15 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t16=int6464#8 # asm 2: movdqa t16=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t16 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t17=int6464#8 # asm 2: movdqa t17=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 t17 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r18=int6464#8 # asm 2: movdqa r18=%xmm7 movdqa %xmm12,%xmm7 # qhasm: float6464 r18 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(ab8=int6464#9 # asm 2: movdqa 128(ab8=%xmm8 movdqa 128(%rsi),%xmm8 # qhasm: ab8six = ab8 # asm 1: movdqa ab8six=int6464#13 # asm 2: movdqa ab8six=%xmm12 movdqa %xmm8,%xmm12 # qhasm: float6464 ab8six *= SIX_SIX # asm 1: mulpd SIX_SIX,t8=int6464#14 # asm 2: movdqa t8=%xmm13 movdqa %xmm8,%xmm13 # qhasm: float6464 t8 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t13=int6464#14 # asm 2: movdqa t13=%xmm13 movdqa %xmm8,%xmm13 # qhasm: float6464 t13 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t14=int6464#14 # asm 2: movdqa t14=%xmm13 movdqa %xmm8,%xmm13 # qhasm: float6464 t14 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(r19=int6464#9 # asm 2: movdqa r19=%xmm8 movdqa %xmm8,%xmm8 # qhasm: float6464 r19 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t9=int6464#14 # asm 2: movdqa t9=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t9 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t10=int6464#14 # asm 2: movdqa t10=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t10 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t12=int6464#14 # asm 2: movdqa t12=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t12 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t15=int6464#14 # asm 2: movdqa t15=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t15 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t16=int6464#14 # asm 2: movdqa t16=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t16 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t17=int6464#14 # asm 2: movdqa t17=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t17 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t18=int6464#13 # asm 2: movdqa t18=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t18 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(ab9=int6464#10 # asm 2: movdqa 144(ab9=%xmm9 movdqa 144(%rsi),%xmm9 # qhasm: ab9six = ab9 # asm 1: movdqa ab9six=int6464#13 # asm 2: movdqa ab9six=%xmm12 movdqa %xmm9,%xmm12 # qhasm: float6464 ab9six *= SIX_SIX # asm 1: mulpd SIX_SIX,t9=int6464#14 # asm 2: movdqa t9=%xmm13 movdqa %xmm9,%xmm13 # qhasm: float6464 t9 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t13=int6464#14 # asm 2: movdqa t13=%xmm13 movdqa %xmm9,%xmm13 # qhasm: float6464 t13 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t14=int6464#14 # asm 2: movdqa t14=%xmm13 movdqa %xmm9,%xmm13 # qhasm: float6464 t14 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t15=int6464#14 # asm 2: movdqa t15=%xmm13 movdqa %xmm9,%xmm13 # qhasm: float6464 t15 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t19=int6464#14 # asm 2: movdqa t19=%xmm13 movdqa %xmm9,%xmm13 # qhasm: float6464 t19 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r20=int6464#10 # asm 2: movdqa r20=%xmm9 movdqa %xmm9,%xmm9 # qhasm: float6464 r20 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t10=int6464#14 # asm 2: movdqa t10=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t10 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t12=int6464#14 # asm 2: movdqa t12=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t12 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t16=int6464#14 # asm 2: movdqa t16=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t16 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t17=int6464#14 # asm 2: movdqa t17=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t17 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t18=int6464#13 # asm 2: movdqa t18=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t18 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(ab10=int6464#11 # asm 2: movdqa 160(ab10=%xmm10 movdqa 160(%rsi),%xmm10 # qhasm: ab10six = ab10 # asm 1: movdqa ab10six=int6464#13 # asm 2: movdqa ab10six=%xmm12 movdqa %xmm10,%xmm12 # qhasm: float6464 ab10six *= SIX_SIX # asm 1: mulpd SIX_SIX,t10=int6464#14 # asm 2: movdqa t10=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t10 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t13=int6464#14 # asm 2: movdqa t13=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t13 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t14=int6464#14 # asm 2: movdqa t14=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t14 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t16=int6464#14 # asm 2: movdqa t16=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t16 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t15=int6464#14 # asm 2: movdqa t15=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t15 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t19=int6464#14 # asm 2: movdqa t19=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t19 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t20=int6464#14 # asm 2: movdqa t20=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t20 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r21=int6464#11 # asm 2: movdqa r21=%xmm10 movdqa %xmm10,%xmm10 # qhasm: float6464 r21 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t12=int6464#14 # asm 2: movdqa t12=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t12 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t17=int6464#14 # asm 2: movdqa t17=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t17 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(t18=int6464#13 # asm 2: movdqa t18=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t18 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(ab11=int6464#12 # asm 2: movdqa 176(ab11=%xmm11 movdqa 176(%rsi),%xmm11 # qhasm: ab11six = ab11 # asm 1: movdqa ab11six=int6464#13 # asm 2: movdqa ab11six=%xmm12 movdqa %xmm11,%xmm12 # qhasm: float6464 ab11six *= SIX_SIX # asm 1: mulpd SIX_SIX,t11=int6464#14 # asm 2: movdqa t11=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t11 *= *(int128 *)(t2p + 0) # asm 1: mulpd 0(t13=int6464#14 # asm 2: movdqa t13=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t13 *= *(int128 *)(t2p + 32) # asm 1: mulpd 32(t14=int6464#14 # asm 2: movdqa t14=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t14 *= *(int128 *)(t2p + 48) # asm 1: mulpd 48(t15=int6464#14 # asm 2: movdqa t15=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t15 *= *(int128 *)(t2p + 64) # asm 1: mulpd 64(t16=int6464#14 # asm 2: movdqa t16=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t16 *= *(int128 *)(t2p + 80) # asm 1: mulpd 80(t17=int6464#14 # asm 2: movdqa t17=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t17 *= *(int128 *)(t2p + 96) # asm 1: mulpd 96(t19=int6464#14 # asm 2: movdqa t19=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t19 *= *(int128 *)(t2p + 128) # asm 1: mulpd 128(t20=int6464#14 # asm 2: movdqa t20=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t20 *= *(int128 *)(t2p + 144) # asm 1: mulpd 144(t21=int6464#14 # asm 2: movdqa t21=%xmm13 movdqa %xmm11,%xmm13 # qhasm: float6464 t21 *= *(int128 *)(t2p + 160) # asm 1: mulpd 160(r22=int6464#12 # asm 2: movdqa r22=%xmm11 movdqa %xmm11,%xmm11 # qhasm: float6464 r22 *= *(int128 *)(t2p + 176) # asm 1: mulpd 176(t12=int6464#14 # asm 2: movdqa t12=%xmm13 movdqa %xmm12,%xmm13 # qhasm: float6464 t12 *= *(int128 *)(t2p + 16) # asm 1: mulpd 16(t18=int6464#13 # asm 2: movdqa t18=%xmm12 movdqa %xmm12,%xmm12 # qhasm: float6464 t18 *= *(int128 *)(t2p + 112) # asm 1: mulpd 112(r0=int6464#1 # asm 2: movdqa 0(r0=%xmm0 movdqa 0(%rdx),%xmm0 # qhasm: float6464 r0 -= r12 # asm 1: subpd t15=int6464#13 # asm 2: movdqa t15=%xmm12 movdqa %xmm4,%xmm12 # qhasm: float6464 t15 *= SIX_SIX # asm 1: mulpd SIX_SIX,t18=int6464#13 # asm 2: movdqa t18=%xmm12 movdqa %xmm7,%xmm12 # qhasm: float6464 t18 *= TWO_TWO # asm 1: mulpd TWO_TWO,t21=int6464#13 # asm 2: movdqa t21=%xmm12 movdqa %xmm10,%xmm12 # qhasm: float6464 t21 *= SIX_SIX # asm 1: mulpd SIX_SIX,r3=int6464#13 # asm 2: movdqa 48(r3=%xmm12 movdqa 48(%rdx),%xmm12 # qhasm: float6464 r3 -= r12 # asm 1: subpd t15=int6464#14 # asm 2: movdqa t15=%xmm13 movdqa %xmm4,%xmm13 # qhasm: float6464 t15 *= FIVE_FIVE # asm 1: mulpd FIVE_FIVE,t21=int6464#14 # asm 2: movdqa t21=%xmm13 movdqa %xmm10,%xmm13 # qhasm: float6464 t21 *= EIGHT_EIGHT # asm 1: mulpd EIGHT_EIGHT,r6=int6464#14 # asm 2: movdqa 96(r6=%xmm13 movdqa 96(%rdx),%xmm13 # qhasm: t12 = r12 # asm 1: movdqa t12=int6464#15 # asm 2: movdqa t12=%xmm14 movdqa %xmm1,%xmm14 # qhasm: float6464 t12 *= FOUR_FOUR # asm 1: mulpd FOUR_FOUR,t15=int6464#15 # asm 2: movdqa t15=%xmm14 movdqa %xmm4,%xmm14 # qhasm: float6464 t15 *= EIGHTEEN_EIGHTEEN # asm 1: mulpd EIGHTEEN_EIGHTEEN,t18=int6464#15 # asm 2: movdqa t18=%xmm14 movdqa %xmm7,%xmm14 # qhasm: float6464 t18 *= THREE_THREE # asm 1: mulpd THREE_THREE,t21=int6464#15 # asm 2: movdqa t21=%xmm14 movdqa %xmm10,%xmm14 # qhasm: float6464 t21 *= THIRTY_THIRTY # asm 1: mulpd THIRTY_THIRTY,r9=int6464#15 # asm 2: movdqa 144(r9=%xmm14 movdqa 144(%rdx),%xmm14 # qhasm: float6464 r9 -= r12 # asm 1: subpd t15=int6464#2 # asm 2: movdqa t15=%xmm1 movdqa %xmm4,%xmm1 # qhasm: float6464 t15 *= TWO_TWO # asm 1: mulpd TWO_TWO,t21=int6464#2 # asm 2: movdqa t21=%xmm1 movdqa %xmm10,%xmm1 # qhasm: float6464 t21 *= NINE_NINE # asm 1: mulpd NINE_NINE,r1=int6464#2 # asm 2: movdqa 16(r1=%xmm1 movdqa 16(%rdx),%xmm1 # qhasm: float6464 r1 -= r13 # asm 1: subpd t19=int6464#5 # asm 2: movdqa t19=%xmm4 movdqa %xmm8,%xmm4 # qhasm: float6464 t19 *= TWO_TWO # asm 1: mulpd TWO_TWO,r4=int6464#5 # asm 2: movdqa 64(r4=%xmm4 movdqa 64(%rdx),%xmm4 # qhasm: t13 = r13 # asm 1: movdqa t13=int6464#8 # asm 2: movdqa t13=%xmm7 movdqa %xmm2,%xmm7 # qhasm: float6464 t13 *= SIX_SIX # asm 1: mulpd SIX_SIX,t16=int6464#8 # asm 2: movdqa t16=%xmm7 movdqa %xmm5,%xmm7 # qhasm: float6464 t16 *= FIVE_FIVE # asm 1: mulpd FIVE_FIVE,t19=int6464#8 # asm 2: movdqa t19=%xmm7 movdqa %xmm8,%xmm7 # qhasm: float6464 t19 *= SIX_SIX # asm 1: mulpd SIX_SIX,t22=int6464#8 # asm 2: movdqa t22=%xmm7 movdqa %xmm11,%xmm7 # qhasm: float6464 t22 *= EIGHT_EIGHT # asm 1: mulpd EIGHT_EIGHT,r7=int6464#8 # asm 2: movdqa 112(r7=%xmm7 movdqa 112(%rdx),%xmm7 # qhasm: t13 = r13 # asm 1: movdqa t13=int6464#11 # asm 2: movdqa t13=%xmm10 movdqa %xmm2,%xmm10 # qhasm: float6464 t13 *= FOUR_FOUR # asm 1: mulpd FOUR_FOUR,t16=int6464#11 # asm 2: movdqa t16=%xmm10 movdqa %xmm5,%xmm10 # qhasm: float6464 t16 *= THREE_THREE # asm 1: mulpd THREE_THREE,t19=int6464#11 # asm 2: movdqa t19=%xmm10 movdqa %xmm8,%xmm10 # qhasm: float6464 t19 *= THREE_THREE # asm 1: mulpd THREE_THREE,t22=int6464#11 # asm 2: movdqa t22=%xmm10 movdqa %xmm11,%xmm10 # qhasm: float6464 t22 *= FIVE_FIVE # asm 1: mulpd FIVE_FIVE,r10=int6464#11 # asm 2: movdqa 160(r10=%xmm10 movdqa 160(%rdx),%xmm10 # qhasm: t13 = r13 # asm 1: movdqa t13=int6464#3 # asm 2: movdqa t13=%xmm2 movdqa %xmm2,%xmm2 # qhasm: float6464 t13 *= SIX_SIX # asm 1: mulpd SIX_SIX,t16=int6464#3 # asm 2: movdqa t16=%xmm2 movdqa %xmm5,%xmm2 # qhasm: float6464 t16 *= TWO_TWO # asm 1: mulpd TWO_TWO,t19=int6464#3 # asm 2: movdqa t19=%xmm2 movdqa %xmm8,%xmm2 # qhasm: float6464 t19 *= SIX_SIX # asm 1: mulpd SIX_SIX,t22=int6464#3 # asm 2: movdqa t22=%xmm2 movdqa %xmm11,%xmm2 # qhasm: float6464 t22 *= NINE_NINE # asm 1: mulpd NINE_NINE,r2=int6464#3 # asm 2: movdqa 32(r2=%xmm2 movdqa 32(%rdx),%xmm2 # qhasm: float6464 r2 -= r14 # asm 1: subpd t20=int6464#6 # asm 2: movdqa t20=%xmm5 movdqa %xmm9,%xmm5 # qhasm: float6464 t20 *= TWO_TWO # asm 1: mulpd TWO_TWO,r5=int6464#6 # asm 2: movdqa 80(r5=%xmm5 movdqa 80(%rdx),%xmm5 # qhasm: t14 = r14 # asm 1: movdqa t14=int6464#9 # asm 2: movdqa t14=%xmm8 movdqa %xmm3,%xmm8 # qhasm: float6464 t14 *= SIX_SIX # asm 1: mulpd SIX_SIX,t17=int6464#9 # asm 2: movdqa t17=%xmm8 movdqa %xmm6,%xmm8 # qhasm: float6464 t17 *= FIVE_FIVE # asm 1: mulpd FIVE_FIVE,t20=int6464#9 # asm 2: movdqa t20=%xmm8 movdqa %xmm9,%xmm8 # qhasm: float6464 t20 *= SIX_SIX # asm 1: mulpd SIX_SIX,r8=int6464#9 # asm 2: movdqa 128(r8=%xmm8 movdqa 128(%rdx),%xmm8 # qhasm: t14 = r14 # asm 1: movdqa t14=int6464#12 # asm 2: movdqa t14=%xmm11 movdqa %xmm3,%xmm11 # qhasm: float6464 t14 *= FOUR_FOUR # asm 1: mulpd FOUR_FOUR,t17=int6464#12 # asm 2: movdqa t17=%xmm11 movdqa %xmm6,%xmm11 # qhasm: float6464 t17 *= THREE_THREE # asm 1: mulpd THREE_THREE,t20=int6464#12 # asm 2: movdqa t20=%xmm11 movdqa %xmm9,%xmm11 # qhasm: float6464 t20 *= THREE_THREE # asm 1: mulpd THREE_THREE,r11=int6464#12 # asm 2: movdqa 176(r11=%xmm11 movdqa 176(%rdx),%xmm11 # qhasm: t14 = r14 # asm 1: movdqa t14=int6464#4 # asm 2: movdqa t14=%xmm3 movdqa %xmm3,%xmm3 # qhasm: float6464 t14 *= SIX_SIX # asm 1: mulpd SIX_SIX,t17=int6464#4 # asm 2: movdqa t17=%xmm3 movdqa %xmm6,%xmm3 # qhasm: float6464 t17 *= TWO_TWO # asm 1: mulpd TWO_TWO,t20=int6464#4 # asm 2: movdqa t20=%xmm3 movdqa %xmm9,%xmm3 # qhasm: float6464 t20 *= SIX_SIX # asm 1: mulpd SIX_SIX,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm1,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm4,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm7,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm10,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm2,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm5,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm8,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm11,%xmm6 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,2t6=int6464#10 # asm 2: movdqa 2t6=%xmm9 movdqa %xmm6,%xmm9 # qhasm: float6464 2t6 *= FOUR_FOUR # asm 1: mulpd FOUR_FOUR,<2t6=int6464#10 # asm 2: mulpd FOUR_FOUR,<2t6=%xmm9 mulpd FOUR_FOUR,%xmm9 # qhasm: float6464 r6 -= 2t6 # asm 1: subpd <2t6=int6464#10,carry=int6464#7 # asm 2: movdqa carry=%xmm6 movdqa %xmm0,%xmm6 # qhasm: float6464 carry *= V6INV_V6INV # asm 1: mulpd V6INV_V6INV,carry=int6464#1 # asm 2: movdqa carry=%xmm0 movdqa %xmm12,%xmm0 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#1 # asm 2: movdqa carry=%xmm0 movdqa %xmm13,%xmm0 # qhasm: float6464 carry *= V6INV_V6INV # asm 1: mulpd V6INV_V6INV,carry=int6464#1 # asm 2: movdqa carry=%xmm0 movdqa %xmm14,%xmm0 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#1 # asm 2: movdqa carry=%xmm0 movdqa %xmm1,%xmm0 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#1 # asm 2: movdqa carry=%xmm0 movdqa %xmm4,%xmm0 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#1 # asm 2: movdqa carry=%xmm0 movdqa %xmm7,%xmm0 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,carry=int6464#1 # asm 2: movdqa carry=%xmm0 movdqa %xmm10,%xmm0 # qhasm: float6464 carry *= VINV_VINV # asm 1: mulpd VINV_VINV,