| 12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964 | /** Math library** Copyright (C) 2016 Intel Corporation. All rights reserved.** Redistribution and use in source and binary forms, with or without* modification, are permitted provided that the following conditions* are met:**   * Redistributions of source code must retain the above copyright*     notice, this list of conditions and the following disclaimer.*   * Redistributions in binary form must reproduce the above copyright*     notice, this list of conditions and the following disclaimer in*     the documentation and/or other materials provided with the*     distribution.*   * Neither the name of Intel Corporation nor the names of its*     contributors may be used to endorse or promote products derived*     from this software without specific prior written permission.** THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.*** Author Name <jingwei.zhang@intel.com>*   History:*   03-14-2016 Initial version. numerics svn rev. 12864*/	.file "dpml_ux_ops_64.c"	.text..TXTST0:# -- Begin  __dpml_extended_multiply__	.text       .align    16,0x90	.globl __dpml_extended_multiply____dpml_extended_multiply__:# parameter 1: %rdi# parameter 2: %rsi# parameter 3: %rdx# parameter 4: %rcx..B1.1:	.cfi_startproc..___tag_value___dpml_extended_multiply__.1:..L2:        pushq     %r12	.cfi_def_cfa_offset 16	.cfi_offset 12, -16        pushq     %r13	.cfi_def_cfa_offset 24	.cfi_offset 13, -24        pushq     %r14	.cfi_def_cfa_offset 32	.cfi_offset 14, -32        pushq     %r15	.cfi_def_cfa_offset 40	.cfi_offset 15, -40        pushq     %rbx	.cfi_def_cfa_offset 48	.cfi_offset 3, -48        pushq     %rbp	.cfi_def_cfa_offset 56	.cfi_offset 6, -56        movq      %rdi, %r15        movq      16(%rsi), %r14        movq      %r14, %r10        movl      %r14d, %r9d        movq      %rcx, %r8        movq      %r9, %r13        movq      16(%r15), %rax        movq      %rax, %rbp        movl      %eax, %ebx        movq      %rax, %rdi        shrq      $32, %rbp        imulq     %rbx, %r13        imulq     %r14, %rdi        movq      %r9, -8(%rsp)        imulq     %rbp, %r9        shrq      $32, %r13        shrq      $32, %r10        addq      %r9, %r13        movq      %rbx, %r9        imulq     %r10, %r9        addq      %r13, %r9        movq      %r9, %rcx        cmpq      %r13, %r9        movq      %rbp, %r9        movl      $0, %r13d        movq      8(%r15), %r12        setb      %r13b        imulq     %r10, %r9        imulq     %r12, %r14        movq      8(%rsi), %r11        imulq     %r11, %rax        shrq      $32, %rcx        shlq      $32, %r13        addq      %r13, %rcx        addq      %r9, %rcx        movl      (%r15), %r9d        addq      %r14, %rcx        movq      %rdi, 16(%r8)        xorl      (%rsi), %r9d        cmpq      %r14, %rcx        movl      4(%rsi), %esi        movq      %r12, %r14        movl      4(%r15), %edi        movl      %r9d, (%rdx)        movl      %r9d, (%r8)        movl      $0, %r9d        setb      %r9b        addq      %rax, %rcx        movq      %rcx, 8(%r8)        cmpq      %rax, %rcx        movl      %r11d, %ecx        lea       (%rdi,%rsi), %r13d        movl      %r13d, 4(%rdx)        lea       -128(%rdi,%rsi), %r15d        movl      %r15d, 4(%r8)        movq      %rcx, %r15        movq      %rcx, %r13        movl      $0, %eax        setb      %al        imulq     %rbx, %r15        imulq     %r11, %r14        imulq     %rbp, %r13        shrq      $32, %r11        imulq     %r11, %rbx        imulq     %r11, %rbp        shrq      $32, %r15        addq      %r13, %r15        addq      %r15, %rbx        cmpq      %r15, %rbx        movq      %rbx, %r8        movl      $0, %ebx        setb      %bl        shrq      $32, %r8        shlq      $32, %rbx        addq      %rbx, %r8        addq      %rbp, %r8        movq      -8(%rsp), %rbp        movq      %rbp, %rdi        movl      %r12d, %r15d        shrq      $32, %r12        movq      %r15, %rsi        imulq     %r15, %rdi        imulq     %r12, %rbp        imulq     %r10, %rsi        imulq     %r12, %r10        addq      %rax, %r9        xorl      %eax, %eax        shrq      $32, %rdi        addq      %r14, %r9        addq      %rbp, %rdi        cmpq      %r14, %r9        movl      $0, %r14d        setb      %r14b        addq      %rdi, %rsi        cmpq      %rdi, %rsi        movq      %rsi, %r13        setb      %al        addq      %r8, %r9        shrq      $32, %r13        shlq      $32, %rax        addq      %rax, %r13        movq      %rcx, %rax        imulq     %r15, %rax        imulq     %r12, %rcx        imulq     %r11, %r15        imulq     %r11, %r12        shrq      $32, %rax        addq      %r10, %r13        addq      %rcx, %rax        xorl      %r10d, %r10d        cmpq      %r8, %r9        setb      %r10b        xorl      %r11d, %r11d        addq      %rax, %r15        cmpq      %rax, %r15        movq      %r15, %rcx        movl      $0, %r15d        setb      %r15b        addq      %r13, %r9        cmpq      %r13, %r9        setb      %r11b        shrq      $32, %rcx        addq      %r10, %r14        shlq      $32, %r15        addq      %r15, %rcx        addq      %r12, %rcx        addq      %r11, %r14        addq      %r14, %rcx        movq      %r9, 16(%rdx)        movq      %rcx, 8(%rdx)	.cfi_restore 6        popq      %rbp	.cfi_def_cfa_offset 48	.cfi_restore 3        popq      %rbx	.cfi_def_cfa_offset 40	.cfi_restore 15        popq      %r15	.cfi_def_cfa_offset 32	.cfi_restore 14        popq      %r14	.cfi_def_cfa_offset 24	.cfi_restore 13        popq      %r13	.cfi_def_cfa_offset 16	.cfi_restore 12        popq      %r12	.cfi_def_cfa_offset 8        ret               .align    16,0x90	.cfi_endproc	.type	__dpml_extended_multiply__,@function	.size	__dpml_extended_multiply__,.-__dpml_extended_multiply__	.data# -- End  __dpml_extended_multiply__	.text# -- Begin  __dpml_evaluate_rational__	.text       .align    16,0x90	.globl __dpml_evaluate_rational____dpml_evaluate_rational__:# parameter 1: %rdi# parameter 2: %rsi# parameter 3: %rdx# parameter 4: %rcx# parameter 5: %r8..B2.1:	.cfi_startproc..___tag_value___dpml_evaluate_rational__.28:..L29:        pushq     %r12	.cfi_def_cfa_offset 16	.cfi_offset 12, -16        pushq     %r13	.cfi_def_cfa_offset 24	.cfi_offset 13, -24        pushq     %r14	.cfi_def_cfa_offset 32	.cfi_offset 14, -32        pushq     %r15	.cfi_def_cfa_offset 40	.cfi_offset 15, -40        pushq     %rbx	.cfi_def_cfa_offset 48	.cfi_offset 3, -48        pushq     %rbp	.cfi_def_cfa_offset 56	.cfi_offset 6, -56        subq      $88, %rsp	.cfi_def_cfa_offset 144        movq      %rdi, %r13        movq      %rcx, %rbx        movq      %r8, %r14        movq      %rbx, %r9        movq      %rsi, %r15        movq      %rdx, 32(%rsp)        movq      %rbx, %rbp        sarq      $58, %r9        movq      %fs:40, %rax        xorq      %rsp, %rax        movq      %rax, 72(%rsp)        addl      %r9d, 4(%r13)        testq     $68, %rbx        jne       ..B2.3..B2.2:        movl      $136, %eax        xorl      %ebp, %ebp        cmpl      $0, (%r13)        movq      %r13, %r12        cmovne    %rax, %rbp        xorq      %rbx, %rbp        jmp       ..B2.4..B2.3:        movq      %r13, %rdi        lea       48(%rsp), %r12        movq      %r13, %rsi        movq      %r12, %rdx..___tag_value___dpml_evaluate_rational__.43:        call      __dpml_multiply__@PLT..___tag_value___dpml_evaluate_rational__.44:..B2.4:        movq      %r12, %rdi        xorl      %esi, %esi..___tag_value___dpml_evaluate_rational__.45:        call      __dpml_ffs_and_shift__@PLT..___tag_value___dpml_evaluate_rational__.46:..B2.5:        movq      32(%rsp), %r9        movq      %r9, %r8        negq      %r8        movslq    4(%r12), %rax        imulq     %rax, %r8        shlq      $4, %r9        movq      %r8, 24(%rsp)        testq     $512, %rbx        movq      %r9, 40(%rsp)        je        ..B2.7..B2.6:        testq     $1024, %rbx        je        ..B2.8..B2.7:        xorl      %eax, %eax        jmp       ..B2.9..B2.8:        movl      $1, %eax..B2.9:        testq     $15, %rbx        lea       (%rax,%rax,2), %rax        lea       (,%rax,8), %r9        lea       (%r14,%rax,8), %r8        je        ..B2.16..B2.10:        testq     $8, %rbp        lea       __eval_neg_poly(%rip), %r11        lea       __eval_pos_poly(%rip), %r10        movq      %rbx, %rax        cmovne    %r11, %r10        negq      %r9        andq      $240, %rax        cmove     %r14, %r8        addq      %r14, %r9        addq      $24, %r9        cmpq      %r11, %r10        jne       ..B2.12..B2.11:        movq      %r12, %rdi        movq      %r15, %rdx        movq      24(%rsp), %rsi        movq      32(%rsp), %rcx        movq      %r8, (%rsp)        movq      %rax, 8(%rsp)        movq      %r9, 16(%rsp)..___tag_value___dpml_evaluate_rational__.47:        call      __eval_neg_poly..___tag_value___dpml_evaluate_rational__.48:        jmp       ..B2.36..B2.12:        movq      %r12, %rdi        movq      %r15, %rdx        movq      24(%rsp), %rsi        movq      32(%rsp), %rcx        movq      %r8, (%rsp)        movq      %rax, 8(%rsp)        movq      %r9, 16(%rsp)..___tag_value___dpml_evaluate_rational__.49:        call      __eval_pos_poly..___tag_value___dpml_evaluate_rational__.50:..B2.36:        movq      16(%rsp), %r9        movq      8(%rsp), %rax        movq      (%rsp), %r8..B2.13:        testq     $2, %rbx        je        ..B2.15..B2.14:        movq      %r13, %rdi        movq      %r8, %rsi        movq      %r8, %rdx        movq      %r8, (%rsp)        movq      %rax, 8(%rsp)        movq      %r9, 16(%rsp)..___tag_value___dpml_evaluate_rational__.51:        call      __dpml_multiply__@PLT..___tag_value___dpml_evaluate_rational__.52:..B2.35:        movq      16(%rsp), %r9        movq      8(%rsp), %rax        movq      (%rsp), %r8..B2.15:        movq      40(%rsp), %r10        lea       24(%r15,%r10), %r15        movl      -8(%r15), %r11d        addl      %r11d, 4(%r8)        jmp       ..B2.17..B2.16:        movq      40(%rsp), %r10        orq       $256, %rbx        testq     $1024, %rbx        movq      %rbx, %rax        movq      %r14, %r9        lea       24(%r15,%r10), %r11        cmovne    %r11, %r15        andq      $240, %rax..B2.17:        testq     %rax, %rax        je        ..B2.26..B2.18:        testq     $128, %rbp        lea       __eval_neg_poly(%rip), %rbp        lea       __eval_pos_poly(%rip), %rax        cmovne    %rbp, %rax        cmpq      %rbp, %rax        jne       ..B2.20..B2.19:        movq      %r12, %rdi        movq      %r15, %rdx        movq      %r9, %r8        movq      24(%rsp), %rsi        movq      32(%rsp), %rcx        movq      %r9, 16(%rsp)..___tag_value___dpml_evaluate_rational__.53:        call      __eval_neg_poly..___tag_value___dpml_evaluate_rational__.54:        jmp       ..B2.39..B2.20:        movq      %r12, %rdi        movq      %r15, %rdx        movq      %r9, %r8        movq      24(%rsp), %rsi        movq      32(%rsp), %rcx        movq      %r9, 16(%rsp)..___tag_value___dpml_evaluate_rational__.55:        call      __eval_pos_poly..___tag_value___dpml_evaluate_rational__.56:..B2.39:        movq      16(%rsp), %r9..B2.21:        testq     $32, %rbx        je        ..B2.23..B2.22:        movq      %r13, %rdi        movq      %r9, %rsi        movq      %r9, %rdx        movq      %r9, 16(%rsp)..___tag_value___dpml_evaluate_rational__.57:        call      __dpml_multiply__@PLT..___tag_value___dpml_evaluate_rational__.58:..B2.38:        movq      16(%rsp), %r9..B2.23:        movq      40(%rsp), %rax        movl      16(%rax,%r15), %ebp        addl      %ebp, 4(%r9)        testq     $1024, %rbx        je        ..B2.27..B2.24:        movq      72(%rsp), %rax        xorq      %rsp, %rax        cmpq      %fs:40, %rax        jne       ..B2.31..B2.25:        addq      $88, %rsp	.cfi_def_cfa_offset 56	.cfi_restore 6        popq      %rbp	.cfi_def_cfa_offset 48	.cfi_restore 3        popq      %rbx	.cfi_def_cfa_offset 40	.cfi_restore 15        popq      %r15	.cfi_def_cfa_offset 32	.cfi_restore 14        popq      %r14	.cfi_def_cfa_offset 24	.cfi_restore 13        popq      %r13	.cfi_def_cfa_offset 16	.cfi_restore 12        popq      %r12	.cfi_def_cfa_offset 8        ret       	.cfi_def_cfa_offset 144	.cfi_offset 3, -48	.cfi_offset 6, -56	.cfi_offset 12, -16	.cfi_offset 13, -24	.cfi_offset 14, -32	.cfi_offset 15, -40..B2.26:        orq       $256, %rbx..B2.27:        testq     $256, %rbx        jne       ..B2.29..B2.28:        movq      %r14, %rdi        lea       24(%r14), %rsi        movl      $2, %edx        movq      %r14, %rcx..___tag_value___dpml_evaluate_rational__.79:        call      __dpml_divide__@PLT..___tag_value___dpml_evaluate_rational__.80:..B2.29:        movq      72(%rsp), %rax        xorq      %rsp, %rax        cmpq      %fs:40, %rax        jne       ..B2.31..B2.30:        addq      $88, %rsp	.cfi_def_cfa_offset 56	.cfi_restore 6        popq      %rbp	.cfi_def_cfa_offset 48	.cfi_restore 3        popq      %rbx	.cfi_def_cfa_offset 40	.cfi_restore 15        popq      %r15	.cfi_def_cfa_offset 32	.cfi_restore 14        popq      %r14	.cfi_def_cfa_offset 24	.cfi_restore 13        popq      %r13	.cfi_def_cfa_offset 16	.cfi_restore 12        popq      %r12	.cfi_def_cfa_offset 8        ret       	.cfi_def_cfa_offset 144	.cfi_offset 3, -48	.cfi_offset 6, -56	.cfi_offset 12, -16	.cfi_offset 13, -24	.cfi_offset 14, -32	.cfi_offset 15, -40..B2.31:        call      __stack_chk_fail@PLT        .align    16,0x90	.cfi_endproc	.type	__dpml_evaluate_rational__,@function	.size	__dpml_evaluate_rational__,.-__dpml_evaluate_rational__	.data# -- End  __dpml_evaluate_rational__	.text# -- Begin  __eval_neg_poly	.text       .align    16,0x90__eval_neg_poly:# parameter 1: %rdi# parameter 2: %rsi# parameter 3: %rdx# parameter 4: %rcx# parameter 5: %r8..B3.1:	.cfi_startproc..___tag_value___eval_neg_poly.102:..L103:        pushq     %r12	.cfi_def_cfa_offset 16	.cfi_offset 12, -16        pushq     %r13	.cfi_def_cfa_offset 24	.cfi_offset 13, -24        pushq     %r14	.cfi_def_cfa_offset 32	.cfi_offset 14, -32        pushq     %r15	.cfi_def_cfa_offset 40	.cfi_offset 15, -40        pushq     %rbx	.cfi_def_cfa_offset 48	.cfi_offset 3, -48        pushq     %rbp	.cfi_def_cfa_offset 56	.cfi_offset 6, -56        movq      %rsi, %r12        xorl      %r10d, %r10d        cmpq      $128, %r12        movq      %rcx, %rax        movl      $1, %ecx        movslq    4(%rdi), %r11        movq      %rdx, %r13        movq      8(%rdi), %r14        movq      16(%rdi), %rdx        movq      %r11, %rdi        cmovl     %ecx, %r10d        xorl      %ebx, %ebx        negq      %rdi        xorl      %r9d, %r9d        testl     %r10d, %r10d        jne       ..B3.5..B3.3:        addq      %r11, %r12        addq      $16, %r13        decq      %rax        cmpq      $128, %r12        jge       ..B3.3..B3.5:        xorl      %ebp, %ebp        cmpq      $64, %r12        cmovge    %ebp, %ecx        testl     %ecx, %ecx        jne       ..B3.12..B3.7:        movl      %r12d, %ecx        decq      %rax        addq      %r11, %r12        movq      8(%r13), %r9        addq      $16, %r13        shrq      %cl, %r9        cmpq      $64, %r12        jl        ..B3.12..B3.8:        testq     %r9, %r9        je        ..B3.7..B3.9:        movq      %r14, %rbp        xorl      %r15d, %r15d        movl      %r14d, %esi        shrq      $32, %rbp        movq      %r8, -72(%rsp)..B3.10:        movl      %r9d, %r10d        movq      %rsi, %r8        imulq     %r10, %r8        imulq     %rbp, %r10        shrq      $32, %r9        movq      %rsi, %rcx        imulq     %r9, %rcx        imulq     %rbp, %r9        shrq      $32, %r8        decq      %rax        addq      %r10, %r8        addq      %r8, %rcx        cmpq      %r8, %rcx        movl      %r15d, %r8d        movq      %rcx, %r10        movl      %r12d, %ecx        setb      %r8b        addq      %r11, %r12        shrq      $32, %r10        shlq      $32, %r8        addq      %r8, %r10        addq      %r9, %r10        movq      8(%r13), %r9        addq      $16, %r13        shrq      %cl, %r9        subq      %r10, %r9        cmpq      $64, %r12        jge       ..B3.10..B3.11:        movq      -72(%rsp), %r8..B3.12:        movq      %r12, %r10        negq      %r10        testq     %r12, %r12        je        ..B3.19..B3.13:        movq      %r14, %rsi        movl      %r14d, %ebp        shrq      $32, %rsi        movq      %r14, -16(%rsp)        xorl      %r14d, %r14d        movq      %r8, -72(%rsp)..B3.14:        movl      %r9d, %ebx        movq      %rbp, %rcx        imulq     %rbx, %rcx        imulq     %rsi, %rbx        shrq      $32, %r9        movq      %rbp, %r15        imulq     %r9, %r15        imulq     %rsi, %r9        shrq      $32, %rcx        decq      %rax        addq      %rbx, %rcx        movl      %r14d, %ebx        addq      %rcx, %r15        cmpq      %rcx, %r15        movq      %r15, %r8        movl      %r12d, %ecx        setb      %bl        shrq      $32, %r8        shlq      $32, %rbx        addq      %rbx, %r8        movq      8(%r13), %rbx        addq      %r9, %r8        movq      (%r13), %r15        movq      %rbx, %r9        shrq      %cl, %r15        movl      %r10d, %ecx        shlq      %cl, %r9        movl      %r12d, %ecx        orq       %r9, %r15        addq      %rdi, %r10        movq      %r15, %r9        addq      $16, %r13        subq      %r8, %r9        movl      %r14d, %r8d        cmpq      %r15, %r9        seta      %r8b        shrq      %cl, %rbx        subq      %r8, %rbx        addq      %r11, %r12        je        ..B3.24..B3.15:        testq     %rbx, %rbx        je        ..B3.14..B3.16:        movq      %rdx, %rcx        movl      %edx, %r15d        shrq      $32, %rcx        movq      %rdx, -8(%rsp)        movq      %rcx, -48(%rsp)        movq      %r15, -40(%rsp)        movq      %rsi, -32(%rsp)        movq      %rdi, -56(%rsp)        movq      %r11, -64(%rsp)..B3.17:        movl      %r12d, %ecx        decq      %rax        movq      8(%r13), %r14        movq      %r14, %r11        movq      (%r13), %rdx        addq      $16, %r13        shrq      %cl, %rdx        movl      %r10d, %ecx        shlq      %cl, %r11        movl      %r12d, %ecx        orq       %r11, %rdx        movq      -40(%rsp), %r11        movq      %r11, %r8        movl      %ebx, %esi        movq      -48(%rsp), %r15        shrq      %cl, %r14        movq      %r15, %rcx        imulq     %rsi, %r8        imulq     %rsi, %rcx        shrq      $32, %r8        movq      -16(%rsp), %rdi        addq      %rcx, %r8        imulq     %rbx, %rdi        shrq      $32, %rbx        movq      %r11, %rcx        imulq     %rbx, %rcx        imulq     %rbx, %r15        addq      %r8, %rcx        negq      %rdi        cmpq      %r8, %rcx        movl      $0, %r8d        movq      %rcx, %r11        movl      %r8d, %ecx        setb      %cl        addq      %rdx, %rdi        cmpq      %rdx, %rdi        movq      %rax, -24(%rsp)        movl      %r8d, %edx        movl      %r9d, %eax        seta      %dl        shrq      $32, %r11        shlq      $32, %rcx        addq      %rcx, %r11        addq      %r15, %r11        movq      %rbp, %r15        imulq     %rax, %r15        movq      -32(%rsp), %rcx        negq      %r11        imulq     %rcx, %rax        shrq      $32, %r15        addq      %rdi, %r11        shrq      $32, %r9        addq      %rax, %r15        movq      %rbp, %rax        imulq     %r9, %rax        imulq     %rcx, %r9        addq      %r15, %rax        cmpq      %r15, %rax        movl      %r8d, %r15d        setb      %r15b        subq      %rdx, %r14        movq      %rax, %rdx        shrq      $32, %rdx        shlq      $32, %r15        addq      %r15, %rdx        xorl      %r15d, %r15d        addq      %r9, %rdx        xorl      %r9d, %r9d        cmpq      %rdi, %r11        movq      %rbp, %rdi        seta      %r9b        imulq     %rsi, %rdi        imulq     %rcx, %rsi        shrq      $32, %rdi        addq      %rsi, %rdi        movq      %rbp, %rsi        imulq     %rbx, %rsi        imulq     %rcx, %rbx        addq      %rdi, %rsi        subq      %r9, %r14        movq      %r11, %r9        cmpq      %rdi, %rsi        movq      %rsi, %rax        setb      %r15b        subq      %rdx, %r9        cmpq      %r11, %r9        seta      %r8b        shrq      $32, %rax        shlq      $32, %r15        addq      %r15, %rax        subq      %r8, %r14        addq      %rbx, %rax        movq      %r14, %rbx        addq      -56(%rsp), %r10        subq      %rax, %rbx        movq      -24(%rsp), %rax        addq      -64(%rsp), %r12        jne       ..B3.17..B3.18:        movq      -8(%rsp), %rdx        movq      -16(%rsp), %r14        movq      -72(%rsp), %r8..B3.19:        testq     %rax, %rax        js        ..B3.23..B3.20:        movl      %edx, %r12d        movq      %r14, %r10        shrq      $32, %rdx        movl      %r14d, %r11d        shrq      $32, %r10        movq      %rdx, -8(%rsp)        movq      %r14, -16(%rsp)        movq      %r8, -72(%rsp)..B3.21:        movl      %ebx, %ecx        movq      %r12, %r8        movq      -8(%rsp), %r15        movq      %r15, %rdi        imulq     %rcx, %r8        imulq     %rcx, %rdi        movq      -16(%rsp), %rdx        movq      %r12, %rbp        imulq     %rbx, %rdx        shrq      $32, %rbx        negq      %rdx        imulq     %rbx, %rbp        imulq     %rbx, %r15        shrq      $32, %r8        addq      %rdi, %r8        addq      %r8, %rbp        movq      %rbp, %rdi        cmpq      %r8, %rbp        movl      $0, %ebp        movl      %ebp, %r8d        setb      %r8b        movq      (%r13), %r14        addq      %r14, %rdx        shrq      $32, %rdi        shlq      $32, %r8        addq      %r8, %rdi        xorl      %r8d, %r8d        cmpq      %r14, %rdx        movq      %r11, %r14        seta      %r8b        addq      %r15, %rdi        movl      %r9d, %r15d        negq      %rdi        addq      %rdx, %rdi        imulq     %r15, %r14        imulq     %r10, %r15        movq      8(%r13), %rsi        addq      $16, %r13        shrq      $32, %r9        subq      %r8, %rsi        movq      %r11, %r8        imulq     %r9, %r8        imulq     %r10, %r9        shrq      $32, %r14        addq      %r15, %r14        addq      %r14, %r8        cmpq      %r14, %r8        movl      %ebp, %r14d        movq      %r8, %r15        setb      %r14b        shrq      $32, %r15        shlq      $32, %r14        addq      %r14, %r15        xorl      %r14d, %r14d        addq      %r9, %r15        xorl      %r9d, %r9d        cmpq      %rdx, %rdi        movq      %r11, %rdx        seta      %r9b        imulq     %rcx, %rdx        imulq     %r10, %rcx        shrq      $32, %rdx        addq      %rcx, %rdx        movq      %r11, %rcx        imulq     %rbx, %rcx        imulq     %r10, %rbx        addq      %rdx, %rcx        subq      %r9, %rsi        movq      %rdi, %r9        cmpq      %rdx, %rcx        movq      %rcx, %r8        setb      %r14b        subq      %r15, %r9        cmpq      %rdi, %r9        seta      %bpl        shrq      $32, %r8        shlq      $32, %r14        addq      %r14, %r8        subq      %rbp, %rsi        addq      %rbx, %r8        movq      %rsi, %rbx        subq      %r8, %rbx        decq      %rax        jns       ..B3.21..B3.22:        movq      -72(%rsp), %r8..B3.23:        xorl      %eax, %eax        movq      %rbx, 8(%r8)        movq      %r9, 16(%r8)        movl      %eax, 4(%r8)        movl      %eax, (%r8)	.cfi_restore 6        popq      %rbp	.cfi_def_cfa_offset 48	.cfi_restore 3        popq      %rbx	.cfi_def_cfa_offset 40	.cfi_restore 15        popq      %r15	.cfi_def_cfa_offset 32	.cfi_restore 14        popq      %r14	.cfi_def_cfa_offset 24	.cfi_restore 13        popq      %r13	.cfi_def_cfa_offset 16	.cfi_restore 12        popq      %r12	.cfi_def_cfa_offset 8        ret       	.cfi_def_cfa_offset 56	.cfi_offset 3, -48	.cfi_offset 6, -56	.cfi_offset 12, -16	.cfi_offset 13, -24	.cfi_offset 14, -32	.cfi_offset 15, -40..B3.24:        movq      -16(%rsp), %r14        movq      -72(%rsp), %r8        jmp       ..B3.19        .align    16,0x90	.cfi_endproc	.type	__eval_neg_poly,@function	.size	__eval_neg_poly,.-__eval_neg_poly	.data# -- End  __eval_neg_poly	.text# -- Begin  __dpml_divide__	.text       .align    16,0x90	.globl __dpml_divide____dpml_divide__:# parameter 1: %rdi# parameter 2: %rsi# parameter 3: %rdx# parameter 4: %rcx..B4.1:	.cfi_startproc..___tag_value___dpml_divide__.136:..L137:        pushq     %r12	.cfi_def_cfa_offset 16	.cfi_offset 12, -16        pushq     %r13	.cfi_def_cfa_offset 24	.cfi_offset 13, -24        pushq     %r14	.cfi_def_cfa_offset 32	.cfi_offset 14, -32        pushq     %r15	.cfi_def_cfa_offset 40	.cfi_offset 15, -40        pushq     %rbx	.cfi_def_cfa_offset 48	.cfi_offset 3, -48        pushq     %rbp	.cfi_def_cfa_offset 56	.cfi_offset 6, -56        subq      $56, %rsp	.cfi_def_cfa_offset 112        movq      %rsi, %rbp        testq     %rbp, %rbp        lea       __ux_one__(%rip), %r8        movq      %rdi, %rbx        movq      %rcx, %r15        jne       ..L151        movq      %r8, %rbp..L151:        testq     %rbx, %rbx        movq      %rdx, %r12        movq      %fs:40, %rax        jne       ..L152        movq      %r8, %rbx..L152:        xorq      %rsp, %rax        movq      %rax, 48(%rsp)        cmpq      %r8, %rbp        movq      8(%rbp), %r13        movq      16(%rbp), %r14        je        ..B4.12..B4.2:        testq     %r13, %r13        jl        ..B4.5..B4.3:        movq      %rbp, %rdi        xorl      %esi, %esi..___tag_value___dpml_divide__.153:        call      __dpml_ffs_and_shift__@PLT..___tag_value___dpml_divide__.154:..B4.4:        movq      8(%rbp), %r13        movq      16(%rbp), %r14..B4.5:        lea       16(%rsp), %rdi..___tag_value___dpml_divide__.155:        call      fegetenv@PLT..___tag_value___dpml_divide__.156:..B4.6:        movq      %r13, %rcx        pxor      %xmm0, %xmm0        shrq      $1, %rcx        movq      $0x3fffffffff, %rax        cvtsi2sdq %rcx, %xmm0        movsd     .L_2il0floatpacket.0(%rip), %xmm3        movq      %r13, %r9        divsd     %xmm0, %xmm3        andq      %rax, %r9        movq      %r14, %r8        shlq      $15, %r9        movq      $0xffffffc000000000, %rdx        shrq      $49, %r8        movq      %r13, %rsi        movl      (%rbx), %r10d        orq       %r8, %r9        movl      4(%rbx), %r8d        andq      %rdx, %rsi        xorl      (%rbp), %r10d        pxor      %xmm1, %xmm1        subl      4(%rbp), %r8d        pxor      %xmm2, %xmm2        cvttsd2si %xmm3, %rbp        cvtsi2sdq %r9, %xmm2        shrq      $1, %rsi        addq      $-1280, %rbp        cvtsi2sdq %rsi, %xmm1        movq      8(%rbx), %rsi        pxor      %xmm7, %xmm7        movq      16(%rbx), %rcx        movq      $0xfffffff000000000, %rbx        andq      %rbx, %rbp        andq      %rsi, %rax        cvtsi2sdq %rbp, %xmm7        movsd     .L_2il0floatpacket.1(%rip), %xmm5        movq      %rcx, %r11        mulsd     %xmm5, %xmm2        pxor      %xmm6, %xmm6        mulsd     %xmm7, %xmm1        mulsd     %xmm7, %xmm2        movsd     .L_2il0floatpacket.0(%rip), %xmm0        movq      %rsi, %rdi        shlq      $15, %rax        andq      %rsi, %rdx        shrq      $49, %r11        pxor      %xmm8, %xmm8        orq       %r11, %rax        pxor      %xmm9, %xmm9        cvtsi2sdq %rax, %xmm6        subsd     %xmm1, %xmm0        mulsd     %xmm6, %xmm5        subsd     %xmm2, %xmm0        movsd     .L_2il0floatpacket.2(%rip), %xmm4        mulsd     %xmm3, %xmm4        shrq      $11, %rdi        shlq      $10, %rdi        shrq      $1, %rdx        cvtsi2sdq %rdi, %xmm8        cvtsi2sdq %rdx, %xmm9        mulsd     %xmm4, %xmm0        movsd     .L_2il0floatpacket.3(%rip), %xmm10        xorl      %edi, %edi        mulsd     %xmm7, %xmm10        mulsd     %xmm0, %xmm8        mulsd     %xmm10, %xmm9        mulsd     %xmm5, %xmm10        mulsd     .L_2il0floatpacket.4(%rip), %xmm9        addsd     %xmm8, %xmm10        cvttsd2si %xmm9, %rdx        cvttsd2si %xmm10, %rbx        movq      %rdx, %rax        lea       (%rbx,%rdx,4), %rdx        cmpq      %rbx, %rdx        movl      %r10d, (%r15)        setb      %dil        xorl      %ebx, %ebx        shrq      $62, %rax        addq      %rdi, %rax        cmpq      $1, %r12        je        ..B4.8..B4.7:        movsd     .L_2il0floatpacket.5(%rip), %xmm1        movq      %r14, %r9        mulsd     %xmm0, %xmm1        xorl      %r11d, %r11d        movq      %r15, (%rsp)        movq      %rax, %r12        cvttsd2si %xmm1, %r15        movl      %edx, %ebx        lea       (%r15,%rbp,4), %r15        testq     %r15, %r15        movq      %rbx, %rdi        movl      %r8d, 8(%rsp)        movq      $-1, %r8        movl      %r14d, %ebp        cmove     %r8, %r15        shrq      $32, %r9        movq      %rdx, %r8        imulq     %rbp, %rdi        movq      %rbx, %r10        negq      %r12        shrq      $32, %r8        andq      %r12, %r14        imulq     %r9, %r10        imulq     %r8, %rbp        imulq     %r8, %r9        shrq      $32, %rdi        andq      %r13, %r12        addq      %r10, %rdi        addq      %rdi, %rbp        cmpq      %rdi, %rbp        movq      %rbp, %r10        movl      %r13d, %ebp        movq      %rbx, %rdi        setb      %r11b        shrq      $32, %r10        shlq      $32, %r11        addq      %r11, %r10        movq      %r13, %r11        addq      %r9, %r10        movq      %r13, %r9        shrq      $32, %r9        addq      %r10, %r14        imulq     %rbp, %rdi        imulq     %r9, %rbx        imulq     %r8, %rbp        imulq     %rdx, %r11        imulq     %r8, %r9        shrq      $32, %rdi        xorl      %r8d, %r8d        addq      %rbx, %rdi        addq      %rdi, %rbp        cmpq      %rdi, %rbp        movq      %rbp, %rbx        movl      $0, %ebp        setb      %bpl        cmpq      %r10, %r14        setb      %r8b        xorl      %r10d, %r10d        addq      %r11, %r14        shrq      $32, %rbx        shlq      $32, %rbp        addq      %rbp, %rbx        xorl      %ebp, %ebp        addq      %r9, %rbx        xorl      %r9d, %r9d        cmpq      %r11, %r14        setb      %r9b        xorl      %r11d, %r11d        addq      %rbx, %r12        cmpq      %rbx, %r12        setb      %r11b        xorl      %ebx, %ebx        addq      %r9, %r8        addq      %r8, %r12        xorl      %r9d, %r9d        cmpq      %r8, %r12        setb      %bl        cmpq      %r14, %rcx        setb      %r10b        addq      %rbx, %r11        xorl      %ebx, %ebx        cmpq      %r12, %rsi        setb      %bl        negq      %r12        addq      %rsi, %r12        negq      %r11        cmpq      %r10, %r12        setb      %bpl        subq      %rbx, %r11        movl      $1, %ebx        cmpq      %r12, %r10        movl      $0, %r12d        cmovne    %ebx, %r12d        subq      %rbp, %r11        orq       %r11, %r12        cmove     %r13, %r11        negq      %r14        addq      %rcx, %r14        xorq      %r11, %r13        subq      %r13, %r14        movl      %r15d, %r13d        movl      %r14d, %edi        movq      %r13, %r8        shrq      $32, %r14        imulq     %rdi, %r8        imulq     %r14, %r13        shrq      $32, %r15        imulq     %r15, %rdi        imulq     %r15, %r14        shrq      $32, %r8        addq      %r13, %r8        addq      %r8, %rdi        cmpq      %r8, %rdi        movq      %rdi, %r13        movl      8(%rsp), %r8d        setb      %r9b        orq       %rcx, %rsi        movl      $0, %ecx        cmove     %ecx, %ebx        xorl      %esi, %esi        shrq      $32, %r13        shlq      $32, %r9        addq      %r9, %r13        addq      %r14, %r13        movq      %r13, %r14        shrq      $63, %r14        addq      %r14, %r12        addq      %r12, %rdx        cmpq      %r12, %rdx        setb      %sil        sarq      $63, %r12        addq      %r12, %rax        movq      (%rsp), %r15        lea       (%rbx,%r13,2), %rbx        addq      %rsi, %rax..B4.8:        movq      %rax, %rsi        movl      %eax, %ecx        movq      %rdx, %rbp        andq      %rax, %rdx        shrq      %cl, %rbp        movl      %eax, %ecx        shlq      $63, %rsi        lea       16(%rsp), %rdi        shlq      $63, %rdx        orq       %rbp, %rsi        shrq      %cl, %rbx        movslq    %r8d, %r8        orq       %rbx, %rdx        addq      %rax, %r8        movq      %rsi, 8(%r15)        movq      %rdx, 16(%r15)        movl      %r8d, 4(%r15)..___tag_value___dpml_divide__.157:        call      fesetenv@PLT..___tag_value___dpml_divide__.158:..B4.9:        movq      48(%rsp), %rax        xorq      %rsp, %rax        cmpq      %fs:40, %rax        jne       ..B4.11..B4.10:        addq      $56, %rsp	.cfi_def_cfa_offset 56	.cfi_restore 6        popq      %rbp	.cfi_def_cfa_offset 48	.cfi_restore 3        popq      %rbx	.cfi_def_cfa_offset 40	.cfi_restore 15        popq      %r15	.cfi_def_cfa_offset 32	.cfi_restore 14        popq      %r14	.cfi_def_cfa_offset 24	.cfi_restore 13        popq      %r13	.cfi_def_cfa_offset 16	.cfi_restore 12        popq      %r12	.cfi_def_cfa_offset 8        ret       	.cfi_def_cfa_offset 112	.cfi_offset 3, -48	.cfi_offset 6, -56	.cfi_offset 12, -16	.cfi_offset 13, -24	.cfi_offset 14, -32	.cfi_offset 15, -40..B4.11:        call      __stack_chk_fail@PLT..B4.12:        movq      8(%rbx), %rcx        movq      %rcx, 8(%r15)        movl      (%rbx), %eax        movl      4(%rbx), %edx        movq      16(%rbx), %rbx        movl      %eax, (%r15)        movl      %edx, 4(%r15)        movq      %rbx, 16(%r15)        movq      48(%rsp), %rbp        xorq      %rsp, %rbp        cmpq      %fs:40, %rbp        jne       ..B4.11        jmp       ..B4.10        .align    16,0x90	.cfi_endproc	.type	__dpml_divide__,@function	.size	__dpml_divide__,.-__dpml_divide__	.data# -- End  __dpml_divide__	.text# -- Begin  __dpml_multiply__	.text       .align    16,0x90	.globl __dpml_multiply____dpml_multiply__:# parameter 1: %rdi# parameter 2: %rsi# parameter 3: %rdx..B5.1:	.cfi_startproc..___tag_value___dpml_multiply__.180:..L181:        pushq     %r13	.cfi_def_cfa_offset 16	.cfi_offset 13, -16        pushq     %r14	.cfi_def_cfa_offset 24	.cfi_offset 14, -24        pushq     %r15	.cfi_def_cfa_offset 32	.cfi_offset 15, -32        movq      %rdi, %r15        movq      8(%rsi), %r9        movq      %rdx, %r11        movl      %r9d, %ecx        movq      %rcx, %r13        movq      %rcx, %rax        movq      16(%r15), %rdi        movl      %edi, %r14d        movq      8(%r15), %r8        movq      %r8, %r10        shrq      $32, %rdi        imulq     %r14, %r13        imulq     %r9, %r10        imulq     %rdi, %rax        shrq      $32, %r9        imulq     %r9, %r14        imulq     %r9, %rdi        shrq      $32, %r13        addq      %rax, %r13        addq      %r13, %r14        cmpq      %r13, %r14        movq      %r14, %rax        movl      $0, %r14d        setb      %r14b        shrq      $32, %rax        shlq      $32, %r14        addq      %r14, %rax        addq      %rdi, %rax        movl      (%r15), %edi        addq      %rax, %r10        movq      16(%rsi), %rdx        xorl      (%rsi), %edi        movl      4(%r15), %r15d        movl      %edx, %r13d        movl      %edi, (%r11)        movq      %r13, %rdi        addl      4(%rsi), %r15d        movl      %r8d, %esi        shrq      $32, %r8        movq      %rsi, %r14        imulq     %rsi, %rdi        imulq     %r8, %r13        shrq      $32, %rdx        imulq     %rdx, %r14        imulq     %r8, %rdx        shrq      $32, %rdi        addq      %r13, %rdi        xorl      %r13d, %r13d        addq      %rdi, %r14        cmpq      %rdi, %r14        movl      %r15d, 4(%r11)        movq      %r14, %r15        setb      %r13b        xorl      %r14d, %r14d        cmpq      %rax, %r10        setb      %r14b        shrq      $32, %r15        shlq      $32, %r13        addq      %r13, %r15        addq      %rdx, %r15        movq      %rcx, %rdx        imulq     %rsi, %rdx        imulq     %r8, %rcx        imulq     %r9, %rsi        imulq     %r9, %r8        shrq      $32, %rdx        addq      %r15, %r10        addq      %rcx, %rdx        xorl      %ecx, %ecx        addq      %rdx, %rsi        cmpq      %rdx, %rsi        movq      %rsi, %r13        setb      %cl        shrq      $32, %r13        shlq      $32, %rcx        addq      %rcx, %r13        addq      %r8, %r13        xorl      %r8d, %r8d        cmpq      %r15, %r10        setb      %r8b        addq      %r14, %r8        addq      %r8, %r13        movq      %r10, 16(%r11)        movq      %r13, 8(%r11)	.cfi_restore 15        popq      %r15	.cfi_def_cfa_offset 24	.cfi_restore 14        popq      %r14	.cfi_def_cfa_offset 16	.cfi_restore 13        popq      %r13	.cfi_def_cfa_offset 8        ret               .align    16,0x90	.cfi_endproc	.type	__dpml_multiply__,@function	.size	__dpml_multiply__,.-__dpml_multiply__	.data# -- End  __dpml_multiply__	.text# -- Begin  __eval_pos_poly	.text       .align    16,0x90__eval_pos_poly:# parameter 1: %rdi# parameter 2: %rsi# parameter 3: %rdx# parameter 4: %rcx# parameter 5: %r8..B6.1:	.cfi_startproc..___tag_value___eval_pos_poly.195:..L196:        pushq     %r12	.cfi_def_cfa_offset 16	.cfi_offset 12, -16        pushq     %r13	.cfi_def_cfa_offset 24	.cfi_offset 13, -24        pushq     %r14	.cfi_def_cfa_offset 32	.cfi_offset 14, -32        pushq     %r15	.cfi_def_cfa_offset 40	.cfi_offset 15, -40        pushq     %rbx	.cfi_def_cfa_offset 48	.cfi_offset 3, -48        pushq     %rbp	.cfi_def_cfa_offset 56	.cfi_offset 6, -56        movq      %rsi, %r11        movq      16(%rdi), %r9        cmpq      $128, %r11        movq      %r9, -32(%rsp)        movl      $0, %r9d        movq      %rcx, %rax        movl      $1, %ecx        movslq    4(%rdi), %rbp        cmovl     %ecx, %r9d        xorl      %ebx, %ebx        xorl      %esi, %esi        movq      %rbp, -64(%rsp)        negq      %rbp        movq      8(%rdi), %r10        testl     %r9d, %r9d        jne       ..B6.5..B6.2:        movq      -64(%rsp), %rdi..B6.3:        addq      %rdi, %r11        addq      $16, %rdx        decq      %rax        cmpq      $128, %r11        jge       ..B6.3..B6.5:        xorl      %edi, %edi        cmpq      $64, %r11        cmovge    %edi, %ecx        testl     %ecx, %ecx        jne       ..B6.12..B6.6:        movq      -64(%rsp), %rdi..B6.7:        movl      %r11d, %ecx        decq      %rax        addq      %rdi, %r11        movq      8(%rdx), %rsi        addq      $16, %rdx        shrq      %cl, %rsi        cmpq      $64, %r11        jl        ..B6.12..B6.8:        testq     %rsi, %rsi        je        ..B6.7..B6.9:        movq      %r10, %r9        xorl      %r12d, %r12d        movl      %r10d, %ebx        shrq      $32, %r9        movq      -64(%rsp), %r13..B6.10:        movl      %esi, %edi        movq      %rbx, %r15        imulq     %rdi, %r15        imulq     %r9, %rdi        shrq      $32, %rsi        movq      %rbx, %r14        imulq     %rsi, %r14        imulq     %r9, %rsi        shrq      $32, %r15        movl      %r12d, %ecx        addq      %rdi, %r15        decq      %rax        addq      %r15, %r14        cmpq      %r15, %r14        movq      %r14, %rdi        setb      %cl        shrq      $32, %rdi        shlq      $32, %rcx        addq      %rcx, %rdi        movl      %r11d, %ecx        addq      %rsi, %rdi        addq      %r13, %r11        movq      8(%rdx), %rsi        addq      $16, %rdx        shrq      %cl, %rsi        addq      %rdi, %rsi        cmpq      $64, %r11        jge       ..B6.10..B6.11:        xorl      %ebx, %ebx        cmpq      %rdi, %rsi        setb      %bl..B6.12:        movq      %r11, %r9        xorl      %edi, %edi        negq      %r9        testq     %r11, %r11        je        ..B6.38..B6.13:        testq     %rbx, %rbx        jne       ..B6.18..B6.14:        movq      %r10, %rbx        xorl      %r13d, %r13d        movq      %r8, -56(%rsp)        movl      %r10d, %r15d        shrq      $32, %rbx        movq      %r10, -40(%rsp)        movq      -64(%rsp), %r8..B6.15:        movl      %esi, %r12d        movq      %r15, %r14        imulq     %r12, %r14        imulq     %rbx, %r12        shrq      $32, %rsi        movq      %r15, %rcx        imulq     %rsi, %rcx        imulq     %rbx, %rsi        shrq      $32, %r14        decq      %rax        addq      %r12, %r14        movl      %r13d, %r12d        addq      %r14, %rcx        cmpq      %r14, %rcx        movq      %rcx, %r10        movl      %r11d, %ecx        setb      %r12b        shrq      $32, %r10        shlq      $32, %r12        addq      %r12, %r10        movq      8(%rdx), %r12        addq      %rsi, %r10        movq      (%rdx), %rsi        movq      %r12, %r14        shrq      %cl, %rsi        movl      %r9d, %ecx        shlq      %cl, %r14        movl      %r11d, %ecx        orq       %r14, %rsi        addq      %rbp, %r9        addq      %r10, %rsi        addq      $16, %rdx        cmpq      %r10, %rsi        movl      %r13d, %r10d        setb      %r10b        shrq      %cl, %r12        addq      %r10, %r12        addq      %r8, %r11        je        ..B6.32..B6.16:        testq     %r12, %r12        je        ..B6.15..B6.17:        movq      -40(%rsp), %r10        movq      %r12, %rbx        movq      -56(%rsp), %r8..B6.18:        testq     %rax, %rax        jge       ..B6.33..B6.19:        movq      %rbx, 8(%r8)        movq      %rsi, 16(%r8)        movl      %edi, 4(%r8)        movl      $0, (%r8)	.cfi_restore 6        popq      %rbp	.cfi_def_cfa_offset 48	.cfi_restore 3        popq      %rbx	.cfi_def_cfa_offset 40	.cfi_restore 15        popq      %r15	.cfi_def_cfa_offset 32	.cfi_restore 14        popq      %r14	.cfi_def_cfa_offset 24	.cfi_restore 13        popq      %r13	.cfi_def_cfa_offset 16	.cfi_restore 12        popq      %r12	.cfi_def_cfa_offset 8        ret       	.cfi_def_cfa_offset 56	.cfi_offset 3, -48	.cfi_offset 6, -56	.cfi_offset 12, -16	.cfi_offset 13, -24	.cfi_offset 14, -32	.cfi_offset 15, -40..B6.38:        movl      %edi, -48(%rsp)        movq      %r10, -40(%rsp)        movq      %r8, -56(%rsp)..B6.21:        testq     %rax, %rax        jl        ..B6.30..B6.22:        movl      %ebx, %edi        decq      %rax        movq      -32(%rsp), %rcx        movq      %rdi, %r15        movl      %ecx, %r13d        movq      %rdi, %r11        movq      -40(%rsp), %rbp        movq      %rbp, %r9        shrq      $32, %rcx        imulq     %r13, %r15        imulq     %rbx, %r9        imulq     %rcx, %r11        shrq      $32, %rbx        movq      %r13, -24(%rsp)        imulq     %rbx, %r13        shrq      $32, %r15        addq      %r11, %r15        xorl      %r11d, %r11d        addq      %r15, %r13        xorl      %r14d, %r14d        movq      %r13, %r8        cmpq      %r15, %r13        movq      %rcx, %r13        setb      %r14b        imulq     %rbx, %r13        shrq      $32, %r8        shlq      $32, %r14        movq      (%rdx), %r12        addq      %r14, %r8        movq      %rax, -16(%rsp)        addq      %r9, %r12        movl      %esi, %eax        addq      %r13, %r8        xorl      %r13d, %r13d        cmpq      %r9, %r12        movq      %rax, %r14        movl      %ebp, %r9d        setb      %r13b        addq      %r8, %r12        shrq      $32, %rbp        imulq     %r9, %r14        imulq     %rbp, %rax        shrq      $32, %r14        shrq      $32, %rsi        addq      %rax, %r14        movq      %r9, %rax        imulq     %rsi, %rax        imulq     %rbp, %rsi        addq      %r14, %rax        cmpq      %r14, %rax        movl      %r11d, %r14d        movq      %rax, %r15        setb      %r14b        shrq      $32, %r15        shlq      $32, %r14        addq      %r14, %r15        xorl      %r14d, %r14d        addq      %rsi, %r15        xorl      %esi, %esi        cmpq      %r8, %r12        movq      %rdi, %r8        setb      %sil        imulq     %r9, %r8        imulq     %rbp, %rdi        shrq      $32, %r8        addq      %rdi, %r8        movq      %r9, %rdi        imulq     %rbx, %rdi        imulq     %rbp, %rbx        addq      %r8, %rdi        cmpq      %r8, %rdi        movq      %rdi, %rax        setb      %r14b        xorl      %edi, %edi        shrq      $32, %rax        addq      %rsi, %r13        shlq      $32, %r14        lea       (%r12,%r15), %rsi        addq      %r14, %rax        addq      %rbx, %rax        xorl      %ebx, %ebx        cmpq      %r15, %rsi        setb      %bl        addq      %r13, %rbx        movq      8(%rdx), %r10        addq      $16, %rdx        addq      %rbx, %r10        cmpq      %rbx, %r10        setb      %dil        lea       (%r10,%rax), %rbx        cmpq      %rax, %rbx        movq      -16(%rsp), %rax        setb      %r11b        addl      %r11d, %edi        je        ..B6.21..B6.23:        movq      %rbx, %rdi        movl      $1, %r11d        shrq      $1, %rsi        shlq      $63, %rdi        shrq      $1, %rbx        orq       %rdi, %rsi        incl      -48(%rsp)        btsq      $63, %rbx        testq     %rax, %rax        js        ..B6.30..B6.39:        movq      %rcx, -72(%rsp)..B6.26:        movl      %r11d, %ecx        decq      %rax        movq      (%rdx), %r15        shrq      %cl, %r15        movq      %r11, %rcx        negq      %rcx        movq      8(%rdx), %r12        movq      %r12, %r13        movq      -72(%rsp), %r10        movq      %r10, %r8        shlq      %cl, %r13        addq      $16, %rdx        movl      %r11d, %ecx        orq       %r13, %r15        shrq      %cl, %r12        movq      -24(%rsp), %rcx        movq      %rcx, %r14        movl      %ebx, %r13d        imulq     %r13, %r14        imulq     %r13, %r8        movq      -40(%rsp), %rdi        imulq     %rbx, %rdi        shrq      $32, %rbx        addq      %rdi, %r15        imulq     %rbx, %rcx        imulq     %rbx, %r10        shrq      $32, %r14        addq      %r8, %r14        addq      %r14, %rcx        cmpq      %r14, %rcx        movl      $0, %r14d        movq      %rcx, %r8        movl      %r14d, %ecx        setb      %cl        shrq      $32, %r8        shlq      $32, %rcx        addq      %rcx, %r8        movq      %rax, -16(%rsp)        addq      %r10, %r8        movl      %esi, %eax        xorl      %r10d, %r10d        cmpq      %rdi, %r15        movq      %r9, %rdi        setb      %r10b        addq      %r8, %r15        imulq     %rax, %rdi        imulq     %rbp, %rax        shrq      $32, %rdi        shrq      $32, %rsi        addq      %rax, %rdi        movq      %r9, %rax        imulq     %rsi, %rax        imulq     %rbp, %rsi        addq      %rdi, %rax        cmpq      %rdi, %rax        movl      %r14d, %edi        movq      %rax, %rcx        setb      %dil        shrq      $32, %rcx        shlq      $32, %rdi        addq      %rdi, %rcx        movq      %r9, %rdi        imulq     %r13, %rdi        imulq     %rbp, %r13        shrq      $32, %rdi        addq      %rsi, %rcx        addq      %r13, %rdi        movq      %r9, %r13        imulq     %rbx, %r13        imulq     %rbp, %rbx        xorl      %esi, %esi        cmpq      %r8, %r15        setb      %sil        xorl      %r8d, %r8d        addq      %rdi, %r13        cmpq      %rdi, %r13        movq      %r13, %rax        setb      %r8b        shrq      $32, %rax        addq      %rsi, %r10        shlq      $32, %r8        lea       (%r15,%rcx), %rsi        addq      %r8, %rax        xorl      %r15d, %r15d        addq      %rbx, %rax        xorl      %ebx, %ebx        cmpq      %rcx, %rsi        setb      %bl        addq      %r10, %rbx        addq      %rbx, %r12        cmpq      %rbx, %r12        setb      %r15b        addq      -64(%rsp), %r11        lea       (%r12,%rax), %rbx        cmpq      %rax, %rbx        movq      -16(%rsp), %rax        setb      %r14b        addl      %r14d, %r15d        je        ..B6.28..B6.27:        movq      %rbx, %rcx        incq      %r11        shrq      $1, %rsi        shlq      $63, %rcx        shrq      $1, %rbx        orq       %rcx, %rsi        incl      -48(%rsp)        btsq      $63, %rbx..B6.28:        testq     %r11, %r11        je        ..B6.21..B6.29:        testq     %rax, %rax        jns       ..B6.26..B6.30:        movl      -48(%rsp), %edi        movq      -56(%rsp), %r8        jmp       ..B6.19..B6.32:        movq      -40(%rsp), %r10        movq      %r12, %rbx        movq      -56(%rsp), %r8        jmp       ..B6.38..B6.33:        movq      -32(%rsp), %rcx        movq      %r10, %rbp        movl      %ecx, %r15d        movl      %r10d, %r9d        shrq      $32, %rbp        shrq      $32, %rcx        movq      %r15, -24(%rsp)        movl      %edi, -48(%rsp)        movq      %r10, -40(%rsp)        movq      %r8, -56(%rsp)        jmp       ..B6.39        .align    16,0x90	.cfi_endproc	.type	__eval_pos_poly,@function	.size	__eval_pos_poly,.-__eval_pos_poly	.data# -- End  __eval_pos_poly	.section .rodata, "a"	.align 8	.align 8.L_2il0floatpacket.0:	.long	0x00000000,0x47b00000	.type	.L_2il0floatpacket.0,@object	.size	.L_2il0floatpacket.0,8	.align 8.L_2il0floatpacket.1:	.long	0x00000000,0x3ef00000	.type	.L_2il0floatpacket.1,@object	.size	.L_2il0floatpacket.1,8	.align 8.L_2il0floatpacket.2:	.long	0x00000000,0x34700000	.type	.L_2il0floatpacket.2,@object	.size	.L_2il0floatpacket.2,8	.align 8.L_2il0floatpacket.3:	.long	0x00000000,0x3c300000	.type	.L_2il0floatpacket.3,@object	.size	.L_2il0floatpacket.3,8	.align 8.L_2il0floatpacket.4:	.long	0x00000000,0x3fd00000	.type	.L_2il0floatpacket.4,@object	.size	.L_2il0floatpacket.4,8	.align 8.L_2il0floatpacket.5:	.long	0x00000000,0x43d00000	.type	.L_2il0floatpacket.5,@object	.size	.L_2il0floatpacket.5,8	.align 8__ux_one__:	.long	0	.long	1	.long	0x00000000,0x80000000	.long	0x00000000,0x00000000	.type	__ux_one__,@object	.size	__ux_one__,24	.data	.section .note.GNU-stack, ""// -- Begin DWARF2 SEGMENT .eh_frame	.section .eh_frame,"a",@progbits.eh_frame_seg:	.align 1# End
 |