// Copyright 2011 the V8 project authors. All rights reserved. // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following // disclaimer in the documentation and/or other materials provided // with the distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived // from this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. #include #include "src/v8.h" #include "src/code-factory.h" #include "src/debug/debug.h" #include "src/disasm.h" #include "src/disassembler.h" #include "src/frames-inl.h" #include "src/macro-assembler.h" #include "test/cctest/cctest.h" namespace v8 { namespace internal { #define __ assm. static void DummyStaticFunction(Object* result) { } TEST(DisasmIa320) { CcTest::InitializeVM(); Isolate* isolate = CcTest::i_isolate(); HandleScope scope(isolate); v8::internal::byte buffer[8192]; Assembler assm(AssemblerOptions{}, buffer, sizeof buffer); DummyStaticFunction(nullptr); // just bloody use it (DELETE; debugging) // Short immediate instructions __ adc(eax, 12345678); __ add(eax, Immediate(12345678)); __ or_(eax, 12345678); __ sub(eax, Immediate(12345678)); __ xor_(eax, 12345678); __ and_(eax, 12345678); Handle foo = isolate->factory()->NewFixedArray(10, TENURED); __ cmp(eax, foo); // ---- This one caused crash __ mov(ebx, Operand(esp, ecx, times_2, 0)); // [esp+ecx*4] // ---- All instructions that I can think of __ add(edx, ebx); __ add(edx, Operand(12, RelocInfo::NONE)); __ add(edx, Operand(ebx, 0)); __ add(edx, Operand(ebx, 16)); __ add(edx, Operand(ebx, 1999)); __ add(edx, Operand(ebx, -4)); __ add(edx, Operand(ebx, -1999)); __ add(edx, Operand(esp, 0)); __ add(edx, Operand(esp, 16)); __ add(edx, Operand(esp, 1999)); __ add(edx, Operand(esp, -4)); __ add(edx, Operand(esp, -1999)); __ nop(); __ add(esi, Operand(ecx, times_4, 0)); __ add(esi, Operand(ecx, times_4, 24)); __ add(esi, Operand(ecx, times_4, -4)); __ add(esi, Operand(ecx, times_4, -1999)); __ nop(); __ add(edi, Operand(ebp, ecx, times_4, 0)); __ add(edi, Operand(ebp, ecx, times_4, 12)); __ add(edi, Operand(ebp, ecx, times_4, -8)); __ add(edi, Operand(ebp, ecx, times_4, -3999)); __ add(Operand(ebp, ecx, times_4, 12), Immediate(12)); __ bswap(eax); __ nop(); __ add(ebx, Immediate(12)); __ nop(); __ adc(edx, Operand(ebx)); __ adc(ecx, 12); __ adc(ecx, 1000); __ nop(); __ and_(edx, 3); __ and_(edx, Operand(esp, 4)); __ cmp(edx, 3); __ cmp(edx, Operand(esp, 4)); __ cmp(Operand(ebp, ecx, times_4, 0), Immediate(1000)); Handle foo2 = isolate->factory()->NewFixedArray(10, TENURED); __ cmp(ebx, foo2); __ cmpb(ebx, Operand(ebp, ecx, times_2, 0)); __ cmpb(Operand(ebp, ecx, times_2, 0), ebx); __ or_(edx, 3); __ xor_(edx, 3); __ nop(); __ cpuid(); __ movsx_b(edx, ecx); __ movsx_w(edx, ecx); __ movzx_b(edx, ecx); __ movzx_w(edx, ecx); __ nop(); __ imul(edx, ecx); __ shld(edx, ecx, 10); __ shld_cl(edx, ecx); __ shrd(edx, ecx, 10); __ shrd_cl(edx, ecx); __ bts(edx, ecx); __ bts(Operand(ebx, ecx, times_4, 0), ecx); __ nop(); __ pushad(); __ popad(); __ pushfd(); __ popfd(); __ push(Immediate(12)); __ push(Immediate(23456)); __ push(ecx); __ push(esi); __ push(Operand(ebp, JavaScriptFrameConstants::kFunctionOffset)); __ push(Operand(ebx, ecx, times_4, 0)); __ push(Operand(ebx, ecx, times_4, 0)); __ push(Operand(ebx, ecx, times_4, 10000)); __ pop(edx); __ pop(eax); __ pop(Operand(ebx, ecx, times_4, 0)); __ nop(); __ add(edx, Operand(esp, 16)); __ add(edx, ecx); __ mov_b(edx, ecx); __ mov_b(ecx, 6); __ mov_b(Operand(ebx, ecx, times_4, 10000), 6); __ mov_b(Operand(esp, 16), edx); __ mov_w(edx, Operand(esp, 16)); __ mov_w(Operand(esp, 16), edx); __ nop(); __ movsx_w(edx, Operand(esp, 12)); __ movsx_b(edx, Operand(esp, 12)); __ movzx_w(edx, Operand(esp, 12)); __ movzx_b(edx, Operand(esp, 12)); __ nop(); __ mov(edx, 1234567); __ mov(edx, Operand(esp, 12)); __ mov(Operand(ebx, ecx, times_4, 10000), Immediate(12345)); __ mov(Operand(ebx, ecx, times_4, 10000), edx); __ nop(); __ dec_b(edx); __ dec_b(Operand(eax, 10)); __ dec_b(Operand(ebx, ecx, times_4, 10000)); __ dec(edx); __ cdq(); __ nop(); __ idiv(edx); __ idiv(Operand(edx, ecx, times_1, 1)); __ idiv(Operand(esp, 12)); __ div(edx); __ div(Operand(edx, ecx, times_1, 1)); __ div(Operand(esp, 12)); __ mul(edx); __ neg(edx); __ not_(edx); __ test(Operand(ebx, ecx, times_4, 10000), Immediate(123456)); __ imul(edx, Operand(ebx, ecx, times_4, 10000)); __ imul(edx, ecx, 12); __ imul(edx, Operand(edx, eax, times_2, 42), 8); __ imul(edx, ecx, 1000); __ imul(edx, Operand(ebx, ecx, times_4, 1), 9000); __ inc(edx); __ inc(Operand(ebx, ecx, times_4, 10000)); __ push(Operand(ebx, ecx, times_4, 10000)); __ pop(Operand(ebx, ecx, times_4, 10000)); __ call(Operand(ebx, ecx, times_4, 10000)); __ jmp(Operand(ebx, ecx, times_4, 10000)); __ lea(edx, Operand(ebx, ecx, times_4, 10000)); __ or_(edx, 12345); __ or_(edx, Operand(ebx, ecx, times_4, 10000)); __ nop(); __ rcl(edx, 1); __ rcl(edx, 7); __ rcr(edx, 1); __ rcr(edx, 7); __ ror(edx, 1); __ ror(edx, 6); __ ror_cl(edx); __ ror(Operand(ebx, ecx, times_4, 10000), 1); __ ror(Operand(ebx, ecx, times_4, 10000), 6); __ ror_cl(Operand(ebx, ecx, times_4, 10000)); __ sar(edx, 1); __ sar(edx, 6); __ sar_cl(edx); __ sar(Operand(ebx, ecx, times_4, 10000), 1); __ sar(Operand(ebx, ecx, times_4, 10000), 6); __ sar_cl(Operand(ebx, ecx, times_4, 10000)); __ sbb(edx, Operand(ebx, ecx, times_4, 10000)); __ shl(edx, 1); __ shl(edx, 6); __ shl_cl(edx); __ shl(Operand(ebx, ecx, times_4, 10000), 1); __ shl(Operand(ebx, ecx, times_4, 10000), 6); __ shl_cl(Operand(ebx, ecx, times_4, 10000)); __ shrd_cl(Operand(ebx, ecx, times_4, 10000), edx); __ shr(edx, 1); __ shr(edx, 7); __ shr_cl(edx); __ shr(Operand(ebx, ecx, times_4, 10000), 1); __ shr(Operand(ebx, ecx, times_4, 10000), 6); __ shr_cl(Operand(ebx, ecx, times_4, 10000)); // Immediates __ adc(edx, 12345); __ add(ebx, Immediate(12)); __ add(Operand(edx, ecx, times_4, 10000), Immediate(12)); __ and_(ebx, 12345); __ cmp(ebx, 12345); __ cmp(ebx, Immediate(12)); __ cmp(Operand(edx, ecx, times_4, 10000), Immediate(12)); __ cmpb(eax, Immediate(100)); __ or_(ebx, 12345); __ sub(ebx, Immediate(12)); __ sub(Operand(edx, ecx, times_4, 10000), Immediate(12)); __ xor_(ebx, 12345); __ imul(edx, ecx, 12); __ imul(edx, ecx, 1000); __ cld(); __ rep_movs(); __ rep_stos(); __ stos(); __ sub(edx, Operand(ebx, ecx, times_4, 10000)); __ sub(edx, ebx); __ test(edx, Immediate(12345)); __ test(edx, Operand(ebx, ecx, times_8, 10000)); __ test(Operand(esi, edi, times_1, -20000000), Immediate(300000000)); __ test_b(edx, Operand(ecx, ebx, times_2, 1000)); __ test_b(Operand(eax, -20), Immediate(0x9A)); __ nop(); __ xor_(edx, 12345); __ xor_(edx, Operand(ebx, ecx, times_8, 10000)); __ bts(Operand(ebx, ecx, times_8, 10000), edx); __ hlt(); __ int3(); __ ret(0); __ ret(8); // Calls Label L1, L2; __ bind(&L1); __ nop(); __ call(&L1); __ call(&L2); __ nop(); __ bind(&L2); __ call(Operand(ebx, ecx, times_4, 10000)); __ nop(); Handle ic = BUILTIN_CODE(isolate, LoadIC); __ call(ic, RelocInfo::CODE_TARGET); __ nop(); __ call(FUNCTION_ADDR(DummyStaticFunction), RelocInfo::RUNTIME_ENTRY); __ nop(); __ jmp(&L1); __ jmp(Operand(ebx, ecx, times_4, 10000)); __ jmp(ic, RelocInfo::CODE_TARGET); __ nop(); Label Ljcc; __ nop(); // long jumps __ j(overflow, &Ljcc); __ j(no_overflow, &Ljcc); __ j(below, &Ljcc); __ j(above_equal, &Ljcc); __ j(equal, &Ljcc); __ j(not_equal, &Ljcc); __ j(below_equal, &Ljcc); __ j(above, &Ljcc); __ j(sign, &Ljcc); __ j(not_sign, &Ljcc); __ j(parity_even, &Ljcc); __ j(parity_odd, &Ljcc); __ j(less, &Ljcc); __ j(greater_equal, &Ljcc); __ j(less_equal, &Ljcc); __ j(greater, &Ljcc); __ nop(); __ bind(&Ljcc); // short jumps __ j(overflow, &Ljcc); __ j(no_overflow, &Ljcc); __ j(below, &Ljcc); __ j(above_equal, &Ljcc); __ j(equal, &Ljcc); __ j(not_equal, &Ljcc); __ j(below_equal, &Ljcc); __ j(above, &Ljcc); __ j(sign, &Ljcc); __ j(not_sign, &Ljcc); __ j(parity_even, &Ljcc); __ j(parity_odd, &Ljcc); __ j(less, &Ljcc); __ j(greater_equal, &Ljcc); __ j(less_equal, &Ljcc); __ j(greater, &Ljcc); // 0xD9 instructions __ nop(); __ fld(1); __ fld1(); __ fldz(); __ fldpi(); __ fabs(); __ fchs(); __ fprem(); __ fprem1(); __ fincstp(); __ ftst(); __ fxch(3); __ fld_s(Operand(ebx, ecx, times_4, 10000)); __ fstp_s(Operand(ebx, ecx, times_4, 10000)); __ ffree(3); __ fld_d(Operand(ebx, ecx, times_4, 10000)); __ fstp_d(Operand(ebx, ecx, times_4, 10000)); __ nop(); __ fild_s(Operand(ebx, ecx, times_4, 10000)); __ fistp_s(Operand(ebx, ecx, times_4, 10000)); __ fild_d(Operand(ebx, ecx, times_4, 10000)); __ fistp_d(Operand(ebx, ecx, times_4, 10000)); __ fnstsw_ax(); __ nop(); __ fadd(3); __ fsub(3); __ fmul(3); __ fdiv(3); __ faddp(3); __ fsubp(3); __ fmulp(3); __ fdivp(3); __ fcompp(); __ fwait(); __ frndint(); __ fninit(); __ nop(); // SSE instruction { // Move operation __ movaps(xmm0, xmm1); __ movups(xmm0, xmm1); __ movups(xmm0, Operand(edx, 4)); __ movups(Operand(edx, 4), xmm0); __ shufps(xmm0, xmm0, 0x0); __ cvtsd2ss(xmm0, xmm1); __ cvtsd2ss(xmm0, Operand(ebx, ecx, times_4, 10000)); __ movq(xmm0, Operand(edx, 4)); // logic operation __ andps(xmm0, xmm1); __ andps(xmm0, Operand(ebx, ecx, times_4, 10000)); __ orps(xmm0, xmm1); __ orps(xmm0, Operand(ebx, ecx, times_4, 10000)); __ xorps(xmm0, xmm1); __ xorps(xmm0, Operand(ebx, ecx, times_4, 10000)); // Arithmetic operation __ addss(xmm1, xmm0); __ addss(xmm1, Operand(ebx, ecx, times_4, 10000)); __ mulss(xmm1, xmm0); __ mulss(xmm1, Operand(ebx, ecx, times_4, 10000)); __ subss(xmm1, xmm0); __ subss(xmm1, Operand(ebx, ecx, times_4, 10000)); __ divss(xmm1, xmm0); __ divss(xmm1, Operand(ebx, ecx, times_4, 10000)); __ maxss(xmm1, xmm0); __ maxss(xmm1, Operand(ebx, ecx, times_4, 10000)); __ minss(xmm1, xmm0); __ minss(xmm1, Operand(ebx, ecx, times_4, 10000)); __ sqrtss(xmm1, xmm0); __ sqrtss(xmm1, Operand(ebx, ecx, times_4, 10000)); __ addps(xmm1, xmm0); __ addps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ subps(xmm1, xmm0); __ subps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ mulps(xmm1, xmm0); __ mulps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ divps(xmm1, xmm0); __ divps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ minps(xmm1, xmm0); __ minps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ maxps(xmm1, xmm0); __ maxps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ rcpps(xmm1, xmm0); __ rcpps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ rsqrtps(xmm1, xmm0); __ rsqrtps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ cmpeqps(xmm5, xmm1); __ cmpeqps(xmm5, Operand(ebx, ecx, times_4, 10000)); __ cmpltps(xmm5, xmm1); __ cmpltps(xmm5, Operand(ebx, ecx, times_4, 10000)); __ cmpleps(xmm5, xmm1); __ cmpleps(xmm5, Operand(ebx, ecx, times_4, 10000)); __ cmpneqps(xmm5, xmm1); __ cmpneqps(xmm5, Operand(ebx, ecx, times_4, 10000)); __ ucomiss(xmm0, xmm1); __ ucomiss(xmm0, Operand(ebx, ecx, times_4, 10000)); } { __ cvttss2si(edx, Operand(ebx, ecx, times_4, 10000)); __ cvtsi2sd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ cvtss2sd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ cvtss2sd(xmm1, xmm0); __ cvtdq2ps(xmm1, xmm0); __ cvtdq2ps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ cvttps2dq(xmm1, xmm0); __ cvttps2dq(xmm1, Operand(ebx, ecx, times_4, 10000)); __ movsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ movsd(Operand(ebx, ecx, times_4, 10000), xmm1); // 128 bit move instructions. __ movdqa(xmm0, Operand(ebx, ecx, times_4, 10000)); __ movdqa(Operand(ebx, ecx, times_4, 10000), xmm0); __ movdqu(xmm0, Operand(ebx, ecx, times_4, 10000)); __ movdqu(Operand(ebx, ecx, times_4, 10000), xmm0); __ movd(xmm0, edi); __ movd(xmm0, Operand(ebx, ecx, times_4, 10000)); __ movd(eax, xmm1); __ movd(Operand(ebx, ecx, times_4, 10000), xmm1); __ addsd(xmm1, xmm0); __ addsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ mulsd(xmm1, xmm0); __ mulsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ subsd(xmm1, xmm0); __ subsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ divsd(xmm1, xmm0); __ divsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ minsd(xmm1, xmm0); __ minsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ maxsd(xmm1, xmm0); __ maxsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ sqrtsd(xmm1, xmm0); __ sqrtsd(xmm1, Operand(ebx, ecx, times_4, 10000)); __ ucomisd(xmm0, xmm1); __ cmpltsd(xmm0, xmm1); __ andpd(xmm0, xmm1); __ psllw(xmm0, 17); __ pslld(xmm0, 17); __ psrlw(xmm0, 17); __ psrld(xmm0, 17); __ psraw(xmm0, 17); __ psrad(xmm0, 17); __ psllq(xmm0, 17); __ psllq(xmm0, xmm1); __ psrlq(xmm0, 17); __ psrlq(xmm0, xmm1); __ pshufhw(xmm5, xmm1, 5); __ pshufhw(xmm5, Operand(edx, 4), 5); __ pshuflw(xmm5, xmm1, 5); __ pshuflw(xmm5, Operand(edx, 4), 5); __ pshufd(xmm5, xmm1, 5); __ pshufd(xmm5, Operand(edx, 4), 5); __ pinsrw(xmm5, edx, 5); __ pinsrw(xmm5, Operand(edx, 4), 5); #define EMIT_SSE2_INSTR(instruction, notUsed1, notUsed2, notUsed3) \ __ instruction(xmm5, xmm1); \ __ instruction(xmm5, Operand(edx, 4)); SSE2_INSTRUCTION_LIST(EMIT_SSE2_INSTR) #undef EMIT_SSE2_INSTR } // cmov. { __ cmov(overflow, eax, Operand(eax, 0)); __ cmov(no_overflow, eax, Operand(eax, 1)); __ cmov(below, eax, Operand(eax, 2)); __ cmov(above_equal, eax, Operand(eax, 3)); __ cmov(equal, eax, Operand(ebx, 0)); __ cmov(not_equal, eax, Operand(ebx, 1)); __ cmov(below_equal, eax, Operand(ebx, 2)); __ cmov(above, eax, Operand(ebx, 3)); __ cmov(sign, eax, Operand(ecx, 0)); __ cmov(not_sign, eax, Operand(ecx, 1)); __ cmov(parity_even, eax, Operand(ecx, 2)); __ cmov(parity_odd, eax, Operand(ecx, 3)); __ cmov(less, eax, Operand(edx, 0)); __ cmov(greater_equal, eax, Operand(edx, 1)); __ cmov(less_equal, eax, Operand(edx, 2)); __ cmov(greater, eax, Operand(edx, 3)); } { if (CpuFeatures::IsSupported(SSE3)) { CpuFeatureScope scope(&assm, SSE3); __ haddps(xmm1, xmm0); __ haddps(xmm1, Operand(ebx, ecx, times_4, 10000)); } } #define EMIT_SSE34_INSTR(instruction, notUsed1, notUsed2, notUsed3, notUsed4) \ __ instruction(xmm5, xmm1); \ __ instruction(xmm5, Operand(edx, 4)); { if (CpuFeatures::IsSupported(SSSE3)) { CpuFeatureScope scope(&assm, SSSE3); SSSE3_INSTRUCTION_LIST(EMIT_SSE34_INSTR) __ palignr(xmm5, xmm1, 5); __ palignr(xmm5, Operand(edx, 4), 5); } } { if (CpuFeatures::IsSupported(SSE4_1)) { CpuFeatureScope scope(&assm, SSE4_1); __ pblendw(xmm5, xmm1, 5); __ pblendw(xmm5, Operand(edx, 4), 5); __ pextrb(eax, xmm0, 1); __ pextrb(Operand(edx, 4), xmm0, 1); __ pextrw(eax, xmm0, 1); __ pextrw(Operand(edx, 4), xmm0, 1); __ pextrd(eax, xmm0, 1); __ pextrd(Operand(edx, 4), xmm0, 1); __ insertps(xmm1, xmm2, 0); __ insertps(xmm1, Operand(edx, 4), 0); __ pinsrb(xmm1, eax, 0); __ pinsrb(xmm1, Operand(edx, 4), 0); __ pinsrd(xmm1, eax, 0); __ pinsrd(xmm1, Operand(edx, 4), 0); __ extractps(eax, xmm1, 0); SSE4_INSTRUCTION_LIST(EMIT_SSE34_INSTR) SSE4_RM_INSTRUCTION_LIST(EMIT_SSE34_INSTR) } } #undef EMIT_SSE34_INSTR // AVX instruction { if (CpuFeatures::IsSupported(AVX)) { CpuFeatureScope scope(&assm, AVX); __ vaddsd(xmm0, xmm1, xmm2); __ vaddsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmulsd(xmm0, xmm1, xmm2); __ vmulsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vsubsd(xmm0, xmm1, xmm2); __ vsubsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vdivsd(xmm0, xmm1, xmm2); __ vdivsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vminsd(xmm0, xmm1, xmm2); __ vminsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmaxsd(xmm0, xmm1, xmm2); __ vmaxsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vsqrtsd(xmm0, xmm1, xmm2); __ vsqrtsd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vaddss(xmm0, xmm1, xmm2); __ vaddss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmulss(xmm0, xmm1, xmm2); __ vmulss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vsubss(xmm0, xmm1, xmm2); __ vsubss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vdivss(xmm0, xmm1, xmm2); __ vdivss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vminss(xmm0, xmm1, xmm2); __ vminss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmaxss(xmm0, xmm1, xmm2); __ vmaxss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vsqrtss(xmm0, xmm1, xmm2); __ vsqrtss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vandps(xmm0, xmm1, xmm2); __ vandps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vxorps(xmm0, xmm1, xmm2); __ vxorps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vaddps(xmm0, xmm1, xmm2); __ vaddps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmulps(xmm0, xmm1, xmm2); __ vmulps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vsubps(xmm0, xmm1, xmm2); __ vsubps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vminps(xmm0, xmm1, xmm2); __ vminps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vdivps(xmm0, xmm1, xmm2); __ vdivps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmaxps(xmm0, xmm1, xmm2); __ vmaxps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vrcpps(xmm1, xmm0); __ vrcpps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ vrsqrtps(xmm1, xmm0); __ vrsqrtps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmovaps(xmm0, xmm1); __ vshufps(xmm0, xmm1, xmm2, 3); __ vshufps(xmm0, xmm1, Operand(edx, 4), 3); __ vhaddps(xmm0, xmm1, xmm2); __ vhaddps(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vcmpeqps(xmm5, xmm4, xmm1); __ vcmpeqps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000)); __ vcmpltps(xmm5, xmm4, xmm1); __ vcmpltps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000)); __ vcmpleps(xmm5, xmm4, xmm1); __ vcmpleps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000)); __ vcmpneqps(xmm5, xmm4, xmm1); __ vcmpneqps(xmm5, xmm4, Operand(ebx, ecx, times_4, 10000)); __ vandpd(xmm0, xmm1, xmm2); __ vandpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vxorpd(xmm0, xmm1, xmm2); __ vxorpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vaddpd(xmm0, xmm1, xmm2); __ vaddpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmulpd(xmm0, xmm1, xmm2); __ vmulpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vsubpd(xmm0, xmm1, xmm2); __ vsubpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vminpd(xmm0, xmm1, xmm2); __ vminpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vdivpd(xmm0, xmm1, xmm2); __ vdivpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmaxpd(xmm0, xmm1, xmm2); __ vmaxpd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vpsllw(xmm0, xmm7, 21); __ vpslld(xmm0, xmm7, 21); __ vpsrlw(xmm0, xmm7, 21); __ vpsrld(xmm0, xmm7, 21); __ vpsraw(xmm0, xmm7, 21); __ vpsrad(xmm0, xmm7, 21); __ vpshufhw(xmm5, xmm1, 5); __ vpshufhw(xmm5, Operand(edx, 4), 5); __ vpshuflw(xmm5, xmm1, 5); __ vpshuflw(xmm5, Operand(edx, 4), 5); __ vpshufd(xmm5, xmm1, 5); __ vpshufd(xmm5, Operand(edx, 4), 5); __ vpblendw(xmm5, xmm1, xmm0, 5); __ vpblendw(xmm5, xmm1, Operand(edx, 4), 5); __ vpalignr(xmm5, xmm1, xmm0, 5); __ vpalignr(xmm5, xmm1, Operand(edx, 4), 5); __ vpextrb(eax, xmm0, 1); __ vpextrb(Operand(edx, 4), xmm0, 1); __ vpextrw(eax, xmm0, 1); __ vpextrw(Operand(edx, 4), xmm0, 1); __ vpextrd(eax, xmm0, 1); __ vpextrd(Operand(edx, 4), xmm0, 1); __ vinsertps(xmm0, xmm1, xmm2, 0); __ vinsertps(xmm0, xmm1, Operand(edx, 4), 0); __ vpinsrb(xmm0, xmm1, eax, 0); __ vpinsrb(xmm0, xmm1, Operand(edx, 4), 0); __ vpinsrw(xmm0, xmm1, eax, 0); __ vpinsrw(xmm0, xmm1, Operand(edx, 4), 0); __ vpinsrd(xmm0, xmm1, eax, 0); __ vpinsrd(xmm0, xmm1, Operand(edx, 4), 0); __ vcvtdq2ps(xmm1, xmm0); __ vcvtdq2ps(xmm1, Operand(ebx, ecx, times_4, 10000)); __ vcvttps2dq(xmm1, xmm0); __ vcvttps2dq(xmm1, Operand(ebx, ecx, times_4, 10000)); __ vmovdqu(xmm0, Operand(ebx, ecx, times_4, 10000)); __ vmovdqu(Operand(ebx, ecx, times_4, 10000), xmm0); __ vmovd(xmm0, edi); __ vmovd(xmm0, Operand(ebx, ecx, times_4, 10000)); __ vmovd(eax, xmm1); __ vmovd(Operand(ebx, ecx, times_4, 10000), xmm1); #define EMIT_SSE2_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3) \ __ v##instruction(xmm7, xmm5, xmm1); \ __ v##instruction(xmm7, xmm5, Operand(edx, 4)); SSE2_INSTRUCTION_LIST(EMIT_SSE2_AVXINSTR) #undef EMIT_SSE2_AVXINSTR #define EMIT_SSE34_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3, \ notUsed4) \ __ v##instruction(xmm7, xmm5, xmm1); \ __ v##instruction(xmm7, xmm5, Operand(edx, 4)); SSSE3_INSTRUCTION_LIST(EMIT_SSE34_AVXINSTR) SSE4_INSTRUCTION_LIST(EMIT_SSE34_AVXINSTR) #undef EMIT_SSE34_AVXINSTR #define EMIT_SSE4_RM_AVXINSTR(instruction, notUsed1, notUsed2, notUsed3, \ notUsed4) \ __ v##instruction(xmm5, xmm1); \ __ v##instruction(xmm5, Operand(edx, 4)); SSE4_RM_INSTRUCTION_LIST(EMIT_SSE4_RM_AVXINSTR) #undef EMIT_SSE4_RM_AVXINSTR } } // FMA3 instruction { if (CpuFeatures::IsSupported(FMA3)) { CpuFeatureScope scope(&assm, FMA3); __ vfmadd132sd(xmm0, xmm1, xmm2); __ vfmadd132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmadd213sd(xmm0, xmm1, xmm2); __ vfmadd213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmadd231sd(xmm0, xmm1, xmm2); __ vfmadd231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmsub132sd(xmm0, xmm1, xmm2); __ vfmsub132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmsub213sd(xmm0, xmm1, xmm2); __ vfmsub213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmsub231sd(xmm0, xmm1, xmm2); __ vfmsub231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmadd132sd(xmm0, xmm1, xmm2); __ vfnmadd132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmadd213sd(xmm0, xmm1, xmm2); __ vfnmadd213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmadd231sd(xmm0, xmm1, xmm2); __ vfnmadd231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmsub132sd(xmm0, xmm1, xmm2); __ vfnmsub132sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmsub213sd(xmm0, xmm1, xmm2); __ vfnmsub213sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmsub231sd(xmm0, xmm1, xmm2); __ vfnmsub231sd(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmadd132ss(xmm0, xmm1, xmm2); __ vfmadd132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmadd213ss(xmm0, xmm1, xmm2); __ vfmadd213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmadd231ss(xmm0, xmm1, xmm2); __ vfmadd231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmsub132ss(xmm0, xmm1, xmm2); __ vfmsub132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmsub213ss(xmm0, xmm1, xmm2); __ vfmsub213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfmsub231ss(xmm0, xmm1, xmm2); __ vfmsub231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmadd132ss(xmm0, xmm1, xmm2); __ vfnmadd132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmadd213ss(xmm0, xmm1, xmm2); __ vfnmadd213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmadd231ss(xmm0, xmm1, xmm2); __ vfnmadd231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmsub132ss(xmm0, xmm1, xmm2); __ vfnmsub132ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmsub213ss(xmm0, xmm1, xmm2); __ vfnmsub213ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); __ vfnmsub231ss(xmm0, xmm1, xmm2); __ vfnmsub231ss(xmm0, xmm1, Operand(ebx, ecx, times_4, 10000)); } } // BMI1 instructions { if (CpuFeatures::IsSupported(BMI1)) { CpuFeatureScope scope(&assm, BMI1); __ andn(eax, ebx, ecx); __ andn(eax, ebx, Operand(ebx, ecx, times_4, 10000)); __ bextr(eax, ebx, ecx); __ bextr(eax, Operand(ebx, ecx, times_4, 10000), ebx); __ blsi(eax, ebx); __ blsi(eax, Operand(ebx, ecx, times_4, 10000)); __ blsmsk(eax, ebx); __ blsmsk(eax, Operand(ebx, ecx, times_4, 10000)); __ blsr(eax, ebx); __ blsr(eax, Operand(ebx, ecx, times_4, 10000)); __ tzcnt(eax, ebx); __ tzcnt(eax, Operand(ebx, ecx, times_4, 10000)); } } // LZCNT instructions { if (CpuFeatures::IsSupported(LZCNT)) { CpuFeatureScope scope(&assm, LZCNT); __ lzcnt(eax, ebx); __ lzcnt(eax, Operand(ebx, ecx, times_4, 10000)); } } // POPCNT instructions { if (CpuFeatures::IsSupported(POPCNT)) { CpuFeatureScope scope(&assm, POPCNT); __ popcnt(eax, ebx); __ popcnt(eax, Operand(ebx, ecx, times_4, 10000)); } } // BMI2 instructions { if (CpuFeatures::IsSupported(BMI2)) { CpuFeatureScope scope(&assm, BMI2); __ bzhi(eax, ebx, ecx); __ bzhi(eax, Operand(ebx, ecx, times_4, 10000), ebx); __ mulx(eax, ebx, ecx); __ mulx(eax, ebx, Operand(ebx, ecx, times_4, 10000)); __ pdep(eax, ebx, ecx); __ pdep(eax, ebx, Operand(ebx, ecx, times_4, 10000)); __ pext(eax, ebx, ecx); __ pext(eax, ebx, Operand(ebx, ecx, times_4, 10000)); __ sarx(eax, ebx, ecx); __ sarx(eax, Operand(ebx, ecx, times_4, 10000), ebx); __ shlx(eax, ebx, ecx); __ shlx(eax, Operand(ebx, ecx, times_4, 10000), ebx); __ shrx(eax, ebx, ecx); __ shrx(eax, Operand(ebx, ecx, times_4, 10000), ebx); __ rorx(eax, ebx, 31); __ rorx(eax, Operand(ebx, ecx, times_4, 10000), 31); } } // xchg. { __ xchg_b(eax, Operand(eax, 8)); __ xchg_w(eax, Operand(ebx, 8)); __ xchg(eax, eax); __ xchg(eax, ebx); __ xchg(ebx, ebx); __ xchg(ebx, Operand(esp, 12)); } // cmpxchg. { __ cmpxchg_b(Operand(esp, 12), eax); __ cmpxchg_w(Operand(ebx, ecx, times_4, 10000), eax); __ cmpxchg(Operand(ebx, ecx, times_4, 10000), eax); __ cmpxchg(Operand(ebx, ecx, times_4, 10000), eax); __ cmpxchg8b(Operand(ebx, ecx, times_8, 10000)); } // lock prefix. { __ lock(); __ cmpxchg(Operand(esp, 12), ebx); __ lock(); __ xchg_w(eax, Operand(ecx, 8)); } // Nop instructions for (int i = 0; i < 16; i++) { __ Nop(i); } __ pause(); __ ret(0); CodeDesc desc; assm.GetCode(isolate, &desc); Handle code = isolate->factory()->NewCode(desc, Code::STUB, Handle()); USE(code); #ifdef OBJECT_PRINT StdoutStream os; code->Print(os); Address begin = code->raw_instruction_start(); Address end = code->raw_instruction_end(); disasm::Disassembler::Disassemble(stdout, reinterpret_cast(begin), reinterpret_cast(end)); #endif } #undef __ } // namespace internal } // namespace v8