diff options
Diffstat (limited to 'deps/v8/test/cctest/test-macro-assembler-arm.cc')
-rw-r--r-- | deps/v8/test/cctest/test-macro-assembler-arm.cc | 356 |
1 files changed, 356 insertions, 0 deletions
diff --git a/deps/v8/test/cctest/test-macro-assembler-arm.cc b/deps/v8/test/cctest/test-macro-assembler-arm.cc index 06efc58cfa..63919a46cc 100644 --- a/deps/v8/test/cctest/test-macro-assembler-arm.cc +++ b/deps/v8/test/cctest/test-macro-assembler-arm.cc @@ -42,6 +42,7 @@ typedef void* (*F)(int x, int y, int p2, int p3, int p4); #define __ masm-> +typedef Object* (*F3)(void* p0, int p1, int p2, int p3, int p4); typedef int (*F5)(void*, void*, void*, void*, void*); @@ -134,4 +135,359 @@ TEST(LoadAndStoreWithRepresentation) { CHECK(!CALL_GENERATED_CODE(isolate, f, 0, 0, 0, 0, 0)); } +TEST(ExtractLane) { + if (!CpuFeatures::IsSupported(NEON)) return; + + // Allocate an executable page of memory. + size_t actual_size; + byte* buffer = static_cast<byte*>(v8::base::OS::Allocate( + Assembler::kMinimalBufferSize, &actual_size, true)); + CHECK(buffer); + Isolate* isolate = CcTest::i_isolate(); + HandleScope handles(isolate); + MacroAssembler assembler(isolate, buffer, static_cast<int>(actual_size), + v8::internal::CodeObjectRequired::kYes); + MacroAssembler* masm = &assembler; // Create a pointer for the __ macro. + + typedef struct { + int32_t i32x4_low[4]; + int32_t i32x4_high[4]; + int32_t i16x8_low[8]; + int32_t i16x8_high[8]; + int32_t i8x16_low[16]; + int32_t i8x16_high[16]; + int32_t f32x4_low[4]; + int32_t f32x4_high[4]; + } T; + T t; + + __ stm(db_w, sp, r4.bit() | r5.bit() | lr.bit()); + + for (int i = 0; i < 4; i++) { + __ mov(r4, Operand(i)); + __ vdup(Neon32, q1, r4); + __ ExtractLane(r5, q1, NeonS32, i); + __ str(r5, MemOperand(r0, offsetof(T, i32x4_low) + 4 * i)); + SwVfpRegister si = SwVfpRegister::from_code(i); + __ ExtractLane(si, q1, r4, i); + __ vstr(si, r0, offsetof(T, f32x4_low) + 4 * i); + } + + for (int i = 0; i < 8; i++) { + __ mov(r4, Operand(i)); + __ vdup(Neon16, q1, r4); + __ ExtractLane(r5, q1, NeonS16, i); + __ str(r5, MemOperand(r0, offsetof(T, i16x8_low) + 4 * i)); + } + + for (int i = 0; i < 16; i++) { + __ mov(r4, Operand(i)); + __ vdup(Neon8, q1, r4); + __ ExtractLane(r5, q1, NeonS8, i); + __ str(r5, MemOperand(r0, offsetof(T, i8x16_low) + 4 * i)); + } + + if (CpuFeatures::IsSupported(VFP32DREGS)) { + for (int i = 0; i < 4; i++) { + __ mov(r4, Operand(-i)); + __ vdup(Neon32, q15, r4); + __ ExtractLane(r5, q15, NeonS32, i); + __ str(r5, MemOperand(r0, offsetof(T, i32x4_high) + 4 * i)); + SwVfpRegister si = SwVfpRegister::from_code(i); + __ ExtractLane(si, q15, r4, i); + __ vstr(si, r0, offsetof(T, f32x4_high) + 4 * i); + } + + for (int i = 0; i < 8; i++) { + __ mov(r4, Operand(-i)); + __ vdup(Neon16, q15, r4); + __ ExtractLane(r5, q15, NeonS16, i); + __ str(r5, MemOperand(r0, offsetof(T, i16x8_high) + 4 * i)); + } + + for (int i = 0; i < 16; i++) { + __ mov(r4, Operand(-i)); + __ vdup(Neon8, q15, r4); + __ ExtractLane(r5, q15, NeonS8, i); + __ str(r5, MemOperand(r0, offsetof(T, i8x16_high) + 4 * i)); + } + } + + __ ldm(ia_w, sp, r4.bit() | r5.bit() | pc.bit()); + + CodeDesc desc; + masm->GetCode(&desc); + Handle<Code> code = isolate->factory()->NewCode( + desc, Code::ComputeFlags(Code::STUB), Handle<Code>()); +#ifdef DEBUG + OFStream os(stdout); + code->Print(os); +#endif + F3 f = FUNCTION_CAST<F3>(code->entry()); + Object* dummy = CALL_GENERATED_CODE(isolate, f, &t, 0, 0, 0, 0); + USE(dummy); + for (int i = 0; i < 4; i++) { + CHECK_EQ(i, t.i32x4_low[i]); + CHECK_EQ(i, t.f32x4_low[i]); + } + for (int i = 0; i < 8; i++) { + CHECK_EQ(i, t.i16x8_low[i]); + } + for (int i = 0; i < 16; i++) { + CHECK_EQ(i, t.i8x16_low[i]); + } + if (CpuFeatures::IsSupported(VFP32DREGS)) { + for (int i = 0; i < 4; i++) { + CHECK_EQ(-i, t.i32x4_high[i]); + CHECK_EQ(-i, t.f32x4_high[i]); + } + for (int i = 0; i < 8; i++) { + CHECK_EQ(-i, t.i16x8_high[i]); + } + for (int i = 0; i < 16; i++) { + CHECK_EQ(-i, t.i8x16_high[i]); + } + } +} + +TEST(ReplaceLane) { + if (!CpuFeatures::IsSupported(NEON)) return; + + // Allocate an executable page of memory. + size_t actual_size; + byte* buffer = static_cast<byte*>(v8::base::OS::Allocate( + Assembler::kMinimalBufferSize, &actual_size, true)); + CHECK(buffer); + Isolate* isolate = CcTest::i_isolate(); + HandleScope handles(isolate); + MacroAssembler assembler(isolate, buffer, static_cast<int>(actual_size), + v8::internal::CodeObjectRequired::kYes); + MacroAssembler* masm = &assembler; // Create a pointer for the __ macro. + + typedef struct { + int32_t i32x4_low[4]; + int32_t i32x4_high[4]; + int16_t i16x8_low[8]; + int16_t i16x8_high[8]; + int8_t i8x16_low[16]; + int8_t i8x16_high[16]; + int32_t f32x4_low[4]; + int32_t f32x4_high[4]; + } T; + T t; + + __ stm(db_w, sp, r4.bit() | r5.bit() | r6.bit() | r7.bit() | lr.bit()); + + const Register kScratch = r5; + + __ veor(q0, q0, q0); // Zero + __ veor(q1, q1, q1); // Zero + for (int i = 0; i < 4; i++) { + __ mov(r4, Operand(i)); + __ ReplaceLane(q0, q0, r4, NeonS32, i); + SwVfpRegister si = SwVfpRegister::from_code(i); + __ vmov(si, r4); + __ ReplaceLane(q1, q1, si, kScratch, i); + } + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i32x4_low)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, f32x4_low)))); + __ vst1(Neon8, NeonListOperand(q1), NeonMemOperand(r4)); + + __ veor(q0, q0, q0); // Zero + for (int i = 0; i < 8; i++) { + __ mov(r4, Operand(i)); + __ ReplaceLane(q0, q0, r4, NeonS16, i); + } + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i16x8_low)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ veor(q0, q0, q0); // Zero + for (int i = 0; i < 16; i++) { + __ mov(r4, Operand(i)); + __ ReplaceLane(q0, q0, r4, NeonS8, i); + } + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i8x16_low)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + if (CpuFeatures::IsSupported(VFP32DREGS)) { + __ veor(q14, q14, q14); // Zero + __ veor(q15, q15, q15); // Zero + for (int i = 0; i < 4; i++) { + __ mov(r4, Operand(-i)); + __ ReplaceLane(q14, q14, r4, NeonS32, i); + SwVfpRegister si = SwVfpRegister::from_code(i); + __ vmov(si, r4); + __ ReplaceLane(q15, q15, si, kScratch, i); + } + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i32x4_high)))); + __ vst1(Neon8, NeonListOperand(q14), NeonMemOperand(r4)); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, f32x4_high)))); + __ vst1(Neon8, NeonListOperand(q15), NeonMemOperand(r4)); + + __ veor(q14, q14, q14); // Zero + for (int i = 0; i < 8; i++) { + __ mov(r4, Operand(-i)); + __ ReplaceLane(q14, q14, r4, NeonS16, i); + } + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i16x8_high)))); + __ vst1(Neon8, NeonListOperand(q14), NeonMemOperand(r4)); + + __ veor(q14, q14, q14); // Zero + for (int i = 0; i < 16; i++) { + __ mov(r4, Operand(-i)); + __ ReplaceLane(q14, q14, r4, NeonS8, i); + } + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, i8x16_high)))); + __ vst1(Neon8, NeonListOperand(q14), NeonMemOperand(r4)); + } + + __ ldm(ia_w, sp, r4.bit() | r5.bit() | r6.bit() | r7.bit() | pc.bit()); + + CodeDesc desc; + masm->GetCode(&desc); + Handle<Code> code = isolate->factory()->NewCode( + desc, Code::ComputeFlags(Code::STUB), Handle<Code>()); +#ifdef DEBUG + OFStream os(stdout); + code->Print(os); +#endif + F3 f = FUNCTION_CAST<F3>(code->entry()); + Object* dummy = CALL_GENERATED_CODE(isolate, f, &t, 0, 0, 0, 0); + USE(dummy); + for (int i = 0; i < 4; i++) { + CHECK_EQ(i, t.i32x4_low[i]); + CHECK_EQ(i, t.f32x4_low[i]); + } + for (int i = 0; i < 8; i++) { + CHECK_EQ(i, t.i16x8_low[i]); + } + for (int i = 0; i < 16; i++) { + CHECK_EQ(i, t.i8x16_low[i]); + } + if (CpuFeatures::IsSupported(VFP32DREGS)) { + for (int i = 0; i < 4; i++) { + CHECK_EQ(-i, t.i32x4_high[i]); + CHECK_EQ(-i, t.f32x4_high[i]); + } + for (int i = 0; i < 8; i++) { + CHECK_EQ(-i, t.i16x8_high[i]); + } + for (int i = 0; i < 16; i++) { + CHECK_EQ(-i, t.i8x16_high[i]); + } + } +} + +#define CHECK_EQ_32X4(field, v0, v1, v2, v3) \ + CHECK_EQ(v0, t.field[0]); \ + CHECK_EQ(v1, t.field[1]); \ + CHECK_EQ(v2, t.field[2]); \ + CHECK_EQ(v3, t.field[3]); + +TEST(Swizzle) { + if (!CpuFeatures::IsSupported(NEON)) return; + + // Allocate an executable page of memory. + size_t actual_size; + byte* buffer = static_cast<byte*>(v8::base::OS::Allocate( + Assembler::kMinimalBufferSize, &actual_size, true)); + CHECK(buffer); + Isolate* isolate = CcTest::i_isolate(); + HandleScope handles(isolate); + MacroAssembler assembler(isolate, buffer, static_cast<int>(actual_size), + v8::internal::CodeObjectRequired::kYes); + MacroAssembler* masm = &assembler; // Create a pointer for the __ macro. + + typedef struct { + int32_t _32x4_3210[4]; // identity + int32_t _32x4_1032[4]; // high / low swap + int32_t _32x4_0000[4]; // vdup's + int32_t _32x4_1111[4]; + int32_t _32x4_2222[4]; + int32_t _32x4_3333[4]; + int32_t _32x4_2103[4]; // rotate left + int32_t _32x4_0321[4]; // rotate right + int32_t _32x4_1132[4]; // irregular + int32_t _32x4_1132_in_place[4]; // irregular, in-place + } T; + T t; + + __ stm(db_w, sp, r4.bit() | r5.bit() | r6.bit() | r7.bit() | lr.bit()); + + const Register kScratch = r5; + + // Make test vector [0, 1, 2, 3] + __ veor(q1, q1, q1); // Zero + for (int i = 0; i < 4; i++) { + __ mov(r4, Operand(i)); + __ ReplaceLane(q1, q1, r4, NeonS32, i); + } + __ Swizzle(q0, q1, kScratch, Neon32, 0x3210); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_3210)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x1032); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_1032)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x0000); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_0000)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x1111); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_1111)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x2222); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_2222)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x3333); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_3333)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x2103); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_2103)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x0321); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_0321)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ Swizzle(q0, q1, kScratch, Neon32, 0x1132); + __ add(r4, r0, Operand(static_cast<int32_t>(offsetof(T, _32x4_1132)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ vmov(q0, q1); + __ Swizzle(q0, q0, kScratch, Neon32, 0x1132); + __ add(r4, r0, + Operand(static_cast<int32_t>(offsetof(T, _32x4_1132_in_place)))); + __ vst1(Neon8, NeonListOperand(q0), NeonMemOperand(r4)); + + __ ldm(ia_w, sp, r4.bit() | r5.bit() | r6.bit() | r7.bit() | pc.bit()); + + CodeDesc desc; + masm->GetCode(&desc); + Handle<Code> code = isolate->factory()->NewCode( + desc, Code::ComputeFlags(Code::STUB), Handle<Code>()); +#ifdef DEBUG + OFStream os(stdout); + code->Print(os); +#endif + F3 f = FUNCTION_CAST<F3>(code->entry()); + Object* dummy = CALL_GENERATED_CODE(isolate, f, &t, 0, 0, 0, 0); + USE(dummy); + CHECK_EQ_32X4(_32x4_3210, 0, 1, 2, 3); + CHECK_EQ_32X4(_32x4_1032, 2, 3, 0, 1); + CHECK_EQ_32X4(_32x4_0000, 0, 0, 0, 0); + CHECK_EQ_32X4(_32x4_1111, 1, 1, 1, 1); + CHECK_EQ_32X4(_32x4_2222, 2, 2, 2, 2); + CHECK_EQ_32X4(_32x4_3333, 3, 3, 3, 3); + CHECK_EQ_32X4(_32x4_2103, 3, 0, 1, 2); + CHECK_EQ_32X4(_32x4_0321, 1, 2, 3, 0); + CHECK_EQ_32X4(_32x4_1132, 2, 3, 1, 1); + CHECK_EQ_32X4(_32x4_1132_in_place, 2, 3, 1, 1); +} + #undef __ |