diff options
Diffstat (limited to 'deps/v8/src/mips/code-stubs-mips.cc')
-rw-r--r-- | deps/v8/src/mips/code-stubs-mips.cc | 400 |
1 files changed, 330 insertions, 70 deletions
diff --git a/deps/v8/src/mips/code-stubs-mips.cc b/deps/v8/src/mips/code-stubs-mips.cc index c4fc383552..837b5a495b 100644 --- a/deps/v8/src/mips/code-stubs-mips.cc +++ b/deps/v8/src/mips/code-stubs-mips.cc @@ -12,6 +12,7 @@ #include "src/codegen.h" #include "src/ic/handler-compiler.h" #include "src/ic/ic.h" +#include "src/ic/stub-cache.h" #include "src/isolate.h" #include "src/jsregexp.h" #include "src/regexp-macro-assembler.h" @@ -1121,13 +1122,12 @@ void CEntryStub::Generate(MacroAssembler* masm) { __ LoadRoot(t0, Heap::kExceptionRootIndex); __ Branch(&exception_returned, eq, t0, Operand(v0)); - ExternalReference pending_exception_address( - Isolate::kPendingExceptionAddress, isolate()); - // Check that there is no pending exception, otherwise we // should have returned the exception sentinel. if (FLAG_debug_code) { Label okay; + ExternalReference pending_exception_address( + Isolate::kPendingExceptionAddress, isolate()); __ li(a2, Operand(pending_exception_address)); __ lw(a2, MemOperand(a2)); __ LoadRoot(t0, Heap::kTheHoleValueRootIndex); @@ -1147,25 +1147,52 @@ void CEntryStub::Generate(MacroAssembler* masm) { // Handling of exception. __ bind(&exception_returned); - // Retrieve the pending exception. - __ li(a2, Operand(pending_exception_address)); - __ lw(v0, MemOperand(a2)); - - // Clear the pending exception. - __ li(a3, Operand(isolate()->factory()->the_hole_value())); - __ sw(a3, MemOperand(a2)); - - // Special handling of termination exceptions which are uncatchable - // by javascript code. - Label throw_termination_exception; - __ LoadRoot(t0, Heap::kTerminationExceptionRootIndex); - __ Branch(&throw_termination_exception, eq, v0, Operand(t0)); - - // Handle normal exception. - __ Throw(v0); + ExternalReference pending_handler_context_address( + Isolate::kPendingHandlerContextAddress, isolate()); + ExternalReference pending_handler_code_address( + Isolate::kPendingHandlerCodeAddress, isolate()); + ExternalReference pending_handler_offset_address( + Isolate::kPendingHandlerOffsetAddress, isolate()); + ExternalReference pending_handler_fp_address( + Isolate::kPendingHandlerFPAddress, isolate()); + ExternalReference pending_handler_sp_address( + Isolate::kPendingHandlerSPAddress, isolate()); + + // Ask the runtime for help to determine the handler. This will set v0 to + // contain the current pending exception, don't clobber it. + ExternalReference find_handler(Runtime::kFindExceptionHandler, isolate()); + { + FrameScope scope(masm, StackFrame::MANUAL); + __ PrepareCallCFunction(3, 0, a0); + __ mov(a0, zero_reg); + __ mov(a1, zero_reg); + __ li(a2, Operand(ExternalReference::isolate_address(isolate()))); + __ CallCFunction(find_handler, 3); + } - __ bind(&throw_termination_exception); - __ ThrowUncatchable(v0); + // Retrieve the handler context, SP and FP. + __ li(cp, Operand(pending_handler_context_address)); + __ lw(cp, MemOperand(cp)); + __ li(sp, Operand(pending_handler_sp_address)); + __ lw(sp, MemOperand(sp)); + __ li(fp, Operand(pending_handler_fp_address)); + __ lw(fp, MemOperand(fp)); + + // If the handler is a JS frame, restore the context to the frame. Note that + // the context will be set to (cp == 0) for non-JS frames. + Label zero; + __ Branch(&zero, eq, cp, Operand(zero_reg)); + __ sw(cp, MemOperand(fp, StandardFrameConstants::kContextOffset)); + __ bind(&zero); + + // Compute the handler entry address and jump to it. + __ li(a1, Operand(pending_handler_code_address)); + __ lw(a1, MemOperand(a1)); + __ li(a2, Operand(pending_handler_offset_address)); + __ lw(a2, MemOperand(a2)); + __ Addu(a1, a1, Operand(Code::kHeaderSize - kHeapObjectTag)); + __ Addu(t9, a1, a2); + __ Jump(t9); } @@ -1252,7 +1279,7 @@ void JSEntryStub::Generate(MacroAssembler* masm) { handler_offset_ = handler_entry.pos(); // Caught exception: Store result (exception) in the pending exception // field in the JSEnv and return a failure sentinel. Coming in here the - // fp will be invalid because the PushTryHandler below sets it to 0 to + // fp will be invalid because the PushStackHandler below sets it to 0 to // signal the existence of the JSEntry frame. __ li(t0, Operand(ExternalReference(Isolate::kPendingExceptionAddress, isolate))); @@ -1261,10 +1288,9 @@ void JSEntryStub::Generate(MacroAssembler* masm) { __ b(&exit); // b exposes branch delay slot. __ nop(); // Branch delay slot nop. - // Invoke: Link this frame into the handler chain. There's only one - // handler block in this code object, so its index is 0. + // Invoke: Link this frame into the handler chain. __ bind(&invoke); - __ PushTryHandler(StackHandler::JS_ENTRY, 0); + __ PushStackHandler(); // If an exception not caught by another handler occurs, this handler // returns control to the code after the bal(&invoke) above, which // restores all kCalleeSaved registers (including cp and fp) to their @@ -1309,7 +1335,7 @@ void JSEntryStub::Generate(MacroAssembler* masm) { __ Call(t9); // Unlink this frame from the handler chain. - __ PopTryHandler(); + __ PopStackHandler(); __ bind(&exit); // v0 holds result // Check if the current stack frame is marked as the outermost JS frame. @@ -1352,12 +1378,8 @@ void LoadIndexedStringStub::Generate(MacroAssembler* masm) { Register result = v0; DCHECK(!scratch.is(receiver) && !scratch.is(index)); DCHECK(!FLAG_vector_ics || - (!scratch.is(VectorLoadICDescriptor::VectorRegister()) && - result.is(VectorLoadICDescriptor::SlotRegister()))); + !scratch.is(VectorLoadICDescriptor::VectorRegister())); - // StringCharAtGenerator doesn't use the result register until it's passed - // the different miss possibilities. If it did, we would have a conflict - // when FLAG_vector_ics is true. StringCharAtGenerator char_at_generator(receiver, index, scratch, result, &miss, // When not a string. &miss, // When not a number. @@ -1368,7 +1390,7 @@ void LoadIndexedStringStub::Generate(MacroAssembler* masm) { __ Ret(); StubRuntimeCallHelper call_helper; - char_at_generator.GenerateSlow(masm, call_helper); + char_at_generator.GenerateSlow(masm, PART_OF_IC_HANDLER, call_helper); __ bind(&miss); PropertyAccessCompiler::TailCallBuiltin( @@ -1938,8 +1960,11 @@ void ArgumentsAccessStub::GenerateNewStrict(MacroAssembler* masm) { __ bind(&adaptor_frame); __ lw(a1, MemOperand(a2, ArgumentsAdaptorFrameConstants::kLengthOffset)); if (has_new_target()) { + Label skip_decrement; + __ Branch(&skip_decrement, eq, a1, Operand(Smi::FromInt(0))); // Subtract 1 from smi-tagged arguments count. __ Subu(a1, a1, Operand(2)); + __ bind(&skip_decrement); } __ sw(a1, MemOperand(sp, 0)); __ sll(at, a1, kPointerSizeLog2 - kSmiTagSize); @@ -2054,7 +2079,7 @@ void RegExpExecStub::Generate(MacroAssembler* masm) { // time or if regexp entry in generated code is turned off runtime switch or // at compilation. #ifdef V8_INTERPRETED_REGEXP - __ TailCallRuntime(Runtime::kRegExpExecRT, 4, 1); + __ TailCallRuntime(Runtime::kRegExpExec, 4, 1); #else // V8_INTERPRETED_REGEXP // Stack frame on entry. @@ -2338,17 +2363,8 @@ void RegExpExecStub::Generate(MacroAssembler* masm) { __ lw(v0, MemOperand(a2, 0)); __ Branch(&runtime, eq, v0, Operand(a1)); - __ sw(a1, MemOperand(a2, 0)); // Clear pending exception. - - // Check if the exception is a termination. If so, throw as uncatchable. - __ LoadRoot(a0, Heap::kTerminationExceptionRootIndex); - Label termination_exception; - __ Branch(&termination_exception, eq, v0, Operand(a0)); - - __ Throw(v0); - - __ bind(&termination_exception); - __ ThrowUncatchable(v0); + // For exception, throw the exception again. + __ TailCallRuntime(Runtime::kRegExpExecReThrow, 4, 1); __ bind(&failure); // For failure and exception return null. @@ -2444,7 +2460,7 @@ void RegExpExecStub::Generate(MacroAssembler* masm) { // Do the runtime call to execute the regexp. __ bind(&runtime); - __ TailCallRuntime(Runtime::kRegExpExecRT, 4, 1); + __ TailCallRuntime(Runtime::kRegExpExec, 4, 1); // Deferred code for string handling. // (6) Not a long external string? If yes, go to (8). @@ -3031,7 +3047,7 @@ void StringCharCodeAtGenerator::GenerateFast(MacroAssembler* masm) { void StringCharCodeAtGenerator::GenerateSlow( - MacroAssembler* masm, + MacroAssembler* masm, EmbedMode embed_mode, const RuntimeCallHelper& call_helper) { __ Abort(kUnexpectedFallthroughToCharCodeAtSlowCase); @@ -3045,7 +3061,12 @@ void StringCharCodeAtGenerator::GenerateSlow( DONT_DO_SMI_CHECK); call_helper.BeforeCall(masm); // Consumed by runtime conversion function: - __ Push(object_, index_); + if (FLAG_vector_ics && embed_mode == PART_OF_IC_HANDLER) { + __ Push(VectorLoadICDescriptor::VectorRegister(), + VectorLoadICDescriptor::SlotRegister(), object_, index_); + } else { + __ Push(object_, index_); + } if (index_flags_ == STRING_INDEX_IS_NUMBER) { __ CallRuntime(Runtime::kNumberToIntegerMapMinusZero, 1); } else { @@ -3056,9 +3077,13 @@ void StringCharCodeAtGenerator::GenerateSlow( // Save the conversion result before the pop instructions below // have a chance to overwrite it. - __ Move(index_, v0); - __ pop(object_); + if (FLAG_vector_ics && embed_mode == PART_OF_IC_HANDLER) { + __ Pop(VectorLoadICDescriptor::SlotRegister(), + VectorLoadICDescriptor::VectorRegister(), object_); + } else { + __ pop(object_); + } // Reload the instance type. __ lw(result_, FieldMemOperand(object_, HeapObject::kMapOffset)); __ lbu(result_, FieldMemOperand(result_, Map::kInstanceTypeOffset)); @@ -3385,7 +3410,7 @@ void SubStringStub::Generate(MacroAssembler* masm) { // Just jump to runtime to create the sub string. __ bind(&runtime); - __ TailCallRuntime(Runtime::kSubString, 3, 1); + __ TailCallRuntime(Runtime::kSubStringRT, 3, 1); __ bind(&single_char); // v0: original string @@ -3579,7 +3604,7 @@ void StringCompareStub::Generate(MacroAssembler* masm) { StringHelper::GenerateCompareFlatOneByteStrings(masm, a1, a0, a2, a3, t0, t1); __ bind(&runtime); - __ TailCallRuntime(Runtime::kStringCompare, 2, 1); + __ TailCallRuntime(Runtime::kStringCompareRT, 2, 1); } @@ -3889,7 +3914,7 @@ void CompareICStub::GenerateStrings(MacroAssembler* masm) { if (equality) { __ TailCallRuntime(Runtime::kStringEquals, 2, 1); } else { - __ TailCallRuntime(Runtime::kStringCompare, 2, 1); + __ TailCallRuntime(Runtime::kStringCompareRT, 2, 1); } __ bind(&miss); @@ -4529,15 +4554,15 @@ void StubFailureTrampolineStub::Generate(MacroAssembler* masm) { void LoadICTrampolineStub::Generate(MacroAssembler* masm) { EmitLoadTypeFeedbackVector(masm, VectorLoadICDescriptor::VectorRegister()); - VectorLoadStub stub(isolate(), state()); - __ Jump(stub.GetCode(), RelocInfo::CODE_TARGET); + VectorRawLoadStub stub(isolate(), state()); + stub.GenerateForTrampoline(masm); } void KeyedLoadICTrampolineStub::Generate(MacroAssembler* masm) { EmitLoadTypeFeedbackVector(masm, VectorLoadICDescriptor::VectorRegister()); - VectorKeyedLoadStub stub(isolate()); - __ Jump(stub.GetCode(), RelocInfo::CODE_TARGET); + VectorRawKeyedLoadStub stub(isolate()); + stub.GenerateForTrampoline(masm); } @@ -4555,6 +4580,243 @@ void CallIC_ArrayTrampolineStub::Generate(MacroAssembler* masm) { } +void VectorRawLoadStub::Generate(MacroAssembler* masm) { + GenerateImpl(masm, false); +} + + +void VectorRawLoadStub::GenerateForTrampoline(MacroAssembler* masm) { + GenerateImpl(masm, true); +} + + +static void HandleArrayCases(MacroAssembler* masm, Register receiver, + Register key, Register vector, Register slot, + Register feedback, Register scratch1, + Register scratch2, Register scratch3, + bool is_polymorphic, Label* miss) { + // feedback initially contains the feedback array + Label next_loop, prepare_next; + Label load_smi_map, compare_map; + Label start_polymorphic; + + Register receiver_map = scratch1; + Register cached_map = scratch2; + + // Receiver might not be a heap object. + __ JumpIfSmi(receiver, &load_smi_map); + __ lw(receiver_map, FieldMemOperand(receiver, HeapObject::kMapOffset)); + __ bind(&compare_map); + __ lw(cached_map, + FieldMemOperand(feedback, FixedArray::OffsetOfElementAt(0))); + __ lw(cached_map, FieldMemOperand(cached_map, WeakCell::kValueOffset)); + __ Branch(&start_polymorphic, ne, receiver_map, Operand(cached_map)); + // found, now call handler. + Register handler = feedback; + __ lw(handler, FieldMemOperand(feedback, FixedArray::OffsetOfElementAt(1))); + __ Addu(t9, handler, Operand(Code::kHeaderSize - kHeapObjectTag)); + __ Jump(t9); + + + Register length = scratch3; + __ bind(&start_polymorphic); + __ lw(length, FieldMemOperand(feedback, FixedArray::kLengthOffset)); + if (!is_polymorphic) { + // If the IC could be monomorphic we have to make sure we don't go past the + // end of the feedback array. + __ Branch(miss, eq, length, Operand(Smi::FromInt(2))); + } + + Register too_far = length; + Register pointer_reg = feedback; + + // +-----+------+------+-----+-----+ ... ----+ + // | map | len | wm0 | h0 | wm1 | hN | + // +-----+------+------+-----+-----+ ... ----+ + // 0 1 2 len-1 + // ^ ^ + // | | + // pointer_reg too_far + // aka feedback scratch3 + // also need receiver_map (aka scratch1) + // use cached_map (scratch2) to look in the weak map values. + __ sll(at, length, kPointerSizeLog2 - kSmiTagSize); + __ Addu(too_far, feedback, Operand(at)); + __ Addu(too_far, too_far, Operand(FixedArray::kHeaderSize - kHeapObjectTag)); + __ Addu(pointer_reg, feedback, + Operand(FixedArray::OffsetOfElementAt(2) - kHeapObjectTag)); + + __ bind(&next_loop); + __ lw(cached_map, MemOperand(pointer_reg)); + __ lw(cached_map, FieldMemOperand(cached_map, WeakCell::kValueOffset)); + __ Branch(&prepare_next, ne, receiver_map, Operand(cached_map)); + __ lw(handler, MemOperand(pointer_reg, kPointerSize)); + __ Addu(t9, handler, Operand(Code::kHeaderSize - kHeapObjectTag)); + __ Jump(t9); + + __ bind(&prepare_next); + __ Addu(pointer_reg, pointer_reg, Operand(kPointerSize * 2)); + __ Branch(&next_loop, lt, pointer_reg, Operand(too_far)); + + // We exhausted our array of map handler pairs. + __ jmp(miss); + + __ bind(&load_smi_map); + __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); + __ jmp(&compare_map); +} + + +static void HandleMonomorphicCase(MacroAssembler* masm, Register receiver, + Register key, Register vector, Register slot, + Register weak_cell, Register scratch, + Label* miss) { + // feedback initially contains the feedback array + Label compare_smi_map; + Register receiver_map = scratch; + Register cached_map = weak_cell; + + // Move the weak map into the weak_cell register. + __ lw(cached_map, FieldMemOperand(weak_cell, WeakCell::kValueOffset)); + + // Receiver might not be a heap object. + __ JumpIfSmi(receiver, &compare_smi_map); + __ lw(receiver_map, FieldMemOperand(receiver, HeapObject::kMapOffset)); + __ Branch(miss, ne, cached_map, Operand(receiver_map)); + + Register handler = weak_cell; + __ sll(at, slot, kPointerSizeLog2 - kSmiTagSize); + __ Addu(handler, vector, Operand(at)); + __ lw(handler, + FieldMemOperand(handler, FixedArray::kHeaderSize + kPointerSize)); + __ Addu(t9, handler, Operand(Code::kHeaderSize - kHeapObjectTag)); + __ Jump(t9); + + // In microbenchmarks, it made sense to unroll this code so that the call to + // the handler is duplicated for a HeapObject receiver and a Smi receiver. + __ bind(&compare_smi_map); + __ LoadRoot(at, Heap::kHeapNumberMapRootIndex); + __ Branch(miss, ne, at, Operand(weak_cell)); + __ sll(at, slot, kPointerSizeLog2 - kSmiTagSize); + __ Addu(handler, vector, Operand(at)); + __ lw(handler, + FieldMemOperand(handler, FixedArray::kHeaderSize + kPointerSize)); + __ Addu(t9, handler, Operand(Code::kHeaderSize - kHeapObjectTag)); + __ Jump(t9); +} + + +void VectorRawLoadStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { + Register receiver = VectorLoadICDescriptor::ReceiverRegister(); // a1 + Register name = VectorLoadICDescriptor::NameRegister(); // a2 + Register vector = VectorLoadICDescriptor::VectorRegister(); // a3 + Register slot = VectorLoadICDescriptor::SlotRegister(); // a0 + Register feedback = t0; + Register scratch1 = t1; + + __ sll(at, slot, kPointerSizeLog2 - kSmiTagSize); + __ Addu(feedback, vector, Operand(at)); + __ lw(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); + + // Is it a weak cell? + Label try_array; + Label not_array, smi_key, key_okay, miss; + __ lw(scratch1, FieldMemOperand(feedback, HeapObject::kMapOffset)); + __ LoadRoot(at, Heap::kWeakCellMapRootIndex); + __ Branch(&try_array, ne, at, Operand(scratch1)); + HandleMonomorphicCase(masm, receiver, name, vector, slot, feedback, scratch1, + &miss); + + // Is it a fixed array? + __ bind(&try_array); + __ LoadRoot(at, Heap::kFixedArrayMapRootIndex); + __ Branch(¬_array, ne, at, Operand(scratch1)); + HandleArrayCases(masm, receiver, name, vector, slot, feedback, scratch1, t4, + t5, true, &miss); + + __ bind(¬_array); + __ LoadRoot(at, Heap::kmegamorphic_symbolRootIndex); + __ Branch(&miss, ne, at, Operand(feedback)); + Code::Flags code_flags = Code::RemoveTypeAndHolderFromFlags( + Code::ComputeHandlerFlags(Code::LOAD_IC)); + masm->isolate()->stub_cache()->GenerateProbe(masm, Code::LOAD_IC, code_flags, + false, receiver, name, feedback, + scratch1, t4, t5); + + __ bind(&miss); + LoadIC::GenerateMiss(masm); +} + + +void VectorRawKeyedLoadStub::Generate(MacroAssembler* masm) { + GenerateImpl(masm, false); +} + + +void VectorRawKeyedLoadStub::GenerateForTrampoline(MacroAssembler* masm) { + GenerateImpl(masm, true); +} + + +void VectorRawKeyedLoadStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { + Register receiver = VectorLoadICDescriptor::ReceiverRegister(); // a1 + Register key = VectorLoadICDescriptor::NameRegister(); // a2 + Register vector = VectorLoadICDescriptor::VectorRegister(); // a3 + Register slot = VectorLoadICDescriptor::SlotRegister(); // a0 + Register feedback = t0; + Register scratch1 = t1; + + __ sll(at, slot, kPointerSizeLog2 - kSmiTagSize); + __ Addu(feedback, vector, Operand(at)); + __ lw(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); + + // Is it a weak cell? + Label try_array; + Label not_array, smi_key, key_okay, miss; + __ lw(scratch1, FieldMemOperand(feedback, HeapObject::kMapOffset)); + __ LoadRoot(at, Heap::kWeakCellMapRootIndex); + __ Branch(&try_array, ne, at, Operand(scratch1)); + __ JumpIfNotSmi(key, &miss); + HandleMonomorphicCase(masm, receiver, key, vector, slot, feedback, scratch1, + &miss); + + __ bind(&try_array); + // Is it a fixed array? + __ LoadRoot(at, Heap::kFixedArrayMapRootIndex); + __ Branch(¬_array, ne, at, Operand(scratch1)); + // We have a polymorphic element handler. + __ JumpIfNotSmi(key, &miss); + + Label polymorphic, try_poly_name; + __ bind(&polymorphic); + HandleArrayCases(masm, receiver, key, vector, slot, feedback, scratch1, t4, + t5, true, &miss); + + __ bind(¬_array); + // Is it generic? + __ LoadRoot(at, Heap::kmegamorphic_symbolRootIndex); + __ Branch(&try_poly_name, ne, at, Operand(feedback)); + Handle<Code> megamorphic_stub = + KeyedLoadIC::ChooseMegamorphicStub(masm->isolate()); + __ Jump(megamorphic_stub, RelocInfo::CODE_TARGET); + + __ bind(&try_poly_name); + // We might have a name in feedback, and a fixed array in the next slot. + __ Branch(&miss, ne, key, Operand(feedback)); + // If the name comparison succeeded, we know we have a fixed array with + // at least one map/handler pair. + __ sll(at, slot, kPointerSizeLog2 - kSmiTagSize); + __ Addu(feedback, vector, Operand(at)); + __ lw(feedback, + FieldMemOperand(feedback, FixedArray::kHeaderSize + kPointerSize)); + HandleArrayCases(masm, receiver, key, vector, slot, feedback, scratch1, t4, + t5, false, &miss); + + __ bind(&miss); + KeyedLoadIC::GenerateMiss(masm); +} + + void ProfileEntryHookStub::MaybeCallEntryHook(MacroAssembler* masm) { if (masm->isolate()->function_entry_hook() != NULL) { ProfileEntryHookStub stub(masm->isolate()); @@ -5011,7 +5273,6 @@ static void CallApiFunctionAndReturn( } Label promote_scheduled_exception; - Label exception_handled; Label delete_allocated_handles; Label leave_exit_frame; Label return_value_loaded; @@ -5032,13 +5293,8 @@ static void CallApiFunctionAndReturn( __ lw(at, MemOperand(s3, kLimitOffset)); __ Branch(&delete_allocated_handles, ne, s1, Operand(at)); - // Check if the function scheduled an exception. + // Leave the API exit frame. __ bind(&leave_exit_frame); - __ LoadRoot(t0, Heap::kTheHoleValueRootIndex); - __ li(at, Operand(ExternalReference::scheduled_exception_address(isolate))); - __ lw(t1, MemOperand(at)); - __ Branch(&promote_scheduled_exception, ne, t0, Operand(t1)); - __ bind(&exception_handled); bool restore_context = context_restore_operand != NULL; if (restore_context) { @@ -5051,16 +5307,20 @@ static void CallApiFunctionAndReturn( } else { __ li(s0, Operand(stack_space)); } - __ LeaveExitFrame(false, s0, !restore_context, EMIT_RETURN, + __ LeaveExitFrame(false, s0, !restore_context, NO_EMIT_RETURN, stack_space_offset != kInvalidStackOffset); + // Check if the function scheduled an exception. + __ LoadRoot(t0, Heap::kTheHoleValueRootIndex); + __ li(at, Operand(ExternalReference::scheduled_exception_address(isolate))); + __ lw(t1, MemOperand(at)); + __ Branch(&promote_scheduled_exception, ne, t0, Operand(t1)); + + __ Ret(); + + // Re-throw by promoting a scheduled exception. __ bind(&promote_scheduled_exception); - { - FrameScope frame(masm, StackFrame::INTERNAL); - __ CallExternalReference( - ExternalReference(Runtime::kPromoteScheduledException, isolate), 0); - } - __ jmp(&exception_handled); + __ TailCallRuntime(Runtime::kPromoteScheduledException, 0, 1); // HandleScope limit has changed. Delete allocated extensions. __ bind(&delete_allocated_handles); |