9 #if V8_TARGET_ARCH_MIPS
36 patcher.masm()->li(v8::internal::t9, Operand(
reinterpret_cast<int32_t>(
37 debug_info_->GetIsolate()->builtins()->Return_DebugBreak()->entry())));
38 patcher.masm()->Call(v8::internal::t9);
39 patcher.masm()->nop();
40 patcher.masm()->nop();
41 patcher.masm()->nop();
59 return rinfo->IsPatchedReturnSequence();
66 return rinfo()->IsPatchedDebugBreakSlotSequence();
81 patcher.masm()->li(v8::internal::t9, Operand(
reinterpret_cast<int32_t>(
82 debug_info_->GetIsolate()->builtins()->Slot_DebugBreak()->entry())));
83 patcher.masm()->Call(v8::internal::t9);
94 #define __ ACCESS_MASM(masm)
98 static void Generate_DebugBreakCallHelper(MacroAssembler* masm,
119 DCHECK((object_regs & non_object_regs) == 0);
120 if ((object_regs | non_object_regs) != 0) {
123 Register reg = { r };
124 if ((non_object_regs & (1 << r)) != 0) {
125 if (FLAG_debug_code) {
126 __ And(at, reg, 0xc0000000);
127 __ Assert(
eq, kUnableToEncodeValueAsSmi, at, Operand(zero_reg));
132 __ MultiPush(object_regs | non_object_regs);
136 __ RecordComment(
"// Calling from debug break to runtime - come in - over");
138 __ PrepareCEntryArgs(0);
139 __ PrepareCEntryFunction(ExternalReference::debug_break(masm->isolate()));
141 CEntryStub ceb(masm->isolate(), 1);
145 if ((object_regs | non_object_regs) != 0) {
146 __ MultiPop(object_regs | non_object_regs);
149 Register reg = { r };
150 if ((non_object_regs & (1 << r)) != 0) {
153 if (FLAG_debug_code &&
154 (((object_regs |non_object_regs) & (1 << r)) == 0)) {
169 ExternalReference after_break_target =
170 ExternalReference::debug_after_break_target_address(masm->isolate());
171 __ li(t9, Operand(after_break_target));
183 Generate_DebugBreakCallHelper(masm, a1.bit() | a3.bit(), 0);
190 Generate_DebugBreakCallHelper(masm, receiver.bit() |
name.bit(), 0);
199 Generate_DebugBreakCallHelper(
200 masm, receiver.bit() |
name.bit() | value.bit(), 0);
215 Generate_DebugBreakCallHelper(
216 masm, receiver.bit() |
name.bit() | value.bit(), 0);
225 Generate_DebugBreakCallHelper(masm, a0.bit(), 0);
233 Generate_DebugBreakCallHelper(masm, v0.bit(), 0);
242 Generate_DebugBreakCallHelper(masm, a1.bit(), 0);
252 Generate_DebugBreakCallHelper(masm, a1.bit() , a0.bit());
257 MacroAssembler* masm) {
265 Generate_DebugBreakCallHelper(masm, a1.bit() | a2.bit() | a3.bit(), a0.bit());
273 Label check_codesize;
274 __ bind(&check_codesize);
275 __ RecordDebugBreakSlot();
280 masm->InstructionsGeneratedSince(&check_codesize));
287 Generate_DebugBreakCallHelper(masm, 0, 0);
297 ExternalReference restarter_frame_function_slot =
298 ExternalReference::debug_restarter_frame_function_pointer_address(
300 __ li(at, Operand(restarter_frame_function_slot));
friend class BlockTrampolinePoolScope
static const int kDebugBreakSlotInstructions
static const int kJSReturnSequenceInstructions
Handle< DebugInfo > debug_info_
bool IsDebugBreakAtSlot()
void SetDebugBreakAtSlot()
void ClearDebugBreakAtSlot()
bool IsDebugBreakAtReturn()
void ClearDebugBreakAtReturn()
RelocInfo * original_rinfo()
void SetDebugBreakAtReturn()
static const int kHeaderSize
static void GenerateSlot(MacroAssembler *masm)
static void GenerateStoreICDebugBreak(MacroAssembler *masm)
static void GenerateLoadICDebugBreak(MacroAssembler *masm)
static void GeneratePlainReturnLiveEdit(MacroAssembler *masm)
static void GenerateCallConstructStubDebugBreak(MacroAssembler *masm)
static void GenerateCallConstructStubRecordDebugBreak(MacroAssembler *masm)
static void GenerateCallFunctionStubDebugBreak(MacroAssembler *masm)
static void GenerateCompareNilICDebugBreak(MacroAssembler *masm)
static void GenerateKeyedStoreICDebugBreak(MacroAssembler *masm)
static void GenerateCallICStubDebugBreak(MacroAssembler *masm)
static void GenerateFrameDropperLiveEdit(MacroAssembler *masm)
static void GenerateKeyedLoadICDebugBreak(MacroAssembler *masm)
static void GenerateReturnDebugBreak(MacroAssembler *masm)
static void GenerateSlotDebugBreak(MacroAssembler *masm)
static bool IsDebugBreakAtReturn(RelocInfo *rinfo)
static const int kSharedFunctionInfoOffset
static const int kContextOffset
static const int kFramePaddingValue
static const bool kFrameDropperSupported
static const int kFramePaddingInitialSize
static const Register ReceiverRegister()
static const Register NameRegister()
static bool IsJSReturn(Mode mode)
void PatchCode(byte *instructions, int instruction_count)
static const int kCodeOffset
static Smi * FromInt(int value)
static const Register ReceiverRegister()
static const Register NameRegister()
static const Register ValueRegister()
enable harmony numeric enable harmony object literal extensions Optimize object Array DOM strings and string trace pretenuring decisions of HAllocate instructions Enables optimizations which favor memory size over execution speed maximum source size in bytes considered for a single inlining maximum cumulative number of AST nodes considered for inlining trace the tracking of allocation sites deoptimize every n garbage collections perform array bounds checks elimination analyze liveness of environment slots and zap dead values flushes the cache of optimized code for closures on every GC allow uint32 values on optimize frames if they are used only in safe operations track concurrent recompilation artificial compilation delay in ms do not emit check maps for constant values that have a leaf deoptimize the optimized code if the layout of the maps changes enable context specialization in TurboFan execution budget before interrupt is triggered max percentage of megamorphic generic ICs to allow optimization enable use of SAHF instruction if enable use of VFP3 instructions if available enable use of NEON instructions if enable use of SDIV and UDIV instructions if enable use of MLS instructions if enable loading bit constant by means of movw movt instruction enable unaligned accesses for enable use of d16 d31 registers on ARM this requires VFP3 force all emitted branches to be in long enable alignment of csp to bytes on platforms which prefer the register to always be expose gc extension under the specified name show built in functions in stack traces use random jit cookie to mask large constants minimum length for automatic enable preparsing CPU profiler sampling interval in microseconds trace out of bounds accesses to external arrays default size of stack region v8 is allowed to maximum length of function source code printed in a stack trace min size of a semi the new space consists of two semi spaces print one trace line following each garbage collection do not print trace line after scavenger collection print cumulative GC statistics in name
#define DCHECK(condition)
#define DCHECK_EQ(v1, v2)
const uint32_t kDebugZapValue
const RegList kJSCallerSaved
int JSCallerSavedCode(int n)
MemOperand FieldMemOperand(Register object, int offset)
const int kNumJSCallerSaved
Debugger support for the V8 JavaScript engine.