/aosp12/art/compiler/utils/arm/ |
H A D | assembler_arm_vixl.cc | 384 void ArmVIXLAssembler::StoreRegisterList(RegList regs, size_t stack_offset) { in StoreRegisterList() argument 390 if (stack_offset != 0) { in StoreRegisterList() 393 ___ Add(base, sp, Operand::From(stack_offset)); in StoreRegisterList() 398 ___ Str(vixl32::Register(i), MemOperand(sp, stack_offset)); in StoreRegisterList() 399 stack_offset += kRegSizeInBytes; in StoreRegisterList() 405 void ArmVIXLAssembler::LoadRegisterList(RegList regs, size_t stack_offset) { in LoadRegisterList() argument 411 if (stack_offset != 0) { in LoadRegisterList() 413 ___ Add(base, sp, Operand::From(stack_offset)); in LoadRegisterList() 418 ___ Ldr(vixl32::Register(i), MemOperand(sp, stack_offset)); in LoadRegisterList() 419 stack_offset += kRegSizeInBytes; in LoadRegisterList()
|
H A D | assembler_arm_vixl.h | 233 void LoadRegisterList(RegList regs, size_t stack_offset); 234 void StoreRegisterList(RegList regs, size_t stack_offset);
|
/aosp12/art/compiler/optimizing/ |
H A D | code_generator.cc | 677 locations->SetInAt(i, Location::StackSlot(stack_offset)); in CreateStringBuilderAppendLocations() 681 stack_offset = RoundUp(stack_offset, sizeof(uint64_t)); in CreateStringBuilderAppendLocations() 684 stack_offset += sizeof(uint32_t); in CreateStringBuilderAppendLocations() 692 stack_offset += sizeof(uint32_t); in CreateStringBuilderAppendLocations() 1727 locations->SetStackBit(stack_offset / kVRegSize); in SaveLiveRegisters() 1731 saved_core_stack_offsets_[i] = stack_offset; in SaveLiveRegisters() 1732 stack_offset += codegen->SaveCoreRegister(stack_offset, i); in SaveLiveRegisters() 1739 saved_fpu_stack_offsets_[i] = stack_offset; in SaveLiveRegisters() 1740 stack_offset += codegen->SaveFloatingPointRegister(stack_offset, i); in SaveLiveRegisters() 1751 stack_offset += codegen->RestoreCoreRegister(stack_offset, i); in RestoreLiveRegisters() [all …]
|
H A D | code_generator_arm_vixl.cc | 206 if (stack_offset != 0) { in SaveContiguousSRegisterList() 220 return stack_offset; in SaveContiguousSRegisterList() 254 if (stack_offset != 0) { in RestoreContiguousSRegisterList() 268 return stack_offset; in RestoreContiguousSRegisterList() 336 stack_offset += kArmWordSize; in SaveLiveRegisters() 343 orig_offset = stack_offset; in SaveLiveRegisters() 347 stack_offset += kArmWordSize; in SaveLiveRegisters() 350 stack_offset = orig_offset; in SaveLiveRegisters() 356 stack_offset = SaveContiguousSRegisterList(begin, end - 1, codegen, stack_offset); in SaveLiveRegisters() 369 stack_offset += kArmWordSize; in RestoreLiveRegisters() [all …]
|
H A D | code_generator_x86.cc | 5201 __ movl(temp, Address(ESP, stack_offset)); in GetInvokeStaticOrDirectExtraParameter() 6772 int stack_offset = ensure_scratch.IsSpilled() ? kX86WordSize : 0; in MoveMemoryToMemory() local 6776 __ movl(temp_reg, Address(ESP, src + stack_offset)); in MoveMemoryToMemory() 6777 __ movl(Address(ESP, dst + stack_offset), temp_reg); in MoveMemoryToMemory() 6778 stack_offset += kX86WordSize; in MoveMemoryToMemory() 6953 int stack_offset = ensure_scratch.IsSpilled() ? kX86WordSize : 0; in Exchange() local 6955 __ movl(Address(ESP, mem + stack_offset), reg); in Exchange() 6965 __ movl(temp_reg, Address(ESP, mem + stack_offset)); in Exchange32() 6966 __ movss(Address(ESP, mem + stack_offset), reg); in Exchange32() 6988 stack_offset += ensure_scratch2.IsSpilled() ? kX86WordSize : 0; in ExchangeMemory() [all …]
|
H A D | code_generator_arm64.cc | 173 size_t stack_offset = codegen->GetFirstRegisterSlotInSlowPath(); in SaveLiveRegisters() local 178 locations->SetStackBit(stack_offset / kVRegSize); in SaveLiveRegisters() 180 DCHECK_LT(stack_offset, codegen->GetFrameSize() - codegen->FrameEntrySpillSize()); in SaveLiveRegisters() 182 saved_core_stack_offsets_[i] = stack_offset; in SaveLiveRegisters() 183 stack_offset += kXRegSizeInBytes; in SaveLiveRegisters() 189 DCHECK_LT(stack_offset, codegen->GetFrameSize() - codegen->FrameEntrySpillSize()); in SaveLiveRegisters() 191 saved_fpu_stack_offsets_[i] = stack_offset; in SaveLiveRegisters() 192 stack_offset += fp_reg_size; in SaveLiveRegisters() 836 size_t stack_offset = calling_convention.GetStackOffsetOf(stack_index_); in GetNextLocation() local 837 next_location = DataType::Is64BitType(type) ? Location::DoubleStackSlot(stack_offset) in GetNextLocation() [all …]
|
H A D | code_generator_x86_64.cc | 6094 int stack_offset = ensure_scratch.IsSpilled() ? kX86_64WordSize : 0; in ExchangeMemory32() local 6095 __ movl(CpuRegister(TMP), Address(CpuRegister(RSP), mem1 + stack_offset)); in ExchangeMemory32() 6097 Address(CpuRegister(RSP), mem2 + stack_offset)); in ExchangeMemory32() 6098 __ movl(Address(CpuRegister(RSP), mem2 + stack_offset), CpuRegister(TMP)); in ExchangeMemory32() 6099 __ movl(Address(CpuRegister(RSP), mem1 + stack_offset), in ExchangeMemory32() 6107 int stack_offset = ensure_scratch.IsSpilled() ? kX86_64WordSize : 0; in ExchangeMemory64() local 6112 Address(CpuRegister(RSP), mem1 + stack_offset)); in ExchangeMemory64() 6114 Address(CpuRegister(RSP), mem2 + stack_offset)); in ExchangeMemory64() 6115 __ movq(Address(CpuRegister(RSP), mem2 + stack_offset), in ExchangeMemory64() 6117 __ movq(Address(CpuRegister(RSP), mem1 + stack_offset), in ExchangeMemory64() [all …]
|
/aosp12/art/libelffile/dwarf/ |
H A D | expression.h | 69 void WriteOpFbreg(int32_t stack_offset) { in WriteOpFbreg() argument 71 PushSleb128(stack_offset); in WriteOpFbreg()
|
/aosp12/art/compiler/utils/arm64/ |
H A D | assembler_arm64.h | 104 void SaveRestoreZRegisterList(uint32_t vreg_bit_vector, int64_t stack_offset) { in SaveRestoreZRegisterList() argument 110 vixl_masm_.Add(temp, vixl::aarch64::sp, stack_offset); in SaveRestoreZRegisterList()
|
/aosp12/art/runtime/ |
H A D | thread.cc | 4076 const size_t stack_offset = stack_index * kFrameSlotSize; in VisitQuickFramePrecise() local 4077 FindWithType(stack_offset, in VisitQuickFramePrecise()
|