Lines Matching refs:to_ref
186 mirror::Object* to_ref = Mark(self, from_ref); in MarkHeapReference() local
187 if (from_ref != to_ref) { in MarkHeapReference()
193 } while (!field->CasWeakRelaxed(from_ref, to_ref)); in MarkHeapReference()
501 mirror::Object* to_ref = concurrent_copying_->Mark(self, ref); in VisitRoots() local
502 if (to_ref != ref) { in VisitRoots()
503 *root = to_ref; in VisitRoots()
518 mirror::Object* to_ref = concurrent_copying_->Mark(self, ref); in VisitRoots() local
519 if (to_ref != ref) { in VisitRoots()
520 root->Assign(to_ref); in VisitRoots()
1813 void ConcurrentCopying::PushOntoMarkStack(Thread* const self, mirror::Object* to_ref) { in PushOntoMarkStack() argument
1815 << " " << to_ref << " " << mirror::Object::PrettyTypeOf(to_ref); in PushOntoMarkStack()
1825 gc_mark_stack_->PushBack(to_ref); in PushOntoMarkStack()
1845 new_tl_mark_stack->PushBack(to_ref); in PushOntoMarkStack()
1852 tl_mark_stack->PushBack(to_ref); in PushOntoMarkStack()
1861 gc_mark_stack_->PushBack(to_ref); in PushOntoMarkStack()
1865 << "ref=" << to_ref in PushOntoMarkStack()
1875 gc_mark_stack_->PushBack(to_ref); in PushOntoMarkStack()
2127 mirror::Object* to_ref = gc_mark_stack_->PopBack(); in ProcessMarkStackOnce() local
2128 ProcessMarkStackRef(to_ref); in ProcessMarkStackOnce()
2172 mirror::Object* to_ref = gc_mark_stack_->PopBack(); in ProcessMarkStackOnce() local
2173 ProcessMarkStackRef(to_ref); in ProcessMarkStackOnce()
2203 mirror::Object* to_ref = p->AsMirrorPtr(); in ProcessThreadLocalMarkStacks() local
2204 processor(to_ref); in ProcessThreadLocalMarkStacks()
2227 inline void ConcurrentCopying::ProcessMarkStackRef(mirror::Object* to_ref) { in ProcessMarkStackRef() argument
2228 DCHECK(!region_space_->IsInFromSpace(to_ref)); in ProcessMarkStackRef()
2230 space::RegionSpace::RegionType rtype = region_space_->GetRegionType(to_ref); in ProcessMarkStackRef()
2232 DCHECK(to_ref->GetReadBarrierState() == ReadBarrier::GrayState()) in ProcessMarkStackRef()
2233 << " to_ref=" << to_ref in ProcessMarkStackRef()
2234 << " rb_state=" << to_ref->GetReadBarrierState() in ProcessMarkStackRef()
2235 << " is_marked=" << IsMarked(to_ref) in ProcessMarkStackRef()
2236 << " type=" << to_ref->PrettyTypeOf() in ProcessMarkStackRef()
2238 << " space=" << heap_->DumpSpaceNameFromAddress(to_ref) in ProcessMarkStackRef()
2246 DCHECK(!region_space_->IsInNewlyAllocatedRegion(to_ref)) << to_ref; in ProcessMarkStackRef()
2251 if (!kUseBakerReadBarrier || !region_space_bitmap_->Set(to_ref)) { in ProcessMarkStackRef()
2255 CHECK(region_space_->IsLargeObject(to_ref)); in ProcessMarkStackRef()
2256 region_space_->ZeroLiveBytesForLargeObject(to_ref); in ProcessMarkStackRef()
2273 region_space_bitmap_->Set(to_ref); in ProcessMarkStackRef()
2278 DCHECK(!region_space_->HasAddress(to_ref)) << to_ref; in ProcessMarkStackRef()
2279 DCHECK(!immune_spaces_.ContainsObject(to_ref)); in ProcessMarkStackRef()
2284 const bool is_los = !mark_bitmap->HasAddress(to_ref); in ProcessMarkStackRef()
2286 if (!IsAligned<kPageSize>(to_ref)) { in ProcessMarkStackRef()
2294 to_ref, in ProcessMarkStackRef()
2298 << "ref=" << to_ref in ProcessMarkStackRef()
2302 DCHECK(los_bitmap->HasAddress(to_ref)); in ProcessMarkStackRef()
2305 perform_scan = !los_bitmap->Set(to_ref); in ProcessMarkStackRef()
2309 perform_scan = !mark_bitmap->Set(to_ref); in ProcessMarkStackRef()
2316 obj_size = to_ref->SizeOf<kDefaultVerifyFlags>(); in ProcessMarkStackRef()
2318 Scan<true>(to_ref, obj_size); in ProcessMarkStackRef()
2320 Scan<false>(to_ref, obj_size); in ProcessMarkStackRef()
2324 DCHECK(to_ref->GetReadBarrierState() == ReadBarrier::GrayState()) in ProcessMarkStackRef()
2325 << " to_ref=" << to_ref in ProcessMarkStackRef()
2326 << " rb_state=" << to_ref->GetReadBarrierState() in ProcessMarkStackRef()
2327 << " is_marked=" << IsMarked(to_ref) in ProcessMarkStackRef()
2328 << " type=" << to_ref->PrettyTypeOf() in ProcessMarkStackRef()
2330 << " space=" << heap_->DumpSpaceNameFromAddress(to_ref) in ProcessMarkStackRef()
2337 if (UNLIKELY((to_ref->GetClass<kVerifyNone, kWithoutReadBarrier>()->IsTypeOfReferenceClass() && in ProcessMarkStackRef()
2338 (referent = to_ref->AsReference()->GetReferent<kWithoutReadBarrier>()) != nullptr && in ProcessMarkStackRef()
2342 DCHECK(to_ref->AsReference()->GetPendingNext() != nullptr) in ProcessMarkStackRef()
2343 << "Left unenqueued ref gray " << to_ref; in ProcessMarkStackRef()
2350 bool success = to_ref->AtomicSetReadBarrierState<std::memory_order_release>( in ProcessMarkStackRef()
2363 DCHECK(region_space_bitmap_->Test(to_ref)); in ProcessMarkStackRef()
2365 obj_size = to_ref->SizeOf<kDefaultVerifyFlags>(); in ProcessMarkStackRef()
2367 region_space_->AddLiveBytes(to_ref, RoundUp(obj_size, space::RegionSpace::kAlignment)); in ProcessMarkStackRef()
2370 CHECK(to_ref != nullptr); in ProcessMarkStackRef()
2372 CHECK(!region_space->IsInFromSpace(to_ref)) << "Scanning object " << to_ref << " in from space"; in ProcessMarkStackRef()
2373 AssertToSpaceInvariant(nullptr, MemberOffset(0), to_ref); in ProcessMarkStackRef()
2375 to_ref->VisitReferences</*kVisitNativeRoots=*/true, kDefaultVerifyFlags, kWithoutReadBarrier>( in ProcessMarkStackRef()
3172 inline void ConcurrentCopying::Scan(mirror::Object* to_ref, size_t obj_size) { in Scan() argument
3182 obj_size = to_ref->SizeOf<kDefaultVerifyFlags>(); in Scan()
3186 DCHECK(!region_space_->IsInFromSpace(to_ref)); in Scan()
3190 to_ref->VisitReferences</*kVisitNativeRoots=*/true, kDefaultVerifyFlags, kWithoutReadBarrier>( in Scan()
3204 mirror::Object* to_ref = Mark</*kGrayImmuneObject=*/false, kNoUnEvac, /*kFromGCThread=*/true>( in Process() local
3209 if (to_ref == ref) { in Process()
3214 mirror::Object* new_ref = to_ref; in Process()
3237 mirror::Object* to_ref = Mark(self, ref); in VisitRoots() local
3238 if (to_ref == ref) { in VisitRoots()
3243 mirror::Object* new_ref = to_ref; in VisitRoots()
3258 mirror::Object* to_ref = Mark<kGrayImmuneObject>(self, ref); in MarkRoot() local
3259 if (to_ref != ref) { in MarkRoot()
3262 auto new_ref = mirror::CompressedReference<mirror::Object>::FromMirrorPtr(to_ref); in MarkRoot()
3444 mirror::Object* to_ref = region_space_->AllocNonvirtual</*kForEvac=*/ true>( in Copy() local
3447 if (LIKELY(to_ref != nullptr)) { in Copy()
3451 to_ref = AllocateInSkippedBlock(self, region_space_alloc_size); in Copy()
3452 if (to_ref != nullptr) { in Copy()
3456 region_space_->RecordAlloc(to_ref); in Copy()
3471 to_ref = heap_->non_moving_space_->Alloc( in Copy()
3473 if (UNLIKELY(to_ref == nullptr)) { in Copy()
3482 DCHECK(to_ref != nullptr); in Copy()
3485 to_ref->SetClass(klass); in Copy()
3493 memcpy(reinterpret_cast<uint8_t*>(to_ref) + kObjectHeaderSize, in Copy()
3507 FillWithFakeObject(self, to_ref, bytes_allocated); in Copy()
3509 DCHECK(region_space_->IsInToSpace(to_ref)); in Copy()
3512 region_space_->FreeLarge</*kForEvac=*/ true>(to_ref, bytes_allocated); in Copy()
3520 reinterpret_cast<uint8_t*>(to_ref))); in Copy()
3523 DCHECK(heap_->non_moving_space_->HasAddress(to_ref)); in Copy()
3526 heap_->non_moving_space_->Free(self, to_ref); in Copy()
3530 mirror::Object* lost_fwd_ptr = to_ref; in Copy()
3531 to_ref = reinterpret_cast<mirror::Object*>(old_lock_word.ForwardingAddress()); in Copy()
3532 CHECK(to_ref != nullptr); in Copy()
3533 CHECK_NE(to_ref, lost_fwd_ptr); in Copy()
3534 CHECK(region_space_->IsInToSpace(to_ref) || heap_->non_moving_space_->HasAddress(to_ref)) in Copy()
3535 << "to_ref=" << to_ref << " " << heap_->DumpSpaces(); in Copy()
3536 CHECK_NE(to_ref->GetLockWord(false).GetState(), LockWord::kForwardingAddress); in Copy()
3537 return to_ref; in Copy()
3541 to_ref->SetLockWord(old_lock_word, false); in Copy()
3544 to_ref->SetReadBarrierState(ReadBarrier::GrayState()); in Copy()
3547 LockWord new_lock_word = LockWord::FromForwardingAddress(reinterpret_cast<size_t>(to_ref)); in Copy()
3567 DCHECK(region_space_->IsInToSpace(to_ref)); in Copy()
3569 DCHECK(heap_->non_moving_space_->HasAddress(to_ref)); in Copy()
3573 CHECK(!heap_->non_moving_space_->GetLiveBitmap()->AtomicTestAndSet(to_ref)); in Copy()
3577 CHECK(!heap_->non_moving_space_->GetMarkBitmap()->AtomicTestAndSet(to_ref)); in Copy()
3581 DCHECK(to_ref->GetReadBarrierState() == ReadBarrier::GrayState()); in Copy()
3583 DCHECK(GetFwdPtr(from_ref) == to_ref); in Copy()
3584 CHECK_NE(to_ref->GetLockWord(false).GetState(), LockWord::kForwardingAddress); in Copy()
3588 PushOntoMarkStack(self, to_ref); in Copy()
3589 return to_ref; in Copy()
3604 mirror::Object* to_ref; in IsMarked() local
3606 to_ref = GetFwdPtr(from_ref); in IsMarked()
3607 DCHECK(to_ref == nullptr || region_space_->IsInToSpace(to_ref) || in IsMarked()
3608 heap_->non_moving_space_->HasAddress(to_ref)) in IsMarked()
3609 << "from_ref=" << from_ref << " to_ref=" << to_ref; in IsMarked()
3612 to_ref = from_ref; in IsMarked()
3614 to_ref = nullptr; in IsMarked()
3623 to_ref = from_ref; in IsMarked()
3628 to_ref = from_ref; in IsMarked()
3630 to_ref = nullptr; in IsMarked()
3634 return to_ref; in IsMarked()
3800 mirror::Object* to_ref = IsMarked(from_ref); in IsNullOrMarkedHeapReference() local
3801 if (to_ref == nullptr) { in IsNullOrMarkedHeapReference()
3804 if (from_ref != to_ref) { in IsNullOrMarkedHeapReference()
3811 } while (!field->CasWeakRelaxed(from_ref, to_ref)); in IsNullOrMarkedHeapReference()
3813 field->Assign(to_ref); in IsNullOrMarkedHeapReference()