%PDF- %PDF-
Direktori : /home/vacivi36/vittasync.vacivitta.com.br/vittasync/node/deps/v8/src/heap/ |
Current File : /home/vacivi36/vittasync.vacivitta.com.br/vittasync/node/deps/v8/src/heap/marking-barrier-inl.h |
// Copyright 2020 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef V8_HEAP_MARKING_BARRIER_INL_H_ #define V8_HEAP_MARKING_BARRIER_INL_H_ #include "src/base/logging.h" #include "src/heap/incremental-marking-inl.h" #include "src/heap/incremental-marking.h" #include "src/heap/marking-barrier.h" namespace v8 { namespace internal { void MarkingBarrier::MarkValue(Tagged<HeapObject> host, Tagged<HeapObject> value) { if (value.InReadOnlySpace()) return; DCHECK(IsCurrentMarkingBarrier(host)); DCHECK(is_activated_ || shared_heap_worklist_.has_value()); // When shared heap isn't enabled all objects are local, we can just run the // local marking barrier. Also from the point-of-view of the shared space // isolate (= main isolate) also shared objects are considered local. if (V8_UNLIKELY(uses_shared_heap_) && !is_shared_space_isolate_) { // Check whether incremental marking is enabled for that object's space. if (!MemoryChunk::FromHeapObject(host)->IsMarking()) { return; } if (host.InWritableSharedSpace()) { // Invoking shared marking barrier when storing into shared objects. MarkValueShared(value); return; } else if (value.InWritableSharedSpace()) { // No marking needed when storing shared objects in local objects. return; } } DCHECK_IMPLIES(host.InWritableSharedSpace(), is_shared_space_isolate_); DCHECK_IMPLIES(value.InWritableSharedSpace(), is_shared_space_isolate_); DCHECK(is_activated_); MarkValueLocal(value); } void MarkingBarrier::MarkValueShared(Tagged<HeapObject> value) { // Value is either in read-only space or shared heap. DCHECK(value.InAnySharedSpace()); // We should only reach this on client isolates (= worker isolates). DCHECK(!is_shared_space_isolate_); DCHECK(shared_heap_worklist_.has_value()); // Mark shared object and push it onto shared heap worklist. if (marking_state_.TryMark(value)) { shared_heap_worklist_->Push(value); } } void MarkingBarrier::MarkValueLocal(Tagged<HeapObject> value) { DCHECK(!value.InReadOnlySpace()); if (is_minor()) { // We do not need to insert into RememberedSet<OLD_TO_NEW> here because the // C++ marking barrier already does this for us. // TODO(v8:13012): Consider updating C++ barriers to respect // POINTERS_TO_HERE_ARE_INTERESTING and POINTERS_FROM_HERE_ARE_INTERESTING // page flags and make the following branch a DCHECK. if (Heap::InYoungGeneration(value)) { WhiteToGreyAndPush(value); // NEW->NEW } } else { if (WhiteToGreyAndPush(value)) { if (V8_UNLIKELY(v8_flags.track_retaining_path)) { heap_->AddRetainingRoot(Root::kWriteBarrier, value); } } } } template <typename TSlot> inline void MarkingBarrier::MarkRange(Tagged<HeapObject> host, TSlot start, TSlot end) { auto* isolate = heap_->isolate(); const bool record_slots = IsCompacting(host) && !MemoryChunk::FromHeapObject(host)->ShouldSkipEvacuationSlotRecording(); for (TSlot slot = start; slot < end; ++slot) { typename TSlot::TObject object = slot.Relaxed_Load(); Tagged<HeapObject> heap_object; // Mark both, weak and strong edges. if (object.GetHeapObject(isolate, &heap_object)) { MarkValue(host, heap_object); if (record_slots) { major_collector_->RecordSlot(host, HeapObjectSlot(slot), heap_object); } } } } bool MarkingBarrier::IsCompacting(Tagged<HeapObject> object) const { if (is_compacting_) { DCHECK(is_major()); return true; } return shared_heap_worklist_.has_value() && object.InWritableSharedSpace(); } bool MarkingBarrier::WhiteToGreyAndPush(Tagged<HeapObject> obj) { if (marking_state_.TryMark(obj)) { current_worklist_->Push(obj); return true; } return false; } } // namespace internal } // namespace v8 #endif // V8_HEAP_MARKING_BARRIER_INL_H_