// Copyright 2018 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef V8_OBJECTS_HEAP_OBJECT_H_ #define V8_OBJECTS_HEAP_OBJECT_H_ #include "src/common/globals.h" #include "src/roots/roots.h" #include "src/objects/objects.h" #include "src/objects/tagged-field.h" // Has to be the last include (doesn't have include guards): #include "src/objects/object-macros.h" namespace v8 { namespace internal { class Heap; // HeapObject is the superclass for all classes describing heap allocated // objects. class HeapObject : public Object { public: bool is_null() const { return static_cast(ptr()) == static_cast(kNullAddress); } // [map]: Contains a map which contains the object's reflective // information. DECL_GETTER(map, Map) inline void set_map(Map value); inline ObjectSlot map_slot() const; // The no-write-barrier version. This is OK if the object is white and in // new space, or if the value is an immortal immutable object, like the maps // of primitive (non-JS) objects like strings, heap numbers etc. inline void set_map_no_write_barrier(Map value); // Access the map using acquire load and release store. DECL_GETTER(synchronized_map, Map) inline void synchronized_set_map(Map value); // Compare-and-swaps map word using release store, returns true if the map // word was actually swapped. inline bool synchronized_compare_and_swap_map_word(MapWord old_map_word, MapWord new_map_word); // Initialize the map immediately after the object is allocated. // Do not use this outside Heap. inline void set_map_after_allocation( Map value, WriteBarrierMode mode = UPDATE_WRITE_BARRIER); // During garbage collection, the map word of a heap object does not // necessarily contain a map pointer. DECL_GETTER(map_word, MapWord) inline void set_map_word(MapWord map_word); // Access the map word using acquire load and release store. DECL_GETTER(synchronized_map_word, MapWord) inline void synchronized_set_map_word(MapWord map_word); // TODO(v8:7464): Once RO_SPACE is shared between isolates, this method can be // removed as ReadOnlyRoots will be accessible from a global variable. For now // this method exists to help remove GetIsolate/GetHeap from HeapObject, in a // way that doesn't require passing Isolate/Heap down huge call chains or to // places where it might not be safe to access it. inline ReadOnlyRoots GetReadOnlyRoots() const; // This version is intended to be used for the isolate values produced by // i::GetIsolateForPtrCompr(HeapObject) function which may return nullptr. inline ReadOnlyRoots GetReadOnlyRoots(Isolate* isolate) const; #define IS_TYPE_FUNCTION_DECL(Type) \ V8_INLINE bool Is##Type() const; \ V8_INLINE bool Is##Type(Isolate* isolate) const; HEAP_OBJECT_TYPE_LIST(IS_TYPE_FUNCTION_DECL) IS_TYPE_FUNCTION_DECL(HashTableBase) IS_TYPE_FUNCTION_DECL(SmallOrderedHashTable) #undef IS_TYPE_FUNCTION_DECL bool IsExternal(Isolate* isolate) const; // Oddball checks are faster when they are raw pointer comparisons, so the // isolate/read-only roots overloads should be preferred where possible. #define IS_TYPE_FUNCTION_DECL(Type, Value) \ V8_INLINE bool Is##Type(Isolate* isolate) const; \ V8_INLINE bool Is##Type(ReadOnlyRoots roots) const; \ V8_INLINE bool Is##Type() const; ODDBALL_LIST(IS_TYPE_FUNCTION_DECL) IS_TYPE_FUNCTION_DECL(NullOrUndefined, /* unused */) #undef IS_TYPE_FUNCTION_DECL #define DECL_STRUCT_PREDICATE(NAME, Name, name) \ V8_INLINE bool Is##Name() const; \ V8_INLINE bool Is##Name(Isolate* isolate) const; STRUCT_LIST(DECL_STRUCT_PREDICATE) #undef DECL_STRUCT_PREDICATE // Converts an address to a HeapObject pointer. static inline HeapObject FromAddress(Address address) { DCHECK_TAG_ALIGNED(address); return HeapObject(address + kHeapObjectTag); } // Returns the address of this HeapObject. inline Address address() const { return ptr() - kHeapObjectTag; } // Iterates over pointers contained in the object (including the Map). // If it's not performance critical iteration use the non-templatized // version. void Iterate(ObjectVisitor* v); template inline void IterateFast(ObjectVisitor* v); // Iterates over all pointers contained in the object except the // first map pointer. The object type is given in the first // parameter. This function does not access the map pointer in the // object, and so is safe to call while the map pointer is modified. // If it's not performance critical iteration use the non-templatized // version. void IterateBody(ObjectVisitor* v); void IterateBody(Map map, int object_size, ObjectVisitor* v); template inline void IterateBodyFast(ObjectVisitor* v); template inline void IterateBodyFast(Map map, int object_size, ObjectVisitor* v); // Returns true if the object contains a tagged value at given offset. // It is used for invalid slots filtering. If the offset points outside // of the object or to the map word, the result is UNDEFINED (!!!). V8_EXPORT_PRIVATE bool IsValidSlot(Map map, int offset); // Returns the heap object's size in bytes inline int Size() const; // Given a heap object's map pointer, returns the heap size in bytes // Useful when the map pointer field is used for other purposes. // GC internal. V8_EXPORT_PRIVATE int SizeFromMap(Map map) const; // Returns the field at offset in obj, as a read/write Object reference. // Does no checking, and is safe to use during GC, while maps are invalid. // Does not invoke write barrier, so should only be assigned to // during marking GC. inline ObjectSlot RawField(int byte_offset) const; inline MaybeObjectSlot RawMaybeWeakField(int byte_offset) const; DECL_CAST(HeapObject) // Return the write barrier mode for this. Callers of this function // must be able to present a reference to an DisallowHeapAllocation // object as a sign that they are not going to use this function // from code that allocates and thus invalidates the returned write // barrier mode. inline WriteBarrierMode GetWriteBarrierMode( const DisallowHeapAllocation& promise); // Dispatched behavior. void HeapObjectShortPrint(std::ostream& os); // NOLINT #ifdef OBJECT_PRINT void PrintHeader(std::ostream& os, const char* id); // NOLINT #endif DECL_PRINTER(HeapObject) EXPORT_DECL_VERIFIER(HeapObject) #ifdef VERIFY_HEAP inline void VerifyObjectField(Isolate* isolate, int offset); inline void VerifySmiField(int offset); inline void VerifyMaybeObjectField(Isolate* isolate, int offset); // Verify a pointer is a valid HeapObject pointer that points to object // areas in the heap. static void VerifyHeapPointer(Isolate* isolate, Object p); #endif static inline AllocationAlignment RequiredAlignment(Map map); // Whether the object needs rehashing. That is the case if the object's // content depends on FLAG_hash_seed. When the object is deserialized into // a heap with a different hash seed, these objects need to adapt. bool NeedsRehashing() const; // Rehashing support is not implemented for all objects that need rehashing. // With objects that need rehashing but cannot be rehashed, rehashing has to // be disabled. bool CanBeRehashed() const; // Rehash the object based on the layout inferred from its map. void RehashBasedOnMap(ReadOnlyRoots root); // Layout description. #define HEAP_OBJECT_FIELDS(V) \ V(kMapOffset, kTaggedSize) \ /* Header size. */ \ V(kHeaderSize, 0) DEFINE_FIELD_OFFSET_CONSTANTS(Object::kHeaderSize, HEAP_OBJECT_FIELDS) #undef HEAP_OBJECT_FIELDS STATIC_ASSERT(kMapOffset == Internals::kHeapObjectMapOffset); using MapField = TaggedField; inline Address GetFieldAddress(int field_offset) const; protected: // Special-purpose constructor for subclasses that have fast paths where // their ptr() is a Smi. enum class AllowInlineSmiStorage { kRequireHeapObjectTag, kAllowBeingASmi }; inline HeapObject(Address ptr, AllowInlineSmiStorage allow_smi); OBJECT_CONSTRUCTORS(HeapObject, Object); }; OBJECT_CONSTRUCTORS_IMPL(HeapObject, Object) CAST_ACCESSOR(HeapObject) } // namespace internal } // namespace v8 #include "src/objects/object-macros-undef.h" #endif // V8_OBJECTS_HEAP_OBJECT_H_