Source code
Revision control
Copy as Markdown
Other Tools
/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*-
* vim: set ts=8 sts=2 et sw=2 tw=80:
* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
#include "vm/NativeObject-inl.h"
#include "mozilla/Casting.h"
#include "mozilla/CheckedInt.h"
#include "mozilla/Maybe.h"
#include <algorithm>
#include <iterator>
#include "gc/MaybeRooted.h"
#include "gc/StableCellHasher.h"
#include "js/friend/ErrorMessages.h" // js::GetErrorMessage, JSMSG_*
#include "js/friend/StackLimits.h" // js::AutoCheckRecursionLimit
#include "js/Value.h"
#include "vm/EqualityOperations.h" // js::SameValue
#include "vm/GetterSetter.h" // js::GetterSetter
#include "vm/Interpreter.h" // js::CallGetter, js::CallSetter
#include "vm/PlainObject.h" // js::PlainObject
#include "vm/TypedArrayObject.h"
#ifdef ENABLE_RECORD_TUPLE
# include "builtin/RecordObject.h"
# include "builtin/TupleObject.h"
# include "vm/RecordTupleShared.h"
#endif
#include "gc/Nursery-inl.h"
#include "vm/JSObject-inl.h"
#include "vm/Shape-inl.h"
using namespace js;
using JS::AutoCheckCannotGC;
using mozilla::CheckedInt;
using mozilla::PodCopy;
using mozilla::RoundUpPow2;
struct EmptyObjectElements {
const ObjectElements emptyElementsHeader;
// Add an extra (unused) Value to make sure an out-of-bounds index when
// masked (resulting in index 0) accesses valid memory.
const Value val;
public:
constexpr EmptyObjectElements()
: emptyElementsHeader(0, 0), val(UndefinedValue()) {}
explicit constexpr EmptyObjectElements(ObjectElements::SharedMemory shmem)
: emptyElementsHeader(0, 0, shmem), val(UndefinedValue()) {}
};
static constexpr EmptyObjectElements emptyElementsHeader;
/* Objects with no elements share one empty set of elements. */
HeapSlot* const js::emptyObjectElements = reinterpret_cast<HeapSlot*>(
uintptr_t(&emptyElementsHeader) + sizeof(ObjectElements));
static constexpr EmptyObjectElements emptyElementsHeaderShared(
ObjectElements::SharedMemory::IsShared);
/* Objects with no elements share one empty set of elements. */
HeapSlot* const js::emptyObjectElementsShared = reinterpret_cast<HeapSlot*>(
uintptr_t(&emptyElementsHeaderShared) + sizeof(ObjectElements));
struct EmptyObjectSlots : public ObjectSlots {
explicit constexpr EmptyObjectSlots(size_t dictionarySlotSpan)
: ObjectSlots(0, dictionarySlotSpan, NoUniqueIdInSharedEmptySlots) {}
};
static constexpr EmptyObjectSlots emptyObjectSlotsHeaders[17] = {
EmptyObjectSlots(0), EmptyObjectSlots(1), EmptyObjectSlots(2),
EmptyObjectSlots(3), EmptyObjectSlots(4), EmptyObjectSlots(5),
EmptyObjectSlots(6), EmptyObjectSlots(7), EmptyObjectSlots(8),
EmptyObjectSlots(9), EmptyObjectSlots(10), EmptyObjectSlots(11),
EmptyObjectSlots(12), EmptyObjectSlots(13), EmptyObjectSlots(14),
EmptyObjectSlots(15), EmptyObjectSlots(16)};
static_assert(std::size(emptyObjectSlotsHeaders) ==
NativeObject::MAX_FIXED_SLOTS + 1);
HeapSlot* const js::emptyObjectSlotsForDictionaryObject[17] = {
emptyObjectSlotsHeaders[0].slots(), emptyObjectSlotsHeaders[1].slots(),
emptyObjectSlotsHeaders[2].slots(), emptyObjectSlotsHeaders[3].slots(),
emptyObjectSlotsHeaders[4].slots(), emptyObjectSlotsHeaders[5].slots(),
emptyObjectSlotsHeaders[6].slots(), emptyObjectSlotsHeaders[7].slots(),
emptyObjectSlotsHeaders[8].slots(), emptyObjectSlotsHeaders[9].slots(),
emptyObjectSlotsHeaders[10].slots(), emptyObjectSlotsHeaders[11].slots(),
emptyObjectSlotsHeaders[12].slots(), emptyObjectSlotsHeaders[13].slots(),
emptyObjectSlotsHeaders[14].slots(), emptyObjectSlotsHeaders[15].slots(),
emptyObjectSlotsHeaders[16].slots()};
static_assert(std::size(emptyObjectSlotsForDictionaryObject) ==
NativeObject::MAX_FIXED_SLOTS + 1);
HeapSlot* const js::emptyObjectSlots = emptyObjectSlotsForDictionaryObject[0];
#ifdef DEBUG
bool NativeObject::canHaveNonEmptyElements() {
return !this->is<TypedArrayObject>();
}
#endif // DEBUG
/* static */
void ObjectElements::PrepareForPreventExtensions(JSContext* cx,
NativeObject* obj) {
if (!obj->hasEmptyElements()) {
obj->shrinkCapacityToInitializedLength(cx);
}
// shrinkCapacityToInitializedLength ensures there are no shifted elements.
MOZ_ASSERT(obj->getElementsHeader()->numShiftedElements() == 0);
}
/* static */
void ObjectElements::PreventExtensions(NativeObject* obj) {
MOZ_ASSERT(!obj->isExtensible());
MOZ_ASSERT(obj->getElementsHeader()->numShiftedElements() == 0);
MOZ_ASSERT(obj->getDenseInitializedLength() == obj->getDenseCapacity());
if (!obj->hasEmptyElements()) {
obj->getElementsHeader()->setNotExtensible();
}
}
/* static */
bool ObjectElements::FreezeOrSeal(JSContext* cx, Handle<NativeObject*> obj,
IntegrityLevel level) {
MOZ_ASSERT_IF(level == IntegrityLevel::Frozen && obj->is<ArrayObject>(),
!obj->as<ArrayObject>().lengthIsWritable());
MOZ_ASSERT(!obj->isExtensible());
MOZ_ASSERT(obj->getElementsHeader()->numShiftedElements() == 0);
if (obj->hasEmptyElements() || obj->denseElementsAreFrozen()) {
return true;
}
if (level == IntegrityLevel::Frozen) {
if (!JSObject::setFlag(cx, obj, ObjectFlag::FrozenElements)) {
return false;
}
}
if (!obj->denseElementsAreSealed()) {
obj->getElementsHeader()->seal();
}
if (level == IntegrityLevel::Frozen) {
obj->getElementsHeader()->freeze();
}
return true;
}
#ifdef DEBUG
static mozilla::Atomic<bool, mozilla::Relaxed> gShapeConsistencyChecksEnabled(
false);
/* static */
void js::NativeObject::enableShapeConsistencyChecks() {
gShapeConsistencyChecksEnabled = true;
}
void js::NativeObject::checkShapeConsistency() {
if (!gShapeConsistencyChecksEnabled) {
return;
}
MOZ_ASSERT(is<NativeObject>());
if (PropMap* map = shape()->propMap()) {
map->checkConsistency(this);
} else {
MOZ_ASSERT(shape()->propMapLength() == 0);
}
}
#endif
#ifdef DEBUG
bool js::NativeObject::slotInRange(uint32_t slot,
SentinelAllowed sentinel) const {
MOZ_ASSERT(!gc::IsForwarded(shape()));
uint32_t capacity = numFixedSlots() + numDynamicSlots();
if (sentinel == SENTINEL_ALLOWED) {
return slot <= capacity;
}
return slot < capacity;
}
bool js::NativeObject::slotIsFixed(uint32_t slot) const {
// We call numFixedSlotsMaybeForwarded() to allow reading slots of
// associated objects in trace hooks that may be called during a moving GC.
return slot < numFixedSlotsMaybeForwarded();
}
bool js::NativeObject::isNumFixedSlots(uint32_t nfixed) const {
// We call numFixedSlotsMaybeForwarded() to allow reading slots of
// associated objects in trace hooks that may be called during a moving GC.
return nfixed == numFixedSlotsMaybeForwarded();
}
uint32_t js::NativeObject::outOfLineNumDynamicSlots() const {
return numDynamicSlots();
}
#endif /* DEBUG */
mozilla::Maybe<PropertyInfo> js::NativeObject::lookup(JSContext* cx, jsid id) {
MOZ_ASSERT(is<NativeObject>());
uint32_t index;
if (PropMap* map = shape()->lookup(cx, id, &index)) {
return mozilla::Some(map->getPropertyInfo(index));
}
return mozilla::Nothing();
}
mozilla::Maybe<PropertyInfo> js::NativeObject::lookupPure(jsid id) {
MOZ_ASSERT(is<NativeObject>());
uint32_t index;
if (PropMap* map = shape()->lookupPure(id, &index)) {
return mozilla::Some(map->getPropertyInfo(index));
}
return mozilla::Nothing();
}
bool NativeObject::setUniqueId(JSContext* cx, uint64_t uid) {
MOZ_ASSERT(!hasUniqueId());
MOZ_ASSERT(!gc::HasUniqueId(this));
return setOrUpdateUniqueId(cx, uid);
}
bool NativeObject::setOrUpdateUniqueId(JSContext* cx, uint64_t uid) {
if (!hasDynamicSlots() && !allocateSlots(cx, 0)) {
return false;
}
getSlotsHeader()->setUniqueId(uid);
return true;
}
bool NativeObject::growSlots(JSContext* cx, uint32_t oldCapacity,
uint32_t newCapacity) {
MOZ_ASSERT(newCapacity > oldCapacity);
/*
* Slot capacities are determined by the span of allocated objects. Due to
* the limited number of bits to store shape slots, object growth is
* throttled well before the slot capacity can overflow.
*/
NativeObject::slotsSizeMustNotOverflow();
MOZ_ASSERT(newCapacity <= MAX_SLOTS_COUNT);
if (!hasDynamicSlots()) {
return allocateSlots(cx, newCapacity);
}
uint64_t uid = maybeUniqueId();
uint32_t newAllocated = ObjectSlots::allocCount(newCapacity);
uint32_t dictionarySpan = getSlotsHeader()->dictionarySlotSpan();
uint32_t oldAllocated = ObjectSlots::allocCount(oldCapacity);
ObjectSlots* oldHeaderSlots = ObjectSlots::fromSlots(slots_);
MOZ_ASSERT(oldHeaderSlots->capacity() == oldCapacity);
HeapSlot* allocation = ReallocateObjectBuffer<HeapSlot>(
cx, this, reinterpret_cast<HeapSlot*>(oldHeaderSlots), oldAllocated,
newAllocated);
if (!allocation) {
return false; /* Leave slots at its old size. */
}
auto* newHeaderSlots =
new (allocation) ObjectSlots(newCapacity, dictionarySpan, uid);
slots_ = newHeaderSlots->slots();
Debug_SetSlotRangeToCrashOnTouch(slots_ + oldCapacity,
newCapacity - oldCapacity);
RemoveCellMemory(this, ObjectSlots::allocSize(oldCapacity),
MemoryUse::ObjectSlots);
AddCellMemory(this, ObjectSlots::allocSize(newCapacity),
MemoryUse::ObjectSlots);
MOZ_ASSERT(hasDynamicSlots());
return true;
}
bool NativeObject::growSlotsForNewSlot(JSContext* cx, uint32_t numFixed,
uint32_t slot) {
MOZ_ASSERT(slotSpan() == slot);
MOZ_ASSERT(shape()->numFixedSlots() == numFixed);
MOZ_ASSERT(slot >= numFixed);
uint32_t newCapacity = calculateDynamicSlots(numFixed, slot + 1, getClass());
uint32_t oldCapacity = numDynamicSlots();
MOZ_ASSERT(oldCapacity < newCapacity);
return growSlots(cx, oldCapacity, newCapacity);
}
bool NativeObject::allocateInitialSlots(JSContext* cx, uint32_t capacity) {
uint32_t count = ObjectSlots::allocCount(capacity);
HeapSlot* allocation = AllocateObjectBuffer<HeapSlot>(cx, this, count);
if (!allocation) {
// The new object will be unreachable, but we still have to make it safe
// for finalization. Also we must check for it during GC compartment
// checks (see IsPartiallyInitializedObject).
initEmptyDynamicSlots();
return false;
}
auto* headerSlots = new (allocation)
ObjectSlots(capacity, 0, ObjectSlots::NoUniqueIdInDynamicSlots);
slots_ = headerSlots->slots();
Debug_SetSlotRangeToCrashOnTouch(slots_, capacity);
if (!IsInsideNursery(this)) {
AddCellMemory(this, ObjectSlots::allocSize(capacity),
MemoryUse::ObjectSlots);
}
MOZ_ASSERT(hasDynamicSlots());
return true;
}
bool NativeObject::allocateSlots(JSContext* cx, uint32_t newCapacity) {
MOZ_ASSERT(!hasUniqueId());
MOZ_ASSERT(!hasDynamicSlots());
uint32_t newAllocated = ObjectSlots::allocCount(newCapacity);
uint32_t dictionarySpan = getSlotsHeader()->dictionarySlotSpan();
HeapSlot* allocation = AllocateObjectBuffer<HeapSlot>(cx, this, newAllocated);
if (!allocation) {
return false;
}
auto* newHeaderSlots = new (allocation) ObjectSlots(
newCapacity, dictionarySpan, ObjectSlots::NoUniqueIdInDynamicSlots);
slots_ = newHeaderSlots->slots();
Debug_SetSlotRangeToCrashOnTouch(slots_, newCapacity);
AddCellMemory(this, ObjectSlots::allocSize(newCapacity),
MemoryUse::ObjectSlots);
MOZ_ASSERT(hasDynamicSlots());
return true;
}
/* static */
bool NativeObject::growSlotsPure(JSContext* cx, NativeObject* obj,
uint32_t newCapacity) {
// IC code calls this directly.
AutoUnsafeCallWithABI unsafe;
if (!obj->growSlots(cx, obj->numDynamicSlots(), newCapacity)) {
cx->recoverFromOutOfMemory();
return false;
}
return true;
}
/* static */
bool NativeObject::addDenseElementPure(JSContext* cx, NativeObject* obj) {
// IC code calls this directly.
AutoUnsafeCallWithABI unsafe;
MOZ_ASSERT(obj->getDenseInitializedLength() == obj->getDenseCapacity());
MOZ_ASSERT(obj->isExtensible());
MOZ_ASSERT(!obj->isIndexed());
MOZ_ASSERT(!obj->is<TypedArrayObject>());
MOZ_ASSERT_IF(obj->is<ArrayObject>(),
obj->as<ArrayObject>().lengthIsWritable());
// growElements will report OOM also if the number of dense elements will
// exceed MAX_DENSE_ELEMENTS_COUNT. See goodElementsAllocationAmount.
uint32_t oldCapacity = obj->getDenseCapacity();
if (MOZ_UNLIKELY(!obj->growElements(cx, oldCapacity + 1))) {
cx->recoverFromOutOfMemory();
return false;
}
MOZ_ASSERT(obj->getDenseCapacity() > oldCapacity);
MOZ_ASSERT(obj->getDenseCapacity() <= MAX_DENSE_ELEMENTS_COUNT);
return true;
}
static inline void FreeSlots(JSContext* cx, NativeObject* obj,
ObjectSlots* slots, size_t nbytes) {
// Note: this is called when shrinking slots, not from the finalizer.
MOZ_ASSERT(cx->isMainThreadContext());
if (obj->isTenured()) {
MOZ_ASSERT(!cx->nursery().isInside(slots));
js_free(slots);
} else {
cx->nursery().freeBuffer(slots, nbytes);
}
}
void NativeObject::shrinkSlots(JSContext* cx, uint32_t oldCapacity,
uint32_t newCapacity) {
MOZ_ASSERT(hasDynamicSlots());
MOZ_ASSERT(newCapacity < oldCapacity);
MOZ_ASSERT(oldCapacity == getSlotsHeader()->capacity());
ObjectSlots* oldHeaderSlots = ObjectSlots::fromSlots(slots_);
MOZ_ASSERT(oldHeaderSlots->capacity() == oldCapacity);
uint64_t uid = maybeUniqueId();
uint32_t oldAllocated = ObjectSlots::allocCount(oldCapacity);
if (newCapacity == 0 && uid == 0) {
size_t nbytes = ObjectSlots::allocSize(oldCapacity);
RemoveCellMemory(this, nbytes, MemoryUse::ObjectSlots);
FreeSlots(cx, this, oldHeaderSlots, nbytes);
// dictionarySlotSpan is initialized to the correct value by the callers.
setEmptyDynamicSlots(0);
return;
}
MOZ_ASSERT_IF(!is<ArrayObject>() && !hasUniqueId(),
newCapacity >= SLOT_CAPACITY_MIN);
uint32_t dictionarySpan = getSlotsHeader()->dictionarySlotSpan();
uint32_t newAllocated = ObjectSlots::allocCount(newCapacity);
HeapSlot* allocation = ReallocateObjectBuffer<HeapSlot>(
cx, this, reinterpret_cast<HeapSlot*>(oldHeaderSlots), oldAllocated,
newAllocated);
if (!allocation) {
// It's possible for realloc to fail when shrinking an allocation. In this
// case we continue using the original allocation but still update the
// capacity to the new requested capacity, which is smaller than the actual
// capacity.
cx->recoverFromOutOfMemory();
allocation = reinterpret_cast<HeapSlot*>(getSlotsHeader());
}
RemoveCellMemory(this, ObjectSlots::allocSize(oldCapacity),
MemoryUse::ObjectSlots);
AddCellMemory(this, ObjectSlots::allocSize(newCapacity),
MemoryUse::ObjectSlots);
auto* newHeaderSlots =
new (allocation) ObjectSlots(newCapacity, dictionarySpan, uid);
slots_ = newHeaderSlots->slots();
}
void NativeObject::initFixedElements(gc::AllocKind kind, uint32_t length) {
uint32_t capacity =
gc::GetGCKindSlots(kind) - ObjectElements::VALUES_PER_HEADER;
setFixedElements();
new (getElementsHeader()) ObjectElements(capacity, length);
getElementsHeader()->flags |= ObjectElements::FIXED;
MOZ_ASSERT(hasFixedElements());
}
bool NativeObject::willBeSparseElements(uint32_t requiredCapacity,
uint32_t newElementsHint) {
MOZ_ASSERT(is<NativeObject>());
MOZ_ASSERT(requiredCapacity > MIN_SPARSE_INDEX);
uint32_t cap = getDenseCapacity();
MOZ_ASSERT(requiredCapacity >= cap);
if (requiredCapacity > MAX_DENSE_ELEMENTS_COUNT) {
return true;
}
uint32_t minimalDenseCount = requiredCapacity / SPARSE_DENSITY_RATIO;
if (newElementsHint >= minimalDenseCount) {
return false;
}
minimalDenseCount -= newElementsHint;
if (minimalDenseCount > cap) {
return true;
}
uint32_t len = getDenseInitializedLength();
const Value* elems = getDenseElements();
for (uint32_t i = 0; i < len; i++) {
if (!elems[i].isMagic(JS_ELEMENTS_HOLE) && !--minimalDenseCount) {
return false;
}
}
return true;
}
/* static */
DenseElementResult NativeObject::maybeDensifySparseElements(
JSContext* cx, Handle<NativeObject*> obj) {
/*
* Wait until after the object goes into dictionary mode, which must happen
* when sparsely packing any array with more than MIN_SPARSE_INDEX elements
* (see PropertyTree::MAX_HEIGHT).
*/
if (!obj->inDictionaryMode()) {
return DenseElementResult::Incomplete;
}
/*
* Only measure the number of indexed properties every log(n) times when
* populating the object.
*/
uint32_t slotSpan = obj->slotSpan();
if (slotSpan != RoundUpPow2(slotSpan)) {
return DenseElementResult::Incomplete;
}
/* Watch for conditions under which an object's elements cannot be dense. */
if (!obj->isExtensible()) {
return DenseElementResult::Incomplete;
}
/*
* The indexes in the object need to be sufficiently dense before they can
* be converted to dense mode.
*/
uint32_t numDenseElements = 0;
uint32_t newInitializedLength = 0;
for (ShapePropertyIter<NoGC> iter(obj->shape()); !iter.done(); iter++) {
uint32_t index;
if (!IdIsIndex(iter->key(), &index)) {
continue;
}
if (iter->flags() != PropertyFlags::defaultDataPropFlags) {
// For simplicity, only densify the object if all indexed properties can
// be converted to dense elements.
return DenseElementResult::Incomplete;
}
MOZ_ASSERT(iter->isDataProperty());
numDenseElements++;
newInitializedLength = std::max(newInitializedLength, index + 1);
}
if (numDenseElements * SPARSE_DENSITY_RATIO < newInitializedLength) {
return DenseElementResult::Incomplete;
}
if (newInitializedLength > MAX_DENSE_ELEMENTS_COUNT) {
return DenseElementResult::Incomplete;
}
/*
* This object meets all necessary restrictions, convert all indexed
* properties into dense elements.
*/
if (newInitializedLength > obj->getDenseCapacity()) {
if (!obj->growElements(cx, newInitializedLength)) {
return DenseElementResult::Failure;
}
}
obj->ensureDenseInitializedLength(newInitializedLength, 0);
if (obj->compartment()->objectMaybeInIteration(obj)) {
// Mark the densified elements as maybe-in-iteration. See also the comment
// in GetIterator.
obj->markDenseElementsMaybeInIteration();
}
if (!NativeObject::densifySparseElements(cx, obj)) {
return DenseElementResult::Failure;
}
return DenseElementResult::Success;
}
void NativeObject::moveShiftedElements() {
MOZ_ASSERT(isExtensible());
ObjectElements* header = getElementsHeader();
uint32_t numShifted = header->numShiftedElements();
MOZ_ASSERT(numShifted > 0);
uint32_t initLength = header->initializedLength;
ObjectElements* newHeader =
static_cast<ObjectElements*>(getUnshiftedElementsHeader());
memmove(newHeader, header, sizeof(ObjectElements));
newHeader->clearShiftedElements();
newHeader->capacity += numShifted;
elements_ = newHeader->elements();
// To move the elements, temporarily update initializedLength to include
// the shifted elements.
newHeader->initializedLength += numShifted;
// Move the elements. Initialize to |undefined| to ensure pre-barriers
// don't see garbage.
for (size_t i = 0; i < numShifted; i++) {
initDenseElement(i, UndefinedValue());
}
moveDenseElements(0, numShifted, initLength);
// Restore the initialized length. We use setDenseInitializedLength to
// make sure prepareElementRangeForOverwrite is called on the shifted
// elements.
setDenseInitializedLength(initLength);
}
void NativeObject::maybeMoveShiftedElements() {
MOZ_ASSERT(isExtensible());
ObjectElements* header = getElementsHeader();
MOZ_ASSERT(header->numShiftedElements() > 0);
// Move the elements if less than a third of the allocated space is in use.
if (header->capacity < header->numAllocatedElements() / 3) {
moveShiftedElements();
}
}
bool NativeObject::tryUnshiftDenseElements(uint32_t count) {
MOZ_ASSERT(isExtensible());
MOZ_ASSERT(count > 0);
ObjectElements* header = getElementsHeader();
uint32_t numShifted = header->numShiftedElements();
if (count > numShifted) {
// We need more elements than are easily available. Try to make space
// for more elements than we need (and shift the remaining ones) so
// that unshifting more elements later will be fast.
// Don't bother reserving elements if the number of elements is small.
// Note that there's no technical reason for using this particular
// limit.
if (header->initializedLength <= 10 ||
header->hasNonwritableArrayLength() ||
MOZ_UNLIKELY(count > ObjectElements::MaxShiftedElements)) {
return false;
}
MOZ_ASSERT(header->capacity >= header->initializedLength);
uint32_t unusedCapacity = header->capacity - header->initializedLength;
// Determine toShift, the number of extra elements we want to make
// available.
uint32_t toShift = count - numShifted;
MOZ_ASSERT(toShift <= ObjectElements::MaxShiftedElements,
"count <= MaxShiftedElements so toShift <= MaxShiftedElements");
// Give up if we need to allocate more elements.
if (toShift > unusedCapacity) {
return false;
}
// Move more elements than we need, so that other unshift calls will be
// fast. We just have to make sure we don't exceed unusedCapacity.
toShift = std::min(toShift + unusedCapacity / 2, unusedCapacity);
// Ensure |numShifted + toShift| does not exceed MaxShiftedElements.
if (numShifted + toShift > ObjectElements::MaxShiftedElements) {
toShift = ObjectElements::MaxShiftedElements - numShifted;
}
MOZ_ASSERT(count <= numShifted + toShift);
MOZ_ASSERT(numShifted + toShift <= ObjectElements::MaxShiftedElements);
MOZ_ASSERT(toShift <= unusedCapacity);
// Now move/unshift the elements.
uint32_t initLen = header->initializedLength;
setDenseInitializedLength(initLen + toShift);
for (uint32_t i = 0; i < toShift; i++) {
initDenseElement(initLen + i, UndefinedValue());
}
moveDenseElements(toShift, 0, initLen);
// Shift the elements we just prepended.
shiftDenseElementsUnchecked(toShift);
// We can now fall-through to the fast path below.
header = getElementsHeader();
MOZ_ASSERT(header->numShiftedElements() == numShifted + toShift);
numShifted = header->numShiftedElements();
MOZ_ASSERT(count <= numShifted);
}
elements_ -= count;
ObjectElements* newHeader = getElementsHeader();
memmove(newHeader, header, sizeof(ObjectElements));
newHeader->unshiftShiftedElements(count);
// Initialize to |undefined| to ensure pre-barriers don't see garbage.
for (uint32_t i = 0; i < count; i++) {
initDenseElement(i, UndefinedValue());
}
return true;
}
// Given a requested capacity (in elements) and (potentially) the length of an
// array for which elements are being allocated, compute an actual allocation
// amount (in elements). (Allocation amounts include space for an
// ObjectElements instance, so a return value of |N| implies
// |N - ObjectElements::VALUES_PER_HEADER| usable elements.)
//
// The requested/actual allocation distinction is meant to:
//
// * preserve amortized O(N) time to add N elements;
// * minimize the number of unused elements beyond an array's length, and
// * provide at least ELEMENT_CAPACITY_MIN elements no matter what (so adding
// the first several elements to small arrays only needs one allocation).
//
// Note: the structure and behavior of this method follow along with
// UnboxedArrayObject::chooseCapacityIndex. Changes to the allocation strategy
// in one should generally be matched by the other.
/* static */
bool NativeObject::goodElementsAllocationAmount(JSContext* cx,
uint32_t reqCapacity,
uint32_t length,
uint32_t* goodAmount) {
if (reqCapacity > MAX_DENSE_ELEMENTS_COUNT) {
ReportOutOfMemory(cx);
return false;
}
uint32_t reqAllocated = reqCapacity + ObjectElements::VALUES_PER_HEADER;
// Handle "small" requests primarily by doubling.
const uint32_t Mebi = 1 << 20;
if (reqAllocated < Mebi) {
uint32_t amount =
mozilla::AssertedCast<uint32_t>(RoundUpPow2(reqAllocated));
// If |amount| would be 2/3 or more of the array's length, adjust
// it (up or down) to be equal to the array's length. This avoids
// allocating excess elements that aren't likely to be needed, either
// in this resizing or a subsequent one. The 2/3 factor is chosen so
// that exceptional resizings will at most triple the capacity, as
// opposed to the usual doubling.
uint32_t goodCapacity = amount - ObjectElements::VALUES_PER_HEADER;
if (length >= reqCapacity && goodCapacity > (length / 3) * 2) {
amount = length + ObjectElements::VALUES_PER_HEADER;
}
if (amount < ELEMENT_CAPACITY_MIN) {
amount = ELEMENT_CAPACITY_MIN;
}
*goodAmount = amount;
return true;
}
// The almost-doubling above wastes a lot of space for larger bucket sizes.
// For large amounts, switch to bucket sizes that obey this formula:
//
// count(n+1) = Math.ceil(count(n) * 1.125)
//
// where |count(n)| is the size of the nth bucket, measured in 2**20 slots.
// These bucket sizes still preserve amortized O(N) time to add N elements,
// just with a larger constant factor.
//
// The bucket size table below was generated with this JavaScript (and
// manual reformatting):
//
// for (let n = 1, i = 0; i < 34; i++) {
// print('0x' + (n * (1 << 20)).toString(16) + ', ');
// n = Math.ceil(n * 1.125);
// }
static constexpr uint32_t BigBuckets[] = {
0x100000, 0x200000, 0x300000, 0x400000, 0x500000, 0x600000,
0x700000, 0x800000, 0x900000, 0xb00000, 0xd00000, 0xf00000,
0x1100000, 0x1400000, 0x1700000, 0x1a00000, 0x1e00000, 0x2200000,
0x2700000, 0x2c00000, 0x3200000, 0x3900000, 0x4100000, 0x4a00000,
0x5400000, 0x5f00000, 0x6b00000, 0x7900000, 0x8900000, 0x9b00000,
0xaf00000, 0xc500000, 0xde00000, 0xfa00000};
static_assert(BigBuckets[std::size(BigBuckets) - 1] <=
MAX_DENSE_ELEMENTS_ALLOCATION);
// Pick the first bucket that'll fit |reqAllocated|.
for (uint32_t b : BigBuckets) {
if (b >= reqAllocated) {
*goodAmount = b;
return true;
}
}
// Otherwise, return the maximum bucket size.
*goodAmount = MAX_DENSE_ELEMENTS_ALLOCATION;
return true;
}
bool NativeObject::growElements(JSContext* cx, uint32_t reqCapacity) {
MOZ_ASSERT(isExtensible());
MOZ_ASSERT(canHaveNonEmptyElements());
// If there are shifted elements, consider moving them first. If we don't
// move them here, the code below will include the shifted elements in the
// resize.
uint32_t numShifted = getElementsHeader()->numShiftedElements();
if (numShifted > 0) {
// If the number of elements is small, it's cheaper to just move them as
// it may avoid a malloc/realloc. Note that there's no technical reason
// for using this particular value, but it works well in real-world use
// cases.
static const size_t MaxElementsToMoveEagerly = 20;
if (getElementsHeader()->initializedLength <= MaxElementsToMoveEagerly) {
moveShiftedElements();
} else {
maybeMoveShiftedElements();
}
if (getDenseCapacity() >= reqCapacity) {
return true;
}
// moveShiftedElements() may have changed the number of shifted elements;
// update `numShifted` accordingly.
numShifted = getElementsHeader()->numShiftedElements();
// If |reqCapacity + numShifted| overflows, we just move all shifted
// elements to avoid the problem.
CheckedInt<uint32_t> checkedReqCapacity(reqCapacity);
checkedReqCapacity += numShifted;
if (MOZ_UNLIKELY(!checkedReqCapacity.isValid())) {
moveShiftedElements();
numShifted = 0;
}
}
uint32_t oldCapacity = getDenseCapacity();
MOZ_ASSERT(oldCapacity < reqCapacity);
uint32_t newAllocated = 0;
if (is<ArrayObject>() && !as<ArrayObject>().lengthIsWritable()) {
// Preserve the |capacity <= length| invariant for arrays with
// non-writable length. See also js::ArraySetLength which initially
// enforces this requirement.
MOZ_ASSERT(reqCapacity <= as<ArrayObject>().length());
// Adding to reqCapacity must not overflow uint32_t.
MOZ_ASSERT(reqCapacity <= MAX_DENSE_ELEMENTS_COUNT);
// Then, add the header and shifted elements sizes to the new capacity
// to get the overall amount to allocate.
newAllocated = reqCapacity + numShifted + ObjectElements::VALUES_PER_HEADER;
} else {
// For arrays with writable length, and all non-Array objects, call
// `NativeObject::goodElementsAllocationAmount()` to determine the
// amount to allocate from the the requested capacity and existing length.
if (!goodElementsAllocationAmount(cx, reqCapacity + numShifted,
getElementsHeader()->length,
&newAllocated)) {
return false;
}
}
// newAllocated now contains the size of the buffer we need to allocate;
// subtract off the header and shifted elements size to get the new capacity
uint32_t newCapacity =
newAllocated - ObjectElements::VALUES_PER_HEADER - numShifted;
// If the new capacity isn't strictly greater than the old capacity, then this
// method shouldn't have been called; if the new capacity doesn't satisfy
// what was requested, then one of the calculations above must have been
// wrong.
MOZ_ASSERT(newCapacity > oldCapacity && newCapacity >= reqCapacity);
// If newCapacity exceeds MAX_DENSE_ELEMENTS_COUNT, the array should become
// sparse.
MOZ_ASSERT(newCapacity <= MAX_DENSE_ELEMENTS_COUNT);
uint32_t initlen = getDenseInitializedLength();
HeapSlot* oldHeaderSlots =
reinterpret_cast<HeapSlot*>(getUnshiftedElementsHeader());
HeapSlot* newHeaderSlots;
uint32_t oldAllocated = 0;
if (hasDynamicElements()) {
// If the object has dynamic elements, then we might be able to resize the
// buffer in-place.
// First, check that adding to oldCapacity won't overflow uint32_t
MOZ_ASSERT(oldCapacity <= MAX_DENSE_ELEMENTS_COUNT);
// Then, add the header and shifted elements sizes to get the overall size
// of the existing buffer
oldAllocated = oldCapacity + ObjectElements::VALUES_PER_HEADER + numShifted;
// Finally, try to resize the buffer.
newHeaderSlots = ReallocateObjectBuffer<HeapSlot>(
cx, this, oldHeaderSlots, oldAllocated, newAllocated);
if (!newHeaderSlots) {
return false; // If the resizing failed, then we leave elements at its
// old size.
}
} else {
// If the object has fixed elements, then we always need to allocate a new
// buffer, because if we've reached this code, then the requested capacity
// is greater than the existing inline space available within the object
newHeaderSlots = AllocateObjectBuffer<HeapSlot>(cx, this, newAllocated);
if (!newHeaderSlots) {
return false; // Leave elements at its old size.
}
// Copy the initialized elements into the new buffer,
PodCopy(newHeaderSlots, oldHeaderSlots,
ObjectElements::VALUES_PER_HEADER + initlen + numShifted);
}
// If the object already had dynamic elements, then we have to account
// for freeing the old elements buffer.
if (oldAllocated) {
RemoveCellMemory(this, oldAllocated * sizeof(HeapSlot),
MemoryUse::ObjectElements);
}
ObjectElements* newheader = reinterpret_cast<ObjectElements*>(newHeaderSlots);
// Update the elements pointer to point to the new elements buffer.
elements_ = newheader->elements() + numShifted;
// Clear the "fixed elements" flag, because if this code has been reached,
// this object now has dynamic elements.
getElementsHeader()->flags &= ~ObjectElements::FIXED;
getElementsHeader()->capacity = newCapacity;
// Poison the uninitialized portion of the new elements buffer.
Debug_SetSlotRangeToCrashOnTouch(elements_ + initlen, newCapacity - initlen);
// Account for allocating the new elements buffer.
AddCellMemory(this, newAllocated * sizeof(HeapSlot),
MemoryUse::ObjectElements);
return true;
}
void NativeObject::shrinkElements(JSContext* cx, uint32_t reqCapacity) {
MOZ_ASSERT(canHaveNonEmptyElements());
MOZ_ASSERT(reqCapacity >= getDenseInitializedLength());
if (!hasDynamicElements()) {
return;
}
// If we have shifted elements, consider moving them.
uint32_t numShifted = getElementsHeader()->numShiftedElements();
if (numShifted > 0) {
maybeMoveShiftedElements();
numShifted = getElementsHeader()->numShiftedElements();
}
uint32_t oldCapacity = getDenseCapacity();
MOZ_ASSERT(reqCapacity < oldCapacity);
uint32_t newAllocated = 0;
MOZ_ALWAYS_TRUE(goodElementsAllocationAmount(cx, reqCapacity + numShifted, 0,
&newAllocated));
MOZ_ASSERT(oldCapacity <= MAX_DENSE_ELEMENTS_COUNT);
uint32_t oldAllocated =
oldCapacity + ObjectElements::VALUES_PER_HEADER + numShifted;
if (newAllocated == oldAllocated) {
return; // Leave elements at its old size.
}
MOZ_ASSERT(newAllocated > ObjectElements::VALUES_PER_HEADER);
uint32_t newCapacity =
newAllocated - ObjectElements::VALUES_PER_HEADER - numShifted;
MOZ_ASSERT(newCapacity <= MAX_DENSE_ELEMENTS_COUNT);
HeapSlot* oldHeaderSlots =
reinterpret_cast<HeapSlot*>(getUnshiftedElementsHeader());
HeapSlot* newHeaderSlots = ReallocateObjectBuffer<HeapSlot>(
cx, this, oldHeaderSlots, oldAllocated, newAllocated);
if (!newHeaderSlots) {
cx->recoverFromOutOfMemory();
return; // Leave elements at its old size.
}
RemoveCellMemory(this, oldAllocated * sizeof(HeapSlot),
MemoryUse::ObjectElements);
ObjectElements* newheader = reinterpret_cast<ObjectElements*>(newHeaderSlots);
elements_ = newheader->elements() + numShifted;
getElementsHeader()->capacity = newCapacity;
AddCellMemory(this, newAllocated * sizeof(HeapSlot),
MemoryUse::ObjectElements);
}
void NativeObject::shrinkCapacityToInitializedLength(JSContext* cx) {
// When an array's length becomes non-writable, writes to indexes greater
// greater than or equal to the length don't change the array. We handle this
// with a check for non-writable length in most places. But in JIT code every
// check counts -- so we piggyback the check on the already-required range
// check for |index < capacity| by making capacity of arrays with non-writable
// length never exceed the length. This mechanism is also used when an object
// becomes non-extensible.
if (getElementsHeader()->numShiftedElements() > 0) {
moveShiftedElements();
}
ObjectElements* header = getElementsHeader();
uint32_t len = header->initializedLength;
MOZ_ASSERT(header->capacity >= len);
if (header->capacity == len) {
return;
}
shrinkElements(cx, len);
header = getElementsHeader();
uint32_t oldAllocated = header->numAllocatedElements();
header->capacity = len;
// The size of the memory allocation hasn't changed but we lose the actual
// capacity information. Make the associated size match the updated capacity.
if (!hasFixedElements()) {
uint32_t newAllocated = header->numAllocatedElements();
RemoveCellMemory(this, oldAllocated * sizeof(HeapSlot),
MemoryUse::ObjectElements);
AddCellMemory(this, newAllocated * sizeof(HeapSlot),
MemoryUse::ObjectElements);
}
}
/* static */
bool NativeObject::allocDictionarySlot(JSContext* cx, Handle<NativeObject*> obj,
uint32_t* slotp) {
MOZ_ASSERT(obj->inDictionaryMode());
uint32_t slotSpan = obj->slotSpan();
MOZ_ASSERT(slotSpan >= JSSLOT_FREE(obj->getClass()));
// Try to pull a free slot from the slot-number free list.
DictionaryPropMap* map = obj->dictionaryShape()->propMap();
uint32_t last = map->freeList();
if (last != SHAPE_INVALID_SLOT) {
#ifdef DEBUG
MOZ_ASSERT(last < slotSpan);
uint32_t next = obj->getSlot(last).toPrivateUint32();
MOZ_ASSERT_IF(next != SHAPE_INVALID_SLOT, next < slotSpan);
#endif
*slotp = last;
const Value& vref = obj->getSlot(last);
map->setFreeList(vref.toPrivateUint32());
obj->setSlot(last, UndefinedValue());
return true;
}
if (MOZ_UNLIKELY(slotSpan >= SHAPE_MAXIMUM_SLOT)) {
ReportOutOfMemory(cx);
return false;
}
*slotp = slotSpan;
uint32_t numFixed = obj->numFixedSlots();
if (slotSpan < numFixed) {
obj->initFixedSlot(slotSpan, UndefinedValue());
obj->setDictionaryModeSlotSpan(slotSpan + 1);
return true;
}
uint32_t dynamicSlotIndex = slotSpan - numFixed;
if (dynamicSlotIndex >= obj->numDynamicSlots()) {
if (MOZ_UNLIKELY(!obj->growSlotsForNewSlot(cx, numFixed, slotSpan))) {
return false;
}
}
obj->initDynamicSlot(numFixed, slotSpan, UndefinedValue());
obj->setDictionaryModeSlotSpan(slotSpan + 1);
return true;
}
void NativeObject::freeDictionarySlot(uint32_t slot) {
MOZ_ASSERT(inDictionaryMode());
MOZ_ASSERT(slot < slotSpan());
DictionaryPropMap* map = dictionaryShape()->propMap();
uint32_t last = map->freeList();
// Can't afford to check the whole free list, but let's check the head.
MOZ_ASSERT_IF(last != SHAPE_INVALID_SLOT, last < slotSpan() && last != slot);
// dictionary's free list.
if (JSSLOT_FREE(getClass()) <= slot) {
MOZ_ASSERT_IF(last != SHAPE_INVALID_SLOT, last < slotSpan());
setSlot(slot, PrivateUint32Value(last));
map->setFreeList(slot);
} else {
setSlot(slot, UndefinedValue());
}
}
template <AllowGC allowGC>
bool js::NativeLookupOwnProperty(
JSContext* cx, typename MaybeRooted<NativeObject*, allowGC>::HandleType obj,
typename MaybeRooted<jsid, allowGC>::HandleType id, PropertyResult* propp) {
return NativeLookupOwnPropertyInline<allowGC>(cx, obj, id, propp);
}
template bool js::NativeLookupOwnProperty<CanGC>(JSContext* cx,
Handle<NativeObject*> obj,
HandleId id,
PropertyResult* propp);
template bool js::NativeLookupOwnProperty<NoGC>(JSContext* cx,
NativeObject* const& obj,
const jsid& id,
PropertyResult* propp);
/*** [[DefineOwnProperty]] **************************************************/
static bool CallJSAddPropertyOp(JSContext* cx, JSAddPropertyOp op,
HandleObject obj, HandleId id, HandleValue v) {
AutoCheckRecursionLimit recursion(cx);
if (!recursion.check(cx)) {
return false;
}
cx->check(obj, id, v);
return op(cx, obj, id, v);
}
static MOZ_ALWAYS_INLINE bool CallAddPropertyHook(JSContext* cx,
Handle<NativeObject*> obj,
HandleId id,
HandleValue value) {
JSAddPropertyOp addProperty = obj->getClass()->getAddProperty();
if (MOZ_UNLIKELY(addProperty)) {
MOZ_ASSERT(!cx->isHelperThreadContext());
if (!CallJSAddPropertyOp(cx, addProperty, obj, id, value)) {
NativeObject::removeProperty(cx, obj, id);
return false;
}
}
return true;
}
static MOZ_ALWAYS_INLINE bool CallAddPropertyHookDense(
JSContext* cx, Handle<NativeObject*> obj, uint32_t index,
HandleValue value) {
// Inline addProperty for array objects.
if (obj->is<ArrayObject>()) {
ArrayObject* arr = &obj->as<ArrayObject>();
uint32_t length = arr->length();
if (index >= length) {
arr->setLength(index + 1);
}
return true;
}
JSAddPropertyOp addProperty = obj->getClass()->getAddProperty();
if (MOZ_UNLIKELY(addProperty)) {
MOZ_ASSERT(!cx->isHelperThreadContext());
RootedId id(cx, PropertyKey::Int(index));
if (!CallJSAddPropertyOp(cx, addProperty, obj, id, value)) {
obj->setDenseElementHole(index);
return false;
}
}
return true;
}
/**
* Determines whether a write to the given element on |arr| should fail
* because |arr| has a non-writable length, and writing that element would
* increase the length of the array.
*/
static bool WouldDefinePastNonwritableLength(ArrayObject* arr, uint32_t index) {
return !arr->lengthIsWritable() && index >= arr->length();
}
static bool ChangeProperty(JSContext* cx, Handle<NativeObject*> obj,
HandleId id, HandleObject getter,
HandleObject setter, PropertyFlags flags,
PropertyResult* existing, uint32_t* slotOut) {
MOZ_ASSERT(existing);
Rooted<GetterSetter*> gs(cx);
// If we're redefining a getter/setter property but the getter and setter
// objects are still the same, use the existing GetterSetter.
if (existing->isNativeProperty()) {
PropertyInfo prop = existing->propertyInfo();
if (prop.isAccessorProperty()) {
GetterSetter* current = obj->getGetterSetter(prop);
if (current->getter() == getter && current->setter() == setter) {
gs = current;
}
}
}
if (!gs) {
gs = GetterSetter::create(cx, getter, setter);
if (!gs) {
return false;
}
}
if (existing->isNativeProperty()) {
if (!NativeObject::changeProperty(cx, obj, id, flags, slotOut)) {
return false;
}
} else {
if (!NativeObject::addProperty(cx, obj, id, flags, slotOut)) {
return false;
}
}
obj->setSlot(*slotOut, PrivateGCThingValue(gs));
return true;
}
static PropertyFlags ComputePropertyFlags(const PropertyDescriptor& desc) {
desc.assertComplete();
PropertyFlags flags;
flags.setFlag(PropertyFlag::Configurable, desc.configurable());
flags.setFlag(PropertyFlag::Enumerable, desc.enumerable());
if (desc.isDataDescriptor()) {
flags.setFlag(PropertyFlag::Writable, desc.writable());
} else {
MOZ_ASSERT(desc.isAccessorDescriptor());
flags.setFlag(PropertyFlag::AccessorProperty);
}
return flags;
}
// Whether we're adding a new property or changing an existing property (this
// can be either a property stored in the shape tree or a dense element).
enum class IsAddOrChange { Add, Change };
template <IsAddOrChange AddOrChange>
static MOZ_ALWAYS_INLINE bool AddOrChangeProperty(
JSContext* cx, Handle<NativeObject*> obj, HandleId id,
Handle<PropertyDescriptor> desc, PropertyResult* existing = nullptr) {
desc.assertComplete();
#ifdef DEBUG
if constexpr (AddOrChange == IsAddOrChange::Add) {
MOZ_ASSERT(existing == nullptr);
MOZ_ASSERT(!obj->containsPure(id));
} else {
static_assert(AddOrChange == IsAddOrChange::Change);
MOZ_ASSERT(existing);
MOZ_ASSERT(existing->isNativeProperty() || existing->isDenseElement());
}
#endif
// Use dense storage for indexed properties where possible: when we have an
// integer key with default property attributes and are either adding a new
// property or changing a dense element.
PropertyFlags flags = ComputePropertyFlags(desc);
if (id.isInt() && flags == PropertyFlags::defaultDataPropFlags &&
(AddOrChange == IsAddOrChange::Add || existing->isDenseElement())) {
MOZ_ASSERT(!desc.isAccessorDescriptor());
MOZ_ASSERT(!obj->is<TypedArrayObject>());
uint32_t index = id.toInt();
DenseElementResult edResult = obj->ensureDenseElements(cx, index, 1);
if (edResult == DenseElementResult::Failure) {
return false;
}
if (edResult == DenseElementResult::Success) {
obj->setDenseElement(index, desc.value());
if (!CallAddPropertyHookDense(cx, obj, index, desc.value())) {
return false;
}
return true;
}
}
uint32_t slot;
if constexpr (AddOrChange == IsAddOrChange::Add) {
if (desc.isAccessorDescriptor()) {
Rooted<GetterSetter*> gs(
cx, GetterSetter::create(cx, desc.getter(), desc.setter()));
if (!gs) {
return false;
}
if (!NativeObject::addProperty(cx, obj, id, flags, &slot)) {
return false;
}
obj->initSlot(slot, PrivateGCThingValue(gs));
} else {
if (!NativeObject::addProperty(cx, obj, id, flags, &slot)) {
return false;
}
obj->initSlot(slot, desc.value());
}
} else {
if (desc.isAccessorDescriptor()) {
if (!ChangeProperty(cx, obj, id, desc.getter(), desc.setter(), flags,
existing, &slot)) {
return false;
}
} else {
if (existing->isNativeProperty()) {
if (!NativeObject::changeProperty(cx, obj, id, flags, &slot)) {
return false;
}
} else {
if (!NativeObject::addProperty(cx, obj, id, flags, &slot)) {
return false;
}
}
obj->setSlot(slot, desc.value());
}
}
MOZ_ASSERT(slot < obj->slotSpan());
// Clear any existing dense index after adding a sparse indexed property,
// and investigate converting the object to dense indexes.
if (id.isInt()) {
uint32_t index = id.toInt();
if constexpr (AddOrChange == IsAddOrChange::Add) {
MOZ_ASSERT(!obj->containsDenseElement(index));
} else {
obj->removeDenseElementForSparseIndex(index);
}
// Only try to densify sparse elements if the property we just added/changed
// is in the last slot. This avoids a perf cliff in pathological cases: in
// maybeDensifySparseElements we densify if the slot span is a power-of-two,
// but if we get slots from the free list, the slot span will stay the same
// until the free list is empty. This means we'd get quadratic behavior by
if (slot == obj->slotSpan() - 1) {
DenseElementResult edResult =
NativeObject::maybeDensifySparseElements(cx, obj);
if (edResult == DenseElementResult::Failure) {
return false;
}
if (edResult == DenseElementResult::Success) {
MOZ_ASSERT(!desc.isAccessorDescriptor());
return CallAddPropertyHookDense(cx, obj, index, desc.value());
}
}
}
if (desc.isDataDescriptor()) {
return CallAddPropertyHook(cx, obj, id, desc.value());
}
return CallAddPropertyHook(cx, obj, id, UndefinedHandleValue);
}
// Versions of AddOrChangeProperty optimized for adding a plain data property.
// This function doesn't handle integer ids as we may have to store them in
// dense elements.
static MOZ_ALWAYS_INLINE bool AddDataProperty(JSContext* cx,
Handle<NativeObject*> obj,
HandleId id, HandleValue v) {
MOZ_ASSERT(!id.isInt());
uint32_t slot;
if (!NativeObject::addProperty(cx, obj, id,
PropertyFlags::defaultDataPropFlags, &slot)) {
return false;
}
obj->initSlot(slot, v);
return CallAddPropertyHook(cx, obj, id, v);
}
bool js::AddSlotAndCallAddPropHook(JSContext* cx, Handle<NativeObject*> obj,
HandleValue v, Handle<Shape*> newShape) {
MOZ_ASSERT(obj->getClass()->getAddProperty());
MOZ_ASSERT(newShape->asShared().lastProperty().isDataProperty());
RootedId id(cx, newShape->asShared().lastProperty().key());
MOZ_ASSERT(!id.isInt());
uint32_t slot = newShape->asShared().lastProperty().slot();
if (!obj->setShapeAndAddNewSlot(cx, &newShape->asShared(), slot)) {
return false;
}
obj->initSlot(slot, v);
return CallAddPropertyHook(cx, obj, id, v);
}
static bool IsAccessorDescriptor(const PropertyResult& prop) {
if (prop.isNativeProperty()) {
return prop.propertyInfo().isAccessorProperty();
}
MOZ_ASSERT(prop.isDenseElement() || prop.isTypedArrayElement());
return false;
}
static bool IsDataDescriptor(const PropertyResult& prop) {
return !IsAccessorDescriptor(prop);
}
static bool GetCustomDataProperty(JSContext* cx, HandleObject obj, HandleId id,
MutableHandleValue vp);
static bool GetExistingDataProperty(JSContext* cx, Handle<NativeObject*> obj,
HandleId id, const PropertyResult& prop,
MutableHandleValue vp) {
if (prop.isDenseElement()) {
vp.set(obj->getDenseElement(prop.denseElementIndex()));
return true;
}
if (prop.isTypedArrayElement()) {
size_t idx = prop.typedArrayElementIndex();
return obj->as<TypedArrayObject>().getElement<CanGC>(cx, idx, vp);
}
PropertyInfo propInfo = prop.propertyInfo();
if (propInfo.isDataProperty()) {
vp.set(obj->getSlot(propInfo.slot()));
return true;
}
MOZ_ASSERT(!cx->isHelperThreadContext());
MOZ_RELEASE_ASSERT(propInfo.isCustomDataProperty());
return GetCustomDataProperty(cx, obj, id, vp);
}
/*
* If desc is redundant with an existing own property obj[id], then set
* |*redundant = true| and return true.
*/
static bool DefinePropertyIsRedundant(JSContext* cx, Handle<NativeObject*> obj,
HandleId id, const PropertyResult& prop,
JS::PropertyAttributes attrs,
Handle<PropertyDescriptor> desc,
bool* redundant) {
*redundant = false;
if (desc.hasConfigurable() && desc.configurable() != attrs.configurable()) {
return true;
}
if (desc.hasEnumerable() && desc.enumerable() != attrs.enumerable()) {
return true;
}
if (desc.isDataDescriptor()) {
if (IsAccessorDescriptor(prop)) {
return true;
}
if (desc.hasWritable() && desc.writable() != attrs.writable()) {
return true;
}
if (desc.hasValue()) {
// Get the current value of the existing property.
RootedValue currentValue(cx);
if (!GetExistingDataProperty(cx, obj, id, prop, ¤tValue)) {
return false;
}
// Don't call SameValue here to ensure we properly update distinct
// NaN values.
if (desc.value() != currentValue) {
return true;
}
}
// Check for custom data properties for ArrayObject/ArgumentsObject.
// PropertyDescriptor can't represent these properties so they're never
// redundant.
if (prop.isNativeProperty() && prop.propertyInfo().isCustomDataProperty()) {
return true;
}
} else if (desc.isAccessorDescriptor()) {
if (!prop.isNativeProperty()) {
return true;
}
PropertyInfo propInfo = prop.propertyInfo();
if (desc.hasGetter() && (!propInfo.isAccessorProperty() ||
desc.getter() != obj->getGetter(propInfo))) {
return true;
}
if (desc.hasSetter() && (!propInfo.isAccessorProperty() ||
desc.setter() != obj->getSetter(propInfo))) {
return true;
}
}
*redundant = true;
return true;
}
bool js::NativeDefineProperty(JSContext* cx, Handle<NativeObject*> obj,
HandleId id, Handle<PropertyDescriptor> desc_,
ObjectOpResult& result) {
desc_.assertValid();
// Section numbers and step numbers below refer to ES2018, draft rev
// 540b827fccf6122a984be99ab9af7be20e3b5562.
//
// This function aims to implement 9.1.6 [[DefineOwnProperty]] as well as
// the [[DefineOwnProperty]] methods described in 9.4.2.1 (arrays), 9.4.4.2
// (arguments), and 9.4.5.3 (typed array views).
// Dispense with custom behavior of exotic native objects first.
if (obj->is<ArrayObject>()) {
// 9.4.2.1 step 2. Redefining an array's length is very special.
Rooted<ArrayObject*> arr(cx, &obj->as<ArrayObject>());
if (id == NameToId(cx->names().length)) {
// 9.1.6.3 ValidateAndApplyPropertyDescriptor, step 7.a.
if (desc_.isAccessorDescriptor()) {
return result.fail(JSMSG_CANT_REDEFINE_PROP);
}
MOZ_ASSERT(!cx->isHelperThreadContext());
return ArraySetLength(cx, arr, id, desc_, result);
}
// 9.4.2.1 step 3. Don't extend a fixed-length array.
uint32_t index;
if (IdIsIndex(id, &index)) {
if (WouldDefinePastNonwritableLength(arr, index)) {
return result.fail(JSMSG_CANT_DEFINE_PAST_ARRAY_LENGTH);
}
}
} else if (obj->is<TypedArrayObject>()) {
// 9.4.5.3 step 3. Indexed properties of typed arrays are special.
if (mozilla::Maybe<uint64_t> index = ToTypedArrayIndex(id)) {
MOZ_ASSERT(!cx->isHelperThreadContext());
Rooted<TypedArrayObject*> tobj(cx, &obj->as<TypedArrayObject>());
return DefineTypedArrayElement(cx, tobj, index.value(), desc_, result);
}
} else if (obj->is<ArgumentsObject>()) {
Rooted<ArgumentsObject*> argsobj(cx, &obj->as<ArgumentsObject>());
if (id.isAtom(cx->names().length)) {
// Either we are resolving the .length property on this object,
// or redefining it. In the latter case only, we must reify the
// property.
if (!desc_.resolving()) {
if (!ArgumentsObject::reifyLength(cx, argsobj)) {
return false;
}
}
} else if (id.isAtom(cx->names().callee) &&
argsobj->is<MappedArgumentsObject>()) {
// Do same thing as .length for .callee on MappedArgumentsObject.