Commit de672226 authored by ulan@chromium.org's avatar ulan@chromium.org

Clear old backing store of WeakCollection on updates.

Not clearing can lead to a crash under following conditions:
1. Backing store of a weak map is allocated in large object space.
2. The backing store is marked incrementaly via the weak map.
3. The weak map is updated and gets a new backing store.
4. The store buffer overflows and marks the chunk of the old backing store as
"scan on scavenge."
5. Mark-compact collection kills some elements of the weak map. Note that the
old backing store survives because it was marked incrementally, but its dead
elements are not cleared.
6. Scavenger iterates over the old backing store, tries to move a dead object
and crashes.

BUG=v8:3631
LOG=N
TEST=cctest/test-heap/Regress3631
R=jkummerow@chromium.org

Review URL: https://codereview.chromium.org/686783003

Cr-Commit-Position: refs/heads/master@{#25032}
git-svn-id: https://v8.googlecode.com/svn/branches/bleeding_edge@25032 ce2b1a6d-e550-0410-aec6-3dcde31c8c00
parent 1bb79539
...@@ -288,6 +288,10 @@ RUNTIME_FUNCTION(Runtime_WeakCollectionDelete) { ...@@ -288,6 +288,10 @@ RUNTIME_FUNCTION(Runtime_WeakCollectionDelete) {
Handle<ObjectHashTable> new_table = Handle<ObjectHashTable> new_table =
ObjectHashTable::Remove(table, key, &was_present); ObjectHashTable::Remove(table, key, &was_present);
weak_collection->set_table(*new_table); weak_collection->set_table(*new_table);
if (*table != *new_table) {
// Zap the old table since we didn't record slots for its elements.
table->FillWithHoles(0, table->length());
}
return isolate->heap()->ToBoolean(was_present); return isolate->heap()->ToBoolean(was_present);
} }
...@@ -304,6 +308,10 @@ RUNTIME_FUNCTION(Runtime_WeakCollectionSet) { ...@@ -304,6 +308,10 @@ RUNTIME_FUNCTION(Runtime_WeakCollectionSet) {
RUNTIME_ASSERT(table->IsKey(*key)); RUNTIME_ASSERT(table->IsKey(*key));
Handle<ObjectHashTable> new_table = ObjectHashTable::Put(table, key, value); Handle<ObjectHashTable> new_table = ObjectHashTable::Put(table, key, value);
weak_collection->set_table(*new_table); weak_collection->set_table(*new_table);
if (*table != *new_table) {
// Zap the old table since we didn't record slots for its elements.
table->FillWithHoles(0, table->length());
}
return *weak_collection; return *weak_collection;
} }
......
...@@ -4608,6 +4608,46 @@ TEST(Regress388880) { ...@@ -4608,6 +4608,46 @@ TEST(Regress388880) {
} }
TEST(Regress3631) {
i::FLAG_expose_gc = true;
CcTest::InitializeVM();
v8::HandleScope scope(CcTest::isolate());
Isolate* isolate = CcTest::i_isolate();
Heap* heap = isolate->heap();
IncrementalMarking* marking = CcTest::heap()->incremental_marking();
v8::Local<v8::Value> result = CompileRun(
"var weak_map = new WeakMap();"
"var future_keys = [];"
"for (var i = 0; i < 50; i++) {"
" var key = {'k' : i + 0.1};"
" weak_map.set(key, 1);"
" future_keys.push({'x' : i + 0.2});"
"}"
"weak_map");
if (marking->IsStopped()) {
marking->Start();
}
// Incrementally mark the backing store.
Handle<JSObject> obj =
v8::Utils::OpenHandle(*v8::Handle<v8::Object>::Cast(result));
Handle<JSWeakCollection> weak_map(reinterpret_cast<JSWeakCollection*>(*obj));
while (!Marking::IsBlack(
Marking::MarkBitFrom(HeapObject::cast(weak_map->table()))) &&
!marking->IsStopped()) {
marking->Step(MB, IncrementalMarking::NO_GC_VIA_STACK_GUARD);
}
// Stash the backing store in a handle.
Handle<Object> save(weak_map->table(), isolate);
// The following line will update the backing store.
CompileRun(
"for (var i = 0; i < 50; i++) {"
" weak_map.set(future_keys[i], i);"
"}");
heap->incremental_marking()->set_should_hurry(true);
heap->CollectGarbage(OLD_POINTER_SPACE);
}
#ifdef DEBUG #ifdef DEBUG
TEST(PathTracer) { TEST(PathTracer) {
CcTest::InitializeVM(); CcTest::InitializeVM();
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment