OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 854 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
865 | 865 |
866 mark_compact_collector_.CollectGarbage(); | 866 mark_compact_collector_.CollectGarbage(); |
867 | 867 |
868 LOG(isolate_, ResourceEvent("markcompact", "end")); | 868 LOG(isolate_, ResourceEvent("markcompact", "end")); |
869 | 869 |
870 gc_state_ = NOT_IN_GC; | 870 gc_state_ = NOT_IN_GC; |
871 | 871 |
872 isolate_->counters()->objs_since_last_full()->Set(0); | 872 isolate_->counters()->objs_since_last_full()->Set(0); |
873 | 873 |
874 contexts_disposed_ = 0; | 874 contexts_disposed_ = 0; |
| 875 |
| 876 isolate_->set_context_exit_happened(false); |
875 } | 877 } |
876 | 878 |
877 | 879 |
878 void Heap::MarkCompactPrologue() { | 880 void Heap::MarkCompactPrologue() { |
879 // At any old GC clear the keyed lookup cache to enable collection of unused | 881 // At any old GC clear the keyed lookup cache to enable collection of unused |
880 // maps. | 882 // maps. |
881 isolate_->keyed_lookup_cache()->Clear(); | 883 isolate_->keyed_lookup_cache()->Clear(); |
882 isolate_->context_slot_cache()->Clear(); | 884 isolate_->context_slot_cache()->Clear(); |
883 isolate_->descriptor_lookup_cache()->Clear(); | 885 isolate_->descriptor_lookup_cache()->Clear(); |
884 StringSplitCache::Clear(string_split_cache()); | 886 StringSplitCache::Clear(string_split_cache()); |
(...skipping 5769 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6654 isolate_->heap()->store_buffer()->Compact(); | 6656 isolate_->heap()->store_buffer()->Compact(); |
6655 isolate_->heap()->store_buffer()->Filter(MemoryChunk::ABOUT_TO_BE_FREED); | 6657 isolate_->heap()->store_buffer()->Filter(MemoryChunk::ABOUT_TO_BE_FREED); |
6656 for (chunk = chunks_queued_for_free_; chunk != NULL; chunk = next) { | 6658 for (chunk = chunks_queued_for_free_; chunk != NULL; chunk = next) { |
6657 next = chunk->next_chunk(); | 6659 next = chunk->next_chunk(); |
6658 isolate_->memory_allocator()->Free(chunk); | 6660 isolate_->memory_allocator()->Free(chunk); |
6659 } | 6661 } |
6660 chunks_queued_for_free_ = NULL; | 6662 chunks_queued_for_free_ = NULL; |
6661 } | 6663 } |
6662 | 6664 |
6663 } } // namespace v8::internal | 6665 } } // namespace v8::internal |
OLD | NEW |