OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
11 // with the distribution. | 11 // with the distribution. |
(...skipping 44 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
56 // MarkCompactCollector | 56 // MarkCompactCollector |
57 | 57 |
58 MarkCompactCollector::MarkCompactCollector() : // NOLINT | 58 MarkCompactCollector::MarkCompactCollector() : // NOLINT |
59 #ifdef DEBUG | 59 #ifdef DEBUG |
60 state_(IDLE), | 60 state_(IDLE), |
61 #endif | 61 #endif |
62 sweep_precisely_(false), | 62 sweep_precisely_(false), |
63 compacting_(false), | 63 compacting_(false), |
64 was_marked_incrementally_(false), | 64 was_marked_incrementally_(false), |
65 collect_maps_(FLAG_collect_maps), | 65 collect_maps_(FLAG_collect_maps), |
| 66 flush_monomorphic_ics_(false), |
66 tracer_(NULL), | 67 tracer_(NULL), |
67 migration_slots_buffer_(NULL), | 68 migration_slots_buffer_(NULL), |
68 heap_(NULL), | 69 heap_(NULL), |
69 code_flusher_(NULL), | 70 code_flusher_(NULL), |
70 encountered_weak_maps_(NULL) { } | 71 encountered_weak_maps_(NULL) { } |
71 | 72 |
72 | 73 |
73 #ifdef DEBUG | 74 #ifdef DEBUG |
74 class VerifyMarkingVisitor: public ObjectVisitor { | 75 class VerifyMarkingVisitor: public ObjectVisitor { |
75 public: | 76 public: |
(...skipping 432 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
508 | 509 |
509 | 510 |
510 void MarkCompactCollector::Prepare(GCTracer* tracer) { | 511 void MarkCompactCollector::Prepare(GCTracer* tracer) { |
511 was_marked_incrementally_ = heap()->incremental_marking()->IsMarking(); | 512 was_marked_incrementally_ = heap()->incremental_marking()->IsMarking(); |
512 | 513 |
513 // Disable collection of maps if incremental marking is enabled. | 514 // Disable collection of maps if incremental marking is enabled. |
514 // Map collection algorithm relies on a special map transition tree traversal | 515 // Map collection algorithm relies on a special map transition tree traversal |
515 // order which is not implemented for incremental marking. | 516 // order which is not implemented for incremental marking. |
516 collect_maps_ = FLAG_collect_maps && !was_marked_incrementally_; | 517 collect_maps_ = FLAG_collect_maps && !was_marked_incrementally_; |
517 | 518 |
| 519 // Monomorphic ICs are preserved when possible, but need to be flushed |
| 520 // when they might be keeping a Context alive, or when the heap is about |
| 521 // to be serialized. |
| 522 flush_monomorphic_ics_ = |
| 523 heap()->isolate()->context_exit_happened() || Serializer::enabled(); |
| 524 |
518 // Rather than passing the tracer around we stash it in a static member | 525 // Rather than passing the tracer around we stash it in a static member |
519 // variable. | 526 // variable. |
520 tracer_ = tracer; | 527 tracer_ = tracer; |
521 | 528 |
522 #ifdef DEBUG | 529 #ifdef DEBUG |
523 ASSERT(state_ == IDLE); | 530 ASSERT(state_ == IDLE); |
524 state_ = PREPARE_GC; | 531 state_ = PREPARE_GC; |
525 #endif | 532 #endif |
526 | 533 |
527 ASSERT(!FLAG_never_compact || !FLAG_always_compact); | 534 ASSERT(!FLAG_never_compact || !FLAG_always_compact); |
(...skipping 346 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
874 // that there can be no such embedded pointers and add assertion here. | 881 // that there can be no such embedded pointers and add assertion here. |
875 HeapObject* object = HeapObject::cast(rinfo->target_object()); | 882 HeapObject* object = HeapObject::cast(rinfo->target_object()); |
876 heap->mark_compact_collector()->RecordRelocSlot(rinfo, object); | 883 heap->mark_compact_collector()->RecordRelocSlot(rinfo, object); |
877 MarkBit mark = Marking::MarkBitFrom(object); | 884 MarkBit mark = Marking::MarkBitFrom(object); |
878 heap->mark_compact_collector()->MarkObject(object, mark); | 885 heap->mark_compact_collector()->MarkObject(object, mark); |
879 } | 886 } |
880 | 887 |
881 static inline void VisitCodeTarget(Heap* heap, RelocInfo* rinfo) { | 888 static inline void VisitCodeTarget(Heap* heap, RelocInfo* rinfo) { |
882 ASSERT(RelocInfo::IsCodeTarget(rinfo->rmode())); | 889 ASSERT(RelocInfo::IsCodeTarget(rinfo->rmode())); |
883 Code* target = Code::GetCodeFromTargetAddress(rinfo->target_address()); | 890 Code* target = Code::GetCodeFromTargetAddress(rinfo->target_address()); |
884 if (FLAG_cleanup_code_caches_at_gc && target->is_inline_cache_stub()) { | 891 if (FLAG_cleanup_code_caches_at_gc && target->is_inline_cache_stub() |
| 892 && (target->ic_state() == MEGAMORPHIC || |
| 893 heap->mark_compact_collector()->flush_monomorphic_ics_)) { |
885 IC::Clear(rinfo->pc()); | 894 IC::Clear(rinfo->pc()); |
886 target = Code::GetCodeFromTargetAddress(rinfo->target_address()); | 895 target = Code::GetCodeFromTargetAddress(rinfo->target_address()); |
887 } else { | 896 } else { |
888 if (FLAG_cleanup_code_caches_at_gc && | 897 if (FLAG_cleanup_code_caches_at_gc && |
889 target->kind() == Code::STUB && | 898 target->kind() == Code::STUB && |
890 target->major_key() == CodeStub::CallFunction && | 899 target->major_key() == CodeStub::CallFunction && |
891 target->has_function_cache()) { | 900 target->has_function_cache()) { |
892 CallFunctionStub::Clear(heap, rinfo->pc()); | 901 CallFunctionStub::Clear(heap, rinfo->pc()); |
893 } | 902 } |
894 } | 903 } |
(...skipping 3012 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3907 while (buffer != NULL) { | 3916 while (buffer != NULL) { |
3908 SlotsBuffer* next_buffer = buffer->next(); | 3917 SlotsBuffer* next_buffer = buffer->next(); |
3909 DeallocateBuffer(buffer); | 3918 DeallocateBuffer(buffer); |
3910 buffer = next_buffer; | 3919 buffer = next_buffer; |
3911 } | 3920 } |
3912 *buffer_address = NULL; | 3921 *buffer_address = NULL; |
3913 } | 3922 } |
3914 | 3923 |
3915 | 3924 |
3916 } } // namespace v8::internal | 3925 } } // namespace v8::internal |
OLD | NEW |