OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 193 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
204 | 204 |
205 | 205 |
206 MaybeObject* Heap::CopyFixedDoubleArray(FixedDoubleArray* src) { | 206 MaybeObject* Heap::CopyFixedDoubleArray(FixedDoubleArray* src) { |
207 return CopyFixedDoubleArrayWithMap(src, src->map()); | 207 return CopyFixedDoubleArrayWithMap(src, src->map()); |
208 } | 208 } |
209 | 209 |
210 | 210 |
211 MaybeObject* Heap::AllocateRaw(int size_in_bytes, | 211 MaybeObject* Heap::AllocateRaw(int size_in_bytes, |
212 AllocationSpace space, | 212 AllocationSpace space, |
213 AllocationSpace retry_space) { | 213 AllocationSpace retry_space) { |
| 214 SLOW_ASSERT(!isolate_->optimizing_compiler_thread()->IsOptimizerThread()); |
214 ASSERT(allocation_allowed_ && gc_state_ == NOT_IN_GC); | 215 ASSERT(allocation_allowed_ && gc_state_ == NOT_IN_GC); |
215 ASSERT(space != NEW_SPACE || | 216 ASSERT(space != NEW_SPACE || |
216 retry_space == OLD_POINTER_SPACE || | 217 retry_space == OLD_POINTER_SPACE || |
217 retry_space == OLD_DATA_SPACE || | 218 retry_space == OLD_DATA_SPACE || |
218 retry_space == LO_SPACE); | 219 retry_space == LO_SPACE); |
219 #ifdef DEBUG | 220 #ifdef DEBUG |
220 if (FLAG_gc_interval >= 0 && | 221 if (FLAG_gc_interval >= 0 && |
221 !disallow_allocation_failure_ && | 222 !disallow_allocation_failure_ && |
222 Heap::allocation_timeout_-- <= 0) { | 223 Heap::allocation_timeout_-- <= 0) { |
223 return Failure::RetryAfterGC(space); | 224 return Failure::RetryAfterGC(space); |
(...skipping 672 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
896 AssertNoAllocation::~AssertNoAllocation() { } | 897 AssertNoAllocation::~AssertNoAllocation() { } |
897 DisableAssertNoAllocation::DisableAssertNoAllocation() { } | 898 DisableAssertNoAllocation::DisableAssertNoAllocation() { } |
898 DisableAssertNoAllocation::~DisableAssertNoAllocation() { } | 899 DisableAssertNoAllocation::~DisableAssertNoAllocation() { } |
899 | 900 |
900 #endif | 901 #endif |
901 | 902 |
902 | 903 |
903 } } // namespace v8::internal | 904 } } // namespace v8::internal |
904 | 905 |
905 #endif // V8_HEAP_INL_H_ | 906 #endif // V8_HEAP_INL_H_ |
OLD | NEW |