OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 54 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
65 // Number of times a function has to be seen on the stack before it is | 65 // Number of times a function has to be seen on the stack before it is |
66 // optimized. | 66 // optimized. |
67 static const int kProfilerTicksBeforeOptimization = 2; | 67 static const int kProfilerTicksBeforeOptimization = 2; |
68 | 68 |
69 // Maximum size in bytes of generated code for a function to be optimized | 69 // Maximum size in bytes of generated code for a function to be optimized |
70 // the very first time it is seen on the stack. | 70 // the very first time it is seen on the stack. |
71 static const int kMaxSizeEarlyOpt = 500; | 71 static const int kMaxSizeEarlyOpt = 500; |
72 | 72 |
73 | 73 |
74 Atomic32 RuntimeProfiler::state_ = 0; | 74 Atomic32 RuntimeProfiler::state_ = 0; |
75 | 75 // TODO(isolates): Create the semaphore lazily and clean it up when no |
76 // TODO(isolates): Clean up the semaphore when it is no longer required. | 76 // longer required. |
77 static LazySemaphore<0>::type semaphore = LAZY_SEMAPHORE_INITIALIZER; | 77 Semaphore* RuntimeProfiler::semaphore_ = OS::CreateSemaphore(0); |
78 | 78 |
79 #ifdef DEBUG | 79 #ifdef DEBUG |
80 bool RuntimeProfiler::has_been_globally_set_up_ = false; | 80 bool RuntimeProfiler::has_been_globally_set_up_ = false; |
81 #endif | 81 #endif |
82 bool RuntimeProfiler::enabled_ = false; | 82 bool RuntimeProfiler::enabled_ = false; |
83 | 83 |
84 | 84 |
85 RuntimeProfiler::RuntimeProfiler(Isolate* isolate) | 85 RuntimeProfiler::RuntimeProfiler(Isolate* isolate) |
86 : isolate_(isolate), | 86 : isolate_(isolate), |
87 sampler_threshold_(kSamplerThresholdInit), | 87 sampler_threshold_(kSamplerThresholdInit), |
(...skipping 317 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
405 } | 405 } |
406 | 406 |
407 | 407 |
408 void RuntimeProfiler::HandleWakeUp(Isolate* isolate) { | 408 void RuntimeProfiler::HandleWakeUp(Isolate* isolate) { |
409 // The profiler thread must still be waiting. | 409 // The profiler thread must still be waiting. |
410 ASSERT(NoBarrier_Load(&state_) >= 0); | 410 ASSERT(NoBarrier_Load(&state_) >= 0); |
411 // In IsolateEnteredJS we have already incremented the counter and | 411 // In IsolateEnteredJS we have already incremented the counter and |
412 // undid the decrement done by the profiler thread. Increment again | 412 // undid the decrement done by the profiler thread. Increment again |
413 // to get the right count of active isolates. | 413 // to get the right count of active isolates. |
414 NoBarrier_AtomicIncrement(&state_, 1); | 414 NoBarrier_AtomicIncrement(&state_, 1); |
415 semaphore.Pointer()->Signal(); | 415 semaphore_->Signal(); |
416 } | 416 } |
417 | 417 |
418 | 418 |
419 bool RuntimeProfiler::IsSomeIsolateInJS() { | 419 bool RuntimeProfiler::IsSomeIsolateInJS() { |
420 return NoBarrier_Load(&state_) > 0; | 420 return NoBarrier_Load(&state_) > 0; |
421 } | 421 } |
422 | 422 |
423 | 423 |
424 bool RuntimeProfiler::WaitForSomeIsolateToEnterJS() { | 424 bool RuntimeProfiler::WaitForSomeIsolateToEnterJS() { |
425 Atomic32 old_state = NoBarrier_CompareAndSwap(&state_, 0, -1); | 425 Atomic32 old_state = NoBarrier_CompareAndSwap(&state_, 0, -1); |
426 ASSERT(old_state >= -1); | 426 ASSERT(old_state >= -1); |
427 if (old_state != 0) return false; | 427 if (old_state != 0) return false; |
428 semaphore.Pointer()->Wait(); | 428 semaphore_->Wait(); |
429 return true; | 429 return true; |
430 } | 430 } |
431 | 431 |
432 | 432 |
433 void RuntimeProfiler::StopRuntimeProfilerThreadBeforeShutdown(Thread* thread) { | 433 void RuntimeProfiler::StopRuntimeProfilerThreadBeforeShutdown(Thread* thread) { |
434 // Do a fake increment. If the profiler is waiting on the semaphore, | 434 // Do a fake increment. If the profiler is waiting on the semaphore, |
435 // the returned state is 0, which can be left as an initial state in | 435 // the returned state is 0, which can be left as an initial state in |
436 // case profiling is restarted later. If the profiler is not | 436 // case profiling is restarted later. If the profiler is not |
437 // waiting, the increment will prevent it from waiting, but has to | 437 // waiting, the increment will prevent it from waiting, but has to |
438 // be undone after the profiler is stopped. | 438 // be undone after the profiler is stopped. |
439 Atomic32 new_state = NoBarrier_AtomicIncrement(&state_, 1); | 439 Atomic32 new_state = NoBarrier_AtomicIncrement(&state_, 1); |
440 ASSERT(new_state >= 0); | 440 ASSERT(new_state >= 0); |
441 if (new_state == 0) { | 441 if (new_state == 0) { |
442 // The profiler thread is waiting. Wake it up. It must check for | 442 // The profiler thread is waiting. Wake it up. It must check for |
443 // stop conditions before attempting to wait again. | 443 // stop conditions before attempting to wait again. |
444 semaphore.Pointer()->Signal(); | 444 semaphore_->Signal(); |
445 } | 445 } |
446 thread->Join(); | 446 thread->Join(); |
447 // The profiler thread is now stopped. Undo the increment in case it | 447 // The profiler thread is now stopped. Undo the increment in case it |
448 // was not waiting. | 448 // was not waiting. |
449 if (new_state != 0) { | 449 if (new_state != 0) { |
450 NoBarrier_AtomicIncrement(&state_, -1); | 450 NoBarrier_AtomicIncrement(&state_, -1); |
451 } | 451 } |
452 } | 452 } |
453 | 453 |
454 | 454 |
(...skipping 17 matching lines...) Expand all Loading... |
472 | 472 |
473 bool RuntimeProfilerRateLimiter::SuspendIfNecessary() { | 473 bool RuntimeProfilerRateLimiter::SuspendIfNecessary() { |
474 if (!RuntimeProfiler::IsSomeIsolateInJS()) { | 474 if (!RuntimeProfiler::IsSomeIsolateInJS()) { |
475 return RuntimeProfiler::WaitForSomeIsolateToEnterJS(); | 475 return RuntimeProfiler::WaitForSomeIsolateToEnterJS(); |
476 } | 476 } |
477 return false; | 477 return false; |
478 } | 478 } |
479 | 479 |
480 | 480 |
481 } } // namespace v8::internal | 481 } } // namespace v8::internal |
OLD | NEW |