Index: base/message_pump_win.cc |
diff --git a/base/message_pump_win.cc b/base/message_pump_win.cc |
index fb962caaa2f45e2f83792f10dc3fe7bd147e5d23..28421714cc678993dbcb0f356c598332e17d01b5 100644 |
--- a/base/message_pump_win.cc |
+++ b/base/message_pump_win.cc |
@@ -9,8 +9,6 @@ |
#include "base/debug/trace_event.h" |
#include "base/message_loop.h" |
#include "base/metrics/histogram.h" |
-#include "base/process_util.h" |
-#include "base/win/wrapped_window_proc.h" |
namespace { |
@@ -21,19 +19,34 @@ enum MessageLoopProblems { |
MESSAGE_LOOP_PROBLEM_MAX, |
}; |
+// Possible states of the message pump: |
+// - kPumpIdle: the thread is sleeping, waiting for work to be posted. |
+// - kPumpHaveWork: a window message or completion packet has been posted. |
+// - kPumpDisabled: the pump is disabled, no more work can be posted. |
+static const LONG kPumpIdle = 0; |
+static const LONG kPumpHaveWork = 1; |
+static const LONG kPumpDisabled = 2; |
+ |
+// Used to wake up the message loop thread in the case if it is waiting for |
+// another thread to exit MessagePumpForUI::ScheduleWork(). |
+VOID CALLBACK DummyApc(ULONG_PTR) {} |
+ |
} // namespace |
namespace base { |
-static const wchar_t kWndClass[] = L"Chrome_MessagePumpWindow"; |
- |
// Message sent to get an additional time slice for pumping (processing) another |
// task (a series of such messages creates a continuous task pump). |
static const int kMsgHaveWork = WM_USER + 1; |
+// Used by MessagePumpUI to wake up the thread and check any pending timers. |
+static const int kTimerId = 1; |
+ |
//----------------------------------------------------------------------------- |
// MessagePumpWin public: |
+MessagePumpWin::MessagePumpWin() : pump_state_(kPumpIdle), state_(NULL) {} |
+ |
void MessagePumpWin::AddObserver(MessagePumpObserver* observer) { |
observers_.AddObserver(observer); |
} |
@@ -96,25 +109,31 @@ int MessagePumpWin::GetCurrentDelay() const { |
// MessagePumpForUI public: |
MessagePumpForUI::MessagePumpForUI() |
- : instance_(NULL), |
- message_filter_(new MessageFilter) { |
- InitMessageWnd(); |
+ : message_filter_(new MessageFilter), |
+ window_(new win::MessageWindow()) { |
+ CHECK(DuplicateHandle(GetCurrentProcess(), |
+ GetCurrentThread(), |
+ GetCurrentProcess(), |
+ thread_.Receive(), |
+ THREAD_SET_CONTEXT, |
+ FALSE, // no not inherit this handle |
+ 0)); |
+ CHECK(window_->Create(this)); |
} |
MessagePumpForUI::~MessagePumpForUI() { |
- DestroyWindow(message_hwnd_); |
- UnregisterClass(kWndClass, instance_); |
} |
void MessagePumpForUI::ScheduleWork() { |
- if (InterlockedExchange(&have_work_, 1)) |
- return; // Someone else continued the pumping. |
+ if (InterlockedCompareExchange(&pump_state_, kPumpHaveWork, |
+ kPumpIdle) != kPumpIdle) { |
+ // Either someone else continued the pumping or the pump is disabled. |
+ return; |
+ } |
// Make sure the MessagePump does some work for us. |
- BOOL ret = PostMessage(message_hwnd_, kMsgHaveWork, |
- reinterpret_cast<WPARAM>(this), 0); |
- if (ret) |
- return; // There was room in the Window Message queue. |
+ if (PostMessage(window_->hwnd(), kMsgHaveWork, 0, 0)) |
+ return; |
// We have failed to insert a have-work message, so there is a chance that we |
// will starve tasks/timers while sitting in a nested message loop. Nested |
@@ -124,12 +143,27 @@ void MessagePumpForUI::ScheduleWork() { |
// common (queue is full, of about 2000 messages), so we'll do a near-graceful |
// recovery. Nested loops are pretty transient (we think), so this will |
// probably be recoverable. |
- InterlockedExchange(&have_work_, 0); // Clarify that we didn't really insert. |
UMA_HISTOGRAM_ENUMERATION("Chrome.MessageLoopProblem", MESSAGE_POST_ERROR, |
MESSAGE_LOOP_PROBLEM_MAX); |
+ |
+ // Clarify that we didn't really insert. |
+ InterlockedExchange(&pump_state_, kPumpIdle); |
darin (slow to review)
2013/06/10 19:28:45
nit: Shouldn't this use InterlockedCompareExchange
alexeypa (please no reviews)
2013/06/10 20:16:12
No, at this point |pump_state_| is guaranteed to b
|
+ |
+ // Try to wake up the message loop thread by posting an APC. This might not |
darin (slow to review)
2013/06/10 19:28:45
I'm not sure I fully understand the conditions tha
alexeypa (please no reviews)
2013/06/10 20:16:12
I don't have any data on hand but two reasons that
|
+ // work while a nested loop still running (see the comment above) but this |
+ // will unblock WillDestroyCurrentMessageLoop() if it is waiting for |
+ // ScheduleWork() to complete. |
+ // |
+ // According to the UMA metrics posting an I/O completion packet has very low |
+ // error rate. Queuing an APC hits roughly the same path in the kernel so |
+ // the error rate should be low as well. Given that we do it only when |
+ // PostMessage() fails it should be safe to CHECK() here. |
+ CHECK(QueueUserAPC(&DummyApc, thread_, 0)); |
} |
void MessagePumpForUI::ScheduleDelayedWork(const TimeTicks& delayed_work_time) { |
+ DCHECK(window_->CalledOnValidThread()); |
+ |
// |
// We would *like* to provide high resolution timers. Windows timers using |
// SetTimer() have a 10ms granularity. We have to use WM_TIMER as a wakeup |
@@ -159,8 +193,7 @@ void MessagePumpForUI::ScheduleDelayedWork(const TimeTicks& delayed_work_time) { |
// Create a WM_TIMER event that will wake us up to check for any pending |
// timers (in case we are running within a nested, external sub-pump). |
- BOOL ret = SetTimer(message_hwnd_, reinterpret_cast<UINT_PTR>(this), |
- delay_msec, NULL); |
+ BOOL ret = SetTimer(window_->hwnd(), kTimerId, delay_msec, NULL); |
if (ret) |
return; |
// If we can't set timers, we are in big trouble... but cross our fingers for |
@@ -170,6 +203,31 @@ void MessagePumpForUI::ScheduleDelayedWork(const TimeTicks& delayed_work_time) { |
MESSAGE_LOOP_PROBLEM_MAX); |
} |
+void MessagePumpForUI::WillDestroyCurrentMessageLoop() { |
+ DCHECK(window_->CalledOnValidThread()); |
+ |
+ // Disable the message pump. If |pump_state_ == kPumpHaveWork| then |
+ // ScheduleWork() might be still running on a different thread. Wait until |
+ // |kMsgHaveWork| is received or |pump_state_| is reset back to |kPumpIdle|. |
+ while (InterlockedCompareExchange(&pump_state_, kPumpDisabled, |
+ kPumpIdle) == kPumpHaveWork) { |
+ MSG msg; |
+ if (PeekMessage(&msg, NULL, 0, 0, PM_REMOVE)) { |
+ if (msg.message == kMsgHaveWork && msg.hwnd == window_->hwnd()) { |
+ // Now that we received |kMsgHaveWork| the pump can be safely disabled. |
+ InterlockedExchange(&pump_state_, kPumpDisabled); |
+ break; |
+ } |
+ } |
+ |
+ // Wait until |kMsgHaveWork| is posted or an APC is received. |
+ WaitForWork(); |
+ } |
+ |
+ // At this point the pump is disabled and other threads exited ScheduleWork(). |
+ window_.reset(); |
+} |
+ |
void MessagePumpForUI::PumpOutPendingPaintMessages() { |
// If we are being called outside of the context of Run, then don't try to do |
// any work. |
@@ -197,21 +255,28 @@ void MessagePumpForUI::PumpOutPendingPaintMessages() { |
//----------------------------------------------------------------------------- |
// MessagePumpForUI private: |
-// static |
-LRESULT CALLBACK MessagePumpForUI::WndProcThunk( |
- HWND hwnd, UINT message, WPARAM wparam, LPARAM lparam) { |
+bool MessagePumpForUI::HandleMessage(HWND hwnd, |
+ UINT message, |
+ WPARAM wparam, |
+ LPARAM lparam, |
+ LRESULT* result) { |
switch (message) { |
case kMsgHaveWork: |
- reinterpret_cast<MessagePumpForUI*>(wparam)->HandleWorkMessage(); |
+ HandleWorkMessage(); |
break; |
+ |
case WM_TIMER: |
- reinterpret_cast<MessagePumpForUI*>(wparam)->HandleTimerMessage(); |
+ HandleTimerMessage(); |
break; |
} |
- return DefWindowProc(hwnd, message, wparam, lparam); |
+ |
+ // Do default processing for all messages. |
+ return false; |
} |
void MessagePumpForUI::DoRunLoop() { |
+ DCHECK(window_->CalledOnValidThread()); |
+ |
// IF this was just a simple PeekMessage() loop (servicing all possible work |
// queues), then Windows would try to achieve the following order according |
// to MSDN documentation about PeekMessage with no filter): |
@@ -249,7 +314,7 @@ void MessagePumpForUI::DoRunLoop() { |
// don't want to disturb that timer if it is already in flight. However, |
// if we did do all remaining delayed work, then lets kill the WM_TIMER. |
if (more_work_is_plausible && delayed_work_time_.is_null()) |
- KillTimer(message_hwnd_, reinterpret_cast<UINT_PTR>(this)); |
+ KillTimer(window_->hwnd(), kTimerId); |
if (state_->should_quit) |
break; |
@@ -267,20 +332,6 @@ void MessagePumpForUI::DoRunLoop() { |
} |
} |
-void MessagePumpForUI::InitMessageWnd() { |
- WNDCLASSEX wc = {0}; |
- wc.cbSize = sizeof(wc); |
- wc.lpfnWndProc = base::win::WrappedWindowProc<WndProcThunk>; |
- wc.hInstance = base::GetModuleFromAddress(wc.lpfnWndProc); |
- wc.lpszClassName = kWndClass; |
- instance_ = wc.hInstance; |
- RegisterClassEx(&wc); |
- |
- message_hwnd_ = |
- CreateWindow(kWndClass, 0, 0, 0, 0, 0, 0, HWND_MESSAGE, 0, instance_, 0); |
- DCHECK(message_hwnd_); |
-} |
- |
void MessagePumpForUI::WaitForWork() { |
// Wait until a message is available, up to the time needed by the timer |
// manager to fire the next set of timers. |
@@ -290,7 +341,7 @@ void MessagePumpForUI::WaitForWork() { |
DWORD result; |
result = MsgWaitForMultipleObjectsEx(0, NULL, delay, QS_ALLINPUT, |
- MWMO_INPUTAVAILABLE); |
+ MWMO_ALERTABLE | MWMO_INPUTAVAILABLE); |
darin (slow to review)
2013/06/10 19:28:45
I'm a little worried that making this call alertab
alexeypa (please no reviews)
2013/06/10 20:16:12
It means that it will be possible to post APCs to
|
if (WAIT_OBJECT_0 == result) { |
// A WM_* message is available. |
@@ -311,6 +362,11 @@ void MessagePumpForUI::WaitForWork() { |
WaitMessage(); |
} |
return; |
+ } else if (WAIT_IO_COMPLETION == result) { |
+ // The wait was ended by one or more APCs. It could be cause |
+ // MessagePumpUI::ScheduleWork() is trying to wake up the message loop |
+ // thread. |
+ return; |
} |
DCHECK_NE(WAIT_FAILED, result) << GetLastError(); |
@@ -322,7 +378,7 @@ void MessagePumpForUI::HandleWorkMessage() { |
// sort. |
if (!state_) { |
// Since we handled a kMsgHaveWork message, we must still update this flag. |
- InterlockedExchange(&have_work_, 0); |
+ InterlockedExchange(&pump_state_, kPumpIdle); |
return; |
} |
@@ -338,7 +394,7 @@ void MessagePumpForUI::HandleWorkMessage() { |
} |
void MessagePumpForUI::HandleTimerMessage() { |
- KillTimer(message_hwnd_, reinterpret_cast<UINT_PTR>(this)); |
+ KillTimer(window_->hwnd(), kTimerId); |
// If we are being called outside of the context of Run, then don't do |
// anything. This could correspond to a MessageBox call or something of |
@@ -382,7 +438,7 @@ bool MessagePumpForUI::ProcessMessageHelper(const MSG& msg) { |
} |
// While running our main message pump, we discard kMsgHaveWork messages. |
- if (msg.message == kMsgHaveWork && msg.hwnd == message_hwnd_) |
+ if (msg.message == kMsgHaveWork && msg.hwnd == window_->hwnd()) |
return ProcessPumpReplacementMessage(); |
if (CallMsgFilter(const_cast<MSG*>(&msg), kMessageFilterCode)) |
@@ -410,7 +466,7 @@ bool MessagePumpForUI::ProcessPumpReplacementMessage() { |
// goal is to make the kMsgHaveWork as non-intrusive as possible, even though |
// a continuous stream of such messages are posted. This method carefully |
// peeks a message while there is no chance for a kMsgHaveWork to be pending, |
- // then resets the have_work_ flag (allowing a replacement kMsgHaveWork to |
+ // then resets the pump_state_ flag (allowing a replacement kMsgHaveWork to |
// possibly be posted), and finally dispatches that peeked replacement. Note |
// that the re-post of kMsgHaveWork may be asynchronous to this thread!! |
@@ -429,11 +485,11 @@ bool MessagePumpForUI::ProcessPumpReplacementMessage() { |
} |
DCHECK(!have_message || kMsgHaveWork != msg.message || |
- msg.hwnd != message_hwnd_); |
+ msg.hwnd != window_->hwnd()); |
// Since we discarded a kMsgHaveWork message, we must update the flag. |
- int old_have_work = InterlockedExchange(&have_work_, 0); |
- DCHECK(old_have_work); |
+ int old_pump_state = InterlockedExchange(&pump_state_, kPumpIdle); |
+ DCHECK(old_pump_state); |
// We don't need a special time slice if we didn't have_message to process. |
if (!have_message) |
@@ -461,7 +517,7 @@ MessagePumpForIO::MessagePumpForIO() { |
} |
void MessagePumpForIO::ScheduleWork() { |
- if (InterlockedExchange(&have_work_, 1)) |
+ if (InterlockedExchange(&pump_state_, kPumpHaveWork)) |
return; // Someone else continued the pumping. |
// Make sure the MessagePump does some work for us. |
@@ -472,7 +528,7 @@ void MessagePumpForIO::ScheduleWork() { |
return; // Post worked perfectly. |
// See comment in MessagePumpForUI::ScheduleWork() for this error recovery. |
- InterlockedExchange(&have_work_, 0); // Clarify that we didn't succeed. |
+ InterlockedExchange(&pump_state_, kPumpIdle); |
UMA_HISTOGRAM_ENUMERATION("Chrome.MessageLoopProblem", COMPLETION_POST_ERROR, |
MESSAGE_LOOP_PROBLEM_MAX); |
} |
@@ -617,7 +673,7 @@ bool MessagePumpForIO::ProcessInternalIOItem(const IOItem& item) { |
this == reinterpret_cast<MessagePumpForIO*>(item.handler)) { |
// This is our internal completion. |
DCHECK(!item.bytes_transfered); |
- InterlockedExchange(&have_work_, 0); |
+ InterlockedExchange(&pump_state_, kPumpIdle); |
return true; |
} |
return false; |