| // Copyright 2014 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "platform/scheduler/base/task_queue_manager.h" |
| |
| #include <memory> |
| #include <queue> |
| #include <set> |
| |
| #include "base/bind.h" |
| #include "base/trace_event/trace_event.h" |
| #include "platform/scheduler/base/real_time_domain.h" |
| #include "platform/scheduler/base/task_queue_impl.h" |
| #include "platform/scheduler/base/task_queue_manager_delegate.h" |
| #include "platform/scheduler/base/task_queue_selector.h" |
| #include "platform/scheduler/base/task_time_observer.h" |
| #include "platform/scheduler/base/work_queue.h" |
| #include "platform/scheduler/base/work_queue_sets.h" |
| |
| static const double kLongTaskTraceEventThreshold = 0.05; |
| |
| namespace blink { |
| namespace scheduler { |
| |
| namespace { |
| double MonotonicTimeInSeconds(base::TimeTicks time_ticks) { |
| return (time_ticks - base::TimeTicks()).InSecondsF(); |
| } |
| |
| // Converts a OnceClosure to a RepeatingClosure. It hits CHECK failure to run |
| // the resulting RepeatingClosure more than once. |
| // TODO(tzik): This will be unneeded after the Closure-to-OnceClosure migration |
| // on TaskRunner finished. Remove it once it gets unneeded. |
| base::RepeatingClosure UnsafeConvertOnceClosureToRepeating( |
| base::OnceClosure cb) { |
| return base::BindRepeating([](base::OnceClosure cb) { std::move(cb).Run(); }, |
| base::Passed(&cb)); |
| } |
| } // namespace |
| |
| TaskQueueManager::TaskQueueManager( |
| scoped_refptr<TaskQueueManagerDelegate> delegate) |
| : real_time_domain_(new RealTimeDomain()), |
| graceful_shutdown_helper_(new internal::GracefulQueueShutdownHelper()), |
| delegate_(delegate), |
| task_was_run_on_quiescence_monitored_queue_(false), |
| work_batch_size_(1), |
| task_count_(0), |
| currently_executing_task_queue_(nullptr), |
| observer_(nullptr), |
| weak_factory_(this) { |
| DCHECK(delegate->RunsTasksInCurrentSequence()); |
| TRACE_EVENT_OBJECT_CREATED_WITH_ID( |
| TRACE_DISABLED_BY_DEFAULT("renderer.scheduler"), "TaskQueueManager", |
| this); |
| selector_.SetTaskQueueSelectorObserver(this); |
| |
| delayed_do_work_closure_ = base::BindRepeating( |
| &TaskQueueManager::DoWork, weak_factory_.GetWeakPtr(), true); |
| immediate_do_work_closure_ = base::BindRepeating( |
| &TaskQueueManager::DoWork, weak_factory_.GetWeakPtr(), false); |
| |
| // TODO(alexclarke): Change this to be a parameter that's passed in. |
| RegisterTimeDomain(real_time_domain_.get()); |
| |
| delegate_->AddNestingObserver(this); |
| } |
| |
| TaskQueueManager::~TaskQueueManager() { |
| TRACE_EVENT_OBJECT_DELETED_WITH_ID( |
| TRACE_DISABLED_BY_DEFAULT("renderer.scheduler"), "TaskQueueManager", |
| this); |
| |
| for (internal::TaskQueueImpl* queue : active_queues_) { |
| selector_.RemoveQueue(queue); |
| queue->UnregisterTaskQueue(); |
| } |
| |
| active_queues_.clear(); |
| queues_to_gracefully_shutdown_.clear(); |
| |
| graceful_shutdown_helper_->OnTaskQueueManagerDeleted(); |
| |
| selector_.SetTaskQueueSelectorObserver(nullptr); |
| |
| delegate_->RemoveNestingObserver(this); |
| } |
| |
| TaskQueueManager::AnyThread::AnyThread() |
| : do_work_running_count(0), |
| immediate_do_work_posted_count(0), |
| is_nested(false) {} |
| |
| void TaskQueueManager::RegisterTimeDomain(TimeDomain* time_domain) { |
| time_domains_.insert(time_domain); |
| time_domain->OnRegisterWithTaskQueueManager(this); |
| } |
| |
| void TaskQueueManager::UnregisterTimeDomain(TimeDomain* time_domain) { |
| time_domains_.erase(time_domain); |
| } |
| |
| std::unique_ptr<internal::TaskQueueImpl> TaskQueueManager::CreateTaskQueueImpl( |
| const TaskQueue::Spec& spec) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| TimeDomain* time_domain = |
| spec.time_domain ? spec.time_domain : real_time_domain_.get(); |
| DCHECK(time_domains_.find(time_domain) != time_domains_.end()); |
| std::unique_ptr<internal::TaskQueueImpl> task_queue = |
| std::make_unique<internal::TaskQueueImpl>(this, time_domain, spec); |
| active_queues_.insert(task_queue.get()); |
| selector_.AddQueue(task_queue.get()); |
| return task_queue; |
| } |
| |
| void TaskQueueManager::SetObserver(Observer* observer) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| observer_ = observer; |
| } |
| |
| void TaskQueueManager::UnregisterTaskQueueImpl( |
| std::unique_ptr<internal::TaskQueueImpl> task_queue) { |
| TRACE_EVENT1("renderer.scheduler", "TaskQueueManager::UnregisterTaskQueue", |
| "queue_name", task_queue->GetName()); |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| |
| selector_.RemoveQueue(task_queue.get()); |
| |
| { |
| base::AutoLock lock(any_thread_lock_); |
| any_thread().has_incoming_immediate_work.erase(task_queue.get()); |
| } |
| |
| task_queue->UnregisterTaskQueue(); |
| |
| // Add |task_queue| to |queues_to_delete_| so we can prevent it from being |
| // freed while any of our structures hold hold a raw pointer to it. |
| active_queues_.erase(task_queue.get()); |
| queues_to_delete_[task_queue.get()] = std::move(task_queue); |
| } |
| |
| void TaskQueueManager::ReloadEmptyWorkQueues( |
| const IncomingImmediateWorkMap& queues_to_reload) const { |
| // There are two cases where a queue needs reloading. First, it might be |
| // completely empty and we've just posted a task (this method handles that |
| // case). Secondly if the work queue becomes empty in when calling |
| // WorkQueue::TakeTaskFromWorkQueue (handled there). |
| for (const auto& pair : queues_to_reload) { |
| pair.first->ReloadImmediateWorkQueueIfEmpty(); |
| } |
| } |
| |
| void TaskQueueManager::WakeUpReadyDelayedQueues(LazyNow* lazy_now) { |
| TRACE_EVENT0(TRACE_DISABLED_BY_DEFAULT("renderer.scheduler"), |
| "TaskQueueManager::WakeUpReadyDelayedQueues"); |
| |
| for (TimeDomain* time_domain : time_domains_) { |
| if (time_domain == real_time_domain_.get()) { |
| time_domain->WakeUpReadyDelayedQueues(lazy_now); |
| } else { |
| LazyNow time_domain_lazy_now = time_domain->CreateLazyNow(); |
| time_domain->WakeUpReadyDelayedQueues(&time_domain_lazy_now); |
| } |
| } |
| } |
| |
| void TaskQueueManager::OnBeginNestedRunLoop() { |
| // We just entered a nested run loop, make sure there's a DoWork posted or |
| // the system will grind to a halt. |
| { |
| base::AutoLock lock(any_thread_lock_); |
| any_thread().immediate_do_work_posted_count++; |
| any_thread().is_nested = true; |
| } |
| if (observer_) |
| observer_->OnBeginNestedRunLoop(); |
| |
| delegate_->PostTask(FROM_HERE, immediate_do_work_closure_); |
| } |
| |
| void TaskQueueManager::OnQueueHasIncomingImmediateWork( |
| internal::TaskQueueImpl* queue, |
| internal::EnqueueOrder enqueue_order, |
| bool queue_is_blocked) { |
| MoveableAutoLock lock(any_thread_lock_); |
| any_thread().has_incoming_immediate_work.insert( |
| std::make_pair(queue, enqueue_order)); |
| if (!queue_is_blocked) |
| MaybeScheduleImmediateWorkLocked(FROM_HERE, std::move(lock)); |
| } |
| |
| void TaskQueueManager::MaybeScheduleImmediateWork( |
| const base::Location& from_here) { |
| MoveableAutoLock lock(any_thread_lock_); |
| MaybeScheduleImmediateWorkLocked(from_here, std::move(lock)); |
| } |
| |
| void TaskQueueManager::MaybeScheduleImmediateWorkLocked( |
| const base::Location& from_here, |
| MoveableAutoLock lock) { |
| { |
| MoveableAutoLock auto_lock(std::move(lock)); |
| // Unless we're nested, try to avoid posting redundant DoWorks. |
| if (!any_thread().is_nested && |
| (any_thread().do_work_running_count == 1 || |
| any_thread().immediate_do_work_posted_count > 0)) { |
| return; |
| } |
| |
| any_thread().immediate_do_work_posted_count++; |
| } |
| |
| TRACE_EVENT0(TRACE_DISABLED_BY_DEFAULT("renderer.scheduler"), |
| "TaskQueueManager::MaybeScheduleImmediateWorkLocked::PostTask"); |
| delegate_->PostTask(from_here, immediate_do_work_closure_); |
| } |
| |
| void TaskQueueManager::MaybeScheduleDelayedWork( |
| const base::Location& from_here, |
| TimeDomain* requesting_time_domain, |
| base::TimeTicks now, |
| base::TimeTicks run_time) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| // Make sure we don't cancel another TimeDomain's wake-up. |
| DCHECK(!next_delayed_do_work_ || |
| next_delayed_do_work_.time_domain() == requesting_time_domain); |
| { |
| base::AutoLock lock(any_thread_lock_); |
| |
| // Unless we're nested, don't post a delayed DoWork if there's an immediate |
| // DoWork in flight or we're inside a DoWork. We can rely on DoWork posting |
| // a delayed continuation as needed. |
| if (!any_thread().is_nested && |
| (any_thread().immediate_do_work_posted_count > 0 || |
| any_thread().do_work_running_count == 1)) { |
| return; |
| } |
| } |
| |
| // If there's a delayed DoWork scheduled to run sooner, we don't need to do |
| // anything because DoWork will post a delayed continuation as needed. |
| if (next_delayed_do_work_ && next_delayed_do_work_.run_time() <= run_time) |
| return; |
| |
| cancelable_delayed_do_work_closure_.Reset(delayed_do_work_closure_); |
| |
| base::TimeDelta delay = std::max(base::TimeDelta(), run_time - now); |
| TRACE_EVENT1(TRACE_DISABLED_BY_DEFAULT("renderer.scheduler"), |
| "TaskQueueManager::MaybeScheduleDelayedWork::PostDelayedTask", |
| "delay_ms", delay.InMillisecondsF()); |
| |
| cancelable_delayed_do_work_closure_.Reset(delayed_do_work_closure_); |
| next_delayed_do_work_ = NextDelayedDoWork(run_time, requesting_time_domain); |
| delegate_->PostDelayedTask( |
| from_here, cancelable_delayed_do_work_closure_.callback(), delay); |
| } |
| |
| void TaskQueueManager::CancelDelayedWork(TimeDomain* requesting_time_domain, |
| base::TimeTicks run_time) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| if (next_delayed_do_work_.run_time() != run_time) |
| return; |
| |
| DCHECK_EQ(next_delayed_do_work_.time_domain(), requesting_time_domain); |
| cancelable_delayed_do_work_closure_.Cancel(); |
| next_delayed_do_work_.Clear(); |
| } |
| |
| void TaskQueueManager::DoWork(bool delayed) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| TRACE_EVENT1("renderer.scheduler", "TaskQueueManager::DoWork", "delayed", |
| delayed); |
| |
| LazyNow lazy_now(real_time_domain()->CreateLazyNow()); |
| bool is_nested = delegate_->IsNested(); |
| |
| // This must be done before running any tasks because they could invoke a |
| // nested run loop and we risk having a stale |next_delayed_do_work_|. |
| if (delayed) |
| next_delayed_do_work_.Clear(); |
| |
| for (int i = 0; i < work_batch_size_; i++) { |
| IncomingImmediateWorkMap queues_to_reload; |
| |
| bool was_nested = false; |
| { |
| base::AutoLock lock(any_thread_lock_); |
| if (i == 0) { |
| any_thread().do_work_running_count++; |
| |
| if (!delayed) { |
| any_thread().immediate_do_work_posted_count--; |
| DCHECK_GE(any_thread().immediate_do_work_posted_count, 0); |
| } |
| } else { |
| // Ideally we'd have an OnNestedMessageloopExit observer, but in it's |
| // absence we may need to clear this flag after running a task (which |
| // ran a nested messageloop). |
| if (any_thread().is_nested && !is_nested) |
| was_nested = true; |
| any_thread().is_nested = is_nested; |
| } |
| DCHECK_EQ(any_thread().is_nested, delegate_->IsNested()); |
| std::swap(queues_to_reload, any_thread().has_incoming_immediate_work); |
| } |
| |
| if (observer_ && was_nested) |
| observer_->OnExitNestedRunLoop(); |
| |
| // It's important we call ReloadEmptyWorkQueues out side of the lock to |
| // avoid a lock order inversion. |
| ReloadEmptyWorkQueues(queues_to_reload); |
| |
| WakeUpReadyDelayedQueues(&lazy_now); |
| |
| internal::WorkQueue* work_queue = nullptr; |
| if (!SelectWorkQueueToService(&work_queue)) |
| break; |
| |
| // NB this may unregister |work_queue|. |
| base::TimeTicks time_after_task; |
| switch (ProcessTaskFromWorkQueue(work_queue, is_nested, lazy_now, |
| &time_after_task)) { |
| case ProcessTaskResult::kDeferred: |
| // If a task was deferred, try again with another task. |
| continue; |
| case ProcessTaskResult::kExecuted: |
| break; |
| case ProcessTaskResult::kTaskQueueManagerDeleted: |
| return; // The TaskQueueManager got deleted, we must bail out. |
| } |
| |
| lazy_now = time_after_task.is_null() ? real_time_domain()->CreateLazyNow() |
| : LazyNow(time_after_task); |
| |
| // Only run a single task per batch in nested run loops so that we can |
| // properly exit the nested loop when someone calls RunLoop::Quit(). |
| if (is_nested) |
| break; |
| } |
| |
| if (!is_nested) |
| CleanUpQueues(); |
| |
| // TODO(alexclarke): Consider refactoring the above loop to terminate only |
| // when there's no more work left to be done, rather than posting a |
| // continuation task. |
| |
| bool was_nested = false; |
| { |
| MoveableAutoLock lock(any_thread_lock_); |
| base::Optional<NextTaskDelay> next_delay = |
| ComputeDelayTillNextTaskLocked(&lazy_now); |
| |
| any_thread().do_work_running_count--; |
| DCHECK_GE(any_thread().do_work_running_count, 0); |
| |
| if (any_thread().is_nested && !is_nested) |
| was_nested = true; |
| any_thread().is_nested = is_nested; |
| DCHECK_EQ(any_thread().is_nested, delegate_->IsNested()); |
| |
| PostDoWorkContinuationLocked(next_delay, &lazy_now, std::move(lock)); |
| } |
| |
| if (observer_ && was_nested) |
| observer_->OnExitNestedRunLoop(); |
| } |
| |
| void TaskQueueManager::PostDoWorkContinuationLocked( |
| base::Optional<NextTaskDelay> next_delay, |
| LazyNow* lazy_now, |
| MoveableAutoLock lock) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| |
| { |
| MoveableAutoLock auto_lock(std::move(lock)); |
| |
| // If there are no tasks left then we don't need to post a continuation. |
| if (!next_delay) { |
| // If there's a pending delayed DoWork, cancel it because it's not needed. |
| if (next_delayed_do_work_) { |
| next_delayed_do_work_.Clear(); |
| cancelable_delayed_do_work_closure_.Cancel(); |
| } |
| return; |
| } |
| |
| // If an immediate DoWork is posted, we don't need to post a continuation. |
| if (any_thread().immediate_do_work_posted_count > 0) |
| return; |
| |
| if (next_delay->Delay() <= base::TimeDelta()) { |
| // If a delayed DoWork is pending then we don't need to post a |
| // continuation because it should run immediately. |
| if (next_delayed_do_work_ && |
| next_delayed_do_work_.run_time() <= lazy_now->Now()) { |
| return; |
| } |
| |
| any_thread().immediate_do_work_posted_count++; |
| } |
| } |
| |
| // We avoid holding |any_thread_lock_| while posting the task. |
| if (next_delay->Delay() <= base::TimeDelta()) { |
| delegate_->PostTask(FROM_HERE, immediate_do_work_closure_); |
| } else { |
| base::TimeTicks run_time = lazy_now->Now() + next_delay->Delay(); |
| |
| if (next_delayed_do_work_.run_time() == run_time) |
| return; |
| |
| next_delayed_do_work_ = |
| NextDelayedDoWork(run_time, next_delay->time_domain()); |
| cancelable_delayed_do_work_closure_.Reset(delayed_do_work_closure_); |
| delegate_->PostDelayedTask(FROM_HERE, |
| cancelable_delayed_do_work_closure_.callback(), |
| next_delay->Delay()); |
| } |
| } |
| |
| base::Optional<TaskQueueManager::NextTaskDelay> |
| TaskQueueManager::ComputeDelayTillNextTaskLocked(LazyNow* lazy_now) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| |
| // Unfortunately because |any_thread_lock_| is held it's not safe to call |
| // ReloadEmptyWorkQueues here (possible lock order inversion), however this |
| // check is equavalent to calling ReloadEmptyWorkQueues first. |
| for (const auto& pair : any_thread().has_incoming_immediate_work) { |
| if (pair.first->CouldTaskRun(pair.second)) |
| return NextTaskDelay(); |
| } |
| |
| // If the selector has non-empty queues we trivially know there is immediate |
| // work to be done. |
| if (!selector_.EnabledWorkQueuesEmpty()) |
| return NextTaskDelay(); |
| |
| // Otherwise we need to find the shortest delay, if any. NB we don't need to |
| // call WakeUpReadyDelayedQueues because it's assumed DelayTillNextTask will |
| // return base::TimeDelta>() if the delayed task is due to run now. |
| base::Optional<NextTaskDelay> delay_till_next_task; |
| for (TimeDomain* time_domain : time_domains_) { |
| base::Optional<base::TimeDelta> delay = |
| time_domain->DelayTillNextTask(lazy_now); |
| if (!delay) |
| continue; |
| |
| NextTaskDelay task_delay = (delay.value() == base::TimeDelta()) |
| ? NextTaskDelay() |
| : NextTaskDelay(delay.value(), time_domain); |
| |
| if (!delay_till_next_task || delay_till_next_task > task_delay) |
| delay_till_next_task = task_delay; |
| } |
| return delay_till_next_task; |
| } |
| |
| bool TaskQueueManager::SelectWorkQueueToService( |
| internal::WorkQueue** out_work_queue) { |
| bool should_run = selector_.SelectWorkQueueToService(out_work_queue); |
| TRACE_EVENT_OBJECT_SNAPSHOT_WITH_ID( |
| TRACE_DISABLED_BY_DEFAULT("renderer.scheduler.debug"), "TaskQueueManager", |
| this, AsValueWithSelectorResult(should_run, *out_work_queue)); |
| return should_run; |
| } |
| |
| void TaskQueueManager::DidQueueTask( |
| const internal::TaskQueueImpl::Task& pending_task) { |
| task_annotator_.DidQueueTask("TaskQueueManager::PostTask", pending_task); |
| } |
| |
| TaskQueueManager::ProcessTaskResult TaskQueueManager::ProcessTaskFromWorkQueue( |
| internal::WorkQueue* work_queue, |
| bool is_nested, |
| LazyNow time_before_task, |
| base::TimeTicks* time_after_task) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| base::WeakPtr<TaskQueueManager> protect = GetWeakPtr(); |
| internal::TaskQueueImpl::Task pending_task = |
| work_queue->TakeTaskFromWorkQueue(); |
| |
| // It's possible the task was canceled, if so bail out. |
| if (pending_task.task.IsCancelled()) |
| return ProcessTaskResult::kExecuted; |
| |
| internal::TaskQueueImpl* queue = work_queue->task_queue(); |
| if (queue->GetQuiescenceMonitored()) |
| task_was_run_on_quiescence_monitored_queue_ = true; |
| |
| if (pending_task.nestable == base::Nestable::kNonNestable && is_nested) { |
| // Defer non-nestable work to the main task runner. NOTE these tasks can be |
| // arbitrarily delayed so the additional delay should not be a problem. |
| // TODO(skyostil): Figure out a way to not forget which task queue the |
| // task is associated with. See http://crbug.com/522843. |
| // TODO(tzik): Remove base::UnsafeConvertOnceClosureToRepeating once |
| // TaskRunners have migrated to OnceClosure. |
| delegate_->PostNonNestableTask( |
| pending_task.posted_from, |
| UnsafeConvertOnceClosureToRepeating(std::move(pending_task.task))); |
| return ProcessTaskResult::kDeferred; |
| } |
| |
| double task_start_time_sec = 0; |
| base::TimeTicks task_start_time; |
| TRACE_TASK_EXECUTION("TaskQueueManager::ProcessTaskFromWorkQueue", |
| pending_task); |
| if (queue->GetShouldNotifyObservers()) { |
| for (auto& observer : task_observers_) |
| observer.WillProcessTask(pending_task); |
| queue->NotifyWillProcessTask(pending_task); |
| |
| bool notify_time_observers = !delegate_->IsNested() && |
| (task_time_observers_.might_have_observers() || |
| queue->RequiresTaskTiming()); |
| if (notify_time_observers) { |
| task_start_time = time_before_task.Now(); |
| task_start_time_sec = MonotonicTimeInSeconds(task_start_time); |
| for (auto& observer : task_time_observers_) |
| observer.WillProcessTask(task_start_time_sec); |
| queue->OnTaskStarted(pending_task, task_start_time); |
| } |
| } |
| |
| TRACE_EVENT1("renderer.scheduler", "TaskQueueManager::RunTask", "queue", |
| queue->GetName()); |
| // NOTE when TaskQueues get unregistered a reference ends up getting retained |
| // by |queues_to_delete_| which is cleared at the top of |DoWork|. This means |
| // we are OK to use raw pointers here. |
| internal::TaskQueueImpl* prev_executing_task_queue = |
| currently_executing_task_queue_; |
| currently_executing_task_queue_ = queue; |
| task_annotator_.RunTask("TaskQueueManager::PostTask", &pending_task); |
| // Detect if the TaskQueueManager just got deleted. If this happens we must |
| // not access any member variables after this point. |
| if (!protect) |
| return ProcessTaskResult::kTaskQueueManagerDeleted; |
| |
| currently_executing_task_queue_ = prev_executing_task_queue; |
| |
| double task_end_time_sec = 0; |
| if (queue->GetShouldNotifyObservers()) { |
| if (task_start_time_sec) { |
| *time_after_task = real_time_domain()->Now(); |
| task_end_time_sec = MonotonicTimeInSeconds(*time_after_task); |
| |
| for (auto& observer : task_time_observers_) |
| observer.DidProcessTask(task_start_time_sec, task_end_time_sec); |
| } |
| |
| for (auto& observer : task_observers_) |
| observer.DidProcessTask(pending_task); |
| queue->NotifyDidProcessTask(pending_task); |
| } |
| |
| if (task_start_time_sec && task_end_time_sec) |
| queue->OnTaskCompleted(pending_task, task_start_time, *time_after_task); |
| |
| if (task_start_time_sec && task_end_time_sec && |
| task_end_time_sec - task_start_time_sec > kLongTaskTraceEventThreshold) { |
| TRACE_EVENT_INSTANT1("blink", "LongTask", TRACE_EVENT_SCOPE_THREAD, |
| "duration", task_end_time_sec - task_start_time_sec); |
| } |
| |
| return ProcessTaskResult::kExecuted; |
| } |
| |
| bool TaskQueueManager::RunsTasksInCurrentSequence() const { |
| return delegate_->RunsTasksInCurrentSequence(); |
| } |
| |
| void TaskQueueManager::SetWorkBatchSize(int work_batch_size) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| DCHECK_GE(work_batch_size, 1); |
| work_batch_size_ = work_batch_size; |
| } |
| |
| void TaskQueueManager::AddTaskObserver( |
| base::MessageLoop::TaskObserver* task_observer) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| task_observers_.AddObserver(task_observer); |
| } |
| |
| void TaskQueueManager::RemoveTaskObserver( |
| base::MessageLoop::TaskObserver* task_observer) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| task_observers_.RemoveObserver(task_observer); |
| } |
| |
| void TaskQueueManager::AddTaskTimeObserver( |
| TaskTimeObserver* task_time_observer) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| task_time_observers_.AddObserver(task_time_observer); |
| } |
| |
| void TaskQueueManager::RemoveTaskTimeObserver( |
| TaskTimeObserver* task_time_observer) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| task_time_observers_.RemoveObserver(task_time_observer); |
| } |
| |
| bool TaskQueueManager::GetAndClearSystemIsQuiescentBit() { |
| bool task_was_run = task_was_run_on_quiescence_monitored_queue_; |
| task_was_run_on_quiescence_monitored_queue_ = false; |
| return !task_was_run; |
| } |
| |
| const scoped_refptr<TaskQueueManagerDelegate>& TaskQueueManager::Delegate() |
| const { |
| return delegate_; |
| } |
| |
| internal::EnqueueOrder TaskQueueManager::GetNextSequenceNumber() { |
| return enqueue_order_generator_.GenerateNext(); |
| } |
| |
| LazyNow TaskQueueManager::CreateLazyNow() const { |
| return LazyNow(delegate_.get()); |
| } |
| |
| size_t TaskQueueManager::GetNumberOfPendingTasks() const { |
| size_t task_count = 0; |
| for (auto& queue : active_queues_) |
| task_count += queue->GetNumberOfPendingTasks(); |
| return task_count; |
| } |
| |
| std::unique_ptr<base::trace_event::ConvertableToTraceFormat> |
| TaskQueueManager::AsValueWithSelectorResult( |
| bool should_run, |
| internal::WorkQueue* selected_work_queue) const { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| std::unique_ptr<base::trace_event::TracedValue> state( |
| new base::trace_event::TracedValue()); |
| base::TimeTicks now = real_time_domain()->CreateLazyNow().Now(); |
| state->BeginArray("active_queues"); |
| for (auto& queue : active_queues_) |
| queue->AsValueInto(now, state.get()); |
| state->EndArray(); |
| state->BeginArray("queues_to_gracefully_shutdown"); |
| for (const auto& pair : queues_to_gracefully_shutdown_) |
| pair.first->AsValueInto(now, state.get()); |
| state->EndArray(); |
| state->BeginArray("queues_to_delete"); |
| for (const auto& pair : queues_to_delete_) |
| pair.first->AsValueInto(now, state.get()); |
| state->EndArray(); |
| state->BeginDictionary("selector"); |
| selector_.AsValueInto(state.get()); |
| state->EndDictionary(); |
| if (should_run) { |
| state->SetString("selected_queue", |
| selected_work_queue->task_queue()->GetName()); |
| state->SetString("work_queue_name", selected_work_queue->name()); |
| } |
| |
| state->BeginArray("time_domains"); |
| for (auto* time_domain : time_domains_) |
| time_domain->AsValueInto(state.get()); |
| state->EndArray(); |
| { |
| base::AutoLock lock(any_thread_lock_); |
| state->SetBoolean("is_nested", any_thread().is_nested); |
| state->SetInteger("do_work_running_count", |
| any_thread().do_work_running_count); |
| state->SetInteger("immediate_do_work_posted_count", |
| any_thread().immediate_do_work_posted_count); |
| |
| state->BeginArray("has_incoming_immediate_work"); |
| for (const auto& pair : any_thread().has_incoming_immediate_work) { |
| state->AppendString(pair.first->GetName()); |
| } |
| state->EndArray(); |
| } |
| return std::move(state); |
| } |
| |
| void TaskQueueManager::OnTaskQueueEnabled(internal::TaskQueueImpl* queue) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| DCHECK(queue->IsQueueEnabled()); |
| // Only schedule DoWork if there's something to do. |
| if (queue->HasTaskToRunImmediately() && !queue->BlockedByFence()) |
| MaybeScheduleImmediateWork(FROM_HERE); |
| } |
| |
| void TaskQueueManager::OnTriedToSelectBlockedWorkQueue( |
| internal::WorkQueue* work_queue) { |
| DCHECK(main_thread_checker_.CalledOnValidThread()); |
| DCHECK(!work_queue->Empty()); |
| if (observer_) |
| observer_->OnTriedToExecuteBlockedTask(); |
| } |
| |
| bool TaskQueueManager::HasImmediateWorkForTesting() const { |
| return !selector_.EnabledWorkQueuesEmpty(); |
| } |
| |
| namespace { |
| |
| void SweepCanceledDelayedTasksInQueue( |
| internal::TaskQueueImpl* queue, |
| std::map<TimeDomain*, base::TimeTicks>* time_domain_now) { |
| TimeDomain* time_domain = queue->GetTimeDomain(); |
| if (time_domain_now->find(time_domain) == time_domain_now->end()) |
| time_domain_now->insert(std::make_pair(time_domain, time_domain->Now())); |
| queue->SweepCanceledDelayedTasks(time_domain_now->at(time_domain)); |
| } |
| |
| } // namespace |
| |
| void TaskQueueManager::SweepCanceledDelayedTasks() { |
| std::map<TimeDomain*, base::TimeTicks> time_domain_now; |
| for (const auto& queue : active_queues_) |
| SweepCanceledDelayedTasksInQueue(queue, &time_domain_now); |
| for (const auto& pair : queues_to_gracefully_shutdown_) |
| SweepCanceledDelayedTasksInQueue(pair.first, &time_domain_now); |
| } |
| |
| void TaskQueueManager::TakeQueuesToGracefullyShutdownFromHelper() { |
| std::vector<std::unique_ptr<internal::TaskQueueImpl>> queues = |
| graceful_shutdown_helper_->TakeQueues(); |
| for (std::unique_ptr<internal::TaskQueueImpl>& queue : queues) { |
| queues_to_gracefully_shutdown_[queue.get()] = std::move(queue); |
| } |
| } |
| |
| void TaskQueueManager::CleanUpQueues() { |
| TakeQueuesToGracefullyShutdownFromHelper(); |
| |
| for (auto it = queues_to_gracefully_shutdown_.begin(); |
| it != queues_to_gracefully_shutdown_.end();) { |
| if (it->first->IsEmpty()) { |
| UnregisterTaskQueueImpl(std::move(it->second)); |
| active_queues_.erase(it->first); |
| queues_to_gracefully_shutdown_.erase(it++); |
| } else { |
| ++it; |
| } |
| } |
| queues_to_delete_.clear(); |
| } |
| |
| scoped_refptr<internal::GracefulQueueShutdownHelper> |
| TaskQueueManager::GetGracefulQueueShutdownHelper() const { |
| return graceful_shutdown_helper_; |
| } |
| |
| base::WeakPtr<TaskQueueManager> TaskQueueManager::GetWeakPtr() { |
| return weak_factory_.GetWeakPtr(); |
| } |
| |
| } // namespace scheduler |
| } // namespace blink |