1. 程式人生 > 實用技巧 >Threading and Tasks in Chrome

Threading and Tasks in Chrome

Contents

Note: SeeThreading and Tasks FAQfor more examples.

Overview

Chrome has amulti-process architectureand each process is heavily multi-threaded. In this document we will go over the basic threading system shared by each process. The main goal is to keep the main thread (a.k.a. “UI” thread in the browser process) and IO thread (each process' thread for handlingIPC) responsive. This means offloading any blocking I/O or other expensive operations to other threads. Our approach is to use message passing as the way of communicating between threads. We discourage locking and thread-safe objects. Instead, objects live on only one (often virtual -- we'll get to that later!) thread and we pass messages between those threads for communication.

This documentation assumes familiarity with computer sciencethreading concepts.

Nomenclature

Core Concepts

  • Task: A unit of work to be processed. Effectively a function pointer with optionally associated state. In Chrome this isbase::Callbackcreated viabase::Bind(documentation).
  • Task queue: A queue of tasks to be processed.
  • Physical thread: An operating system provided thread (e.g. pthread on POSIX or CreateThread() on Windows). The Chrome cross-platform abstraction isbase::PlatformThread. You should pretty much never use this directly.
  • base::Thread: A physical thread forever processing messages from a dedicated task queue until Quit(). You should pretty much never be creating your ownbase::Thread's.
  • Thread pool: A pool of physical threads with a shared task queue. In Chrome, this isbase::ThreadPoolInstance. There's exactly one instance per Chrome process, it serves tasks posted throughbase/task/post_task.hand as such you should rarely need to use thebase::ThreadPoolInstanceAPI directly (more on posting tasks later).
  • SequenceorVirtual thread: A chrome-managed thread of execution. Like a physical thread, only one task can run on a given sequence / virtual thread at any given moment and each task sees the side-effects of the preceding tasks. Tasks are executed sequentially but may hop physical threads between each one.
  • Task runner: An interface through which tasks can be posted. In Chrome this isbase::TaskRunner.
  • Sequenced task runner: A task runner which guarantees that tasks posted to it will run sequentially, in posted order. Each such task is guaranteed to see the side-effects of the task preceding it. Tasks posted to a sequenced task runner are typically processed by a single thread (virtual or physical). In Chrome this isbase::SequencedTaskRunnerwhich is-abase::TaskRunner.
  • Single-thread task runner: A sequenced task runner which guarantees that all tasks will be processed by the same physical thread. In Chrome this isbase::SingleThreadTaskRunnerwhich is-abase::SequencedTaskRunner. Weprefer sequences to threadswhenever possible.

Threading Lexicon

Note to the reader: the following terms are an attempt to bridge the gap between common threading nomenclature and the way we use them in Chrome. It might be a bit heavy if you're just getting started. Should this be hard to parse, consider skipping to the more detailed sections below and referring back to this as necessary.

  • Thread-unsafe: The vast majority of types in Chrome are thread-unsafe (by design). Access to such types/methods must be externally synchronized. Typically thread-unsafe types require that all tasks accessing their state be posted to the samebase::SequencedTaskRunnerand they verify this in debug builds with aSEQUENCE_CHECKERmember. Locks are also an option to synchronize access but in Chrome we stronglyprefer sequences to locks.
  • Thread-affine: Such types/methods need to be always accessed from the same physical thread (i.e. from the samebase::SingleThreadTaskRunner) and typically have aTHREAD_CHECKERmember to verify that they are. Short of using a third-party API or having a leaf dependency which is thread-affine: there's pretty much no reason for a type to be thread-affine in Chrome. Note thatbase::SingleThreadTaskRunneris-abase::SequencedTaskRunnerso thread-affine is a subset of thread-unsafe. Thread-affine is also sometimes referred to asthread-hostile.
  • Thread-safe: Such types/methods can be safely accessed concurrently.
  • Thread-compatible: Such types provide safe concurrent access to const methods but require synchronization for non-const (or mixed const/non-const access). Chrome doesn't expose reader-writer locks; as such, the only use case for this is objects (typically globals) which are initialized once in a thread-safe manner (either in the single-threaded phase of startup or lazily through a thread-safe static-local-initialization paradigm a labase::NoDestructor) and forever after immutable.
  • Immutable: A subset of thread-compatible types which cannot be modified after construction.
  • Sequence-friendly: Such types/methods are thread-unsafe types which support being invoked from abase::SequencedTaskRunner. Ideally this would be the case for all thread-unsafe types but legacy code sometimes has overzealous checks that enforce thread-affinity in mere thread-unsafe scenarios. SeePrefer Sequences to Threadsbelow for more details.

Threads

Every Chrome process has

  • a main thread
    • in the browser process (BrowserThread::UI): updates the UI
    • in renderer processes (Blink main thread): runs most of Blink
  • an IO thread
    • in the browser process (BrowserThread::IO): handles IPCs and network requests
    • in renderer processes: handles IPCs
  • a few more special-purpose threads
  • and a pool of general-purpose threads

Most threads have a loop that gets tasks from a queue and runs them (the queue may be shared between multiple threads).

Tasks

A task is abase::OnceClosureadded to a queue for asynchronous execution.

Abase::OnceClosurestores a function pointer and arguments. It has aRun()method that invokes the function pointer using the bound arguments. It is created usingbase::BindOnce. (ref.Callback<> and Bind() documentation).

void TaskA() {}
void TaskB(int v) {}

auto task_a = base::BindOnce(&TaskA);
auto task_b = base::BindOnce(&TaskB, 42);

A group of tasks can be executed in one of the following ways:

  • Parallel: No task execution ordering, possibly all at once on any thread
  • Sequenced: Tasks executed in posting order, one at a time on any thread.
  • Single Threaded: Tasks executed in posting order, one at a time on a single thread.

Prefer Sequences to Physical Threads

Sequenced execution (on virtual threads) is strongly preferred to single-threaded execution (on physical threads). Except for types/methods bound to the main thread (UI) or IO threads: thread-safety is better achieved viabase::SequencedTaskRunnerthan through managing your own physical threads (ref.Posting a Sequenced Taskbelow).

All APIs which are exposed for “current physical thread” have an equivalent for “current sequence” (mapping).

If you find yourself writing a sequence-friendly type and it fails thread-affinity checks (e.g.,THREAD_CHECKER) in a leaf dependency: consider making that dependency sequence-friendly as well. Most core APIs in Chrome are sequence-friendly, but some legacy types may still over-zealously use ThreadChecker/ThreadTaskRunnerHandle/SingleThreadTaskRunner when they could instead rely on the “current sequence” and no longer be thread-affine.

Posting a Parallel Task

Direct Posting to the Thread Pool

A task that can run on any thread and doesn’t have ordering or mutual exclusion requirements with other tasks should be posted using one of thebase::ThreadPool::PostTask*()functions defined inbase/task/thread_pool.h.

base::ThreadPool::PostTask(FROM_HERE, base::BindOnce(&Task));

This posts tasks with default traits.

Thebase::ThreadPool::PostTask*()functions allow the caller to provide additional details about the task via TaskTraits (ref.Annotating Tasks with TaskTraits).

base::ThreadPool::PostTask(
    FROM_HERE, {base::TaskPriority::BEST_EFFORT, MayBlock()},
    base::BindOnce(&Task));

Posting via a TaskRunner

A parallelbase::TaskRunneris an alternative to callingbase::ThreadPool::PostTask*()directly. This is mainly useful when it isn’t known in advance whether tasks will be posted in parallel, in sequence, or to a single-thread (ref.Posting a Sequenced Task,Posting Multiple Tasks to the Same Thread). Sincebase::TaskRunneris the base class ofbase::SequencedTaskRunnerandbase::SingleThreadTaskRunner, ascoped_refptr<TaskRunner>member can hold abase::TaskRunner, abase::SequencedTaskRunneror abase::SingleThreadTaskRunner.

class A {
 public:
  A() = default;

  void PostSomething() {
    task_runner_->PostTask(FROM_HERE, base::BindOnce(&A, &DoSomething));
  }

  void DoSomething() {
  }

 private:
  scoped_refptr<base::TaskRunner> task_runner_ =
      base::ThreadPool::CreateTaskRunner({base::TaskPriority::USER_VISIBLE});
};

Unless a test needs to control precisely how tasks are executed, it is preferred to callbase::ThreadPool::PostTask*()directly (ref.Testingfor less invasive ways of controlling tasks in tests).

Posting a Sequenced Task

A sequence is a set of tasks that run one at a time in posting order (not necessarily on the same thread). To post tasks as part of a sequence, use abase::SequencedTaskRunner.

Posting to a New Sequence

Abase::SequencedTaskRunnercan be created bybase::ThreadPool::CreateSequencedTaskRunner().

scoped_refptr<SequencedTaskRunner> sequenced_task_runner =
    base::ThreadPool::CreateSequencedTaskRunner(...);

// TaskB runs after TaskA completes.
sequenced_task_runner->PostTask(FROM_HERE, base::BindOnce(&TaskA));
sequenced_task_runner->PostTask(FROM_HERE, base::BindOnce(&TaskB));

Posting to the Current (Virtual) Thread

The preferred way of posting to the current (virtual) thread is viabase::SequencedTaskRunnerHandle::Get().

// The task will run on the current (virtual) thread's default task queue.
base::SequencedTaskRunnerHandle::Get()->PostTask(
    FROM_HERE, base::BindOnce(&Task);

Note that SequencedTaskRunnerHandle::Get() returns the default queue for the current virtual thread. On threads with multiple task queues (e.g. BrowserThread::UI) this can be a different queue than the one the current task belongs to. The “current” task runner is intentionally not exposed via a static getter. Either you know it already and can post to it directly or you don't and the only sensible destination is the default queue.

Using Sequences Instead of Locks

Usage of locks is discouraged in Chrome. Sequences inherently provide thread-safety. Prefer classes that are always accessed from the same sequence to managing your own thread-safety with locks.

Thread-safe but not thread-affine; how so?Tasks posted to the same sequence will run in sequential order. After a sequenced task completes, the next task may be picked up by a different worker thread, but that task is guaranteed to see any side-effects caused by the previous one(s) on its sequence.

class A {
 public:
  A() {
    // Do not require accesses to be on the creation sequence.
    DETACH_FROM_SEQUENCE(sequence_checker_);
  }

  void AddValue(int v) {
    // Check that all accesses are on the same sequence.
    DCHECK_CALLED_ON_VALID_SEQUENCE(sequence_checker_);
    values_.push_back(v);
}

 private:
  SEQUENCE_CHECKER(sequence_checker_);

  // No lock required, because all accesses are on the
  // same sequence.
  std::vector<int> values_;
};

A a;
scoped_refptr<SequencedTaskRunner> task_runner_for_a = ...;
task_runner_for_a->PostTask(FROM_HERE,
                      base::BindOnce(&A::AddValue, base::Unretained(&a), 42));
task_runner_for_a->PostTask(FROM_HERE,
                      base::BindOnce(&A::AddValue, base::Unretained(&a), 27));

// Access from a different sequence causes a DCHECK failure.
scoped_refptr<SequencedTaskRunner> other_task_runner = ...;
other_task_runner->PostTask(FROM_HERE,
                            base::BindOnce(&A::AddValue, base::Unretained(&a), 1));

Locks should only be used to swap in a shared data structure that can be accessed on multiple threads. If one thread updates it based on expensive computation or through disk access, then that slow work should be done without holding the lock. Only when the result is available should the lock be used to swap in the new data. An example of this is in PluginList::LoadPlugins (content/browser/plugin_list.cc. If you must use locks,hereare some best practices and pitfalls to avoid.

In order to write non-blocking code, many APIs in Chrome are asynchronous. Usually this means that they either need to be executed on a particular thread/sequence and will return results via a custom delegate interface, or they take abase::Callback<>object that is called when the requested operation is completed. Executing work on a specific thread/sequence is covered in the PostTask sections above.

Posting Multiple Tasks to the Same Thread

If multiple tasks need to run on the same thread, post them to abase::SingleThreadTaskRunner. All tasks posted to the samebase::SingleThreadTaskRunnerrun on the same thread in posting order.

Posting to the Main Thread or to the IO Thread in the Browser Process

To post tasks to the main thread or to the IO thread, usecontent::GetUIThreadTaskRunner({})orcontent::GetIOThreadTaskRunner({})fromcontent/public/browser/browser_thread.h

You may provide additional BrowserTaskTraits as a parameter to those methods though this is generally still uncommon in BrowserThreads and should be reserved for advanced use cases.

There‘s an ongoing migration (task APIs v3) away from the previous base-API-with-traits which you may still find throughout the codebase (it’s equivalent):

base::PostTask(FROM_HERE, {content::BrowserThread::UI}, ...);

base::CreateSingleThreadTaskRunner({content::BrowserThread::IO})
    ->PostTask(FROM_HERE, ...);

Note: For the duration of the migration, you'll unfortunately need to continue manually includingcontent/public/browser/browser_task_traits.h. to use the browser_thread.h API.

The main thread and the IO thread are already super busy. Therefore, prefer posting to a general purpose thread when possible (ref.Posting a Parallel Task,Posting a Sequenced task). Good reasons to post to the main thread are to update the UI or access objects that are bound to it (e.g.Profile). A good reason to post to the IO thread is to access the internals of components that are bound to it (e.g. IPCs, network). Note: It is not necessary to have an explicit post task to the IO thread to send/receive an IPC or send/receive data on the network.

Posting to the Main Thread in a Renderer Process

TODO(blink-dev)

Posting to a Custom SingleThreadTaskRunner

If multiple tasks need to run on the same thread and that thread doesn’t have to be the main thread or the IO thread, post them to abase::SingleThreadTaskRunnercreated bybase::Threadpool::CreateSingleThreadTaskRunner.

scoped_refptr<SingleThreadTaskRunner> single_thread_task_runner =
    base::Threadpool::CreateSingleThreadTaskRunner(...);

// TaskB runs after TaskA completes. Both tasks run on the same thread.
single_thread_task_runner->PostTask(FROM_HERE, base::BindOnce(&TaskA));
single_thread_task_runner->PostTask(FROM_HERE, base::BindOnce(&TaskB));

Remember that weprefer sequences to physical threadsand that this thus should rarely be necessary.

Posting to the Current Thread

IMPORTANT:To post a task that needs mutual exclusion with the current sequence of tasks but doesn’t absolutely need to run on the current physical thread, usebase::SequencedTaskRunnerHandle::Get()instead ofbase::ThreadTaskRunnerHandle::Get()(ref.Posting to the Current Sequence). That will better document the requirements of the posted task and will avoid unnecessarily making your API physical thread-affine. In a single-thread task,base::SequencedTaskRunnerHandle::Get()is equivalent tobase::ThreadTaskRunnerHandle::Get().

If you must post a task to the current physical thread nonetheless, usebase::ThreadTaskRunnerHandle.

// The task will run on the current thread in the future.
base::ThreadTaskRunnerHandle::Get()->PostTask(
    FROM_HERE, base::BindOnce(&Task));

Posting Tasks to a COM Single-Thread Apartment (STA) Thread (Windows)

Tasks that need to run on a COM Single-Thread Apartment (STA) thread must be posted to abase::SingleThreadTaskRunnerreturned bybase::ThreadPool::CreateCOMSTATaskRunner(). As mentioned inPosting Multiple Tasks to the Same Thread, all tasks posted to the samebase::SingleThreadTaskRunnerrun on the same thread in posting order.

// Task(A|B|C)UsingCOMSTA will run on the same COM STA thread.

void TaskAUsingCOMSTA() {
  // [ This runs on a COM STA thread. ]

  // Make COM STA calls.
  // ...

  // Post another task to the current COM STA thread.
  base::ThreadTaskRunnerHandle::Get()->PostTask(
      FROM_HERE, base::BindOnce(&TaskCUsingCOMSTA));
}
void TaskBUsingCOMSTA() { }
void TaskCUsingCOMSTA() { }

auto com_sta_task_runner = base::ThreadPool::CreateCOMSTATaskRunner(...);
com_sta_task_runner->PostTask(FROM_HERE, base::BindOnce(&TaskAUsingCOMSTA));
com_sta_task_runner->PostTask(FROM_HERE, base::BindOnce(&TaskBUsingCOMSTA));

Annotating Tasks with TaskTraits

base::TaskTraitsencapsulate information about a task that helps the thread pool make better scheduling decisions.

Methods that takebase::TaskTraitscan be be passed{}when default traits are sufficient. Default traits are appropriate for tasks that:

  • Don’t block (ref. MayBlock and WithBaseSyncPrimitives);
  • Pertain to user-blocking activity; (explicitly or implicitly by having an ordering dependency with a component that does)
  • Can either block shutdown or be skipped on shutdown (thread pool is free to choose a fitting default). Tasks that don’t match this description must be posted with explicit TaskTraits.

base/task/task_traits.hprovides exhaustive documentation of available traits. The content layer also provides additional traits incontent/public/browser/browser_task_traits.hto facilitate posting a task onto a BrowserThread.

Below are some examples of how to specifybase::TaskTraits.

// This task has no explicit TaskTraits. It cannot block. Its priority is
// USER_BLOCKING. It will either block shutdown or be skipped on shutdown.
base::ThreadPool::PostTask(FROM_HERE, base::BindOnce(...));

// This task has the highest priority. The thread pool will schedule it before
// USER_VISIBLE and BEST_EFFORT tasks.
base::ThreadPool::PostTask(
    FROM_HERE, {base::TaskPriority::USER_BLOCKING},
    base::BindOnce(...));

// This task has the lowest priority and is allowed to block (e.g. it
// can read a file from disk).
base::ThreadPool::PostTask(
    FROM_HERE, {base::TaskPriority::BEST_EFFORT, base::MayBlock()},
    base::BindOnce(...));

// This task blocks shutdown. The process won't exit before its
// execution is complete.
base::ThreadPool::PostTask(
    FROM_HERE, {base::TaskShutdownBehavior::BLOCK_SHUTDOWN},
    base::BindOnce(...));

Keeping the Browser Responsive

Do not perform expensive work on the main thread, the IO thread or any sequence that is expected to run tasks with a low latency. Instead, perform expensive work asynchronously usingbase::ThreadPool::PostTaskAndReply*()orbase::SequencedTaskRunner::PostTaskAndReply(). Note that asynchronous/overlapped I/O on the IO thread are fine.

Example: Running the code below on the main thread will prevent the browser from responding to user input for a long time.

// GetHistoryItemsFromDisk() may block for a long time.
// AddHistoryItemsToOmniboxDropDown() updates the UI and therefore must
// be called on the main thread.
AddHistoryItemsToOmniboxDropdown(GetHistoryItemsFromDisk("keyword"));

The code below solves the problem by scheduling a call toGetHistoryItemsFromDisk()in a thread pool followed by a call toAddHistoryItemsToOmniboxDropdown()on the origin sequence (the main thread in this case). The return value of the first call is automatically provided as argument to the second call.

base::ThreadPool::PostTaskAndReplyWithResult(
    FROM_HERE, {base::MayBlock()},
    base::BindOnce(&GetHistoryItemsFromDisk, "keyword"),
    base::BindOnce(&AddHistoryItemsToOmniboxDropdown));

Posting a Task with a Delay

Posting a One-Off Task with a Delay

To post a task that must run once after a delay expires, usebase::ThreadPool::PostDelayedTask*()orbase::TaskRunner::PostDelayedTask().

base::ThreadPool::PostDelayedTask(
  FROM_HERE, {base::TaskPriority::BEST_EFFORT}, base::BindOnce(&Task),
  base::TimeDelta::FromHours(1));

scoped_refptr<base::SequencedTaskRunner> task_runner =
    base::ThreadPool::CreateSequencedTaskRunner(
        {base::TaskPriority::BEST_EFFORT});
task_runner->PostDelayedTask(
    FROM_HERE, base::BindOnce(&Task), base::TimeDelta::FromHours(1));
NOTE:A task that has a 1-hour delay probably doesn’t have to run right away when its delay expires. Specifybase::TaskPriority::BEST_EFFORTto prevent it from slowing down the browser when its delay expires.

Posting a Repeating Task with a Delay

To post a task that must run at regular intervals, usebase::RepeatingTimer.

class A {
 public:
  ~A() {
    // The timer is stopped automatically when it is deleted.
  }
  void StartDoingStuff() {
    timer_.Start(FROM_HERE, TimeDelta::FromSeconds(1),
                 this, &MyClass::DoStuff);
  }
  void StopDoingStuff() {
    timer_.Stop();
  }
 private:
  void DoStuff() {
    // This method is called every second on the sequence that invoked
    // StartDoingStuff().
  }
  base::RepeatingTimer timer_;
};

Cancelling a Task

Using base::WeakPtr

base::WeakPtrcan be used to ensure that any callback bound to an object is canceled when that object is destroyed.

int Compute() {  }

class A {
 public:
  void ComputeAndStore() {
    // Schedule a call to Compute() in a thread pool followed by
    // a call to A::Store() on the current sequence. The call to
    // A::Store() is canceled when |weak_ptr_factory_| is destroyed.
    // (guarantees that |this| will not be used-after-free).
    base::ThreadPool::PostTaskAndReplyWithResult(
        FROM_HERE, base::BindOnce(&Compute),
        base::BindOnce(&A::Store, weak_ptr_factory_.GetWeakPtr()));
  }

 private:
  void Store(int value) { value_ = value; }

  int value_;
  base::WeakPtrFactory<A> weak_ptr_factory_{this};
};

Note:WeakPtris not thread-safe:GetWeakPtr(),~WeakPtrFactory(), andCompute()(bound to aWeakPtr) must all run on the same sequence.

Using base::CancelableTaskTracker

base::CancelableTaskTrackerallows cancellation to happen on a different sequence than the one on which tasks run. Keep in mind thatCancelableTaskTrackercannot cancel tasks that have already started to run.

auto task_runner = base::ThreadPool::CreateTaskRunner({});
base::CancelableTaskTracker cancelable_task_tracker;
cancelable_task_tracker.PostTask(task_runner.get(), FROM_HERE,
                                 base::DoNothing());
// Cancels Task(), only if it hasn't already started running.
cancelable_task_tracker.TryCancelAll();

Posting a Job to run in parallel

Thebase::PostJobis a power user API to be able to schedule a single base::RepeatingCallback worker task and request that ThreadPool workers invoke it concurrently. This avoids degenerate cases:

  • CallingPostTask()for each work item, causing significant overhead.
  • Fixed number ofPostTask()calls that split the work and might run for a long time. This is problematic when many components post “num cores” tasks and all expect to use all the cores. In these cases, the scheduler lacks context to be fair to multiple same-priority requests and/or ability to request lower priority work to yield when high priority work comes in.

Seebase/task/job_perftest.ccfor a complete example.

// A canonical implementation of |worker_task|.
void WorkerTask(base::JobDelegate* job_delegate) {
  while (!job_delegate->ShouldYield()) {
    auto work_item = TakeWorkItem(); // Smallest unit of work.
    if (!work_item)
      return:
    ProcessWork(work_item);
  }
}

// Returns the latest thread-safe number of incomplete work items.
void NumIncompleteWorkItems(size_t worker_count) {
  // NumIncompleteWorkItems() may use |worker_count| if it needs to account for
  // local work lists, which is easier than doing its own accounting, keeping in
  // mind that the actual number of items may be racily overestimated and thus
  // WorkerTask() may be called when there's no available work.
  return GlobalQueueSize() + worker_count;
}

base::PostJob(FROM_HERE, {},
              base::BindRepeating(&WorkerTask),
              base::BindRepeating(&NumIncompleteWorkItems));

By doing as much work as possible in a loop when invoked, the worker task avoids scheduling overhead. Meanwhilebase::JobDelegate::ShouldYield()is periodically invoked to conditionally exit and let the scheduler prioritize other work. This yield-semantic allows, for example, a user-visible job to use all cores but get out of the way when a user-blocking task comes in.

Adding additional work to a running job.

When new work items are added and the API user wants additional threads to invoke the worker task concurrently,JobHandle/JobDelegate::NotifyConcurrencyIncrease()mustbe invoked shortly after max concurrency increases.

Testing

For more details seeTesting Components Which Post Tasks.

To test code that usesbase::ThreadTaskRunnerHandle,base::SequencedTaskRunnerHandleor a function inbase/task/post_task.h, instantiate abase::test::TaskEnvironmentfor the scope of the test. If you need BrowserThreads, usecontent::BrowserTaskEnvironmentinstead ofbase::test::TaskEnvironment.

Tests can run thebase::test::TaskEnvironment's message pump using abase::RunLoop, which can be made to run untilQuit()(explicitly or viaRunLoop::QuitClosure()), or toRunUntilIdle()ready-to-run tasks and immediately return.

TaskEnvironment configures RunLoop::Run() to GTEST_FAIL() if it hasn't been explicitly quit after TestTimeouts::action_timeout(). This is preferable to having the test hang if the code under test fails to trigger the RunLoop to quit. The timeout can be overridden with base::test::ScopedRunLoopTimeout.

class MyTest : public testing::Test {
 public:
  // ...
 protected:
   base::test::TaskEnvironment task_environment_;
};

TEST(MyTest, MyTest) {
  base::ThreadTaskRunnerHandle::Get()->PostTask(FROM_HERE, base::BindOnce(&A));
  base::SequencedTaskRunnerHandle::Get()->PostTask(FROM_HERE,
                                                   base::BindOnce(&B));
  base::ThreadTaskRunnerHandle::Get()->PostDelayedTask(
      FROM_HERE, base::BindOnce(&C), base::TimeDelta::Max());

  // This runs the (Thread|Sequenced)TaskRunnerHandle queue until it is empty.
  // Delayed tasks are not added to the queue until they are ripe for execution.
  base::RunLoop().RunUntilIdle();
  // A and B have been executed. C is not ripe for execution yet.

  base::RunLoop run_loop;
  base::ThreadTaskRunnerHandle::Get()->PostTask(FROM_HERE, base::BindOnce(&D));
  base::ThreadTaskRunnerHandle::Get()->PostTask(FROM_HERE, run_loop.QuitClosure());
  base::ThreadTaskRunnerHandle::Get()->PostTask(FROM_HERE, base::BindOnce(&E));

  // This runs the (Thread|Sequenced)TaskRunnerHandle queue until QuitClosure is
  // invoked.
  run_loop.Run();
  // D and run_loop.QuitClosure() have been executed. E is still in the queue.

  // Tasks posted to thread pool run asynchronously as they are posted.
  base::ThreadPool::PostTask(FROM_HERE, {}, base::BindOnce(&F));
  auto task_runner =
      base::ThreadPool::CreateSequencedTaskRunner({});
  task_runner->PostTask(FROM_HERE, base::BindOnce(&G));

  // To block until all tasks posted to thread pool are done running:
  base::ThreadPoolInstance::Get()->FlushForTesting();
  // F and G have been executed.

  base::ThreadPool::PostTaskAndReplyWithResult(
      FROM_HERE, {}, base::BindOnce(&H), base::BindOnce(&I));

  // This runs the (Thread|Sequenced)TaskRunnerHandle queue until both the
  // (Thread|Sequenced)TaskRunnerHandle queue and the TaskSchedule queue are
  // empty:
  task_environment_.RunUntilIdle();
  // E, H, I have been executed.
}

Using ThreadPool in a New Process

ThreadPoolInstance needs to be initialized in a process before the functions inbase/task/post_task.hcan be used. Initialization of ThreadPoolInstance in the Chrome browser process and child processes (renderer, GPU, utility) has already been taken care of. To use ThreadPoolInstance in another process, initialize ThreadPoolInstance early in the main function:

// This initializes and starts ThreadPoolInstance with default params.
base::ThreadPoolInstance::CreateAndStartWithDefaultParams(“process_name”);
// The base/task/post_task.h API can now be used with base::ThreadPool trait.
// Tasks will be // scheduled as they are posted.

// This initializes ThreadPoolInstance.
base::ThreadPoolInstance::Create(“process_name”);
// The base/task/post_task.h API can now be used with base::ThreadPool trait. No
// threads will be created and no tasks will be scheduled until after Start() is
// called.
base::ThreadPoolInstance::Get()->Start(params);
// ThreadPool can now create threads and schedule tasks.

And shutdown ThreadPoolInstance late in the main function:

base::ThreadPoolInstance::Get()->Shutdown();
// Tasks posted with TaskShutdownBehavior::BLOCK_SHUTDOWN and
// tasks posted with TaskShutdownBehavior::SKIP_ON_SHUTDOWN that
// have started to run before the Shutdown() call have now completed their
// execution. Tasks posted with
// TaskShutdownBehavior::CONTINUE_ON_SHUTDOWN may still be
// running.

TaskRunner ownership (encourage no dependency injection)

TaskRunners shouldn't be passed through several components. Instead, the components that uses a TaskRunner should be the one that creates it.

Seethis exampleof a refactoring where a TaskRunner was passed through a lot of components only to be used in an eventual leaf. The leaf can and should now obtain its TaskRunner directly frombase/task/post_task.h.

As mentioned above,base::test::TaskEnvironmentallows unit tests to control tasks posted from underlying TaskRunners. In rare cases where a test needs to more precisely control task ordering: dependency injection of TaskRunners can be useful. For such cases the preferred approach is the following:

class Foo {
 public:

  // Overrides |background_task_runner_| in tests.
  void SetBackgroundTaskRunnerForTesting(
      scoped_refptr<base::SequencedTaskRunner> background_task_runner) {
    background_task_runner_ = std::move(background_task_runner);
  }

 private:
  scoped_refptr<base::SequencedTaskRunner> background_task_runner_ =
      base::ThreadPool::CreateSequencedTaskRunner(
          {base::MayBlock(), base::TaskPriority::BEST_EFFORT});
}

Note that this still allows removing all layers of plumbing between //chrome and that component since unit tests will use the leaf layer directly.

FAQ

SeeThreading and Tasks FAQfor more examples.

Internals

SequenceManager

SequenceManagermanages TaskQueues which have different properties (e.g. priority, common task type) multiplexing all posted tasks into a single backing sequence. This will usually be a MessagePump. Depending on the type of message pump used other events such as UI messages may be processed as well. On Windows APC calls (as time permits) and signals sent to a registered set of HANDLEs may also be processed.

MessagePump

MessagePumpsare responsible for processing native messages as well as for giving cycles to their delegate (SequenceManager) periodically. MessagePumps take care to mixing delegate callbacks with native message processing so neither type of event starves the other of cycles.

There are differentMessagePumpTypes, most common are:

  • DEFAULT: Supports tasks and timers only

  • UI: Supports native UI events (e.g. Windows messages)

  • IO: Supports asynchronous IO (not file I/O!)

  • CUSTOM: User provided implementation of MessagePump interface

RunLoop

RunLoop is s helper class to run the RunLoop::Delegate associated with the current thread (usually a SequenceManager). Create a RunLoop on the stack and call Run/Quit to run a nested RunLoop but please avoid nested loops in production code!

Task Reentrancy

SequenceManager has task reentrancy protection. This means that if a task is being processed, a second task cannot start until the first task is finished. Reentrancy can happen when processing a task, and an inner message pump is created. That inner pump then processes native messages which could implicitly start an inner task. Inner message pumps are created with dialogs (DialogBox), common dialogs (GetOpenFileName), OLE functions (DoDragDrop), printer functions (StartDoc) andmanyothers.

Sample workaround when inner task processing is needed:
  HRESULT hr;
  {
    CurrentThread::ScopedNestableTaskAllower allow;
    hr = DoDragDrop(...); // Implicitly runs a modal message loop.
  }
  // Process |hr| (the result returned by DoDragDrop()).

Please be SURE your task is reentrant (nestable) and all global variables are stable and accessible before before using CurrentThread::ScopedNestableTaskAllower.

APIs for general use

User code should hardly ever need to access SequenceManager APIs directly as these are meant for code that deals with scheduling. Instead you should use the following:

  • base::RunLoop: Drive the SequenceManager from the thread it's bound to.

  • base::Thread/SequencedTaskRunnerHandle: Post back to the SequenceManager TaskQueues from a task running on it.

  • SequenceLocalStorageSlot : Bind external state to a sequence.

  • base::CurrentThread : Proxy to a subset of Task related APIs bound to the current thread

  • Embedders may provide their own static accessors to post tasks on specific loops (e.g. content::BrowserThreads).

SingleThreadTaskExecutor and TaskEnvironment

Instead of having to deal with SequenceManager and TaskQueues code that needs a simple task posting environment (one default task queue) can use aSingleThreadTaskExecutor.

Unit tests can useTaskEnvironmentwhich is highly configurable.

MessageLoop and CurrentThread

You might come across references to MessageLoop or CurrentThread in the code or documentation. These classes no longer exist and we are in the process or getting rid of all references to them. base::CurrentThread was replaced by base::CurrentThread and the drop in replacements for base::MessageLoop are base::SingleThreadTaskExecutor and base::Test::TaskEnvironment.