clang 20.0.0 (based on r547379) from build 12806354. Bug: http://b/379133546 Test: N/A Change-Id: I2eb8938af55d809de674be63cb30cf27e801862b Upstream-Commit: ad834e67b1105d15ef907f6255d4c96e8e733f57
290 lines
11 KiB
C++
290 lines
11 KiB
C++
//===-- llvm/Support/ThreadPool.h - A ThreadPool implementation -*- C++ -*-===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file defines a crude C++11 based thread pool.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#ifndef LLVM_SUPPORT_THREADPOOL_H
|
|
#define LLVM_SUPPORT_THREADPOOL_H
|
|
|
|
#include "llvm/ADT/DenseMap.h"
|
|
#include "llvm/Config/llvm-config.h"
|
|
#include "llvm/Support/RWMutex.h"
|
|
#include "llvm/Support/Threading.h"
|
|
#include "llvm/Support/thread.h"
|
|
|
|
#include <future>
|
|
|
|
#include <condition_variable>
|
|
#include <deque>
|
|
#include <functional>
|
|
#include <memory>
|
|
#include <mutex>
|
|
#include <utility>
|
|
|
|
namespace llvm {
|
|
|
|
class ThreadPoolTaskGroup;
|
|
|
|
/// This defines the abstract base interface for a ThreadPool allowing
|
|
/// asynchronous parallel execution on a defined number of threads.
|
|
///
|
|
/// It is possible to reuse one thread pool for different groups of tasks
|
|
/// by grouping tasks using ThreadPoolTaskGroup. All tasks are processed using
|
|
/// the same queue, but it is possible to wait only for a specific group of
|
|
/// tasks to finish.
|
|
///
|
|
/// It is also possible for worker threads to submit new tasks and wait for
|
|
/// them. Note that this may result in a deadlock in cases such as when a task
|
|
/// (directly or indirectly) tries to wait for its own completion, or when all
|
|
/// available threads are used up by tasks waiting for a task that has no thread
|
|
/// left to run on (this includes waiting on the returned future). It should be
|
|
/// generally safe to wait() for a group as long as groups do not form a cycle.
|
|
class ThreadPoolInterface {
|
|
/// The actual method to enqueue a task to be defined by the concrete
|
|
/// implementation.
|
|
virtual void asyncEnqueue(std::function<void()> Task,
|
|
ThreadPoolTaskGroup *Group) = 0;
|
|
|
|
public:
|
|
/// Destroying the pool will drain the pending tasks and wait. The current
|
|
/// thread may participate in the execution of the pending tasks.
|
|
virtual ~ThreadPoolInterface();
|
|
|
|
/// Blocking wait for all the threads to complete and the queue to be empty.
|
|
/// It is an error to try to add new tasks while blocking on this call.
|
|
/// Calling wait() from a task would deadlock waiting for itself.
|
|
virtual void wait() = 0;
|
|
|
|
/// Blocking wait for only all the threads in the given group to complete.
|
|
/// It is possible to wait even inside a task, but waiting (directly or
|
|
/// indirectly) on itself will deadlock. If called from a task running on a
|
|
/// worker thread, the call may process pending tasks while waiting in order
|
|
/// not to waste the thread.
|
|
virtual void wait(ThreadPoolTaskGroup &Group) = 0;
|
|
|
|
/// Returns the maximum number of worker this pool can eventually grow to.
|
|
virtual unsigned getMaxConcurrency() const = 0;
|
|
|
|
/// Asynchronous submission of a task to the pool. The returned future can be
|
|
/// used to wait for the task to finish and is *non-blocking* on destruction.
|
|
template <typename Function, typename... Args>
|
|
auto async(Function &&F, Args &&...ArgList) {
|
|
auto Task =
|
|
std::bind(std::forward<Function>(F), std::forward<Args>(ArgList)...);
|
|
return async(std::move(Task));
|
|
}
|
|
|
|
/// Overload, task will be in the given task group.
|
|
template <typename Function, typename... Args>
|
|
auto async(ThreadPoolTaskGroup &Group, Function &&F, Args &&...ArgList) {
|
|
auto Task =
|
|
std::bind(std::forward<Function>(F), std::forward<Args>(ArgList)...);
|
|
return async(Group, std::move(Task));
|
|
}
|
|
|
|
/// Asynchronous submission of a task to the pool. The returned future can be
|
|
/// used to wait for the task to finish and is *non-blocking* on destruction.
|
|
template <typename Func>
|
|
auto async(Func &&F) -> std::shared_future<decltype(F())> {
|
|
return asyncImpl(std::function<decltype(F())()>(std::forward<Func>(F)),
|
|
nullptr);
|
|
}
|
|
|
|
template <typename Func>
|
|
auto async(ThreadPoolTaskGroup &Group, Func &&F)
|
|
-> std::shared_future<decltype(F())> {
|
|
return asyncImpl(std::function<decltype(F())()>(std::forward<Func>(F)),
|
|
&Group);
|
|
}
|
|
|
|
private:
|
|
/// Asynchronous submission of a task to the pool. The returned future can be
|
|
/// used to wait for the task to finish and is *non-blocking* on destruction.
|
|
template <typename ResTy>
|
|
std::shared_future<ResTy> asyncImpl(std::function<ResTy()> Task,
|
|
ThreadPoolTaskGroup *Group) {
|
|
auto Future = std::async(std::launch::deferred, std::move(Task)).share();
|
|
asyncEnqueue([Future]() { Future.wait(); }, Group);
|
|
return Future;
|
|
}
|
|
};
|
|
|
|
#if LLVM_ENABLE_THREADS
|
|
/// A ThreadPool implementation using std::threads.
|
|
///
|
|
/// The pool keeps a vector of threads alive, waiting on a condition variable
|
|
/// for some work to become available.
|
|
class StdThreadPool : public ThreadPoolInterface {
|
|
public:
|
|
/// Construct a pool using the hardware strategy \p S for mapping hardware
|
|
/// execution resources (threads, cores, CPUs)
|
|
/// Defaults to using the maximum execution resources in the system, but
|
|
/// accounting for the affinity mask.
|
|
StdThreadPool(ThreadPoolStrategy S = hardware_concurrency());
|
|
|
|
/// Blocking destructor: the pool will wait for all the threads to complete.
|
|
~StdThreadPool() override;
|
|
|
|
/// Blocking wait for all the threads to complete and the queue to be empty.
|
|
/// It is an error to try to add new tasks while blocking on this call.
|
|
/// Calling wait() from a task would deadlock waiting for itself.
|
|
void wait() override;
|
|
|
|
/// Blocking wait for only all the threads in the given group to complete.
|
|
/// It is possible to wait even inside a task, but waiting (directly or
|
|
/// indirectly) on itself will deadlock. If called from a task running on a
|
|
/// worker thread, the call may process pending tasks while waiting in order
|
|
/// not to waste the thread.
|
|
void wait(ThreadPoolTaskGroup &Group) override;
|
|
|
|
/// Returns the maximum number of worker threads in the pool, not the current
|
|
/// number of threads!
|
|
unsigned getMaxConcurrency() const override { return MaxThreadCount; }
|
|
|
|
// TODO: Remove, misleading legacy name warning!
|
|
LLVM_DEPRECATED("Use getMaxConcurrency instead", "getMaxConcurrency")
|
|
unsigned getThreadCount() const { return MaxThreadCount; }
|
|
|
|
/// Returns true if the current thread is a worker thread of this thread pool.
|
|
bool isWorkerThread() const;
|
|
|
|
private:
|
|
/// Returns true if all tasks in the given group have finished (nullptr means
|
|
/// all tasks regardless of their group). QueueLock must be locked.
|
|
bool workCompletedUnlocked(ThreadPoolTaskGroup *Group) const;
|
|
|
|
/// Asynchronous submission of a task to the pool. The returned future can be
|
|
/// used to wait for the task to finish and is *non-blocking* on destruction.
|
|
void asyncEnqueue(std::function<void()> Task,
|
|
ThreadPoolTaskGroup *Group) override {
|
|
int requestedThreads;
|
|
{
|
|
// Lock the queue and push the new task
|
|
std::unique_lock<std::mutex> LockGuard(QueueLock);
|
|
|
|
// Don't allow enqueueing after disabling the pool
|
|
assert(EnableFlag && "Queuing a thread during ThreadPool destruction");
|
|
Tasks.emplace_back(std::make_pair(std::move(Task), Group));
|
|
requestedThreads = ActiveThreads + Tasks.size();
|
|
}
|
|
QueueCondition.notify_one();
|
|
grow(requestedThreads);
|
|
}
|
|
|
|
/// Grow to ensure that we have at least `requested` Threads, but do not go
|
|
/// over MaxThreadCount.
|
|
void grow(int requested);
|
|
|
|
void processTasks(ThreadPoolTaskGroup *WaitingForGroup);
|
|
|
|
/// Threads in flight
|
|
std::vector<llvm::thread> Threads;
|
|
/// Lock protecting access to the Threads vector.
|
|
mutable llvm::sys::RWMutex ThreadsLock;
|
|
|
|
/// Tasks waiting for execution in the pool.
|
|
std::deque<std::pair<std::function<void()>, ThreadPoolTaskGroup *>> Tasks;
|
|
|
|
/// Locking and signaling for accessing the Tasks queue.
|
|
std::mutex QueueLock;
|
|
std::condition_variable QueueCondition;
|
|
|
|
/// Signaling for job completion (all tasks or all tasks in a group).
|
|
std::condition_variable CompletionCondition;
|
|
|
|
/// Keep track of the number of thread actually busy
|
|
unsigned ActiveThreads = 0;
|
|
/// Number of threads active for tasks in the given group (only non-zero).
|
|
DenseMap<ThreadPoolTaskGroup *, unsigned> ActiveGroups;
|
|
|
|
/// Signal for the destruction of the pool, asking thread to exit.
|
|
bool EnableFlag = true;
|
|
|
|
const ThreadPoolStrategy Strategy;
|
|
|
|
/// Maximum number of threads to potentially grow this pool to.
|
|
const unsigned MaxThreadCount;
|
|
};
|
|
#endif // LLVM_ENABLE_THREADS
|
|
|
|
/// A non-threaded implementation.
|
|
class SingleThreadExecutor : public ThreadPoolInterface {
|
|
public:
|
|
/// Construct a non-threaded pool, ignoring using the hardware strategy.
|
|
SingleThreadExecutor(ThreadPoolStrategy ignored = {});
|
|
|
|
/// Blocking destructor: the pool will first execute the pending tasks.
|
|
~SingleThreadExecutor() override;
|
|
|
|
/// Blocking wait for all the tasks to execute first
|
|
void wait() override;
|
|
|
|
/// Blocking wait for only all the tasks in the given group to complete.
|
|
void wait(ThreadPoolTaskGroup &Group) override;
|
|
|
|
/// Returns always 1: there is no concurrency.
|
|
unsigned getMaxConcurrency() const override { return 1; }
|
|
|
|
// TODO: Remove, misleading legacy name warning!
|
|
LLVM_DEPRECATED("Use getMaxConcurrency instead", "getMaxConcurrency")
|
|
unsigned getThreadCount() const { return 1; }
|
|
|
|
/// Returns true if the current thread is a worker thread of this thread pool.
|
|
bool isWorkerThread() const;
|
|
|
|
private:
|
|
/// Asynchronous submission of a task to the pool. The returned future can be
|
|
/// used to wait for the task to finish and is *non-blocking* on destruction.
|
|
void asyncEnqueue(std::function<void()> Task,
|
|
ThreadPoolTaskGroup *Group) override {
|
|
Tasks.emplace_back(std::make_pair(std::move(Task), Group));
|
|
}
|
|
|
|
/// Tasks waiting for execution in the pool.
|
|
std::deque<std::pair<std::function<void()>, ThreadPoolTaskGroup *>> Tasks;
|
|
};
|
|
|
|
#if LLVM_ENABLE_THREADS
|
|
using DefaultThreadPool = StdThreadPool;
|
|
#else
|
|
using DefaultThreadPool = SingleThreadExecutor;
|
|
#endif
|
|
|
|
/// A group of tasks to be run on a thread pool. Thread pool tasks in different
|
|
/// groups can run on the same threadpool but can be waited for separately.
|
|
/// It is even possible for tasks of one group to submit and wait for tasks
|
|
/// of another group, as long as this does not form a loop.
|
|
class ThreadPoolTaskGroup {
|
|
public:
|
|
/// The ThreadPool argument is the thread pool to forward calls to.
|
|
ThreadPoolTaskGroup(ThreadPoolInterface &Pool) : Pool(Pool) {}
|
|
|
|
/// Blocking destructor: will wait for all the tasks in the group to complete
|
|
/// by calling ThreadPool::wait().
|
|
~ThreadPoolTaskGroup() { wait(); }
|
|
|
|
/// Calls ThreadPool::async() for this group.
|
|
template <typename Function, typename... Args>
|
|
inline auto async(Function &&F, Args &&...ArgList) {
|
|
return Pool.async(*this, std::forward<Function>(F),
|
|
std::forward<Args>(ArgList)...);
|
|
}
|
|
|
|
/// Calls ThreadPool::wait() for this group.
|
|
void wait() { Pool.wait(*this); }
|
|
|
|
private:
|
|
ThreadPoolInterface &Pool;
|
|
};
|
|
|
|
} // namespace llvm
|
|
|
|
#endif // LLVM_SUPPORT_THREADPOOL_H
|