mirror of
https://github.com/yuzu-emu/yuzu.git
synced 2024-12-04 16:43:51 +01:00
306840a580
Provides a simplified SPSC, MPSC, and MPMC bounded queue implementation using mutexes.
247 lines
5.7 KiB
C++
247 lines
5.7 KiB
C++
// SPDX-FileCopyrightText: Copyright 2023 yuzu Emulator Project
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
|
|
#pragma once
|
|
|
|
#include <atomic>
|
|
#include <condition_variable>
|
|
#include <cstddef>
|
|
#include <mutex>
|
|
#include <new>
|
|
|
|
#include "common/polyfill_thread.h"
|
|
|
|
namespace Common {
|
|
|
|
namespace detail {
|
|
constexpr size_t DefaultCapacity = 0x1000;
|
|
} // namespace detail
|
|
|
|
template <typename T, size_t Capacity = detail::DefaultCapacity>
|
|
class SPSCQueue {
|
|
static_assert((Capacity & (Capacity - 1)) == 0, "Capacity must be a power of two.");
|
|
|
|
public:
|
|
void Push(T&& t) {
|
|
const size_t write_index = m_write_index.load();
|
|
|
|
// Wait until we have free slots to write to.
|
|
while ((write_index - m_read_index.load()) == Capacity) {
|
|
std::this_thread::yield();
|
|
}
|
|
|
|
// Determine the position to write to.
|
|
const size_t pos = write_index % Capacity;
|
|
|
|
// Push into the queue.
|
|
m_data[pos] = std::move(t);
|
|
|
|
// Increment the write index.
|
|
++m_write_index;
|
|
|
|
// Notify the consumer that we have pushed into the queue.
|
|
std::scoped_lock lock{cv_mutex};
|
|
cv.notify_one();
|
|
}
|
|
|
|
template <typename... Args>
|
|
void Push(Args&&... args) {
|
|
const size_t write_index = m_write_index.load();
|
|
|
|
// Wait until we have free slots to write to.
|
|
while ((write_index - m_read_index.load()) == Capacity) {
|
|
std::this_thread::yield();
|
|
}
|
|
|
|
// Determine the position to write to.
|
|
const size_t pos = write_index % Capacity;
|
|
|
|
// Emplace into the queue.
|
|
std::construct_at(std::addressof(m_data[pos]), std::forward<Args>(args)...);
|
|
|
|
// Increment the write index.
|
|
++m_write_index;
|
|
|
|
// Notify the consumer that we have pushed into the queue.
|
|
std::scoped_lock lock{cv_mutex};
|
|
cv.notify_one();
|
|
}
|
|
|
|
bool TryPop(T& t) {
|
|
return Pop(t);
|
|
}
|
|
|
|
void PopWait(T& t, std::stop_token stop_token) {
|
|
Wait(stop_token);
|
|
Pop(t);
|
|
}
|
|
|
|
T PopWait(std::stop_token stop_token) {
|
|
Wait(stop_token);
|
|
T t;
|
|
Pop(t);
|
|
return t;
|
|
}
|
|
|
|
void Clear() {
|
|
while (!Empty()) {
|
|
Pop();
|
|
}
|
|
}
|
|
|
|
bool Empty() const {
|
|
return m_read_index.load() == m_write_index.load();
|
|
}
|
|
|
|
size_t Size() const {
|
|
return m_write_index.load() - m_read_index.load();
|
|
}
|
|
|
|
private:
|
|
void Pop() {
|
|
const size_t read_index = m_read_index.load();
|
|
|
|
// Check if the queue is empty.
|
|
if (read_index == m_write_index.load()) {
|
|
return;
|
|
}
|
|
|
|
// Determine the position to read from.
|
|
const size_t pos = read_index % Capacity;
|
|
|
|
// Pop the data off the queue, deleting it.
|
|
std::destroy_at(std::addressof(m_data[pos]));
|
|
|
|
// Increment the read index.
|
|
++m_read_index;
|
|
}
|
|
|
|
bool Pop(T& t) {
|
|
const size_t read_index = m_read_index.load();
|
|
|
|
// Check if the queue is empty.
|
|
if (read_index == m_write_index.load()) {
|
|
return false;
|
|
}
|
|
|
|
// Determine the position to read from.
|
|
const size_t pos = read_index % Capacity;
|
|
|
|
// Pop the data off the queue, moving it.
|
|
t = std::move(m_data[pos]);
|
|
|
|
// Increment the read index.
|
|
++m_read_index;
|
|
|
|
return true;
|
|
}
|
|
|
|
void Wait(std::stop_token stop_token) {
|
|
std::unique_lock lock{cv_mutex};
|
|
Common::CondvarWait(cv, lock, stop_token, [this] { return !Empty(); });
|
|
}
|
|
|
|
alignas(128) std::atomic_size_t m_read_index{0};
|
|
alignas(128) std::atomic_size_t m_write_index{0};
|
|
|
|
std::array<T, Capacity> m_data;
|
|
|
|
std::condition_variable_any cv;
|
|
std::mutex cv_mutex;
|
|
};
|
|
|
|
template <typename T, size_t Capacity = detail::DefaultCapacity>
|
|
class MPSCQueue {
|
|
public:
|
|
void Push(T&& t) {
|
|
std::scoped_lock lock{write_mutex};
|
|
spsc_queue.Push(std::move(t));
|
|
}
|
|
|
|
template <typename... Args>
|
|
void Push(Args&&... args) {
|
|
std::scoped_lock lock{write_mutex};
|
|
spsc_queue.Push(std::forward<Args>(args)...);
|
|
}
|
|
|
|
bool TryPop(T& t) {
|
|
return spsc_queue.TryPop(t);
|
|
}
|
|
|
|
void PopWait(T& t, std::stop_token stop_token) {
|
|
spsc_queue.PopWait(t, stop_token);
|
|
}
|
|
|
|
T PopWait(std::stop_token stop_token) {
|
|
return spsc_queue.PopWait(stop_token);
|
|
}
|
|
|
|
void Clear() {
|
|
spsc_queue.Clear();
|
|
}
|
|
|
|
bool Empty() {
|
|
return spsc_queue.Empty();
|
|
}
|
|
|
|
size_t Size() {
|
|
return spsc_queue.Size();
|
|
}
|
|
|
|
private:
|
|
SPSCQueue<T, Capacity> spsc_queue;
|
|
std::mutex write_mutex;
|
|
};
|
|
|
|
template <typename T, size_t Capacity = detail::DefaultCapacity>
|
|
class MPMCQueue {
|
|
public:
|
|
void Push(T&& t) {
|
|
std::scoped_lock lock{write_mutex};
|
|
spsc_queue.Push(std::move(t));
|
|
}
|
|
|
|
template <typename... Args>
|
|
void Push(Args&&... args) {
|
|
std::scoped_lock lock{write_mutex};
|
|
spsc_queue.Push(std::forward<Args>(args)...);
|
|
}
|
|
|
|
bool TryPop(T& t) {
|
|
std::scoped_lock lock{read_mutex};
|
|
return spsc_queue.TryPop(t);
|
|
}
|
|
|
|
void PopWait(T& t, std::stop_token stop_token) {
|
|
std::scoped_lock lock{read_mutex};
|
|
spsc_queue.PopWait(t, stop_token);
|
|
}
|
|
|
|
T PopWait(std::stop_token stop_token) {
|
|
std::scoped_lock lock{read_mutex};
|
|
return spsc_queue.PopWait(stop_token);
|
|
}
|
|
|
|
void Clear() {
|
|
std::scoped_lock lock{read_mutex};
|
|
spsc_queue.Clear();
|
|
}
|
|
|
|
bool Empty() {
|
|
std::scoped_lock lock{read_mutex};
|
|
return spsc_queue.Empty();
|
|
}
|
|
|
|
size_t Size() {
|
|
std::scoped_lock lock{read_mutex};
|
|
return spsc_queue.Size();
|
|
}
|
|
|
|
private:
|
|
SPSCQueue<T, Capacity> spsc_queue;
|
|
std::mutex write_mutex;
|
|
std::mutex read_mutex;
|
|
};
|
|
|
|
} // namespace Common
|