Browse Source
Merge pull request #3396 from FernandoS27/prometheus-1
Merge pull request #3396 from FernandoS27/prometheus-1
Implement SpinLocks, Fibers and a Host Timerpull/15/merge
committed by
GitHub
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
22 changed files with 1646 additions and 3 deletions
-
11src/common/CMakeLists.txt
-
226src/common/fiber.cpp
-
92src/common/fiber.h
-
54src/common/spin_lock.cpp
-
21src/common/spin_lock.h
-
4src/common/thread.h
-
26src/common/uint128.cpp
-
3src/common/uint128.h
-
92src/common/wall_clock.cpp
-
51src/common/wall_clock.h
-
33src/common/x64/cpu_detect.cpp
-
12src/common/x64/cpu_detect.h
-
95src/common/x64/native_clock.cpp
-
41src/common/x64/native_clock.h
-
2src/core/CMakeLists.txt
-
15src/core/core_timing_util.cpp
-
3src/core/core_timing_util.h
-
206src/core/host_timing.cpp
-
160src/core/host_timing.h
-
2src/tests/CMakeLists.txt
-
358src/tests/common/fibers.cpp
-
142src/tests/core/host_timing.cpp
@ -0,0 +1,226 @@ |
|||
// Copyright 2020 yuzu Emulator Project
|
|||
// Licensed under GPLv2 or any later version
|
|||
// Refer to the license.txt file included.
|
|||
|
|||
#include "common/assert.h"
|
|||
#include "common/fiber.h"
|
|||
#if defined(_WIN32) || defined(WIN32)
|
|||
#include <windows.h>
|
|||
#else
|
|||
#include <boost/context/detail/fcontext.hpp>
|
|||
#endif
|
|||
|
|||
namespace Common { |
|||
|
|||
constexpr std::size_t default_stack_size = 256 * 1024; // 256kb
|
|||
|
|||
#if defined(_WIN32) || defined(WIN32)
|
|||
|
|||
struct Fiber::FiberImpl { |
|||
LPVOID handle = nullptr; |
|||
LPVOID rewind_handle = nullptr; |
|||
}; |
|||
|
|||
void Fiber::Start() { |
|||
ASSERT(previous_fiber != nullptr); |
|||
previous_fiber->guard.unlock(); |
|||
previous_fiber.reset(); |
|||
entry_point(start_parameter); |
|||
UNREACHABLE(); |
|||
} |
|||
|
|||
void Fiber::OnRewind() { |
|||
ASSERT(impl->handle != nullptr); |
|||
DeleteFiber(impl->handle); |
|||
impl->handle = impl->rewind_handle; |
|||
impl->rewind_handle = nullptr; |
|||
rewind_point(rewind_parameter); |
|||
UNREACHABLE(); |
|||
} |
|||
|
|||
void Fiber::FiberStartFunc(void* fiber_parameter) { |
|||
auto fiber = static_cast<Fiber*>(fiber_parameter); |
|||
fiber->Start(); |
|||
} |
|||
|
|||
void Fiber::RewindStartFunc(void* fiber_parameter) { |
|||
auto fiber = static_cast<Fiber*>(fiber_parameter); |
|||
fiber->OnRewind(); |
|||
} |
|||
|
|||
Fiber::Fiber(std::function<void(void*)>&& entry_point_func, void* start_parameter) |
|||
: entry_point{std::move(entry_point_func)}, start_parameter{start_parameter} { |
|||
impl = std::make_unique<FiberImpl>(); |
|||
impl->handle = CreateFiber(default_stack_size, &FiberStartFunc, this); |
|||
} |
|||
|
|||
Fiber::Fiber() { |
|||
impl = std::make_unique<FiberImpl>(); |
|||
} |
|||
|
|||
Fiber::~Fiber() { |
|||
if (released) { |
|||
return; |
|||
} |
|||
// Make sure the Fiber is not being used
|
|||
const bool locked = guard.try_lock(); |
|||
ASSERT_MSG(locked, "Destroying a fiber that's still running"); |
|||
if (locked) { |
|||
guard.unlock(); |
|||
} |
|||
DeleteFiber(impl->handle); |
|||
} |
|||
|
|||
void Fiber::Exit() { |
|||
ASSERT_MSG(is_thread_fiber, "Exitting non main thread fiber"); |
|||
if (!is_thread_fiber) { |
|||
return; |
|||
} |
|||
ConvertFiberToThread(); |
|||
guard.unlock(); |
|||
released = true; |
|||
} |
|||
|
|||
void Fiber::SetRewindPoint(std::function<void(void*)>&& rewind_func, void* start_parameter) { |
|||
rewind_point = std::move(rewind_func); |
|||
rewind_parameter = start_parameter; |
|||
} |
|||
|
|||
void Fiber::Rewind() { |
|||
ASSERT(rewind_point); |
|||
ASSERT(impl->rewind_handle == nullptr); |
|||
impl->rewind_handle = CreateFiber(default_stack_size, &RewindStartFunc, this); |
|||
SwitchToFiber(impl->rewind_handle); |
|||
} |
|||
|
|||
void Fiber::YieldTo(std::shared_ptr<Fiber>& from, std::shared_ptr<Fiber>& to) { |
|||
ASSERT_MSG(from != nullptr, "Yielding fiber is null!"); |
|||
ASSERT_MSG(to != nullptr, "Next fiber is null!"); |
|||
to->guard.lock(); |
|||
to->previous_fiber = from; |
|||
SwitchToFiber(to->impl->handle); |
|||
ASSERT(from->previous_fiber != nullptr); |
|||
from->previous_fiber->guard.unlock(); |
|||
from->previous_fiber.reset(); |
|||
} |
|||
|
|||
std::shared_ptr<Fiber> Fiber::ThreadToFiber() { |
|||
std::shared_ptr<Fiber> fiber = std::shared_ptr<Fiber>{new Fiber()}; |
|||
fiber->guard.lock(); |
|||
fiber->impl->handle = ConvertThreadToFiber(nullptr); |
|||
fiber->is_thread_fiber = true; |
|||
return fiber; |
|||
} |
|||
|
|||
#else
|
|||
|
|||
struct Fiber::FiberImpl { |
|||
alignas(64) std::array<u8, default_stack_size> stack; |
|||
u8* stack_limit; |
|||
alignas(64) std::array<u8, default_stack_size> rewind_stack; |
|||
u8* rewind_stack_limit; |
|||
boost::context::detail::fcontext_t context; |
|||
boost::context::detail::fcontext_t rewind_context; |
|||
}; |
|||
|
|||
void Fiber::Start(boost::context::detail::transfer_t& transfer) { |
|||
ASSERT(previous_fiber != nullptr); |
|||
previous_fiber->impl->context = transfer.fctx; |
|||
previous_fiber->guard.unlock(); |
|||
previous_fiber.reset(); |
|||
entry_point(start_parameter); |
|||
UNREACHABLE(); |
|||
} |
|||
|
|||
void Fiber::OnRewind([[maybe_unused]] boost::context::detail::transfer_t& transfer) { |
|||
ASSERT(impl->context != nullptr); |
|||
impl->context = impl->rewind_context; |
|||
impl->rewind_context = nullptr; |
|||
u8* tmp = impl->stack_limit; |
|||
impl->stack_limit = impl->rewind_stack_limit; |
|||
impl->rewind_stack_limit = tmp; |
|||
rewind_point(rewind_parameter); |
|||
UNREACHABLE(); |
|||
} |
|||
|
|||
void Fiber::FiberStartFunc(boost::context::detail::transfer_t transfer) { |
|||
auto fiber = static_cast<Fiber*>(transfer.data); |
|||
fiber->Start(transfer); |
|||
} |
|||
|
|||
void Fiber::RewindStartFunc(boost::context::detail::transfer_t transfer) { |
|||
auto fiber = static_cast<Fiber*>(transfer.data); |
|||
fiber->OnRewind(transfer); |
|||
} |
|||
|
|||
Fiber::Fiber(std::function<void(void*)>&& entry_point_func, void* start_parameter) |
|||
: entry_point{std::move(entry_point_func)}, start_parameter{start_parameter} { |
|||
impl = std::make_unique<FiberImpl>(); |
|||
impl->stack_limit = impl->stack.data(); |
|||
impl->rewind_stack_limit = impl->rewind_stack.data(); |
|||
u8* stack_base = impl->stack_limit + default_stack_size; |
|||
impl->context = |
|||
boost::context::detail::make_fcontext(stack_base, impl->stack.size(), FiberStartFunc); |
|||
} |
|||
|
|||
void Fiber::SetRewindPoint(std::function<void(void*)>&& rewind_func, void* start_parameter) { |
|||
rewind_point = std::move(rewind_func); |
|||
rewind_parameter = start_parameter; |
|||
} |
|||
|
|||
Fiber::Fiber() { |
|||
impl = std::make_unique<FiberImpl>(); |
|||
} |
|||
|
|||
Fiber::~Fiber() { |
|||
if (released) { |
|||
return; |
|||
} |
|||
// Make sure the Fiber is not being used
|
|||
const bool locked = guard.try_lock(); |
|||
ASSERT_MSG(locked, "Destroying a fiber that's still running"); |
|||
if (locked) { |
|||
guard.unlock(); |
|||
} |
|||
} |
|||
|
|||
void Fiber::Exit() { |
|||
|
|||
ASSERT_MSG(is_thread_fiber, "Exitting non main thread fiber"); |
|||
if (!is_thread_fiber) { |
|||
return; |
|||
} |
|||
guard.unlock(); |
|||
released = true; |
|||
} |
|||
|
|||
void Fiber::Rewind() { |
|||
ASSERT(rewind_point); |
|||
ASSERT(impl->rewind_context == nullptr); |
|||
u8* stack_base = impl->rewind_stack_limit + default_stack_size; |
|||
impl->rewind_context = |
|||
boost::context::detail::make_fcontext(stack_base, impl->stack.size(), RewindStartFunc); |
|||
boost::context::detail::jump_fcontext(impl->rewind_context, this); |
|||
} |
|||
|
|||
void Fiber::YieldTo(std::shared_ptr<Fiber>& from, std::shared_ptr<Fiber>& to) { |
|||
ASSERT_MSG(from != nullptr, "Yielding fiber is null!"); |
|||
ASSERT_MSG(to != nullptr, "Next fiber is null!"); |
|||
to->guard.lock(); |
|||
to->previous_fiber = from; |
|||
auto transfer = boost::context::detail::jump_fcontext(to->impl->context, to.get()); |
|||
ASSERT(from->previous_fiber != nullptr); |
|||
from->previous_fiber->impl->context = transfer.fctx; |
|||
from->previous_fiber->guard.unlock(); |
|||
from->previous_fiber.reset(); |
|||
} |
|||
|
|||
std::shared_ptr<Fiber> Fiber::ThreadToFiber() { |
|||
std::shared_ptr<Fiber> fiber = std::shared_ptr<Fiber>{new Fiber()}; |
|||
fiber->guard.lock(); |
|||
fiber->is_thread_fiber = true; |
|||
return fiber; |
|||
} |
|||
|
|||
#endif
|
|||
} // namespace Common
|
|||
@ -0,0 +1,92 @@ |
|||
// Copyright 2020 yuzu Emulator Project |
|||
// Licensed under GPLv2 or any later version |
|||
// Refer to the license.txt file included. |
|||
|
|||
#pragma once |
|||
|
|||
#include <functional> |
|||
#include <memory> |
|||
|
|||
#include "common/common_types.h" |
|||
#include "common/spin_lock.h" |
|||
|
|||
#if !defined(_WIN32) && !defined(WIN32) |
|||
namespace boost::context::detail { |
|||
struct transfer_t; |
|||
} |
|||
#endif |
|||
|
|||
namespace Common { |
|||
|
|||
/** |
|||
* Fiber class |
|||
* a fiber is a userspace thread with it's own context. They can be used to |
|||
* implement coroutines, emulated threading systems and certain asynchronous |
|||
* patterns. |
|||
* |
|||
* This class implements fibers at a low level, thus allowing greater freedom |
|||
* to implement such patterns. This fiber class is 'threadsafe' only one fiber |
|||
* can be running at a time and threads will be locked while trying to yield to |
|||
* a running fiber until it yields. WARNING exchanging two running fibers between |
|||
* threads will cause a deadlock. In order to prevent a deadlock, each thread should |
|||
* have an intermediary fiber, you switch to the intermediary fiber of the current |
|||
* thread and then from it switch to the expected fiber. This way you can exchange |
|||
* 2 fibers within 2 different threads. |
|||
*/ |
|||
class Fiber { |
|||
public: |
|||
Fiber(std::function<void(void*)>&& entry_point_func, void* start_parameter); |
|||
~Fiber(); |
|||
|
|||
Fiber(const Fiber&) = delete; |
|||
Fiber& operator=(const Fiber&) = delete; |
|||
|
|||
Fiber(Fiber&&) = default; |
|||
Fiber& operator=(Fiber&&) = default; |
|||
|
|||
/// Yields control from Fiber 'from' to Fiber 'to' |
|||
/// Fiber 'from' must be the currently running fiber. |
|||
static void YieldTo(std::shared_ptr<Fiber>& from, std::shared_ptr<Fiber>& to); |
|||
static std::shared_ptr<Fiber> ThreadToFiber(); |
|||
|
|||
void SetRewindPoint(std::function<void(void*)>&& rewind_func, void* start_parameter); |
|||
|
|||
void Rewind(); |
|||
|
|||
/// Only call from main thread's fiber |
|||
void Exit(); |
|||
|
|||
/// Changes the start parameter of the fiber. Has no effect if the fiber already started |
|||
void SetStartParameter(void* new_parameter) { |
|||
start_parameter = new_parameter; |
|||
} |
|||
|
|||
private: |
|||
Fiber(); |
|||
|
|||
#if defined(_WIN32) || defined(WIN32) |
|||
void OnRewind(); |
|||
void Start(); |
|||
static void FiberStartFunc(void* fiber_parameter); |
|||
static void RewindStartFunc(void* fiber_parameter); |
|||
#else |
|||
void OnRewind(boost::context::detail::transfer_t& transfer); |
|||
void Start(boost::context::detail::transfer_t& transfer); |
|||
static void FiberStartFunc(boost::context::detail::transfer_t transfer); |
|||
static void RewindStartFunc(boost::context::detail::transfer_t transfer); |
|||
#endif |
|||
|
|||
struct FiberImpl; |
|||
|
|||
SpinLock guard{}; |
|||
std::function<void(void*)> entry_point; |
|||
std::function<void(void*)> rewind_point; |
|||
void* rewind_parameter{}; |
|||
void* start_parameter{}; |
|||
std::shared_ptr<Fiber> previous_fiber; |
|||
std::unique_ptr<FiberImpl> impl; |
|||
bool is_thread_fiber{}; |
|||
bool released{}; |
|||
}; |
|||
|
|||
} // namespace Common |
|||
@ -0,0 +1,54 @@ |
|||
// Copyright 2020 yuzu Emulator Project
|
|||
// Licensed under GPLv2 or any later version
|
|||
// Refer to the license.txt file included.
|
|||
|
|||
#include "common/spin_lock.h"
|
|||
|
|||
#if _MSC_VER
|
|||
#include <intrin.h>
|
|||
#if _M_AMD64
|
|||
#define __x86_64__ 1
|
|||
#endif
|
|||
#if _M_ARM64
|
|||
#define __aarch64__ 1
|
|||
#endif
|
|||
#else
|
|||
#if __x86_64__
|
|||
#include <xmmintrin.h>
|
|||
#endif
|
|||
#endif
|
|||
|
|||
namespace { |
|||
|
|||
void thread_pause() { |
|||
#if __x86_64__
|
|||
_mm_pause(); |
|||
#elif __aarch64__ && _MSC_VER
|
|||
__yield(); |
|||
#elif __aarch64__
|
|||
asm("yield"); |
|||
#endif
|
|||
} |
|||
|
|||
} // namespace
|
|||
|
|||
namespace Common { |
|||
|
|||
void SpinLock::lock() { |
|||
while (lck.test_and_set(std::memory_order_acquire)) { |
|||
thread_pause(); |
|||
} |
|||
} |
|||
|
|||
void SpinLock::unlock() { |
|||
lck.clear(std::memory_order_release); |
|||
} |
|||
|
|||
bool SpinLock::try_lock() { |
|||
if (lck.test_and_set(std::memory_order_acquire)) { |
|||
return false; |
|||
} |
|||
return true; |
|||
} |
|||
|
|||
} // namespace Common
|
|||
@ -0,0 +1,21 @@ |
|||
// Copyright 2020 yuzu Emulator Project |
|||
// Licensed under GPLv2 or any later version |
|||
// Refer to the license.txt file included. |
|||
|
|||
#pragma once |
|||
|
|||
#include <atomic> |
|||
|
|||
namespace Common { |
|||
|
|||
class SpinLock { |
|||
public: |
|||
void lock(); |
|||
void unlock(); |
|||
bool try_lock(); |
|||
|
|||
private: |
|||
std::atomic_flag lck = ATOMIC_FLAG_INIT; |
|||
}; |
|||
|
|||
} // namespace Common |
|||
@ -0,0 +1,92 @@ |
|||
// Copyright 2020 yuzu Emulator Project
|
|||
// Licensed under GPLv2 or any later version
|
|||
// Refer to the license.txt file included.
|
|||
|
|||
#include "common/uint128.h"
|
|||
#include "common/wall_clock.h"
|
|||
|
|||
#ifdef ARCHITECTURE_x86_64
|
|||
#include "common/x64/cpu_detect.h"
|
|||
#include "common/x64/native_clock.h"
|
|||
#endif
|
|||
|
|||
namespace Common { |
|||
|
|||
using base_timer = std::chrono::steady_clock; |
|||
using base_time_point = std::chrono::time_point<base_timer>; |
|||
|
|||
class StandardWallClock : public WallClock { |
|||
public: |
|||
StandardWallClock(u64 emulated_cpu_frequency, u64 emulated_clock_frequency) |
|||
: WallClock(emulated_cpu_frequency, emulated_clock_frequency, false) { |
|||
start_time = base_timer::now(); |
|||
} |
|||
|
|||
std::chrono::nanoseconds GetTimeNS() override { |
|||
base_time_point current = base_timer::now(); |
|||
auto elapsed = current - start_time; |
|||
return std::chrono::duration_cast<std::chrono::nanoseconds>(elapsed); |
|||
} |
|||
|
|||
std::chrono::microseconds GetTimeUS() override { |
|||
base_time_point current = base_timer::now(); |
|||
auto elapsed = current - start_time; |
|||
return std::chrono::duration_cast<std::chrono::microseconds>(elapsed); |
|||
} |
|||
|
|||
std::chrono::milliseconds GetTimeMS() override { |
|||
base_time_point current = base_timer::now(); |
|||
auto elapsed = current - start_time; |
|||
return std::chrono::duration_cast<std::chrono::milliseconds>(elapsed); |
|||
} |
|||
|
|||
u64 GetClockCycles() override { |
|||
std::chrono::nanoseconds time_now = GetTimeNS(); |
|||
const u128 temporary = |
|||
Common::Multiply64Into128(time_now.count(), emulated_clock_frequency); |
|||
return Common::Divide128On32(temporary, 1000000000).first; |
|||
} |
|||
|
|||
u64 GetCPUCycles() override { |
|||
std::chrono::nanoseconds time_now = GetTimeNS(); |
|||
const u128 temporary = Common::Multiply64Into128(time_now.count(), emulated_cpu_frequency); |
|||
return Common::Divide128On32(temporary, 1000000000).first; |
|||
} |
|||
|
|||
private: |
|||
base_time_point start_time; |
|||
}; |
|||
|
|||
#ifdef ARCHITECTURE_x86_64
|
|||
|
|||
std::unique_ptr<WallClock> CreateBestMatchingClock(u32 emulated_cpu_frequency, |
|||
u32 emulated_clock_frequency) { |
|||
const auto& caps = GetCPUCaps(); |
|||
u64 rtsc_frequency = 0; |
|||
if (caps.invariant_tsc) { |
|||
if (caps.base_frequency != 0) { |
|||
rtsc_frequency = static_cast<u64>(caps.base_frequency) * 1000000U; |
|||
} |
|||
if (rtsc_frequency == 0) { |
|||
rtsc_frequency = EstimateRDTSCFrequency(); |
|||
} |
|||
} |
|||
if (rtsc_frequency == 0) { |
|||
return std::make_unique<StandardWallClock>(emulated_cpu_frequency, |
|||
emulated_clock_frequency); |
|||
} else { |
|||
return std::make_unique<X64::NativeClock>(emulated_cpu_frequency, emulated_clock_frequency, |
|||
rtsc_frequency); |
|||
} |
|||
} |
|||
|
|||
#else
|
|||
|
|||
std::unique_ptr<WallClock> CreateBestMatchingClock(u32 emulated_cpu_frequency, |
|||
u32 emulated_clock_frequency) { |
|||
return std::make_unique<StandardWallClock>(emulated_cpu_frequency, emulated_clock_frequency); |
|||
} |
|||
|
|||
#endif
|
|||
|
|||
} // namespace Common
|
|||
@ -0,0 +1,51 @@ |
|||
// Copyright 2020 yuzu Emulator Project |
|||
// Licensed under GPLv2 or any later version |
|||
// Refer to the license.txt file included. |
|||
|
|||
#pragma once |
|||
|
|||
#include <chrono> |
|||
#include <memory> |
|||
|
|||
#include "common/common_types.h" |
|||
|
|||
namespace Common { |
|||
|
|||
class WallClock { |
|||
public: |
|||
/// Returns current wall time in nanoseconds |
|||
virtual std::chrono::nanoseconds GetTimeNS() = 0; |
|||
|
|||
/// Returns current wall time in microseconds |
|||
virtual std::chrono::microseconds GetTimeUS() = 0; |
|||
|
|||
/// Returns current wall time in milliseconds |
|||
virtual std::chrono::milliseconds GetTimeMS() = 0; |
|||
|
|||
/// Returns current wall time in emulated clock cycles |
|||
virtual u64 GetClockCycles() = 0; |
|||
|
|||
/// Returns current wall time in emulated cpu cycles |
|||
virtual u64 GetCPUCycles() = 0; |
|||
|
|||
/// Tells if the wall clock, uses the host CPU's hardware clock |
|||
bool IsNative() const { |
|||
return is_native; |
|||
} |
|||
|
|||
protected: |
|||
WallClock(u64 emulated_cpu_frequency, u64 emulated_clock_frequency, bool is_native) |
|||
: emulated_cpu_frequency{emulated_cpu_frequency}, |
|||
emulated_clock_frequency{emulated_clock_frequency}, is_native{is_native} {} |
|||
|
|||
u64 emulated_cpu_frequency; |
|||
u64 emulated_clock_frequency; |
|||
|
|||
private: |
|||
bool is_native; |
|||
}; |
|||
|
|||
std::unique_ptr<WallClock> CreateBestMatchingClock(u32 emulated_cpu_frequency, |
|||
u32 emulated_clock_frequency); |
|||
|
|||
} // namespace Common |
|||
@ -0,0 +1,95 @@ |
|||
// Copyright 2020 yuzu Emulator Project
|
|||
// Licensed under GPLv2 or any later version
|
|||
// Refer to the license.txt file included.
|
|||
|
|||
#include <chrono>
|
|||
#include <thread>
|
|||
|
|||
#ifdef _MSC_VER
|
|||
#include <intrin.h>
|
|||
#else
|
|||
#include <x86intrin.h>
|
|||
#endif
|
|||
|
|||
#include "common/uint128.h"
|
|||
#include "common/x64/native_clock.h"
|
|||
|
|||
namespace Common { |
|||
|
|||
u64 EstimateRDTSCFrequency() { |
|||
const auto milli_10 = std::chrono::milliseconds{10}; |
|||
// get current time
|
|||
_mm_mfence(); |
|||
const u64 tscStart = __rdtsc(); |
|||
const auto startTime = std::chrono::high_resolution_clock::now(); |
|||
// wait roughly 3 seconds
|
|||
while (true) { |
|||
auto milli = std::chrono::duration_cast<std::chrono::milliseconds>( |
|||
std::chrono::high_resolution_clock::now() - startTime); |
|||
if (milli.count() >= 3000) |
|||
break; |
|||
std::this_thread::sleep_for(milli_10); |
|||
} |
|||
const auto endTime = std::chrono::high_resolution_clock::now(); |
|||
_mm_mfence(); |
|||
const u64 tscEnd = __rdtsc(); |
|||
// calculate difference
|
|||
const u64 timer_diff = |
|||
std::chrono::duration_cast<std::chrono::nanoseconds>(endTime - startTime).count(); |
|||
const u64 tsc_diff = tscEnd - tscStart; |
|||
const u64 tsc_freq = MultiplyAndDivide64(tsc_diff, 1000000000ULL, timer_diff); |
|||
return tsc_freq; |
|||
} |
|||
|
|||
namespace X64 { |
|||
NativeClock::NativeClock(u64 emulated_cpu_frequency, u64 emulated_clock_frequency, |
|||
u64 rtsc_frequency) |
|||
: WallClock(emulated_cpu_frequency, emulated_clock_frequency, true), rtsc_frequency{ |
|||
rtsc_frequency} { |
|||
_mm_mfence(); |
|||
last_measure = __rdtsc(); |
|||
accumulated_ticks = 0U; |
|||
} |
|||
|
|||
u64 NativeClock::GetRTSC() { |
|||
rtsc_serialize.lock(); |
|||
_mm_mfence(); |
|||
const u64 current_measure = __rdtsc(); |
|||
u64 diff = current_measure - last_measure; |
|||
diff = diff & ~static_cast<u64>(static_cast<s64>(diff) >> 63); // max(diff, 0)
|
|||
if (current_measure > last_measure) { |
|||
last_measure = current_measure; |
|||
} |
|||
accumulated_ticks += diff; |
|||
rtsc_serialize.unlock(); |
|||
return accumulated_ticks; |
|||
} |
|||
|
|||
std::chrono::nanoseconds NativeClock::GetTimeNS() { |
|||
const u64 rtsc_value = GetRTSC(); |
|||
return std::chrono::nanoseconds{MultiplyAndDivide64(rtsc_value, 1000000000, rtsc_frequency)}; |
|||
} |
|||
|
|||
std::chrono::microseconds NativeClock::GetTimeUS() { |
|||
const u64 rtsc_value = GetRTSC(); |
|||
return std::chrono::microseconds{MultiplyAndDivide64(rtsc_value, 1000000, rtsc_frequency)}; |
|||
} |
|||
|
|||
std::chrono::milliseconds NativeClock::GetTimeMS() { |
|||
const u64 rtsc_value = GetRTSC(); |
|||
return std::chrono::milliseconds{MultiplyAndDivide64(rtsc_value, 1000, rtsc_frequency)}; |
|||
} |
|||
|
|||
u64 NativeClock::GetClockCycles() { |
|||
const u64 rtsc_value = GetRTSC(); |
|||
return MultiplyAndDivide64(rtsc_value, emulated_clock_frequency, rtsc_frequency); |
|||
} |
|||
|
|||
u64 NativeClock::GetCPUCycles() { |
|||
const u64 rtsc_value = GetRTSC(); |
|||
return MultiplyAndDivide64(rtsc_value, emulated_cpu_frequency, rtsc_frequency); |
|||
} |
|||
|
|||
} // namespace X64
|
|||
|
|||
} // namespace Common
|
|||
@ -0,0 +1,41 @@ |
|||
// Copyright 2020 yuzu Emulator Project |
|||
// Licensed under GPLv2 or any later version |
|||
// Refer to the license.txt file included. |
|||
|
|||
#pragma once |
|||
|
|||
#include <optional> |
|||
|
|||
#include "common/spin_lock.h" |
|||
#include "common/wall_clock.h" |
|||
|
|||
namespace Common { |
|||
|
|||
namespace X64 { |
|||
class NativeClock : public WallClock { |
|||
public: |
|||
NativeClock(u64 emulated_cpu_frequency, u64 emulated_clock_frequency, u64 rtsc_frequency); |
|||
|
|||
std::chrono::nanoseconds GetTimeNS() override; |
|||
|
|||
std::chrono::microseconds GetTimeUS() override; |
|||
|
|||
std::chrono::milliseconds GetTimeMS() override; |
|||
|
|||
u64 GetClockCycles() override; |
|||
|
|||
u64 GetCPUCycles() override; |
|||
|
|||
private: |
|||
u64 GetRTSC(); |
|||
|
|||
SpinLock rtsc_serialize{}; |
|||
u64 last_measure{}; |
|||
u64 accumulated_ticks{}; |
|||
u64 rtsc_frequency; |
|||
}; |
|||
} // namespace X64 |
|||
|
|||
u64 EstimateRDTSCFrequency(); |
|||
|
|||
} // namespace Common |
|||
@ -0,0 +1,206 @@ |
|||
// Copyright 2020 yuzu Emulator Project
|
|||
// Licensed under GPLv2 or any later version
|
|||
// Refer to the license.txt file included.
|
|||
|
|||
#include "core/host_timing.h"
|
|||
|
|||
#include <algorithm>
|
|||
#include <mutex>
|
|||
#include <string>
|
|||
#include <tuple>
|
|||
|
|||
#include "common/assert.h"
|
|||
#include "core/core_timing_util.h"
|
|||
|
|||
namespace Core::HostTiming { |
|||
|
|||
std::shared_ptr<EventType> CreateEvent(std::string name, TimedCallback&& callback) { |
|||
return std::make_shared<EventType>(std::move(callback), std::move(name)); |
|||
} |
|||
|
|||
struct CoreTiming::Event { |
|||
u64 time; |
|||
u64 fifo_order; |
|||
u64 userdata; |
|||
std::weak_ptr<EventType> type; |
|||
|
|||
// Sort by time, unless the times are the same, in which case sort by
|
|||
// the order added to the queue
|
|||
friend bool operator>(const Event& left, const Event& right) { |
|||
return std::tie(left.time, left.fifo_order) > std::tie(right.time, right.fifo_order); |
|||
} |
|||
|
|||
friend bool operator<(const Event& left, const Event& right) { |
|||
return std::tie(left.time, left.fifo_order) < std::tie(right.time, right.fifo_order); |
|||
} |
|||
}; |
|||
|
|||
CoreTiming::CoreTiming() { |
|||
clock = |
|||
Common::CreateBestMatchingClock(Core::Hardware::BASE_CLOCK_RATE, Core::Hardware::CNTFREQ); |
|||
} |
|||
|
|||
CoreTiming::~CoreTiming() = default; |
|||
|
|||
void CoreTiming::ThreadEntry(CoreTiming& instance) { |
|||
instance.ThreadLoop(); |
|||
} |
|||
|
|||
void CoreTiming::Initialize() { |
|||
event_fifo_id = 0; |
|||
const auto empty_timed_callback = [](u64, s64) {}; |
|||
ev_lost = CreateEvent("_lost_event", empty_timed_callback); |
|||
timer_thread = std::make_unique<std::thread>(ThreadEntry, std::ref(*this)); |
|||
} |
|||
|
|||
void CoreTiming::Shutdown() { |
|||
paused = true; |
|||
shutting_down = true; |
|||
event.Set(); |
|||
timer_thread->join(); |
|||
ClearPendingEvents(); |
|||
timer_thread.reset(); |
|||
has_started = false; |
|||
} |
|||
|
|||
void CoreTiming::Pause(bool is_paused) { |
|||
paused = is_paused; |
|||
} |
|||
|
|||
void CoreTiming::SyncPause(bool is_paused) { |
|||
if (is_paused == paused && paused_set == paused) { |
|||
return; |
|||
} |
|||
Pause(is_paused); |
|||
event.Set(); |
|||
while (paused_set != is_paused) |
|||
; |
|||
} |
|||
|
|||
bool CoreTiming::IsRunning() const { |
|||
return !paused_set; |
|||
} |
|||
|
|||
bool CoreTiming::HasPendingEvents() const { |
|||
return !(wait_set && event_queue.empty()); |
|||
} |
|||
|
|||
void CoreTiming::ScheduleEvent(s64 ns_into_future, const std::shared_ptr<EventType>& event_type, |
|||
u64 userdata) { |
|||
basic_lock.lock(); |
|||
const u64 timeout = static_cast<u64>(GetGlobalTimeNs().count() + ns_into_future); |
|||
|
|||
event_queue.emplace_back(Event{timeout, event_fifo_id++, userdata, event_type}); |
|||
|
|||
std::push_heap(event_queue.begin(), event_queue.end(), std::greater<>()); |
|||
basic_lock.unlock(); |
|||
event.Set(); |
|||
} |
|||
|
|||
void CoreTiming::UnscheduleEvent(const std::shared_ptr<EventType>& event_type, u64 userdata) { |
|||
basic_lock.lock(); |
|||
const auto itr = std::remove_if(event_queue.begin(), event_queue.end(), [&](const Event& e) { |
|||
return e.type.lock().get() == event_type.get() && e.userdata == userdata; |
|||
}); |
|||
|
|||
// Removing random items breaks the invariant so we have to re-establish it.
|
|||
if (itr != event_queue.end()) { |
|||
event_queue.erase(itr, event_queue.end()); |
|||
std::make_heap(event_queue.begin(), event_queue.end(), std::greater<>()); |
|||
} |
|||
basic_lock.unlock(); |
|||
} |
|||
|
|||
void CoreTiming::AddTicks(std::size_t core_index, u64 ticks) { |
|||
ticks_count[core_index] += ticks; |
|||
} |
|||
|
|||
void CoreTiming::ResetTicks(std::size_t core_index) { |
|||
ticks_count[core_index] = 0; |
|||
} |
|||
|
|||
u64 CoreTiming::GetCPUTicks() const { |
|||
return clock->GetCPUCycles(); |
|||
} |
|||
|
|||
u64 CoreTiming::GetClockTicks() const { |
|||
return clock->GetClockCycles(); |
|||
} |
|||
|
|||
void CoreTiming::ClearPendingEvents() { |
|||
event_queue.clear(); |
|||
} |
|||
|
|||
void CoreTiming::RemoveEvent(const std::shared_ptr<EventType>& event_type) { |
|||
basic_lock.lock(); |
|||
|
|||
const auto itr = std::remove_if(event_queue.begin(), event_queue.end(), [&](const Event& e) { |
|||
return e.type.lock().get() == event_type.get(); |
|||
}); |
|||
|
|||
// Removing random items breaks the invariant so we have to re-establish it.
|
|||
if (itr != event_queue.end()) { |
|||
event_queue.erase(itr, event_queue.end()); |
|||
std::make_heap(event_queue.begin(), event_queue.end(), std::greater<>()); |
|||
} |
|||
basic_lock.unlock(); |
|||
} |
|||
|
|||
std::optional<u64> CoreTiming::Advance() { |
|||
advance_lock.lock(); |
|||
basic_lock.lock(); |
|||
global_timer = GetGlobalTimeNs().count(); |
|||
|
|||
while (!event_queue.empty() && event_queue.front().time <= global_timer) { |
|||
Event evt = std::move(event_queue.front()); |
|||
std::pop_heap(event_queue.begin(), event_queue.end(), std::greater<>()); |
|||
event_queue.pop_back(); |
|||
basic_lock.unlock(); |
|||
|
|||
if (auto event_type{evt.type.lock()}) { |
|||
event_type->callback(evt.userdata, global_timer - evt.time); |
|||
} |
|||
|
|||
basic_lock.lock(); |
|||
} |
|||
|
|||
if (!event_queue.empty()) { |
|||
const u64 next_time = event_queue.front().time - global_timer; |
|||
basic_lock.unlock(); |
|||
advance_lock.unlock(); |
|||
return next_time; |
|||
} else { |
|||
basic_lock.unlock(); |
|||
advance_lock.unlock(); |
|||
return std::nullopt; |
|||
} |
|||
} |
|||
|
|||
void CoreTiming::ThreadLoop() { |
|||
has_started = true; |
|||
while (!shutting_down) { |
|||
while (!paused) { |
|||
paused_set = false; |
|||
const auto next_time = Advance(); |
|||
if (next_time) { |
|||
std::chrono::nanoseconds next_time_ns = std::chrono::nanoseconds(*next_time); |
|||
event.WaitFor(next_time_ns); |
|||
} else { |
|||
wait_set = true; |
|||
event.Wait(); |
|||
} |
|||
wait_set = false; |
|||
} |
|||
paused_set = true; |
|||
} |
|||
} |
|||
|
|||
std::chrono::nanoseconds CoreTiming::GetGlobalTimeNs() const { |
|||
return clock->GetTimeNS(); |
|||
} |
|||
|
|||
std::chrono::microseconds CoreTiming::GetGlobalTimeUs() const { |
|||
return clock->GetTimeUS(); |
|||
} |
|||
|
|||
} // namespace Core::HostTiming
|
|||
@ -0,0 +1,160 @@ |
|||
// Copyright 2020 yuzu Emulator Project |
|||
// Licensed under GPLv2 or any later version |
|||
// Refer to the license.txt file included. |
|||
|
|||
#pragma once |
|||
|
|||
#include <atomic> |
|||
#include <chrono> |
|||
#include <functional> |
|||
#include <memory> |
|||
#include <mutex> |
|||
#include <optional> |
|||
#include <string> |
|||
#include <thread> |
|||
#include <vector> |
|||
|
|||
#include "common/common_types.h" |
|||
#include "common/spin_lock.h" |
|||
#include "common/thread.h" |
|||
#include "common/threadsafe_queue.h" |
|||
#include "common/wall_clock.h" |
|||
#include "core/hardware_properties.h" |
|||
|
|||
namespace Core::HostTiming { |
|||
|
|||
/// A callback that may be scheduled for a particular core timing event. |
|||
using TimedCallback = std::function<void(u64 userdata, s64 cycles_late)>; |
|||
|
|||
/// Contains the characteristics of a particular event. |
|||
struct EventType { |
|||
EventType(TimedCallback&& callback, std::string&& name) |
|||
: callback{std::move(callback)}, name{std::move(name)} {} |
|||
|
|||
/// The event's callback function. |
|||
TimedCallback callback; |
|||
/// A pointer to the name of the event. |
|||
const std::string name; |
|||
}; |
|||
|
|||
/** |
|||
* This is a system to schedule events into the emulated machine's future. Time is measured |
|||
* in main CPU clock cycles. |
|||
* |
|||
* To schedule an event, you first have to register its type. This is where you pass in the |
|||
* callback. You then schedule events using the type id you get back. |
|||
* |
|||
* The int cyclesLate that the callbacks get is how many cycles late it was. |
|||
* So to schedule a new event on a regular basis: |
|||
* inside callback: |
|||
* ScheduleEvent(periodInCycles - cyclesLate, callback, "whatever") |
|||
*/ |
|||
class CoreTiming { |
|||
public: |
|||
CoreTiming(); |
|||
~CoreTiming(); |
|||
|
|||
CoreTiming(const CoreTiming&) = delete; |
|||
CoreTiming(CoreTiming&&) = delete; |
|||
|
|||
CoreTiming& operator=(const CoreTiming&) = delete; |
|||
CoreTiming& operator=(CoreTiming&&) = delete; |
|||
|
|||
/// CoreTiming begins at the boundary of timing slice -1. An initial call to Advance() is |
|||
/// required to end slice - 1 and start slice 0 before the first cycle of code is executed. |
|||
void Initialize(); |
|||
|
|||
/// Tears down all timing related functionality. |
|||
void Shutdown(); |
|||
|
|||
/// Pauses/Unpauses the execution of the timer thread. |
|||
void Pause(bool is_paused); |
|||
|
|||
/// Pauses/Unpauses the execution of the timer thread and waits until paused. |
|||
void SyncPause(bool is_paused); |
|||
|
|||
/// Checks if core timing is running. |
|||
bool IsRunning() const; |
|||
|
|||
/// Checks if the timer thread has started. |
|||
bool HasStarted() const { |
|||
return has_started; |
|||
} |
|||
|
|||
/// Checks if there are any pending time events. |
|||
bool HasPendingEvents() const; |
|||
|
|||
/// Schedules an event in core timing |
|||
void ScheduleEvent(s64 ns_into_future, const std::shared_ptr<EventType>& event_type, |
|||
u64 userdata = 0); |
|||
|
|||
void UnscheduleEvent(const std::shared_ptr<EventType>& event_type, u64 userdata); |
|||
|
|||
/// We only permit one event of each type in the queue at a time. |
|||
void RemoveEvent(const std::shared_ptr<EventType>& event_type); |
|||
|
|||
void AddTicks(std::size_t core_index, u64 ticks); |
|||
|
|||
void ResetTicks(std::size_t core_index); |
|||
|
|||
/// Returns current time in emulated CPU cycles |
|||
u64 GetCPUTicks() const; |
|||
|
|||
/// Returns current time in emulated in Clock cycles |
|||
u64 GetClockTicks() const; |
|||
|
|||
/// Returns current time in microseconds. |
|||
std::chrono::microseconds GetGlobalTimeUs() const; |
|||
|
|||
/// Returns current time in nanoseconds. |
|||
std::chrono::nanoseconds GetGlobalTimeNs() const; |
|||
|
|||
/// Checks for events manually and returns time in nanoseconds for next event, threadsafe. |
|||
std::optional<u64> Advance(); |
|||
|
|||
private: |
|||
struct Event; |
|||
|
|||
/// Clear all pending events. This should ONLY be done on exit. |
|||
void ClearPendingEvents(); |
|||
|
|||
static void ThreadEntry(CoreTiming& instance); |
|||
void ThreadLoop(); |
|||
|
|||
std::unique_ptr<Common::WallClock> clock; |
|||
|
|||
u64 global_timer = 0; |
|||
|
|||
std::chrono::nanoseconds start_point; |
|||
|
|||
// The queue is a min-heap using std::make_heap/push_heap/pop_heap. |
|||
// We don't use std::priority_queue because we need to be able to serialize, unserialize and |
|||
// erase arbitrary events (RemoveEvent()) regardless of the queue order. These aren't |
|||
// accomodated by the standard adaptor class. |
|||
std::vector<Event> event_queue; |
|||
u64 event_fifo_id = 0; |
|||
|
|||
std::shared_ptr<EventType> ev_lost; |
|||
Common::Event event{}; |
|||
Common::SpinLock basic_lock{}; |
|||
Common::SpinLock advance_lock{}; |
|||
std::unique_ptr<std::thread> timer_thread; |
|||
std::atomic<bool> paused{}; |
|||
std::atomic<bool> paused_set{}; |
|||
std::atomic<bool> wait_set{}; |
|||
std::atomic<bool> shutting_down{}; |
|||
std::atomic<bool> has_started{}; |
|||
|
|||
std::array<std::atomic<u64>, Core::Hardware::NUM_CPU_CORES> ticks_count{}; |
|||
}; |
|||
|
|||
/// Creates a core timing event with the given name and callback. |
|||
/// |
|||
/// @param name The name of the core timing event to create. |
|||
/// @param callback The callback to execute for the event. |
|||
/// |
|||
/// @returns An EventType instance representing the created event. |
|||
/// |
|||
std::shared_ptr<EventType> CreateEvent(std::string name, TimedCallback&& callback); |
|||
|
|||
} // namespace Core::HostTiming |
|||
@ -0,0 +1,358 @@ |
|||
// Copyright 2020 yuzu Emulator Project
|
|||
// Licensed under GPLv2 or any later version
|
|||
// Refer to the license.txt file included.
|
|||
|
|||
#include <atomic>
|
|||
#include <cstdlib>
|
|||
#include <functional>
|
|||
#include <memory>
|
|||
#include <thread>
|
|||
#include <unordered_map>
|
|||
#include <vector>
|
|||
|
|||
#include <catch2/catch.hpp>
|
|||
#include <math.h>
|
|||
#include "common/common_types.h"
|
|||
#include "common/fiber.h"
|
|||
#include "common/spin_lock.h"
|
|||
|
|||
namespace Common { |
|||
|
|||
class TestControl1 { |
|||
public: |
|||
TestControl1() = default; |
|||
|
|||
void DoWork(); |
|||
|
|||
void ExecuteThread(u32 id); |
|||
|
|||
std::unordered_map<std::thread::id, u32> ids; |
|||
std::vector<std::shared_ptr<Common::Fiber>> thread_fibers; |
|||
std::vector<std::shared_ptr<Common::Fiber>> work_fibers; |
|||
std::vector<u32> items; |
|||
std::vector<u32> results; |
|||
}; |
|||
|
|||
static void WorkControl1(void* control) { |
|||
auto* test_control = static_cast<TestControl1*>(control); |
|||
test_control->DoWork(); |
|||
} |
|||
|
|||
void TestControl1::DoWork() { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
u32 value = items[id]; |
|||
for (u32 i = 0; i < id; i++) { |
|||
value++; |
|||
} |
|||
results[id] = value; |
|||
Fiber::YieldTo(work_fibers[id], thread_fibers[id]); |
|||
} |
|||
|
|||
void TestControl1::ExecuteThread(u32 id) { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
ids[this_id] = id; |
|||
auto thread_fiber = Fiber::ThreadToFiber(); |
|||
thread_fibers[id] = thread_fiber; |
|||
work_fibers[id] = std::make_shared<Fiber>(std::function<void(void*)>{WorkControl1}, this); |
|||
items[id] = rand() % 256; |
|||
Fiber::YieldTo(thread_fibers[id], work_fibers[id]); |
|||
thread_fibers[id]->Exit(); |
|||
} |
|||
|
|||
static void ThreadStart1(u32 id, TestControl1& test_control) { |
|||
test_control.ExecuteThread(id); |
|||
} |
|||
|
|||
/** This test checks for fiber setup configuration and validates that fibers are
|
|||
* doing all the work required. |
|||
*/ |
|||
TEST_CASE("Fibers::Setup", "[common]") { |
|||
constexpr u32 num_threads = 7; |
|||
TestControl1 test_control{}; |
|||
test_control.thread_fibers.resize(num_threads); |
|||
test_control.work_fibers.resize(num_threads); |
|||
test_control.items.resize(num_threads, 0); |
|||
test_control.results.resize(num_threads, 0); |
|||
std::vector<std::thread> threads; |
|||
for (u32 i = 0; i < num_threads; i++) { |
|||
threads.emplace_back(ThreadStart1, i, std::ref(test_control)); |
|||
} |
|||
for (u32 i = 0; i < num_threads; i++) { |
|||
threads[i].join(); |
|||
} |
|||
for (u32 i = 0; i < num_threads; i++) { |
|||
REQUIRE(test_control.items[i] + i == test_control.results[i]); |
|||
} |
|||
} |
|||
|
|||
class TestControl2 { |
|||
public: |
|||
TestControl2() = default; |
|||
|
|||
void DoWork1() { |
|||
trap2 = false; |
|||
while (trap.load()) |
|||
; |
|||
for (u32 i = 0; i < 12000; i++) { |
|||
value1 += i; |
|||
} |
|||
Fiber::YieldTo(fiber1, fiber3); |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
assert1 = id == 1; |
|||
value2 += 5000; |
|||
Fiber::YieldTo(fiber1, thread_fibers[id]); |
|||
} |
|||
|
|||
void DoWork2() { |
|||
while (trap2.load()) |
|||
; |
|||
value2 = 2000; |
|||
trap = false; |
|||
Fiber::YieldTo(fiber2, fiber1); |
|||
assert3 = false; |
|||
} |
|||
|
|||
void DoWork3() { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
assert2 = id == 0; |
|||
value1 += 1000; |
|||
Fiber::YieldTo(fiber3, thread_fibers[id]); |
|||
} |
|||
|
|||
void ExecuteThread(u32 id); |
|||
|
|||
void CallFiber1() { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
Fiber::YieldTo(thread_fibers[id], fiber1); |
|||
} |
|||
|
|||
void CallFiber2() { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
Fiber::YieldTo(thread_fibers[id], fiber2); |
|||
} |
|||
|
|||
void Exit(); |
|||
|
|||
bool assert1{}; |
|||
bool assert2{}; |
|||
bool assert3{true}; |
|||
u32 value1{}; |
|||
u32 value2{}; |
|||
std::atomic<bool> trap{true}; |
|||
std::atomic<bool> trap2{true}; |
|||
std::unordered_map<std::thread::id, u32> ids; |
|||
std::vector<std::shared_ptr<Common::Fiber>> thread_fibers; |
|||
std::shared_ptr<Common::Fiber> fiber1; |
|||
std::shared_ptr<Common::Fiber> fiber2; |
|||
std::shared_ptr<Common::Fiber> fiber3; |
|||
}; |
|||
|
|||
static void WorkControl2_1(void* control) { |
|||
auto* test_control = static_cast<TestControl2*>(control); |
|||
test_control->DoWork1(); |
|||
} |
|||
|
|||
static void WorkControl2_2(void* control) { |
|||
auto* test_control = static_cast<TestControl2*>(control); |
|||
test_control->DoWork2(); |
|||
} |
|||
|
|||
static void WorkControl2_3(void* control) { |
|||
auto* test_control = static_cast<TestControl2*>(control); |
|||
test_control->DoWork3(); |
|||
} |
|||
|
|||
void TestControl2::ExecuteThread(u32 id) { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
ids[this_id] = id; |
|||
auto thread_fiber = Fiber::ThreadToFiber(); |
|||
thread_fibers[id] = thread_fiber; |
|||
} |
|||
|
|||
void TestControl2::Exit() { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
thread_fibers[id]->Exit(); |
|||
} |
|||
|
|||
static void ThreadStart2_1(u32 id, TestControl2& test_control) { |
|||
test_control.ExecuteThread(id); |
|||
test_control.CallFiber1(); |
|||
test_control.Exit(); |
|||
} |
|||
|
|||
static void ThreadStart2_2(u32 id, TestControl2& test_control) { |
|||
test_control.ExecuteThread(id); |
|||
test_control.CallFiber2(); |
|||
test_control.Exit(); |
|||
} |
|||
|
|||
/** This test checks for fiber thread exchange configuration and validates that fibers are
|
|||
* that a fiber has been succesfully transfered from one thread to another and that the TLS |
|||
* region of the thread is kept while changing fibers. |
|||
*/ |
|||
TEST_CASE("Fibers::InterExchange", "[common]") { |
|||
TestControl2 test_control{}; |
|||
test_control.thread_fibers.resize(2); |
|||
test_control.fiber1 = |
|||
std::make_shared<Fiber>(std::function<void(void*)>{WorkControl2_1}, &test_control); |
|||
test_control.fiber2 = |
|||
std::make_shared<Fiber>(std::function<void(void*)>{WorkControl2_2}, &test_control); |
|||
test_control.fiber3 = |
|||
std::make_shared<Fiber>(std::function<void(void*)>{WorkControl2_3}, &test_control); |
|||
std::thread thread1(ThreadStart2_1, 0, std::ref(test_control)); |
|||
std::thread thread2(ThreadStart2_2, 1, std::ref(test_control)); |
|||
thread1.join(); |
|||
thread2.join(); |
|||
REQUIRE(test_control.assert1); |
|||
REQUIRE(test_control.assert2); |
|||
REQUIRE(test_control.assert3); |
|||
REQUIRE(test_control.value2 == 7000); |
|||
u32 cal_value = 0; |
|||
for (u32 i = 0; i < 12000; i++) { |
|||
cal_value += i; |
|||
} |
|||
cal_value += 1000; |
|||
REQUIRE(test_control.value1 == cal_value); |
|||
} |
|||
|
|||
class TestControl3 { |
|||
public: |
|||
TestControl3() = default; |
|||
|
|||
void DoWork1() { |
|||
value1 += 1; |
|||
Fiber::YieldTo(fiber1, fiber2); |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
value3 += 1; |
|||
Fiber::YieldTo(fiber1, thread_fibers[id]); |
|||
} |
|||
|
|||
void DoWork2() { |
|||
value2 += 1; |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
Fiber::YieldTo(fiber2, thread_fibers[id]); |
|||
} |
|||
|
|||
void ExecuteThread(u32 id); |
|||
|
|||
void CallFiber1() { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
Fiber::YieldTo(thread_fibers[id], fiber1); |
|||
} |
|||
|
|||
void Exit(); |
|||
|
|||
u32 value1{}; |
|||
u32 value2{}; |
|||
u32 value3{}; |
|||
std::unordered_map<std::thread::id, u32> ids; |
|||
std::vector<std::shared_ptr<Common::Fiber>> thread_fibers; |
|||
std::shared_ptr<Common::Fiber> fiber1; |
|||
std::shared_ptr<Common::Fiber> fiber2; |
|||
}; |
|||
|
|||
static void WorkControl3_1(void* control) { |
|||
auto* test_control = static_cast<TestControl3*>(control); |
|||
test_control->DoWork1(); |
|||
} |
|||
|
|||
static void WorkControl3_2(void* control) { |
|||
auto* test_control = static_cast<TestControl3*>(control); |
|||
test_control->DoWork2(); |
|||
} |
|||
|
|||
void TestControl3::ExecuteThread(u32 id) { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
ids[this_id] = id; |
|||
auto thread_fiber = Fiber::ThreadToFiber(); |
|||
thread_fibers[id] = thread_fiber; |
|||
} |
|||
|
|||
void TestControl3::Exit() { |
|||
std::thread::id this_id = std::this_thread::get_id(); |
|||
u32 id = ids[this_id]; |
|||
thread_fibers[id]->Exit(); |
|||
} |
|||
|
|||
static void ThreadStart3(u32 id, TestControl3& test_control) { |
|||
test_control.ExecuteThread(id); |
|||
test_control.CallFiber1(); |
|||
test_control.Exit(); |
|||
} |
|||
|
|||
/** This test checks for one two threads racing for starting the same fiber.
|
|||
* It checks execution occured in an ordered manner and by no time there were |
|||
* two contexts at the same time. |
|||
*/ |
|||
TEST_CASE("Fibers::StartRace", "[common]") { |
|||
TestControl3 test_control{}; |
|||
test_control.thread_fibers.resize(2); |
|||
test_control.fiber1 = |
|||
std::make_shared<Fiber>(std::function<void(void*)>{WorkControl3_1}, &test_control); |
|||
test_control.fiber2 = |
|||
std::make_shared<Fiber>(std::function<void(void*)>{WorkControl3_2}, &test_control); |
|||
std::thread thread1(ThreadStart3, 0, std::ref(test_control)); |
|||
std::thread thread2(ThreadStart3, 1, std::ref(test_control)); |
|||
thread1.join(); |
|||
thread2.join(); |
|||
REQUIRE(test_control.value1 == 1); |
|||
REQUIRE(test_control.value2 == 1); |
|||
REQUIRE(test_control.value3 == 1); |
|||
} |
|||
|
|||
class TestControl4; |
|||
|
|||
static void WorkControl4(void* control); |
|||
|
|||
class TestControl4 { |
|||
public: |
|||
TestControl4() { |
|||
fiber1 = std::make_shared<Fiber>(std::function<void(void*)>{WorkControl4}, this); |
|||
goal_reached = false; |
|||
rewinded = false; |
|||
} |
|||
|
|||
void Execute() { |
|||
thread_fiber = Fiber::ThreadToFiber(); |
|||
Fiber::YieldTo(thread_fiber, fiber1); |
|||
thread_fiber->Exit(); |
|||
} |
|||
|
|||
void DoWork() { |
|||
fiber1->SetRewindPoint(std::function<void(void*)>{WorkControl4}, this); |
|||
if (rewinded) { |
|||
goal_reached = true; |
|||
Fiber::YieldTo(fiber1, thread_fiber); |
|||
} |
|||
rewinded = true; |
|||
fiber1->Rewind(); |
|||
} |
|||
|
|||
std::shared_ptr<Common::Fiber> fiber1; |
|||
std::shared_ptr<Common::Fiber> thread_fiber; |
|||
bool goal_reached; |
|||
bool rewinded; |
|||
}; |
|||
|
|||
static void WorkControl4(void* control) { |
|||
auto* test_control = static_cast<TestControl4*>(control); |
|||
test_control->DoWork(); |
|||
} |
|||
|
|||
TEST_CASE("Fibers::Rewind", "[common]") { |
|||
TestControl4 test_control{}; |
|||
test_control.Execute(); |
|||
REQUIRE(test_control.goal_reached); |
|||
REQUIRE(test_control.rewinded); |
|||
} |
|||
|
|||
} // namespace Common
|
|||
@ -0,0 +1,142 @@ |
|||
// Copyright 2016 Dolphin Emulator Project / 2017 Dolphin Emulator Project
|
|||
// Licensed under GPLv2+
|
|||
// Refer to the license.txt file included.
|
|||
|
|||
#include <catch2/catch.hpp>
|
|||
|
|||
#include <array>
|
|||
#include <bitset>
|
|||
#include <cstdlib>
|
|||
#include <memory>
|
|||
#include <string>
|
|||
|
|||
#include "common/file_util.h"
|
|||
#include "core/core.h"
|
|||
#include "core/host_timing.h"
|
|||
|
|||
// Numbers are chosen randomly to make sure the correct one is given.
|
|||
static constexpr std::array<u64, 5> CB_IDS{{42, 144, 93, 1026, UINT64_C(0xFFFF7FFFF7FFFF)}}; |
|||
static constexpr int MAX_SLICE_LENGTH = 10000; // Copied from CoreTiming internals
|
|||
static constexpr std::array<u64, 5> calls_order{{2, 0, 1, 4, 3}}; |
|||
static std::array<s64, 5> delays{}; |
|||
|
|||
static std::bitset<CB_IDS.size()> callbacks_ran_flags; |
|||
static u64 expected_callback = 0; |
|||
|
|||
template <unsigned int IDX> |
|||
void HostCallbackTemplate(u64 userdata, s64 nanoseconds_late) { |
|||
static_assert(IDX < CB_IDS.size(), "IDX out of range"); |
|||
callbacks_ran_flags.set(IDX); |
|||
REQUIRE(CB_IDS[IDX] == userdata); |
|||
REQUIRE(CB_IDS[IDX] == CB_IDS[calls_order[expected_callback]]); |
|||
delays[IDX] = nanoseconds_late; |
|||
++expected_callback; |
|||
} |
|||
|
|||
struct ScopeInit final { |
|||
ScopeInit() { |
|||
core_timing.Initialize(); |
|||
} |
|||
~ScopeInit() { |
|||
core_timing.Shutdown(); |
|||
} |
|||
|
|||
Core::HostTiming::CoreTiming core_timing; |
|||
}; |
|||
|
|||
#pragma optimize("", off)
|
|||
|
|||
static u64 TestTimerSpeed(Core::HostTiming::CoreTiming& core_timing) { |
|||
u64 start = core_timing.GetGlobalTimeNs().count(); |
|||
u64 placebo = 0; |
|||
for (std::size_t i = 0; i < 1000; i++) { |
|||
placebo += core_timing.GetGlobalTimeNs().count(); |
|||
} |
|||
u64 end = core_timing.GetGlobalTimeNs().count(); |
|||
return (end - start); |
|||
} |
|||
|
|||
#pragma optimize("", on)
|
|||
|
|||
TEST_CASE("HostTiming[BasicOrder]", "[core]") { |
|||
ScopeInit guard; |
|||
auto& core_timing = guard.core_timing; |
|||
std::vector<std::shared_ptr<Core::HostTiming::EventType>> events{ |
|||
Core::HostTiming::CreateEvent("callbackA", HostCallbackTemplate<0>), |
|||
Core::HostTiming::CreateEvent("callbackB", HostCallbackTemplate<1>), |
|||
Core::HostTiming::CreateEvent("callbackC", HostCallbackTemplate<2>), |
|||
Core::HostTiming::CreateEvent("callbackD", HostCallbackTemplate<3>), |
|||
Core::HostTiming::CreateEvent("callbackE", HostCallbackTemplate<4>), |
|||
}; |
|||
|
|||
expected_callback = 0; |
|||
|
|||
core_timing.SyncPause(true); |
|||
|
|||
u64 one_micro = 1000U; |
|||
for (std::size_t i = 0; i < events.size(); i++) { |
|||
u64 order = calls_order[i]; |
|||
core_timing.ScheduleEvent(i * one_micro + 100U, events[order], CB_IDS[order]); |
|||
} |
|||
/// test pause
|
|||
REQUIRE(callbacks_ran_flags.none()); |
|||
|
|||
core_timing.Pause(false); // No need to sync
|
|||
|
|||
while (core_timing.HasPendingEvents()) |
|||
; |
|||
|
|||
REQUIRE(callbacks_ran_flags.all()); |
|||
|
|||
for (std::size_t i = 0; i < delays.size(); i++) { |
|||
const double delay = static_cast<double>(delays[i]); |
|||
const double micro = delay / 1000.0f; |
|||
const double mili = micro / 1000.0f; |
|||
printf("HostTimer Pausing Delay[%zu]: %.3f %.6f\n", i, micro, mili); |
|||
} |
|||
} |
|||
|
|||
TEST_CASE("HostTiming[BasicOrderNoPausing]", "[core]") { |
|||
ScopeInit guard; |
|||
auto& core_timing = guard.core_timing; |
|||
std::vector<std::shared_ptr<Core::HostTiming::EventType>> events{ |
|||
Core::HostTiming::CreateEvent("callbackA", HostCallbackTemplate<0>), |
|||
Core::HostTiming::CreateEvent("callbackB", HostCallbackTemplate<1>), |
|||
Core::HostTiming::CreateEvent("callbackC", HostCallbackTemplate<2>), |
|||
Core::HostTiming::CreateEvent("callbackD", HostCallbackTemplate<3>), |
|||
Core::HostTiming::CreateEvent("callbackE", HostCallbackTemplate<4>), |
|||
}; |
|||
|
|||
core_timing.SyncPause(true); |
|||
core_timing.SyncPause(false); |
|||
|
|||
expected_callback = 0; |
|||
|
|||
u64 start = core_timing.GetGlobalTimeNs().count(); |
|||
u64 one_micro = 1000U; |
|||
for (std::size_t i = 0; i < events.size(); i++) { |
|||
u64 order = calls_order[i]; |
|||
core_timing.ScheduleEvent(i * one_micro + 100U, events[order], CB_IDS[order]); |
|||
} |
|||
u64 end = core_timing.GetGlobalTimeNs().count(); |
|||
const double scheduling_time = static_cast<double>(end - start); |
|||
const double timer_time = static_cast<double>(TestTimerSpeed(core_timing)); |
|||
|
|||
while (core_timing.HasPendingEvents()) |
|||
; |
|||
|
|||
REQUIRE(callbacks_ran_flags.all()); |
|||
|
|||
for (std::size_t i = 0; i < delays.size(); i++) { |
|||
const double delay = static_cast<double>(delays[i]); |
|||
const double micro = delay / 1000.0f; |
|||
const double mili = micro / 1000.0f; |
|||
printf("HostTimer No Pausing Delay[%zu]: %.3f %.6f\n", i, micro, mili); |
|||
} |
|||
|
|||
const double micro = scheduling_time / 1000.0f; |
|||
const double mili = micro / 1000.0f; |
|||
printf("HostTimer No Pausing Scheduling Time: %.3f %.6f\n", micro, mili); |
|||
printf("HostTimer No Pausing Timer Time: %.3f %.6f\n", timer_time / 1000.f, |
|||
timer_time / 1000000.f); |
|||
} |
|||
Write
Preview
Loading…
Cancel
Save
Reference in new issue