mirror of
https://git.eden-emu.dev/eden-emu/eden
synced 2026-04-10 05:28:56 +02:00
for core stuff: just remove unique ptrs that dont need any pointer stability at all (afterall its an allocation within an allocation so yeah) for fibers: Main reasoning behind this is because virtualBuffer<> is stupidly fucking expensive and it also clutters my fstat view ALSO mmap is a syscall, syscalls are bad for performance or whatever ALSO std::vector<> is better suited for handling this kind of "fixed size thing where its like big but not THAT big" (512 KiB isn't going to kill your memory usage for each fiber...) for core.cpp stuff - inlines stuff into std::optional<> as opposed to std::unique_ptr<> (because yknow, we are making the Impl from an unique_ptr, allocating within an allocation is unnecessary) - reorganizes the structures a bit so padding doesnt screw us up (it's not perfect but eh saves a measly 44 bytes) - removes unused/dead code - uses std::vector<> instead of std::deque<> no perf impact expected, maybe some initialisation boost but very minimal impact nonethless lto gets rid of most calls anyways - the heavy issue is with shared_ptr and the cache coherency from the atomics... but i clumped them together because well, they kinda do not suffer from cache coherency - hopefully not a mistake this balloons the size of Impl to about 1.67 MB - which is fine because we throw it in the stack anyways REST OF INTERFACES: most of them ballooned in size as well, but overhead is ok since its an allocation within an alloc, no stack is used (when it comes to storing these i mean) Signed-off-by: lizzie lizzie@eden-emu.dev Reviewed-on: https://git.eden-emu.dev/eden-emu/eden/pulls/3306 Reviewed-by: CamilleLaVey <camillelavey99@gmail.com> Reviewed-by: MaranBr <maranbr@eden-emu.dev> Co-authored-by: lizzie <lizzie@eden-emu.dev> Co-committed-by: lizzie <lizzie@eden-emu.dev>
358 lines
10 KiB
C++
358 lines
10 KiB
C++
// SPDX-FileCopyrightText: Copyright 2025 Eden Emulator Project
|
|
// SPDX-License-Identifier: GPL-3.0-or-later
|
|
|
|
// SPDX-FileCopyrightText: Copyright 2020 yuzu Emulator Project
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
|
|
#include <algorithm>
|
|
#include <mutex>
|
|
#include <string>
|
|
#include <tuple>
|
|
|
|
#ifdef _WIN32
|
|
#include "common/windows/timer_resolution.h"
|
|
#endif
|
|
|
|
#ifdef ARCHITECTURE_x86_64
|
|
#include "common/x64/cpu_wait.h"
|
|
#endif
|
|
|
|
#include "common/settings.h"
|
|
#include "core/core_timing.h"
|
|
#include "core/hardware_properties.h"
|
|
|
|
namespace Core::Timing {
|
|
|
|
constexpr s64 MAX_SLICE_LENGTH = 10000;
|
|
|
|
std::shared_ptr<EventType> CreateEvent(std::string name, TimedCallback&& callback) {
|
|
return std::make_shared<EventType>(std::move(callback), std::move(name));
|
|
}
|
|
|
|
struct CoreTiming::Event {
|
|
s64 time;
|
|
u64 fifo_order;
|
|
std::weak_ptr<EventType> type;
|
|
s64 reschedule_time;
|
|
heap_t::handle_type handle{};
|
|
|
|
// Sort by time, unless the times are the same, in which case sort by
|
|
// the order added to the queue
|
|
friend bool operator>(const Event& left, const Event& right) {
|
|
return std::tie(left.time, left.fifo_order) > std::tie(right.time, right.fifo_order);
|
|
}
|
|
|
|
friend bool operator<(const Event& left, const Event& right) {
|
|
return std::tie(left.time, left.fifo_order) < std::tie(right.time, right.fifo_order);
|
|
}
|
|
};
|
|
|
|
CoreTiming::CoreTiming() : clock{Common::CreateOptimalClock()} {}
|
|
|
|
CoreTiming::~CoreTiming() {
|
|
Reset();
|
|
}
|
|
|
|
void CoreTiming::Initialize(std::function<void()>&& on_thread_init_) {
|
|
Reset();
|
|
on_thread_init = std::move(on_thread_init_);
|
|
event_fifo_id = 0;
|
|
shutting_down = false;
|
|
cpu_ticks = 0;
|
|
if (is_multicore) {
|
|
timer_thread.emplace([](CoreTiming& instance) {
|
|
Common::SetCurrentThreadName("HostTiming");
|
|
Common::SetCurrentThreadPriority(Common::ThreadPriority::High);
|
|
instance.on_thread_init();
|
|
instance.ThreadLoop();
|
|
}, std::ref(*this));
|
|
}
|
|
}
|
|
|
|
void CoreTiming::ClearPendingEvents() {
|
|
std::scoped_lock lock{advance_lock, basic_lock};
|
|
event_queue.clear();
|
|
event.Set();
|
|
}
|
|
|
|
void CoreTiming::Pause(bool is_paused) {
|
|
paused = is_paused;
|
|
pause_event.Set();
|
|
|
|
if (!is_paused) {
|
|
pause_end_time = GetGlobalTimeNs().count();
|
|
}
|
|
}
|
|
|
|
void CoreTiming::SyncPause(bool is_paused) {
|
|
if (is_paused == paused && paused_set == paused) {
|
|
return;
|
|
}
|
|
|
|
Pause(is_paused);
|
|
if (timer_thread) {
|
|
if (!is_paused) {
|
|
pause_event.Set();
|
|
}
|
|
event.Set();
|
|
while (paused_set != is_paused)
|
|
;
|
|
}
|
|
|
|
if (!is_paused) {
|
|
pause_end_time = GetGlobalTimeNs().count();
|
|
}
|
|
}
|
|
|
|
bool CoreTiming::IsRunning() const {
|
|
return !paused_set;
|
|
}
|
|
|
|
bool CoreTiming::HasPendingEvents() const {
|
|
std::scoped_lock lock{basic_lock};
|
|
return !(wait_set && event_queue.empty());
|
|
}
|
|
|
|
void CoreTiming::ScheduleEvent(std::chrono::nanoseconds ns_into_future,
|
|
const std::shared_ptr<EventType>& event_type, bool absolute_time) {
|
|
{
|
|
std::scoped_lock scope{basic_lock};
|
|
const auto next_time{absolute_time ? ns_into_future : GetGlobalTimeNs() + ns_into_future};
|
|
|
|
auto h{event_queue.emplace(Event{next_time.count(), event_fifo_id++, event_type, 0})};
|
|
(*h).handle = h;
|
|
}
|
|
|
|
event.Set();
|
|
}
|
|
|
|
void CoreTiming::ScheduleLoopingEvent(std::chrono::nanoseconds start_time,
|
|
std::chrono::nanoseconds resched_time,
|
|
const std::shared_ptr<EventType>& event_type,
|
|
bool absolute_time) {
|
|
{
|
|
std::scoped_lock scope{basic_lock};
|
|
const auto next_time{absolute_time ? start_time : GetGlobalTimeNs() + start_time};
|
|
|
|
auto h{event_queue.emplace(
|
|
Event{next_time.count(), event_fifo_id++, event_type, resched_time.count()})};
|
|
(*h).handle = h;
|
|
}
|
|
|
|
event.Set();
|
|
}
|
|
|
|
void CoreTiming::UnscheduleEvent(const std::shared_ptr<EventType>& event_type,
|
|
UnscheduleEventType type) {
|
|
{
|
|
std::scoped_lock lk{basic_lock};
|
|
|
|
std::vector<heap_t::handle_type> to_remove;
|
|
for (auto itr = event_queue.begin(); itr != event_queue.end(); itr++) {
|
|
const Event& e = *itr;
|
|
if (e.type.lock().get() == event_type.get()) {
|
|
to_remove.push_back(itr->handle);
|
|
}
|
|
}
|
|
|
|
for (auto& h : to_remove) {
|
|
event_queue.erase(h);
|
|
}
|
|
|
|
event_type->sequence_number++;
|
|
}
|
|
|
|
// Force any in-progress events to finish
|
|
if (type == UnscheduleEventType::Wait) {
|
|
std::scoped_lock lk{advance_lock};
|
|
}
|
|
}
|
|
|
|
static u64 GetNextTickCount(u64 next_ticks) {
|
|
if (Settings::values.use_custom_cpu_ticks.GetValue()) {
|
|
return Settings::values.cpu_ticks.GetValue();
|
|
}
|
|
return next_ticks;
|
|
}
|
|
|
|
void CoreTiming::AddTicks(u64 ticks_to_add) {
|
|
const u64 ticks = GetNextTickCount(ticks_to_add);
|
|
cpu_ticks += ticks;
|
|
downcount -= static_cast<s64>(ticks);
|
|
}
|
|
|
|
void CoreTiming::Idle() {
|
|
AddTicks(1000U);
|
|
}
|
|
|
|
void CoreTiming::ResetTicks() {
|
|
downcount = MAX_SLICE_LENGTH;
|
|
}
|
|
|
|
u64 CoreTiming::GetClockTicks() const {
|
|
u64 fres;
|
|
if (is_multicore) [[likely]] {
|
|
fres = clock->GetCNTPCT();
|
|
} else {
|
|
fres = Common::WallClock::CPUTickToCNTPCT(cpu_ticks);
|
|
}
|
|
|
|
const auto overclock = Settings::values.fast_cpu_time.GetValue();
|
|
|
|
if (overclock != Settings::CpuClock::Off) {
|
|
fres = (u64) ((double) fres * (1.7 + 0.3 * u32(overclock)));
|
|
}
|
|
|
|
if (Settings::values.sync_core_speed.GetValue()) {
|
|
const auto ticks = double(fres);
|
|
const auto speed_limit = double(Settings::values.speed_limit.GetValue())*0.01;
|
|
return u64(ticks/speed_limit);
|
|
} else {
|
|
return fres;
|
|
}
|
|
}
|
|
|
|
u64 CoreTiming::GetGPUTicks() const {
|
|
if (is_multicore) [[likely]] {
|
|
return clock->GetGPUTick();
|
|
}
|
|
return Common::WallClock::CPUTickToGPUTick(cpu_ticks);
|
|
}
|
|
|
|
std::optional<s64> CoreTiming::Advance() {
|
|
std::scoped_lock lock{advance_lock, basic_lock};
|
|
global_timer = GetGlobalTimeNs().count();
|
|
|
|
while (!event_queue.empty() && event_queue.top().time <= global_timer) {
|
|
const Event& evt = event_queue.top();
|
|
|
|
if (const auto event_type{evt.type.lock()}) {
|
|
const auto evt_time = evt.time;
|
|
const auto evt_sequence_num = event_type->sequence_number;
|
|
|
|
if (evt.reschedule_time == 0) {
|
|
event_queue.pop();
|
|
|
|
basic_lock.unlock();
|
|
|
|
event_type->callback(
|
|
evt_time, std::chrono::nanoseconds{GetGlobalTimeNs().count() - evt_time});
|
|
|
|
basic_lock.lock();
|
|
} else {
|
|
basic_lock.unlock();
|
|
|
|
const auto new_schedule_time{event_type->callback(
|
|
evt_time, std::chrono::nanoseconds{GetGlobalTimeNs().count() - evt_time})};
|
|
|
|
basic_lock.lock();
|
|
|
|
if (evt_sequence_num != event_type->sequence_number) {
|
|
// Heap handle is invalidated after external modification.
|
|
continue;
|
|
}
|
|
|
|
const auto next_schedule_time{new_schedule_time.has_value()
|
|
? new_schedule_time.value().count()
|
|
: evt.reschedule_time};
|
|
|
|
// If this event was scheduled into a pause, its time now is going to be way
|
|
// behind. Re-set this event to continue from the end of the pause.
|
|
auto next_time{evt.time + next_schedule_time};
|
|
if (evt.time < pause_end_time) {
|
|
next_time = pause_end_time + next_schedule_time;
|
|
}
|
|
|
|
event_queue.update(evt.handle, Event{next_time, event_fifo_id++, evt.type,
|
|
next_schedule_time, evt.handle});
|
|
}
|
|
}
|
|
|
|
global_timer = GetGlobalTimeNs().count();
|
|
}
|
|
|
|
if (!event_queue.empty()) {
|
|
return event_queue.top().time;
|
|
} else {
|
|
return std::nullopt;
|
|
}
|
|
}
|
|
|
|
void CoreTiming::ThreadLoop() {
|
|
has_started = true;
|
|
while (!shutting_down) {
|
|
while (!paused) {
|
|
paused_set = false;
|
|
const auto next_time = Advance();
|
|
if (next_time) {
|
|
// There are more events left in the queue, wait until the next event.
|
|
auto wait_time = *next_time - GetGlobalTimeNs().count();
|
|
if (wait_time > 0) {
|
|
#ifdef _WIN32
|
|
while (!paused && !event.IsSet() && wait_time > 0) {
|
|
wait_time = *next_time - GetGlobalTimeNs().count();
|
|
if (wait_time >= timer_resolution_ns) {
|
|
Common::Windows::SleepForOneTick();
|
|
} else {
|
|
#ifdef ARCHITECTURE_x86_64
|
|
Common::X64::MicroSleep();
|
|
#else
|
|
std::this_thread::yield();
|
|
#endif
|
|
}
|
|
}
|
|
|
|
if (event.IsSet()) {
|
|
event.Reset();
|
|
}
|
|
#else
|
|
event.WaitFor(std::chrono::nanoseconds(wait_time));
|
|
#endif
|
|
}
|
|
} else {
|
|
// Queue is empty, wait until another event is scheduled and signals us to
|
|
// continue.
|
|
wait_set = true;
|
|
event.Wait();
|
|
}
|
|
wait_set = false;
|
|
}
|
|
|
|
paused_set = true;
|
|
pause_event.Wait();
|
|
}
|
|
}
|
|
|
|
void CoreTiming::Reset() {
|
|
paused = true;
|
|
shutting_down = true;
|
|
pause_event.Set();
|
|
event.Set();
|
|
if (timer_thread) {
|
|
timer_thread->join();
|
|
}
|
|
timer_thread.reset();
|
|
has_started = false;
|
|
}
|
|
|
|
std::chrono::nanoseconds CoreTiming::GetGlobalTimeNs() const {
|
|
if (is_multicore) [[likely]] {
|
|
return clock->GetTimeNS();
|
|
}
|
|
return std::chrono::nanoseconds{Common::WallClock::CPUTickToNS(cpu_ticks)};
|
|
}
|
|
|
|
std::chrono::microseconds CoreTiming::GetGlobalTimeUs() const {
|
|
if (is_multicore) [[likely]] {
|
|
return clock->GetTimeUS();
|
|
}
|
|
return std::chrono::microseconds{Common::WallClock::CPUTickToUS(cpu_ticks)};
|
|
}
|
|
|
|
#ifdef _WIN32
|
|
void CoreTiming::SetTimerResolutionNs(std::chrono::nanoseconds ns) {
|
|
timer_resolution_ns = ns.count();
|
|
}
|
|
#endif
|
|
|
|
} // namespace Core::Timing
|