kernel: fix single-core preemption points
This commit is contained in:
parent
21945ae127
commit
da07e13e07
|
@ -144,11 +144,8 @@ void CpuManager::SingleCoreRunIdleThread() {
|
||||||
}
|
}
|
||||||
|
|
||||||
void CpuManager::PreemptSingleCore(bool from_running_environment) {
|
void CpuManager::PreemptSingleCore(bool from_running_environment) {
|
||||||
{
|
|
||||||
auto& kernel = system.Kernel();
|
auto& kernel = system.Kernel();
|
||||||
auto& scheduler = kernel.Scheduler(current_core);
|
|
||||||
|
|
||||||
Kernel::KThread* current_thread = scheduler.GetSchedulerCurrentThread();
|
|
||||||
if (idle_count >= 4 || from_running_environment) {
|
if (idle_count >= 4 || from_running_environment) {
|
||||||
if (!from_running_environment) {
|
if (!from_running_environment) {
|
||||||
system.CoreTiming().Idle();
|
system.CoreTiming().Idle();
|
||||||
|
@ -160,25 +157,14 @@ void CpuManager::PreemptSingleCore(bool from_running_environment) {
|
||||||
}
|
}
|
||||||
current_core.store((current_core + 1) % Core::Hardware::NUM_CPU_CORES);
|
current_core.store((current_core + 1) % Core::Hardware::NUM_CPU_CORES);
|
||||||
system.CoreTiming().ResetTicks();
|
system.CoreTiming().ResetTicks();
|
||||||
scheduler.Unload(scheduler.GetSchedulerCurrentThread());
|
kernel.Scheduler(current_core).PreemptSingleCore();
|
||||||
|
|
||||||
auto& next_scheduler = kernel.Scheduler(current_core);
|
|
||||||
|
|
||||||
// Disable dispatch. We're about to preempt this thread.
|
|
||||||
Kernel::KScopedDisableDispatch dd{kernel};
|
|
||||||
Common::Fiber::YieldTo(current_thread->GetHostContext(), *next_scheduler.GetSwitchFiber());
|
|
||||||
}
|
|
||||||
|
|
||||||
// We've now been scheduled again, and we may have exchanged schedulers.
|
// We've now been scheduled again, and we may have exchanged schedulers.
|
||||||
// Reload the scheduler in case it's different.
|
// Reload the scheduler in case it's different.
|
||||||
{
|
if (!kernel.Scheduler(current_core).IsIdle()) {
|
||||||
auto& scheduler = system.Kernel().Scheduler(current_core);
|
|
||||||
scheduler.Reload(scheduler.GetSchedulerCurrentThread());
|
|
||||||
if (!scheduler.IsIdle()) {
|
|
||||||
idle_count = 0;
|
idle_count = 0;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
void CpuManager::GuestActivate() {
|
void CpuManager::GuestActivate() {
|
||||||
// Similar to the HorizonKernelMain callback in HOS
|
// Similar to the HorizonKernelMain callback in HOS
|
||||||
|
|
|
@ -5,7 +5,6 @@
|
||||||
|
|
||||||
#include <array>
|
#include <array>
|
||||||
#include <atomic>
|
#include <atomic>
|
||||||
#include <csetjmp>
|
|
||||||
#include <functional>
|
#include <functional>
|
||||||
#include <memory>
|
#include <memory>
|
||||||
#include <thread>
|
#include <thread>
|
||||||
|
|
|
@ -103,7 +103,20 @@ void KScheduler::ScheduleOnInterrupt() {
|
||||||
GetCurrentThread(kernel).EnableDispatch();
|
GetCurrentThread(kernel).EnableDispatch();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void KScheduler::PreemptSingleCore() {
|
||||||
|
GetCurrentThread(kernel).DisableDispatch();
|
||||||
|
|
||||||
|
auto* thread = GetCurrentThreadPointer(kernel);
|
||||||
|
auto& previous_scheduler = kernel.Scheduler(thread->GetCurrentCore());
|
||||||
|
previous_scheduler.Unload(thread);
|
||||||
|
|
||||||
|
Common::Fiber::YieldTo(thread->GetHostContext(), *m_switch_fiber);
|
||||||
|
|
||||||
|
GetCurrentThread(kernel).EnableDispatch();
|
||||||
|
}
|
||||||
|
|
||||||
void KScheduler::RescheduleCurrentCore() {
|
void KScheduler::RescheduleCurrentCore() {
|
||||||
|
ASSERT(!kernel.IsPhantomModeForSingleCore());
|
||||||
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() == 1);
|
ASSERT(GetCurrentThread(kernel).GetDisableDispatchCount() == 1);
|
||||||
|
|
||||||
GetCurrentThread(kernel).EnableDispatch();
|
GetCurrentThread(kernel).EnableDispatch();
|
||||||
|
|
|
@ -49,6 +49,7 @@ public:
|
||||||
|
|
||||||
void SetInterruptTaskRunnable();
|
void SetInterruptTaskRunnable();
|
||||||
void RequestScheduleOnInterrupt();
|
void RequestScheduleOnInterrupt();
|
||||||
|
void PreemptSingleCore();
|
||||||
|
|
||||||
u64 GetIdleCount() {
|
u64 GetIdleCount() {
|
||||||
return m_state.idle_count;
|
return m_state.idle_count;
|
||||||
|
@ -62,10 +63,6 @@ public:
|
||||||
return m_current_thread.load() == m_idle_thread;
|
return m_current_thread.load() == m_idle_thread;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::shared_ptr<Common::Fiber> GetSwitchFiber() {
|
|
||||||
return m_switch_fiber;
|
|
||||||
}
|
|
||||||
|
|
||||||
KThread* GetPreviousThread() const {
|
KThread* GetPreviousThread() const {
|
||||||
return m_state.prev_thread;
|
return m_state.prev_thread;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1204,12 +1204,6 @@ KScopedDisableDispatch::~KScopedDisableDispatch() {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Skip the reschedule if single-core.
|
|
||||||
if (!Settings::values.use_multi_core.GetValue()) {
|
|
||||||
GetCurrentThread(kernel).EnableDispatch();
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (GetCurrentThread(kernel).GetDisableDispatchCount() <= 1) {
|
if (GetCurrentThread(kernel).GetDisableDispatchCount() <= 1) {
|
||||||
auto scheduler = kernel.CurrentScheduler();
|
auto scheduler = kernel.CurrentScheduler();
|
||||||
|
|
||||||
|
|
|
@ -110,7 +110,6 @@ void SetCurrentThread(KernelCore& kernel, KThread* thread);
|
||||||
[[nodiscard]] KThread* GetCurrentThreadPointer(KernelCore& kernel);
|
[[nodiscard]] KThread* GetCurrentThreadPointer(KernelCore& kernel);
|
||||||
[[nodiscard]] KThread& GetCurrentThread(KernelCore& kernel);
|
[[nodiscard]] KThread& GetCurrentThread(KernelCore& kernel);
|
||||||
[[nodiscard]] s32 GetCurrentCoreId(KernelCore& kernel);
|
[[nodiscard]] s32 GetCurrentCoreId(KernelCore& kernel);
|
||||||
size_t CaptureBacktrace(void** buffer, size_t max);
|
|
||||||
|
|
||||||
class KThread final : public KAutoObjectWithSlabHeapAndContainer<KThread, KWorkerTask>,
|
class KThread final : public KAutoObjectWithSlabHeapAndContainer<KThread, KWorkerTask>,
|
||||||
public boost::intrusive::list_base_hook<> {
|
public boost::intrusive::list_base_hook<> {
|
||||||
|
|
Loading…
Reference in New Issue