// SPDX-FileCopyrightText: Copyright 2023 yuzu Emulator Project // SPDX-License-Identifier: GPL-2.0-or-later #include #include #include "common/signal_chain.h" #include "core/arm/nce/arm_nce.h" #include "core/arm/nce/guest_context.h" #include "core/arm/nce/patcher.h" #include "core/core.h" #include "core/memory.h" #include "core/hle/kernel/k_process.h" #include #include #include namespace Core { namespace { struct sigaction g_orig_action; // Verify assembly offsets. using NativeExecutionParameters = Kernel::KThread::NativeExecutionParameters; static_assert(offsetof(NativeExecutionParameters, native_context) == TpidrEl0NativeContext); static_assert(offsetof(NativeExecutionParameters, lock) == TpidrEl0Lock); static_assert(offsetof(NativeExecutionParameters, magic) == TpidrEl0TlsMagic); fpsimd_context* GetFloatingPointState(mcontext_t& host_ctx) { _aarch64_ctx* header = reinterpret_cast<_aarch64_ctx*>(&host_ctx.__reserved); while (header->magic != FPSIMD_MAGIC) { header = reinterpret_cast<_aarch64_ctx*>(reinterpret_cast(header) + header->size); } return reinterpret_cast(header); } } // namespace void* ArmNce::RestoreGuestContext(void* raw_context) { // Retrieve the host context. auto& host_ctx = static_cast(raw_context)->uc_mcontext; // Thread-local parameters will be located in x9. auto* tpidr = reinterpret_cast(host_ctx.regs[9]); auto* guest_ctx = static_cast(tpidr->native_context); // Retrieve the host floating point state. auto* fpctx = GetFloatingPointState(host_ctx); // Save host callee-saved registers. std::memcpy(guest_ctx->host_ctx.host_saved_vregs.data(), &fpctx->vregs[8], sizeof(guest_ctx->host_ctx.host_saved_vregs)); std::memcpy(guest_ctx->host_ctx.host_saved_regs.data(), &host_ctx.regs[19], sizeof(guest_ctx->host_ctx.host_saved_regs)); // Save stack pointer. guest_ctx->host_ctx.host_sp = host_ctx.sp; // Restore all guest state except tpidr_el0. host_ctx.sp = guest_ctx->sp; host_ctx.pc = guest_ctx->pc; host_ctx.pstate = guest_ctx->pstate; fpctx->fpcr = guest_ctx->fpcr; fpctx->fpsr = guest_ctx->fpsr; std::memcpy(host_ctx.regs, guest_ctx->cpu_registers.data(), sizeof(host_ctx.regs)); std::memcpy(fpctx->vregs, guest_ctx->vector_registers.data(), sizeof(fpctx->vregs)); // Return the new thread-local storage pointer. return tpidr; } void ArmNce::SaveGuestContext(GuestContext* guest_ctx, void* raw_context) { // Retrieve the host context. auto& host_ctx = static_cast(raw_context)->uc_mcontext; // Retrieve the host floating point state. auto* fpctx = GetFloatingPointState(host_ctx); // Save all guest registers except tpidr_el0. std::memcpy(guest_ctx->cpu_registers.data(), host_ctx.regs, sizeof(host_ctx.regs)); std::memcpy(guest_ctx->vector_registers.data(), fpctx->vregs, sizeof(fpctx->vregs)); guest_ctx->fpsr = fpctx->fpsr; guest_ctx->fpcr = fpctx->fpcr; guest_ctx->pstate = static_cast(host_ctx.pstate); guest_ctx->pc = host_ctx.pc; guest_ctx->sp = host_ctx.sp; // Restore stack pointer. host_ctx.sp = guest_ctx->host_ctx.host_sp; // Restore host callee-saved registers. std::memcpy(&host_ctx.regs[19], guest_ctx->host_ctx.host_saved_regs.data(), sizeof(guest_ctx->host_ctx.host_saved_regs)); std::memcpy(&fpctx->vregs[8], guest_ctx->host_ctx.host_saved_vregs.data(), sizeof(guest_ctx->host_ctx.host_saved_vregs)); // Return from the call on exit by setting pc to x30. host_ctx.pc = guest_ctx->host_ctx.host_saved_regs[11]; // Clear esr_el1 and return it. host_ctx.regs[0] = guest_ctx->esr_el1.exchange(0); } bool ArmNce::HandleGuestFault(GuestContext* guest_ctx, void* raw_info, void* raw_context) { auto& host_ctx = static_cast(raw_context)->uc_mcontext; auto* info = static_cast(raw_info); // Try to handle an invalid access. // TODO: handle accesses which split a page? const Common::ProcessAddress addr = (reinterpret_cast(info->si_addr) & ~Memory::YUZU_PAGEMASK); if (guest_ctx->system->ApplicationMemory().InvalidateNCE(addr, Memory::YUZU_PAGESIZE)) { // We handled the access successfully and are returning to guest code. return true; } // We can't handle the access, so determine why we crashed. const bool is_prefetch_abort = host_ctx.pc == reinterpret_cast(info->si_addr); // For data aborts, skip the instruction and return to guest code. // This will allow games to continue in many scenarios where they would otherwise crash. if (!is_prefetch_abort) { host_ctx.pc += 4; return true; } // This is a prefetch abort. guest_ctx->esr_el1.fetch_or(static_cast(HaltReason::PrefetchAbort)); // Forcibly mark the context as locked. We are still running. // We may race with SignalInterrupt here: // - If we lose the race, then SignalInterrupt will send us a signal we are masking, // and it will do nothing when it is unmasked, as we have already left guest code. // - If we win the race, then SignalInterrupt will wait for us to unlock first. auto& thread_params = guest_ctx->parent->m_running_thread->GetNativeExecutionParameters(); thread_params.lock.store(SpinLockLocked); // Return to host. SaveGuestContext(guest_ctx, raw_context); return false; } void ArmNce::HandleHostFault(int sig, void* raw_info, void* raw_context) { return g_orig_action.sa_sigaction(sig, static_cast(raw_info), raw_context); } void ArmNce::LockThread(Kernel::KThread* thread) { auto* thread_params = &thread->GetNativeExecutionParameters(); LockThreadParameters(thread_params); } void ArmNce::UnlockThread(Kernel::KThread* thread) { auto* thread_params = &thread->GetNativeExecutionParameters(); UnlockThreadParameters(thread_params); } HaltReason ArmNce::RunThread(Kernel::KThread* thread) { // Check if we're already interrupted. // If we are, we can just return immediately. HaltReason hr = static_cast(m_guest_ctx.esr_el1.exchange(0)); if (True(hr)) { return hr; } // Get the thread context. auto* thread_params = &thread->GetNativeExecutionParameters(); auto* process = thread->GetOwnerProcess(); // Assign current members. m_running_thread = thread; m_guest_ctx.parent = this; thread_params->native_context = &m_guest_ctx; thread_params->tpidr_el0 = m_guest_ctx.tpidr_el0; thread_params->tpidrro_el0 = m_guest_ctx.tpidrro_el0; thread_params->is_running = true; // TODO: finding and creating the post handler needs to be locked // to deal with dynamic loading of NROs. const auto& post_handlers = process->GetPostHandlers(); if (auto it = post_handlers.find(m_guest_ctx.pc); it != post_handlers.end()) { hr = ReturnToRunCodeByTrampoline(thread_params, &m_guest_ctx, it->second); } else { hr = ReturnToRunCodeByExceptionLevelChange(m_thread_id, thread_params); } // Unload members. // The thread does not change, so we can persist the old reference. m_running_thread = nullptr; m_guest_ctx.tpidr_el0 = thread_params->tpidr_el0; thread_params->native_context = nullptr; thread_params->is_running = false; // Return the halt reason. return hr; } HaltReason ArmNce::StepThread(Kernel::KThread* thread) { return HaltReason::StepThread; } u32 ArmNce::GetSvcNumber() const { return m_guest_ctx.svc; } void ArmNce::GetSvcArguments(std::span args) const { for (size_t i = 0; i < 8; i++) { args[i] = m_guest_ctx.cpu_registers[i]; } } void ArmNce::SetSvcArguments(std::span args) { for (size_t i = 0; i < 8; i++) { m_guest_ctx.cpu_registers[i] = args[i]; } } ArmNce::ArmNce(System& system, bool uses_wall_clock, std::size_t core_index) : ArmInterface{uses_wall_clock}, m_system{system}, m_core_index{core_index} { m_guest_ctx.system = &m_system; } ArmNce::~ArmNce() = default; void ArmNce::Initialize() { m_thread_id = gettid(); // Setup our signals static std::once_flag signals; std::call_once(signals, [] { using HandlerType = decltype(sigaction::sa_sigaction); sigset_t signal_mask; sigemptyset(&signal_mask); sigaddset(&signal_mask, ReturnToRunCodeByExceptionLevelChangeSignal); sigaddset(&signal_mask, BreakFromRunCodeSignal); sigaddset(&signal_mask, GuestFaultSignal); struct sigaction return_to_run_code_action {}; return_to_run_code_action.sa_flags = SA_SIGINFO | SA_ONSTACK; return_to_run_code_action.sa_sigaction = reinterpret_cast( &ArmNce::ReturnToRunCodeByExceptionLevelChangeSignalHandler); return_to_run_code_action.sa_mask = signal_mask; Common::SigAction(ReturnToRunCodeByExceptionLevelChangeSignal, &return_to_run_code_action, nullptr); struct sigaction break_from_run_code_action {}; break_from_run_code_action.sa_flags = SA_SIGINFO | SA_ONSTACK; break_from_run_code_action.sa_sigaction = reinterpret_cast(&ArmNce::BreakFromRunCodeSignalHandler); break_from_run_code_action.sa_mask = signal_mask; Common::SigAction(BreakFromRunCodeSignal, &break_from_run_code_action, nullptr); struct sigaction fault_action {}; fault_action.sa_flags = SA_SIGINFO | SA_ONSTACK | SA_RESTART; fault_action.sa_sigaction = reinterpret_cast(&ArmNce::GuestFaultSignalHandler); fault_action.sa_mask = signal_mask; Common::SigAction(GuestFaultSignal, &fault_action, &g_orig_action); // Simplify call for g_orig_action. // These fields occupy the same space in memory, so this should be a no-op in practice. if (!(g_orig_action.sa_flags & SA_SIGINFO)) { g_orig_action.sa_sigaction = reinterpret_cast(g_orig_action.sa_handler); } }); } void ArmNce::SetTpidrroEl0(u64 value) { m_guest_ctx.tpidrro_el0 = value; } void ArmNce::GetContext(Kernel::Svc::ThreadContext& ctx) const { for (size_t i = 0; i < 29; i++) { ctx.r[i] = m_guest_ctx.cpu_registers[i]; } ctx.fp = m_guest_ctx.cpu_registers[29]; ctx.lr = m_guest_ctx.cpu_registers[30]; ctx.sp = m_guest_ctx.sp; ctx.pc = m_guest_ctx.pc; ctx.pstate = m_guest_ctx.pstate; ctx.v = m_guest_ctx.vector_registers; ctx.fpcr = m_guest_ctx.fpcr; ctx.fpsr = m_guest_ctx.fpsr; ctx.tpidr = m_guest_ctx.tpidr_el0; } void ArmNce::SetContext(const Kernel::Svc::ThreadContext& ctx) { for (size_t i = 0; i < 29; i++) { m_guest_ctx.cpu_registers[i] = ctx.r[i]; } m_guest_ctx.cpu_registers[29] = ctx.fp; m_guest_ctx.cpu_registers[30] = ctx.lr; m_guest_ctx.sp = ctx.sp; m_guest_ctx.pc = ctx.pc; m_guest_ctx.pstate = ctx.pstate; m_guest_ctx.vector_registers = ctx.v; m_guest_ctx.fpcr = ctx.fpcr; m_guest_ctx.fpsr = ctx.fpsr; m_guest_ctx.tpidr_el0 = ctx.tpidr; } void ArmNce::SignalInterrupt(Kernel::KThread* thread) { // Add break loop condition. m_guest_ctx.esr_el1.fetch_or(static_cast(HaltReason::BreakLoop)); // Lock the thread context. auto* params = &thread->GetNativeExecutionParameters(); LockThreadParameters(params); if (params->is_running) { // We should signal to the running thread. // The running thread will unlock the thread context. syscall(SYS_tkill, m_thread_id, BreakFromRunCodeSignal); } else { // If the thread is no longer running, we have nothing to do. UnlockThreadParameters(params); } } void ArmNce::ClearInstructionCache() { // TODO: This is not possible to implement correctly on Linux because // we do not have any access to ic iallu. // Require accesses to complete. std::atomic_thread_fence(std::memory_order_seq_cst); } void ArmNce::InvalidateCacheRange(u64 addr, std::size_t size) { this->ClearInstructionCache(); } } // namespace Core