2020-01-31 06:46:18 +00:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2018-2020 Atmosphère-NX
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
|
|
* under the terms and conditions of the GNU General Public License,
|
|
|
|
* version 2, as published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
|
|
* more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
*/
|
|
|
|
#include <mesosphere.hpp>
|
|
|
|
|
2020-02-15 02:22:55 +00:00
|
|
|
namespace ams::kern::arch::arm64 {
|
2020-01-31 06:46:18 +00:00
|
|
|
|
|
|
|
/* These are implemented elsewhere (asm). */
|
|
|
|
void UserModeThreadStarter();
|
|
|
|
void SupervisorModeThreadStarter();
|
|
|
|
|
|
|
|
void OnThreadStart() {
|
2020-02-20 05:21:44 +00:00
|
|
|
MESOSPHERE_ASSERT(!KInterruptManager::AreInterruptsEnabled());
|
|
|
|
/* Send KDebug event for this thread's creation. */
|
|
|
|
{
|
|
|
|
KScopedInterruptEnable ei;
|
2020-12-01 23:19:29 +00:00
|
|
|
KDebug::OnDebugEvent(ams::svc::DebugEvent_CreateThread, GetCurrentThread().GetId(), GetInteger(GetCurrentThread().GetThreadLocalRegionAddress()));
|
2020-02-20 05:21:44 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Handle any pending dpc. */
|
|
|
|
while (GetCurrentThread().HasDpc()) {
|
|
|
|
KDpcManager::HandleDpc();
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Clear our status as in an exception handler */
|
|
|
|
GetCurrentThread().ClearInExceptionHandler();
|
2020-01-31 06:46:18 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
|
2020-07-31 12:52:59 +00:00
|
|
|
constexpr inline u32 El0PsrMask = 0xFF0FFE20;
|
|
|
|
|
2020-02-08 10:49:32 +00:00
|
|
|
ALWAYS_INLINE bool IsFpuEnabled() {
|
|
|
|
return cpu::ArchitecturalFeatureAccessControlRegisterAccessor().IsFpEnabled();
|
|
|
|
}
|
|
|
|
|
|
|
|
ALWAYS_INLINE void EnableFpu() {
|
|
|
|
cpu::ArchitecturalFeatureAccessControlRegisterAccessor().SetFpEnabled(true).Store();
|
|
|
|
cpu::InstructionMemoryBarrier();
|
|
|
|
}
|
|
|
|
|
2020-01-31 06:46:18 +00:00
|
|
|
uintptr_t SetupStackForUserModeThreadStarter(KVirtualAddress pc, KVirtualAddress k_sp, KVirtualAddress u_sp, uintptr_t arg, bool is_64_bit) {
|
|
|
|
/* NOTE: Stack layout on entry looks like following: */
|
|
|
|
/* SP */
|
|
|
|
/* | */
|
|
|
|
/* v */
|
|
|
|
/* | KExceptionContext (size 0x120) | KThread::StackParameters (size 0x30) | */
|
|
|
|
KExceptionContext *ctx = GetPointer<KExceptionContext>(k_sp) - 1;
|
|
|
|
|
|
|
|
/* Clear context. */
|
|
|
|
std::memset(ctx, 0, sizeof(*ctx));
|
|
|
|
|
|
|
|
/* Set PC and argument. */
|
2020-07-30 00:48:51 +00:00
|
|
|
ctx->pc = GetInteger(pc) & ~(UINT64_C(1));
|
2020-01-31 06:46:18 +00:00
|
|
|
ctx->x[0] = arg;
|
|
|
|
|
|
|
|
/* Set PSR. */
|
|
|
|
if (is_64_bit) {
|
|
|
|
ctx->psr = 0;
|
|
|
|
} else {
|
2020-07-29 22:29:01 +00:00
|
|
|
constexpr u64 PsrArmValue = 0x00;
|
|
|
|
constexpr u64 PsrThumbValue = 0x20;
|
2020-01-31 06:46:18 +00:00
|
|
|
ctx->psr = ((pc & 1) == 0 ? PsrArmValue : PsrThumbValue) | (0x10);
|
2020-07-29 22:29:01 +00:00
|
|
|
MESOSPHERE_LOG("Creating User 32-Thread, %016lx\n", GetInteger(pc));
|
2020-01-31 06:46:18 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Set stack pointer. */
|
|
|
|
if (is_64_bit) {
|
|
|
|
ctx->sp = GetInteger(u_sp);
|
|
|
|
} else {
|
|
|
|
ctx->x[13] = GetInteger(u_sp);
|
|
|
|
}
|
|
|
|
|
|
|
|
return reinterpret_cast<uintptr_t>(ctx);
|
|
|
|
}
|
|
|
|
|
|
|
|
uintptr_t SetupStackForSupervisorModeThreadStarter(KVirtualAddress pc, KVirtualAddress sp, uintptr_t arg) {
|
|
|
|
/* NOTE: Stack layout on entry looks like following: */
|
|
|
|
/* SP */
|
|
|
|
/* | */
|
|
|
|
/* v */
|
|
|
|
/* | u64 argument | u64 entrypoint | KThread::StackParameters (size 0x30) | */
|
|
|
|
static_assert(sizeof(KThread::StackParameters) == 0x30);
|
|
|
|
|
|
|
|
u64 *stack = GetPointer<u64>(sp);
|
|
|
|
*(--stack) = GetInteger(pc);
|
|
|
|
*(--stack) = arg;
|
|
|
|
return reinterpret_cast<uintptr_t>(stack);
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
Result KThreadContext::Initialize(KVirtualAddress u_pc, KVirtualAddress k_sp, KVirtualAddress u_sp, uintptr_t arg, bool is_user, bool is_64_bit, bool is_main) {
|
|
|
|
MESOSPHERE_ASSERT(k_sp != Null<KVirtualAddress>);
|
|
|
|
|
|
|
|
/* Ensure that the stack pointers are aligned. */
|
|
|
|
k_sp = util::AlignDown(GetInteger(k_sp), 16);
|
|
|
|
u_sp = util::AlignDown(GetInteger(u_sp), 16);
|
|
|
|
|
|
|
|
/* Determine LR and SP. */
|
|
|
|
if (is_user) {
|
|
|
|
/* Usermode thread. */
|
2020-02-15 02:22:55 +00:00
|
|
|
this->lr = reinterpret_cast<uintptr_t>(::ams::kern::arch::arm64::UserModeThreadStarter);
|
2020-01-31 06:46:18 +00:00
|
|
|
this->sp = SetupStackForUserModeThreadStarter(u_pc, k_sp, u_sp, arg, is_64_bit);
|
|
|
|
} else {
|
|
|
|
/* Kernel thread. */
|
|
|
|
MESOSPHERE_ASSERT(is_64_bit);
|
|
|
|
|
|
|
|
if (is_main) {
|
|
|
|
/* Main thread. */
|
|
|
|
this->lr = GetInteger(u_pc);
|
|
|
|
this->sp = GetInteger(k_sp);
|
|
|
|
} else {
|
|
|
|
/* Generic Kernel thread. */
|
2020-02-15 02:22:55 +00:00
|
|
|
this->lr = reinterpret_cast<uintptr_t>(::ams::kern::arch::arm64::SupervisorModeThreadStarter);
|
2020-01-31 06:46:18 +00:00
|
|
|
this->sp = SetupStackForSupervisorModeThreadStarter(u_pc, k_sp, arg);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Clear callee-saved registers. */
|
|
|
|
for (size_t i = 0; i < util::size(this->callee_saved.registers); i++) {
|
|
|
|
this->callee_saved.registers[i] = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Clear FPU state. */
|
|
|
|
this->fpcr = 0;
|
|
|
|
this->fpsr = 0;
|
|
|
|
this->cpacr = 0;
|
|
|
|
for (size_t i = 0; i < util::size(this->fpu_registers); i++) {
|
|
|
|
this->fpu_registers[i] = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Lock the context, if we're a main thread. */
|
|
|
|
this->locked = is_main;
|
|
|
|
|
|
|
|
return ResultSuccess();
|
|
|
|
}
|
|
|
|
|
|
|
|
Result KThreadContext::Finalize() {
|
|
|
|
/* This doesn't actually do anything. */
|
|
|
|
return ResultSuccess();
|
|
|
|
}
|
|
|
|
|
2020-02-20 03:38:20 +00:00
|
|
|
void KThreadContext::SetArguments(uintptr_t arg0, uintptr_t arg1) {
|
|
|
|
u64 *stack = reinterpret_cast<u64 *>(this->sp);
|
|
|
|
stack[0] = arg0;
|
|
|
|
stack[1] = arg1;
|
|
|
|
}
|
|
|
|
|
2020-02-08 10:49:32 +00:00
|
|
|
void KThreadContext::FpuContextSwitchHandler(KThread *thread) {
|
|
|
|
MESOSPHERE_ASSERT(!KInterruptManager::AreInterruptsEnabled());
|
|
|
|
MESOSPHERE_ASSERT(!IsFpuEnabled());
|
|
|
|
|
|
|
|
/* Enable the FPU. */
|
|
|
|
EnableFpu();
|
|
|
|
|
|
|
|
/* Restore the FPU registers. */
|
|
|
|
KProcess *process = thread->GetOwnerProcess();
|
|
|
|
MESOSPHERE_ASSERT(process != nullptr);
|
|
|
|
if (process->Is64Bit()) {
|
2020-02-20 03:38:20 +00:00
|
|
|
RestoreFpuRegisters64(thread->GetContext());
|
2020-02-08 10:49:32 +00:00
|
|
|
} else {
|
2020-02-20 03:38:20 +00:00
|
|
|
RestoreFpuRegisters32(thread->GetContext());
|
2020-02-08 10:49:32 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-10 01:54:09 +00:00
|
|
|
void KThreadContext::CloneFpuStatus() {
|
|
|
|
u64 pcr, psr;
|
|
|
|
cpu::InstructionMemoryBarrier();
|
|
|
|
if (IsFpuEnabled()) {
|
|
|
|
__asm__ __volatile__("mrs %[pcr], fpcr" : [pcr]"=r"(pcr) :: "memory");
|
|
|
|
__asm__ __volatile__("mrs %[psr], fpsr" : [psr]"=r"(psr) :: "memory");
|
|
|
|
} else {
|
|
|
|
pcr = GetCurrentThread().GetContext().GetFpcr();
|
|
|
|
psr = GetCurrentThread().GetContext().GetFpsr();
|
|
|
|
}
|
|
|
|
|
|
|
|
this->SetFpcr(pcr);
|
|
|
|
this->SetFpsr(psr);
|
|
|
|
}
|
|
|
|
|
2020-07-31 09:49:43 +00:00
|
|
|
void KThreadContext::SetFpuRegisters(const u128 *v, bool is_64_bit) {
|
|
|
|
if (is_64_bit) {
|
|
|
|
for (size_t i = 0; i < KThreadContext::NumFpuRegisters; ++i) {
|
|
|
|
this->fpu_registers[i] = v[i];
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for (size_t i = 0; i < KThreadContext::NumFpuRegisters / 2; ++i) {
|
|
|
|
this->fpu_registers[i] = v[i];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-28 10:56:47 +00:00
|
|
|
void GetUserContext(ams::svc::ThreadContext *out, const KThread *thread) {
|
|
|
|
MESOSPHERE_ASSERT(KScheduler::IsSchedulerLockedByCurrentThread());
|
|
|
|
MESOSPHERE_ASSERT(thread->IsSuspended());
|
|
|
|
MESOSPHERE_ASSERT(thread->GetOwnerProcess() != nullptr);
|
|
|
|
|
|
|
|
/* Get the contexts. */
|
|
|
|
const KExceptionContext *e_ctx = GetExceptionContext(thread);
|
|
|
|
const KThreadContext *t_ctx = std::addressof(thread->GetContext());
|
|
|
|
|
|
|
|
if (thread->GetOwnerProcess()->Is64Bit()) {
|
|
|
|
/* Set special registers. */
|
|
|
|
out->fp = e_ctx->x[29];
|
|
|
|
out->lr = e_ctx->x[30];
|
|
|
|
out->sp = e_ctx->sp;
|
|
|
|
out->pc = e_ctx->pc;
|
2020-07-31 12:52:59 +00:00
|
|
|
out->pstate = e_ctx->psr & El0PsrMask;
|
2020-07-28 10:56:47 +00:00
|
|
|
|
|
|
|
/* Get the thread's general purpose registers. */
|
|
|
|
if (thread->IsCallingSvc()) {
|
|
|
|
for (size_t i = 19; i < 29; ++i) {
|
|
|
|
out->r[i] = e_ctx->x[i];
|
|
|
|
}
|
|
|
|
if (e_ctx->write == 0) {
|
|
|
|
out->pc -= sizeof(u32);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for (size_t i = 0; i < 29; ++i) {
|
|
|
|
out->r[i] = e_ctx->x[i];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Copy tpidr. */
|
|
|
|
out->tpidr = e_ctx->tpidr;
|
|
|
|
|
|
|
|
/* Copy fpu registers. */
|
|
|
|
static_assert(util::size(ams::svc::ThreadContext{}.v) == KThreadContext::NumFpuRegisters);
|
|
|
|
const u128 *f = t_ctx->GetFpuRegisters();
|
|
|
|
for (size_t i = 0; i < KThreadContext::NumFpuRegisters; ++i) {
|
|
|
|
out->v[i] = f[i];
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/* Set special registers. */
|
|
|
|
out->pc = static_cast<u32>(e_ctx->pc);
|
|
|
|
out->pstate = e_ctx->psr & 0xFF0FFE20;
|
|
|
|
|
|
|
|
/* Get the thread's general purpose registers. */
|
|
|
|
for (size_t i = 0; i < 15; ++i) {
|
|
|
|
out->r[i] = static_cast<u32>(e_ctx->x[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Adjust PC, if the thread is calling svc. */
|
|
|
|
if (thread->IsCallingSvc()) {
|
|
|
|
if (e_ctx->write == 0) {
|
|
|
|
/* Adjust by 2 if thumb mode, 4 if arm mode. */
|
|
|
|
out->pc -= ((e_ctx->psr & 0x20) == 0) ? sizeof(u32) : sizeof(u16);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Copy tpidr. */
|
|
|
|
out->tpidr = static_cast<u32>(e_ctx->tpidr);
|
|
|
|
|
|
|
|
/* Copy fpu registers. */
|
|
|
|
static_assert(util::size(ams::svc::ThreadContext{}.v) == KThreadContext::NumFpuRegisters);
|
|
|
|
const u128 *f = t_ctx->GetFpuRegisters();
|
|
|
|
for (size_t i = 0; i < KThreadContext::NumFpuRegisters / 2; ++i) {
|
|
|
|
out->v[i] = f[i];
|
|
|
|
}
|
2020-07-31 09:49:43 +00:00
|
|
|
for (size_t i = KThreadContext::NumFpuRegisters / 2; i < KThreadContext::NumFpuRegisters; ++i) {
|
|
|
|
out->v[i] = 0;
|
|
|
|
}
|
2020-07-28 10:56:47 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Copy fpcr/fpsr. */
|
|
|
|
out->fpcr = t_ctx->GetFpcr();
|
|
|
|
out->fpsr = t_ctx->GetFpsr();
|
|
|
|
}
|
|
|
|
|
2020-07-31 08:27:09 +00:00
|
|
|
void KThreadContext::OnThreadTerminating(const KThread *thread) {
|
2020-08-17 21:20:24 +00:00
|
|
|
MESOSPHERE_UNUSED(thread);
|
2020-07-31 08:27:09 +00:00
|
|
|
/* ... */
|
|
|
|
}
|
|
|
|
|
2020-01-31 06:46:18 +00:00
|
|
|
}
|