diff --git a/libraries/libmesosphere/include/mesosphere/arch/arm64/init/kern_k_init_arguments.hpp b/libraries/libmesosphere/include/mesosphere/arch/arm64/init/kern_k_init_arguments.hpp index 04abbd3da..204f34a7d 100644 --- a/libraries/libmesosphere/include/mesosphere/arch/arm64/init/kern_k_init_arguments.hpp +++ b/libraries/libmesosphere/include/mesosphere/arch/arm64/init/kern_k_init_arguments.hpp @@ -14,6 +14,7 @@ * along with this program. If not, see . */ #pragma once +#include namespace ams::kern::init { @@ -31,5 +32,19 @@ namespace ams::kern::init { u64 setup_function; u64 exception_stack; }; + static_assert(sizeof(KInitArguments) == INIT_ARGUMENTS_SIZE); + + static_assert(__builtin_offsetof(KInitArguments, ttbr0) == INIT_ARGUMENTS_TTBR0); + static_assert(__builtin_offsetof(KInitArguments, ttbr1) == INIT_ARGUMENTS_TTBR1); + static_assert(__builtin_offsetof(KInitArguments, tcr) == INIT_ARGUMENTS_TCR); + static_assert(__builtin_offsetof(KInitArguments, mair) == INIT_ARGUMENTS_MAIR); + static_assert(__builtin_offsetof(KInitArguments, cpuactlr) == INIT_ARGUMENTS_CPUACTLR); + static_assert(__builtin_offsetof(KInitArguments, cpuectlr) == INIT_ARGUMENTS_CPUECTLR); + static_assert(__builtin_offsetof(KInitArguments, sctlr) == INIT_ARGUMENTS_SCTLR); + static_assert(__builtin_offsetof(KInitArguments, sp) == INIT_ARGUMENTS_SP); + static_assert(__builtin_offsetof(KInitArguments, entrypoint) == INIT_ARGUMENTS_ENTRYPOINT); + static_assert(__builtin_offsetof(KInitArguments, argument) == INIT_ARGUMENTS_ARGUMENT); + static_assert(__builtin_offsetof(KInitArguments, setup_function) == INIT_ARGUMENTS_SETUP_FUNCTION); + static_assert(__builtin_offsetof(KInitArguments, exception_stack) == INIT_ARGUMENTS_EXCEPTION_STACK); } \ No newline at end of file diff --git a/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_assembly_offsets.h b/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_assembly_offsets.h index f756d85c1..2b9e4f01e 100644 --- a/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_assembly_offsets.h +++ b/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_assembly_offsets.h @@ -15,6 +15,9 @@ */ #pragma once +/* TODO: Different header for this? */ +#define AMS_KERN_NUM_SUPERVISOR_CALLS 0xC0 + #define THREAD_STACK_PARAMETERS_SIZE 0x30 #define THREAD_STACK_PARAMETERS_SVC_PERMISSION 0x00 #define THREAD_STACK_PARAMETERS_CONTEXT 0x18 @@ -24,4 +27,124 @@ #define THREAD_STACK_PARAMETERS_CURRENT_SVC_ID 0x2B #define THREAD_STACK_PARAMETERS_IS_CALLING_SVC 0x2C #define THREAD_STACK_PARAMETERS_IS_IN_EXCEPTION_HANDLER 0x2D -#define THREAD_STACK_PARAMETERS_IS_PINNED 0x2E \ No newline at end of file +#define THREAD_STACK_PARAMETERS_IS_PINNED 0x2E + +#define THREAD_CONTEXT_SIZE 0x290 +#define THREAD_CONTEXT_CPU_REGISTERS 0x000 +#define THREAD_CONTEXT_X19 0x000 +#define THREAD_CONTEXT_X20 0x008 +#define THREAD_CONTEXT_X21 0x010 +#define THREAD_CONTEXT_X22 0x018 +#define THREAD_CONTEXT_X23 0x020 +#define THREAD_CONTEXT_X24 0x028 +#define THREAD_CONTEXT_X25 0x030 +#define THREAD_CONTEXT_X26 0x038 +#define THREAD_CONTEXT_X27 0x040 +#define THREAD_CONTEXT_X28 0x048 +#define THREAD_CONTEXT_X29 0x050 +#define THREAD_CONTEXT_LR 0x058 +#define THREAD_CONTEXT_SP 0x060 +#define THREAD_CONTEXT_CPACR 0x068 +#define THREAD_CONTEXT_FPCR 0x070 +#define THREAD_CONTEXT_FPSR 0x078 +#define THREAD_CONTEXT_FPU_REGISTERS 0x080 +#define THREAD_CONTEXT_LOCKED 0x280 + +#define THREAD_CONTEXT_X19_X20 THREAD_CONTEXT_X19 +#define THREAD_CONTEXT_X21_X22 THREAD_CONTEXT_X21 +#define THREAD_CONTEXT_X23_X24 THREAD_CONTEXT_X23 +#define THREAD_CONTEXT_X25_X26 THREAD_CONTEXT_X25 +#define THREAD_CONTEXT_X27_X28 THREAD_CONTEXT_X27 +#define THREAD_CONTEXT_X29_X30 THREAD_CONTEXT_X29 +#define THREAD_CONTEXT_LR_SP THREAD_CONTEXT_LR +#define THREAD_CONTEXT_SP_CPACR THREAD_CONTEXT_SP +#define THREAD_CONTEXT_FPCR_FPSR THREAD_CONTEXT_FPCR + +#define EXCEPTION_CONTEXT_SIZE 0x120 +#define EXCEPTION_CONTEXT_X0 0x000 +#define EXCEPTION_CONTEXT_X1 0x008 +#define EXCEPTION_CONTEXT_X2 0x010 +#define EXCEPTION_CONTEXT_X3 0x018 +#define EXCEPTION_CONTEXT_X4 0x020 +#define EXCEPTION_CONTEXT_X5 0x028 +#define EXCEPTION_CONTEXT_X6 0x030 +#define EXCEPTION_CONTEXT_X7 0x038 +#define EXCEPTION_CONTEXT_X8 0x040 +#define EXCEPTION_CONTEXT_X9 0x048 +#define EXCEPTION_CONTEXT_X10 0x050 +#define EXCEPTION_CONTEXT_X11 0x058 +#define EXCEPTION_CONTEXT_X12 0x060 +#define EXCEPTION_CONTEXT_X13 0x068 +#define EXCEPTION_CONTEXT_X14 0x070 +#define EXCEPTION_CONTEXT_X15 0x078 +#define EXCEPTION_CONTEXT_X16 0x080 +#define EXCEPTION_CONTEXT_X17 0x088 +#define EXCEPTION_CONTEXT_X18 0x090 +#define EXCEPTION_CONTEXT_X19 0x098 +#define EXCEPTION_CONTEXT_X20 0x0A0 +#define EXCEPTION_CONTEXT_X21 0x0A8 +#define EXCEPTION_CONTEXT_X22 0x0B0 +#define EXCEPTION_CONTEXT_X23 0x0B8 +#define EXCEPTION_CONTEXT_X24 0x0C0 +#define EXCEPTION_CONTEXT_X25 0x0C8 +#define EXCEPTION_CONTEXT_X26 0x0D0 +#define EXCEPTION_CONTEXT_X27 0x0D8 +#define EXCEPTION_CONTEXT_X28 0x0E0 +#define EXCEPTION_CONTEXT_X29 0x0E8 +#define EXCEPTION_CONTEXT_X30 0x0F0 +#define EXCEPTION_CONTEXT_SP 0x0F8 +#define EXCEPTION_CONTEXT_PC 0x100 +#define EXCEPTION_CONTEXT_PSR 0x108 +#define EXCEPTION_CONTEXT_TPIDR 0x110 + +#define EXCEPTION_CONTEXT_X0_X1 EXCEPTION_CONTEXT_X0 +#define EXCEPTION_CONTEXT_X2_X3 EXCEPTION_CONTEXT_X2 +#define EXCEPTION_CONTEXT_X4_X5 EXCEPTION_CONTEXT_X4 +#define EXCEPTION_CONTEXT_X6_X7 EXCEPTION_CONTEXT_X6 +#define EXCEPTION_CONTEXT_X8_X9 EXCEPTION_CONTEXT_X8 +#define EXCEPTION_CONTEXT_X10_X11 EXCEPTION_CONTEXT_X10 +#define EXCEPTION_CONTEXT_X12_X13 EXCEPTION_CONTEXT_X12 +#define EXCEPTION_CONTEXT_X14_X15 EXCEPTION_CONTEXT_X14 +#define EXCEPTION_CONTEXT_X16_X17 EXCEPTION_CONTEXT_X16 +#define EXCEPTION_CONTEXT_X18_X19 EXCEPTION_CONTEXT_X18 +#define EXCEPTION_CONTEXT_X20_X21 EXCEPTION_CONTEXT_X20 +#define EXCEPTION_CONTEXT_X22_X23 EXCEPTION_CONTEXT_X22 +#define EXCEPTION_CONTEXT_X24_X25 EXCEPTION_CONTEXT_X24 +#define EXCEPTION_CONTEXT_X26_X27 EXCEPTION_CONTEXT_X26 +#define EXCEPTION_CONTEXT_X28_X29 EXCEPTION_CONTEXT_X28 +#define EXCEPTION_CONTEXT_X30_SP EXCEPTION_CONTEXT_X30 +#define EXCEPTION_CONTEXT_PC_PSR EXCEPTION_CONTEXT_PC + +#define EXCEPTION_CONTEXT_X9_X10 EXCEPTION_CONTEXT_X9 +#define EXCEPTION_CONTEXT_X19_X20 EXCEPTION_CONTEXT_X19 +#define EXCEPTION_CONTEXT_X21_X22 EXCEPTION_CONTEXT_X21 +#define EXCEPTION_CONTEXT_X23_X24 EXCEPTION_CONTEXT_X23 +#define EXCEPTION_CONTEXT_X25_X26 EXCEPTION_CONTEXT_X25 +#define EXCEPTION_CONTEXT_X27_X28 EXCEPTION_CONTEXT_X27 +#define EXCEPTION_CONTEXT_X29_X30 EXCEPTION_CONTEXT_X29 +#define EXCEPTION_CONTEXT_SP_PC EXCEPTION_CONTEXT_SP +#define EXCEPTION_CONTEXT_PSR_TPIDR EXCEPTION_CONTEXT_PSR + +#define THREAD_LOCAL_REGION_MESSAGE_BUFFER 0x000 +#define THREAD_LOCAL_REGION_DISABLE_COUNT 0x100 +#define THREAD_LOCAL_REGION_INTERRUPT_FLAG 0x102 +#define THREAD_LOCAL_REGION_SIZE 0x200 + +#define INIT_ARGUMENTS_SIZE 0x60 +#define INIT_ARGUMENTS_TTBR0 0x00 +#define INIT_ARGUMENTS_TTBR1 0x08 +#define INIT_ARGUMENTS_TCR 0x10 +#define INIT_ARGUMENTS_MAIR 0x18 +#define INIT_ARGUMENTS_CPUACTLR 0x20 +#define INIT_ARGUMENTS_CPUECTLR 0x28 +#define INIT_ARGUMENTS_SCTLR 0x30 +#define INIT_ARGUMENTS_SP 0x38 +#define INIT_ARGUMENTS_ENTRYPOINT 0x40 +#define INIT_ARGUMENTS_ARGUMENT 0x48 +#define INIT_ARGUMENTS_SETUP_FUNCTION 0x50 +#define INIT_ARGUMENTS_EXCEPTION_STACK 0x58 + +#define KSCHEDULER_NEEDS_SCHEDULING 0x00 +#define KSCHEDULER_INTERRUPT_TASK_THREAD_RUNNABLE 0x01 +#define KSCHEDULER_HIGHEST_PRIORITY_THREAD 0x10 +#define KSCHEDULER_IDLE_THREAD_STACK 0x18 diff --git a/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_exception_context.hpp b/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_exception_context.hpp index 40982b697..255fec7e3 100644 --- a/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_exception_context.hpp +++ b/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_exception_context.hpp @@ -43,6 +43,42 @@ namespace ams::kern::arch::arm64 { } } }; - static_assert(sizeof(KExceptionContext) == 0x120); + static_assert(sizeof(KExceptionContext) == EXCEPTION_CONTEXT_SIZE); + + static_assert(__builtin_offsetof(KExceptionContext, x[ 0]) == EXCEPTION_CONTEXT_X0); + static_assert(__builtin_offsetof(KExceptionContext, x[ 1]) == EXCEPTION_CONTEXT_X1); + static_assert(__builtin_offsetof(KExceptionContext, x[ 2]) == EXCEPTION_CONTEXT_X2); + static_assert(__builtin_offsetof(KExceptionContext, x[ 3]) == EXCEPTION_CONTEXT_X3); + static_assert(__builtin_offsetof(KExceptionContext, x[ 4]) == EXCEPTION_CONTEXT_X4); + static_assert(__builtin_offsetof(KExceptionContext, x[ 5]) == EXCEPTION_CONTEXT_X5); + static_assert(__builtin_offsetof(KExceptionContext, x[ 6]) == EXCEPTION_CONTEXT_X6); + static_assert(__builtin_offsetof(KExceptionContext, x[ 7]) == EXCEPTION_CONTEXT_X7); + static_assert(__builtin_offsetof(KExceptionContext, x[ 8]) == EXCEPTION_CONTEXT_X8); + static_assert(__builtin_offsetof(KExceptionContext, x[ 9]) == EXCEPTION_CONTEXT_X9); + static_assert(__builtin_offsetof(KExceptionContext, x[10]) == EXCEPTION_CONTEXT_X10); + static_assert(__builtin_offsetof(KExceptionContext, x[11]) == EXCEPTION_CONTEXT_X11); + static_assert(__builtin_offsetof(KExceptionContext, x[12]) == EXCEPTION_CONTEXT_X12); + static_assert(__builtin_offsetof(KExceptionContext, x[13]) == EXCEPTION_CONTEXT_X13); + static_assert(__builtin_offsetof(KExceptionContext, x[14]) == EXCEPTION_CONTEXT_X14); + static_assert(__builtin_offsetof(KExceptionContext, x[15]) == EXCEPTION_CONTEXT_X15); + static_assert(__builtin_offsetof(KExceptionContext, x[16]) == EXCEPTION_CONTEXT_X16); + static_assert(__builtin_offsetof(KExceptionContext, x[17]) == EXCEPTION_CONTEXT_X17); + static_assert(__builtin_offsetof(KExceptionContext, x[18]) == EXCEPTION_CONTEXT_X18); + static_assert(__builtin_offsetof(KExceptionContext, x[19]) == EXCEPTION_CONTEXT_X19); + static_assert(__builtin_offsetof(KExceptionContext, x[20]) == EXCEPTION_CONTEXT_X20); + static_assert(__builtin_offsetof(KExceptionContext, x[21]) == EXCEPTION_CONTEXT_X21); + static_assert(__builtin_offsetof(KExceptionContext, x[22]) == EXCEPTION_CONTEXT_X22); + static_assert(__builtin_offsetof(KExceptionContext, x[23]) == EXCEPTION_CONTEXT_X23); + static_assert(__builtin_offsetof(KExceptionContext, x[24]) == EXCEPTION_CONTEXT_X24); + static_assert(__builtin_offsetof(KExceptionContext, x[25]) == EXCEPTION_CONTEXT_X25); + static_assert(__builtin_offsetof(KExceptionContext, x[26]) == EXCEPTION_CONTEXT_X26); + static_assert(__builtin_offsetof(KExceptionContext, x[27]) == EXCEPTION_CONTEXT_X27); + static_assert(__builtin_offsetof(KExceptionContext, x[28]) == EXCEPTION_CONTEXT_X28); + static_assert(__builtin_offsetof(KExceptionContext, x[29]) == EXCEPTION_CONTEXT_X29); + static_assert(__builtin_offsetof(KExceptionContext, x[30]) == EXCEPTION_CONTEXT_X30); + static_assert(__builtin_offsetof(KExceptionContext, sp) == EXCEPTION_CONTEXT_SP); + static_assert(__builtin_offsetof(KExceptionContext, pc) == EXCEPTION_CONTEXT_PC); + static_assert(__builtin_offsetof(KExceptionContext, psr) == EXCEPTION_CONTEXT_PSR); + static_assert(__builtin_offsetof(KExceptionContext, tpidr) == EXCEPTION_CONTEXT_TPIDR); } \ No newline at end of file diff --git a/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_thread_context.hpp b/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_thread_context.hpp index d2de2c88f..e09a48790 100644 --- a/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_thread_context.hpp +++ b/libraries/libmesosphere/include/mesosphere/arch/arm64/kern_k_thread_context.hpp @@ -79,8 +79,38 @@ namespace ams::kern::arch::arm64 { const u128 *GetFpuRegisters() const { return m_fpu_registers; } public: static void OnThreadTerminating(const KThread *thread); + public: + static consteval bool ValidateOffsets(); }; + consteval bool KThreadContext::ValidateOffsets() { + static_assert(sizeof(KThreadContext) == THREAD_CONTEXT_SIZE); + + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.registers) == THREAD_CONTEXT_CPU_REGISTERS); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x19) == THREAD_CONTEXT_X19); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x20) == THREAD_CONTEXT_X20); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x21) == THREAD_CONTEXT_X21); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x22) == THREAD_CONTEXT_X22); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x23) == THREAD_CONTEXT_X23); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x24) == THREAD_CONTEXT_X24); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x25) == THREAD_CONTEXT_X25); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x26) == THREAD_CONTEXT_X26); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x27) == THREAD_CONTEXT_X27); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x28) == THREAD_CONTEXT_X28); + static_assert(__builtin_offsetof(KThreadContext, m_callee_saved.x29) == THREAD_CONTEXT_X29); + static_assert(__builtin_offsetof(KThreadContext, m_lr) == THREAD_CONTEXT_LR); + static_assert(__builtin_offsetof(KThreadContext, m_sp) == THREAD_CONTEXT_SP); + static_assert(__builtin_offsetof(KThreadContext, m_cpacr) == THREAD_CONTEXT_CPACR); + static_assert(__builtin_offsetof(KThreadContext, m_fpcr) == THREAD_CONTEXT_FPCR); + static_assert(__builtin_offsetof(KThreadContext, m_fpsr) == THREAD_CONTEXT_FPSR); + static_assert(__builtin_offsetof(KThreadContext, m_fpu_registers) == THREAD_CONTEXT_FPU_REGISTERS); + static_assert(__builtin_offsetof(KThreadContext, m_locked) == THREAD_CONTEXT_LOCKED); + + return true; + } + static_assert(KThreadContext::ValidateOffsets()); + + void GetUserContext(ams::svc::ThreadContext *out, const KThread *thread); } \ No newline at end of file diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp index 5bcbab66f..942365e22 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp @@ -194,8 +194,20 @@ namespace ams::kern { static bool s_scheduler_update_needed; static KSchedulerPriorityQueue s_priority_queue; static LockType s_scheduler_lock; + public: + static consteval bool ValidateAssemblyOffsets(); }; + consteval bool KScheduler::ValidateAssemblyOffsets() { + static_assert(__builtin_offsetof(KScheduler, m_state.needs_scheduling) == KSCHEDULER_NEEDS_SCHEDULING); + static_assert(__builtin_offsetof(KScheduler, m_state.interrupt_task_thread_runnable) == KSCHEDULER_INTERRUPT_TASK_THREAD_RUNNABLE); + static_assert(__builtin_offsetof(KScheduler, m_state.highest_priority_thread) == KSCHEDULER_HIGHEST_PRIORITY_THREAD); + static_assert(__builtin_offsetof(KScheduler, m_state.idle_thread_stack) == KSCHEDULER_IDLE_THREAD_STACK); + + return true; + } + static_assert(KScheduler::ValidateAssemblyOffsets()); + class KScopedSchedulerLock : KScopedLock { public: explicit ALWAYS_INLINE KScopedSchedulerLock() : KScopedLock(KScheduler::s_scheduler_lock) { /* ... */ } diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_thread_local_page.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_thread_local_page.hpp index a41c49694..489f0ca02 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_thread_local_page.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_thread_local_page.hpp @@ -105,4 +105,10 @@ namespace ams::kern { } }; + /* Miscellaneous sanity checking. */ + static_assert(ams::svc::ThreadLocalRegionSize == THREAD_LOCAL_REGION_SIZE); + static_assert(__builtin_offsetof(ams::svc::ThreadLocalRegion, message_buffer) == THREAD_LOCAL_REGION_MESSAGE_BUFFER); + static_assert(__builtin_offsetof(ams::svc::ThreadLocalRegion, disable_count) == THREAD_LOCAL_REGION_DISABLE_COUNT); + static_assert(__builtin_offsetof(ams::svc::ThreadLocalRegion, interrupt_flag) == THREAD_LOCAL_REGION_INTERRUPT_FLAG); + } diff --git a/libraries/libmesosphere/include/mesosphere/svc/kern_svc_prototypes.hpp b/libraries/libmesosphere/include/mesosphere/svc/kern_svc_prototypes.hpp index 078b2cd2b..a9163b56e 100644 --- a/libraries/libmesosphere/include/mesosphere/svc/kern_svc_prototypes.hpp +++ b/libraries/libmesosphere/include/mesosphere/svc/kern_svc_prototypes.hpp @@ -20,7 +20,7 @@ namespace ams::kern::svc { - static constexpr size_t NumSupervisorCalls = 0xC0; + static constexpr size_t NumSupervisorCalls = AMS_KERN_NUM_SUPERVISOR_CALLS; #define AMS_KERN_SVC_DECLARE_ENUM_ID(ID, RETURN_TYPE, NAME, ...) \ SvcId_##NAME = ID, diff --git a/libraries/libmesosphere/source/arch/arm64/kern_panic_asm.s b/libraries/libmesosphere/source/arch/arm64/kern_panic_asm.s index 7d317ac60..33ece6e0f 100644 --- a/libraries/libmesosphere/source/arch/arm64/kern_panic_asm.s +++ b/libraries/libmesosphere/source/arch/arm64/kern_panic_asm.s @@ -14,6 +14,7 @@ * along with this program. If not, see . */ #include +#include #if defined(MESOSPHERE_ENABLE_PANIC_REGISTER_DUMP) @@ -32,28 +33,28 @@ \ /* Save x0/x1/sp to the context. */ \ ldr x1, [sp, #(8 * 0)]; \ - str x1, [x0, #(8 * 0)]; \ + str x1, [x0, #(EXCEPTION_CONTEXT_X0)]; \ ldr x1, [sp, #(8 * 1)]; \ - str x1, [x0, #(8 * 1)]; \ + str x1, [x0, #(EXCEPTION_CONTEXT_X1)]; \ \ /* Save all other registers to the context. */ \ - stp x2, x3, [x0, #(8 * 2)]; \ - stp x4, x5, [x0, #(8 * 4)]; \ - stp x6, x7, [x0, #(8 * 6)]; \ - stp x8, x9, [x0, #(8 * 8)]; \ - stp x10, x11, [x0, #(8 * 10)]; \ - stp x12, x13, [x0, #(8 * 12)]; \ - stp x14, x15, [x0, #(8 * 14)]; \ - stp x16, x17, [x0, #(8 * 16)]; \ - stp x18, x19, [x0, #(8 * 18)]; \ - stp x20, x21, [x0, #(8 * 20)]; \ - stp x22, x23, [x0, #(8 * 22)]; \ - stp x24, x25, [x0, #(8 * 24)]; \ - stp x26, x27, [x0, #(8 * 26)]; \ - stp x28, x29, [x0, #(8 * 28)]; \ + stp x2, x3, [x0, #(EXCEPTION_CONTEXT_X2_X3)]; \ + stp x4, x5, [x0, #(EXCEPTION_CONTEXT_X4_X5)]; \ + stp x6, x7, [x0, #(EXCEPTION_CONTEXT_X6_X7)]; \ + stp x8, x9, [x0, #(EXCEPTION_CONTEXT_X8_X9)]; \ + stp x10, x11, [x0, #(EXCEPTION_CONTEXT_X10_X11)]; \ + stp x12, x13, [x0, #(EXCEPTION_CONTEXT_X12_X13)]; \ + stp x14, x15, [x0, #(EXCEPTION_CONTEXT_X14_X15)]; \ + stp x16, x17, [x0, #(EXCEPTION_CONTEXT_X16_X17)]; \ + stp x18, x19, [x0, #(EXCEPTION_CONTEXT_X18_X19)]; \ + stp x20, x21, [x0, #(EXCEPTION_CONTEXT_X20_X21)]; \ + stp x22, x23, [x0, #(EXCEPTION_CONTEXT_X22_X23)]; \ + stp x24, x25, [x0, #(EXCEPTION_CONTEXT_X24_X25)]; \ + stp x26, x27, [x0, #(EXCEPTION_CONTEXT_X26_X27)]; \ + stp x28, x29, [x0, #(EXCEPTION_CONTEXT_X28_X29)]; \ \ add x1, sp, #16; \ - stp x30, x1, [x0, #(8 * 30)]; \ + stp x30, x1, [x0, #(EXCEPTION_CONTEXT_X30_SP)]; \ \ /* Restore x0/x1. */ \ ldp x0, x1, [sp], #16; diff --git a/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_exception_asm.s b/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_exception_asm.s index e3f72e410..428a5e185 100644 --- a/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_exception_asm.s +++ b/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_exception_asm.s @@ -21,15 +21,15 @@ .type _ZN3ams4kern3svc25CallReturnFromException64Ev, %function _ZN3ams4kern3svc25CallReturnFromException64Ev: /* Save registers the SVC entry handler didn't. */ - stp x12, x13, [sp, #(8 * 12)] - stp x14, x15, [sp, #(8 * 14)] - stp x16, x17, [sp, #(8 * 16)] - str x19, [sp, #(8 * 19)] - stp x20, x21, [sp, #(8 * 20)] - stp x22, x23, [sp, #(8 * 22)] - stp x24, x25, [sp, #(8 * 24)] - stp x26, x26, [sp, #(8 * 26)] - stp x28, x29, [sp, #(8 * 28)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + str x19, [sp, #(EXCEPTION_CONTEXT_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp x26, x26, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] /* Call ams::kern::arch::arm64::ReturnFromException(result). */ bl _ZN3ams4kern4arch5arm6419ReturnFromExceptionENS_6ResultE @@ -63,7 +63,7 @@ _ZN3ams4kern3svc14RestoreContextEm: 0: /* We should handle DPC. */ /* Check the dpc flags. */ - ldrb w8, [sp, #(0x120 + THREAD_STACK_PARAMETERS_DPC_FLAGS)] + ldrb w8, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_DPC_FLAGS)] cbz w8, 1f /* We have DPC to do! */ @@ -83,32 +83,32 @@ _ZN3ams4kern3svc14RestoreContextEm: 1: /* We're done with DPC, and should return from the svc. */ /* Clear our in-SVC note. */ - strb wzr, [sp, #(0x120 + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] + strb wzr, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] /* Restore registers. */ - ldp x30, x8, [sp, #(8 * 30)] - ldp x9, x10, [sp, #(8 * 32)] - ldr x11, [sp, #(8 * 34)] + ldp x30, x8, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + ldp x9, x10, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x11, [sp, #(EXCEPTION_CONTEXT_TPIDR)] msr sp_el0, x8 msr elr_el1, x9 msr spsr_el1, x10 msr tpidr_el0, x11 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] - ldp x16, x17, [sp, #(8 * 16)] - ldp x18, x19, [sp, #(8 * 18)] - ldp x20, x21, [sp, #(8 * 20)] - ldp x22, x23, [sp, #(8 * 22)] - ldp x24, x25, [sp, #(8 * 24)] - ldp x26, x27, [sp, #(8 * 26)] - ldp x28, x29, [sp, #(8 * 28)] + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + ldp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + ldp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + ldp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + ldp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + ldp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] /* Return. */ - add sp, sp, #0x120 + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) eret diff --git a/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_handlers_asm.s b/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_handlers_asm.s index eaffea9cd..429ccfed2 100644 --- a/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_handlers_asm.s +++ b/libraries/libmesosphere/source/arch/arm64/svc/kern_svc_handlers_asm.s @@ -22,45 +22,45 @@ .type _ZN3ams4kern4arch5arm6412SvcHandler64Ev, %function _ZN3ams4kern4arch5arm6412SvcHandler64Ev: /* Create a KExceptionContext for the exception. */ - sub sp, sp, #0x120 + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) /* Save registers needed for ReturnFromException */ - stp x9, x10, [sp, #(8 * 9)] - str x11, [sp, #(8 * 11)] - str x18, [sp, #(8 * 18)] + stp x9, x10, [sp, #(EXCEPTION_CONTEXT_X9_X10)] + str x11, [sp, #(EXCEPTION_CONTEXT_X11)] + str x18, [sp, #(EXCEPTION_CONTEXT_X18)] mrs x8, sp_el0 mrs x9, elr_el1 mrs x10, spsr_el1 mrs x11, tpidr_el0 - ldr x18, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CUR_THREAD)] + ldr x18, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CUR_THREAD)] /* Save callee-saved registers. */ - stp x19, x20, [sp, #(8 * 19)] - stp x21, x22, [sp, #(8 * 21)] - stp x23, x24, [sp, #(8 * 23)] - stp x25, x26, [sp, #(8 * 25)] - stp x27, x28, [sp, #(8 * 27)] + stp x19, x20, [sp, #(EXCEPTION_CONTEXT_X19_X20)] + stp x21, x22, [sp, #(EXCEPTION_CONTEXT_X21_X22)] + stp x23, x24, [sp, #(EXCEPTION_CONTEXT_X23_X24)] + stp x25, x26, [sp, #(EXCEPTION_CONTEXT_X25_X26)] + stp x27, x28, [sp, #(EXCEPTION_CONTEXT_X27_X28)] /* Save miscellaneous registers. */ - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x29, x30, [sp, #(8 * 29)] - stp x8, x9, [sp, #(8 * 31)] - stp x10, x11, [sp, #(8 * 33)] + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x29, x30, [sp, #(EXCEPTION_CONTEXT_X29_X30)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_SP_PC)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_PSR_TPIDR)] /* Check if the SVC index is out of range. */ mrs x8, esr_el1 and x8, x8, #0xFF - cmp x8, #0x80 + cmp x8, #(AMS_KERN_NUM_SUPERVISOR_CALLS) b.ge 3f /* Check the specific SVC permission bit for allowal. */ mov x9, sp add x9, x9, x8, lsr#3 - ldrb w9, [x9, #(0x120 + THREAD_STACK_PARAMETERS_SVC_PERMISSION)] + ldrb w9, [x9, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_SVC_PERMISSION)] and x10, x8, #0x7 lsr x10, x9, x10 tst x10, #1 @@ -68,11 +68,11 @@ _ZN3ams4kern4arch5arm6412SvcHandler64Ev: /* Check if our disable count allows us to call SVCs. */ mrs x10, tpidrro_el0 - ldrh w10, [x10, #0x100] + ldrh w10, [x10, #(THREAD_LOCAL_REGION_DISABLE_COUNT)] cbz w10, 1f /* It might not, so check the stack params to see if we must not allow the SVC. */ - ldrb w10, [sp, #(0x120 + THREAD_STACK_PARAMETERS_IS_PINNED)] + ldrb w10, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_IS_PINNED)] cbz w10, 3f 1: /* We can call the SVC. */ @@ -82,8 +82,8 @@ _ZN3ams4kern4arch5arm6412SvcHandler64Ev: /* Note that we're calling the SVC. */ mov w10, #1 - strb w10, [sp, #(0x120 + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] - strb w8, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CURRENT_SVC_ID)] + strb w10, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] + strb w8, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CURRENT_SVC_ID)] /* If we should, trace the svc entry. */ #if defined(MESOSPHERE_BUILD_FOR_TRACING) @@ -110,7 +110,7 @@ _ZN3ams4kern4arch5arm6412SvcHandler64Ev: 2: /* We completed the SVC, and we should handle DPC. */ /* Check the dpc flags. */ - ldrb w8, [sp, #(0x120 + THREAD_STACK_PARAMETERS_DPC_FLAGS)] + ldrb w8, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_DPC_FLAGS)] cbz w8, 4f /* We have DPC to do! */ @@ -130,57 +130,57 @@ _ZN3ams4kern4arch5arm6412SvcHandler64Ev: 3: /* Invalid SVC. */ /* Setup the context to call into HandleException. */ - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp xzr, xzr, [sp, #(8 * 8)] - stp xzr, xzr, [sp, #(8 * 10)] - stp xzr, xzr, [sp, #(8 * 12)] - stp xzr, xzr, [sp, #(8 * 14)] - stp xzr, xzr, [sp, #(8 * 16)] - str x19, [sp, #(8 * 19)] - stp x20, x21, [sp, #(8 * 20)] - stp x22, x23, [sp, #(8 * 22)] - stp x24, x25, [sp, #(8 * 24)] - stp x26, x27, [sp, #(8 * 26)] - stp x28, x29, [sp, #(8 * 28)] + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + str x19, [sp, #(EXCEPTION_CONTEXT_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] /* Call ams::kern::arch::arm64::HandleException(ams::kern::arch::arm64::KExceptionContext *) */ mov x0, sp bl _ZN3ams4kern4arch5arm6415HandleExceptionEPNS2_17KExceptionContextE /* Restore registers. */ - ldp x30, x8, [sp, #(8 * 30)] - ldp x9, x10, [sp, #(8 * 32)] - ldr x11, [sp, #(8 * 34)] + ldp x30, x8, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + ldp x9, x10, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x11, [sp, #(EXCEPTION_CONTEXT_TPIDR)] msr sp_el0, x8 msr elr_el1, x9 msr spsr_el1, x10 msr tpidr_el0, x11 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] - ldp x16, x17, [sp, #(8 * 16)] - ldp x18, x19, [sp, #(8 * 18)] - ldp x20, x21, [sp, #(8 * 20)] - ldp x22, x23, [sp, #(8 * 22)] - ldp x24, x25, [sp, #(8 * 24)] - ldp x26, x27, [sp, #(8 * 26)] - ldp x28, x29, [sp, #(8 * 28)] + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + ldp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + ldp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + ldp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + ldp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + ldp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] /* Return. */ - add sp, sp, #0x120 + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) eret 4: /* Return from SVC. */ /* Clear our in-SVC note. */ - strb wzr, [sp, #(0x120 + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] + strb wzr, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] /* If we should, trace the svc exit. */ #if defined(MESOSPHERE_BUILD_FOR_TRACING) @@ -199,10 +199,10 @@ _ZN3ams4kern4arch5arm6412SvcHandler64Ev: #endif /* Restore registers. */ - ldp x30, x8, [sp, #(8 * 30)] - ldp x9, x10, [sp, #(8 * 32)] - ldr x11, [sp, #(8 * 34)] - ldr x18, [sp, #(8 * 18)] + ldp x30, x8, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + ldp x9, x10, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x11, [sp, #(EXCEPTION_CONTEXT_TPIDR)] + ldr x18, [sp, #(EXCEPTION_CONTEXT_X18)] msr sp_el0, x8 msr elr_el1, x9 msr spsr_el1, x10 @@ -221,7 +221,7 @@ _ZN3ams4kern4arch5arm6412SvcHandler64Ev: mov x17, xzr /* Return. */ - add sp, sp, #0x120 + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) eret /* ams::kern::arch::arm64::SvcHandler32() */ @@ -240,36 +240,36 @@ _ZN3ams4kern4arch5arm6412SvcHandler32Ev: mov w7, w7 /* Create a KExceptionContext for the exception. */ - sub sp, sp, #0x120 + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) /* Save system registers */ mrs x17, elr_el1 mrs x20, spsr_el1 mrs x19, tpidr_el0 - ldr x18, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CUR_THREAD)] - stp x17, x20, [sp, #(8 * 32)] - str x19, [sp, #(8 * 34)] + ldr x18, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CUR_THREAD)] + stp x17, x20, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + str x19, [sp, #(EXCEPTION_CONTEXT_TPIDR)] /* Save registers. */ - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x8, x9, [sp, #(8 * 8)] - stp x10, x11, [sp, #(8 * 10)] - stp x12, x13, [sp, #(8 * 12)] - stp x14, xzr, [sp, #(8 * 14)] + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, xzr, [sp, #(EXCEPTION_CONTEXT_X14_X15)] /* Check if the SVC index is out of range. */ mrs x16, esr_el1 and x16, x16, #0xFF - cmp x16, #0x80 + cmp x16, #(AMS_KERN_NUM_SUPERVISOR_CALLS) b.ge 3f /* Check the specific SVC permission bit for allowal. */ mov x20, sp add x20, x20, x16, lsr#3 - ldrb w20, [x20, #(0x120 + THREAD_STACK_PARAMETERS_SVC_PERMISSION)] + ldrb w20, [x20, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_SVC_PERMISSION)] and x17, x16, #0x7 lsr x17, x20, x17 tst x17, #1 @@ -277,11 +277,11 @@ _ZN3ams4kern4arch5arm6412SvcHandler32Ev: /* Check if our disable count allows us to call SVCs. */ mrs x15, tpidrro_el0 - ldrh w15, [x15, #0x100] + ldrh w15, [x15, #(THREAD_LOCAL_REGION_DISABLE_COUNT)] cbz w15, 1f /* It might not, so check the stack params to see if we must not allow the SVC. */ - ldrb w15, [sp, #(0x120 + THREAD_STACK_PARAMETERS_IS_PINNED)] + ldrb w15, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_IS_PINNED)] cbz w15, 3f 1: /* We can call the SVC. */ @@ -291,8 +291,8 @@ _ZN3ams4kern4arch5arm6412SvcHandler32Ev: /* Note that we're calling the SVC. */ mov w15, #1 - strb w15, [sp, #(0x120 + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] - strb w16, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CURRENT_SVC_ID)] + strb w15, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] + strb w16, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CURRENT_SVC_ID)] /* If we should, trace the svc entry. */ #if defined(MESOSPHERE_BUILD_FOR_TRACING) @@ -319,7 +319,7 @@ _ZN3ams4kern4arch5arm6412SvcHandler32Ev: 2: /* We completed the SVC, and we should handle DPC. */ /* Check the dpc flags. */ - ldrb w16, [sp, #(0x120 + THREAD_STACK_PARAMETERS_DPC_FLAGS)] + ldrb w16, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_DPC_FLAGS)] cbz w16, 4f /* We have DPC to do! */ @@ -339,45 +339,45 @@ _ZN3ams4kern4arch5arm6412SvcHandler32Ev: 3: /* Invalid SVC. */ /* Setup the context to call into HandleException. */ - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp xzr, xzr, [sp, #(8 * 16)] - stp xzr, xzr, [sp, #(8 * 18)] - stp xzr, xzr, [sp, #(8 * 20)] - stp xzr, xzr, [sp, #(8 * 22)] - stp xzr, xzr, [sp, #(8 * 24)] - stp xzr, xzr, [sp, #(8 * 26)] - stp xzr, xzr, [sp, #(8 * 28)] - stp xzr, xzr, [sp, #(8 * 30)] + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + stp xzr, xzr, [sp, #(EXCEPTION_CONTEXT_X30_SP)] /* Call ams::kern::arch::arm64::HandleException(ams::kern::arch::arm64::KExceptionContext *) */ mov x0, sp bl _ZN3ams4kern4arch5arm6415HandleExceptionEPNS2_17KExceptionContextE /* Restore registers. */ - ldp x17, x20, [sp, #(8 * 32)] - ldr x19, [sp, #(8 * 34)] + ldp x17, x20, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x19, [sp, #(EXCEPTION_CONTEXT_TPIDR)] msr elr_el1, x17 msr spsr_el1, x20 msr tpidr_el0, x19 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] /* Return. */ - add sp, sp, #0x120 + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) eret 4: /* Return from SVC. */ /* Clear our in-SVC note. */ - strb wzr, [sp, #(0x120 + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] + strb wzr, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_IS_CALLING_SVC)] /* If we should, trace the svc exit. */ #if defined(MESOSPHERE_BUILD_FOR_TRACING) @@ -396,16 +396,16 @@ _ZN3ams4kern4arch5arm6412SvcHandler32Ev: #endif /* Restore registers. */ - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, xzr, [sp, #(8 * 14)] - ldp x17, x20, [sp, #(8 * 32)] - ldr x19, [sp, #(8 * 34)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, xzr, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x17, x20, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x19, [sp, #(EXCEPTION_CONTEXT_TPIDR)] msr elr_el1, x17 msr spsr_el1, x20 msr tpidr_el0, x19 /* Return. */ - add sp, sp, #0x120 + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) eret diff --git a/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager_asm.s b/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager_asm.s index db198a294..d38c476fd 100644 --- a/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager_asm.s +++ b/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager_asm.s @@ -14,7 +14,6 @@ * along with this program. If not, see . */ - /* For some reason GAS doesn't know about it, even with .cpu cortex-a57 */ #define cpuactlr_el1 s3_1_c15_c2_0 #define cpuectlr_el1 s3_1_c15_c2_1 diff --git a/mesosphere/kernel/source/arch/arm64/init/start.s b/mesosphere/kernel/source/arch/arm64/init/start.s index 6300c50bd..6b513d0eb 100644 --- a/mesosphere/kernel/source/arch/arm64/init/start.s +++ b/mesosphere/kernel/source/arch/arm64/init/start.s @@ -13,6 +13,7 @@ * You should have received a copy of the GNU General Public License * along with this program. If not, see . */ +#include /* For some reason GAS doesn't know about it, even with .cpu cortex-a57 */ #define cpuactlr_el1 s3_1_c15_c2_0 @@ -157,13 +158,13 @@ othercore_el1: bl _ZN3ams4kern4init19DisableMmuAndCachesEv /* Setup system registers using values from our KInitArguments. */ - ldr x1, [x20, #0x00] + ldr x1, [x20, #(INIT_ARGUMENTS_TTBR0)] msr ttbr0_el1, x1 - ldr x1, [x20, #0x08] + ldr x1, [x20, #(INIT_ARGUMENTS_TTBR1)] msr ttbr1_el1, x1 - ldr x1, [x20, #0x10] + ldr x1, [x20, #(INIT_ARGUMENTS_TCR)] msr tcr_el1, x1 - ldr x1, [x20, #0x18] + ldr x1, [x20, #(INIT_ARGUMENTS_MAIR)] msr mair_el1, x1 /* Perform cpu-specific setup. */ @@ -179,9 +180,9 @@ othercore_el1: b othercore_cpu_specific_setup_end othercore_cpu_specific_setup_cortex_a57: othercore_cpu_specific_setup_cortex_a53: - ldr x1, [x20, #0x20] + ldr x1, [x20, #(INIT_ARGUMENTS_CPUACTLR)] msr cpuactlr_el1, x1 - ldr x1, [x20, #0x28] + ldr x1, [x20, #(INIT_ARGUMENTS_CPUECTLR)] msr cpuectlr_el1, x1 othercore_cpu_specific_setup_end: @@ -190,14 +191,14 @@ othercore_cpu_specific_setup_end: isb /* Set sctlr_el1 and ensure instruction consistency. */ - ldr x1, [x20, #0x30] + ldr x1, [x20, #(INIT_ARGUMENTS_SCTLR)] msr sctlr_el1, x1 dsb sy isb /* Jump to the virtual address equivalent to ams::kern::init::InvokeEntrypoint */ - ldr x1, [x20, #0x50] + ldr x1, [x20, #(INIT_ARGUMENTS_SETUP_FUNCTION)] adr x2, _ZN3ams4kern4init14StartOtherCoreEPKNS1_14KInitArgumentsE sub x1, x1, x2 adr x2, _ZN3ams4kern4init16InvokeEntrypointEPKNS1_14KInitArgumentsE @@ -218,7 +219,7 @@ _ZN3ams4kern4init16InvokeEntrypointEPKNS1_14KInitArgumentsE: isb /* Setup the stack pointer. */ - ldr x1, [x20, #0x38] + ldr x1, [x20, #(INIT_ARGUMENTS_SP)] mov sp, x1 /* Ensure that system debug registers are setup. */ @@ -228,12 +229,12 @@ _ZN3ams4kern4init16InvokeEntrypointEPKNS1_14KInitArgumentsE: bl _ZN3ams4kern4init26InitializeExceptionVectorsEv /* Setup the exception stack in cntv_cval_el0. */ - ldr x1, [x20, #0x58] + ldr x1, [x20, #(INIT_ARGUMENTS_EXCEPTION_STACK)] msr cntv_cval_el0, x1 /* Jump to the entrypoint. */ - ldr x1, [x20, #0x40] - ldr x0, [x20, #0x48] + ldr x1, [x20, #(INIT_ARGUMENTS_ENTRYPOINT)] + ldr x0, [x20, #(INIT_ARGUMENTS_ARGUMENT)] br x1 diff --git a/mesosphere/kernel/source/arch/arm64/kern_exception_handlers_asm.s b/mesosphere/kernel/source/arch/arm64/kern_exception_handlers_asm.s index 2c8f1936d..8b8da0f98 100644 --- a/mesosphere/kernel/source/arch/arm64/kern_exception_handlers_asm.s +++ b/mesosphere/kernel/source/arch/arm64/kern_exception_handlers_asm.s @@ -72,62 +72,66 @@ _ZN3ams4kern4arch5arm6422EL1IrqExceptionHandlerEv: .type _ZN3ams4kern4arch5arm6422EL0IrqExceptionHandlerEv, %function _ZN3ams4kern4arch5arm6422EL0IrqExceptionHandlerEv: /* Save registers that need saving. */ - sub sp, sp, #0x120 + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x8, x9, [sp, #(8 * 8)] - stp x10, x11, [sp, #(8 * 10)] - stp x12, x13, [sp, #(8 * 12)] - stp x14, x15, [sp, #(8 * 14)] - stp x16, x17, [sp, #(8 * 16)] - stp x18, x19, [sp, #(8 * 18)] - stp x20, x21, [sp, #(8 * 20)] - stp x22, x23, [sp, #(8 * 22)] - stp x24, x25, [sp, #(8 * 24)] - stp x26, x27, [sp, #(8 * 26)] - stp x28, x29, [sp, #(8 * 28)] + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + stp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] mrs x20, sp_el0 mrs x21, elr_el1 mrs x22, spsr_el1 mrs x23, tpidr_el0 mov w22, w22 - stp x30, x20, [sp, #(8 * 30)] - stp x21, x22, [sp, #(8 * 32)] - str x23, [sp, #(8 * 34)] + + stp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + stp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + str x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] /* Invoke KInterruptManager::HandleInterrupt(bool user_mode). */ - ldr x18, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CUR_THREAD)] + ldr x18, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CUR_THREAD)] mov x0, #1 bl _ZN3ams4kern4arch5arm6417KInterruptManager15HandleInterruptEb /* Restore state from the context. */ - ldp x30, x20, [sp, #(8 * 30)] - ldp x21, x22, [sp, #(8 * 32)] - ldr x23, [sp, #(8 * 34)] + ldp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + ldp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] + msr sp_el0, x20 msr elr_el1, x21 msr spsr_el1, x22 msr tpidr_el0, x23 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] - ldp x16, x17, [sp, #(8 * 16)] - ldp x18, x19, [sp, #(8 * 18)] - ldp x20, x21, [sp, #(8 * 20)] - ldp x22, x23, [sp, #(8 * 22)] - ldp x24, x25, [sp, #(8 * 24)] - ldp x26, x27, [sp, #(8 * 26)] - ldp x28, x29, [sp, #(8 * 28)] - add sp, sp, #0x120 + + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + ldp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + ldp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + ldp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + ldp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + ldp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) /* Return from the exception. */ eret @@ -171,60 +175,65 @@ _ZN3ams4kern4arch5arm6430EL0SynchronousExceptionHandlerEv: ldp x16, x17, [sp], 16 /* Create a KExceptionContext to pass to HandleException. */ - sub sp, sp, #0x120 - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x8, x9, [sp, #(8 * 8)] - stp x10, x11, [sp, #(8 * 10)] - stp x12, x13, [sp, #(8 * 12)] - stp x14, x15, [sp, #(8 * 14)] - stp x16, x17, [sp, #(8 * 16)] - stp x18, x19, [sp, #(8 * 18)] - stp x20, x21, [sp, #(8 * 20)] - stp x22, x23, [sp, #(8 * 22)] - stp x24, x25, [sp, #(8 * 24)] - stp x26, x27, [sp, #(8 * 26)] - stp x28, x29, [sp, #(8 * 28)] + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + stp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + mrs x20, sp_el0 mrs x21, elr_el1 mrs x22, spsr_el1 mrs x23, tpidr_el0 mov w22, w22 - stp x30, x20, [sp, #(8 * 30)] - stp x21, x22, [sp, #(8 * 32)] - str x23, [sp, #(8 * 34)] + + stp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + stp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + str x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] /* Call ams::kern::arch::arm64::HandleException(ams::kern::arch::arm64::KExceptionContext *) */ - ldr x18, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CUR_THREAD)] + ldr x18, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CUR_THREAD)] mov x0, sp bl _ZN3ams4kern4arch5arm6415HandleExceptionEPNS2_17KExceptionContextE /* Restore state from the context. */ - ldp x30, x20, [sp, #(8 * 30)] - ldp x21, x22, [sp, #(8 * 32)] - ldr x23, [sp, #(8 * 34)] + ldp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + ldp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] + msr sp_el0, x20 msr elr_el1, x21 msr spsr_el1, x22 msr tpidr_el0, x23 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] - ldp x16, x17, [sp, #(8 * 16)] - ldp x18, x19, [sp, #(8 * 18)] - ldp x20, x21, [sp, #(8 * 20)] - ldp x22, x23, [sp, #(8 * 22)] - ldp x24, x25, [sp, #(8 * 24)] - ldp x26, x27, [sp, #(8 * 26)] - ldp x28, x29, [sp, #(8 * 28)] - add sp, sp, #0x120 + + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + ldp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + ldp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + ldp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + ldp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + ldp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) /* Return from the exception. */ eret @@ -344,30 +353,33 @@ _ZN3ams4kern4arch5arm6430EL1SynchronousExceptionHandlerEv: ldr x1, [sp, #16] /* Create a KExceptionContext to pass to HandleException. */ - sub sp, sp, #0x120 - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x8, x9, [sp, #(8 * 8)] - stp x10, x11, [sp, #(8 * 10)] - stp x12, x13, [sp, #(8 * 12)] - stp x14, x15, [sp, #(8 * 14)] - stp x16, x17, [sp, #(8 * 16)] - stp x18, x19, [sp, #(8 * 18)] - stp x20, x21, [sp, #(8 * 20)] - stp x22, x23, [sp, #(8 * 22)] - stp x24, x25, [sp, #(8 * 24)] - stp x26, x27, [sp, #(8 * 26)] - stp x28, x29, [sp, #(8 * 28)] + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) + + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + stp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + mrs x20, sp_el0 mrs x21, elr_el1 mrs x22, spsr_el1 mrs x23, tpidr_el0 mov w22, w22 - stp x30, x20, [sp, #(8 * 30)] - stp x21, x22, [sp, #(8 * 32)] - str x23, [sp, #(8 * 34)] + + stp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + stp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + str x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] /* Call ams::kern::arch::arm64::HandleException(ams::kern::arch::arm64::KExceptionContext *) */ mov x0, sp @@ -421,53 +433,53 @@ _ZN3ams4kern4arch5arm6430EL1SynchronousExceptionHandlerEv: .type _ZN3ams4kern4arch5arm6425FpuAccessExceptionHandlerEv, %function _ZN3ams4kern4arch5arm6425FpuAccessExceptionHandlerEv: /* Save registers that need saving. */ - sub sp, sp, #0x120 + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x8, x9, [sp, #(8 * 8)] - stp x10, x11, [sp, #(8 * 10)] - stp x12, x13, [sp, #(8 * 12)] - stp x14, x15, [sp, #(8 * 14)] - stp x16, x17, [sp, #(8 * 16)] - stp x18, x19, [sp, #(8 * 18)] - stp x20, x21, [sp, #(8 * 20)] + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + stp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] mrs x19, sp_el0 mrs x20, elr_el1 mrs x21, spsr_el1 mov w21, w21 - stp x30, x19, [sp, #(8 * 30)] - stp x20, x21, [sp, #(8 * 32)] + stp x30, x19, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] /* Invoke the FPU context switch handler. */ - ldr x18, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CUR_THREAD)] + ldr x18, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CUR_THREAD)] bl _ZN3ams4kern4arch5arm6423FpuContextSwitchHandlerEv /* Restore registers that we saved. */ - ldp x30, x19, [sp, #(8 * 30)] - ldp x20, x21, [sp, #(8 * 32)] + ldp x30, x19, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] msr sp_el0, x19 msr elr_el1, x20 msr spsr_el1, x21 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] - ldp x16, x17, [sp, #(8 * 16)] - ldp x18, x19, [sp, #(8 * 18)] - ldp x20, x21, [sp, #(8 * 20)] + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + ldp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] - add sp, sp, #0x120 + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) /* Return from the exception. */ eret @@ -495,30 +507,32 @@ _ZN3ams4kern4arch5arm6421EL1SystemErrorHandlerEv: mrs x0, tpidr_el1 /* Create a KExceptionContext to pass to HandleException. */ - sub sp, sp, #0x120 - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x8, x9, [sp, #(8 * 8)] - stp x10, x11, [sp, #(8 * 10)] - stp x12, x13, [sp, #(8 * 12)] - stp x14, x15, [sp, #(8 * 14)] - stp x16, x17, [sp, #(8 * 16)] - stp x18, x19, [sp, #(8 * 18)] - stp x20, x21, [sp, #(8 * 20)] - stp x22, x23, [sp, #(8 * 22)] - stp x24, x25, [sp, #(8 * 24)] - stp x26, x27, [sp, #(8 * 26)] - stp x28, x29, [sp, #(8 * 28)] + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + stp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + mrs x20, sp_el0 mrs x21, elr_el1 mrs x22, spsr_el1 mrs x23, tpidr_el0 mov w22, w22 - stp x30, x20, [sp, #(8 * 30)] - stp x21, x22, [sp, #(8 * 32)] - str x23, [sp, #(8 * 34)] + + stp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + stp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + str x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] /* Invoke ams::kern::arch::arm64::HandleException(ams::kern::arch::arm64::KExceptionContext *). */ mov x0, sp @@ -533,60 +547,66 @@ _ZN3ams4kern4arch5arm6421EL1SystemErrorHandlerEv: .type _ZN3ams4kern4arch5arm6421EL0SystemErrorHandlerEv, %function _ZN3ams4kern4arch5arm6421EL0SystemErrorHandlerEv: /* Create a KExceptionContext to pass to HandleException. */ - sub sp, sp, #0x120 - stp x0, x1, [sp, #(8 * 0)] - stp x2, x3, [sp, #(8 * 2)] - stp x4, x5, [sp, #(8 * 4)] - stp x6, x7, [sp, #(8 * 6)] - stp x8, x9, [sp, #(8 * 8)] - stp x10, x11, [sp, #(8 * 10)] - stp x12, x13, [sp, #(8 * 12)] - stp x14, x15, [sp, #(8 * 14)] - stp x16, x17, [sp, #(8 * 16)] - stp x18, x19, [sp, #(8 * 18)] - stp x20, x21, [sp, #(8 * 20)] - stp x22, x23, [sp, #(8 * 22)] - stp x24, x25, [sp, #(8 * 24)] - stp x26, x27, [sp, #(8 * 26)] - stp x28, x29, [sp, #(8 * 28)] + sub sp, sp, #(EXCEPTION_CONTEXT_SIZE) + + stp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + stp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + stp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + stp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + stp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + stp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + stp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + stp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + stp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + stp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + stp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + stp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + stp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + stp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + stp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + mrs x20, sp_el0 mrs x21, elr_el1 mrs x22, spsr_el1 mrs x23, tpidr_el0 mov w22, w22 - stp x30, x20, [sp, #(8 * 30)] - stp x21, x22, [sp, #(8 * 32)] - str x23, [sp, #(8 * 34)] + + stp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + stp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + str x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] /* Invoke ams::kern::arch::arm64::HandleException(ams::kern::arch::arm64::KExceptionContext *). */ - ldr x18, [sp, #(0x120 + THREAD_STACK_PARAMETERS_CUR_THREAD)] + ldr x18, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_CUR_THREAD)] mov x0, sp bl _ZN3ams4kern4arch5arm6415HandleExceptionEPNS2_17KExceptionContextE /* Restore state from the context. */ - ldp x30, x20, [sp, #(8 * 30)] - ldp x21, x22, [sp, #(8 * 32)] - ldr x23, [sp, #(8 * 34)] + ldp x30, x20, [sp, #(EXCEPTION_CONTEXT_X30_SP)] + ldp x21, x22, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] + ldr x23, [sp, #(EXCEPTION_CONTEXT_TPIDR)] + msr sp_el0, x20 msr elr_el1, x21 msr spsr_el1, x22 msr tpidr_el0, x23 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] - ldp x16, x17, [sp, #(8 * 16)] - ldp x18, x19, [sp, #(8 * 18)] - ldp x20, x21, [sp, #(8 * 20)] - ldp x22, x23, [sp, #(8 * 22)] - ldp x24, x25, [sp, #(8 * 24)] - ldp x26, x27, [sp, #(8 * 26)] - ldp x28, x29, [sp, #(8 * 28)] - add sp, sp, #0x120 + + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + ldp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + ldp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + ldp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + ldp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + ldp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] + + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) /* Return from the exception. */ eret diff --git a/mesosphere/kernel/source/arch/arm64/kern_k_scheduler_asm.s b/mesosphere/kernel/source/arch/arm64/kern_k_scheduler_asm.s index 5ac0af70c..e19400650 100644 --- a/mesosphere/kernel/source/arch/arm64/kern_k_scheduler_asm.s +++ b/mesosphere/kernel/source/arch/arm64/kern_k_scheduler_asm.s @@ -15,88 +15,88 @@ */ #include -#define SAVE_THREAD_CONTEXT(ctx, tmp0, tmp1, done_label) \ - /* Save the callee save registers + SP and cpacr. */ \ - mov tmp0, sp; \ - mrs tmp1, cpacr_el1; \ - stp x19, x20, [ctx, #(8 * 0)]; \ - stp x21, x22, [ctx, #(8 * 2)]; \ - stp x23, x24, [ctx, #(8 * 4)]; \ - stp x25, x26, [ctx, #(8 * 6)]; \ - stp x27, x28, [ctx, #(8 * 8)]; \ - stp x29, x30, [ctx, #(8 * 10)]; \ - \ - stp tmp0, tmp1, [ctx, #0x60]; \ - \ - /* Check whether the FPU is enabled. */ \ - /* If it isn't, skip saving FPU state. */ \ - and tmp1, tmp1, #0x300000; \ - cbz tmp1, done_label; \ - \ - /* Save fpcr and fpsr. */ \ - mrs tmp0, fpcr; \ - mrs tmp1, fpsr; \ - stp tmp0, tmp1, [ctx, #0x70]; \ - \ - /* Save the FPU registers. */ \ - stp q0, q1, [ctx, #(16 * 0 + 0x80)]; \ - stp q2, q3, [ctx, #(16 * 2 + 0x80)]; \ - stp q4, q5, [ctx, #(16 * 4 + 0x80)]; \ - stp q6, q7, [ctx, #(16 * 6 + 0x80)]; \ - stp q8, q9, [ctx, #(16 * 8 + 0x80)]; \ - stp q10, q11, [ctx, #(16 * 10 + 0x80)]; \ - stp q12, q13, [ctx, #(16 * 12 + 0x80)]; \ - stp q14, q15, [ctx, #(16 * 14 + 0x80)]; \ - stp q16, q17, [ctx, #(16 * 16 + 0x80)]; \ - stp q18, q19, [ctx, #(16 * 18 + 0x80)]; \ - stp q20, q21, [ctx, #(16 * 20 + 0x80)]; \ - stp q22, q23, [ctx, #(16 * 22 + 0x80)]; \ - stp q24, q25, [ctx, #(16 * 24 + 0x80)]; \ - stp q26, q27, [ctx, #(16 * 26 + 0x80)]; \ - stp q28, q29, [ctx, #(16 * 28 + 0x80)]; \ - stp q30, q31, [ctx, #(16 * 30 + 0x80)]; +#define SAVE_THREAD_CONTEXT(ctx, tmp0, tmp1, done_label) \ + /* Save the callee save registers + SP and cpacr. */ \ + mov tmp0, sp; \ + mrs tmp1, cpacr_el1; \ + stp x19, x20, [ctx, #(THREAD_CONTEXT_X19_X20)]; \ + stp x21, x22, [ctx, #(THREAD_CONTEXT_X21_X22)]; \ + stp x23, x24, [ctx, #(THREAD_CONTEXT_X23_X24)]; \ + stp x25, x26, [ctx, #(THREAD_CONTEXT_X25_X26)]; \ + stp x27, x28, [ctx, #(THREAD_CONTEXT_X27_X28)]; \ + stp x29, x30, [ctx, #(THREAD_CONTEXT_X29_X30)]; \ + \ + stp tmp0, tmp1, [ctx, #(THREAD_CONTEXT_SP_CPACR)]; \ + \ + /* Check whether the FPU is enabled. */ \ + /* If it isn't, skip saving FPU state. */ \ + and tmp1, tmp1, #0x300000; \ + cbz tmp1, done_label; \ + \ + /* Save fpcr and fpsr. */ \ + mrs tmp0, fpcr; \ + mrs tmp1, fpsr; \ + stp tmp0, tmp1, [ctx, #(THREAD_CONTEXT_FPCR_FPSR)]; \ + \ + /* Save the FPU registers. */ \ + stp q0, q1, [ctx, #(16 * 0 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q2, q3, [ctx, #(16 * 2 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q4, q5, [ctx, #(16 * 4 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q6, q7, [ctx, #(16 * 6 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q8, q9, [ctx, #(16 * 8 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q10, q11, [ctx, #(16 * 10 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q12, q13, [ctx, #(16 * 12 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q14, q15, [ctx, #(16 * 14 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q16, q17, [ctx, #(16 * 16 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q18, q19, [ctx, #(16 * 18 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q20, q21, [ctx, #(16 * 20 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q22, q23, [ctx, #(16 * 22 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q24, q25, [ctx, #(16 * 24 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q26, q27, [ctx, #(16 * 26 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q28, q29, [ctx, #(16 * 28 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + stp q30, q31, [ctx, #(16 * 30 + THREAD_CONTEXT_FPU_REGISTERS)]; -#define RESTORE_THREAD_CONTEXT(ctx, tmp0, tmp1, done_label) \ - /* Restore the callee save registers + SP and cpacr. */ \ - ldp tmp0, tmp1, [ctx, #0x60]; \ - mov sp, tmp0; \ - ldp x19, x20, [ctx, #(8 * 0)]; \ - ldp x21, x22, [ctx, #(8 * 2)]; \ - ldp x23, x24, [ctx, #(8 * 4)]; \ - ldp x25, x26, [ctx, #(8 * 6)]; \ - ldp x27, x28, [ctx, #(8 * 8)]; \ - ldp x29, x30, [ctx, #(8 * 10)]; \ - \ - msr cpacr_el1, tmp1; \ - isb; \ - \ - /* Check whether the FPU is enabled. */ \ - /* If it isn't, skip saving FPU state. */ \ - and tmp1, tmp1, #0x300000; \ - cbz tmp1, done_label; \ - \ - /* Save fpcr and fpsr. */ \ - ldp tmp0, tmp1, [ctx, #0x70]; \ - msr fpcr, tmp0; \ - msr fpsr, tmp1; \ - \ - /* Save the FPU registers. */ \ - ldp q0, q1, [ctx, #(16 * 0 + 0x80)]; \ - ldp q2, q3, [ctx, #(16 * 2 + 0x80)]; \ - ldp q4, q5, [ctx, #(16 * 4 + 0x80)]; \ - ldp q6, q7, [ctx, #(16 * 6 + 0x80)]; \ - ldp q8, q9, [ctx, #(16 * 8 + 0x80)]; \ - ldp q10, q11, [ctx, #(16 * 10 + 0x80)]; \ - ldp q12, q13, [ctx, #(16 * 12 + 0x80)]; \ - ldp q14, q15, [ctx, #(16 * 14 + 0x80)]; \ - ldp q16, q17, [ctx, #(16 * 16 + 0x80)]; \ - ldp q18, q19, [ctx, #(16 * 18 + 0x80)]; \ - ldp q20, q21, [ctx, #(16 * 20 + 0x80)]; \ - ldp q22, q23, [ctx, #(16 * 22 + 0x80)]; \ - ldp q24, q25, [ctx, #(16 * 24 + 0x80)]; \ - ldp q26, q27, [ctx, #(16 * 26 + 0x80)]; \ - ldp q28, q29, [ctx, #(16 * 28 + 0x80)]; \ - ldp q30, q31, [ctx, #(16 * 30 + 0x80)]; +#define RESTORE_THREAD_CONTEXT(ctx, tmp0, tmp1, done_label) \ + /* Restore the callee save registers + SP and cpacr. */ \ + ldp tmp0, tmp1, [ctx, #(THREAD_CONTEXT_SP_CPACR)]; \ + mov sp, tmp0; \ + ldp x19, x20, [ctx, #(THREAD_CONTEXT_X19_X20)]; \ + ldp x21, x22, [ctx, #(THREAD_CONTEXT_X21_X22)]; \ + ldp x23, x24, [ctx, #(THREAD_CONTEXT_X23_X24)]; \ + ldp x25, x26, [ctx, #(THREAD_CONTEXT_X25_X26)]; \ + ldp x27, x28, [ctx, #(THREAD_CONTEXT_X27_X28)]; \ + ldp x29, x30, [ctx, #(THREAD_CONTEXT_X29_X30)]; \ + \ + msr cpacr_el1, tmp1; \ + isb; \ + \ + /* Check whether the FPU is enabled. */ \ + /* If it isn't, skip saving FPU state. */ \ + and tmp1, tmp1, #0x300000; \ + cbz tmp1, done_label; \ + \ + /* Save fpcr and fpsr. */ \ + ldp tmp0, tmp1, [ctx, #(THREAD_CONTEXT_FPCR_FPSR)]; \ + msr fpcr, tmp0; \ + msr fpsr, tmp1; \ + \ + /* Save the FPU registers. */ \ + ldp q0, q1, [ctx, #(16 * 0 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q2, q3, [ctx, #(16 * 2 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q4, q5, [ctx, #(16 * 4 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q6, q7, [ctx, #(16 * 6 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q8, q9, [ctx, #(16 * 8 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q10, q11, [ctx, #(16 * 10 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q12, q13, [ctx, #(16 * 12 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q14, q15, [ctx, #(16 * 14 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q16, q17, [ctx, #(16 * 16 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q18, q19, [ctx, #(16 * 18 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q20, q21, [ctx, #(16 * 20 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q22, q23, [ctx, #(16 * 22 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q24, q25, [ctx, #(16 * 24 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q26, q27, [ctx, #(16 * 26 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q28, q29, [ctx, #(16 * 28 + THREAD_CONTEXT_FPU_REGISTERS)]; \ + ldp q30, q31, [ctx, #(16 * 30 + THREAD_CONTEXT_FPU_REGISTERS)]; /* ams::kern::KScheduler::ScheduleImpl() */ @@ -109,11 +109,11 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: /* Right now, x0 contains (this). We want x1 to point to the scheduling state, */ - /* Current KScheduler layout has state at +0x0. */ + /* KScheduler layout has state at +0x0, this is guaranteed statically by assembly offsets. */ mov x1, x0 /* First thing we want to do is check whether the interrupt task thread is runnable. */ - ldrb w3, [x1, #1] + ldrb w3, [x1, #(KSCHEDULER_INTERRUPT_TASK_THREAD_RUNNABLE)] cbz w3, 0f /* If it is, we want to call KScheduler::InterruptTaskThreadToRunnable() to change its state to runnable. */ @@ -124,7 +124,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: ldp x0, x1, [sp], 16 /* Clear the interrupt task thread as runnable. */ - strb wzr, [x1, #1] + strb wzr, [x1, #(KSCHEDULER_INTERRUPT_TASK_THREAD_RUNNABLE)] 0: /* Interrupt task thread runnable checked. */ /* Now we want to check if there's any scheduling to do. */ @@ -135,7 +135,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: dmb ish /* Check if the highest priority thread is the same as the current thread. */ - ldr x7, [x1, 16] + ldr x7, [x1, #(KSCHEDULER_HIGHEST_PRIORITY_THREAD)] ldr x2, [x18] cmp x7, x2 b.ne 1f @@ -158,7 +158,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: 2: /* We're done saving this thread's context, so we need to unlock it. */ /* We can just do an atomic write to the relevant KThreadContext member. */ - add x2, x2, #0x280 + add x2, x2, #(THREAD_CONTEXT_LOCKED) stlrb wzr, [x2] 3: /* The current thread's context has been entirely taken care of. */ @@ -169,7 +169,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: mov x21, x7 /* highest priority thread */ /* Set our stack to the idle thread stack. */ - ldr x3, [x20, #0x18] + ldr x3, [x20, #(KSCHEDULER_IDLE_THREAD_STACK)] mov sp, x3 b 5f @@ -183,7 +183,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: dmb ish /* Refresh the highest priority thread. */ - ldr x21, [x20, 16] + ldr x21, [x20, #(KSCHEDULER_HIGHEST_PRIORITY_THREAD)] 5: /* We're starting to try to do the context switch. */ /* Check if the highest priority thread if null. */ @@ -197,7 +197,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: mov x22, x0 /* Prepare to try to acquire the context lock. */ - add x1, x22, #0x280 + add x1, x22, #(THREAD_CONTEXT_LOCKED) mov w2, #1 6: /* We want to try to lock the highest priority thread's context. */ @@ -214,7 +214,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: 7: /* The highest priority thread's context is already locked. */ /* Check if we need scheduling. If we don't, we can retry directly. */ - ldarb w3, [x20] + ldarb w3, [x20] // ldarb w3, [x20, #(KSCHEDULER_NEEDS_SCHEDULING)] cbz w3, 6b /* If we do, another core is interfering, and we must start from the top. */ @@ -229,7 +229,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: bl _ZN3ams4kern10KScheduler12SwitchThreadEPNS0_7KThreadE /* Check if we need scheduling. If we don't, then we can't complete the switch and should retry. */ - ldarb w1, [x20] + ldarb w1, [x20] // ldarb w1, [x20, #(KSCHEDULER_NEEDS_SCHEDULING)] cbnz w1, 10f /* Restore the thread context. */ @@ -241,7 +241,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: 10: /* Our switch failed. */ /* We should unlock the thread context, and then retry. */ - add x1, x22, #0x280 + add x1, x22, #(THREAD_CONTEXT_LOCKED) stlrb wzr, [x1] b 4b @@ -255,7 +255,7 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: 12: /* We've switched to the idle thread, so we want to loop until we schedule a non-idle thread. */ /* Check if we need scheduling. */ - ldarb w3, [x20] + ldarb w3, [x20] // ldarb w3, [x20, #(KSCHEDULER_NEEDS_SCHEDULING)] cbnz w3, 13f /* If we don't, wait for an interrupt and check again. */ @@ -268,14 +268,14 @@ _ZN3ams4kern10KScheduler12ScheduleImplEv: 13: /* We need scheduling again! */ /* Check whether the interrupt task thread needs to be set runnable. */ - ldrb w3, [x20, #1] + ldrb w3, [x20, #(KSCHEDULER_INTERRUPT_TASK_THREAD_RUNNABLE)] cbz w3, 4b /* It does, so do so. We're using the idle thread stack so no register state preserve needed. */ bl _ZN3ams4kern10KScheduler29InterruptTaskThreadToRunnableEv /* Clear the interrupt task thread as runnable. */ - strb wzr, [x20, #1] + strb wzr, [x20, #(KSCHEDULER_INTERRUPT_TASK_THREAD_RUNNABLE)] /* Retry the scheduling loop. */ b 4b diff --git a/mesosphere/kernel/source/arch/arm64/kern_k_thread_context_asm.s b/mesosphere/kernel/source/arch/arm64/kern_k_thread_context_asm.s index 0e914eb60..d7af9df78 100644 --- a/mesosphere/kernel/source/arch/arm64/kern_k_thread_context_asm.s +++ b/mesosphere/kernel/source/arch/arm64/kern_k_thread_context_asm.s @@ -27,39 +27,39 @@ _ZN3ams4kern4arch5arm6421UserModeThreadStarterEv: /* | KExceptionContext (size 0x120) | KThread::StackParameters (size 0x30) | */ /* Clear the disable count for this thread's stack parameters. */ - strh wzr, [sp, #(0x120 + THREAD_STACK_PARAMETERS_DISABLE_COUNT)] + strh wzr, [sp, #(EXCEPTION_CONTEXT_SIZE + THREAD_STACK_PARAMETERS_DISABLE_COUNT)] /* Call ams::kern::arch::arm64::OnThreadStart() */ bl _ZN3ams4kern4arch5arm6413OnThreadStartEv /* Restore thread state from the KExceptionContext on stack */ - ldp x30, x19, [sp, #(8 * 30)] /* x30 = lr, x19 = sp */ - ldp x20, x21, [sp, #(8 * 30 + 16)] /* x20 = pc, x21 = psr */ - ldr x22, [sp, #(8 * 30 + 32)] /* x22 = tpidr */ + ldp x30, x19, [sp, #(EXCEPTION_CONTEXT_X30_SP)] /* x30 = lr, x19 = sp */ + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_PC_PSR)] /* x20 = pc, x21 = psr */ + ldr x22, [sp, #(EXCEPTION_CONTEXT_TPIDR)] /* x22 = tpidr */ msr sp_el0, x19 msr elr_el1, x20 msr spsr_el1, x21 msr tpidr_el0, x22 - ldp x0, x1, [sp, #(8 * 0)] - ldp x2, x3, [sp, #(8 * 2)] - ldp x4, x5, [sp, #(8 * 4)] - ldp x6, x7, [sp, #(8 * 6)] - ldp x8, x9, [sp, #(8 * 8)] - ldp x10, x11, [sp, #(8 * 10)] - ldp x12, x13, [sp, #(8 * 12)] - ldp x14, x15, [sp, #(8 * 14)] - ldp x16, x17, [sp, #(8 * 16)] - ldp x18, x19, [sp, #(8 * 18)] - ldp x20, x21, [sp, #(8 * 20)] - ldp x22, x23, [sp, #(8 * 22)] - ldp x24, x25, [sp, #(8 * 24)] - ldp x26, x27, [sp, #(8 * 26)] - ldp x28, x29, [sp, #(8 * 28)] + ldp x0, x1, [sp, #(EXCEPTION_CONTEXT_X0_X1)] + ldp x2, x3, [sp, #(EXCEPTION_CONTEXT_X2_X3)] + ldp x4, x5, [sp, #(EXCEPTION_CONTEXT_X4_X5)] + ldp x6, x7, [sp, #(EXCEPTION_CONTEXT_X6_X7)] + ldp x8, x9, [sp, #(EXCEPTION_CONTEXT_X8_X9)] + ldp x10, x11, [sp, #(EXCEPTION_CONTEXT_X10_X11)] + ldp x12, x13, [sp, #(EXCEPTION_CONTEXT_X12_X13)] + ldp x14, x15, [sp, #(EXCEPTION_CONTEXT_X14_X15)] + ldp x16, x17, [sp, #(EXCEPTION_CONTEXT_X16_X17)] + ldp x18, x19, [sp, #(EXCEPTION_CONTEXT_X18_X19)] + ldp x20, x21, [sp, #(EXCEPTION_CONTEXT_X20_X21)] + ldp x22, x23, [sp, #(EXCEPTION_CONTEXT_X22_X23)] + ldp x24, x25, [sp, #(EXCEPTION_CONTEXT_X24_X25)] + ldp x26, x27, [sp, #(EXCEPTION_CONTEXT_X26_X27)] + ldp x28, x29, [sp, #(EXCEPTION_CONTEXT_X28_X29)] /* Increment stack pointer above the KExceptionContext */ - add sp, sp, #0x120 + add sp, sp, #(EXCEPTION_CONTEXT_SIZE) /* Return to EL0 */ eret @@ -95,28 +95,28 @@ _ZN3ams4kern4arch5arm6427SupervisorModeThreadStarterEv: .type _ZN3ams4kern4arch5arm6414KThreadContext21RestoreFpuRegisters64ERKS3_, %function _ZN3ams4kern4arch5arm6414KThreadContext21RestoreFpuRegisters64ERKS3_: /* Load and restore FPCR and FPSR from the context. */ - ldr x1, [x0, #0x70] + ldr x1, [x0, #(THREAD_CONTEXT_FPCR)] msr fpcr, x1 - ldr x1, [x0, #0x78] + ldr x1, [x0, #(THREAD_CONTEXT_FPSR)] msr fpsr, x1 /* Restore the FPU registers. */ - ldp q0, q1, [x0, #(16 * 0 + 0x80)] - ldp q2, q3, [x0, #(16 * 2 + 0x80)] - ldp q4, q5, [x0, #(16 * 4 + 0x80)] - ldp q6, q7, [x0, #(16 * 6 + 0x80)] - ldp q8, q9, [x0, #(16 * 8 + 0x80)] - ldp q10, q11, [x0, #(16 * 10 + 0x80)] - ldp q12, q13, [x0, #(16 * 12 + 0x80)] - ldp q14, q15, [x0, #(16 * 14 + 0x80)] - ldp q16, q17, [x0, #(16 * 16 + 0x80)] - ldp q18, q19, [x0, #(16 * 18 + 0x80)] - ldp q20, q21, [x0, #(16 * 20 + 0x80)] - ldp q22, q23, [x0, #(16 * 22 + 0x80)] - ldp q24, q25, [x0, #(16 * 24 + 0x80)] - ldp q26, q27, [x0, #(16 * 26 + 0x80)] - ldp q28, q29, [x0, #(16 * 28 + 0x80)] - ldp q30, q31, [x0, #(16 * 30 + 0x80)] + ldp q0, q1, [x0, #(16 * 0 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q2, q3, [x0, #(16 * 2 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q4, q5, [x0, #(16 * 4 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q6, q7, [x0, #(16 * 6 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q8, q9, [x0, #(16 * 8 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q10, q11, [x0, #(16 * 10 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q12, q13, [x0, #(16 * 12 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q14, q15, [x0, #(16 * 14 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q16, q17, [x0, #(16 * 16 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q18, q19, [x0, #(16 * 18 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q20, q21, [x0, #(16 * 20 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q22, q23, [x0, #(16 * 22 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q24, q25, [x0, #(16 * 24 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q26, q27, [x0, #(16 * 26 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q28, q29, [x0, #(16 * 28 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q30, q31, [x0, #(16 * 30 + THREAD_CONTEXT_FPU_REGISTERS)] ret @@ -126,19 +126,19 @@ _ZN3ams4kern4arch5arm6414KThreadContext21RestoreFpuRegisters64ERKS3_: .type _ZN3ams4kern4arch5arm6414KThreadContext21RestoreFpuRegisters32ERKS3_, %function _ZN3ams4kern4arch5arm6414KThreadContext21RestoreFpuRegisters32ERKS3_: /* Load and restore FPCR and FPSR from the context. */ - ldr x1, [x0, #0x70] + ldr x1, [x0, #(THREAD_CONTEXT_FPCR)] msr fpcr, x1 - ldr x1, [x0, #0x78] + ldr x1, [x0, #(THREAD_CONTEXT_FPSR)] msr fpsr, x1 /* Restore the FPU registers. */ - ldp q0, q1, [x0, #(16 * 0 + 0x80)] - ldp q2, q3, [x0, #(16 * 2 + 0x80)] - ldp q4, q5, [x0, #(16 * 4 + 0x80)] - ldp q6, q7, [x0, #(16 * 6 + 0x80)] - ldp q8, q9, [x0, #(16 * 8 + 0x80)] - ldp q10, q11, [x0, #(16 * 10 + 0x80)] - ldp q12, q13, [x0, #(16 * 12 + 0x80)] - ldp q14, q15, [x0, #(16 * 14 + 0x80)] + ldp q0, q1, [x0, #(16 * 0 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q2, q3, [x0, #(16 * 2 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q4, q5, [x0, #(16 * 4 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q6, q7, [x0, #(16 * 6 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q8, q9, [x0, #(16 * 8 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q10, q11, [x0, #(16 * 10 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q12, q13, [x0, #(16 * 12 + THREAD_CONTEXT_FPU_REGISTERS)] + ldp q14, q15, [x0, #(16 * 14 + THREAD_CONTEXT_FPU_REGISTERS)] ret