mirror of
https://github.com/Atmosphere-NX/Atmosphere
synced 2024-12-22 20:31:14 +00:00
kern: push code through call to kernelldr
This commit is contained in:
parent
bce7133128
commit
24d41ce55e
2 changed files with 311 additions and 14 deletions
|
@ -15,10 +15,10 @@ def main(argc, argv):
|
||||||
kernel_ldr = f.read()
|
kernel_ldr = f.read()
|
||||||
with open('kernel/kernel.bin', 'rb') as f:
|
with open('kernel/kernel.bin', 'rb') as f:
|
||||||
kernel = f.read()
|
kernel = f.read()
|
||||||
kernel_metadata_offset = up('<I', kernel[4:8])[0]
|
kernel_metadata_offset = 4
|
||||||
assert (kernel_metadata_offset <= len(kernel) - 0x40)
|
assert (kernel_metadata_offset <= len(kernel) - 0x40)
|
||||||
assert (kernel[kernel_metadata_offset:kernel_metadata_offset + 4] == b'MSS0')
|
assert (kernel[kernel_metadata_offset:kernel_metadata_offset + 4] == b'MSS0')
|
||||||
kernel_end = up('<I', kernel[kernel_metadata_offset + 0x30:kernel_metadata_offset + 0x34])[0]
|
kernel_end = up('<I', kernel[kernel_metadata_offset + 0x34:kernel_metadata_offset + 0x38])[0]
|
||||||
assert (kernel_end >= len(kernel))
|
assert (kernel_end >= len(kernel))
|
||||||
|
|
||||||
embedded_ini = b''
|
embedded_ini = b''
|
||||||
|
@ -29,9 +29,9 @@ def main(argc, argv):
|
||||||
kernel_ldr_end = kernel_ldr_offset + len(kernel_ldr)
|
kernel_ldr_end = kernel_ldr_offset + len(kernel_ldr)
|
||||||
|
|
||||||
with open('mesosphere.bin', 'wb') as f:
|
with open('mesosphere.bin', 'wb') as f:
|
||||||
f.write(kernel[:kernel_metadata_offset + 8])
|
f.write(kernel[:kernel_metadata_offset + 4])
|
||||||
f.write(pk('<II', embedded_ini_offset, kernel_ldr_offset))
|
f.write(pk('<QQ', embedded_ini_offset, kernel_ldr_offset))
|
||||||
f.write(kernel[kernel_metadata_offset + 0x10:])
|
f.write(kernel[kernel_metadata_offset + 0x14:])
|
||||||
f.seek(embedded_ini_offset)
|
f.seek(embedded_ini_offset)
|
||||||
f.write(embedded_ini)
|
f.write(embedded_ini)
|
||||||
f.seek(embedded_ini_end)
|
f.seek(embedded_ini_end)
|
||||||
|
|
|
@ -18,18 +18,30 @@
|
||||||
#define cpuactlr_el1 s3_1_c15_c2_0
|
#define cpuactlr_el1 s3_1_c15_c2_0
|
||||||
#define cpuectlr_el1 s3_1_c15_c2_1
|
#define cpuectlr_el1 s3_1_c15_c2_1
|
||||||
|
|
||||||
|
#define LOAD_IMMEDIATE_32(reg, val) \
|
||||||
|
mov reg, #(((val) >> 0x00) & 0xFFFF); \
|
||||||
|
movk reg, #(((val) >> 0x10) & 0xFFFF), lsl#16
|
||||||
|
|
||||||
|
#define LOAD_IMMEDIATE_64(reg, val) \
|
||||||
|
mov reg, #(((val) >> 0x00) & 0xFFFF); \
|
||||||
|
movk reg, #(((val) >> 0x10) & 0xFFFF), lsl#16; \
|
||||||
|
movk reg, #(((val) >> 0x20) & 0xFFFF), lsl#32; \
|
||||||
|
movk reg, #(((val) >> 0x30) & 0xFFFF), lsl#48
|
||||||
|
|
||||||
|
#define LOAD_FROM_LABEL(reg, label) \
|
||||||
|
adr reg, label; \
|
||||||
|
ldr reg, [reg]
|
||||||
|
|
||||||
.section .crt0.text.start, "ax", %progbits
|
.section .crt0.text.start, "ax", %progbits
|
||||||
.global _start
|
.global _start
|
||||||
_start:
|
_start:
|
||||||
/* TODO */
|
b _ZN3ams4kern4init10StartCore0Emm
|
||||||
b _start
|
|
||||||
.word (__metadata_begin - _start)
|
|
||||||
|
|
||||||
__metadata_begin:
|
__metadata_begin:
|
||||||
.ascii "MSS0" /* Magic */
|
.ascii "MSS0" /* Magic */
|
||||||
.word 0 /* KInitArguments */
|
__metadata_ini_offset:
|
||||||
.word 0 /* INI1 base address. */
|
.quad 0 /* INI1 base address. */
|
||||||
.word 0 /* Kernel Loader base address. */
|
__metadata_kernelldr_offset:
|
||||||
|
.quad 0 /* Kernel Loader base address. */
|
||||||
__metadata_kernel_layout:
|
__metadata_kernel_layout:
|
||||||
.word _start - _start /* rx_offset */
|
.word _start - _start /* rx_offset */
|
||||||
.word __rodata_start - _start /* rx_end_offset */
|
.word __rodata_start - _start /* rx_end_offset */
|
||||||
|
@ -43,6 +55,291 @@ __metadata_kernel_layout:
|
||||||
.word _DYNAMIC - _start /* dynamic_offset */
|
.word _DYNAMIC - _start /* dynamic_offset */
|
||||||
.word __init_array_start - _start /* init_array_offset */
|
.word __init_array_start - _start /* init_array_offset */
|
||||||
.word __init_array_end - _start /* init_array_end_offset */
|
.word __init_array_end - _start /* init_array_end_offset */
|
||||||
.if (. - __metadata_begin) != 0x40
|
.if (. - __metadata_begin) != 0x44
|
||||||
.error "Incorrect Mesosphere Metadata"
|
.error "Incorrect Mesosphere Metadata"
|
||||||
.endif
|
.endif
|
||||||
|
|
||||||
|
/* ams::kern::init::StartCore0(uintptr_t, uintptr_t) */
|
||||||
|
.section .crt0.text._ZN3ams4kern4init10StartCore0Emm, "ax", %progbits
|
||||||
|
.global _ZN3ams4kern4init10StartCore0Emm
|
||||||
|
.type _ZN3ams4kern4init10StartCore0Emm, %function
|
||||||
|
_ZN3ams4kern4init10StartCore0Emm:
|
||||||
|
/* Mask all interrupts. */
|
||||||
|
msr daifset, #0xF
|
||||||
|
|
||||||
|
/* Save arguments for later use. */
|
||||||
|
mov x19, x0
|
||||||
|
mov x20, x1
|
||||||
|
|
||||||
|
/* Check our current EL. We want to be executing out of EL1. */
|
||||||
|
/* If we're in EL2, we'll need to deprivilege ourselves. */
|
||||||
|
mrs x1, currentel
|
||||||
|
cmp x1, #0x4
|
||||||
|
b.eq core0_el1
|
||||||
|
cmp x1, #0x8
|
||||||
|
b.eq core0_el2
|
||||||
|
core0_el3:
|
||||||
|
b core0_el3
|
||||||
|
core0_el2:
|
||||||
|
bl _ZN3ams4kern4init16JumpFromEL2ToEL1Ev
|
||||||
|
core0_el1:
|
||||||
|
bl _ZN3ams4kern4init19DisableMmuAndCachesEv
|
||||||
|
|
||||||
|
/* We want to invoke kernel loader. */
|
||||||
|
adr x0, _start
|
||||||
|
adr x1, __metadata_kernel_layout
|
||||||
|
LOAD_FROM_LABEL(x2, __metadata_ini_offset)
|
||||||
|
add x2, x0, x2
|
||||||
|
LOAD_FROM_LABEL(x3, __metadata_kernelldr_offset)
|
||||||
|
add x3, x0, x3
|
||||||
|
blr x3
|
||||||
|
|
||||||
|
/* TODO: Finish post-kernelldr init code. */
|
||||||
|
1:
|
||||||
|
b 1b
|
||||||
|
|
||||||
|
|
||||||
|
/* ams::kern::init::JumpFromEL2ToEL1() */
|
||||||
|
.section .crt0.text._ZN3ams4kern4init16JumpFromEL2ToEL1Ev, "ax", %progbits
|
||||||
|
.global _ZN3ams4kern4init16JumpFromEL2ToEL1Ev
|
||||||
|
.type _ZN3ams4kern4init16JumpFromEL2ToEL1Ev, %function
|
||||||
|
_ZN3ams4kern4init16JumpFromEL2ToEL1Ev:
|
||||||
|
/* We're going to want to ERET to our caller. */
|
||||||
|
msr elr_el2, x30
|
||||||
|
|
||||||
|
/* Ensure that the cache is coherent. */
|
||||||
|
bl _ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv
|
||||||
|
dsb sy
|
||||||
|
|
||||||
|
bl _ZN3ams4kern5arm643cpu38FlushEntireDataCacheSharedWithoutStackEv
|
||||||
|
dsb sy
|
||||||
|
|
||||||
|
bl _ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv
|
||||||
|
dsb sy
|
||||||
|
|
||||||
|
/* Invalidate the entire TLB, and ensure instruction consistency. */
|
||||||
|
tlbi vmalle1is
|
||||||
|
dsb sy
|
||||||
|
isb
|
||||||
|
|
||||||
|
/* Setup system registers for deprivileging. */
|
||||||
|
/* ACTLR_EL2: */
|
||||||
|
/* - CPUACTLR access control = 1 */
|
||||||
|
/* - CPUECTLR access control = 1 */
|
||||||
|
/* - L2CTLR access control = 1 */
|
||||||
|
/* - L2ECTLR access control = 1 */
|
||||||
|
/* - L2ACTLR access control = 1 */
|
||||||
|
mov x0, #0x73
|
||||||
|
msr actlr_el2, x0
|
||||||
|
|
||||||
|
/* HCR_EL2: */
|
||||||
|
/* - RW = 1 (el1 is aarch64) */
|
||||||
|
mov x0, #0x80000000
|
||||||
|
msr hcr_el2, x0
|
||||||
|
|
||||||
|
/* SCTLR_EL1: */
|
||||||
|
/* - EOS = 1 */
|
||||||
|
/* - EIS = 1 */
|
||||||
|
/* - SPAN = 1 */
|
||||||
|
LOAD_IMMEDIATE_32(x0, 0x00C00800)
|
||||||
|
msr sctlr_el1, x0
|
||||||
|
|
||||||
|
/* DACR32_EL2: */
|
||||||
|
/* - Manager access for all D<n> */
|
||||||
|
mov x0, #0xFFFFFFFF
|
||||||
|
msr dacr32_el2, x0
|
||||||
|
|
||||||
|
/* SPSR_EL2: */
|
||||||
|
/* - EL1h */
|
||||||
|
/* - IRQ masked */
|
||||||
|
/* - FIQ masked */
|
||||||
|
mov x0, #0xC5
|
||||||
|
msr spsr_el2, x0
|
||||||
|
|
||||||
|
eret
|
||||||
|
|
||||||
|
/* ams::kern::init::DisableMmuAndCaches() */
|
||||||
|
.section .crt0.text._ZN3ams4kern4init19DisableMmuAndCachesEv, "ax", %progbits
|
||||||
|
.global _ZN3ams4kern4init19DisableMmuAndCachesEv
|
||||||
|
.type _ZN3ams4kern4init19DisableMmuAndCachesEv, %function
|
||||||
|
_ZN3ams4kern4init19DisableMmuAndCachesEv:
|
||||||
|
/* The stack isn't set up, so we'll need to trash a register. */
|
||||||
|
mov x22, x30
|
||||||
|
|
||||||
|
/* Ensure that the cache is coherent. */
|
||||||
|
bl _ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv
|
||||||
|
dsb sy
|
||||||
|
|
||||||
|
bl _ZN3ams4kern5arm643cpu38FlushEntireDataCacheSharedWithoutStackEv
|
||||||
|
dsb sy
|
||||||
|
|
||||||
|
bl _ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv
|
||||||
|
dsb sy
|
||||||
|
|
||||||
|
/* Invalidate the entire TLB, and ensure instruction consistency. */
|
||||||
|
tlbi vmalle1is
|
||||||
|
dsb sy
|
||||||
|
isb
|
||||||
|
|
||||||
|
/* Invalidate the instruction cache, and ensure instruction consistency. */
|
||||||
|
ic ialluis
|
||||||
|
dsb sy
|
||||||
|
isb
|
||||||
|
|
||||||
|
/* Set SCTLR_EL1 to disable the caches and mmu. */
|
||||||
|
/* SCTLR_EL1: */
|
||||||
|
/* - M = 0 */
|
||||||
|
/* - C = 0 */
|
||||||
|
/* - I = 0 */
|
||||||
|
mrs x0, sctlr_el1
|
||||||
|
LOAD_IMMEDIATE_64(x1, ~0x1005)
|
||||||
|
and x0, x0, x1
|
||||||
|
msr sctlr_el1, x0
|
||||||
|
|
||||||
|
mov x30, x22
|
||||||
|
ret
|
||||||
|
|
||||||
|
/* ams::kern::arm64::cpu::FlushEntireDataCacheLocalWithoutStack() */
|
||||||
|
.section .crt0.text._ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv, "ax", %progbits
|
||||||
|
.global _ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv
|
||||||
|
.type _ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv, %function
|
||||||
|
_ZN3ams4kern5arm643cpu37FlushEntireDataCacheLocalWithoutStackEv:
|
||||||
|
/* The stack isn't set up, so we'll need to trash a register. */
|
||||||
|
mov x23, x30
|
||||||
|
|
||||||
|
/* CacheLineIdAccessor clidr_el1; */
|
||||||
|
mrs x10, clidr_el1
|
||||||
|
/* const int levels_of_unification = clidr_el1.GetLevelsOfUnification(); */
|
||||||
|
ubfx x10, x10, #0x15, 3
|
||||||
|
|
||||||
|
/* int level = levels_of_unification - 1 */
|
||||||
|
sub w9, w10, #1
|
||||||
|
|
||||||
|
/* while (level >= 0) { */
|
||||||
|
begin_flush_cache_local_loop:
|
||||||
|
cmn w9, #1
|
||||||
|
b.eq done_flush_cache_local_loop
|
||||||
|
|
||||||
|
/* FlushEntireDataCacheImplWithoutStack(level); */
|
||||||
|
mov w0, w9
|
||||||
|
bl _ZN3ams4kern5arm643cpu36FlushEntireDataCacheImplWithoutStackEv
|
||||||
|
|
||||||
|
/* level--; */
|
||||||
|
sub w9, w9, #1
|
||||||
|
|
||||||
|
/* } */
|
||||||
|
b begin_flush_cache_local_loop
|
||||||
|
|
||||||
|
done_flush_cache_local_loop:
|
||||||
|
mov x30, x23
|
||||||
|
ret
|
||||||
|
|
||||||
|
/* ams::kern::arm64::cpu::FlushEntireDataCacheSharedWithoutStack() */
|
||||||
|
.section .crt0.text._ZN3ams4kern5arm643cpu38FlushEntireDataCacheSharedWithoutStackEv, "ax", %progbits
|
||||||
|
.global _ZN3ams4kern5arm643cpu38FlushEntireDataCacheSharedWithoutStackEv
|
||||||
|
.type _ZN3ams4kern5arm643cpu38FlushEntireDataCacheSharedWithoutStackEv, %function
|
||||||
|
_ZN3ams4kern5arm643cpu38FlushEntireDataCacheSharedWithoutStackEv:
|
||||||
|
/* The stack isn't set up, so we'll need to trash a register. */
|
||||||
|
mov x23, x30
|
||||||
|
|
||||||
|
/* CacheLineIdAccessor clidr_el1; */
|
||||||
|
mrs x10, clidr_el1
|
||||||
|
/* const int levels_of_coherency = clidr_el1.GetLevelsOfCoherency(); */
|
||||||
|
ubfx x9, x10, #0x18, 3
|
||||||
|
/* const int levels_of_unification = clidr_el1.GetLevelsOfUnification(); */
|
||||||
|
ubfx x10, x10, #0x15, 3
|
||||||
|
|
||||||
|
/* int level = levels_of_coherency */
|
||||||
|
|
||||||
|
/* while (level >= levels_of_unification) { */
|
||||||
|
begin_flush_cache_shared_loop:
|
||||||
|
cmp w10, w9
|
||||||
|
b.gt done_flush_cache_shared_loop
|
||||||
|
|
||||||
|
/* FlushEntireDataCacheImplWithoutStack(level); */
|
||||||
|
mov w0, w9
|
||||||
|
bl _ZN3ams4kern5arm643cpu36FlushEntireDataCacheImplWithoutStackEv
|
||||||
|
|
||||||
|
/* level--; */
|
||||||
|
sub w9, w9, #1
|
||||||
|
|
||||||
|
/* } */
|
||||||
|
b begin_flush_cache_shared_loop
|
||||||
|
|
||||||
|
done_flush_cache_shared_loop:
|
||||||
|
mov x30, x23
|
||||||
|
ret
|
||||||
|
|
||||||
|
/* ams::kern::arm64::cpu::FlushEntireDataCacheImplWithoutStack() */
|
||||||
|
.section .crt0.text._ZN3ams4kern5arm643cpu36FlushEntireDataCacheImplWithoutStackEv, "ax", %progbits
|
||||||
|
.global _ZN3ams4kern5arm643cpu36FlushEntireDataCacheImplWithoutStackEv
|
||||||
|
.type _ZN3ams4kern5arm643cpu36FlushEntireDataCacheImplWithoutStackEv, %function
|
||||||
|
_ZN3ams4kern5arm643cpu36FlushEntireDataCacheImplWithoutStackEv:
|
||||||
|
/* const u64 level_sel_value = static_cast<u64>(level << 1); */
|
||||||
|
lsl w6, w0, #1
|
||||||
|
sxtw x6, w6
|
||||||
|
|
||||||
|
/* cpu::SetCsselrEl1(level_sel_value); */
|
||||||
|
msr csselr_el1, x6
|
||||||
|
|
||||||
|
/* cpu::InstructionMemoryBarrier(); */
|
||||||
|
isb
|
||||||
|
|
||||||
|
/* CacheSizeIdAccessor ccsidr_el1; */
|
||||||
|
mrs x3, ccsidr_el1
|
||||||
|
|
||||||
|
/* const int num_ways = ccsidr_el1.GetAssociativity(); */
|
||||||
|
ubfx x7, x3, #3, #0xA
|
||||||
|
mov w8, w7
|
||||||
|
|
||||||
|
/* const int line_size = ccsidr_el1.GetLineSize(); */
|
||||||
|
and x4, x3, #7
|
||||||
|
|
||||||
|
/* const u64 way_shift = static_cast<u64>(__builtin_clz(num_ways)); */
|
||||||
|
clz w7, w7
|
||||||
|
|
||||||
|
/* const u64 set_shift = static_cast<u64>(line_size + 4); */
|
||||||
|
add w4, w4, #4
|
||||||
|
|
||||||
|
/* const int num_sets = ccsidr_el1.GetNumberOfSets(); */
|
||||||
|
ubfx w3, w3, #0xD, #0xF
|
||||||
|
|
||||||
|
/* int way = 0; */
|
||||||
|
mov x5, #0
|
||||||
|
|
||||||
|
/* while (way <= num_ways) { */
|
||||||
|
begin_flush_cache_impl_way_loop:
|
||||||
|
cmp w8, w5
|
||||||
|
b.lt done_flush_cache_impl_way_loop
|
||||||
|
|
||||||
|
/* int set = 0; */
|
||||||
|
mov x0, #0
|
||||||
|
|
||||||
|
/* while (set <= num_sets) { */
|
||||||
|
begin_flush_cache_impl_set_loop:
|
||||||
|
cmp w3, w0
|
||||||
|
b.lt done_flush_cache_impl_set_loop
|
||||||
|
|
||||||
|
/* const u64 cisw_value = (static_cast<u64>(way) << way_shift) | (static_cast<u64>(set) << set_shift) | level_sel_value; */
|
||||||
|
lsl x2, x5, x7
|
||||||
|
lsl x1, x0, x4
|
||||||
|
orr x1, x1, x2
|
||||||
|
orr x1, x1, x6
|
||||||
|
|
||||||
|
/* __asm__ __volatile__("dc cisw, %0" :: "r"(cisw_value) : "memory"); */
|
||||||
|
dc cisw, x1
|
||||||
|
|
||||||
|
/* set++; */
|
||||||
|
add x0, x0, #1
|
||||||
|
|
||||||
|
/* } */
|
||||||
|
b begin_flush_cache_impl_set_loop
|
||||||
|
done_flush_cache_impl_set_loop:
|
||||||
|
|
||||||
|
/* way++; */
|
||||||
|
add x5, x5, 1
|
||||||
|
|
||||||
|
/* } */
|
||||||
|
b begin_flush_cache_impl_way_loop
|
||||||
|
done_flush_cache_impl_way_loop:
|
||||||
|
ret
|
||||||
|
|
Loading…
Reference in a new issue