mirror of
https://github.com/Atmosphere-NX/Atmosphere
synced 2024-11-15 17:39:28 +00:00
thermosphere: implement reading and writing guest memory
This commit is contained in:
parent
0f0228e240
commit
217c1ad054
22 changed files with 467 additions and 118 deletions
|
@ -19,6 +19,12 @@
|
||||||
#include "utils.h"
|
#include "utils.h"
|
||||||
#include "sysreg.h"
|
#include "sysreg.h"
|
||||||
|
|
||||||
|
static inline u32 cacheGetInstructionCachePolicy(void)
|
||||||
|
{
|
||||||
|
u32 ctr = (u32)GET_SYSREG(ctr_el0);
|
||||||
|
return (ctr >> 14) & 3;
|
||||||
|
}
|
||||||
|
|
||||||
static inline u32 cacheGetSmallestInstructionCacheLineSize(void)
|
static inline u32 cacheGetSmallestInstructionCacheLineSize(void)
|
||||||
{
|
{
|
||||||
u32 ctr = (u32)GET_SYSREG(ctr_el0);
|
u32 ctr = (u32)GET_SYSREG(ctr_el0);
|
||||||
|
|
|
@ -140,7 +140,7 @@ vector_entry _synchSp0
|
||||||
msr elr_el2, x18
|
msr elr_el2, x18
|
||||||
// Note: non-broadcasting TLB maintenance op
|
// Note: non-broadcasting TLB maintenance op
|
||||||
tlbi alle2
|
tlbi alle2
|
||||||
dsb ish
|
dsb nsh
|
||||||
isb
|
isb
|
||||||
eret
|
eret
|
||||||
check_vector_size _synchSp0
|
check_vector_size _synchSp0
|
||||||
|
@ -271,11 +271,30 @@ doSmcIndirectCallImplSize:
|
||||||
|
|
||||||
/* Current EL, SPx */
|
/* Current EL, SPx */
|
||||||
|
|
||||||
EXCEPTION_HANDLER_START _synchSpx, EXCEPTION_TYPE_HOST_CRASH
|
vector_entry _synchSpx
|
||||||
|
// Ignore crash if x18 is 0, when we're copying memory from the guest (w/ irq masked)
|
||||||
|
cbz x18, _synchSpxIgnoreCrash
|
||||||
|
|
||||||
|
PIVOT_STACK_FOR_CRASH
|
||||||
|
SAVE_MOST_REGISTERS
|
||||||
|
|
||||||
|
mov x0, sp
|
||||||
|
mov w1, #0
|
||||||
|
|
||||||
|
bl exceptionEntryPostprocess
|
||||||
|
|
||||||
mov x0, sp
|
mov x0, sp
|
||||||
mrs x1, esr_el2
|
mrs x1, esr_el2
|
||||||
bl handleSameElSyncException
|
bl handleSameElSyncException
|
||||||
EXCEPTION_HANDLER_END _synchSpx, EXCEPTION_TYPE_HOST_CRASH
|
|
||||||
|
b .
|
||||||
|
|
||||||
|
_synchSpxIgnoreCrash:
|
||||||
|
mrs x18, elr_el2
|
||||||
|
add x18, x18, #4
|
||||||
|
msr elr_el2, x18
|
||||||
|
eret
|
||||||
|
check_vector_size _synchSpx
|
||||||
|
|
||||||
EXCEPTION_HANDLER_START _irqSpx, EXCEPTION_TYPE_HOST
|
EXCEPTION_HANDLER_START _irqSpx, EXCEPTION_TYPE_HOST
|
||||||
mov x0, sp
|
mov x0, sp
|
||||||
|
|
|
@ -25,7 +25,7 @@ void executeFunctionOnCores(ExecutedFunction fun, void *args, bool sync, u32 cor
|
||||||
currentCoreCtx->executedFunction = fun;
|
currentCoreCtx->executedFunction = fun;
|
||||||
currentCoreCtx->executedFunctionArgs = args;
|
currentCoreCtx->executedFunctionArgs = args;
|
||||||
currentCoreCtx->executedFunctionSync = sync;
|
currentCoreCtx->executedFunctionSync = sync;
|
||||||
|
__compiler_barrier();
|
||||||
generateSgiForList(ThermosphereSgi_ExecuteFunction, coreList);
|
generateSgiForList(ThermosphereSgi_ExecuteFunction, coreList);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -42,13 +42,13 @@ FUNCTION fpuLoadRegistersFromStorage
|
||||||
msr fpsr, x1
|
msr fpsr, x1
|
||||||
msr fpcr, x2
|
msr fpcr, x2
|
||||||
dsb ish
|
dsb ish
|
||||||
isb ish
|
isb
|
||||||
ret
|
ret
|
||||||
END_FUNCTION
|
END_FUNCTION
|
||||||
|
|
||||||
FUNCTION fpuStoreRegistersToStorage
|
FUNCTION fpuStoreRegistersToStorage
|
||||||
dsb ish
|
dsb ish
|
||||||
isb ish
|
isb
|
||||||
LDSTORE_QREGS stp
|
LDSTORE_QREGS stp
|
||||||
mrs x1, fpsr
|
mrs x1, fpsr
|
||||||
mrs x2, fpcr
|
mrs x2, fpcr
|
||||||
|
|
292
thermosphere/src/guest_memory.c
Normal file
292
thermosphere/src/guest_memory.c
Normal file
|
@ -0,0 +1,292 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2019 Atmosphère-NX
|
||||||
|
*
|
||||||
|
* This program is free software; you can redistribute it and/or modify it
|
||||||
|
* under the terms and conditions of the GNU General Public License,
|
||||||
|
* version 2, as published by the Free Software Foundation.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
||||||
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||||
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||||
|
* more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU General Public License
|
||||||
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <string.h>
|
||||||
|
|
||||||
|
#include "guest_memory.h"
|
||||||
|
#include "memory_map.h"
|
||||||
|
#include "mmu.h"
|
||||||
|
#include "spinlock.h"
|
||||||
|
#include "core_ctx.h"
|
||||||
|
#include "sysreg.h"
|
||||||
|
#include "vgic.h"
|
||||||
|
#include "irq.h"
|
||||||
|
#include "caches.h"
|
||||||
|
|
||||||
|
static size_t guestReadWriteGicd(size_t offset, size_t size, void *readBuf, const void *writeBuf)
|
||||||
|
{
|
||||||
|
recursiveSpinlockLock(&g_irqManager.lock);
|
||||||
|
|
||||||
|
if (readBuf != NULL) {
|
||||||
|
size_t readOffset = 0;
|
||||||
|
size_t rem = size;
|
||||||
|
while (rem > 0) {
|
||||||
|
if ((offset + readOffset) % 4 == 0 && rem >= 4) {
|
||||||
|
// All accesses of this kind are valid
|
||||||
|
*(u32 *)((uintptr_t)readBuf + readOffset) = vgicReadGicdRegister(offset + readOffset, 4);
|
||||||
|
readOffset += 4;
|
||||||
|
rem -= 4;
|
||||||
|
} else if ((offset + readOffset) % 2 == 0 && rem >= 2) {
|
||||||
|
// All accesses of this kind would be translated to ldrh and are thus invalid. Abort.
|
||||||
|
size = readOffset;
|
||||||
|
goto end;
|
||||||
|
} else if (vgicValidateGicdRegisterAccess(offset + readOffset, 1)) {
|
||||||
|
// Valid byte access
|
||||||
|
*(u8 *)((uintptr_t)readBuf + readOffset) = vgicReadGicdRegister(offset + readOffset, 1);
|
||||||
|
readOffset += 1;
|
||||||
|
rem -= 1;
|
||||||
|
} else {
|
||||||
|
// Invalid byte access
|
||||||
|
size = readOffset;
|
||||||
|
goto end;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (writeBuf != NULL) {
|
||||||
|
size_t writeOffset = 0;
|
||||||
|
size_t rem = size;
|
||||||
|
while (rem > 0) {
|
||||||
|
if ((offset + writeOffset) % 4 == 0 && rem >= 4) {
|
||||||
|
// All accesses of this kind are valid
|
||||||
|
vgicWriteGicdRegister(*(u32 *)((uintptr_t)writeBuf + writeOffset), offset + writeOffset, 4);
|
||||||
|
writeOffset += 4;
|
||||||
|
rem -= 4;
|
||||||
|
} else if ((offset + writeOffset) % 2 == 0 && rem >= 2) {
|
||||||
|
// All accesses of this kind would be translated to ldrh and are thus invalid. Abort.
|
||||||
|
size = writeOffset;
|
||||||
|
goto end;
|
||||||
|
} else if (vgicValidateGicdRegisterAccess(offset + writeOffset, 1)) {
|
||||||
|
// Valid byte access
|
||||||
|
vgicWriteGicdRegister(*(u32 *)((uintptr_t)writeBuf + writeOffset), offset + writeOffset, 1);
|
||||||
|
writeOffset += 1;
|
||||||
|
rem -= 1;
|
||||||
|
} else {
|
||||||
|
// Invalid byte access
|
||||||
|
size = writeOffset;
|
||||||
|
goto end;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
end:
|
||||||
|
recursiveSpinlockUnlock(&g_irqManager.lock);
|
||||||
|
return size;
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t guestReadWriteDeviceMemory(void *addr, size_t size, void *readBuf, const void *writeBuf)
|
||||||
|
{
|
||||||
|
// We might trigger bus errors... ignore the exception and return early if that's the case
|
||||||
|
|
||||||
|
CoreCtx *curCtxBackup = currentCoreCtx;
|
||||||
|
__compiler_barrier();
|
||||||
|
currentCoreCtx = NULL;
|
||||||
|
__compiler_barrier();
|
||||||
|
|
||||||
|
uintptr_t addri = (uintptr_t)addr;
|
||||||
|
|
||||||
|
if (readBuf != NULL) {
|
||||||
|
size_t readOffset = 0;
|
||||||
|
size_t rem = size;
|
||||||
|
while (rem > 0 && (__compiler_barrier(), currentCoreCtx == NULL)) {
|
||||||
|
if ((addri + readOffset) % 4 == 0 && rem >= 4) {
|
||||||
|
*(vu32 *)((uintptr_t)readBuf + readOffset) = *(vu32 *)(addri + readOffset);
|
||||||
|
readOffset += 4;
|
||||||
|
rem -= 4;
|
||||||
|
} else if (readOffset % 2 == 0 && rem >= 2) {
|
||||||
|
*(vu16 *)((uintptr_t)readBuf + readOffset) = *(vu16 *)(addri + readOffset);
|
||||||
|
readOffset += 2;
|
||||||
|
rem -= 2;
|
||||||
|
} else {
|
||||||
|
*(vu8 *)((uintptr_t)readBuf + readOffset) = *(vu8 *)(addri + readOffset);
|
||||||
|
readOffset += 1;
|
||||||
|
rem -= 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (rem != 0) {
|
||||||
|
size = readOffset;
|
||||||
|
goto end;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (writeBuf != NULL) {
|
||||||
|
size_t writeOffset = 0;
|
||||||
|
size_t rem = size;
|
||||||
|
while (rem > 0 && (__compiler_barrier(), currentCoreCtx == NULL)) {
|
||||||
|
if ((addri + writeOffset) % 4 == 0 && rem >= 4) {
|
||||||
|
*(vu32 *)(addri + writeOffset) = *(vu32 *)((uintptr_t)writeBuf + writeOffset);
|
||||||
|
writeOffset += 4;
|
||||||
|
rem -= 4;
|
||||||
|
} else if (writeOffset % 2 == 0 && rem >= 2) {
|
||||||
|
*(vu16 *)(addri + writeOffset) = *(vu16 *)((uintptr_t)writeBuf + writeOffset);
|
||||||
|
writeOffset += 2;
|
||||||
|
rem -= 2;
|
||||||
|
} else {
|
||||||
|
*(vu8 *)(addri + writeOffset) = *(vu8 *)((uintptr_t)writeBuf + writeOffset);
|
||||||
|
writeOffset += 1;
|
||||||
|
rem -= 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (rem != 0) {
|
||||||
|
size = writeOffset;
|
||||||
|
goto end;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
end:
|
||||||
|
__compiler_barrier();
|
||||||
|
currentCoreCtx = curCtxBackup;
|
||||||
|
__compiler_barrier();
|
||||||
|
return size;
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t guestReadWriteNormalMemory(void *addr, size_t size, void *readBuf, const void *writeBuf)
|
||||||
|
{
|
||||||
|
if (readBuf != NULL) {
|
||||||
|
memcpy(readBuf, addr, size);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (writeBuf != NULL) {
|
||||||
|
memcpy(addr, writeBuf, size);
|
||||||
|
|
||||||
|
// We may have written to executable memory or to translation tables...
|
||||||
|
// & the page may have various aliases.
|
||||||
|
// We need to ensure cache & TLB coherency.
|
||||||
|
cacheCleanDataCacheRangePoU(addr, size);
|
||||||
|
u32 policy = cacheGetInstructionCachePolicy();
|
||||||
|
if (policy == 1 || policy == 2) {
|
||||||
|
// AVIVT, VIVT
|
||||||
|
cacheInvalidateInstructionCache();
|
||||||
|
} else {
|
||||||
|
// VPIPT, PIPT
|
||||||
|
// Ez coherency, just do range operations...
|
||||||
|
cacheInvalidateInstructionCacheRangePoU(addr, size);
|
||||||
|
}
|
||||||
|
__tlb_invalidate_el1();
|
||||||
|
__dsb();
|
||||||
|
__isb();
|
||||||
|
}
|
||||||
|
|
||||||
|
return size;
|
||||||
|
}
|
||||||
|
|
||||||
|
static size_t guestReadWriteMemoryPage(uintptr_t addr, size_t size, void *readBuf, const void *writeBuf)
|
||||||
|
{
|
||||||
|
u64 irqFlags = maskIrq();
|
||||||
|
size_t offset = addr & 0xFFFull;
|
||||||
|
|
||||||
|
// Translate the VA, stages 1&2
|
||||||
|
__asm__ __volatile__ ("at s12e1r, %0" :: "r"(addr) : "memory");
|
||||||
|
u64 par = GET_SYSREG(par_el1);
|
||||||
|
if (par & PAR_F) {
|
||||||
|
// The translation failed. Why?
|
||||||
|
if (par & PAR_S) {
|
||||||
|
// Stage 2 fault. Could be an attempt to access the GICD, let's see what the IPA is...
|
||||||
|
__asm__ __volatile__ ("at s1e1r, %0" :: "r"(addr) : "memory");
|
||||||
|
par = GET_SYSREG(par_el1);
|
||||||
|
if ((par & PAR_F) != 0 || (par & PAR_PA_MASK) != MEMORY_MAP_VA_GICD) {
|
||||||
|
// The guest doesn't have access to it...
|
||||||
|
// Read as 0, write ignored
|
||||||
|
if (readBuf != NULL) {
|
||||||
|
memset(readBuf, 0, size);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// GICD mmio
|
||||||
|
size = guestReadWriteGicd(offset, size, readBuf, writeBuf);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// Oops, couldn't read/write anything (stage 1 fault)
|
||||||
|
size = 0;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
/*
|
||||||
|
Translation didn't fail.
|
||||||
|
|
||||||
|
To avoid "B2.8 Mismatched memory attributes" we must use the same effective
|
||||||
|
attributes & shareability as the guest.
|
||||||
|
|
||||||
|
Note that par_el1 reports the effective shareablity of device and noncacheable memory as inner shareable.
|
||||||
|
In fact, the VMSAv8-64 section in the Armv8 ARM reads:
|
||||||
|
"The shareability field is only relevant if the memory is a Normal Cacheable memory type. All Device and Normal
|
||||||
|
Non-cacheable memory regions are always treated as Outer Shareable, regardless of the translation table
|
||||||
|
shareability attributes."
|
||||||
|
|
||||||
|
There's one corner case where we can't avoid it: another core is running,
|
||||||
|
changes the attributes (other than permissions) of the page, and issues
|
||||||
|
a broadcasting TLB maintenance instructions and/or accesses the page with the altered
|
||||||
|
attribute itself. We don't handle this corner case -- just don't read/write that kind of memory...
|
||||||
|
*/
|
||||||
|
u64 memAttribs = (par >> PAR_ATTR_SHIFT) & PAR_ATTR_MASK;
|
||||||
|
u32 shrb = (par >> PAR_SH_SHIFT) & PAR_SH_MASK;
|
||||||
|
uintptr_t pa = par & PAR_PA_MASK;
|
||||||
|
uintptr_t va = MEMORY_MAP_VA_GUEST_MEM + 0x2000 * currentCoreCtx->coreId;
|
||||||
|
|
||||||
|
u64 mair = GET_SYSREG(mair_el2);
|
||||||
|
mair |= memAttribs << (8 * MEMORY_MAP_MEMTYPE_NORMAL_GUEST_SLOT);
|
||||||
|
SET_SYSREG(mair_el2, mair);
|
||||||
|
__isb();
|
||||||
|
|
||||||
|
u64 attribs = MMU_PTE_BLOCK_XN | MMU_PTE_BLOCK_SH(shrb) | MMU_PTE_BLOCK_MEMTYPE(MEMORY_MAP_MEMTYPE_NORMAL_GUEST_SLOT);
|
||||||
|
mmu_map_page((uintptr_t *)MEMORY_MAP_VA_TTBL, va, pa, attribs);
|
||||||
|
// Note: no need to broadcast here
|
||||||
|
__tlb_invalidate_el2_page_local(pa);
|
||||||
|
__dsb_local();
|
||||||
|
|
||||||
|
void *vaddr = (void *)(va + offset);
|
||||||
|
if (memAttribs & 0xF0) {
|
||||||
|
// Normal memory, or unpredictable
|
||||||
|
size = guestReadWriteNormalMemory(vaddr, size, readBuf, writeBuf);
|
||||||
|
} else {
|
||||||
|
// Device memory, or unpredictable
|
||||||
|
size = guestReadWriteDeviceMemory(vaddr, size, readBuf, writeBuf);
|
||||||
|
}
|
||||||
|
|
||||||
|
__dsb_local();
|
||||||
|
__isb();
|
||||||
|
mmu_unmap_page((uintptr_t *)MEMORY_MAP_VA_TTBL, va);
|
||||||
|
// Note: no need to broadcast here
|
||||||
|
__tlb_invalidate_el2_page_local(pa);
|
||||||
|
__dsb_local();
|
||||||
|
|
||||||
|
mair &= ~(0xFFul << (8 * MEMORY_MAP_MEMTYPE_NORMAL_GUEST_SLOT));
|
||||||
|
SET_SYSREG(mair_el2, mair);
|
||||||
|
__isb();
|
||||||
|
}
|
||||||
|
|
||||||
|
restoreInterruptFlags(irqFlags);
|
||||||
|
return size;
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t guestReadWriteMemory(uintptr_t addr, size_t size, void *readBuf, const void *writeBuf)
|
||||||
|
{
|
||||||
|
uintptr_t curAddr = addr;
|
||||||
|
size_t remainingAmount = size;
|
||||||
|
u8 *rb8 = (u8 *)readBuf;
|
||||||
|
const u8 *wb8 = (const u8*)writeBuf;
|
||||||
|
while (remainingAmount > 0) {
|
||||||
|
size_t expectedAmount = ((curAddr & ~0xFFFul) + 0x1000) - curAddr;
|
||||||
|
expectedAmount = expectedAmount > remainingAmount ? remainingAmount : expectedAmount;
|
||||||
|
size_t actualAmount = guestReadWriteMemoryPage(curAddr, expectedAmount, rb8, wb8);
|
||||||
|
curAddr += actualAmount;
|
||||||
|
rb8 += actualAmount;
|
||||||
|
wb8 += actualAmount;
|
||||||
|
remainingAmount -= actualAmount;
|
||||||
|
if (actualAmount != expectedAmount) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return curAddr - addr;
|
||||||
|
}
|
31
thermosphere/src/guest_memory.h
Normal file
31
thermosphere/src/guest_memory.h
Normal file
|
@ -0,0 +1,31 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2019 Atmosphère-NX
|
||||||
|
*
|
||||||
|
* This program is free software; you can redistribute it and/or modify it
|
||||||
|
* under the terms and conditions of the GNU General Public License,
|
||||||
|
* version 2, as published by the Free Software Foundation.
|
||||||
|
*
|
||||||
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
||||||
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||||
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||||
|
* more details.
|
||||||
|
*
|
||||||
|
* You should have received a copy of the GNU General Public License
|
||||||
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include "utils.h"
|
||||||
|
|
||||||
|
size_t guestReadWriteMemory(uintptr_t addr, size_t size, void *readBuf, const void *writeBuf);
|
||||||
|
|
||||||
|
static inline size_t guestReadMemory(uintptr_t addr, size_t size, void *buf)
|
||||||
|
{
|
||||||
|
return guestReadWriteMemory(addr, size, buf, NULL);
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline size_t guestWriteMemory(uintptr_t addr, size_t size, void *buf)
|
||||||
|
{
|
||||||
|
return guestReadWriteMemory(addr, size, NULL, buf);
|
||||||
|
}
|
|
@ -41,7 +41,7 @@ static void initSysregs(void)
|
||||||
SET_SYSREG(cntp_ctl_el0, 0x00000000);
|
SET_SYSREG(cntp_ctl_el0, 0x00000000);
|
||||||
SET_SYSREG(cntv_ctl_el0, 0x00000000);
|
SET_SYSREG(cntv_ctl_el0, 0x00000000);
|
||||||
|
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -14,6 +14,7 @@
|
||||||
#include "timer.h"
|
#include "timer.h"
|
||||||
#include "irq.h"
|
#include "irq.h"
|
||||||
#include "transport_interface.h"
|
#include "transport_interface.h"
|
||||||
|
#include "guest_memory.h"
|
||||||
|
|
||||||
#include "memory_map.h"
|
#include "memory_map.h"
|
||||||
#include "mmu.h"
|
#include "mmu.h"
|
||||||
|
@ -32,8 +33,8 @@ static void loadKernelViaSemihosting(void)
|
||||||
MMU_PTE_BLOCK_XN | MMU_PTE_BLOCK_INNER_SHAREBLE | MMU_PTE_BLOCK_MEMTYPE(MEMORY_MAP_MEMTYPE_NORMAL_UNCACHEABLE)
|
MMU_PTE_BLOCK_XN | MMU_PTE_BLOCK_INNER_SHAREBLE | MMU_PTE_BLOCK_MEMTYPE(MEMORY_MAP_MEMTYPE_NORMAL_UNCACHEABLE)
|
||||||
);
|
);
|
||||||
|
|
||||||
__tlb_invalidate_el2();
|
__tlb_invalidate_el2_local();
|
||||||
__dsb();
|
__dsb_local();
|
||||||
|
|
||||||
DEBUG("Loading kernel via semihosted file I/O... ");
|
DEBUG("Loading kernel via semihosted file I/O... ");
|
||||||
handle = semihosting_file_open("test_kernel.bin", FOPEN_MODE_RB);
|
handle = semihosting_file_open("test_kernel.bin", FOPEN_MODE_RB);
|
||||||
|
@ -49,8 +50,8 @@ static void loadKernelViaSemihosting(void)
|
||||||
semihosting_file_close(handle);
|
semihosting_file_close(handle);
|
||||||
|
|
||||||
mmu_unmap_range(1, mmuTable, 0x40000000, 0x40000000);
|
mmu_unmap_range(1, mmuTable, 0x40000000, 0x40000000);
|
||||||
__tlb_invalidate_el2();
|
__tlb_invalidate_el2_local();
|
||||||
__dsb();
|
__dsb_local();
|
||||||
|
|
||||||
currentCoreCtx->kernelEntrypoint = buf;
|
currentCoreCtx->kernelEntrypoint = buf;
|
||||||
}
|
}
|
||||||
|
|
|
@ -143,20 +143,20 @@ void memoryMapEnableMmu(const LoadImageLayout *layout)
|
||||||
SET_SYSREG(ttbr0_el2, mmuTable);
|
SET_SYSREG(ttbr0_el2, mmuTable);
|
||||||
SET_SYSREG(tcr_el2, tcr);
|
SET_SYSREG(tcr_el2, tcr);
|
||||||
SET_SYSREG(mair_el2, mair);
|
SET_SYSREG(mair_el2, mair);
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
|
|
||||||
// TLB invalidation
|
// TLB invalidation
|
||||||
// Whether this does anything before MMU is enabled is impldef, apparently
|
// Whether this does anything before MMU is enabled is impldef, apparently
|
||||||
__tlb_invalidate_el2_local();
|
__tlb_invalidate_el2_local();
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
|
|
||||||
// Enable MMU & enable caching. We will crash.
|
// Enable MMU & enable caching. We will crash.
|
||||||
u64 sctlr = GET_SYSREG(sctlr_el2);
|
u64 sctlr = GET_SYSREG(sctlr_el2);
|
||||||
sctlr |= SCTLR_ELx_I | SCTLR_ELx_C | SCTLR_ELx_M;
|
sctlr |= SCTLR_ELx_I | SCTLR_ELx_C | SCTLR_ELx_M;
|
||||||
SET_SYSREG(sctlr_el2, sctlr);
|
SET_SYSREG(sctlr_el2, sctlr);
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -22,6 +22,7 @@
|
||||||
#define MEMORY_MAP_MEMTYPE_NORMAL 1ul
|
#define MEMORY_MAP_MEMTYPE_NORMAL 1ul
|
||||||
#define MEMORY_MAP_MEMTYPE_DEVICE_NGNRE 2ul
|
#define MEMORY_MAP_MEMTYPE_DEVICE_NGNRE 2ul
|
||||||
#define MEMORY_MAP_MEMTYPE_NORMAL_UNCACHEABLE 3ul
|
#define MEMORY_MAP_MEMTYPE_NORMAL_UNCACHEABLE 3ul
|
||||||
|
#define MEMORY_MAP_MEMTYPE_NORMAL_GUEST_SLOT 4ul
|
||||||
|
|
||||||
#define MEMORY_MAP_VA_SPACE_SIZE 39ul
|
#define MEMORY_MAP_VA_SPACE_SIZE 39ul
|
||||||
|
|
||||||
|
|
|
@ -72,6 +72,7 @@
|
||||||
*/
|
*/
|
||||||
#define MMU_PTE_BLOCK_MEMTYPE(x) ((uint64_t)((x) << 2))
|
#define MMU_PTE_BLOCK_MEMTYPE(x) ((uint64_t)((x) << 2))
|
||||||
#define MMU_PTE_BLOCK_NS BITL(5)
|
#define MMU_PTE_BLOCK_NS BITL(5)
|
||||||
|
#define MMU_PTE_BLOCK_SH(x) ((x) << 8)
|
||||||
#define MMU_PTE_BLOCK_NON_SHAREABLE (0ull << 8)
|
#define MMU_PTE_BLOCK_NON_SHAREABLE (0ull << 8)
|
||||||
#define MMU_PTE_BLOCK_OUTER_SHAREABLE (2ull << 8)
|
#define MMU_PTE_BLOCK_OUTER_SHAREABLE (2ull << 8)
|
||||||
#define MMU_PTE_BLOCK_INNER_SHAREBLE (3ull << 8)
|
#define MMU_PTE_BLOCK_INNER_SHAREBLE (3ull << 8)
|
||||||
|
|
|
@ -26,6 +26,6 @@ void devicesMapAllExtra(void)
|
||||||
|
|
||||||
// Don't broadcast, since it's only ran once per boot by only one core, before the others are started...
|
// Don't broadcast, since it's only ran once per boot by only one core, before the others are started...
|
||||||
__tlb_invalidate_el2_local();
|
__tlb_invalidate_el2_local();
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
}
|
}
|
||||||
|
|
|
@ -39,18 +39,18 @@ void stage2ConfigureAndEnable(void)
|
||||||
// Stage2 regs config
|
// Stage2 regs config
|
||||||
SET_SYSREG(vttbr_el2, vttbr);
|
SET_SYSREG(vttbr_el2, vttbr);
|
||||||
SET_SYSREG(vtcr_el2, vtcr);
|
SET_SYSREG(vtcr_el2, vtcr);
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
|
|
||||||
// Enable stage 2
|
// Enable stage 2
|
||||||
u64 hcr = GET_SYSREG(hcr_el2);
|
u64 hcr = GET_SYSREG(hcr_el2);
|
||||||
hcr |= HCR_VM;
|
hcr |= HCR_VM;
|
||||||
SET_SYSREG(hcr_el2, hcr);
|
SET_SYSREG(hcr_el2, hcr);
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
|
|
||||||
// TLB invalidation
|
// TLB invalidation
|
||||||
__tlb_invalidate_el1_stage12_local();
|
__tlb_invalidate_el1_stage12_local();
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
}
|
}
|
||||||
|
|
|
@ -33,6 +33,6 @@ void devicesMapAllExtra(void)
|
||||||
|
|
||||||
// Don't broadcast, since it's only ran once per boot by only one core, before the others are started...
|
// Don't broadcast, since it's only ran once per boot by only one core, before the others are started...
|
||||||
__tlb_invalidate_el2_local();
|
__tlb_invalidate_el2_local();
|
||||||
__dsb();
|
__dsb_local();
|
||||||
__isb();
|
__isb();
|
||||||
}
|
}
|
||||||
|
|
|
@ -16,6 +16,8 @@ void doSmcIndirectCall(ExceptionStackFrame *frame, u32 smcId)
|
||||||
|
|
||||||
cacheHandleSelfModifyingCodePoU(codebuf, doSmcIndirectCallImplSize/4);
|
cacheHandleSelfModifyingCodePoU(codebuf, doSmcIndirectCallImplSize/4);
|
||||||
|
|
||||||
|
__dsb_sy();
|
||||||
|
__isb();
|
||||||
((void (*)(ExceptionStackFrame *))codebuf)(frame);
|
((void (*)(ExceptionStackFrame *))codebuf)(frame);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -62,10 +62,10 @@ static inline bool doApplySoftwareBreakpoint(size_t id)
|
||||||
|
|
||||||
u32 brkInst = 0xF2000000 | bp->uid;
|
u32 brkInst = 0xF2000000 | bp->uid;
|
||||||
|
|
||||||
if (readEl1Memory(&bp->savedInstruction, bp->address, 4) && writeEl1Memory(bp->address, &brkInst, 4)) {
|
/*if (readEl1Memory(&bp->savedInstruction, bp->address, 4) && writeEl1Memory(bp->address, &brkInst, 4)) {
|
||||||
bp->applied = true;
|
bp->applied = true;
|
||||||
return true;
|
return true;
|
||||||
}
|
}*/
|
||||||
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
@ -91,10 +91,10 @@ static inline bool doRevertSoftwareBreakpoint(size_t id)
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (writeEl1Memory(bp->address, &bp->savedInstruction, 4)) {
|
/*if (writeEl1Memory(bp->address, &bp->savedInstruction, 4)) {
|
||||||
bp->applied = false;
|
bp->applied = false;
|
||||||
return true;
|
return true;
|
||||||
}
|
}*/
|
||||||
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
|
@ -52,6 +52,10 @@ _startCommon:
|
||||||
|
|
||||||
// "Boot core only" stuff:
|
// "Boot core only" stuff:
|
||||||
bl cacheClearSharedDataCachesOnBoot
|
bl cacheClearSharedDataCachesOnBoot
|
||||||
|
ic iallu
|
||||||
|
dsb nsh
|
||||||
|
isb
|
||||||
|
|
||||||
// Temporarily use temp end region as stack, then create the translation table
|
// Temporarily use temp end region as stack, then create the translation table
|
||||||
// The stack top is also equal to the mmu table address...
|
// The stack top is also equal to the mmu table address...
|
||||||
adr x0, g_loadImageLayout
|
adr x0, g_loadImageLayout
|
||||||
|
@ -89,9 +93,6 @@ _postMmuEnableReturnAddr:
|
||||||
stp x18, xzr, [sp, #-0x10]!
|
stp x18, xzr, [sp, #-0x10]!
|
||||||
sub sp, sp, #EXCEP_STACK_FRAME_SIZE
|
sub sp, sp, #EXCEP_STACK_FRAME_SIZE
|
||||||
|
|
||||||
dsb sy
|
|
||||||
isb
|
|
||||||
|
|
||||||
mov x0, sp
|
mov x0, sp
|
||||||
mov x1, x20
|
mov x1, x20
|
||||||
bl thermosphereMain
|
bl thermosphereMain
|
||||||
|
|
|
@ -412,13 +412,28 @@
|
||||||
#define MDSCR_SS BITL(0)
|
#define MDSCR_SS BITL(0)
|
||||||
|
|
||||||
// Common CNTHCTL_EL2 flags
|
// Common CNTHCTL_EL2 flags
|
||||||
#define CNTHCTL_EVNTI_MASK 0xFll
|
#define CNTHCTL_EVNTI_MASK 0xFul
|
||||||
#define CNTHCTL_EVNTI_SHIFT 4
|
#define CNTHCTL_EVNTI_SHIFT 4
|
||||||
#define CNTHCTL_EVNTDIR BITL(3)
|
#define CNTHCTL_EVNTDIR BITL(3)
|
||||||
#define CNTHCTL_EVNTEN BITL(2)
|
#define CNTHCTL_EVNTEN BITL(2)
|
||||||
#define CNTHCTL_EL1PCEN BITL(1)
|
#define CNTHCTL_EL1PCEN BITL(1)
|
||||||
#define CNTHCTL_EL1PCTEN BITL(0)
|
#define CNTHCTL_EL1PCTEN BITL(0)
|
||||||
|
|
||||||
|
// PAR_EL1
|
||||||
|
#define PAR_F BITL(0)
|
||||||
|
// Successful translation:
|
||||||
|
#define PAR_ATTR_SHIFT 56
|
||||||
|
#define PAR_ATTR_MASK 0xFFul
|
||||||
|
#define PAR_PA_MASK MASK2L(51, 12) // bits 51-48 RES0 if not implemented
|
||||||
|
#define PAR_NS BITL(9)
|
||||||
|
#define PAR_SH_SHIFT 7
|
||||||
|
#define PAR_SH_MASK 3ul
|
||||||
|
// Faulting translation:
|
||||||
|
#define PAR_S BITL(9)
|
||||||
|
#define PAR_PTW BITL(8)
|
||||||
|
#define PAR_FST_SHIFT 1
|
||||||
|
#define PAR_FST_MASK 0x3Ful
|
||||||
|
|
||||||
#define ENCODE_SYSREG_FIELDS_MOV(op0, op1, crn, crm, op2) (((op0) << 19) | ((op1) << 16) | ((crn) << 12) | ((crm) << 8) | ((op2) << 5))
|
#define ENCODE_SYSREG_FIELDS_MOV(op0, op1, crn, crm, op2) (((op0) << 19) | ((op1) << 16) | ((crn) << 12) | ((crm) << 8) | ((op2) << 5))
|
||||||
#define ENCODE_SYSREG_MOV(name) EVAL(ENCODE_SYSREG_FIELDS_MOV CAT(TUP_, name))
|
#define ENCODE_SYSREG_MOV(name) EVAL(ENCODE_SYSREG_FIELDS_MOV CAT(TUP_, name))
|
||||||
#define MAKE_MSR(name, Rt) (0xD5000000 | ENCODE_SYSREG_MOV(name) | ((Rt) & 0x1F))
|
#define MAKE_MSR(name, Rt) (0xD5000000 | ENCODE_SYSREG_MOV(name) | ((Rt) & 0x1F))
|
||||||
|
|
|
@ -27,44 +27,3 @@ __attribute__((noinline)) bool overlaps(u64 as, u64 ae, u64 bs, u64 be)
|
||||||
return true;
|
return true;
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: put that elsewhere
|
|
||||||
bool readEl1Memory(void *dst, uintptr_t addr, size_t size)
|
|
||||||
{
|
|
||||||
// Note: what if we read uncached regions/not shared?
|
|
||||||
bool valid;
|
|
||||||
|
|
||||||
u64 flags = maskIrq();
|
|
||||||
uintptr_t pa = get_physical_address_el1_stage12(&valid, addr);
|
|
||||||
restoreInterruptFlags(flags);
|
|
||||||
|
|
||||||
if (!valid) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
memcpy(dst, (const void *)pa, size);
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
bool writeEl1Memory(uintptr_t addr, const void *src, size_t size)
|
|
||||||
{
|
|
||||||
bool valid;
|
|
||||||
|
|
||||||
u64 flags = maskIrq();
|
|
||||||
uintptr_t pa = get_physical_address_el1_stage12(&valid, addr);
|
|
||||||
restoreInterruptFlags(flags);
|
|
||||||
|
|
||||||
if (!valid) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
memcpy((void *)pa, src, size);
|
|
||||||
cacheHandleSelfModifyingCodePoU((const void *)pa, size);
|
|
||||||
|
|
||||||
__tlb_invalidate_el1_stage12_local(); //FIXME FIXME FIXME
|
|
||||||
__dsb_sy();
|
|
||||||
__isb();
|
|
||||||
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
|
@ -62,6 +62,11 @@ typedef enum ReadWriteDirection {
|
||||||
DIRECTION_READWRITE = DIRECTION_READ | DIRECTION_WRITE,
|
DIRECTION_READWRITE = DIRECTION_READ | DIRECTION_WRITE,
|
||||||
} ReadWriteDirection;
|
} ReadWriteDirection;
|
||||||
|
|
||||||
|
static inline void __compiler_barrier(void)
|
||||||
|
{
|
||||||
|
__asm__ __volatile__ ("" ::: "memory");
|
||||||
|
}
|
||||||
|
|
||||||
static inline void __wfi(void)
|
static inline void __wfi(void)
|
||||||
{
|
{
|
||||||
__asm__ __volatile__ ("wfi" ::: "memory");
|
__asm__ __volatile__ ("wfi" ::: "memory");
|
||||||
|
@ -100,6 +105,11 @@ static inline void __dsb(void)
|
||||||
__asm__ __volatile__ ("dsb ish" ::: "memory");
|
__asm__ __volatile__ ("dsb ish" ::: "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void __dsb_local(void)
|
||||||
|
{
|
||||||
|
__asm__ __volatile__ ("dsb nsh" ::: "memory");
|
||||||
|
}
|
||||||
|
|
||||||
static inline void __dmb_sy(void)
|
static inline void __dmb_sy(void)
|
||||||
{
|
{
|
||||||
__asm__ __volatile__ ("dmb sy" ::: "memory");
|
__asm__ __volatile__ ("dmb sy" ::: "memory");
|
||||||
|
@ -125,11 +135,27 @@ static inline void __tlb_invalidate_el2_local(void)
|
||||||
__asm__ __volatile__ ("tlbi alle2" ::: "memory");
|
__asm__ __volatile__ ("tlbi alle2" ::: "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void __tlb_invalidate_el2_page(uintptr_t addr)
|
||||||
|
{
|
||||||
|
__asm__ __volatile__ ("tlbi vae2is, %0" :: "r"(addr) : "memory");
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline void __tlb_invalidate_el2_page_local(uintptr_t addr)
|
||||||
|
{
|
||||||
|
__asm__ __volatile__ ("tlbi vae2is, %0" :: "r"(addr) : "memory");
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
static inline void __tlb_invalidate_el1_stage12_local(void)
|
static inline void __tlb_invalidate_el1_stage12_local(void)
|
||||||
{
|
{
|
||||||
__asm__ __volatile__ ("tlbi alle1" ::: "memory");
|
__asm__ __volatile__ ("tlbi alle1" ::: "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void __tlb_invalidate_el1(void)
|
||||||
|
{
|
||||||
|
__asm__ __volatile__ ("tlbi vmalle1is" ::: "memory");
|
||||||
|
}
|
||||||
|
|
||||||
bool overlaps(u64 as, u64 ae, u64 bs, u64 be);
|
bool overlaps(u64 as, u64 ae, u64 bs, u64 be);
|
||||||
|
|
||||||
// Assumes addr is valid, must be called with interrupts masked
|
// Assumes addr is valid, must be called with interrupts masked
|
||||||
|
@ -138,23 +164,11 @@ static inline uintptr_t va2pa(const void *el2_vaddr) {
|
||||||
// For debug purposes only
|
// For debug purposes only
|
||||||
uintptr_t PAR;
|
uintptr_t PAR;
|
||||||
uintptr_t va = (uintptr_t)el2_vaddr;
|
uintptr_t va = (uintptr_t)el2_vaddr;
|
||||||
__asm__ __volatile__ ("at s1e2r, %0" :: "r"(va));
|
__asm__ __volatile__ ("at s1e2r, %0" :: "r"(va) : "memory");
|
||||||
__asm__ __volatile__ ("mrs %0, par_el1" : "=r"(PAR));
|
__asm__ __volatile__ ("mrs %0, par_el1" : "=r"(PAR) :: "memory");
|
||||||
return (PAR & MASK2L(47, 12)) | (va & MASKL(12));
|
return (PAR & MASK2L(47, 12)) | (va & MASKL(12));
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline uintptr_t get_physical_address_el1_stage12(bool *valid, uintptr_t el1_vaddr) {
|
|
||||||
// NOTE: interrupt must be disabled when calling this func
|
|
||||||
uintptr_t PAR;
|
|
||||||
__asm__ __volatile__ ("at s12e1r, %0" :: "r"(el1_vaddr)); // note: we don't care whether it's writable in EL1&0 translation regime
|
|
||||||
__asm__ __volatile__ ("mrs %0, par_el1" : "=r"(PAR));
|
|
||||||
*valid = (PAR & 1) == 0ull;
|
|
||||||
return (PAR & 1) ? 0ull : (PAR & MASK2L(47, 12)) | ((uintptr_t)el1_vaddr & MASKL(12));
|
|
||||||
}
|
|
||||||
|
|
||||||
bool readEl1Memory(void *dst, uintptr_t addr, size_t size);
|
|
||||||
bool writeEl1Memory(uintptr_t addr, const void *src, size_t size);
|
|
||||||
|
|
||||||
static inline void panic(void) {
|
static inline void panic(void) {
|
||||||
#ifndef PLATFORM_QEMU
|
#ifndef PLATFORM_QEMU
|
||||||
__builtin_trap();
|
__builtin_trap();
|
||||||
|
|
|
@ -526,12 +526,23 @@ static inline u32 vgicGetPeripheralId2Register(void)
|
||||||
return 2 << 4;
|
return 2 << 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void handleVgicMmioWrite(ExceptionStackFrame *frame, DataAbortIss dabtIss, size_t offset)
|
bool vgicValidateGicdRegisterAccess(size_t offset, size_t sz)
|
||||||
{
|
{
|
||||||
size_t sz = BITL(dabtIss.sas);
|
// ipriorityr, itargetsr, *pendsgir are byte-accessible
|
||||||
u32 val = (u32)(readFrameRegisterZ(frame, dabtIss.srt) & MASKL(8 * sz));
|
if (
|
||||||
uintptr_t addr = (uintptr_t)gicd + offset;
|
!(offset >= GICDOFF(ipriorityr) && offset < GICDOFF(ipriorityr) + GIC_IRQID_MAX) &&
|
||||||
|
!(offset >= GICDOFF(itargetsr) && offset < GICDOFF(itargetsr) + GIC_IRQID_MAX) &&
|
||||||
|
!(offset >= GICDOFF(cpendsgir) && offset < GICDOFF(cpendsgir) + 16) &&
|
||||||
|
!(offset >= GICDOFF(spendsgir) && offset < GICDOFF(spendsgir) + 16)
|
||||||
|
) {
|
||||||
|
return (offset & 3) == 0 && sz == 4;
|
||||||
|
} else {
|
||||||
|
return sz == 1 || (sz == 4 && ((offset & 3) != 0));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void vgicWriteGicdRegister(u32 val, size_t offset, size_t sz)
|
||||||
|
{
|
||||||
//DEBUG("gicd write off 0x%03llx sz %lx val %x w%d\n", offset, sz, val, (int)dabtIss.srt);
|
//DEBUG("gicd write off 0x%03llx sz %lx val %x w%d\n", offset, sz, val, (int)dabtIss.srt);
|
||||||
|
|
||||||
switch (offset) {
|
switch (offset) {
|
||||||
|
@ -607,19 +618,15 @@ static void handleVgicMmioWrite(ExceptionStackFrame *frame, DataAbortIss dabtIss
|
||||||
break;
|
break;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
dumpUnhandledDataAbort(dabtIss, addr, "GICD reserved/implementation-defined register");
|
DEBUG("Write to GICD reserved/implementation-defined register offset=0x%03lx value=0x%08lx", offset, val);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void handleVgicMmioRead(ExceptionStackFrame *frame, DataAbortIss dabtIss, size_t offset)
|
u32 vgicReadGicdRegister(size_t offset, size_t sz)
|
||||||
{
|
{
|
||||||
size_t sz = BITL(dabtIss.sas);
|
|
||||||
uintptr_t addr = (uintptr_t)gicd + offset;
|
|
||||||
|
|
||||||
u32 val = 0;
|
|
||||||
|
|
||||||
//DEBUG("gicd read off 0x%03llx sz %lx\n", offset, sz);
|
//DEBUG("gicd read off 0x%03llx sz %lx\n", offset, sz);
|
||||||
|
u32 val = 0;
|
||||||
|
|
||||||
switch (offset) {
|
switch (offset) {
|
||||||
case GICDOFF(icfgr) ... GICDOFF(icfgr) + 31/4:
|
case GICDOFF(icfgr) ... GICDOFF(icfgr) + 31/4:
|
||||||
|
@ -682,7 +689,7 @@ static void handleVgicMmioRead(ExceptionStackFrame *frame, DataAbortIss dabtIss,
|
||||||
|
|
||||||
case GICDOFF(sgir):
|
case GICDOFF(sgir):
|
||||||
// Write-only register
|
// Write-only register
|
||||||
dumpUnhandledDataAbort(dabtIss, addr, "GICD read to write-only register GCID_SGIR");
|
DEBUG("Read from write-only register GCID_SGIR\n");
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case GICDOFF(icpidr2):
|
case GICDOFF(icpidr2):
|
||||||
|
@ -690,10 +697,24 @@ static void handleVgicMmioRead(ExceptionStackFrame *frame, DataAbortIss dabtIss,
|
||||||
break;
|
break;
|
||||||
|
|
||||||
default:
|
default:
|
||||||
dumpUnhandledDataAbort(dabtIss, addr, "GICD reserved/implementation-defined register");
|
DEBUG("Read from GICD reserved/implementation-defined register offset=0x%03lx\n", offset);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
return val;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void handleVgicMmioWrite(ExceptionStackFrame *frame, DataAbortIss dabtIss, size_t offset)
|
||||||
|
{
|
||||||
|
size_t sz = BITL(dabtIss.sas);
|
||||||
|
u32 val = (u32)(readFrameRegisterZ(frame, dabtIss.srt) & MASKL(8 * sz));
|
||||||
|
vgicWriteGicdRegister(val, offset, sz);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void handleVgicMmioRead(ExceptionStackFrame *frame, DataAbortIss dabtIss, size_t offset)
|
||||||
|
{
|
||||||
|
size_t sz = BITL(dabtIss.sas);
|
||||||
|
u32 val = vgicReadGicdRegister(offset, sz);
|
||||||
writeFrameRegisterZ(frame, dabtIss.srt, val);
|
writeFrameRegisterZ(frame, dabtIss.srt, val);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -993,28 +1014,10 @@ void handleVgicdMmio(ExceptionStackFrame *frame, DataAbortIss dabtIss, size_t of
|
||||||
{
|
{
|
||||||
size_t sz = BITL(dabtIss.sas);
|
size_t sz = BITL(dabtIss.sas);
|
||||||
uintptr_t addr = (uintptr_t)gicd + offset;
|
uintptr_t addr = (uintptr_t)gicd + offset;
|
||||||
bool oops = true;
|
bool oops = !vgicValidateGicdRegisterAccess(offset, sz);
|
||||||
|
|
||||||
// ipriorityr, itargetsr, *pendsgir are byte-accessible
|
if (oops) {
|
||||||
if (
|
dumpUnhandledDataAbort(dabtIss, addr, "invalid GICD register access");
|
||||||
!(offset >= GICDOFF(ipriorityr) && offset < GICDOFF(ipriorityr) + GIC_IRQID_MAX) &&
|
|
||||||
!(offset >= GICDOFF(itargetsr) && offset < GICDOFF(itargetsr) + GIC_IRQID_MAX) &&
|
|
||||||
!(offset >= GICDOFF(cpendsgir) && offset < GICDOFF(cpendsgir) + 16) &&
|
|
||||||
!(offset >= GICDOFF(spendsgir) && offset < GICDOFF(spendsgir) + 16)
|
|
||||||
) {
|
|
||||||
if ((offset & 3) != 0 || sz != 4) {
|
|
||||||
dumpUnhandledDataAbort(dabtIss, addr, "GICD non-word aligned MMIO");
|
|
||||||
} else {
|
|
||||||
oops = false;
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
if (sz != 1 && sz != 4) {
|
|
||||||
dumpUnhandledDataAbort(dabtIss, addr, "GICD 16 or 64-bit access");
|
|
||||||
} else if (sz == 4 && (offset & 3) != 0) {
|
|
||||||
dumpUnhandledDataAbort(dabtIss, addr, "GICD misaligned MMIO");
|
|
||||||
} else {
|
|
||||||
oops = false;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
recursiveSpinlockLock(&g_irqManager.lock);
|
recursiveSpinlockLock(&g_irqManager.lock);
|
||||||
|
|
|
@ -19,6 +19,10 @@
|
||||||
#include "types.h"
|
#include "types.h"
|
||||||
#include "data_abort.h"
|
#include "data_abort.h"
|
||||||
|
|
||||||
|
bool vgicValidateGicdRegisterAccess(size_t offset, size_t sz);
|
||||||
|
void vgicWriteGicdRegister(u32 val, size_t offset, size_t sz);
|
||||||
|
u32 vgicReadGicdRegister(size_t offset, size_t sz);
|
||||||
|
|
||||||
void handleVgicdMmio(ExceptionStackFrame *frame, DataAbortIss dabtIss, size_t offset);
|
void handleVgicdMmio(ExceptionStackFrame *frame, DataAbortIss dabtIss, size_t offset);
|
||||||
|
|
||||||
void vgicInit(void);
|
void vgicInit(void);
|
||||||
|
|
Loading…
Reference in a new issue