mirror of
https://github.com/Atmosphere-NX/Atmosphere
synced 2024-12-22 04:11:18 +00:00
kern: start KPageTable(Impl) refactor, use array-with-levels for KPageTableImpl
This commit is contained in:
parent
7aa0bed869
commit
02e837d82e
3 changed files with 205 additions and 230 deletions
|
@ -170,9 +170,17 @@ namespace ams::kern::arch::arm64 {
|
||||||
constexpr ALWAYS_INLINE bool IsReadOnly() const { return this->GetBits(7, 1) != 0; }
|
constexpr ALWAYS_INLINE bool IsReadOnly() const { return this->GetBits(7, 1) != 0; }
|
||||||
constexpr ALWAYS_INLINE bool IsUserAccessible() const { return this->GetBits(6, 1) != 0; }
|
constexpr ALWAYS_INLINE bool IsUserAccessible() const { return this->GetBits(6, 1) != 0; }
|
||||||
constexpr ALWAYS_INLINE bool IsNonSecure() const { return this->GetBits(5, 1) != 0; }
|
constexpr ALWAYS_INLINE bool IsNonSecure() const { return this->GetBits(5, 1) != 0; }
|
||||||
|
|
||||||
|
constexpr ALWAYS_INLINE u64 GetTestTableMask() const { return (m_attributes & ExtensionFlag_TestTableMask); }
|
||||||
|
|
||||||
constexpr ALWAYS_INLINE bool IsBlock() const { return (m_attributes & ExtensionFlag_TestTableMask) == ExtensionFlag_Valid; }
|
constexpr ALWAYS_INLINE bool IsBlock() const { return (m_attributes & ExtensionFlag_TestTableMask) == ExtensionFlag_Valid; }
|
||||||
|
constexpr ALWAYS_INLINE bool IsPage() const { return (m_attributes & ExtensionFlag_TestTableMask) == ExtensionFlag_TestTableMask; }
|
||||||
constexpr ALWAYS_INLINE bool IsTable() const { return (m_attributes & ExtensionFlag_TestTableMask) == 2; }
|
constexpr ALWAYS_INLINE bool IsTable() const { return (m_attributes & ExtensionFlag_TestTableMask) == 2; }
|
||||||
constexpr ALWAYS_INLINE bool IsEmpty() const { return (m_attributes & ExtensionFlag_TestTableMask) == 0; }
|
constexpr ALWAYS_INLINE bool IsEmpty() const { return (m_attributes & ExtensionFlag_TestTableMask) == 0; }
|
||||||
|
|
||||||
|
constexpr ALWAYS_INLINE KPhysicalAddress GetTable() const { return this->SelectBits(12, 36); }
|
||||||
|
|
||||||
|
constexpr ALWAYS_INLINE bool IsMappedTable() const { return this->GetBits(0, 2) == 3; }
|
||||||
constexpr ALWAYS_INLINE bool IsMapped() const { return this->GetBits(0, 1) != 0; }
|
constexpr ALWAYS_INLINE bool IsMapped() const { return this->GetBits(0, 1) != 0; }
|
||||||
|
|
||||||
constexpr ALWAYS_INLINE decltype(auto) SetUserExecuteNever(bool en) { this->SetBit(54, en); return *this; }
|
constexpr ALWAYS_INLINE decltype(auto) SetUserExecuteNever(bool en) { this->SetBit(54, en); return *this; }
|
||||||
|
@ -196,10 +204,13 @@ namespace ams::kern::arch::arm64 {
|
||||||
return (m_attributes & BaseMaskForMerge) == attr;
|
return (m_attributes & BaseMaskForMerge) == attr;
|
||||||
}
|
}
|
||||||
|
|
||||||
constexpr ALWAYS_INLINE u64 GetRawAttributesUnsafeForSwap() const {
|
constexpr ALWAYS_INLINE u64 GetRawAttributesUnsafe() const {
|
||||||
return m_attributes;
|
return m_attributes;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
constexpr ALWAYS_INLINE u64 GetRawAttributesUnsafeForSwap() const {
|
||||||
|
return m_attributes;
|
||||||
|
}
|
||||||
protected:
|
protected:
|
||||||
constexpr ALWAYS_INLINE u64 GetRawAttributes() const {
|
constexpr ALWAYS_INLINE u64 GetRawAttributes() const {
|
||||||
return m_attributes;
|
return m_attributes;
|
||||||
|
|
|
@ -37,10 +37,17 @@ namespace ams::kern::arch::arm64 {
|
||||||
constexpr bool IsTailMergeDisabled() const { return (this->sw_reserved_bits & PageTableEntry::SoftwareReservedBit_DisableMergeHeadTail) != 0; }
|
constexpr bool IsTailMergeDisabled() const { return (this->sw_reserved_bits & PageTableEntry::SoftwareReservedBit_DisableMergeHeadTail) != 0; }
|
||||||
};
|
};
|
||||||
|
|
||||||
|
enum EntryLevel : u32 {
|
||||||
|
EntryLevel_L3 = 0,
|
||||||
|
EntryLevel_L2 = 1,
|
||||||
|
EntryLevel_L1 = 2,
|
||||||
|
EntryLevel_Count = 3,
|
||||||
|
};
|
||||||
|
|
||||||
struct TraversalContext {
|
struct TraversalContext {
|
||||||
const L1PageTableEntry *l1_entry;
|
const PageTableEntry *level_entries[EntryLevel_Count];
|
||||||
const L2PageTableEntry *l2_entry;
|
EntryLevel level;
|
||||||
const L3PageTableEntry *l3_entry;
|
bool is_contiguous;
|
||||||
};
|
};
|
||||||
private:
|
private:
|
||||||
static constexpr size_t PageBits = util::CountTrailingZeros(PageSize);
|
static constexpr size_t PageBits = util::CountTrailingZeros(PageSize);
|
||||||
|
@ -53,16 +60,26 @@ namespace ams::kern::arch::arm64 {
|
||||||
return (value >> Offset) & ((1ul << Count) - 1);
|
return (value >> Offset) & ((1ul << Count) - 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static constexpr ALWAYS_INLINE u64 GetBits(u64 value, size_t offset, size_t count) {
|
||||||
|
return (value >> offset) & ((1ul << count) - 1);
|
||||||
|
}
|
||||||
|
|
||||||
template<size_t Offset, size_t Count>
|
template<size_t Offset, size_t Count>
|
||||||
constexpr ALWAYS_INLINE u64 SelectBits(u64 value) {
|
static constexpr ALWAYS_INLINE u64 SelectBits(u64 value) {
|
||||||
return value & (((1ul << Count) - 1) << Offset);
|
return value & (((1ul << Count) - 1) << Offset);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static constexpr ALWAYS_INLINE u64 SelectBits(u64 value, size_t offset, size_t count) {
|
||||||
|
return value & (((1ul << count) - 1) << offset);
|
||||||
|
}
|
||||||
|
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetL0Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 0), LevelBits>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetL0Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 0), LevelBits>(GetInteger(addr)); }
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetL1Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 1), LevelBits>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetL1Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 1), LevelBits>(GetInteger(addr)); }
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetL2Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 2), LevelBits>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetL2Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 2), LevelBits>(GetInteger(addr)); }
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetL3Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 3), LevelBits>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetL3Index(KProcessAddress addr) { return GetBits<PageBits + LevelBits * (NumLevels - 3), LevelBits>(GetInteger(addr)); }
|
||||||
|
|
||||||
|
static constexpr ALWAYS_INLINE uintptr_t GetLevelIndex(KProcessAddress addr, EntryLevel level) { return GetBits(GetInteger(addr), PageBits + LevelBits * level, LevelBits); }
|
||||||
|
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetL1Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 1)>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetL1Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 1)>(GetInteger(addr)); }
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetL2Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 2)>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetL2Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 2)>(GetInteger(addr)); }
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetL3Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 3)>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetL3Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 3)>(GetInteger(addr)); }
|
||||||
|
@ -70,13 +87,16 @@ namespace ams::kern::arch::arm64 {
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetContiguousL2Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 2) + 4>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetContiguousL2Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 2) + 4>(GetInteger(addr)); }
|
||||||
static constexpr ALWAYS_INLINE uintptr_t GetContiguousL3Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 3) + 4>(GetInteger(addr)); }
|
static constexpr ALWAYS_INLINE uintptr_t GetContiguousL3Offset(KProcessAddress addr) { return GetBits<0, PageBits + LevelBits * (NumLevels - 3) + 4>(GetInteger(addr)); }
|
||||||
|
|
||||||
|
static constexpr ALWAYS_INLINE uintptr_t GetBlock(const PageTableEntry *pte, EntryLevel level) { return SelectBits(pte->GetRawAttributesUnsafe(), PageBits + LevelBits * level, LevelBits * (NumLevels + 1 - level)); }
|
||||||
|
static constexpr ALWAYS_INLINE uintptr_t GetOffset(KProcessAddress addr, EntryLevel level) { return GetBits(GetInteger(addr), 0, PageBits + LevelBits * level); }
|
||||||
|
|
||||||
static ALWAYS_INLINE KVirtualAddress GetPageTableVirtualAddress(KPhysicalAddress addr) {
|
static ALWAYS_INLINE KVirtualAddress GetPageTableVirtualAddress(KPhysicalAddress addr) {
|
||||||
return KMemoryLayout::GetLinearVirtualAddress(addr);
|
return KMemoryLayout::GetLinearVirtualAddress(addr);
|
||||||
}
|
}
|
||||||
|
|
||||||
ALWAYS_INLINE bool ExtractL1Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L1PageTableEntry *l1_entry, KProcessAddress virt_addr) const;
|
//ALWAYS_INLINE bool ExtractL1Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L1PageTableEntry *l1_entry, KProcessAddress virt_addr) const;
|
||||||
ALWAYS_INLINE bool ExtractL2Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L2PageTableEntry *l2_entry, KProcessAddress virt_addr) const;
|
//ALWAYS_INLINE bool ExtractL2Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L2PageTableEntry *l2_entry, KProcessAddress virt_addr) const;
|
||||||
ALWAYS_INLINE bool ExtractL3Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L3PageTableEntry *l3_entry, KProcessAddress virt_addr) const;
|
//ALWAYS_INLINE bool ExtractL3Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L3PageTableEntry *l3_entry, KProcessAddress virt_addr) const;
|
||||||
private:
|
private:
|
||||||
L1PageTableEntry *m_table;
|
L1PageTableEntry *m_table;
|
||||||
bool m_is_kernel;
|
bool m_is_kernel;
|
||||||
|
|
|
@ -33,103 +33,98 @@ namespace ams::kern::arch::arm64 {
|
||||||
return m_table;
|
return m_table;
|
||||||
}
|
}
|
||||||
|
|
||||||
bool KPageTableImpl::ExtractL3Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L3PageTableEntry *l3_entry, KProcessAddress virt_addr) const {
|
// bool KPageTableImpl::ExtractL3Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L3PageTableEntry *l3_entry, KProcessAddress virt_addr) const {
|
||||||
/* Set the L3 entry. */
|
// /* Set the L3 entry. */
|
||||||
out_context->l3_entry = l3_entry;
|
// out_context->l3_entry = l3_entry;
|
||||||
|
//
|
||||||
if (l3_entry->IsBlock()) {
|
// if (l3_entry->IsBlock()) {
|
||||||
/* Set the output entry. */
|
// /* Set the output entry. */
|
||||||
out_entry->phys_addr = l3_entry->GetBlock() + (virt_addr & (L3BlockSize - 1));
|
// out_entry->phys_addr = l3_entry->GetBlock() + (virt_addr & (L3BlockSize - 1));
|
||||||
if (l3_entry->IsContiguous()) {
|
// if (l3_entry->IsContiguous()) {
|
||||||
out_entry->block_size = L3ContiguousBlockSize;
|
// out_entry->block_size = L3ContiguousBlockSize;
|
||||||
} else {
|
// } else {
|
||||||
out_entry->block_size = L3BlockSize;
|
// out_entry->block_size = L3BlockSize;
|
||||||
}
|
// }
|
||||||
out_entry->sw_reserved_bits = l3_entry->GetSoftwareReservedBits();
|
// out_entry->sw_reserved_bits = l3_entry->GetSoftwareReservedBits();
|
||||||
out_entry->attr = 0;
|
// out_entry->attr = 0;
|
||||||
|
//
|
||||||
return true;
|
// return true;
|
||||||
} else {
|
// } else {
|
||||||
out_entry->phys_addr = Null<KPhysicalAddress>;
|
// out_entry->phys_addr = Null<KPhysicalAddress>;
|
||||||
out_entry->block_size = L3BlockSize;
|
// out_entry->block_size = L3BlockSize;
|
||||||
out_entry->sw_reserved_bits = 0;
|
// out_entry->sw_reserved_bits = 0;
|
||||||
out_entry->attr = 0;
|
// out_entry->attr = 0;
|
||||||
return false;
|
// return false;
|
||||||
}
|
// }
|
||||||
}
|
// }
|
||||||
|
//
|
||||||
bool KPageTableImpl::ExtractL2Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L2PageTableEntry *l2_entry, KProcessAddress virt_addr) const {
|
// bool KPageTableImpl::ExtractL2Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L2PageTableEntry *l2_entry, KProcessAddress virt_addr) const {
|
||||||
/* Set the L2 entry. */
|
// /* Set the L2 entry. */
|
||||||
out_context->l2_entry = l2_entry;
|
// out_context->l2_entry = l2_entry;
|
||||||
|
//
|
||||||
if (l2_entry->IsBlock()) {
|
// if (l2_entry->IsBlock()) {
|
||||||
/* Set the output entry. */
|
// /* Set the output entry. */
|
||||||
out_entry->phys_addr = l2_entry->GetBlock() + (virt_addr & (L2BlockSize - 1));
|
// out_entry->phys_addr = l2_entry->GetBlock() + (virt_addr & (L2BlockSize - 1));
|
||||||
if (l2_entry->IsContiguous()) {
|
// if (l2_entry->IsContiguous()) {
|
||||||
out_entry->block_size = L2ContiguousBlockSize;
|
// out_entry->block_size = L2ContiguousBlockSize;
|
||||||
} else {
|
// } else {
|
||||||
out_entry->block_size = L2BlockSize;
|
// out_entry->block_size = L2BlockSize;
|
||||||
}
|
// }
|
||||||
out_entry->sw_reserved_bits = l2_entry->GetSoftwareReservedBits();
|
// out_entry->sw_reserved_bits = l2_entry->GetSoftwareReservedBits();
|
||||||
out_entry->attr = 0;
|
// out_entry->attr = 0;
|
||||||
|
//
|
||||||
/* Set the output context. */
|
// /* Set the output context. */
|
||||||
out_context->l3_entry = nullptr;
|
// out_context->l3_entry = nullptr;
|
||||||
return true;
|
// return true;
|
||||||
} else if (l2_entry->IsTable()) {
|
// } else if (l2_entry->IsTable()) {
|
||||||
return this->ExtractL3Entry(out_entry, out_context, this->GetL3EntryFromTable(GetPageTableVirtualAddress(l2_entry->GetTable()), virt_addr), virt_addr);
|
// return this->ExtractL3Entry(out_entry, out_context, this->GetL3EntryFromTable(GetPageTableVirtualAddress(l2_entry->GetTable()), virt_addr), virt_addr);
|
||||||
} else {
|
// } else {
|
||||||
out_entry->phys_addr = Null<KPhysicalAddress>;
|
// out_entry->phys_addr = Null<KPhysicalAddress>;
|
||||||
out_entry->block_size = L2BlockSize;
|
// out_entry->block_size = L2BlockSize;
|
||||||
out_entry->sw_reserved_bits = 0;
|
// out_entry->sw_reserved_bits = 0;
|
||||||
out_entry->attr = 0;
|
// out_entry->attr = 0;
|
||||||
|
//
|
||||||
out_context->l3_entry = nullptr;
|
// out_context->l3_entry = nullptr;
|
||||||
return false;
|
// return false;
|
||||||
}
|
// }
|
||||||
}
|
// }
|
||||||
|
//
|
||||||
bool KPageTableImpl::ExtractL1Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L1PageTableEntry *l1_entry, KProcessAddress virt_addr) const {
|
// bool KPageTableImpl::ExtractL1Entry(TraversalEntry *out_entry, TraversalContext *out_context, const L1PageTableEntry *l1_entry, KProcessAddress virt_addr) const {
|
||||||
/* Set the L1 entry. */
|
// /* Set the L1 entry. */
|
||||||
out_context->l1_entry = l1_entry;
|
// out_context->level_entries[EntryLevel_L1] = l1_entry;
|
||||||
|
//
|
||||||
if (l1_entry->IsBlock()) {
|
// if (l1_entry->IsBlock()) {
|
||||||
/* Set the output entry. */
|
// /* Set the output entry. */
|
||||||
out_entry->phys_addr = l1_entry->GetBlock() + (virt_addr & (L1BlockSize - 1));
|
// out_entry->phys_addr = l1_entry->GetBlock() + (virt_addr & (L1BlockSize - 1));
|
||||||
if (l1_entry->IsContiguous()) {
|
// if (l1_entry->IsContiguous()) {
|
||||||
out_entry->block_size = L1ContiguousBlockSize;
|
// out_entry->block_size = L1ContiguousBlockSize;
|
||||||
} else {
|
// } else {
|
||||||
out_entry->block_size = L1BlockSize;
|
// out_entry->block_size = L1BlockSize;
|
||||||
}
|
// }
|
||||||
out_entry->sw_reserved_bits = l1_entry->GetSoftwareReservedBits();
|
// out_entry->sw_reserved_bits = l1_entry->GetSoftwareReservedBits();
|
||||||
|
//
|
||||||
/* Set the output context. */
|
// /* Set the output context. */
|
||||||
out_context->l2_entry = nullptr;
|
// out_context->l2_entry = nullptr;
|
||||||
out_context->l3_entry = nullptr;
|
// out_context->l3_entry = nullptr;
|
||||||
return true;
|
// return true;
|
||||||
} else if (l1_entry->IsTable()) {
|
// } else if (l1_entry->IsTable()) {
|
||||||
return this->ExtractL2Entry(out_entry, out_context, this->GetL2EntryFromTable(GetPageTableVirtualAddress(l1_entry->GetTable()), virt_addr), virt_addr);
|
// return this->ExtractL2Entry(out_entry, out_context, this->GetL2EntryFromTable(GetPageTableVirtualAddress(l1_entry->GetTable()), virt_addr), virt_addr);
|
||||||
} else {
|
// } else {
|
||||||
out_entry->phys_addr = Null<KPhysicalAddress>;
|
// out_entry->phys_addr = Null<KPhysicalAddress>;
|
||||||
out_entry->block_size = L1BlockSize;
|
// out_entry->block_size = L1BlockSize;
|
||||||
out_entry->sw_reserved_bits = 0;
|
// out_entry->sw_reserved_bits = 0;
|
||||||
out_entry->attr = 0;
|
// out_entry->attr = 0;
|
||||||
|
//
|
||||||
out_context->l2_entry = nullptr;
|
// out_context->l2_entry = nullptr;
|
||||||
out_context->l3_entry = nullptr;
|
// out_context->l3_entry = nullptr;
|
||||||
return false;
|
// return false;
|
||||||
}
|
// }
|
||||||
}
|
// }
|
||||||
|
|
||||||
bool KPageTableImpl::BeginTraversal(TraversalEntry *out_entry, TraversalContext *out_context, KProcessAddress address) const {
|
bool KPageTableImpl::BeginTraversal(TraversalEntry *out_entry, TraversalContext *out_context, KProcessAddress address) const {
|
||||||
/* Setup invalid defaults. */
|
/* Setup invalid defaults. */
|
||||||
out_entry->phys_addr = Null<KPhysicalAddress>;
|
*out_entry = {};
|
||||||
out_entry->block_size = L1BlockSize;
|
*out_context = {};
|
||||||
out_entry->sw_reserved_bits = 0;
|
|
||||||
out_entry->attr = 0;
|
|
||||||
out_context->l1_entry = m_table + m_num_entries;
|
|
||||||
out_context->l2_entry = nullptr;
|
|
||||||
out_context->l3_entry = nullptr;
|
|
||||||
|
|
||||||
/* Validate that we can read the actual entry. */
|
/* Validate that we can read the actual entry. */
|
||||||
const size_t l0_index = GetL0Index(address);
|
const size_t l0_index = GetL0Index(address);
|
||||||
|
@ -146,125 +141,79 @@ namespace ams::kern::arch::arm64 {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Extract the entry. */
|
/* Get the L1 entry, and check if it's a table. */
|
||||||
const bool valid = this->ExtractL1Entry(out_entry, out_context, this->GetL1Entry(address), address);
|
out_context->level_entries[EntryLevel_L1] = this->GetL1Entry(address);
|
||||||
|
if (out_context->level_entries[EntryLevel_L1]->IsMappedTable()) {
|
||||||
|
/* Get the L2 entry, and check if it's a table. */
|
||||||
|
out_context->level_entries[EntryLevel_L2] = this->GetL2EntryFromTable(GetPageTableVirtualAddress(out_context->level_entries[EntryLevel_L1]->GetTable()), address);
|
||||||
|
if (out_context->level_entries[EntryLevel_L2]->IsMappedTable()) {
|
||||||
|
/* Get the L3 entry. */
|
||||||
|
out_context->level_entries[EntryLevel_L3] = this->GetL3EntryFromTable(GetPageTableVirtualAddress(out_context->level_entries[EntryLevel_L2]->GetTable()), address);
|
||||||
|
|
||||||
/* Update the context for next traversal. */
|
/* It's either a page or not. */
|
||||||
switch (out_entry->block_size) {
|
out_context->level = EntryLevel_L3;
|
||||||
case L1ContiguousBlockSize:
|
} else {
|
||||||
out_context->l1_entry += (L1ContiguousBlockSize / L1BlockSize) - GetContiguousL1Offset(address) / L1BlockSize;
|
/* Not a L2 table, so possibly an L2 block. */
|
||||||
break;
|
out_context->level = EntryLevel_L2;
|
||||||
case L1BlockSize:
|
}
|
||||||
out_context->l1_entry += 1;
|
} else {
|
||||||
break;
|
/* Not a L1 table, so possibly an L1 block. */
|
||||||
case L2ContiguousBlockSize:
|
out_context->level = EntryLevel_L1;
|
||||||
out_context->l1_entry += 1;
|
|
||||||
out_context->l2_entry += (L2ContiguousBlockSize / L2BlockSize) - GetContiguousL2Offset(address) / L2BlockSize;
|
|
||||||
break;
|
|
||||||
case L2BlockSize:
|
|
||||||
out_context->l1_entry += 1;
|
|
||||||
out_context->l2_entry += 1;
|
|
||||||
break;
|
|
||||||
case L3ContiguousBlockSize:
|
|
||||||
out_context->l1_entry += 1;
|
|
||||||
out_context->l2_entry += 1;
|
|
||||||
out_context->l3_entry += (L3ContiguousBlockSize / L3BlockSize) - GetContiguousL3Offset(address) / L3BlockSize;
|
|
||||||
break;
|
|
||||||
case L3BlockSize:
|
|
||||||
out_context->l1_entry += 1;
|
|
||||||
out_context->l2_entry += 1;
|
|
||||||
out_context->l3_entry += 1;
|
|
||||||
break;
|
|
||||||
MESOSPHERE_UNREACHABLE_DEFAULT_CASE();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return valid;
|
/* Determine other fields. */
|
||||||
|
const auto *pte = out_context->level_entries[out_context->level];
|
||||||
|
|
||||||
|
out_context->is_contiguous = pte->IsContiguous();
|
||||||
|
|
||||||
|
out_entry->sw_reserved_bits = pte->GetSoftwareReservedBits();
|
||||||
|
out_entry->attr = 0;
|
||||||
|
out_entry->phys_addr = this->GetBlock(pte, out_context->level) + this->GetOffset(address, out_context->level);
|
||||||
|
out_entry->block_size = static_cast<size_t>(1) << (PageBits + LevelBits * out_context->level + 4 * out_context->is_contiguous);
|
||||||
|
|
||||||
|
return out_context->level == EntryLevel_L3 ? pte->IsPage() : pte->IsBlock();
|
||||||
}
|
}
|
||||||
|
|
||||||
bool KPageTableImpl::ContinueTraversal(TraversalEntry *out_entry, TraversalContext *context) const {
|
bool KPageTableImpl::ContinueTraversal(TraversalEntry *out_entry, TraversalContext *context) const {
|
||||||
bool valid = false;
|
/* Advance entry. */
|
||||||
|
|
||||||
/* Check if we're not at the end of an L3 table. */
|
auto *cur_pte = context->level_entries[context->level];
|
||||||
if (!util::IsAligned(reinterpret_cast<uintptr_t>(context->l3_entry), PageSize)) {
|
auto *next_pte = reinterpret_cast<PageTableEntry *>(context->is_contiguous ? util::AlignDown(reinterpret_cast<uintptr_t>(cur_pte), 0x10 * sizeof(PageTableEntry)) + 0x10 * sizeof(PageTableEntry) : reinterpret_cast<uintptr_t>(cur_pte) + sizeof(PageTableEntry));
|
||||||
valid = this->ExtractL3Entry(out_entry, context, context->l3_entry, Null<KProcessAddress>);
|
|
||||||
|
|
||||||
switch (out_entry->block_size) {
|
/* Set the pte. */
|
||||||
case L3ContiguousBlockSize:
|
context->level_entries[context->level] = next_pte;
|
||||||
context->l3_entry += (L3ContiguousBlockSize / L3BlockSize);
|
|
||||||
break;
|
|
||||||
case L3BlockSize:
|
|
||||||
context->l3_entry += 1;
|
|
||||||
break;
|
|
||||||
MESOSPHERE_UNREACHABLE_DEFAULT_CASE();
|
|
||||||
}
|
|
||||||
} else if (!util::IsAligned(reinterpret_cast<uintptr_t>(context->l2_entry), PageSize)) {
|
|
||||||
/* We're not at the end of an L2 table. */
|
|
||||||
valid = this->ExtractL2Entry(out_entry, context, context->l2_entry, Null<KProcessAddress>);
|
|
||||||
|
|
||||||
switch (out_entry->block_size) {
|
/* Advance appropriately. */
|
||||||
case L2ContiguousBlockSize:
|
while (context->level < EntryLevel_L1 && util::IsAligned(reinterpret_cast<uintptr_t>(context->level_entries[context->level]), PageSize)) {
|
||||||
context->l2_entry += (L2ContiguousBlockSize / L2BlockSize);
|
/* Advance the above table by one entry. */
|
||||||
break;
|
context->level_entries[context->level + 1]++;
|
||||||
case L2BlockSize:
|
context->level = static_cast<EntryLevel>(util::ToUnderlying(context->level) + 1);
|
||||||
context->l2_entry += 1;
|
}
|
||||||
break;
|
|
||||||
case L3ContiguousBlockSize:
|
/* Check if we've hit the end of the L1 table. */
|
||||||
context->l2_entry += 1;
|
if (context->level == EntryLevel_L1) {
|
||||||
context->l3_entry += (L3ContiguousBlockSize / L3BlockSize);
|
if (context->level_entries[EntryLevel_L1] - static_cast<const PageTableEntry *>(m_table) >= m_num_entries) {
|
||||||
break;
|
*context = {};
|
||||||
case L3BlockSize:
|
*out_entry = {};
|
||||||
context->l2_entry += 1;
|
|
||||||
context->l3_entry += 1;
|
|
||||||
break;
|
|
||||||
MESOSPHERE_UNREACHABLE_DEFAULT_CASE();
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
/* We need to update the l1 entry. */
|
|
||||||
const size_t l1_index = context->l1_entry - m_table;
|
|
||||||
if (l1_index < m_num_entries) {
|
|
||||||
valid = this->ExtractL1Entry(out_entry, context, context->l1_entry, Null<KProcessAddress>);
|
|
||||||
} else {
|
|
||||||
/* Invalid, end traversal. */
|
|
||||||
out_entry->phys_addr = Null<KPhysicalAddress>;
|
|
||||||
out_entry->block_size = L1BlockSize;
|
|
||||||
out_entry->sw_reserved_bits = 0;
|
|
||||||
out_entry->attr = 0;
|
|
||||||
context->l1_entry = m_table + m_num_entries;
|
|
||||||
context->l2_entry = nullptr;
|
|
||||||
context->l3_entry = nullptr;
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
switch (out_entry->block_size) {
|
|
||||||
case L1ContiguousBlockSize:
|
|
||||||
context->l1_entry += (L1ContiguousBlockSize / L1BlockSize);
|
|
||||||
break;
|
|
||||||
case L1BlockSize:
|
|
||||||
context->l1_entry += 1;
|
|
||||||
break;
|
|
||||||
case L2ContiguousBlockSize:
|
|
||||||
context->l1_entry += 1;
|
|
||||||
context->l2_entry += (L2ContiguousBlockSize / L2BlockSize);
|
|
||||||
break;
|
|
||||||
case L2BlockSize:
|
|
||||||
context->l1_entry += 1;
|
|
||||||
context->l2_entry += 1;
|
|
||||||
break;
|
|
||||||
case L3ContiguousBlockSize:
|
|
||||||
context->l1_entry += 1;
|
|
||||||
context->l2_entry += 1;
|
|
||||||
context->l3_entry += (L3ContiguousBlockSize / L3BlockSize);
|
|
||||||
break;
|
|
||||||
case L3BlockSize:
|
|
||||||
context->l1_entry += 1;
|
|
||||||
context->l2_entry += 1;
|
|
||||||
context->l3_entry += 1;
|
|
||||||
break;
|
|
||||||
MESOSPHERE_UNREACHABLE_DEFAULT_CASE();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return valid;
|
/* We may have advanced to a new table, and if we have we should descend. */
|
||||||
|
while (context->level > EntryLevel_L3 && context->level_entries[context->level]->IsMappedTable()) {
|
||||||
|
context->level_entries[context->level - 1] = GetPointer<PageTableEntry>(GetPageTableVirtualAddress(context->level_entries[context->level]->GetTable()));
|
||||||
|
context->level = static_cast<EntryLevel>(util::ToUnderlying(context->level) - 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
const auto *pte = context->level_entries[context->level];
|
||||||
|
|
||||||
|
context->is_contiguous = pte->IsContiguous();
|
||||||
|
|
||||||
|
out_entry->sw_reserved_bits = pte->GetSoftwareReservedBits();
|
||||||
|
out_entry->attr = 0;
|
||||||
|
out_entry->phys_addr = this->GetBlock(pte, context->level);
|
||||||
|
out_entry->block_size = static_cast<size_t>(1) << (PageBits + LevelBits * context->level + 4 * context->is_contiguous);
|
||||||
|
return context->level == EntryLevel_L3 ? pte->IsPage() : pte->IsBlock();
|
||||||
}
|
}
|
||||||
|
|
||||||
bool KPageTableImpl::GetPhysicalAddress(KPhysicalAddress *out, KProcessAddress address) const {
|
bool KPageTableImpl::GetPhysicalAddress(KPhysicalAddress *out, KProcessAddress address) const {
|
||||||
|
@ -283,32 +232,27 @@ namespace ams::kern::arch::arm64 {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Try to get from l1 table. */
|
/* Get the L1 entry, and check if it's a table. */
|
||||||
const L1PageTableEntry *l1_entry = this->GetL1Entry(address);
|
const PageTableEntry *pte = this->GetL1Entry(address);
|
||||||
if (l1_entry->IsBlock()) {
|
EntryLevel level = EntryLevel_L1;
|
||||||
*out = l1_entry->GetBlock() + GetL1Offset(address);
|
if (pte->IsMappedTable()) {
|
||||||
return true;
|
/* Get the L2 entry, and check if it's a table. */
|
||||||
} else if (!l1_entry->IsTable()) {
|
pte = this->GetL2EntryFromTable(GetPageTableVirtualAddress(pte->GetTable()), address);
|
||||||
return false;
|
level = EntryLevel_L2;
|
||||||
|
if (pte->IsMappedTable()) {
|
||||||
|
pte = this->GetL3EntryFromTable(GetPageTableVirtualAddress(pte->GetTable()), address);
|
||||||
|
level = EntryLevel_L3;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Try to get from l2 table. */
|
const bool is_block = level == EntryLevel_L3 ? pte->IsPage() : pte->IsBlock();
|
||||||
const L2PageTableEntry *l2_entry = this->GetL2Entry(l1_entry, address);
|
if (is_block) {
|
||||||
if (l2_entry->IsBlock()) {
|
*out = this->GetBlock(pte, level) + this->GetOffset(address, level);
|
||||||
*out = l2_entry->GetBlock() + GetL2Offset(address);
|
} else {
|
||||||
return true;
|
*out = Null<KPhysicalAddress>;
|
||||||
} else if (!l2_entry->IsTable()) {
|
|
||||||
return false;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Try to get from l3 table. */
|
return is_block;
|
||||||
const L3PageTableEntry *l3_entry = this->GetL3Entry(l2_entry, address);
|
|
||||||
if (l3_entry->IsBlock()) {
|
|
||||||
*out = l3_entry->GetBlock() + GetL3Offset(address);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
return false;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void KPageTableImpl::Dump(uintptr_t start, size_t size) const {
|
void KPageTableImpl::Dump(uintptr_t start, size_t size) const {
|
||||||
|
|
Loading…
Reference in a new issue