diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_auto_object.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_auto_object.hpp index cdd385046..f798ca806 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_auto_object.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_auto_object.hpp @@ -85,7 +85,7 @@ namespace ams::kern { virtual KProcess *GetOwner() const { return nullptr; } u32 GetReferenceCount() const { - return m_ref_count; + return m_ref_count.load(); } ALWAYS_INLINE bool IsDerivedFrom(const TypeObj &rhs) const { diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_dynamic_slab_heap.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_dynamic_slab_heap.hpp index 8d4e05e42..c7e0ad45c 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_dynamic_slab_heap.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_dynamic_slab_heap.hpp @@ -49,9 +49,9 @@ namespace ams::kern { constexpr KVirtualAddress GetAddress() const { return m_address; } constexpr size_t GetSize() const { return m_size; } - constexpr size_t GetUsed() const { return m_used; } - constexpr size_t GetPeak() const { return m_peak; } - constexpr size_t GetCount() const { return m_count; } + constexpr size_t GetUsed() const { return m_used.load(); } + constexpr size_t GetPeak() const { return m_peak.load(); } + constexpr size_t GetCount() const { return m_count.load(); } constexpr bool IsInRange(KVirtualAddress addr) const { return this->GetAddress() <= addr && addr <= this->GetAddress() + this->GetSize() - 1; @@ -65,7 +65,7 @@ namespace ams::kern { /* Free blocks to memory. */ u8 *cur = GetPointer(m_address + m_size); - for (size_t i = 0; i < m_count; i++) { + for (size_t i = 0; i < sz / sizeof(T); i++) { cur -= sizeof(T); this->GetImpl()->Free(cur); } @@ -84,13 +84,13 @@ namespace ams::kern { this->Initialize(page_allocator); /* Allocate until we have the correct number of objects. */ - while (m_count < num_objects) { + while (m_count.load() < num_objects) { auto *allocated = reinterpret_cast(m_page_allocator->Allocate()); MESOSPHERE_ABORT_UNLESS(allocated != nullptr); for (size_t i = 0; i < sizeof(PageBuffer) / sizeof(T); i++) { this->GetImpl()->Free(allocated + i); } - m_count += sizeof(PageBuffer) / sizeof(T); + m_count.fetch_add(sizeof(PageBuffer) / sizeof(T)); } } @@ -106,7 +106,7 @@ namespace ams::kern { for (size_t i = 1; i < sizeof(PageBuffer) / sizeof(T); i++) { this->GetImpl()->Free(allocated + i); } - m_count += sizeof(PageBuffer) / sizeof(T); + m_count.fetch_add(sizeof(PageBuffer) / sizeof(T)); } } } @@ -116,8 +116,8 @@ namespace ams::kern { new (allocated) T(); /* Update our tracking. */ - size_t used = ++m_used; - size_t peak = m_peak; + size_t used = m_used.fetch_add(1) + 1; + size_t peak = m_peak.load(); while (peak < used) { if (m_peak.compare_exchange_weak(peak, used, std::memory_order_relaxed)) { break; @@ -130,7 +130,7 @@ namespace ams::kern { void Free(T *t) { this->GetImpl()->Free(t); - --m_used; + m_used.fetch_sub(1); } }; diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_handle_table.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_handle_table.hpp index 5e5863116..d62b2c85e 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_handle_table.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_handle_table.hpp @@ -303,6 +303,7 @@ namespace ams::kern { const auto linear_id = handle_pack.Get(); const auto reserved = handle_pack.Get(); MESOSPHERE_ASSERT(reserved == 0); + MESOSPHERE_UNUSED(reserved); /* Validate our indexing information. */ if (raw_value == 0) { diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_light_lock.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_light_lock.hpp index c25c963a2..2f1734cdb 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_light_lock.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_light_lock.hpp @@ -50,7 +50,7 @@ namespace ams::kern { } } - void Unlock() { + ALWAYS_INLINE void Unlock() { MESOSPHERE_ASSERT_THIS(); const uintptr_t cur_thread = reinterpret_cast(GetCurrentThreadPointer()); @@ -65,8 +65,8 @@ namespace ams::kern { void LockSlowPath(uintptr_t owner, uintptr_t cur_thread); void UnlockSlowPath(uintptr_t cur_thread); - bool IsLocked() const { return m_tag != 0; } - bool IsLockedByCurrentThread() const { return (m_tag | 0x1ul) == (reinterpret_cast(GetCurrentThreadPointer()) | 0x1ul); } + ALWAYS_INLINE bool IsLocked() const { return m_tag.load() != 0; } + ALWAYS_INLINE bool IsLockedByCurrentThread() const { return (m_tag.load() | 0x1ul) == (reinterpret_cast(GetCurrentThreadPointer()) | 0x1ul); } }; using KScopedLightLock = KScopedLock; diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_page_table_base.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_page_table_base.hpp index 9192e5c77..2eafc4c49 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_page_table_base.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_page_table_base.hpp @@ -203,54 +203,54 @@ namespace ams::kern { virtual Result Operate(PageLinkedList *page_list, KProcessAddress virt_addr, size_t num_pages, const KPageGroup &page_group, const KPageProperties properties, OperationType operation, bool reuse_ll) = 0; virtual void FinalizeUpdate(PageLinkedList *page_list) = 0; - KPageTableImpl &GetImpl() { return m_impl; } - const KPageTableImpl &GetImpl() const { return m_impl; } + ALWAYS_INLINE KPageTableImpl &GetImpl() { return m_impl; } + ALWAYS_INLINE const KPageTableImpl &GetImpl() const { return m_impl; } - bool IsLockedByCurrentThread() const { return m_general_lock.IsLockedByCurrentThread(); } + ALWAYS_INLINE bool IsLockedByCurrentThread() const { return m_general_lock.IsLockedByCurrentThread(); } - bool IsLinearMappedPhysicalAddress(KPhysicalAddress phys_addr) { + ALWAYS_INLINE bool IsLinearMappedPhysicalAddress(KPhysicalAddress phys_addr) { MESOSPHERE_ASSERT(this->IsLockedByCurrentThread()); return KMemoryLayout::IsLinearMappedPhysicalAddress(m_cached_physical_linear_region, phys_addr); } - bool IsLinearMappedPhysicalAddress(KPhysicalAddress phys_addr, size_t size) { + ALWAYS_INLINE bool IsLinearMappedPhysicalAddress(KPhysicalAddress phys_addr, size_t size) { MESOSPHERE_ASSERT(this->IsLockedByCurrentThread()); return KMemoryLayout::IsLinearMappedPhysicalAddress(m_cached_physical_linear_region, phys_addr, size); } - bool IsHeapPhysicalAddress(KPhysicalAddress phys_addr) { + ALWAYS_INLINE bool IsHeapPhysicalAddress(KPhysicalAddress phys_addr) { MESOSPHERE_ASSERT(this->IsLockedByCurrentThread()); return KMemoryLayout::IsHeapPhysicalAddress(m_cached_physical_heap_region, phys_addr); } - bool IsHeapPhysicalAddress(KPhysicalAddress phys_addr, size_t size) { + ALWAYS_INLINE bool IsHeapPhysicalAddress(KPhysicalAddress phys_addr, size_t size) { MESOSPHERE_ASSERT(this->IsLockedByCurrentThread()); return KMemoryLayout::IsHeapPhysicalAddress(m_cached_physical_heap_region, phys_addr, size); } - bool IsHeapPhysicalAddressForFinalize(KPhysicalAddress phys_addr) { + ALWAYS_INLINE bool IsHeapPhysicalAddressForFinalize(KPhysicalAddress phys_addr) { MESOSPHERE_ASSERT(!this->IsLockedByCurrentThread()); return KMemoryLayout::IsHeapPhysicalAddress(m_cached_physical_heap_region, phys_addr); } - bool IsHeapVirtualAddress(KVirtualAddress virt_addr) { + ALWAYS_INLINE bool IsHeapVirtualAddress(KVirtualAddress virt_addr) { MESOSPHERE_ASSERT(this->IsLockedByCurrentThread()); return KMemoryLayout::IsHeapVirtualAddress(m_cached_virtual_heap_region, virt_addr); } - bool IsHeapVirtualAddress(KVirtualAddress virt_addr, size_t size) { + ALWAYS_INLINE bool IsHeapVirtualAddress(KVirtualAddress virt_addr, size_t size) { MESOSPHERE_ASSERT(this->IsLockedByCurrentThread()); return KMemoryLayout::IsHeapVirtualAddress(m_cached_virtual_heap_region, virt_addr, size); } - bool ContainsPages(KProcessAddress addr, size_t num_pages) const { + ALWAYS_INLINE bool ContainsPages(KProcessAddress addr, size_t num_pages) const { return (m_address_space_start <= addr) && (num_pages <= (m_address_space_end - m_address_space_start) / PageSize) && (addr + num_pages * PageSize - 1 <= m_address_space_end - 1); } private: diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_process.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_process.hpp index fce982912..9e5dd82f7 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_process.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_process.hpp @@ -135,6 +135,7 @@ namespace ams::kern { } void UnpinThread(s32 core_id, KThread *thread) { + MESOSPHERE_UNUSED(thread); MESOSPHERE_ASSERT(0 <= core_id && core_id < static_cast(cpu::NumCores)); MESOSPHERE_ASSERT(thread != nullptr); MESOSPHERE_ASSERT(m_pinned_threads[core_id] == thread); diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp index 02b267ac2..5bcbab66f 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_scheduler.hpp @@ -38,7 +38,7 @@ namespace ams::kern { static_assert(ams::svc::HighestThreadPriority <= HighestCoreMigrationAllowedPriority); struct SchedulingState { - std::atomic needs_scheduling; + std::atomic needs_scheduling; bool interrupt_task_thread_runnable; bool should_count_idle; u64 idle_count; @@ -181,7 +181,7 @@ namespace ams::kern { KScopedInterruptDisable intr_disable; ON_SCOPE_EXIT { GetCurrentThread().EnableDispatch(); }; - if (m_state.needs_scheduling) { + if (m_state.needs_scheduling.load()) { Schedule(); } } diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_shared_memory_info.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_shared_memory_info.hpp index 26cd1f3bf..0330d4914 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_shared_memory_info.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_shared_memory_info.hpp @@ -36,8 +36,8 @@ namespace ams::kern { } constexpr void Open() { - const size_t ref_count = ++m_reference_count; - MESOSPHERE_ASSERT(ref_count > 0); + ++m_reference_count; + MESOSPHERE_ASSERT(m_reference_count > 0); } constexpr bool Close() { diff --git a/libraries/libmesosphere/include/mesosphere/kern_k_thread.hpp b/libraries/libmesosphere/include/mesosphere/kern_k_thread.hpp index b985c6537..184d55f37 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_k_thread.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_k_thread.hpp @@ -207,7 +207,7 @@ namespace ams::kern { s32 m_original_physical_ideal_core_id{}; s32 m_num_core_migration_disables{}; ThreadState m_thread_state{}; - std::atomic m_termination_requested{}; + std::atomic m_termination_requested{}; bool m_wait_cancelled{}; bool m_cancellable{}; bool m_signaled{}; @@ -486,7 +486,7 @@ namespace ams::kern { MESOSPHERE_UNUSED(core_id); } - s64 GetCpuTime() const { return m_cpu_time; } + s64 GetCpuTime() const { return m_cpu_time.load(); } s64 GetCpuTime(s32 core_id) const { MESOSPHERE_ABORT_UNLESS(0 <= core_id && core_id < static_cast(cpu::NumCores)); @@ -530,7 +530,7 @@ namespace ams::kern { ALWAYS_INLINE void *GetKernelStackTop() const { return m_kernel_stack_top; } ALWAYS_INLINE bool IsTerminationRequested() const { - return m_termination_requested || this->GetRawState() == ThreadState_Terminated; + return m_termination_requested.load() || this->GetRawState() == ThreadState_Terminated; } size_t GetKernelStackUsage() const; diff --git a/libraries/libmesosphere/include/mesosphere/kern_panic.hpp b/libraries/libmesosphere/include/mesosphere/kern_panic.hpp index 9382de11e..4a99ba0d5 100644 --- a/libraries/libmesosphere/include/mesosphere/kern_panic.hpp +++ b/libraries/libmesosphere/include/mesosphere/kern_panic.hpp @@ -40,8 +40,10 @@ namespace ams::kern { MESOSPHERE_PANIC(__VA_ARGS__); \ } \ }) -#else +#elif defined(MESOSPHERE_PRESERVE_ASSERTION_EXPRESSIONS) #define MESOSPHERE_ASSERT_IMPL(expr, ...) do { static_cast(expr); } while (0) +#else +#define MESOSPHERE_ASSERT_IMPL(expr, ...) static_cast(0) #endif #define MESOSPHERE_ASSERT(expr) MESOSPHERE_ASSERT_IMPL(expr, "Assertion failed: %s\n", #expr) @@ -56,8 +58,10 @@ namespace ams::kern { #ifdef MESOSPHERE_BUILD_FOR_AUDITING #define MESOSPHERE_AUDIT(expr) MESOSPHERE_ASSERT(expr) -#else +#elif defined(MESOSPHERE_PRESERVE_AUDIT_EXPRESSIONS) #define MESOSPHERE_AUDIT(expr) do { static_cast(expr); } while (0) +#else +#define MESOSPHERE_AUDIT(expr) static_cast(0) #endif #define MESOSPHERE_TODO(arg) ({ constexpr const char *__mesosphere_todo = arg; static_cast(__mesosphere_todo); MESOSPHERE_PANIC("TODO (%s): %s\n", __PRETTY_FUNCTION__, __mesosphere_todo); }) diff --git a/libraries/libmesosphere/source/arch/arm64/kern_cpu.cpp b/libraries/libmesosphere/source/arch/arm64/kern_cpu.cpp index 8b31ccb44..e46c6285d 100644 --- a/libraries/libmesosphere/source/arch/arm64/kern_cpu.cpp +++ b/libraries/libmesosphere/source/arch/arm64/kern_cpu.cpp @@ -109,7 +109,7 @@ namespace ams::kern::arch::arm64::cpu { /* Wait for a request to come in. */ { KScopedLightLock lk(m_cv_lock); - while ((m_target_cores & (1ul << core_id)) == 0) { + while ((m_target_cores.load() & (1ul << core_id)) == 0) { m_cv.Wait(std::addressof(m_cv_lock)); } } @@ -120,7 +120,7 @@ namespace ams::kern::arch::arm64::cpu { /* Broadcast, if there's nothing pending. */ { KScopedLightLock lk(m_cv_lock); - if (m_target_cores == 0) { + if (m_target_cores.load() == 0) { m_cv.Broadcast(); } } @@ -163,7 +163,7 @@ namespace ams::kern::arch::arm64::cpu { if ((op == Operation::InstructionMemoryBarrier) || (Kernel::GetState() == Kernel::State::Initializing)) { /* Check that there's no on-going operation. */ MESOSPHERE_ABORT_UNLESS(m_operation == Operation::Idle); - MESOSPHERE_ABORT_UNLESS(m_target_cores == 0); + MESOSPHERE_ABORT_UNLESS(m_target_cores.load() == 0); /* Set operation. */ m_operation = op; @@ -171,12 +171,13 @@ namespace ams::kern::arch::arm64::cpu { /* For certain operations, we want to send an interrupt. */ m_target_cores = other_cores_mask; - const u64 target_mask = m_target_cores; + const u64 target_mask = m_target_cores.load(); + DataSynchronizationBarrier(); Kernel::GetInterruptManager().SendInterProcessorInterrupt(KInterruptName_CacheOperation, target_mask); this->ProcessOperation(); - while (m_target_cores != 0) { + while (m_target_cores.load() != 0) { cpu::Yield(); } @@ -188,7 +189,7 @@ namespace ams::kern::arch::arm64::cpu { /* Check that there's no on-going operation. */ MESOSPHERE_ABORT_UNLESS(m_operation == Operation::Idle); - MESOSPHERE_ABORT_UNLESS(m_target_cores == 0); + MESOSPHERE_ABORT_UNLESS(m_target_cores.load() == 0); /* Set operation. */ m_operation = op; @@ -198,7 +199,7 @@ namespace ams::kern::arch::arm64::cpu { /* Use the condvar. */ m_cv.Broadcast(); - while (m_target_cores != 0) { + while (m_target_cores.load() != 0) { m_cv.Wait(std::addressof(m_cv_lock)); } diff --git a/libraries/libmesosphere/source/arch/arm64/kern_k_interrupt_manager.cpp b/libraries/libmesosphere/source/arch/arm64/kern_k_interrupt_manager.cpp index efb5e87a8..a554dcf3c 100644 --- a/libraries/libmesosphere/source/arch/arm64/kern_k_interrupt_manager.cpp +++ b/libraries/libmesosphere/source/arch/arm64/kern_k_interrupt_manager.cpp @@ -208,6 +208,8 @@ namespace ams::kern::arch::arm64 { } Result KInterruptManager::BindHandler(KInterruptHandler *handler, s32 irq, s32 core_id, s32 priority, bool manual_clear, bool level) { + MESOSPHERE_UNUSED(core_id); + R_UNLESS(KInterruptController::IsGlobal(irq) || KInterruptController::IsLocal(irq), svc::ResultOutOfRange()); KScopedInterruptDisable di; @@ -222,6 +224,8 @@ namespace ams::kern::arch::arm64 { } Result KInterruptManager::UnbindHandler(s32 irq, s32 core_id) { + MESOSPHERE_UNUSED(core_id); + R_UNLESS(KInterruptController::IsGlobal(irq) || KInterruptController::IsLocal(irq), svc::ResultOutOfRange()); KScopedInterruptDisable di; @@ -244,6 +248,8 @@ namespace ams::kern::arch::arm64 { } Result KInterruptManager::ClearInterrupt(s32 irq, s32 core_id) { + MESOSPHERE_UNUSED(core_id); + R_UNLESS(KInterruptController::IsGlobal(irq) || KInterruptController::IsLocal(irq), svc::ResultOutOfRange()); KScopedInterruptDisable di; diff --git a/libraries/libmesosphere/source/board/nintendo/nx/kern_k_device_page_table.cpp b/libraries/libmesosphere/source/board/nintendo/nx/kern_k_device_page_table.cpp index 3ca4b4c8b..97b95ce8e 100644 --- a/libraries/libmesosphere/source/board/nintendo/nx/kern_k_device_page_table.cpp +++ b/libraries/libmesosphere/source/board/nintendo/nx/kern_k_device_page_table.cpp @@ -1163,6 +1163,7 @@ namespace ams::kern::board::nintendo::nx { } void KDevicePageTable::UnmapImpl(KDeviceVirtualAddress address, u64 size, bool force) { + MESOSPHERE_UNUSED(force); MESOSPHERE_ASSERT((address & ~DeviceVirtualAddressMask) == 0); MESOSPHERE_ASSERT(((address + size - 1) & ~DeviceVirtualAddressMask) == 0); diff --git a/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager.cpp b/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager.cpp index 913137d5f..e6860023d 100644 --- a/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager.cpp +++ b/libraries/libmesosphere/source/board/nintendo/nx/kern_k_sleep_manager.cpp @@ -84,6 +84,7 @@ namespace ams::kern::board::nintendo::nx { do { bool res = smc::ReadWriteRegister(std::addressof(value), PmcPhysicalAddress + APBDEV_PMC_PWRGATE_STATUS, 0, 0); MESOSPHERE_ASSERT(res); + MESOSPHERE_UNUSED(res); } while ((value & PWRGATE_STATUS_CE123_MASK) != 0); } diff --git a/libraries/libmesosphere/source/kern_k_handle_table.cpp b/libraries/libmesosphere/source/kern_k_handle_table.cpp index 11e84ff50..7b3733fc4 100644 --- a/libraries/libmesosphere/source/kern_k_handle_table.cpp +++ b/libraries/libmesosphere/source/kern_k_handle_table.cpp @@ -121,6 +121,7 @@ namespace ams::kern { MESOSPHERE_ASSERT(reserved == 0); MESOSPHERE_ASSERT(linear_id != 0); MESOSPHERE_ASSERT(index < m_table_size); + MESOSPHERE_UNUSED(linear_id, reserved); /* Free the entry. */ /* NOTE: This code does not check the linear id. */ @@ -143,6 +144,7 @@ namespace ams::kern { MESOSPHERE_ASSERT(reserved == 0); MESOSPHERE_ASSERT(linear_id != 0); MESOSPHERE_ASSERT(index < m_table_size); + MESOSPHERE_UNUSED(reserved); /* Set the entry. */ Entry *entry = std::addressof(m_table[index]); diff --git a/libraries/libmesosphere/source/kern_k_page_table_base.cpp b/libraries/libmesosphere/source/kern_k_page_table_base.cpp index 6b7cbebca..28a2a0386 100644 --- a/libraries/libmesosphere/source/kern_k_page_table_base.cpp +++ b/libraries/libmesosphere/source/kern_k_page_table_base.cpp @@ -16,9 +16,6 @@ #include #include -#undef ALWAYS_INLINE_LAMBDA -#define ALWAYS_INLINE_LAMBDA - namespace ams::kern { Result KPageTableBase::InitializeForKernel(bool is_64_bit, void *table, KVirtualAddress start, KVirtualAddress end) { @@ -3288,6 +3285,7 @@ namespace ams::kern { TraversalEntry next_entry; bool traverse_valid = src_impl.BeginTraversal(std::addressof(next_entry), std::addressof(context), aligned_src_start); MESOSPHERE_ASSERT(traverse_valid); + MESOSPHERE_UNUSED(traverse_valid); /* Prepare tracking variables. */ KPhysicalAddress cur_block_addr = next_entry.phys_addr; diff --git a/libraries/libmesosphere/source/kern_k_scheduler.cpp b/libraries/libmesosphere/source/kern_k_scheduler.cpp index 22951af90..9534a6e95 100644 --- a/libraries/libmesosphere/source/kern_k_scheduler.cpp +++ b/libraries/libmesosphere/source/kern_k_scheduler.cpp @@ -17,6 +17,9 @@ namespace ams::kern { + #pragma GCC push_options + #pragma GCC optimize ("-O3") + bool KScheduler::s_scheduler_update_needed; KScheduler::LockType KScheduler::s_scheduler_lock; KSchedulerPriorityQueue KScheduler::s_priority_queue; @@ -607,4 +610,6 @@ namespace ams::kern { } } + #pragma GCC pop_options + } diff --git a/libraries/libmesosphere/source/kern_k_thread.cpp b/libraries/libmesosphere/source/kern_k_thread.cpp index 2a89ebb28..d2d5581dd 100644 --- a/libraries/libmesosphere/source/kern_k_thread.cpp +++ b/libraries/libmesosphere/source/kern_k_thread.cpp @@ -63,6 +63,7 @@ namespace ams::kern { m_tls_address = Null; const uintptr_t kern_stack_top_address = reinterpret_cast(kern_stack_top); + MESOSPHERE_UNUSED(kern_stack_top_address); /* Next, assert things based on the type. */ switch (type) { @@ -1161,7 +1162,7 @@ namespace ams::kern { /* Determine if this is the first termination request. */ const bool first_request = [&] ALWAYS_INLINE_LAMBDA () -> bool { /* Perform an atomic compare-and-swap from false to true. */ - bool expected = false; + u8 expected = false; return m_termination_requested.compare_exchange_strong(expected, true); }(); diff --git a/libraries/libvapours/include/vapours/svc/svc_tick.hpp b/libraries/libvapours/include/vapours/svc/svc_tick.hpp index ecdfd53bd..c81767833 100644 --- a/libraries/libvapours/include/vapours/svc/svc_tick.hpp +++ b/libraries/libvapours/include/vapours/svc/svc_tick.hpp @@ -27,7 +27,7 @@ namespace ams::svc { private: s64 tick; private: - static constexpr s64 NanoSecondsPerSecond = INT64_C(1'000'000'000); + static constexpr s64 NanoSecondsPerSecond = TimeSpan::FromSeconds(1).GetNanoSeconds(); static constexpr void DivNs(s64 &out, const s64 value) { out = value / NanoSecondsPerSecond;