2024-05-16 14:55:50 +02:00
|
|
|
// SPDX-FileCopyrightText: Copyright 2024 shadPS4 Emulator Project
|
|
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
|
|
|
|
|
|
#include "common/alignment.h"
|
|
|
|
#include "common/assert.h"
|
2024-06-11 12:14:33 +02:00
|
|
|
#include "common/debug.h"
|
2024-05-16 14:55:50 +02:00
|
|
|
#include "core/libraries/error_codes.h"
|
2024-05-30 17:07:36 +02:00
|
|
|
#include "core/libraries/kernel/memory_management.h"
|
2024-05-16 14:55:50 +02:00
|
|
|
#include "core/memory.h"
|
2024-08-08 14:02:10 +02:00
|
|
|
#include "video_core/renderer_vulkan/vk_rasterizer.h"
|
2024-05-16 14:55:50 +02:00
|
|
|
|
|
|
|
namespace Core {
|
|
|
|
|
|
|
|
MemoryManager::MemoryManager() {
|
2024-06-10 01:13:44 +02:00
|
|
|
// Insert an area that covers direct memory physical block.
|
|
|
|
dmem_map.emplace(0, DirectMemoryArea{0, SCE_KERNEL_MAIN_DMEM_SIZE});
|
2024-05-16 14:55:50 +02:00
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
// Insert a virtual memory area that covers the entire area we manage.
|
2024-07-15 23:34:54 +02:00
|
|
|
const VAddr system_managed_base = impl.SystemManagedVirtualBase();
|
|
|
|
const size_t system_managed_size = impl.SystemManagedVirtualSize();
|
|
|
|
const VAddr system_reserved_base = impl.SystemReservedVirtualBase();
|
|
|
|
const size_t system_reserved_size = impl.SystemReservedVirtualSize();
|
|
|
|
const VAddr user_base = impl.UserVirtualBase();
|
|
|
|
const size_t user_size = impl.UserVirtualSize();
|
|
|
|
vma_map.emplace(system_managed_base,
|
|
|
|
VirtualMemoryArea{system_managed_base, system_managed_size});
|
|
|
|
vma_map.emplace(system_reserved_base,
|
|
|
|
VirtualMemoryArea{system_reserved_base, system_reserved_size});
|
|
|
|
vma_map.emplace(user_base, VirtualMemoryArea{user_base, user_size});
|
2024-07-01 12:35:35 +02:00
|
|
|
|
|
|
|
// Log initialization.
|
2024-07-15 23:34:54 +02:00
|
|
|
LOG_INFO(Kernel_Vmm, "Usable memory address space: {}_GB",
|
|
|
|
(system_managed_size + system_reserved_size + user_size) >> 30);
|
2024-05-16 14:55:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
MemoryManager::~MemoryManager() = default;
|
|
|
|
|
|
|
|
PAddr MemoryManager::Allocate(PAddr search_start, PAddr search_end, size_t size, u64 alignment,
|
|
|
|
int memory_type) {
|
2024-06-10 01:13:44 +02:00
|
|
|
std::scoped_lock lk{mutex};
|
2024-05-16 14:55:50 +02:00
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
auto dmem_area = FindDmemArea(search_start);
|
|
|
|
|
|
|
|
const auto is_suitable = [&] {
|
|
|
|
return dmem_area->second.is_free && dmem_area->second.size >= size;
|
|
|
|
};
|
|
|
|
while (!is_suitable() && dmem_area->second.GetEnd() <= search_end) {
|
|
|
|
dmem_area++;
|
2024-05-16 14:55:50 +02:00
|
|
|
}
|
2024-06-10 01:13:44 +02:00
|
|
|
ASSERT_MSG(is_suitable(), "Unable to find free direct memory area");
|
2024-05-16 14:55:50 +02:00
|
|
|
|
|
|
|
// Align free position
|
2024-06-10 01:13:44 +02:00
|
|
|
PAddr free_addr = dmem_area->second.base;
|
2024-06-07 15:26:43 +02:00
|
|
|
free_addr = alignment > 0 ? Common::AlignUp(free_addr, alignment) : free_addr;
|
2024-05-16 14:55:50 +02:00
|
|
|
|
|
|
|
// Add the allocated region to the list and commit its pages.
|
2024-08-13 08:05:30 +02:00
|
|
|
auto& area = CarveDmemArea(free_addr, size)->second;
|
2024-06-10 01:13:44 +02:00
|
|
|
area.memory_type = memory_type;
|
|
|
|
area.is_free = false;
|
2024-05-16 14:55:50 +02:00
|
|
|
return free_addr;
|
|
|
|
}
|
|
|
|
|
|
|
|
void MemoryManager::Free(PAddr phys_addr, size_t size) {
|
2024-06-10 01:13:44 +02:00
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
2024-08-13 08:05:30 +02:00
|
|
|
auto dmem_area = CarveDmemArea(phys_addr, size);
|
|
|
|
ASSERT(dmem_area != dmem_map.end() && dmem_area->second.size >= size);
|
2024-06-10 01:13:44 +02:00
|
|
|
|
|
|
|
// Release any dmem mappings that reference this physical block.
|
|
|
|
std::vector<std::pair<VAddr, u64>> remove_list;
|
|
|
|
for (const auto& [addr, mapping] : vma_map) {
|
|
|
|
if (mapping.type != VMAType::Direct) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (mapping.phys_base <= phys_addr && phys_addr < mapping.phys_base + mapping.size) {
|
2024-08-13 08:05:30 +02:00
|
|
|
auto vma_segment_start_addr = phys_addr - mapping.phys_base + addr;
|
|
|
|
LOG_INFO(Kernel_Vmm, "Unmaping direct mapping {:#x} with size {:#x}",
|
|
|
|
vma_segment_start_addr, size);
|
2024-06-10 01:13:44 +02:00
|
|
|
// Unmaping might erase from vma_map. We can't do it here.
|
2024-08-13 08:05:30 +02:00
|
|
|
remove_list.emplace_back(vma_segment_start_addr, size);
|
2024-06-10 01:13:44 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
for (const auto& [addr, size] : remove_list) {
|
|
|
|
UnmapMemory(addr, size);
|
|
|
|
}
|
2024-05-16 14:55:50 +02:00
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
// Mark region as free and attempt to coalesce it with neighbours.
|
|
|
|
auto& area = dmem_area->second;
|
|
|
|
area.is_free = true;
|
|
|
|
area.memory_type = 0;
|
|
|
|
MergeAdjacent(dmem_map, dmem_area);
|
2024-05-16 14:55:50 +02:00
|
|
|
}
|
|
|
|
|
2024-07-11 14:35:58 +02:00
|
|
|
int MemoryManager::Reserve(void** out_addr, VAddr virtual_addr, size_t size, MemoryMapFlags flags,
|
|
|
|
u64 alignment) {
|
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
2024-07-21 22:16:46 +02:00
|
|
|
virtual_addr = (virtual_addr == 0) ? impl.SystemManagedVirtualBase() : virtual_addr;
|
2024-07-21 20:29:24 +02:00
|
|
|
alignment = alignment > 0 ? alignment : 16_KB;
|
2024-07-11 14:35:58 +02:00
|
|
|
VAddr mapped_addr = alignment > 0 ? Common::AlignUp(virtual_addr, alignment) : virtual_addr;
|
|
|
|
|
2024-07-25 22:01:12 +02:00
|
|
|
// Fixed mapping means the virtual address must exactly match the provided one.
|
|
|
|
if (True(flags & MemoryMapFlags::Fixed)) {
|
|
|
|
const auto& vma = FindVMA(mapped_addr)->second;
|
|
|
|
// If the VMA is mapped, unmap the region first.
|
|
|
|
if (vma.IsMapped()) {
|
|
|
|
UnmapMemory(mapped_addr, size);
|
|
|
|
}
|
|
|
|
const size_t remaining_size = vma.base + vma.size - mapped_addr;
|
|
|
|
ASSERT_MSG(vma.type == VMAType::Free && remaining_size >= size);
|
|
|
|
}
|
|
|
|
|
2024-07-21 20:29:24 +02:00
|
|
|
// Find the first free area starting with provided virtual address.
|
|
|
|
if (False(flags & MemoryMapFlags::Fixed)) {
|
2024-07-25 22:01:12 +02:00
|
|
|
mapped_addr = SearchFree(mapped_addr, size, alignment);
|
2024-07-21 20:29:24 +02:00
|
|
|
}
|
|
|
|
|
2024-07-11 14:35:58 +02:00
|
|
|
// Add virtual memory area
|
2024-07-25 22:01:12 +02:00
|
|
|
const auto new_vma_handle = CarveVMA(mapped_addr, size);
|
|
|
|
auto& new_vma = new_vma_handle->second;
|
2024-07-11 14:35:58 +02:00
|
|
|
new_vma.disallow_merge = True(flags & MemoryMapFlags::NoCoalesce);
|
|
|
|
new_vma.prot = MemoryProt::NoAccess;
|
|
|
|
new_vma.name = "";
|
|
|
|
new_vma.type = VMAType::Reserved;
|
2024-07-25 22:01:12 +02:00
|
|
|
MergeAdjacent(vma_map, new_vma_handle);
|
2024-07-11 14:35:58 +02:00
|
|
|
|
|
|
|
*out_addr = std::bit_cast<void*>(mapped_addr);
|
|
|
|
return ORBIS_OK;
|
|
|
|
}
|
|
|
|
|
2024-05-16 14:55:50 +02:00
|
|
|
int MemoryManager::MapMemory(void** out_addr, VAddr virtual_addr, size_t size, MemoryProt prot,
|
|
|
|
MemoryMapFlags flags, VMAType type, std::string_view name,
|
2024-06-10 01:13:44 +02:00
|
|
|
bool is_exec, PAddr phys_addr, u64 alignment) {
|
|
|
|
std::scoped_lock lk{mutex};
|
2024-07-25 22:01:12 +02:00
|
|
|
|
|
|
|
// Certain games perform flexible mappings on loop to determine
|
|
|
|
// the available flexible memory size. Questionable but we need to handle this.
|
2024-06-21 17:22:37 +02:00
|
|
|
if (type == VMAType::Flexible && flexible_usage + size > total_flexible_size) {
|
2024-06-15 13:36:07 +02:00
|
|
|
return SCE_KERNEL_ERROR_ENOMEM;
|
|
|
|
}
|
2024-06-10 01:13:44 +02:00
|
|
|
|
|
|
|
// When virtual addr is zero, force it to virtual_base. The guest cannot pass Fixed
|
|
|
|
// flag so we will take the branch that searches for free (or reserved) mappings.
|
2024-07-15 23:34:54 +02:00
|
|
|
virtual_addr = (virtual_addr == 0) ? impl.SystemManagedVirtualBase() : virtual_addr;
|
2024-06-15 13:36:07 +02:00
|
|
|
alignment = alignment > 0 ? alignment : 16_KB;
|
2024-06-05 21:08:18 +02:00
|
|
|
VAddr mapped_addr = alignment > 0 ? Common::AlignUp(virtual_addr, alignment) : virtual_addr;
|
2024-05-16 14:55:50 +02:00
|
|
|
|
|
|
|
// Fixed mapping means the virtual address must exactly match the provided one.
|
2024-07-25 22:01:12 +02:00
|
|
|
if (True(flags & MemoryMapFlags::Fixed)) {
|
2024-05-16 14:55:50 +02:00
|
|
|
// This should return SCE_KERNEL_ERROR_ENOMEM but shouldn't normally happen.
|
|
|
|
const auto& vma = FindVMA(mapped_addr)->second;
|
2024-05-30 17:07:36 +02:00
|
|
|
const size_t remaining_size = vma.base + vma.size - mapped_addr;
|
2024-07-25 22:01:12 +02:00
|
|
|
ASSERT_MSG(!vma.IsMapped() && remaining_size >= size);
|
2024-05-16 14:55:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Find the first free area starting with provided virtual address.
|
|
|
|
if (False(flags & MemoryMapFlags::Fixed)) {
|
2024-07-25 22:01:12 +02:00
|
|
|
mapped_addr = SearchFree(mapped_addr, size, alignment);
|
2024-05-16 14:55:50 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Perform the mapping.
|
2024-06-10 01:13:44 +02:00
|
|
|
*out_addr = impl.Map(mapped_addr, size, alignment, phys_addr, is_exec);
|
2024-06-11 12:14:33 +02:00
|
|
|
TRACK_ALLOC(*out_addr, size, "VMEM");
|
2024-07-25 22:01:12 +02:00
|
|
|
|
|
|
|
auto& new_vma = CarveVMA(mapped_addr, size)->second;
|
|
|
|
new_vma.disallow_merge = True(flags & MemoryMapFlags::NoCoalesce);
|
|
|
|
new_vma.prot = prot;
|
|
|
|
new_vma.name = name;
|
|
|
|
new_vma.type = type;
|
2024-08-13 08:05:30 +02:00
|
|
|
new_vma.is_exec = is_exec;
|
2024-07-25 22:01:12 +02:00
|
|
|
|
|
|
|
if (type == VMAType::Direct) {
|
|
|
|
new_vma.phys_base = phys_addr;
|
2024-08-08 14:02:10 +02:00
|
|
|
rasterizer->MapMemory(mapped_addr, size);
|
2024-07-25 22:01:12 +02:00
|
|
|
}
|
|
|
|
if (type == VMAType::Flexible) {
|
|
|
|
flexible_usage += size;
|
|
|
|
}
|
|
|
|
|
2024-05-16 14:55:50 +02:00
|
|
|
return ORBIS_OK;
|
|
|
|
}
|
|
|
|
|
2024-06-15 13:36:07 +02:00
|
|
|
int MemoryManager::MapFile(void** out_addr, VAddr virtual_addr, size_t size, MemoryProt prot,
|
|
|
|
MemoryMapFlags flags, uintptr_t fd, size_t offset) {
|
2024-07-25 22:01:12 +02:00
|
|
|
VAddr mapped_addr = (virtual_addr == 0) ? impl.SystemManagedVirtualBase() : virtual_addr;
|
2024-06-15 13:36:07 +02:00
|
|
|
const size_t size_aligned = Common::AlignUp(size, 16_KB);
|
|
|
|
|
|
|
|
// Find first free area to map the file.
|
2024-07-11 14:35:58 +02:00
|
|
|
if (False(flags & MemoryMapFlags::Fixed)) {
|
2024-07-29 18:08:06 +02:00
|
|
|
mapped_addr = SearchFree(mapped_addr, size_aligned, 1);
|
2024-07-11 14:35:58 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
if (True(flags & MemoryMapFlags::Fixed)) {
|
|
|
|
const auto& vma = FindVMA(virtual_addr)->second;
|
|
|
|
const size_t remaining_size = vma.base + vma.size - virtual_addr;
|
2024-07-25 22:01:12 +02:00
|
|
|
ASSERT_MSG(!vma.IsMapped() && remaining_size >= size);
|
2024-06-15 13:36:07 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// Map the file.
|
2024-07-11 14:35:58 +02:00
|
|
|
impl.MapFile(mapped_addr, size, offset, std::bit_cast<u32>(prot), fd);
|
2024-06-15 13:36:07 +02:00
|
|
|
|
|
|
|
// Add virtual memory area
|
2024-07-25 22:01:12 +02:00
|
|
|
auto& new_vma = CarveVMA(mapped_addr, size_aligned)->second;
|
2024-06-15 13:36:07 +02:00
|
|
|
new_vma.disallow_merge = True(flags & MemoryMapFlags::NoCoalesce);
|
|
|
|
new_vma.prot = prot;
|
|
|
|
new_vma.name = "File";
|
|
|
|
new_vma.fd = fd;
|
|
|
|
new_vma.type = VMAType::File;
|
|
|
|
|
|
|
|
*out_addr = std::bit_cast<void*>(mapped_addr);
|
|
|
|
return ORBIS_OK;
|
|
|
|
}
|
|
|
|
|
2024-05-16 14:55:50 +02:00
|
|
|
void MemoryManager::UnmapMemory(VAddr virtual_addr, size_t size) {
|
2024-06-10 01:13:44 +02:00
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
2024-07-25 22:01:12 +02:00
|
|
|
const auto it = FindVMA(virtual_addr);
|
2024-08-13 08:05:30 +02:00
|
|
|
const auto& vma_base = it->second;
|
|
|
|
ASSERT_MSG(vma_base.Contains(virtual_addr, size),
|
2024-07-25 22:01:12 +02:00
|
|
|
"Existing mapping does not contain requested unmap range");
|
2024-05-16 14:55:50 +02:00
|
|
|
|
2024-08-13 08:05:30 +02:00
|
|
|
const auto vma_base_addr = vma_base.base;
|
|
|
|
const auto vma_base_size = vma_base.size;
|
|
|
|
const auto phys_base = vma_base.phys_base;
|
|
|
|
const bool is_exec = vma_base.is_exec;
|
|
|
|
const auto start_in_vma = virtual_addr - vma_base_addr;
|
|
|
|
const auto type = vma_base.type;
|
2024-06-15 13:36:07 +02:00
|
|
|
const bool has_backing = type == VMAType::Direct || type == VMAType::File;
|
2024-06-10 01:13:44 +02:00
|
|
|
if (type == VMAType::Direct) {
|
2024-08-08 14:02:10 +02:00
|
|
|
rasterizer->UnmapMemory(virtual_addr, size);
|
2024-05-25 14:33:15 +02:00
|
|
|
}
|
2024-06-15 13:36:07 +02:00
|
|
|
if (type == VMAType::Flexible) {
|
2024-06-21 17:22:37 +02:00
|
|
|
flexible_usage -= size;
|
2024-06-15 13:36:07 +02:00
|
|
|
}
|
2024-05-25 14:33:15 +02:00
|
|
|
|
2024-05-16 14:55:50 +02:00
|
|
|
// Mark region as free and attempt to coalesce it with neighbours.
|
2024-07-25 22:01:12 +02:00
|
|
|
const auto new_it = CarveVMA(virtual_addr, size);
|
|
|
|
auto& vma = new_it->second;
|
2024-05-16 14:55:50 +02:00
|
|
|
vma.type = VMAType::Free;
|
|
|
|
vma.prot = MemoryProt::NoAccess;
|
|
|
|
vma.phys_base = 0;
|
2024-07-25 22:01:12 +02:00
|
|
|
vma.disallow_merge = false;
|
2024-07-29 18:08:06 +02:00
|
|
|
vma.name = "";
|
2024-07-25 22:01:12 +02:00
|
|
|
MergeAdjacent(vma_map, new_it);
|
2024-05-16 14:55:50 +02:00
|
|
|
|
|
|
|
// Unmap the memory region.
|
2024-08-13 08:05:30 +02:00
|
|
|
impl.Unmap(vma_base_addr, vma_base_size, start_in_vma, start_in_vma + size, phys_base, is_exec,
|
|
|
|
has_backing);
|
2024-06-11 12:14:33 +02:00
|
|
|
TRACK_FREE(virtual_addr, "VMEM");
|
2024-05-16 14:55:50 +02:00
|
|
|
}
|
|
|
|
|
2024-05-26 14:51:35 +02:00
|
|
|
int MemoryManager::QueryProtection(VAddr addr, void** start, void** end, u32* prot) {
|
2024-06-10 01:13:44 +02:00
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
2024-05-26 14:51:35 +02:00
|
|
|
const auto it = FindVMA(addr);
|
|
|
|
const auto& vma = it->second;
|
|
|
|
ASSERT_MSG(vma.type != VMAType::Free, "Provided address is not mapped");
|
|
|
|
|
2024-06-26 17:04:28 +02:00
|
|
|
if (start != nullptr) {
|
|
|
|
*start = reinterpret_cast<void*>(vma.base);
|
|
|
|
}
|
|
|
|
if (end != nullptr) {
|
|
|
|
*end = reinterpret_cast<void*>(vma.base + vma.size);
|
|
|
|
}
|
|
|
|
if (prot != nullptr) {
|
|
|
|
*prot = static_cast<u32>(vma.prot);
|
|
|
|
}
|
2024-05-30 17:07:36 +02:00
|
|
|
return ORBIS_OK;
|
|
|
|
}
|
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
int MemoryManager::VirtualQuery(VAddr addr, int flags,
|
2024-08-08 14:02:10 +02:00
|
|
|
::Libraries::Kernel::OrbisVirtualQueryInfo* info) {
|
2024-06-10 21:59:12 +02:00
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
auto it = FindVMA(addr);
|
2024-08-14 15:18:46 +02:00
|
|
|
if (it->second.type == VMAType::Free && flags == 1) {
|
2024-06-10 01:13:44 +02:00
|
|
|
it++;
|
|
|
|
}
|
2024-08-14 15:18:46 +02:00
|
|
|
if (it->second.type == VMAType::Free) {
|
2024-06-10 01:13:44 +02:00
|
|
|
LOG_WARNING(Kernel_Vmm, "VirtualQuery on free memory region");
|
|
|
|
return ORBIS_KERNEL_ERROR_EACCES;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto& vma = it->second;
|
|
|
|
info->start = vma.base;
|
|
|
|
info->end = vma.base + vma.size;
|
|
|
|
info->is_flexible.Assign(vma.type == VMAType::Flexible);
|
|
|
|
info->is_direct.Assign(vma.type == VMAType::Direct);
|
|
|
|
info->is_commited.Assign(vma.type != VMAType::Free);
|
2024-07-29 18:08:06 +02:00
|
|
|
vma.name.copy(info->name.data(), std::min(info->name.size(), vma.name.size()));
|
2024-06-10 01:13:44 +02:00
|
|
|
if (vma.type == VMAType::Direct) {
|
|
|
|
const auto dmem_it = FindDmemArea(vma.phys_base);
|
|
|
|
ASSERT(dmem_it != dmem_map.end());
|
2024-07-07 00:26:17 +02:00
|
|
|
info->offset = vma.phys_base;
|
2024-06-10 01:13:44 +02:00
|
|
|
info->memory_type = dmem_it->second.memory_type;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ORBIS_OK;
|
|
|
|
}
|
|
|
|
|
2024-05-30 17:07:36 +02:00
|
|
|
int MemoryManager::DirectMemoryQuery(PAddr addr, bool find_next,
|
2024-08-08 14:02:10 +02:00
|
|
|
::Libraries::Kernel::OrbisQueryInfo* out_info) {
|
2024-06-10 01:13:44 +02:00
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
|
|
|
auto dmem_area = FindDmemArea(addr);
|
2024-06-15 13:36:07 +02:00
|
|
|
while (dmem_area != dmem_map.end() && dmem_area->second.is_free && find_next) {
|
2024-06-10 01:13:44 +02:00
|
|
|
dmem_area++;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (dmem_area == dmem_map.end() || dmem_area->second.is_free) {
|
|
|
|
LOG_ERROR(Core, "Unable to find allocated direct memory region to query!");
|
|
|
|
return ORBIS_KERNEL_ERROR_EACCES;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto& area = dmem_area->second;
|
|
|
|
out_info->start = area.base;
|
|
|
|
out_info->end = area.GetEnd();
|
|
|
|
out_info->memoryType = area.memory_type;
|
|
|
|
return ORBIS_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
int MemoryManager::DirectQueryAvailable(PAddr search_start, PAddr search_end, size_t alignment,
|
|
|
|
PAddr* phys_addr_out, size_t* size_out) {
|
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
|
|
|
auto dmem_area = FindDmemArea(search_start);
|
|
|
|
PAddr paddr{};
|
|
|
|
size_t max_size{};
|
|
|
|
while (dmem_area != dmem_map.end() && dmem_area->second.GetEnd() <= search_end) {
|
|
|
|
if (dmem_area->second.size > max_size) {
|
|
|
|
paddr = dmem_area->second.base;
|
|
|
|
max_size = dmem_area->second.size;
|
|
|
|
}
|
|
|
|
dmem_area++;
|
2024-05-30 17:07:36 +02:00
|
|
|
}
|
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
*phys_addr_out = alignment > 0 ? Common::AlignUp(paddr, alignment) : paddr;
|
|
|
|
*size_out = max_size;
|
2024-05-30 17:07:36 +02:00
|
|
|
return ORBIS_OK;
|
2024-05-26 14:51:35 +02:00
|
|
|
}
|
|
|
|
|
2024-07-29 18:08:06 +02:00
|
|
|
void MemoryManager::NameVirtualRange(VAddr virtual_addr, size_t size, std::string_view name) {
|
|
|
|
auto it = FindVMA(virtual_addr);
|
|
|
|
|
|
|
|
ASSERT_MSG(it->second.Contains(virtual_addr, size),
|
|
|
|
"Range provided is not fully containted in vma");
|
|
|
|
it->second.name = name;
|
|
|
|
}
|
2024-07-25 22:01:12 +02:00
|
|
|
VAddr MemoryManager::SearchFree(VAddr virtual_addr, size_t size, u32 alignment) {
|
2024-07-30 21:40:42 +02:00
|
|
|
// If the requested address is below the mapped range, start search from the lowest address
|
|
|
|
auto min_search_address = impl.SystemManagedVirtualBase();
|
|
|
|
if (virtual_addr < min_search_address) {
|
|
|
|
virtual_addr = min_search_address;
|
|
|
|
}
|
|
|
|
|
2024-07-25 22:01:12 +02:00
|
|
|
auto it = FindVMA(virtual_addr);
|
2024-07-30 21:40:42 +02:00
|
|
|
ASSERT_MSG(it != vma_map.end(), "Specified mapping address was not found!");
|
|
|
|
|
2024-07-25 22:01:12 +02:00
|
|
|
// If the VMA is free and contains the requested mapping we are done.
|
|
|
|
if (it->second.IsFree() && it->second.Contains(virtual_addr, size)) {
|
|
|
|
return virtual_addr;
|
|
|
|
}
|
|
|
|
// Search for the first free VMA that fits our mapping.
|
|
|
|
const auto is_suitable = [&] {
|
|
|
|
if (!it->second.IsFree()) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
const auto& vma = it->second;
|
|
|
|
virtual_addr = Common::AlignUp(vma.base, alignment);
|
|
|
|
// Sometimes the alignment itself might be larger than the VMA.
|
|
|
|
if (virtual_addr > vma.base + vma.size) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
const size_t remaining_size = vma.base + vma.size - virtual_addr;
|
|
|
|
return remaining_size >= size;
|
|
|
|
};
|
|
|
|
while (!is_suitable()) {
|
|
|
|
it++;
|
|
|
|
}
|
|
|
|
return virtual_addr;
|
|
|
|
}
|
|
|
|
|
|
|
|
MemoryManager::VMAHandle MemoryManager::CarveVMA(VAddr virtual_addr, size_t size) {
|
2024-05-16 14:55:50 +02:00
|
|
|
auto vma_handle = FindVMA(virtual_addr);
|
|
|
|
ASSERT_MSG(vma_handle != vma_map.end(), "Virtual address not in vm_map");
|
|
|
|
|
|
|
|
const VirtualMemoryArea& vma = vma_handle->second;
|
2024-07-25 22:01:12 +02:00
|
|
|
ASSERT_MSG(vma.base <= virtual_addr, "Adding a mapping to already mapped region");
|
2024-05-16 14:55:50 +02:00
|
|
|
|
|
|
|
const VAddr start_in_vma = virtual_addr - vma.base;
|
|
|
|
const VAddr end_in_vma = start_in_vma + size;
|
|
|
|
ASSERT_MSG(end_in_vma <= vma.size, "Mapping cannot fit inside free region");
|
|
|
|
|
|
|
|
if (end_in_vma != vma.size) {
|
|
|
|
// Split VMA at the end of the allocated region
|
|
|
|
Split(vma_handle, end_in_vma);
|
|
|
|
}
|
|
|
|
if (start_in_vma != 0) {
|
|
|
|
// Split VMA at the start of the allocated region
|
|
|
|
vma_handle = Split(vma_handle, start_in_vma);
|
|
|
|
}
|
|
|
|
|
2024-07-25 22:01:12 +02:00
|
|
|
return vma_handle;
|
2024-05-16 14:55:50 +02:00
|
|
|
}
|
|
|
|
|
2024-08-13 08:05:30 +02:00
|
|
|
MemoryManager::DMemHandle MemoryManager::CarveDmemArea(PAddr addr, size_t size) {
|
2024-06-10 01:13:44 +02:00
|
|
|
auto dmem_handle = FindDmemArea(addr);
|
|
|
|
ASSERT_MSG(dmem_handle != dmem_map.end(), "Physical address not in dmem_map");
|
|
|
|
|
|
|
|
const DirectMemoryArea& area = dmem_handle->second;
|
2024-08-13 08:05:30 +02:00
|
|
|
ASSERT_MSG(area.base <= addr, "Adding an allocation to already allocated region");
|
2024-06-10 01:13:44 +02:00
|
|
|
|
|
|
|
const PAddr start_in_area = addr - area.base;
|
|
|
|
const PAddr end_in_vma = start_in_area + size;
|
|
|
|
ASSERT_MSG(end_in_vma <= area.size, "Mapping cannot fit inside free region");
|
|
|
|
|
|
|
|
if (end_in_vma != area.size) {
|
|
|
|
// Split VMA at the end of the allocated region
|
|
|
|
Split(dmem_handle, end_in_vma);
|
|
|
|
}
|
|
|
|
if (start_in_area != 0) {
|
|
|
|
// Split VMA at the start of the allocated region
|
|
|
|
dmem_handle = Split(dmem_handle, start_in_area);
|
|
|
|
}
|
|
|
|
|
2024-08-13 08:05:30 +02:00
|
|
|
return dmem_handle;
|
2024-06-10 01:13:44 +02:00
|
|
|
}
|
|
|
|
|
2024-05-30 17:07:36 +02:00
|
|
|
MemoryManager::VMAHandle MemoryManager::Split(VMAHandle vma_handle, size_t offset_in_vma) {
|
2024-05-16 14:55:50 +02:00
|
|
|
auto& old_vma = vma_handle->second;
|
|
|
|
ASSERT(offset_in_vma < old_vma.size && offset_in_vma > 0);
|
|
|
|
|
|
|
|
auto new_vma = old_vma;
|
|
|
|
old_vma.size = offset_in_vma;
|
|
|
|
new_vma.base += offset_in_vma;
|
|
|
|
new_vma.size -= offset_in_vma;
|
|
|
|
|
|
|
|
if (new_vma.type == VMAType::Direct) {
|
|
|
|
new_vma.phys_base += offset_in_vma;
|
|
|
|
}
|
|
|
|
return vma_map.emplace_hint(std::next(vma_handle), new_vma.base, new_vma);
|
|
|
|
}
|
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
MemoryManager::DMemHandle MemoryManager::Split(DMemHandle dmem_handle, size_t offset_in_area) {
|
|
|
|
auto& old_area = dmem_handle->second;
|
|
|
|
ASSERT(offset_in_area < old_area.size && offset_in_area > 0);
|
2024-05-16 14:55:50 +02:00
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
auto new_area = old_area;
|
|
|
|
old_area.size = offset_in_area;
|
|
|
|
new_area.base += offset_in_area;
|
|
|
|
new_area.size -= offset_in_area;
|
2024-05-16 14:55:50 +02:00
|
|
|
|
2024-06-10 01:13:44 +02:00
|
|
|
return dmem_map.emplace_hint(std::next(dmem_handle), new_area.base, new_area);
|
|
|
|
};
|
2024-05-16 14:55:50 +02:00
|
|
|
|
2024-07-21 21:35:01 +02:00
|
|
|
int MemoryManager::GetDirectMemoryType(PAddr addr, int* directMemoryTypeOut,
|
|
|
|
void** directMemoryStartOut, void** directMemoryEndOut) {
|
2024-07-21 20:29:24 +02:00
|
|
|
std::scoped_lock lk{mutex};
|
|
|
|
|
|
|
|
auto dmem_area = FindDmemArea(addr);
|
|
|
|
|
|
|
|
if (dmem_area == dmem_map.end() || dmem_area->second.is_free) {
|
|
|
|
LOG_ERROR(Core, "Unable to find allocated direct memory region to check type!");
|
|
|
|
return ORBIS_KERNEL_ERROR_ENOENT;
|
|
|
|
}
|
|
|
|
|
|
|
|
const auto& area = dmem_area->second;
|
|
|
|
*directMemoryStartOut = reinterpret_cast<void*>(area.base);
|
|
|
|
*directMemoryEndOut = reinterpret_cast<void*>(area.GetEnd());
|
|
|
|
*directMemoryTypeOut = area.memory_type;
|
|
|
|
return ORBIS_OK;
|
|
|
|
}
|
|
|
|
|
2024-05-16 14:55:50 +02:00
|
|
|
} // namespace Core
|