mirror of
https://github.com/yuzu-emu/yuzu-android.git
synced 2024-11-22 16:45:39 +01:00
k_page_table: add MapFirstGroup
This commit is contained in:
parent
b456af31e6
commit
f21058a6c0
@ -183,12 +183,17 @@ private:
|
|||||||
|
|
||||||
class KScopedPageGroup {
|
class KScopedPageGroup {
|
||||||
public:
|
public:
|
||||||
explicit KScopedPageGroup(const KPageGroup* gp) : m_pg(gp) {
|
explicit KScopedPageGroup(const KPageGroup* gp, bool not_first = true) : m_pg(gp) {
|
||||||
if (m_pg) {
|
if (m_pg) {
|
||||||
m_pg->Open();
|
if (not_first) {
|
||||||
|
m_pg->Open();
|
||||||
|
} else {
|
||||||
|
m_pg->OpenFirst();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
explicit KScopedPageGroup(const KPageGroup& gp) : KScopedPageGroup(std::addressof(gp)) {}
|
explicit KScopedPageGroup(const KPageGroup& gp, bool not_first = true)
|
||||||
|
: KScopedPageGroup(std::addressof(gp), not_first) {}
|
||||||
~KScopedPageGroup() {
|
~KScopedPageGroup() {
|
||||||
if (m_pg) {
|
if (m_pg) {
|
||||||
m_pg->Close();
|
m_pg->Close();
|
||||||
|
@ -1724,29 +1724,43 @@ Result KPageTable::MapPhysicalMemory(KProcessAddress address, size_t size) {
|
|||||||
PageSize;
|
PageSize;
|
||||||
|
|
||||||
// While we have pages to map, map them.
|
// While we have pages to map, map them.
|
||||||
while (map_pages > 0) {
|
{
|
||||||
// Check if we're at the end of the physical block.
|
// Create a page group for the current mapping range.
|
||||||
if (pg_pages == 0) {
|
KPageGroup cur_pg(m_kernel, m_block_info_manager);
|
||||||
// Ensure there are more pages to map.
|
{
|
||||||
ASSERT(pg_it != pg.end());
|
ON_RESULT_FAILURE_2 {
|
||||||
|
cur_pg.OpenFirst();
|
||||||
|
cur_pg.Close();
|
||||||
|
};
|
||||||
|
|
||||||
// Advance our physical block.
|
size_t remain_pages = map_pages;
|
||||||
++pg_it;
|
while (remain_pages > 0) {
|
||||||
pg_phys_addr = pg_it->GetAddress();
|
// Check if we're at the end of the physical block.
|
||||||
pg_pages = pg_it->GetNumPages();
|
if (pg_pages == 0) {
|
||||||
|
// Ensure there are more pages to map.
|
||||||
|
ASSERT(pg_it != pg.end());
|
||||||
|
|
||||||
|
// Advance our physical block.
|
||||||
|
++pg_it;
|
||||||
|
pg_phys_addr = pg_it->GetAddress();
|
||||||
|
pg_pages = pg_it->GetNumPages();
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add whatever we can to the current block.
|
||||||
|
const size_t cur_pages = std::min(pg_pages, remain_pages);
|
||||||
|
R_TRY(cur_pg.AddBlock(pg_phys_addr +
|
||||||
|
((pg_pages - cur_pages) * PageSize),
|
||||||
|
cur_pages));
|
||||||
|
|
||||||
|
// Advance.
|
||||||
|
remain_pages -= cur_pages;
|
||||||
|
pg_pages -= cur_pages;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Map whatever we can.
|
// Map the pages.
|
||||||
const size_t cur_pages = std::min(pg_pages, map_pages);
|
R_TRY(this->Operate(cur_address, map_pages, cur_pg,
|
||||||
R_TRY(Operate(cur_address, cur_pages, KMemoryPermission::UserReadWrite,
|
OperationType::MapFirstGroup));
|
||||||
OperationType::MapFirst, pg_phys_addr));
|
|
||||||
|
|
||||||
// Advance.
|
|
||||||
cur_address += cur_pages * PageSize;
|
|
||||||
map_pages -= cur_pages;
|
|
||||||
|
|
||||||
pg_phys_addr += cur_pages * PageSize;
|
|
||||||
pg_pages -= cur_pages;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -3037,9 +3051,10 @@ Result KPageTable::Operate(KProcessAddress addr, size_t num_pages, const KPageGr
|
|||||||
ASSERT(num_pages == page_group.GetNumPages());
|
ASSERT(num_pages == page_group.GetNumPages());
|
||||||
|
|
||||||
switch (operation) {
|
switch (operation) {
|
||||||
case OperationType::MapGroup: {
|
case OperationType::MapGroup:
|
||||||
|
case OperationType::MapFirstGroup: {
|
||||||
// We want to maintain a new reference to every page in the group.
|
// We want to maintain a new reference to every page in the group.
|
||||||
KScopedPageGroup spg(page_group);
|
KScopedPageGroup spg(page_group, operation != OperationType::MapFirstGroup);
|
||||||
|
|
||||||
for (const auto& node : page_group) {
|
for (const auto& node : page_group) {
|
||||||
const size_t size{node.GetNumPages() * PageSize};
|
const size_t size{node.GetNumPages() * PageSize};
|
||||||
@ -3081,7 +3096,6 @@ Result KPageTable::Operate(KProcessAddress addr, size_t num_pages, KMemoryPermis
|
|||||||
m_memory->UnmapRegion(*m_page_table_impl, addr, num_pages * PageSize);
|
m_memory->UnmapRegion(*m_page_table_impl, addr, num_pages * PageSize);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case OperationType::MapFirst:
|
|
||||||
case OperationType::Map: {
|
case OperationType::Map: {
|
||||||
ASSERT(map_addr);
|
ASSERT(map_addr);
|
||||||
ASSERT(Common::IsAligned(GetInteger(map_addr), PageSize));
|
ASSERT(Common::IsAligned(GetInteger(map_addr), PageSize));
|
||||||
@ -3089,11 +3103,7 @@ Result KPageTable::Operate(KProcessAddress addr, size_t num_pages, KMemoryPermis
|
|||||||
|
|
||||||
// Open references to pages, if we should.
|
// Open references to pages, if we should.
|
||||||
if (IsHeapPhysicalAddress(m_kernel.MemoryLayout(), map_addr)) {
|
if (IsHeapPhysicalAddress(m_kernel.MemoryLayout(), map_addr)) {
|
||||||
if (operation == OperationType::MapFirst) {
|
m_kernel.MemoryManager().Open(map_addr, num_pages);
|
||||||
m_kernel.MemoryManager().OpenFirst(map_addr, num_pages);
|
|
||||||
} else {
|
|
||||||
m_kernel.MemoryManager().Open(map_addr, num_pages);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -217,8 +217,8 @@ protected:
|
|||||||
private:
|
private:
|
||||||
enum class OperationType : u32 {
|
enum class OperationType : u32 {
|
||||||
Map = 0,
|
Map = 0,
|
||||||
MapFirst = 1,
|
MapGroup = 1,
|
||||||
MapGroup = 2,
|
MapFirstGroup = 2,
|
||||||
Unmap = 3,
|
Unmap = 3,
|
||||||
ChangePermissions = 4,
|
ChangePermissions = 4,
|
||||||
ChangePermissionsAndRefresh = 5,
|
ChangePermissionsAndRefresh = 5,
|
||||||
|
Loading…
Reference in New Issue
Block a user