Merge pull request #4122 from FearlessTobi/port-905
Port #905 from yuzu: "kernel/vm_manager: Minor changes"
This commit is contained in:
commit
cf4f8463f1
|
@ -2,6 +2,7 @@
|
||||||
// Licensed under GPLv2 or any later version
|
// Licensed under GPLv2 or any later version
|
||||||
// Refer to the license.txt file included.
|
// Refer to the license.txt file included.
|
||||||
|
|
||||||
|
#include <algorithm>
|
||||||
#include <iterator>
|
#include <iterator>
|
||||||
#include "common/assert.h"
|
#include "common/assert.h"
|
||||||
#include "core/hle/kernel/errors.h"
|
#include "core/hle/kernel/errors.h"
|
||||||
|
@ -209,9 +210,9 @@ VMManager::VMAIter VMManager::Unmap(VMAIter vma_handle) {
|
||||||
|
|
||||||
ResultCode VMManager::UnmapRange(VAddr target, u32 size) {
|
ResultCode VMManager::UnmapRange(VAddr target, u32 size) {
|
||||||
CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size));
|
CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size));
|
||||||
VAddr target_end = target + size;
|
const VAddr target_end = target + size;
|
||||||
|
|
||||||
VMAIter end = vma_map.end();
|
const VMAIter end = vma_map.end();
|
||||||
// The comparison against the end of the range must be done using addresses since VMAs can be
|
// The comparison against the end of the range must be done using addresses since VMAs can be
|
||||||
// merged during this process, causing invalidation of the iterators.
|
// merged during this process, causing invalidation of the iterators.
|
||||||
while (vma != end && vma->second.base < target_end) {
|
while (vma != end && vma->second.base < target_end) {
|
||||||
|
@ -234,9 +235,9 @@ VMManager::VMAHandle VMManager::Reprotect(VMAHandle vma_handle, VMAPermission ne
|
||||||
|
|
||||||
ResultCode VMManager::ReprotectRange(VAddr target, u32 size, VMAPermission new_perms) {
|
ResultCode VMManager::ReprotectRange(VAddr target, u32 size, VMAPermission new_perms) {
|
||||||
CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size));
|
CASCADE_RESULT(VMAIter vma, CarveVMARange(target, size));
|
||||||
VAddr target_end = target + size;
|
const VAddr target_end = target + size;
|
||||||
|
|
||||||
VMAIter end = vma_map.end();
|
const VMAIter end = vma_map.end();
|
||||||
// The comparison against the end of the range must be done using addresses since VMAs can be
|
// The comparison against the end of the range must be done using addresses since VMAs can be
|
||||||
// merged during this process, causing invalidation of the iterators.
|
// merged during this process, causing invalidation of the iterators.
|
||||||
while (vma != end && vma->second.base < target_end) {
|
while (vma != end && vma->second.base < target_end) {
|
||||||
|
@ -285,14 +286,14 @@ ResultVal<VMManager::VMAIter> VMManager::CarveVMA(VAddr base, u32 size) {
|
||||||
return ERR_INVALID_ADDRESS;
|
return ERR_INVALID_ADDRESS;
|
||||||
}
|
}
|
||||||
|
|
||||||
VirtualMemoryArea& vma = vma_handle->second;
|
const VirtualMemoryArea& vma = vma_handle->second;
|
||||||
if (vma.type != VMAType::Free) {
|
if (vma.type != VMAType::Free) {
|
||||||
// Region is already allocated
|
// Region is already allocated
|
||||||
return ERR_INVALID_ADDRESS_STATE;
|
return ERR_INVALID_ADDRESS_STATE;
|
||||||
}
|
}
|
||||||
|
|
||||||
u32 start_in_vma = base - vma.base;
|
const VAddr start_in_vma = base - vma.base;
|
||||||
u32 end_in_vma = start_in_vma + size;
|
const VAddr end_in_vma = start_in_vma + size;
|
||||||
|
|
||||||
if (end_in_vma > vma.size) {
|
if (end_in_vma > vma.size) {
|
||||||
// Requested allocation doesn't fit inside VMA
|
// Requested allocation doesn't fit inside VMA
|
||||||
|
@ -315,17 +316,16 @@ ResultVal<VMManager::VMAIter> VMManager::CarveVMARange(VAddr target, u32 size) {
|
||||||
ASSERT_MSG((size & Memory::PAGE_MASK) == 0, "non-page aligned size: {:#10X}", size);
|
ASSERT_MSG((size & Memory::PAGE_MASK) == 0, "non-page aligned size: {:#10X}", size);
|
||||||
ASSERT_MSG((target & Memory::PAGE_MASK) == 0, "non-page aligned base: {:#010X}", target);
|
ASSERT_MSG((target & Memory::PAGE_MASK) == 0, "non-page aligned base: {:#010X}", target);
|
||||||
|
|
||||||
VAddr target_end = target + size;
|
const VAddr target_end = target + size;
|
||||||
ASSERT(target_end >= target);
|
ASSERT(target_end >= target);
|
||||||
ASSERT(target_end <= MAX_ADDRESS);
|
ASSERT(target_end <= MAX_ADDRESS);
|
||||||
ASSERT(size > 0);
|
ASSERT(size > 0);
|
||||||
|
|
||||||
VMAIter begin_vma = StripIterConstness(FindVMA(target));
|
VMAIter begin_vma = StripIterConstness(FindVMA(target));
|
||||||
VMAIter i_end = vma_map.lower_bound(target_end);
|
const VMAIter i_end = vma_map.lower_bound(target_end);
|
||||||
for (auto i = begin_vma; i != i_end; ++i) {
|
if (std::any_of(begin_vma, i_end,
|
||||||
if (i->second.type == VMAType::Free) {
|
[](const auto& entry) { return entry.second.type == VMAType::Free; })) {
|
||||||
return ERR_INVALID_ADDRESS_STATE;
|
return ERR_INVALID_ADDRESS_STATE;
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (target != begin_vma->second.base) {
|
if (target != begin_vma->second.base) {
|
||||||
|
@ -373,7 +373,7 @@ VMManager::VMAIter VMManager::SplitVMA(VMAIter vma_handle, u32 offset_in_vma) {
|
||||||
}
|
}
|
||||||
|
|
||||||
VMManager::VMAIter VMManager::MergeAdjacent(VMAIter iter) {
|
VMManager::VMAIter VMManager::MergeAdjacent(VMAIter iter) {
|
||||||
VMAIter next_vma = std::next(iter);
|
const VMAIter next_vma = std::next(iter);
|
||||||
if (next_vma != vma_map.end() && iter->second.CanBeMergedWith(next_vma->second)) {
|
if (next_vma != vma_map.end() && iter->second.CanBeMergedWith(next_vma->second)) {
|
||||||
iter->second.size += next_vma->second.size;
|
iter->second.size += next_vma->second.size;
|
||||||
vma_map.erase(next_vma);
|
vma_map.erase(next_vma);
|
||||||
|
|
Loading…
Reference in New Issue