aboutsummaryrefslogtreecommitdiff
path: root/Ryujinx.HLE/HOS/Kernel/KMemoryRegionManager.cs
diff options
context:
space:
mode:
authorgdkchan <gab.dark.100@gmail.com>2018-11-28 20:18:09 -0200
committerGitHub <noreply@github.com>2018-11-28 20:18:09 -0200
commit00579927e43bf55ee06ae02933c1e755fb4120eb (patch)
tree0fd06db7b28e0accf87b465ec6f4dc74691febab /Ryujinx.HLE/HOS/Kernel/KMemoryRegionManager.cs
parente7fe7d724778535f8eff390abef54274a343c0b7 (diff)
Better process implementation (#491)
* Initial implementation of KProcess * Some improvements to the memory manager, implement back guest stack trace printing * Better GetInfo implementation, improve checking in some places with information from process capabilities * Allow the cpu to read/write from the correct memory locations for accesses crossing a page boundary * Change long -> ulong for address/size on memory related methods to avoid unnecessary casts * Attempt at implementing ldr:ro with new KProcess * Allow BSS with size 0 on ldr:ro * Add checking for memory block slab heap usage, return errors if full, exit gracefully * Use KMemoryBlockSize const from KMemoryManager * Allow all methods to read from non-contiguous locations * Fix for TransactParcelAuto * Address PR feedback, additionally fix some small issues related to the KIP loader and implement SVCs GetProcessId, GetProcessList, GetSystemInfo, CreatePort and ManageNamedPort * Fix wrong check for source pages count from page list on MapPhysicalMemory * Fix some issues with UnloadNro on ldr:ro
Diffstat (limited to 'Ryujinx.HLE/HOS/Kernel/KMemoryRegionManager.cs')
-rw-r--r--Ryujinx.HLE/HOS/Kernel/KMemoryRegionManager.cs428
1 files changed, 428 insertions, 0 deletions
diff --git a/Ryujinx.HLE/HOS/Kernel/KMemoryRegionManager.cs b/Ryujinx.HLE/HOS/Kernel/KMemoryRegionManager.cs
new file mode 100644
index 00000000..10db0753
--- /dev/null
+++ b/Ryujinx.HLE/HOS/Kernel/KMemoryRegionManager.cs
@@ -0,0 +1,428 @@
+using Ryujinx.Common;
+
+namespace Ryujinx.HLE.HOS.Kernel
+{
+ class KMemoryRegionManager
+ {
+ private static readonly int[] BlockOrders = new int[] { 12, 16, 21, 22, 25, 29, 30 };
+
+ public ulong Address { get; private set; }
+ public ulong EndAddr { get; private set; }
+ public ulong Size { get; private set; }
+
+ private int BlockOrdersCount;
+
+ private KMemoryRegionBlock[] Blocks;
+
+ public KMemoryRegionManager(ulong Address, ulong Size, ulong EndAddr)
+ {
+ Blocks = new KMemoryRegionBlock[BlockOrders.Length];
+
+ this.Address = Address;
+ this.Size = Size;
+ this.EndAddr = EndAddr;
+
+ BlockOrdersCount = BlockOrders.Length;
+
+ for (int BlockIndex = 0; BlockIndex < BlockOrdersCount; BlockIndex++)
+ {
+ Blocks[BlockIndex] = new KMemoryRegionBlock();
+
+ Blocks[BlockIndex].Order = BlockOrders[BlockIndex];
+
+ int NextOrder = BlockIndex == BlockOrdersCount - 1 ? 0 : BlockOrders[BlockIndex + 1];
+
+ Blocks[BlockIndex].NextOrder = NextOrder;
+
+ int CurrBlockSize = 1 << BlockOrders[BlockIndex];
+ int NextBlockSize = CurrBlockSize;
+
+ if (NextOrder != 0)
+ {
+ NextBlockSize = 1 << NextOrder;
+ }
+
+ ulong StartAligned = BitUtils.AlignDown(Address, NextBlockSize);
+ ulong EndAddrAligned = BitUtils.AlignDown(EndAddr, CurrBlockSize);
+
+ ulong SizeInBlocksTruncated = (EndAddrAligned - StartAligned) >> BlockOrders[BlockIndex];
+
+ ulong EndAddrRounded = BitUtils.AlignUp(Address + Size, NextBlockSize);
+
+ ulong SizeInBlocksRounded = (EndAddrRounded - StartAligned) >> BlockOrders[BlockIndex];
+
+ Blocks[BlockIndex].StartAligned = StartAligned;
+ Blocks[BlockIndex].SizeInBlocksTruncated = SizeInBlocksTruncated;
+ Blocks[BlockIndex].SizeInBlocksRounded = SizeInBlocksRounded;
+
+ ulong CurrSizeInBlocks = SizeInBlocksRounded;
+
+ int MaxLevel = 0;
+
+ do
+ {
+ MaxLevel++;
+ }
+ while ((CurrSizeInBlocks /= 64) != 0);
+
+ Blocks[BlockIndex].MaxLevel = MaxLevel;
+
+ Blocks[BlockIndex].Masks = new long[MaxLevel][];
+
+ CurrSizeInBlocks = SizeInBlocksRounded;
+
+ for (int Level = MaxLevel - 1; Level >= 0; Level--)
+ {
+ CurrSizeInBlocks = (CurrSizeInBlocks + 63) / 64;
+
+ Blocks[BlockIndex].Masks[Level] = new long[CurrSizeInBlocks];
+ }
+ }
+
+ if (Size != 0)
+ {
+ FreePages(Address, Size / KMemoryManager.PageSize);
+ }
+ }
+
+ public KernelResult AllocatePages(ulong PagesCount, bool Backwards, out KPageList PageList)
+ {
+ lock (Blocks)
+ {
+ return AllocatePagesImpl(PagesCount, Backwards, out PageList);
+ }
+ }
+
+ private KernelResult AllocatePagesImpl(ulong PagesCount, bool Backwards, out KPageList PageList)
+ {
+ PageList = new KPageList();
+
+ if (BlockOrdersCount > 0)
+ {
+ if (GetFreePagesImpl() < PagesCount)
+ {
+ return KernelResult.OutOfMemory;
+ }
+ }
+ else if (PagesCount != 0)
+ {
+ return KernelResult.OutOfMemory;
+ }
+
+ for (int BlockIndex = BlockOrdersCount - 1; BlockIndex >= 0; BlockIndex--)
+ {
+ KMemoryRegionBlock Block = Blocks[BlockIndex];
+
+ ulong BestFitBlockSize = 1UL << Block.Order;
+
+ ulong BlockPagesCount = BestFitBlockSize / KMemoryManager.PageSize;
+
+ //Check if this is the best fit for this page size.
+ //If so, try allocating as much requested pages as possible.
+ while (BlockPagesCount <= PagesCount)
+ {
+ ulong Address = 0;
+
+ for (int CurrBlockIndex = BlockIndex;
+ CurrBlockIndex < BlockOrdersCount && Address == 0;
+ CurrBlockIndex++)
+ {
+ Block = Blocks[CurrBlockIndex];
+
+ int Index = 0;
+
+ bool ZeroMask = false;
+
+ for (int Level = 0; Level < Block.MaxLevel; Level++)
+ {
+ long Mask = Block.Masks[Level][Index];
+
+ if (Mask == 0)
+ {
+ ZeroMask = true;
+
+ break;
+ }
+
+ if (Backwards)
+ {
+ Index = (Index * 64 + 63) - BitUtils.CountLeadingZeros64(Mask);
+ }
+ else
+ {
+ Index = Index * 64 + BitUtils.CountLeadingZeros64(BitUtils.ReverseBits64(Mask));
+ }
+ }
+
+ if (Block.SizeInBlocksTruncated <= (ulong)Index || ZeroMask)
+ {
+ continue;
+ }
+
+ Block.FreeCount--;
+
+ int TempIdx = Index;
+
+ for (int Level = Block.MaxLevel - 1; Level >= 0; Level--, TempIdx /= 64)
+ {
+ Block.Masks[Level][TempIdx / 64] &= ~(1L << (TempIdx & 63));
+
+ if (Block.Masks[Level][TempIdx / 64] != 0)
+ {
+ break;
+ }
+ }
+
+ Address = Block.StartAligned + ((ulong)Index << Block.Order);
+ }
+
+ for (int CurrBlockIndex = BlockIndex;
+ CurrBlockIndex < BlockOrdersCount && Address == 0;
+ CurrBlockIndex++)
+ {
+ Block = Blocks[CurrBlockIndex];
+
+ int Index = 0;
+
+ bool ZeroMask = false;
+
+ for (int Level = 0; Level < Block.MaxLevel; Level++)
+ {
+ long Mask = Block.Masks[Level][Index];
+
+ if (Mask == 0)
+ {
+ ZeroMask = true;
+
+ break;
+ }
+
+ if (Backwards)
+ {
+ Index = Index * 64 + BitUtils.CountLeadingZeros64(BitUtils.ReverseBits64(Mask));
+ }
+ else
+ {
+ Index = (Index * 64 + 63) - BitUtils.CountLeadingZeros64(Mask);
+ }
+ }
+
+ if (Block.SizeInBlocksTruncated <= (ulong)Index || ZeroMask)
+ {
+ continue;
+ }
+
+ Block.FreeCount--;
+
+ int TempIdx = Index;
+
+ for (int Level = Block.MaxLevel - 1; Level >= 0; Level--, TempIdx /= 64)
+ {
+ Block.Masks[Level][TempIdx / 64] &= ~(1L << (TempIdx & 63));
+
+ if (Block.Masks[Level][TempIdx / 64] != 0)
+ {
+ break;
+ }
+ }
+
+ Address = Block.StartAligned + ((ulong)Index << Block.Order);
+ }
+
+ //The address being zero means that no free space was found on that order,
+ //just give up and try with the next one.
+ if (Address == 0)
+ {
+ break;
+ }
+
+ //If we are using a larger order than best fit, then we should
+ //split it into smaller blocks.
+ ulong FirstFreeBlockSize = 1UL << Block.Order;
+
+ if (FirstFreeBlockSize > BestFitBlockSize)
+ {
+ FreePages(Address + BestFitBlockSize, (FirstFreeBlockSize - BestFitBlockSize) / KMemoryManager.PageSize);
+ }
+
+ //Add new allocated page(s) to the pages list.
+ //If an error occurs, then free all allocated pages and fail.
+ KernelResult Result = PageList.AddRange(Address, BlockPagesCount);
+
+ if (Result != KernelResult.Success)
+ {
+ FreePages(Address, BlockPagesCount);
+
+ foreach (KPageNode PageNode in PageList)
+ {
+ FreePages(PageNode.Address, PageNode.PagesCount);
+ }
+
+ return Result;
+ }
+
+ PagesCount -= BlockPagesCount;
+ }
+ }
+
+ //Success case, all requested pages were allocated successfully.
+ if (PagesCount == 0)
+ {
+ return KernelResult.Success;
+ }
+
+ //Error case, free allocated pages and return out of memory.
+ foreach (KPageNode PageNode in PageList)
+ {
+ FreePages(PageNode.Address, PageNode.PagesCount);
+ }
+
+ PageList = null;
+
+ return KernelResult.OutOfMemory;
+ }
+
+ public void FreePages(KPageList PageList)
+ {
+ lock (Blocks)
+ {
+ foreach (KPageNode PageNode in PageList)
+ {
+ FreePages(PageNode.Address, PageNode.PagesCount);
+ }
+ }
+ }
+
+ private void FreePages(ulong Address, ulong PagesCount)
+ {
+ ulong EndAddr = Address + PagesCount * KMemoryManager.PageSize;
+
+ int BlockIndex = BlockOrdersCount - 1;
+
+ ulong AddressRounded = 0;
+ ulong EndAddrTruncated = 0;
+
+ for (; BlockIndex >= 0; BlockIndex--)
+ {
+ KMemoryRegionBlock AllocInfo = Blocks[BlockIndex];
+
+ int BlockSize = 1 << AllocInfo.Order;
+
+ AddressRounded = BitUtils.AlignUp (Address, BlockSize);
+ EndAddrTruncated = BitUtils.AlignDown(EndAddr, BlockSize);
+
+ if (AddressRounded < EndAddrTruncated)
+ {
+ break;
+ }
+ }
+
+ void FreeRegion(ulong CurrAddress)
+ {
+ for (int CurrBlockIndex = BlockIndex;
+ CurrBlockIndex < BlockOrdersCount && CurrAddress != 0;
+ CurrBlockIndex++)
+ {
+ KMemoryRegionBlock Block = Blocks[CurrBlockIndex];
+
+ Block.FreeCount++;
+
+ ulong FreedBlocks = (CurrAddress - Block.StartAligned) >> Block.Order;
+
+ int Index = (int)FreedBlocks;
+
+ for (int Level = Block.MaxLevel - 1; Level >= 0; Level--, Index /= 64)
+ {
+ long Mask = Block.Masks[Level][Index / 64];
+
+ Block.Masks[Level][Index / 64] = Mask | (1L << (Index & 63));
+
+ if (Mask != 0)
+ {
+ break;
+ }
+ }
+
+ int BlockSizeDelta = 1 << (Block.NextOrder - Block.Order);
+
+ int FreedBlocksTruncated = BitUtils.AlignDown((int)FreedBlocks, BlockSizeDelta);
+
+ if (!Block.TryCoalesce(FreedBlocksTruncated, BlockSizeDelta))
+ {
+ break;
+ }
+
+ CurrAddress = Block.StartAligned + ((ulong)FreedBlocksTruncated << Block.Order);
+ }
+ }
+
+ //Free inside aligned region.
+ ulong BaseAddress = AddressRounded;
+
+ while (BaseAddress < EndAddrTruncated)
+ {
+ ulong BlockSize = 1UL << Blocks[BlockIndex].Order;
+
+ FreeRegion(BaseAddress);
+
+ BaseAddress += BlockSize;
+ }
+
+ int NextBlockIndex = BlockIndex - 1;
+
+ //Free region between Address and aligned region start.
+ BaseAddress = AddressRounded;
+
+ for (BlockIndex = NextBlockIndex; BlockIndex >= 0; BlockIndex--)
+ {
+ ulong BlockSize = 1UL << Blocks[BlockIndex].Order;
+
+ while (BaseAddress - BlockSize >= Address)
+ {
+ BaseAddress -= BlockSize;
+
+ FreeRegion(BaseAddress);
+ }
+ }
+
+ //Free region between aligned region end and End Address.
+ BaseAddress = EndAddrTruncated;
+
+ for (BlockIndex = NextBlockIndex; BlockIndex >= 0; BlockIndex--)
+ {
+ ulong BlockSize = 1UL << Blocks[BlockIndex].Order;
+
+ while (BaseAddress + BlockSize <= EndAddr)
+ {
+ FreeRegion(BaseAddress);
+
+ BaseAddress += BlockSize;
+ }
+ }
+ }
+
+ public ulong GetFreePages()
+ {
+ lock (Blocks)
+ {
+ return GetFreePagesImpl();
+ }
+ }
+
+ private ulong GetFreePagesImpl()
+ {
+ ulong AvailablePages = 0;
+
+ for (int BlockIndex = 0; BlockIndex < BlockOrdersCount; BlockIndex++)
+ {
+ KMemoryRegionBlock Block = Blocks[BlockIndex];
+
+ ulong BlockPagesCount = (1UL << Block.Order) / KMemoryManager.PageSize;
+
+ AvailablePages += BlockPagesCount * Block.FreeCount;
+ }
+
+ return AvailablePages;
+ }
+ }
+} \ No newline at end of file