2014-04-09 04:45:46 +05:30
|
|
|
// Copyright 2014 Citra Emulator Project
|
|
|
|
// Licensed under GPLv2
|
|
|
|
// Refer to the license.txt file included.
|
2013-09-19 09:22:51 +05:30
|
|
|
|
2014-04-18 08:35:31 +05:30
|
|
|
#include <map>
|
|
|
|
|
2014-04-09 05:45:08 +05:30
|
|
|
#include "common/common.h"
|
2013-09-19 09:22:51 +05:30
|
|
|
|
2014-04-09 05:45:08 +05:30
|
|
|
#include "core/mem_map.h"
|
|
|
|
#include "core/hw/hw.h"
|
2014-04-13 07:25:36 +05:30
|
|
|
#include "hle/hle.h"
|
2014-05-07 09:02:04 +05:30
|
|
|
#include "hle/config_mem.h"
|
2013-09-19 09:22:51 +05:30
|
|
|
|
|
|
|
namespace Memory {
|
|
|
|
|
2014-04-25 08:02:26 +05:30
|
|
|
std::map<u32, MemoryBlock> g_heap_map;
|
|
|
|
std::map<u32, MemoryBlock> g_heap_gsp_map;
|
|
|
|
std::map<u32, MemoryBlock> g_shared_map;
|
2014-04-18 08:35:31 +05:30
|
|
|
|
2014-04-28 08:02:51 +05:30
|
|
|
/// Convert a physical address (or firmware-specific virtual address) to primary virtual address
|
|
|
|
u32 _VirtualAddress(const u32 addr) {
|
2014-04-18 04:10:42 +05:30
|
|
|
// Our memory interface read/write functions assume virtual addresses. Put any physical address
|
|
|
|
// to virtual address translations here. This is obviously quite hacky... But we're not doing
|
|
|
|
// any MMU emulation yet or anything
|
2014-04-26 23:51:40 +05:30
|
|
|
if ((addr >= FCRAM_PADDR) && (addr < FCRAM_PADDR_END)) {
|
2014-04-27 22:10:31 +05:30
|
|
|
return VirtualAddressFromPhysical_FCRAM(addr);
|
2014-04-26 23:51:40 +05:30
|
|
|
|
2014-04-28 08:02:51 +05:30
|
|
|
// Virtual address mapping FW0B
|
2014-04-30 08:01:24 +05:30
|
|
|
} else if ((addr >= FCRAM_VADDR_FW0B) && (addr < FCRAM_VADDR_FW0B_END)) {
|
2014-04-28 08:02:51 +05:30
|
|
|
return VirtualAddressFromPhysical_FCRAM(addr);
|
|
|
|
|
2014-04-26 23:51:40 +05:30
|
|
|
// Hardware IO
|
|
|
|
// TODO(bunnei): FixMe
|
|
|
|
// This isn't going to work... The physical address of HARDWARE_IO conflicts with the virtual
|
|
|
|
// address of shared memory.
|
|
|
|
//} else if ((addr >= HARDWARE_IO_PADDR) && (addr < HARDWARE_IO_PADDR_END)) {
|
|
|
|
// return (addr + 0x0EB00000);
|
|
|
|
|
2014-04-18 04:10:42 +05:30
|
|
|
}
|
|
|
|
return addr;
|
|
|
|
}
|
|
|
|
|
2013-09-19 09:22:51 +05:30
|
|
|
template <typename T>
|
2014-07-05 09:16:16 +05:30
|
|
|
inline void Read(T &var, const u32 addr) {
|
2014-04-02 03:48:02 +05:30
|
|
|
// TODO: Figure out the fastest order of tests for both read and write (they are probably different).
|
|
|
|
// TODO: Make sure this represents the mirrors in a correct way.
|
|
|
|
// Could just do a base-relative read, too.... TODO
|
2013-09-19 09:22:51 +05:30
|
|
|
|
2014-04-28 08:02:51 +05:30
|
|
|
const u32 vaddr = _VirtualAddress(addr);
|
2014-05-07 09:02:04 +05:30
|
|
|
|
2014-05-08 06:34:55 +05:30
|
|
|
// Kernel memory command buffer
|
|
|
|
if (vaddr >= KERNEL_MEMORY_VADDR && vaddr < KERNEL_MEMORY_VADDR_END) {
|
|
|
|
var = *((const T*)&g_kernel_mem[vaddr & KERNEL_MEMORY_MASK]);
|
2014-04-13 07:25:36 +05:30
|
|
|
|
2014-04-05 09:31:07 +05:30
|
|
|
// Hardware I/O register reads
|
|
|
|
// 0x10XXXXXX- is physical address space, 0x1EXXXXXX is virtual address space
|
2014-04-25 09:26:06 +05:30
|
|
|
} else if ((vaddr >= HARDWARE_IO_VADDR) && (vaddr < HARDWARE_IO_VADDR_END)) {
|
2014-04-18 04:10:42 +05:30
|
|
|
HW::Read<T>(var, vaddr);
|
|
|
|
|
2014-04-30 08:46:12 +05:30
|
|
|
// ExeFS:/.code is loaded here
|
|
|
|
} else if ((vaddr >= EXEFS_CODE_VADDR) && (vaddr < EXEFS_CODE_VADDR_END)) {
|
|
|
|
var = *((const T*)&g_exefs_code[vaddr & EXEFS_CODE_MASK]);
|
|
|
|
|
2014-04-18 06:45:40 +05:30
|
|
|
// FCRAM - GSP heap
|
2014-04-25 09:26:06 +05:30
|
|
|
} else if ((vaddr >= HEAP_GSP_VADDR) && (vaddr < HEAP_GSP_VADDR_END)) {
|
2014-04-18 06:45:40 +05:30
|
|
|
var = *((const T*)&g_heap_gsp[vaddr & HEAP_GSP_MASK]);
|
|
|
|
|
2014-04-18 06:35:34 +05:30
|
|
|
// FCRAM - application heap
|
2014-04-25 09:26:06 +05:30
|
|
|
} else if ((vaddr >= HEAP_VADDR) && (vaddr < HEAP_VADDR_END)) {
|
2014-04-18 06:35:34 +05:30
|
|
|
var = *((const T*)&g_heap[vaddr & HEAP_MASK]);
|
2014-04-18 04:10:42 +05:30
|
|
|
|
2014-04-25 09:26:06 +05:30
|
|
|
// Shared memory
|
|
|
|
} else if ((vaddr >= SHARED_MEMORY_VADDR) && (vaddr < SHARED_MEMORY_VADDR_END)) {
|
|
|
|
var = *((const T*)&g_shared_mem[vaddr & SHARED_MEMORY_MASK]);
|
|
|
|
|
2014-05-16 04:26:28 +05:30
|
|
|
// System memory
|
|
|
|
} else if ((vaddr >= SYSTEM_MEMORY_VADDR) && (vaddr < SYSTEM_MEMORY_VADDR_END)) {
|
|
|
|
var = *((const T*)&g_system_mem[vaddr & SYSTEM_MEMORY_MASK]);
|
|
|
|
|
2014-05-07 09:02:04 +05:30
|
|
|
// Config memory
|
|
|
|
} else if ((vaddr >= CONFIG_MEMORY_VADDR) && (vaddr < CONFIG_MEMORY_VADDR_END)) {
|
|
|
|
ConfigMem::Read<T>(var, vaddr);
|
|
|
|
|
2014-04-26 10:57:25 +05:30
|
|
|
// VRAM
|
|
|
|
} else if ((vaddr >= VRAM_VADDR) && (vaddr < VRAM_VADDR_END)) {
|
|
|
|
var = *((const T*)&g_vram[vaddr & VRAM_MASK]);
|
|
|
|
|
2014-04-04 07:34:50 +05:30
|
|
|
} else {
|
2014-05-30 08:56:58 +05:30
|
|
|
ERROR_LOG(MEMMAP, "unknown Read%d @ 0x%08X", sizeof(var) * 8, vaddr);
|
2014-04-02 03:48:02 +05:30
|
|
|
}
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
|
|
|
template <typename T>
|
2014-07-05 09:16:16 +05:30
|
|
|
inline void Write(u32 addr, const T data) {
|
2014-04-28 08:02:51 +05:30
|
|
|
u32 vaddr = _VirtualAddress(addr);
|
2014-04-05 09:31:07 +05:30
|
|
|
|
2014-05-08 06:34:55 +05:30
|
|
|
// Kernel memory command buffer
|
|
|
|
if (vaddr >= KERNEL_MEMORY_VADDR && vaddr < KERNEL_MEMORY_VADDR_END) {
|
|
|
|
*(T*)&g_kernel_mem[vaddr & KERNEL_MEMORY_MASK] = data;
|
2014-04-13 07:25:36 +05:30
|
|
|
|
2014-04-05 09:31:07 +05:30
|
|
|
// Hardware I/O register writes
|
|
|
|
// 0x10XXXXXX- is physical address space, 0x1EXXXXXX is virtual address space
|
2014-04-25 09:26:06 +05:30
|
|
|
} else if ((vaddr >= HARDWARE_IO_VADDR) && (vaddr < HARDWARE_IO_VADDR_END)) {
|
2014-04-18 04:10:42 +05:30
|
|
|
HW::Write<T>(vaddr, data);
|
2014-04-18 06:35:34 +05:30
|
|
|
|
2014-04-30 08:46:12 +05:30
|
|
|
// ExeFS:/.code is loaded here
|
|
|
|
} else if ((vaddr >= EXEFS_CODE_VADDR) && (vaddr < EXEFS_CODE_VADDR_END)) {
|
|
|
|
*(T*)&g_exefs_code[vaddr & EXEFS_CODE_MASK] = data;
|
|
|
|
|
2014-04-18 06:35:34 +05:30
|
|
|
// FCRAM - GSP heap
|
2014-04-25 09:26:06 +05:30
|
|
|
} else if ((vaddr >= HEAP_GSP_VADDR) && (vaddr < HEAP_GSP_VADDR_END)) {
|
2014-04-18 06:45:40 +05:30
|
|
|
*(T*)&g_heap_gsp[vaddr & HEAP_GSP_MASK] = data;
|
2014-04-18 06:35:34 +05:30
|
|
|
|
|
|
|
// FCRAM - application heap
|
2014-04-25 09:26:06 +05:30
|
|
|
} else if ((vaddr >= HEAP_VADDR) && (vaddr < HEAP_VADDR_END)) {
|
2014-04-18 06:35:34 +05:30
|
|
|
*(T*)&g_heap[vaddr & HEAP_MASK] = data;
|
2014-04-04 07:34:50 +05:30
|
|
|
|
2014-04-25 09:26:06 +05:30
|
|
|
// Shared memory
|
|
|
|
} else if ((vaddr >= SHARED_MEMORY_VADDR) && (vaddr < SHARED_MEMORY_VADDR_END)) {
|
|
|
|
*(T*)&g_shared_mem[vaddr & SHARED_MEMORY_MASK] = data;
|
|
|
|
|
2014-05-16 04:26:28 +05:30
|
|
|
// System memory
|
|
|
|
} else if ((vaddr >= SYSTEM_MEMORY_VADDR) && (vaddr < SYSTEM_MEMORY_VADDR_END)) {
|
|
|
|
*(T*)&g_system_mem[vaddr & SYSTEM_MEMORY_MASK] = data;
|
|
|
|
|
2014-04-26 10:57:25 +05:30
|
|
|
// VRAM
|
|
|
|
} else if ((vaddr >= VRAM_VADDR) && (vaddr < VRAM_VADDR_END)) {
|
|
|
|
*(T*)&g_vram[vaddr & VRAM_MASK] = data;
|
|
|
|
|
2014-05-07 09:02:04 +05:30
|
|
|
//} else if ((vaddr & 0xFFF00000) == 0x1FF00000) {
|
|
|
|
// _assert_msg_(MEMMAP, false, "umimplemented write to DSP memory");
|
|
|
|
//} else if ((vaddr & 0xFFFF0000) == 0x1FF80000) {
|
|
|
|
// _assert_msg_(MEMMAP, false, "umimplemented write to Configuration Memory");
|
|
|
|
//} else if ((vaddr & 0xFFFFF000) == 0x1FF81000) {
|
|
|
|
// _assert_msg_(MEMMAP, false, "umimplemented write to shared page");
|
2014-04-04 07:34:50 +05:30
|
|
|
|
|
|
|
// Error out...
|
2014-04-02 03:48:02 +05:30
|
|
|
} else {
|
2014-05-30 08:56:58 +05:30
|
|
|
ERROR_LOG(MEMMAP, "unknown Write%d 0x%08X @ 0x%08X", sizeof(data) * 8, data, vaddr);
|
2014-04-02 03:48:02 +05:30
|
|
|
}
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2014-03-25 20:20:34 +05:30
|
|
|
u8 *GetPointer(const u32 addr) {
|
2014-04-28 08:02:51 +05:30
|
|
|
const u32 vaddr = _VirtualAddress(addr);
|
2014-04-18 04:10:42 +05:30
|
|
|
|
2014-05-08 06:34:55 +05:30
|
|
|
// Kernel memory command buffer
|
|
|
|
if (vaddr >= KERNEL_MEMORY_VADDR && vaddr < KERNEL_MEMORY_VADDR_END) {
|
|
|
|
return g_kernel_mem + (vaddr & KERNEL_MEMORY_MASK);
|
|
|
|
|
2014-04-30 08:46:12 +05:30
|
|
|
// ExeFS:/.code is loaded here
|
2014-05-08 06:34:55 +05:30
|
|
|
} else if ((vaddr >= EXEFS_CODE_VADDR) && (vaddr < EXEFS_CODE_VADDR_END)) {
|
2014-04-30 08:46:12 +05:30
|
|
|
return g_exefs_code + (vaddr & EXEFS_CODE_MASK);
|
|
|
|
|
2014-04-18 06:45:40 +05:30
|
|
|
// FCRAM - GSP heap
|
2014-04-30 08:46:12 +05:30
|
|
|
} else if ((vaddr >= HEAP_GSP_VADDR) && (vaddr < HEAP_GSP_VADDR_END)) {
|
2014-04-18 06:45:40 +05:30
|
|
|
return g_heap_gsp + (vaddr & HEAP_GSP_MASK);
|
|
|
|
|
2014-04-18 06:35:34 +05:30
|
|
|
// FCRAM - application heap
|
2014-04-18 07:10:42 +05:30
|
|
|
} else if ((vaddr >= HEAP_VADDR) && (vaddr < HEAP_VADDR_END)) {
|
2014-04-18 06:35:34 +05:30
|
|
|
return g_heap + (vaddr & HEAP_MASK);
|
2014-04-18 04:10:42 +05:30
|
|
|
|
2014-04-25 09:26:06 +05:30
|
|
|
// Shared memory
|
2014-05-16 04:26:28 +05:30
|
|
|
} else if ((vaddr >= SHARED_MEMORY_VADDR) && (vaddr < SHARED_MEMORY_VADDR_END)) {
|
2014-04-25 09:26:06 +05:30
|
|
|
return g_shared_mem + (vaddr & SHARED_MEMORY_MASK);
|
|
|
|
|
2014-05-16 04:26:28 +05:30
|
|
|
// System memory
|
|
|
|
} else if ((vaddr >= SYSTEM_MEMORY_VADDR) && (vaddr < SYSTEM_MEMORY_VADDR_END)) {
|
|
|
|
return g_system_mem + (vaddr & SYSTEM_MEMORY_MASK);
|
|
|
|
|
2014-04-26 10:57:25 +05:30
|
|
|
// VRAM
|
|
|
|
} else if ((vaddr > VRAM_VADDR) && (vaddr < VRAM_VADDR_END)) {
|
|
|
|
return g_vram + (vaddr & VRAM_MASK);
|
|
|
|
|
2014-04-07 08:26:08 +05:30
|
|
|
} else {
|
2014-04-26 23:51:40 +05:30
|
|
|
ERROR_LOG(MEMMAP, "unknown GetPointer @ 0x%08x", vaddr);
|
2014-04-02 03:48:02 +05:30
|
|
|
return 0;
|
|
|
|
}
|
2014-03-25 20:20:34 +05:30
|
|
|
}
|
|
|
|
|
2014-04-25 08:02:26 +05:30
|
|
|
/**
|
|
|
|
* Maps a block of memory on the heap
|
|
|
|
* @param size Size of block in bytes
|
|
|
|
* @param operation Memory map operation type
|
|
|
|
* @param flags Memory allocation flags
|
|
|
|
*/
|
|
|
|
u32 MapBlock_Heap(u32 size, u32 operation, u32 permissions) {
|
|
|
|
MemoryBlock block;
|
|
|
|
|
|
|
|
block.base_address = HEAP_VADDR;
|
|
|
|
block.size = size;
|
|
|
|
block.operation = operation;
|
|
|
|
block.permissions = permissions;
|
|
|
|
|
|
|
|
if (g_heap_map.size() > 0) {
|
|
|
|
const MemoryBlock last_block = g_heap_map.rbegin()->second;
|
|
|
|
block.address = last_block.address + last_block.size;
|
|
|
|
}
|
|
|
|
g_heap_map[block.GetVirtualAddress()] = block;
|
|
|
|
|
|
|
|
return block.GetVirtualAddress();
|
|
|
|
}
|
|
|
|
|
2014-04-18 08:35:31 +05:30
|
|
|
/**
|
|
|
|
* Maps a block of memory on the GSP heap
|
|
|
|
* @param size Size of block in bytes
|
2014-04-25 08:02:26 +05:30
|
|
|
* @param operation Memory map operation type
|
2014-04-18 08:35:31 +05:30
|
|
|
* @param flags Memory allocation flags
|
|
|
|
*/
|
|
|
|
u32 MapBlock_HeapGSP(u32 size, u32 operation, u32 permissions) {
|
2014-04-25 08:02:26 +05:30
|
|
|
MemoryBlock block;
|
2014-04-18 08:35:31 +05:30
|
|
|
|
|
|
|
block.base_address = HEAP_GSP_VADDR;
|
|
|
|
block.size = size;
|
|
|
|
block.operation = operation;
|
|
|
|
block.permissions = permissions;
|
|
|
|
|
|
|
|
if (g_heap_gsp_map.size() > 0) {
|
2014-04-25 08:02:26 +05:30
|
|
|
const MemoryBlock last_block = g_heap_gsp_map.rbegin()->second;
|
2014-04-18 08:35:31 +05:30
|
|
|
block.address = last_block.address + last_block.size;
|
|
|
|
}
|
|
|
|
g_heap_gsp_map[block.GetVirtualAddress()] = block;
|
|
|
|
|
|
|
|
return block.GetVirtualAddress();
|
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
u8 Read8(const u32 addr) {
|
2014-07-25 05:16:10 +05:30
|
|
|
u8 data = 0;
|
|
|
|
Read<u8>(data, addr);
|
|
|
|
return (u8)data;
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
u16 Read16(const u32 addr) {
|
2014-07-25 05:16:10 +05:30
|
|
|
u16_le data = 0;
|
|
|
|
Read<u16_le>(data, addr);
|
2014-08-06 05:35:06 +05:30
|
|
|
|
|
|
|
// Check for 16-bit unaligned memory reads...
|
|
|
|
if (addr & 1) {
|
|
|
|
// TODO(bunnei): Implement 16-bit unaligned memory reads
|
|
|
|
ERROR_LOG(MEMMAP, "16-bit unaligned memory reads are not implemented!");
|
|
|
|
}
|
|
|
|
|
2014-07-25 05:16:10 +05:30
|
|
|
return (u16)data;
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
u32 Read32(const u32 addr) {
|
2014-07-25 05:16:10 +05:30
|
|
|
u32_le data = 0;
|
|
|
|
Read<u32_le>(data, addr);
|
|
|
|
|
|
|
|
// Check for 32-bit unaligned memory reads...
|
|
|
|
if (addr & 3) {
|
|
|
|
// ARM allows for unaligned memory reads, however older ARM architectures read out memory
|
|
|
|
// from unaligned addresses in a shifted way. Our ARM CPU core (SkyEye) corrects for this,
|
|
|
|
// so therefore expects the memory to be read out in this manner.
|
|
|
|
// TODO(bunnei): Determine if this is necessary - perhaps it is OK to remove this from both
|
|
|
|
// SkyEye and here?
|
|
|
|
int shift = (addr & 3) * 8;
|
|
|
|
data = (data << shift) | (data >> (32 - shift));
|
|
|
|
}
|
|
|
|
return (u32)data;
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
u32 Read8_ZX(const u32 addr) {
|
2014-04-02 03:48:02 +05:30
|
|
|
return (u32)Read8(addr);
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
u32 Read16_ZX(const u32 addr) {
|
2014-04-02 03:48:02 +05:30
|
|
|
return (u32)Read16(addr);
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
void Write8(const u32 addr, const u8 data) {
|
2014-07-05 09:16:16 +05:30
|
|
|
Write<u8>(addr, data);
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
void Write16(const u32 addr, const u16 data) {
|
2014-07-05 09:16:16 +05:30
|
|
|
Write<u16_le>(addr, data);
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
void Write32(const u32 addr, const u32 data) {
|
2014-07-05 09:16:16 +05:30
|
|
|
Write<u32_le>(addr, data);
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2013-09-20 08:43:33 +05:30
|
|
|
void Write64(const u32 addr, const u64 data) {
|
2014-07-05 09:16:16 +05:30
|
|
|
Write<u64_le>(addr, data);
|
2013-09-19 09:22:51 +05:30
|
|
|
}
|
|
|
|
|
2014-06-25 04:21:31 +05:30
|
|
|
void WriteBlock(const u32 addr, const u8* data, const int size) {
|
|
|
|
int offset = 0;
|
|
|
|
while (offset < (size & ~3))
|
|
|
|
Write32(addr + offset, *(u32*)&data[offset += 4]);
|
|
|
|
|
|
|
|
if (size & 2)
|
|
|
|
Write16(addr + offset, *(u16*)&data[offset += 2]);
|
|
|
|
|
|
|
|
if (size & 1)
|
|
|
|
Write8(addr + offset, data[offset]);
|
|
|
|
}
|
|
|
|
|
2013-09-19 09:22:51 +05:30
|
|
|
} // namespace
|