mirror of
https://github.com/qemu/qemu.git
synced 2025-08-14 11:43:46 +00:00
memory: make phys_page_find() return a MemoryRegionSection
We no longer describe memory in terms of individual pages; use sections throughout instead. PhysPageDesc no longer used - remove. Signed-off-by: Avi Kivity <avi@redhat.com>
This commit is contained in:
parent
117712c3e4
commit
06ef3525e1
299
exec.c
299
exec.c
@ -186,12 +186,6 @@ unsigned long qemu_host_page_mask;
|
|||||||
static void *l1_map[V_L1_SIZE];
|
static void *l1_map[V_L1_SIZE];
|
||||||
|
|
||||||
#if !defined(CONFIG_USER_ONLY)
|
#if !defined(CONFIG_USER_ONLY)
|
||||||
typedef struct PhysPageDesc {
|
|
||||||
/* offset in host memory of the page + io_index in the low bits */
|
|
||||||
ram_addr_t phys_offset;
|
|
||||||
ram_addr_t region_offset;
|
|
||||||
} PhysPageDesc;
|
|
||||||
|
|
||||||
typedef struct PhysPageEntry PhysPageEntry;
|
typedef struct PhysPageEntry PhysPageEntry;
|
||||||
|
|
||||||
static MemoryRegionSection *phys_sections;
|
static MemoryRegionSection *phys_sections;
|
||||||
@ -212,7 +206,7 @@ static unsigned phys_map_nodes_nb, phys_map_nodes_nb_alloc;
|
|||||||
#define PHYS_MAP_NODE_NIL ((uint16_t)~0)
|
#define PHYS_MAP_NODE_NIL ((uint16_t)~0)
|
||||||
|
|
||||||
/* This is a multi-level map on the physical address space.
|
/* This is a multi-level map on the physical address space.
|
||||||
The bottom level has pointers to PhysPageDesc. */
|
The bottom level has pointers to MemoryRegionSections. */
|
||||||
static PhysPageEntry phys_map = { .u.node = PHYS_MAP_NODE_NIL };
|
static PhysPageEntry phys_map = { .u.node = PHYS_MAP_NODE_NIL };
|
||||||
|
|
||||||
static void io_mem_init(void);
|
static void io_mem_init(void);
|
||||||
@ -463,33 +457,25 @@ static uint16_t *phys_page_find_alloc(target_phys_addr_t index, int alloc)
|
|||||||
return &lp->u.leaf;
|
return &lp->u.leaf;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline PhysPageDesc phys_page_find(target_phys_addr_t index)
|
static MemoryRegionSection phys_page_find(target_phys_addr_t index)
|
||||||
{
|
{
|
||||||
uint16_t *p = phys_page_find_alloc(index, 0);
|
uint16_t *p = phys_page_find_alloc(index, 0);
|
||||||
uint16_t s_index = phys_section_unassigned;
|
uint16_t s_index = phys_section_unassigned;
|
||||||
MemoryRegionSection *section;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc pd;
|
target_phys_addr_t delta;
|
||||||
|
|
||||||
if (p) {
|
if (p) {
|
||||||
s_index = *p;
|
s_index = *p;
|
||||||
}
|
}
|
||||||
section = &phys_sections[s_index];
|
section = phys_sections[s_index];
|
||||||
index <<= TARGET_PAGE_BITS;
|
index <<= TARGET_PAGE_BITS;
|
||||||
assert(section->offset_within_address_space <= index
|
assert(section.offset_within_address_space <= index
|
||||||
&& index <= section->offset_within_address_space + section->size-1);
|
&& index <= section.offset_within_address_space + section.size-1);
|
||||||
pd.phys_offset = section->mr->ram_addr;
|
delta = index - section.offset_within_address_space;
|
||||||
pd.region_offset = (index - section->offset_within_address_space)
|
section.offset_within_address_space += delta;
|
||||||
+ section->offset_within_region;
|
section.offset_within_region += delta;
|
||||||
if (memory_region_is_ram(section->mr)) {
|
section.size -= delta;
|
||||||
pd.phys_offset += pd.region_offset;
|
return section;
|
||||||
pd.region_offset = 0;
|
|
||||||
} else if (section->mr->rom_device) {
|
|
||||||
pd.phys_offset += pd.region_offset;
|
|
||||||
}
|
|
||||||
if (section->readonly) {
|
|
||||||
pd.phys_offset |= io_mem_rom.ram_addr;
|
|
||||||
}
|
|
||||||
return pd;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void tlb_protect_code(ram_addr_t ram_addr);
|
static void tlb_protect_code(ram_addr_t ram_addr);
|
||||||
@ -1449,14 +1435,18 @@ static void breakpoint_invalidate(CPUState *env, target_ulong pc)
|
|||||||
static void breakpoint_invalidate(CPUState *env, target_ulong pc)
|
static void breakpoint_invalidate(CPUState *env, target_ulong pc)
|
||||||
{
|
{
|
||||||
target_phys_addr_t addr;
|
target_phys_addr_t addr;
|
||||||
target_ulong pd;
|
|
||||||
ram_addr_t ram_addr;
|
ram_addr_t ram_addr;
|
||||||
PhysPageDesc p;
|
MemoryRegionSection section;
|
||||||
|
|
||||||
addr = cpu_get_phys_page_debug(env, pc);
|
addr = cpu_get_phys_page_debug(env, pc);
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
if (!(memory_region_is_ram(section.mr)
|
||||||
ram_addr = (pd & TARGET_PAGE_MASK) | (pc & ~TARGET_PAGE_MASK);
|
|| (section.mr->rom_device && section.mr->readable))) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
ram_addr = (memory_region_get_ram_addr(section.mr)
|
||||||
|
+ section.offset_within_region) & TARGET_PAGE_MASK;
|
||||||
|
ram_addr |= (pc & ~TARGET_PAGE_MASK);
|
||||||
tb_invalidate_phys_page_range(ram_addr, ram_addr + 1, 0);
|
tb_invalidate_phys_page_range(ram_addr, ram_addr + 1, 0);
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
@ -2134,24 +2124,21 @@ static void tlb_add_large_page(CPUState *env, target_ulong vaddr,
|
|||||||
env->tlb_flush_mask = mask;
|
env->tlb_flush_mask = mask;
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool is_ram_rom(ram_addr_t pd)
|
static bool is_ram_rom(MemoryRegionSection *s)
|
||||||
{
|
{
|
||||||
pd &= ~TARGET_PAGE_MASK;
|
return memory_region_is_ram(s->mr);
|
||||||
return pd == io_mem_ram.ram_addr || pd == io_mem_rom.ram_addr;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool is_romd(ram_addr_t pd)
|
static bool is_romd(MemoryRegionSection *s)
|
||||||
{
|
{
|
||||||
MemoryRegion *mr;
|
MemoryRegion *mr = s->mr;
|
||||||
|
|
||||||
pd &= ~TARGET_PAGE_MASK;
|
|
||||||
mr = io_mem_region[pd];
|
|
||||||
return mr->rom_device && mr->readable;
|
return mr->rom_device && mr->readable;
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool is_ram_rom_romd(ram_addr_t pd)
|
static bool is_ram_rom_romd(MemoryRegionSection *s)
|
||||||
{
|
{
|
||||||
return is_ram_rom(pd) || is_romd(pd);
|
return is_ram_rom(s) || is_romd(s);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Add a new TLB entry. At most one entry for a given virtual address
|
/* Add a new TLB entry. At most one entry for a given virtual address
|
||||||
@ -2161,8 +2148,7 @@ void tlb_set_page(CPUState *env, target_ulong vaddr,
|
|||||||
target_phys_addr_t paddr, int prot,
|
target_phys_addr_t paddr, int prot,
|
||||||
int mmu_idx, target_ulong size)
|
int mmu_idx, target_ulong size)
|
||||||
{
|
{
|
||||||
PhysPageDesc p;
|
MemoryRegionSection section;
|
||||||
unsigned long pd;
|
|
||||||
unsigned int index;
|
unsigned int index;
|
||||||
target_ulong address;
|
target_ulong address;
|
||||||
target_ulong code_address;
|
target_ulong code_address;
|
||||||
@ -2175,8 +2161,7 @@ void tlb_set_page(CPUState *env, target_ulong vaddr,
|
|||||||
if (size != TARGET_PAGE_SIZE) {
|
if (size != TARGET_PAGE_SIZE) {
|
||||||
tlb_add_large_page(env, vaddr, size);
|
tlb_add_large_page(env, vaddr, size);
|
||||||
}
|
}
|
||||||
p = phys_page_find(paddr >> TARGET_PAGE_BITS);
|
section = phys_page_find(paddr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
#if defined(DEBUG_TLB)
|
#if defined(DEBUG_TLB)
|
||||||
printf("tlb_set_page: vaddr=" TARGET_FMT_lx " paddr=0x" TARGET_FMT_plx
|
printf("tlb_set_page: vaddr=" TARGET_FMT_lx " paddr=0x" TARGET_FMT_plx
|
||||||
" prot=%x idx=%d pd=0x%08lx\n",
|
" prot=%x idx=%d pd=0x%08lx\n",
|
||||||
@ -2184,15 +2169,21 @@ void tlb_set_page(CPUState *env, target_ulong vaddr,
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
address = vaddr;
|
address = vaddr;
|
||||||
if (!is_ram_rom_romd(pd)) {
|
if (!is_ram_rom_romd(§ion)) {
|
||||||
/* IO memory case (romd handled later) */
|
/* IO memory case (romd handled later) */
|
||||||
address |= TLB_MMIO;
|
address |= TLB_MMIO;
|
||||||
}
|
}
|
||||||
addend = (unsigned long)qemu_get_ram_ptr(pd & TARGET_PAGE_MASK);
|
if (is_ram_rom_romd(§ion)) {
|
||||||
if (is_ram_rom(pd)) {
|
addend = (unsigned long)(memory_region_get_ram_ptr(section.mr)
|
||||||
|
+ section.offset_within_region);
|
||||||
|
} else {
|
||||||
|
addend = 0;
|
||||||
|
}
|
||||||
|
if (is_ram_rom(§ion)) {
|
||||||
/* Normal RAM. */
|
/* Normal RAM. */
|
||||||
iotlb = pd & TARGET_PAGE_MASK;
|
iotlb = (memory_region_get_ram_addr(section.mr)
|
||||||
if ((pd & ~TARGET_PAGE_MASK) == io_mem_ram.ram_addr)
|
+ section.offset_within_region) & TARGET_PAGE_MASK;
|
||||||
|
if (!section.readonly)
|
||||||
iotlb |= io_mem_notdirty.ram_addr;
|
iotlb |= io_mem_notdirty.ram_addr;
|
||||||
else
|
else
|
||||||
iotlb |= io_mem_rom.ram_addr;
|
iotlb |= io_mem_rom.ram_addr;
|
||||||
@ -2203,8 +2194,8 @@ void tlb_set_page(CPUState *env, target_ulong vaddr,
|
|||||||
and avoid full address decoding in every device.
|
and avoid full address decoding in every device.
|
||||||
We can't use the high bits of pd for this because
|
We can't use the high bits of pd for this because
|
||||||
IO_MEM_ROMD uses these as a ram address. */
|
IO_MEM_ROMD uses these as a ram address. */
|
||||||
iotlb = (pd & ~TARGET_PAGE_MASK);
|
iotlb = memory_region_get_ram_addr(section.mr) & ~TARGET_PAGE_MASK;
|
||||||
iotlb += p.region_offset;
|
iotlb += section.offset_within_region;
|
||||||
}
|
}
|
||||||
|
|
||||||
code_address = address;
|
code_address = address;
|
||||||
@ -2237,11 +2228,14 @@ void tlb_set_page(CPUState *env, target_ulong vaddr,
|
|||||||
te->addr_code = -1;
|
te->addr_code = -1;
|
||||||
}
|
}
|
||||||
if (prot & PAGE_WRITE) {
|
if (prot & PAGE_WRITE) {
|
||||||
if ((pd & ~TARGET_PAGE_MASK) == io_mem_rom.ram_addr || is_romd(pd)) {
|
if ((memory_region_is_ram(section.mr) && section.readonly)
|
||||||
|
|| is_romd(§ion)) {
|
||||||
/* Write access calls the I/O callback. */
|
/* Write access calls the I/O callback. */
|
||||||
te->addr_write = address | TLB_MMIO;
|
te->addr_write = address | TLB_MMIO;
|
||||||
} else if ((pd & ~TARGET_PAGE_MASK) == io_mem_ram.ram_addr &&
|
} else if (memory_region_is_ram(section.mr)
|
||||||
!cpu_physical_memory_is_dirty(pd)) {
|
&& !cpu_physical_memory_is_dirty(
|
||||||
|
section.mr->ram_addr
|
||||||
|
+ section.offset_within_region)) {
|
||||||
te->addr_write = address | TLB_NOTDIRTY;
|
te->addr_write = address | TLB_NOTDIRTY;
|
||||||
} else {
|
} else {
|
||||||
te->addr_write = address;
|
te->addr_write = address;
|
||||||
@ -3788,22 +3782,22 @@ void cpu_physical_memory_rw(target_phys_addr_t addr, uint8_t *buf,
|
|||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
uint32_t val;
|
uint32_t val;
|
||||||
target_phys_addr_t page;
|
target_phys_addr_t page;
|
||||||
ram_addr_t pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
while (len > 0) {
|
while (len > 0) {
|
||||||
page = addr & TARGET_PAGE_MASK;
|
page = addr & TARGET_PAGE_MASK;
|
||||||
l = (page + TARGET_PAGE_SIZE) - addr;
|
l = (page + TARGET_PAGE_SIZE) - addr;
|
||||||
if (l > len)
|
if (l > len)
|
||||||
l = len;
|
l = len;
|
||||||
p = phys_page_find(page >> TARGET_PAGE_BITS);
|
section = phys_page_find(page >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if (is_write) {
|
if (is_write) {
|
||||||
if ((pd & ~TARGET_PAGE_MASK) != io_mem_ram.ram_addr) {
|
if (!memory_region_is_ram(section.mr)) {
|
||||||
target_phys_addr_t addr1;
|
target_phys_addr_t addr1;
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
addr1 = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
addr1 = (addr & ~TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region;
|
||||||
/* XXX: could force cpu_single_env to NULL to avoid
|
/* XXX: could force cpu_single_env to NULL to avoid
|
||||||
potential bugs */
|
potential bugs */
|
||||||
if (l >= 4 && ((addr1 & 3) == 0)) {
|
if (l >= 4 && ((addr1 & 3) == 0)) {
|
||||||
@ -3822,9 +3816,11 @@ void cpu_physical_memory_rw(target_phys_addr_t addr, uint8_t *buf,
|
|||||||
io_mem_write(io_index, addr1, val, 1);
|
io_mem_write(io_index, addr1, val, 1);
|
||||||
l = 1;
|
l = 1;
|
||||||
}
|
}
|
||||||
} else {
|
} else if (!section.readonly) {
|
||||||
ram_addr_t addr1;
|
ram_addr_t addr1;
|
||||||
addr1 = (pd & TARGET_PAGE_MASK) + (addr & ~TARGET_PAGE_MASK);
|
addr1 = (memory_region_get_ram_addr(section.mr)
|
||||||
|
+ section.offset_within_region)
|
||||||
|
| (addr & ~TARGET_PAGE_MASK);
|
||||||
/* RAM case */
|
/* RAM case */
|
||||||
ptr = qemu_get_ram_ptr(addr1);
|
ptr = qemu_get_ram_ptr(addr1);
|
||||||
memcpy(ptr, buf, l);
|
memcpy(ptr, buf, l);
|
||||||
@ -3838,11 +3834,13 @@ void cpu_physical_memory_rw(target_phys_addr_t addr, uint8_t *buf,
|
|||||||
qemu_put_ram_ptr(ptr);
|
qemu_put_ram_ptr(ptr);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
if (!is_ram_rom_romd(pd)) {
|
if (!is_ram_rom_romd(§ion)) {
|
||||||
target_phys_addr_t addr1;
|
target_phys_addr_t addr1;
|
||||||
/* I/O case */
|
/* I/O case */
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
addr1 = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
addr1 = (addr & ~TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region;
|
||||||
if (l >= 4 && ((addr1 & 3) == 0)) {
|
if (l >= 4 && ((addr1 & 3) == 0)) {
|
||||||
/* 32 bit read access */
|
/* 32 bit read access */
|
||||||
val = io_mem_read(io_index, addr1, 4);
|
val = io_mem_read(io_index, addr1, 4);
|
||||||
@ -3861,7 +3859,8 @@ void cpu_physical_memory_rw(target_phys_addr_t addr, uint8_t *buf,
|
|||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
/* RAM case */
|
/* RAM case */
|
||||||
ptr = qemu_get_ram_ptr(pd & TARGET_PAGE_MASK);
|
ptr = qemu_get_ram_ptr(section.mr->ram_addr
|
||||||
|
+ section.offset_within_region);
|
||||||
memcpy(buf, ptr + (addr & ~TARGET_PAGE_MASK), l);
|
memcpy(buf, ptr + (addr & ~TARGET_PAGE_MASK), l);
|
||||||
qemu_put_ram_ptr(ptr);
|
qemu_put_ram_ptr(ptr);
|
||||||
}
|
}
|
||||||
@ -3879,22 +3878,22 @@ void cpu_physical_memory_write_rom(target_phys_addr_t addr,
|
|||||||
int l;
|
int l;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
target_phys_addr_t page;
|
target_phys_addr_t page;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
while (len > 0) {
|
while (len > 0) {
|
||||||
page = addr & TARGET_PAGE_MASK;
|
page = addr & TARGET_PAGE_MASK;
|
||||||
l = (page + TARGET_PAGE_SIZE) - addr;
|
l = (page + TARGET_PAGE_SIZE) - addr;
|
||||||
if (l > len)
|
if (l > len)
|
||||||
l = len;
|
l = len;
|
||||||
p = phys_page_find(page >> TARGET_PAGE_BITS);
|
section = phys_page_find(page >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if (!is_ram_rom_romd(pd)) {
|
if (!is_ram_rom_romd(§ion)) {
|
||||||
/* do nothing */
|
/* do nothing */
|
||||||
} else {
|
} else {
|
||||||
unsigned long addr1;
|
unsigned long addr1;
|
||||||
addr1 = (pd & TARGET_PAGE_MASK) + (addr & ~TARGET_PAGE_MASK);
|
addr1 = (memory_region_get_ram_addr(section.mr)
|
||||||
|
+ section.offset_within_region)
|
||||||
|
+ (addr & ~TARGET_PAGE_MASK);
|
||||||
/* ROM/RAM case */
|
/* ROM/RAM case */
|
||||||
ptr = qemu_get_ram_ptr(addr1);
|
ptr = qemu_get_ram_ptr(addr1);
|
||||||
memcpy(ptr, buf, l);
|
memcpy(ptr, buf, l);
|
||||||
@ -3967,8 +3966,7 @@ void *cpu_physical_memory_map(target_phys_addr_t addr,
|
|||||||
target_phys_addr_t todo = 0;
|
target_phys_addr_t todo = 0;
|
||||||
int l;
|
int l;
|
||||||
target_phys_addr_t page;
|
target_phys_addr_t page;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
ram_addr_t raddr = RAM_ADDR_MAX;
|
ram_addr_t raddr = RAM_ADDR_MAX;
|
||||||
ram_addr_t rlen;
|
ram_addr_t rlen;
|
||||||
void *ret;
|
void *ret;
|
||||||
@ -3978,10 +3976,9 @@ void *cpu_physical_memory_map(target_phys_addr_t addr,
|
|||||||
l = (page + TARGET_PAGE_SIZE) - addr;
|
l = (page + TARGET_PAGE_SIZE) - addr;
|
||||||
if (l > len)
|
if (l > len)
|
||||||
l = len;
|
l = len;
|
||||||
p = phys_page_find(page >> TARGET_PAGE_BITS);
|
section = phys_page_find(page >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if ((pd & ~TARGET_PAGE_MASK) != io_mem_ram.ram_addr) {
|
if (!(memory_region_is_ram(section.mr) && !section.readonly)) {
|
||||||
if (todo || bounce.buffer) {
|
if (todo || bounce.buffer) {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@ -3996,7 +3993,9 @@ void *cpu_physical_memory_map(target_phys_addr_t addr,
|
|||||||
return bounce.buffer;
|
return bounce.buffer;
|
||||||
}
|
}
|
||||||
if (!todo) {
|
if (!todo) {
|
||||||
raddr = (pd & TARGET_PAGE_MASK) + (addr & ~TARGET_PAGE_MASK);
|
raddr = memory_region_get_ram_addr(section.mr)
|
||||||
|
+ section.offset_within_region
|
||||||
|
+ (addr & ~TARGET_PAGE_MASK);
|
||||||
}
|
}
|
||||||
|
|
||||||
len -= l;
|
len -= l;
|
||||||
@ -4055,16 +4054,15 @@ static inline uint32_t ldl_phys_internal(target_phys_addr_t addr,
|
|||||||
int io_index;
|
int io_index;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
uint32_t val;
|
uint32_t val;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if (!is_ram_rom_romd(pd)) {
|
if (!is_ram_rom_romd(§ion)) {
|
||||||
/* I/O case */
|
/* I/O case */
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
addr = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
addr = (addr & ~TARGET_PAGE_MASK) + section.offset_within_region;
|
||||||
val = io_mem_read(io_index, addr, 4);
|
val = io_mem_read(io_index, addr, 4);
|
||||||
#if defined(TARGET_WORDS_BIGENDIAN)
|
#if defined(TARGET_WORDS_BIGENDIAN)
|
||||||
if (endian == DEVICE_LITTLE_ENDIAN) {
|
if (endian == DEVICE_LITTLE_ENDIAN) {
|
||||||
@ -4077,7 +4075,9 @@ static inline uint32_t ldl_phys_internal(target_phys_addr_t addr,
|
|||||||
#endif
|
#endif
|
||||||
} else {
|
} else {
|
||||||
/* RAM case */
|
/* RAM case */
|
||||||
ptr = qemu_get_ram_ptr(pd & TARGET_PAGE_MASK) +
|
ptr = qemu_get_ram_ptr((memory_region_get_ram_addr(section.mr)
|
||||||
|
& TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region) +
|
||||||
(addr & ~TARGET_PAGE_MASK);
|
(addr & ~TARGET_PAGE_MASK);
|
||||||
switch (endian) {
|
switch (endian) {
|
||||||
case DEVICE_LITTLE_ENDIAN:
|
case DEVICE_LITTLE_ENDIAN:
|
||||||
@ -4116,16 +4116,15 @@ static inline uint64_t ldq_phys_internal(target_phys_addr_t addr,
|
|||||||
int io_index;
|
int io_index;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
uint64_t val;
|
uint64_t val;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if (!is_ram_rom_romd(pd)) {
|
if (!is_ram_rom_romd(§ion)) {
|
||||||
/* I/O case */
|
/* I/O case */
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
addr = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
addr = (addr & ~TARGET_PAGE_MASK) + section.offset_within_region;
|
||||||
|
|
||||||
/* XXX This is broken when device endian != cpu endian.
|
/* XXX This is broken when device endian != cpu endian.
|
||||||
Fix and add "endian" variable check */
|
Fix and add "endian" variable check */
|
||||||
@ -4138,8 +4137,10 @@ static inline uint64_t ldq_phys_internal(target_phys_addr_t addr,
|
|||||||
#endif
|
#endif
|
||||||
} else {
|
} else {
|
||||||
/* RAM case */
|
/* RAM case */
|
||||||
ptr = qemu_get_ram_ptr(pd & TARGET_PAGE_MASK) +
|
ptr = qemu_get_ram_ptr((memory_region_get_ram_addr(section.mr)
|
||||||
(addr & ~TARGET_PAGE_MASK);
|
& TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region)
|
||||||
|
+ (addr & ~TARGET_PAGE_MASK);
|
||||||
switch (endian) {
|
switch (endian) {
|
||||||
case DEVICE_LITTLE_ENDIAN:
|
case DEVICE_LITTLE_ENDIAN:
|
||||||
val = ldq_le_p(ptr);
|
val = ldq_le_p(ptr);
|
||||||
@ -4185,16 +4186,15 @@ static inline uint32_t lduw_phys_internal(target_phys_addr_t addr,
|
|||||||
int io_index;
|
int io_index;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
uint64_t val;
|
uint64_t val;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if (!is_ram_rom_romd(pd)) {
|
if (!is_ram_rom_romd(§ion)) {
|
||||||
/* I/O case */
|
/* I/O case */
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
addr = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
addr = (addr & ~TARGET_PAGE_MASK) + section.offset_within_region;
|
||||||
val = io_mem_read(io_index, addr, 2);
|
val = io_mem_read(io_index, addr, 2);
|
||||||
#if defined(TARGET_WORDS_BIGENDIAN)
|
#if defined(TARGET_WORDS_BIGENDIAN)
|
||||||
if (endian == DEVICE_LITTLE_ENDIAN) {
|
if (endian == DEVICE_LITTLE_ENDIAN) {
|
||||||
@ -4207,8 +4207,10 @@ static inline uint32_t lduw_phys_internal(target_phys_addr_t addr,
|
|||||||
#endif
|
#endif
|
||||||
} else {
|
} else {
|
||||||
/* RAM case */
|
/* RAM case */
|
||||||
ptr = qemu_get_ram_ptr(pd & TARGET_PAGE_MASK) +
|
ptr = qemu_get_ram_ptr((memory_region_get_ram_addr(section.mr)
|
||||||
(addr & ~TARGET_PAGE_MASK);
|
& TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region)
|
||||||
|
+ (addr & ~TARGET_PAGE_MASK);
|
||||||
switch (endian) {
|
switch (endian) {
|
||||||
case DEVICE_LITTLE_ENDIAN:
|
case DEVICE_LITTLE_ENDIAN:
|
||||||
val = lduw_le_p(ptr);
|
val = lduw_le_p(ptr);
|
||||||
@ -4246,18 +4248,23 @@ void stl_phys_notdirty(target_phys_addr_t addr, uint32_t val)
|
|||||||
{
|
{
|
||||||
int io_index;
|
int io_index;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if ((pd & ~TARGET_PAGE_MASK) != io_mem_ram.ram_addr) {
|
if (!memory_region_is_ram(section.mr) || section.readonly) {
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
if (memory_region_is_ram(section.mr)) {
|
||||||
addr = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
io_index = io_mem_rom.ram_addr;
|
||||||
|
} else {
|
||||||
|
io_index = memory_region_get_ram_addr(section.mr);
|
||||||
|
}
|
||||||
|
addr = (addr & ~TARGET_PAGE_MASK) + section.offset_within_region;
|
||||||
io_mem_write(io_index, addr, val, 4);
|
io_mem_write(io_index, addr, val, 4);
|
||||||
} else {
|
} else {
|
||||||
unsigned long addr1 = (pd & TARGET_PAGE_MASK) + (addr & ~TARGET_PAGE_MASK);
|
unsigned long addr1 = (memory_region_get_ram_addr(section.mr)
|
||||||
|
& TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region
|
||||||
|
+ (addr & ~TARGET_PAGE_MASK);
|
||||||
ptr = qemu_get_ram_ptr(addr1);
|
ptr = qemu_get_ram_ptr(addr1);
|
||||||
stl_p(ptr, val);
|
stl_p(ptr, val);
|
||||||
|
|
||||||
@ -4277,15 +4284,18 @@ void stq_phys_notdirty(target_phys_addr_t addr, uint64_t val)
|
|||||||
{
|
{
|
||||||
int io_index;
|
int io_index;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if ((pd & ~TARGET_PAGE_MASK) != io_mem_ram.ram_addr) {
|
if (!memory_region_is_ram(section.mr) || section.readonly) {
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
if (memory_region_is_ram(section.mr)) {
|
||||||
addr = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
io_index = io_mem_rom.ram_addr;
|
||||||
|
} else {
|
||||||
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
}
|
||||||
|
addr = (addr & ~TARGET_PAGE_MASK) + section.offset_within_region;
|
||||||
#ifdef TARGET_WORDS_BIGENDIAN
|
#ifdef TARGET_WORDS_BIGENDIAN
|
||||||
io_mem_write(io_index, addr, val >> 32, 4);
|
io_mem_write(io_index, addr, val >> 32, 4);
|
||||||
io_mem_write(io_index, addr + 4, (uint32_t)val, 4);
|
io_mem_write(io_index, addr + 4, (uint32_t)val, 4);
|
||||||
@ -4294,8 +4304,10 @@ void stq_phys_notdirty(target_phys_addr_t addr, uint64_t val)
|
|||||||
io_mem_write(io_index, addr + 4, val >> 32, 4);
|
io_mem_write(io_index, addr + 4, val >> 32, 4);
|
||||||
#endif
|
#endif
|
||||||
} else {
|
} else {
|
||||||
ptr = qemu_get_ram_ptr(pd & TARGET_PAGE_MASK) +
|
ptr = qemu_get_ram_ptr((memory_region_get_ram_addr(section.mr)
|
||||||
(addr & ~TARGET_PAGE_MASK);
|
& TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region)
|
||||||
|
+ (addr & ~TARGET_PAGE_MASK);
|
||||||
stq_p(ptr, val);
|
stq_p(ptr, val);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -4306,15 +4318,18 @@ static inline void stl_phys_internal(target_phys_addr_t addr, uint32_t val,
|
|||||||
{
|
{
|
||||||
int io_index;
|
int io_index;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if ((pd & ~TARGET_PAGE_MASK) != io_mem_ram.ram_addr) {
|
if (!memory_region_is_ram(section.mr) || section.readonly) {
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
if (memory_region_is_ram(section.mr)) {
|
||||||
addr = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
io_index = io_mem_rom.ram_addr;
|
||||||
|
} else {
|
||||||
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
}
|
||||||
|
addr = (addr & ~TARGET_PAGE_MASK) + section.offset_within_region;
|
||||||
#if defined(TARGET_WORDS_BIGENDIAN)
|
#if defined(TARGET_WORDS_BIGENDIAN)
|
||||||
if (endian == DEVICE_LITTLE_ENDIAN) {
|
if (endian == DEVICE_LITTLE_ENDIAN) {
|
||||||
val = bswap32(val);
|
val = bswap32(val);
|
||||||
@ -4327,7 +4342,9 @@ static inline void stl_phys_internal(target_phys_addr_t addr, uint32_t val,
|
|||||||
io_mem_write(io_index, addr, val, 4);
|
io_mem_write(io_index, addr, val, 4);
|
||||||
} else {
|
} else {
|
||||||
unsigned long addr1;
|
unsigned long addr1;
|
||||||
addr1 = (pd & TARGET_PAGE_MASK) + (addr & ~TARGET_PAGE_MASK);
|
addr1 = (memory_region_get_ram_addr(section.mr) & TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region
|
||||||
|
+ (addr & ~TARGET_PAGE_MASK);
|
||||||
/* RAM case */
|
/* RAM case */
|
||||||
ptr = qemu_get_ram_ptr(addr1);
|
ptr = qemu_get_ram_ptr(addr1);
|
||||||
switch (endian) {
|
switch (endian) {
|
||||||
@ -4379,15 +4396,18 @@ static inline void stw_phys_internal(target_phys_addr_t addr, uint32_t val,
|
|||||||
{
|
{
|
||||||
int io_index;
|
int io_index;
|
||||||
uint8_t *ptr;
|
uint8_t *ptr;
|
||||||
unsigned long pd;
|
MemoryRegionSection section;
|
||||||
PhysPageDesc p;
|
|
||||||
|
|
||||||
p = phys_page_find(addr >> TARGET_PAGE_BITS);
|
section = phys_page_find(addr >> TARGET_PAGE_BITS);
|
||||||
pd = p.phys_offset;
|
|
||||||
|
|
||||||
if ((pd & ~TARGET_PAGE_MASK) != io_mem_ram.ram_addr) {
|
if (!memory_region_is_ram(section.mr) || section.readonly) {
|
||||||
io_index = pd & (IO_MEM_NB_ENTRIES - 1);
|
if (memory_region_is_ram(section.mr)) {
|
||||||
addr = (addr & ~TARGET_PAGE_MASK) + p.region_offset;
|
io_index = io_mem_rom.ram_addr;
|
||||||
|
} else {
|
||||||
|
io_index = memory_region_get_ram_addr(section.mr)
|
||||||
|
& (IO_MEM_NB_ENTRIES - 1);
|
||||||
|
}
|
||||||
|
addr = (addr & ~TARGET_PAGE_MASK) + section.offset_within_region;
|
||||||
#if defined(TARGET_WORDS_BIGENDIAN)
|
#if defined(TARGET_WORDS_BIGENDIAN)
|
||||||
if (endian == DEVICE_LITTLE_ENDIAN) {
|
if (endian == DEVICE_LITTLE_ENDIAN) {
|
||||||
val = bswap16(val);
|
val = bswap16(val);
|
||||||
@ -4400,7 +4420,8 @@ static inline void stw_phys_internal(target_phys_addr_t addr, uint32_t val,
|
|||||||
io_mem_write(io_index, addr, val, 2);
|
io_mem_write(io_index, addr, val, 2);
|
||||||
} else {
|
} else {
|
||||||
unsigned long addr1;
|
unsigned long addr1;
|
||||||
addr1 = (pd & TARGET_PAGE_MASK) + (addr & ~TARGET_PAGE_MASK);
|
addr1 = (memory_region_get_ram_addr(section.mr) & TARGET_PAGE_MASK)
|
||||||
|
+ section.offset_within_region + (addr & ~TARGET_PAGE_MASK);
|
||||||
/* RAM case */
|
/* RAM case */
|
||||||
ptr = qemu_get_ram_ptr(addr1);
|
ptr = qemu_get_ram_ptr(addr1);
|
||||||
switch (endian) {
|
switch (endian) {
|
||||||
@ -4617,7 +4638,7 @@ tb_page_addr_t get_page_addr_code(CPUState *env1, target_ulong addr)
|
|||||||
}
|
}
|
||||||
pd = env1->tlb_table[mmu_idx][page_index].addr_code & ~TARGET_PAGE_MASK;
|
pd = env1->tlb_table[mmu_idx][page_index].addr_code & ~TARGET_PAGE_MASK;
|
||||||
if (pd != io_mem_ram.ram_addr && pd != io_mem_rom.ram_addr
|
if (pd != io_mem_ram.ram_addr && pd != io_mem_rom.ram_addr
|
||||||
&& !is_romd(pd)) {
|
&& !io_mem_region[pd]->rom_device) {
|
||||||
#if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_SPARC)
|
#if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_SPARC)
|
||||||
cpu_unassigned_access(env1, addr, 0, 1, 0, 4);
|
cpu_unassigned_access(env1, addr, 0, 1, 0, 4);
|
||||||
#else
|
#else
|
||||||
|
Loading…
Reference in New Issue
Block a user