mirror of
https://github.com/yuzu-emu/unicorn.git
synced 2024-12-25 17:05:38 +00:00
d46e52d9d0
In the user emulation code path, tlb_vaddr_to_host erronesously passed vaddr as the guest address to be translated, instead of addr, the parameter which actually contained the guest address. This resulted in incorrect addresses being used when emulating block copy (mvc/mvpg) and block clear (xc) instructions for the s390x target. Backports commit c2a85316902e67530da9d6548139fcce73c0cac6 from qemu
500 lines
13 KiB
C
500 lines
13 KiB
C
/*
|
|
* Software MMU support
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
|
*
|
|
*/
|
|
|
|
/*
|
|
* Generate inline load/store functions for all MMU modes (typically
|
|
* at least _user and _kernel) as well as _data versions, for all data
|
|
* sizes.
|
|
*
|
|
* Used by target op helpers.
|
|
*
|
|
* MMU mode suffixes are defined in target cpu.h.
|
|
*/
|
|
#ifndef CPU_LDST_H
|
|
#define CPU_LDST_H
|
|
|
|
#if defined(CONFIG_USER_ONLY)
|
|
/* All direct uses of g2h and h2g need to go away for usermode softmmu. */
|
|
#define g2h(x) ((void *)((unsigned long)(target_ulong)(x) + GUEST_BASE))
|
|
|
|
#if HOST_LONG_BITS <= TARGET_VIRT_ADDR_SPACE_BITS
|
|
#define h2g_valid(x) 1
|
|
#else
|
|
#define h2g_valid(x) ({ \
|
|
unsigned long __guest = (unsigned long)(x) - GUEST_BASE; \
|
|
(__guest < (1ul << TARGET_VIRT_ADDR_SPACE_BITS)) && \
|
|
(!RESERVED_VA || (__guest < RESERVED_VA)); \
|
|
})
|
|
#endif
|
|
|
|
#define h2g_nocheck(x) ({ \
|
|
unsigned long __ret = (unsigned long)(x) - GUEST_BASE; \
|
|
(abi_ulong)__ret; \
|
|
})
|
|
|
|
#define h2g(x) ({ \
|
|
/* Check if given address fits target address space */ \
|
|
assert(h2g_valid(x)); \
|
|
h2g_nocheck(x); \
|
|
})
|
|
|
|
#define saddr(x) g2h(x)
|
|
#define laddr(x) g2h(x)
|
|
|
|
#else /* !CONFIG_USER_ONLY */
|
|
/* NOTE: we use double casts if pointers and target_ulong have
|
|
different sizes */
|
|
#define saddr(x) (uint8_t *)(intptr_t)(x)
|
|
#define laddr(x) (uint8_t *)(intptr_t)(x)
|
|
#endif
|
|
|
|
#define ldub_raw(p) ldub_p(laddr((p)))
|
|
#define ldsb_raw(p) ldsb_p(laddr((p)))
|
|
#define lduw_raw(p) lduw_p(laddr((p)))
|
|
#define ldsw_raw(p) ldsw_p(laddr((p)))
|
|
#define ldl_raw(p) ldl_p(laddr((p)))
|
|
#define ldq_raw(p) ldq_p(laddr((p)))
|
|
#define ldfl_raw(p) ldfl_p(laddr((p)))
|
|
#define ldfq_raw(p) ldfq_p(laddr((p)))
|
|
#define stb_raw(p, v) stb_p(saddr((p)), v)
|
|
#define stw_raw(p, v) stw_p(saddr((p)), v)
|
|
#define stl_raw(p, v) stl_p(saddr((p)), v)
|
|
#define stq_raw(p, v) stq_p(saddr((p)), v)
|
|
#define stfl_raw(p, v) stfl_p(saddr((p)), v)
|
|
#define stfq_raw(p, v) stfq_p(saddr((p)), v)
|
|
|
|
|
|
#if defined(CONFIG_USER_ONLY)
|
|
|
|
/* if user mode, no other memory access functions */
|
|
#define ldub(p) ldub_raw(p)
|
|
#define ldsb(p) ldsb_raw(p)
|
|
#define lduw(p) lduw_raw(p)
|
|
#define ldsw(p) ldsw_raw(p)
|
|
#define ldl(p) ldl_raw(p)
|
|
#define ldq(p) ldq_raw(p)
|
|
#define ldfl(p) ldfl_raw(p)
|
|
#define ldfq(p) ldfq_raw(p)
|
|
#define stb(p, v) stb_raw(p, v)
|
|
#define stw(p, v) stw_raw(p, v)
|
|
#define stl(p, v) stl_raw(p, v)
|
|
#define stq(p, v) stq_raw(p, v)
|
|
#define stfl(p, v) stfl_raw(p, v)
|
|
#define stfq(p, v) stfq_raw(p, v)
|
|
|
|
#define cpu_ldub_code(env1, p) ldub_raw(p)
|
|
#define cpu_ldsb_code(env1, p) ldsb_raw(p)
|
|
#define cpu_lduw_code(env1, p) lduw_raw(p)
|
|
#define cpu_ldsw_code(env1, p) ldsw_raw(p)
|
|
#define cpu_ldl_code(env1, p) ldl_raw(p)
|
|
#define cpu_ldq_code(env1, p) ldq_raw(p)
|
|
|
|
#define cpu_ldub_data(env, addr) ldub_raw(addr)
|
|
#define cpu_lduw_data(env, addr) lduw_raw(addr)
|
|
#define cpu_ldsw_data(env, addr) ldsw_raw(addr)
|
|
#define cpu_ldl_data(env, addr) ldl_raw(addr)
|
|
#define cpu_ldq_data(env, addr) ldq_raw(addr)
|
|
|
|
#define cpu_stb_data(env, addr, data) stb_raw(addr, data)
|
|
#define cpu_stw_data(env, addr, data) stw_raw(addr, data)
|
|
#define cpu_stl_data(env, addr, data) stl_raw(addr, data)
|
|
#define cpu_stq_data(env, addr, data) stq_raw(addr, data)
|
|
|
|
#define cpu_ldub_kernel(env, addr) ldub_raw(addr)
|
|
#define cpu_lduw_kernel(env, addr) lduw_raw(addr)
|
|
#define cpu_ldsw_kernel(env, addr) ldsw_raw(addr)
|
|
#define cpu_ldl_kernel(env, addr) ldl_raw(addr)
|
|
#define cpu_ldq_kernel(env, addr) ldq_raw(addr)
|
|
|
|
#define cpu_stb_kernel(env, addr, data) stb_raw(addr, data)
|
|
#define cpu_stw_kernel(env, addr, data) stw_raw(addr, data)
|
|
#define cpu_stl_kernel(env, addr, data) stl_raw(addr, data)
|
|
#define cpu_stq_kernel(env, addr, data) stq_raw(addr, data)
|
|
|
|
#define ldub_kernel(p) ldub_raw(p)
|
|
#define ldsb_kernel(p) ldsb_raw(p)
|
|
#define lduw_kernel(p) lduw_raw(p)
|
|
#define ldsw_kernel(p) ldsw_raw(p)
|
|
#define ldl_kernel(p) ldl_raw(p)
|
|
#define ldq_kernel(p) ldq_raw(p)
|
|
#define ldfl_kernel(p) ldfl_raw(p)
|
|
#define ldfq_kernel(p) ldfq_raw(p)
|
|
#define stb_kernel(p, v) stb_raw(p, v)
|
|
#define stw_kernel(p, v) stw_raw(p, v)
|
|
#define stl_kernel(p, v) stl_raw(p, v)
|
|
#define stq_kernel(p, v) stq_raw(p, v)
|
|
#define stfl_kernel(p, v) stfl_raw(p, v)
|
|
#define stfq_kernel(p, vt) stfq_raw(p, v)
|
|
|
|
#define cpu_ldub_data(env, addr) ldub_raw(addr)
|
|
#define cpu_lduw_data(env, addr) lduw_raw(addr)
|
|
#define cpu_ldl_data(env, addr) ldl_raw(addr)
|
|
|
|
#define cpu_stb_data(env, addr, data) stb_raw(addr, data)
|
|
#define cpu_stw_data(env, addr, data) stw_raw(addr, data)
|
|
#define cpu_stl_data(env, addr, data) stl_raw(addr, data)
|
|
|
|
#else
|
|
|
|
/* XXX: find something cleaner.
|
|
* Furthermore, this is false for 64 bits targets
|
|
*/
|
|
#define ldul_user ldl_user
|
|
#define ldul_kernel ldl_kernel
|
|
#define ldul_hypv ldl_hypv
|
|
#define ldul_executive ldl_executive
|
|
#define ldul_supervisor ldl_supervisor
|
|
|
|
/* The memory helpers for tcg-generated code need tcg_target_long etc. */
|
|
#include "tcg.h"
|
|
|
|
#define CPU_MMU_INDEX 0
|
|
#define MEMSUFFIX MMU_MODE0_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
|
|
#define CPU_MMU_INDEX 1
|
|
#define MEMSUFFIX MMU_MODE1_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
|
|
#if (NB_MMU_MODES >= 3)
|
|
|
|
#define CPU_MMU_INDEX 2
|
|
#define MEMSUFFIX MMU_MODE2_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 3) */
|
|
|
|
#if (NB_MMU_MODES >= 4)
|
|
|
|
#define CPU_MMU_INDEX 3
|
|
#define MEMSUFFIX MMU_MODE3_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 4) */
|
|
|
|
#if (NB_MMU_MODES >= 5)
|
|
|
|
#define CPU_MMU_INDEX 4
|
|
#define MEMSUFFIX MMU_MODE4_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 5) */
|
|
|
|
#if (NB_MMU_MODES >= 6)
|
|
|
|
#define CPU_MMU_INDEX 5
|
|
#define MEMSUFFIX MMU_MODE5_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 6) */
|
|
|
|
#if (NB_MMU_MODES >= 7) && defined(MMU_MODE6_SUFFIX)
|
|
|
|
#define CPU_MMU_INDEX 6
|
|
#define MEMSUFFIX MMU_MODE6_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 7) */
|
|
|
|
#if (NB_MMU_MODES >= 8) && defined(MMU_MODE7_SUFFIX)
|
|
|
|
#define CPU_MMU_INDEX 7
|
|
#define MEMSUFFIX MMU_MODE7_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 8) */
|
|
|
|
#if (NB_MMU_MODES >= 9) && defined(MMU_MODE8_SUFFIX)
|
|
|
|
#define CPU_MMU_INDEX 8
|
|
#define MEMSUFFIX MMU_MODE8_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 9) */
|
|
|
|
#if (NB_MMU_MODES >= 10) && defined(MMU_MODE9_SUFFIX)
|
|
|
|
#define CPU_MMU_INDEX 9
|
|
#define MEMSUFFIX MMU_MODE9_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 10) */
|
|
|
|
#if (NB_MMU_MODES >= 11) && defined(MMU_MODE10_SUFFIX)
|
|
|
|
#define CPU_MMU_INDEX 10
|
|
#define MEMSUFFIX MMU_MODE10_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 11) */
|
|
|
|
#if (NB_MMU_MODES >= 12) && defined(MMU_MODE11_SUFFIX)
|
|
|
|
#define CPU_MMU_INDEX 11
|
|
#define MEMSUFFIX MMU_MODE11_SUFFIX
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#endif /* (NB_MMU_MODES >= 12) */
|
|
|
|
#if (NB_MMU_MODES > 12)
|
|
#error "NB_MMU_MODES > 12 is not supported for now"
|
|
#endif /* (NB_MMU_MODES > 12) */
|
|
|
|
/* these access are slower, they must be as rare as possible */
|
|
#define CPU_MMU_INDEX (cpu_mmu_index(env, false))
|
|
#define MEMSUFFIX _data
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
|
|
#define ldub(p) ldub_data(p)
|
|
#define ldsb(p) ldsb_data(p)
|
|
#define lduw(p) lduw_data(p)
|
|
#define ldsw(p) ldsw_data(p)
|
|
#define ldl(p) ldl_data(p)
|
|
#define ldq(p) ldq_data(p)
|
|
|
|
#define stb(p, v) stb_data(p, v)
|
|
#define stw(p, v) stw_data(p, v)
|
|
#define stl(p, v) stl_data(p, v)
|
|
#define stq(p, v) stq_data(p, v)
|
|
|
|
#define CPU_MMU_INDEX (cpu_mmu_index(env, true))
|
|
#define MEMSUFFIX _code
|
|
#define SOFTMMU_CODE_ACCESS
|
|
|
|
#define DATA_SIZE 1
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 2
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 4
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#define DATA_SIZE 8
|
|
#include "exec/cpu_ldst_template.h"
|
|
|
|
#undef CPU_MMU_INDEX
|
|
#undef MEMSUFFIX
|
|
#undef SOFTMMU_CODE_ACCESS
|
|
|
|
#endif /* defined(CONFIG_USER_ONLY) */
|
|
|
|
/**
|
|
* tlb_vaddr_to_host:
|
|
* @env: CPUArchState
|
|
* @addr: guest virtual address to look up
|
|
* @access_type: 0 for read, 1 for write, 2 for execute
|
|
* @mmu_idx: MMU index to use for lookup
|
|
*
|
|
* Look up the specified guest virtual index in the TCG softmmu TLB.
|
|
* If the TLB contains a host virtual address suitable for direct RAM
|
|
* access, then return it. Otherwise (TLB miss, TLB entry is for an
|
|
* I/O access, etc) return NULL.
|
|
*
|
|
* This is the equivalent of the initial fast-path code used by
|
|
* TCG backends for guest load and store accesses.
|
|
*/
|
|
static inline void *tlb_vaddr_to_host(CPUArchState *env, target_ulong addr,
|
|
int access_type, int mmu_idx)
|
|
{
|
|
#if defined(CONFIG_USER_ONLY)
|
|
return g2h(addr);
|
|
#else
|
|
int index = (addr >> TARGET_PAGE_BITS) & (CPU_TLB_SIZE - 1);
|
|
CPUTLBEntry *tlbentry = &env->tlb_table[mmu_idx][index];
|
|
target_ulong tlb_addr;
|
|
uintptr_t haddr;
|
|
|
|
switch (access_type) {
|
|
case 0:
|
|
tlb_addr = tlbentry->addr_read;
|
|
break;
|
|
case 1:
|
|
tlb_addr = tlbentry->addr_write;
|
|
break;
|
|
case 2:
|
|
tlb_addr = tlbentry->addr_code;
|
|
break;
|
|
default:
|
|
g_assert_not_reached();
|
|
}
|
|
|
|
if ((addr & TARGET_PAGE_MASK)
|
|
!= (tlb_addr & (TARGET_PAGE_MASK | TLB_INVALID_MASK))) {
|
|
/* TLB entry is for a different page */
|
|
return NULL;
|
|
}
|
|
|
|
if (tlb_addr & ~TARGET_PAGE_MASK) {
|
|
/* IO access */
|
|
return NULL;
|
|
}
|
|
|
|
haddr = (uintptr_t)(addr + env->tlb_table[mmu_idx][index].addend);
|
|
return (void *)haddr;
|
|
#endif /* defined(CONFIG_USER_ONLY) */
|
|
}
|
|
|
|
#endif /* CPU_LDST_H */
|