summaryrefslogtreecommitdiff
path: root/target-ppc
diff options
context:
space:
mode:
authoraurel32 <aurel32@c046a42c-6fe2-441c-8c8c-71466251a162>2008-10-15 17:00:37 +0000
committeraurel32 <aurel32@c046a42c-6fe2-441c-8c8c-71466251a162>2008-10-15 17:00:37 +0000
commitb61f2753a7239f784b5c2f1d3a0249371cd2e164 (patch)
tree0c0fb2d558442db5b7361eb3958480c487f55b44 /target-ppc
parent19f98ff634d82a73b25de8f4dbbce2e0dc6de7e3 (diff)
downloadqemu-b61f2753a7239f784b5c2f1d3a0249371cd2e164.tar.gz
ppc: convert integer load/store to TCG
Signed-off-by: Aurelien Jarno <aurelien@aurel32.net> git-svn-id: svn://svn.savannah.nongnu.org/qemu/trunk@5493 c046a42c-6fe2-441c-8c8c-71466251a162
Diffstat (limited to 'target-ppc')
-rw-r--r--target-ppc/op_mem.h109
-rw-r--r--target-ppc/translate.c573
2 files changed, 413 insertions, 269 deletions
diff --git a/target-ppc/op_mem.h b/target-ppc/op_mem.h
index eae9f4ea5c..b504555cfd 100644
--- a/target-ppc/op_mem.h
+++ b/target-ppc/op_mem.h
@@ -20,111 +20,6 @@
#include "op_mem_access.h"
-/*** Integer load ***/
-#define PPC_LD_OP(name, op) \
-void OPPROTO glue(glue(op_l, name), MEMSUFFIX) (void) \
-{ \
- T1 = glue(op, MEMSUFFIX)((uint32_t)T0); \
- RETURN(); \
-}
-
-#if defined(TARGET_PPC64)
-#define PPC_LD_OP_64(name, op) \
-void OPPROTO glue(glue(glue(op_l, name), _64), MEMSUFFIX) (void) \
-{ \
- T1 = glue(op, MEMSUFFIX)((uint64_t)T0); \
- RETURN(); \
-}
-#endif
-
-#define PPC_ST_OP(name, op) \
-void OPPROTO glue(glue(op_st, name), MEMSUFFIX) (void) \
-{ \
- glue(op, MEMSUFFIX)((uint32_t)T0, T1); \
- RETURN(); \
-}
-
-#if defined(TARGET_PPC64)
-#define PPC_ST_OP_64(name, op) \
-void OPPROTO glue(glue(glue(op_st, name), _64), MEMSUFFIX) (void) \
-{ \
- glue(op, MEMSUFFIX)((uint64_t)T0, T1); \
- RETURN(); \
-}
-#endif
-
-PPC_LD_OP(bz, ldu8);
-PPC_LD_OP(ha, lds16);
-PPC_LD_OP(hz, ldu16);
-PPC_LD_OP(wz, ldu32);
-#if defined(TARGET_PPC64)
-PPC_LD_OP(wa, lds32);
-PPC_LD_OP(d, ldu64);
-PPC_LD_OP_64(bz, ldu8);
-PPC_LD_OP_64(ha, lds16);
-PPC_LD_OP_64(hz, ldu16);
-PPC_LD_OP_64(wz, ldu32);
-PPC_LD_OP_64(wa, lds32);
-PPC_LD_OP_64(d, ldu64);
-#endif
-
-PPC_LD_OP(ha_le, lds16r);
-PPC_LD_OP(hz_le, ldu16r);
-PPC_LD_OP(wz_le, ldu32r);
-#if defined(TARGET_PPC64)
-PPC_LD_OP(wa_le, lds32r);
-PPC_LD_OP(d_le, ldu64r);
-PPC_LD_OP_64(ha_le, lds16r);
-PPC_LD_OP_64(hz_le, ldu16r);
-PPC_LD_OP_64(wz_le, ldu32r);
-PPC_LD_OP_64(wa_le, lds32r);
-PPC_LD_OP_64(d_le, ldu64r);
-#endif
-
-/*** Integer store ***/
-PPC_ST_OP(b, st8);
-PPC_ST_OP(h, st16);
-PPC_ST_OP(w, st32);
-#if defined(TARGET_PPC64)
-PPC_ST_OP(d, st64);
-PPC_ST_OP_64(b, st8);
-PPC_ST_OP_64(h, st16);
-PPC_ST_OP_64(w, st32);
-PPC_ST_OP_64(d, st64);
-#endif
-
-PPC_ST_OP(h_le, st16r);
-PPC_ST_OP(w_le, st32r);
-#if defined(TARGET_PPC64)
-PPC_ST_OP(d_le, st64r);
-PPC_ST_OP_64(h_le, st16r);
-PPC_ST_OP_64(w_le, st32r);
-PPC_ST_OP_64(d_le, st64r);
-#endif
-
-/*** Integer load and store with byte reverse ***/
-PPC_LD_OP(hbr, ldu16r);
-PPC_LD_OP(wbr, ldu32r);
-PPC_ST_OP(hbr, st16r);
-PPC_ST_OP(wbr, st32r);
-#if defined(TARGET_PPC64)
-PPC_LD_OP_64(hbr, ldu16r);
-PPC_LD_OP_64(wbr, ldu32r);
-PPC_ST_OP_64(hbr, st16r);
-PPC_ST_OP_64(wbr, st32r);
-#endif
-
-PPC_LD_OP(hbr_le, ldu16);
-PPC_LD_OP(wbr_le, ldu32);
-PPC_ST_OP(hbr_le, st16);
-PPC_ST_OP(wbr_le, st32);
-#if defined(TARGET_PPC64)
-PPC_LD_OP_64(hbr_le, ldu16);
-PPC_LD_OP_64(wbr_le, ldu32);
-PPC_ST_OP_64(hbr_le, st16);
-PPC_ST_OP_64(wbr_le, st32);
-#endif
-
/*** Integer load and store multiple ***/
void OPPROTO glue(op_lmw, MEMSUFFIX) (void)
{
@@ -985,12 +880,10 @@ _PPC_SPE_ST_OP_64(name, op)
_PPC_SPE_ST_OP(name, op)
#endif
-#if !defined(TARGET_PPC64)
PPC_SPE_LD_OP(dd, ldu64);
PPC_SPE_ST_OP(dd, st64);
PPC_SPE_LD_OP(dd_le, ldu64r);
PPC_SPE_ST_OP(dd_le, st64r);
-#endif
static always_inline uint64_t glue(spe_ldw, MEMSUFFIX) (target_ulong EA)
{
uint64_t ret;
@@ -1135,7 +1028,6 @@ static always_inline void glue(spe_stwho_le, MEMSUFFIX) (target_ulong EA,
glue(st16r, MEMSUFFIX)(EA + 2, data);
}
PPC_SPE_ST_OP(who_le, spe_stwho_le);
-#if !defined(TARGET_PPC64)
static always_inline void glue(spe_stwwo, MEMSUFFIX) (target_ulong EA,
uint64_t data)
{
@@ -1148,7 +1040,6 @@ static always_inline void glue(spe_stwwo_le, MEMSUFFIX) (target_ulong EA,
glue(st32r, MEMSUFFIX)(EA, data);
}
PPC_SPE_ST_OP(wwo_le, spe_stwwo_le);
-#endif
static always_inline uint64_t glue(spe_lh, MEMSUFFIX) (target_ulong EA)
{
uint16_t tmp;
diff --git a/target-ppc/translate.c b/target-ppc/translate.c
index ac9037a325..63a3052858 100644
--- a/target-ppc/translate.c
+++ b/target-ppc/translate.c
@@ -2172,23 +2172,6 @@ static always_inline void gen_addr_register (TCGv EA,
/*** Integer load ***/
#define op_ldst(name) (*gen_op_##name[ctx->mem_idx])()
-/* Byte access routine are endian safe */
-#define gen_op_lbz_le_raw gen_op_lbz_raw
-#define gen_op_lbz_le_user gen_op_lbz_user
-#define gen_op_lbz_le_kernel gen_op_lbz_kernel
-#define gen_op_lbz_le_hypv gen_op_lbz_hypv
-#define gen_op_lbz_le_64_raw gen_op_lbz_64_raw
-#define gen_op_lbz_le_64_user gen_op_lbz_64_user
-#define gen_op_lbz_le_64_kernel gen_op_lbz_64_kernel
-#define gen_op_lbz_le_64_hypv gen_op_lbz_64_hypv
-#define gen_op_stb_le_raw gen_op_stb_raw
-#define gen_op_stb_le_user gen_op_stb_user
-#define gen_op_stb_le_kernel gen_op_stb_kernel
-#define gen_op_stb_le_hypv gen_op_stb_hypv
-#define gen_op_stb_le_64_raw gen_op_stb_64_raw
-#define gen_op_stb_le_64_user gen_op_stb_64_user
-#define gen_op_stb_le_64_kernel gen_op_stb_64_kernel
-#define gen_op_stb_le_64_hypv gen_op_stb_64_hypv
#define OP_LD_TABLE(width) \
static GenOpFunc *gen_op_l##width[NB_MEM_FUNCS] = { \
GEN_MEM_FUNCS(l##width), \
@@ -2198,81 +2181,354 @@ static GenOpFunc *gen_op_st##width[NB_MEM_FUNCS] = { \
GEN_MEM_FUNCS(st##width), \
};
+
+#if defined(TARGET_PPC64)
+#define GEN_QEMU_LD_PPC64(width) \
+static always_inline void gen_qemu_ld##width##_ppc64(TCGv t0, TCGv t1, int flags)\
+{ \
+ if (likely(flags & 2)) \
+ tcg_gen_qemu_ld##width(t0, t1, flags >> 2); \
+ else { \
+ TCGv addr = tcg_temp_new(TCG_TYPE_TL); \
+ tcg_gen_ext32u_tl(addr, t1); \
+ tcg_gen_qemu_ld##width(t0, addr, flags >> 2); \
+ tcg_temp_free(addr); \
+ } \
+}
+GEN_QEMU_LD_PPC64(8u)
+GEN_QEMU_LD_PPC64(8s)
+GEN_QEMU_LD_PPC64(16u)
+GEN_QEMU_LD_PPC64(16s)
+GEN_QEMU_LD_PPC64(32u)
+GEN_QEMU_LD_PPC64(32s)
+GEN_QEMU_LD_PPC64(64)
+
+#define GEN_QEMU_ST_PPC64(width) \
+static always_inline void gen_qemu_st##width##_ppc64(TCGv t0, TCGv t1, int flags)\
+{ \
+ if (likely(flags & 2)) \
+ tcg_gen_qemu_st##width(t0, t1, flags >> 2); \
+ else { \
+ TCGv addr = tcg_temp_new(TCG_TYPE_TL); \
+ tcg_gen_ext32u_tl(addr, t1); \
+ tcg_gen_qemu_st##width(t0, addr, flags >> 2); \
+ tcg_temp_free(addr); \
+ } \
+}
+GEN_QEMU_ST_PPC64(8)
+GEN_QEMU_ST_PPC64(16)
+GEN_QEMU_ST_PPC64(32)
+GEN_QEMU_ST_PPC64(64)
+
+static always_inline void gen_qemu_ld8u(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld8u_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_ld8s(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld8s_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_ld16u(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv t0_32;
+ gen_qemu_ld16u_ppc64(t0, t1, flags);
+ t0_32 = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(t0_32, t0);
+ tcg_gen_bswap16_i32(t0_32, t0_32);
+ tcg_gen_extu_i32_tl(t0, t0_32);
+ tcg_temp_free(t0_32);
+ } else
+ gen_qemu_ld16u_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_ld16s(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv t0_32;
+ gen_qemu_ld16u_ppc64(t0, t1, flags);
+ t0_32 = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(t0_32, t0);
+ tcg_gen_bswap16_i32(t0_32, t0_32);
+ tcg_gen_extu_i32_tl(t0, t0_32);
+ tcg_gen_ext16s_tl(t0, t0);
+ tcg_temp_free(t0_32);
+ } else
+ gen_qemu_ld16s_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_ld32u(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv t0_32;
+ gen_qemu_ld32u_ppc64(t0, t1, flags);
+ t0_32 = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(t0_32, t0);
+ tcg_gen_bswap_i32(t0_32, t0_32);
+ tcg_gen_extu_i32_tl(t0, t0_32);
+ tcg_temp_free(t0_32);
+ } else
+ gen_qemu_ld32u_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_ld32s(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv t0_32;
+ gen_qemu_ld32u_ppc64(t0, t1, flags);
+ t0_32 = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(t0_32, t0);
+ tcg_gen_bswap_i32(t0_32, t0_32);
+ tcg_gen_ext_i32_tl(t0, t0_32);
+ tcg_temp_free(t0_32);
+ } else
+ gen_qemu_ld32s_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_ld64(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld64_ppc64(t0, t1, flags);
+ if (unlikely(flags & 1))
+ tcg_gen_bswap_i64(t0, t0);
+}
+
+static always_inline void gen_qemu_st8(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_st8_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_st16(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv temp1, temp2;
+ temp1 = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(temp1, t0);
+ tcg_gen_ext16u_i32(temp1, temp1);
+ tcg_gen_bswap16_i32(temp1, temp1);
+ temp2 = tcg_temp_new(TCG_TYPE_I64);
+ tcg_gen_extu_i32_tl(temp2, temp1);
+ tcg_temp_free(temp1);
+ gen_qemu_st16_ppc64(temp2, t1, flags);
+ tcg_temp_free(temp2);
+ } else
+ gen_qemu_st16_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_st32(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv temp1, temp2;
+ temp1 = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(temp1, t0);
+ tcg_gen_bswap_i32(temp1, temp1);
+ temp2 = tcg_temp_new(TCG_TYPE_I64);
+ tcg_gen_extu_i32_tl(temp2, temp1);
+ tcg_temp_free(temp1);
+ gen_qemu_st32_ppc64(temp2, t1, flags);
+ tcg_temp_free(temp2);
+ } else
+ gen_qemu_st32_ppc64(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_st64(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv temp = tcg_temp_new(TCG_TYPE_I64);
+ tcg_gen_bswap_i64(temp, t0);
+ gen_qemu_st64_ppc64(temp, t1, flags);
+ tcg_temp_free(temp);
+ } else
+ gen_qemu_st64_ppc64(t0, t1, flags);
+}
+
+
+#else /* defined(TARGET_PPC64) */
+#define GEN_QEMU_LD_PPC32(width) \
+static always_inline void gen_qemu_ld##width##_ppc32(TCGv t0, TCGv t1, int flags)\
+{ \
+ tcg_gen_qemu_ld##width(t0, t1, flags >> 1); \
+}
+GEN_QEMU_LD_PPC32(8u)
+GEN_QEMU_LD_PPC32(8s)
+GEN_QEMU_LD_PPC32(16u)
+GEN_QEMU_LD_PPC32(16s)
+GEN_QEMU_LD_PPC32(32u)
+GEN_QEMU_LD_PPC32(32s)
+GEN_QEMU_LD_PPC32(64)
+
+#define GEN_QEMU_ST_PPC32(width) \
+static always_inline void gen_qemu_st##width##_ppc32(TCGv t0, TCGv t1, int flags)\
+{ \
+ tcg_gen_qemu_st##width(t0, t1, flags >> 1); \
+}
+GEN_QEMU_ST_PPC32(8)
+GEN_QEMU_ST_PPC32(16)
+GEN_QEMU_ST_PPC32(32)
+GEN_QEMU_ST_PPC32(64)
+
+static always_inline void gen_qemu_ld8u(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld8u_ppc32(t0, t1, flags >> 1);
+}
+
+static always_inline void gen_qemu_ld8s(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld8s_ppc32(t0, t1, flags >> 1);
+}
+
+static always_inline void gen_qemu_ld16u(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld16u_ppc32(t0, t1, flags >> 1);
+ if (unlikely(flags & 1))
+ tcg_gen_bswap16_i32(t0, t0);
+}
+
+static always_inline void gen_qemu_ld16s(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ gen_qemu_ld16u_ppc32(t0, t1, flags);
+ tcg_gen_bswap16_i32(t0, t0);
+ tcg_gen_ext16s_i32(t0, t0);
+ } else
+ gen_qemu_ld16s_ppc32(t0, t1, flags);
+}
+
+static always_inline void gen_qemu_ld32u(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld32u_ppc32(t0, t1, flags);
+ if (unlikely(flags & 1))
+ tcg_gen_bswap_i32(t0, t0);
+}
+
+static always_inline void gen_qemu_ld64(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_ld64_ppc32(t0, t1, flags);
+ if (unlikely(flags & 1))
+ tcg_gen_bswap_i64(t0, t0);
+}
+
+static always_inline void gen_qemu_st8(TCGv t0, TCGv t1, int flags)
+{
+ gen_qemu_st8_ppc32(t0, t1, flags >> 1);
+}
+
+static always_inline void gen_qemu_st16(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv temp = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_ext16u_i32(temp, t0);
+ tcg_gen_bswap16_i32(temp, temp);
+ gen_qemu_st16_ppc32(temp, t1, flags >> 1);
+ tcg_temp_free(temp);
+ } else
+ gen_qemu_st16_ppc32(t0, t1, flags >> 1);
+}
+
+static always_inline void gen_qemu_st32(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv temp = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_bswap_i32(temp, t0);
+ gen_qemu_st32_ppc32(temp, t1, flags >> 1);
+ tcg_temp_free(temp);
+ } else
+ gen_qemu_st32_ppc32(t0, t1, flags >> 1);
+}
+
+static always_inline void gen_qemu_st64(TCGv t0, TCGv t1, int flags)
+{
+ if (unlikely(flags & 1)) {
+ TCGv temp = tcg_temp_new(TCG_TYPE_I64);
+ tcg_gen_bswap_i64(temp, t0);
+ gen_qemu_st64_ppc32(temp, t1, flags >> 1);
+ tcg_temp_free(temp);
+ } else
+ gen_qemu_st64_ppc32(t0, t1, flags >> 1);
+}
+
+#endif
+
#define GEN_LD(width, opc, type) \
GEN_HANDLER(l##width, opc, 0xFF, 0xFF, 0x00000000, type) \
{ \
- gen_addr_imm_index(cpu_T[0], ctx, 0); \
- op_ldst(l##width); \
- tcg_gen_mov_tl(cpu_gpr[rD(ctx->opcode)], cpu_T[1]); \
+ TCGv EA = tcg_temp_new(TCG_TYPE_TL); \
+ gen_addr_imm_index(EA, ctx, 0); \
+ gen_qemu_ld##width(cpu_gpr[rD(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_temp_free(EA); \
}
#define GEN_LDU(width, opc, type) \
GEN_HANDLER(l##width##u, opc, 0xFF, 0xFF, 0x00000000, type) \
{ \
+ TCGv EA; \
if (unlikely(rA(ctx->opcode) == 0 || \
rA(ctx->opcode) == rD(ctx->opcode))) { \
GEN_EXCP_INVAL(ctx); \
return; \
} \
+ EA = tcg_temp_new(TCG_TYPE_TL); \
if (type == PPC_64B) \
- gen_addr_imm_index(cpu_T[0], ctx, 0x03); \
+ gen_addr_imm_index(EA, ctx, 0x03); \
else \
- gen_addr_imm_index(cpu_T[0], ctx, 0); \
- op_ldst(l##width); \
- tcg_gen_mov_tl(cpu_gpr[rD(ctx->opcode)], cpu_T[1]); \
- tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], cpu_T[0]); \
+ gen_addr_imm_index(EA, ctx, 0); \
+ gen_qemu_ld##width(cpu_gpr[rD(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], EA); \
+ tcg_temp_free(EA); \
}
#define GEN_LDUX(width, opc2, opc3, type) \
GEN_HANDLER(l##width##ux, 0x1F, opc2, opc3, 0x00000001, type) \
{ \
+ TCGv EA; \
if (unlikely(rA(ctx->opcode) == 0 || \
rA(ctx->opcode) == rD(ctx->opcode))) { \
GEN_EXCP_INVAL(ctx); \
return; \
} \
- gen_addr_reg_index(cpu_T[0], ctx); \
- op_ldst(l##width); \
- tcg_gen_mov_tl(cpu_gpr[rD(ctx->opcode)], cpu_T[1]); \
- tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], cpu_T[0]); \
+ EA = tcg_temp_new(TCG_TYPE_TL); \
+ gen_addr_reg_index(EA, ctx); \
+ gen_qemu_ld##width(cpu_gpr[rD(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], EA); \
+ tcg_temp_free(EA); \
}
#define GEN_LDX(width, opc2, opc3, type) \
GEN_HANDLER(l##width##x, 0x1F, opc2, opc3, 0x00000001, type) \
{ \
- gen_addr_reg_index(cpu_T[0], ctx); \
- op_ldst(l##width); \
- tcg_gen_mov_tl(cpu_gpr[rD(ctx->opcode)], cpu_T[1]); \
+ TCGv EA = tcg_temp_new(TCG_TYPE_TL); \
+ gen_addr_reg_index(EA, ctx); \
+ gen_qemu_ld##width(cpu_gpr[rD(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_temp_free(EA); \
}
#define GEN_LDS(width, op, type) \
-OP_LD_TABLE(width); \
GEN_LD(width, op | 0x20, type); \
GEN_LDU(width, op | 0x21, type); \
GEN_LDUX(width, 0x17, op | 0x01, type); \
GEN_LDX(width, 0x17, op | 0x00, type)
/* lbz lbzu lbzux lbzx */
-GEN_LDS(bz, 0x02, PPC_INTEGER);
+GEN_LDS(8u, 0x02, PPC_INTEGER);
/* lha lhau lhaux lhax */
-GEN_LDS(ha, 0x0A, PPC_INTEGER);
+GEN_LDS(16s, 0x0A, PPC_INTEGER);
/* lhz lhzu lhzux lhzx */
-GEN_LDS(hz, 0x08, PPC_INTEGER);
+GEN_LDS(16u, 0x08, PPC_INTEGER);
/* lwz lwzu lwzux lwzx */
-GEN_LDS(wz, 0x00, PPC_INTEGER);
+GEN_LDS(32u, 0x00, PPC_INTEGER);
#if defined(TARGET_PPC64)
-OP_LD_TABLE(wa);
-OP_LD_TABLE(d);
/* lwaux */
-GEN_LDUX(wa, 0x15, 0x0B, PPC_64B);
+GEN_LDUX(32s, 0x15, 0x0B, PPC_64B);
/* lwax */
-GEN_LDX(wa, 0x15, 0x0A, PPC_64B);
+GEN_LDX(32s, 0x15, 0x0A, PPC_64B);
/* ldux */
-GEN_LDUX(d, 0x15, 0x01, PPC_64B);
+GEN_LDUX(64, 0x15, 0x01, PPC_64B);
/* ldx */
-GEN_LDX(d, 0x15, 0x00, PPC_64B);
+GEN_LDX(64, 0x15, 0x00, PPC_64B);
GEN_HANDLER(ld, 0x3A, 0xFF, 0xFF, 0x00000000, PPC_64B)
{
+ TCGv EA;
if (Rc(ctx->opcode)) {
if (unlikely(rA(ctx->opcode) == 0 ||
rA(ctx->opcode) == rD(ctx->opcode))) {
@@ -2280,17 +2536,18 @@ GEN_HANDLER(ld, 0x3A, 0xFF, 0xFF, 0x00000000, PPC_64B)
return;
}
}
- gen_addr_imm_index(cpu_T[0], ctx, 0x03);
+ EA = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_imm_index(EA, ctx, 0x03);
if (ctx->opcode & 0x02) {
/* lwa (lwau is undefined) */
- op_ldst(lwa);
+ gen_qemu_ld32s(cpu_gpr[rD(ctx->opcode)], EA, ctx->mem_idx);
} else {
/* ld - ldu */
- op_ldst(ld);
+ gen_qemu_ld64(cpu_gpr[rD(ctx->opcode)], EA, ctx->mem_idx);
}
- tcg_gen_mov_tl(cpu_gpr[rD(ctx->opcode)], cpu_T[1]);
if (Rc(ctx->opcode))
- tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], cpu_T[0]);
+ tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], EA);
+ tcg_temp_free(EA);
}
/* lq */
GEN_HANDLER(lq, 0x38, 0xFF, 0xFF, 0x00000000, PPC_64BX)
@@ -2299,6 +2556,7 @@ GEN_HANDLER(lq, 0x38, 0xFF, 0xFF, 0x00000000, PPC_64BX)
GEN_EXCP_PRIVOPC(ctx);
#else
int ra, rd;
+ TCGv EA;
/* Restore CPU state */
if (unlikely(ctx->supervisor == 0)) {
@@ -2316,12 +2574,12 @@ GEN_HANDLER(lq, 0x38, 0xFF, 0xFF, 0x00000000, PPC_64BX)
GEN_EXCP(ctx, POWERPC_EXCP_ALIGN, POWERPC_EXCP_ALIGN_LE);
return;
}
- gen_addr_imm_index(cpu_T[0], ctx, 0x0F);
- op_ldst(ld);
- tcg_gen_mov_tl(cpu_gpr[rd], cpu_T[1]);
- tcg_gen_addi_tl(cpu_T[0], cpu_T[0], 8);
- op_ldst(ld);
- tcg_gen_mov_tl(cpu_gpr[rd + 1], cpu_T[1]);
+ EA = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_imm_index(EA, ctx, 0x0F);
+ gen_qemu_ld64(cpu_gpr[rd], EA, ctx->mem_idx);
+ tcg_gen_addi_tl(EA, EA, 8);
+ gen_qemu_ld64(cpu_gpr[rd+1], EA, ctx->mem_idx);
+ tcg_temp_free(EA);
#endif
}
#endif
@@ -2330,68 +2588,73 @@ GEN_HANDLER(lq, 0x38, 0xFF, 0xFF, 0x00000000, PPC_64BX)
#define GEN_ST(width, opc, type) \
GEN_HANDLER(st##width, opc, 0xFF, 0xFF, 0x00000000, type) \
{ \
- gen_addr_imm_index(cpu_T[0], ctx, 0); \
- tcg_gen_mov_tl(cpu_T[1], cpu_gpr[rS(ctx->opcode)]); \
- op_ldst(st##width); \
+ TCGv EA = tcg_temp_new(TCG_TYPE_TL); \
+ gen_addr_imm_index(EA, ctx, 0); \
+ gen_qemu_st##width(cpu_gpr[rS(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_temp_free(EA); \
}
#define GEN_STU(width, opc, type) \
GEN_HANDLER(st##width##u, opc, 0xFF, 0xFF, 0x00000000, type) \
{ \
+ TCGv EA; \
if (unlikely(rA(ctx->opcode) == 0)) { \
GEN_EXCP_INVAL(ctx); \
return; \
} \
+ EA = tcg_temp_new(TCG_TYPE_TL); \
if (type == PPC_64B) \
- gen_addr_imm_index(cpu_T[0], ctx, 0x03); \
+ gen_addr_imm_index(EA, ctx, 0x03); \
else \
- gen_addr_imm_index(cpu_T[0], ctx, 0); \
- tcg_gen_mov_tl(cpu_T[1], cpu_gpr[rS(ctx->opcode)]); \
- op_ldst(st##width); \
- tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], cpu_T[0]); \
+ gen_addr_imm_index(EA, ctx, 0); \
+ gen_qemu_st##width(cpu_gpr[rS(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], EA); \
+ tcg_temp_free(EA); \
}
#define GEN_STUX(width, opc2, opc3, type) \
GEN_HANDLER(st##width##ux, 0x1F, opc2, opc3, 0x00000001, type) \
{ \
+ TCGv EA; \
if (unlikely(rA(ctx->opcode) == 0)) { \
GEN_EXCP_INVAL(ctx); \
return; \
} \
- gen_addr_reg_index(cpu_T[0], ctx); \
- tcg_gen_mov_tl(cpu_T[1], cpu_gpr[rS(ctx->opcode)]); \
- op_ldst(st##width); \
- tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], cpu_T[0]); \
+ EA = tcg_temp_new(TCG_TYPE_TL); \
+ gen_addr_reg_index(EA, ctx); \
+ gen_qemu_st##width(cpu_gpr[rS(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], EA); \
+ tcg_temp_free(EA); \
}
#define GEN_STX(width, opc2, opc3, type) \
GEN_HANDLER(st##width##x, 0x1F, opc2, opc3, 0x00000001, type) \
{ \
- gen_addr_reg_index(cpu_T[0], ctx); \
- tcg_gen_mov_tl(cpu_T[1], cpu_gpr[rS(ctx->opcode)]); \
- op_ldst(st##width); \
+ TCGv EA = tcg_temp_new(TCG_TYPE_TL); \
+ gen_addr_reg_index(EA, ctx); \
+ gen_qemu_st##width(cpu_gpr[rS(ctx->opcode)], EA, ctx->mem_idx); \
+ tcg_temp_free(EA); \
}
#define GEN_STS(width, op, type) \
-OP_ST_TABLE(width); \
GEN_ST(width, op | 0x20, type); \
GEN_STU(width, op | 0x21, type); \
GEN_STUX(width, 0x17, op | 0x01, type); \
GEN_STX(width, 0x17, op | 0x00, type)
/* stb stbu stbux stbx */
-GEN_STS(b, 0x06, PPC_INTEGER);
+GEN_STS(8, 0x06, PPC_INTEGER);
/* sth sthu sthux sthx */
-GEN_STS(h, 0x0C, PPC_INTEGER);
+GEN_STS(16, 0x0C, PPC_INTEGER);
/* stw stwu stwux stwx */
-GEN_STS(w, 0x04, PPC_INTEGER);
+GEN_STS(32, 0x04, PPC_INTEGER);
#if defined(TARGET_PPC64)
-OP_ST_TABLE(d);
-GEN_STUX(d, 0x15, 0x05, PPC_64B);
-GEN_STX(d, 0x15, 0x04, PPC_64B);
+GEN_STUX(64, 0x15, 0x05, PPC_64B);
+GEN_STX(64, 0x15, 0x04, PPC_64B);
GEN_HANDLER(std, 0x3E, 0xFF, 0xFF, 0x00000000, PPC_64B)
{
int rs;
+ TCGv EA;
rs = rS(ctx->opcode);
if ((ctx->opcode & 0x3) == 0x2) {
@@ -2412,12 +2675,12 @@ GEN_HANDLER(std, 0x3E, 0xFF, 0xFF, 0x00000000, PPC_64B)
GEN_EXCP(ctx, POWERPC_EXCP_ALIGN, POWERPC_EXCP_ALIGN_LE);
return;
}
- gen_addr_imm_index(cpu_T[0], ctx, 0x03);
- tcg_gen_mov_tl(cpu_T[1], cpu_gpr[rs]);
- op_ldst(std);
- tcg_gen_addi_tl(cpu_T[0], cpu_T[0], 8);
- tcg_gen_mov_tl(cpu_T[1], cpu_gpr[rs + 1]);
- op_ldst(std);
+ EA = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_imm_index(EA, ctx, 0x03);
+ gen_qemu_st64(cpu_gpr[rs], EA, ctx->mem_idx);
+ tcg_gen_addi_tl(EA, EA, 8);
+ gen_qemu_st64(cpu_gpr[rs+1], EA, ctx->mem_idx);
+ tcg_temp_free(EA);
#endif
} else {
/* std / stdu */
@@ -2427,27 +2690,60 @@ GEN_HANDLER(std, 0x3E, 0xFF, 0xFF, 0x00000000, PPC_64B)
return;
}
}
- gen_addr_imm_index(cpu_T[0], ctx, 0x03);
- tcg_gen_mov_tl(cpu_T[1], cpu_gpr[rs]);
- op_ldst(std);
+ EA = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_imm_index(EA, ctx, 0x03);
+ gen_qemu_st64(cpu_gpr[rs], EA, ctx->mem_idx);
if (Rc(ctx->opcode))
- tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], cpu_T[0]);
+ tcg_gen_mov_tl(cpu_gpr[rA(ctx->opcode)], EA);
+ tcg_temp_free(EA);
}
}
#endif
/*** Integer load and store with byte reverse ***/
/* lhbrx */
-OP_LD_TABLE(hbr);
-GEN_LDX(hbr, 0x16, 0x18, PPC_INTEGER);
+void always_inline gen_qemu_ld16ur(TCGv t0, TCGv t1, int flags)
+{
+ TCGv temp = tcg_temp_new(TCG_TYPE_I32);
+ gen_qemu_ld16u(temp, t1, flags);
+ tcg_gen_bswap16_i32(temp, temp);
+ tcg_gen_extu_i32_tl(t0, temp);
+ tcg_temp_free(temp);
+}
+GEN_LDX(16ur, 0x16, 0x18, PPC_INTEGER);
+
/* lwbrx */
-OP_LD_TABLE(wbr);
-GEN_LDX(wbr, 0x16, 0x10, PPC_INTEGER);
+void always_inline gen_qemu_ld32ur(TCGv t0, TCGv t1, int flags)
+{
+ TCGv temp = tcg_temp_new(TCG_TYPE_I32);
+ gen_qemu_ld32u(temp, t1, flags);
+ tcg_gen_bswap_i32(temp, temp);
+ tcg_gen_extu_i32_tl(t0, temp);
+ tcg_temp_free(temp);
+}
+GEN_LDX(32ur, 0x16, 0x10, PPC_INTEGER);
+
/* sthbrx */
-OP_ST_TABLE(hbr);
-GEN_STX(hbr, 0x16, 0x1C, PPC_INTEGER);
+void always_inline gen_qemu_st16r(TCGv t0, TCGv t1, int flags)
+{
+ TCGv temp = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(temp, t0);
+ tcg_gen_ext16u_i32(temp, temp);
+ tcg_gen_bswap16_i32(temp, temp);
+ gen_qemu_st16(temp, t1, flags);
+ tcg_temp_free(temp);
+}
+GEN_STX(16r, 0x16, 0x1C, PPC_INTEGER);
+
/* stwbrx */
-OP_ST_TABLE(wbr);
-GEN_STX(wbr, 0x16, 0x14, PPC_INTEGER);
+void always_inline gen_qemu_st32r(TCGv t0, TCGv t1, int flags)
+{
+ TCGv temp = tcg_temp_new(TCG_TYPE_I32);
+ tcg_gen_trunc_tl_i32(temp, t0);
+ tcg_gen_bswap_i32(temp, temp);
+ gen_qemu_st32(temp, t1, flags);
+ tcg_temp_free(temp);
+}
+GEN_STX(32r, 0x16, 0x14, PPC_INTEGER);
/*** Integer load and store multiple ***/
#define op_ldstm(name, reg) (*gen_op_##name[ctx->mem_idx])(reg)
@@ -3444,8 +3740,10 @@ GEN_HANDLER(mtspr, 0x1F, 0x13, 0x0E, 0x00000001, PPC_MISC)
GEN_HANDLER(dcbf, 0x1F, 0x16, 0x02, 0x03C00001, PPC_CACHE)
{
/* XXX: specification says this is treated as a load by the MMU */
- gen_addr_reg_index(cpu_T[0], ctx);
- op_ldst(lbz);
+ TCGv temp = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_reg_index(temp, ctx);
+ gen_qemu_ld8u(temp, temp, ctx->mem_idx);
+ tcg_temp_free(temp);
}
/* dcbi (Supervisor only) */
@@ -3454,14 +3752,18 @@ GEN_HANDLER(dcbi, 0x1F, 0x16, 0x0E, 0x03E00001, PPC_CACHE)
#if defined(CONFIG_USER_ONLY)
GEN_EXCP_PRIVOPC(ctx);
#else
+ TCGv EA, val;
if (unlikely(!ctx->supervisor)) {
GEN_EXCP_PRIVOPC(ctx);
return;
}
- gen_addr_reg_index(cpu_T[0], ctx);
+ EA = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_reg_index(EA, ctx);
/* XXX: specification says this should be treated as a store by the MMU */
- op_ldst(lbz);
- op_ldst(stb);
+ gen_qemu_ld8u(val, EA, ctx->mem_idx);
+ gen_qemu_st8(val, EA, ctx->mem_idx);
+ tcg_temp_free(val);
+ tcg_temp_free(EA);
#endif
}
@@ -3469,8 +3771,10 @@ GEN_HANDLER(dcbi, 0x1F, 0x16, 0x0E, 0x03E00001, PPC_CACHE)
GEN_HANDLER(dcbst, 0x1F, 0x16, 0x01, 0x03E00001, PPC_CACHE)
{
/* XXX: specification say this is treated as a load by the MMU */
- gen_addr_reg_index(cpu_T[0], ctx);
- op_ldst(lbz);
+ TCGv temp = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_reg_index(temp, ctx);
+ gen_qemu_ld8u(temp, temp, ctx->mem_idx);
+ tcg_temp_free(temp);
}
/* dcbt */
@@ -4889,13 +5193,18 @@ GEN_HANDLER(dcread, 0x1F, 0x06, 0x0F, 0x00000001, PPC_4xx_COMMON)
#if defined(CONFIG_USER_ONLY)
GEN_EXCP_PRIVOPC(ctx);
#else
+ TCGv EA, val;
if (unlikely(!ctx->supervisor)) {
GEN_EXCP_PRIVOPC(ctx);
return;
}
- gen_addr_reg_index(cpu_T[0], ctx);
- op_ldst(lwz);
- tcg_gen_mov_tl(cpu_gpr[rD(ctx->opcode)], cpu_T[0]);
+ EA = tcg_temp_new(TCG_TYPE_TL);
+ gen_addr_reg_index(EA, ctx);
+ val = tcg_temp_new(TCG_TYPE_TL);
+ gen_qemu_ld32u(val, EA, ctx->mem_idx);
+ tcg_temp_free(val);
+ tcg_gen_mov_tl(cpu_gpr[rD(ctx->opcode)], EA);
+ tcg_temp_free(EA);
#endif
}
@@ -5593,43 +5902,6 @@ GEN_HANDLER2(evsel3, "evsel", 0x04, 0x1f, 0x09, 0x00000000, PPC_SPE)
}
/* Load and stores */
-#if defined(TARGET_PPC64)
-/* In that case, we already have 64 bits load & stores
- * so, spe_ldd is equivalent to ld and spe_std is equivalent to std
- */
-#define gen_op_spe_ldd_raw gen_op_ld_raw
-#define gen_op_spe_ldd_user gen_op_ld_user
-#define gen_op_spe_ldd_kernel gen_op_ld_kernel
-#define gen_op_spe_ldd_hypv gen_op_ld_hypv
-#define gen_op_spe_ldd_64_raw gen_op_ld_64_raw
-#define gen_op_spe_ldd_64_user gen_op_ld_64_user
-#define gen_op_spe_ldd_64_kernel gen_op_ld_64_kernel
-#define gen_op_spe_ldd_64_hypv gen_op_ld_64_hypv
-#define gen_op_spe_ldd_le_raw gen_op_ld_le_raw
-#define gen_op_spe_ldd_le_user gen_op_ld_le_user
-#define gen_op_spe_ldd_le_kernel gen_op_ld_le_kernel
-#define gen_op_spe_ldd_le_hypv gen_op_ld_le_hypv
-#define gen_op_spe_ldd_le_64_raw gen_op_ld_le_64_raw
-#define gen_op_spe_ldd_le_64_user gen_op_ld_le_64_user
-#define gen_op_spe_ldd_le_64_kernel gen_op_ld_le_64_kernel
-#define gen_op_spe_ldd_le_64_hypv gen_op_ld_le_64_hypv
-#define gen_op_spe_stdd_raw gen_op_std_raw
-#define gen_op_spe_stdd_user gen_op_std_user
-#define gen_op_spe_stdd_kernel gen_op_std_kernel
-#define gen_op_spe_stdd_hypv gen_op_std_hypv
-#define gen_op_spe_stdd_64_raw gen_op_std_64_raw
-#define gen_op_spe_stdd_64_user gen_op_std_64_user
-#define gen_op_spe_stdd_64_kernel gen_op_std_64_kernel
-#define gen_op_spe_stdd_64_hypv gen_op_std_64_hypv
-#define gen_op_spe_stdd_le_raw gen_op_std_le_raw
-#define gen_op_spe_stdd_le_user gen_op_std_le_user
-#define gen_op_spe_stdd_le_kernel gen_op_std_le_kernel
-#define gen_op_spe_stdd_le_hypv gen_op_std_le_hypv
-#define gen_op_spe_stdd_le_64_raw gen_op_std_le_64_raw
-#define gen_op_spe_stdd_le_64_user gen_op_std_le_64_user
-#define gen_op_spe_stdd_le_64_kernel gen_op_std_le_64_kernel
-#define gen_op_spe_stdd_le_64_hypv gen_op_std_le_64_hypv
-#endif /* defined(TARGET_PPC64) */
GEN_SPEOP_LDST(dd, 3);
GEN_SPEOP_LDST(dw, 3);
GEN_SPEOP_LDST(dh, 3);
@@ -5638,25 +5910,6 @@ GEN_SPEOP_LD(whou, 2);
GEN_SPEOP_LD(whos, 2);
GEN_SPEOP_ST(who, 2);
-#if defined(TARGET_PPC64)
-/* In that case, spe_stwwo is equivalent to stw */
-#define gen_op_spe_stwwo_raw gen_op_stw_raw
-#define gen_op_spe_stwwo_user gen_op_stw_user
-#define gen_op_spe_stwwo_kernel gen_op_stw_kernel
-#define gen_op_spe_stwwo_hypv gen_op_stw_hypv
-#define gen_op_spe_stwwo_le_raw gen_op_stw_le_raw
-#define gen_op_spe_stwwo_le_user gen_op_stw_le_user
-#define gen_op_spe_stwwo_le_kernel gen_op_stw_le_kernel
-#define gen_op_spe_stwwo_le_hypv gen_op_stw_le_hypv
-#define gen_op_spe_stwwo_64_raw gen_op_stw_64_raw
-#define gen_op_spe_stwwo_64_user gen_op_stw_64_user
-#define gen_op_spe_stwwo_64_kernel gen_op_stw_64_kernel
-#define gen_op_spe_stwwo_64_hypv gen_op_stw_64_hypv
-#define gen_op_spe_stwwo_le_64_raw gen_op_stw_le_64_raw
-#define gen_op_spe_stwwo_le_64_user gen_op_stw_le_64_user
-#define gen_op_spe_stwwo_le_64_kernel gen_op_stw_le_64_kernel
-#define gen_op_spe_stwwo_le_64_hypv gen_op_stw_le_64_hypv
-#endif
#define _GEN_OP_SPE_STWWE(suffix) \
static always_inline void gen_op_spe_stwwe_##suffix (void) \
{ \