diff mbox

target-tilegx: Finish parsing all opcode for bundle

Message ID BLU436-SMTP16536D2938445A2332E6778B91A0@phx.gbl
State New
Headers show

Commit Message

Chen Gang March 8, 2015, 12:54 a.m. UTC
According to "arch/tile/include/uapi/arch/opcode_tilegx.h" in Linux
kernel source code, qemu can parse all opcode:

  "grep _OPCODE_X0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
  "grep _OPCODE_X1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
  "grep _OPCODE_Y0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
  "grep _OPCODE_Y1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
  "grep _OPCODE_Y2 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"

Signed-off-by: Chen Gang <gang.chen.5i5j@gmail.com>
---
 target-tilegx/cpu-qom.h       |    2 +
 target-tilegx/cpu.c           |    4 -
 target-tilegx/cpu.h           |   22 +-
 target-tilegx/opcode_tilegx.h | 1406 +++++++++++++++++++++++++++++++++++++++++
 target-tilegx/translate.c     |  612 +++++++++++++++++-
 5 files changed, 2026 insertions(+), 20 deletions(-)
 create mode 100644 target-tilegx/opcode_tilegx.h

Comments

Chen Gang March 12, 2015, 3:52 p.m. UTC | #1
I shall send a new patch, next. This patch will be obsoleted.

Thanks.


On 3/8/15 08:54, Chen Gang wrote:
> According to "arch/tile/include/uapi/arch/opcode_tilegx.h" in Linux
> kernel source code, qemu can parse all opcode:
> 
>   "grep _OPCODE_X0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
>   "grep _OPCODE_X1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
>   "grep _OPCODE_Y0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
>   "grep _OPCODE_Y1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
>   "grep _OPCODE_Y2 opcode_tilegx.h | awk '{print $3, $1}' | sort -n"
> 
> Signed-off-by: Chen Gang <gang.chen.5i5j@gmail.com>
> ---
>  target-tilegx/cpu-qom.h       |    2 +
>  target-tilegx/cpu.c           |    4 -
>  target-tilegx/cpu.h           |   22 +-
>  target-tilegx/opcode_tilegx.h | 1406 +++++++++++++++++++++++++++++++++++++++++
>  target-tilegx/translate.c     |  612 +++++++++++++++++-
>  5 files changed, 2026 insertions(+), 20 deletions(-)
>  create mode 100644 target-tilegx/opcode_tilegx.h
> 
> diff --git a/target-tilegx/cpu-qom.h b/target-tilegx/cpu-qom.h
> index 4ee11e1..5615c3b 100644
> --- a/target-tilegx/cpu-qom.h
> +++ b/target-tilegx/cpu-qom.h
> @@ -68,4 +68,6 @@ static inline TileGXCPU *tilegx_env_get_cpu(CPUTLGState *env)
>  
>  #define ENV_GET_CPU(e) CPU(tilegx_env_get_cpu(e))
>  
> +#define ENV_OFFSET offsetof(TileGXCPU, env)
> +
>  #endif
> diff --git a/target-tilegx/cpu.c b/target-tilegx/cpu.c
> index cf46b8b..8255fdc 100644
> --- a/target-tilegx/cpu.c
> +++ b/target-tilegx/cpu.c
> @@ -69,10 +69,6 @@ static void tilegx_cpu_realizefn(DeviceState *dev, Error **errp)
>      tcc->parent_realize(dev, errp);
>  }
>  
> -static void tilegx_tcg_init(void)
> -{
> -}
> -
>  static void tilegx_cpu_initfn(Object *obj)
>  {
>      CPUState *cs = CPU(obj);
> diff --git a/target-tilegx/cpu.h b/target-tilegx/cpu.h
> index 87dc56b..93e16c3 100644
> --- a/target-tilegx/cpu.h
> +++ b/target-tilegx/cpu.h
> @@ -30,16 +30,21 @@
>  #include "fpu/softfloat.h"
>  
>  /* TILE-Gx register alias */
> -#define TILEGX_R_RE  0   /*  0 register, for function/syscall return value */
> -#define TILEGX_R_NR  10  /* 10 register, for syscall number */
> -#define TILEGX_R_BP  52  /* 52 register, optional frame pointer */
> -#define TILEGX_R_TP  53  /* TP register, thread local storage data */
> -#define TILEGX_R_SP  54  /* SP register, stack pointer */
> -#define TILEGX_R_LR  55  /* LR register, may save pc, but it is not pc */
> +#define TILEGX_R_RE    0   /*  0 register, for function/syscall return value */
> +#define TILEGX_R_NR    10  /* 10 register, for syscall number */
> +#define TILEGX_R_BP    52  /* 52 register, optional frame pointer */
> +#define TILEGX_R_TP    53  /* TP register, thread local storage data */
> +#define TILEGX_R_SP    54  /* SP register, stack pointer */
> +#define TILEGX_R_LR    55  /* LR register, may save pc, but it is not pc */
> +#define TILEGX_R_ZERO  63  /* Zero register, always zero */
> +#define TILEGX_R_COUNT 56  /* Only 56 registers are really useful */
> +#define TILEGX_R_NOREG 255 /* Invalid register value */
> +
>  
>  typedef struct CPUTLGState {
> -    uint64_t regs[56];
> -    uint64_t pc;
> +    uint64_t regs[TILEGX_R_COUNT]; /* Common used registers by outside */
> +    uint64_t pc;                   /* Current pc */
> +
>      CPU_COMMON
>  } CPUTLGState;
>  
> @@ -54,6 +59,7 @@ typedef struct CPUTLGState {
>  
>  #include "exec/cpu-all.h"
>  
> +void tilegx_tcg_init(void);
>  int cpu_tilegx_exec(CPUTLGState *s);
>  int cpu_tilegx_signal_handler(int host_signum, void *pinfo, void *puc);
>  
> diff --git a/target-tilegx/opcode_tilegx.h b/target-tilegx/opcode_tilegx.h
> new file mode 100644
> index 0000000..d76ff2d
> --- /dev/null
> +++ b/target-tilegx/opcode_tilegx.h
> @@ -0,0 +1,1406 @@
> +/* TILE-Gx opcode information.
> + *
> + * Copyright 2011 Tilera Corporation. All Rights Reserved.
> + *
> + *   This program is free software; you can redistribute it and/or
> + *   modify it under the terms of the GNU General Public License
> + *   as published by the Free Software Foundation, version 2.
> + *
> + *   This program is distributed in the hope that it will be useful, but
> + *   WITHOUT ANY WARRANTY; without even the implied warranty of
> + *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
> + *   NON INFRINGEMENT.  See the GNU General Public License for
> + *   more details.
> + *
> + *
> + *
> + *
> + *
> + */
> +
> +#ifndef __ARCH_OPCODE_H__
> +#define __ARCH_OPCODE_H__
> +
> +#ifndef __ASSEMBLER__
> +
> +typedef unsigned long long tilegx_bundle_bits;
> +
> +/* These are the bits that determine if a bundle is in the X encoding. */
> +#define TILEGX_BUNDLE_MODE_MASK ((tilegx_bundle_bits)3 << 62)
> +
> +enum
> +{
> +  /* Maximum number of instructions in a bundle (2 for X, 3 for Y). */
> +  TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE = 3,
> +
> +  /* How many different pipeline encodings are there? X0, X1, Y0, Y1, Y2. */
> +  TILEGX_NUM_PIPELINE_ENCODINGS = 5,
> +
> +  /* Log base 2 of TILEGX_BUNDLE_SIZE_IN_BYTES. */
> +  TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES = 3,
> +
> +  /* Instructions take this many bytes. */
> +  TILEGX_BUNDLE_SIZE_IN_BYTES = 1 << TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES,
> +
> +  /* Log base 2 of TILEGX_BUNDLE_ALIGNMENT_IN_BYTES. */
> +  TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES = 3,
> +
> +  /* Bundles should be aligned modulo this number of bytes. */
> +  TILEGX_BUNDLE_ALIGNMENT_IN_BYTES =
> +    (1 << TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES),
> +
> +  /* Number of registers (some are magic, such as network I/O). */
> +  TILEGX_NUM_REGISTERS = 64,
> +};
> +
> +/* Make a few "tile_" variables to simplify common code between
> +   architectures.  */
> +
> +typedef tilegx_bundle_bits tile_bundle_bits;
> +#define TILE_BUNDLE_SIZE_IN_BYTES TILEGX_BUNDLE_SIZE_IN_BYTES
> +#define TILE_BUNDLE_ALIGNMENT_IN_BYTES TILEGX_BUNDLE_ALIGNMENT_IN_BYTES
> +#define TILE_LOG2_BUNDLE_ALIGNMENT_IN_BYTES \
> +  TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES
> +#define TILE_BPT_BUNDLE TILEGX_BPT_BUNDLE
> +
> +/* 64-bit pattern for a { bpt ; nop } bundle. */
> +#define TILEGX_BPT_BUNDLE 0x286a44ae51485000ULL
> +
> +static __inline unsigned int
> +get_BFEnd_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_BFOpcodeExtension_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 24)) & 0xf);
> +}
> +
> +static __inline unsigned int
> +get_BFStart_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 18)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_BrOff_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 31)) & 0x0000003f) |
> +         (((unsigned int)(n >> 37)) & 0x0001ffc0);
> +}
> +
> +static __inline unsigned int
> +get_BrType_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 54)) & 0x1f);
> +}
> +
> +static __inline unsigned int
> +get_Dest_Imm8_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 31)) & 0x0000003f) |
> +         (((unsigned int)(n >> 43)) & 0x000000c0);
> +}
> +
> +static __inline unsigned int
> +get_Dest_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 0)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_Dest_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 31)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_Dest_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 0)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_Dest_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 31)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_Imm16_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0xffff);
> +}
> +
> +static __inline unsigned int
> +get_Imm16_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0xffff);
> +}
> +
> +static __inline unsigned int
> +get_Imm8OpcodeExtension_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 20)) & 0xff);
> +}
> +
> +static __inline unsigned int
> +get_Imm8OpcodeExtension_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 51)) & 0xff);
> +}
> +
> +static __inline unsigned int
> +get_Imm8_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0xff);
> +}
> +
> +static __inline unsigned int
> +get_Imm8_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0xff);
> +}
> +
> +static __inline unsigned int
> +get_Imm8_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0xff);
> +}
> +
> +static __inline unsigned int
> +get_Imm8_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0xff);
> +}
> +
> +static __inline unsigned int
> +get_JumpOff_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 31)) & 0x7ffffff);
> +}
> +
> +static __inline unsigned int
> +get_JumpOpcodeExtension_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 58)) & 0x1);
> +}
> +
> +static __inline unsigned int
> +get_MF_Imm14_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 37)) & 0x3fff);
> +}
> +
> +static __inline unsigned int
> +get_MT_Imm14_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 31)) & 0x0000003f) |
> +         (((unsigned int)(n >> 37)) & 0x00003fc0);
> +}
> +
> +static __inline unsigned int
> +get_Mode(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 62)) & 0x3);
> +}
> +
> +static __inline unsigned int
> +get_Opcode_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 28)) & 0x7);
> +}
> +
> +static __inline unsigned int
> +get_Opcode_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 59)) & 0x7);
> +}
> +
> +static __inline unsigned int
> +get_Opcode_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 27)) & 0xf);
> +}
> +
> +static __inline unsigned int
> +get_Opcode_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 58)) & 0xf);
> +}
> +
> +static __inline unsigned int
> +get_Opcode_Y2(tilegx_bundle_bits n)
> +{
> +  return (((n >> 26)) & 0x00000001) |
> +         (((unsigned int)(n >> 56)) & 0x00000002);
> +}
> +
> +static __inline unsigned int
> +get_RRROpcodeExtension_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 18)) & 0x3ff);
> +}
> +
> +static __inline unsigned int
> +get_RRROpcodeExtension_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 49)) & 0x3ff);
> +}
> +
> +static __inline unsigned int
> +get_RRROpcodeExtension_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 18)) & 0x3);
> +}
> +
> +static __inline unsigned int
> +get_RRROpcodeExtension_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 49)) & 0x3);
> +}
> +
> +static __inline unsigned int
> +get_ShAmt_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_ShAmt_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_ShAmt_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_ShAmt_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_ShiftOpcodeExtension_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 18)) & 0x3ff);
> +}
> +
> +static __inline unsigned int
> +get_ShiftOpcodeExtension_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 49)) & 0x3ff);
> +}
> +
> +static __inline unsigned int
> +get_ShiftOpcodeExtension_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 18)) & 0x3);
> +}
> +
> +static __inline unsigned int
> +get_ShiftOpcodeExtension_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 49)) & 0x3);
> +}
> +
> +static __inline unsigned int
> +get_SrcA_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 6)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcA_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 37)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcA_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 6)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcA_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 37)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcA_Y2(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 20)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcBDest_Y2(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 51)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcB_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcB_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcB_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_SrcB_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_UnaryOpcodeExtension_X0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_UnaryOpcodeExtension_X1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_UnaryOpcodeExtension_Y0(tilegx_bundle_bits num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((n >> 12)) & 0x3f);
> +}
> +
> +static __inline unsigned int
> +get_UnaryOpcodeExtension_Y1(tilegx_bundle_bits n)
> +{
> +  return (((unsigned int)(n >> 43)) & 0x3f);
> +}
> +
> +
> +static __inline int
> +sign_extend(int n, int num_bits)
> +{
> +  int shift = (int)(sizeof(int) * 8 - num_bits);
> +  return (n << shift) >> shift;
> +}
> +
> +
> +
> +static __inline tilegx_bundle_bits
> +create_BFEnd_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_BFOpcodeExtension_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0xf) << 24);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_BFStart_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 18);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_BrOff_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
> +         (((tilegx_bundle_bits)(n & 0x0001ffc0)) << 37);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_BrType_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x1f)) << 54);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Dest_Imm8_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
> +         (((tilegx_bundle_bits)(n & 0x000000c0)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Dest_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 0);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Dest_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 31);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Dest_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 0);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Dest_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 31);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm16_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0xffff) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm16_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0xffff)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm8OpcodeExtension_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0xff) << 20);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm8OpcodeExtension_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0xff)) << 51);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm8_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0xff) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm8_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0xff)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm8_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0xff) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Imm8_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0xff)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_JumpOff_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x7ffffff)) << 31);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_JumpOpcodeExtension_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x1)) << 58);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_MF_Imm14_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3fff)) << 37);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_MT_Imm14_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
> +         (((tilegx_bundle_bits)(n & 0x00003fc0)) << 37);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Mode(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3)) << 62);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Opcode_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x7) << 28);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Opcode_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x7)) << 59);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Opcode_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0xf) << 27);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Opcode_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0xf)) << 58);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_Opcode_Y2(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x00000001) << 26) |
> +         (((tilegx_bundle_bits)(n & 0x00000002)) << 56);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_RRROpcodeExtension_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3ff) << 18);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_RRROpcodeExtension_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3ff)) << 49);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_RRROpcodeExtension_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3) << 18);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_RRROpcodeExtension_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3)) << 49);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShAmt_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShAmt_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShAmt_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShAmt_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShiftOpcodeExtension_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3ff) << 18);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShiftOpcodeExtension_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3ff)) << 49);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShiftOpcodeExtension_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3) << 18);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_ShiftOpcodeExtension_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3)) << 49);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcA_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 6);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcA_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 37);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcA_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 6);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcA_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 37);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcA_Y2(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 20);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcBDest_Y2(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 51);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcB_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcB_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcB_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_SrcB_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_UnaryOpcodeExtension_X0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_UnaryOpcodeExtension_X1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_UnaryOpcodeExtension_Y0(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return ((n & 0x3f) << 12);
> +}
> +
> +static __inline tilegx_bundle_bits
> +create_UnaryOpcodeExtension_Y1(int num)
> +{
> +  const unsigned int n = (unsigned int)num;
> +  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
> +}
> +
> +
> +enum
> +{
> +  ADDI_IMM8_OPCODE_X0 = 1,
> +  ADDI_IMM8_OPCODE_X1 = 1,
> +  ADDI_OPCODE_Y0 = 0,
> +  ADDI_OPCODE_Y1 = 1,
> +  ADDLI_OPCODE_X0 = 1,
> +  ADDLI_OPCODE_X1 = 0,
> +  ADDXI_IMM8_OPCODE_X0 = 2,
> +  ADDXI_IMM8_OPCODE_X1 = 2,
> +  ADDXI_OPCODE_Y0 = 1,
> +  ADDXI_OPCODE_Y1 = 2,
> +  ADDXLI_OPCODE_X0 = 2,
> +  ADDXLI_OPCODE_X1 = 1,
> +  ADDXSC_RRR_0_OPCODE_X0 = 1,
> +  ADDXSC_RRR_0_OPCODE_X1 = 1,
> +  ADDX_RRR_0_OPCODE_X0 = 2,
> +  ADDX_RRR_0_OPCODE_X1 = 2,
> +  ADDX_RRR_0_OPCODE_Y0 = 0,
> +  ADDX_SPECIAL_0_OPCODE_Y1 = 0,
> +  ADD_RRR_0_OPCODE_X0 = 3,
> +  ADD_RRR_0_OPCODE_X1 = 3,
> +  ADD_RRR_0_OPCODE_Y0 = 1,
> +  ADD_SPECIAL_0_OPCODE_Y1 = 1,
> +  ANDI_IMM8_OPCODE_X0 = 3,
> +  ANDI_IMM8_OPCODE_X1 = 3,
> +  ANDI_OPCODE_Y0 = 2,
> +  ANDI_OPCODE_Y1 = 3,
> +  AND_RRR_0_OPCODE_X0 = 4,
> +  AND_RRR_0_OPCODE_X1 = 4,
> +  AND_RRR_5_OPCODE_Y0 = 0,
> +  AND_RRR_5_OPCODE_Y1 = 0,
> +  BEQZT_BRANCH_OPCODE_X1 = 16,
> +  BEQZ_BRANCH_OPCODE_X1 = 17,
> +  BFEXTS_BF_OPCODE_X0 = 4,
> +  BFEXTU_BF_OPCODE_X0 = 5,
> +  BFINS_BF_OPCODE_X0 = 6,
> +  BF_OPCODE_X0 = 3,
> +  BGEZT_BRANCH_OPCODE_X1 = 18,
> +  BGEZ_BRANCH_OPCODE_X1 = 19,
> +  BGTZT_BRANCH_OPCODE_X1 = 20,
> +  BGTZ_BRANCH_OPCODE_X1 = 21,
> +  BLBCT_BRANCH_OPCODE_X1 = 22,
> +  BLBC_BRANCH_OPCODE_X1 = 23,
> +  BLBST_BRANCH_OPCODE_X1 = 24,
> +  BLBS_BRANCH_OPCODE_X1 = 25,
> +  BLEZT_BRANCH_OPCODE_X1 = 26,
> +  BLEZ_BRANCH_OPCODE_X1 = 27,
> +  BLTZT_BRANCH_OPCODE_X1 = 28,
> +  BLTZ_BRANCH_OPCODE_X1 = 29,
> +  BNEZT_BRANCH_OPCODE_X1 = 30,
> +  BNEZ_BRANCH_OPCODE_X1 = 31,
> +  BRANCH_OPCODE_X1 = 2,
> +  CMOVEQZ_RRR_0_OPCODE_X0 = 5,
> +  CMOVEQZ_RRR_4_OPCODE_Y0 = 0,
> +  CMOVNEZ_RRR_0_OPCODE_X0 = 6,
> +  CMOVNEZ_RRR_4_OPCODE_Y0 = 1,
> +  CMPEQI_IMM8_OPCODE_X0 = 4,
> +  CMPEQI_IMM8_OPCODE_X1 = 4,
> +  CMPEQI_OPCODE_Y0 = 3,
> +  CMPEQI_OPCODE_Y1 = 4,
> +  CMPEQ_RRR_0_OPCODE_X0 = 7,
> +  CMPEQ_RRR_0_OPCODE_X1 = 5,
> +  CMPEQ_RRR_3_OPCODE_Y0 = 0,
> +  CMPEQ_RRR_3_OPCODE_Y1 = 2,
> +  CMPEXCH4_RRR_0_OPCODE_X1 = 6,
> +  CMPEXCH_RRR_0_OPCODE_X1 = 7,
> +  CMPLES_RRR_0_OPCODE_X0 = 8,
> +  CMPLES_RRR_0_OPCODE_X1 = 8,
> +  CMPLES_RRR_2_OPCODE_Y0 = 0,
> +  CMPLES_RRR_2_OPCODE_Y1 = 0,
> +  CMPLEU_RRR_0_OPCODE_X0 = 9,
> +  CMPLEU_RRR_0_OPCODE_X1 = 9,
> +  CMPLEU_RRR_2_OPCODE_Y0 = 1,
> +  CMPLEU_RRR_2_OPCODE_Y1 = 1,
> +  CMPLTSI_IMM8_OPCODE_X0 = 5,
> +  CMPLTSI_IMM8_OPCODE_X1 = 5,
> +  CMPLTSI_OPCODE_Y0 = 4,
> +  CMPLTSI_OPCODE_Y1 = 5,
> +  CMPLTS_RRR_0_OPCODE_X0 = 10,
> +  CMPLTS_RRR_0_OPCODE_X1 = 10,
> +  CMPLTS_RRR_2_OPCODE_Y0 = 2,
> +  CMPLTS_RRR_2_OPCODE_Y1 = 2,
> +  CMPLTUI_IMM8_OPCODE_X0 = 6,
> +  CMPLTUI_IMM8_OPCODE_X1 = 6,
> +  CMPLTU_RRR_0_OPCODE_X0 = 11,
> +  CMPLTU_RRR_0_OPCODE_X1 = 11,
> +  CMPLTU_RRR_2_OPCODE_Y0 = 3,
> +  CMPLTU_RRR_2_OPCODE_Y1 = 3,
> +  CMPNE_RRR_0_OPCODE_X0 = 12,
> +  CMPNE_RRR_0_OPCODE_X1 = 12,
> +  CMPNE_RRR_3_OPCODE_Y0 = 1,
> +  CMPNE_RRR_3_OPCODE_Y1 = 3,
> +  CMULAF_RRR_0_OPCODE_X0 = 13,
> +  CMULA_RRR_0_OPCODE_X0 = 14,
> +  CMULFR_RRR_0_OPCODE_X0 = 15,
> +  CMULF_RRR_0_OPCODE_X0 = 16,
> +  CMULHR_RRR_0_OPCODE_X0 = 17,
> +  CMULH_RRR_0_OPCODE_X0 = 18,
> +  CMUL_RRR_0_OPCODE_X0 = 19,
> +  CNTLZ_UNARY_OPCODE_X0 = 1,
> +  CNTLZ_UNARY_OPCODE_Y0 = 1,
> +  CNTTZ_UNARY_OPCODE_X0 = 2,
> +  CNTTZ_UNARY_OPCODE_Y0 = 2,
> +  CRC32_32_RRR_0_OPCODE_X0 = 20,
> +  CRC32_8_RRR_0_OPCODE_X0 = 21,
> +  DBLALIGN2_RRR_0_OPCODE_X0 = 22,
> +  DBLALIGN2_RRR_0_OPCODE_X1 = 13,
> +  DBLALIGN4_RRR_0_OPCODE_X0 = 23,
> +  DBLALIGN4_RRR_0_OPCODE_X1 = 14,
> +  DBLALIGN6_RRR_0_OPCODE_X0 = 24,
> +  DBLALIGN6_RRR_0_OPCODE_X1 = 15,
> +  DBLALIGN_RRR_0_OPCODE_X0 = 25,
> +  DRAIN_UNARY_OPCODE_X1 = 1,
> +  DTLBPR_UNARY_OPCODE_X1 = 2,
> +  EXCH4_RRR_0_OPCODE_X1 = 16,
> +  EXCH_RRR_0_OPCODE_X1 = 17,
> +  FDOUBLE_ADDSUB_RRR_0_OPCODE_X0 = 26,
> +  FDOUBLE_ADD_FLAGS_RRR_0_OPCODE_X0 = 27,
> +  FDOUBLE_MUL_FLAGS_RRR_0_OPCODE_X0 = 28,
> +  FDOUBLE_PACK1_RRR_0_OPCODE_X0 = 29,
> +  FDOUBLE_PACK2_RRR_0_OPCODE_X0 = 30,
> +  FDOUBLE_SUB_FLAGS_RRR_0_OPCODE_X0 = 31,
> +  FDOUBLE_UNPACK_MAX_RRR_0_OPCODE_X0 = 32,
> +  FDOUBLE_UNPACK_MIN_RRR_0_OPCODE_X0 = 33,
> +  FETCHADD4_RRR_0_OPCODE_X1 = 18,
> +  FETCHADDGEZ4_RRR_0_OPCODE_X1 = 19,
> +  FETCHADDGEZ_RRR_0_OPCODE_X1 = 20,
> +  FETCHADD_RRR_0_OPCODE_X1 = 21,
> +  FETCHAND4_RRR_0_OPCODE_X1 = 22,
> +  FETCHAND_RRR_0_OPCODE_X1 = 23,
> +  FETCHOR4_RRR_0_OPCODE_X1 = 24,
> +  FETCHOR_RRR_0_OPCODE_X1 = 25,
> +  FINV_UNARY_OPCODE_X1 = 3,
> +  FLUSHWB_UNARY_OPCODE_X1 = 4,
> +  FLUSH_UNARY_OPCODE_X1 = 5,
> +  FNOP_UNARY_OPCODE_X0 = 3,
> +  FNOP_UNARY_OPCODE_X1 = 6,
> +  FNOP_UNARY_OPCODE_Y0 = 3,
> +  FNOP_UNARY_OPCODE_Y1 = 8,
> +  FSINGLE_ADD1_RRR_0_OPCODE_X0 = 34,
> +  FSINGLE_ADDSUB2_RRR_0_OPCODE_X0 = 35,
> +  FSINGLE_MUL1_RRR_0_OPCODE_X0 = 36,
> +  FSINGLE_MUL2_RRR_0_OPCODE_X0 = 37,
> +  FSINGLE_PACK1_UNARY_OPCODE_X0 = 4,
> +  FSINGLE_PACK1_UNARY_OPCODE_Y0 = 4,
> +  FSINGLE_PACK2_RRR_0_OPCODE_X0 = 38,
> +  FSINGLE_SUB1_RRR_0_OPCODE_X0 = 39,
> +  ICOH_UNARY_OPCODE_X1 = 7,
> +  ILL_UNARY_OPCODE_X1 = 8,
> +  ILL_UNARY_OPCODE_Y1 = 9,
> +  IMM8_OPCODE_X0 = 4,
> +  IMM8_OPCODE_X1 = 3,
> +  INV_UNARY_OPCODE_X1 = 9,
> +  IRET_UNARY_OPCODE_X1 = 10,
> +  JALRP_UNARY_OPCODE_X1 = 11,
> +  JALRP_UNARY_OPCODE_Y1 = 10,
> +  JALR_UNARY_OPCODE_X1 = 12,
> +  JALR_UNARY_OPCODE_Y1 = 11,
> +  JAL_JUMP_OPCODE_X1 = 0,
> +  JRP_UNARY_OPCODE_X1 = 13,
> +  JRP_UNARY_OPCODE_Y1 = 12,
> +  JR_UNARY_OPCODE_X1 = 14,
> +  JR_UNARY_OPCODE_Y1 = 13,
> +  JUMP_OPCODE_X1 = 4,
> +  J_JUMP_OPCODE_X1 = 1,
> +  LD1S_ADD_IMM8_OPCODE_X1 = 7,
> +  LD1S_OPCODE_Y2 = 0,
> +  LD1S_UNARY_OPCODE_X1 = 15,
> +  LD1U_ADD_IMM8_OPCODE_X1 = 8,
> +  LD1U_OPCODE_Y2 = 1,
> +  LD1U_UNARY_OPCODE_X1 = 16,
> +  LD2S_ADD_IMM8_OPCODE_X1 = 9,
> +  LD2S_OPCODE_Y2 = 2,
> +  LD2S_UNARY_OPCODE_X1 = 17,
> +  LD2U_ADD_IMM8_OPCODE_X1 = 10,
> +  LD2U_OPCODE_Y2 = 3,
> +  LD2U_UNARY_OPCODE_X1 = 18,
> +  LD4S_ADD_IMM8_OPCODE_X1 = 11,
> +  LD4S_OPCODE_Y2 = 1,
> +  LD4S_UNARY_OPCODE_X1 = 19,
> +  LD4U_ADD_IMM8_OPCODE_X1 = 12,
> +  LD4U_OPCODE_Y2 = 2,
> +  LD4U_UNARY_OPCODE_X1 = 20,
> +  LDNA_UNARY_OPCODE_X1 = 21,
> +  LDNT1S_ADD_IMM8_OPCODE_X1 = 13,
> +  LDNT1S_UNARY_OPCODE_X1 = 22,
> +  LDNT1U_ADD_IMM8_OPCODE_X1 = 14,
> +  LDNT1U_UNARY_OPCODE_X1 = 23,
> +  LDNT2S_ADD_IMM8_OPCODE_X1 = 15,
> +  LDNT2S_UNARY_OPCODE_X1 = 24,
> +  LDNT2U_ADD_IMM8_OPCODE_X1 = 16,
> +  LDNT2U_UNARY_OPCODE_X1 = 25,
> +  LDNT4S_ADD_IMM8_OPCODE_X1 = 17,
> +  LDNT4S_UNARY_OPCODE_X1 = 26,
> +  LDNT4U_ADD_IMM8_OPCODE_X1 = 18,
> +  LDNT4U_UNARY_OPCODE_X1 = 27,
> +  LDNT_ADD_IMM8_OPCODE_X1 = 19,
> +  LDNT_UNARY_OPCODE_X1 = 28,
> +  LD_ADD_IMM8_OPCODE_X1 = 20,
> +  LD_OPCODE_Y2 = 3,
> +  LD_UNARY_OPCODE_X1 = 29,
> +  LNK_UNARY_OPCODE_X1 = 30,
> +  LNK_UNARY_OPCODE_Y1 = 14,
> +  LWNA_ADD_IMM8_OPCODE_X1 = 21,
> +  MFSPR_IMM8_OPCODE_X1 = 22,
> +  MF_UNARY_OPCODE_X1 = 31,
> +  MM_BF_OPCODE_X0 = 7,
> +  MNZ_RRR_0_OPCODE_X0 = 40,
> +  MNZ_RRR_0_OPCODE_X1 = 26,
> +  MNZ_RRR_4_OPCODE_Y0 = 2,
> +  MNZ_RRR_4_OPCODE_Y1 = 2,
> +  MODE_OPCODE_YA2 = 1,
> +  MODE_OPCODE_YB2 = 2,
> +  MODE_OPCODE_YC2 = 3,
> +  MTSPR_IMM8_OPCODE_X1 = 23,
> +  MULAX_RRR_0_OPCODE_X0 = 41,
> +  MULAX_RRR_3_OPCODE_Y0 = 2,
> +  MULA_HS_HS_RRR_0_OPCODE_X0 = 42,
> +  MULA_HS_HS_RRR_9_OPCODE_Y0 = 0,
> +  MULA_HS_HU_RRR_0_OPCODE_X0 = 43,
> +  MULA_HS_LS_RRR_0_OPCODE_X0 = 44,
> +  MULA_HS_LU_RRR_0_OPCODE_X0 = 45,
> +  MULA_HU_HU_RRR_0_OPCODE_X0 = 46,
> +  MULA_HU_HU_RRR_9_OPCODE_Y0 = 1,
> +  MULA_HU_LS_RRR_0_OPCODE_X0 = 47,
> +  MULA_HU_LU_RRR_0_OPCODE_X0 = 48,
> +  MULA_LS_LS_RRR_0_OPCODE_X0 = 49,
> +  MULA_LS_LS_RRR_9_OPCODE_Y0 = 2,
> +  MULA_LS_LU_RRR_0_OPCODE_X0 = 50,
> +  MULA_LU_LU_RRR_0_OPCODE_X0 = 51,
> +  MULA_LU_LU_RRR_9_OPCODE_Y0 = 3,
> +  MULX_RRR_0_OPCODE_X0 = 52,
> +  MULX_RRR_3_OPCODE_Y0 = 3,
> +  MUL_HS_HS_RRR_0_OPCODE_X0 = 53,
> +  MUL_HS_HS_RRR_8_OPCODE_Y0 = 0,
> +  MUL_HS_HU_RRR_0_OPCODE_X0 = 54,
> +  MUL_HS_LS_RRR_0_OPCODE_X0 = 55,
> +  MUL_HS_LU_RRR_0_OPCODE_X0 = 56,
> +  MUL_HU_HU_RRR_0_OPCODE_X0 = 57,
> +  MUL_HU_HU_RRR_8_OPCODE_Y0 = 1,
> +  MUL_HU_LS_RRR_0_OPCODE_X0 = 58,
> +  MUL_HU_LU_RRR_0_OPCODE_X0 = 59,
> +  MUL_LS_LS_RRR_0_OPCODE_X0 = 60,
> +  MUL_LS_LS_RRR_8_OPCODE_Y0 = 2,
> +  MUL_LS_LU_RRR_0_OPCODE_X0 = 61,
> +  MUL_LU_LU_RRR_0_OPCODE_X0 = 62,
> +  MUL_LU_LU_RRR_8_OPCODE_Y0 = 3,
> +  MZ_RRR_0_OPCODE_X0 = 63,
> +  MZ_RRR_0_OPCODE_X1 = 27,
> +  MZ_RRR_4_OPCODE_Y0 = 3,
> +  MZ_RRR_4_OPCODE_Y1 = 3,
> +  NAP_UNARY_OPCODE_X1 = 32,
> +  NOP_UNARY_OPCODE_X0 = 5,
> +  NOP_UNARY_OPCODE_X1 = 33,
> +  NOP_UNARY_OPCODE_Y0 = 5,
> +  NOP_UNARY_OPCODE_Y1 = 15,
> +  NOR_RRR_0_OPCODE_X0 = 64,
> +  NOR_RRR_0_OPCODE_X1 = 28,
> +  NOR_RRR_5_OPCODE_Y0 = 1,
> +  NOR_RRR_5_OPCODE_Y1 = 1,
> +  ORI_IMM8_OPCODE_X0 = 7,
> +  ORI_IMM8_OPCODE_X1 = 24,
> +  OR_RRR_0_OPCODE_X0 = 65,
> +  OR_RRR_0_OPCODE_X1 = 29,
> +  OR_RRR_5_OPCODE_Y0 = 2,
> +  OR_RRR_5_OPCODE_Y1 = 2,
> +  PCNT_UNARY_OPCODE_X0 = 6,
> +  PCNT_UNARY_OPCODE_Y0 = 6,
> +  REVBITS_UNARY_OPCODE_X0 = 7,
> +  REVBITS_UNARY_OPCODE_Y0 = 7,
> +  REVBYTES_UNARY_OPCODE_X0 = 8,
> +  REVBYTES_UNARY_OPCODE_Y0 = 8,
> +  ROTLI_SHIFT_OPCODE_X0 = 1,
> +  ROTLI_SHIFT_OPCODE_X1 = 1,
> +  ROTLI_SHIFT_OPCODE_Y0 = 0,
> +  ROTLI_SHIFT_OPCODE_Y1 = 0,
> +  ROTL_RRR_0_OPCODE_X0 = 66,
> +  ROTL_RRR_0_OPCODE_X1 = 30,
> +  ROTL_RRR_6_OPCODE_Y0 = 0,
> +  ROTL_RRR_6_OPCODE_Y1 = 0,
> +  RRR_0_OPCODE_X0 = 5,
> +  RRR_0_OPCODE_X1 = 5,
> +  RRR_0_OPCODE_Y0 = 5,
> +  RRR_0_OPCODE_Y1 = 6,
> +  RRR_1_OPCODE_Y0 = 6,
> +  RRR_1_OPCODE_Y1 = 7,
> +  RRR_2_OPCODE_Y0 = 7,
> +  RRR_2_OPCODE_Y1 = 8,
> +  RRR_3_OPCODE_Y0 = 8,
> +  RRR_3_OPCODE_Y1 = 9,
> +  RRR_4_OPCODE_Y0 = 9,
> +  RRR_4_OPCODE_Y1 = 10,
> +  RRR_5_OPCODE_Y0 = 10,
> +  RRR_5_OPCODE_Y1 = 11,
> +  RRR_6_OPCODE_Y0 = 11,
> +  RRR_6_OPCODE_Y1 = 12,
> +  RRR_7_OPCODE_Y0 = 12,
> +  RRR_7_OPCODE_Y1 = 13,
> +  RRR_8_OPCODE_Y0 = 13,
> +  RRR_9_OPCODE_Y0 = 14,
> +  SHIFT_OPCODE_X0 = 6,
> +  SHIFT_OPCODE_X1 = 6,
> +  SHIFT_OPCODE_Y0 = 15,
> +  SHIFT_OPCODE_Y1 = 14,
> +  SHL16INSLI_OPCODE_X0 = 7,
> +  SHL16INSLI_OPCODE_X1 = 7,
> +  SHL1ADDX_RRR_0_OPCODE_X0 = 67,
> +  SHL1ADDX_RRR_0_OPCODE_X1 = 31,
> +  SHL1ADDX_RRR_7_OPCODE_Y0 = 1,
> +  SHL1ADDX_RRR_7_OPCODE_Y1 = 1,
> +  SHL1ADD_RRR_0_OPCODE_X0 = 68,
> +  SHL1ADD_RRR_0_OPCODE_X1 = 32,
> +  SHL1ADD_RRR_1_OPCODE_Y0 = 0,
> +  SHL1ADD_RRR_1_OPCODE_Y1 = 0,
> +  SHL2ADDX_RRR_0_OPCODE_X0 = 69,
> +  SHL2ADDX_RRR_0_OPCODE_X1 = 33,
> +  SHL2ADDX_RRR_7_OPCODE_Y0 = 2,
> +  SHL2ADDX_RRR_7_OPCODE_Y1 = 2,
> +  SHL2ADD_RRR_0_OPCODE_X0 = 70,
> +  SHL2ADD_RRR_0_OPCODE_X1 = 34,
> +  SHL2ADD_RRR_1_OPCODE_Y0 = 1,
> +  SHL2ADD_RRR_1_OPCODE_Y1 = 1,
> +  SHL3ADDX_RRR_0_OPCODE_X0 = 71,
> +  SHL3ADDX_RRR_0_OPCODE_X1 = 35,
> +  SHL3ADDX_RRR_7_OPCODE_Y0 = 3,
> +  SHL3ADDX_RRR_7_OPCODE_Y1 = 3,
> +  SHL3ADD_RRR_0_OPCODE_X0 = 72,
> +  SHL3ADD_RRR_0_OPCODE_X1 = 36,
> +  SHL3ADD_RRR_1_OPCODE_Y0 = 2,
> +  SHL3ADD_RRR_1_OPCODE_Y1 = 2,
> +  SHLI_SHIFT_OPCODE_X0 = 2,
> +  SHLI_SHIFT_OPCODE_X1 = 2,
> +  SHLI_SHIFT_OPCODE_Y0 = 1,
> +  SHLI_SHIFT_OPCODE_Y1 = 1,
> +  SHLXI_SHIFT_OPCODE_X0 = 3,
> +  SHLXI_SHIFT_OPCODE_X1 = 3,
> +  SHLX_RRR_0_OPCODE_X0 = 73,
> +  SHLX_RRR_0_OPCODE_X1 = 37,
> +  SHL_RRR_0_OPCODE_X0 = 74,
> +  SHL_RRR_0_OPCODE_X1 = 38,
> +  SHL_RRR_6_OPCODE_Y0 = 1,
> +  SHL_RRR_6_OPCODE_Y1 = 1,
> +  SHRSI_SHIFT_OPCODE_X0 = 4,
> +  SHRSI_SHIFT_OPCODE_X1 = 4,
> +  SHRSI_SHIFT_OPCODE_Y0 = 2,
> +  SHRSI_SHIFT_OPCODE_Y1 = 2,
> +  SHRS_RRR_0_OPCODE_X0 = 75,
> +  SHRS_RRR_0_OPCODE_X1 = 39,
> +  SHRS_RRR_6_OPCODE_Y0 = 2,
> +  SHRS_RRR_6_OPCODE_Y1 = 2,
> +  SHRUI_SHIFT_OPCODE_X0 = 5,
> +  SHRUI_SHIFT_OPCODE_X1 = 5,
> +  SHRUI_SHIFT_OPCODE_Y0 = 3,
> +  SHRUI_SHIFT_OPCODE_Y1 = 3,
> +  SHRUXI_SHIFT_OPCODE_X0 = 6,
> +  SHRUXI_SHIFT_OPCODE_X1 = 6,
> +  SHRUX_RRR_0_OPCODE_X0 = 76,
> +  SHRUX_RRR_0_OPCODE_X1 = 40,
> +  SHRU_RRR_0_OPCODE_X0 = 77,
> +  SHRU_RRR_0_OPCODE_X1 = 41,
> +  SHRU_RRR_6_OPCODE_Y0 = 3,
> +  SHRU_RRR_6_OPCODE_Y1 = 3,
> +  SHUFFLEBYTES_RRR_0_OPCODE_X0 = 78,
> +  ST1_ADD_IMM8_OPCODE_X1 = 25,
> +  ST1_OPCODE_Y2 = 0,
> +  ST1_RRR_0_OPCODE_X1 = 42,
> +  ST2_ADD_IMM8_OPCODE_X1 = 26,
> +  ST2_OPCODE_Y2 = 1,
> +  ST2_RRR_0_OPCODE_X1 = 43,
> +  ST4_ADD_IMM8_OPCODE_X1 = 27,
> +  ST4_OPCODE_Y2 = 2,
> +  ST4_RRR_0_OPCODE_X1 = 44,
> +  STNT1_ADD_IMM8_OPCODE_X1 = 28,
> +  STNT1_RRR_0_OPCODE_X1 = 45,
> +  STNT2_ADD_IMM8_OPCODE_X1 = 29,
> +  STNT2_RRR_0_OPCODE_X1 = 46,
> +  STNT4_ADD_IMM8_OPCODE_X1 = 30,
> +  STNT4_RRR_0_OPCODE_X1 = 47,
> +  STNT_ADD_IMM8_OPCODE_X1 = 31,
> +  STNT_RRR_0_OPCODE_X1 = 48,
> +  ST_ADD_IMM8_OPCODE_X1 = 32,
> +  ST_OPCODE_Y2 = 3,
> +  ST_RRR_0_OPCODE_X1 = 49,
> +  SUBXSC_RRR_0_OPCODE_X0 = 79,
> +  SUBXSC_RRR_0_OPCODE_X1 = 50,
> +  SUBX_RRR_0_OPCODE_X0 = 80,
> +  SUBX_RRR_0_OPCODE_X1 = 51,
> +  SUBX_RRR_0_OPCODE_Y0 = 2,
> +  SUBX_RRR_0_OPCODE_Y1 = 2,
> +  SUB_RRR_0_OPCODE_X0 = 81,
> +  SUB_RRR_0_OPCODE_X1 = 52,
> +  SUB_RRR_0_OPCODE_Y0 = 3,
> +  SUB_RRR_0_OPCODE_Y1 = 3,
> +  SWINT0_UNARY_OPCODE_X1 = 34,
> +  SWINT1_UNARY_OPCODE_X1 = 35,
> +  SWINT2_UNARY_OPCODE_X1 = 36,
> +  SWINT3_UNARY_OPCODE_X1 = 37,
> +  TBLIDXB0_UNARY_OPCODE_X0 = 9,
> +  TBLIDXB0_UNARY_OPCODE_Y0 = 9,
> +  TBLIDXB1_UNARY_OPCODE_X0 = 10,
> +  TBLIDXB1_UNARY_OPCODE_Y0 = 10,
> +  TBLIDXB2_UNARY_OPCODE_X0 = 11,
> +  TBLIDXB2_UNARY_OPCODE_Y0 = 11,
> +  TBLIDXB3_UNARY_OPCODE_X0 = 12,
> +  TBLIDXB3_UNARY_OPCODE_Y0 = 12,
> +  UNARY_RRR_0_OPCODE_X0 = 82,
> +  UNARY_RRR_0_OPCODE_X1 = 53,
> +  UNARY_RRR_1_OPCODE_Y0 = 3,
> +  UNARY_RRR_1_OPCODE_Y1 = 3,
> +  V1ADDI_IMM8_OPCODE_X0 = 8,
> +  V1ADDI_IMM8_OPCODE_X1 = 33,
> +  V1ADDUC_RRR_0_OPCODE_X0 = 83,
> +  V1ADDUC_RRR_0_OPCODE_X1 = 54,
> +  V1ADD_RRR_0_OPCODE_X0 = 84,
> +  V1ADD_RRR_0_OPCODE_X1 = 55,
> +  V1ADIFFU_RRR_0_OPCODE_X0 = 85,
> +  V1AVGU_RRR_0_OPCODE_X0 = 86,
> +  V1CMPEQI_IMM8_OPCODE_X0 = 9,
> +  V1CMPEQI_IMM8_OPCODE_X1 = 34,
> +  V1CMPEQ_RRR_0_OPCODE_X0 = 87,
> +  V1CMPEQ_RRR_0_OPCODE_X1 = 56,
> +  V1CMPLES_RRR_0_OPCODE_X0 = 88,
> +  V1CMPLES_RRR_0_OPCODE_X1 = 57,
> +  V1CMPLEU_RRR_0_OPCODE_X0 = 89,
> +  V1CMPLEU_RRR_0_OPCODE_X1 = 58,
> +  V1CMPLTSI_IMM8_OPCODE_X0 = 10,
> +  V1CMPLTSI_IMM8_OPCODE_X1 = 35,
> +  V1CMPLTS_RRR_0_OPCODE_X0 = 90,
> +  V1CMPLTS_RRR_0_OPCODE_X1 = 59,
> +  V1CMPLTUI_IMM8_OPCODE_X0 = 11,
> +  V1CMPLTUI_IMM8_OPCODE_X1 = 36,
> +  V1CMPLTU_RRR_0_OPCODE_X0 = 91,
> +  V1CMPLTU_RRR_0_OPCODE_X1 = 60,
> +  V1CMPNE_RRR_0_OPCODE_X0 = 92,
> +  V1CMPNE_RRR_0_OPCODE_X1 = 61,
> +  V1DDOTPUA_RRR_0_OPCODE_X0 = 161,
> +  V1DDOTPUSA_RRR_0_OPCODE_X0 = 93,
> +  V1DDOTPUS_RRR_0_OPCODE_X0 = 94,
> +  V1DDOTPU_RRR_0_OPCODE_X0 = 162,
> +  V1DOTPA_RRR_0_OPCODE_X0 = 95,
> +  V1DOTPUA_RRR_0_OPCODE_X0 = 163,
> +  V1DOTPUSA_RRR_0_OPCODE_X0 = 96,
> +  V1DOTPUS_RRR_0_OPCODE_X0 = 97,
> +  V1DOTPU_RRR_0_OPCODE_X0 = 164,
> +  V1DOTP_RRR_0_OPCODE_X0 = 98,
> +  V1INT_H_RRR_0_OPCODE_X0 = 99,
> +  V1INT_H_RRR_0_OPCODE_X1 = 62,
> +  V1INT_L_RRR_0_OPCODE_X0 = 100,
> +  V1INT_L_RRR_0_OPCODE_X1 = 63,
> +  V1MAXUI_IMM8_OPCODE_X0 = 12,
> +  V1MAXUI_IMM8_OPCODE_X1 = 37,
> +  V1MAXU_RRR_0_OPCODE_X0 = 101,
> +  V1MAXU_RRR_0_OPCODE_X1 = 64,
> +  V1MINUI_IMM8_OPCODE_X0 = 13,
> +  V1MINUI_IMM8_OPCODE_X1 = 38,
> +  V1MINU_RRR_0_OPCODE_X0 = 102,
> +  V1MINU_RRR_0_OPCODE_X1 = 65,
> +  V1MNZ_RRR_0_OPCODE_X0 = 103,
> +  V1MNZ_RRR_0_OPCODE_X1 = 66,
> +  V1MULTU_RRR_0_OPCODE_X0 = 104,
> +  V1MULUS_RRR_0_OPCODE_X0 = 105,
> +  V1MULU_RRR_0_OPCODE_X0 = 106,
> +  V1MZ_RRR_0_OPCODE_X0 = 107,
> +  V1MZ_RRR_0_OPCODE_X1 = 67,
> +  V1SADAU_RRR_0_OPCODE_X0 = 108,
> +  V1SADU_RRR_0_OPCODE_X0 = 109,
> +  V1SHLI_SHIFT_OPCODE_X0 = 7,
> +  V1SHLI_SHIFT_OPCODE_X1 = 7,
> +  V1SHL_RRR_0_OPCODE_X0 = 110,
> +  V1SHL_RRR_0_OPCODE_X1 = 68,
> +  V1SHRSI_SHIFT_OPCODE_X0 = 8,
> +  V1SHRSI_SHIFT_OPCODE_X1 = 8,
> +  V1SHRS_RRR_0_OPCODE_X0 = 111,
> +  V1SHRS_RRR_0_OPCODE_X1 = 69,
> +  V1SHRUI_SHIFT_OPCODE_X0 = 9,
> +  V1SHRUI_SHIFT_OPCODE_X1 = 9,
> +  V1SHRU_RRR_0_OPCODE_X0 = 112,
> +  V1SHRU_RRR_0_OPCODE_X1 = 70,
> +  V1SUBUC_RRR_0_OPCODE_X0 = 113,
> +  V1SUBUC_RRR_0_OPCODE_X1 = 71,
> +  V1SUB_RRR_0_OPCODE_X0 = 114,
> +  V1SUB_RRR_0_OPCODE_X1 = 72,
> +  V2ADDI_IMM8_OPCODE_X0 = 14,
> +  V2ADDI_IMM8_OPCODE_X1 = 39,
> +  V2ADDSC_RRR_0_OPCODE_X0 = 115,
> +  V2ADDSC_RRR_0_OPCODE_X1 = 73,
> +  V2ADD_RRR_0_OPCODE_X0 = 116,
> +  V2ADD_RRR_0_OPCODE_X1 = 74,
> +  V2ADIFFS_RRR_0_OPCODE_X0 = 117,
> +  V2AVGS_RRR_0_OPCODE_X0 = 118,
> +  V2CMPEQI_IMM8_OPCODE_X0 = 15,
> +  V2CMPEQI_IMM8_OPCODE_X1 = 40,
> +  V2CMPEQ_RRR_0_OPCODE_X0 = 119,
> +  V2CMPEQ_RRR_0_OPCODE_X1 = 75,
> +  V2CMPLES_RRR_0_OPCODE_X0 = 120,
> +  V2CMPLES_RRR_0_OPCODE_X1 = 76,
> +  V2CMPLEU_RRR_0_OPCODE_X0 = 121,
> +  V2CMPLEU_RRR_0_OPCODE_X1 = 77,
> +  V2CMPLTSI_IMM8_OPCODE_X0 = 16,
> +  V2CMPLTSI_IMM8_OPCODE_X1 = 41,
> +  V2CMPLTS_RRR_0_OPCODE_X0 = 122,
> +  V2CMPLTS_RRR_0_OPCODE_X1 = 78,
> +  V2CMPLTUI_IMM8_OPCODE_X0 = 17,
> +  V2CMPLTUI_IMM8_OPCODE_X1 = 42,
> +  V2CMPLTU_RRR_0_OPCODE_X0 = 123,
> +  V2CMPLTU_RRR_0_OPCODE_X1 = 79,
> +  V2CMPNE_RRR_0_OPCODE_X0 = 124,
> +  V2CMPNE_RRR_0_OPCODE_X1 = 80,
> +  V2DOTPA_RRR_0_OPCODE_X0 = 125,
> +  V2DOTP_RRR_0_OPCODE_X0 = 126,
> +  V2INT_H_RRR_0_OPCODE_X0 = 127,
> +  V2INT_H_RRR_0_OPCODE_X1 = 81,
> +  V2INT_L_RRR_0_OPCODE_X0 = 128,
> +  V2INT_L_RRR_0_OPCODE_X1 = 82,
> +  V2MAXSI_IMM8_OPCODE_X0 = 18,
> +  V2MAXSI_IMM8_OPCODE_X1 = 43,
> +  V2MAXS_RRR_0_OPCODE_X0 = 129,
> +  V2MAXS_RRR_0_OPCODE_X1 = 83,
> +  V2MINSI_IMM8_OPCODE_X0 = 19,
> +  V2MINSI_IMM8_OPCODE_X1 = 44,
> +  V2MINS_RRR_0_OPCODE_X0 = 130,
> +  V2MINS_RRR_0_OPCODE_X1 = 84,
> +  V2MNZ_RRR_0_OPCODE_X0 = 131,
> +  V2MNZ_RRR_0_OPCODE_X1 = 85,
> +  V2MULFSC_RRR_0_OPCODE_X0 = 132,
> +  V2MULS_RRR_0_OPCODE_X0 = 133,
> +  V2MULTS_RRR_0_OPCODE_X0 = 134,
> +  V2MZ_RRR_0_OPCODE_X0 = 135,
> +  V2MZ_RRR_0_OPCODE_X1 = 86,
> +  V2PACKH_RRR_0_OPCODE_X0 = 136,
> +  V2PACKH_RRR_0_OPCODE_X1 = 87,
> +  V2PACKL_RRR_0_OPCODE_X0 = 137,
> +  V2PACKL_RRR_0_OPCODE_X1 = 88,
> +  V2PACKUC_RRR_0_OPCODE_X0 = 138,
> +  V2PACKUC_RRR_0_OPCODE_X1 = 89,
> +  V2SADAS_RRR_0_OPCODE_X0 = 139,
> +  V2SADAU_RRR_0_OPCODE_X0 = 140,
> +  V2SADS_RRR_0_OPCODE_X0 = 141,
> +  V2SADU_RRR_0_OPCODE_X0 = 142,
> +  V2SHLI_SHIFT_OPCODE_X0 = 10,
> +  V2SHLI_SHIFT_OPCODE_X1 = 10,
> +  V2SHLSC_RRR_0_OPCODE_X0 = 143,
> +  V2SHLSC_RRR_0_OPCODE_X1 = 90,
> +  V2SHL_RRR_0_OPCODE_X0 = 144,
> +  V2SHL_RRR_0_OPCODE_X1 = 91,
> +  V2SHRSI_SHIFT_OPCODE_X0 = 11,
> +  V2SHRSI_SHIFT_OPCODE_X1 = 11,
> +  V2SHRS_RRR_0_OPCODE_X0 = 145,
> +  V2SHRS_RRR_0_OPCODE_X1 = 92,
> +  V2SHRUI_SHIFT_OPCODE_X0 = 12,
> +  V2SHRUI_SHIFT_OPCODE_X1 = 12,
> +  V2SHRU_RRR_0_OPCODE_X0 = 146,
> +  V2SHRU_RRR_0_OPCODE_X1 = 93,
> +  V2SUBSC_RRR_0_OPCODE_X0 = 147,
> +  V2SUBSC_RRR_0_OPCODE_X1 = 94,
> +  V2SUB_RRR_0_OPCODE_X0 = 148,
> +  V2SUB_RRR_0_OPCODE_X1 = 95,
> +  V4ADDSC_RRR_0_OPCODE_X0 = 149,
> +  V4ADDSC_RRR_0_OPCODE_X1 = 96,
> +  V4ADD_RRR_0_OPCODE_X0 = 150,
> +  V4ADD_RRR_0_OPCODE_X1 = 97,
> +  V4INT_H_RRR_0_OPCODE_X0 = 151,
> +  V4INT_H_RRR_0_OPCODE_X1 = 98,
> +  V4INT_L_RRR_0_OPCODE_X0 = 152,
> +  V4INT_L_RRR_0_OPCODE_X1 = 99,
> +  V4PACKSC_RRR_0_OPCODE_X0 = 153,
> +  V4PACKSC_RRR_0_OPCODE_X1 = 100,
> +  V4SHLSC_RRR_0_OPCODE_X0 = 154,
> +  V4SHLSC_RRR_0_OPCODE_X1 = 101,
> +  V4SHL_RRR_0_OPCODE_X0 = 155,
> +  V4SHL_RRR_0_OPCODE_X1 = 102,
> +  V4SHRS_RRR_0_OPCODE_X0 = 156,
> +  V4SHRS_RRR_0_OPCODE_X1 = 103,
> +  V4SHRU_RRR_0_OPCODE_X0 = 157,
> +  V4SHRU_RRR_0_OPCODE_X1 = 104,
> +  V4SUBSC_RRR_0_OPCODE_X0 = 158,
> +  V4SUBSC_RRR_0_OPCODE_X1 = 105,
> +  V4SUB_RRR_0_OPCODE_X0 = 159,
> +  V4SUB_RRR_0_OPCODE_X1 = 106,
> +  WH64_UNARY_OPCODE_X1 = 38,
> +  XORI_IMM8_OPCODE_X0 = 20,
> +  XORI_IMM8_OPCODE_X1 = 45,
> +  XOR_RRR_0_OPCODE_X0 = 160,
> +  XOR_RRR_0_OPCODE_X1 = 107,
> +  XOR_RRR_5_OPCODE_Y0 = 3,
> +  XOR_RRR_5_OPCODE_Y1 = 3
> +};
> +
> +
> +#endif /* __ASSEMBLER__ */
> +
> +#endif /* __ARCH_OPCODE_H__ */
> diff --git a/target-tilegx/translate.c b/target-tilegx/translate.c
> index 9aa82a9..1fd16e4 100644
> --- a/target-tilegx/translate.c
> +++ b/target-tilegx/translate.c
> @@ -23,19 +23,615 @@
>  #include "disas/disas.h"
>  #include "tcg-op.h"
>  #include "exec/cpu_ldst.h"
> +#include "opcode_tilegx.h"
> +
> +#define TILEGX_BUNDLE_TYPE_MASK       0xc000000000000000ULL
> +#define TILEGX_BUNDLE_TYPE_Y(bundle)  ((bundle) & TILEGX_BUNDLE_TYPE_MASK)
> +#define TILEGX_BUNDLE_TYPE_X(bundle)  (!TILEGX_BUNDLE_TYPE_Y(bundle))
> +#define TILEGX_OPCODE_MAX_X0          164  /* include 164 */
> +#define TILEGX_OPCODE_MAX_X1          107  /* include 107 */
> +#define TILEGX_OPCODE_MAX_Y0           15  /* include 15 */
> +#define TILEGX_OPCODE_MAX_Y1           15  /* include 15 */
> +#define TILEGX_OPCODE_MAX_Y2            3  /* include 3 */
> +
> +#define TILEGX_EXCP_OPCODE_UNKNOWN    0x1
> +#define TILEGX_EXCP_OPCODE_UNIMPLEMNT 0x2
> +
> +
> +static TCGv_ptr cpu_env;
> +static TCGv_i64 cpu_pc;
> +static TCGv_i64 cpu_regs[TILEGX_R_COUNT];
> +
> +static const char *reg_names[] = {
> +     "r0",  "r1",  "r2",  "r3",  "r4",  "r5",  "r6",  "r7",
> +     "r8",  "r9", "r10", "r11", "r12", "r13", "r14", "r15",
> +    "r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23",
> +    "r24", "r25", "r26", "r27", "r28", "r29", "r30", "r31",
> +    "r32", "r33", "r34", "r35", "r36", "r37", "r38", "r39",
> +    "r40", "r41", "r42", "r43", "r44", "r45", "r46", "r47",
> +    "r48", "r49", "r50", "r51",  "bp",  "tp",  "sp",  "lr"
> +};
> +
> +/* This is the state at translation time.  */
> +typedef struct DisasContext {
> +    uint64_t pc;                   /* Current pc */
> +    uint64_t exception;            /* Current exception, 0 means empty */
> +
> +    TCGv zero;                     /* For zero register */
> +
> +    struct {
> +        unsigned char idx;         /* index */
> +        TCGv val;                  /* value */
> +    } *tmp_regcur,                 /* Current temporary registers */
> +    tmp_regs[TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE]; /* All temporary registers */
> +
> +    struct {
> +        TCGCond cond;              /* Branch condition */
> +        TCGv dest;                 /* pc jump destination, if will jump */
> +        TCGv val1;                 /* Firt value for condition comparing */
> +        TCGv val2;                 /* Second value for condition comparing */
> +    } jmp;                         /* Jump object, only once in each TB block */
> +} DisasContext;
> +
> +#include "exec/gen-icount.h"
> +
> +typedef void (*TileGXDecodeInsn)(struct DisasContext *, tilegx_bundle_bits);
> +
> +/* by "grep _OPCODE_X0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
> +static TileGXDecodeInsn decode_insns_x0[] = {
> +    NULL, /* Nothing */
> +    NULL, /* 1, ADDI_IMM8_OPCODE_X0 1, ADDLI_OPCODE_X0 1, ADDXSC_RRR_0_OPCODE_X0 1, CNTLZ_UNARY_OPCODE_X0 1, ROTLI_SHIFT_OPCODE_X0 */
> +    NULL, /* 2, ADDXI_IMM8_OPCODE_X0 2, ADDXLI_OPCODE_X0 2, ADDX_RRR_0_OPCODE_X0 2, CNTTZ_UNARY_OPCODE_X0 2, SHLI_SHIFT_OPCODE_X0 */
> +    NULL, /* 3, ADD_RRR_0_OPCODE_X0 3, ANDI_IMM8_OPCODE_X0 3, BF_OPCODE_X0 3, FNOP_UNARY_OPCODE_X0 3, SHLXI_SHIFT_OPCODE_X0 */
> +    NULL, /* 4, AND_RRR_0_OPCODE_X0 4, BFEXTS_BF_OPCODE_X0 4, CMPEQI_IMM8_OPCODE_X0 4, FSINGLE_PACK1_UNARY_OPCODE_X0 4, IMM8_OPCODE_X0 4, SHRSI_SHIFT_OPCODE_X0 */
> +    NULL, /* 5, BFEXTU_BF_OPCODE_X0 5, CMOVEQZ_RRR_0_OPCODE_X0 5, CMPLTSI_IMM8_OPCODE_X0 5, NOP_UNARY_OPCODE_X0 5, RRR_0_OPCODE_X0 5, SHRUI_SHIFT_OPCODE_X0 */
> +    NULL, /* 6, BFINS_BF_OPCODE_X0 6, CMOVNEZ_RRR_0_OPCODE_X0 6, CMPLTUI_IMM8_OPCODE_X0 6, PCNT_UNARY_OPCODE_X0 6, SHIFT_OPCODE_X0 6, SHRUXI_SHIFT_OPCODE_X0 */
> +    NULL, /* 7, CMPEQ_RRR_0_OPCODE_X0 7, MM_BF_OPCODE_X0 7, ORI_IMM8_OPCODE_X0 7, REVBITS_UNARY_OPCODE_X0 7, SHL16INSLI_OPCODE_X0 7, V1SHLI_SHIFT_OPCODE_X0 */
> +    NULL, /* 8, CMPLES_RRR_0_OPCODE_X0 8, REVBYTES_UNARY_OPCODE_X0 8, V1ADDI_IMM8_OPCODE_X0 8, V1SHRSI_SHIFT_OPCODE_X0 */
> +    NULL, /* 9, CMPLEU_RRR_0_OPCODE_X0 9, TBLIDXB0_UNARY_OPCODE_X0 9, V1CMPEQI_IMM8_OPCODE_X0 9, V1SHRUI_SHIFT_OPCODE_X0 */
> +    NULL, /* 10, CMPLTS_RRR_0_OPCODE_X0 10, TBLIDXB1_UNARY_OPCODE_X0 10, V1CMPLTSI_IMM8_OPCODE_X0 10, V2SHLI_SHIFT_OPCODE_X0 */
> +    NULL, /* 11, CMPLTU_RRR_0_OPCODE_X0 11, TBLIDXB2_UNARY_OPCODE_X0 11, V1CMPLTUI_IMM8_OPCODE_X0 11, V2SHRSI_SHIFT_OPCODE_X0 */
> +    NULL, /* 12, CMPNE_RRR_0_OPCODE_X0 12, TBLIDXB3_UNARY_OPCODE_X0 12, V1MAXUI_IMM8_OPCODE_X0 12, V2SHRUI_SHIFT_OPCODE_X0 */
> +    NULL, /* 13, CMULAF_RRR_0_OPCODE_X0 13, V1MINUI_IMM8_OPCODE_X0 */
> +    NULL, /* 14, CMULA_RRR_0_OPCODE_X0 14, V2ADDI_IMM8_OPCODE_X0 */
> +    NULL, /* 15, CMULFR_RRR_0_OPCODE_X0 15, V2CMPEQI_IMM8_OPCODE_X0 */
> +    NULL, /* 16, CMULF_RRR_0_OPCODE_X0 16, V2CMPLTSI_IMM8_OPCODE_X0 */
> +    NULL, /* 17, CMULHR_RRR_0_OPCODE_X0 17, V2CMPLTUI_IMM8_OPCODE_X0 */
> +    NULL, /* 18, CMULH_RRR_0_OPCODE_X0 18, V2MAXSI_IMM8_OPCODE_X0 */
> +    NULL, /* 19, CMUL_RRR_0_OPCODE_X0 19, V2MINSI_IMM8_OPCODE_X0 */
> +    NULL, /* 20, CRC32_32_RRR_0_OPCODE_X0 20, XORI_IMM8_OPCODE_X0 */
> +    NULL, /* 21, CRC32_8_RRR_0_OPCODE_X0 */
> +    NULL, /* 22, DBLALIGN2_RRR_0_OPCODE_X0 */
> +    NULL, /* 23, DBLALIGN4_RRR_0_OPCODE_X0 */
> +    NULL, /* 24, DBLALIGN6_RRR_0_OPCODE_X0 */
> +    NULL, /* 25, DBLALIGN_RRR_0_OPCODE_X0 */
> +    NULL, /* 26, FDOUBLE_ADDSUB_RRR_0_OPCODE_X0 */
> +    NULL, /* 27, FDOUBLE_ADD_FLAGS_RRR_0_OPCODE_X0 */
> +    NULL, /* 28, FDOUBLE_MUL_FLAGS_RRR_0_OPCODE_X0 */
> +    NULL, /* 29, FDOUBLE_PACK1_RRR_0_OPCODE_X0 */
> +    NULL, /* 30, FDOUBLE_PACK2_RRR_0_OPCODE_X0 */
> +    NULL, /* 31, FDOUBLE_SUB_FLAGS_RRR_0_OPCODE_X0 */
> +    NULL, /* 32, FDOUBLE_UNPACK_MAX_RRR_0_OPCODE_X0 */
> +    NULL, /* 33, FDOUBLE_UNPACK_MIN_RRR_0_OPCODE_X0 */
> +    NULL, /* 34, FSINGLE_ADD1_RRR_0_OPCODE_X0 */
> +    NULL, /* 35, FSINGLE_ADDSUB2_RRR_0_OPCODE_X0 */
> +    NULL, /* 36, FSINGLE_MUL1_RRR_0_OPCODE_X0 */
> +    NULL, /* 37, FSINGLE_MUL2_RRR_0_OPCODE_X0 */
> +    NULL, /* 38, FSINGLE_PACK2_RRR_0_OPCODE_X0 */
> +    NULL, /* 39, FSINGLE_SUB1_RRR_0_OPCODE_X0 */
> +    NULL, /* 40, MNZ_RRR_0_OPCODE_X0 */
> +    NULL, /* 41, MULAX_RRR_0_OPCODE_X0 */
> +    NULL, /* 42, MULA_HS_HS_RRR_0_OPCODE_X0 */
> +    NULL, /* 43, MULA_HS_HU_RRR_0_OPCODE_X0 */
> +    NULL, /* 44, MULA_HS_LS_RRR_0_OPCODE_X0 */
> +    NULL, /* 45, MULA_HS_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 46, MULA_HU_HU_RRR_0_OPCODE_X0 */
> +    NULL, /* 47, MULA_HU_LS_RRR_0_OPCODE_X0 */
> +    NULL, /* 48, MULA_HU_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 49, MULA_LS_LS_RRR_0_OPCODE_X0 */
> +    NULL, /* 50, MULA_LS_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 51, MULA_LU_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 52, MULX_RRR_0_OPCODE_X0 */
> +    NULL, /* 53, MUL_HS_HS_RRR_0_OPCODE_X0 */
> +    NULL, /* 54, MUL_HS_HU_RRR_0_OPCODE_X0 */
> +    NULL, /* 55, MUL_HS_LS_RRR_0_OPCODE_X0 */
> +    NULL, /* 56, MUL_HS_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 57, MUL_HU_HU_RRR_0_OPCODE_X0 */
> +    NULL, /* 58, MUL_HU_LS_RRR_0_OPCODE_X0 */
> +    NULL, /* 59, MUL_HU_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 60, MUL_LS_LS_RRR_0_OPCODE_X0 */
> +    NULL, /* 61, MUL_LS_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 62, MUL_LU_LU_RRR_0_OPCODE_X0 */
> +    NULL, /* 63, MZ_RRR_0_OPCODE_X0 */
> +    NULL, /* 64, NOR_RRR_0_OPCODE_X0 */
> +    NULL, /* 65, OR_RRR_0_OPCODE_X0 */
> +    NULL, /* 66, ROTL_RRR_0_OPCODE_X0 */
> +    NULL, /* 67, SHL1ADDX_RRR_0_OPCODE_X0 */
> +    NULL, /* 68, SHL1ADD_RRR_0_OPCODE_X0 */
> +    NULL, /* 69, SHL2ADDX_RRR_0_OPCODE_X0 */
> +    NULL, /* 70, SHL2ADD_RRR_0_OPCODE_X0 */
> +    NULL, /* 71, SHL3ADDX_RRR_0_OPCODE_X0 */
> +    NULL, /* 72, SHL3ADD_RRR_0_OPCODE_X0 */
> +    NULL, /* 73, SHLX_RRR_0_OPCODE_X0 */
> +    NULL, /* 74, SHL_RRR_0_OPCODE_X0 */
> +    NULL, /* 75, SHRS_RRR_0_OPCODE_X0 */
> +    NULL, /* 76, SHRUX_RRR_0_OPCODE_X0 */
> +    NULL, /* 77, SHRU_RRR_0_OPCODE_X0 */
> +    NULL, /* 78, SHUFFLEBYTES_RRR_0_OPCODE_X0 */
> +    NULL, /* 79, SUBXSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 80, SUBX_RRR_0_OPCODE_X0 */
> +    NULL, /* 81, SUB_RRR_0_OPCODE_X0 */
> +    NULL, /* 82, UNARY_RRR_0_OPCODE_X0 */
> +    NULL, /* 83, V1ADDUC_RRR_0_OPCODE_X0 */
> +    NULL, /* 84, V1ADD_RRR_0_OPCODE_X0 */
> +    NULL, /* 85, V1ADIFFU_RRR_0_OPCODE_X0 */
> +    NULL, /* 86, V1AVGU_RRR_0_OPCODE_X0 */
> +    NULL, /* 87, V1CMPEQ_RRR_0_OPCODE_X0 */
> +    NULL, /* 88, V1CMPLES_RRR_0_OPCODE_X0 */
> +    NULL, /* 89, V1CMPLEU_RRR_0_OPCODE_X0 */
> +    NULL, /* 90, V1CMPLTS_RRR_0_OPCODE_X0 */
> +    NULL, /* 91, V1CMPLTU_RRR_0_OPCODE_X0 */
> +    NULL, /* 92, V1CMPNE_RRR_0_OPCODE_X0 */
> +    NULL, /* 93, V1DDOTPUSA_RRR_0_OPCODE_X0 */
> +    NULL, /* 94, V1DDOTPUS_RRR_0_OPCODE_X0 */
> +    NULL, /* 95, V1DOTPA_RRR_0_OPCODE_X0 */
> +    NULL, /* 96, V1DOTPUSA_RRR_0_OPCODE_X0 */
> +    NULL, /* 97, V1DOTPUS_RRR_0_OPCODE_X0 */
> +    NULL, /* 98, V1DOTP_RRR_0_OPCODE_X0 */
> +    NULL, /* 99, V1INT_H_RRR_0_OPCODE_X0 */
> +    NULL, /* 100, V1INT_L_RRR_0_OPCODE_X0 */
> +    NULL, /* 101, V1MAXU_RRR_0_OPCODE_X0 */
> +    NULL, /* 102, V1MINU_RRR_0_OPCODE_X0 */
> +    NULL, /* 103, V1MNZ_RRR_0_OPCODE_X0 */
> +    NULL, /* 104, V1MULTU_RRR_0_OPCODE_X0 */
> +    NULL, /* 105, V1MULUS_RRR_0_OPCODE_X0 */
> +    NULL, /* 106, V1MULU_RRR_0_OPCODE_X0 */
> +    NULL, /* 107, V1MZ_RRR_0_OPCODE_X0 */
> +    NULL, /* 108, V1SADAU_RRR_0_OPCODE_X0 */
> +    NULL, /* 109, V1SADU_RRR_0_OPCODE_X0 */
> +    NULL, /* 110, V1SHL_RRR_0_OPCODE_X0 */
> +    NULL, /* 111, V1SHRS_RRR_0_OPCODE_X0 */
> +    NULL, /* 112, V1SHRU_RRR_0_OPCODE_X0 */
> +    NULL, /* 113, V1SUBUC_RRR_0_OPCODE_X0 */
> +    NULL, /* 114, V1SUB_RRR_0_OPCODE_X0 */
> +    NULL, /* 115, V2ADDSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 116, V2ADD_RRR_0_OPCODE_X0 */
> +    NULL, /* 117, V2ADIFFS_RRR_0_OPCODE_X0 */
> +    NULL, /* 118, V2AVGS_RRR_0_OPCODE_X0 */
> +    NULL, /* 119, V2CMPEQ_RRR_0_OPCODE_X0 */
> +    NULL, /* 120, V2CMPLES_RRR_0_OPCODE_X0 */
> +    NULL, /* 121, V2CMPLEU_RRR_0_OPCODE_X0 */
> +    NULL, /* 122, V2CMPLTS_RRR_0_OPCODE_X0 */
> +    NULL, /* 123, V2CMPLTU_RRR_0_OPCODE_X0 */
> +    NULL, /* 124, V2CMPNE_RRR_0_OPCODE_X0 */
> +    NULL, /* 125, V2DOTPA_RRR_0_OPCODE_X0 */
> +    NULL, /* 126, V2DOTP_RRR_0_OPCODE_X0 */
> +    NULL, /* 127, V2INT_H_RRR_0_OPCODE_X0 */
> +    NULL, /* 128, V2INT_L_RRR_0_OPCODE_X0 */
> +    NULL, /* 129, V2MAXS_RRR_0_OPCODE_X0 */
> +    NULL, /* 130, V2MINS_RRR_0_OPCODE_X0 */
> +    NULL, /* 131, V2MNZ_RRR_0_OPCODE_X0 */
> +    NULL, /* 132, V2MULFSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 133, V2MULS_RRR_0_OPCODE_X0 */
> +    NULL, /* 134, V2MULTS_RRR_0_OPCODE_X0 */
> +    NULL, /* 135, V2MZ_RRR_0_OPCODE_X0 */
> +    NULL, /* 136, V2PACKH_RRR_0_OPCODE_X0 */
> +    NULL, /* 137, V2PACKL_RRR_0_OPCODE_X0 */
> +    NULL, /* 138, V2PACKUC_RRR_0_OPCODE_X0 */
> +    NULL, /* 139, V2SADAS_RRR_0_OPCODE_X0 */
> +    NULL, /* 140, V2SADAU_RRR_0_OPCODE_X0 */
> +    NULL, /* 141, V2SADS_RRR_0_OPCODE_X0 */
> +    NULL, /* 142, V2SADU_RRR_0_OPCODE_X0 */
> +    NULL, /* 143, V2SHLSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 144, V2SHL_RRR_0_OPCODE_X0 */
> +    NULL, /* 145, V2SHRS_RRR_0_OPCODE_X0 */
> +    NULL, /* 146, V2SHRU_RRR_0_OPCODE_X0 */
> +    NULL, /* 147, V2SUBSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 148, V2SUB_RRR_0_OPCODE_X0 */
> +    NULL, /* 149, V4ADDSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 150, V4ADD_RRR_0_OPCODE_X0 */
> +    NULL, /* 151, V4INT_H_RRR_0_OPCODE_X0 */
> +    NULL, /* 152, V4INT_L_RRR_0_OPCODE_X0 */
> +    NULL, /* 153, V4PACKSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 154, V4SHLSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 155, V4SHL_RRR_0_OPCODE_X0 */
> +    NULL, /* 156, V4SHRS_RRR_0_OPCODE_X0 */
> +    NULL, /* 157, V4SHRU_RRR_0_OPCODE_X0 */
> +    NULL, /* 158, V4SUBSC_RRR_0_OPCODE_X0 */
> +    NULL, /* 159, V4SUB_RRR_0_OPCODE_X0 */
> +    NULL, /* 160, XOR_RRR_0_OPCODE_X0 */
> +    NULL, /* 161, V1DDOTPUA_RRR_0_OPCODE_X0 */
> +    NULL, /* 162, V1DDOTPU_RRR_0_OPCODE_X0 */
> +    NULL, /* 163, V1DOTPUA_RRR_0_OPCODE_X0 */
> +    NULL  /* 164, V1DOTPU_RRR_0_OPCODE_X0 */
> +};
> +
> +/* by "grep _OPCODE_X1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
> +static TileGXDecodeInsn decode_insns_x1[] = {
> +    NULL, /* 0, ADDLI_OPCODE_X1 0, JAL_JUMP_OPCODE_X1 */
> +    NULL, /* 1, ADDI_IMM8_OPCODE_X1 1, ADDXLI_OPCODE_X1 1, ADDXSC_RRR_0_OPCODE_X1 1, DRAIN_UNARY_OPCODE_X1 1, J_JUMP_OPCODE_X1 1, ROTLI_SHIFT_OPCODE_X1 */
> +    NULL, /* 2, ADDXI_IMM8_OPCODE_X1 2, ADDX_RRR_0_OPCODE_X1 2, BRANCH_OPCODE_X1 2, DTLBPR_UNARY_OPCODE_X1 2, SHLI_SHIFT_OPCODE_X1 */
> +    NULL, /* 3, ADD_RRR_0_OPCODE_X1 3, ANDI_IMM8_OPCODE_X1 3, FINV_UNARY_OPCODE_X1 3, IMM8_OPCODE_X1 3, SHLXI_SHIFT_OPCODE_X1 */
> +    NULL, /* 4, AND_RRR_0_OPCODE_X1 4, CMPEQI_IMM8_OPCODE_X1 4, FLUSHWB_UNARY_OPCODE_X1 4, JUMP_OPCODE_X1 4, SHRSI_SHIFT_OPCODE_X1 */
> +    NULL, /* 5, CMPEQ_RRR_0_OPCODE_X1 5, CMPLTSI_IMM8_OPCODE_X1 5, FLUSH_UNARY_OPCODE_X1 5, RRR_0_OPCODE_X1 5, SHRUI_SHIFT_OPCODE_X1 */
> +    NULL, /* 6, CMPEXCH4_RRR_0_OPCODE_X1 6, CMPLTUI_IMM8_OPCODE_X1 6, FNOP_UNARY_OPCODE_X1 6, SHIFT_OPCODE_X1 6, SHRUXI_SHIFT_OPCODE_X1 */
> +    NULL, /* 7, CMPEXCH_RRR_0_OPCODE_X1 7, ICOH_UNARY_OPCODE_X1 7, LD1S_ADD_IMM8_OPCODE_X1 7, SHL16INSLI_OPCODE_X1 7, V1SHLI_SHIFT_OPCODE_X1 */
> +    NULL, /* 8, CMPLES_RRR_0_OPCODE_X1 8, ILL_UNARY_OPCODE_X1 8, LD1U_ADD_IMM8_OPCODE_X1 8, V1SHRSI_SHIFT_OPCODE_X1 */
> +    NULL, /* 9, CMPLEU_RRR_0_OPCODE_X1 9, INV_UNARY_OPCODE_X1 9, LD2S_ADD_IMM8_OPCODE_X1 9, V1SHRUI_SHIFT_OPCODE_X1 */
> +    NULL, /* 10, CMPLTS_RRR_0_OPCODE_X1 10, IRET_UNARY_OPCODE_X1 10, LD2U_ADD_IMM8_OPCODE_X1 10, V2SHLI_SHIFT_OPCODE_X1 */
> +    NULL, /* 11, CMPLTU_RRR_0_OPCODE_X1 11, JALRP_UNARY_OPCODE_X1 11, LD4S_ADD_IMM8_OPCODE_X1 11, V2SHRSI_SHIFT_OPCODE_X1 */
> +    NULL, /* 12, CMPNE_RRR_0_OPCODE_X1 12, JALR_UNARY_OPCODE_X1 12, LD4U_ADD_IMM8_OPCODE_X1 12, V2SHRUI_SHIFT_OPCODE_X1 */
> +    NULL, /* 13, DBLALIGN2_RRR_0_OPCODE_X1 13, JRP_UNARY_OPCODE_X1 13, LDNT1S_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 14, DBLALIGN4_RRR_0_OPCODE_X1 14, JR_UNARY_OPCODE_X1 14, LDNT1U_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 15, DBLALIGN6_RRR_0_OPCODE_X1 15, LD1S_UNARY_OPCODE_X1 15, LDNT2S_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 16, BEQZT_BRANCH_OPCODE_X1 16, EXCH4_RRR_0_OPCODE_X1 16, LD1U_UNARY_OPCODE_X1 16, LDNT2U_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 17, BEQZ_BRANCH_OPCODE_X1 17, EXCH_RRR_0_OPCODE_X1 17, LD2S_UNARY_OPCODE_X1 17, LDNT4S_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 18, BGEZT_BRANCH_OPCODE_X1 18, FETCHADD4_RRR_0_OPCODE_X1 18, LD2U_UNARY_OPCODE_X1 18, LDNT4U_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 19, BGEZ_BRANCH_OPCODE_X1 19, FETCHADDGEZ4_RRR_0_OPCODE_X1 19, LD4S_UNARY_OPCODE_X1 19, LDNT_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 20, BGTZT_BRANCH_OPCODE_X1 20, FETCHADDGEZ_RRR_0_OPCODE_X1 20, LD4U_UNARY_OPCODE_X1 20, LD_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 21, BGTZ_BRANCH_OPCODE_X1 21, FETCHADD_RRR_0_OPCODE_X1 21, LDNA_UNARY_OPCODE_X1 21, LWNA_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 22, BLBCT_BRANCH_OPCODE_X1 22, FETCHAND4_RRR_0_OPCODE_X1 22, LDNT1S_UNARY_OPCODE_X1 22, MFSPR_IMM8_OPCODE_X1 */
> +    NULL, /* 23, BLBC_BRANCH_OPCODE_X1 23, FETCHAND_RRR_0_OPCODE_X1 23, LDNT1U_UNARY_OPCODE_X1 23, MTSPR_IMM8_OPCODE_X1 */
> +    NULL, /* 24, BLBST_BRANCH_OPCODE_X1 24, FETCHOR4_RRR_0_OPCODE_X1 24, LDNT2S_UNARY_OPCODE_X1 24, ORI_IMM8_OPCODE_X1 */
> +    NULL, /* 25, BLBS_BRANCH_OPCODE_X1 25, FETCHOR_RRR_0_OPCODE_X1 25, LDNT2U_UNARY_OPCODE_X1 25, ST1_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 26, BLEZT_BRANCH_OPCODE_X1 26, LDNT4S_UNARY_OPCODE_X1 26, MNZ_RRR_0_OPCODE_X1 26, ST2_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 27, BLEZ_BRANCH_OPCODE_X1 27, LDNT4U_UNARY_OPCODE_X1 27, MZ_RRR_0_OPCODE_X1 27, ST4_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 28, BLTZT_BRANCH_OPCODE_X1 28, LDNT_UNARY_OPCODE_X1 28, NOR_RRR_0_OPCODE_X1 28, STNT1_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 29, BLTZ_BRANCH_OPCODE_X1 29, LD_UNARY_OPCODE_X1 29, OR_RRR_0_OPCODE_X1 29, STNT2_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 30, BNEZT_BRANCH_OPCODE_X1 30, LNK_UNARY_OPCODE_X1 30, ROTL_RRR_0_OPCODE_X1 30, STNT4_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 31, BNEZ_BRANCH_OPCODE_X1 31, MF_UNARY_OPCODE_X1 31, SHL1ADDX_RRR_0_OPCODE_X1 31, STNT_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 32, NAP_UNARY_OPCODE_X1 32, SHL1ADD_RRR_0_OPCODE_X1 32, ST_ADD_IMM8_OPCODE_X1 */
> +    NULL, /* 33, NOP_UNARY_OPCODE_X1 33, SHL2ADDX_RRR_0_OPCODE_X1 33, V1ADDI_IMM8_OPCODE_X1 */
> +    NULL, /* 34, SHL2ADD_RRR_0_OPCODE_X1 34, SWINT0_UNARY_OPCODE_X1 34, V1CMPEQI_IMM8_OPCODE_X1 */
> +    NULL, /* 35, SHL3ADDX_RRR_0_OPCODE_X1 35, SWINT1_UNARY_OPCODE_X1 35, V1CMPLTSI_IMM8_OPCODE_X1 */
> +    NULL, /* 36, SHL3ADD_RRR_0_OPCODE_X1 36, SWINT2_UNARY_OPCODE_X1 36, V1CMPLTUI_IMM8_OPCODE_X1 */
> +    NULL, /* 37, SHLX_RRR_0_OPCODE_X1 37, SWINT3_UNARY_OPCODE_X1 37, V1MAXUI_IMM8_OPCODE_X1 */
> +    NULL, /* 38, SHL_RRR_0_OPCODE_X1 38, V1MINUI_IMM8_OPCODE_X1 38, WH64_UNARY_OPCODE_X1 */
> +    NULL, /* 39, SHRS_RRR_0_OPCODE_X1 39, V2ADDI_IMM8_OPCODE_X1 */
> +    NULL, /* 40, SHRUX_RRR_0_OPCODE_X1 40, V2CMPEQI_IMM8_OPCODE_X1 */
> +    NULL, /* 41, SHRU_RRR_0_OPCODE_X1 41, V2CMPLTSI_IMM8_OPCODE_X1 */
> +    NULL, /* 42, ST1_RRR_0_OPCODE_X1 42, V2CMPLTUI_IMM8_OPCODE_X1 */
> +    NULL, /* 43, ST2_RRR_0_OPCODE_X1 43, V2MAXSI_IMM8_OPCODE_X1 */
> +    NULL, /* 44, ST4_RRR_0_OPCODE_X1 44, V2MINSI_IMM8_OPCODE_X1 */
> +    NULL, /* 45, STNT1_RRR_0_OPCODE_X1 45, XORI_IMM8_OPCODE_X1 */
> +    NULL, /* 46, STNT2_RRR_0_OPCODE_X1 */
> +    NULL, /* 47, STNT4_RRR_0_OPCODE_X1 */
> +    NULL, /* 48, STNT_RRR_0_OPCODE_X1 */
> +    NULL, /* 49, ST_RRR_0_OPCODE_X1 */
> +    NULL, /* 50, SUBXSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 51, SUBX_RRR_0_OPCODE_X1 */
> +    NULL, /* 52, SUB_RRR_0_OPCODE_X1 */
> +    NULL, /* 53, UNARY_RRR_0_OPCODE_X1 */
> +    NULL, /* 54, V1ADDUC_RRR_0_OPCODE_X1 */
> +    NULL, /* 55, V1ADD_RRR_0_OPCODE_X1 */
> +    NULL, /* 56, V1CMPEQ_RRR_0_OPCODE_X1 */
> +    NULL, /* 57, V1CMPLES_RRR_0_OPCODE_X1 */
> +    NULL, /* 58, V1CMPLEU_RRR_0_OPCODE_X1 */
> +    NULL, /* 59, V1CMPLTS_RRR_0_OPCODE_X1 */
> +    NULL, /* 60, V1CMPLTU_RRR_0_OPCODE_X1 */
> +    NULL, /* 61, V1CMPNE_RRR_0_OPCODE_X1 */
> +    NULL, /* 62, V1INT_H_RRR_0_OPCODE_X1 */
> +    NULL, /* 63, V1INT_L_RRR_0_OPCODE_X1 */
> +    NULL, /* 64, V1MAXU_RRR_0_OPCODE_X1 */
> +    NULL, /* 65, V1MINU_RRR_0_OPCODE_X1 */
> +    NULL, /* 66, V1MNZ_RRR_0_OPCODE_X1 */
> +    NULL, /* 67, V1MZ_RRR_0_OPCODE_X1 */
> +    NULL, /* 68, V1SHL_RRR_0_OPCODE_X1 */
> +    NULL, /* 69, V1SHRS_RRR_0_OPCODE_X1 */
> +    NULL, /* 70, V1SHRU_RRR_0_OPCODE_X1 */
> +    NULL, /* 71, V1SUBUC_RRR_0_OPCODE_X1 */
> +    NULL, /* 72, V1SUB_RRR_0_OPCODE_X1 */
> +    NULL, /* 73, V2ADDSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 74, V2ADD_RRR_0_OPCODE_X1 */
> +    NULL, /* 75, V2CMPEQ_RRR_0_OPCODE_X1 */
> +    NULL, /* 76, V2CMPLES_RRR_0_OPCODE_X1 */
> +    NULL, /* 77, V2CMPLEU_RRR_0_OPCODE_X1 */
> +    NULL, /* 78, V2CMPLTS_RRR_0_OPCODE_X1 */
> +    NULL, /* 79, V2CMPLTU_RRR_0_OPCODE_X1 */
> +    NULL, /* 80, V2CMPNE_RRR_0_OPCODE_X1 */
> +    NULL, /* 81, V2INT_H_RRR_0_OPCODE_X1 */
> +    NULL, /* 82, V2INT_L_RRR_0_OPCODE_X1 */
> +    NULL, /* 83, V2MAXS_RRR_0_OPCODE_X1 */
> +    NULL, /* 84, V2MINS_RRR_0_OPCODE_X1 */
> +    NULL, /* 85, V2MNZ_RRR_0_OPCODE_X1 */
> +    NULL, /* 86, V2MZ_RRR_0_OPCODE_X1 */
> +    NULL, /* 87, V2PACKH_RRR_0_OPCODE_X1 */
> +    NULL, /* 88, V2PACKL_RRR_0_OPCODE_X1 */
> +    NULL, /* 89, V2PACKUC_RRR_0_OPCODE_X1 */
> +    NULL, /* 90, V2SHLSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 91, V2SHL_RRR_0_OPCODE_X1 */
> +    NULL, /* 92, V2SHRS_RRR_0_OPCODE_X1 */
> +    NULL, /* 93, V2SHRU_RRR_0_OPCODE_X1 */
> +    NULL, /* 94, V2SUBSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 95, V2SUB_RRR_0_OPCODE_X1 */
> +    NULL, /* 96, V4ADDSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 97, V4ADD_RRR_0_OPCODE_X1 */
> +    NULL, /* 98, V4INT_H_RRR_0_OPCODE_X1 */
> +    NULL, /* 99, V4INT_L_RRR_0_OPCODE_X1 */
> +    NULL, /* 100, V4PACKSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 101, V4SHLSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 102, V4SHL_RRR_0_OPCODE_X1 */
> +    NULL, /* 103, V4SHRS_RRR_0_OPCODE_X1 */
> +    NULL, /* 104, V4SHRU_RRR_0_OPCODE_X1 */
> +    NULL, /* 105, V4SUBSC_RRR_0_OPCODE_X1 */
> +    NULL, /* 106, V4SUB_RRR_0_OPCODE_X1 */
> +    NULL  /* 107, XOR_RRR_0_OPCODE_X1 */
> +};
> +
> +/* by "grep _OPCODE_Y0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
> +static TileGXDecodeInsn decode_insns_y0[] = {
> +    NULL, /* 0, ADDI_OPCODE_Y0 0, ADDX_RRR_0_OPCODE_Y0 0, AND_RRR_5_OPCODE_Y0 0, CMOVEQZ_RRR_4_OPCODE_Y0 0, CMPEQ_RRR_3_OPCODE_Y0 0, CMPLES_RRR_2_OPCODE_Y0 0, MULA_HS_HS_RRR_9_OPCODE_Y0 0, MUL_HS_HS_RRR_8_OPCODE_Y0 0, ROTLI_SHIFT_OPCODE_Y0 0, ROTL_RRR_6_OPCODE_Y0 0, SHL1ADD_RRR_1_OPCODE_Y0 */
> +    NULL, /* 1, ADD_RRR_0_OPCODE_Y0 1, ADDXI_OPCODE_Y0 1, CMOVNEZ_RRR_4_OPCODE_Y0 1, CMPLEU_RRR_2_OPCODE_Y0 1, CMPNE_RRR_3_OPCODE_Y0 1, CNTLZ_UNARY_OPCODE_Y0 1, MULA_HU_HU_RRR_9_OPCODE_Y0 1, MUL_HU_HU_RRR_8_OPCODE_Y0 1, NOR_RRR_5_OPCODE_Y0 1, SHL1ADDX_RRR_7_OPCODE_Y0 1, SHL2ADD_RRR_1_OPCODE_Y0 1, SHLI_SHIFT_OPCODE_Y0 1, SHL_RRR_6_OPCODE_Y0 */
> +    NULL, /* 2, ANDI_OPCODE_Y0 2, CMPLTS_RRR_2_OPCODE_Y0 2, CNTTZ_UNARY_OPCODE_Y0 2, MNZ_RRR_4_OPCODE_Y0 2, MULA_LS_LS_RRR_9_OPCODE_Y0 2, MULAX_RRR_3_OPCODE_Y0 2, MUL_LS_LS_RRR_8_OPCODE_Y0 2, OR_RRR_5_OPCODE_Y0 2, SHL2ADDX_RRR_7_OPCODE_Y0 2, SHL3ADD_RRR_1_OPCODE_Y0 2, SHRSI_SHIFT_OPCODE_Y0 2, SHRS_RRR_6_OPCODE_Y0 2, SUBX_RRR_0_OPCODE_Y0 */
> +    NULL, /* 3, CMPEQI_OPCODE_Y0 3, CMPLTU_RRR_2_OPCODE_Y0 3, FNOP_UNARY_OPCODE_Y0 3, MULA_LU_LU_RRR_9_OPCODE_Y0 3, MUL_LU_LU_RRR_8_OPCODE_Y0 3, MULX_RRR_3_OPCODE_Y0 3, MZ_RRR_4_OPCODE_Y0 3, SHL3ADDX_RRR_7_OPCODE_Y0 3, SHRUI_SHIFT_OPCODE_Y0 3, SHRU_RRR_6_OPCODE_Y0 3, SUB_RRR_0_OPCODE_Y0 3, UNARY_RRR_1_OPCODE_Y0 3, XOR_RRR_5_OPCODE_Y0 */
> +    NULL, /* 4, CMPLTSI_OPCODE_Y0 4, FSINGLE_PACK1_UNARY_OPCODE_Y0 */
> +    NULL, /* 5, NOP_UNARY_OPCODE_Y0 5, RRR_0_OPCODE_Y0 */
> +    NULL, /* 6, PCNT_UNARY_OPCODE_Y0 6, RRR_1_OPCODE_Y0 */
> +    NULL, /* 7, REVBITS_UNARY_OPCODE_Y0 7, RRR_2_OPCODE_Y0 */
> +    NULL, /* 8, REVBYTES_UNARY_OPCODE_Y0 8, RRR_3_OPCODE_Y0 */
> +    NULL, /* 9, RRR_4_OPCODE_Y0 9, TBLIDXB0_UNARY_OPCODE_Y0 */
> +    NULL, /* 10, RRR_5_OPCODE_Y0 10, TBLIDXB1_UNARY_OPCODE_Y0 */
> +    NULL, /* 11, RRR_6_OPCODE_Y0 11, TBLIDXB2_UNARY_OPCODE_Y0 */
> +    NULL, /* 12, RRR_7_OPCODE_Y0 12, TBLIDXB3_UNARY_OPCODE_Y0 */
> +    NULL, /* 13, RRR_8_OPCODE_Y0 */
> +    NULL, /* 14, RRR_9_OPCODE_Y0 */
> +    NULL  /* 15, SHIFT_OPCODE_Y0 */
> +};
> +
> +/* by "grep _OPCODE_Y1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
> +static TileGXDecodeInsn decode_insns_y1[] = {
> +    NULL, /* 0, ADDX_SPECIAL_0_OPCODE_Y1 0, AND_RRR_5_OPCODE_Y1 0, CMPLES_RRR_2_OPCODE_Y1 0, ROTLI_SHIFT_OPCODE_Y1 0, ROTL_RRR_6_OPCODE_Y1 0, SHL1ADD_RRR_1_OPCODE_Y1 */
> +    NULL, /* 1, ADDI_OPCODE_Y1 1, ADD_SPECIAL_0_OPCODE_Y1 1, CMPLEU_RRR_2_OPCODE_Y1 1, NOR_RRR_5_OPCODE_Y1 1, SHL1ADDX_RRR_7_OPCODE_Y1 1, SHL2ADD_RRR_1_OPCODE_Y1 1, SHLI_SHIFT_OPCODE_Y1 1, SHL_RRR_6_OPCODE_Y1 */
> +    NULL, /* 2, ADDXI_OPCODE_Y1 2, CMPEQ_RRR_3_OPCODE_Y1 2, CMPLTS_RRR_2_OPCODE_Y1 2, MNZ_RRR_4_OPCODE_Y1 2, OR_RRR_5_OPCODE_Y1 2, SHL2ADDX_RRR_7_OPCODE_Y1 2, SHL3ADD_RRR_1_OPCODE_Y1 2, SHRSI_SHIFT_OPCODE_Y1 2, SHRS_RRR_6_OPCODE_Y1 2, SUBX_RRR_0_OPCODE_Y1 */
> +    NULL, /* 3, ANDI_OPCODE_Y1 3, CMPLTU_RRR_2_OPCODE_Y1 3, CMPNE_RRR_3_OPCODE_Y1 3, MZ_RRR_4_OPCODE_Y1 3, SHL3ADDX_RRR_7_OPCODE_Y1 3, SHRUI_SHIFT_OPCODE_Y1 3, SHRU_RRR_6_OPCODE_Y1 3, SUB_RRR_0_OPCODE_Y1 3, UNARY_RRR_1_OPCODE_Y1 3 XOR_RRR_5_OPCODE_Y1 */
> +    NULL, /* 4, CMPEQI_OPCODE_Y1 */
> +    NULL, /* 5, CMPLTSI_OPCODE_Y1 */
> +    NULL, /* 6, RRR_0_OPCODE_Y1 */
> +    NULL, /* 7, RRR_1_OPCODE_Y1 */
> +    NULL, /* 8, FNOP_UNARY_OPCODE_Y1 8, RRR_2_OPCODE_Y1 */
> +    NULL, /* 9, ILL_UNARY_OPCODE_Y1 9, RRR_3_OPCODE_Y1 */
> +    NULL, /* 10, JALRP_UNARY_OPCODE_Y1 10, RRR_4_OPCODE_Y1 */
> +    NULL, /* 11, JALR_UNARY_OPCODE_Y1 11, RRR_5_OPCODE_Y1 */
> +    NULL, /* 12, JRP_UNARY_OPCODE_Y1 12, RRR_6_OPCODE_Y1 */
> +    NULL, /* 13, JR_UNARY_OPCODE_Y1 13, RRR_7_OPCODE_Y1 */
> +    NULL, /* 14, LNK_UNARY_OPCODE_Y1 14, SHIFT_OPCODE_Y1 */
> +    NULL  /* 15, NOP_UNARY_OPCODE_Y1 */
> +};
> +
> +/* by "grep _OPCODE_Y2 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
> +static TileGXDecodeInsn decode_insns_y2[] = {
> +    NULL, /* 0, LD1S_OPCODE_Y2 0, ST1_OPCODE_Y2 */
> +    NULL, /* 1, LD1U_OPCODE_Y2 1, LD4S_OPCODE_Y2 1, ST2_OPCODE_Y2 */
> +    NULL, /* 2, LD2S_OPCODE_Y2 2, LD4U_OPCODE_Y2 2, ST4_OPCODE_Y2 */
> +    NULL  /* 3, LD2U_OPCODE_Y2 3, LD_OPCODE_Y2 3, ST_OPCODE_Y2 */
> +};
> +
> +void tilegx_tcg_init(void)
> +{
> +    int i;
> +
> +    cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
> +    cpu_pc = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUTLGState, pc), "pc");
> +    for (i = 0; i < TILEGX_R_COUNT; i++) {
> +        cpu_regs[i] = tcg_global_mem_new_i64(TCG_AREG0,
> +                                             offsetof(CPUTLGState, regs[i]),
> +                                             reg_names[i]);
> +    }
> +}
> +
> +static void translate_x0(struct DisasContext *dc, tilegx_bundle_bits bundle)
> +{
> +    unsigned int opcode = get_Opcode_X0(bundle);
> +
> +    if (opcode > TILEGX_OPCODE_MAX_X0) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
> +        qemu_log("unknown x0 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +    if (!decode_insns_x0[opcode]) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
> +        qemu_log("unimplement x0 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +
> +    dc->tmp_regcur = dc->tmp_regs + 0;
> +    decode_insns_x0[opcode](dc, bundle);
> +}
> +
> +static void translate_x1(struct DisasContext *dc, tilegx_bundle_bits bundle)
> +{
> +    unsigned int opcode = get_Opcode_X1(bundle);
> +
> +    if (opcode > TILEGX_OPCODE_MAX_X1) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
> +        qemu_log("unknown x1 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +    if (!decode_insns_x1[opcode]) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
> +        qemu_log("unimplement x1 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +
> +    dc->tmp_regcur = dc->tmp_regs + 1;
> +    decode_insns_x1[opcode](dc, bundle);
> +}
> +
> +static void translate_y0(struct DisasContext *dc, tilegx_bundle_bits bundle)
> +{
> +    unsigned int opcode = get_Opcode_Y0(bundle);
> +
> +    if (opcode > TILEGX_OPCODE_MAX_Y0) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
> +        qemu_log("unknown y0 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +    if (!decode_insns_y0[opcode]) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
> +        qemu_log("unimplement y0 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +
> +    dc->tmp_regcur = dc->tmp_regs + 0;
> +    decode_insns_y0[opcode](dc, bundle);
> +}
> +
> +static void translate_y1(struct DisasContext *dc, tilegx_bundle_bits bundle)
> +{
> +    unsigned int opcode = get_Opcode_Y1(bundle);
> +
> +    if (opcode > TILEGX_OPCODE_MAX_Y1) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
> +        qemu_log("unknown y1 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +    if (!decode_insns_y1[opcode]) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
> +        qemu_log("unimplement y1 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +
> +    dc->tmp_regcur = dc->tmp_regs + 1;
> +    decode_insns_y1[opcode](dc, bundle);
> +}
> +
> +static void translate_y2(struct DisasContext *dc, tilegx_bundle_bits bundle)
> +{
> +    unsigned int opcode = get_Opcode_Y2(bundle);
> +
> +    if (opcode > TILEGX_OPCODE_MAX_Y2) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
> +        qemu_log("unknown y2 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +    if (!decode_insns_y2[opcode]) {
> +        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
> +        qemu_log("unimplement y2 opcode: %u for bundle: %16.16llx\n",
> +                 opcode, bundle);
> +        return;
> +    }
> +
> +    dc->tmp_regcur = dc->tmp_regs + 2;
> +    decode_insns_y2[opcode](dc, bundle);
> +}
> +
> +static void translate_one_bundle(struct DisasContext *dc, uint64_t bundle)
> +{
> +    int i;
> +    TCGv tmp;
> +
> +    for (i = 0; i < TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE; i++) {
> +        dc->tmp_regs[i].idx = TILEGX_R_NOREG;
> +        TCGV_UNUSED_I64(dc->tmp_regs[i].val);
> +    }
> +
> +    if (TILEGX_BUNDLE_TYPE_Y(bundle)) {
> +        translate_y0(dc, bundle);
> +        translate_y1(dc, bundle);
> +        translate_y2(dc, bundle);
> +    } else {
> +        translate_x0(dc, bundle);
> +        translate_x1(dc, bundle);
> +    }
> +
> +    for (i = 0; i < TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE; i++) {
> +        if (dc->tmp_regs[i].idx == TILEGX_R_NOREG) {
> +            continue;
> +        }
> +        tcg_gen_mov_i64(cpu_regs[dc->tmp_regs[i].idx], dc->tmp_regs[i].val);
> +        tcg_temp_free_i64(dc->tmp_regs[i].val);
> +    }
> +
> +    if (dc->jmp.cond != TCG_COND_NEVER) {
> +        if (dc->jmp.cond == TCG_COND_ALWAYS) {
> +            tcg_gen_mov_i64(cpu_pc, dc->jmp.dest);
> +        } else {
> +            tmp = tcg_const_i64(dc->pc + TILEGX_BUNDLE_SIZE_IN_BYTES);
> +            tcg_gen_movcond_i64(dc->jmp.cond, cpu_pc,
> +                                dc->jmp.val1, dc->jmp.val2,
> +                                dc->jmp.dest, tmp);
> +            tcg_temp_free_i64(dc->jmp.val1);
> +            tcg_temp_free_i64(dc->jmp.val2);
> +            tcg_temp_free_i64(tmp);
> +        }
> +        tcg_temp_free_i64(dc->jmp.dest);
> +        tcg_gen_exit_tb(0);
> +    }
> +}
>  
>  static inline void gen_intermediate_code_internal(TileGXCPU *cpu,
>                                                    TranslationBlock *tb,
>                                                    bool search_pc)
>  {
> -    /*
> -     * FIXME: after load elf64 tilegx binary successfully, it will quit, at
> -     * present, and will implement the related features next.
> -     */
> -    qemu_log("\nLoaded elf64 tilegx successfully\n");
> -    qemu_log("reached code start position: [" TARGET_FMT_lx "] %s\n\n",
> -             tb->pc, lookup_symbol(tb->pc));
> -    exit(0);
> +    DisasContext ctx;
> +    DisasContext *dc = &ctx;
> +
> +    CPUTLGState *env = &cpu->env;
> +    uint64_t pc_start = tb->pc;
> +    uint64_t next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
> +    int j, lj = -1;
> +    int num_insns = 0;
> +    int max_insns = tb->cflags & CF_COUNT_MASK;
> +
> +    dc->pc = pc_start;
> +    dc->exception = 0;
> +    dc->jmp.cond = TCG_COND_NEVER;
> +    TCGV_UNUSED_I64(dc->jmp.dest);
> +    TCGV_UNUSED_I64(dc->jmp.val1);
> +    TCGV_UNUSED_I64(dc->jmp.val2);
> +
> +    if (!max_insns) {
> +        max_insns = CF_COUNT_MASK;
> +    }
> +    gen_tb_start(tb);
> +
> +    do {
> +        TCGV_UNUSED_I64(dc->zero);
> +        if (search_pc) {
> +            j = tcg_op_buf_count();
> +            if (lj < j) {
> +                lj++;
> +                while (lj < j) {
> +                    tcg_ctx.gen_opc_instr_start[lj++] = 0;
> +                }
> +            }
> +            tcg_ctx.gen_opc_pc[lj] = dc->pc;
> +            tcg_ctx.gen_opc_instr_start[lj] = 1;
> +            tcg_ctx.gen_opc_icount[lj] = num_insns;
> +        }
> +        translate_one_bundle(dc, cpu_ldq_data(env, dc->pc));
> +        if (dc->exception) {
> +            exit(-1);
> +        }
> +        num_insns++;
> +        dc->pc += TILEGX_BUNDLE_SIZE_IN_BYTES;
> +    } while (dc->jmp.cond == TCG_COND_NEVER
> +             && dc->pc <= next_page_start - TILEGX_BUNDLE_SIZE_IN_BYTES
> +             && num_insns <= max_insns - TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE
> +             && !tcg_op_buf_full());
> +
> +    gen_tb_end(tb, num_insns);
> +    if (search_pc) {
> +        j = tcg_op_buf_count();
> +        lj++;
> +        while (lj <= j) {
> +            tcg_ctx.gen_opc_instr_start[lj++] = 0;
> +        }
> +    } else {
> +        tb->size = dc->pc - pc_start;
> +        tb->icount = num_insns;
> +    }
> +
> +    return;
>  }
>  
>  void gen_intermediate_code(CPUTLGState *env, struct TranslationBlock *tb)
>
diff mbox

Patch

diff --git a/target-tilegx/cpu-qom.h b/target-tilegx/cpu-qom.h
index 4ee11e1..5615c3b 100644
--- a/target-tilegx/cpu-qom.h
+++ b/target-tilegx/cpu-qom.h
@@ -68,4 +68,6 @@  static inline TileGXCPU *tilegx_env_get_cpu(CPUTLGState *env)
 
 #define ENV_GET_CPU(e) CPU(tilegx_env_get_cpu(e))
 
+#define ENV_OFFSET offsetof(TileGXCPU, env)
+
 #endif
diff --git a/target-tilegx/cpu.c b/target-tilegx/cpu.c
index cf46b8b..8255fdc 100644
--- a/target-tilegx/cpu.c
+++ b/target-tilegx/cpu.c
@@ -69,10 +69,6 @@  static void tilegx_cpu_realizefn(DeviceState *dev, Error **errp)
     tcc->parent_realize(dev, errp);
 }
 
-static void tilegx_tcg_init(void)
-{
-}
-
 static void tilegx_cpu_initfn(Object *obj)
 {
     CPUState *cs = CPU(obj);
diff --git a/target-tilegx/cpu.h b/target-tilegx/cpu.h
index 87dc56b..93e16c3 100644
--- a/target-tilegx/cpu.h
+++ b/target-tilegx/cpu.h
@@ -30,16 +30,21 @@ 
 #include "fpu/softfloat.h"
 
 /* TILE-Gx register alias */
-#define TILEGX_R_RE  0   /*  0 register, for function/syscall return value */
-#define TILEGX_R_NR  10  /* 10 register, for syscall number */
-#define TILEGX_R_BP  52  /* 52 register, optional frame pointer */
-#define TILEGX_R_TP  53  /* TP register, thread local storage data */
-#define TILEGX_R_SP  54  /* SP register, stack pointer */
-#define TILEGX_R_LR  55  /* LR register, may save pc, but it is not pc */
+#define TILEGX_R_RE    0   /*  0 register, for function/syscall return value */
+#define TILEGX_R_NR    10  /* 10 register, for syscall number */
+#define TILEGX_R_BP    52  /* 52 register, optional frame pointer */
+#define TILEGX_R_TP    53  /* TP register, thread local storage data */
+#define TILEGX_R_SP    54  /* SP register, stack pointer */
+#define TILEGX_R_LR    55  /* LR register, may save pc, but it is not pc */
+#define TILEGX_R_ZERO  63  /* Zero register, always zero */
+#define TILEGX_R_COUNT 56  /* Only 56 registers are really useful */
+#define TILEGX_R_NOREG 255 /* Invalid register value */
+
 
 typedef struct CPUTLGState {
-    uint64_t regs[56];
-    uint64_t pc;
+    uint64_t regs[TILEGX_R_COUNT]; /* Common used registers by outside */
+    uint64_t pc;                   /* Current pc */
+
     CPU_COMMON
 } CPUTLGState;
 
@@ -54,6 +59,7 @@  typedef struct CPUTLGState {
 
 #include "exec/cpu-all.h"
 
+void tilegx_tcg_init(void);
 int cpu_tilegx_exec(CPUTLGState *s);
 int cpu_tilegx_signal_handler(int host_signum, void *pinfo, void *puc);
 
diff --git a/target-tilegx/opcode_tilegx.h b/target-tilegx/opcode_tilegx.h
new file mode 100644
index 0000000..d76ff2d
--- /dev/null
+++ b/target-tilegx/opcode_tilegx.h
@@ -0,0 +1,1406 @@ 
+/* TILE-Gx opcode information.
+ *
+ * Copyright 2011 Tilera Corporation. All Rights Reserved.
+ *
+ *   This program is free software; you can redistribute it and/or
+ *   modify it under the terms of the GNU General Public License
+ *   as published by the Free Software Foundation, version 2.
+ *
+ *   This program is distributed in the hope that it will be useful, but
+ *   WITHOUT ANY WARRANTY; without even the implied warranty of
+ *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
+ *   NON INFRINGEMENT.  See the GNU General Public License for
+ *   more details.
+ *
+ *
+ *
+ *
+ *
+ */
+
+#ifndef __ARCH_OPCODE_H__
+#define __ARCH_OPCODE_H__
+
+#ifndef __ASSEMBLER__
+
+typedef unsigned long long tilegx_bundle_bits;
+
+/* These are the bits that determine if a bundle is in the X encoding. */
+#define TILEGX_BUNDLE_MODE_MASK ((tilegx_bundle_bits)3 << 62)
+
+enum
+{
+  /* Maximum number of instructions in a bundle (2 for X, 3 for Y). */
+  TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE = 3,
+
+  /* How many different pipeline encodings are there? X0, X1, Y0, Y1, Y2. */
+  TILEGX_NUM_PIPELINE_ENCODINGS = 5,
+
+  /* Log base 2 of TILEGX_BUNDLE_SIZE_IN_BYTES. */
+  TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES = 3,
+
+  /* Instructions take this many bytes. */
+  TILEGX_BUNDLE_SIZE_IN_BYTES = 1 << TILEGX_LOG2_BUNDLE_SIZE_IN_BYTES,
+
+  /* Log base 2 of TILEGX_BUNDLE_ALIGNMENT_IN_BYTES. */
+  TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES = 3,
+
+  /* Bundles should be aligned modulo this number of bytes. */
+  TILEGX_BUNDLE_ALIGNMENT_IN_BYTES =
+    (1 << TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES),
+
+  /* Number of registers (some are magic, such as network I/O). */
+  TILEGX_NUM_REGISTERS = 64,
+};
+
+/* Make a few "tile_" variables to simplify common code between
+   architectures.  */
+
+typedef tilegx_bundle_bits tile_bundle_bits;
+#define TILE_BUNDLE_SIZE_IN_BYTES TILEGX_BUNDLE_SIZE_IN_BYTES
+#define TILE_BUNDLE_ALIGNMENT_IN_BYTES TILEGX_BUNDLE_ALIGNMENT_IN_BYTES
+#define TILE_LOG2_BUNDLE_ALIGNMENT_IN_BYTES \
+  TILEGX_LOG2_BUNDLE_ALIGNMENT_IN_BYTES
+#define TILE_BPT_BUNDLE TILEGX_BPT_BUNDLE
+
+/* 64-bit pattern for a { bpt ; nop } bundle. */
+#define TILEGX_BPT_BUNDLE 0x286a44ae51485000ULL
+
+static __inline unsigned int
+get_BFEnd_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0x3f);
+}
+
+static __inline unsigned int
+get_BFOpcodeExtension_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 24)) & 0xf);
+}
+
+static __inline unsigned int
+get_BFStart_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 18)) & 0x3f);
+}
+
+static __inline unsigned int
+get_BrOff_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 31)) & 0x0000003f) |
+         (((unsigned int)(n >> 37)) & 0x0001ffc0);
+}
+
+static __inline unsigned int
+get_BrType_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 54)) & 0x1f);
+}
+
+static __inline unsigned int
+get_Dest_Imm8_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 31)) & 0x0000003f) |
+         (((unsigned int)(n >> 43)) & 0x000000c0);
+}
+
+static __inline unsigned int
+get_Dest_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 0)) & 0x3f);
+}
+
+static __inline unsigned int
+get_Dest_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 31)) & 0x3f);
+}
+
+static __inline unsigned int
+get_Dest_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 0)) & 0x3f);
+}
+
+static __inline unsigned int
+get_Dest_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 31)) & 0x3f);
+}
+
+static __inline unsigned int
+get_Imm16_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0xffff);
+}
+
+static __inline unsigned int
+get_Imm16_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0xffff);
+}
+
+static __inline unsigned int
+get_Imm8OpcodeExtension_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 20)) & 0xff);
+}
+
+static __inline unsigned int
+get_Imm8OpcodeExtension_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 51)) & 0xff);
+}
+
+static __inline unsigned int
+get_Imm8_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0xff);
+}
+
+static __inline unsigned int
+get_Imm8_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0xff);
+}
+
+static __inline unsigned int
+get_Imm8_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0xff);
+}
+
+static __inline unsigned int
+get_Imm8_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0xff);
+}
+
+static __inline unsigned int
+get_JumpOff_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 31)) & 0x7ffffff);
+}
+
+static __inline unsigned int
+get_JumpOpcodeExtension_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 58)) & 0x1);
+}
+
+static __inline unsigned int
+get_MF_Imm14_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 37)) & 0x3fff);
+}
+
+static __inline unsigned int
+get_MT_Imm14_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 31)) & 0x0000003f) |
+         (((unsigned int)(n >> 37)) & 0x00003fc0);
+}
+
+static __inline unsigned int
+get_Mode(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 62)) & 0x3);
+}
+
+static __inline unsigned int
+get_Opcode_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 28)) & 0x7);
+}
+
+static __inline unsigned int
+get_Opcode_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 59)) & 0x7);
+}
+
+static __inline unsigned int
+get_Opcode_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 27)) & 0xf);
+}
+
+static __inline unsigned int
+get_Opcode_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 58)) & 0xf);
+}
+
+static __inline unsigned int
+get_Opcode_Y2(tilegx_bundle_bits n)
+{
+  return (((n >> 26)) & 0x00000001) |
+         (((unsigned int)(n >> 56)) & 0x00000002);
+}
+
+static __inline unsigned int
+get_RRROpcodeExtension_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 18)) & 0x3ff);
+}
+
+static __inline unsigned int
+get_RRROpcodeExtension_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 49)) & 0x3ff);
+}
+
+static __inline unsigned int
+get_RRROpcodeExtension_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 18)) & 0x3);
+}
+
+static __inline unsigned int
+get_RRROpcodeExtension_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 49)) & 0x3);
+}
+
+static __inline unsigned int
+get_ShAmt_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0x3f);
+}
+
+static __inline unsigned int
+get_ShAmt_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0x3f);
+}
+
+static __inline unsigned int
+get_ShAmt_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0x3f);
+}
+
+static __inline unsigned int
+get_ShAmt_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0x3f);
+}
+
+static __inline unsigned int
+get_ShiftOpcodeExtension_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 18)) & 0x3ff);
+}
+
+static __inline unsigned int
+get_ShiftOpcodeExtension_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 49)) & 0x3ff);
+}
+
+static __inline unsigned int
+get_ShiftOpcodeExtension_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 18)) & 0x3);
+}
+
+static __inline unsigned int
+get_ShiftOpcodeExtension_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 49)) & 0x3);
+}
+
+static __inline unsigned int
+get_SrcA_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 6)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcA_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 37)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcA_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 6)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcA_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 37)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcA_Y2(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 20)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcBDest_Y2(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 51)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcB_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcB_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcB_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0x3f);
+}
+
+static __inline unsigned int
+get_SrcB_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0x3f);
+}
+
+static __inline unsigned int
+get_UnaryOpcodeExtension_X0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0x3f);
+}
+
+static __inline unsigned int
+get_UnaryOpcodeExtension_X1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0x3f);
+}
+
+static __inline unsigned int
+get_UnaryOpcodeExtension_Y0(tilegx_bundle_bits num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((n >> 12)) & 0x3f);
+}
+
+static __inline unsigned int
+get_UnaryOpcodeExtension_Y1(tilegx_bundle_bits n)
+{
+  return (((unsigned int)(n >> 43)) & 0x3f);
+}
+
+
+static __inline int
+sign_extend(int n, int num_bits)
+{
+  int shift = (int)(sizeof(int) * 8 - num_bits);
+  return (n << shift) >> shift;
+}
+
+
+
+static __inline tilegx_bundle_bits
+create_BFEnd_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_BFOpcodeExtension_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0xf) << 24);
+}
+
+static __inline tilegx_bundle_bits
+create_BFStart_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 18);
+}
+
+static __inline tilegx_bundle_bits
+create_BrOff_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
+         (((tilegx_bundle_bits)(n & 0x0001ffc0)) << 37);
+}
+
+static __inline tilegx_bundle_bits
+create_BrType_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x1f)) << 54);
+}
+
+static __inline tilegx_bundle_bits
+create_Dest_Imm8_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
+         (((tilegx_bundle_bits)(n & 0x000000c0)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_Dest_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 0);
+}
+
+static __inline tilegx_bundle_bits
+create_Dest_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 31);
+}
+
+static __inline tilegx_bundle_bits
+create_Dest_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 0);
+}
+
+static __inline tilegx_bundle_bits
+create_Dest_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 31);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm16_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0xffff) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm16_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0xffff)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm8OpcodeExtension_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0xff) << 20);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm8OpcodeExtension_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0xff)) << 51);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm8_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0xff) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm8_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0xff)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm8_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0xff) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_Imm8_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0xff)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_JumpOff_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x7ffffff)) << 31);
+}
+
+static __inline tilegx_bundle_bits
+create_JumpOpcodeExtension_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x1)) << 58);
+}
+
+static __inline tilegx_bundle_bits
+create_MF_Imm14_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3fff)) << 37);
+}
+
+static __inline tilegx_bundle_bits
+create_MT_Imm14_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x0000003f)) << 31) |
+         (((tilegx_bundle_bits)(n & 0x00003fc0)) << 37);
+}
+
+static __inline tilegx_bundle_bits
+create_Mode(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3)) << 62);
+}
+
+static __inline tilegx_bundle_bits
+create_Opcode_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x7) << 28);
+}
+
+static __inline tilegx_bundle_bits
+create_Opcode_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x7)) << 59);
+}
+
+static __inline tilegx_bundle_bits
+create_Opcode_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0xf) << 27);
+}
+
+static __inline tilegx_bundle_bits
+create_Opcode_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0xf)) << 58);
+}
+
+static __inline tilegx_bundle_bits
+create_Opcode_Y2(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x00000001) << 26) |
+         (((tilegx_bundle_bits)(n & 0x00000002)) << 56);
+}
+
+static __inline tilegx_bundle_bits
+create_RRROpcodeExtension_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3ff) << 18);
+}
+
+static __inline tilegx_bundle_bits
+create_RRROpcodeExtension_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3ff)) << 49);
+}
+
+static __inline tilegx_bundle_bits
+create_RRROpcodeExtension_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3) << 18);
+}
+
+static __inline tilegx_bundle_bits
+create_RRROpcodeExtension_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3)) << 49);
+}
+
+static __inline tilegx_bundle_bits
+create_ShAmt_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_ShAmt_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_ShAmt_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_ShAmt_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_ShiftOpcodeExtension_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3ff) << 18);
+}
+
+static __inline tilegx_bundle_bits
+create_ShiftOpcodeExtension_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3ff)) << 49);
+}
+
+static __inline tilegx_bundle_bits
+create_ShiftOpcodeExtension_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3) << 18);
+}
+
+static __inline tilegx_bundle_bits
+create_ShiftOpcodeExtension_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3)) << 49);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcA_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 6);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcA_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 37);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcA_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 6);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcA_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 37);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcA_Y2(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 20);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcBDest_Y2(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 51);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcB_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcB_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcB_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_SrcB_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_UnaryOpcodeExtension_X0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_UnaryOpcodeExtension_X1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
+}
+
+static __inline tilegx_bundle_bits
+create_UnaryOpcodeExtension_Y0(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return ((n & 0x3f) << 12);
+}
+
+static __inline tilegx_bundle_bits
+create_UnaryOpcodeExtension_Y1(int num)
+{
+  const unsigned int n = (unsigned int)num;
+  return (((tilegx_bundle_bits)(n & 0x3f)) << 43);
+}
+
+
+enum
+{
+  ADDI_IMM8_OPCODE_X0 = 1,
+  ADDI_IMM8_OPCODE_X1 = 1,
+  ADDI_OPCODE_Y0 = 0,
+  ADDI_OPCODE_Y1 = 1,
+  ADDLI_OPCODE_X0 = 1,
+  ADDLI_OPCODE_X1 = 0,
+  ADDXI_IMM8_OPCODE_X0 = 2,
+  ADDXI_IMM8_OPCODE_X1 = 2,
+  ADDXI_OPCODE_Y0 = 1,
+  ADDXI_OPCODE_Y1 = 2,
+  ADDXLI_OPCODE_X0 = 2,
+  ADDXLI_OPCODE_X1 = 1,
+  ADDXSC_RRR_0_OPCODE_X0 = 1,
+  ADDXSC_RRR_0_OPCODE_X1 = 1,
+  ADDX_RRR_0_OPCODE_X0 = 2,
+  ADDX_RRR_0_OPCODE_X1 = 2,
+  ADDX_RRR_0_OPCODE_Y0 = 0,
+  ADDX_SPECIAL_0_OPCODE_Y1 = 0,
+  ADD_RRR_0_OPCODE_X0 = 3,
+  ADD_RRR_0_OPCODE_X1 = 3,
+  ADD_RRR_0_OPCODE_Y0 = 1,
+  ADD_SPECIAL_0_OPCODE_Y1 = 1,
+  ANDI_IMM8_OPCODE_X0 = 3,
+  ANDI_IMM8_OPCODE_X1 = 3,
+  ANDI_OPCODE_Y0 = 2,
+  ANDI_OPCODE_Y1 = 3,
+  AND_RRR_0_OPCODE_X0 = 4,
+  AND_RRR_0_OPCODE_X1 = 4,
+  AND_RRR_5_OPCODE_Y0 = 0,
+  AND_RRR_5_OPCODE_Y1 = 0,
+  BEQZT_BRANCH_OPCODE_X1 = 16,
+  BEQZ_BRANCH_OPCODE_X1 = 17,
+  BFEXTS_BF_OPCODE_X0 = 4,
+  BFEXTU_BF_OPCODE_X0 = 5,
+  BFINS_BF_OPCODE_X0 = 6,
+  BF_OPCODE_X0 = 3,
+  BGEZT_BRANCH_OPCODE_X1 = 18,
+  BGEZ_BRANCH_OPCODE_X1 = 19,
+  BGTZT_BRANCH_OPCODE_X1 = 20,
+  BGTZ_BRANCH_OPCODE_X1 = 21,
+  BLBCT_BRANCH_OPCODE_X1 = 22,
+  BLBC_BRANCH_OPCODE_X1 = 23,
+  BLBST_BRANCH_OPCODE_X1 = 24,
+  BLBS_BRANCH_OPCODE_X1 = 25,
+  BLEZT_BRANCH_OPCODE_X1 = 26,
+  BLEZ_BRANCH_OPCODE_X1 = 27,
+  BLTZT_BRANCH_OPCODE_X1 = 28,
+  BLTZ_BRANCH_OPCODE_X1 = 29,
+  BNEZT_BRANCH_OPCODE_X1 = 30,
+  BNEZ_BRANCH_OPCODE_X1 = 31,
+  BRANCH_OPCODE_X1 = 2,
+  CMOVEQZ_RRR_0_OPCODE_X0 = 5,
+  CMOVEQZ_RRR_4_OPCODE_Y0 = 0,
+  CMOVNEZ_RRR_0_OPCODE_X0 = 6,
+  CMOVNEZ_RRR_4_OPCODE_Y0 = 1,
+  CMPEQI_IMM8_OPCODE_X0 = 4,
+  CMPEQI_IMM8_OPCODE_X1 = 4,
+  CMPEQI_OPCODE_Y0 = 3,
+  CMPEQI_OPCODE_Y1 = 4,
+  CMPEQ_RRR_0_OPCODE_X0 = 7,
+  CMPEQ_RRR_0_OPCODE_X1 = 5,
+  CMPEQ_RRR_3_OPCODE_Y0 = 0,
+  CMPEQ_RRR_3_OPCODE_Y1 = 2,
+  CMPEXCH4_RRR_0_OPCODE_X1 = 6,
+  CMPEXCH_RRR_0_OPCODE_X1 = 7,
+  CMPLES_RRR_0_OPCODE_X0 = 8,
+  CMPLES_RRR_0_OPCODE_X1 = 8,
+  CMPLES_RRR_2_OPCODE_Y0 = 0,
+  CMPLES_RRR_2_OPCODE_Y1 = 0,
+  CMPLEU_RRR_0_OPCODE_X0 = 9,
+  CMPLEU_RRR_0_OPCODE_X1 = 9,
+  CMPLEU_RRR_2_OPCODE_Y0 = 1,
+  CMPLEU_RRR_2_OPCODE_Y1 = 1,
+  CMPLTSI_IMM8_OPCODE_X0 = 5,
+  CMPLTSI_IMM8_OPCODE_X1 = 5,
+  CMPLTSI_OPCODE_Y0 = 4,
+  CMPLTSI_OPCODE_Y1 = 5,
+  CMPLTS_RRR_0_OPCODE_X0 = 10,
+  CMPLTS_RRR_0_OPCODE_X1 = 10,
+  CMPLTS_RRR_2_OPCODE_Y0 = 2,
+  CMPLTS_RRR_2_OPCODE_Y1 = 2,
+  CMPLTUI_IMM8_OPCODE_X0 = 6,
+  CMPLTUI_IMM8_OPCODE_X1 = 6,
+  CMPLTU_RRR_0_OPCODE_X0 = 11,
+  CMPLTU_RRR_0_OPCODE_X1 = 11,
+  CMPLTU_RRR_2_OPCODE_Y0 = 3,
+  CMPLTU_RRR_2_OPCODE_Y1 = 3,
+  CMPNE_RRR_0_OPCODE_X0 = 12,
+  CMPNE_RRR_0_OPCODE_X1 = 12,
+  CMPNE_RRR_3_OPCODE_Y0 = 1,
+  CMPNE_RRR_3_OPCODE_Y1 = 3,
+  CMULAF_RRR_0_OPCODE_X0 = 13,
+  CMULA_RRR_0_OPCODE_X0 = 14,
+  CMULFR_RRR_0_OPCODE_X0 = 15,
+  CMULF_RRR_0_OPCODE_X0 = 16,
+  CMULHR_RRR_0_OPCODE_X0 = 17,
+  CMULH_RRR_0_OPCODE_X0 = 18,
+  CMUL_RRR_0_OPCODE_X0 = 19,
+  CNTLZ_UNARY_OPCODE_X0 = 1,
+  CNTLZ_UNARY_OPCODE_Y0 = 1,
+  CNTTZ_UNARY_OPCODE_X0 = 2,
+  CNTTZ_UNARY_OPCODE_Y0 = 2,
+  CRC32_32_RRR_0_OPCODE_X0 = 20,
+  CRC32_8_RRR_0_OPCODE_X0 = 21,
+  DBLALIGN2_RRR_0_OPCODE_X0 = 22,
+  DBLALIGN2_RRR_0_OPCODE_X1 = 13,
+  DBLALIGN4_RRR_0_OPCODE_X0 = 23,
+  DBLALIGN4_RRR_0_OPCODE_X1 = 14,
+  DBLALIGN6_RRR_0_OPCODE_X0 = 24,
+  DBLALIGN6_RRR_0_OPCODE_X1 = 15,
+  DBLALIGN_RRR_0_OPCODE_X0 = 25,
+  DRAIN_UNARY_OPCODE_X1 = 1,
+  DTLBPR_UNARY_OPCODE_X1 = 2,
+  EXCH4_RRR_0_OPCODE_X1 = 16,
+  EXCH_RRR_0_OPCODE_X1 = 17,
+  FDOUBLE_ADDSUB_RRR_0_OPCODE_X0 = 26,
+  FDOUBLE_ADD_FLAGS_RRR_0_OPCODE_X0 = 27,
+  FDOUBLE_MUL_FLAGS_RRR_0_OPCODE_X0 = 28,
+  FDOUBLE_PACK1_RRR_0_OPCODE_X0 = 29,
+  FDOUBLE_PACK2_RRR_0_OPCODE_X0 = 30,
+  FDOUBLE_SUB_FLAGS_RRR_0_OPCODE_X0 = 31,
+  FDOUBLE_UNPACK_MAX_RRR_0_OPCODE_X0 = 32,
+  FDOUBLE_UNPACK_MIN_RRR_0_OPCODE_X0 = 33,
+  FETCHADD4_RRR_0_OPCODE_X1 = 18,
+  FETCHADDGEZ4_RRR_0_OPCODE_X1 = 19,
+  FETCHADDGEZ_RRR_0_OPCODE_X1 = 20,
+  FETCHADD_RRR_0_OPCODE_X1 = 21,
+  FETCHAND4_RRR_0_OPCODE_X1 = 22,
+  FETCHAND_RRR_0_OPCODE_X1 = 23,
+  FETCHOR4_RRR_0_OPCODE_X1 = 24,
+  FETCHOR_RRR_0_OPCODE_X1 = 25,
+  FINV_UNARY_OPCODE_X1 = 3,
+  FLUSHWB_UNARY_OPCODE_X1 = 4,
+  FLUSH_UNARY_OPCODE_X1 = 5,
+  FNOP_UNARY_OPCODE_X0 = 3,
+  FNOP_UNARY_OPCODE_X1 = 6,
+  FNOP_UNARY_OPCODE_Y0 = 3,
+  FNOP_UNARY_OPCODE_Y1 = 8,
+  FSINGLE_ADD1_RRR_0_OPCODE_X0 = 34,
+  FSINGLE_ADDSUB2_RRR_0_OPCODE_X0 = 35,
+  FSINGLE_MUL1_RRR_0_OPCODE_X0 = 36,
+  FSINGLE_MUL2_RRR_0_OPCODE_X0 = 37,
+  FSINGLE_PACK1_UNARY_OPCODE_X0 = 4,
+  FSINGLE_PACK1_UNARY_OPCODE_Y0 = 4,
+  FSINGLE_PACK2_RRR_0_OPCODE_X0 = 38,
+  FSINGLE_SUB1_RRR_0_OPCODE_X0 = 39,
+  ICOH_UNARY_OPCODE_X1 = 7,
+  ILL_UNARY_OPCODE_X1 = 8,
+  ILL_UNARY_OPCODE_Y1 = 9,
+  IMM8_OPCODE_X0 = 4,
+  IMM8_OPCODE_X1 = 3,
+  INV_UNARY_OPCODE_X1 = 9,
+  IRET_UNARY_OPCODE_X1 = 10,
+  JALRP_UNARY_OPCODE_X1 = 11,
+  JALRP_UNARY_OPCODE_Y1 = 10,
+  JALR_UNARY_OPCODE_X1 = 12,
+  JALR_UNARY_OPCODE_Y1 = 11,
+  JAL_JUMP_OPCODE_X1 = 0,
+  JRP_UNARY_OPCODE_X1 = 13,
+  JRP_UNARY_OPCODE_Y1 = 12,
+  JR_UNARY_OPCODE_X1 = 14,
+  JR_UNARY_OPCODE_Y1 = 13,
+  JUMP_OPCODE_X1 = 4,
+  J_JUMP_OPCODE_X1 = 1,
+  LD1S_ADD_IMM8_OPCODE_X1 = 7,
+  LD1S_OPCODE_Y2 = 0,
+  LD1S_UNARY_OPCODE_X1 = 15,
+  LD1U_ADD_IMM8_OPCODE_X1 = 8,
+  LD1U_OPCODE_Y2 = 1,
+  LD1U_UNARY_OPCODE_X1 = 16,
+  LD2S_ADD_IMM8_OPCODE_X1 = 9,
+  LD2S_OPCODE_Y2 = 2,
+  LD2S_UNARY_OPCODE_X1 = 17,
+  LD2U_ADD_IMM8_OPCODE_X1 = 10,
+  LD2U_OPCODE_Y2 = 3,
+  LD2U_UNARY_OPCODE_X1 = 18,
+  LD4S_ADD_IMM8_OPCODE_X1 = 11,
+  LD4S_OPCODE_Y2 = 1,
+  LD4S_UNARY_OPCODE_X1 = 19,
+  LD4U_ADD_IMM8_OPCODE_X1 = 12,
+  LD4U_OPCODE_Y2 = 2,
+  LD4U_UNARY_OPCODE_X1 = 20,
+  LDNA_UNARY_OPCODE_X1 = 21,
+  LDNT1S_ADD_IMM8_OPCODE_X1 = 13,
+  LDNT1S_UNARY_OPCODE_X1 = 22,
+  LDNT1U_ADD_IMM8_OPCODE_X1 = 14,
+  LDNT1U_UNARY_OPCODE_X1 = 23,
+  LDNT2S_ADD_IMM8_OPCODE_X1 = 15,
+  LDNT2S_UNARY_OPCODE_X1 = 24,
+  LDNT2U_ADD_IMM8_OPCODE_X1 = 16,
+  LDNT2U_UNARY_OPCODE_X1 = 25,
+  LDNT4S_ADD_IMM8_OPCODE_X1 = 17,
+  LDNT4S_UNARY_OPCODE_X1 = 26,
+  LDNT4U_ADD_IMM8_OPCODE_X1 = 18,
+  LDNT4U_UNARY_OPCODE_X1 = 27,
+  LDNT_ADD_IMM8_OPCODE_X1 = 19,
+  LDNT_UNARY_OPCODE_X1 = 28,
+  LD_ADD_IMM8_OPCODE_X1 = 20,
+  LD_OPCODE_Y2 = 3,
+  LD_UNARY_OPCODE_X1 = 29,
+  LNK_UNARY_OPCODE_X1 = 30,
+  LNK_UNARY_OPCODE_Y1 = 14,
+  LWNA_ADD_IMM8_OPCODE_X1 = 21,
+  MFSPR_IMM8_OPCODE_X1 = 22,
+  MF_UNARY_OPCODE_X1 = 31,
+  MM_BF_OPCODE_X0 = 7,
+  MNZ_RRR_0_OPCODE_X0 = 40,
+  MNZ_RRR_0_OPCODE_X1 = 26,
+  MNZ_RRR_4_OPCODE_Y0 = 2,
+  MNZ_RRR_4_OPCODE_Y1 = 2,
+  MODE_OPCODE_YA2 = 1,
+  MODE_OPCODE_YB2 = 2,
+  MODE_OPCODE_YC2 = 3,
+  MTSPR_IMM8_OPCODE_X1 = 23,
+  MULAX_RRR_0_OPCODE_X0 = 41,
+  MULAX_RRR_3_OPCODE_Y0 = 2,
+  MULA_HS_HS_RRR_0_OPCODE_X0 = 42,
+  MULA_HS_HS_RRR_9_OPCODE_Y0 = 0,
+  MULA_HS_HU_RRR_0_OPCODE_X0 = 43,
+  MULA_HS_LS_RRR_0_OPCODE_X0 = 44,
+  MULA_HS_LU_RRR_0_OPCODE_X0 = 45,
+  MULA_HU_HU_RRR_0_OPCODE_X0 = 46,
+  MULA_HU_HU_RRR_9_OPCODE_Y0 = 1,
+  MULA_HU_LS_RRR_0_OPCODE_X0 = 47,
+  MULA_HU_LU_RRR_0_OPCODE_X0 = 48,
+  MULA_LS_LS_RRR_0_OPCODE_X0 = 49,
+  MULA_LS_LS_RRR_9_OPCODE_Y0 = 2,
+  MULA_LS_LU_RRR_0_OPCODE_X0 = 50,
+  MULA_LU_LU_RRR_0_OPCODE_X0 = 51,
+  MULA_LU_LU_RRR_9_OPCODE_Y0 = 3,
+  MULX_RRR_0_OPCODE_X0 = 52,
+  MULX_RRR_3_OPCODE_Y0 = 3,
+  MUL_HS_HS_RRR_0_OPCODE_X0 = 53,
+  MUL_HS_HS_RRR_8_OPCODE_Y0 = 0,
+  MUL_HS_HU_RRR_0_OPCODE_X0 = 54,
+  MUL_HS_LS_RRR_0_OPCODE_X0 = 55,
+  MUL_HS_LU_RRR_0_OPCODE_X0 = 56,
+  MUL_HU_HU_RRR_0_OPCODE_X0 = 57,
+  MUL_HU_HU_RRR_8_OPCODE_Y0 = 1,
+  MUL_HU_LS_RRR_0_OPCODE_X0 = 58,
+  MUL_HU_LU_RRR_0_OPCODE_X0 = 59,
+  MUL_LS_LS_RRR_0_OPCODE_X0 = 60,
+  MUL_LS_LS_RRR_8_OPCODE_Y0 = 2,
+  MUL_LS_LU_RRR_0_OPCODE_X0 = 61,
+  MUL_LU_LU_RRR_0_OPCODE_X0 = 62,
+  MUL_LU_LU_RRR_8_OPCODE_Y0 = 3,
+  MZ_RRR_0_OPCODE_X0 = 63,
+  MZ_RRR_0_OPCODE_X1 = 27,
+  MZ_RRR_4_OPCODE_Y0 = 3,
+  MZ_RRR_4_OPCODE_Y1 = 3,
+  NAP_UNARY_OPCODE_X1 = 32,
+  NOP_UNARY_OPCODE_X0 = 5,
+  NOP_UNARY_OPCODE_X1 = 33,
+  NOP_UNARY_OPCODE_Y0 = 5,
+  NOP_UNARY_OPCODE_Y1 = 15,
+  NOR_RRR_0_OPCODE_X0 = 64,
+  NOR_RRR_0_OPCODE_X1 = 28,
+  NOR_RRR_5_OPCODE_Y0 = 1,
+  NOR_RRR_5_OPCODE_Y1 = 1,
+  ORI_IMM8_OPCODE_X0 = 7,
+  ORI_IMM8_OPCODE_X1 = 24,
+  OR_RRR_0_OPCODE_X0 = 65,
+  OR_RRR_0_OPCODE_X1 = 29,
+  OR_RRR_5_OPCODE_Y0 = 2,
+  OR_RRR_5_OPCODE_Y1 = 2,
+  PCNT_UNARY_OPCODE_X0 = 6,
+  PCNT_UNARY_OPCODE_Y0 = 6,
+  REVBITS_UNARY_OPCODE_X0 = 7,
+  REVBITS_UNARY_OPCODE_Y0 = 7,
+  REVBYTES_UNARY_OPCODE_X0 = 8,
+  REVBYTES_UNARY_OPCODE_Y0 = 8,
+  ROTLI_SHIFT_OPCODE_X0 = 1,
+  ROTLI_SHIFT_OPCODE_X1 = 1,
+  ROTLI_SHIFT_OPCODE_Y0 = 0,
+  ROTLI_SHIFT_OPCODE_Y1 = 0,
+  ROTL_RRR_0_OPCODE_X0 = 66,
+  ROTL_RRR_0_OPCODE_X1 = 30,
+  ROTL_RRR_6_OPCODE_Y0 = 0,
+  ROTL_RRR_6_OPCODE_Y1 = 0,
+  RRR_0_OPCODE_X0 = 5,
+  RRR_0_OPCODE_X1 = 5,
+  RRR_0_OPCODE_Y0 = 5,
+  RRR_0_OPCODE_Y1 = 6,
+  RRR_1_OPCODE_Y0 = 6,
+  RRR_1_OPCODE_Y1 = 7,
+  RRR_2_OPCODE_Y0 = 7,
+  RRR_2_OPCODE_Y1 = 8,
+  RRR_3_OPCODE_Y0 = 8,
+  RRR_3_OPCODE_Y1 = 9,
+  RRR_4_OPCODE_Y0 = 9,
+  RRR_4_OPCODE_Y1 = 10,
+  RRR_5_OPCODE_Y0 = 10,
+  RRR_5_OPCODE_Y1 = 11,
+  RRR_6_OPCODE_Y0 = 11,
+  RRR_6_OPCODE_Y1 = 12,
+  RRR_7_OPCODE_Y0 = 12,
+  RRR_7_OPCODE_Y1 = 13,
+  RRR_8_OPCODE_Y0 = 13,
+  RRR_9_OPCODE_Y0 = 14,
+  SHIFT_OPCODE_X0 = 6,
+  SHIFT_OPCODE_X1 = 6,
+  SHIFT_OPCODE_Y0 = 15,
+  SHIFT_OPCODE_Y1 = 14,
+  SHL16INSLI_OPCODE_X0 = 7,
+  SHL16INSLI_OPCODE_X1 = 7,
+  SHL1ADDX_RRR_0_OPCODE_X0 = 67,
+  SHL1ADDX_RRR_0_OPCODE_X1 = 31,
+  SHL1ADDX_RRR_7_OPCODE_Y0 = 1,
+  SHL1ADDX_RRR_7_OPCODE_Y1 = 1,
+  SHL1ADD_RRR_0_OPCODE_X0 = 68,
+  SHL1ADD_RRR_0_OPCODE_X1 = 32,
+  SHL1ADD_RRR_1_OPCODE_Y0 = 0,
+  SHL1ADD_RRR_1_OPCODE_Y1 = 0,
+  SHL2ADDX_RRR_0_OPCODE_X0 = 69,
+  SHL2ADDX_RRR_0_OPCODE_X1 = 33,
+  SHL2ADDX_RRR_7_OPCODE_Y0 = 2,
+  SHL2ADDX_RRR_7_OPCODE_Y1 = 2,
+  SHL2ADD_RRR_0_OPCODE_X0 = 70,
+  SHL2ADD_RRR_0_OPCODE_X1 = 34,
+  SHL2ADD_RRR_1_OPCODE_Y0 = 1,
+  SHL2ADD_RRR_1_OPCODE_Y1 = 1,
+  SHL3ADDX_RRR_0_OPCODE_X0 = 71,
+  SHL3ADDX_RRR_0_OPCODE_X1 = 35,
+  SHL3ADDX_RRR_7_OPCODE_Y0 = 3,
+  SHL3ADDX_RRR_7_OPCODE_Y1 = 3,
+  SHL3ADD_RRR_0_OPCODE_X0 = 72,
+  SHL3ADD_RRR_0_OPCODE_X1 = 36,
+  SHL3ADD_RRR_1_OPCODE_Y0 = 2,
+  SHL3ADD_RRR_1_OPCODE_Y1 = 2,
+  SHLI_SHIFT_OPCODE_X0 = 2,
+  SHLI_SHIFT_OPCODE_X1 = 2,
+  SHLI_SHIFT_OPCODE_Y0 = 1,
+  SHLI_SHIFT_OPCODE_Y1 = 1,
+  SHLXI_SHIFT_OPCODE_X0 = 3,
+  SHLXI_SHIFT_OPCODE_X1 = 3,
+  SHLX_RRR_0_OPCODE_X0 = 73,
+  SHLX_RRR_0_OPCODE_X1 = 37,
+  SHL_RRR_0_OPCODE_X0 = 74,
+  SHL_RRR_0_OPCODE_X1 = 38,
+  SHL_RRR_6_OPCODE_Y0 = 1,
+  SHL_RRR_6_OPCODE_Y1 = 1,
+  SHRSI_SHIFT_OPCODE_X0 = 4,
+  SHRSI_SHIFT_OPCODE_X1 = 4,
+  SHRSI_SHIFT_OPCODE_Y0 = 2,
+  SHRSI_SHIFT_OPCODE_Y1 = 2,
+  SHRS_RRR_0_OPCODE_X0 = 75,
+  SHRS_RRR_0_OPCODE_X1 = 39,
+  SHRS_RRR_6_OPCODE_Y0 = 2,
+  SHRS_RRR_6_OPCODE_Y1 = 2,
+  SHRUI_SHIFT_OPCODE_X0 = 5,
+  SHRUI_SHIFT_OPCODE_X1 = 5,
+  SHRUI_SHIFT_OPCODE_Y0 = 3,
+  SHRUI_SHIFT_OPCODE_Y1 = 3,
+  SHRUXI_SHIFT_OPCODE_X0 = 6,
+  SHRUXI_SHIFT_OPCODE_X1 = 6,
+  SHRUX_RRR_0_OPCODE_X0 = 76,
+  SHRUX_RRR_0_OPCODE_X1 = 40,
+  SHRU_RRR_0_OPCODE_X0 = 77,
+  SHRU_RRR_0_OPCODE_X1 = 41,
+  SHRU_RRR_6_OPCODE_Y0 = 3,
+  SHRU_RRR_6_OPCODE_Y1 = 3,
+  SHUFFLEBYTES_RRR_0_OPCODE_X0 = 78,
+  ST1_ADD_IMM8_OPCODE_X1 = 25,
+  ST1_OPCODE_Y2 = 0,
+  ST1_RRR_0_OPCODE_X1 = 42,
+  ST2_ADD_IMM8_OPCODE_X1 = 26,
+  ST2_OPCODE_Y2 = 1,
+  ST2_RRR_0_OPCODE_X1 = 43,
+  ST4_ADD_IMM8_OPCODE_X1 = 27,
+  ST4_OPCODE_Y2 = 2,
+  ST4_RRR_0_OPCODE_X1 = 44,
+  STNT1_ADD_IMM8_OPCODE_X1 = 28,
+  STNT1_RRR_0_OPCODE_X1 = 45,
+  STNT2_ADD_IMM8_OPCODE_X1 = 29,
+  STNT2_RRR_0_OPCODE_X1 = 46,
+  STNT4_ADD_IMM8_OPCODE_X1 = 30,
+  STNT4_RRR_0_OPCODE_X1 = 47,
+  STNT_ADD_IMM8_OPCODE_X1 = 31,
+  STNT_RRR_0_OPCODE_X1 = 48,
+  ST_ADD_IMM8_OPCODE_X1 = 32,
+  ST_OPCODE_Y2 = 3,
+  ST_RRR_0_OPCODE_X1 = 49,
+  SUBXSC_RRR_0_OPCODE_X0 = 79,
+  SUBXSC_RRR_0_OPCODE_X1 = 50,
+  SUBX_RRR_0_OPCODE_X0 = 80,
+  SUBX_RRR_0_OPCODE_X1 = 51,
+  SUBX_RRR_0_OPCODE_Y0 = 2,
+  SUBX_RRR_0_OPCODE_Y1 = 2,
+  SUB_RRR_0_OPCODE_X0 = 81,
+  SUB_RRR_0_OPCODE_X1 = 52,
+  SUB_RRR_0_OPCODE_Y0 = 3,
+  SUB_RRR_0_OPCODE_Y1 = 3,
+  SWINT0_UNARY_OPCODE_X1 = 34,
+  SWINT1_UNARY_OPCODE_X1 = 35,
+  SWINT2_UNARY_OPCODE_X1 = 36,
+  SWINT3_UNARY_OPCODE_X1 = 37,
+  TBLIDXB0_UNARY_OPCODE_X0 = 9,
+  TBLIDXB0_UNARY_OPCODE_Y0 = 9,
+  TBLIDXB1_UNARY_OPCODE_X0 = 10,
+  TBLIDXB1_UNARY_OPCODE_Y0 = 10,
+  TBLIDXB2_UNARY_OPCODE_X0 = 11,
+  TBLIDXB2_UNARY_OPCODE_Y0 = 11,
+  TBLIDXB3_UNARY_OPCODE_X0 = 12,
+  TBLIDXB3_UNARY_OPCODE_Y0 = 12,
+  UNARY_RRR_0_OPCODE_X0 = 82,
+  UNARY_RRR_0_OPCODE_X1 = 53,
+  UNARY_RRR_1_OPCODE_Y0 = 3,
+  UNARY_RRR_1_OPCODE_Y1 = 3,
+  V1ADDI_IMM8_OPCODE_X0 = 8,
+  V1ADDI_IMM8_OPCODE_X1 = 33,
+  V1ADDUC_RRR_0_OPCODE_X0 = 83,
+  V1ADDUC_RRR_0_OPCODE_X1 = 54,
+  V1ADD_RRR_0_OPCODE_X0 = 84,
+  V1ADD_RRR_0_OPCODE_X1 = 55,
+  V1ADIFFU_RRR_0_OPCODE_X0 = 85,
+  V1AVGU_RRR_0_OPCODE_X0 = 86,
+  V1CMPEQI_IMM8_OPCODE_X0 = 9,
+  V1CMPEQI_IMM8_OPCODE_X1 = 34,
+  V1CMPEQ_RRR_0_OPCODE_X0 = 87,
+  V1CMPEQ_RRR_0_OPCODE_X1 = 56,
+  V1CMPLES_RRR_0_OPCODE_X0 = 88,
+  V1CMPLES_RRR_0_OPCODE_X1 = 57,
+  V1CMPLEU_RRR_0_OPCODE_X0 = 89,
+  V1CMPLEU_RRR_0_OPCODE_X1 = 58,
+  V1CMPLTSI_IMM8_OPCODE_X0 = 10,
+  V1CMPLTSI_IMM8_OPCODE_X1 = 35,
+  V1CMPLTS_RRR_0_OPCODE_X0 = 90,
+  V1CMPLTS_RRR_0_OPCODE_X1 = 59,
+  V1CMPLTUI_IMM8_OPCODE_X0 = 11,
+  V1CMPLTUI_IMM8_OPCODE_X1 = 36,
+  V1CMPLTU_RRR_0_OPCODE_X0 = 91,
+  V1CMPLTU_RRR_0_OPCODE_X1 = 60,
+  V1CMPNE_RRR_0_OPCODE_X0 = 92,
+  V1CMPNE_RRR_0_OPCODE_X1 = 61,
+  V1DDOTPUA_RRR_0_OPCODE_X0 = 161,
+  V1DDOTPUSA_RRR_0_OPCODE_X0 = 93,
+  V1DDOTPUS_RRR_0_OPCODE_X0 = 94,
+  V1DDOTPU_RRR_0_OPCODE_X0 = 162,
+  V1DOTPA_RRR_0_OPCODE_X0 = 95,
+  V1DOTPUA_RRR_0_OPCODE_X0 = 163,
+  V1DOTPUSA_RRR_0_OPCODE_X0 = 96,
+  V1DOTPUS_RRR_0_OPCODE_X0 = 97,
+  V1DOTPU_RRR_0_OPCODE_X0 = 164,
+  V1DOTP_RRR_0_OPCODE_X0 = 98,
+  V1INT_H_RRR_0_OPCODE_X0 = 99,
+  V1INT_H_RRR_0_OPCODE_X1 = 62,
+  V1INT_L_RRR_0_OPCODE_X0 = 100,
+  V1INT_L_RRR_0_OPCODE_X1 = 63,
+  V1MAXUI_IMM8_OPCODE_X0 = 12,
+  V1MAXUI_IMM8_OPCODE_X1 = 37,
+  V1MAXU_RRR_0_OPCODE_X0 = 101,
+  V1MAXU_RRR_0_OPCODE_X1 = 64,
+  V1MINUI_IMM8_OPCODE_X0 = 13,
+  V1MINUI_IMM8_OPCODE_X1 = 38,
+  V1MINU_RRR_0_OPCODE_X0 = 102,
+  V1MINU_RRR_0_OPCODE_X1 = 65,
+  V1MNZ_RRR_0_OPCODE_X0 = 103,
+  V1MNZ_RRR_0_OPCODE_X1 = 66,
+  V1MULTU_RRR_0_OPCODE_X0 = 104,
+  V1MULUS_RRR_0_OPCODE_X0 = 105,
+  V1MULU_RRR_0_OPCODE_X0 = 106,
+  V1MZ_RRR_0_OPCODE_X0 = 107,
+  V1MZ_RRR_0_OPCODE_X1 = 67,
+  V1SADAU_RRR_0_OPCODE_X0 = 108,
+  V1SADU_RRR_0_OPCODE_X0 = 109,
+  V1SHLI_SHIFT_OPCODE_X0 = 7,
+  V1SHLI_SHIFT_OPCODE_X1 = 7,
+  V1SHL_RRR_0_OPCODE_X0 = 110,
+  V1SHL_RRR_0_OPCODE_X1 = 68,
+  V1SHRSI_SHIFT_OPCODE_X0 = 8,
+  V1SHRSI_SHIFT_OPCODE_X1 = 8,
+  V1SHRS_RRR_0_OPCODE_X0 = 111,
+  V1SHRS_RRR_0_OPCODE_X1 = 69,
+  V1SHRUI_SHIFT_OPCODE_X0 = 9,
+  V1SHRUI_SHIFT_OPCODE_X1 = 9,
+  V1SHRU_RRR_0_OPCODE_X0 = 112,
+  V1SHRU_RRR_0_OPCODE_X1 = 70,
+  V1SUBUC_RRR_0_OPCODE_X0 = 113,
+  V1SUBUC_RRR_0_OPCODE_X1 = 71,
+  V1SUB_RRR_0_OPCODE_X0 = 114,
+  V1SUB_RRR_0_OPCODE_X1 = 72,
+  V2ADDI_IMM8_OPCODE_X0 = 14,
+  V2ADDI_IMM8_OPCODE_X1 = 39,
+  V2ADDSC_RRR_0_OPCODE_X0 = 115,
+  V2ADDSC_RRR_0_OPCODE_X1 = 73,
+  V2ADD_RRR_0_OPCODE_X0 = 116,
+  V2ADD_RRR_0_OPCODE_X1 = 74,
+  V2ADIFFS_RRR_0_OPCODE_X0 = 117,
+  V2AVGS_RRR_0_OPCODE_X0 = 118,
+  V2CMPEQI_IMM8_OPCODE_X0 = 15,
+  V2CMPEQI_IMM8_OPCODE_X1 = 40,
+  V2CMPEQ_RRR_0_OPCODE_X0 = 119,
+  V2CMPEQ_RRR_0_OPCODE_X1 = 75,
+  V2CMPLES_RRR_0_OPCODE_X0 = 120,
+  V2CMPLES_RRR_0_OPCODE_X1 = 76,
+  V2CMPLEU_RRR_0_OPCODE_X0 = 121,
+  V2CMPLEU_RRR_0_OPCODE_X1 = 77,
+  V2CMPLTSI_IMM8_OPCODE_X0 = 16,
+  V2CMPLTSI_IMM8_OPCODE_X1 = 41,
+  V2CMPLTS_RRR_0_OPCODE_X0 = 122,
+  V2CMPLTS_RRR_0_OPCODE_X1 = 78,
+  V2CMPLTUI_IMM8_OPCODE_X0 = 17,
+  V2CMPLTUI_IMM8_OPCODE_X1 = 42,
+  V2CMPLTU_RRR_0_OPCODE_X0 = 123,
+  V2CMPLTU_RRR_0_OPCODE_X1 = 79,
+  V2CMPNE_RRR_0_OPCODE_X0 = 124,
+  V2CMPNE_RRR_0_OPCODE_X1 = 80,
+  V2DOTPA_RRR_0_OPCODE_X0 = 125,
+  V2DOTP_RRR_0_OPCODE_X0 = 126,
+  V2INT_H_RRR_0_OPCODE_X0 = 127,
+  V2INT_H_RRR_0_OPCODE_X1 = 81,
+  V2INT_L_RRR_0_OPCODE_X0 = 128,
+  V2INT_L_RRR_0_OPCODE_X1 = 82,
+  V2MAXSI_IMM8_OPCODE_X0 = 18,
+  V2MAXSI_IMM8_OPCODE_X1 = 43,
+  V2MAXS_RRR_0_OPCODE_X0 = 129,
+  V2MAXS_RRR_0_OPCODE_X1 = 83,
+  V2MINSI_IMM8_OPCODE_X0 = 19,
+  V2MINSI_IMM8_OPCODE_X1 = 44,
+  V2MINS_RRR_0_OPCODE_X0 = 130,
+  V2MINS_RRR_0_OPCODE_X1 = 84,
+  V2MNZ_RRR_0_OPCODE_X0 = 131,
+  V2MNZ_RRR_0_OPCODE_X1 = 85,
+  V2MULFSC_RRR_0_OPCODE_X0 = 132,
+  V2MULS_RRR_0_OPCODE_X0 = 133,
+  V2MULTS_RRR_0_OPCODE_X0 = 134,
+  V2MZ_RRR_0_OPCODE_X0 = 135,
+  V2MZ_RRR_0_OPCODE_X1 = 86,
+  V2PACKH_RRR_0_OPCODE_X0 = 136,
+  V2PACKH_RRR_0_OPCODE_X1 = 87,
+  V2PACKL_RRR_0_OPCODE_X0 = 137,
+  V2PACKL_RRR_0_OPCODE_X1 = 88,
+  V2PACKUC_RRR_0_OPCODE_X0 = 138,
+  V2PACKUC_RRR_0_OPCODE_X1 = 89,
+  V2SADAS_RRR_0_OPCODE_X0 = 139,
+  V2SADAU_RRR_0_OPCODE_X0 = 140,
+  V2SADS_RRR_0_OPCODE_X0 = 141,
+  V2SADU_RRR_0_OPCODE_X0 = 142,
+  V2SHLI_SHIFT_OPCODE_X0 = 10,
+  V2SHLI_SHIFT_OPCODE_X1 = 10,
+  V2SHLSC_RRR_0_OPCODE_X0 = 143,
+  V2SHLSC_RRR_0_OPCODE_X1 = 90,
+  V2SHL_RRR_0_OPCODE_X0 = 144,
+  V2SHL_RRR_0_OPCODE_X1 = 91,
+  V2SHRSI_SHIFT_OPCODE_X0 = 11,
+  V2SHRSI_SHIFT_OPCODE_X1 = 11,
+  V2SHRS_RRR_0_OPCODE_X0 = 145,
+  V2SHRS_RRR_0_OPCODE_X1 = 92,
+  V2SHRUI_SHIFT_OPCODE_X0 = 12,
+  V2SHRUI_SHIFT_OPCODE_X1 = 12,
+  V2SHRU_RRR_0_OPCODE_X0 = 146,
+  V2SHRU_RRR_0_OPCODE_X1 = 93,
+  V2SUBSC_RRR_0_OPCODE_X0 = 147,
+  V2SUBSC_RRR_0_OPCODE_X1 = 94,
+  V2SUB_RRR_0_OPCODE_X0 = 148,
+  V2SUB_RRR_0_OPCODE_X1 = 95,
+  V4ADDSC_RRR_0_OPCODE_X0 = 149,
+  V4ADDSC_RRR_0_OPCODE_X1 = 96,
+  V4ADD_RRR_0_OPCODE_X0 = 150,
+  V4ADD_RRR_0_OPCODE_X1 = 97,
+  V4INT_H_RRR_0_OPCODE_X0 = 151,
+  V4INT_H_RRR_0_OPCODE_X1 = 98,
+  V4INT_L_RRR_0_OPCODE_X0 = 152,
+  V4INT_L_RRR_0_OPCODE_X1 = 99,
+  V4PACKSC_RRR_0_OPCODE_X0 = 153,
+  V4PACKSC_RRR_0_OPCODE_X1 = 100,
+  V4SHLSC_RRR_0_OPCODE_X0 = 154,
+  V4SHLSC_RRR_0_OPCODE_X1 = 101,
+  V4SHL_RRR_0_OPCODE_X0 = 155,
+  V4SHL_RRR_0_OPCODE_X1 = 102,
+  V4SHRS_RRR_0_OPCODE_X0 = 156,
+  V4SHRS_RRR_0_OPCODE_X1 = 103,
+  V4SHRU_RRR_0_OPCODE_X0 = 157,
+  V4SHRU_RRR_0_OPCODE_X1 = 104,
+  V4SUBSC_RRR_0_OPCODE_X0 = 158,
+  V4SUBSC_RRR_0_OPCODE_X1 = 105,
+  V4SUB_RRR_0_OPCODE_X0 = 159,
+  V4SUB_RRR_0_OPCODE_X1 = 106,
+  WH64_UNARY_OPCODE_X1 = 38,
+  XORI_IMM8_OPCODE_X0 = 20,
+  XORI_IMM8_OPCODE_X1 = 45,
+  XOR_RRR_0_OPCODE_X0 = 160,
+  XOR_RRR_0_OPCODE_X1 = 107,
+  XOR_RRR_5_OPCODE_Y0 = 3,
+  XOR_RRR_5_OPCODE_Y1 = 3
+};
+
+
+#endif /* __ASSEMBLER__ */
+
+#endif /* __ARCH_OPCODE_H__ */
diff --git a/target-tilegx/translate.c b/target-tilegx/translate.c
index 9aa82a9..1fd16e4 100644
--- a/target-tilegx/translate.c
+++ b/target-tilegx/translate.c
@@ -23,19 +23,615 @@ 
 #include "disas/disas.h"
 #include "tcg-op.h"
 #include "exec/cpu_ldst.h"
+#include "opcode_tilegx.h"
+
+#define TILEGX_BUNDLE_TYPE_MASK       0xc000000000000000ULL
+#define TILEGX_BUNDLE_TYPE_Y(bundle)  ((bundle) & TILEGX_BUNDLE_TYPE_MASK)
+#define TILEGX_BUNDLE_TYPE_X(bundle)  (!TILEGX_BUNDLE_TYPE_Y(bundle))
+#define TILEGX_OPCODE_MAX_X0          164  /* include 164 */
+#define TILEGX_OPCODE_MAX_X1          107  /* include 107 */
+#define TILEGX_OPCODE_MAX_Y0           15  /* include 15 */
+#define TILEGX_OPCODE_MAX_Y1           15  /* include 15 */
+#define TILEGX_OPCODE_MAX_Y2            3  /* include 3 */
+
+#define TILEGX_EXCP_OPCODE_UNKNOWN    0x1
+#define TILEGX_EXCP_OPCODE_UNIMPLEMNT 0x2
+
+
+static TCGv_ptr cpu_env;
+static TCGv_i64 cpu_pc;
+static TCGv_i64 cpu_regs[TILEGX_R_COUNT];
+
+static const char *reg_names[] = {
+     "r0",  "r1",  "r2",  "r3",  "r4",  "r5",  "r6",  "r7",
+     "r8",  "r9", "r10", "r11", "r12", "r13", "r14", "r15",
+    "r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23",
+    "r24", "r25", "r26", "r27", "r28", "r29", "r30", "r31",
+    "r32", "r33", "r34", "r35", "r36", "r37", "r38", "r39",
+    "r40", "r41", "r42", "r43", "r44", "r45", "r46", "r47",
+    "r48", "r49", "r50", "r51",  "bp",  "tp",  "sp",  "lr"
+};
+
+/* This is the state at translation time.  */
+typedef struct DisasContext {
+    uint64_t pc;                   /* Current pc */
+    uint64_t exception;            /* Current exception, 0 means empty */
+
+    TCGv zero;                     /* For zero register */
+
+    struct {
+        unsigned char idx;         /* index */
+        TCGv val;                  /* value */
+    } *tmp_regcur,                 /* Current temporary registers */
+    tmp_regs[TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE]; /* All temporary registers */
+
+    struct {
+        TCGCond cond;              /* Branch condition */
+        TCGv dest;                 /* pc jump destination, if will jump */
+        TCGv val1;                 /* Firt value for condition comparing */
+        TCGv val2;                 /* Second value for condition comparing */
+    } jmp;                         /* Jump object, only once in each TB block */
+} DisasContext;
+
+#include "exec/gen-icount.h"
+
+typedef void (*TileGXDecodeInsn)(struct DisasContext *, tilegx_bundle_bits);
+
+/* by "grep _OPCODE_X0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
+static TileGXDecodeInsn decode_insns_x0[] = {
+    NULL, /* Nothing */
+    NULL, /* 1, ADDI_IMM8_OPCODE_X0 1, ADDLI_OPCODE_X0 1, ADDXSC_RRR_0_OPCODE_X0 1, CNTLZ_UNARY_OPCODE_X0 1, ROTLI_SHIFT_OPCODE_X0 */
+    NULL, /* 2, ADDXI_IMM8_OPCODE_X0 2, ADDXLI_OPCODE_X0 2, ADDX_RRR_0_OPCODE_X0 2, CNTTZ_UNARY_OPCODE_X0 2, SHLI_SHIFT_OPCODE_X0 */
+    NULL, /* 3, ADD_RRR_0_OPCODE_X0 3, ANDI_IMM8_OPCODE_X0 3, BF_OPCODE_X0 3, FNOP_UNARY_OPCODE_X0 3, SHLXI_SHIFT_OPCODE_X0 */
+    NULL, /* 4, AND_RRR_0_OPCODE_X0 4, BFEXTS_BF_OPCODE_X0 4, CMPEQI_IMM8_OPCODE_X0 4, FSINGLE_PACK1_UNARY_OPCODE_X0 4, IMM8_OPCODE_X0 4, SHRSI_SHIFT_OPCODE_X0 */
+    NULL, /* 5, BFEXTU_BF_OPCODE_X0 5, CMOVEQZ_RRR_0_OPCODE_X0 5, CMPLTSI_IMM8_OPCODE_X0 5, NOP_UNARY_OPCODE_X0 5, RRR_0_OPCODE_X0 5, SHRUI_SHIFT_OPCODE_X0 */
+    NULL, /* 6, BFINS_BF_OPCODE_X0 6, CMOVNEZ_RRR_0_OPCODE_X0 6, CMPLTUI_IMM8_OPCODE_X0 6, PCNT_UNARY_OPCODE_X0 6, SHIFT_OPCODE_X0 6, SHRUXI_SHIFT_OPCODE_X0 */
+    NULL, /* 7, CMPEQ_RRR_0_OPCODE_X0 7, MM_BF_OPCODE_X0 7, ORI_IMM8_OPCODE_X0 7, REVBITS_UNARY_OPCODE_X0 7, SHL16INSLI_OPCODE_X0 7, V1SHLI_SHIFT_OPCODE_X0 */
+    NULL, /* 8, CMPLES_RRR_0_OPCODE_X0 8, REVBYTES_UNARY_OPCODE_X0 8, V1ADDI_IMM8_OPCODE_X0 8, V1SHRSI_SHIFT_OPCODE_X0 */
+    NULL, /* 9, CMPLEU_RRR_0_OPCODE_X0 9, TBLIDXB0_UNARY_OPCODE_X0 9, V1CMPEQI_IMM8_OPCODE_X0 9, V1SHRUI_SHIFT_OPCODE_X0 */
+    NULL, /* 10, CMPLTS_RRR_0_OPCODE_X0 10, TBLIDXB1_UNARY_OPCODE_X0 10, V1CMPLTSI_IMM8_OPCODE_X0 10, V2SHLI_SHIFT_OPCODE_X0 */
+    NULL, /* 11, CMPLTU_RRR_0_OPCODE_X0 11, TBLIDXB2_UNARY_OPCODE_X0 11, V1CMPLTUI_IMM8_OPCODE_X0 11, V2SHRSI_SHIFT_OPCODE_X0 */
+    NULL, /* 12, CMPNE_RRR_0_OPCODE_X0 12, TBLIDXB3_UNARY_OPCODE_X0 12, V1MAXUI_IMM8_OPCODE_X0 12, V2SHRUI_SHIFT_OPCODE_X0 */
+    NULL, /* 13, CMULAF_RRR_0_OPCODE_X0 13, V1MINUI_IMM8_OPCODE_X0 */
+    NULL, /* 14, CMULA_RRR_0_OPCODE_X0 14, V2ADDI_IMM8_OPCODE_X0 */
+    NULL, /* 15, CMULFR_RRR_0_OPCODE_X0 15, V2CMPEQI_IMM8_OPCODE_X0 */
+    NULL, /* 16, CMULF_RRR_0_OPCODE_X0 16, V2CMPLTSI_IMM8_OPCODE_X0 */
+    NULL, /* 17, CMULHR_RRR_0_OPCODE_X0 17, V2CMPLTUI_IMM8_OPCODE_X0 */
+    NULL, /* 18, CMULH_RRR_0_OPCODE_X0 18, V2MAXSI_IMM8_OPCODE_X0 */
+    NULL, /* 19, CMUL_RRR_0_OPCODE_X0 19, V2MINSI_IMM8_OPCODE_X0 */
+    NULL, /* 20, CRC32_32_RRR_0_OPCODE_X0 20, XORI_IMM8_OPCODE_X0 */
+    NULL, /* 21, CRC32_8_RRR_0_OPCODE_X0 */
+    NULL, /* 22, DBLALIGN2_RRR_0_OPCODE_X0 */
+    NULL, /* 23, DBLALIGN4_RRR_0_OPCODE_X0 */
+    NULL, /* 24, DBLALIGN6_RRR_0_OPCODE_X0 */
+    NULL, /* 25, DBLALIGN_RRR_0_OPCODE_X0 */
+    NULL, /* 26, FDOUBLE_ADDSUB_RRR_0_OPCODE_X0 */
+    NULL, /* 27, FDOUBLE_ADD_FLAGS_RRR_0_OPCODE_X0 */
+    NULL, /* 28, FDOUBLE_MUL_FLAGS_RRR_0_OPCODE_X0 */
+    NULL, /* 29, FDOUBLE_PACK1_RRR_0_OPCODE_X0 */
+    NULL, /* 30, FDOUBLE_PACK2_RRR_0_OPCODE_X0 */
+    NULL, /* 31, FDOUBLE_SUB_FLAGS_RRR_0_OPCODE_X0 */
+    NULL, /* 32, FDOUBLE_UNPACK_MAX_RRR_0_OPCODE_X0 */
+    NULL, /* 33, FDOUBLE_UNPACK_MIN_RRR_0_OPCODE_X0 */
+    NULL, /* 34, FSINGLE_ADD1_RRR_0_OPCODE_X0 */
+    NULL, /* 35, FSINGLE_ADDSUB2_RRR_0_OPCODE_X0 */
+    NULL, /* 36, FSINGLE_MUL1_RRR_0_OPCODE_X0 */
+    NULL, /* 37, FSINGLE_MUL2_RRR_0_OPCODE_X0 */
+    NULL, /* 38, FSINGLE_PACK2_RRR_0_OPCODE_X0 */
+    NULL, /* 39, FSINGLE_SUB1_RRR_0_OPCODE_X0 */
+    NULL, /* 40, MNZ_RRR_0_OPCODE_X0 */
+    NULL, /* 41, MULAX_RRR_0_OPCODE_X0 */
+    NULL, /* 42, MULA_HS_HS_RRR_0_OPCODE_X0 */
+    NULL, /* 43, MULA_HS_HU_RRR_0_OPCODE_X0 */
+    NULL, /* 44, MULA_HS_LS_RRR_0_OPCODE_X0 */
+    NULL, /* 45, MULA_HS_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 46, MULA_HU_HU_RRR_0_OPCODE_X0 */
+    NULL, /* 47, MULA_HU_LS_RRR_0_OPCODE_X0 */
+    NULL, /* 48, MULA_HU_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 49, MULA_LS_LS_RRR_0_OPCODE_X0 */
+    NULL, /* 50, MULA_LS_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 51, MULA_LU_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 52, MULX_RRR_0_OPCODE_X0 */
+    NULL, /* 53, MUL_HS_HS_RRR_0_OPCODE_X0 */
+    NULL, /* 54, MUL_HS_HU_RRR_0_OPCODE_X0 */
+    NULL, /* 55, MUL_HS_LS_RRR_0_OPCODE_X0 */
+    NULL, /* 56, MUL_HS_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 57, MUL_HU_HU_RRR_0_OPCODE_X0 */
+    NULL, /* 58, MUL_HU_LS_RRR_0_OPCODE_X0 */
+    NULL, /* 59, MUL_HU_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 60, MUL_LS_LS_RRR_0_OPCODE_X0 */
+    NULL, /* 61, MUL_LS_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 62, MUL_LU_LU_RRR_0_OPCODE_X0 */
+    NULL, /* 63, MZ_RRR_0_OPCODE_X0 */
+    NULL, /* 64, NOR_RRR_0_OPCODE_X0 */
+    NULL, /* 65, OR_RRR_0_OPCODE_X0 */
+    NULL, /* 66, ROTL_RRR_0_OPCODE_X0 */
+    NULL, /* 67, SHL1ADDX_RRR_0_OPCODE_X0 */
+    NULL, /* 68, SHL1ADD_RRR_0_OPCODE_X0 */
+    NULL, /* 69, SHL2ADDX_RRR_0_OPCODE_X0 */
+    NULL, /* 70, SHL2ADD_RRR_0_OPCODE_X0 */
+    NULL, /* 71, SHL3ADDX_RRR_0_OPCODE_X0 */
+    NULL, /* 72, SHL3ADD_RRR_0_OPCODE_X0 */
+    NULL, /* 73, SHLX_RRR_0_OPCODE_X0 */
+    NULL, /* 74, SHL_RRR_0_OPCODE_X0 */
+    NULL, /* 75, SHRS_RRR_0_OPCODE_X0 */
+    NULL, /* 76, SHRUX_RRR_0_OPCODE_X0 */
+    NULL, /* 77, SHRU_RRR_0_OPCODE_X0 */
+    NULL, /* 78, SHUFFLEBYTES_RRR_0_OPCODE_X0 */
+    NULL, /* 79, SUBXSC_RRR_0_OPCODE_X0 */
+    NULL, /* 80, SUBX_RRR_0_OPCODE_X0 */
+    NULL, /* 81, SUB_RRR_0_OPCODE_X0 */
+    NULL, /* 82, UNARY_RRR_0_OPCODE_X0 */
+    NULL, /* 83, V1ADDUC_RRR_0_OPCODE_X0 */
+    NULL, /* 84, V1ADD_RRR_0_OPCODE_X0 */
+    NULL, /* 85, V1ADIFFU_RRR_0_OPCODE_X0 */
+    NULL, /* 86, V1AVGU_RRR_0_OPCODE_X0 */
+    NULL, /* 87, V1CMPEQ_RRR_0_OPCODE_X0 */
+    NULL, /* 88, V1CMPLES_RRR_0_OPCODE_X0 */
+    NULL, /* 89, V1CMPLEU_RRR_0_OPCODE_X0 */
+    NULL, /* 90, V1CMPLTS_RRR_0_OPCODE_X0 */
+    NULL, /* 91, V1CMPLTU_RRR_0_OPCODE_X0 */
+    NULL, /* 92, V1CMPNE_RRR_0_OPCODE_X0 */
+    NULL, /* 93, V1DDOTPUSA_RRR_0_OPCODE_X0 */
+    NULL, /* 94, V1DDOTPUS_RRR_0_OPCODE_X0 */
+    NULL, /* 95, V1DOTPA_RRR_0_OPCODE_X0 */
+    NULL, /* 96, V1DOTPUSA_RRR_0_OPCODE_X0 */
+    NULL, /* 97, V1DOTPUS_RRR_0_OPCODE_X0 */
+    NULL, /* 98, V1DOTP_RRR_0_OPCODE_X0 */
+    NULL, /* 99, V1INT_H_RRR_0_OPCODE_X0 */
+    NULL, /* 100, V1INT_L_RRR_0_OPCODE_X0 */
+    NULL, /* 101, V1MAXU_RRR_0_OPCODE_X0 */
+    NULL, /* 102, V1MINU_RRR_0_OPCODE_X0 */
+    NULL, /* 103, V1MNZ_RRR_0_OPCODE_X0 */
+    NULL, /* 104, V1MULTU_RRR_0_OPCODE_X0 */
+    NULL, /* 105, V1MULUS_RRR_0_OPCODE_X0 */
+    NULL, /* 106, V1MULU_RRR_0_OPCODE_X0 */
+    NULL, /* 107, V1MZ_RRR_0_OPCODE_X0 */
+    NULL, /* 108, V1SADAU_RRR_0_OPCODE_X0 */
+    NULL, /* 109, V1SADU_RRR_0_OPCODE_X0 */
+    NULL, /* 110, V1SHL_RRR_0_OPCODE_X0 */
+    NULL, /* 111, V1SHRS_RRR_0_OPCODE_X0 */
+    NULL, /* 112, V1SHRU_RRR_0_OPCODE_X0 */
+    NULL, /* 113, V1SUBUC_RRR_0_OPCODE_X0 */
+    NULL, /* 114, V1SUB_RRR_0_OPCODE_X0 */
+    NULL, /* 115, V2ADDSC_RRR_0_OPCODE_X0 */
+    NULL, /* 116, V2ADD_RRR_0_OPCODE_X0 */
+    NULL, /* 117, V2ADIFFS_RRR_0_OPCODE_X0 */
+    NULL, /* 118, V2AVGS_RRR_0_OPCODE_X0 */
+    NULL, /* 119, V2CMPEQ_RRR_0_OPCODE_X0 */
+    NULL, /* 120, V2CMPLES_RRR_0_OPCODE_X0 */
+    NULL, /* 121, V2CMPLEU_RRR_0_OPCODE_X0 */
+    NULL, /* 122, V2CMPLTS_RRR_0_OPCODE_X0 */
+    NULL, /* 123, V2CMPLTU_RRR_0_OPCODE_X0 */
+    NULL, /* 124, V2CMPNE_RRR_0_OPCODE_X0 */
+    NULL, /* 125, V2DOTPA_RRR_0_OPCODE_X0 */
+    NULL, /* 126, V2DOTP_RRR_0_OPCODE_X0 */
+    NULL, /* 127, V2INT_H_RRR_0_OPCODE_X0 */
+    NULL, /* 128, V2INT_L_RRR_0_OPCODE_X0 */
+    NULL, /* 129, V2MAXS_RRR_0_OPCODE_X0 */
+    NULL, /* 130, V2MINS_RRR_0_OPCODE_X0 */
+    NULL, /* 131, V2MNZ_RRR_0_OPCODE_X0 */
+    NULL, /* 132, V2MULFSC_RRR_0_OPCODE_X0 */
+    NULL, /* 133, V2MULS_RRR_0_OPCODE_X0 */
+    NULL, /* 134, V2MULTS_RRR_0_OPCODE_X0 */
+    NULL, /* 135, V2MZ_RRR_0_OPCODE_X0 */
+    NULL, /* 136, V2PACKH_RRR_0_OPCODE_X0 */
+    NULL, /* 137, V2PACKL_RRR_0_OPCODE_X0 */
+    NULL, /* 138, V2PACKUC_RRR_0_OPCODE_X0 */
+    NULL, /* 139, V2SADAS_RRR_0_OPCODE_X0 */
+    NULL, /* 140, V2SADAU_RRR_0_OPCODE_X0 */
+    NULL, /* 141, V2SADS_RRR_0_OPCODE_X0 */
+    NULL, /* 142, V2SADU_RRR_0_OPCODE_X0 */
+    NULL, /* 143, V2SHLSC_RRR_0_OPCODE_X0 */
+    NULL, /* 144, V2SHL_RRR_0_OPCODE_X0 */
+    NULL, /* 145, V2SHRS_RRR_0_OPCODE_X0 */
+    NULL, /* 146, V2SHRU_RRR_0_OPCODE_X0 */
+    NULL, /* 147, V2SUBSC_RRR_0_OPCODE_X0 */
+    NULL, /* 148, V2SUB_RRR_0_OPCODE_X0 */
+    NULL, /* 149, V4ADDSC_RRR_0_OPCODE_X0 */
+    NULL, /* 150, V4ADD_RRR_0_OPCODE_X0 */
+    NULL, /* 151, V4INT_H_RRR_0_OPCODE_X0 */
+    NULL, /* 152, V4INT_L_RRR_0_OPCODE_X0 */
+    NULL, /* 153, V4PACKSC_RRR_0_OPCODE_X0 */
+    NULL, /* 154, V4SHLSC_RRR_0_OPCODE_X0 */
+    NULL, /* 155, V4SHL_RRR_0_OPCODE_X0 */
+    NULL, /* 156, V4SHRS_RRR_0_OPCODE_X0 */
+    NULL, /* 157, V4SHRU_RRR_0_OPCODE_X0 */
+    NULL, /* 158, V4SUBSC_RRR_0_OPCODE_X0 */
+    NULL, /* 159, V4SUB_RRR_0_OPCODE_X0 */
+    NULL, /* 160, XOR_RRR_0_OPCODE_X0 */
+    NULL, /* 161, V1DDOTPUA_RRR_0_OPCODE_X0 */
+    NULL, /* 162, V1DDOTPU_RRR_0_OPCODE_X0 */
+    NULL, /* 163, V1DOTPUA_RRR_0_OPCODE_X0 */
+    NULL  /* 164, V1DOTPU_RRR_0_OPCODE_X0 */
+};
+
+/* by "grep _OPCODE_X1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
+static TileGXDecodeInsn decode_insns_x1[] = {
+    NULL, /* 0, ADDLI_OPCODE_X1 0, JAL_JUMP_OPCODE_X1 */
+    NULL, /* 1, ADDI_IMM8_OPCODE_X1 1, ADDXLI_OPCODE_X1 1, ADDXSC_RRR_0_OPCODE_X1 1, DRAIN_UNARY_OPCODE_X1 1, J_JUMP_OPCODE_X1 1, ROTLI_SHIFT_OPCODE_X1 */
+    NULL, /* 2, ADDXI_IMM8_OPCODE_X1 2, ADDX_RRR_0_OPCODE_X1 2, BRANCH_OPCODE_X1 2, DTLBPR_UNARY_OPCODE_X1 2, SHLI_SHIFT_OPCODE_X1 */
+    NULL, /* 3, ADD_RRR_0_OPCODE_X1 3, ANDI_IMM8_OPCODE_X1 3, FINV_UNARY_OPCODE_X1 3, IMM8_OPCODE_X1 3, SHLXI_SHIFT_OPCODE_X1 */
+    NULL, /* 4, AND_RRR_0_OPCODE_X1 4, CMPEQI_IMM8_OPCODE_X1 4, FLUSHWB_UNARY_OPCODE_X1 4, JUMP_OPCODE_X1 4, SHRSI_SHIFT_OPCODE_X1 */
+    NULL, /* 5, CMPEQ_RRR_0_OPCODE_X1 5, CMPLTSI_IMM8_OPCODE_X1 5, FLUSH_UNARY_OPCODE_X1 5, RRR_0_OPCODE_X1 5, SHRUI_SHIFT_OPCODE_X1 */
+    NULL, /* 6, CMPEXCH4_RRR_0_OPCODE_X1 6, CMPLTUI_IMM8_OPCODE_X1 6, FNOP_UNARY_OPCODE_X1 6, SHIFT_OPCODE_X1 6, SHRUXI_SHIFT_OPCODE_X1 */
+    NULL, /* 7, CMPEXCH_RRR_0_OPCODE_X1 7, ICOH_UNARY_OPCODE_X1 7, LD1S_ADD_IMM8_OPCODE_X1 7, SHL16INSLI_OPCODE_X1 7, V1SHLI_SHIFT_OPCODE_X1 */
+    NULL, /* 8, CMPLES_RRR_0_OPCODE_X1 8, ILL_UNARY_OPCODE_X1 8, LD1U_ADD_IMM8_OPCODE_X1 8, V1SHRSI_SHIFT_OPCODE_X1 */
+    NULL, /* 9, CMPLEU_RRR_0_OPCODE_X1 9, INV_UNARY_OPCODE_X1 9, LD2S_ADD_IMM8_OPCODE_X1 9, V1SHRUI_SHIFT_OPCODE_X1 */
+    NULL, /* 10, CMPLTS_RRR_0_OPCODE_X1 10, IRET_UNARY_OPCODE_X1 10, LD2U_ADD_IMM8_OPCODE_X1 10, V2SHLI_SHIFT_OPCODE_X1 */
+    NULL, /* 11, CMPLTU_RRR_0_OPCODE_X1 11, JALRP_UNARY_OPCODE_X1 11, LD4S_ADD_IMM8_OPCODE_X1 11, V2SHRSI_SHIFT_OPCODE_X1 */
+    NULL, /* 12, CMPNE_RRR_0_OPCODE_X1 12, JALR_UNARY_OPCODE_X1 12, LD4U_ADD_IMM8_OPCODE_X1 12, V2SHRUI_SHIFT_OPCODE_X1 */
+    NULL, /* 13, DBLALIGN2_RRR_0_OPCODE_X1 13, JRP_UNARY_OPCODE_X1 13, LDNT1S_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 14, DBLALIGN4_RRR_0_OPCODE_X1 14, JR_UNARY_OPCODE_X1 14, LDNT1U_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 15, DBLALIGN6_RRR_0_OPCODE_X1 15, LD1S_UNARY_OPCODE_X1 15, LDNT2S_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 16, BEQZT_BRANCH_OPCODE_X1 16, EXCH4_RRR_0_OPCODE_X1 16, LD1U_UNARY_OPCODE_X1 16, LDNT2U_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 17, BEQZ_BRANCH_OPCODE_X1 17, EXCH_RRR_0_OPCODE_X1 17, LD2S_UNARY_OPCODE_X1 17, LDNT4S_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 18, BGEZT_BRANCH_OPCODE_X1 18, FETCHADD4_RRR_0_OPCODE_X1 18, LD2U_UNARY_OPCODE_X1 18, LDNT4U_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 19, BGEZ_BRANCH_OPCODE_X1 19, FETCHADDGEZ4_RRR_0_OPCODE_X1 19, LD4S_UNARY_OPCODE_X1 19, LDNT_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 20, BGTZT_BRANCH_OPCODE_X1 20, FETCHADDGEZ_RRR_0_OPCODE_X1 20, LD4U_UNARY_OPCODE_X1 20, LD_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 21, BGTZ_BRANCH_OPCODE_X1 21, FETCHADD_RRR_0_OPCODE_X1 21, LDNA_UNARY_OPCODE_X1 21, LWNA_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 22, BLBCT_BRANCH_OPCODE_X1 22, FETCHAND4_RRR_0_OPCODE_X1 22, LDNT1S_UNARY_OPCODE_X1 22, MFSPR_IMM8_OPCODE_X1 */
+    NULL, /* 23, BLBC_BRANCH_OPCODE_X1 23, FETCHAND_RRR_0_OPCODE_X1 23, LDNT1U_UNARY_OPCODE_X1 23, MTSPR_IMM8_OPCODE_X1 */
+    NULL, /* 24, BLBST_BRANCH_OPCODE_X1 24, FETCHOR4_RRR_0_OPCODE_X1 24, LDNT2S_UNARY_OPCODE_X1 24, ORI_IMM8_OPCODE_X1 */
+    NULL, /* 25, BLBS_BRANCH_OPCODE_X1 25, FETCHOR_RRR_0_OPCODE_X1 25, LDNT2U_UNARY_OPCODE_X1 25, ST1_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 26, BLEZT_BRANCH_OPCODE_X1 26, LDNT4S_UNARY_OPCODE_X1 26, MNZ_RRR_0_OPCODE_X1 26, ST2_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 27, BLEZ_BRANCH_OPCODE_X1 27, LDNT4U_UNARY_OPCODE_X1 27, MZ_RRR_0_OPCODE_X1 27, ST4_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 28, BLTZT_BRANCH_OPCODE_X1 28, LDNT_UNARY_OPCODE_X1 28, NOR_RRR_0_OPCODE_X1 28, STNT1_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 29, BLTZ_BRANCH_OPCODE_X1 29, LD_UNARY_OPCODE_X1 29, OR_RRR_0_OPCODE_X1 29, STNT2_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 30, BNEZT_BRANCH_OPCODE_X1 30, LNK_UNARY_OPCODE_X1 30, ROTL_RRR_0_OPCODE_X1 30, STNT4_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 31, BNEZ_BRANCH_OPCODE_X1 31, MF_UNARY_OPCODE_X1 31, SHL1ADDX_RRR_0_OPCODE_X1 31, STNT_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 32, NAP_UNARY_OPCODE_X1 32, SHL1ADD_RRR_0_OPCODE_X1 32, ST_ADD_IMM8_OPCODE_X1 */
+    NULL, /* 33, NOP_UNARY_OPCODE_X1 33, SHL2ADDX_RRR_0_OPCODE_X1 33, V1ADDI_IMM8_OPCODE_X1 */
+    NULL, /* 34, SHL2ADD_RRR_0_OPCODE_X1 34, SWINT0_UNARY_OPCODE_X1 34, V1CMPEQI_IMM8_OPCODE_X1 */
+    NULL, /* 35, SHL3ADDX_RRR_0_OPCODE_X1 35, SWINT1_UNARY_OPCODE_X1 35, V1CMPLTSI_IMM8_OPCODE_X1 */
+    NULL, /* 36, SHL3ADD_RRR_0_OPCODE_X1 36, SWINT2_UNARY_OPCODE_X1 36, V1CMPLTUI_IMM8_OPCODE_X1 */
+    NULL, /* 37, SHLX_RRR_0_OPCODE_X1 37, SWINT3_UNARY_OPCODE_X1 37, V1MAXUI_IMM8_OPCODE_X1 */
+    NULL, /* 38, SHL_RRR_0_OPCODE_X1 38, V1MINUI_IMM8_OPCODE_X1 38, WH64_UNARY_OPCODE_X1 */
+    NULL, /* 39, SHRS_RRR_0_OPCODE_X1 39, V2ADDI_IMM8_OPCODE_X1 */
+    NULL, /* 40, SHRUX_RRR_0_OPCODE_X1 40, V2CMPEQI_IMM8_OPCODE_X1 */
+    NULL, /* 41, SHRU_RRR_0_OPCODE_X1 41, V2CMPLTSI_IMM8_OPCODE_X1 */
+    NULL, /* 42, ST1_RRR_0_OPCODE_X1 42, V2CMPLTUI_IMM8_OPCODE_X1 */
+    NULL, /* 43, ST2_RRR_0_OPCODE_X1 43, V2MAXSI_IMM8_OPCODE_X1 */
+    NULL, /* 44, ST4_RRR_0_OPCODE_X1 44, V2MINSI_IMM8_OPCODE_X1 */
+    NULL, /* 45, STNT1_RRR_0_OPCODE_X1 45, XORI_IMM8_OPCODE_X1 */
+    NULL, /* 46, STNT2_RRR_0_OPCODE_X1 */
+    NULL, /* 47, STNT4_RRR_0_OPCODE_X1 */
+    NULL, /* 48, STNT_RRR_0_OPCODE_X1 */
+    NULL, /* 49, ST_RRR_0_OPCODE_X1 */
+    NULL, /* 50, SUBXSC_RRR_0_OPCODE_X1 */
+    NULL, /* 51, SUBX_RRR_0_OPCODE_X1 */
+    NULL, /* 52, SUB_RRR_0_OPCODE_X1 */
+    NULL, /* 53, UNARY_RRR_0_OPCODE_X1 */
+    NULL, /* 54, V1ADDUC_RRR_0_OPCODE_X1 */
+    NULL, /* 55, V1ADD_RRR_0_OPCODE_X1 */
+    NULL, /* 56, V1CMPEQ_RRR_0_OPCODE_X1 */
+    NULL, /* 57, V1CMPLES_RRR_0_OPCODE_X1 */
+    NULL, /* 58, V1CMPLEU_RRR_0_OPCODE_X1 */
+    NULL, /* 59, V1CMPLTS_RRR_0_OPCODE_X1 */
+    NULL, /* 60, V1CMPLTU_RRR_0_OPCODE_X1 */
+    NULL, /* 61, V1CMPNE_RRR_0_OPCODE_X1 */
+    NULL, /* 62, V1INT_H_RRR_0_OPCODE_X1 */
+    NULL, /* 63, V1INT_L_RRR_0_OPCODE_X1 */
+    NULL, /* 64, V1MAXU_RRR_0_OPCODE_X1 */
+    NULL, /* 65, V1MINU_RRR_0_OPCODE_X1 */
+    NULL, /* 66, V1MNZ_RRR_0_OPCODE_X1 */
+    NULL, /* 67, V1MZ_RRR_0_OPCODE_X1 */
+    NULL, /* 68, V1SHL_RRR_0_OPCODE_X1 */
+    NULL, /* 69, V1SHRS_RRR_0_OPCODE_X1 */
+    NULL, /* 70, V1SHRU_RRR_0_OPCODE_X1 */
+    NULL, /* 71, V1SUBUC_RRR_0_OPCODE_X1 */
+    NULL, /* 72, V1SUB_RRR_0_OPCODE_X1 */
+    NULL, /* 73, V2ADDSC_RRR_0_OPCODE_X1 */
+    NULL, /* 74, V2ADD_RRR_0_OPCODE_X1 */
+    NULL, /* 75, V2CMPEQ_RRR_0_OPCODE_X1 */
+    NULL, /* 76, V2CMPLES_RRR_0_OPCODE_X1 */
+    NULL, /* 77, V2CMPLEU_RRR_0_OPCODE_X1 */
+    NULL, /* 78, V2CMPLTS_RRR_0_OPCODE_X1 */
+    NULL, /* 79, V2CMPLTU_RRR_0_OPCODE_X1 */
+    NULL, /* 80, V2CMPNE_RRR_0_OPCODE_X1 */
+    NULL, /* 81, V2INT_H_RRR_0_OPCODE_X1 */
+    NULL, /* 82, V2INT_L_RRR_0_OPCODE_X1 */
+    NULL, /* 83, V2MAXS_RRR_0_OPCODE_X1 */
+    NULL, /* 84, V2MINS_RRR_0_OPCODE_X1 */
+    NULL, /* 85, V2MNZ_RRR_0_OPCODE_X1 */
+    NULL, /* 86, V2MZ_RRR_0_OPCODE_X1 */
+    NULL, /* 87, V2PACKH_RRR_0_OPCODE_X1 */
+    NULL, /* 88, V2PACKL_RRR_0_OPCODE_X1 */
+    NULL, /* 89, V2PACKUC_RRR_0_OPCODE_X1 */
+    NULL, /* 90, V2SHLSC_RRR_0_OPCODE_X1 */
+    NULL, /* 91, V2SHL_RRR_0_OPCODE_X1 */
+    NULL, /* 92, V2SHRS_RRR_0_OPCODE_X1 */
+    NULL, /* 93, V2SHRU_RRR_0_OPCODE_X1 */
+    NULL, /* 94, V2SUBSC_RRR_0_OPCODE_X1 */
+    NULL, /* 95, V2SUB_RRR_0_OPCODE_X1 */
+    NULL, /* 96, V4ADDSC_RRR_0_OPCODE_X1 */
+    NULL, /* 97, V4ADD_RRR_0_OPCODE_X1 */
+    NULL, /* 98, V4INT_H_RRR_0_OPCODE_X1 */
+    NULL, /* 99, V4INT_L_RRR_0_OPCODE_X1 */
+    NULL, /* 100, V4PACKSC_RRR_0_OPCODE_X1 */
+    NULL, /* 101, V4SHLSC_RRR_0_OPCODE_X1 */
+    NULL, /* 102, V4SHL_RRR_0_OPCODE_X1 */
+    NULL, /* 103, V4SHRS_RRR_0_OPCODE_X1 */
+    NULL, /* 104, V4SHRU_RRR_0_OPCODE_X1 */
+    NULL, /* 105, V4SUBSC_RRR_0_OPCODE_X1 */
+    NULL, /* 106, V4SUB_RRR_0_OPCODE_X1 */
+    NULL  /* 107, XOR_RRR_0_OPCODE_X1 */
+};
+
+/* by "grep _OPCODE_Y0 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
+static TileGXDecodeInsn decode_insns_y0[] = {
+    NULL, /* 0, ADDI_OPCODE_Y0 0, ADDX_RRR_0_OPCODE_Y0 0, AND_RRR_5_OPCODE_Y0 0, CMOVEQZ_RRR_4_OPCODE_Y0 0, CMPEQ_RRR_3_OPCODE_Y0 0, CMPLES_RRR_2_OPCODE_Y0 0, MULA_HS_HS_RRR_9_OPCODE_Y0 0, MUL_HS_HS_RRR_8_OPCODE_Y0 0, ROTLI_SHIFT_OPCODE_Y0 0, ROTL_RRR_6_OPCODE_Y0 0, SHL1ADD_RRR_1_OPCODE_Y0 */
+    NULL, /* 1, ADD_RRR_0_OPCODE_Y0 1, ADDXI_OPCODE_Y0 1, CMOVNEZ_RRR_4_OPCODE_Y0 1, CMPLEU_RRR_2_OPCODE_Y0 1, CMPNE_RRR_3_OPCODE_Y0 1, CNTLZ_UNARY_OPCODE_Y0 1, MULA_HU_HU_RRR_9_OPCODE_Y0 1, MUL_HU_HU_RRR_8_OPCODE_Y0 1, NOR_RRR_5_OPCODE_Y0 1, SHL1ADDX_RRR_7_OPCODE_Y0 1, SHL2ADD_RRR_1_OPCODE_Y0 1, SHLI_SHIFT_OPCODE_Y0 1, SHL_RRR_6_OPCODE_Y0 */
+    NULL, /* 2, ANDI_OPCODE_Y0 2, CMPLTS_RRR_2_OPCODE_Y0 2, CNTTZ_UNARY_OPCODE_Y0 2, MNZ_RRR_4_OPCODE_Y0 2, MULA_LS_LS_RRR_9_OPCODE_Y0 2, MULAX_RRR_3_OPCODE_Y0 2, MUL_LS_LS_RRR_8_OPCODE_Y0 2, OR_RRR_5_OPCODE_Y0 2, SHL2ADDX_RRR_7_OPCODE_Y0 2, SHL3ADD_RRR_1_OPCODE_Y0 2, SHRSI_SHIFT_OPCODE_Y0 2, SHRS_RRR_6_OPCODE_Y0 2, SUBX_RRR_0_OPCODE_Y0 */
+    NULL, /* 3, CMPEQI_OPCODE_Y0 3, CMPLTU_RRR_2_OPCODE_Y0 3, FNOP_UNARY_OPCODE_Y0 3, MULA_LU_LU_RRR_9_OPCODE_Y0 3, MUL_LU_LU_RRR_8_OPCODE_Y0 3, MULX_RRR_3_OPCODE_Y0 3, MZ_RRR_4_OPCODE_Y0 3, SHL3ADDX_RRR_7_OPCODE_Y0 3, SHRUI_SHIFT_OPCODE_Y0 3, SHRU_RRR_6_OPCODE_Y0 3, SUB_RRR_0_OPCODE_Y0 3, UNARY_RRR_1_OPCODE_Y0 3, XOR_RRR_5_OPCODE_Y0 */
+    NULL, /* 4, CMPLTSI_OPCODE_Y0 4, FSINGLE_PACK1_UNARY_OPCODE_Y0 */
+    NULL, /* 5, NOP_UNARY_OPCODE_Y0 5, RRR_0_OPCODE_Y0 */
+    NULL, /* 6, PCNT_UNARY_OPCODE_Y0 6, RRR_1_OPCODE_Y0 */
+    NULL, /* 7, REVBITS_UNARY_OPCODE_Y0 7, RRR_2_OPCODE_Y0 */
+    NULL, /* 8, REVBYTES_UNARY_OPCODE_Y0 8, RRR_3_OPCODE_Y0 */
+    NULL, /* 9, RRR_4_OPCODE_Y0 9, TBLIDXB0_UNARY_OPCODE_Y0 */
+    NULL, /* 10, RRR_5_OPCODE_Y0 10, TBLIDXB1_UNARY_OPCODE_Y0 */
+    NULL, /* 11, RRR_6_OPCODE_Y0 11, TBLIDXB2_UNARY_OPCODE_Y0 */
+    NULL, /* 12, RRR_7_OPCODE_Y0 12, TBLIDXB3_UNARY_OPCODE_Y0 */
+    NULL, /* 13, RRR_8_OPCODE_Y0 */
+    NULL, /* 14, RRR_9_OPCODE_Y0 */
+    NULL  /* 15, SHIFT_OPCODE_Y0 */
+};
+
+/* by "grep _OPCODE_Y1 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
+static TileGXDecodeInsn decode_insns_y1[] = {
+    NULL, /* 0, ADDX_SPECIAL_0_OPCODE_Y1 0, AND_RRR_5_OPCODE_Y1 0, CMPLES_RRR_2_OPCODE_Y1 0, ROTLI_SHIFT_OPCODE_Y1 0, ROTL_RRR_6_OPCODE_Y1 0, SHL1ADD_RRR_1_OPCODE_Y1 */
+    NULL, /* 1, ADDI_OPCODE_Y1 1, ADD_SPECIAL_0_OPCODE_Y1 1, CMPLEU_RRR_2_OPCODE_Y1 1, NOR_RRR_5_OPCODE_Y1 1, SHL1ADDX_RRR_7_OPCODE_Y1 1, SHL2ADD_RRR_1_OPCODE_Y1 1, SHLI_SHIFT_OPCODE_Y1 1, SHL_RRR_6_OPCODE_Y1 */
+    NULL, /* 2, ADDXI_OPCODE_Y1 2, CMPEQ_RRR_3_OPCODE_Y1 2, CMPLTS_RRR_2_OPCODE_Y1 2, MNZ_RRR_4_OPCODE_Y1 2, OR_RRR_5_OPCODE_Y1 2, SHL2ADDX_RRR_7_OPCODE_Y1 2, SHL3ADD_RRR_1_OPCODE_Y1 2, SHRSI_SHIFT_OPCODE_Y1 2, SHRS_RRR_6_OPCODE_Y1 2, SUBX_RRR_0_OPCODE_Y1 */
+    NULL, /* 3, ANDI_OPCODE_Y1 3, CMPLTU_RRR_2_OPCODE_Y1 3, CMPNE_RRR_3_OPCODE_Y1 3, MZ_RRR_4_OPCODE_Y1 3, SHL3ADDX_RRR_7_OPCODE_Y1 3, SHRUI_SHIFT_OPCODE_Y1 3, SHRU_RRR_6_OPCODE_Y1 3, SUB_RRR_0_OPCODE_Y1 3, UNARY_RRR_1_OPCODE_Y1 3 XOR_RRR_5_OPCODE_Y1 */
+    NULL, /* 4, CMPEQI_OPCODE_Y1 */
+    NULL, /* 5, CMPLTSI_OPCODE_Y1 */
+    NULL, /* 6, RRR_0_OPCODE_Y1 */
+    NULL, /* 7, RRR_1_OPCODE_Y1 */
+    NULL, /* 8, FNOP_UNARY_OPCODE_Y1 8, RRR_2_OPCODE_Y1 */
+    NULL, /* 9, ILL_UNARY_OPCODE_Y1 9, RRR_3_OPCODE_Y1 */
+    NULL, /* 10, JALRP_UNARY_OPCODE_Y1 10, RRR_4_OPCODE_Y1 */
+    NULL, /* 11, JALR_UNARY_OPCODE_Y1 11, RRR_5_OPCODE_Y1 */
+    NULL, /* 12, JRP_UNARY_OPCODE_Y1 12, RRR_6_OPCODE_Y1 */
+    NULL, /* 13, JR_UNARY_OPCODE_Y1 13, RRR_7_OPCODE_Y1 */
+    NULL, /* 14, LNK_UNARY_OPCODE_Y1 14, SHIFT_OPCODE_Y1 */
+    NULL  /* 15, NOP_UNARY_OPCODE_Y1 */
+};
+
+/* by "grep _OPCODE_Y2 opcode_tilegx.h | awk '{print $3, $1}' | sort -n" */
+static TileGXDecodeInsn decode_insns_y2[] = {
+    NULL, /* 0, LD1S_OPCODE_Y2 0, ST1_OPCODE_Y2 */
+    NULL, /* 1, LD1U_OPCODE_Y2 1, LD4S_OPCODE_Y2 1, ST2_OPCODE_Y2 */
+    NULL, /* 2, LD2S_OPCODE_Y2 2, LD4U_OPCODE_Y2 2, ST4_OPCODE_Y2 */
+    NULL  /* 3, LD2U_OPCODE_Y2 3, LD_OPCODE_Y2 3, ST_OPCODE_Y2 */
+};
+
+void tilegx_tcg_init(void)
+{
+    int i;
+
+    cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
+    cpu_pc = tcg_global_mem_new_i64(TCG_AREG0, offsetof(CPUTLGState, pc), "pc");
+    for (i = 0; i < TILEGX_R_COUNT; i++) {
+        cpu_regs[i] = tcg_global_mem_new_i64(TCG_AREG0,
+                                             offsetof(CPUTLGState, regs[i]),
+                                             reg_names[i]);
+    }
+}
+
+static void translate_x0(struct DisasContext *dc, tilegx_bundle_bits bundle)
+{
+    unsigned int opcode = get_Opcode_X0(bundle);
+
+    if (opcode > TILEGX_OPCODE_MAX_X0) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
+        qemu_log("unknown x0 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+    if (!decode_insns_x0[opcode]) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
+        qemu_log("unimplement x0 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+
+    dc->tmp_regcur = dc->tmp_regs + 0;
+    decode_insns_x0[opcode](dc, bundle);
+}
+
+static void translate_x1(struct DisasContext *dc, tilegx_bundle_bits bundle)
+{
+    unsigned int opcode = get_Opcode_X1(bundle);
+
+    if (opcode > TILEGX_OPCODE_MAX_X1) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
+        qemu_log("unknown x1 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+    if (!decode_insns_x1[opcode]) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
+        qemu_log("unimplement x1 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+
+    dc->tmp_regcur = dc->tmp_regs + 1;
+    decode_insns_x1[opcode](dc, bundle);
+}
+
+static void translate_y0(struct DisasContext *dc, tilegx_bundle_bits bundle)
+{
+    unsigned int opcode = get_Opcode_Y0(bundle);
+
+    if (opcode > TILEGX_OPCODE_MAX_Y0) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
+        qemu_log("unknown y0 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+    if (!decode_insns_y0[opcode]) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
+        qemu_log("unimplement y0 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+
+    dc->tmp_regcur = dc->tmp_regs + 0;
+    decode_insns_y0[opcode](dc, bundle);
+}
+
+static void translate_y1(struct DisasContext *dc, tilegx_bundle_bits bundle)
+{
+    unsigned int opcode = get_Opcode_Y1(bundle);
+
+    if (opcode > TILEGX_OPCODE_MAX_Y1) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
+        qemu_log("unknown y1 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+    if (!decode_insns_y1[opcode]) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
+        qemu_log("unimplement y1 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+
+    dc->tmp_regcur = dc->tmp_regs + 1;
+    decode_insns_y1[opcode](dc, bundle);
+}
+
+static void translate_y2(struct DisasContext *dc, tilegx_bundle_bits bundle)
+{
+    unsigned int opcode = get_Opcode_Y2(bundle);
+
+    if (opcode > TILEGX_OPCODE_MAX_Y2) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNKNOWN;
+        qemu_log("unknown y2 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+    if (!decode_insns_y2[opcode]) {
+        dc->exception = TILEGX_EXCP_OPCODE_UNIMPLEMNT;
+        qemu_log("unimplement y2 opcode: %u for bundle: %16.16llx\n",
+                 opcode, bundle);
+        return;
+    }
+
+    dc->tmp_regcur = dc->tmp_regs + 2;
+    decode_insns_y2[opcode](dc, bundle);
+}
+
+static void translate_one_bundle(struct DisasContext *dc, uint64_t bundle)
+{
+    int i;
+    TCGv tmp;
+
+    for (i = 0; i < TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE; i++) {
+        dc->tmp_regs[i].idx = TILEGX_R_NOREG;
+        TCGV_UNUSED_I64(dc->tmp_regs[i].val);
+    }
+
+    if (TILEGX_BUNDLE_TYPE_Y(bundle)) {
+        translate_y0(dc, bundle);
+        translate_y1(dc, bundle);
+        translate_y2(dc, bundle);
+    } else {
+        translate_x0(dc, bundle);
+        translate_x1(dc, bundle);
+    }
+
+    for (i = 0; i < TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE; i++) {
+        if (dc->tmp_regs[i].idx == TILEGX_R_NOREG) {
+            continue;
+        }
+        tcg_gen_mov_i64(cpu_regs[dc->tmp_regs[i].idx], dc->tmp_regs[i].val);
+        tcg_temp_free_i64(dc->tmp_regs[i].val);
+    }
+
+    if (dc->jmp.cond != TCG_COND_NEVER) {
+        if (dc->jmp.cond == TCG_COND_ALWAYS) {
+            tcg_gen_mov_i64(cpu_pc, dc->jmp.dest);
+        } else {
+            tmp = tcg_const_i64(dc->pc + TILEGX_BUNDLE_SIZE_IN_BYTES);
+            tcg_gen_movcond_i64(dc->jmp.cond, cpu_pc,
+                                dc->jmp.val1, dc->jmp.val2,
+                                dc->jmp.dest, tmp);
+            tcg_temp_free_i64(dc->jmp.val1);
+            tcg_temp_free_i64(dc->jmp.val2);
+            tcg_temp_free_i64(tmp);
+        }
+        tcg_temp_free_i64(dc->jmp.dest);
+        tcg_gen_exit_tb(0);
+    }
+}
 
 static inline void gen_intermediate_code_internal(TileGXCPU *cpu,
                                                   TranslationBlock *tb,
                                                   bool search_pc)
 {
-    /*
-     * FIXME: after load elf64 tilegx binary successfully, it will quit, at
-     * present, and will implement the related features next.
-     */
-    qemu_log("\nLoaded elf64 tilegx successfully\n");
-    qemu_log("reached code start position: [" TARGET_FMT_lx "] %s\n\n",
-             tb->pc, lookup_symbol(tb->pc));
-    exit(0);
+    DisasContext ctx;
+    DisasContext *dc = &ctx;
+
+    CPUTLGState *env = &cpu->env;
+    uint64_t pc_start = tb->pc;
+    uint64_t next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
+    int j, lj = -1;
+    int num_insns = 0;
+    int max_insns = tb->cflags & CF_COUNT_MASK;
+
+    dc->pc = pc_start;
+    dc->exception = 0;
+    dc->jmp.cond = TCG_COND_NEVER;
+    TCGV_UNUSED_I64(dc->jmp.dest);
+    TCGV_UNUSED_I64(dc->jmp.val1);
+    TCGV_UNUSED_I64(dc->jmp.val2);
+
+    if (!max_insns) {
+        max_insns = CF_COUNT_MASK;
+    }
+    gen_tb_start(tb);
+
+    do {
+        TCGV_UNUSED_I64(dc->zero);
+        if (search_pc) {
+            j = tcg_op_buf_count();
+            if (lj < j) {
+                lj++;
+                while (lj < j) {
+                    tcg_ctx.gen_opc_instr_start[lj++] = 0;
+                }
+            }
+            tcg_ctx.gen_opc_pc[lj] = dc->pc;
+            tcg_ctx.gen_opc_instr_start[lj] = 1;
+            tcg_ctx.gen_opc_icount[lj] = num_insns;
+        }
+        translate_one_bundle(dc, cpu_ldq_data(env, dc->pc));
+        if (dc->exception) {
+            exit(-1);
+        }
+        num_insns++;
+        dc->pc += TILEGX_BUNDLE_SIZE_IN_BYTES;
+    } while (dc->jmp.cond == TCG_COND_NEVER
+             && dc->pc <= next_page_start - TILEGX_BUNDLE_SIZE_IN_BYTES
+             && num_insns <= max_insns - TILEGX_MAX_INSTRUCTIONS_PER_BUNDLE
+             && !tcg_op_buf_full());
+
+    gen_tb_end(tb, num_insns);
+    if (search_pc) {
+        j = tcg_op_buf_count();
+        lj++;
+        while (lj <= j) {
+            tcg_ctx.gen_opc_instr_start[lj++] = 0;
+        }
+    } else {
+        tb->size = dc->pc - pc_start;
+        tb->icount = num_insns;
+    }
+
+    return;
 }
 
 void gen_intermediate_code(CPUTLGState *env, struct TranslationBlock *tb)