| # CS_ARCH_ARM, CS_MODE_THUMB+CS_MODE_V8+CS_MODE_MCLASS, None |
| 0x3f,0xee,0x09,0x2e = vcvtb.f16.f32 q1, q4 |
| 0x3f,0xfe,0x03,0x1e = vcvtt.f32.f16 q0, q1 |
| 0xb2,0xee,0xc0,0x0b = vcvtt.f64.f16 d0, s0 |
| 0xf3,0xee,0xc2,0x0b = vcvtt.f16.f64 s1, d2 |
| 0x3f,0xee,0x09,0x3e = vcvtt.f16.f32 q1, q4 |
| 0x0c,0xee,0x0c,0x3e = vqdmladhx.s8 q1, q6, q6 |
| 0x12,0xee,0x08,0x1e = vqdmladhx.s16 q0, q1, q4 |
| 0x26,0xee,0x0e,0x1e = vqdmladhx.s32 q0, q3, q7 |
| 0x02,0xee,0x02,0x0e = vqdmladh.s8 q0, q1, q1 |
| 0x14,0xee,0x04,0x0e = vqdmladh.s16 q0, q2, q2 |
| 0x2a,0xee,0x0e,0x2e = vqdmladh.s32 q1, q5, q7 |
| 0x0e,0xee,0x01,0x1e = vqrdmladhx.s8 q0, q7, q0 |
| 0x10,0xee,0x03,0x1e = vqrdmladhx.s16 q0, q0, q1 |
| 0x20,0xee,0x09,0x3e = vqrdmladhx.s32 q1, q0, q4 |
| 0x22,0xee,0x01,0x3e = vqrdmladhx.s32 q1, q1, q0 |
| 0x20,0xee,0x03,0x3e = vqrdmladhx.s32 q1, q0, q1 |
| 0x0c,0xee,0x05,0x0e = vqrdmladh.s8 q0, q6, q2 |
| 0x1a,0xee,0x09,0x2e = vqrdmladh.s16 q1, q5, q4 |
| 0x24,0xee,0x05,0x0e = vqrdmladh.s32 q0, q2, q2 |
| 0x08,0xfe,0x0e,0x3e = vqdmlsdhx.s8 q1, q4, q7 |
| 0x14,0xfe,0x0a,0x1e = vqdmlsdhx.s16 q0, q2, q5 |
| 0x28,0xfe,0x0c,0x7e = vqdmlsdhx.s32 q3, q4, q6 |
| 0x06,0xfe,0x0c,0x0e = vqdmlsdh.s8 q0, q3, q6 |
| 0x18,0xfe,0x02,0x0e = vqdmlsdh.s16 q0, q4, q1 |
| 0x2a,0xfe,0x00,0x4e = vqdmlsdh.s32 q2, q5, q0 |
| 0x06,0xfe,0x03,0x1e = vqrdmlsdhx.s8 q0, q3, q1 |
| 0x12,0xfe,0x09,0x1e = vqrdmlsdhx.s16 q0, q1, q4 |
| 0x2c,0xfe,0x07,0x3e = vqrdmlsdhx.s32 q1, q6, q3 |
| 0x06,0xfe,0x01,0x6e = vqrdmlsdh.s8 q3, q3, q0 |
| 0x1e,0xfe,0x09,0x0e = vqrdmlsdh.s16 q0, q7, q4 |
| 0x2c,0xfe,0x0f,0x0e = vqrdmlsdh.s32 q0, q6, q7 |
| 0x20,0xfe,0x0f,0x0e = vqrdmlsdh.s32 q0, q0, q7 |
| 0x2c,0xfe,0x01,0x0e = vqrdmlsdh.s32 q0, q6, q0 |
| 0x32,0xee,0x05,0x0e = vcmul.f16 q0, q1, q2, #0x5a |
| 0x34,0xee,0x0a,0xce = vcmul.f16 q6, q2, q5, #0 |
| 0x30,0xee,0x0b,0x2e = vcmul.f16 q1, q0, q5, #0x5a |
| 0x30,0xee,0x0a,0x3e = vcmul.f16 q1, q0, q5, #0xb4 |
| 0x30,0xee,0x0b,0x3e = vcmul.f16 q1, q0, q5, #0x10e |
| 0x30,0xee,0x03,0x3e = vcmul.f16 q1, q0, q1, #0x10e |
| 0x3e,0xfe,0x0a,0x2e = vcmul.f32 q1, q7, q5, #0 |
| 0x38,0xfe,0x05,0x6e = vcmul.f32 q3, q4, q2, #0x5a |
| 0x32,0xfe,0x06,0xbe = vcmul.f32 q5, q1, q3, #0xb4 |
| 0x3e,0xfe,0x09,0x1e = vcmul.f32 q0, q7, q4, #0x10e |
| 0x0d,0xee,0x00,0x4e = vmullb.s8 q2, q6, q0 |
| 0x19,0xee,0x06,0x6e = vmullb.s16 q3, q4, q3 |
| 0x2b,0xee,0x0c,0x6e = vmullb.s32 q3, q5, q6 |
| 0x0d,0xee,0x04,0x1e = vmullt.s8 q0, q6, q2 |
| 0x11,0xee,0x04,0x1e = vmullt.s16 q0, q0, q2 |
| 0x29,0xee,0x08,0x5e = vmullt.s32 q2, q4, q4 |
| 0x37,0xee,0x0e,0x4e = vmullb.p8 q2, q3, q7 |
| 0x33,0xfe,0x06,0x0e = vmullb.p16 q0, q1, q3 |
| 0x33,0xee,0x0e,0x3e = vmullt.p8 q1, q1, q7 |
| 0x3f,0xfe,0x0e,0x1e = vmullt.p16 q0, q7, q7 |
| 0x09,0xee,0x0b,0x0e = vmulh.s8 q0, q4, q5 |
| 0x1f,0xee,0x09,0x0e = vmulh.s16 q0, q7, q4 |
| 0x2f,0xee,0x09,0x0e = vmulh.s32 q0, q7, q4 |
| 0x0b,0xfe,0x05,0x6e = vmulh.u8 q3, q5, q2 |
| 0x1f,0xfe,0x09,0x4e = vmulh.u16 q2, q7, q4 |
| 0x27,0xfe,0x05,0x2e = vmulh.u32 q1, q3, q2 |
| 0x03,0xee,0x05,0x3e = vrmulh.s8 q1, q1, q2 |
| 0x13,0xee,0x05,0x3e = vrmulh.s16 q1, q1, q2 |
| 0x23,0xee,0x01,0x7e = vrmulh.s32 q3, q1, q0 |
| 0x0d,0xfe,0x01,0x3e = vrmulh.u8 q1, q6, q0 |
| 0x17,0xfe,0x0d,0x9e = vrmulh.u16 q4, q3, q6 |
| 0x25,0xfe,0x05,0x3e = vrmulh.u32 q1, q2, q2 |
| 0x33,0xee,0x03,0x0e = vqmovnb.s16 q0, q1 |
| 0x33,0xee,0x01,0x5e = vqmovnt.s16 q2, q0 |
| 0x37,0xee,0x0b,0x0e = vqmovnb.s32 q0, q5 |
| 0x37,0xee,0x03,0x1e = vqmovnt.s32 q0, q1 |
| 0x33,0xfe,0x09,0x0e = vqmovnb.u16 q0, q4 |
| 0x33,0xfe,0x0f,0x1e = vqmovnt.u16 q0, q7 |
| 0x37,0xfe,0x09,0x0e = vqmovnb.u32 q0, q4 |
| 0x37,0xfe,0x05,0x1e = vqmovnt.u32 q0, q2 |
| 0x3f,0xee,0x09,0x2e = vcvtb.f16.f32 q1, q4 |
| 0x3f,0xee,0x09,0x3e = vcvtt.f16.f32 q1, q4 |
| 0x3f,0xfe,0x07,0x0e = vcvtb.f32.f16 q0, q3 |
| 0x3f,0xfe,0x03,0x1e = vcvtt.f32.f16 q0, q1 |
| 0x31,0xee,0x87,0x0e = vqmovunb.s16 q0, q3 |
| 0x31,0xee,0x83,0x9e = vqmovunt.s16 q4, q1 |
| 0x35,0xee,0x8f,0x2e = vqmovunb.s32 q1, q7 |
| 0x35,0xee,0x85,0x1e = vqmovunt.s32 q0, q2 |
| 0x31,0xfe,0x8b,0x2e = vmovnb.i16 q1, q5 |
| 0x31,0xfe,0x81,0x1e = vmovnt.i16 q0, q0 |
| 0x35,0xfe,0x81,0x2e = vmovnb.i32 q1, q0 |
| 0x35,0xfe,0x87,0x7e = vmovnt.i32 q3, q3 |
| 0x0e,0xee,0x0a,0x6f = vhcadd.s8 q3, q7, q5, #0x5a |
| 0x10,0xee,0x0c,0x0f = vhcadd.s16 q0, q0, q6, #0x5a |
| 0x10,0xee,0x0c,0x0f = vhcadd.s16 q0, q0, q6, #0x5a |
| 0x12,0xee,0x00,0x7f = vhcadd.s16 q3, q1, q0, #0x10e |
| 0x28,0xee,0x0a,0x6f = vhcadd.s32 q3, q4, q5, #0x5a |
| 0x2e,0xee,0x04,0xdf = vhcadd.s32 q6, q7, q2, #0x10e |
| 0x30,0xee,0x04,0x2f = vadc.i32 q1, q0, q2 |
| 0x32,0xee,0x02,0x1f = vadci.i32 q0, q1, q1 |
| 0x00,0xfe,0x04,0x2f = vcadd.i8 q1, q0, q2, #0x5a |
| 0x14,0xfe,0x06,0x0f = vcadd.i16 q0, q2, q3, #0x5a |
| 0x1a,0xfe,0x0a,0x1f = vcadd.i16 q0, q5, q5, #0x10e |
| 0x24,0xfe,0x0a,0x8f = vcadd.i32 q4, q2, q5, #0x5a |
| 0x2a,0xfe,0x00,0xbf = vcadd.i32 q5, q5, q0, #0x10e |
| 0x32,0xfe,0x02,0x6f = vsbc.i32 q3, q1, q1 |
| 0x3c,0xfe,0x04,0x5f = vsbci.i32 q2, q6, q2 |
| 0x38,0xee,0x0b,0x0f = vqdmullb.s16 q0, q4, q5 |
| 0x3c,0xee,0x0b,0x1f = vqdmullt.s16 q0, q6, q5 |
| 0x36,0xfe,0x0f,0x0f = vqdmullb.s32 q0, q3, q7 |
| 0x3e,0xfe,0x0b,0x1f = vqdmullt.s32 q0, q7, q5 |
| 0x32,0xee,0x01,0x0f = vqdmullb.s16 q0, q1, q0 |
| 0x30,0xee,0x0b,0x1f = vqdmullt.s16 q0, q0, q5 |
| 0x32,0xee,0x05,0x1f = vqdmullt.s16 q0, q1, q2 |
| 0x30,0xee,0x60,0x0f = vqdmullb.s16 q0, q0, r0 |
| 0x20,0xfe,0x02,0x1f = vcadd.i32 q0, q0, q1, #0x10e |
| 0x90,0xfd,0x42,0x08 = vcadd.f32 q0, q0, q1, #0x10e |
| 0x20,0xee,0x02,0x1f = vhcadd.s32 q0, q0, q1, #0x10e |
| 0x10,0xee,0x02,0x1f = vhcadd.s16 q0, q0, q1, #0x10e |
| 0xb0,0xff,0xc0,0x00 = vrev32.8 q0, q0 |
| 0x71,0xfe,0x4d,0x8f = vpste |
| 0x32,0xfe,0x05,0x1f = vqdmulltt.s32 q0, q1, q2 |
| 0x32,0xee,0x05,0x0f = vqdmullbe.s16 q0, q1, q2 |
| 0x71,0xfe,0x4d,0x8f = vpste |
| 0x33,0xee,0x04,0x1e = vmulltt.p8 q0, q1, q2 |
| 0x33,0xfe,0x04,0x0e = vmullbe.p16 q0, q1, q2 |
| 0x71,0xfe,0x4d,0x8f = vpste |
| 0x32,0xee,0x04,0x1e = vcmult.f16 q0, q1, q2, #0xb4 |
| 0x32,0xee,0x04,0x1e = vcmule.f16 q0, q1, q2, #0xb4 |
| 0x71,0xfe,0x4d,0xcf = vpstet |
| 0x3f,0xee,0x03,0x0e = vcvtbt.f16.f32 q0, q1 |
| 0xb7,0xff,0x42,0x01 = vcvtne.s16.f16 q0, q1 |
| 0x77,0xee,0xc1,0x9f = vpte.f32 lt, q3, r1 |
| 0x3f,0xee,0x01,0x5e = vcvttt.f16.f32 q2, q0 |
| 0x3f,0xfe,0x01,0x3e = vcvtte.f32.f16 q1, q0 |
| 0x77,0xee,0xc1,0x9f = vpte.f32 lt, q3, r1 |
| 0x3f,0xee,0x01,0x4e = vcvtbt.f16.f32 q2, q0 |
| 0x3f,0xfe,0x01,0x2e = vcvtbe.f32.f16 q1, q0 |
| 0x0c,0xbf = ite eq |
| 0xb3,0xee,0xe0,0x0a = vcvtteq.f16.f32 s0, s1 |
| 0xb3,0xee,0xe0,0x0a = vcvttne.f16.f32 s0, s1 |