| # CS_ARCH_ARM, CS_MODE_THUMB+CS_MODE_V8+CS_MODE_MCLASS, None |
| 0x90,0xed,0x00,0x1e = vldrb.u8 q0, [r0] |
| 0x90,0xed,0x00,0x3e = vldrb.u8 q1, [r0] |
| 0x9b,0xed,0x00,0x1e = vldrb.u8 q0, [r11] |
| 0x9b,0xed,0x00,0x7e = vldrb.u8 q3, [r11] |
| 0x94,0xed,0x38,0x1e = vldrb.u8 q0, [r4, #0x38] |
| 0x94,0xed,0x38,0x9e = vldrb.u8 q4, [r4, #0x38] |
| 0x98,0xed,0x38,0x1e = vldrb.u8 q0, [r8, #0x38] |
| 0xb4,0xed,0x38,0xbe = vldrb.u8 q5, [r4, #0x38]! |
| 0xb4,0xed,0x38,0xbe = vldrb.u8 q5, [r4, #0x38]! |
| 0x34,0xec,0x19,0xbe = vldrb.u8 q5, [r4], #-0x19 |
| 0x3a,0xec,0x19,0xbe = vldrb.u8 q5, [r10], #-0x19 |
| 0x1d,0xed,0x19,0xbe = vldrb.u8 q5, [sp, #-0x19] |
| 0x1d,0xed,0x7f,0xbe = vldrb.u8 q5, [sp, #-0x7f] |
| 0x80,0xed,0x00,0x1e = vstrb.8 q0, [r0] |
| 0x80,0xed,0x00,0x3e = vstrb.8 q1, [r0] |
| 0x8b,0xed,0x00,0x1e = vstrb.8 q0, [r11] |
| 0x8b,0xed,0x00,0x7e = vstrb.8 q3, [r11] |
| 0x84,0xed,0x38,0x1e = vstrb.8 q0, [r4, #0x38] |
| 0x84,0xed,0x38,0x9e = vstrb.8 q4, [r4, #0x38] |
| 0x88,0xed,0x38,0x1e = vstrb.8 q0, [r8, #0x38] |
| 0xa4,0xed,0x38,0xbe = vstrb.8 q5, [r4, #0x38]! |
| 0xa4,0xed,0x38,0xbe = vstrb.8 q5, [r4, #0x38]! |
| 0x24,0xec,0x19,0xbe = vstrb.8 q5, [r4], #-0x19 |
| 0x2a,0xec,0x19,0xbe = vstrb.8 q5, [r10], #-0x19 |
| 0x0d,0xed,0x19,0xbe = vstrb.8 q5, [sp, #-0x19] |
| 0x8d,0xed,0x7f,0xbe = vstrb.8 q5, [sp, #0x7f] |
| 0x90,0xfd,0x80,0x0e = vldrb.u16 q0, [r0] |
| 0x90,0xfd,0x80,0x2e = vldrb.u16 q1, [r0] |
| 0x97,0xfd,0x80,0x0e = vldrb.u16 q0, [r7] |
| 0x97,0xfd,0x80,0x6e = vldrb.u16 q3, [r7] |
| 0x94,0xfd,0xb8,0x0e = vldrb.u16 q0, [r4, #0x38] |
| 0x94,0xfd,0xb8,0x8e = vldrb.u16 q4, [r4, #0x38] |
| 0x92,0xfd,0xb8,0x0e = vldrb.u16 q0, [r2, #0x38] |
| 0xb4,0xfd,0xb8,0xae = vldrb.u16 q5, [r4, #0x38]! |
| 0xb4,0xfd,0xb8,0xae = vldrb.u16 q5, [r4, #0x38]! |
| 0x34,0xfc,0x81,0xae = vldrb.u16 q5, [r4], #-1 |
| 0x33,0xfc,0x99,0xae = vldrb.u16 q5, [r3], #-0x19 |
| 0x16,0xfd,0x99,0xae = vldrb.u16 q5, [r6, #-0x19] |
| 0x16,0xfd,0xc0,0xae = vldrb.u16 q5, [r6, #-0x40] |
| 0x90,0xed,0x80,0x0e = vldrb.s16 q0, [r0] |
| 0x90,0xed,0x80,0x2e = vldrb.s16 q1, [r0] |
| 0x97,0xed,0x80,0x0e = vldrb.s16 q0, [r7] |
| 0x97,0xed,0x80,0x6e = vldrb.s16 q3, [r7] |
| 0x94,0xed,0xb8,0x0e = vldrb.s16 q0, [r4, #0x38] |
| 0x94,0xed,0xb8,0x8e = vldrb.s16 q4, [r4, #0x38] |
| 0x92,0xed,0xb8,0x0e = vldrb.s16 q0, [r2, #0x38] |
| 0xb4,0xed,0xb8,0xae = vldrb.s16 q5, [r4, #0x38]! |
| 0xb4,0xed,0xb8,0xae = vldrb.s16 q5, [r4, #0x38]! |
| 0x34,0xec,0x99,0xae = vldrb.s16 q5, [r4], #-0x19 |
| 0x33,0xec,0x99,0xae = vldrb.s16 q5, [r3], #-0x19 |
| 0x16,0xed,0x99,0xae = vldrb.s16 q5, [r6, #-0x19] |
| 0x16,0xed,0xc0,0xae = vldrb.s16 q5, [r6, #-0x40] |
| 0x80,0xed,0x80,0x0e = vstrb.16 q0, [r0] |
| 0x80,0xed,0x80,0x2e = vstrb.16 q1, [r0] |
| 0x87,0xed,0x80,0x0e = vstrb.16 q0, [r7] |
| 0x87,0xed,0x80,0x6e = vstrb.16 q3, [r7] |
| 0x84,0xed,0xb8,0x0e = vstrb.16 q0, [r4, #0x38] |
| 0x84,0xed,0xb8,0x8e = vstrb.16 q4, [r4, #0x38] |
| 0x85,0xed,0xb8,0x0e = vstrb.16 q0, [r5, #0x38] |
| 0xa4,0xed,0xb8,0xae = vstrb.16 q5, [r4, #0x38]! |
| 0xa4,0xed,0xb8,0xae = vstrb.16 q5, [r4, #0x38]! |
| 0x24,0xec,0x99,0xae = vstrb.16 q5, [r4], #-0x19 |
| 0x23,0xec,0x99,0xae = vstrb.16 q5, [r3], #-0x19 |
| 0x02,0xed,0x99,0xae = vstrb.16 q5, [r2, #-0x19] |
| 0x02,0xed,0xc0,0xae = vstrb.16 q5, [r2, #-0x40] |
| 0x90,0xfd,0x00,0x0f = vldrb.u32 q0, [r0] |
| 0x90,0xfd,0x00,0x2f = vldrb.u32 q1, [r0] |
| 0x97,0xfd,0x00,0x0f = vldrb.u32 q0, [r7] |
| 0x97,0xfd,0x00,0x6f = vldrb.u32 q3, [r7] |
| 0x94,0xfd,0x38,0x0f = vldrb.u32 q0, [r4, #0x38] |
| 0x94,0xfd,0x38,0x8f = vldrb.u32 q4, [r4, #0x38] |
| 0x92,0xfd,0x38,0x0f = vldrb.u32 q0, [r2, #0x38] |
| 0xb4,0xfd,0x38,0xaf = vldrb.u32 q5, [r4, #0x38]! |
| 0xb4,0xfd,0x38,0xaf = vldrb.u32 q5, [r4, #0x38]! |
| 0x34,0xfc,0x19,0xaf = vldrb.u32 q5, [r4], #-0x19 |
| 0x33,0xfc,0x19,0xaf = vldrb.u32 q5, [r3], #-0x19 |
| 0x16,0xfd,0x19,0xaf = vldrb.u32 q5, [r6, #-0x19] |
| 0x16,0xfd,0x40,0xaf = vldrb.u32 q5, [r6, #-0x40] |
| 0x90,0xed,0x00,0x0f = vldrb.s32 q0, [r0] |
| 0x90,0xed,0x00,0x2f = vldrb.s32 q1, [r0] |
| 0x97,0xed,0x00,0x0f = vldrb.s32 q0, [r7] |
| 0x97,0xed,0x00,0x6f = vldrb.s32 q3, [r7] |
| 0x94,0xed,0x38,0x0f = vldrb.s32 q0, [r4, #0x38] |
| 0x94,0xed,0x38,0x8f = vldrb.s32 q4, [r4, #0x38] |
| 0x92,0xed,0x38,0x0f = vldrb.s32 q0, [r2, #0x38] |
| 0xb4,0xed,0x38,0xaf = vldrb.s32 q5, [r4, #0x38]! |
| 0xb4,0xed,0x38,0xaf = vldrb.s32 q5, [r4, #0x38]! |
| 0x34,0xec,0x19,0xaf = vldrb.s32 q5, [r4], #-0x19 |
| 0x33,0xec,0x19,0xaf = vldrb.s32 q5, [r3], #-0x19 |
| 0x16,0xed,0x19,0xaf = vldrb.s32 q5, [r6, #-0x19] |
| 0x16,0xed,0x40,0xaf = vldrb.s32 q5, [r6, #-0x40] |
| 0x80,0xed,0x00,0x0f = vstrb.32 q0, [r0] |
| 0x80,0xed,0x00,0x2f = vstrb.32 q1, [r0] |
| 0x87,0xed,0x00,0x0f = vstrb.32 q0, [r7] |
| 0x87,0xed,0x00,0x6f = vstrb.32 q3, [r7] |
| 0x84,0xed,0x38,0x0f = vstrb.32 q0, [r4, #0x38] |
| 0x84,0xed,0x38,0x8f = vstrb.32 q4, [r4, #0x38] |
| 0x85,0xed,0x38,0x0f = vstrb.32 q0, [r5, #0x38] |
| 0xa4,0xed,0x38,0xaf = vstrb.32 q5, [r4, #0x38]! |
| 0xa4,0xed,0x38,0xaf = vstrb.32 q5, [r4, #0x38]! |
| 0x24,0xec,0x19,0xaf = vstrb.32 q5, [r4], #-0x19 |
| 0x23,0xec,0x19,0xaf = vstrb.32 q5, [r3], #-0x19 |
| 0x02,0xed,0x19,0xaf = vstrb.32 q5, [r2, #-0x19] |
| 0x02,0xed,0x40,0xaf = vstrb.32 q5, [r2, #-0x40] |
| 0x90,0xed,0x80,0x1e = vldrh.u16 q0, [r0] |
| 0x90,0xed,0x80,0x3e = vldrh.u16 q1, [r0] |
| 0x9b,0xed,0x80,0x1e = vldrh.u16 q0, [r11] |
| 0x9b,0xed,0x80,0x7e = vldrh.u16 q3, [r11] |
| 0x94,0xed,0x9c,0x1e = vldrh.u16 q0, [r4, #0x38] |
| 0x94,0xed,0x9c,0x9e = vldrh.u16 q4, [r4, #0x38] |
| 0x98,0xed,0x9c,0x1e = vldrh.u16 q0, [r8, #0x38] |
| 0xb4,0xed,0x9c,0xbe = vldrh.u16 q5, [r4, #0x38]! |
| 0xb4,0xed,0x9c,0xbe = vldrh.u16 q5, [r4, #0x38]! |
| 0x34,0xec,0x8d,0xbe = vldrh.u16 q5, [r4], #-0x1a |
| 0x3a,0xec,0x8d,0xbe = vldrh.u16 q5, [r10], #-0x1a |
| 0x1d,0xed,0x8d,0xbe = vldrh.u16 q5, [sp, #-0x1a] |
| 0x1d,0xed,0xa0,0xbe = vldrh.u16 q5, [sp, #-0x40] |
| 0x1d,0xed,0xff,0xbe = vldrh.u16 q5, [sp, #-0xfe] |
| 0xba,0xec,0xff,0xbe = vldrh.u16 q5, [r10], #0xfe |
| 0x80,0xed,0x80,0x1e = vstrh.16 q0, [r0] |
| 0x80,0xed,0x80,0x3e = vstrh.16 q1, [r0] |
| 0x8b,0xed,0x80,0x1e = vstrh.16 q0, [r11] |
| 0x8b,0xed,0x80,0x7e = vstrh.16 q3, [r11] |
| 0x84,0xed,0x9c,0x1e = vstrh.16 q0, [r4, #0x38] |
| 0x84,0xed,0x9c,0x9e = vstrh.16 q4, [r4, #0x38] |
| 0x88,0xed,0x9c,0x1e = vstrh.16 q0, [r8, #0x38] |
| 0xa4,0xed,0x9c,0xbe = vstrh.16 q5, [r4, #0x38]! |
| 0xa4,0xed,0x9c,0xbe = vstrh.16 q5, [r4, #0x38]! |
| 0x24,0xec,0x8d,0xbe = vstrh.16 q5, [r4], #-0x1a |
| 0x2a,0xec,0x8d,0xbe = vstrh.16 q5, [r10], #-0x1a |
| 0x0d,0xed,0x8d,0xbe = vstrh.16 q5, [sp, #-0x1a] |
| 0x0d,0xed,0xa0,0xbe = vstrh.16 q5, [sp, #-0x40] |
| 0x0d,0xed,0xff,0xbe = vstrh.16 q5, [sp, #-0xfe] |
| 0xaa,0xec,0xff,0xbe = vstrh.16 q5, [r10], #0xfe |
| 0x98,0xfd,0x00,0x0f = vldrh.u32 q0, [r0] |
| 0x98,0xfd,0x00,0x2f = vldrh.u32 q1, [r0] |
| 0x9f,0xfd,0x00,0x0f = vldrh.u32 q0, [r7] |
| 0x9f,0xfd,0x00,0x6f = vldrh.u32 q3, [r7] |
| 0x9c,0xfd,0x1c,0x0f = vldrh.u32 q0, [r4, #0x38] |
| 0x9c,0xfd,0x1c,0x8f = vldrh.u32 q4, [r4, #0x38] |
| 0x9a,0xfd,0x1c,0x0f = vldrh.u32 q0, [r2, #0x38] |
| 0xbc,0xfd,0x1c,0xaf = vldrh.u32 q5, [r4, #0x38]! |
| 0xbc,0xfd,0x1c,0xaf = vldrh.u32 q5, [r4, #0x38]! |
| 0x3c,0xfc,0x0d,0xaf = vldrh.u32 q5, [r4], #-0x1a |
| 0x3b,0xfc,0x0d,0xaf = vldrh.u32 q5, [r3], #-0x1a |
| 0x1e,0xfd,0x0d,0xaf = vldrh.u32 q5, [r6, #-0x1a] |
| 0x1e,0xfd,0x20,0xaf = vldrh.u32 q5, [r6, #-0x40] |
| 0x1e,0xfd,0x7f,0xaf = vldrh.u32 q5, [r6, #-0xfe] |
| 0xbc,0xfd,0x7f,0xaf = vldrh.u32 q5, [r4, #0xfe]! |
| 0x98,0xed,0x00,0x0f = vldrh.s32 q0, [r0] |
| 0x98,0xed,0x00,0x2f = vldrh.s32 q1, [r0] |
| 0x9f,0xed,0x00,0x0f = vldrh.s32 q0, [r7] |
| 0x9f,0xed,0x00,0x6f = vldrh.s32 q3, [r7] |
| 0x9c,0xed,0x1c,0x0f = vldrh.s32 q0, [r4, #0x38] |
| 0x9c,0xed,0x1c,0x8f = vldrh.s32 q4, [r4, #0x38] |
| 0x9a,0xed,0x1c,0x0f = vldrh.s32 q0, [r2, #0x38] |
| 0xbc,0xed,0x1c,0xaf = vldrh.s32 q5, [r4, #0x38]! |
| 0xbc,0xed,0x1c,0xaf = vldrh.s32 q5, [r4, #0x38]! |
| 0x3c,0xec,0x0d,0xaf = vldrh.s32 q5, [r4], #-0x1a |
| 0x3b,0xec,0x0d,0xaf = vldrh.s32 q5, [r3], #-0x1a |
| 0x1e,0xed,0x0d,0xaf = vldrh.s32 q5, [r6, #-0x1a] |
| 0x1e,0xed,0x20,0xaf = vldrh.s32 q5, [r6, #-0x40] |
| 0x1e,0xed,0x7f,0xaf = vldrh.s32 q5, [r6, #-0xfe] |
| 0xbc,0xed,0x7f,0xaf = vldrh.s32 q5, [r4, #0xfe]! |
| 0x88,0xed,0x00,0x0f = vstrh.32 q0, [r0] |
| 0x88,0xed,0x00,0x2f = vstrh.32 q1, [r0] |
| 0x8f,0xed,0x00,0x0f = vstrh.32 q0, [r7] |
| 0x8f,0xed,0x00,0x6f = vstrh.32 q3, [r7] |
| 0x8c,0xed,0x1c,0x0f = vstrh.32 q0, [r4, #0x38] |
| 0x8c,0xed,0x1c,0x8f = vstrh.32 q4, [r4, #0x38] |
| 0x8d,0xed,0x1c,0x0f = vstrh.32 q0, [r5, #0x38] |
| 0xac,0xed,0x1c,0xaf = vstrh.32 q5, [r4, #0x38]! |
| 0xac,0xed,0x1c,0xaf = vstrh.32 q5, [r4, #0x38]! |
| 0x2c,0xec,0x0d,0xaf = vstrh.32 q5, [r4], #-0x1a |
| 0x2b,0xec,0x0d,0xaf = vstrh.32 q5, [r3], #-0x1a |
| 0x0a,0xed,0x0d,0xaf = vstrh.32 q5, [r2, #-0x1a] |
| 0x0a,0xed,0x20,0xaf = vstrh.32 q5, [r2, #-0x40] |
| 0x0a,0xed,0x7f,0xaf = vstrh.32 q5, [r2, #-0xfe] |
| 0xac,0xed,0x7f,0xaf = vstrh.32 q5, [r4, #0xfe]! |
| 0x90,0xed,0x00,0x1f = vldrw.u32 q0, [r0] |
| 0x90,0xed,0x00,0x3f = vldrw.u32 q1, [r0] |
| 0x9b,0xed,0x00,0x1f = vldrw.u32 q0, [r11] |
| 0x9b,0xed,0x00,0x7f = vldrw.u32 q3, [r11] |
| 0x94,0xed,0x0e,0x1f = vldrw.u32 q0, [r4, #0x38] |
| 0x94,0xed,0x0e,0x9f = vldrw.u32 q4, [r4, #0x38] |
| 0x98,0xed,0x0e,0x1f = vldrw.u32 q0, [r8, #0x38] |
| 0xb4,0xed,0x0e,0xbf = vldrw.u32 q5, [r4, #0x38]! |
| 0xb4,0xed,0x0e,0xbf = vldrw.u32 q5, [r4, #0x38]! |
| 0x34,0xec,0x07,0xbf = vldrw.u32 q5, [r4], #-0x1c |
| 0x3a,0xec,0x07,0xbf = vldrw.u32 q5, [r10], #-0x1c |
| 0x1d,0xed,0x07,0xbf = vldrw.u32 q5, [sp, #-0x1c] |
| 0x1d,0xed,0x10,0xbf = vldrw.u32 q5, [sp, #-0x40] |
| 0x1d,0xed,0x7f,0xbf = vldrw.u32 q5, [sp, #-0x1fc] |
| 0xb4,0xed,0x7f,0xbf = vldrw.u32 q5, [r4, #0x1fc]! |
| 0x80,0xed,0x00,0x1f = vstrw.32 q0, [r0] |
| 0x80,0xed,0x00,0x3f = vstrw.32 q1, [r0] |
| 0x8b,0xed,0x00,0x1f = vstrw.32 q0, [r11] |
| 0x8b,0xed,0x00,0x7f = vstrw.32 q3, [r11] |
| 0x84,0xed,0x0e,0x1f = vstrw.32 q0, [r4, #0x38] |
| 0x84,0xed,0x0e,0x9f = vstrw.32 q4, [r4, #0x38] |
| 0x88,0xed,0x0e,0x1f = vstrw.32 q0, [r8, #0x38] |
| 0xa4,0xed,0x0e,0xbf = vstrw.32 q5, [r4, #0x38]! |
| 0xa4,0xed,0x0e,0xbf = vstrw.32 q5, [r4, #0x38]! |
| 0x24,0xec,0x07,0xbf = vstrw.32 q5, [r4], #-0x1c |
| 0x2a,0xec,0x07,0xbf = vstrw.32 q5, [r10], #-0x1c |
| 0x0d,0xed,0x07,0xbf = vstrw.32 q5, [sp, #-0x1c] |
| 0x0d,0xed,0x10,0xbf = vstrw.32 q5, [sp, #-0x40] |
| 0x0d,0xed,0x7f,0xbf = vstrw.32 q5, [sp, #-0x1fc] |
| 0xa4,0xed,0x7f,0xbf = vstrw.32 q5, [r4, #0x1fc]! |
| 0x90,0xfc,0x02,0x0e = vldrb.u8 q0, [r0, q1] |
| 0x9a,0xfc,0x02,0x6e = vldrb.u8 q3, [r10, q1] |
| 0x90,0xfc,0x82,0x0e = vldrb.u16 q0, [r0, q1] |
| 0x99,0xfc,0x82,0x6e = vldrb.u16 q3, [r9, q1] |
| 0x90,0xec,0x82,0x0e = vldrb.s16 q0, [r0, q1] |
| 0x9d,0xec,0x82,0x6e = vldrb.s16 q3, [sp, q1] |
| 0x90,0xfc,0x02,0x0f = vldrb.u32 q0, [r0, q1] |
| 0x90,0xfc,0x02,0x6f = vldrb.u32 q3, [r0, q1] |
| 0x90,0xec,0x02,0x0f = vldrb.s32 q0, [r0, q1] |
| 0x90,0xec,0x02,0x6f = vldrb.s32 q3, [r0, q1] |
| 0x90,0xfc,0x92,0x0e = vldrh.u16 q0, [r0, q1] |
| 0x90,0xfc,0x92,0x6e = vldrh.u16 q3, [r0, q1] |
| 0x90,0xfc,0x12,0x0f = vldrh.u32 q0, [r0, q1] |
| 0x90,0xfc,0x12,0x6f = vldrh.u32 q3, [r0, q1] |
| 0x90,0xec,0x12,0x0f = vldrh.s32 q0, [r0, q1] |
| 0x90,0xec,0x12,0x6f = vldrh.s32 q3, [r0, q1] |
| 0x90,0xfc,0x93,0x0e = vldrh.u16 q0, [r0, q1, uxtw #1] |
| 0x90,0xfc,0x42,0x0f = vldrw.u32 q0, [r0, q1] |
| 0x90,0xfc,0x42,0x6f = vldrw.u32 q3, [r0, q1] |
| 0x90,0xfc,0x43,0x0f = vldrw.u32 q0, [r0, q1, uxtw #2] |
| 0x9d,0xfc,0x43,0x0f = vldrw.u32 q0, [sp, q1, uxtw #2] |
| 0x90,0xfc,0xd2,0x0f = vldrd.u64 q0, [r0, q1] |
| 0x90,0xfc,0xd2,0x6f = vldrd.u64 q3, [r0, q1] |
| 0x90,0xfc,0xd3,0x0f = vldrd.u64 q0, [r0, q1, uxtw #3] |
| 0x9d,0xfc,0xd3,0x0f = vldrd.u64 q0, [sp, q1, uxtw #3] |
| 0x80,0xec,0x02,0x0e = vstrb.8 q0, [r0, q1] |
| 0x8a,0xec,0x02,0x6e = vstrb.8 q3, [r10, q1] |
| 0x80,0xec,0x06,0x6e = vstrb.8 q3, [r0, q3] |
| 0x80,0xec,0x82,0x0e = vstrb.16 q0, [r0, q1] |
| 0x8d,0xec,0x82,0x6e = vstrb.16 q3, [sp, q1] |
| 0x80,0xec,0x86,0x6e = vstrb.16 q3, [r0, q3] |
| 0x80,0xec,0x02,0x0f = vstrb.32 q0, [r0, q1] |
| 0x80,0xec,0x02,0x6f = vstrb.32 q3, [r0, q1] |
| 0x80,0xec,0x06,0x6f = vstrb.32 q3, [r0, q3] |
| 0x80,0xec,0x92,0x0e = vstrh.16 q0, [r0, q1] |
| 0x80,0xec,0x92,0x6e = vstrh.16 q3, [r0, q1] |
| 0x80,0xec,0x96,0x6e = vstrh.16 q3, [r0, q3] |
| 0x80,0xec,0x12,0x0f = vstrh.32 q0, [r0, q1] |
| 0x80,0xec,0x12,0x6f = vstrh.32 q3, [r0, q1] |
| 0x80,0xec,0x16,0x6f = vstrh.32 q3, [r0, q3] |
| 0x80,0xec,0x93,0x0e = vstrh.16 q0, [r0, q1, uxtw #1] |
| 0x88,0xec,0x17,0x6f = vstrh.32 q3, [r8, q3, uxtw #1] |
| 0x80,0xec,0x42,0x0f = vstrw.32 q0, [r0, q1] |
| 0x80,0xec,0x42,0x6f = vstrw.32 q3, [r0, q1] |
| 0x80,0xec,0x46,0x6f = vstrw.32 q3, [r0, q3] |
| 0x80,0xec,0x43,0x0f = vstrw.32 q0, [r0, q1, uxtw #2] |
| 0x8d,0xec,0x43,0x0f = vstrw.32 q0, [sp, q1, uxtw #2] |
| 0x80,0xec,0xd2,0x0f = vstrd.64 q0, [r0, q1] |
| 0x80,0xec,0xd2,0x6f = vstrd.64 q3, [r0, q1] |
| 0x80,0xec,0xd6,0x6f = vstrd.64 q3, [r0, q3] |
| 0x80,0xec,0xd3,0x0f = vstrd.64 q0, [r0, q1, uxtw #3] |
| 0x8d,0xec,0xd3,0x0f = vstrd.64 q0, [sp, q1, uxtw #3] |
| 0x92,0xfd,0x00,0x1e = vldrw.u32 q0, [q1] |
| 0x92,0xfd,0x00,0xfe = vldrw.u32 q7, [q1] |
| 0xb2,0xfd,0x00,0xfe = vldrw.u32 q7, [q1]! |
| 0x92,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4] |
| 0x12,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #-4] |
| 0x92,0xfd,0x7f,0xfe = vldrw.u32 q7, [q1, #0x1fc] |
| 0x12,0xfd,0x7f,0xfe = vldrw.u32 q7, [q1, #-0x1fc] |
| 0x92,0xfd,0x42,0xfe = vldrw.u32 q7, [q1, #0x108] |
| 0xb2,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4]! |
| 0x82,0xfd,0x00,0x1e = vstrw.32 q0, [q1] |
| 0x82,0xfd,0x00,0x3e = vstrw.32 q1, [q1] |
| 0x82,0xfd,0x00,0xfe = vstrw.32 q7, [q1] |
| 0xa2,0xfd,0x00,0xfe = vstrw.32 q7, [q1]! |
| 0x8e,0xfd,0x00,0xfe = vstrw.32 q7, [q7] |
| 0x82,0xfd,0x01,0xfe = vstrw.32 q7, [q1, #4] |
| 0x02,0xfd,0x01,0xfe = vstrw.32 q7, [q1, #-4] |
| 0x82,0xfd,0x7f,0xfe = vstrw.32 q7, [q1, #0x1fc] |
| 0x02,0xfd,0x7f,0xfe = vstrw.32 q7, [q1, #-0x1fc] |
| 0xa2,0xfd,0x42,0xfe = vstrw.32 q7, [q1, #0x108]! |
| 0x92,0xfd,0x00,0x1f = vldrd.u64 q0, [q1] |
| 0x92,0xfd,0x00,0xff = vldrd.u64 q7, [q1] |
| 0xb2,0xfd,0x00,0xff = vldrd.u64 q7, [q1]! |
| 0x92,0xfd,0x01,0xff = vldrd.u64 q7, [q1, #8] |
| 0x12,0xfd,0x01,0xff = vldrd.u64 q7, [q1, #-8] |
| 0x92,0xfd,0x7f,0xff = vldrd.u64 q7, [q1, #0x3f8] |
| 0x12,0xfd,0x7f,0xff = vldrd.u64 q7, [q1, #-0x3f8] |
| 0x92,0xfd,0x21,0xff = vldrd.u64 q7, [q1, #0x108] |
| 0x92,0xfd,0x4e,0xff = vldrd.u64 q7, [q1, #0x270] |
| 0x92,0xfd,0x21,0xff = vldrd.u64 q7, [q1, #0x108] |
| 0x32,0xfd,0x7f,0xff = vldrd.u64 q7, [q1, #-0x3f8]! |
| 0x82,0xfd,0x00,0x1f = vstrd.64 q0, [q1] |
| 0x82,0xfd,0x00,0x3f = vstrd.64 q1, [q1] |
| 0x82,0xfd,0x00,0xff = vstrd.64 q7, [q1] |
| 0xa2,0xfd,0x00,0xff = vstrd.64 q7, [q1]! |
| 0x8e,0xfd,0x00,0xff = vstrd.64 q7, [q7] |
| 0x82,0xfd,0x01,0xff = vstrd.64 q7, [q1, #8] |
| 0x22,0xfd,0x01,0xff = vstrd.64 q7, [q1, #-8]! |
| 0x82,0xfd,0x7f,0xff = vstrd.64 q7, [q1, #0x3f8] |
| 0x02,0xfd,0x7f,0xff = vstrd.64 q7, [q1, #-0x3f8] |
| 0x82,0xfd,0x21,0xff = vstrd.64 q7, [q1, #0x108] |
| 0x82,0xfd,0x4e,0xff = vstrd.64 q7, [q1, #0x270] |
| 0x82,0xfd,0x21,0xff = vstrd.64 q7, [q1, #0x108] |
| 0x90,0xed,0x00,0x1e = vldrb.u8 q0, [r0] |
| 0x90,0xed,0x00,0x1e = vldrb.u8 q0, [r0] |
| 0x98,0xed,0x38,0x1e = vldrb.u8 q0, [r8, #0x38] |
| 0x98,0xed,0x38,0x1e = vldrb.u8 q0, [r8, #0x38] |
| 0xb4,0xed,0x38,0xbe = vldrb.u8 q5, [r4, #0x38]! |
| 0xb4,0xed,0x38,0xbe = vldrb.u8 q5, [r4, #0x38]! |
| 0x80,0xed,0x00,0x1e = vstrb.8 q0, [r0] |
| 0x80,0xed,0x00,0x1e = vstrb.8 q0, [r0] |
| 0x84,0xed,0x38,0x9e = vstrb.8 q4, [r4, #0x38] |
| 0x84,0xed,0x38,0x9e = vstrb.8 q4, [r4, #0x38] |
| 0xa4,0xed,0x38,0xbe = vstrb.8 q5, [r4, #0x38]! |
| 0xa4,0xed,0x38,0xbe = vstrb.8 q5, [r4, #0x38]! |
| 0x90,0xed,0x80,0x1e = vldrh.u16 q0, [r0] |
| 0x90,0xed,0x80,0x1e = vldrh.u16 q0, [r0] |
| 0x90,0xed,0x80,0x1e = vldrh.u16 q0, [r0] |
| 0x94,0xed,0x9c,0x1e = vldrh.u16 q0, [r4, #0x38] |
| 0x94,0xed,0x9c,0x1e = vldrh.u16 q0, [r4, #0x38] |
| 0x94,0xed,0x9c,0x1e = vldrh.u16 q0, [r4, #0x38] |
| 0xb4,0xed,0x9c,0xbe = vldrh.u16 q5, [r4, #0x38]! |
| 0xb4,0xed,0x9c,0xbe = vldrh.u16 q5, [r4, #0x38]! |
| 0xb4,0xed,0x9c,0xbe = vldrh.u16 q5, [r4, #0x38]! |
| 0x80,0xed,0x80,0x1e = vstrh.16 q0, [r0] |
| 0x80,0xed,0x80,0x1e = vstrh.16 q0, [r0] |
| 0x80,0xed,0x80,0x1e = vstrh.16 q0, [r0] |
| 0x84,0xed,0x9c,0x1e = vstrh.16 q0, [r4, #0x38] |
| 0x84,0xed,0x9c,0x1e = vstrh.16 q0, [r4, #0x38] |
| 0x84,0xed,0x9c,0x1e = vstrh.16 q0, [r4, #0x38] |
| 0xa4,0xed,0x9c,0xbe = vstrh.16 q5, [r4, #0x38]! |
| 0xa4,0xed,0x9c,0xbe = vstrh.16 q5, [r4, #0x38]! |
| 0xa4,0xed,0x9c,0xbe = vstrh.16 q5, [r4, #0x38]! |
| 0x90,0xed,0x00,0x1f = vldrw.u32 q0, [r0] |
| 0x90,0xed,0x00,0x1f = vldrw.u32 q0, [r0] |
| 0x90,0xed,0x00,0x1f = vldrw.u32 q0, [r0] |
| 0x94,0xed,0x0e,0x1f = vldrw.u32 q0, [r4, #0x38] |
| 0x94,0xed,0x0e,0x1f = vldrw.u32 q0, [r4, #0x38] |
| 0x94,0xed,0x0e,0x1f = vldrw.u32 q0, [r4, #0x38] |
| 0xb4,0xed,0x0e,0xbf = vldrw.u32 q5, [r4, #0x38]! |
| 0xb4,0xed,0x0e,0xbf = vldrw.u32 q5, [r4, #0x38]! |
| 0xb4,0xed,0x0e,0xbf = vldrw.u32 q5, [r4, #0x38]! |
| 0x80,0xed,0x00,0x1f = vstrw.32 q0, [r0] |
| 0x80,0xed,0x00,0x1f = vstrw.32 q0, [r0] |
| 0x80,0xed,0x00,0x1f = vstrw.32 q0, [r0] |
| 0x84,0xed,0x0e,0x1f = vstrw.32 q0, [r4, #0x38] |
| 0x84,0xed,0x0e,0x1f = vstrw.32 q0, [r4, #0x38] |
| 0x84,0xed,0x0e,0x1f = vstrw.32 q0, [r4, #0x38] |
| 0xa4,0xed,0x0e,0xbf = vstrw.32 q5, [r4, #0x38]! |
| 0xa4,0xed,0x0e,0xbf = vstrw.32 q5, [r4, #0x38]! |
| 0xa4,0xed,0x0e,0xbf = vstrw.32 q5, [r4, #0x38]! |
| 0x90,0xfc,0x02,0x0e = vldrb.u8 q0, [r0, q1] |
| 0x90,0xfc,0x02,0x0e = vldrb.u8 q0, [r0, q1] |
| 0x90,0xfc,0x92,0x6e = vldrh.u16 q3, [r0, q1] |
| 0x90,0xfc,0x92,0x6e = vldrh.u16 q3, [r0, q1] |
| 0x90,0xfc,0x92,0x6e = vldrh.u16 q3, [r0, q1] |
| 0x90,0xfc,0x93,0x0e = vldrh.u16 q0, [r0, q1, uxtw #1] |
| 0x90,0xfc,0x93,0x0e = vldrh.u16 q0, [r0, q1, uxtw #1] |
| 0x90,0xfc,0x93,0x0e = vldrh.u16 q0, [r0, q1, uxtw #1] |
| 0x90,0xfc,0x42,0x0f = vldrw.u32 q0, [r0, q1] |
| 0x90,0xfc,0x42,0x0f = vldrw.u32 q0, [r0, q1] |
| 0x90,0xfc,0x42,0x0f = vldrw.u32 q0, [r0, q1] |
| 0x90,0xfc,0x43,0x0f = vldrw.u32 q0, [r0, q1, uxtw #2] |
| 0x90,0xfc,0x43,0x0f = vldrw.u32 q0, [r0, q1, uxtw #2] |
| 0x90,0xfc,0x43,0x0f = vldrw.u32 q0, [r0, q1, uxtw #2] |
| 0x90,0xfc,0xd2,0x0f = vldrd.u64 q0, [r0, q1] |
| 0x90,0xfc,0xd2,0x0f = vldrd.u64 q0, [r0, q1] |
| 0x90,0xfc,0xd2,0x0f = vldrd.u64 q0, [r0, q1] |
| 0x90,0xfc,0xd3,0x0f = vldrd.u64 q0, [r0, q1, uxtw #3] |
| 0x90,0xfc,0xd3,0x0f = vldrd.u64 q0, [r0, q1, uxtw #3] |
| 0x90,0xfc,0xd3,0x0f = vldrd.u64 q0, [r0, q1, uxtw #3] |
| 0x80,0xec,0x02,0x0e = vstrb.8 q0, [r0, q1] |
| 0x80,0xec,0x02,0x0e = vstrb.8 q0, [r0, q1] |
| 0x80,0xec,0x92,0x6e = vstrh.16 q3, [r0, q1] |
| 0x80,0xec,0x92,0x6e = vstrh.16 q3, [r0, q1] |
| 0x80,0xec,0x92,0x6e = vstrh.16 q3, [r0, q1] |
| 0x80,0xec,0x93,0x0e = vstrh.16 q0, [r0, q1, uxtw #1] |
| 0x80,0xec,0x93,0x0e = vstrh.16 q0, [r0, q1, uxtw #1] |
| 0x80,0xec,0x93,0x0e = vstrh.16 q0, [r0, q1, uxtw #1] |
| 0x80,0xec,0x42,0x0f = vstrw.32 q0, [r0, q1] |
| 0x80,0xec,0x42,0x0f = vstrw.32 q0, [r0, q1] |
| 0x80,0xec,0x42,0x0f = vstrw.32 q0, [r0, q1] |
| 0x80,0xec,0x43,0x0f = vstrw.32 q0, [r0, q1, uxtw #2] |
| 0x80,0xec,0x43,0x0f = vstrw.32 q0, [r0, q1, uxtw #2] |
| 0x80,0xec,0x43,0x0f = vstrw.32 q0, [r0, q1, uxtw #2] |
| 0x80,0xec,0xd2,0x6f = vstrd.64 q3, [r0, q1] |
| 0x80,0xec,0xd2,0x6f = vstrd.64 q3, [r0, q1] |
| 0x80,0xec,0xd2,0x6f = vstrd.64 q3, [r0, q1] |
| 0x80,0xec,0xd3,0x0f = vstrd.64 q0, [r0, q1, uxtw #3] |
| 0x80,0xec,0xd3,0x0f = vstrd.64 q0, [r0, q1, uxtw #3] |
| 0x80,0xec,0xd3,0x0f = vstrd.64 q0, [r0, q1, uxtw #3] |
| 0x92,0xfd,0x00,0x1e = vldrw.u32 q0, [q1] |
| 0x92,0xfd,0x00,0x1e = vldrw.u32 q0, [q1] |
| 0x92,0xfd,0x00,0x1e = vldrw.u32 q0, [q1] |
| 0xb2,0xfd,0x00,0xfe = vldrw.u32 q7, [q1]! |
| 0xb2,0xfd,0x00,0xfe = vldrw.u32 q7, [q1]! |
| 0xb2,0xfd,0x00,0xfe = vldrw.u32 q7, [q1]! |
| 0x92,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4] |
| 0x92,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4] |
| 0x92,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4] |
| 0xb2,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4]! |
| 0xb2,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4]! |
| 0xb2,0xfd,0x01,0xfe = vldrw.u32 q7, [q1, #4]! |
| 0x82,0xfd,0x00,0x1e = vstrw.32 q0, [q1] |
| 0x82,0xfd,0x00,0x1e = vstrw.32 q0, [q1] |
| 0x82,0xfd,0x00,0x1e = vstrw.32 q0, [q1] |
| 0xa2,0xfd,0x00,0xfe = vstrw.32 q7, [q1]! |
| 0xa2,0xfd,0x00,0xfe = vstrw.32 q7, [q1]! |
| 0xa2,0xfd,0x00,0xfe = vstrw.32 q7, [q1]! |
| 0x82,0xfd,0x7f,0xfe = vstrw.32 q7, [q1, #0x1fc] |
| 0x82,0xfd,0x7f,0xfe = vstrw.32 q7, [q1, #0x1fc] |
| 0x82,0xfd,0x7f,0xfe = vstrw.32 q7, [q1, #0x1fc] |
| 0xa2,0xfd,0x42,0xfe = vstrw.32 q7, [q1, #0x108]! |
| 0xa2,0xfd,0x42,0xfe = vstrw.32 q7, [q1, #0x108]! |
| 0xa2,0xfd,0x42,0xfe = vstrw.32 q7, [q1, #0x108]! |
| 0x92,0xfd,0x00,0x1f = vldrd.u64 q0, [q1] |
| 0x92,0xfd,0x00,0x1f = vldrd.u64 q0, [q1] |
| 0x92,0xfd,0x00,0x1f = vldrd.u64 q0, [q1] |
| 0xb2,0xfd,0x00,0xff = vldrd.u64 q7, [q1]! |
| 0xb2,0xfd,0x00,0xff = vldrd.u64 q7, [q1]! |
| 0xb2,0xfd,0x00,0xff = vldrd.u64 q7, [q1]! |
| 0x92,0xfd,0x01,0xff = vldrd.u64 q7, [q1, #8] |
| 0x92,0xfd,0x01,0xff = vldrd.u64 q7, [q1, #8] |
| 0x92,0xfd,0x01,0xff = vldrd.u64 q7, [q1, #8] |
| 0x32,0xfd,0x7f,0xff = vldrd.u64 q7, [q1, #-0x3f8]! |
| 0x32,0xfd,0x7f,0xff = vldrd.u64 q7, [q1, #-0x3f8]! |
| 0x32,0xfd,0x7f,0xff = vldrd.u64 q7, [q1, #-0x3f8]! |
| 0x82,0xfd,0x00,0x1f = vstrd.64 q0, [q1] |
| 0x82,0xfd,0x00,0x1f = vstrd.64 q0, [q1] |
| 0x82,0xfd,0x00,0x1f = vstrd.64 q0, [q1] |
| 0xa2,0xfd,0x00,0xff = vstrd.64 q7, [q1]! |
| 0xa2,0xfd,0x00,0xff = vstrd.64 q7, [q1]! |
| 0xa2,0xfd,0x00,0xff = vstrd.64 q7, [q1]! |
| 0x82,0xfd,0x7f,0xff = vstrd.64 q7, [q1, #0x3f8] |
| 0x82,0xfd,0x7f,0xff = vstrd.64 q7, [q1, #0x3f8] |
| 0x82,0xfd,0x7f,0xff = vstrd.64 q7, [q1, #0x3f8] |
| 0x22,0xfd,0x01,0xff = vstrd.64 q7, [q1, #-8]! |
| 0x22,0xfd,0x01,0xff = vstrd.64 q7, [q1, #-8]! |
| 0x22,0xfd,0x01,0xff = vstrd.64 q7, [q1, #-8]! |
| 0x71,0xfe,0x4d,0x8f = vpste |
| 0xa2,0xfd,0x42,0xfe = vstrwt.32 q7, [q1, #0x108]! |
| 0x92,0xfd,0x01,0xff = vldrde.u64 q7, [q1, #8] |