Statistics
| Branch: | Revision:

root / target-arm / translate-a64.c @ e4b998d4

History | View | Annotate | Download (283 kB)

1
/*
2
 *  AArch64 translation
3
 *
4
 *  Copyright (c) 2013 Alexander Graf <agraf@suse.de>
5
 *
6
 * This library is free software; you can redistribute it and/or
7
 * modify it under the terms of the GNU Lesser General Public
8
 * License as published by the Free Software Foundation; either
9
 * version 2 of the License, or (at your option) any later version.
10
 *
11
 * This library is distributed in the hope that it will be useful,
12
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14
 * Lesser General Public License for more details.
15
 *
16
 * You should have received a copy of the GNU Lesser General Public
17
 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18
 */
19
#include <stdarg.h>
20
#include <stdlib.h>
21
#include <stdio.h>
22
#include <string.h>
23
#include <inttypes.h>
24

    
25
#include "cpu.h"
26
#include "tcg-op.h"
27
#include "qemu/log.h"
28
#include "translate.h"
29
#include "qemu/host-utils.h"
30

    
31
#include "exec/gen-icount.h"
32

    
33
#include "helper.h"
34
#define GEN_HELPER 1
35
#include "helper.h"
36

    
37
static TCGv_i64 cpu_X[32];
38
static TCGv_i64 cpu_pc;
39
static TCGv_i32 cpu_NF, cpu_ZF, cpu_CF, cpu_VF;
40

    
41
/* Load/store exclusive handling */
42
static TCGv_i64 cpu_exclusive_addr;
43
static TCGv_i64 cpu_exclusive_val;
44
static TCGv_i64 cpu_exclusive_high;
45
#ifdef CONFIG_USER_ONLY
46
static TCGv_i64 cpu_exclusive_test;
47
static TCGv_i32 cpu_exclusive_info;
48
#endif
49

    
50
static const char *regnames[] = {
51
    "x0", "x1", "x2", "x3", "x4", "x5", "x6", "x7",
52
    "x8", "x9", "x10", "x11", "x12", "x13", "x14", "x15",
53
    "x16", "x17", "x18", "x19", "x20", "x21", "x22", "x23",
54
    "x24", "x25", "x26", "x27", "x28", "x29", "lr", "sp"
55
};
56

    
57
enum a64_shift_type {
58
    A64_SHIFT_TYPE_LSL = 0,
59
    A64_SHIFT_TYPE_LSR = 1,
60
    A64_SHIFT_TYPE_ASR = 2,
61
    A64_SHIFT_TYPE_ROR = 3
62
};
63

    
64
/* Table based decoder typedefs - used when the relevant bits for decode
65
 * are too awkwardly scattered across the instruction (eg SIMD).
66
 */
67
typedef void AArch64DecodeFn(DisasContext *s, uint32_t insn);
68

    
69
typedef struct AArch64DecodeTable {
70
    uint32_t pattern;
71
    uint32_t mask;
72
    AArch64DecodeFn *disas_fn;
73
} AArch64DecodeTable;
74

    
75
/* Function prototype for gen_ functions for calling Neon helpers */
76
typedef void NeonGenTwoOpFn(TCGv_i32, TCGv_i32, TCGv_i32);
77
typedef void NeonGenTwoOpEnvFn(TCGv_i32, TCGv_ptr, TCGv_i32, TCGv_i32);
78
typedef void NeonGenTwo64OpFn(TCGv_i64, TCGv_i64, TCGv_i64);
79
typedef void NeonGenNarrowFn(TCGv_i32, TCGv_i64);
80
typedef void NeonGenNarrowEnvFn(TCGv_i32, TCGv_ptr, TCGv_i64);
81
typedef void NeonGenWidenFn(TCGv_i64, TCGv_i32);
82
typedef void NeonGenTwoSingleOPFn(TCGv_i32, TCGv_i32, TCGv_i32, TCGv_ptr);
83
typedef void NeonGenTwoDoubleOPFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGv_ptr);
84

    
85
/* initialize TCG globals.  */
86
void a64_translate_init(void)
87
{
88
    int i;
89

    
90
    cpu_pc = tcg_global_mem_new_i64(TCG_AREG0,
91
                                    offsetof(CPUARMState, pc),
92
                                    "pc");
93
    for (i = 0; i < 32; i++) {
94
        cpu_X[i] = tcg_global_mem_new_i64(TCG_AREG0,
95
                                          offsetof(CPUARMState, xregs[i]),
96
                                          regnames[i]);
97
    }
98

    
99
    cpu_NF = tcg_global_mem_new_i32(TCG_AREG0, offsetof(CPUARMState, NF), "NF");
100
    cpu_ZF = tcg_global_mem_new_i32(TCG_AREG0, offsetof(CPUARMState, ZF), "ZF");
101
    cpu_CF = tcg_global_mem_new_i32(TCG_AREG0, offsetof(CPUARMState, CF), "CF");
102
    cpu_VF = tcg_global_mem_new_i32(TCG_AREG0, offsetof(CPUARMState, VF), "VF");
103

    
104
    cpu_exclusive_addr = tcg_global_mem_new_i64(TCG_AREG0,
105
        offsetof(CPUARMState, exclusive_addr), "exclusive_addr");
106
    cpu_exclusive_val = tcg_global_mem_new_i64(TCG_AREG0,
107
        offsetof(CPUARMState, exclusive_val), "exclusive_val");
108
    cpu_exclusive_high = tcg_global_mem_new_i64(TCG_AREG0,
109
        offsetof(CPUARMState, exclusive_high), "exclusive_high");
110
#ifdef CONFIG_USER_ONLY
111
    cpu_exclusive_test = tcg_global_mem_new_i64(TCG_AREG0,
112
        offsetof(CPUARMState, exclusive_test), "exclusive_test");
113
    cpu_exclusive_info = tcg_global_mem_new_i32(TCG_AREG0,
114
        offsetof(CPUARMState, exclusive_info), "exclusive_info");
115
#endif
116
}
117

    
118
void aarch64_cpu_dump_state(CPUState *cs, FILE *f,
119
                            fprintf_function cpu_fprintf, int flags)
120
{
121
    ARMCPU *cpu = ARM_CPU(cs);
122
    CPUARMState *env = &cpu->env;
123
    uint32_t psr = pstate_read(env);
124
    int i;
125

    
126
    cpu_fprintf(f, "PC=%016"PRIx64"  SP=%016"PRIx64"\n",
127
            env->pc, env->xregs[31]);
128
    for (i = 0; i < 31; i++) {
129
        cpu_fprintf(f, "X%02d=%016"PRIx64, i, env->xregs[i]);
130
        if ((i % 4) == 3) {
131
            cpu_fprintf(f, "\n");
132
        } else {
133
            cpu_fprintf(f, " ");
134
        }
135
    }
136
    cpu_fprintf(f, "PSTATE=%08x (flags %c%c%c%c)\n",
137
                psr,
138
                psr & PSTATE_N ? 'N' : '-',
139
                psr & PSTATE_Z ? 'Z' : '-',
140
                psr & PSTATE_C ? 'C' : '-',
141
                psr & PSTATE_V ? 'V' : '-');
142
    cpu_fprintf(f, "\n");
143

    
144
    if (flags & CPU_DUMP_FPU) {
145
        int numvfpregs = 32;
146
        for (i = 0; i < numvfpregs; i += 2) {
147
            uint64_t vlo = float64_val(env->vfp.regs[i * 2]);
148
            uint64_t vhi = float64_val(env->vfp.regs[(i * 2) + 1]);
149
            cpu_fprintf(f, "q%02d=%016" PRIx64 ":%016" PRIx64 " ",
150
                        i, vhi, vlo);
151
            vlo = float64_val(env->vfp.regs[(i + 1) * 2]);
152
            vhi = float64_val(env->vfp.regs[((i + 1) * 2) + 1]);
153
            cpu_fprintf(f, "q%02d=%016" PRIx64 ":%016" PRIx64 "\n",
154
                        i + 1, vhi, vlo);
155
        }
156
        cpu_fprintf(f, "FPCR: %08x  FPSR: %08x\n",
157
                    vfp_get_fpcr(env), vfp_get_fpsr(env));
158
    }
159
}
160

    
161
static int get_mem_index(DisasContext *s)
162
{
163
#ifdef CONFIG_USER_ONLY
164
    return 1;
165
#else
166
    return s->user;
167
#endif
168
}
169

    
170
void gen_a64_set_pc_im(uint64_t val)
171
{
172
    tcg_gen_movi_i64(cpu_pc, val);
173
}
174

    
175
static void gen_exception(int excp)
176
{
177
    TCGv_i32 tmp = tcg_temp_new_i32();
178
    tcg_gen_movi_i32(tmp, excp);
179
    gen_helper_exception(cpu_env, tmp);
180
    tcg_temp_free_i32(tmp);
181
}
182

    
183
static void gen_exception_insn(DisasContext *s, int offset, int excp)
184
{
185
    gen_a64_set_pc_im(s->pc - offset);
186
    gen_exception(excp);
187
    s->is_jmp = DISAS_EXC;
188
}
189

    
190
static inline bool use_goto_tb(DisasContext *s, int n, uint64_t dest)
191
{
192
    /* No direct tb linking with singlestep or deterministic io */
193
    if (s->singlestep_enabled || (s->tb->cflags & CF_LAST_IO)) {
194
        return false;
195
    }
196

    
197
    /* Only link tbs from inside the same guest page */
198
    if ((s->tb->pc & TARGET_PAGE_MASK) != (dest & TARGET_PAGE_MASK)) {
199
        return false;
200
    }
201

    
202
    return true;
203
}
204

    
205
static inline void gen_goto_tb(DisasContext *s, int n, uint64_t dest)
206
{
207
    TranslationBlock *tb;
208

    
209
    tb = s->tb;
210
    if (use_goto_tb(s, n, dest)) {
211
        tcg_gen_goto_tb(n);
212
        gen_a64_set_pc_im(dest);
213
        tcg_gen_exit_tb((tcg_target_long)tb + n);
214
        s->is_jmp = DISAS_TB_JUMP;
215
    } else {
216
        gen_a64_set_pc_im(dest);
217
        if (s->singlestep_enabled) {
218
            gen_exception(EXCP_DEBUG);
219
        }
220
        tcg_gen_exit_tb(0);
221
        s->is_jmp = DISAS_JUMP;
222
    }
223
}
224

    
225
static void unallocated_encoding(DisasContext *s)
226
{
227
    gen_exception_insn(s, 4, EXCP_UDEF);
228
}
229

    
230
#define unsupported_encoding(s, insn)                                    \
231
    do {                                                                 \
232
        qemu_log_mask(LOG_UNIMP,                                         \
233
                      "%s:%d: unsupported instruction encoding 0x%08x "  \
234
                      "at pc=%016" PRIx64 "\n",                          \
235
                      __FILE__, __LINE__, insn, s->pc - 4);              \
236
        unallocated_encoding(s);                                         \
237
    } while (0);
238

    
239
static void init_tmp_a64_array(DisasContext *s)
240
{
241
#ifdef CONFIG_DEBUG_TCG
242
    int i;
243
    for (i = 0; i < ARRAY_SIZE(s->tmp_a64); i++) {
244
        TCGV_UNUSED_I64(s->tmp_a64[i]);
245
    }
246
#endif
247
    s->tmp_a64_count = 0;
248
}
249

    
250
static void free_tmp_a64(DisasContext *s)
251
{
252
    int i;
253
    for (i = 0; i < s->tmp_a64_count; i++) {
254
        tcg_temp_free_i64(s->tmp_a64[i]);
255
    }
256
    init_tmp_a64_array(s);
257
}
258

    
259
static TCGv_i64 new_tmp_a64(DisasContext *s)
260
{
261
    assert(s->tmp_a64_count < TMP_A64_MAX);
262
    return s->tmp_a64[s->tmp_a64_count++] = tcg_temp_new_i64();
263
}
264

    
265
static TCGv_i64 new_tmp_a64_zero(DisasContext *s)
266
{
267
    TCGv_i64 t = new_tmp_a64(s);
268
    tcg_gen_movi_i64(t, 0);
269
    return t;
270
}
271

    
272
/*
273
 * Register access functions
274
 *
275
 * These functions are used for directly accessing a register in where
276
 * changes to the final register value are likely to be made. If you
277
 * need to use a register for temporary calculation (e.g. index type
278
 * operations) use the read_* form.
279
 *
280
 * B1.2.1 Register mappings
281
 *
282
 * In instruction register encoding 31 can refer to ZR (zero register) or
283
 * the SP (stack pointer) depending on context. In QEMU's case we map SP
284
 * to cpu_X[31] and ZR accesses to a temporary which can be discarded.
285
 * This is the point of the _sp forms.
286
 */
287
static TCGv_i64 cpu_reg(DisasContext *s, int reg)
288
{
289
    if (reg == 31) {
290
        return new_tmp_a64_zero(s);
291
    } else {
292
        return cpu_X[reg];
293
    }
294
}
295

    
296
/* register access for when 31 == SP */
297
static TCGv_i64 cpu_reg_sp(DisasContext *s, int reg)
298
{
299
    return cpu_X[reg];
300
}
301

    
302
/* read a cpu register in 32bit/64bit mode. Returns a TCGv_i64
303
 * representing the register contents. This TCGv is an auto-freed
304
 * temporary so it need not be explicitly freed, and may be modified.
305
 */
306
static TCGv_i64 read_cpu_reg(DisasContext *s, int reg, int sf)
307
{
308
    TCGv_i64 v = new_tmp_a64(s);
309
    if (reg != 31) {
310
        if (sf) {
311
            tcg_gen_mov_i64(v, cpu_X[reg]);
312
        } else {
313
            tcg_gen_ext32u_i64(v, cpu_X[reg]);
314
        }
315
    } else {
316
        tcg_gen_movi_i64(v, 0);
317
    }
318
    return v;
319
}
320

    
321
static TCGv_i64 read_cpu_reg_sp(DisasContext *s, int reg, int sf)
322
{
323
    TCGv_i64 v = new_tmp_a64(s);
324
    if (sf) {
325
        tcg_gen_mov_i64(v, cpu_X[reg]);
326
    } else {
327
        tcg_gen_ext32u_i64(v, cpu_X[reg]);
328
    }
329
    return v;
330
}
331

    
332
/* Return the offset into CPUARMState of an element of specified
333
 * size, 'element' places in from the least significant end of
334
 * the FP/vector register Qn.
335
 */
336
static inline int vec_reg_offset(int regno, int element, TCGMemOp size)
337
{
338
    int offs = offsetof(CPUARMState, vfp.regs[regno * 2]);
339
#ifdef HOST_WORDS_BIGENDIAN
340
    /* This is complicated slightly because vfp.regs[2n] is
341
     * still the low half and  vfp.regs[2n+1] the high half
342
     * of the 128 bit vector, even on big endian systems.
343
     * Calculate the offset assuming a fully bigendian 128 bits,
344
     * then XOR to account for the order of the two 64 bit halves.
345
     */
346
    offs += (16 - ((element + 1) * (1 << size)));
347
    offs ^= 8;
348
#else
349
    offs += element * (1 << size);
350
#endif
351
    return offs;
352
}
353

    
354
/* Return the offset into CPUARMState of a slice (from
355
 * the least significant end) of FP register Qn (ie
356
 * Dn, Sn, Hn or Bn).
357
 * (Note that this is not the same mapping as for A32; see cpu.h)
358
 */
359
static inline int fp_reg_offset(int regno, TCGMemOp size)
360
{
361
    int offs = offsetof(CPUARMState, vfp.regs[regno * 2]);
362
#ifdef HOST_WORDS_BIGENDIAN
363
    offs += (8 - (1 << size));
364
#endif
365
    return offs;
366
}
367

    
368
/* Offset of the high half of the 128 bit vector Qn */
369
static inline int fp_reg_hi_offset(int regno)
370
{
371
    return offsetof(CPUARMState, vfp.regs[regno * 2 + 1]);
372
}
373

    
374
/* Convenience accessors for reading and writing single and double
375
 * FP registers. Writing clears the upper parts of the associated
376
 * 128 bit vector register, as required by the architecture.
377
 * Note that unlike the GP register accessors, the values returned
378
 * by the read functions must be manually freed.
379
 */
380
static TCGv_i64 read_fp_dreg(DisasContext *s, int reg)
381
{
382
    TCGv_i64 v = tcg_temp_new_i64();
383

    
384
    tcg_gen_ld_i64(v, cpu_env, fp_reg_offset(reg, MO_64));
385
    return v;
386
}
387

    
388
static TCGv_i32 read_fp_sreg(DisasContext *s, int reg)
389
{
390
    TCGv_i32 v = tcg_temp_new_i32();
391

    
392
    tcg_gen_ld_i32(v, cpu_env, fp_reg_offset(reg, MO_32));
393
    return v;
394
}
395

    
396
static void write_fp_dreg(DisasContext *s, int reg, TCGv_i64 v)
397
{
398
    TCGv_i64 tcg_zero = tcg_const_i64(0);
399

    
400
    tcg_gen_st_i64(v, cpu_env, fp_reg_offset(reg, MO_64));
401
    tcg_gen_st_i64(tcg_zero, cpu_env, fp_reg_hi_offset(reg));
402
    tcg_temp_free_i64(tcg_zero);
403
}
404

    
405
static void write_fp_sreg(DisasContext *s, int reg, TCGv_i32 v)
406
{
407
    TCGv_i64 tmp = tcg_temp_new_i64();
408

    
409
    tcg_gen_extu_i32_i64(tmp, v);
410
    write_fp_dreg(s, reg, tmp);
411
    tcg_temp_free_i64(tmp);
412
}
413

    
414
static TCGv_ptr get_fpstatus_ptr(void)
415
{
416
    TCGv_ptr statusptr = tcg_temp_new_ptr();
417
    int offset;
418

    
419
    /* In A64 all instructions (both FP and Neon) use the FPCR;
420
     * there is no equivalent of the A32 Neon "standard FPSCR value"
421
     * and all operations use vfp.fp_status.
422
     */
423
    offset = offsetof(CPUARMState, vfp.fp_status);
424
    tcg_gen_addi_ptr(statusptr, cpu_env, offset);
425
    return statusptr;
426
}
427

    
428
/* Set ZF and NF based on a 64 bit result. This is alas fiddlier
429
 * than the 32 bit equivalent.
430
 */
431
static inline void gen_set_NZ64(TCGv_i64 result)
432
{
433
    TCGv_i64 flag = tcg_temp_new_i64();
434

    
435
    tcg_gen_setcondi_i64(TCG_COND_NE, flag, result, 0);
436
    tcg_gen_trunc_i64_i32(cpu_ZF, flag);
437
    tcg_gen_shri_i64(flag, result, 32);
438
    tcg_gen_trunc_i64_i32(cpu_NF, flag);
439
    tcg_temp_free_i64(flag);
440
}
441

    
442
/* Set NZCV as for a logical operation: NZ as per result, CV cleared. */
443
static inline void gen_logic_CC(int sf, TCGv_i64 result)
444
{
445
    if (sf) {
446
        gen_set_NZ64(result);
447
    } else {
448
        tcg_gen_trunc_i64_i32(cpu_ZF, result);
449
        tcg_gen_trunc_i64_i32(cpu_NF, result);
450
    }
451
    tcg_gen_movi_i32(cpu_CF, 0);
452
    tcg_gen_movi_i32(cpu_VF, 0);
453
}
454

    
455
/* dest = T0 + T1; compute C, N, V and Z flags */
456
static void gen_add_CC(int sf, TCGv_i64 dest, TCGv_i64 t0, TCGv_i64 t1)
457
{
458
    if (sf) {
459
        TCGv_i64 result, flag, tmp;
460
        result = tcg_temp_new_i64();
461
        flag = tcg_temp_new_i64();
462
        tmp = tcg_temp_new_i64();
463

    
464
        tcg_gen_movi_i64(tmp, 0);
465
        tcg_gen_add2_i64(result, flag, t0, tmp, t1, tmp);
466

    
467
        tcg_gen_trunc_i64_i32(cpu_CF, flag);
468

    
469
        gen_set_NZ64(result);
470

    
471
        tcg_gen_xor_i64(flag, result, t0);
472
        tcg_gen_xor_i64(tmp, t0, t1);
473
        tcg_gen_andc_i64(flag, flag, tmp);
474
        tcg_temp_free_i64(tmp);
475
        tcg_gen_shri_i64(flag, flag, 32);
476
        tcg_gen_trunc_i64_i32(cpu_VF, flag);
477

    
478
        tcg_gen_mov_i64(dest, result);
479
        tcg_temp_free_i64(result);
480
        tcg_temp_free_i64(flag);
481
    } else {
482
        /* 32 bit arithmetic */
483
        TCGv_i32 t0_32 = tcg_temp_new_i32();
484
        TCGv_i32 t1_32 = tcg_temp_new_i32();
485
        TCGv_i32 tmp = tcg_temp_new_i32();
486

    
487
        tcg_gen_movi_i32(tmp, 0);
488
        tcg_gen_trunc_i64_i32(t0_32, t0);
489
        tcg_gen_trunc_i64_i32(t1_32, t1);
490
        tcg_gen_add2_i32(cpu_NF, cpu_CF, t0_32, tmp, t1_32, tmp);
491
        tcg_gen_mov_i32(cpu_ZF, cpu_NF);
492
        tcg_gen_xor_i32(cpu_VF, cpu_NF, t0_32);
493
        tcg_gen_xor_i32(tmp, t0_32, t1_32);
494
        tcg_gen_andc_i32(cpu_VF, cpu_VF, tmp);
495
        tcg_gen_extu_i32_i64(dest, cpu_NF);
496

    
497
        tcg_temp_free_i32(tmp);
498
        tcg_temp_free_i32(t0_32);
499
        tcg_temp_free_i32(t1_32);
500
    }
501
}
502

    
503
/* dest = T0 - T1; compute C, N, V and Z flags */
504
static void gen_sub_CC(int sf, TCGv_i64 dest, TCGv_i64 t0, TCGv_i64 t1)
505
{
506
    if (sf) {
507
        /* 64 bit arithmetic */
508
        TCGv_i64 result, flag, tmp;
509

    
510
        result = tcg_temp_new_i64();
511
        flag = tcg_temp_new_i64();
512
        tcg_gen_sub_i64(result, t0, t1);
513

    
514
        gen_set_NZ64(result);
515

    
516
        tcg_gen_setcond_i64(TCG_COND_GEU, flag, t0, t1);
517
        tcg_gen_trunc_i64_i32(cpu_CF, flag);
518

    
519
        tcg_gen_xor_i64(flag, result, t0);
520
        tmp = tcg_temp_new_i64();
521
        tcg_gen_xor_i64(tmp, t0, t1);
522
        tcg_gen_and_i64(flag, flag, tmp);
523
        tcg_temp_free_i64(tmp);
524
        tcg_gen_shri_i64(flag, flag, 32);
525
        tcg_gen_trunc_i64_i32(cpu_VF, flag);
526
        tcg_gen_mov_i64(dest, result);
527
        tcg_temp_free_i64(flag);
528
        tcg_temp_free_i64(result);
529
    } else {
530
        /* 32 bit arithmetic */
531
        TCGv_i32 t0_32 = tcg_temp_new_i32();
532
        TCGv_i32 t1_32 = tcg_temp_new_i32();
533
        TCGv_i32 tmp;
534

    
535
        tcg_gen_trunc_i64_i32(t0_32, t0);
536
        tcg_gen_trunc_i64_i32(t1_32, t1);
537
        tcg_gen_sub_i32(cpu_NF, t0_32, t1_32);
538
        tcg_gen_mov_i32(cpu_ZF, cpu_NF);
539
        tcg_gen_setcond_i32(TCG_COND_GEU, cpu_CF, t0_32, t1_32);
540
        tcg_gen_xor_i32(cpu_VF, cpu_NF, t0_32);
541
        tmp = tcg_temp_new_i32();
542
        tcg_gen_xor_i32(tmp, t0_32, t1_32);
543
        tcg_temp_free_i32(t0_32);
544
        tcg_temp_free_i32(t1_32);
545
        tcg_gen_and_i32(cpu_VF, cpu_VF, tmp);
546
        tcg_temp_free_i32(tmp);
547
        tcg_gen_extu_i32_i64(dest, cpu_NF);
548
    }
549
}
550

    
551
/* dest = T0 + T1 + CF; do not compute flags. */
552
static void gen_adc(int sf, TCGv_i64 dest, TCGv_i64 t0, TCGv_i64 t1)
553
{
554
    TCGv_i64 flag = tcg_temp_new_i64();
555
    tcg_gen_extu_i32_i64(flag, cpu_CF);
556
    tcg_gen_add_i64(dest, t0, t1);
557
    tcg_gen_add_i64(dest, dest, flag);
558
    tcg_temp_free_i64(flag);
559

    
560
    if (!sf) {
561
        tcg_gen_ext32u_i64(dest, dest);
562
    }
563
}
564

    
565
/* dest = T0 + T1 + CF; compute C, N, V and Z flags. */
566
static void gen_adc_CC(int sf, TCGv_i64 dest, TCGv_i64 t0, TCGv_i64 t1)
567
{
568
    if (sf) {
569
        TCGv_i64 result, cf_64, vf_64, tmp;
570
        result = tcg_temp_new_i64();
571
        cf_64 = tcg_temp_new_i64();
572
        vf_64 = tcg_temp_new_i64();
573
        tmp = tcg_const_i64(0);
574

    
575
        tcg_gen_extu_i32_i64(cf_64, cpu_CF);
576
        tcg_gen_add2_i64(result, cf_64, t0, tmp, cf_64, tmp);
577
        tcg_gen_add2_i64(result, cf_64, result, cf_64, t1, tmp);
578
        tcg_gen_trunc_i64_i32(cpu_CF, cf_64);
579
        gen_set_NZ64(result);
580

    
581
        tcg_gen_xor_i64(vf_64, result, t0);
582
        tcg_gen_xor_i64(tmp, t0, t1);
583
        tcg_gen_andc_i64(vf_64, vf_64, tmp);
584
        tcg_gen_shri_i64(vf_64, vf_64, 32);
585
        tcg_gen_trunc_i64_i32(cpu_VF, vf_64);
586

    
587
        tcg_gen_mov_i64(dest, result);
588

    
589
        tcg_temp_free_i64(tmp);
590
        tcg_temp_free_i64(vf_64);
591
        tcg_temp_free_i64(cf_64);
592
        tcg_temp_free_i64(result);
593
    } else {
594
        TCGv_i32 t0_32, t1_32, tmp;
595
        t0_32 = tcg_temp_new_i32();
596
        t1_32 = tcg_temp_new_i32();
597
        tmp = tcg_const_i32(0);
598

    
599
        tcg_gen_trunc_i64_i32(t0_32, t0);
600
        tcg_gen_trunc_i64_i32(t1_32, t1);
601
        tcg_gen_add2_i32(cpu_NF, cpu_CF, t0_32, tmp, cpu_CF, tmp);
602
        tcg_gen_add2_i32(cpu_NF, cpu_CF, cpu_NF, cpu_CF, t1_32, tmp);
603

    
604
        tcg_gen_mov_i32(cpu_ZF, cpu_NF);
605
        tcg_gen_xor_i32(cpu_VF, cpu_NF, t0_32);
606
        tcg_gen_xor_i32(tmp, t0_32, t1_32);
607
        tcg_gen_andc_i32(cpu_VF, cpu_VF, tmp);
608
        tcg_gen_extu_i32_i64(dest, cpu_NF);
609

    
610
        tcg_temp_free_i32(tmp);
611
        tcg_temp_free_i32(t1_32);
612
        tcg_temp_free_i32(t0_32);
613
    }
614
}
615

    
616
/*
617
 * Load/Store generators
618
 */
619

    
620
/*
621
 * Store from GPR register to memory
622
 */
623
static void do_gpr_st(DisasContext *s, TCGv_i64 source,
624
                      TCGv_i64 tcg_addr, int size)
625
{
626
    g_assert(size <= 3);
627
    tcg_gen_qemu_st_i64(source, tcg_addr, get_mem_index(s), MO_TE + size);
628
}
629

    
630
/*
631
 * Load from memory to GPR register
632
 */
633
static void do_gpr_ld(DisasContext *s, TCGv_i64 dest, TCGv_i64 tcg_addr,
634
                      int size, bool is_signed, bool extend)
635
{
636
    TCGMemOp memop = MO_TE + size;
637

    
638
    g_assert(size <= 3);
639

    
640
    if (is_signed) {
641
        memop += MO_SIGN;
642
    }
643

    
644
    tcg_gen_qemu_ld_i64(dest, tcg_addr, get_mem_index(s), memop);
645

    
646
    if (extend && is_signed) {
647
        g_assert(size < 3);
648
        tcg_gen_ext32u_i64(dest, dest);
649
    }
650
}
651

    
652
/*
653
 * Store from FP register to memory
654
 */
655
static void do_fp_st(DisasContext *s, int srcidx, TCGv_i64 tcg_addr, int size)
656
{
657
    /* This writes the bottom N bits of a 128 bit wide vector to memory */
658
    TCGv_i64 tmp = tcg_temp_new_i64();
659
    tcg_gen_ld_i64(tmp, cpu_env, fp_reg_offset(srcidx, MO_64));
660
    if (size < 4) {
661
        tcg_gen_qemu_st_i64(tmp, tcg_addr, get_mem_index(s), MO_TE + size);
662
    } else {
663
        TCGv_i64 tcg_hiaddr = tcg_temp_new_i64();
664
        tcg_gen_qemu_st_i64(tmp, tcg_addr, get_mem_index(s), MO_TEQ);
665
        tcg_gen_qemu_st64(tmp, tcg_addr, get_mem_index(s));
666
        tcg_gen_ld_i64(tmp, cpu_env, fp_reg_hi_offset(srcidx));
667
        tcg_gen_addi_i64(tcg_hiaddr, tcg_addr, 8);
668
        tcg_gen_qemu_st_i64(tmp, tcg_hiaddr, get_mem_index(s), MO_TEQ);
669
        tcg_temp_free_i64(tcg_hiaddr);
670
    }
671

    
672
    tcg_temp_free_i64(tmp);
673
}
674

    
675
/*
676
 * Load from memory to FP register
677
 */
678
static void do_fp_ld(DisasContext *s, int destidx, TCGv_i64 tcg_addr, int size)
679
{
680
    /* This always zero-extends and writes to a full 128 bit wide vector */
681
    TCGv_i64 tmplo = tcg_temp_new_i64();
682
    TCGv_i64 tmphi;
683

    
684
    if (size < 4) {
685
        TCGMemOp memop = MO_TE + size;
686
        tmphi = tcg_const_i64(0);
687
        tcg_gen_qemu_ld_i64(tmplo, tcg_addr, get_mem_index(s), memop);
688
    } else {
689
        TCGv_i64 tcg_hiaddr;
690
        tmphi = tcg_temp_new_i64();
691
        tcg_hiaddr = tcg_temp_new_i64();
692

    
693
        tcg_gen_qemu_ld_i64(tmplo, tcg_addr, get_mem_index(s), MO_TEQ);
694
        tcg_gen_addi_i64(tcg_hiaddr, tcg_addr, 8);
695
        tcg_gen_qemu_ld_i64(tmphi, tcg_hiaddr, get_mem_index(s), MO_TEQ);
696
        tcg_temp_free_i64(tcg_hiaddr);
697
    }
698

    
699
    tcg_gen_st_i64(tmplo, cpu_env, fp_reg_offset(destidx, MO_64));
700
    tcg_gen_st_i64(tmphi, cpu_env, fp_reg_hi_offset(destidx));
701

    
702
    tcg_temp_free_i64(tmplo);
703
    tcg_temp_free_i64(tmphi);
704
}
705

    
706
/*
707
 * Vector load/store helpers.
708
 *
709
 * The principal difference between this and a FP load is that we don't
710
 * zero extend as we are filling a partial chunk of the vector register.
711
 * These functions don't support 128 bit loads/stores, which would be
712
 * normal load/store operations.
713
 *
714
 * The _i32 versions are useful when operating on 32 bit quantities
715
 * (eg for floating point single or using Neon helper functions).
716
 */
717

    
718
/* Get value of an element within a vector register */
719
static void read_vec_element(DisasContext *s, TCGv_i64 tcg_dest, int srcidx,
720
                             int element, TCGMemOp memop)
721
{
722
    int vect_off = vec_reg_offset(srcidx, element, memop & MO_SIZE);
723
    switch (memop) {
724
    case MO_8:
725
        tcg_gen_ld8u_i64(tcg_dest, cpu_env, vect_off);
726
        break;
727
    case MO_16:
728
        tcg_gen_ld16u_i64(tcg_dest, cpu_env, vect_off);
729
        break;
730
    case MO_32:
731
        tcg_gen_ld32u_i64(tcg_dest, cpu_env, vect_off);
732
        break;
733
    case MO_8|MO_SIGN:
734
        tcg_gen_ld8s_i64(tcg_dest, cpu_env, vect_off);
735
        break;
736
    case MO_16|MO_SIGN:
737
        tcg_gen_ld16s_i64(tcg_dest, cpu_env, vect_off);
738
        break;
739
    case MO_32|MO_SIGN:
740
        tcg_gen_ld32s_i64(tcg_dest, cpu_env, vect_off);
741
        break;
742
    case MO_64:
743
    case MO_64|MO_SIGN:
744
        tcg_gen_ld_i64(tcg_dest, cpu_env, vect_off);
745
        break;
746
    default:
747
        g_assert_not_reached();
748
    }
749
}
750

    
751
static void read_vec_element_i32(DisasContext *s, TCGv_i32 tcg_dest, int srcidx,
752
                                 int element, TCGMemOp memop)
753
{
754
    int vect_off = vec_reg_offset(srcidx, element, memop & MO_SIZE);
755
    switch (memop) {
756
    case MO_8:
757
        tcg_gen_ld8u_i32(tcg_dest, cpu_env, vect_off);
758
        break;
759
    case MO_16:
760
        tcg_gen_ld16u_i32(tcg_dest, cpu_env, vect_off);
761
        break;
762
    case MO_8|MO_SIGN:
763
        tcg_gen_ld8s_i32(tcg_dest, cpu_env, vect_off);
764
        break;
765
    case MO_16|MO_SIGN:
766
        tcg_gen_ld16s_i32(tcg_dest, cpu_env, vect_off);
767
        break;
768
    case MO_32:
769
    case MO_32|MO_SIGN:
770
        tcg_gen_ld_i32(tcg_dest, cpu_env, vect_off);
771
        break;
772
    default:
773
        g_assert_not_reached();
774
    }
775
}
776

    
777
/* Set value of an element within a vector register */
778
static void write_vec_element(DisasContext *s, TCGv_i64 tcg_src, int destidx,
779
                              int element, TCGMemOp memop)
780
{
781
    int vect_off = vec_reg_offset(destidx, element, memop & MO_SIZE);
782
    switch (memop) {
783
    case MO_8:
784
        tcg_gen_st8_i64(tcg_src, cpu_env, vect_off);
785
        break;
786
    case MO_16:
787
        tcg_gen_st16_i64(tcg_src, cpu_env, vect_off);
788
        break;
789
    case MO_32:
790
        tcg_gen_st32_i64(tcg_src, cpu_env, vect_off);
791
        break;
792
    case MO_64:
793
        tcg_gen_st_i64(tcg_src, cpu_env, vect_off);
794
        break;
795
    default:
796
        g_assert_not_reached();
797
    }
798
}
799

    
800
static void write_vec_element_i32(DisasContext *s, TCGv_i32 tcg_src,
801
                                  int destidx, int element, TCGMemOp memop)
802
{
803
    int vect_off = vec_reg_offset(destidx, element, memop & MO_SIZE);
804
    switch (memop) {
805
    case MO_8:
806
        tcg_gen_st8_i32(tcg_src, cpu_env, vect_off);
807
        break;
808
    case MO_16:
809
        tcg_gen_st16_i32(tcg_src, cpu_env, vect_off);
810
        break;
811
    case MO_32:
812
        tcg_gen_st_i32(tcg_src, cpu_env, vect_off);
813
        break;
814
    default:
815
        g_assert_not_reached();
816
    }
817
}
818

    
819
/* Clear the high 64 bits of a 128 bit vector (in general non-quad
820
 * vector ops all need to do this).
821
 */
822
static void clear_vec_high(DisasContext *s, int rd)
823
{
824
    TCGv_i64 tcg_zero = tcg_const_i64(0);
825

    
826
    write_vec_element(s, tcg_zero, rd, 1, MO_64);
827
    tcg_temp_free_i64(tcg_zero);
828
}
829

    
830
/* Store from vector register to memory */
831
static void do_vec_st(DisasContext *s, int srcidx, int element,
832
                      TCGv_i64 tcg_addr, int size)
833
{
834
    TCGMemOp memop = MO_TE + size;
835
    TCGv_i64 tcg_tmp = tcg_temp_new_i64();
836

    
837
    read_vec_element(s, tcg_tmp, srcidx, element, size);
838
    tcg_gen_qemu_st_i64(tcg_tmp, tcg_addr, get_mem_index(s), memop);
839

    
840
    tcg_temp_free_i64(tcg_tmp);
841
}
842

    
843
/* Load from memory to vector register */
844
static void do_vec_ld(DisasContext *s, int destidx, int element,
845
                      TCGv_i64 tcg_addr, int size)
846
{
847
    TCGMemOp memop = MO_TE + size;
848
    TCGv_i64 tcg_tmp = tcg_temp_new_i64();
849

    
850
    tcg_gen_qemu_ld_i64(tcg_tmp, tcg_addr, get_mem_index(s), memop);
851
    write_vec_element(s, tcg_tmp, destidx, element, size);
852

    
853
    tcg_temp_free_i64(tcg_tmp);
854
}
855

    
856
/*
857
 * This utility function is for doing register extension with an
858
 * optional shift. You will likely want to pass a temporary for the
859
 * destination register. See DecodeRegExtend() in the ARM ARM.
860
 */
861
static void ext_and_shift_reg(TCGv_i64 tcg_out, TCGv_i64 tcg_in,
862
                              int option, unsigned int shift)
863
{
864
    int extsize = extract32(option, 0, 2);
865
    bool is_signed = extract32(option, 2, 1);
866

    
867
    if (is_signed) {
868
        switch (extsize) {
869
        case 0:
870
            tcg_gen_ext8s_i64(tcg_out, tcg_in);
871
            break;
872
        case 1:
873
            tcg_gen_ext16s_i64(tcg_out, tcg_in);
874
            break;
875
        case 2:
876
            tcg_gen_ext32s_i64(tcg_out, tcg_in);
877
            break;
878
        case 3:
879
            tcg_gen_mov_i64(tcg_out, tcg_in);
880
            break;
881
        }
882
    } else {
883
        switch (extsize) {
884
        case 0:
885
            tcg_gen_ext8u_i64(tcg_out, tcg_in);
886
            break;
887
        case 1:
888
            tcg_gen_ext16u_i64(tcg_out, tcg_in);
889
            break;
890
        case 2:
891
            tcg_gen_ext32u_i64(tcg_out, tcg_in);
892
            break;
893
        case 3:
894
            tcg_gen_mov_i64(tcg_out, tcg_in);
895
            break;
896
        }
897
    }
898

    
899
    if (shift) {
900
        tcg_gen_shli_i64(tcg_out, tcg_out, shift);
901
    }
902
}
903

    
904
static inline void gen_check_sp_alignment(DisasContext *s)
905
{
906
    /* The AArch64 architecture mandates that (if enabled via PSTATE
907
     * or SCTLR bits) there is a check that SP is 16-aligned on every
908
     * SP-relative load or store (with an exception generated if it is not).
909
     * In line with general QEMU practice regarding misaligned accesses,
910
     * we omit these checks for the sake of guest program performance.
911
     * This function is provided as a hook so we can more easily add these
912
     * checks in future (possibly as a "favour catching guest program bugs
913
     * over speed" user selectable option).
914
     */
915
}
916

    
917
/*
918
 * This provides a simple table based table lookup decoder. It is
919
 * intended to be used when the relevant bits for decode are too
920
 * awkwardly placed and switch/if based logic would be confusing and
921
 * deeply nested. Since it's a linear search through the table, tables
922
 * should be kept small.
923
 *
924
 * It returns the first handler where insn & mask == pattern, or
925
 * NULL if there is no match.
926
 * The table is terminated by an empty mask (i.e. 0)
927
 */
928
static inline AArch64DecodeFn *lookup_disas_fn(const AArch64DecodeTable *table,
929
                                               uint32_t insn)
930
{
931
    const AArch64DecodeTable *tptr = table;
932

    
933
    while (tptr->mask) {
934
        if ((insn & tptr->mask) == tptr->pattern) {
935
            return tptr->disas_fn;
936
        }
937
        tptr++;
938
    }
939
    return NULL;
940
}
941

    
942
/*
943
 * the instruction disassembly implemented here matches
944
 * the instruction encoding classifications in chapter 3 (C3)
945
 * of the ARM Architecture Reference Manual (DDI0487A_a)
946
 */
947

    
948
/* C3.2.7 Unconditional branch (immediate)
949
 *   31  30       26 25                                  0
950
 * +----+-----------+-------------------------------------+
951
 * | op | 0 0 1 0 1 |                 imm26               |
952
 * +----+-----------+-------------------------------------+
953
 */
954
static void disas_uncond_b_imm(DisasContext *s, uint32_t insn)
955
{
956
    uint64_t addr = s->pc + sextract32(insn, 0, 26) * 4 - 4;
957

    
958
    if (insn & (1 << 31)) {
959
        /* C5.6.26 BL Branch with link */
960
        tcg_gen_movi_i64(cpu_reg(s, 30), s->pc);
961
    }
962

    
963
    /* C5.6.20 B Branch / C5.6.26 BL Branch with link */
964
    gen_goto_tb(s, 0, addr);
965
}
966

    
967
/* C3.2.1 Compare & branch (immediate)
968
 *   31  30         25  24  23                  5 4      0
969
 * +----+-------------+----+---------------------+--------+
970
 * | sf | 0 1 1 0 1 0 | op |         imm19       |   Rt   |
971
 * +----+-------------+----+---------------------+--------+
972
 */
973
static void disas_comp_b_imm(DisasContext *s, uint32_t insn)
974
{
975
    unsigned int sf, op, rt;
976
    uint64_t addr;
977
    int label_match;
978
    TCGv_i64 tcg_cmp;
979

    
980
    sf = extract32(insn, 31, 1);
981
    op = extract32(insn, 24, 1); /* 0: CBZ; 1: CBNZ */
982
    rt = extract32(insn, 0, 5);
983
    addr = s->pc + sextract32(insn, 5, 19) * 4 - 4;
984

    
985
    tcg_cmp = read_cpu_reg(s, rt, sf);
986
    label_match = gen_new_label();
987

    
988
    tcg_gen_brcondi_i64(op ? TCG_COND_NE : TCG_COND_EQ,
989
                        tcg_cmp, 0, label_match);
990

    
991
    gen_goto_tb(s, 0, s->pc);
992
    gen_set_label(label_match);
993
    gen_goto_tb(s, 1, addr);
994
}
995

    
996
/* C3.2.5 Test & branch (immediate)
997
 *   31  30         25  24  23   19 18          5 4    0
998
 * +----+-------------+----+-------+-------------+------+
999
 * | b5 | 0 1 1 0 1 1 | op |  b40  |    imm14    |  Rt  |
1000
 * +----+-------------+----+-------+-------------+------+
1001
 */
1002
static void disas_test_b_imm(DisasContext *s, uint32_t insn)
1003
{
1004
    unsigned int bit_pos, op, rt;
1005
    uint64_t addr;
1006
    int label_match;
1007
    TCGv_i64 tcg_cmp;
1008

    
1009
    bit_pos = (extract32(insn, 31, 1) << 5) | extract32(insn, 19, 5);
1010
    op = extract32(insn, 24, 1); /* 0: TBZ; 1: TBNZ */
1011
    addr = s->pc + sextract32(insn, 5, 14) * 4 - 4;
1012
    rt = extract32(insn, 0, 5);
1013

    
1014
    tcg_cmp = tcg_temp_new_i64();
1015
    tcg_gen_andi_i64(tcg_cmp, cpu_reg(s, rt), (1ULL << bit_pos));
1016
    label_match = gen_new_label();
1017
    tcg_gen_brcondi_i64(op ? TCG_COND_NE : TCG_COND_EQ,
1018
                        tcg_cmp, 0, label_match);
1019
    tcg_temp_free_i64(tcg_cmp);
1020
    gen_goto_tb(s, 0, s->pc);
1021
    gen_set_label(label_match);
1022
    gen_goto_tb(s, 1, addr);
1023
}
1024

    
1025
/* C3.2.2 / C5.6.19 Conditional branch (immediate)
1026
 *  31           25  24  23                  5   4  3    0
1027
 * +---------------+----+---------------------+----+------+
1028
 * | 0 1 0 1 0 1 0 | o1 |         imm19       | o0 | cond |
1029
 * +---------------+----+---------------------+----+------+
1030
 */
1031
static void disas_cond_b_imm(DisasContext *s, uint32_t insn)
1032
{
1033
    unsigned int cond;
1034
    uint64_t addr;
1035

    
1036
    if ((insn & (1 << 4)) || (insn & (1 << 24))) {
1037
        unallocated_encoding(s);
1038
        return;
1039
    }
1040
    addr = s->pc + sextract32(insn, 5, 19) * 4 - 4;
1041
    cond = extract32(insn, 0, 4);
1042

    
1043
    if (cond < 0x0e) {
1044
        /* genuinely conditional branches */
1045
        int label_match = gen_new_label();
1046
        arm_gen_test_cc(cond, label_match);
1047
        gen_goto_tb(s, 0, s->pc);
1048
        gen_set_label(label_match);
1049
        gen_goto_tb(s, 1, addr);
1050
    } else {
1051
        /* 0xe and 0xf are both "always" conditions */
1052
        gen_goto_tb(s, 0, addr);
1053
    }
1054
}
1055

    
1056
/* C5.6.68 HINT */
1057
static void handle_hint(DisasContext *s, uint32_t insn,
1058
                        unsigned int op1, unsigned int op2, unsigned int crm)
1059
{
1060
    unsigned int selector = crm << 3 | op2;
1061

    
1062
    if (op1 != 3) {
1063
        unallocated_encoding(s);
1064
        return;
1065
    }
1066

    
1067
    switch (selector) {
1068
    case 0: /* NOP */
1069
        return;
1070
    case 1: /* YIELD */
1071
    case 2: /* WFE */
1072
    case 3: /* WFI */
1073
    case 4: /* SEV */
1074
    case 5: /* SEVL */
1075
        /* we treat all as NOP at least for now */
1076
        return;
1077
    default:
1078
        /* default specified as NOP equivalent */
1079
        return;
1080
    }
1081
}
1082

    
1083
static void gen_clrex(DisasContext *s, uint32_t insn)
1084
{
1085
    tcg_gen_movi_i64(cpu_exclusive_addr, -1);
1086
}
1087

    
1088
/* CLREX, DSB, DMB, ISB */
1089
static void handle_sync(DisasContext *s, uint32_t insn,
1090
                        unsigned int op1, unsigned int op2, unsigned int crm)
1091
{
1092
    if (op1 != 3) {
1093
        unallocated_encoding(s);
1094
        return;
1095
    }
1096

    
1097
    switch (op2) {
1098
    case 2: /* CLREX */
1099
        gen_clrex(s, insn);
1100
        return;
1101
    case 4: /* DSB */
1102
    case 5: /* DMB */
1103
    case 6: /* ISB */
1104
        /* We don't emulate caches so barriers are no-ops */
1105
        return;
1106
    default:
1107
        unallocated_encoding(s);
1108
        return;
1109
    }
1110
}
1111

    
1112
/* C5.6.130 MSR (immediate) - move immediate to processor state field */
1113
static void handle_msr_i(DisasContext *s, uint32_t insn,
1114
                         unsigned int op1, unsigned int op2, unsigned int crm)
1115
{
1116
    unsupported_encoding(s, insn);
1117
}
1118

    
1119
static void gen_get_nzcv(TCGv_i64 tcg_rt)
1120
{
1121
    TCGv_i32 tmp = tcg_temp_new_i32();
1122
    TCGv_i32 nzcv = tcg_temp_new_i32();
1123

    
1124
    /* build bit 31, N */
1125
    tcg_gen_andi_i32(nzcv, cpu_NF, (1 << 31));
1126
    /* build bit 30, Z */
1127
    tcg_gen_setcondi_i32(TCG_COND_EQ, tmp, cpu_ZF, 0);
1128
    tcg_gen_deposit_i32(nzcv, nzcv, tmp, 30, 1);
1129
    /* build bit 29, C */
1130
    tcg_gen_deposit_i32(nzcv, nzcv, cpu_CF, 29, 1);
1131
    /* build bit 28, V */
1132
    tcg_gen_shri_i32(tmp, cpu_VF, 31);
1133
    tcg_gen_deposit_i32(nzcv, nzcv, tmp, 28, 1);
1134
    /* generate result */
1135
    tcg_gen_extu_i32_i64(tcg_rt, nzcv);
1136

    
1137
    tcg_temp_free_i32(nzcv);
1138
    tcg_temp_free_i32(tmp);
1139
}
1140

    
1141
static void gen_set_nzcv(TCGv_i64 tcg_rt)
1142

    
1143
{
1144
    TCGv_i32 nzcv = tcg_temp_new_i32();
1145

    
1146
    /* take NZCV from R[t] */
1147
    tcg_gen_trunc_i64_i32(nzcv, tcg_rt);
1148

    
1149
    /* bit 31, N */
1150
    tcg_gen_andi_i32(cpu_NF, nzcv, (1 << 31));
1151
    /* bit 30, Z */
1152
    tcg_gen_andi_i32(cpu_ZF, nzcv, (1 << 30));
1153
    tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_ZF, cpu_ZF, 0);
1154
    /* bit 29, C */
1155
    tcg_gen_andi_i32(cpu_CF, nzcv, (1 << 29));
1156
    tcg_gen_shri_i32(cpu_CF, cpu_CF, 29);
1157
    /* bit 28, V */
1158
    tcg_gen_andi_i32(cpu_VF, nzcv, (1 << 28));
1159
    tcg_gen_shli_i32(cpu_VF, cpu_VF, 3);
1160
    tcg_temp_free_i32(nzcv);
1161
}
1162

    
1163
/* C5.6.129 MRS - move from system register
1164
 * C5.6.131 MSR (register) - move to system register
1165
 * C5.6.204 SYS
1166
 * C5.6.205 SYSL
1167
 * These are all essentially the same insn in 'read' and 'write'
1168
 * versions, with varying op0 fields.
1169
 */
1170
static void handle_sys(DisasContext *s, uint32_t insn, bool isread,
1171
                       unsigned int op0, unsigned int op1, unsigned int op2,
1172
                       unsigned int crn, unsigned int crm, unsigned int rt)
1173
{
1174
    const ARMCPRegInfo *ri;
1175
    TCGv_i64 tcg_rt;
1176

    
1177
    ri = get_arm_cp_reginfo(s->cp_regs,
1178
                            ENCODE_AA64_CP_REG(CP_REG_ARM64_SYSREG_CP,
1179
                                               crn, crm, op0, op1, op2));
1180

    
1181
    if (!ri) {
1182
        /* Unknown register; this might be a guest error or a QEMU
1183
         * unimplemented feature.
1184
         */
1185
        qemu_log_mask(LOG_UNIMP, "%s access to unsupported AArch64 "
1186
                      "system register op0:%d op1:%d crn:%d crm:%d op2:%d\n",
1187
                      isread ? "read" : "write", op0, op1, crn, crm, op2);
1188
        unallocated_encoding(s);
1189
        return;
1190
    }
1191

    
1192
    /* Check access permissions */
1193
    if (!cp_access_ok(s->current_pl, ri, isread)) {
1194
        unallocated_encoding(s);
1195
        return;
1196
    }
1197

    
1198
    if (ri->accessfn) {
1199
        /* Emit code to perform further access permissions checks at
1200
         * runtime; this may result in an exception.
1201
         */
1202
        TCGv_ptr tmpptr;
1203
        gen_a64_set_pc_im(s->pc - 4);
1204
        tmpptr = tcg_const_ptr(ri);
1205
        gen_helper_access_check_cp_reg(cpu_env, tmpptr);
1206
        tcg_temp_free_ptr(tmpptr);
1207
    }
1208

    
1209
    /* Handle special cases first */
1210
    switch (ri->type & ~(ARM_CP_FLAG_MASK & ~ARM_CP_SPECIAL)) {
1211
    case ARM_CP_NOP:
1212
        return;
1213
    case ARM_CP_NZCV:
1214
        tcg_rt = cpu_reg(s, rt);
1215
        if (isread) {
1216
            gen_get_nzcv(tcg_rt);
1217
        } else {
1218
            gen_set_nzcv(tcg_rt);
1219
        }
1220
        return;
1221
    default:
1222
        break;
1223
    }
1224

    
1225
    if (use_icount && (ri->type & ARM_CP_IO)) {
1226
        gen_io_start();
1227
    }
1228

    
1229
    tcg_rt = cpu_reg(s, rt);
1230

    
1231
    if (isread) {
1232
        if (ri->type & ARM_CP_CONST) {
1233
            tcg_gen_movi_i64(tcg_rt, ri->resetvalue);
1234
        } else if (ri->readfn) {
1235
            TCGv_ptr tmpptr;
1236
            tmpptr = tcg_const_ptr(ri);
1237
            gen_helper_get_cp_reg64(tcg_rt, cpu_env, tmpptr);
1238
            tcg_temp_free_ptr(tmpptr);
1239
        } else {
1240
            tcg_gen_ld_i64(tcg_rt, cpu_env, ri->fieldoffset);
1241
        }
1242
    } else {
1243
        if (ri->type & ARM_CP_CONST) {
1244
            /* If not forbidden by access permissions, treat as WI */
1245
            return;
1246
        } else if (ri->writefn) {
1247
            TCGv_ptr tmpptr;
1248
            tmpptr = tcg_const_ptr(ri);
1249
            gen_helper_set_cp_reg64(cpu_env, tmpptr, tcg_rt);
1250
            tcg_temp_free_ptr(tmpptr);
1251
        } else {
1252
            tcg_gen_st_i64(tcg_rt, cpu_env, ri->fieldoffset);
1253
        }
1254
    }
1255

    
1256
    if (use_icount && (ri->type & ARM_CP_IO)) {
1257
        /* I/O operations must end the TB here (whether read or write) */
1258
        gen_io_end();
1259
        s->is_jmp = DISAS_UPDATE;
1260
    } else if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
1261
        /* We default to ending the TB on a coprocessor register write,
1262
         * but allow this to be suppressed by the register definition
1263
         * (usually only necessary to work around guest bugs).
1264
         */
1265
        s->is_jmp = DISAS_UPDATE;
1266
    }
1267
}
1268

    
1269
/* C3.2.4 System
1270
 *  31                 22 21  20 19 18 16 15   12 11    8 7   5 4    0
1271
 * +---------------------+---+-----+-----+-------+-------+-----+------+
1272
 * | 1 1 0 1 0 1 0 1 0 0 | L | op0 | op1 |  CRn  |  CRm  | op2 |  Rt  |
1273
 * +---------------------+---+-----+-----+-------+-------+-----+------+
1274
 */
1275
static void disas_system(DisasContext *s, uint32_t insn)
1276
{
1277
    unsigned int l, op0, op1, crn, crm, op2, rt;
1278
    l = extract32(insn, 21, 1);
1279
    op0 = extract32(insn, 19, 2);
1280
    op1 = extract32(insn, 16, 3);
1281
    crn = extract32(insn, 12, 4);
1282
    crm = extract32(insn, 8, 4);
1283
    op2 = extract32(insn, 5, 3);
1284
    rt = extract32(insn, 0, 5);
1285

    
1286
    if (op0 == 0) {
1287
        if (l || rt != 31) {
1288
            unallocated_encoding(s);
1289
            return;
1290
        }
1291
        switch (crn) {
1292
        case 2: /* C5.6.68 HINT */
1293
            handle_hint(s, insn, op1, op2, crm);
1294
            break;
1295
        case 3: /* CLREX, DSB, DMB, ISB */
1296
            handle_sync(s, insn, op1, op2, crm);
1297
            break;
1298
        case 4: /* C5.6.130 MSR (immediate) */
1299
            handle_msr_i(s, insn, op1, op2, crm);
1300
            break;
1301
        default:
1302
            unallocated_encoding(s);
1303
            break;
1304
        }
1305
        return;
1306
    }
1307
    handle_sys(s, insn, l, op0, op1, op2, crn, crm, rt);
1308
}
1309

    
1310
/* C3.2.3 Exception generation
1311
 *
1312
 *  31             24 23 21 20                     5 4   2 1  0
1313
 * +-----------------+-----+------------------------+-----+----+
1314
 * | 1 1 0 1 0 1 0 0 | opc |          imm16         | op2 | LL |
1315
 * +-----------------------+------------------------+----------+
1316
 */
1317
static void disas_exc(DisasContext *s, uint32_t insn)
1318
{
1319
    int opc = extract32(insn, 21, 3);
1320
    int op2_ll = extract32(insn, 0, 5);
1321

    
1322
    switch (opc) {
1323
    case 0:
1324
        /* SVC, HVC, SMC; since we don't support the Virtualization
1325
         * or TrustZone extensions these all UNDEF except SVC.
1326
         */
1327
        if (op2_ll != 1) {
1328
            unallocated_encoding(s);
1329
            break;
1330
        }
1331
        gen_exception_insn(s, 0, EXCP_SWI);
1332
        break;
1333
    case 1:
1334
        if (op2_ll != 0) {
1335
            unallocated_encoding(s);
1336
            break;
1337
        }
1338
        /* BRK */
1339
        gen_exception_insn(s, 0, EXCP_BKPT);
1340
        break;
1341
    case 2:
1342
        if (op2_ll != 0) {
1343
            unallocated_encoding(s);
1344
            break;
1345
        }
1346
        /* HLT */
1347
        unsupported_encoding(s, insn);
1348
        break;
1349
    case 5:
1350
        if (op2_ll < 1 || op2_ll > 3) {
1351
            unallocated_encoding(s);
1352
            break;
1353
        }
1354
        /* DCPS1, DCPS2, DCPS3 */
1355
        unsupported_encoding(s, insn);
1356
        break;
1357
    default:
1358
        unallocated_encoding(s);
1359
        break;
1360
    }
1361
}
1362

    
1363
/* C3.2.7 Unconditional branch (register)
1364
 *  31           25 24   21 20   16 15   10 9    5 4     0
1365
 * +---------------+-------+-------+-------+------+-------+
1366
 * | 1 1 0 1 0 1 1 |  opc  |  op2  |  op3  |  Rn  |  op4  |
1367
 * +---------------+-------+-------+-------+------+-------+
1368
 */
1369
static void disas_uncond_b_reg(DisasContext *s, uint32_t insn)
1370
{
1371
    unsigned int opc, op2, op3, rn, op4;
1372

    
1373
    opc = extract32(insn, 21, 4);
1374
    op2 = extract32(insn, 16, 5);
1375
    op3 = extract32(insn, 10, 6);
1376
    rn = extract32(insn, 5, 5);
1377
    op4 = extract32(insn, 0, 5);
1378

    
1379
    if (op4 != 0x0 || op3 != 0x0 || op2 != 0x1f) {
1380
        unallocated_encoding(s);
1381
        return;
1382
    }
1383

    
1384
    switch (opc) {
1385
    case 0: /* BR */
1386
    case 2: /* RET */
1387
        break;
1388
    case 1: /* BLR */
1389
        tcg_gen_movi_i64(cpu_reg(s, 30), s->pc);
1390
        break;
1391
    case 4: /* ERET */
1392
    case 5: /* DRPS */
1393
        if (rn != 0x1f) {
1394
            unallocated_encoding(s);
1395
        } else {
1396
            unsupported_encoding(s, insn);
1397
        }
1398
        return;
1399
    default:
1400
        unallocated_encoding(s);
1401
        return;
1402
    }
1403

    
1404
    tcg_gen_mov_i64(cpu_pc, cpu_reg(s, rn));
1405
    s->is_jmp = DISAS_JUMP;
1406
}
1407

    
1408
/* C3.2 Branches, exception generating and system instructions */
1409
static void disas_b_exc_sys(DisasContext *s, uint32_t insn)
1410
{
1411
    switch (extract32(insn, 25, 7)) {
1412
    case 0x0a: case 0x0b:
1413
    case 0x4a: case 0x4b: /* Unconditional branch (immediate) */
1414
        disas_uncond_b_imm(s, insn);
1415
        break;
1416
    case 0x1a: case 0x5a: /* Compare & branch (immediate) */
1417
        disas_comp_b_imm(s, insn);
1418
        break;
1419
    case 0x1b: case 0x5b: /* Test & branch (immediate) */
1420
        disas_test_b_imm(s, insn);
1421
        break;
1422
    case 0x2a: /* Conditional branch (immediate) */
1423
        disas_cond_b_imm(s, insn);
1424
        break;
1425
    case 0x6a: /* Exception generation / System */
1426
        if (insn & (1 << 24)) {
1427
            disas_system(s, insn);
1428
        } else {
1429
            disas_exc(s, insn);
1430
        }
1431
        break;
1432
    case 0x6b: /* Unconditional branch (register) */
1433
        disas_uncond_b_reg(s, insn);
1434
        break;
1435
    default:
1436
        unallocated_encoding(s);
1437
        break;
1438
    }
1439
}
1440

    
1441
/*
1442
 * Load/Store exclusive instructions are implemented by remembering
1443
 * the value/address loaded, and seeing if these are the same
1444
 * when the store is performed. This is not actually the architecturally
1445
 * mandated semantics, but it works for typical guest code sequences
1446
 * and avoids having to monitor regular stores.
1447
 *
1448
 * In system emulation mode only one CPU will be running at once, so
1449
 * this sequence is effectively atomic.  In user emulation mode we
1450
 * throw an exception and handle the atomic operation elsewhere.
1451
 */
1452
static void gen_load_exclusive(DisasContext *s, int rt, int rt2,
1453
                               TCGv_i64 addr, int size, bool is_pair)
1454
{
1455
    TCGv_i64 tmp = tcg_temp_new_i64();
1456
    TCGMemOp memop = MO_TE + size;
1457

    
1458
    g_assert(size <= 3);
1459
    tcg_gen_qemu_ld_i64(tmp, addr, get_mem_index(s), memop);
1460

    
1461
    if (is_pair) {
1462
        TCGv_i64 addr2 = tcg_temp_new_i64();
1463
        TCGv_i64 hitmp = tcg_temp_new_i64();
1464

    
1465
        g_assert(size >= 2);
1466
        tcg_gen_addi_i64(addr2, addr, 1 << size);
1467
        tcg_gen_qemu_ld_i64(hitmp, addr2, get_mem_index(s), memop);
1468
        tcg_temp_free_i64(addr2);
1469
        tcg_gen_mov_i64(cpu_exclusive_high, hitmp);
1470
        tcg_gen_mov_i64(cpu_reg(s, rt2), hitmp);
1471
        tcg_temp_free_i64(hitmp);
1472
    }
1473

    
1474
    tcg_gen_mov_i64(cpu_exclusive_val, tmp);
1475
    tcg_gen_mov_i64(cpu_reg(s, rt), tmp);
1476

    
1477
    tcg_temp_free_i64(tmp);
1478
    tcg_gen_mov_i64(cpu_exclusive_addr, addr);
1479
}
1480

    
1481
#ifdef CONFIG_USER_ONLY
1482
static void gen_store_exclusive(DisasContext *s, int rd, int rt, int rt2,
1483
                                TCGv_i64 addr, int size, int is_pair)
1484
{
1485
    tcg_gen_mov_i64(cpu_exclusive_test, addr);
1486
    tcg_gen_movi_i32(cpu_exclusive_info,
1487
                     size | is_pair << 2 | (rd << 4) | (rt << 9) | (rt2 << 14));
1488
    gen_exception_insn(s, 4, EXCP_STREX);
1489
}
1490
#else
1491
static void gen_store_exclusive(DisasContext *s, int rd, int rt, int rt2,
1492
                                TCGv_i64 inaddr, int size, int is_pair)
1493
{
1494
    /* if (env->exclusive_addr == addr && env->exclusive_val == [addr]
1495
     *     && (!is_pair || env->exclusive_high == [addr + datasize])) {
1496
     *     [addr] = {Rt};
1497
     *     if (is_pair) {
1498
     *         [addr + datasize] = {Rt2};
1499
     *     }
1500
     *     {Rd} = 0;
1501
     * } else {
1502
     *     {Rd} = 1;
1503
     * }
1504
     * env->exclusive_addr = -1;
1505
     */
1506
    int fail_label = gen_new_label();
1507
    int done_label = gen_new_label();
1508
    TCGv_i64 addr = tcg_temp_local_new_i64();
1509
    TCGv_i64 tmp;
1510

    
1511
    /* Copy input into a local temp so it is not trashed when the
1512
     * basic block ends at the branch insn.
1513
     */
1514
    tcg_gen_mov_i64(addr, inaddr);
1515
    tcg_gen_brcond_i64(TCG_COND_NE, addr, cpu_exclusive_addr, fail_label);
1516

    
1517
    tmp = tcg_temp_new_i64();
1518
    tcg_gen_qemu_ld_i64(tmp, addr, get_mem_index(s), MO_TE + size);
1519
    tcg_gen_brcond_i64(TCG_COND_NE, tmp, cpu_exclusive_val, fail_label);
1520
    tcg_temp_free_i64(tmp);
1521

    
1522
    if (is_pair) {
1523
        TCGv_i64 addrhi = tcg_temp_new_i64();
1524
        TCGv_i64 tmphi = tcg_temp_new_i64();
1525

    
1526
        tcg_gen_addi_i64(addrhi, addr, 1 << size);
1527
        tcg_gen_qemu_ld_i64(tmphi, addrhi, get_mem_index(s), MO_TE + size);
1528
        tcg_gen_brcond_i64(TCG_COND_NE, tmphi, cpu_exclusive_high, fail_label);
1529

    
1530
        tcg_temp_free_i64(tmphi);
1531
        tcg_temp_free_i64(addrhi);
1532
    }
1533

    
1534
    /* We seem to still have the exclusive monitor, so do the store */
1535
    tcg_gen_qemu_st_i64(cpu_reg(s, rt), addr, get_mem_index(s), MO_TE + size);
1536
    if (is_pair) {
1537
        TCGv_i64 addrhi = tcg_temp_new_i64();
1538

    
1539
        tcg_gen_addi_i64(addrhi, addr, 1 << size);
1540
        tcg_gen_qemu_st_i64(cpu_reg(s, rt2), addrhi,
1541
                            get_mem_index(s), MO_TE + size);
1542
        tcg_temp_free_i64(addrhi);
1543
    }
1544

    
1545
    tcg_temp_free_i64(addr);
1546

    
1547
    tcg_gen_movi_i64(cpu_reg(s, rd), 0);
1548
    tcg_gen_br(done_label);
1549
    gen_set_label(fail_label);
1550
    tcg_gen_movi_i64(cpu_reg(s, rd), 1);
1551
    gen_set_label(done_label);
1552
    tcg_gen_movi_i64(cpu_exclusive_addr, -1);
1553

    
1554
}
1555
#endif
1556

    
1557
/* C3.3.6 Load/store exclusive
1558
 *
1559
 *  31 30 29         24  23  22   21  20  16  15  14   10 9    5 4    0
1560
 * +-----+-------------+----+---+----+------+----+-------+------+------+
1561
 * | sz  | 0 0 1 0 0 0 | o2 | L | o1 |  Rs  | o0 |  Rt2  |  Rn  | Rt   |
1562
 * +-----+-------------+----+---+----+------+----+-------+------+------+
1563
 *
1564
 *  sz: 00 -> 8 bit, 01 -> 16 bit, 10 -> 32 bit, 11 -> 64 bit
1565
 *   L: 0 -> store, 1 -> load
1566
 *  o2: 0 -> exclusive, 1 -> not
1567
 *  o1: 0 -> single register, 1 -> register pair
1568
 *  o0: 1 -> load-acquire/store-release, 0 -> not
1569
 *
1570
 *  o0 == 0 AND o2 == 1 is un-allocated
1571
 *  o1 == 1 is un-allocated except for 32 and 64 bit sizes
1572
 */
1573
static void disas_ldst_excl(DisasContext *s, uint32_t insn)
1574
{
1575
    int rt = extract32(insn, 0, 5);
1576
    int rn = extract32(insn, 5, 5);
1577
    int rt2 = extract32(insn, 10, 5);
1578
    int is_lasr = extract32(insn, 15, 1);
1579
    int rs = extract32(insn, 16, 5);
1580
    int is_pair = extract32(insn, 21, 1);
1581
    int is_store = !extract32(insn, 22, 1);
1582
    int is_excl = !extract32(insn, 23, 1);
1583
    int size = extract32(insn, 30, 2);
1584
    TCGv_i64 tcg_addr;
1585

    
1586
    if ((!is_excl && !is_lasr) ||
1587
        (is_pair && size < 2)) {
1588
        unallocated_encoding(s);
1589
        return;
1590
    }
1591

    
1592
    if (rn == 31) {
1593
        gen_check_sp_alignment(s);
1594
    }
1595
    tcg_addr = read_cpu_reg_sp(s, rn, 1);
1596

    
1597
    /* Note that since TCG is single threaded load-acquire/store-release
1598
     * semantics require no extra if (is_lasr) { ... } handling.
1599
     */
1600

    
1601
    if (is_excl) {
1602
        if (!is_store) {
1603
            gen_load_exclusive(s, rt, rt2, tcg_addr, size, is_pair);
1604
        } else {
1605
            gen_store_exclusive(s, rs, rt, rt2, tcg_addr, size, is_pair);
1606
        }
1607
    } else {
1608
        TCGv_i64 tcg_rt = cpu_reg(s, rt);
1609
        if (is_store) {
1610
            do_gpr_st(s, tcg_rt, tcg_addr, size);
1611
        } else {
1612
            do_gpr_ld(s, tcg_rt, tcg_addr, size, false, false);
1613
        }
1614
        if (is_pair) {
1615
            TCGv_i64 tcg_rt2 = cpu_reg(s, rt);
1616
            tcg_gen_addi_i64(tcg_addr, tcg_addr, 1 << size);
1617
            if (is_store) {
1618
                do_gpr_st(s, tcg_rt2, tcg_addr, size);
1619
            } else {
1620
                do_gpr_ld(s, tcg_rt2, tcg_addr, size, false, false);
1621
            }
1622
        }
1623
    }
1624
}
1625

    
1626
/*
1627
 * C3.3.5 Load register (literal)
1628
 *
1629
 *  31 30 29   27  26 25 24 23                5 4     0
1630
 * +-----+-------+---+-----+-------------------+-------+
1631
 * | opc | 0 1 1 | V | 0 0 |     imm19         |  Rt   |
1632
 * +-----+-------+---+-----+-------------------+-------+
1633
 *
1634
 * V: 1 -> vector (simd/fp)
1635
 * opc (non-vector): 00 -> 32 bit, 01 -> 64 bit,
1636
 *                   10-> 32 bit signed, 11 -> prefetch
1637
 * opc (vector): 00 -> 32 bit, 01 -> 64 bit, 10 -> 128 bit (11 unallocated)
1638
 */
1639
static void disas_ld_lit(DisasContext *s, uint32_t insn)
1640
{
1641
    int rt = extract32(insn, 0, 5);
1642
    int64_t imm = sextract32(insn, 5, 19) << 2;
1643
    bool is_vector = extract32(insn, 26, 1);
1644
    int opc = extract32(insn, 30, 2);
1645
    bool is_signed = false;
1646
    int size = 2;
1647
    TCGv_i64 tcg_rt, tcg_addr;
1648

    
1649
    if (is_vector) {
1650
        if (opc == 3) {
1651
            unallocated_encoding(s);
1652
            return;
1653
        }
1654
        size = 2 + opc;
1655
    } else {
1656
        if (opc == 3) {
1657
            /* PRFM (literal) : prefetch */
1658
            return;
1659
        }
1660
        size = 2 + extract32(opc, 0, 1);
1661
        is_signed = extract32(opc, 1, 1);
1662
    }
1663

    
1664
    tcg_rt = cpu_reg(s, rt);
1665

    
1666
    tcg_addr = tcg_const_i64((s->pc - 4) + imm);
1667
    if (is_vector) {
1668
        do_fp_ld(s, rt, tcg_addr, size);
1669
    } else {
1670
        do_gpr_ld(s, tcg_rt, tcg_addr, size, is_signed, false);
1671
    }
1672
    tcg_temp_free_i64(tcg_addr);
1673
}
1674

    
1675
/*
1676
 * C5.6.80 LDNP (Load Pair - non-temporal hint)
1677
 * C5.6.81 LDP (Load Pair - non vector)
1678
 * C5.6.82 LDPSW (Load Pair Signed Word - non vector)
1679
 * C5.6.176 STNP (Store Pair - non-temporal hint)
1680
 * C5.6.177 STP (Store Pair - non vector)
1681
 * C6.3.165 LDNP (Load Pair of SIMD&FP - non-temporal hint)
1682
 * C6.3.165 LDP (Load Pair of SIMD&FP)
1683
 * C6.3.284 STNP (Store Pair of SIMD&FP - non-temporal hint)
1684
 * C6.3.284 STP (Store Pair of SIMD&FP)
1685
 *
1686
 *  31 30 29   27  26  25 24   23  22 21   15 14   10 9    5 4    0
1687
 * +-----+-------+---+---+-------+---+-----------------------------+
1688
 * | opc | 1 0 1 | V | 0 | index | L |  imm7 |  Rt2  |  Rn  | Rt   |
1689
 * +-----+-------+---+---+-------+---+-------+-------+------+------+
1690
 *
1691
 * opc: LDP/STP/LDNP/STNP        00 -> 32 bit, 10 -> 64 bit
1692
 *      LDPSW                    01
1693
 *      LDP/STP/LDNP/STNP (SIMD) 00 -> 32 bit, 01 -> 64 bit, 10 -> 128 bit
1694
 *   V: 0 -> GPR, 1 -> Vector
1695
 * idx: 00 -> signed offset with non-temporal hint, 01 -> post-index,
1696
 *      10 -> signed offset, 11 -> pre-index
1697
 *   L: 0 -> Store 1 -> Load
1698
 *
1699
 * Rt, Rt2 = GPR or SIMD registers to be stored
1700
 * Rn = general purpose register containing address
1701
 * imm7 = signed offset (multiple of 4 or 8 depending on size)
1702
 */
1703
static void disas_ldst_pair(DisasContext *s, uint32_t insn)
1704
{
1705
    int rt = extract32(insn, 0, 5);
1706
    int rn = extract32(insn, 5, 5);
1707
    int rt2 = extract32(insn, 10, 5);
1708
    int64_t offset = sextract32(insn, 15, 7);
1709
    int index = extract32(insn, 23, 2);
1710
    bool is_vector = extract32(insn, 26, 1);
1711
    bool is_load = extract32(insn, 22, 1);
1712
    int opc = extract32(insn, 30, 2);
1713

    
1714
    bool is_signed = false;
1715
    bool postindex = false;
1716
    bool wback = false;
1717

    
1718
    TCGv_i64 tcg_addr; /* calculated address */
1719
    int size;
1720

    
1721
    if (opc == 3) {
1722
        unallocated_encoding(s);
1723
        return;
1724
    }
1725

    
1726
    if (is_vector) {
1727
        size = 2 + opc;
1728
    } else {
1729
        size = 2 + extract32(opc, 1, 1);
1730
        is_signed = extract32(opc, 0, 1);
1731
        if (!is_load && is_signed) {
1732
            unallocated_encoding(s);
1733
            return;
1734
        }
1735
    }
1736

    
1737
    switch (index) {
1738
    case 1: /* post-index */
1739
        postindex = true;
1740
        wback = true;
1741
        break;
1742
    case 0:
1743
        /* signed offset with "non-temporal" hint. Since we don't emulate
1744
         * caches we don't care about hints to the cache system about
1745
         * data access patterns, and handle this identically to plain
1746
         * signed offset.
1747
         */
1748
        if (is_signed) {
1749
            /* There is no non-temporal-hint version of LDPSW */
1750
            unallocated_encoding(s);
1751
            return;
1752
        }
1753
        postindex = false;
1754
        break;
1755
    case 2: /* signed offset, rn not updated */
1756
        postindex = false;
1757
        break;
1758
    case 3: /* pre-index */
1759
        postindex = false;
1760
        wback = true;
1761
        break;
1762
    }
1763

    
1764
    offset <<= size;
1765

    
1766
    if (rn == 31) {
1767
        gen_check_sp_alignment(s);
1768
    }
1769

    
1770
    tcg_addr = read_cpu_reg_sp(s, rn, 1);
1771

    
1772
    if (!postindex) {
1773
        tcg_gen_addi_i64(tcg_addr, tcg_addr, offset);
1774
    }
1775

    
1776
    if (is_vector) {
1777
        if (is_load) {
1778
            do_fp_ld(s, rt, tcg_addr, size);
1779
        } else {
1780
            do_fp_st(s, rt, tcg_addr, size);
1781
        }
1782
    } else {
1783
        TCGv_i64 tcg_rt = cpu_reg(s, rt);
1784
        if (is_load) {
1785
            do_gpr_ld(s, tcg_rt, tcg_addr, size, is_signed, false);
1786
        } else {
1787
            do_gpr_st(s, tcg_rt, tcg_addr, size);
1788
        }
1789
    }
1790
    tcg_gen_addi_i64(tcg_addr, tcg_addr, 1 << size);
1791
    if (is_vector) {
1792
        if (is_load) {
1793
            do_fp_ld(s, rt2, tcg_addr, size);
1794
        } else {
1795
            do_fp_st(s, rt2, tcg_addr, size);
1796
        }
1797
    } else {
1798
        TCGv_i64 tcg_rt2 = cpu_reg(s, rt2);
1799
        if (is_load) {
1800
            do_gpr_ld(s, tcg_rt2, tcg_addr, size, is_signed, false);
1801
        } else {
1802
            do_gpr_st(s, tcg_rt2, tcg_addr, size);
1803
        }
1804
    }
1805

    
1806
    if (wback) {
1807
        if (postindex) {
1808
            tcg_gen_addi_i64(tcg_addr, tcg_addr, offset - (1 << size));
1809
        } else {
1810
            tcg_gen_subi_i64(tcg_addr, tcg_addr, 1 << size);
1811
        }
1812
        tcg_gen_mov_i64(cpu_reg_sp(s, rn), tcg_addr);
1813
    }
1814
}
1815

    
1816
/*
1817
 * C3.3.8 Load/store (immediate post-indexed)
1818
 * C3.3.9 Load/store (immediate pre-indexed)
1819
 * C3.3.12 Load/store (unscaled immediate)
1820
 *
1821
 * 31 30 29   27  26 25 24 23 22 21  20    12 11 10 9    5 4    0
1822
 * +----+-------+---+-----+-----+---+--------+-----+------+------+
1823
 * |size| 1 1 1 | V | 0 0 | opc | 0 |  imm9  | idx |  Rn  |  Rt  |
1824
 * +----+-------+---+-----+-----+---+--------+-----+------+------+
1825
 *
1826
 * idx = 01 -> post-indexed, 11 pre-indexed, 00 unscaled imm. (no writeback)
1827
 * V = 0 -> non-vector
1828
 * size: 00 -> 8 bit, 01 -> 16 bit, 10 -> 32 bit, 11 -> 64bit
1829
 * opc: 00 -> store, 01 -> loadu, 10 -> loads 64, 11 -> loads 32
1830
 */
1831
static void disas_ldst_reg_imm9(DisasContext *s, uint32_t insn)
1832
{
1833
    int rt = extract32(insn, 0, 5);
1834
    int rn = extract32(insn, 5, 5);
1835
    int imm9 = sextract32(insn, 12, 9);
1836
    int opc = extract32(insn, 22, 2);
1837
    int size = extract32(insn, 30, 2);
1838
    int idx = extract32(insn, 10, 2);
1839
    bool is_signed = false;
1840
    bool is_store = false;
1841
    bool is_extended = false;
1842
    bool is_vector = extract32(insn, 26, 1);
1843
    bool post_index;
1844
    bool writeback;
1845

    
1846
    TCGv_i64 tcg_addr;
1847

    
1848
    if (is_vector) {
1849
        size |= (opc & 2) << 1;
1850
        if (size > 4) {
1851
            unallocated_encoding(s);
1852
            return;
1853
        }
1854
        is_store = ((opc & 1) == 0);
1855
    } else {
1856
        if (size == 3 && opc == 2) {
1857
            /* PRFM - prefetch */
1858
            return;
1859
        }
1860
        if (opc == 3 && size > 1) {
1861
            unallocated_encoding(s);
1862
            return;
1863
        }
1864
        is_store = (opc == 0);
1865
        is_signed = opc & (1<<1);
1866
        is_extended = (size < 3) && (opc & 1);
1867
    }
1868

    
1869
    switch (idx) {
1870
    case 0:
1871
        post_index = false;
1872
        writeback = false;
1873
        break;
1874
    case 1:
1875
        post_index = true;
1876
        writeback = true;
1877
        break;
1878
    case 3:
1879
        post_index = false;
1880
        writeback = true;
1881
        break;
1882
    case 2:
1883
        g_assert(false);
1884
        break;
1885
    }
1886

    
1887
    if (rn == 31) {
1888
        gen_check_sp_alignment(s);
1889
    }
1890
    tcg_addr = read_cpu_reg_sp(s, rn, 1);
1891

    
1892
    if (!post_index) {
1893
        tcg_gen_addi_i64(tcg_addr, tcg_addr, imm9);
1894
    }
1895

    
1896
    if (is_vector) {
1897
        if (is_store) {
1898
            do_fp_st(s, rt, tcg_addr, size);
1899
        } else {
1900
            do_fp_ld(s, rt, tcg_addr, size);
1901
        }
1902
    } else {
1903
        TCGv_i64 tcg_rt = cpu_reg(s, rt);
1904
        if (is_store) {
1905
            do_gpr_st(s, tcg_rt, tcg_addr, size);
1906
        } else {
1907
            do_gpr_ld(s, tcg_rt, tcg_addr, size, is_signed, is_extended);
1908
        }
1909
    }
1910

    
1911
    if (writeback) {
1912
        TCGv_i64 tcg_rn = cpu_reg_sp(s, rn);
1913
        if (post_index) {
1914
            tcg_gen_addi_i64(tcg_addr, tcg_addr, imm9);
1915
        }
1916
        tcg_gen_mov_i64(tcg_rn, tcg_addr);
1917
    }
1918
}
1919

    
1920
/*
1921
 * C3.3.10 Load/store (register offset)
1922
 *
1923
 * 31 30 29   27  26 25 24 23 22 21  20  16 15 13 12 11 10 9  5 4  0
1924
 * +----+-------+---+-----+-----+---+------+-----+--+-----+----+----+
1925
 * |size| 1 1 1 | V | 0 0 | opc | 1 |  Rm  | opt | S| 1 0 | Rn | Rt |
1926
 * +----+-------+---+-----+-----+---+------+-----+--+-----+----+----+
1927
 *
1928
 * For non-vector:
1929
 *   size: 00-> byte, 01 -> 16 bit, 10 -> 32bit, 11 -> 64bit
1930
 *   opc: 00 -> store, 01 -> loadu, 10 -> loads 64, 11 -> loads 32
1931
 * For vector:
1932
 *   size is opc<1>:size<1:0> so 100 -> 128 bit; 110 and 111 unallocated
1933
 *   opc<0>: 0 -> store, 1 -> load
1934
 * V: 1 -> vector/simd
1935
 * opt: extend encoding (see DecodeRegExtend)
1936
 * S: if S=1 then scale (essentially index by sizeof(size))
1937
 * Rt: register to transfer into/out of
1938
 * Rn: address register or SP for base
1939
 * Rm: offset register or ZR for offset
1940
 */
1941
static void disas_ldst_reg_roffset(DisasContext *s, uint32_t insn)
1942
{
1943
    int rt = extract32(insn, 0, 5);
1944
    int rn = extract32(insn, 5, 5);
1945
    int shift = extract32(insn, 12, 1);
1946
    int rm = extract32(insn, 16, 5);
1947
    int opc = extract32(insn, 22, 2);
1948
    int opt = extract32(insn, 13, 3);
1949
    int size = extract32(insn, 30, 2);
1950
    bool is_signed = false;
1951
    bool is_store = false;
1952
    bool is_extended = false;
1953
    bool is_vector = extract32(insn, 26, 1);
1954

    
1955
    TCGv_i64 tcg_rm;
1956
    TCGv_i64 tcg_addr;
1957

    
1958
    if (extract32(opt, 1, 1) == 0) {
1959
        unallocated_encoding(s);
1960
        return;
1961
    }
1962

    
1963
    if (is_vector) {
1964
        size |= (opc & 2) << 1;
1965
        if (size > 4) {
1966
            unallocated_encoding(s);
1967
            return;
1968
        }
1969
        is_store = !extract32(opc, 0, 1);
1970
    } else {
1971
        if (size == 3 && opc == 2) {
1972
            /* PRFM - prefetch */
1973
            return;
1974
        }
1975
        if (opc == 3 && size > 1) {
1976
            unallocated_encoding(s);
1977
            return;
1978
        }
1979
        is_store = (opc == 0);
1980
        is_signed = extract32(opc, 1, 1);
1981
        is_extended = (size < 3) && extract32(opc, 0, 1);
1982
    }
1983

    
1984
    if (rn == 31) {
1985
        gen_check_sp_alignment(s);
1986
    }
1987
    tcg_addr = read_cpu_reg_sp(s, rn, 1);
1988

    
1989
    tcg_rm = read_cpu_reg(s, rm, 1);
1990
    ext_and_shift_reg(tcg_rm, tcg_rm, opt, shift ? size : 0);
1991

    
1992
    tcg_gen_add_i64(tcg_addr, tcg_addr, tcg_rm);
1993

    
1994
    if (is_vector) {
1995
        if (is_store) {
1996
            do_fp_st(s, rt, tcg_addr, size);
1997
        } else {
1998
            do_fp_ld(s, rt, tcg_addr, size);
1999
        }
2000
    } else {
2001
        TCGv_i64 tcg_rt = cpu_reg(s, rt);
2002
        if (is_store) {
2003
            do_gpr_st(s, tcg_rt, tcg_addr, size);
2004
        } else {
2005
            do_gpr_ld(s, tcg_rt, tcg_addr, size, is_signed, is_extended);
2006
        }
2007
    }
2008
}
2009

    
2010
/*
2011
 * C3.3.13 Load/store (unsigned immediate)
2012
 *
2013
 * 31 30 29   27  26 25 24 23 22 21        10 9     5
2014
 * +----+-------+---+-----+-----+------------+-------+------+
2015
 * |size| 1 1 1 | V | 0 1 | opc |   imm12    |  Rn   |  Rt  |
2016
 * +----+-------+---+-----+-----+------------+-------+------+
2017
 *
2018
 * For non-vector:
2019
 *   size: 00-> byte, 01 -> 16 bit, 10 -> 32bit, 11 -> 64bit
2020
 *   opc: 00 -> store, 01 -> loadu, 10 -> loads 64, 11 -> loads 32
2021
 * For vector:
2022
 *   size is opc<1>:size<1:0> so 100 -> 128 bit; 110 and 111 unallocated
2023
 *   opc<0>: 0 -> store, 1 -> load
2024
 * Rn: base address register (inc SP)
2025
 * Rt: target register
2026
 */
2027
static void disas_ldst_reg_unsigned_imm(DisasContext *s, uint32_t insn)
2028
{
2029
    int rt = extract32(insn, 0, 5);
2030
    int rn = extract32(insn, 5, 5);
2031
    unsigned int imm12 = extract32(insn, 10, 12);
2032
    bool is_vector = extract32(insn, 26, 1);
2033
    int size = extract32(insn, 30, 2);
2034
    int opc = extract32(insn, 22, 2);
2035
    unsigned int offset;
2036

    
2037
    TCGv_i64 tcg_addr;
2038

    
2039
    bool is_store;
2040
    bool is_signed = false;
2041
    bool is_extended = false;
2042

    
2043
    if (is_vector) {
2044
        size |= (opc & 2) << 1;
2045
        if (size > 4) {
2046
            unallocated_encoding(s);
2047
            return;
2048
        }
2049
        is_store = !extract32(opc, 0, 1);
2050
    } else {
2051
        if (size == 3 && opc == 2) {
2052
            /* PRFM - prefetch */
2053
            return;
2054
        }
2055
        if (opc == 3 && size > 1) {
2056
            unallocated_encoding(s);
2057
            return;
2058
        }
2059
        is_store = (opc == 0);
2060
        is_signed = extract32(opc, 1, 1);
2061
        is_extended = (size < 3) && extract32(opc, 0, 1);
2062
    }
2063

    
2064
    if (rn == 31) {
2065
        gen_check_sp_alignment(s);
2066
    }
2067
    tcg_addr = read_cpu_reg_sp(s, rn, 1);
2068
    offset = imm12 << size;
2069
    tcg_gen_addi_i64(tcg_addr, tcg_addr, offset);
2070

    
2071
    if (is_vector) {
2072
        if (is_store) {
2073
            do_fp_st(s, rt, tcg_addr, size);
2074
        } else {
2075
            do_fp_ld(s, rt, tcg_addr, size);
2076
        }
2077
    } else {
2078
        TCGv_i64 tcg_rt = cpu_reg(s, rt);
2079
        if (is_store) {
2080
            do_gpr_st(s, tcg_rt, tcg_addr, size);
2081
        } else {
2082
            do_gpr_ld(s, tcg_rt, tcg_addr, size, is_signed, is_extended);
2083
        }
2084
    }
2085
}
2086

    
2087
/* Load/store register (immediate forms) */
2088
static void disas_ldst_reg_imm(DisasContext *s, uint32_t insn)
2089
{
2090
    switch (extract32(insn, 10, 2)) {
2091
    case 0: case 1: case 3:
2092
        /* Load/store register (unscaled immediate) */
2093
        /* Load/store immediate pre/post-indexed */
2094
        disas_ldst_reg_imm9(s, insn);
2095
        break;
2096
    case 2:
2097
        /* Load/store register unprivileged */
2098
        unsupported_encoding(s, insn);
2099
        break;
2100
    default:
2101
        unallocated_encoding(s);
2102
        break;
2103
    }
2104
}
2105

    
2106
/* Load/store register (all forms) */
2107
static void disas_ldst_reg(DisasContext *s, uint32_t insn)
2108
{
2109
    switch (extract32(insn, 24, 2)) {
2110
    case 0:
2111
        if (extract32(insn, 21, 1) == 1 && extract32(insn, 10, 2) == 2) {
2112
            disas_ldst_reg_roffset(s, insn);
2113
        } else {
2114
            disas_ldst_reg_imm(s, insn);
2115
        }
2116
        break;
2117
    case 1:
2118
        disas_ldst_reg_unsigned_imm(s, insn);
2119
        break;
2120
    default:
2121
        unallocated_encoding(s);
2122
        break;
2123
    }
2124
}
2125

    
2126
/* C3.3.1 AdvSIMD load/store multiple structures
2127
 *
2128
 *  31  30  29           23 22  21         16 15    12 11  10 9    5 4    0
2129
 * +---+---+---------------+---+-------------+--------+------+------+------+
2130
 * | 0 | Q | 0 0 1 1 0 0 0 | L | 0 0 0 0 0 0 | opcode | size |  Rn  |  Rt  |
2131
 * +---+---+---------------+---+-------------+--------+------+------+------+
2132
 *
2133
 * C3.3.2 AdvSIMD load/store multiple structures (post-indexed)
2134
 *
2135
 *  31  30  29           23 22  21  20     16 15    12 11  10 9    5 4    0
2136
 * +---+---+---------------+---+---+---------+--------+------+------+------+
2137
 * | 0 | Q | 0 0 1 1 0 0 1 | L | 0 |   Rm    | opcode | size |  Rn  |  Rt  |
2138
 * +---+---+---------------+---+---+---------+--------+------+------+------+
2139
 *
2140
 * Rt: first (or only) SIMD&FP register to be transferred
2141
 * Rn: base address or SP
2142
 * Rm (post-index only): post-index register (when !31) or size dependent #imm
2143
 */
2144
static void disas_ldst_multiple_struct(DisasContext *s, uint32_t insn)
2145
{
2146
    int rt = extract32(insn, 0, 5);
2147
    int rn = extract32(insn, 5, 5);
2148
    int size = extract32(insn, 10, 2);
2149
    int opcode = extract32(insn, 12, 4);
2150
    bool is_store = !extract32(insn, 22, 1);
2151
    bool is_postidx = extract32(insn, 23, 1);
2152
    bool is_q = extract32(insn, 30, 1);
2153
    TCGv_i64 tcg_addr, tcg_rn;
2154

    
2155
    int ebytes = 1 << size;
2156
    int elements = (is_q ? 128 : 64) / (8 << size);
2157
    int rpt;    /* num iterations */
2158
    int selem;  /* structure elements */
2159
    int r;
2160

    
2161
    if (extract32(insn, 31, 1) || extract32(insn, 21, 1)) {
2162
        unallocated_encoding(s);
2163
        return;
2164
    }
2165

    
2166
    /* From the shared decode logic */
2167
    switch (opcode) {
2168
    case 0x0:
2169
        rpt = 1;
2170
        selem = 4;
2171
        break;
2172
    case 0x2:
2173
        rpt = 4;
2174
        selem = 1;
2175
        break;
2176
    case 0x4:
2177
        rpt = 1;
2178
        selem = 3;
2179
        break;
2180
    case 0x6:
2181
        rpt = 3;
2182
        selem = 1;
2183
        break;
2184
    case 0x7:
2185
        rpt = 1;
2186
        selem = 1;
2187
        break;
2188
    case 0x8:
2189
        rpt = 1;
2190
        selem = 2;
2191
        break;
2192
    case 0xa:
2193
        rpt = 2;
2194
        selem = 1;
2195
        break;
2196
    default:
2197
        unallocated_encoding(s);
2198
        return;
2199
    }
2200

    
2201
    if (size == 3 && !is_q && selem != 1) {
2202
        /* reserved */
2203
        unallocated_encoding(s);
2204
        return;
2205
    }
2206

    
2207
    if (rn == 31) {
2208
        gen_check_sp_alignment(s);
2209
    }
2210

    
2211
    tcg_rn = cpu_reg_sp(s, rn);
2212
    tcg_addr = tcg_temp_new_i64();
2213
    tcg_gen_mov_i64(tcg_addr, tcg_rn);
2214

    
2215
    for (r = 0; r < rpt; r++) {
2216
        int e;
2217
        for (e = 0; e < elements; e++) {
2218
            int tt = (rt + r) % 32;
2219
            int xs;
2220
            for (xs = 0; xs < selem; xs++) {
2221
                if (is_store) {
2222
                    do_vec_st(s, tt, e, tcg_addr, size);
2223
                } else {
2224
                    do_vec_ld(s, tt, e, tcg_addr, size);
2225

    
2226
                    /* For non-quad operations, setting a slice of the low
2227
                     * 64 bits of the register clears the high 64 bits (in
2228
                     * the ARM ARM pseudocode this is implicit in the fact
2229
                     * that 'rval' is a 64 bit wide variable). We optimize
2230
                     * by noticing that we only need to do this the first
2231
                     * time we touch a register.
2232
                     */
2233
                    if (!is_q && e == 0 && (r == 0 || xs == selem - 1)) {
2234
                        clear_vec_high(s, tt);
2235
                    }
2236
                }
2237
                tcg_gen_addi_i64(tcg_addr, tcg_addr, ebytes);
2238
                tt = (tt + 1) % 32;
2239
            }
2240
        }
2241
    }
2242

    
2243
    if (is_postidx) {
2244
        int rm = extract32(insn, 16, 5);
2245
        if (rm == 31) {
2246
            tcg_gen_mov_i64(tcg_rn, tcg_addr);
2247
        } else {
2248
            tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
2249
        }
2250
    }
2251
    tcg_temp_free_i64(tcg_addr);
2252
}
2253

    
2254
/* C3.3.3 AdvSIMD load/store single structure
2255
 *
2256
 *  31  30  29           23 22 21 20       16 15 13 12  11  10 9    5 4    0
2257
 * +---+---+---------------+-----+-----------+-----+---+------+------+------+
2258
 * | 0 | Q | 0 0 1 1 0 1 0 | L R | 0 0 0 0 0 | opc | S | size |  Rn  |  Rt  |
2259
 * +---+---+---------------+-----+-----------+-----+---+------+------+------+
2260
 *
2261
 * C3.3.4 AdvSIMD load/store single structure (post-indexed)
2262
 *
2263
 *  31  30  29           23 22 21 20       16 15 13 12  11  10 9    5 4    0
2264
 * +---+---+---------------+-----+-----------+-----+---+------+------+------+
2265
 * | 0 | Q | 0 0 1 1 0 1 1 | L R |     Rm    | opc | S | size |  Rn  |  Rt  |
2266
 * +---+---+---------------+-----+-----------+-----+---+------+------+------+
2267
 *
2268
 * Rt: first (or only) SIMD&FP register to be transferred
2269
 * Rn: base address or SP
2270
 * Rm (post-index only): post-index register (when !31) or size dependent #imm
2271
 * index = encoded in Q:S:size dependent on size
2272
 *
2273
 * lane_size = encoded in R, opc
2274
 * transfer width = encoded in opc, S, size
2275
 */
2276
static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
2277
{
2278
    int rt = extract32(insn, 0, 5);
2279
    int rn = extract32(insn, 5, 5);
2280
    int size = extract32(insn, 10, 2);
2281
    int S = extract32(insn, 12, 1);
2282
    int opc = extract32(insn, 13, 3);
2283
    int R = extract32(insn, 21, 1);
2284
    int is_load = extract32(insn, 22, 1);
2285
    int is_postidx = extract32(insn, 23, 1);
2286
    int is_q = extract32(insn, 30, 1);
2287

    
2288
    int scale = extract32(opc, 1, 2);
2289
    int selem = (extract32(opc, 0, 1) << 1 | R) + 1;
2290
    bool replicate = false;
2291
    int index = is_q << 3 | S << 2 | size;
2292
    int ebytes, xs;
2293
    TCGv_i64 tcg_addr, tcg_rn;
2294

    
2295
    switch (scale) {
2296
    case 3:
2297
        if (!is_load || S) {
2298
            unallocated_encoding(s);
2299
            return;
2300
        }
2301
        scale = size;
2302
        replicate = true;
2303
        break;
2304
    case 0:
2305
        break;
2306
    case 1:
2307
        if (extract32(size, 0, 1)) {
2308
            unallocated_encoding(s);
2309
            return;
2310
        }
2311
        index >>= 1;
2312
        break;
2313
    case 2:
2314
        if (extract32(size, 1, 1)) {
2315
            unallocated_encoding(s);
2316
            return;
2317
        }
2318
        if (!extract32(size, 0, 1)) {
2319
            index >>= 2;
2320
        } else {
2321
            if (S) {
2322
                unallocated_encoding(s);
2323
                return;
2324
            }
2325
            index >>= 3;
2326
            scale = 3;
2327
        }
2328
        break;
2329
    default:
2330
        g_assert_not_reached();
2331
    }
2332

    
2333
    ebytes = 1 << scale;
2334

    
2335
    if (rn == 31) {
2336
        gen_check_sp_alignment(s);
2337
    }
2338

    
2339
    tcg_rn = cpu_reg_sp(s, rn);
2340
    tcg_addr = tcg_temp_new_i64();
2341
    tcg_gen_mov_i64(tcg_addr, tcg_rn);
2342

    
2343
    for (xs = 0; xs < selem; xs++) {
2344
        if (replicate) {
2345
            /* Load and replicate to all elements */
2346
            uint64_t mulconst;
2347
            TCGv_i64 tcg_tmp = tcg_temp_new_i64();
2348

    
2349
            tcg_gen_qemu_ld_i64(tcg_tmp, tcg_addr,
2350
                                get_mem_index(s), MO_TE + scale);
2351
            switch (scale) {
2352
            case 0:
2353
                mulconst = 0x0101010101010101ULL;
2354
                break;
2355
            case 1:
2356
                mulconst = 0x0001000100010001ULL;
2357
                break;
2358
            case 2:
2359
                mulconst = 0x0000000100000001ULL;
2360
                break;
2361
            case 3:
2362
                mulconst = 0;
2363
                break;
2364
            default:
2365
                g_assert_not_reached();
2366
            }
2367
            if (mulconst) {
2368
                tcg_gen_muli_i64(tcg_tmp, tcg_tmp, mulconst);
2369
            }
2370
            write_vec_element(s, tcg_tmp, rt, 0, MO_64);
2371
            if (is_q) {
2372
                write_vec_element(s, tcg_tmp, rt, 1, MO_64);
2373
            } else {
2374
                clear_vec_high(s, rt);
2375
            }
2376
            tcg_temp_free_i64(tcg_tmp);
2377
        } else {
2378
            /* Load/store one element per register */
2379
            if (is_load) {
2380
                do_vec_ld(s, rt, index, tcg_addr, MO_TE + scale);
2381
            } else {
2382
                do_vec_st(s, rt, index, tcg_addr, MO_TE + scale);
2383
            }
2384
        }
2385
        tcg_gen_addi_i64(tcg_addr, tcg_addr, ebytes);
2386
        rt = (rt + 1) % 32;
2387
    }
2388

    
2389
    if (is_postidx) {
2390
        int rm = extract32(insn, 16, 5);
2391
        if (rm == 31) {
2392
            tcg_gen_mov_i64(tcg_rn, tcg_addr);
2393
        } else {
2394
            tcg_gen_add_i64(tcg_rn, tcg_rn, cpu_reg(s, rm));
2395
        }
2396
    }
2397
    tcg_temp_free_i64(tcg_addr);
2398
}
2399

    
2400
/* C3.3 Loads and stores */
2401
static void disas_ldst(DisasContext *s, uint32_t insn)
2402
{
2403
    switch (extract32(insn, 24, 6)) {
2404
    case 0x08: /* Load/store exclusive */
2405
        disas_ldst_excl(s, insn);
2406
        break;
2407
    case 0x18: case 0x1c: /* Load register (literal) */
2408
        disas_ld_lit(s, insn);
2409
        break;
2410
    case 0x28: case 0x29:
2411
    case 0x2c: case 0x2d: /* Load/store pair (all forms) */
2412
        disas_ldst_pair(s, insn);
2413
        break;
2414
    case 0x38: case 0x39:
2415
    case 0x3c: case 0x3d: /* Load/store register (all forms) */
2416
        disas_ldst_reg(s, insn);
2417
        break;
2418
    case 0x0c: /* AdvSIMD load/store multiple structures */
2419
        disas_ldst_multiple_struct(s, insn);
2420
        break;
2421
    case 0x0d: /* AdvSIMD load/store single structure */
2422
        disas_ldst_single_struct(s, insn);
2423
        break;
2424
    default:
2425
        unallocated_encoding(s);
2426
        break;
2427
    }
2428
}
2429

    
2430
/* C3.4.6 PC-rel. addressing
2431
 *   31  30   29 28       24 23                5 4    0
2432
 * +----+-------+-----------+-------------------+------+
2433
 * | op | immlo | 1 0 0 0 0 |       immhi       |  Rd  |
2434
 * +----+-------+-----------+-------------------+------+
2435
 */
2436
static void disas_pc_rel_adr(DisasContext *s, uint32_t insn)
2437
{
2438
    unsigned int page, rd;
2439
    uint64_t base;
2440
    int64_t offset;
2441

    
2442
    page = extract32(insn, 31, 1);
2443
    /* SignExtend(immhi:immlo) -> offset */
2444
    offset = ((int64_t)sextract32(insn, 5, 19) << 2) | extract32(insn, 29, 2);
2445
    rd = extract32(insn, 0, 5);
2446
    base = s->pc - 4;
2447

    
2448
    if (page) {
2449
        /* ADRP (page based) */
2450
        base &= ~0xfff;
2451
        offset <<= 12;
2452
    }
2453

    
2454
    tcg_gen_movi_i64(cpu_reg(s, rd), base + offset);
2455
}
2456

    
2457
/*
2458
 * C3.4.1 Add/subtract (immediate)
2459
 *
2460
 *  31 30 29 28       24 23 22 21         10 9   5 4   0
2461
 * +--+--+--+-----------+-----+-------------+-----+-----+
2462
 * |sf|op| S| 1 0 0 0 1 |shift|    imm12    |  Rn | Rd  |
2463
 * +--+--+--+-----------+-----+-------------+-----+-----+
2464
 *
2465
 *    sf: 0 -> 32bit, 1 -> 64bit
2466
 *    op: 0 -> add  , 1 -> sub
2467
 *     S: 1 -> set flags
2468
 * shift: 00 -> LSL imm by 0, 01 -> LSL imm by 12
2469
 */
2470
static void disas_add_sub_imm(DisasContext *s, uint32_t insn)
2471
{
2472
    int rd = extract32(insn, 0, 5);
2473
    int rn = extract32(insn, 5, 5);
2474
    uint64_t imm = extract32(insn, 10, 12);
2475
    int shift = extract32(insn, 22, 2);
2476
    bool setflags = extract32(insn, 29, 1);
2477
    bool sub_op = extract32(insn, 30, 1);
2478
    bool is_64bit = extract32(insn, 31, 1);
2479

    
2480
    TCGv_i64 tcg_rn = cpu_reg_sp(s, rn);
2481
    TCGv_i64 tcg_rd = setflags ? cpu_reg(s, rd) : cpu_reg_sp(s, rd);
2482
    TCGv_i64 tcg_result;
2483

    
2484
    switch (shift) {
2485
    case 0x0:
2486
        break;
2487
    case 0x1:
2488
        imm <<= 12;
2489
        break;
2490
    default:
2491
        unallocated_encoding(s);
2492
        return;
2493
    }
2494

    
2495
    tcg_result = tcg_temp_new_i64();
2496
    if (!setflags) {
2497
        if (sub_op) {
2498
            tcg_gen_subi_i64(tcg_result, tcg_rn, imm);
2499
        } else {
2500
            tcg_gen_addi_i64(tcg_result, tcg_rn, imm);
2501
        }
2502
    } else {
2503
        TCGv_i64 tcg_imm = tcg_const_i64(imm);
2504
        if (sub_op) {
2505
            gen_sub_CC(is_64bit, tcg_result, tcg_rn, tcg_imm);
2506
        } else {
2507
            gen_add_CC(is_64bit, tcg_result, tcg_rn, tcg_imm);
2508
        }
2509
        tcg_temp_free_i64(tcg_imm);
2510
    }
2511

    
2512
    if (is_64bit) {
2513
        tcg_gen_mov_i64(tcg_rd, tcg_result);
2514
    } else {
2515
        tcg_gen_ext32u_i64(tcg_rd, tcg_result);
2516
    }
2517

    
2518
    tcg_temp_free_i64(tcg_result);
2519
}
2520

    
2521
/* The input should be a value in the bottom e bits (with higher
2522
 * bits zero); returns that value replicated into every element
2523
 * of size e in a 64 bit integer.
2524
 */
2525
static uint64_t bitfield_replicate(uint64_t mask, unsigned int e)
2526
{
2527
    assert(e != 0);
2528
    while (e < 64) {
2529
        mask |= mask << e;
2530
        e *= 2;
2531
    }
2532
    return mask;
2533
}
2534

    
2535
/* Return a value with the bottom len bits set (where 0 < len <= 64) */
2536
static inline uint64_t bitmask64(unsigned int length)
2537
{
2538
    assert(length > 0 && length <= 64);
2539
    return ~0ULL >> (64 - length);
2540
}
2541

    
2542
/* Simplified variant of pseudocode DecodeBitMasks() for the case where we
2543
 * only require the wmask. Returns false if the imms/immr/immn are a reserved
2544
 * value (ie should cause a guest UNDEF exception), and true if they are
2545
 * valid, in which case the decoded bit pattern is written to result.
2546
 */
2547
static bool logic_imm_decode_wmask(uint64_t *result, unsigned int immn,
2548
                                   unsigned int imms, unsigned int immr)
2549
{
2550
    uint64_t mask;
2551
    unsigned e, levels, s, r;
2552
    int len;
2553

    
2554
    assert(immn < 2 && imms < 64 && immr < 64);
2555

    
2556
    /* The bit patterns we create here are 64 bit patterns which
2557
     * are vectors of identical elements of size e = 2, 4, 8, 16, 32 or
2558
     * 64 bits each. Each element contains the same value: a run
2559
     * of between 1 and e-1 non-zero bits, rotated within the
2560
     * element by between 0 and e-1 bits.
2561
     *
2562
     * The element size and run length are encoded into immn (1 bit)
2563
     * and imms (6 bits) as follows:
2564
     * 64 bit elements: immn = 1, imms = <length of run - 1>
2565
     * 32 bit elements: immn = 0, imms = 0 : <length of run - 1>
2566
     * 16 bit elements: immn = 0, imms = 10 : <length of run - 1>
2567
     *  8 bit elements: immn = 0, imms = 110 : <length of run - 1>
2568
     *  4 bit elements: immn = 0, imms = 1110 : <length of run - 1>
2569
     *  2 bit elements: immn = 0, imms = 11110 : <length of run - 1>
2570
     * Notice that immn = 0, imms = 11111x is the only combination
2571
     * not covered by one of the above options; this is reserved.
2572
     * Further, <length of run - 1> all-ones is a reserved pattern.
2573
     *
2574
     * In all cases the rotation is by immr % e (and immr is 6 bits).
2575
     */
2576

    
2577
    /* First determine the element size */
2578
    len = 31 - clz32((immn << 6) | (~imms & 0x3f));
2579
    if (len < 1) {
2580
        /* This is the immn == 0, imms == 0x11111x case */
2581
        return false;
2582
    }
2583
    e = 1 << len;
2584

    
2585
    levels = e - 1;
2586
    s = imms & levels;
2587
    r = immr & levels;
2588

    
2589
    if (s == levels) {
2590
        /* <length of run - 1> mustn't be all-ones. */
2591
        return false;
2592
    }
2593

    
2594
    /* Create the value of one element: s+1 set bits rotated
2595
     * by r within the element (which is e bits wide)...
2596
     */
2597
    mask = bitmask64(s + 1);
2598
    mask = (mask >> r) | (mask << (e - r));
2599
    /* ...then replicate the element over the whole 64 bit value */
2600
    mask = bitfield_replicate(mask, e);
2601
    *result = mask;
2602
    return true;
2603
}
2604

    
2605
/* C3.4.4 Logical (immediate)
2606
 *   31  30 29 28         23 22  21  16 15  10 9    5 4    0
2607
 * +----+-----+-------------+---+------+------+------+------+
2608
 * | sf | opc | 1 0 0 1 0 0 | N | immr | imms |  Rn  |  Rd  |
2609
 * +----+-----+-------------+---+------+------+------+------+
2610
 */
2611
static void disas_logic_imm(DisasContext *s, uint32_t insn)
2612
{
2613
    unsigned int sf, opc, is_n, immr, imms, rn, rd;
2614
    TCGv_i64 tcg_rd, tcg_rn;
2615
    uint64_t wmask;
2616
    bool is_and = false;
2617

    
2618
    sf = extract32(insn, 31, 1);
2619
    opc = extract32(insn, 29, 2);
2620
    is_n = extract32(insn, 22, 1);
2621
    immr = extract32(insn, 16, 6);
2622
    imms = extract32(insn, 10, 6);
2623
    rn = extract32(insn, 5, 5);
2624
    rd = extract32(insn, 0, 5);
2625

    
2626
    if (!sf && is_n) {
2627
        unallocated_encoding(s);
2628
        return;
2629
    }
2630

    
2631
    if (opc == 0x3) { /* ANDS */
2632
        tcg_rd = cpu_reg(s, rd);
2633
    } else {
2634
        tcg_rd = cpu_reg_sp(s, rd);
2635
    }
2636
    tcg_rn = cpu_reg(s, rn);
2637

    
2638
    if (!logic_imm_decode_wmask(&wmask, is_n, imms, immr)) {
2639
        /* some immediate field values are reserved */
2640
        unallocated_encoding(s);
2641
        return;
2642
    }
2643

    
2644
    if (!sf) {
2645
        wmask &= 0xffffffff;
2646
    }
2647

    
2648
    switch (opc) {
2649
    case 0x3: /* ANDS */
2650
    case 0x0: /* AND */
2651
        tcg_gen_andi_i64(tcg_rd, tcg_rn, wmask);
2652
        is_and = true;
2653
        break;
2654
    case 0x1: /* ORR */
2655
        tcg_gen_ori_i64(tcg_rd, tcg_rn, wmask);
2656
        break;
2657
    case 0x2: /* EOR */
2658
        tcg_gen_xori_i64(tcg_rd, tcg_rn, wmask);
2659
        break;
2660
    default:
2661
        assert(FALSE); /* must handle all above */
2662
        break;
2663
    }
2664

    
2665
    if (!sf && !is_and) {
2666
        /* zero extend final result; we know we can skip this for AND
2667
         * since the immediate had the high 32 bits clear.
2668
         */
2669
        tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
2670
    }
2671

    
2672
    if (opc == 3) { /* ANDS */
2673
        gen_logic_CC(sf, tcg_rd);
2674
    }
2675
}
2676

    
2677
/*
2678
 * C3.4.5 Move wide (immediate)
2679
 *
2680
 *  31 30 29 28         23 22 21 20             5 4    0
2681
 * +--+-----+-------------+-----+----------------+------+
2682
 * |sf| opc | 1 0 0 1 0 1 |  hw |  imm16         |  Rd  |
2683
 * +--+-----+-------------+-----+----------------+------+
2684
 *
2685
 * sf: 0 -> 32 bit, 1 -> 64 bit
2686
 * opc: 00 -> N, 10 -> Z, 11 -> K
2687
 * hw: shift/16 (0,16, and sf only 32, 48)
2688
 */
2689
static void disas_movw_imm(DisasContext *s, uint32_t insn)
2690
{
2691
    int rd = extract32(insn, 0, 5);
2692
    uint64_t imm = extract32(insn, 5, 16);
2693
    int sf = extract32(insn, 31, 1);
2694
    int opc = extract32(insn, 29, 2);
2695
    int pos = extract32(insn, 21, 2) << 4;
2696
    TCGv_i64 tcg_rd = cpu_reg(s, rd);
2697
    TCGv_i64 tcg_imm;
2698

    
2699
    if (!sf && (pos >= 32)) {
2700
        unallocated_encoding(s);
2701
        return;
2702
    }
2703

    
2704
    switch (opc) {
2705
    case 0: /* MOVN */
2706
    case 2: /* MOVZ */
2707
        imm <<= pos;
2708
        if (opc == 0) {
2709
            imm = ~imm;
2710
        }
2711
        if (!sf) {
2712
            imm &= 0xffffffffu;
2713
        }
2714
        tcg_gen_movi_i64(tcg_rd, imm);
2715
        break;
2716
    case 3: /* MOVK */
2717
        tcg_imm = tcg_const_i64(imm);
2718
        tcg_gen_deposit_i64(tcg_rd, tcg_rd, tcg_imm, pos, 16);
2719
        tcg_temp_free_i64(tcg_imm);
2720
        if (!sf) {
2721
            tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
2722
        }
2723
        break;
2724
    default:
2725
        unallocated_encoding(s);
2726
        break;
2727
    }
2728
}
2729

    
2730
/* C3.4.2 Bitfield
2731
 *   31  30 29 28         23 22  21  16 15  10 9    5 4    0
2732
 * +----+-----+-------------+---+------+------+------+------+
2733
 * | sf | opc | 1 0 0 1 1 0 | N | immr | imms |  Rn  |  Rd  |
2734
 * +----+-----+-------------+---+------+------+------+------+
2735
 */
2736
static void disas_bitfield(DisasContext *s, uint32_t insn)
2737
{
2738
    unsigned int sf, n, opc, ri, si, rn, rd, bitsize, pos, len;
2739
    TCGv_i64 tcg_rd, tcg_tmp;
2740

    
2741
    sf = extract32(insn, 31, 1);
2742
    opc = extract32(insn, 29, 2);
2743
    n = extract32(insn, 22, 1);
2744
    ri = extract32(insn, 16, 6);
2745
    si = extract32(insn, 10, 6);
2746
    rn = extract32(insn, 5, 5);
2747
    rd = extract32(insn, 0, 5);
2748
    bitsize = sf ? 64 : 32;
2749

    
2750
    if (sf != n || ri >= bitsize || si >= bitsize || opc > 2) {
2751
        unallocated_encoding(s);
2752
        return;
2753
    }
2754

    
2755
    tcg_rd = cpu_reg(s, rd);
2756
    tcg_tmp = read_cpu_reg(s, rn, sf);
2757

    
2758
    /* OPTME: probably worth recognizing common cases of ext{8,16,32}{u,s} */
2759

    
2760
    if (opc != 1) { /* SBFM or UBFM */
2761
        tcg_gen_movi_i64(tcg_rd, 0);
2762
    }
2763

    
2764
    /* do the bit move operation */
2765
    if (si >= ri) {
2766
        /* Wd<s-r:0> = Wn<s:r> */
2767
        tcg_gen_shri_i64(tcg_tmp, tcg_tmp, ri);
2768
        pos = 0;
2769
        len = (si - ri) + 1;
2770
    } else {
2771
        /* Wd<32+s-r,32-r> = Wn<s:0> */
2772
        pos = bitsize - ri;
2773
        len = si + 1;
2774
    }
2775

    
2776
    tcg_gen_deposit_i64(tcg_rd, tcg_rd, tcg_tmp, pos, len);
2777

    
2778
    if (opc == 0) { /* SBFM - sign extend the destination field */
2779
        tcg_gen_shli_i64(tcg_rd, tcg_rd, 64 - (pos + len));
2780
        tcg_gen_sari_i64(tcg_rd, tcg_rd, 64 - (pos + len));
2781
    }
2782

    
2783
    if (!sf) { /* zero extend final result */
2784
        tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
2785
    }
2786
}
2787

    
2788
/* C3.4.3 Extract
2789
 *   31  30  29 28         23 22   21  20  16 15    10 9    5 4    0
2790
 * +----+------+-------------+---+----+------+--------+------+------+
2791
 * | sf | op21 | 1 0 0 1 1 1 | N | o0 |  Rm  |  imms  |  Rn  |  Rd  |
2792
 * +----+------+-------------+---+----+------+--------+------+------+
2793
 */
2794
static void disas_extract(DisasContext *s, uint32_t insn)
2795
{
2796
    unsigned int sf, n, rm, imm, rn, rd, bitsize, op21, op0;
2797

    
2798
    sf = extract32(insn, 31, 1);
2799
    n = extract32(insn, 22, 1);
2800
    rm = extract32(insn, 16, 5);
2801
    imm = extract32(insn, 10, 6);
2802
    rn = extract32(insn, 5, 5);
2803
    rd = extract32(insn, 0, 5);
2804
    op21 = extract32(insn, 29, 2);
2805
    op0 = extract32(insn, 21, 1);
2806
    bitsize = sf ? 64 : 32;
2807

    
2808
    if (sf != n || op21 || op0 || imm >= bitsize) {
2809
        unallocated_encoding(s);
2810
    } else {
2811
        TCGv_i64 tcg_rd, tcg_rm, tcg_rn;
2812

    
2813
        tcg_rd = cpu_reg(s, rd);
2814

    
2815
        if (imm) {
2816
            /* OPTME: we can special case rm==rn as a rotate */
2817
            tcg_rm = read_cpu_reg(s, rm, sf);
2818
            tcg_rn = read_cpu_reg(s, rn, sf);
2819
            tcg_gen_shri_i64(tcg_rm, tcg_rm, imm);
2820
            tcg_gen_shli_i64(tcg_rn, tcg_rn, bitsize - imm);
2821
            tcg_gen_or_i64(tcg_rd, tcg_rm, tcg_rn);
2822
            if (!sf) {
2823
                tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
2824
            }
2825
        } else {
2826
            /* tcg shl_i32/shl_i64 is undefined for 32/64 bit shifts,
2827
             * so an extract from bit 0 is a special case.
2828
             */
2829
            if (sf) {
2830
                tcg_gen_mov_i64(tcg_rd, cpu_reg(s, rm));
2831
            } else {
2832
                tcg_gen_ext32u_i64(tcg_rd, cpu_reg(s, rm));
2833
            }
2834
        }
2835

    
2836
    }
2837
}
2838

    
2839
/* C3.4 Data processing - immediate */
2840
static void disas_data_proc_imm(DisasContext *s, uint32_t insn)
2841
{
2842
    switch (extract32(insn, 23, 6)) {
2843
    case 0x20: case 0x21: /* PC-rel. addressing */
2844
        disas_pc_rel_adr(s, insn);
2845
        break;
2846
    case 0x22: case 0x23: /* Add/subtract (immediate) */
2847
        disas_add_sub_imm(s, insn);
2848
        break;
2849
    case 0x24: /* Logical (immediate) */
2850
        disas_logic_imm(s, insn);
2851
        break;
2852
    case 0x25: /* Move wide (immediate) */
2853
        disas_movw_imm(s, insn);
2854
        break;
2855
    case 0x26: /* Bitfield */
2856
        disas_bitfield(s, insn);
2857
        break;
2858
    case 0x27: /* Extract */
2859
        disas_extract(s, insn);
2860
        break;
2861
    default:
2862
        unallocated_encoding(s);
2863
        break;
2864
    }
2865
}
2866

    
2867
/* Shift a TCGv src by TCGv shift_amount, put result in dst.
2868
 * Note that it is the caller's responsibility to ensure that the
2869
 * shift amount is in range (ie 0..31 or 0..63) and provide the ARM
2870
 * mandated semantics for out of range shifts.
2871
 */
2872
static void shift_reg(TCGv_i64 dst, TCGv_i64 src, int sf,
2873
                      enum a64_shift_type shift_type, TCGv_i64 shift_amount)
2874
{
2875
    switch (shift_type) {
2876
    case A64_SHIFT_TYPE_LSL:
2877
        tcg_gen_shl_i64(dst, src, shift_amount);
2878
        break;
2879
    case A64_SHIFT_TYPE_LSR:
2880
        tcg_gen_shr_i64(dst, src, shift_amount);
2881
        break;
2882
    case A64_SHIFT_TYPE_ASR:
2883
        if (!sf) {
2884
            tcg_gen_ext32s_i64(dst, src);
2885
        }
2886
        tcg_gen_sar_i64(dst, sf ? src : dst, shift_amount);
2887
        break;
2888
    case A64_SHIFT_TYPE_ROR:
2889
        if (sf) {
2890
            tcg_gen_rotr_i64(dst, src, shift_amount);
2891
        } else {
2892
            TCGv_i32 t0, t1;
2893
            t0 = tcg_temp_new_i32();
2894
            t1 = tcg_temp_new_i32();
2895
            tcg_gen_trunc_i64_i32(t0, src);
2896
            tcg_gen_trunc_i64_i32(t1, shift_amount);
2897
            tcg_gen_rotr_i32(t0, t0, t1);
2898
            tcg_gen_extu_i32_i64(dst, t0);
2899
            tcg_temp_free_i32(t0);
2900
            tcg_temp_free_i32(t1);
2901
        }
2902
        break;
2903
    default:
2904
        assert(FALSE); /* all shift types should be handled */
2905
        break;
2906
    }
2907

    
2908
    if (!sf) { /* zero extend final result */
2909
        tcg_gen_ext32u_i64(dst, dst);
2910
    }
2911
}
2912

    
2913
/* Shift a TCGv src by immediate, put result in dst.
2914
 * The shift amount must be in range (this should always be true as the
2915
 * relevant instructions will UNDEF on bad shift immediates).
2916
 */
2917
static void shift_reg_imm(TCGv_i64 dst, TCGv_i64 src, int sf,
2918
                          enum a64_shift_type shift_type, unsigned int shift_i)
2919
{
2920
    assert(shift_i < (sf ? 64 : 32));
2921

    
2922
    if (shift_i == 0) {
2923
        tcg_gen_mov_i64(dst, src);
2924
    } else {
2925
        TCGv_i64 shift_const;
2926

    
2927
        shift_const = tcg_const_i64(shift_i);
2928
        shift_reg(dst, src, sf, shift_type, shift_const);
2929
        tcg_temp_free_i64(shift_const);
2930
    }
2931
}
2932

    
2933
/* C3.5.10 Logical (shifted register)
2934
 *   31  30 29 28       24 23   22 21  20  16 15    10 9    5 4    0
2935
 * +----+-----+-----------+-------+---+------+--------+------+------+
2936
 * | sf | opc | 0 1 0 1 0 | shift | N |  Rm  |  imm6  |  Rn  |  Rd  |
2937
 * +----+-----+-----------+-------+---+------+--------+------+------+
2938
 */
2939
static void disas_logic_reg(DisasContext *s, uint32_t insn)
2940
{
2941
    TCGv_i64 tcg_rd, tcg_rn, tcg_rm;
2942
    unsigned int sf, opc, shift_type, invert, rm, shift_amount, rn, rd;
2943

    
2944
    sf = extract32(insn, 31, 1);
2945
    opc = extract32(insn, 29, 2);
2946
    shift_type = extract32(insn, 22, 2);
2947
    invert = extract32(insn, 21, 1);
2948
    rm = extract32(insn, 16, 5);
2949
    shift_amount = extract32(insn, 10, 6);
2950
    rn = extract32(insn, 5, 5);
2951
    rd = extract32(insn, 0, 5);
2952

    
2953
    if (!sf && (shift_amount & (1 << 5))) {
2954
        unallocated_encoding(s);
2955
        return;
2956
    }
2957

    
2958
    tcg_rd = cpu_reg(s, rd);
2959

    
2960
    if (opc == 1 && shift_amount == 0 && shift_type == 0 && rn == 31) {
2961
        /* Unshifted ORR and ORN with WZR/XZR is the standard encoding for
2962
         * register-register MOV and MVN, so it is worth special casing.
2963
         */
2964
        tcg_rm = cpu_reg(s, rm);
2965
        if (invert) {
2966
            tcg_gen_not_i64(tcg_rd, tcg_rm);
2967
            if (!sf) {
2968
                tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
2969
            }
2970
        } else {
2971
            if (sf) {
2972
                tcg_gen_mov_i64(tcg_rd, tcg_rm);
2973
            } else {
2974
                tcg_gen_ext32u_i64(tcg_rd, tcg_rm);
2975
            }
2976
        }
2977
        return;
2978
    }
2979

    
2980
    tcg_rm = read_cpu_reg(s, rm, sf);
2981

    
2982
    if (shift_amount) {
2983
        shift_reg_imm(tcg_rm, tcg_rm, sf, shift_type, shift_amount);
2984
    }
2985

    
2986
    tcg_rn = cpu_reg(s, rn);
2987

    
2988
    switch (opc | (invert << 2)) {
2989
    case 0: /* AND */
2990
    case 3: /* ANDS */
2991
        tcg_gen_and_i64(tcg_rd, tcg_rn, tcg_rm);
2992
        break;
2993
    case 1: /* ORR */
2994
        tcg_gen_or_i64(tcg_rd, tcg_rn, tcg_rm);
2995
        break;
2996
    case 2: /* EOR */
2997
        tcg_gen_xor_i64(tcg_rd, tcg_rn, tcg_rm);
2998
        break;
2999
    case 4: /* BIC */
3000
    case 7: /* BICS */
3001
        tcg_gen_andc_i64(tcg_rd, tcg_rn, tcg_rm);
3002
        break;
3003
    case 5: /* ORN */
3004
        tcg_gen_orc_i64(tcg_rd, tcg_rn, tcg_rm);
3005
        break;
3006
    case 6: /* EON */
3007
        tcg_gen_eqv_i64(tcg_rd, tcg_rn, tcg_rm);
3008
        break;
3009
    default:
3010
        assert(FALSE);
3011
        break;
3012
    }
3013

    
3014
    if (!sf) {
3015
        tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
3016
    }
3017

    
3018
    if (opc == 3) {
3019
        gen_logic_CC(sf, tcg_rd);
3020
    }
3021
}
3022

    
3023
/*
3024
 * C3.5.1 Add/subtract (extended register)
3025
 *
3026
 *  31|30|29|28       24|23 22|21|20   16|15  13|12  10|9  5|4  0|
3027
 * +--+--+--+-----------+-----+--+-------+------+------+----+----+
3028
 * |sf|op| S| 0 1 0 1 1 | opt | 1|  Rm   |option| imm3 | Rn | Rd |
3029
 * +--+--+--+-----------+-----+--+-------+------+------+----+----+
3030
 *
3031
 *  sf: 0 -> 32bit, 1 -> 64bit
3032
 *  op: 0 -> add  , 1 -> sub
3033
 *   S: 1 -> set flags
3034
 * opt: 00
3035
 * option: extension type (see DecodeRegExtend)
3036
 * imm3: optional shift to Rm
3037
 *
3038
 * Rd = Rn + LSL(extend(Rm), amount)
3039
 */
3040
static void disas_add_sub_ext_reg(DisasContext *s, uint32_t insn)
3041
{
3042
    int rd = extract32(insn, 0, 5);
3043
    int rn = extract32(insn, 5, 5);
3044
    int imm3 = extract32(insn, 10, 3);
3045
    int option = extract32(insn, 13, 3);
3046
    int rm = extract32(insn, 16, 5);
3047
    bool setflags = extract32(insn, 29, 1);
3048
    bool sub_op = extract32(insn, 30, 1);
3049
    bool sf = extract32(insn, 31, 1);
3050

    
3051
    TCGv_i64 tcg_rm, tcg_rn; /* temps */
3052
    TCGv_i64 tcg_rd;
3053
    TCGv_i64 tcg_result;
3054

    
3055
    if (imm3 > 4) {
3056
        unallocated_encoding(s);
3057
        return;
3058
    }
3059

    
3060
    /* non-flag setting ops may use SP */
3061
    if (!setflags) {
3062
        tcg_rn = read_cpu_reg_sp(s, rn, sf);
3063
        tcg_rd = cpu_reg_sp(s, rd);
3064
    } else {
3065
        tcg_rn = read_cpu_reg(s, rn, sf);
3066
        tcg_rd = cpu_reg(s, rd);
3067
    }
3068

    
3069
    tcg_rm = read_cpu_reg(s, rm, sf);
3070
    ext_and_shift_reg(tcg_rm, tcg_rm, option, imm3);
3071

    
3072
    tcg_result = tcg_temp_new_i64();
3073

    
3074
    if (!setflags) {
3075
        if (sub_op) {
3076
            tcg_gen_sub_i64(tcg_result, tcg_rn, tcg_rm);
3077
        } else {
3078
            tcg_gen_add_i64(tcg_result, tcg_rn, tcg_rm);
3079
        }
3080
    } else {
3081
        if (sub_op) {
3082
            gen_sub_CC(sf, tcg_result, tcg_rn, tcg_rm);
3083
        } else {
3084
            gen_add_CC(sf, tcg_result, tcg_rn, tcg_rm);
3085
        }
3086
    }
3087

    
3088
    if (sf) {
3089
        tcg_gen_mov_i64(tcg_rd, tcg_result);
3090
    } else {
3091
        tcg_gen_ext32u_i64(tcg_rd, tcg_result);
3092
    }
3093

    
3094
    tcg_temp_free_i64(tcg_result);
3095
}
3096

    
3097
/*
3098
 * C3.5.2 Add/subtract (shifted register)
3099
 *
3100
 *  31 30 29 28       24 23 22 21 20   16 15     10 9    5 4    0
3101
 * +--+--+--+-----------+-----+--+-------+---------+------+------+
3102
 * |sf|op| S| 0 1 0 1 1 |shift| 0|  Rm   |  imm6   |  Rn  |  Rd  |
3103
 * +--+--+--+-----------+-----+--+-------+---------+------+------+
3104
 *
3105
 *    sf: 0 -> 32bit, 1 -> 64bit
3106
 *    op: 0 -> add  , 1 -> sub
3107
 *     S: 1 -> set flags
3108
 * shift: 00 -> LSL, 01 -> LSR, 10 -> ASR, 11 -> RESERVED
3109
 *  imm6: Shift amount to apply to Rm before the add/sub
3110
 */
3111
static void disas_add_sub_reg(DisasContext *s, uint32_t insn)
3112
{
3113
    int rd = extract32(insn, 0, 5);
3114
    int rn = extract32(insn, 5, 5);
3115
    int imm6 = extract32(insn, 10, 6);
3116
    int rm = extract32(insn, 16, 5);
3117
    int shift_type = extract32(insn, 22, 2);
3118
    bool setflags = extract32(insn, 29, 1);
3119
    bool sub_op = extract32(insn, 30, 1);
3120
    bool sf = extract32(insn, 31, 1);
3121

    
3122
    TCGv_i64 tcg_rd = cpu_reg(s, rd);
3123
    TCGv_i64 tcg_rn, tcg_rm;
3124
    TCGv_i64 tcg_result;
3125

    
3126
    if ((shift_type == 3) || (!sf && (imm6 > 31))) {
3127
        unallocated_encoding(s);
3128
        return;
3129
    }
3130

    
3131
    tcg_rn = read_cpu_reg(s, rn, sf);
3132
    tcg_rm = read_cpu_reg(s, rm, sf);
3133

    
3134
    shift_reg_imm(tcg_rm, tcg_rm, sf, shift_type, imm6);
3135

    
3136
    tcg_result = tcg_temp_new_i64();
3137

    
3138
    if (!setflags) {
3139
        if (sub_op) {
3140
            tcg_gen_sub_i64(tcg_result, tcg_rn, tcg_rm);
3141
        } else {
3142
            tcg_gen_add_i64(tcg_result, tcg_rn, tcg_rm);
3143
        }
3144
    } else {
3145
        if (sub_op) {
3146
            gen_sub_CC(sf, tcg_result, tcg_rn, tcg_rm);
3147
        } else {
3148
            gen_add_CC(sf, tcg_result, tcg_rn, tcg_rm);
3149
        }
3150
    }
3151

    
3152
    if (sf) {
3153
        tcg_gen_mov_i64(tcg_rd, tcg_result);
3154
    } else {
3155
        tcg_gen_ext32u_i64(tcg_rd, tcg_result);
3156
    }
3157

    
3158
    tcg_temp_free_i64(tcg_result);
3159
}
3160

    
3161
/* C3.5.9 Data-processing (3 source)
3162

3163
   31 30  29 28       24 23 21  20  16  15  14  10 9    5 4    0
3164
  +--+------+-----------+------+------+----+------+------+------+
3165
  |sf| op54 | 1 1 0 1 1 | op31 |  Rm  | o0 |  Ra  |  Rn  |  Rd  |
3166
  +--+------+-----------+------+------+----+------+------+------+
3167

3168
 */
3169
static void disas_data_proc_3src(DisasContext *s, uint32_t insn)
3170
{
3171
    int rd = extract32(insn, 0, 5);
3172
    int rn = extract32(insn, 5, 5);
3173
    int ra = extract32(insn, 10, 5);
3174
    int rm = extract32(insn, 16, 5);
3175
    int op_id = (extract32(insn, 29, 3) << 4) |
3176
        (extract32(insn, 21, 3) << 1) |
3177
        extract32(insn, 15, 1);
3178
    bool sf = extract32(insn, 31, 1);
3179
    bool is_sub = extract32(op_id, 0, 1);
3180
    bool is_high = extract32(op_id, 2, 1);
3181
    bool is_signed = false;
3182
    TCGv_i64 tcg_op1;
3183
    TCGv_i64 tcg_op2;
3184
    TCGv_i64 tcg_tmp;
3185

    
3186
    /* Note that op_id is sf:op54:op31:o0 so it includes the 32/64 size flag */
3187
    switch (op_id) {
3188
    case 0x42: /* SMADDL */
3189
    case 0x43: /* SMSUBL */
3190
    case 0x44: /* SMULH */
3191
        is_signed = true;
3192
        break;
3193
    case 0x0: /* MADD (32bit) */
3194
    case 0x1: /* MSUB (32bit) */
3195
    case 0x40: /* MADD (64bit) */
3196
    case 0x41: /* MSUB (64bit) */
3197
    case 0x4a: /* UMADDL */
3198
    case 0x4b: /* UMSUBL */
3199
    case 0x4c: /* UMULH */
3200
        break;
3201
    default:
3202
        unallocated_encoding(s);
3203
        return;
3204
    }
3205

    
3206
    if (is_high) {
3207
        TCGv_i64 low_bits = tcg_temp_new_i64(); /* low bits discarded */
3208
        TCGv_i64 tcg_rd = cpu_reg(s, rd);
3209
        TCGv_i64 tcg_rn = cpu_reg(s, rn);
3210
        TCGv_i64 tcg_rm = cpu_reg(s, rm);
3211

    
3212
        if (is_signed) {
3213
            tcg_gen_muls2_i64(low_bits, tcg_rd, tcg_rn, tcg_rm);
3214
        } else {
3215
            tcg_gen_mulu2_i64(low_bits, tcg_rd, tcg_rn, tcg_rm);
3216
        }
3217

    
3218
        tcg_temp_free_i64(low_bits);
3219
        return;
3220
    }
3221

    
3222
    tcg_op1 = tcg_temp_new_i64();
3223
    tcg_op2 = tcg_temp_new_i64();
3224
    tcg_tmp = tcg_temp_new_i64();
3225

    
3226
    if (op_id < 0x42) {
3227
        tcg_gen_mov_i64(tcg_op1, cpu_reg(s, rn));
3228
        tcg_gen_mov_i64(tcg_op2, cpu_reg(s, rm));
3229
    } else {
3230
        if (is_signed) {
3231
            tcg_gen_ext32s_i64(tcg_op1, cpu_reg(s, rn));
3232
            tcg_gen_ext32s_i64(tcg_op2, cpu_reg(s, rm));
3233
        } else {
3234
            tcg_gen_ext32u_i64(tcg_op1, cpu_reg(s, rn));
3235
            tcg_gen_ext32u_i64(tcg_op2, cpu_reg(s, rm));
3236
        }
3237
    }
3238

    
3239
    if (ra == 31 && !is_sub) {
3240
        /* Special-case MADD with rA == XZR; it is the standard MUL alias */
3241
        tcg_gen_mul_i64(cpu_reg(s, rd), tcg_op1, tcg_op2);
3242
    } else {
3243
        tcg_gen_mul_i64(tcg_tmp, tcg_op1, tcg_op2);
3244
        if (is_sub) {
3245
            tcg_gen_sub_i64(cpu_reg(s, rd), cpu_reg(s, ra), tcg_tmp);
3246
        } else {
3247
            tcg_gen_add_i64(cpu_reg(s, rd), cpu_reg(s, ra), tcg_tmp);
3248
        }
3249
    }
3250

    
3251
    if (!sf) {
3252
        tcg_gen_ext32u_i64(cpu_reg(s, rd), cpu_reg(s, rd));
3253
    }
3254

    
3255
    tcg_temp_free_i64(tcg_op1);
3256
    tcg_temp_free_i64(tcg_op2);
3257
    tcg_temp_free_i64(tcg_tmp);
3258
}
3259

    
3260
/* C3.5.3 - Add/subtract (with carry)
3261
 *  31 30 29 28 27 26 25 24 23 22 21  20  16  15   10  9    5 4   0
3262
 * +--+--+--+------------------------+------+---------+------+-----+
3263
 * |sf|op| S| 1  1  0  1  0  0  0  0 |  rm  | opcode2 |  Rn  |  Rd |
3264
 * +--+--+--+------------------------+------+---------+------+-----+
3265
 *                                            [000000]
3266
 */
3267

    
3268
static void disas_adc_sbc(DisasContext *s, uint32_t insn)
3269
{
3270
    unsigned int sf, op, setflags, rm, rn, rd;
3271
    TCGv_i64 tcg_y, tcg_rn, tcg_rd;
3272

    
3273
    if (extract32(insn, 10, 6) != 0) {
3274
        unallocated_encoding(s);
3275
        return;
3276
    }
3277

    
3278
    sf = extract32(insn, 31, 1);
3279
    op = extract32(insn, 30, 1);
3280
    setflags = extract32(insn, 29, 1);
3281
    rm = extract32(insn, 16, 5);
3282
    rn = extract32(insn, 5, 5);
3283
    rd = extract32(insn, 0, 5);
3284

    
3285
    tcg_rd = cpu_reg(s, rd);
3286
    tcg_rn = cpu_reg(s, rn);
3287

    
3288
    if (op) {
3289
        tcg_y = new_tmp_a64(s);
3290
        tcg_gen_not_i64(tcg_y, cpu_reg(s, rm));
3291
    } else {
3292
        tcg_y = cpu_reg(s, rm);
3293
    }
3294

    
3295
    if (setflags) {
3296
        gen_adc_CC(sf, tcg_rd, tcg_rn, tcg_y);
3297
    } else {
3298
        gen_adc(sf, tcg_rd, tcg_rn, tcg_y);
3299
    }
3300
}
3301

    
3302
/* C3.5.4 - C3.5.5 Conditional compare (immediate / register)
3303
 *  31 30 29 28 27 26 25 24 23 22 21  20    16 15  12  11  10  9   5  4 3   0
3304
 * +--+--+--+------------------------+--------+------+----+--+------+--+-----+
3305
 * |sf|op| S| 1  1  0  1  0  0  1  0 |imm5/rm | cond |i/r |o2|  Rn  |o3|nzcv |
3306
 * +--+--+--+------------------------+--------+------+----+--+------+--+-----+
3307
 *        [1]                             y                [0]       [0]
3308
 */
3309
static void disas_cc(DisasContext *s, uint32_t insn)
3310
{
3311
    unsigned int sf, op, y, cond, rn, nzcv, is_imm;
3312
    int label_continue = -1;
3313
    TCGv_i64 tcg_tmp, tcg_y, tcg_rn;
3314

    
3315
    if (!extract32(insn, 29, 1)) {
3316
        unallocated_encoding(s);
3317
        return;
3318
    }
3319
    if (insn & (1 << 10 | 1 << 4)) {
3320
        unallocated_encoding(s);
3321
        return;
3322
    }
3323
    sf = extract32(insn, 31, 1);
3324
    op = extract32(insn, 30, 1);
3325
    is_imm = extract32(insn, 11, 1);
3326
    y = extract32(insn, 16, 5); /* y = rm (reg) or imm5 (imm) */
3327
    cond = extract32(insn, 12, 4);
3328
    rn = extract32(insn, 5, 5);
3329
    nzcv = extract32(insn, 0, 4);
3330

    
3331
    if (cond < 0x0e) { /* not always */
3332
        int label_match = gen_new_label();
3333
        label_continue = gen_new_label();
3334
        arm_gen_test_cc(cond, label_match);
3335
        /* nomatch: */
3336
        tcg_tmp = tcg_temp_new_i64();
3337
        tcg_gen_movi_i64(tcg_tmp, nzcv << 28);
3338
        gen_set_nzcv(tcg_tmp);
3339
        tcg_temp_free_i64(tcg_tmp);
3340
        tcg_gen_br(label_continue);
3341
        gen_set_label(label_match);
3342
    }
3343
    /* match, or condition is always */
3344
    if (is_imm) {
3345
        tcg_y = new_tmp_a64(s);
3346
        tcg_gen_movi_i64(tcg_y, y);
3347
    } else {
3348
        tcg_y = cpu_reg(s, y);
3349
    }
3350
    tcg_rn = cpu_reg(s, rn);
3351

    
3352
    tcg_tmp = tcg_temp_new_i64();
3353
    if (op) {
3354
        gen_sub_CC(sf, tcg_tmp, tcg_rn, tcg_y);
3355
    } else {
3356
        gen_add_CC(sf, tcg_tmp, tcg_rn, tcg_y);
3357
    }
3358
    tcg_temp_free_i64(tcg_tmp);
3359

    
3360
    if (cond < 0x0e) { /* continue */
3361
        gen_set_label(label_continue);
3362
    }
3363
}
3364

    
3365
/* C3.5.6 Conditional select
3366
 *   31   30  29  28             21 20  16 15  12 11 10 9    5 4    0
3367
 * +----+----+---+-----------------+------+------+-----+------+------+
3368
 * | sf | op | S | 1 1 0 1 0 1 0 0 |  Rm  | cond | op2 |  Rn  |  Rd  |
3369
 * +----+----+---+-----------------+------+------+-----+------+------+
3370
 */
3371
static void disas_cond_select(DisasContext *s, uint32_t insn)
3372
{
3373
    unsigned int sf, else_inv, rm, cond, else_inc, rn, rd;
3374
    TCGv_i64 tcg_rd, tcg_src;
3375

    
3376
    if (extract32(insn, 29, 1) || extract32(insn, 11, 1)) {
3377
        /* S == 1 or op2<1> == 1 */
3378
        unallocated_encoding(s);
3379
        return;
3380
    }
3381
    sf = extract32(insn, 31, 1);
3382
    else_inv = extract32(insn, 30, 1);
3383
    rm = extract32(insn, 16, 5);
3384
    cond = extract32(insn, 12, 4);
3385
    else_inc = extract32(insn, 10, 1);
3386
    rn = extract32(insn, 5, 5);
3387
    rd = extract32(insn, 0, 5);
3388

    
3389
    if (rd == 31) {
3390
        /* silly no-op write; until we use movcond we must special-case
3391
         * this to avoid a dead temporary across basic blocks.
3392
         */
3393
        return;
3394
    }
3395

    
3396
    tcg_rd = cpu_reg(s, rd);
3397

    
3398
    if (cond >= 0x0e) { /* condition "always" */
3399
        tcg_src = read_cpu_reg(s, rn, sf);
3400
        tcg_gen_mov_i64(tcg_rd, tcg_src);
3401
    } else {
3402
        /* OPTME: we could use movcond here, at the cost of duplicating
3403
         * a lot of the arm_gen_test_cc() logic.
3404
         */
3405
        int label_match = gen_new_label();
3406
        int label_continue = gen_new_label();
3407

    
3408
        arm_gen_test_cc(cond, label_match);
3409
        /* nomatch: */
3410
        tcg_src = cpu_reg(s, rm);
3411

    
3412
        if (else_inv && else_inc) {
3413
            tcg_gen_neg_i64(tcg_rd, tcg_src);
3414
        } else if (else_inv) {
3415
            tcg_gen_not_i64(tcg_rd, tcg_src);
3416
        } else if (else_inc) {
3417
            tcg_gen_addi_i64(tcg_rd, tcg_src, 1);
3418
        } else {
3419
            tcg_gen_mov_i64(tcg_rd, tcg_src);
3420
        }
3421
        if (!sf) {
3422
            tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
3423
        }
3424
        tcg_gen_br(label_continue);
3425
        /* match: */
3426
        gen_set_label(label_match);
3427
        tcg_src = read_cpu_reg(s, rn, sf);
3428
        tcg_gen_mov_i64(tcg_rd, tcg_src);
3429
        /* continue: */
3430
        gen_set_label(label_continue);
3431
    }
3432
}
3433

    
3434
static void handle_clz(DisasContext *s, unsigned int sf,
3435
                       unsigned int rn, unsigned int rd)
3436
{
3437
    TCGv_i64 tcg_rd, tcg_rn;
3438
    tcg_rd = cpu_reg(s, rd);
3439
    tcg_rn = cpu_reg(s, rn);
3440

    
3441
    if (sf) {
3442
        gen_helper_clz64(tcg_rd, tcg_rn);
3443
    } else {
3444
        TCGv_i32 tcg_tmp32 = tcg_temp_new_i32();
3445
        tcg_gen_trunc_i64_i32(tcg_tmp32, tcg_rn);
3446
        gen_helper_clz(tcg_tmp32, tcg_tmp32);
3447
        tcg_gen_extu_i32_i64(tcg_rd, tcg_tmp32);
3448
        tcg_temp_free_i32(tcg_tmp32);
3449
    }
3450
}
3451

    
3452
static void handle_cls(DisasContext *s, unsigned int sf,
3453
                       unsigned int rn, unsigned int rd)
3454
{
3455
    TCGv_i64 tcg_rd, tcg_rn;
3456
    tcg_rd = cpu_reg(s, rd);
3457
    tcg_rn = cpu_reg(s, rn);
3458

    
3459
    if (sf) {
3460
        gen_helper_cls64(tcg_rd, tcg_rn);
3461
    } else {
3462
        TCGv_i32 tcg_tmp32 = tcg_temp_new_i32();
3463
        tcg_gen_trunc_i64_i32(tcg_tmp32, tcg_rn);
3464
        gen_helper_cls32(tcg_tmp32, tcg_tmp32);
3465
        tcg_gen_extu_i32_i64(tcg_rd, tcg_tmp32);
3466
        tcg_temp_free_i32(tcg_tmp32);
3467
    }
3468
}
3469

    
3470
static void handle_rbit(DisasContext *s, unsigned int sf,
3471
                        unsigned int rn, unsigned int rd)
3472
{
3473
    TCGv_i64 tcg_rd, tcg_rn;
3474
    tcg_rd = cpu_reg(s, rd);
3475
    tcg_rn = cpu_reg(s, rn);
3476

    
3477
    if (sf) {
3478
        gen_helper_rbit64(tcg_rd, tcg_rn);
3479
    } else {
3480
        TCGv_i32 tcg_tmp32 = tcg_temp_new_i32();
3481
        tcg_gen_trunc_i64_i32(tcg_tmp32, tcg_rn);
3482
        gen_helper_rbit(tcg_tmp32, tcg_tmp32);
3483
        tcg_gen_extu_i32_i64(tcg_rd, tcg_tmp32);
3484
        tcg_temp_free_i32(tcg_tmp32);
3485
    }
3486
}
3487

    
3488
/* C5.6.149 REV with sf==1, opcode==3 ("REV64") */
3489
static void handle_rev64(DisasContext *s, unsigned int sf,
3490
                         unsigned int rn, unsigned int rd)
3491
{
3492
    if (!sf) {
3493
        unallocated_encoding(s);
3494
        return;
3495
    }
3496
    tcg_gen_bswap64_i64(cpu_reg(s, rd), cpu_reg(s, rn));
3497
}
3498

    
3499
/* C5.6.149 REV with sf==0, opcode==2
3500
 * C5.6.151 REV32 (sf==1, opcode==2)
3501
 */
3502
static void handle_rev32(DisasContext *s, unsigned int sf,
3503
                         unsigned int rn, unsigned int rd)
3504
{
3505
    TCGv_i64 tcg_rd = cpu_reg(s, rd);
3506

    
3507
    if (sf) {
3508
        TCGv_i64 tcg_tmp = tcg_temp_new_i64();
3509
        TCGv_i64 tcg_rn = read_cpu_reg(s, rn, sf);
3510

    
3511
        /* bswap32_i64 requires zero high word */
3512
        tcg_gen_ext32u_i64(tcg_tmp, tcg_rn);
3513
        tcg_gen_bswap32_i64(tcg_rd, tcg_tmp);
3514
        tcg_gen_shri_i64(tcg_tmp, tcg_rn, 32);
3515
        tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp);
3516
        tcg_gen_concat32_i64(tcg_rd, tcg_rd, tcg_tmp);
3517

    
3518
        tcg_temp_free_i64(tcg_tmp);
3519
    } else {
3520
        tcg_gen_ext32u_i64(tcg_rd, cpu_reg(s, rn));
3521
        tcg_gen_bswap32_i64(tcg_rd, tcg_rd);
3522
    }
3523
}
3524

    
3525
/* C5.6.150 REV16 (opcode==1) */
3526
static void handle_rev16(DisasContext *s, unsigned int sf,
3527
                         unsigned int rn, unsigned int rd)
3528
{
3529
    TCGv_i64 tcg_rd = cpu_reg(s, rd);
3530
    TCGv_i64 tcg_tmp = tcg_temp_new_i64();
3531
    TCGv_i64 tcg_rn = read_cpu_reg(s, rn, sf);
3532

    
3533
    tcg_gen_andi_i64(tcg_tmp, tcg_rn, 0xffff);
3534
    tcg_gen_bswap16_i64(tcg_rd, tcg_tmp);
3535

    
3536
    tcg_gen_shri_i64(tcg_tmp, tcg_rn, 16);
3537
    tcg_gen_andi_i64(tcg_tmp, tcg_tmp, 0xffff);
3538
    tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp);
3539
    tcg_gen_deposit_i64(tcg_rd, tcg_rd, tcg_tmp, 16, 16);
3540

    
3541
    if (sf) {
3542
        tcg_gen_shri_i64(tcg_tmp, tcg_rn, 32);
3543
        tcg_gen_andi_i64(tcg_tmp, tcg_tmp, 0xffff);
3544
        tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp);
3545
        tcg_gen_deposit_i64(tcg_rd, tcg_rd, tcg_tmp, 32, 16);
3546

    
3547
        tcg_gen_shri_i64(tcg_tmp, tcg_rn, 48);
3548
        tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp);
3549
        tcg_gen_deposit_i64(tcg_rd, tcg_rd, tcg_tmp, 48, 16);
3550
    }
3551

    
3552
    tcg_temp_free_i64(tcg_tmp);
3553
}
3554

    
3555
/* C3.5.7 Data-processing (1 source)
3556
 *   31  30  29  28             21 20     16 15    10 9    5 4    0
3557
 * +----+---+---+-----------------+---------+--------+------+------+
3558
 * | sf | 1 | S | 1 1 0 1 0 1 1 0 | opcode2 | opcode |  Rn  |  Rd  |
3559
 * +----+---+---+-----------------+---------+--------+------+------+
3560
 */
3561
static void disas_data_proc_1src(DisasContext *s, uint32_t insn)
3562
{
3563
    unsigned int sf, opcode, rn, rd;
3564

    
3565
    if (extract32(insn, 29, 1) || extract32(insn, 16, 5)) {
3566
        unallocated_encoding(s);
3567
        return;
3568
    }
3569

    
3570
    sf = extract32(insn, 31, 1);
3571
    opcode = extract32(insn, 10, 6);
3572
    rn = extract32(insn, 5, 5);
3573
    rd = extract32(insn, 0, 5);
3574

    
3575
    switch (opcode) {
3576
    case 0: /* RBIT */
3577
        handle_rbit(s, sf, rn, rd);
3578
        break;
3579
    case 1: /* REV16 */
3580
        handle_rev16(s, sf, rn, rd);
3581
        break;
3582
    case 2: /* REV32 */
3583
        handle_rev32(s, sf, rn, rd);
3584
        break;
3585
    case 3: /* REV64 */
3586
        handle_rev64(s, sf, rn, rd);
3587
        break;
3588
    case 4: /* CLZ */
3589
        handle_clz(s, sf, rn, rd);
3590
        break;
3591
    case 5: /* CLS */
3592
        handle_cls(s, sf, rn, rd);
3593
        break;
3594
    }
3595
}
3596

    
3597
static void handle_div(DisasContext *s, bool is_signed, unsigned int sf,
3598
                       unsigned int rm, unsigned int rn, unsigned int rd)
3599
{
3600
    TCGv_i64 tcg_n, tcg_m, tcg_rd;
3601
    tcg_rd = cpu_reg(s, rd);
3602

    
3603
    if (!sf && is_signed) {
3604
        tcg_n = new_tmp_a64(s);
3605
        tcg_m = new_tmp_a64(s);
3606
        tcg_gen_ext32s_i64(tcg_n, cpu_reg(s, rn));
3607
        tcg_gen_ext32s_i64(tcg_m, cpu_reg(s, rm));
3608
    } else {
3609
        tcg_n = read_cpu_reg(s, rn, sf);
3610
        tcg_m = read_cpu_reg(s, rm, sf);
3611
    }
3612

    
3613
    if (is_signed) {
3614
        gen_helper_sdiv64(tcg_rd, tcg_n, tcg_m);
3615
    } else {
3616
        gen_helper_udiv64(tcg_rd, tcg_n, tcg_m);
3617
    }
3618

    
3619
    if (!sf) { /* zero extend final result */
3620
        tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
3621
    }
3622
}
3623

    
3624
/* C5.6.115 LSLV, C5.6.118 LSRV, C5.6.17 ASRV, C5.6.154 RORV */
3625
static void handle_shift_reg(DisasContext *s,
3626
                             enum a64_shift_type shift_type, unsigned int sf,
3627
                             unsigned int rm, unsigned int rn, unsigned int rd)
3628
{
3629
    TCGv_i64 tcg_shift = tcg_temp_new_i64();
3630
    TCGv_i64 tcg_rd = cpu_reg(s, rd);
3631
    TCGv_i64 tcg_rn = read_cpu_reg(s, rn, sf);
3632

    
3633
    tcg_gen_andi_i64(tcg_shift, cpu_reg(s, rm), sf ? 63 : 31);
3634
    shift_reg(tcg_rd, tcg_rn, sf, shift_type, tcg_shift);
3635
    tcg_temp_free_i64(tcg_shift);
3636
}
3637

    
3638
/* C3.5.8 Data-processing (2 source)
3639
 *   31   30  29 28             21 20  16 15    10 9    5 4    0
3640
 * +----+---+---+-----------------+------+--------+------+------+
3641
 * | sf | 0 | S | 1 1 0 1 0 1 1 0 |  Rm  | opcode |  Rn  |  Rd  |
3642
 * +----+---+---+-----------------+------+--------+------+------+
3643
 */
3644
static void disas_data_proc_2src(DisasContext *s, uint32_t insn)
3645
{
3646
    unsigned int sf, rm, opcode, rn, rd;
3647
    sf = extract32(insn, 31, 1);
3648
    rm = extract32(insn, 16, 5);
3649
    opcode = extract32(insn, 10, 6);
3650
    rn = extract32(insn, 5, 5);
3651
    rd = extract32(insn, 0, 5);
3652

    
3653
    if (extract32(insn, 29, 1)) {
3654
        unallocated_encoding(s);
3655
        return;
3656
    }
3657

    
3658
    switch (opcode) {
3659
    case 2: /* UDIV */
3660
        handle_div(s, false, sf, rm, rn, rd);
3661
        break;
3662
    case 3: /* SDIV */
3663
        handle_div(s, true, sf, rm, rn, rd);
3664
        break;
3665
    case 8: /* LSLV */
3666
        handle_shift_reg(s, A64_SHIFT_TYPE_LSL, sf, rm, rn, rd);
3667
        break;
3668
    case 9: /* LSRV */
3669
        handle_shift_reg(s, A64_SHIFT_TYPE_LSR, sf, rm, rn, rd);
3670
        break;
3671
    case 10: /* ASRV */
3672
        handle_shift_reg(s, A64_SHIFT_TYPE_ASR, sf, rm, rn, rd);
3673
        break;
3674
    case 11: /* RORV */
3675
        handle_shift_reg(s, A64_SHIFT_TYPE_ROR, sf, rm, rn, rd);
3676
        break;
3677
    case 16:
3678
    case 17:
3679
    case 18:
3680
    case 19:
3681
    case 20:
3682
    case 21:
3683
    case 22:
3684
    case 23: /* CRC32 */
3685
        unsupported_encoding(s, insn);
3686
        break;
3687
    default:
3688
        unallocated_encoding(s);
3689
        break;
3690
    }
3691
}
3692

    
3693
/* C3.5 Data processing - register */
3694
static void disas_data_proc_reg(DisasContext *s, uint32_t insn)
3695
{
3696
    switch (extract32(insn, 24, 5)) {
3697
    case 0x0a: /* Logical (shifted register) */
3698
        disas_logic_reg(s, insn);
3699
        break;
3700
    case 0x0b: /* Add/subtract */
3701
        if (insn & (1 << 21)) { /* (extended register) */
3702
            disas_add_sub_ext_reg(s, insn);
3703
        } else {
3704
            disas_add_sub_reg(s, insn);
3705
        }
3706
        break;
3707
    case 0x1b: /* Data-processing (3 source) */
3708
        disas_data_proc_3src(s, insn);
3709
        break;
3710
    case 0x1a:
3711
        switch (extract32(insn, 21, 3)) {
3712
        case 0x0: /* Add/subtract (with carry) */
3713
            disas_adc_sbc(s, insn);
3714
            break;
3715
        case 0x2: /* Conditional compare */
3716
            disas_cc(s, insn); /* both imm and reg forms */
3717
            break;
3718
        case 0x4: /* Conditional select */
3719
            disas_cond_select(s, insn);
3720
            break;
3721
        case 0x6: /* Data-processing */
3722
            if (insn & (1 << 30)) { /* (1 source) */
3723
                disas_data_proc_1src(s, insn);
3724
            } else {            /* (2 source) */
3725
                disas_data_proc_2src(s, insn);
3726
            }
3727
            break;
3728
        default:
3729
            unallocated_encoding(s);
3730
            break;
3731
        }
3732
        break;
3733
    default:
3734
        unallocated_encoding(s);
3735
        break;
3736
    }
3737
}
3738

    
3739
static void handle_fp_compare(DisasContext *s, bool is_double,
3740
                              unsigned int rn, unsigned int rm,
3741
                              bool cmp_with_zero, bool signal_all_nans)
3742
{
3743
    TCGv_i64 tcg_flags = tcg_temp_new_i64();
3744
    TCGv_ptr fpst = get_fpstatus_ptr();
3745

    
3746
    if (is_double) {
3747
        TCGv_i64 tcg_vn, tcg_vm;
3748

    
3749
        tcg_vn = read_fp_dreg(s, rn);
3750
        if (cmp_with_zero) {
3751
            tcg_vm = tcg_const_i64(0);
3752
        } else {
3753
            tcg_vm = read_fp_dreg(s, rm);
3754
        }
3755
        if (signal_all_nans) {
3756
            gen_helper_vfp_cmped_a64(tcg_flags, tcg_vn, tcg_vm, fpst);
3757
        } else {
3758
            gen_helper_vfp_cmpd_a64(tcg_flags, tcg_vn, tcg_vm, fpst);
3759
        }
3760
        tcg_temp_free_i64(tcg_vn);
3761
        tcg_temp_free_i64(tcg_vm);
3762
    } else {
3763
        TCGv_i32 tcg_vn, tcg_vm;
3764

    
3765
        tcg_vn = read_fp_sreg(s, rn);
3766
        if (cmp_with_zero) {
3767
            tcg_vm = tcg_const_i32(0);
3768
        } else {
3769
            tcg_vm = read_fp_sreg(s, rm);
3770
        }
3771
        if (signal_all_nans) {
3772
            gen_helper_vfp_cmpes_a64(tcg_flags, tcg_vn, tcg_vm, fpst);
3773
        } else {
3774
            gen_helper_vfp_cmps_a64(tcg_flags, tcg_vn, tcg_vm, fpst);
3775
        }
3776
        tcg_temp_free_i32(tcg_vn);
3777
        tcg_temp_free_i32(tcg_vm);
3778
    }
3779

    
3780
    tcg_temp_free_ptr(fpst);
3781

    
3782
    gen_set_nzcv(tcg_flags);
3783

    
3784
    tcg_temp_free_i64(tcg_flags);
3785
}
3786

    
3787
/* C3.6.22 Floating point compare
3788
 *   31  30  29 28       24 23  22  21 20  16 15 14 13  10    9    5 4     0
3789
 * +---+---+---+-----------+------+---+------+-----+---------+------+-------+
3790
 * | M | 0 | S | 1 1 1 1 0 | type | 1 |  Rm  | op  | 1 0 0 0 |  Rn  |  op2  |
3791
 * +---+---+---+-----------+------+---+------+-----+---------+------+-------+
3792
 */
3793
static void disas_fp_compare(DisasContext *s, uint32_t insn)
3794
{
3795
    unsigned int mos, type, rm, op, rn, opc, op2r;
3796

    
3797
    mos = extract32(insn, 29, 3);
3798
    type = extract32(insn, 22, 2); /* 0 = single, 1 = double */
3799
    rm = extract32(insn, 16, 5);
3800
    op = extract32(insn, 14, 2);
3801
    rn = extract32(insn, 5, 5);
3802
    opc = extract32(insn, 3, 2);
3803
    op2r = extract32(insn, 0, 3);
3804

    
3805
    if (mos || op || op2r || type > 1) {
3806
        unallocated_encoding(s);
3807
        return;
3808
    }
3809

    
3810
    handle_fp_compare(s, type, rn, rm, opc & 1, opc & 2);
3811
}
3812

    
3813
/* C3.6.23 Floating point conditional compare
3814
 *   31  30  29 28       24 23  22  21 20  16 15  12 11 10 9    5  4   3    0
3815
 * +---+---+---+-----------+------+---+------+------+-----+------+----+------+
3816
 * | M | 0 | S | 1 1 1 1 0 | type | 1 |  Rm  | cond | 0 1 |  Rn  | op | nzcv |
3817
 * +---+---+---+-----------+------+---+------+------+-----+------+----+------+
3818
 */
3819
static void disas_fp_ccomp(DisasContext *s, uint32_t insn)
3820
{
3821
    unsigned int mos, type, rm, cond, rn, op, nzcv;
3822
    TCGv_i64 tcg_flags;
3823
    int label_continue = -1;
3824

    
3825
    mos = extract32(insn, 29, 3);
3826
    type = extract32(insn, 22, 2); /* 0 = single, 1 = double */
3827
    rm = extract32(insn, 16, 5);
3828
    cond = extract32(insn, 12, 4);
3829
    rn = extract32(insn, 5, 5);
3830
    op = extract32(insn, 4, 1);
3831
    nzcv = extract32(insn, 0, 4);
3832

    
3833
    if (mos || type > 1) {
3834
        unallocated_encoding(s);
3835
        return;
3836
    }
3837

    
3838
    if (cond < 0x0e) { /* not always */
3839
        int label_match = gen_new_label();
3840
        label_continue = gen_new_label();
3841
        arm_gen_test_cc(cond, label_match);
3842
        /* nomatch: */
3843
        tcg_flags = tcg_const_i64(nzcv << 28);
3844
        gen_set_nzcv(tcg_flags);
3845
        tcg_temp_free_i64(tcg_flags);
3846
        tcg_gen_br(label_continue);
3847
        gen_set_label(label_match);
3848
    }
3849

    
3850
    handle_fp_compare(s, type, rn, rm, false, op);
3851

    
3852
    if (cond < 0x0e) {
3853
        gen_set_label(label_continue);
3854
    }
3855
}
3856

    
3857
/* copy src FP register to dst FP register; type specifies single or double */
3858
static void gen_mov_fp2fp(DisasContext *s, int type, int dst, int src)
3859
{
3860
    if (type) {
3861
        TCGv_i64 v = read_fp_dreg(s, src);
3862
        write_fp_dreg(s, dst, v);
3863
        tcg_temp_free_i64(v);
3864
    } else {
3865
        TCGv_i32 v = read_fp_sreg(s, src);
3866
        write_fp_sreg(s, dst, v);
3867
        tcg_temp_free_i32(v);
3868
    }
3869
}
3870

    
3871
/* C3.6.24 Floating point conditional select
3872
 *   31  30  29 28       24 23  22  21 20  16 15  12 11 10 9    5 4    0
3873
 * +---+---+---+-----------+------+---+------+------+-----+------+------+
3874
 * | M | 0 | S | 1 1 1 1 0 | type | 1 |  Rm  | cond | 1 1 |  Rn  |  Rd  |
3875
 * +---+---+---+-----------+------+---+------+------+-----+------+------+
3876
 */
3877
static void disas_fp_csel(DisasContext *s, uint32_t insn)
3878
{
3879
    unsigned int mos, type, rm, cond, rn, rd;
3880
    int label_continue = -1;
3881

    
3882
    mos = extract32(insn, 29, 3);
3883
    type = extract32(insn, 22, 2); /* 0 = single, 1 = double */
3884
    rm = extract32(insn, 16, 5);
3885
    cond = extract32(insn, 12, 4);
3886
    rn = extract32(insn, 5, 5);
3887
    rd = extract32(insn, 0, 5);
3888

    
3889
    if (mos || type > 1) {
3890
        unallocated_encoding(s);
3891
        return;
3892
    }
3893

    
3894
    if (cond < 0x0e) { /* not always */
3895
        int label_match = gen_new_label();
3896
        label_continue = gen_new_label();
3897
        arm_gen_test_cc(cond, label_match);
3898
        /* nomatch: */
3899
        gen_mov_fp2fp(s, type, rd, rm);
3900
        tcg_gen_br(label_continue);
3901
        gen_set_label(label_match);
3902
    }
3903

    
3904
    gen_mov_fp2fp(s, type, rd, rn);
3905

    
3906
    if (cond < 0x0e) { /* continue */
3907
        gen_set_label(label_continue);
3908
    }
3909
}
3910

    
3911
/* C3.6.25 Floating-point data-processing (1 source) - single precision */
3912
static void handle_fp_1src_single(DisasContext *s, int opcode, int rd, int rn)
3913
{
3914
    TCGv_ptr fpst;
3915
    TCGv_i32 tcg_op;
3916
    TCGv_i32 tcg_res;
3917

    
3918
    fpst = get_fpstatus_ptr();
3919
    tcg_op = read_fp_sreg(s, rn);
3920
    tcg_res = tcg_temp_new_i32();
3921

    
3922
    switch (opcode) {
3923
    case 0x0: /* FMOV */
3924
        tcg_gen_mov_i32(tcg_res, tcg_op);
3925
        break;
3926
    case 0x1: /* FABS */
3927
        gen_helper_vfp_abss(tcg_res, tcg_op);
3928
        break;
3929
    case 0x2: /* FNEG */
3930
        gen_helper_vfp_negs(tcg_res, tcg_op);
3931
        break;
3932
    case 0x3: /* FSQRT */
3933
        gen_helper_vfp_sqrts(tcg_res, tcg_op, cpu_env);
3934
        break;
3935
    case 0x8: /* FRINTN */
3936
    case 0x9: /* FRINTP */
3937
    case 0xa: /* FRINTM */
3938
    case 0xb: /* FRINTZ */
3939
    case 0xc: /* FRINTA */
3940
    {
3941
        TCGv_i32 tcg_rmode = tcg_const_i32(arm_rmode_to_sf(opcode & 7));
3942

    
3943
        gen_helper_set_rmode(tcg_rmode, tcg_rmode, cpu_env);
3944
        gen_helper_rints(tcg_res, tcg_op, fpst);
3945

    
3946
        gen_helper_set_rmode(tcg_rmode, tcg_rmode, cpu_env);
3947
        tcg_temp_free_i32(tcg_rmode);
3948
        break;
3949
    }
3950
    case 0xe: /* FRINTX */
3951
        gen_helper_rints_exact(tcg_res, tcg_op, fpst);
3952
        break;
3953
    case 0xf: /* FRINTI */
3954
        gen_helper_rints(tcg_res, tcg_op, fpst);
3955
        break;
3956
    default:
3957
        abort();
3958
    }
3959

    
3960
    write_fp_sreg(s, rd, tcg_res);
3961

    
3962
    tcg_temp_free_ptr(fpst);
3963
    tcg_temp_free_i32(tcg_op);
3964
    tcg_temp_free_i32(tcg_res);
3965
}
3966

    
3967
/* C3.6.25 Floating-point data-processing (1 source) - double precision */
3968
static void handle_fp_1src_double(DisasContext *s, int opcode, int rd, int rn)
3969
{
3970
    TCGv_ptr fpst;
3971
    TCGv_i64 tcg_op;
3972
    TCGv_i64 tcg_res;
3973

    
3974
    fpst = get_fpstatus_ptr();
3975
    tcg_op = read_fp_dreg(s, rn);
3976
    tcg_res = tcg_temp_new_i64();
3977

    
3978
    switch (opcode) {
3979
    case 0x0: /* FMOV */
3980
        tcg_gen_mov_i64(tcg_res, tcg_op);
3981
        break;
3982
    case 0x1: /* FABS */
3983
        gen_helper_vfp_absd(tcg_res, tcg_op);
3984
        break;
3985
    case 0x2: /* FNEG */
3986
        gen_helper_vfp_negd(tcg_res, tcg_op);
3987
        break;
3988
    case 0x3: /* FSQRT */
3989
        gen_helper_vfp_sqrtd(tcg_res, tcg_op, cpu_env);
3990
        break;
3991
    case 0x8: /* FRINTN */
3992
    case 0x9: /* FRINTP */
3993
    case 0xa: /* FRINTM */
3994
    case 0xb: /* FRINTZ */
3995
    case 0xc: /* FRINTA */
3996
    {
3997
        TCGv_i32 tcg_rmode = tcg_const_i32(arm_rmode_to_sf(opcode & 7));
3998

    
3999
        gen_helper_set_rmode(tcg_rmode, tcg_rmode, cpu_env);
4000
        gen_helper_rintd(tcg_res, tcg_op, fpst);
4001

    
4002
        gen_helper_set_rmode(tcg_rmode, tcg_rmode, cpu_env);
4003
        tcg_temp_free_i32(tcg_rmode);
4004
        break;
4005
    }
4006
    case 0xe: /* FRINTX */
4007
        gen_helper_rintd_exact(tcg_res, tcg_op, fpst);
4008
        break;
4009
    case 0xf: /* FRINTI */
4010
        gen_helper_rintd(tcg_res, tcg_op, fpst);
4011
        break;
4012
    default:
4013
        abort();
4014
    }
4015

    
4016
    write_fp_dreg(s, rd, tcg_res);
4017

    
4018
    tcg_temp_free_ptr(fpst);
4019
    tcg_temp_free_i64(tcg_op);
4020
    tcg_temp_free_i64(tcg_res);
4021
}
4022

    
4023
static void handle_fp_fcvt(DisasContext *s, int opcode,
4024
                           int rd, int rn, int dtype, int ntype)
4025
{
4026
    switch (ntype) {
4027
    case 0x0:
4028
    {
4029
        TCGv_i32 tcg_rn = read_fp_sreg(s, rn);
4030
        if (dtype == 1) {
4031
            /* Single to double */
4032
            TCGv_i64 tcg_rd = tcg_temp_new_i64();
4033
            gen_helper_vfp_fcvtds(tcg_rd, tcg_rn, cpu_env);
4034
            write_fp_dreg(s, rd, tcg_rd);
4035
            tcg_temp_free_i64(tcg_rd);
4036
        } else {
4037
            /* Single to half */
4038
            TCGv_i32 tcg_rd = tcg_temp_new_i32();
4039
            gen_helper_vfp_fcvt_f32_to_f16(tcg_rd, tcg_rn, cpu_env);
4040
            /* write_fp_sreg is OK here because top half of tcg_rd is zero */
4041
            write_fp_sreg(s, rd, tcg_rd);
4042
            tcg_temp_free_i32(tcg_rd);
4043
        }
4044
        tcg_temp_free_i32(tcg_rn);
4045
        break;
4046
    }
4047
    case 0x1:
4048
    {
4049
        TCGv_i64 tcg_rn = read_fp_dreg(s, rn);
4050
        TCGv_i32 tcg_rd = tcg_temp_new_i32();
4051
        if (dtype == 0) {
4052
            /* Double to single */
4053
            gen_helper_vfp_fcvtsd(tcg_rd, tcg_rn, cpu_env);
4054
        } else {
4055
            /* Double to half */
4056
            gen_helper_vfp_fcvt_f64_to_f16(tcg_rd, tcg_rn, cpu_env);
4057
            /* write_fp_sreg is OK here because top half of tcg_rd is zero */
4058
        }
4059
        write_fp_sreg(s, rd, tcg_rd);
4060
        tcg_temp_free_i32(tcg_rd);
4061
        tcg_temp_free_i64(tcg_rn);
4062
        break;
4063
    }
4064
    case 0x3:
4065
    {
4066
        TCGv_i32 tcg_rn = read_fp_sreg(s, rn);
4067
        tcg_gen_ext16u_i32(tcg_rn, tcg_rn);
4068
        if (dtype == 0) {
4069
            /* Half to single */
4070
            TCGv_i32 tcg_rd = tcg_temp_new_i32();
4071
            gen_helper_vfp_fcvt_f16_to_f32(tcg_rd, tcg_rn, cpu_env);
4072
            write_fp_sreg(s, rd, tcg_rd);
4073
            tcg_temp_free_i32(tcg_rd);
4074
        } else {
4075
            /* Half to double */
4076
            TCGv_i64 tcg_rd = tcg_temp_new_i64();
4077
            gen_helper_vfp_fcvt_f16_to_f64(tcg_rd, tcg_rn, cpu_env);
4078
            write_fp_dreg(s, rd, tcg_rd);
4079
            tcg_temp_free_i64(tcg_rd);
4080
        }
4081
        tcg_temp_free_i32(tcg_rn);
4082
        break;
4083
    }
4084
    default:
4085
        abort();
4086
    }
4087
}
4088

    
4089
/* C3.6.25 Floating point data-processing (1 source)
4090
 *   31  30  29 28       24 23  22  21 20    15 14       10 9    5 4    0
4091
 * +---+---+---+-----------+------+---+--------+-----------+------+------+
4092
 * | M | 0 | S | 1 1 1 1 0 | type | 1 | opcode | 1 0 0 0 0 |  Rn  |  Rd  |
4093
 * +---+---+---+-----------+------+---+--------+-----------+------+------+
4094
 */
4095
static void disas_fp_1src(DisasContext *s, uint32_t insn)
4096
{
4097
    int type = extract32(insn, 22, 2);
4098
    int opcode = extract32(insn, 15, 6);
4099
    int rn = extract32(insn, 5, 5);
4100
    int rd = extract32(insn, 0, 5);
4101

    
4102
    switch (opcode) {
4103
    case 0x4: case 0x5: case 0x7:
4104
    {
4105
        /* FCVT between half, single and double precision */
4106
        int dtype = extract32(opcode, 0, 2);
4107
        if (type == 2 || dtype == type) {
4108
            unallocated_encoding(s);
4109
            return;
4110
        }
4111
        handle_fp_fcvt(s, opcode, rd, rn, dtype, type);
4112
        break;
4113
    }
4114
    case 0x0 ... 0x3:
4115
    case 0x8 ... 0xc:
4116
    case 0xe ... 0xf:
4117
        /* 32-to-32 and 64-to-64 ops */
4118
        switch (type) {
4119
        case 0:
4120
            handle_fp_1src_single(s, opcode, rd, rn);
4121
            break;
4122
        case 1:
4123
            handle_fp_1src_double(s, opcode, rd, rn);
4124
            break;
4125
        default:
4126
            unallocated_encoding(s);
4127
        }
4128
        break;
4129
    default:
4130
        unallocated_encoding(s);
4131
        break;
4132
    }
4133
}
4134

    
4135
/* C3.6.26 Floating-point data-processing (2 source) - single precision */
4136
static void handle_fp_2src_single(DisasContext *s, int opcode,
4137
                                  int rd, int rn, int rm)
4138
{
4139
    TCGv_i32 tcg_op1;
4140
    TCGv_i32 tcg_op2;
4141
    TCGv_i32 tcg_res;
4142
    TCGv_ptr fpst;
4143

    
4144
    tcg_res = tcg_temp_new_i32();
4145
    fpst = get_fpstatus_ptr();
4146
    tcg_op1 = read_fp_sreg(s, rn);
4147
    tcg_op2 = read_fp_sreg(s, rm);
4148

    
4149
    switch (opcode) {
4150
    case 0x0: /* FMUL */
4151
        gen_helper_vfp_muls(tcg_res, tcg_op1, tcg_op2, fpst);
4152
        break;
4153
    case 0x1: /* FDIV */
4154
        gen_helper_vfp_divs(tcg_res, tcg_op1, tcg_op2, fpst);
4155
        break;
4156
    case 0x2: /* FADD */
4157
        gen_helper_vfp_adds(tcg_res, tcg_op1, tcg_op2, fpst);
4158
        break;
4159
    case 0x3: /* FSUB */
4160
        gen_helper_vfp_subs(tcg_res, tcg_op1, tcg_op2, fpst);
4161
        break;
4162
    case 0x4: /* FMAX */
4163
        gen_helper_vfp_maxs(tcg_res, tcg_op1, tcg_op2, fpst);
4164
        break;
4165
    case 0x5: /* FMIN */
4166
        gen_helper_vfp_mins(tcg_res, tcg_op1, tcg_op2, fpst);
4167
        break;
4168
    case 0x6: /* FMAXNM */
4169
        gen_helper_vfp_maxnums(tcg_res, tcg_op1, tcg_op2, fpst);
4170
        break;
4171
    case 0x7: /* FMINNM */
4172
        gen_helper_vfp_minnums(tcg_res, tcg_op1, tcg_op2, fpst);
4173
        break;
4174
    case 0x8: /* FNMUL */
4175
        gen_helper_vfp_muls(tcg_res, tcg_op1, tcg_op2, fpst);
4176
        gen_helper_vfp_negs(tcg_res, tcg_res);
4177
        break;
4178
    }
4179

    
4180
    write_fp_sreg(s, rd, tcg_res);
4181

    
4182
    tcg_temp_free_ptr(fpst);
4183
    tcg_temp_free_i32(tcg_op1);
4184
    tcg_temp_free_i32(tcg_op2);
4185
    tcg_temp_free_i32(tcg_res);
4186
}
4187

    
4188
/* C3.6.26 Floating-point data-processing (2 source) - double precision */
4189
static void handle_fp_2src_double(DisasContext *s, int opcode,
4190
                                  int rd, int rn, int rm)
4191
{
4192
    TCGv_i64 tcg_op1;
4193
    TCGv_i64 tcg_op2;
4194
    TCGv_i64 tcg_res;
4195
    TCGv_ptr fpst;
4196

    
4197
    tcg_res = tcg_temp_new_i64();
4198
    fpst = get_fpstatus_ptr();
4199
    tcg_op1 = read_fp_dreg(s, rn);
4200
    tcg_op2 = read_fp_dreg(s, rm);
4201

    
4202
    switch (opcode) {
4203
    case 0x0: /* FMUL */
4204
        gen_helper_vfp_muld(tcg_res, tcg_op1, tcg_op2, fpst);
4205
        break;
4206
    case 0x1: /* FDIV */
4207
        gen_helper_vfp_divd(tcg_res, tcg_op1, tcg_op2, fpst);
4208
        break;
4209
    case 0x2: /* FADD */
4210
        gen_helper_vfp_addd(tcg_res, tcg_op1, tcg_op2, fpst);
4211
        break;
4212
    case 0x3: /* FSUB */
4213
        gen_helper_vfp_subd(tcg_res, tcg_op1, tcg_op2, fpst);
4214
        break;
4215
    case 0x4: /* FMAX */
4216
        gen_helper_vfp_maxd(tcg_res, tcg_op1, tcg_op2, fpst);
4217
        break;
4218
    case 0x5: /* FMIN */
4219
        gen_helper_vfp_mind(tcg_res, tcg_op1, tcg_op2, fpst);
4220
        break;
4221
    case 0x6: /* FMAXNM */
4222
        gen_helper_vfp_maxnumd(tcg_res, tcg_op1, tcg_op2, fpst);
4223
        break;
4224
    case 0x7: /* FMINNM */
4225
        gen_helper_vfp_minnumd(tcg_res, tcg_op1, tcg_op2, fpst);
4226
        break;
4227
    case 0x8: /* FNMUL */
4228
        gen_helper_vfp_muld(tcg_res, tcg_op1, tcg_op2, fpst);
4229
        gen_helper_vfp_negd(tcg_res, tcg_res);
4230
        break;
4231
    }
4232

    
4233
    write_fp_dreg(s, rd, tcg_res);
4234

    
4235
    tcg_temp_free_ptr(fpst);
4236
    tcg_temp_free_i64(tcg_op1);
4237
    tcg_temp_free_i64(tcg_op2);
4238
    tcg_temp_free_i64(tcg_res);
4239
}
4240

    
4241
/* C3.6.26 Floating point data-processing (2 source)
4242
 *   31  30  29 28       24 23  22  21 20  16 15    12 11 10 9    5 4    0
4243
 * +---+---+---+-----------+------+---+------+--------+-----+------+------+
4244
 * | M | 0 | S | 1 1 1 1 0 | type | 1 |  Rm  | opcode | 1 0 |  Rn  |  Rd  |
4245
 * +---+---+---+-----------+------+---+------+--------+-----+------+------+
4246
 */
4247
static void disas_fp_2src(DisasContext *s, uint32_t insn)
4248
{
4249
    int type = extract32(insn, 22, 2);
4250
    int rd = extract32(insn, 0, 5);
4251
    int rn = extract32(insn, 5, 5);
4252
    int rm = extract32(insn, 16, 5);
4253
    int opcode = extract32(insn, 12, 4);
4254

    
4255
    if (opcode > 8) {
4256
        unallocated_encoding(s);
4257
        return;
4258
    }
4259

    
4260
    switch (type) {
4261
    case 0:
4262
        handle_fp_2src_single(s, opcode, rd, rn, rm);
4263
        break;
4264
    case 1:
4265
        handle_fp_2src_double(s, opcode, rd, rn, rm);
4266
        break;
4267
    default:
4268
        unallocated_encoding(s);
4269
    }
4270
}
4271

    
4272
/* C3.6.27 Floating-point data-processing (3 source) - single precision */
4273
static void handle_fp_3src_single(DisasContext *s, bool o0, bool o1,
4274
                                  int rd, int rn, int rm, int ra)
4275
{
4276
    TCGv_i32 tcg_op1, tcg_op2, tcg_op3;
4277
    TCGv_i32 tcg_res = tcg_temp_new_i32();
4278
    TCGv_ptr fpst = get_fpstatus_ptr();
4279

    
4280
    tcg_op1 = read_fp_sreg(s, rn);
4281
    tcg_op2 = read_fp_sreg(s, rm);
4282
    tcg_op3 = read_fp_sreg(s, ra);
4283

    
4284
    /* These are fused multiply-add, and must be done as one
4285
     * floating point operation with no rounding between the
4286
     * multiplication and addition steps.
4287
     * NB that doing the negations here as separate steps is
4288
     * correct : an input NaN should come out with its sign bit
4289
     * flipped if it is a negated-input.
4290
     */
4291
    if (o1 == true) {
4292
        gen_helper_vfp_negs(tcg_op3, tcg_op3);
4293
    }
4294

    
4295
    if (o0 != o1) {
4296
        gen_helper_vfp_negs(tcg_op1, tcg_op1);
4297
    }
4298

    
4299
    gen_helper_vfp_muladds(tcg_res, tcg_op1, tcg_op2, tcg_op3, fpst);
4300

    
4301
    write_fp_sreg(s, rd, tcg_res);
4302

    
4303
    tcg_temp_free_ptr(fpst);
4304
    tcg_temp_free_i32(tcg_op1);
4305
    tcg_temp_free_i32(tcg_op2);
4306
    tcg_temp_free_i32(tcg_op3);
4307
    tcg_temp_free_i32(tcg_res);
4308
}
4309

    
4310
/* C3.6.27 Floating-point data-processing (3 source) - double precision */
4311
static void handle_fp_3src_double(DisasContext *s, bool o0, bool o1,
4312
                                  int rd, int rn, int rm, int ra)
4313
{
4314
    TCGv_i64 tcg_op1, tcg_op2, tcg_op3;
4315
    TCGv_i64 tcg_res = tcg_temp_new_i64();
4316
    TCGv_ptr fpst = get_fpstatus_ptr();
4317

    
4318
    tcg_op1 = read_fp_dreg(s, rn);
4319
    tcg_op2 = read_fp_dreg(s, rm);
4320
    tcg_op3 = read_fp_dreg(s, ra);
4321

    
4322
    /* These are fused multiply-add, and must be done as one
4323
     * floating point operation with no rounding between the
4324
     * multiplication and addition steps.
4325
     * NB that doing the negations here as separate steps is
4326
     * correct : an input NaN should come out with its sign bit
4327
     * flipped if it is a negated-input.
4328
     */
4329
    if (o1 == true) {
4330
        gen_helper_vfp_negd(tcg_op3, tcg_op3);
4331
    }
4332

    
4333
    if (o0 != o1) {
4334
        gen_helper_vfp_negd(tcg_op1, tcg_op1);
4335
    }
4336

    
4337
    gen_helper_vfp_muladdd(tcg_res, tcg_op1, tcg_op2, tcg_op3, fpst);
4338

    
4339
    write_fp_dreg(s, rd, tcg_res);
4340

    
4341
    tcg_temp_free_ptr(fpst);
4342
    tcg_temp_free_i64(tcg_op1);
4343
    tcg_temp_free_i64(tcg_op2);
4344
    tcg_temp_free_i64(tcg_op3);
4345
    tcg_temp_free_i64(tcg_res);
4346
}
4347

    
4348
/* C3.6.27 Floating point data-processing (3 source)
4349
 *   31  30  29 28       24 23  22  21  20  16  15  14  10 9    5 4    0
4350
 * +---+---+---+-----------+------+----+------+----+------+------+------+
4351
 * | M | 0 | S | 1 1 1 1 1 | type | o1 |  Rm  | o0 |  Ra  |  Rn  |  Rd  |
4352
 * +---+---+---+-----------+------+----+------+----+------+------+------+
4353
 */
4354
static void disas_fp_3src(DisasContext *s, uint32_t insn)
4355
{
4356
    int type = extract32(insn, 22, 2);
4357
    int rd = extract32(insn, 0, 5);
4358
    int rn = extract32(insn, 5, 5);
4359
    int ra = extract32(insn, 10, 5);
4360
    int rm = extract32(insn, 16, 5);
4361
    bool o0 = extract32(insn, 15, 1);
4362
    bool o1 = extract32(insn, 21, 1);
4363

    
4364
    switch (type) {
4365
    case 0:
4366
        handle_fp_3src_single(s, o0, o1, rd, rn, rm, ra);
4367
        break;
4368
    case 1:
4369
        handle_fp_3src_double(s, o0, o1, rd, rn, rm, ra);
4370
        break;
4371
    default:
4372
        unallocated_encoding(s);
4373
    }
4374
}
4375

    
4376
/* C3.6.28 Floating point immediate
4377
 *   31  30  29 28       24 23  22  21 20        13 12   10 9    5 4    0
4378
 * +---+---+---+-----------+------+---+------------+-------+------+------+
4379
 * | M | 0 | S | 1 1 1 1 0 | type | 1 |    imm8    | 1 0 0 | imm5 |  Rd  |
4380
 * +---+---+---+-----------+------+---+------------+-------+------+------+
4381
 */
4382
static void disas_fp_imm(DisasContext *s, uint32_t insn)
4383
{
4384
    int rd = extract32(insn, 0, 5);
4385
    int imm8 = extract32(insn, 13, 8);
4386
    int is_double = extract32(insn, 22, 2);
4387
    uint64_t imm;
4388
    TCGv_i64 tcg_res;
4389

    
4390
    if (is_double > 1) {
4391
        unallocated_encoding(s);
4392
        return;
4393
    }
4394

    
4395
    /* The imm8 encodes the sign bit, enough bits to represent
4396
     * an exponent in the range 01....1xx to 10....0xx,
4397
     * and the most significant 4 bits of the mantissa; see
4398
     * VFPExpandImm() in the v8 ARM ARM.
4399
     */
4400
    if (is_double) {
4401
        imm = (extract32(imm8, 7, 1) ? 0x8000 : 0) |
4402
            (extract32(imm8, 6, 1) ? 0x3fc0 : 0x4000) |
4403
            extract32(imm8, 0, 6);
4404
        imm <<= 48;
4405
    } else {
4406
        imm = (extract32(imm8, 7, 1) ? 0x8000 : 0) |
4407
            (extract32(imm8, 6, 1) ? 0x3e00 : 0x4000) |
4408
            (extract32(imm8, 0, 6) << 3);
4409
        imm <<= 16;
4410
    }
4411

    
4412
    tcg_res = tcg_const_i64(imm);
4413
    write_fp_dreg(s, rd, tcg_res);
4414
    tcg_temp_free_i64(tcg_res);
4415
}
4416

    
4417
/* Handle floating point <=> fixed point conversions. Note that we can
4418
 * also deal with fp <=> integer conversions as a special case (scale == 64)
4419
 * OPTME: consider handling that special case specially or at least skipping
4420
 * the call to scalbn in the helpers for zero shifts.
4421
 */
4422
static void handle_fpfpcvt(DisasContext *s, int rd, int rn, int opcode,
4423
                           bool itof, int rmode, int scale, int sf, int type)
4424
{
4425
    bool is_signed = !(opcode & 1);
4426
    bool is_double = type;
4427
    TCGv_ptr tcg_fpstatus;
4428
    TCGv_i32 tcg_shift;
4429

    
4430
    tcg_fpstatus = get_fpstatus_ptr();
4431

    
4432
    tcg_shift = tcg_const_i32(64 - scale);
4433

    
4434
    if (itof) {
4435
        TCGv_i64 tcg_int = cpu_reg(s, rn);
4436
        if (!sf) {
4437
            TCGv_i64 tcg_extend = new_tmp_a64(s);
4438

    
4439
            if (is_signed) {
4440
                tcg_gen_ext32s_i64(tcg_extend, tcg_int);
4441
            } else {
4442
                tcg_gen_ext32u_i64(tcg_extend, tcg_int);
4443
            }
4444

    
4445
            tcg_int = tcg_extend;
4446
        }
4447

    
4448
        if (is_double) {
4449
            TCGv_i64 tcg_double = tcg_temp_new_i64();
4450
            if (is_signed) {
4451
                gen_helper_vfp_sqtod(tcg_double, tcg_int,
4452
                                     tcg_shift, tcg_fpstatus);
4453
            } else {
4454
                gen_helper_vfp_uqtod(tcg_double, tcg_int,
4455
                                     tcg_shift, tcg_fpstatus);
4456
            }
4457
            write_fp_dreg(s, rd, tcg_double);
4458
            tcg_temp_free_i64(tcg_double);
4459
        } else {
4460
            TCGv_i32 tcg_single = tcg_temp_new_i32();
4461
            if (is_signed) {
4462
                gen_helper_vfp_sqtos(tcg_single, tcg_int,
4463
                                     tcg_shift, tcg_fpstatus);
4464
            } else {
4465
                gen_helper_vfp_uqtos(tcg_single, tcg_int,
4466
                                     tcg_shift, tcg_fpstatus);
4467
            }
4468
            write_fp_sreg(s, rd, tcg_single);
4469
            tcg_temp_free_i32(tcg_single);
4470
        }
4471
    } else {
4472
        TCGv_i64 tcg_int = cpu_reg(s, rd);
4473
        TCGv_i32 tcg_rmode;
4474

    
4475
        if (extract32(opcode, 2, 1)) {
4476
            /* There are too many rounding modes to all fit into rmode,
4477
             * so FCVTA[US] is a special case.
4478
             */
4479
            rmode = FPROUNDING_TIEAWAY;
4480
        }
4481

    
4482
        tcg_rmode = tcg_const_i32(arm_rmode_to_sf(rmode));
4483

    
4484
        gen_helper_set_rmode(tcg_rmode, tcg_rmode, cpu_env);
4485

    
4486
        if (is_double) {
4487
            TCGv_i64 tcg_double = read_fp_dreg(s, rn);
4488
            if (is_signed) {
4489
                if (!sf) {
4490
                    gen_helper_vfp_tosld(tcg_int, tcg_double,
4491
                                         tcg_shift, tcg_fpstatus);
4492
                } else {
4493
                    gen_helper_vfp_tosqd(tcg_int, tcg_double,
4494
                                         tcg_shift, tcg_fpstatus);
4495
                }
4496
            } else {
4497
                if (!sf) {
4498
                    gen_helper_vfp_tould(tcg_int, tcg_double,
4499
                                         tcg_shift, tcg_fpstatus);
4500
                } else {
4501
                    gen_helper_vfp_touqd(tcg_int, tcg_double,
4502
                                         tcg_shift, tcg_fpstatus);
4503
                }
4504
            }
4505
            tcg_temp_free_i64(tcg_double);
4506
        } else {
4507
            TCGv_i32 tcg_single = read_fp_sreg(s, rn);
4508
            if (sf) {
4509
                if (is_signed) {
4510
                    gen_helper_vfp_tosqs(tcg_int, tcg_single,
4511
                                         tcg_shift, tcg_fpstatus);
4512
                } else {
4513
                    gen_helper_vfp_touqs(tcg_int, tcg_single,
4514
                                         tcg_shift, tcg_fpstatus);
4515
                }
4516
            } else {
4517
                TCGv_i32 tcg_dest = tcg_temp_new_i32();
4518
                if (is_signed) {
4519
                    gen_helper_vfp_tosls(tcg_dest, tcg_single,
4520
                                         tcg_shift, tcg_fpstatus);
4521
                } else {
4522
                    gen_helper_vfp_touls(tcg_dest, tcg_single,
4523
                                         tcg_shift, tcg_fpstatus);
4524
                }
4525
                tcg_gen_extu_i32_i64(tcg_int, tcg_dest);
4526
                tcg_temp_free_i32(tcg_dest);
4527
            }
4528
            tcg_temp_free_i32(tcg_single);
4529
        }
4530

    
4531
        gen_helper_set_rmode(tcg_rmode, tcg_rmode, cpu_env);
4532
        tcg_temp_free_i32(tcg_rmode);
4533

    
4534
        if (!sf) {
4535
            tcg_gen_ext32u_i64(tcg_int, tcg_int);
4536
        }
4537
    }
4538

    
4539
    tcg_temp_free_ptr(tcg_fpstatus);
4540
    tcg_temp_free_i32(tcg_shift);
4541
}
4542

    
4543
/* C3.6.29 Floating point <-> fixed point conversions
4544
 *   31   30  29 28       24 23  22  21 20   19 18    16 15   10 9    5 4    0
4545
 * +----+---+---+-----------+------+---+-------+--------+-------+------+------+
4546
 * | sf | 0 | S | 1 1 1 1 0 | type | 0 | rmode | opcode | scale |  Rn  |  Rd  |
4547
 * +----+---+---+-----------+------+---+-------+--------+-------+------+------+
4548
 */
4549
static void disas_fp_fixed_conv(DisasContext *s, uint32_t insn)
4550
{
4551
    int rd = extract32(insn, 0, 5);
4552
    int rn = extract32(insn, 5, 5);
4553
    int scale = extract32(insn, 10, 6);
4554
    int opcode = extract32(insn, 16, 3);
4555
    int rmode = extract32(insn, 19, 2);
4556
    int type = extract32(insn, 22, 2);
4557
    bool sbit = extract32(insn, 29, 1);
4558
    bool sf = extract32(insn, 31, 1);
4559
    bool itof;
4560

    
4561
    if (sbit || (type > 1)
4562
        || (!sf && scale < 32)) {
4563
        unallocated_encoding(s);
4564
        return;
4565
    }
4566

    
4567
    switch ((rmode << 3) | opcode) {
4568
    case 0x2: /* SCVTF */
4569
    case 0x3: /* UCVTF */
4570
        itof = true;
4571
        break;
4572
    case 0x18: /* FCVTZS */
4573
    case 0x19: /* FCVTZU */
4574
        itof = false;
4575
        break;
4576
    default:
4577
        unallocated_encoding(s);
4578
        return;
4579
    }
4580

    
4581
    handle_fpfpcvt(s, rd, rn, opcode, itof, FPROUNDING_ZERO, scale, sf, type);
4582
}
4583

    
4584
static void handle_fmov(DisasContext *s, int rd, int rn, int type, bool itof)
4585
{
4586
    /* FMOV: gpr to or from float, double, or top half of quad fp reg,
4587
     * without conversion.
4588
     */
4589

    
4590
    if (itof) {
4591
        TCGv_i64 tcg_rn = cpu_reg(s, rn);
4592

    
4593
        switch (type) {
4594
        case 0:
4595
        {
4596
            /* 32 bit */
4597
            TCGv_i64 tmp = tcg_temp_new_i64();
4598
            tcg_gen_ext32u_i64(tmp, tcg_rn);
4599
            tcg_gen_st_i64(tmp, cpu_env, fp_reg_offset(rd, MO_64));
4600
            tcg_gen_movi_i64(tmp, 0);
4601
            tcg_gen_st_i64(tmp, cpu_env, fp_reg_hi_offset(rd));
4602
            tcg_temp_free_i64(tmp);
4603
            break;
4604
        }
4605
        case 1:
4606
        {
4607
            /* 64 bit */
4608
            TCGv_i64 tmp = tcg_const_i64(0);
4609
            tcg_gen_st_i64(tcg_rn, cpu_env, fp_reg_offset(rd, MO_64));
4610
            tcg_gen_st_i64(tmp, cpu_env, fp_reg_hi_offset(rd));
4611
            tcg_temp_free_i64(tmp);
4612
            break;
4613
        }
4614
        case 2:
4615
            /* 64 bit to top half. */
4616
            tcg_gen_st_i64(tcg_rn, cpu_env, fp_reg_hi_offset(rd));
4617
            break;
4618
        }
4619
    } else {
4620
        TCGv_i64 tcg_rd = cpu_reg(s, rd);
4621

    
4622
        switch (type) {
4623
        case 0:
4624
            /* 32 bit */
4625
            tcg_gen_ld32u_i64(tcg_rd, cpu_env, fp_reg_offset(rn, MO_32));
4626
            break;
4627
        case 1:
4628
            /* 64 bit */
4629
            tcg_gen_ld_i64(tcg_rd, cpu_env, fp_reg_offset(rn, MO_64));
4630
            break;
4631
        case 2:
4632
            /* 64 bits from top half */
4633
            tcg_gen_ld_i64(tcg_rd, cpu_env, fp_reg_hi_offset(rn));
4634
            break;
4635
        }
4636
    }
4637
}
4638

    
4639
/* C3.6.30 Floating point <-> integer conversions
4640
 *   31   30  29 28       24 23  22  21 20   19 18 16 15         10 9  5 4  0
4641
 * +----+---+---+-----------+------+---+-------+-----+-------------+----+----+
4642
 * | sf | 0 | S | 1 1 1 1 0 | type | 1 | rmode | opc | 0 0 0 0 0 0 | Rn | Rd |
4643
 * +----+---+---+-----------+------+---+-------+-----+-------------+----+----+
4644
 */
4645
static void disas_fp_int_conv(DisasContext *s, uint32_t insn)
4646
{
4647
    int rd = extract32(insn, 0, 5);
4648
    int rn = extract32(insn, 5, 5);
4649
    int opcode = extract32(insn, 16, 3);
4650
    int rmode = extract32(insn, 19, 2);
4651
    int type = extract32(insn, 22, 2);
4652
    bool sbit = extract32(insn, 29, 1);
4653
    bool sf = extract32(insn, 31, 1);
4654

    
4655
    if (sbit) {
4656
        unallocated_encoding(s);
4657
        return;
4658
    }
4659

    
4660
    if (opcode > 5) {
4661
        /* FMOV */
4662
        bool itof = opcode & 1;
4663

    
4664
        if (rmode >= 2) {
4665
            unallocated_encoding(s);
4666
            return;
4667
        }
4668

    
4669
        switch (sf << 3 | type << 1 | rmode) {
4670
        case 0x0: /* 32 bit */
4671
        case 0xa: /* 64 bit */
4672
        case 0xd: /* 64 bit to top half of quad */
4673
            break;
4674
        default:
4675
            /* all other sf/type/rmode combinations are invalid */
4676
            unallocated_encoding(s);
4677
            break;
4678
        }
4679

    
4680
        handle_fmov(s, rd, rn, type, itof);
4681
    } else {
4682
        /* actual FP conversions */
4683
        bool itof = extract32(opcode, 1, 1);
4684

    
4685
        if (type > 1 || (rmode != 0 && opcode > 1)) {
4686
            unallocated_encoding(s);
4687
            return;
4688
        }
4689

    
4690
        handle_fpfpcvt(s, rd, rn, opcode, itof, rmode, 64, sf, type);
4691
    }
4692
}
4693

    
4694
/* FP-specific subcases of table C3-6 (SIMD and FP data processing)
4695
 *   31  30  29 28     25 24                          0
4696
 * +---+---+---+---------+-----------------------------+
4697
 * |   | 0 |   | 1 1 1 1 |                             |
4698
 * +---+---+---+---------+-----------------------------+
4699
 */
4700
static void disas_data_proc_fp(DisasContext *s, uint32_t insn)
4701
{
4702
    if (extract32(insn, 24, 1)) {
4703
        /* Floating point data-processing (3 source) */
4704
        disas_fp_3src(s, insn);
4705
    } else if (extract32(insn, 21, 1) == 0) {
4706
        /* Floating point to fixed point conversions */
4707
        disas_fp_fixed_conv(s, insn);
4708
    } else {
4709
        switch (extract32(insn, 10, 2)) {
4710
        case 1:
4711
            /* Floating point conditional compare */
4712
            disas_fp_ccomp(s, insn);
4713
            break;
4714
        case 2:
4715
            /* Floating point data-processing (2 source) */
4716
            disas_fp_2src(s, insn);
4717
            break;
4718
        case 3:
4719
            /* Floating point conditional select */
4720
            disas_fp_csel(s, insn);
4721
            break;
4722
        case 0:
4723
            switch (ctz32(extract32(insn, 12, 4))) {
4724
            case 0: /* [15:12] == xxx1 */
4725
                /* Floating point immediate */
4726
                disas_fp_imm(s, insn);
4727
                break;
4728
            case 1: /* [15:12] == xx10 */
4729
                /* Floating point compare */
4730
                disas_fp_compare(s, insn);
4731
                break;
4732
            case 2: /* [15:12] == x100 */
4733
                /* Floating point data-processing (1 source) */
4734
                disas_fp_1src(s, insn);
4735
                break;
4736
            case 3: /* [15:12] == 1000 */
4737
                unallocated_encoding(s);
4738
                break;
4739
            default: /* [15:12] == 0000 */
4740
                /* Floating point <-> integer conversions */
4741
                disas_fp_int_conv(s, insn);
4742
                break;
4743
            }
4744
            break;
4745
        }
4746
    }
4747
}
4748

    
4749
static void do_ext64(DisasContext *s, TCGv_i64 tcg_left, TCGv_i64 tcg_right,
4750
                     int pos)
4751
{
4752
    /* Extract 64 bits from the middle of two concatenated 64 bit
4753
     * vector register slices left:right. The extracted bits start
4754
     * at 'pos' bits into the right (least significant) side.
4755
     * We return the result in tcg_right, and guarantee not to
4756
     * trash tcg_left.
4757
     */
4758
    TCGv_i64 tcg_tmp = tcg_temp_new_i64();
4759
    assert(pos > 0 && pos < 64);
4760

    
4761
    tcg_gen_shri_i64(tcg_right, tcg_right, pos);
4762
    tcg_gen_shli_i64(tcg_tmp, tcg_left, 64 - pos);
4763
    tcg_gen_or_i64(tcg_right, tcg_right, tcg_tmp);
4764

    
4765
    tcg_temp_free_i64(tcg_tmp);
4766
}
4767

    
4768
/* C3.6.1 EXT
4769
 *   31  30 29         24 23 22  21 20  16 15  14  11 10  9    5 4    0
4770
 * +---+---+-------------+-----+---+------+---+------+---+------+------+
4771
 * | 0 | Q | 1 0 1 1 1 0 | op2 | 0 |  Rm  | 0 | imm4 | 0 |  Rn  |  Rd  |
4772
 * +---+---+-------------+-----+---+------+---+------+---+------+------+
4773
 */
4774
static void disas_simd_ext(DisasContext *s, uint32_t insn)
4775
{
4776
    int is_q = extract32(insn, 30, 1);
4777
    int op2 = extract32(insn, 22, 2);
4778
    int imm4 = extract32(insn, 11, 4);
4779
    int rm = extract32(insn, 16, 5);
4780
    int rn = extract32(insn, 5, 5);
4781
    int rd = extract32(insn, 0, 5);
4782
    int pos = imm4 << 3;
4783
    TCGv_i64 tcg_resl, tcg_resh;
4784

    
4785
    if (op2 != 0 || (!is_q && extract32(imm4, 3, 1))) {
4786
        unallocated_encoding(s);
4787
        return;
4788
    }
4789

    
4790
    tcg_resh = tcg_temp_new_i64();
4791
    tcg_resl = tcg_temp_new_i64();
4792

    
4793
    /* Vd gets bits starting at pos bits into Vm:Vn. This is
4794
     * either extracting 128 bits from a 128:128 concatenation, or
4795
     * extracting 64 bits from a 64:64 concatenation.
4796
     */
4797
    if (!is_q) {
4798
        read_vec_element(s, tcg_resl, rn, 0, MO_64);
4799
        if (pos != 0) {
4800
            read_vec_element(s, tcg_resh, rm, 0, MO_64);
4801
            do_ext64(s, tcg_resh, tcg_resl, pos);
4802
        }
4803
        tcg_gen_movi_i64(tcg_resh, 0);
4804
    } else {
4805
        TCGv_i64 tcg_hh;
4806
        typedef struct {
4807
            int reg;
4808
            int elt;
4809
        } EltPosns;
4810
        EltPosns eltposns[] = { {rn, 0}, {rn, 1}, {rm, 0}, {rm, 1} };
4811
        EltPosns *elt = eltposns;
4812

    
4813
        if (pos >= 64) {
4814
            elt++;
4815
            pos -= 64;
4816
        }
4817

    
4818
        read_vec_element(s, tcg_resl, elt->reg, elt->elt, MO_64);
4819
        elt++;
4820
        read_vec_element(s, tcg_resh, elt->reg, elt->elt, MO_64);
4821
        elt++;
4822
        if (pos != 0) {
4823
            do_ext64(s, tcg_resh, tcg_resl, pos);
4824
            tcg_hh = tcg_temp_new_i64();
4825
            read_vec_element(s, tcg_hh, elt->reg, elt->elt, MO_64);
4826
            do_ext64(s, tcg_hh, tcg_resh, pos);
4827
            tcg_temp_free_i64(tcg_hh);
4828
        }
4829
    }
4830

    
4831
    write_vec_element(s, tcg_resl, rd, 0, MO_64);
4832
    tcg_temp_free_i64(tcg_resl);
4833
    write_vec_element(s, tcg_resh, rd, 1, MO_64);
4834
    tcg_temp_free_i64(tcg_resh);
4835
}
4836

    
4837
/* C3.6.2 TBL/TBX
4838
 *   31  30 29         24 23 22  21 20  16 15  14 13  12  11 10 9    5 4    0
4839
 * +---+---+-------------+-----+---+------+---+-----+----+-----+------+------+
4840
 * | 0 | Q | 0 0 1 1 1 0 | op2 | 0 |  Rm  | 0 | len | op | 0 0 |  Rn  |  Rd  |
4841
 * +---+---+-------------+-----+---+------+---+-----+----+-----+------+------+
4842
 */
4843
static void disas_simd_tb(DisasContext *s, uint32_t insn)
4844
{
4845
    int op2 = extract32(insn, 22, 2);
4846
    int is_q = extract32(insn, 30, 1);
4847
    int rm = extract32(insn, 16, 5);
4848
    int rn = extract32(insn, 5, 5);
4849
    int rd = extract32(insn, 0, 5);
4850
    int is_tblx = extract32(insn, 12, 1);
4851
    int len = extract32(insn, 13, 2);
4852
    TCGv_i64 tcg_resl, tcg_resh, tcg_idx;
4853
    TCGv_i32 tcg_regno, tcg_numregs;
4854

    
4855
    if (op2 != 0) {
4856
        unallocated_encoding(s);
4857
        return;
4858
    }
4859

    
4860
    /* This does a table lookup: for every byte element in the input
4861
     * we index into a table formed from up to four vector registers,
4862
     * and then the output is the result of the lookups. Our helper
4863
     * function does the lookup operation for a single 64 bit part of
4864
     * the input.
4865
     */
4866
    tcg_resl = tcg_temp_new_i64();
4867
    tcg_resh = tcg_temp_new_i64();
4868

    
4869
    if (is_tblx) {
4870
        read_vec_element(s, tcg_resl, rd, 0, MO_64);
4871
    } else {
4872
        tcg_gen_movi_i64(tcg_resl, 0);
4873
    }
4874
    if (is_tblx && is_q) {
4875
        read_vec_element(s, tcg_resh, rd, 1, MO_64);
4876
    } else {
4877
        tcg_gen_movi_i64(tcg_resh, 0);
4878
    }
4879

    
4880
    tcg_idx = tcg_temp_new_i64();
4881
    tcg_regno = tcg_const_i32(rn);
4882
    tcg_numregs = tcg_const_i32(len + 1);
4883
    read_vec_element(s, tcg_idx, rm, 0, MO_64);
4884
    gen_helper_simd_tbl(tcg_resl, cpu_env, tcg_resl, tcg_idx,
4885
                        tcg_regno, tcg_numregs);
4886
    if (is_q) {
4887
        read_vec_element(s, tcg_idx, rm, 1, MO_64);
4888
        gen_helper_simd_tbl(tcg_resh, cpu_env, tcg_resh, tcg_idx,
4889
                            tcg_regno, tcg_numregs);
4890
    }
4891
    tcg_temp_free_i64(tcg_idx);
4892
    tcg_temp_free_i32(tcg_regno);
4893
    tcg_temp_free_i32(tcg_numregs);
4894

    
4895
    write_vec_element(s, tcg_resl, rd, 0, MO_64);
4896
    tcg_temp_free_i64(tcg_resl);
4897
    write_vec_element(s, tcg_resh, rd, 1, MO_64);
4898
    tcg_temp_free_i64(tcg_resh);
4899
}
4900

    
4901
/* C3.6.3 ZIP/UZP/TRN
4902
 *   31  30 29         24 23  22  21 20   16 15 14 12 11 10 9    5 4    0
4903
 * +---+---+-------------+------+---+------+---+------------------+------+
4904
 * | 0 | Q | 0 0 1 1 1 0 | size | 0 |  Rm  | 0 | opc | 1 0 |  Rn  |  Rd  |
4905
 * +---+---+-------------+------+---+------+---+------------------+------+
4906
 */
4907
static void disas_simd_zip_trn(DisasContext *s, uint32_t insn)
4908
{
4909
    int rd = extract32(insn, 0, 5);
4910
    int rn = extract32(insn, 5, 5);
4911
    int rm = extract32(insn, 16, 5);
4912
    int size = extract32(insn, 22, 2);
4913
    /* opc field bits [1:0] indicate ZIP/UZP/TRN;
4914
     * bit 2 indicates 1 vs 2 variant of the insn.
4915
     */
4916
    int opcode = extract32(insn, 12, 2);
4917
    bool part = extract32(insn, 14, 1);
4918
    bool is_q = extract32(insn, 30, 1);
4919
    int esize = 8 << size;
4920
    int i, ofs;
4921
    int datasize = is_q ? 128 : 64;
4922
    int elements = datasize / esize;
4923
    TCGv_i64 tcg_res, tcg_resl, tcg_resh;
4924

    
4925
    if (opcode == 0 || (size == 3 && !is_q)) {
4926
        unallocated_encoding(s);
4927
        return;
4928
    }
4929

    
4930
    tcg_resl = tcg_const_i64(0);
4931
    tcg_resh = tcg_const_i64(0);
4932
    tcg_res = tcg_temp_new_i64();
4933

    
4934
    for (i = 0; i < elements; i++) {
4935
        switch (opcode) {
4936
        case 1: /* UZP1/2 */
4937
        {
4938
            int midpoint = elements / 2;
4939
            if (i < midpoint) {
4940
                read_vec_element(s, tcg_res, rn, 2 * i + part, size);
4941
            } else {
4942
                read_vec_element(s, tcg_res, rm,
4943
                                 2 * (i - midpoint) + part, size);
4944
            }
4945
            break;
4946
        }
4947
        case 2: /* TRN1/2 */
4948
            if (i & 1) {
4949
                read_vec_element(s, tcg_res, rm, (i & ~1) + part, size);
4950
            } else {
4951
                read_vec_element(s, tcg_res, rn, (i & ~1) + part, size);
4952
            }
4953
            break;
4954
        case 3: /* ZIP1/2 */
4955
        {
4956
            int base = part * elements / 2;
4957
            if (i & 1) {
4958
                read_vec_element(s, tcg_res, rm, base + (i >> 1), size);
4959
            } else {
4960
                read_vec_element(s, tcg_res, rn, base + (i >> 1), size);
4961
            }
4962
            break;
4963
        }
4964
        default:
4965
            g_assert_not_reached();
4966
        }
4967

    
4968
        ofs = i * esize;
4969
        if (ofs < 64) {
4970
            tcg_gen_shli_i64(tcg_res, tcg_res, ofs);
4971
            tcg_gen_or_i64(tcg_resl, tcg_resl, tcg_res);
4972
        } else {
4973
            tcg_gen_shli_i64(tcg_res, tcg_res, ofs - 64);
4974
            tcg_gen_or_i64(tcg_resh, tcg_resh, tcg_res);
4975
        }
4976
    }
4977

    
4978
    tcg_temp_free_i64(tcg_res);
4979

    
4980
    write_vec_element(s, tcg_resl, rd, 0, MO_64);
4981
    tcg_temp_free_i64(tcg_resl);
4982
    write_vec_element(s, tcg_resh, rd, 1, MO_64);
4983
    tcg_temp_free_i64(tcg_resh);
4984
}
4985

    
4986
static void do_minmaxop(DisasContext *s, TCGv_i32 tcg_elt1, TCGv_i32 tcg_elt2,
4987
                        int opc, bool is_min, TCGv_ptr fpst)
4988
{
4989
    /* Helper function for disas_simd_across_lanes: do a single precision
4990
     * min/max operation on the specified two inputs,
4991
     * and return the result in tcg_elt1.
4992
     */
4993
    if (opc == 0xc) {
4994
        if (is_min) {
4995
            gen_helper_vfp_minnums(tcg_elt1, tcg_elt1, tcg_elt2, fpst);
4996
        } else {
4997
            gen_helper_vfp_maxnums(tcg_elt1, tcg_elt1, tcg_elt2, fpst);
4998
        }
4999
    } else {
5000
        assert(opc == 0xf);
5001
        if (is_min) {
5002
            gen_helper_vfp_mins(tcg_elt1, tcg_elt1, tcg_elt2, fpst);
5003
        } else {
5004
            gen_helper_vfp_maxs(tcg_elt1, tcg_elt1, tcg_elt2, fpst);
5005
        }
5006
    }
5007
}
5008

    
5009
/* C3.6.4 AdvSIMD across lanes
5010
 *   31  30  29 28       24 23  22 21       17 16    12 11 10 9    5 4    0
5011
 * +---+---+---+-----------+------+-----------+--------+-----+------+------+
5012
 * | 0 | Q | U | 0 1 1 1 0 | size | 1 1 0 0 0 | opcode | 1 0 |  Rn  |  Rd  |
5013
 * +---+---+---+-----------+------+-----------+--------+-----+------+------+
5014
 */
5015
static void disas_simd_across_lanes(DisasContext *s, uint32_t insn)
5016
{
5017
    int rd = extract32(insn, 0, 5);
5018
    int rn = extract32(insn, 5, 5);
5019
    int size = extract32(insn, 22, 2);
5020
    int opcode = extract32(insn, 12, 5);
5021
    bool is_q = extract32(insn, 30, 1);
5022
    bool is_u = extract32(insn, 29, 1);
5023
    bool is_fp = false;
5024
    bool is_min = false;
5025
    int esize;
5026
    int elements;
5027
    int i;
5028
    TCGv_i64 tcg_res, tcg_elt;
5029

    
5030
    switch (opcode) {
5031
    case 0x1b: /* ADDV */
5032
        if (is_u) {
5033
            unallocated_encoding(s);
5034
            return;
5035
        }
5036
        /* fall through */
5037
    case 0x3: /* SADDLV, UADDLV */
5038
    case 0xa: /* SMAXV, UMAXV */
5039
    case 0x1a: /* SMINV, UMINV */
5040
        if (size == 3 || (size == 2 && !is_q)) {
5041
            unallocated_encoding(s);
5042
            return;
5043
        }
5044
        break;
5045
    case 0xc: /* FMAXNMV, FMINNMV */
5046
    case 0xf: /* FMAXV, FMINV */
5047
        if (!is_u || !is_q || extract32(size, 0, 1)) {
5048
            unallocated_encoding(s);
5049
            return;
5050
        }
5051
        /* Bit 1 of size field encodes min vs max, and actual size is always
5052
         * 32 bits: adjust the size variable so following code can rely on it
5053
         */
5054
        is_min = extract32(size, 1, 1);
5055
        is_fp = true;
5056
        size = 2;
5057
        break;
5058
    default:
5059
        unallocated_encoding(s);
5060
        return;
5061
    }
5062

    
5063
    esize = 8 << size;
5064
    elements = (is_q ? 128 : 64) / esize;
5065

    
5066
    tcg_res = tcg_temp_new_i64();
5067
    tcg_elt = tcg_temp_new_i64();
5068

    
5069
    /* These instructions operate across all lanes of a vector
5070
     * to produce a single result. We can guarantee that a 64
5071
     * bit intermediate is sufficient:
5072
     *  + for [US]ADDLV the maximum element size is 32 bits, and
5073
     *    the result type is 64 bits
5074
     *  + for FMAX*V, FMIN*V, ADDV the intermediate type is the
5075
     *    same as the element size, which is 32 bits at most
5076
     * For the integer operations we can choose to work at 64
5077
     * or 32 bits and truncate at the end; for simplicity
5078
     * we use 64 bits always. The floating point
5079
     * ops do require 32 bit intermediates, though.
5080
     */
5081
    if (!is_fp) {
5082
        read_vec_element(s, tcg_res, rn, 0, size | (is_u ? 0 : MO_SIGN));
5083

    
5084
        for (i = 1; i < elements; i++) {
5085
            read_vec_element(s, tcg_elt, rn, i, size | (is_u ? 0 : MO_SIGN));
5086

    
5087
            switch (opcode) {
5088
            case 0x03: /* SADDLV / UADDLV */
5089
            case 0x1b: /* ADDV */
5090
                tcg_gen_add_i64(tcg_res, tcg_res, tcg_elt);
5091
                break;
5092
            case 0x0a: /* SMAXV / UMAXV */
5093
                tcg_gen_movcond_i64(is_u ? TCG_COND_GEU : TCG_COND_GE,
5094
                                    tcg_res,
5095
                                    tcg_res, tcg_elt, tcg_res, tcg_elt);
5096
                break;
5097
            case 0x1a: /* SMINV / UMINV */
5098
                tcg_gen_movcond_i64(is_u ? TCG_COND_LEU : TCG_COND_LE,
5099
                                    tcg_res,
5100
                                    tcg_res, tcg_elt, tcg_res, tcg_elt);
5101
                break;
5102
                break;
5103
            default:
5104
                g_assert_not_reached();
5105
            }
5106

    
5107
        }
5108
    } else {
5109
        /* Floating point ops which work on 32 bit (single) intermediates.
5110
         * Note that correct NaN propagation requires that we do these
5111
         * operations in exactly the order specified by the pseudocode.
5112
         */
5113
        TCGv_i32 tcg_elt1 = tcg_temp_new_i32();
5114
        TCGv_i32 tcg_elt2 = tcg_temp_new_i32();
5115
        TCGv_i32 tcg_elt3 = tcg_temp_new_i32();
5116
        TCGv_ptr fpst = get_fpstatus_ptr();
5117

    
5118
        assert(esize == 32);
5119
        assert(elements == 4);
5120

    
5121
        read_vec_element(s, tcg_elt, rn, 0, MO_32);
5122
        tcg_gen_trunc_i64_i32(tcg_elt1, tcg_elt);
5123
        read_vec_element(s, tcg_elt, rn, 1, MO_32);
5124
        tcg_gen_trunc_i64_i32(tcg_elt2, tcg_elt);
5125

    
5126
        do_minmaxop(s, tcg_elt1, tcg_elt2, opcode, is_min, fpst);
5127

    
5128
        read_vec_element(s, tcg_elt, rn, 2, MO_32);
5129
        tcg_gen_trunc_i64_i32(tcg_elt2, tcg_elt);
5130
        read_vec_element(s, tcg_elt, rn, 3, MO_32);
5131
        tcg_gen_trunc_i64_i32(tcg_elt3, tcg_elt);
5132

    
5133
        do_minmaxop(s, tcg_elt2, tcg_elt3, opcode, is_min, fpst);
5134

    
5135
        do_minmaxop(s, tcg_elt1, tcg_elt2, opcode, is_min, fpst);
5136

    
5137
        tcg_gen_extu_i32_i64(tcg_res, tcg_elt1);
5138
        tcg_temp_free_i32(tcg_elt1);
5139
        tcg_temp_free_i32(tcg_elt2);
5140
        tcg_temp_free_i32(tcg_elt3);
5141
        tcg_temp_free_ptr(fpst);
5142
    }
5143

    
5144
    tcg_temp_free_i64(tcg_elt);
5145

    
5146
    /* Now truncate the result to the width required for the final output */
5147
    if (opcode == 0x03) {
5148
        /* SADDLV, UADDLV: result is 2*esize */
5149
        size++;
5150
    }
5151

    
5152
    switch (size) {
5153
    case 0:
5154
        tcg_gen_ext8u_i64(tcg_res, tcg_res);
5155
        break;
5156
    case 1:
5157
        tcg_gen_ext16u_i64(tcg_res, tcg_res);
5158
        break;
5159
    case 2:
5160
        tcg_gen_ext32u_i64(tcg_res, tcg_res);
5161
        break;
5162
    case 3:
5163
        break;
5164
    default:
5165
        g_assert_not_reached();
5166
    }
5167

    
5168
    write_fp_dreg(s, rd, tcg_res);
5169
    tcg_temp_free_i64(tcg_res);
5170
}
5171

    
5172
/* C6.3.31 DUP (Element, Vector)
5173
 *
5174
 *  31  30   29              21 20    16 15        10  9    5 4    0
5175
 * +---+---+-------------------+--------+-------------+------+------+
5176
 * | 0 | Q | 0 0 1 1 1 0 0 0 0 |  imm5  | 0 0 0 0 0 1 |  Rn  |  Rd  |
5177
 * +---+---+-------------------+--------+-------------+------+------+
5178
 *
5179
 * size: encoded in imm5 (see ARM ARM LowestSetBit())
5180
 */
5181
static void handle_simd_dupe(DisasContext *s, int is_q, int rd, int rn,
5182
                             int imm5)
5183
{
5184
    int size = ctz32(imm5);
5185
    int esize = 8 << size;
5186
    int elements = (is_q ? 128 : 64) / esize;
5187
    int index, i;
5188
    TCGv_i64 tmp;
5189

    
5190
    if (size > 3 || (size == 3 && !is_q)) {
5191
        unallocated_encoding(s);
5192
        return;
5193
    }
5194

    
5195
    index = imm5 >> (size + 1);
5196

    
5197
    tmp = tcg_temp_new_i64();
5198
    read_vec_element(s, tmp, rn, index, size);
5199

    
5200
    for (i = 0; i < elements; i++) {
5201
        write_vec_element(s, tmp, rd, i, size);
5202
    }
5203

    
5204
    if (!is_q) {
5205
        clear_vec_high(s, rd);
5206
    }
5207

    
5208
    tcg_temp_free_i64(tmp);
5209
}
5210

    
5211
/* C6.3.31 DUP (element, scalar)
5212
 *  31                   21 20    16 15        10  9    5 4    0
5213
 * +-----------------------+--------+-------------+------+------+
5214
 * | 0 1 0 1 1 1 1 0 0 0 0 |  imm5  | 0 0 0 0 0 1 |  Rn  |  Rd  |
5215
 * +-----------------------+--------+-------------+------+------+
5216
 */
5217
static void handle_simd_dupes(DisasContext *s, int rd, int rn,
5218
                              int imm5)
5219
{
5220
    int size = ctz32(imm5);
5221
    int index;
5222
    TCGv_i64 tmp;
5223

    
5224
    if (size > 3) {
5225
        unallocated_encoding(s);
5226
        return;
5227
    }
5228

    
5229
    index = imm5 >> (size + 1);
5230

    
5231
    /* This instruction just extracts the specified element and
5232
     * zero-extends it into the bottom of the destination register.
5233
     */
5234
    tmp = tcg_temp_new_i64();
5235
    read_vec_element(s, tmp, rn, index, size);
5236
    write_fp_dreg(s, rd, tmp);
5237
    tcg_temp_free_i64(tmp);
5238
}
5239

    
5240
/* C6.3.32 DUP (General)
5241
 *
5242
 *  31  30   29              21 20    16 15        10  9    5 4    0
5243
 * +---+---+-------------------+--------+-------------+------+------+
5244
 * | 0 | Q | 0 0 1 1 1 0 0 0 0 |  imm5  | 0 0 0 0 1 1 |  Rn  |  Rd  |
5245
 * +---+---+-------------------+--------+-------------+------+------+
5246
 *
5247
 * size: encoded in imm5 (see ARM ARM LowestSetBit())
5248
 */
5249
static void handle_simd_dupg(DisasContext *s, int is_q, int rd, int rn,
5250
                             int imm5)
5251
{
5252
    int size = ctz32(imm5);
5253
    int esize = 8 << size;
5254
    int elements = (is_q ? 128 : 64)/esize;
5255
    int i = 0;
5256

    
5257
    if (size > 3 || ((size == 3) && !is_q)) {
5258
        unallocated_encoding(s);
5259
        return;
5260
    }
5261
    for (i = 0; i < elements; i++) {
5262
        write_vec_element(s, cpu_reg(s, rn), rd, i, size);
5263
    }
5264
    if (!is_q) {
5265
        clear_vec_high(s, rd);
5266
    }
5267
}
5268

    
5269
/* C6.3.150 INS (Element)
5270
 *
5271
 *  31                   21 20    16 15  14    11  10 9    5 4    0
5272
 * +-----------------------+--------+------------+---+------+------+
5273
 * | 0 1 1 0 1 1 1 0 0 0 0 |  imm5  | 0 |  imm4  | 1 |  Rn  |  Rd  |
5274
 * +-----------------------+--------+------------+---+------+------+
5275
 *
5276
 * size: encoded in imm5 (see ARM ARM LowestSetBit())
5277
 * index: encoded in imm5<4:size+1>
5278
 */
5279
static void handle_simd_inse(DisasContext *s, int rd, int rn,
5280
                             int imm4, int imm5)
5281
{
5282
    int size = ctz32(imm5);
5283
    int src_index, dst_index;
5284
    TCGv_i64 tmp;
5285

    
5286
    if (size > 3) {
5287
        unallocated_encoding(s);
5288
        return;
5289
    }
5290
    dst_index = extract32(imm5, 1+size, 5);
5291
    src_index = extract32(imm4, size, 4);
5292

    
5293
    tmp = tcg_temp_new_i64();
5294

    
5295
    read_vec_element(s, tmp, rn, src_index, size);
5296
    write_vec_element(s, tmp, rd, dst_index, size);
5297

    
5298
    tcg_temp_free_i64(tmp);
5299
}
5300

    
5301

    
5302
/* C6.3.151 INS (General)
5303
 *
5304
 *  31                   21 20    16 15        10  9    5 4    0
5305
 * +-----------------------+--------+-------------+------+------+
5306
 * | 0 1 0 0 1 1 1 0 0 0 0 |  imm5  | 0 0 0 1 1 1 |  Rn  |  Rd  |
5307
 * +-----------------------+--------+-------------+------+------+
5308
 *
5309
 * size: encoded in imm5 (see ARM ARM LowestSetBit())
5310
 * index: encoded in imm5<4:size+1>
5311
 */
5312
static void handle_simd_insg(DisasContext *s, int rd, int rn, int imm5)
5313
{
5314
    int size = ctz32(imm5);
5315
    int idx;
5316

    
5317
    if (size > 3) {
5318
        unallocated_encoding(s);
5319
        return;
5320
    }
5321

    
5322
    idx = extract32(imm5, 1 + size, 4 - size);
5323
    write_vec_element(s, cpu_reg(s, rn), rd, idx, size);
5324
}
5325

    
5326
/*
5327
 * C6.3.321 UMOV (General)
5328
 * C6.3.237 SMOV (General)
5329
 *
5330
 *  31  30   29              21 20    16 15    12   10 9    5 4    0
5331
 * +---+---+-------------------+--------+-------------+------+------+
5332
 * | 0 | Q | 0 0 1 1 1 0 0 0 0 |  imm5  | 0 0 1 U 1 1 |  Rn  |  Rd  |
5333
 * +---+---+-------------------+--------+-------------+------+------+
5334
 *
5335
 * U: unsigned when set
5336
 * size: encoded in imm5 (see ARM ARM LowestSetBit())
5337
 */
5338
static void handle_simd_umov_smov(DisasContext *s, int is_q, int is_signed,
5339
                                  int rn, int rd, int imm5)
5340
{
5341
    int size = ctz32(imm5);
5342
    int element;
5343
    TCGv_i64 tcg_rd;
5344

    
5345
    /* Check for UnallocatedEncodings */
5346
    if (is_signed) {
5347
        if (size > 2 || (size == 2 && !is_q)) {
5348
            unallocated_encoding(s);
5349
            return;
5350
        }
5351
    } else {
5352
        if (size > 3
5353
            || (size < 3 && is_q)
5354
            || (size == 3 && !is_q)) {
5355
            unallocated_encoding(s);
5356
            return;
5357
        }
5358
    }
5359
    element = extract32(imm5, 1+size, 4);
5360

    
5361
    tcg_rd = cpu_reg(s, rd);
5362
    read_vec_element(s, tcg_rd, rn, element, size | (is_signed ? MO_SIGN : 0));
5363
    if (is_signed && !is_q) {
5364
        tcg_gen_ext32u_i64(tcg_rd, tcg_rd);
5365
    }
5366
}
5367

    
5368
/* C3.6.5 AdvSIMD copy
5369
 *   31  30  29  28             21 20  16 15  14  11 10  9    5 4    0
5370
 * +---+---+----+-----------------+------+---+------+---+------+------+
5371
 * | 0 | Q | op | 0 1 1 1 0 0 0 0 | imm5 | 0 | imm4 | 1 |  Rn  |  Rd  |
5372
 * +---+---+----+-----------------+------+---+------+---+------+------+
5373
 */
5374
static void disas_simd_copy(DisasContext *s, uint32_t insn)
5375
{
5376
    int rd = extract32(insn, 0, 5);
5377
    int rn = extract32(insn, 5, 5);
5378
    int imm4 = extract32(insn, 11, 4);
5379
    int op = extract32(insn, 29, 1);
5380
    int is_q = extract32(insn, 30, 1);
5381
    int imm5 = extract32(insn, 16, 5);
5382

    
5383
    if (op) {
5384
        if (is_q) {
5385
            /* INS (element) */
5386
            handle_simd_inse(s, rd, rn, imm4, imm5);
5387
        } else {
5388
            unallocated_encoding(s);
5389
        }
5390
    } else {
5391
        switch (imm4) {
5392
        case 0:
5393
            /* DUP (element - vector) */
5394
            handle_simd_dupe(s, is_q, rd, rn, imm5);
5395
            break;
5396
        case 1:
5397
            /* DUP (general) */
5398
            handle_simd_dupg(s, is_q, rd, rn, imm5);
5399
            break;
5400
        case 3:
5401
            if (is_q) {
5402
                /* INS (general) */
5403
                handle_simd_insg(s, rd, rn, imm5);
5404
            } else {
5405
                unallocated_encoding(s);
5406
            }
5407
            break;
5408
        case 5:
5409
        case 7:
5410
            /* UMOV/SMOV (is_q indicates 32/64; imm4 indicates signedness) */
5411
            handle_simd_umov_smov(s, is_q, (imm4 == 5), rn, rd, imm5);
5412
            break;
5413
        default:
5414
            unallocated_encoding(s);
5415
            break;
5416
        }
5417
    }
5418
}
5419

    
5420
/* C3.6.6 AdvSIMD modified immediate
5421
 *  31  30   29  28                 19 18 16 15   12  11  10  9     5 4    0
5422
 * +---+---+----+---------------------+-----+-------+----+---+-------+------+
5423
 * | 0 | Q | op | 0 1 1 1 1 0 0 0 0 0 | abc | cmode | o2 | 1 | defgh |  Rd  |
5424
 * +---+---+----+---------------------+-----+-------+----+---+-------+------+
5425
 *
5426
 * There are a number of operations that can be carried out here:
5427
 *   MOVI - move (shifted) imm into register
5428
 *   MVNI - move inverted (shifted) imm into register
5429
 *   ORR  - bitwise OR of (shifted) imm with register
5430
 *   BIC  - bitwise clear of (shifted) imm with register
5431
 */
5432
static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
5433
{
5434
    int rd = extract32(insn, 0, 5);
5435
    int cmode = extract32(insn, 12, 4);
5436
    int cmode_3_1 = extract32(cmode, 1, 3);
5437
    int cmode_0 = extract32(cmode, 0, 1);
5438
    int o2 = extract32(insn, 11, 1);
5439
    uint64_t abcdefgh = extract32(insn, 5, 5) | (extract32(insn, 16, 3) << 5);
5440
    bool is_neg = extract32(insn, 29, 1);
5441
    bool is_q = extract32(insn, 30, 1);
5442
    uint64_t imm = 0;
5443
    TCGv_i64 tcg_rd, tcg_imm;
5444
    int i;
5445

    
5446
    if (o2 != 0 || ((cmode == 0xf) && is_neg && !is_q)) {
5447
        unallocated_encoding(s);
5448
        return;
5449
    }
5450

    
5451
    /* See AdvSIMDExpandImm() in ARM ARM */
5452
    switch (cmode_3_1) {
5453
    case 0: /* Replicate(Zeros(24):imm8, 2) */
5454
    case 1: /* Replicate(Zeros(16):imm8:Zeros(8), 2) */
5455
    case 2: /* Replicate(Zeros(8):imm8:Zeros(16), 2) */
5456
    case 3: /* Replicate(imm8:Zeros(24), 2) */
5457
    {
5458
        int shift = cmode_3_1 * 8;
5459
        imm = bitfield_replicate(abcdefgh << shift, 32);
5460
        break;
5461
    }
5462
    case 4: /* Replicate(Zeros(8):imm8, 4) */
5463
    case 5: /* Replicate(imm8:Zeros(8), 4) */
5464
    {
5465
        int shift = (cmode_3_1 & 0x1) * 8;
5466
        imm = bitfield_replicate(abcdefgh << shift, 16);
5467
        break;
5468
    }
5469
    case 6:
5470
        if (cmode_0) {
5471
            /* Replicate(Zeros(8):imm8:Ones(16), 2) */
5472
            imm = (abcdefgh << 16) | 0xffff;
5473
        } else {
5474
            /* Replicate(Zeros(16):imm8:Ones(8), 2) */
5475
            imm = (abcdefgh << 8) | 0xff;
5476
        }
5477
        imm = bitfield_replicate(imm, 32);
5478
        break;
5479
    case 7:
5480
        if (!cmode_0 && !is_neg) {
5481
            imm = bitfield_replicate(abcdefgh, 8);
5482
        } else if (!cmode_0 && is_neg) {
5483
            int i;
5484
            imm = 0;
5485
            for (i = 0; i < 8; i++) {
5486
                if ((abcdefgh) & (1 << i)) {
5487
                    imm |= 0xffULL << (i * 8);
5488
                }
5489
            }
5490
        } else if (cmode_0) {
5491
            if (is_neg) {
5492
                imm = (abcdefgh & 0x3f) << 48;
5493
                if (abcdefgh & 0x80) {
5494
                    imm |= 0x8000000000000000ULL;
5495
                }
5496
                if (abcdefgh & 0x40) {
5497
                    imm |= 0x3fc0000000000000ULL;
5498
                } else {
5499
                    imm |= 0x4000000000000000ULL;
5500
                }
5501
            } else {
5502
                imm = (abcdefgh & 0x3f) << 19;
5503
                if (abcdefgh & 0x80) {
5504
                    imm |= 0x80000000;
5505
                }
5506
                if (abcdefgh & 0x40) {
5507
                    imm |= 0x3e000000;
5508
                } else {
5509
                    imm |= 0x40000000;
5510
                }
5511
                imm |= (imm << 32);
5512
            }
5513
        }
5514
        break;
5515
    }
5516

    
5517
    if (cmode_3_1 != 7 && is_neg) {
5518
        imm = ~imm;
5519
    }
5520

    
5521
    tcg_imm = tcg_const_i64(imm);
5522
    tcg_rd = new_tmp_a64(s);
5523

    
5524
    for (i = 0; i < 2; i++) {
5525
        int foffs = i ? fp_reg_hi_offset(rd) : fp_reg_offset(rd, MO_64);
5526

    
5527
        if (i == 1 && !is_q) {
5528
            /* non-quad ops clear high half of vector */
5529
            tcg_gen_movi_i64(tcg_rd, 0);
5530
        } else if ((cmode & 0x9) == 0x1 || (cmode & 0xd) == 0x9) {
5531
            tcg_gen_ld_i64(tcg_rd, cpu_env, foffs);
5532
            if (is_neg) {
5533
                /* AND (BIC) */
5534
                tcg_gen_and_i64(tcg_rd, tcg_rd, tcg_imm);
5535
            } else {
5536
                /* ORR */
5537
                tcg_gen_or_i64(tcg_rd, tcg_rd, tcg_imm);
5538
            }
5539
        } else {
5540
            /* MOVI */
5541
            tcg_gen_mov_i64(tcg_rd, tcg_imm);
5542
        }
5543
        tcg_gen_st_i64(tcg_rd, cpu_env, foffs);
5544
    }
5545

    
5546
    tcg_temp_free_i64(tcg_imm);
5547
}
5548

    
5549
/* C3.6.7 AdvSIMD scalar copy
5550
 *  31 30  29  28             21 20  16 15  14  11 10  9    5 4    0
5551
 * +-----+----+-----------------+------+---+------+---+------+------+
5552
 * | 0 1 | op | 1 1 1 1 0 0 0 0 | imm5 | 0 | imm4 | 1 |  Rn  |  Rd  |
5553
 * +-----+----+-----------------+------+---+------+---+------+------+
5554
 */
5555
static void disas_simd_scalar_copy(DisasContext *s, uint32_t insn)
5556
{
5557
    int rd = extract32(insn, 0, 5);
5558
    int rn = extract32(insn, 5, 5);
5559
    int imm4 = extract32(insn, 11, 4);
5560
    int imm5 = extract32(insn, 16, 5);
5561
    int op = extract32(insn, 29, 1);
5562

    
5563
    if (op != 0 || imm4 != 0) {
5564
        unallocated_encoding(s);
5565
        return;
5566
    }
5567

    
5568
    /* DUP (element, scalar) */
5569
    handle_simd_dupes(s, rd, rn, imm5);
5570
}
5571

    
5572
/* C3.6.8 AdvSIMD scalar pairwise
5573
 *  31 30  29 28       24 23  22 21       17 16    12 11 10 9    5 4    0
5574
 * +-----+---+-----------+------+-----------+--------+-----+------+------+
5575
 * | 0 1 | U | 1 1 1 1 0 | size | 1 1 0 0 0 | opcode | 1 0 |  Rn  |  Rd  |
5576
 * +-----+---+-----------+------+-----------+--------+-----+------+------+
5577
 */
5578
static void disas_simd_scalar_pairwise(DisasContext *s, uint32_t insn)
5579
{
5580
    int u = extract32(insn, 29, 1);
5581
    int size = extract32(insn, 22, 2);
5582
    int opcode = extract32(insn, 12, 5);
5583
    int rn = extract32(insn, 5, 5);
5584
    int rd = extract32(insn, 0, 5);
5585
    TCGv_ptr fpst;
5586

    
5587
    /* For some ops (the FP ones), size[1] is part of the encoding.
5588
     * For ADDP strictly it is not but size[1] is always 1 for valid
5589
     * encodings.
5590
     */
5591
    opcode |= (extract32(size, 1, 1) << 5);
5592

    
5593
    switch (opcode) {
5594
    case 0x3b: /* ADDP */
5595
        if (u || size != 3) {
5596
            unallocated_encoding(s);
5597
            return;
5598
        }
5599
        TCGV_UNUSED_PTR(fpst);
5600
        break;
5601
    case 0xc: /* FMAXNMP */
5602
    case 0xd: /* FADDP */
5603
    case 0xf: /* FMAXP */
5604
    case 0x2c: /* FMINNMP */
5605
    case 0x2f: /* FMINP */
5606
        /* FP op, size[0] is 32 or 64 bit */
5607
        if (!u) {
5608
            unallocated_encoding(s);
5609
            return;
5610
        }
5611
        size = extract32(size, 0, 1) ? 3 : 2;
5612
        fpst = get_fpstatus_ptr();
5613
        break;
5614
    default:
5615
        unallocated_encoding(s);
5616
        return;
5617
    }
5618

    
5619
    if (size == 3) {
5620
        TCGv_i64 tcg_op1 = tcg_temp_new_i64();
5621
        TCGv_i64 tcg_op2 = tcg_temp_new_i64();
5622
        TCGv_i64 tcg_res = tcg_temp_new_i64();
5623

    
5624
        read_vec_element(s, tcg_op1, rn, 0, MO_64);
5625
        read_vec_element(s, tcg_op2, rn, 1, MO_64);
5626

    
5627
        switch (opcode) {
5628
        case 0x3b: /* ADDP */
5629
            tcg_gen_add_i64(tcg_res, tcg_op1, tcg_op2);
5630
            break;
5631
        case 0xc: /* FMAXNMP */
5632
            gen_helper_vfp_maxnumd(tcg_res, tcg_op1, tcg_op2, fpst);
5633
            break;
5634
        case 0xd: /* FADDP */
5635
            gen_helper_vfp_addd(tcg_res, tcg_op1, tcg_op2, fpst);
5636
            break;
5637
        case 0xf: /* FMAXP */
5638
            gen_helper_vfp_maxd(tcg_res, tcg_op1, tcg_op2, fpst);
5639
            break;
5640
        case 0x2c: /* FMINNMP */
5641
            gen_helper_vfp_minnumd(tcg_res, tcg_op1, tcg_op2, fpst);
5642
            break;
5643
        case 0x2f: /* FMINP */
5644
            gen_helper_vfp_mind(tcg_res, tcg_op1, tcg_op2, fpst);
5645
            break;
5646
        default:
5647
            g_assert_not_reached();
5648
        }
5649

    
5650
        write_fp_dreg(s, rd, tcg_res);
5651

    
5652
        tcg_temp_free_i64(tcg_op1);
5653
        tcg_temp_free_i64(tcg_op2);
5654
        tcg_temp_free_i64(tcg_res);
5655
    } else {
5656
        TCGv_i32 tcg_op1 = tcg_temp_new_i32();
5657
        TCGv_i32 tcg_op2 = tcg_temp_new_i32();
5658
        TCGv_i32 tcg_res = tcg_temp_new_i32();
5659

    
5660
        read_vec_element_i32(s, tcg_op1, rn, 0, MO_32);
5661
        read_vec_element_i32(s, tcg_op2, rn, 1, MO_32);
5662

    
5663
        switch (opcode) {
5664
        case 0xc: /* FMAXNMP */
5665
            gen_helper_vfp_maxnums(tcg_res, tcg_op1, tcg_op2, fpst);
5666
            break;
5667
        case 0xd: /* FADDP */
5668
            gen_helper_vfp_adds(tcg_res, tcg_op1, tcg_op2, fpst);
5669
            break;
5670
        case 0xf: /* FMAXP */
5671
            gen_helper_vfp_maxs(tcg_res, tcg_op1, tcg_op2, fpst);
5672
            break;
5673
        case 0x2c: /* FMINNMP */
5674
            gen_helper_vfp_minnums(tcg_res, tcg_op1, tcg_op2, fpst);
5675
            break;
5676
        case 0x2f: /* FMINP */
5677
            gen_helper_vfp_mins(tcg_res, tcg_op1, tcg_op2, fpst);
5678
            break;
5679
        default:
5680
            g_assert_not_reached();
5681
        }
5682

    
5683
        write_fp_sreg(s, rd, tcg_res);
5684

    
5685
        tcg_temp_free_i32(tcg_op1);
5686
        tcg_temp_free_i32(tcg_op2);
5687
        tcg_temp_free_i32(tcg_res);
5688
    }
5689

    
5690
    if (!TCGV_IS_UNUSED_PTR(fpst)) {
5691
        tcg_temp_free_ptr(fpst);
5692
    }
5693
}
5694

    
5695
/*
5696
 * Common SSHR[RA]/USHR[RA] - Shift right (optional rounding/accumulate)
5697
 *
5698
 * This code is handles the common shifting code and is used by both
5699
 * the vector and scalar code.
5700
 */
5701
static void handle_shri_with_rndacc(TCGv_i64 tcg_res, TCGv_i64 tcg_src,
5702
                                    TCGv_i64 tcg_rnd, bool accumulate,
5703
                                    bool is_u, int size, int shift)
5704
{
5705
    bool extended_result = false;
5706
    bool round = !TCGV_IS_UNUSED_I64(tcg_rnd);
5707
    int ext_lshift = 0;
5708
    TCGv_i64 tcg_src_hi;
5709

    
5710
    if (round && size == 3) {
5711
        extended_result = true;
5712
        ext_lshift = 64 - shift;
5713
        tcg_src_hi = tcg_temp_new_i64();
5714
    } else if (shift == 64) {
5715
        if (!accumulate && is_u) {
5716
            /* result is zero */
5717
            tcg_gen_movi_i64(tcg_res, 0);
5718
            return;
5719
        }
5720
    }
5721

    
5722
    /* Deal with the rounding step */
5723
    if (round) {
5724
        if (extended_result) {
5725
            TCGv_i64 tcg_zero = tcg_const_i64(0);
5726
            if (!is_u) {
5727
                /* take care of sign extending tcg_res */
5728
                tcg_gen_sari_i64(tcg_src_hi, tcg_src, 63);
5729
                tcg_gen_add2_i64(tcg_src, tcg_src_hi,
5730
                                 tcg_src, tcg_src_hi,
5731
                                 tcg_rnd, tcg_zero);
5732
            } else {
5733
                tcg_gen_add2_i64(tcg_src, tcg_src_hi,
5734
                                 tcg_src, tcg_zero,
5735
                                 tcg_rnd, tcg_zero);
5736
            }
5737
            tcg_temp_free_i64(tcg_zero);
5738
        } else {
5739
            tcg_gen_add_i64(tcg_src, tcg_src, tcg_rnd);
5740
        }
5741
    }
5742

    
5743
    /* Now do the shift right */
5744
    if (round && extended_result) {
5745
        /* extended case, >64 bit precision required */
5746
        if (ext_lshift == 0) {
5747
            /* special case, only high bits matter */
5748
            tcg_gen_mov_i64(tcg_src, tcg_src_hi);
5749
        } else {
5750
            tcg_gen_shri_i64(tcg_src, tcg_src, shift);
5751
            tcg_gen_shli_i64(tcg_src_hi, tcg_src_hi, ext_lshift);
5752
            tcg_gen_or_i64(tcg_src, tcg_src, tcg_src_hi);
5753
        }
5754
    } else {
5755
        if (is_u) {
5756
            if (shift == 64) {
5757
                /* essentially shifting in 64 zeros */
5758
                tcg_gen_movi_i64(tcg_src, 0);
5759
            } else {
5760
                tcg_gen_shri_i64(tcg_src, tcg_src, shift);
5761
            }
5762
        } else {
5763
            if (shift == 64) {
5764
                /* effectively extending the sign-bit */
5765
                tcg_gen_sari_i64(tcg_src, tcg_src, 63);
5766
            } else {
5767
                tcg_gen_sari_i64(tcg_src, tcg_src, shift);
5768
            }
5769
        }
5770
    }
5771

    
5772
    if (accumulate) {
5773
        tcg_gen_add_i64(tcg_res, tcg_res, tcg_src);
5774
    } else {
5775
        tcg_gen_mov_i64(tcg_res, tcg_src);
5776
    }
5777

    
5778
    if (extended_result) {
5779
        tcg_temp_free_i64(tcg_src_hi);
5780
    }
5781
}
5782

    
5783
/* Common SHL/SLI - Shift left with an optional insert */
5784
static void handle_shli_with_ins(TCGv_i64 tcg_res, TCGv_i64 tcg_src,
5785
                                 bool insert, int shift)
5786
{
5787
    if (insert) { /* SLI */
5788
        tcg_gen_deposit_i64(tcg_res, tcg_res, tcg_src, shift, 64 - shift);
5789
    } else { /* SHL */
5790
        tcg_gen_shli_i64(tcg_res, tcg_src, shift);
5791
    }
5792
}
5793

    
5794
/* SSHR[RA]/USHR[RA] - Scalar shift right (optional rounding/accumulate) */
5795
static void handle_scalar_simd_shri(DisasContext *s,
5796
                                    bool is_u, int immh, int immb,
5797
                                    int opcode, int rn, int rd)
5798
{
5799
    const int size = 3;
5800
    int immhb = immh << 3 | immb;
5801
    int shift = 2 * (8 << size) - immhb;
5802
    bool accumulate = false;
5803
    bool round = false;
5804
    TCGv_i64 tcg_rn;
5805
    TCGv_i64 tcg_rd;
5806
    TCGv_i64 tcg_round;
5807

    
5808
    if (!extract32(immh, 3, 1)) {
5809
        unallocated_encoding(s);
5810
        return;
5811
    }
5812

    
5813
    switch (opcode) {
5814
    case 0x02: /* SSRA / USRA (accumulate) */
5815
        accumulate = true;
5816
        break;
5817
    case 0x04: /* SRSHR / URSHR (rounding) */
5818
        round = true;
5819
        break;
5820
    case 0x06: /* SRSRA / URSRA (accum + rounding) */
5821
        accumulate = round = true;
5822
        break;
5823
    }
5824

    
5825
    if (round) {
5826
        uint64_t round_const = 1ULL << (shift - 1);
5827
        tcg_round = tcg_const_i64(round_const);
5828
    } else {
5829
        TCGV_UNUSED_I64(tcg_round);
5830
    }
5831

    
5832
    tcg_rn = read_fp_dreg(s, rn);
5833
    tcg_rd = accumulate ? read_fp_dreg(s, rd) : tcg_temp_new_i64();
5834

    
5835
    handle_shri_with_rndacc(tcg_rd, tcg_rn, tcg_round,
5836
                               accumulate, is_u, size, shift);
5837

    
5838
    write_fp_dreg(s, rd, tcg_rd);
5839

    
5840
    tcg_temp_free_i64(tcg_rn);
5841
    tcg_temp_free_i64(tcg_rd);
5842
    if (round) {
5843
        tcg_temp_free_i64(tcg_round);
5844
    }
5845
}
5846

    
5847
/* SHL/SLI - Scalar shift left */
5848
static void handle_scalar_simd_shli(DisasContext *s, bool insert,
5849
                                    int immh, int immb, int opcode,
5850
                                    int rn, int rd)
5851
{
5852
    int size = 32 - clz32(immh) - 1;
5853
    int immhb = immh << 3 | immb;
5854
    int shift = immhb - (8 << size);
5855
    TCGv_i64 tcg_rn = new_tmp_a64(s);
5856
    TCGv_i64 tcg_rd = new_tmp_a64(s);
5857

    
5858
    if (!extract32(immh, 3, 1)) {
5859
        unallocated_encoding(s);
5860
        return;
5861
    }
5862

    
5863
    tcg_rn = read_fp_dreg(s, rn);
5864
    tcg_rd = insert ? read_fp_dreg(s, rd) : tcg_temp_new_i64();
5865

    
5866
    handle_shli_with_ins(tcg_rd, tcg_rn, insert, shift);
5867

    
5868
    write_fp_dreg(s, rd, tcg_rd);
5869

    
5870
    tcg_temp_free_i64(tcg_rn);
5871
    tcg_temp_free_i64(tcg_rd);
5872
}
5873

    
5874
/* C3.6.9 AdvSIMD scalar shift by immediate
5875
 *  31 30  29 28         23 22  19 18  16 15    11  10 9    5 4    0
5876
 * +-----+---+-------------+------+------+--------+---+------+------+
5877
 * | 0 1 | U | 1 1 1 1 1 0 | immh | immb | opcode | 1 |  Rn  |  Rd  |
5878
 * +-----+---+-------------+------+------+--------+---+------+------+
5879
 *
5880
 * This is the scalar version so it works on a fixed sized registers
5881
 */
5882
static void disas_simd_scalar_shift_imm(DisasContext *s, uint32_t insn)
5883
{
5884
    int rd = extract32(insn, 0, 5);
5885
    int rn = extract32(insn, 5, 5);
5886
    int opcode = extract32(insn, 11, 5);
5887
    int immb = extract32(insn, 16, 3);
5888
    int immh = extract32(insn, 19, 4);
5889
    bool is_u = extract32(insn, 29, 1);
5890

    
5891
    switch (opcode) {
5892
    case 0x00: /* SSHR / USHR */
5893
    case 0x02: /* SSRA / USRA */
5894
    case 0x04: /* SRSHR / URSHR */
5895
    case 0x06: /* SRSRA / URSRA */
5896
        handle_scalar_simd_shri(s, is_u, immh, immb, opcode, rn, rd);
5897
        break;
5898
    case 0x0a: /* SHL / SLI */
5899
        handle_scalar_simd_shli(s, is_u, immh, immb, opcode, rn, rd);
5900
        break;
5901
    default:
5902
        unsupported_encoding(s, insn);
5903
        break;
5904
    }
5905
}
5906

    
5907
/* C3.6.10 AdvSIMD scalar three different
5908
 *  31 30  29 28       24 23  22  21 20  16 15    12 11 10 9    5 4    0
5909
 * +-----+---+-----------+------+---+------+--------+-----+------+------+
5910
 * | 0 1 | U | 1 1 1 1 0 | size | 1 |  Rm  | opcode | 0 0 |  Rn  |  Rd  |
5911
 * +-----+---+-----------+------+---+------+--------+-----+------+------+
5912
 */
5913
static void disas_simd_scalar_three_reg_diff(DisasContext *s, uint32_t insn)
5914
{
5915
    bool is_u = extract32(insn, 29, 1);
5916
    int size = extract32(insn, 22, 2);
5917
    int opcode = extract32(insn, 12, 4);
5918
    int rm = extract32(insn, 16, 5);
5919
    int rn = extract32(insn, 5, 5);
5920
    int rd = extract32(insn, 0, 5);
5921

    
5922
    if (is_u) {
5923
        unallocated_encoding(s);
5924
        return;
5925
    }
5926

    
5927
    switch (opcode) {
5928
    case 0x9: /* SQDMLAL, SQDMLAL2 */
5929
    case 0xb: /* SQDMLSL, SQDMLSL2 */
5930
    case 0xd: /* SQDMULL, SQDMULL2 */
5931
        if (size == 0 || size == 3) {
5932
            unallocated_encoding(s);
5933
            return;
5934
        }
5935
        break;
5936
    default:
5937
        unallocated_encoding(s);
5938
        return;
5939
    }
5940

    
5941
    if (size == 2) {
5942
        TCGv_i64 tcg_op1 = tcg_temp_new_i64();
5943
        TCGv_i64 tcg_op2 = tcg_temp_new_i64();
5944
        TCGv_i64 tcg_res = tcg_temp_new_i64();
5945

    
5946
        read_vec_element(s, tcg_op1, rn, 0, MO_32 | MO_SIGN);
5947
        read_vec_element(s, tcg_op2, rm, 0, MO_32 | MO_SIGN);
5948

    
5949
        tcg_gen_mul_i64(tcg_res, tcg_op1, tcg_op2);
5950
        gen_helper_neon_addl_saturate_s64(tcg_res, cpu_env, tcg_res, tcg_res);
5951

    
5952
        switch (opcode) {
5953
        case 0xd: /* SQDMULL, SQDMULL2 */
5954
            break;
5955
        case 0xb: /* SQDMLSL, SQDMLSL2 */
5956
            tcg_gen_neg_i64(tcg_res, tcg_res);
5957
            /* fall through */
5958
        case 0x9: /* SQDMLAL, SQDMLAL2 */
5959
            read_vec_element(s, tcg_op1, rd, 0, MO_64);
5960
            gen_helper_neon_addl_saturate_s64(tcg_res, cpu_env,
5961
                                              tcg_res, tcg_op1);
5962
            break;
5963
        default:
5964
            g_assert_not_reached();
5965
        }
5966

    
5967
        write_fp_dreg(s, rd, tcg_res);
5968

    
5969
        tcg_temp_free_i64(tcg_op1);
5970
        tcg_temp_free_i64(tcg_op2);
5971
        tcg_temp_free_i64(tcg_res);
5972
    } else {
5973
        TCGv_i32 tcg_op1 = tcg_temp_new_i32();
5974
        TCGv_i32 tcg_op2 = tcg_temp_new_i32();
5975
        TCGv_i64 tcg_res = tcg_temp_new_i64();
5976

    
5977
        read_vec_element_i32(s, tcg_op1, rn, 0, MO_16);
5978
        read_vec_element_i32(s, tcg_op2, rm, 0, MO_16);
5979

    
5980
        gen_helper_neon_mull_s16(tcg_res, tcg_op1, tcg_op2);
5981
        gen_helper_neon_addl_saturate_s32(tcg_res, cpu_env, tcg_res, tcg_res);
5982

    
5983
        switch (opcode) {
5984
        case 0xd: /* SQDMULL, SQDMULL2 */
5985
            break;
5986
        case 0xb: /* SQDMLSL, SQDMLSL2 */
5987
            gen_helper_neon_negl_u32(tcg_res, tcg_res);
5988
            /* fall through */
5989
        case 0x9: /* SQDMLAL, SQDMLAL2 */
5990
        {
5991
            TCGv_i64 tcg_op3 = tcg_temp_new_i64();
5992
            read_vec_element(s, tcg_op3, rd, 0, MO_32);
5993
            gen_helper_neon_addl_saturate_s32(tcg_res, cpu_env,
5994
                                              tcg_res, tcg_op3);
5995
            tcg_temp_free_i64(tcg_op3);
5996
            break;
5997
        }
5998
        default:
5999
            g_assert_not_reached();
6000
        }
6001

    
6002
        tcg_gen_ext32u_i64(tcg_res, tcg_res);
6003
        write_fp_dreg(s, rd, tcg_res);
6004

    
6005
        tcg_temp_free_i32(tcg_op1);
6006
        tcg_temp_free_i32(tcg_op2);
6007
        tcg_temp_free_i64(tcg_res);
6008
    }
6009
}
6010

    
6011
static void handle_3same_64(DisasContext *s, int opcode, bool u,
6012
                            TCGv_i64 tcg_rd, TCGv_i64 tcg_rn, TCGv_i64 tcg_rm)
6013
{
6014
    /* Handle 64x64->64 opcodes which are shared between the scalar
6015
     * and vector 3-same groups. We cover every opcode where size == 3
6016
     * is valid in either the three-reg-same (integer, not pairwise)
6017
     * or scalar-three-reg-same groups. (Some opcodes are not yet
6018
     * implemented.)
6019
     */
6020
    TCGCond cond;
6021

    
6022
    switch (opcode) {
6023
    case 0x1: /* SQADD */
6024
        if (u) {
6025
            gen_helper_neon_qadd_u64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6026
        } else {
6027
            gen_helper_neon_qadd_s64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6028
        }
6029
        break;
6030
    case 0x5: /* SQSUB */
6031
        if (u) {
6032
            gen_helper_neon_qsub_u64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6033
        } else {
6034
            gen_helper_neon_qsub_s64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6035
        }
6036
        break;
6037
    case 0x6: /* CMGT, CMHI */
6038
        /* 64 bit integer comparison, result = test ? (2^64 - 1) : 0.
6039
         * We implement this using setcond (test) and then negating.
6040
         */
6041
        cond = u ? TCG_COND_GTU : TCG_COND_GT;
6042
    do_cmop:
6043
        tcg_gen_setcond_i64(cond, tcg_rd, tcg_rn, tcg_rm);
6044
        tcg_gen_neg_i64(tcg_rd, tcg_rd);
6045
        break;
6046
    case 0x7: /* CMGE, CMHS */
6047
        cond = u ? TCG_COND_GEU : TCG_COND_GE;
6048
        goto do_cmop;
6049
    case 0x11: /* CMTST, CMEQ */
6050
        if (u) {
6051
            cond = TCG_COND_EQ;
6052
            goto do_cmop;
6053
        }
6054
        /* CMTST : test is "if (X & Y != 0)". */
6055
        tcg_gen_and_i64(tcg_rd, tcg_rn, tcg_rm);
6056
        tcg_gen_setcondi_i64(TCG_COND_NE, tcg_rd, tcg_rd, 0);
6057
        tcg_gen_neg_i64(tcg_rd, tcg_rd);
6058
        break;
6059
    case 0x8: /* SSHL, USHL */
6060
        if (u) {
6061
            gen_helper_neon_shl_u64(tcg_rd, tcg_rn, tcg_rm);
6062
        } else {
6063
            gen_helper_neon_shl_s64(tcg_rd, tcg_rn, tcg_rm);
6064
        }
6065
        break;
6066
    case 0x9: /* SQSHL, UQSHL */
6067
        if (u) {
6068
            gen_helper_neon_qshl_u64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6069
        } else {
6070
            gen_helper_neon_qshl_s64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6071
        }
6072
        break;
6073
    case 0xa: /* SRSHL, URSHL */
6074
        if (u) {
6075
            gen_helper_neon_rshl_u64(tcg_rd, tcg_rn, tcg_rm);
6076
        } else {
6077
            gen_helper_neon_rshl_s64(tcg_rd, tcg_rn, tcg_rm);
6078
        }
6079
        break;
6080
    case 0xb: /* SQRSHL, UQRSHL */
6081
        if (u) {
6082
            gen_helper_neon_qrshl_u64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6083
        } else {
6084
            gen_helper_neon_qrshl_s64(tcg_rd, cpu_env, tcg_rn, tcg_rm);
6085
        }
6086
        break;
6087
    case 0x10: /* ADD, SUB */
6088
        if (u) {
6089
            tcg_gen_sub_i64(tcg_rd, tcg_rn, tcg_rm);
6090
        } else {
6091
            tcg_gen_add_i64(tcg_rd, tcg_rn, tcg_rm);
6092
        }
6093
        break;
6094
    default:
6095
        g_assert_not_reached();
6096
    }
6097
}
6098

    
6099
/* Handle the 3-same-operands float operations; shared by the scalar
6100
 * and vector encodings. The caller must filter out any encodings
6101
 * not allocated for the encoding it is dealing with.
6102
 */
6103
static void handle_3same_float(DisasContext *s, int size, int elements,
6104
                               int fpopcode, int rd, int rn, int rm)
6105
{
6106
    int pass;
6107
    TCGv_ptr fpst = get_fpstatus_ptr();
6108

    
6109
    for (pass = 0; pass < elements; pass++) {
6110
        if (size) {
6111
            /* Double */
6112
            TCGv_i64 tcg_op1 = tcg_temp_new_i64();
6113
            TCGv_i64 tcg_op2 = tcg_temp_new_i64();
6114
            TCGv_i64 tcg_res = tcg_temp_new_i64();
6115

    
6116
            read_vec_element(s, tcg_op1, rn, pass, MO_64);
6117
            read_vec_element(s, tcg_op2, rm, pass, MO_64);
6118

    
6119
            switch (fpopcode) {
6120
            case 0x39: /* FMLS */
6121
                /* As usual for ARM, separate negation for fused multiply-add */
6122
                gen_helper_vfp_negd(tcg_op1, tcg_op1);
6123
                /* fall through */
6124
            case 0x19: /* FMLA */
6125
                read_vec_element(s, tcg_res, rd, pass, MO_64);
6126
                gen_helper_vfp_muladdd(tcg_res, tcg_op1, tcg_op2,
6127
                                       tcg_res, fpst);
6128
                break;
6129
            case 0x18: /* FMAXNM */
6130
                gen_helper_vfp_maxnumd(tcg_res, tcg_op1, tcg_op2, fpst);
6131
                break;
6132
            case 0x1a: /* FADD */
6133
                gen_helper_vfp_addd(tcg_res, tcg_op1, tcg_op2, fpst);
6134
                break;
6135
            case 0x1b: /* FMULX */
6136
                gen_helper_vfp_mulxd(tcg_res, tcg_op1, tcg_op2, fpst);
6137
                break;
6138
            case 0x1c: /* FCMEQ */
6139
                gen_helper_neon_ceq_f64(tcg_res, tcg_op1, tcg_op2, fpst);
6140
                break;
6141
            case 0x1e: /* FMAX */
6142
                gen_helper_vfp_maxd(tcg_res, tcg_op1, tcg_op2, fpst);
6143
                break;
6144
            case 0x1f: /* FRECPS */
6145
                gen_helper_recpsf_f64(tcg_res, tcg_op1, tcg_op2, fpst);
6146
                break;
6147
            case 0x38: /* FMINNM */
6148
                gen_helper_vfp_minnumd(tcg_res, tcg_op1, tcg_op2, fpst);
6149
                break;
6150
            case 0x3a: /* FSUB */
6151
                gen_helper_vfp_subd(tcg_res, tcg_op1, tcg_op2, fpst);
6152
                break;
6153
            case 0x3e: /* FMIN */
6154
                gen_helper_vfp_mind(tcg_res, tcg_op1, tcg_op2, fpst);
6155
                break;
6156
            case 0x3f: /* FRSQRTS */
6157
                gen_helper_rsqrtsf_f64(tcg_res, tcg_op1, tcg_op2, fpst);
6158
                break;
6159
            case 0x5b: /* FMUL */
6160
                gen_helper_vfp_muld(tcg_res, tcg_op1, tcg_op2, fpst);
6161
                break;
6162
            case 0x5c: /* FCMGE */
6163
                gen_helper_neon_cge_f64(tcg_res, tcg_op1, tcg_op2, fpst);
6164
                break;
6165
            case 0x5d: /* FACGE */
6166
                gen_helper_neon_acge_f64(tcg_res, tcg_op1, tcg_op2, fpst);
6167
                break;
6168
            case 0x5f: /* FDIV */
6169
                gen_helper_vfp_divd(tcg_res, tcg_op1, tcg_op2, fpst);
6170
                break;
6171
            case 0x7a: /* FABD */
6172
                gen_helper_vfp_subd(tcg_res, tcg_op1, tcg_op2, fpst);
6173
                gen_helper_vfp_absd(tcg_res, tcg_res);
6174
                break;
6175
            case 0x7c: /* FCMGT */
6176
                gen_helper_neon_cgt_f64(tcg_res, tcg_op1, tcg_op2, fpst);
6177
                break;
6178
            case 0x7d: /* FACGT */
6179
                gen_helper_neon_acgt_f64(tcg_res, tcg_op1, tcg_op2, fpst);
6180
                break;
6181
            default:
6182
                g_assert_not_reached();
6183
            }
6184

    
6185
            write_vec_element(s, tcg_res, rd, pass, MO_64);
6186

    
6187
            tcg_temp_free_i64(tcg_res);
6188
            tcg_temp_free_i64(tcg_op1);
6189
            tcg_temp_free_i64(tcg_op2);
6190
        } else {
6191
            /* Single */
6192
            TCGv_i32 tcg_op1 = tcg_temp_new_i32();
6193
            TCGv_i32 tcg_op2 = tcg_temp_new_i32();
6194
            TCGv_i32 tcg_res = tcg_temp_new_i32();
6195

    
6196
            read_vec_element_i32(s, tcg_op1, rn, pass, MO_32);
6197
            read_vec_element_i32(s, tcg_op2, rm, pass, MO_32);
6198

    
6199
            switch (fpopcode) {
6200
            case 0x39: /* FMLS */
6201
                /* As usual for ARM, separate negation for fused multiply-add */
6202
                gen_helper_vfp_negs(tcg_op1, tcg_op1);
6203
                /* fall through */
6204
            case 0x19: /* FMLA */
6205
                read_vec_element_i32(s, tcg_res, rd, pass, MO_32);
6206
                gen_helper_vfp_muladds(tcg_res, tcg_op1, tcg_op2,
6207
                                       tcg_res, fpst);
6208
                break;
6209
            case 0x1a: /* FADD */
6210
                gen_helper_vfp_adds(tcg_res, tcg_op1, tcg_op2, fpst);
6211
                break;
6212
            case 0x1b: /* FMULX */
6213
                gen_helper_vfp_mulxs(tcg_res, tcg_op1, tcg_op2, fpst);
6214
                break;
6215
            case 0x1c: /* FCMEQ */
6216
                gen_helper_neon_ceq_f32(tcg_res, tcg_op1, tcg_op2, fpst);
6217
                break;
6218
            case 0x1e: /* FMAX */
6219
                gen_helper_vfp_maxs(tcg_res, tcg_op1, tcg_op2, fpst);
6220
                break;
6221
            case 0x1f: /* FRECPS */
6222
                gen_helper_recpsf_f32(tcg_res, tcg_op1, tcg_op2, fpst);
6223
                break;
6224
            case 0x18: /* FMAXNM */
6225
                gen_helper_vfp_maxnums(tcg_res, tcg_op1, tcg_op2, fpst);
6226
                break;
6227
            case 0x38: /* FMINNM */
6228
                gen_helper_vfp_minnums(tcg_res, tcg_op1, tcg_op2, fpst);
6229
                break;
6230
            case 0x3a: /* FSUB */
6231
                gen_helper_vfp_subs(tcg_res, tcg_op1, tcg_op2, fpst);
6232
                break;
6233
            case 0x3e: /* FMIN */
6234
                gen_helper_vfp_mins(tcg_res, tcg_op1, tcg_op2, fpst);
6235
                break;
6236
            case 0x3f: /* FRSQRTS */
6237
                gen_helper_rsqrtsf_f32(tcg_res, tcg_op1, tcg_op2, fpst);
6238
                break;
6239
            case 0x5b: /* FMUL */
6240
                gen_helper_vfp_muls(tcg_res, tcg_op1, tcg_op2, fpst);
6241
                break;
6242
            case 0x5c: /* FCMGE */
6243
                gen_helper_neon_cge_f32(tcg_res, tcg_op1, tcg_op2, fpst);
6244
                break;
6245
            case 0x5d: /* FACGE */
6246
                gen_helper_neon_acge_f32(tcg_res, tcg_op1, tcg_op2, fpst);
6247
                break;
6248
            case 0x5f: /* FDIV */
6249
                gen_helper_vfp_divs(tcg_res, tcg_op1, tcg_op2, fpst);
6250
                break;
6251
            case 0x7a: /* FABD */
6252
                gen_helper_vfp_subs(tcg_res, tcg_op1, tcg_op2, fpst);
6253
                gen_helper_vfp_abss(tcg_res, tcg_res);
6254
                break;
6255
            case 0x7c: /* FCMGT */
6256
                gen_helper_neon_cgt_f32(tcg_res, tcg_op1, tcg_op2, fpst);
6257
                break;
6258
            case 0x7d: /* FACGT */
6259
                gen_helper_neon_acgt_f32(tcg_res, tcg_op1, tcg_op2, fpst);
6260
                break;
6261
            default:
6262
                g_assert_not_reached();
6263
            }
6264

    
6265
            if (elements == 1) {
6266
                /* scalar single so clear high part */
6267
                TCGv_i64 tcg_tmp = tcg_temp_new_i64();
6268

    
6269
                tcg_gen_extu_i32_i64(tcg_tmp, tcg_res);
6270
                write_vec_element(s, tcg_tmp, rd, pass, MO_64);
6271
                tcg_temp_free_i64(tcg_tmp);
6272
            } else {
6273
                write_vec_element_i32(s, tcg_res, rd, pass, MO_32);
6274
            }
6275

    
6276
            tcg_temp_free_i32(tcg_res);
6277
            tcg_temp_free_i32(tcg_op1);
6278
            tcg_temp_free_i32(tcg_op2);
6279
        }
6280
    }
6281

    
6282
    tcg_temp_free_ptr(fpst);
6283

    
6284
    if ((elements << size) < 4) {
6285
        /* scalar, or non-quad vector op */
6286
        clear_vec_high(s, rd);
6287
    }
6288
}
6289

    
6290
/* C3.6.11 AdvSIMD scalar three same
6291
 *  31 30  29 28       24 23  22  21 20  16 15    11  10 9    5 4    0
6292
 * +-----+---+-----------+------+---+------+--------+---+------+------+
6293
 * | 0 1 | U | 1 1 1 1 0 | size | 1 |  Rm  | opcode | 1 |  Rn  |  Rd  |
6294
 * +-----+---+-----------+------+---+------+--------+---+------+------+
6295
 */
6296
static void disas_simd_scalar_three_reg_same(DisasContext *s, uint32_t insn)
6297
{
6298
    int rd = extract32(insn, 0, 5);
6299
    int rn = extract32(insn, 5, 5);
6300
    int opcode = extract32(insn, 11, 5);
6301
    int rm = extract32(insn, 16, 5);
6302
    int size = extract32(insn, 22, 2);
6303
    bool u = extract32(insn, 29, 1);
6304
    TCGv_i64 tcg_rd;
6305

    
6306
    if (opcode >= 0x18) {
6307
        /* Floating point: U, size[1] and opcode indicate operation */
6308
        int fpopcode = opcode | (extract32(size, 1, 1) << 5) | (u << 6);
6309
        switch (fpopcode) {
6310
        case 0x1b: /* FMULX */
6311
        case 0x1f: /* FRECPS */
6312
        case 0x3f: /* FRSQRTS */
6313
        case 0x5d: /* FACGE */
6314
        case 0x7d: /* FACGT */
6315
        case 0x1c: /* FCMEQ */
6316
        case 0x5c: /* FCMGE */
6317
        case 0x7c: /* FCMGT */
6318
        case 0x7a: /* FABD */
6319
            break;
6320
        default:
6321
            unallocated_encoding(s);
6322
            return;
6323
        }
6324

    
6325
        handle_3same_float(s, extract32(size, 0, 1), 1, fpopcode, rd, rn, rm);
6326
        return;
6327
    }
6328

    
6329
    switch (opcode) {
6330
    case 0x1: /* SQADD, UQADD */
6331
    case 0x5: /* SQSUB, UQSUB */
6332
    case 0x9: /* SQSHL, UQSHL */
6333
    case 0xb: /* SQRSHL, UQRSHL */
6334
        break;
6335
    case 0x8: /* SSHL, USHL */
6336
    case 0xa: /* SRSHL, URSHL */
6337
    case 0x6: /* CMGT, CMHI */
6338
    case 0x7: /* CMGE, CMHS */
6339
    case 0x11: /* CMTST, CMEQ */
6340
    case 0x10: /* ADD, SUB (vector) */
6341
        if (size != 3) {
6342
            unallocated_encoding(s);
6343
            return;
6344
        }
6345
        break;
6346
    case 0x16: /* SQDMULH, SQRDMULH (vector) */
6347
        if (size != 1 && size != 2) {
6348
            unallocated_encoding(s);
6349
            return;
6350
        }
6351
        break;
6352
    default:
6353
        unallocated_encoding(s);
6354
        return;
6355
    }
6356

    
6357
    tcg_rd = tcg_temp_new_i64();
6358

    
6359
    if (size == 3) {
6360
        TCGv_i64 tcg_rn = read_fp_dreg(s, rn);
6361
        TCGv_i64 tcg_rm = read_fp_dreg(s, rm);
6362

    
6363
        handle_3same_64(s, opcode, u, tcg_rd, tcg_rn, tcg_rm);
6364
        tcg_temp_free_i64(tcg_rn);
6365
        tcg_temp_free_i64(tcg_rm);
6366
    } else {
6367
        /* Do a single operation on the lowest element in the vector.
6368
         * We use the standard Neon helpers and rely on 0 OP 0 == 0 with
6369
         * no side effects for all these operations.
6370
         * OPTME: special-purpose helpers would avoid doing some
6371
         * unnecessary work in the helper for the 8 and 16 bit cases.
6372
         */
6373
        NeonGenTwoOpEnvFn *genenvfn;
6374
        TCGv_i32 tcg_rn = tcg_temp_new_i32();
6375
        TCGv_i32 tcg_rm = tcg_temp_new_i32();
6376
        TCGv_i32 tcg_rd32 = tcg_temp_new_i32();
6377

    
6378
        read_vec_element_i32(s, tcg_rn, rn, 0, size);
6379
        read_vec_element_i32(s, tcg_rm, rm, 0, size);
6380

    
6381
        switch (opcode) {
6382
        case 0x1: /* SQADD, UQADD */
6383
        {
6384
            static NeonGenTwoOpEnvFn * const fns[3][2] = {
6385
                { gen_helper_neon_qadd_s8, gen_helper_neon_qadd_u8 },
6386
                { gen_helper_neon_qadd_s16, gen_helper_neon_qadd_u16 },
6387
                { gen_helper_neon_qadd_s32, gen_helper_neon_qadd_u32 },
6388
            };
6389
            genenvfn = fns[size][u];
6390
            break;
6391
        }
6392
        case 0x5: /* SQSUB, UQSUB */
6393
        {
6394
            static NeonGenTwoOpEnvFn * const fns[3][2] = {
6395
                { gen_helper_neon_qsub_s8, gen_helper_neon_qsub_u8 },
6396
                { gen_helper_neon_qsub_s16, gen_helper_neon_qsub_u16 },
6397
                { gen_helper_neon_qsub_s32, gen_helper_neon_qsub_u32 },
6398
            };
6399
            genenvfn = fns[size][u];
6400
            break;
6401
        }
6402
        case 0x9: /* SQSHL, UQSHL */
6403
        {
6404
            static NeonGenTwoOpEnvFn * const fns[3][2] = {
6405
                { gen_helper_neon_qshl_s8, gen_helper_neon_qshl_u8 },
6406
                { gen_helper_neon_qshl_s16, gen_helper_neon_qshl_u16 },
6407
                { gen_helper_neon_qshl_s32, gen_helper_neon_qshl_u32 },
6408
            };
6409
            genenvfn = fns[size][u];
6410
            break;
6411
        }
6412
        case 0xb: /* SQRSHL, UQRSHL */
6413
        {
6414
            static NeonGenTwoOpEnvFn * const fns[3][2] = {
6415
                { gen_helper_neon_qrshl_s8, gen_helper_neon_qrshl_u8 },
6416
                { gen_helper_neon_qrshl_s16, gen_helper_neon_qrshl_u16 },
6417
                { gen_helper_neon_qrshl_s32, gen_helper_neon_qrshl_u32 },
6418
            };
6419
            genenvfn = fns[size][u];
6420
            break;
6421
        }
6422
        case 0x16: /* SQDMULH, SQRDMULH */
6423
        {
6424
            static NeonGenTwoOpEnvFn * const fns[2][2] = {
6425
                { gen_helper_neon_qdmulh_s16, gen_helper_neon_qrdmulh_s16 },
6426
                { gen_helper_neon_qdmulh_s32, gen_helper_neon_qrdmulh_s32 },
6427
            };
6428
            assert(size == 1 || size == 2);
6429
            genenvfn = fns[size - 1][u];
6430
            break;
6431
        }
6432
        default:
6433
            g_assert_not_reached();
6434
        }
6435

    
6436
        genenvfn(tcg_rd32, cpu_env, tcg_rn, tcg_rm);
6437
        tcg_gen_extu_i32_i64(tcg_rd, tcg_rd32);
6438
        tcg_temp_free_i32(tcg_rd32);
6439
        tcg_temp_free_i32(tcg_rn);
6440
        tcg_temp_free_i32(tcg_rm);
6441
    }
6442

    
6443
    write_fp_dreg(s, rd, tcg_rd);
6444

    
6445
    tcg_temp_free_i64(tcg_rd);
6446
}
6447

    
6448
static void handle_2misc_64(DisasContext *s, int opcode, bool u,
6449
                            TCGv_i64 tcg_rd, TCGv_i64 tcg_rn)
6450
{
6451
    /* Handle 64->64 opcodes which are shared between the scalar and
6452
     * vector 2-reg-misc groups. We cover every integer opcode where size == 3
6453
     * is valid in either group and also the double-precision fp ops.
6454
     */
6455
    TCGCond cond;
6456

    
6457
    switch (opcode) {
6458
    case 0x5: /* NOT */
6459
        /* This opcode is shared with CNT and RBIT but we have earlier
6460
         * enforced that size == 3 if and only if this is the NOT insn.
6461
         */
6462
        tcg_gen_not_i64(tcg_rd, tcg_rn);
6463
        break;
6464
    case 0xa: /* CMLT */
6465
        /* 64 bit integer comparison against zero, result is
6466
         * test ? (2^64 - 1) : 0. We implement via setcond(!test) and
6467
         * subtracting 1.
6468
         */
6469
        cond = TCG_COND_LT;
6470
    do_cmop:
6471
        tcg_gen_setcondi_i64(cond, tcg_rd, tcg_rn, 0);
6472
        tcg_gen_neg_i64(tcg_rd, tcg_rd);
6473
        break;
6474
    case 0x8: /* CMGT, CMGE */
6475
        cond = u ? TCG_COND_GE : TCG_COND_GT;
6476
        goto do_cmop;
6477
    case 0x9: /* CMEQ, CMLE */
6478
        cond = u ? TCG_COND_LE : TCG_COND_EQ;
6479
        goto do_cmop;
6480
    case 0xb: /* ABS, NEG */
6481
        if (u) {
6482
            tcg_gen_neg_i64(tcg_rd, tcg_rn);
6483
        } else {
6484
            TCGv_i64 tcg_zero = tcg_const_i64(0);
6485
            tcg_gen_neg_i64(tcg_rd, tcg_rn);
6486
            tcg_gen_movcond_i64(TCG_COND_GT, tcg_rd, tcg_rn, tcg_zero,
6487
                                tcg_rn, tcg_rd);
6488
            tcg_temp_free_i64(tcg_zero);
6489
        }
6490
        break;
6491
    case 0x2f: /* FABS */
6492
        gen_helper_vfp_absd(tcg_rd, tcg_rn);
6493
        break;
6494
    case 0x6f: /* FNEG */
6495
        gen_helper_vfp_negd(tcg_rd, tcg_rn);
6496
        break;
6497
    default:
6498
        g_assert_not_reached();
6499
    }
6500
}
6501

    
6502
static void handle_2misc_fcmp_zero(DisasContext *s, int opcode,
6503
                                   bool is_scalar, bool is_u, bool is_q,
6504
                                   int size, int rn, int rd)
6505
{
6506
    bool is_double = (size == 3);
6507
    TCGv_ptr fpst = get_fpstatus_ptr();
6508

    
6509
    if (is_double) {
6510
        TCGv_i64 tcg_op = tcg_temp_new_i64();
6511
        TCGv_i64 tcg_zero = tcg_const_i64(0);
6512
        TCGv_i64 tcg_res = tcg_temp_new_i64();
6513
        NeonGenTwoDoubleOPFn *genfn;
6514
        bool swap = false;
6515
        int pass;
6516

    
6517
        switch (opcode) {
6518
        case 0x2e: /* FCMLT (zero) */
6519
            swap = true;
6520
            /* fallthrough */
6521
        case 0x2c: /* FCMGT (zero) */
6522
            genfn = gen_helper_neon_cgt_f64;
6523
            break;
6524
        case 0x2d: /* FCMEQ (zero) */
6525
            genfn = gen_helper_neon_ceq_f64;
6526
            break;
6527
        case 0x6d: /* FCMLE (zero) */
6528
            swap = true;
6529
            /* fall through */
6530
        case 0x6c: /* FCMGE (zero) */
6531
            genfn = gen_helper_neon_cge_f64;
6532
            break;
6533
        default:
6534
            g_assert_not_reached();
6535
        }
6536

    
6537
        for (pass = 0; pass < (is_scalar ? 1 : 2); pass++) {
6538
            read_vec_element(s, tcg_op, rn, pass, MO_64);
6539
            if (swap) {
6540
                genfn(tcg_res, tcg_zero, tcg_op, fpst);
6541
            } else {
6542
                genfn(tcg_res, tcg_op, tcg_zero, fpst);
6543
            }
6544
            write_vec_element(s, tcg_res, rd, pass, MO_64);
6545
        }
6546
        if (is_scalar) {
6547
            clear_vec_high(s, rd);
6548
        }
6549

    
6550
        tcg_temp_free_i64(tcg_res);
6551
        tcg_temp_free_i64(tcg_zero);
6552
        tcg_temp_free_i64(tcg_op);
6553
    } else {
6554
        TCGv_i32 tcg_op = tcg_temp_new_i32();
6555
        TCGv_i32 tcg_zero = tcg_const_i32(0);
6556
        TCGv_i32 tcg_res = tcg_temp_new_i32();
6557
        NeonGenTwoSingleOPFn *genfn;
6558
        bool swap = false;
6559
        int pass, maxpasses;
6560

    
6561
        switch (opcode) {
6562
        case 0x2e: /* FCMLT (zero) */
6563
            swap = true;
6564
            /* fall through */
6565
        case 0x2c: /* FCMGT (zero) */
6566
            genfn = gen_helper_neon_cgt_f32;
6567
            break;
6568
        case 0x2d: /* FCMEQ (zero) */
6569
            genfn = gen_helper_neon_ceq_f32;
6570
            break;
6571
        case 0x6d: /* FCMLE (zero) */
6572
            swap = true;
6573
            /* fall through */
6574
        case 0x6c: /* FCMGE (zero) */
6575
            genfn = gen_helper_neon_cge_f32;
6576
            break;
6577
        default:
6578
            g_assert_not_reached();
6579
        }
6580

    
6581
        if (is_scalar) {
6582
            maxpasses = 1;
6583
        } else {
6584
            maxpasses = is_q ? 4 : 2;
6585
        }
6586

    
6587
        for (pass = 0; pass < maxpasses; pass++) {
6588
            read_vec_element_i32(s, tcg_op, rn, pass, MO_32);
6589
            if (swap) {
6590
                genfn(tcg_res, tcg_zero, tcg_op, fpst);
6591
            } else {
6592
                genfn(tcg_res, tcg_op, tcg_zero, fpst);
6593
            }
6594
            if (is_scalar) {
6595
                write_fp_sreg(s, rd, tcg_res);
6596
            } else {
6597
                write_vec_element_i32(s, tcg_res, rd, pass, MO_32);
6598
            }
6599
        }
6600
        tcg_temp_free_i32(tcg_res);
6601
        tcg_temp_free_i32(tcg_zero);
6602
        tcg_temp_free_i32(tcg_op);
6603
        if (!is_q && !is_scalar) {
6604
            clear_vec_high(s, rd);
6605
        }
6606
    }
6607

    
6608
    tcg_temp_free_ptr(fpst);
6609
}
6610

    
6611
/* C3.6.12 AdvSIMD scalar two reg misc
6612
 *  31 30  29 28       24 23  22 21       17 16    12 11 10 9    5 4    0
6613
 * +-----+---+-----------+------+-----------+--------+-----+------+------+
6614
 * | 0 1 | U | 1 1 1 1 0 | size | 1 0 0 0 0 | opcode | 1 0 |  Rn  |  Rd  |
6615
 * +-----+---+-----------+------+-----------+--------+-----+------+------+
6616
 */
6617
static void disas_simd_scalar_two_reg_misc(DisasContext *s, uint32_t insn)
6618
{
6619
    int rd = extract32(insn, 0, 5);
6620
    int rn = extract32(insn, 5, 5);
6621
    int opcode = extract32(insn, 12, 5);
6622
    int size = extract32(insn, 22, 2);
6623
    bool u = extract32(insn, 29, 1);
6624

    
6625
    switch (opcode) {
6626
    case 0xa: /* CMLT */
6627
        if (u) {
6628
            unallocated_encoding(s);
6629
            return;
6630
        }
6631
        /* fall through */
6632
    case 0x8: /* CMGT, CMGE */
6633
    case 0x9: /* CMEQ, CMLE */
6634
    case 0xb: /* ABS, NEG */
6635
        if (size != 3) {
6636
            unallocated_encoding(s);
6637
            return;
6638
        }
6639
        break;
6640
    case 0xc ... 0xf:
6641
    case 0x16 ... 0x1d:
6642
    case 0x1f:
6643
        /* Floating point: U, size[1] and opcode indicate operation;
6644
         * size[0] indicates single or double precision.
6645
         */
6646
        opcode |= (extract32(size, 1, 1) << 5) | (u << 6);
6647
        size = extract32(size, 0, 1) ? 3 : 2;
6648
        switch (opcode) {
6649
        case 0x2c: /* FCMGT (zero) */
6650
        case 0x2d: /* FCMEQ (zero) */
6651
        case 0x2e: /* FCMLT (zero) */
6652
        case 0x6c: /* FCMGE (zero) */
6653
        case 0x6d: /* FCMLE (zero) */
6654
            handle_2misc_fcmp_zero(s, opcode, true, u, true, size, rn, rd);
6655
            return;
6656
        case 0x1a: /* FCVTNS */
6657
        case 0x1b: /* FCVTMS */
6658
        case 0x1c: /* FCVTAS */
6659
        case 0x1d: /* SCVTF */
6660
        case 0x3a: /* FCVTPS */
6661
        case 0x3b: /* FCVTZS */
6662
        case 0x3d: /* FRECPE */
6663
        case 0x3f: /* FRECPX */
6664
        case 0x56: /* FCVTXN, FCVTXN2 */
6665
        case 0x5a: /* FCVTNU */
6666
        case 0x5b: /* FCVTMU */
6667
        case 0x5c: /* FCVTAU */
6668
        case 0x5d: /* UCVTF */
6669
        case 0x7a: /* FCVTPU */
6670
        case 0x7b: /* FCVTZU */
6671
        case 0x7d: /* FRSQRTE */
6672
            unsupported_encoding(s, insn);
6673
            return;
6674
        default:
6675
            unallocated_encoding(s);
6676
            return;
6677
        }
6678
        break;
6679
    default:
6680
        /* Other categories of encoding in this class:
6681
         *  + SUQADD/USQADD/SQABS/SQNEG : size 8, 16, 32 or 64
6682
         *  + SQXTN/SQXTN2/SQXTUN/SQXTUN2/UQXTN/UQXTN2:
6683
         *    narrowing saturate ops: size 64/32/16 -> 32/16/8
6684
         */
6685
        unsupported_encoding(s, insn);
6686
        return;
6687
    }
6688

    
6689
    if (size == 3) {
6690
        TCGv_i64 tcg_rn = read_fp_dreg(s, rn);
6691
        TCGv_i64 tcg_rd = tcg_temp_new_i64();
6692

    
6693
        handle_2misc_64(s, opcode, u, tcg_rd, tcg_rn);
6694
        write_fp_dreg(s, rd, tcg_rd);
6695
        tcg_temp_free_i64(tcg_rd);
6696
        tcg_temp_free_i64(tcg_rn);
6697
    } else {
6698
        /* the 'size might not be 64' ops aren't implemented yet */
6699
        g_assert_not_reached();
6700
    }
6701
}
6702

    
6703
/* SSHR[RA]/USHR[RA] - Vector shift right (optional rounding/accumulate) */
6704
static void handle_vec_simd_shri(DisasContext *s, bool is_q, bool is_u,
6705
                                 int immh, int immb, int opcode, int rn, int rd)
6706
{
6707
    int size = 32 - clz32(immh) - 1;
6708
    int immhb = immh << 3 | immb;
6709
    int shift = 2 * (8 << size) - immhb;
6710
    bool accumulate = false;
6711
    bool round = false;
6712
    int dsize = is_q ? 128 : 64;
6713
    int esize = 8 << size;
6714
    int elements = dsize/esize;
6715
    TCGMemOp memop = size | (is_u ? 0 : MO_SIGN);
6716
    TCGv_i64 tcg_rn = new_tmp_a64(s);
6717
    TCGv_i64 tcg_rd = new_tmp_a64(s);
6718
    TCGv_i64 tcg_round;
6719
    int i;
6720

    
6721
    if (extract32(immh, 3, 1) && !is_q) {
6722
        unallocated_encoding(s);
6723
        return;
6724
    }
6725

    
6726
    if (size > 3 && !is_q) {
6727
        unallocated_encoding(s);
6728
        return;
6729
    }
6730

    
6731
    switch (opcode) {
6732
    case 0x02: /* SSRA / USRA (accumulate) */
6733
        accumulate = true;
6734
        break;
6735
    case 0x04: /* SRSHR / URSHR (rounding) */
6736
        round = true;
6737
        break;
6738
    case 0x06: /* SRSRA / URSRA (accum + rounding) */
6739
        accumulate = round = true;
6740
        break;
6741
    }
6742

    
6743
    if (round) {
6744
        uint64_t round_const = 1ULL << (shift - 1);
6745
        tcg_round = tcg_const_i64(round_const);
6746
    } else {
6747
        TCGV_UNUSED_I64(tcg_round);
6748
    }
6749

    
6750
    for (i = 0; i < elements; i++) {
6751
        read_vec_element(s, tcg_rn, rn, i, memop);
6752
        if (accumulate) {
6753
            read_vec_element(s, tcg_rd, rd, i, memop);
6754
        }
6755

    
6756
        handle_shri_with_rndacc(tcg_rd, tcg_rn, tcg_round,
6757
                                accumulate, is_u, size, shift);
6758

    
6759
        write_vec_element(s, tcg_rd, rd, i, size);
6760
    }
6761

    
6762
    if (!is_q) {
6763
        clear_vec_high(s, rd);
6764
    }
6765

    
6766
    if (round) {
6767
        tcg_temp_free_i64(tcg_round);
6768
    }
6769
}
6770

    
6771
/* SHL/SLI - Vector shift left */
6772
static void handle_vec_simd_shli(DisasContext *s, bool is_q, bool insert,
6773
                                int immh, int immb, int opcode, int rn, int rd)
6774
{
6775
    int size = 32 - clz32(immh) - 1;
6776
    int immhb = immh << 3 | immb;
6777
    int shift = immhb - (8 << size);
6778
    int dsize = is_q ? 128 : 64;
6779
    int esize = 8 << size;
6780
    int elements = dsize/esize;
6781
    TCGv_i64 tcg_rn = new_tmp_a64(s);
6782
    TCGv_i64 tcg_rd = new_tmp_a64(s);
6783
    int i;
6784

    
6785
    if (extract32(immh, 3, 1) && !is_q) {
6786
        unallocated_encoding(s);
6787
        return;
6788
    }
6789

    
6790
    if (size > 3 && !is_q) {
6791
        unallocated_encoding(s);
6792
        return;
6793
    }
6794

    
6795
    for (i = 0; i < elements; i++) {
6796
        read_vec_element(s, tcg_rn, rn, i, size);
6797
        if (insert) {
6798
            read_vec_element(s, tcg_rd, rd, i, size);
6799
        }
6800

    
6801
        handle_shli_with_ins(tcg_rd, tcg_rn, insert, shift);
6802

    
6803
        write_vec_element(s, tcg_rd, rd, i, size);
6804
    }
6805

    
6806
    if (!is_q) {
6807
        clear_vec_high(s, rd);
6808
    }
6809
}
6810

    
6811
/* USHLL/SHLL - Vector shift left with widening */
6812
static void handle_vec_simd_wshli(DisasContext *s, bool is_q, bool is_u,
6813
                                 int immh, int immb, int opcode, int rn, int rd)
6814
{
6815
    int size = 32 - clz32(immh) - 1;
6816
    int immhb = immh << 3 | immb;
6817
    int shift = immhb - (8 << size);
6818
    int dsize = 64;
6819
    int esize = 8 << size;
6820
    int elements = dsize/esize;
6821
    TCGv_i64 tcg_rn = new_tmp_a64(s);
6822
    TCGv_i64 tcg_rd = new_tmp_a64(s);
6823
    int i;
6824

    
6825
    if (size >= 3) {
6826
        unallocated_encoding(s);
6827
        return;
6828
    }
6829

    
6830
    /* For the LL variants the store is larger than the load,
6831
     * so if rd == rn we would overwrite parts of our input.
6832
     * So load everything right now and use shifts in the main loop.
6833
     */
6834
    read_vec_element(s, tcg_rn, rn, is_q ? 1 : 0, MO_64);
6835

    
6836
    for (i = 0; i < elements; i++) {
6837
        tcg_gen_shri_i64(tcg_rd, tcg_rn, i * esize);
6838
        ext_and_shift_reg(tcg_rd, tcg_rd, size | (!is_u << 2), 0);
6839
        tcg_gen_shli_i64(tcg_rd, tcg_rd, shift);
6840
        write_vec_element(s, tcg_rd, rd, i, size + 1);
6841
    }
6842
}
6843

    
6844

    
6845
/* C3.6.14 AdvSIMD shift by immediate
6846
 *  31  30   29 28         23 22  19 18  16 15    11  10 9    5 4    0
6847
 * +---+---+---+-------------+------+------+--------+---+------+------+
6848
 * | 0 | Q | U | 0 1 1 1 1 0 | immh | immb | opcode | 1 |  Rn  |  Rd  |
6849
 * +---+---+---+-------------+------+------+--------+---+------+------+
6850
 */
6851
static void disas_simd_shift_imm(DisasContext *s, uint32_t insn)
6852
{
6853
    int rd = extract32(insn, 0, 5);
6854
    int rn = extract32(insn, 5, 5);
6855
    int opcode = extract32(insn, 11, 5);
6856
    int immb = extract32(insn, 16, 3);
6857
    int immh = extract32(insn, 19, 4);
6858
    bool is_u = extract32(insn, 29, 1);
6859
    bool is_q = extract32(insn, 30, 1);
6860

    
6861
    switch (opcode) {
6862
    case 0x00: /* SSHR / USHR */
6863
    case 0x02: /* SSRA / USRA (accumulate) */
6864
    case 0x04: /* SRSHR / URSHR (rounding) */
6865
    case 0x06: /* SRSRA / URSRA (accum + rounding) */
6866
        handle_vec_simd_shri(s, is_q, is_u, immh, immb, opcode, rn, rd);
6867
        break;
6868
    case 0x0a: /* SHL / SLI */
6869
        handle_vec_simd_shli(s, is_q, is_u, immh, immb, opcode, rn, rd);
6870
        break;
6871
    case 0x14: /* SSHLL / USHLL */
6872
        handle_vec_simd_wshli(s, is_q, is_u, immh, immb, opcode, rn, rd);
6873
        break;
6874
    default:
6875
        /* We don't currently implement any of the Narrow or saturating shifts;
6876
         * nor do we implement the fixed-point conversions in this
6877
         * encoding group (SCVTF, FCVTZS, UCVTF, FCVTZU).
6878
         */
6879
        unsupported_encoding(s, insn);
6880
        return;
6881
    }
6882
}
6883

    
6884
/* Generate code to do a "long" addition or subtraction, ie one done in
6885
 * TCGv_i64 on vector lanes twice the width specified by size.
6886
 */
6887
static void gen_neon_addl(int size, bool is_sub, TCGv_i64 tcg_res,
6888
                          TCGv_i64 tcg_op1, TCGv_i64 tcg_op2)
6889
{
6890
    static NeonGenTwo64OpFn * const fns[3][2] = {
6891
        { gen_helper_neon_addl_u16, gen_helper_neon_subl_u16 },
6892
        { gen_helper_neon_addl_u32, gen_helper_neon_subl_u32 },
6893
        { tcg_gen_add_i64, tcg_gen_sub_i64 },
6894
    };
6895
    NeonGenTwo64OpFn *genfn;
6896
    assert(size < 3);
6897

    
6898
    genfn = fns[size][is_sub];
6899
    genfn(tcg_res, tcg_op1, tcg_op2);
6900
}
6901

    
6902
static void handle_3rd_widening(DisasContext *s, int is_q, int is_u, int size,
6903
                                int opcode, int rd, int rn, int rm)
6904
{
6905
    /* 3-reg-different widening insns: 64 x 64 -> 128 */
6906
    TCGv_i64 tcg_res[2];
6907
    int pass, accop;
6908

    
6909
    tcg_res[0] = tcg_temp_new_i64();
6910
    tcg_res[1] = tcg_temp_new_i64();
6911

    
6912
    /* Does this op do an adding accumulate, a subtracting accumulate,
6913
     * or no accumulate at all?
6914
     */
6915
    switch (opcode) {
6916
    case 5:
6917
    case 8:
6918
    case 9:
6919
        accop = 1;
6920
        break;
6921
    case 10:
6922
    case 11:
6923
        accop = -1;
6924
        break;
6925
    default:
6926
        accop = 0;
6927
        break;
6928
    }
6929

    
6930
    if (accop != 0) {
6931
        read_vec_element(s, tcg_res[0], rd, 0, MO_64);
6932
        read_vec_element(s, tcg_res[1], rd, 1, MO_64);
6933
    }
6934

    
6935
    /* size == 2 means two 32x32->64 operations; this is worth special
6936
     * casing because we can generally handle it inline.
6937
     */
6938
    if (size == 2) {
6939
        for (pass = 0; pass < 2; pass++) {
6940
            TCGv_i64 tcg_op1 = tcg_temp_new_i64();
6941
            TCGv_i64 tcg_op2 = tcg_temp_new_i64();
6942
            TCGv_i64 tcg_passres;
6943
            TCGMemOp memop = MO_32 | (is_u ? 0 : MO_SIGN);
6944

    
6945
            int elt = pass + is_q * 2;
6946

    
6947
            read_vec_element(s, tcg_op1, rn, elt, memop);
6948
            read_vec_element(s, tcg_op2, rm, elt, memop);
6949

    
6950
            if (accop == 0) {
6951
                tcg_passres = tcg_res[pass];
6952
            } else {
6953
                tcg_passres = tcg_temp_new_i64();
6954
            }
6955

    
6956
            switch (opcode) {
6957
            case 0: /* SADDL, SADDL2, UADDL, UADDL2 */
6958
                tcg_gen_add_i64(tcg_passres, tcg_op1, tcg_op2);
6959
                break;
6960
            case 2: /* SSUBL, SSUBL2, USUBL, USUBL2 */
6961
                tcg_gen_sub_i64(tcg_passres, tcg_op1, tcg_op2);
6962
                break;
6963
            case 5: /* SABAL, SABAL2, UABAL, UABAL2 */
6964
            case 7: /* SABDL, SABDL2, UABDL, UABDL2 */
6965
            {
6966
                TCGv_i64 tcg_tmp1 = tcg_temp_new_i64();
6967
                TCGv_i64 tcg_tmp2 = tcg_temp_new_i64();
6968

    
6969
                tcg_gen_sub_i64(tcg_tmp1, tcg_op1, tcg_op2);
6970
                tcg_gen_sub_i64(tcg_tmp2, tcg_op2, tcg_op1);
6971
                tcg_gen_movcond_i64(is_u ? TCG_COND_GEU : TCG_COND_GE,
6972
                                    tcg_passres,
6973
                                    tcg_op1, tcg_op2, tcg_tmp1, tcg_tmp2);
6974
                tcg_temp_free_i64(tcg_tmp1);
6975
                tcg_temp_free_i64(tcg_tmp2);
6976
                break;
6977
            }
6978
            case 8: /* SMLAL, SMLAL2, UMLAL, UMLAL2 */
6979
            case 10: /* SMLSL, SMLSL2, UMLSL, UMLSL2 */
6980
            case 12: /* UMULL, UMULL2, SMULL, SMULL2 */
6981
                tcg_gen_mul_i64(tcg_passres, tcg_op1, tcg_op2);
6982
                break;
6983
            case 9: /* SQDMLAL, SQDMLAL2 */
6984
            case 11: /* SQDMLSL, SQDMLSL2 */
6985
            case 13: /* SQDMULL, SQDMULL2 */
6986
                tcg_gen_mul_i64(tcg_passres, tcg_op1, tcg_op2);
6987
                gen_helper_neon_addl_saturate_s64(tcg_passres, cpu_env,
6988
                                                  tcg_passres, tcg_passres);
6989
                break;
6990
            default:
6991
                g_assert_not_reached();
6992
            }
6993

    
6994
            if (opcode == 9 || opcode == 11) {
6995
                /* saturating accumulate ops */
6996
                if (accop < 0) {
6997
                    tcg_gen_neg_i64(tcg_passres, tcg_passres);
6998
                }
6999
                gen_helper_neon_addl_saturate_s64(tcg_res[pass], cpu_env,
7000
                                                  tcg_res[pass], tcg_passres);
7001
            } else if (accop > 0) {
7002
                tcg_gen_add_i64(tcg_res[pass], tcg_res[pass], tcg_passres);
7003
            } else if (accop < 0) {
7004
                tcg_gen_sub_i64(tcg_res[pass], tcg_res[pass], tcg_passres);
7005
            }
7006

    
7007
            if (accop != 0) {
7008
                tcg_temp_free_i64(tcg_passres);
7009
            }
7010

    
7011
            tcg_temp_free_i64(tcg_op1);
7012
            tcg_temp_free_i64(tcg_op2);
7013
        }
7014
    } else {
7015
        /* size 0 or 1, generally helper functions */
7016
        for (pass = 0; pass < 2; pass++) {
7017
            TCGv_i32 tcg_op1 = tcg_temp_new_i32();
7018
            TCGv_i32 tcg_op2 = tcg_temp_new_i32();
7019
            TCGv_i64 tcg_passres;
7020
            int elt = pass + is_q * 2;
7021

    
7022
            read_vec_element_i32(s, tcg_op1, rn, elt, MO_32);
7023
            read_vec_element_i32(s, tcg_op2, rm, elt, MO_32);
7024

    
7025
            if (accop == 0) {
7026
                tcg_passres = tcg_res[pass];
7027
            } else {
7028
                tcg_passres = tcg_temp_new_i64();
7029
            }
7030

    
7031
            switch (opcode) {
7032
            case 0: /* SADDL, SADDL2, UADDL, UADDL2 */
7033
            case 2: /* SSUBL, SSUBL2, USUBL, USUBL2 */
7034
            {
7035
                TCGv_i64 tcg_op2_64 = tcg_temp_new_i64();
7036
                static NeonGenWidenFn * const widenfns[2][2] = {
7037
                    { gen_helper_neon_widen_s8, gen_helper_neon_widen_u8 },
7038
                    { gen_helper_neon_widen_s16, gen_helper_neon_widen_u16 },
7039
                };
7040
                NeonGenWidenFn *widenfn = widenfns[size][is_u];
7041

    
7042
                widenfn(tcg_op2_64, tcg_op2);
7043
                widenfn(tcg_passres, tcg_op1);
7044
                gen_neon_addl(size, (opcode == 2), tcg_passres,
7045
                              tcg_passres, tcg_op2_64);
7046
                tcg_temp_free_i64(tcg_op2_64);
7047
                break;
7048
            }
7049
            case 5: /* SABAL, SABAL2, UABAL, UABAL2 */
7050
            case 7: /* SABDL, SABDL2, UABDL, UABDL2 */
7051
                if (size == 0) {
7052
                    if (is_u) {
7053
                        gen_helper_neon_abdl_u16(tcg_passres, tcg_op1, tcg_op2);
7054
                    } else {
7055
                        gen_helper_neon_abdl_s16(tcg_passres, tcg_op1, tcg_op2);
7056
                    }
7057
                } else {
7058
                    if (is_u) {
7059
                        gen_helper_neon_abdl_u32(tcg_passres, tcg_op1, tcg_op2);
7060
                    } else {
7061
                        gen_helper_neon_abdl_s32(tcg_passres, tcg_op1, tcg_op2);
7062
                    }
7063
                }
7064
                break;
7065
            case 8: /* SMLAL, SMLAL2, UMLAL, UMLAL2 */
7066
            case 10: /* SMLSL, SMLSL2, UMLSL, UMLSL2 */
7067
            case 12: /* UMULL, UMULL2, SMULL, SMULL2 */
7068
                if (size == 0) {
7069
                    if (is_u) {
7070
                        gen_helper_neon_mull_u8(tcg_passres, tcg_op1, tcg_op2);
7071
                    } else {
7072
                        gen_helper_neon_mull_s8(tcg_passres, tcg_op1, tcg_op2);
7073
                    }
7074
                } else {
7075
                    if (is_u) {
7076
                        gen_helper_neon_mull_u16(tcg_passres, tcg_op1, tcg_op2);
7077
                    } else {
7078
                        gen_helper_neon_mull_s16(tcg_passres, tcg_op1, tcg_op2);
7079
                    }
7080
                }
7081
                break;
7082
            case 9: /* SQDMLAL, SQDMLAL2 */
7083
            case 11: /* SQDMLSL, SQDMLSL2 */
7084
            case 13: /* SQDMULL, SQDMULL2 */
7085
                assert(size == 1);
7086
                gen_helper_neon_mull_s16(tcg_passres, tcg_op1, tcg_op2);
7087
                gen_helper_neon_addl_saturate_s32(tcg_passres, cpu_env,
7088
                                                  tcg_passres, tcg_passres);
7089
                break;
7090
            default:
7091
                g_assert_not_reached();
7092
            }
7093
            tcg_temp_free_i32(tcg_op1);
7094
            tcg_temp_free_i32(tcg_op2);
7095

    
7096
            if (accop != 0) {
7097
                if (opcode == 9 || opcode == 11) {
7098
                    /* saturating accumulate ops */
7099
                    if (accop < 0) {
7100
                        gen_helper_neon_negl_u32(tcg_passres, tcg_passres);
7101
                    }
7102
                    gen_helper_neon_addl_saturate_s32(tcg_res[pass], cpu_env,
7103
                                                      tcg_res[pass],
7104
                                                      tcg_passres);
7105
                } else {
7106
                    gen_neon_addl(size, (accop < 0), tcg_res[pass],
7107
                                  tcg_res[pass], tcg_passres);
7108
                }
7109
                tcg_temp_free_i64(tcg_passres);
7110
            }
7111
        }
7112
    }
7113

    
7114
    write_vec_element(s, tcg_res[0], rd, 0, MO_64);
7115
    write_vec_element(s, tcg_res[1], rd, 1, MO_64);
7116
    tcg_temp_free_i64(tcg_res[0]);
7117
    tcg_temp_free_i64(tcg_res[1]);
7118
}
7119

    
7120
static void handle_3rd_wide(DisasContext *s, int is_q, int is_u, int size,
7121
                            int opcode, int rd, int rn, int rm)
7122
{
7123
    TCGv_i64 tcg_res[2];
7124
    int part = is_q ? 2 : 0;
7125
    int pass;
7126

    
7127
    for (pass = 0; pass < 2; pass++) {
7128
        TCGv_i64 tcg_op1 = tcg_temp_new_i64();
7129
        TCGv_i32 tcg_op2 = tcg_temp_new_i32();
7130
        TCGv_i64 tcg_op2_wide = tcg_temp_new_i64();
7131
        static NeonGenWidenFn * const widenfns[3][2] = {
7132
            { gen_helper_neon_widen_s8, gen_helper_neon_widen_u8 },
7133
            { gen_helper_neon_widen_s16, gen_helper_neon_widen_u16 },
7134
            { tcg_gen_ext_i32_i64, tcg_gen_extu_i32_i64 },
7135
        };
7136
        NeonGenWidenFn *widenfn = widenfns[size][is_u];
7137

    
7138
        read_vec_element(s, tcg_op1, rn, pass, MO_64);
7139
        read_vec_element_i32(s, tcg_op2, rm, part + pass, MO_32);
7140
        widenfn(tcg_op2_wide, tcg_op2);
7141
        tcg_temp_free_i32(tcg_op2);
7142
        tcg_res[pass] = tcg_temp_new_i64();
7143
        gen_neon_addl(size, (opcode == 3),
7144
                      tcg_res[pass], tcg_op1, tcg_op2_wide);
7145
        tcg_temp_free_i64(tcg_op1);
7146
        tcg_temp_free_i64(tcg_op2_wide);
7147
    }
7148

    
7149
    for (pass = 0; pass < 2; pass++) {
7150
        write_vec_element(s, tcg_res[pass], rd, pass, MO_64);
7151
        tcg_temp_free_i64(tcg_res[pass]);
7152
    }
7153
}
7154

    
7155
static void do_narrow_high_u32(TCGv_i32 res, TCGv_i64 in)
7156
{
7157
    tcg_gen_shri_i64(in, in, 32);
7158
    tcg_gen_trunc_i64_i32(res, in);
7159
}
7160

    
7161
static void do_narrow_round_high_u32(TCGv_i32 res, TCGv_i64 in)
7162
{
7163
    tcg_gen_addi_i64(in, in, 1U << 31);
7164
    do_narrow_high_u32(res, in);
7165
}
7166

    
7167
static void handle_3rd_narrowing(DisasContext *s, int is_q, int is_u, int size,
7168
                                 int opcode, int rd, int rn, int rm)
7169
{
7170
    TCGv_i32 tcg_res[2];
7171
    int part = is_q ? 2 : 0;
7172
    int pass;
7173

    
7174
    for (pass = 0; pass < 2; pass++) {
7175
        TCGv_i64 tcg_op1 = tcg_temp_new_i64();
7176
        TCGv_i64 tcg_op2 = tcg_temp_new_i64();
7177
        TCGv_i64 tcg_wideres = tcg_temp_new_i64();
7178
        static NeonGenNarrowFn * const narrowfns[3][2] = {
7179
            { gen_helper_neon_narrow_high_u8,
7180
              gen_helper_neon_narrow_round_high_u8 },
7181
            { gen_helper_neon_narrow_high_u16,
7182
              gen_helper_neon_narrow_round_high_u16 },
7183
            { do_narrow_high_u32, do_narrow_round_high_u32 },
7184
        };
7185
        NeonGenNarrowFn *gennarrow = narrowfns[size][is_u];
7186

    
7187
        read_vec_element(s, tcg_op1, rn, pass, MO_64);
7188
        read_vec_element(s, tcg_op2, rm, pass, MO_64);
7189

    
7190
        gen_neon_addl(size, (opcode == 6), tcg_wideres, tcg_op1, tcg_op2);
7191

    
7192
        tcg_temp_free_i64(tcg_op1);
7193
        tcg_temp_free_i64(tcg_op2);
7194

    
7195
        tcg_res[pass] = tcg_temp_new_i32();
7196
        gennarrow(tcg_res[pass], tcg_wideres);
7197
        tcg_temp_free_i64(tcg_wideres);
7198
    }
7199

    
7200
    for (pass = 0; pass < 2; pass++) {
7201
        write_vec_element_i32(s, tcg_res[pass], rd, pass + part, MO_32);
7202
        tcg_temp_free_i32(tcg_res[pass]);
7203
    }
7204
    if (!is_q) {
7205
        clear_vec_high(s, rd);
7206
    }
7207
}
7208

    
7209
/* C3.6.15 AdvSIMD three different
7210
 *   31  30  29 28       24 23  22  21 20  16 15    12 11 10 9    5 4    0
7211
 * +---+---+---+-----------+------+---+------+--------+-----+------+------+
7212
 * | 0 | Q | U | 0 1 1 1 0 | size | 1 |  Rm  | opcode | 0 0 |  Rn  |  Rd  |
7213
 * +---+---+---+-----------+------+---+------+--------+-----+------+------+
7214
 */
7215
static void disas_simd_three_reg_diff(DisasContext *s, uint32_t insn)
7216
{
7217
    /* Instructions in this group fall into three basic classes
7218
     * (in each case with the operation working on each element in
7219
     * the input vectors):
7220
     * (1) widening 64 x 64 -> 128 (with possibly Vd as an extra
7221
     *     128 bit input)
7222
     * (2) wide 64 x 128 -> 128
7223
     * (3) narrowing 128 x 128 -> 64
7224
     * Here we do initial decode, catch unallocated cases and
7225
     * dispatch to separate functions for each class.
7226
     */
7227
    int is_q = extract32(insn, 30, 1);
7228
    int is_u = extract32(insn, 29, 1);
7229
    int size = extract32(insn, 22, 2);
7230
    int opcode = extract32(insn, 12, 4);
7231
    int rm = extract32(insn, 16, 5);
7232
    int rn = extract32(insn, 5, 5);
7233
    int rd = extract32(insn, 0, 5);
7234

    
7235
    switch (opcode) {
7236
    case 1: /* SADDW, SADDW2, UADDW, UADDW2 */
7237
    case 3: /* SSUBW, SSUBW2, USUBW, USUBW2 */
7238
        /* 64 x 128 -> 128 */
7239
        if (size == 3) {
7240
            unallocated_encoding(s);
7241
            return;
7242
        }
7243
        handle_3rd_wide(s, is_q, is_u, size, opcode, rd, rn, rm);
7244
        break;
7245
    case 4: /* ADDHN, ADDHN2, RADDHN, RADDHN2 */
7246
    case 6: /* SUBHN, SUBHN2, RSUBHN, RSUBHN2 */
7247
        /* 128 x 128 -> 64 */
7248
        if (size == 3) {
7249
            unallocated_encoding(s);
7250
            return;
7251
        }
7252
        handle_3rd_narrowing(s, is_q, is_u, size, opcode, rd, rn, rm);
7253
        break;
7254
    case 14: /* PMULL, PMULL2 */
7255
        if (is_u || size == 1 || size == 2) {
7256
            unallocated_encoding(s);
7257
            return;
7258
        }
7259
        unsupported_encoding(s, insn);
7260
        break;
7261
    case 9: /* SQDMLAL, SQDMLAL2 */
7262
    case 11: /* SQDMLSL, SQDMLSL2 */
7263
    case 13: /* SQDMULL, SQDMULL2 */
7264
        if (is_u || size == 0) {
7265
            unallocated_encoding(s);
7266
            return;
7267
        }
7268
        /* fall through */
7269
    case 0: /* SADDL, SADDL2, UADDL, UADDL2 */
7270
    case 2: /* SSUBL, SSUBL2, USUBL, USUBL2 */
7271
    case 5: /* SABAL, SABAL2, UABAL, UABAL2 */
7272
    case 7: /* SABDL, SABDL2, UABDL, UABDL2 */
7273
    case 8: /* SMLAL, SMLAL2, UMLAL, UMLAL2 */
7274
    case 10: /* SMLSL, SMLSL2, UMLSL, UMLSL2 */
7275
    case 12: /* SMULL, SMULL2, UMULL, UMULL2 */
7276
        /* 64 x 64 -> 128 */
7277
        if (size == 3) {
7278
            unallocated_encoding(s);
7279
            return;
7280
        }
7281
        handle_3rd_widening(s, is_q, is_u, size, opcode, rd, rn, rm);
7282
        break;
7283
    default:
7284
        /* opcode 15 not allocated */
7285
        unallocated_encoding(s);
7286
        break;
7287
    }
7288
}
7289

    
7290
/* Logic op (opcode == 3) subgroup of C3.6.16. */
7291
static void disas_simd_3same_logic(DisasContext *s, uint32_t insn)
7292
{
7293
    int rd = extract32(insn, 0, 5);
7294
    int rn = extract32(insn, 5, 5);
7295
    int rm = extract32(insn, 16, 5);
7296
    int size = extract32(insn, 22, 2);
7297
    bool is_u = extract32(insn, 29, 1);
7298
    bool is_q = extract32(insn, 30, 1);
7299
    TCGv_i64 tcg_op1 = tcg_temp_new_i64();
7300
    TCGv_i64 tcg_op2 = tcg_temp_new_i64();
7301
    TCGv_i64 tcg_res[2];
7302
    int pass;
7303

    
7304
    tcg_res[0] = tcg_temp_new_i64();
7305
    tcg_res[1] = tcg_temp_new_i64();
7306

    
7307
    for (pass = 0; pass < (is_q ? 2 : 1); pass++) {
7308
        read_vec_element(s, tcg_op1, rn, pass, MO_64);
7309
        read_vec_element(s, tcg_op2, rm, pass, MO_64);
7310

    
7311
        if (!is_u) {
7312
            switch (size) {
7313
            case 0: /* AND */
7314
                tcg_gen_and_i64(tcg_res[pass], tcg_op1, tcg_op2);
7315
                break;
7316
            case 1: /* BIC */
7317
                tcg_gen_andc_i64(tcg_res[pass], tcg_op1, tcg_op2);
7318
                break;
7319
            case 2: /* ORR */
7320
                tcg_gen_or_i64(tcg_res[pass], tcg_op1, tcg_op2);
7321
                break;
7322
            case 3: /* ORN */
7323
                tcg_gen_orc_i64(tcg_res[pass], tcg_op1, tcg_op2);
7324
                break;
7325
            }
7326
        } else {
7327
            if (size != 0) {
7328
                /* B* ops need res loaded to operate on */
7329
                read_vec_element(s, tcg_res[pass], rd, pass, MO_64);
7330
            }
7331

    
7332
            switch (size) {
7333
            case 0: /* EOR */
7334
                tcg_gen_xor_i64(tcg_res[pass], tcg_op1, tcg_op2);
7335
                break;
7336
            case 1: /* BSL bitwise select */
7337
                tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_op2);
7338
                tcg_gen_and_i64(tcg_op1, tcg_op1, tcg_res[pass]);
7339
                tcg_gen_xor_i64(tcg_res[pass], tcg_op2, tcg_op1);
7340
                break;
7341
            case 2: /* BIT, bitwise insert if true */
7342
                tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_res[pass]);
7343
                tcg_gen_and_i64(tcg_op1, tcg_op1, tcg_op2);
7344
                tcg_gen_xor_i64(tcg_res[pass], tcg_res[pass], tcg_op1);
7345
                break;
7346
            case 3: /* BIF, bitwise insert if false */
7347
                tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_res[pass]);
7348
                tcg_gen_andc_i64(tcg_op1, tcg_op1, tcg_op2);
7349
                tcg_gen_xor_i64(tcg_res[pass], tcg_res[pass], tcg_op1);
7350
                break;
7351
            }
7352
        }
7353
    }
7354

    
7355
    write_vec_element(s, tcg_res[0], rd, 0, MO_64);
7356
    if (!is_q) {
7357
        tcg_gen_movi_i64(tcg_res[1], 0);
7358
    }
7359
    write_vec_element(s, tcg_res[1], rd, 1, MO_64);
7360

    
7361
    tcg_temp_free_i64(tcg_op1);
7362
    tcg_temp_free_i64(tcg_op2);
7363
    tcg_temp_free_i64(tcg_res[0]);
7364
    tcg_temp_free_i64(tcg_res[1]);
7365
}
7366

    
7367
/* Helper functions for 32 bit comparisons */
7368
static void gen_max_s32(TCGv_i32 res, TCGv_i32 op1, TCGv_i32 op2)
7369
{
7370
    tcg_gen_movcond_i32(TCG_COND_GE, res, op1, op2, op1, op2);
7371
}
7372

    
7373
static void gen_max_u32(TCGv_i32 res, TCGv_i32 op1, TCGv_i32 op2)
7374
{
7375
    tcg_gen_movcond_i32(TCG_COND_GEU, res, op1, op2, op1, op2);
7376
}
7377

    
7378
static void gen_min_s32(TCGv_i32 res, TCGv_i32 op1, TCGv_i32 op2)
7379
{
7380
    tcg_gen_movcond_i32(TCG_COND_LE, res, op1, op2, op1, op2);
7381
}
7382

    
7383
static void gen_min_u32(TCGv_i32 res, TCGv_i32 op1, TCGv_i32 op2)
7384
{
7385
    tcg_gen_movcond_i32(TCG_COND_LEU, res, op1, op2, op1, op2);
7386
}
7387

    
7388
/* Pairwise op subgroup of C3.6.16.
7389
 *
7390
 * This is called directly or via the handle_3same_float for float pairwise
7391
 * operations where the opcode and size are calculated differently.
7392
 */
7393
static void handle_simd_3same_pair(DisasContext *s, int is_q, int u, int opcode,
7394
                                   int size, int rn, int rm, int rd)
7395
{
7396
    TCGv_ptr fpst;
7397
    int pass;
7398

    
7399
    /* Floating point operations need fpst */
7400
    if (opcode >= 0x58) {
7401
        fpst = get_fpstatus_ptr();
7402
    } else {
7403
        TCGV_UNUSED_PTR(fpst);
7404
    }
7405

    
7406
    /* These operations work on the concatenated rm:rn, with each pair of
7407
     * adjacent elements being operated on to produce an element in the result.
7408
     */
7409
    if (size == 3) {
7410
        TCGv_i64 tcg_res[2];
7411

    
7412
        for (pass = 0; pass < 2; pass++) {
7413
            TCGv_i64 tcg_op1 = tcg_temp_new_i64();
7414
            TCGv_i64 tcg_op2 = tcg_temp_new_i64();
7415
            int passreg = (pass == 0) ? rn : rm;
7416

    
7417
            read_vec_element(s, tcg_op1, passreg, 0, MO_64);
7418
            read_vec_element(s, tcg_op2, passreg, 1, MO_64);
7419
            tcg_res[pass] = tcg_temp_new_i64();
7420

    
7421
            switch (opcode) {
7422
            case 0x17: /* ADDP */
7423
                tcg_gen_add_i64(tcg_res[pass], tcg_op1, tcg_op2);
7424
                break;
7425
            case 0x58: /* FMAXNMP */
7426
                gen_helper_vfp_maxnumd(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7427
                break;
7428
            case 0x5a: /* FADDP */
7429
                gen_helper_vfp_addd(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7430
                break;
7431
            case 0x5e: /* FMAXP */
7432
                gen_helper_vfp_maxd(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7433
                break;
7434
            case 0x78: /* FMINNMP */
7435
                gen_helper_vfp_minnumd(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7436
                break;
7437
            case 0x7e: /* FMINP */
7438
                gen_helper_vfp_mind(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7439
                break;
7440
            default:
7441
                g_assert_not_reached();
7442
            }
7443

    
7444
            tcg_temp_free_i64(tcg_op1);
7445
            tcg_temp_free_i64(tcg_op2);
7446
        }
7447

    
7448
        for (pass = 0; pass < 2; pass++) {
7449
            write_vec_element(s, tcg_res[pass], rd, pass, MO_64);
7450
            tcg_temp_free_i64(tcg_res[pass]);
7451
        }
7452
    } else {
7453
        int maxpass = is_q ? 4 : 2;
7454
        TCGv_i32 tcg_res[4];
7455

    
7456
        for (pass = 0; pass < maxpass; pass++) {
7457
            TCGv_i32 tcg_op1 = tcg_temp_new_i32();
7458
            TCGv_i32 tcg_op2 = tcg_temp_new_i32();
7459
            NeonGenTwoOpFn *genfn = NULL;
7460
            int passreg = pass < (maxpass / 2) ? rn : rm;
7461
            int passelt = (is_q && (pass & 1)) ? 2 : 0;
7462

    
7463
            read_vec_element_i32(s, tcg_op1, passreg, passelt, MO_32);
7464
            read_vec_element_i32(s, tcg_op2, passreg, passelt + 1, MO_32);
7465
            tcg_res[pass] = tcg_temp_new_i32();
7466

    
7467
            switch (opcode) {
7468
            case 0x17: /* ADDP */
7469
            {
7470
                static NeonGenTwoOpFn * const fns[3] = {
7471
                    gen_helper_neon_padd_u8,
7472
                    gen_helper_neon_padd_u16,
7473
                    tcg_gen_add_i32,
7474
                };
7475
                genfn = fns[size];
7476
                break;
7477
            }
7478
            case 0x14: /* SMAXP, UMAXP */
7479
            {
7480
                static NeonGenTwoOpFn * const fns[3][2] = {
7481
                    { gen_helper_neon_pmax_s8, gen_helper_neon_pmax_u8 },
7482
                    { gen_helper_neon_pmax_s16, gen_helper_neon_pmax_u16 },
7483
                    { gen_max_s32, gen_max_u32 },
7484
                };
7485
                genfn = fns[size][u];
7486
                break;
7487
            }
7488
            case 0x15: /* SMINP, UMINP */
7489
            {
7490
                static NeonGenTwoOpFn * const fns[3][2] = {
7491
                    { gen_helper_neon_pmin_s8, gen_helper_neon_pmin_u8 },
7492
                    { gen_helper_neon_pmin_s16, gen_helper_neon_pmin_u16 },
7493
                    { gen_min_s32, gen_min_u32 },
7494
                };
7495
                genfn = fns[size][u];
7496
                break;
7497
            }
7498
            /* The FP operations are all on single floats (32 bit) */
7499
            case 0x58: /* FMAXNMP */
7500
                gen_helper_vfp_maxnums(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7501
                break;
7502
            case 0x5a: /* FADDP */
7503
                gen_helper_vfp_adds(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7504
                break;
7505
            case 0x5e: /* FMAXP */
7506
                gen_helper_vfp_maxs(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7507
                break;
7508
            case 0x78: /* FMINNMP */
7509
                gen_helper_vfp_minnums(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7510
                break;
7511
            case 0x7e: /* FMINP */
7512
                gen_helper_vfp_mins(tcg_res[pass], tcg_op1, tcg_op2, fpst);
7513
                break;
7514
            default:
7515
                g_assert_not_reached();
7516
            }
7517

    
7518
            /* FP ops called directly, otherwise call now */
7519
            if (genfn) {
7520
                genfn(tcg_res[pass], tcg_op1, tcg_op2);
7521
            }
7522

    
7523
            tcg_temp_free_i32(tcg_op1);
7524
            tcg_temp_free_i32(tcg_op2);
7525
        }
7526

    
7527
        for (pass = 0; pass < maxpass; pass++) {
7528
            write_vec_element_i32(s, tcg_res[pass], rd, pass, MO_32);
7529
            tcg_temp_free_i32(tcg_res[pass]);
7530
        }
7531
        if (!is_q) {
7532
            clear_vec_high(s, rd);
7533
        }
7534
    }
7535

    
7536
    if (!TCGV_IS_UNUSED_PTR(fpst)) {
7537
        tcg_temp_free_ptr(fpst);
7538
    }
7539
}
7540

    
7541
/* Floating point op subgroup of C3.6.16. */
7542
static void disas_simd_3same_float(DisasContext *s, uint32_t insn)
7543
{
7544
    /* For floating point ops, the U, size[1] and opcode bits
7545
     * together indicate the operation. size[0] indicates single
7546
     * or double.
7547
     */
7548
    int fpopcode = extract32(insn, 11, 5)
7549
        | (extract32(insn, 23, 1) << 5)
7550
        | (extract32(insn, 29, 1) << 6);
7551
    int is_q = extract32(insn, 30, 1);
7552
    int size = extract32(insn, 22, 1);
7553
    int rm = extract32(insn, 16, 5);
7554
    int rn = extract32(insn, 5, 5);
7555
    int rd = extract32(insn, 0, 5);
7556

    
7557
    int datasize = is_q ? 128 : 64;
7558
    int esize = 32 << size;
7559
    int elements = datasize / esize;
7560

    
7561
    if (size == 1 && !is_q) {
7562
        unallocated_encoding(s);
7563
        return;
7564
    }
7565

    
7566
    switch (fpopcode) {
7567
    case 0x58: /* FMAXNMP */
7568
    case 0x5a: /* FADDP */
7569
    case 0x5e: /* FMAXP */
7570
    case 0x78: /* FMINNMP */
7571
    case 0x7e: /* FMINP */
7572
        if (size && !is_q) {
7573
            unallocated_encoding(s);
7574
            return;
7575
        }
7576
        handle_simd_3same_pair(s, is_q, 0, fpopcode, size ? MO_64 : MO_32,
7577
                               rn, rm, rd);
7578
        return;
7579
    case 0x1b: /* FMULX */
7580
    case 0x1f: /* FRECPS */
7581
    case 0x3f: /* FRSQRTS */
7582
    case 0x5d: /* FACGE */
7583
    case 0x7d: /* FACGT */
7584
    case 0x19: /* FMLA */
7585
    case 0x39: /* FMLS */
7586
    case 0x18: /* FMAXNM */
7587
    case 0x1a: /* FADD */
7588
    case 0x1c: /* FCMEQ */
7589
    case 0x1e: /* FMAX */
7590
    case 0x38: /* FMINNM */
7591
    case 0x3a: /* FSUB */
7592
    case 0x3e: /* FMIN */
7593
    case 0x5b: /* FMUL */
7594
    case 0x5c: /* FCMGE */
7595
    case 0x5f: /* FDIV */
7596
    case 0x7a: /* FABD */
7597
    case 0x7c: /* FCMGT */
7598
        handle_3same_float(s, size, elements, fpopcode, rd, rn, rm);
7599
        return;
7600
    default:
7601
        unallocated_encoding(s);
7602
        return;
7603
    }
7604
}
7605

    
7606
/* Integer op subgroup of C3.6.16. */
7607
static void disas_simd_3same_int(DisasContext *s, uint32_t insn)
7608
{
7609
    int is_q = extract32(insn, 30, 1);
7610
    int u = extract32(insn, 29, 1);
7611
    int size = extract32(insn, 22, 2);
7612
    int opcode = extract32(insn, 11, 5);
7613
    int rm = extract32(insn, 16, 5);
7614
    int rn = extract32(insn, 5, 5);
7615
    int rd = extract32(insn, 0, 5);
7616
    int pass;
7617

    
7618
    switch (opcode) {
7619
    case 0x13: /* MUL, PMUL */
7620
        if (u && size != 0) {
7621
            unallocated_encoding(s);
7622
            return;
7623
        }
7624
        /* fall through */
7625
    case 0x0: /* SHADD, UHADD */
7626
    case 0x2: /* SRHADD, URHADD */
7627
    case 0x4: /* SHSUB, UHSUB */
7628
    case 0xc: /* SMAX, UMAX */
7629
    case 0xd: /* SMIN, UMIN */
7630
    case 0xe: /* SABD, UABD */
7631
    case 0xf: /* SABA, UABA */
7632
    case 0x12: /* MLA, MLS */
7633
        if (size == 3) {
7634
            unallocated_encoding(s);
7635
            return;
7636
        }
7637
        break;
7638
    case 0x16: /* SQDMULH, SQRDMULH */
7639
        if (size == 0 || size == 3) {
7640
            unallocated_encoding(s);
7641
            return;
7642
        }
7643
        break;
7644
    default:
7645
        if (size == 3 && !is_q) {
7646
            unallocated_encoding(s);
7647
            return;
7648
        }
7649
        break;
7650
    }
7651

    
7652
    if (size == 3) {
7653
        for (pass = 0; pass < (is_q ? 2 : 1); pass++) {
7654
            TCGv_i64 tcg_op1 = tcg_temp_new_i64();
7655
            TCGv_i64 tcg_op2 = tcg_temp_new_i64();
7656
            TCGv_i64 tcg_res = tcg_temp_new_i64();
7657

    
7658
            read_vec_element(s, tcg_op1, rn, pass, MO_64);
7659
            read_vec_element(s, tcg_op2, rm, pass, MO_64);
7660

    
7661
            handle_3same_64(s, opcode, u, tcg_res, tcg_op1, tcg_op2);
7662

    
7663
            write_vec_element(s, tcg_res, rd, pass, MO_64);
7664

    
7665
            tcg_temp_free_i64(tcg_res);
7666
            tcg_temp_free_i64(tcg_op1);
7667
            tcg_temp_free_i64(tcg_op2);
7668
        }
7669
    } else {
7670
        for (pass = 0; pass < (is_q ? 4 : 2); pass++) {
7671
            TCGv_i32 tcg_op1 = tcg_temp_new_i32();
7672
            TCGv_i32 tcg_op2 = tcg_temp_new_i32();
7673
            TCGv_i32 tcg_res = tcg_temp_new_i32();
7674
            NeonGenTwoOpFn *genfn = NULL;
7675
            NeonGenTwoOpEnvFn *genenvfn = NULL;
7676

    
7677
            read_vec_element_i32(s, tcg_op1, rn, pass, MO_32);
7678
            read_vec_element_i32(s, tcg_op2, rm, pass, MO_32);
7679

    
7680
            switch (opcode) {
7681
            case 0x0: /* SHADD, UHADD */
7682
            {
7683
                static NeonGenTwoOpFn * const fns[3][2] = {
7684
                    { gen_helper_neon_hadd_s8, gen_helper_neon_hadd_u8 },
7685
                    { gen_helper_neon_hadd_s16, gen_helper_neon_hadd_u16 },
7686
                    { gen_helper_neon_hadd_s32, gen_helper_neon_hadd_u32 },
7687
                };
7688
                genfn = fns[size][u];
7689
                break;
7690
            }
7691
            case 0x1: /* SQADD, UQADD */
7692
            {
7693
                static NeonGenTwoOpEnvFn * const fns[3][2] = {
7694
                    { gen_helper_neon_qadd_s8, gen_helper_neon_qadd_u8 },
7695
                    { gen_helper_neon_qadd_s16, gen_helper_neon_qadd_u16 },
7696
                    { gen_helper_neon_qadd_s32, gen_helper_neon_qadd_u32 },
7697
                };
7698
                genenvfn = fns[size][u];
7699
                break;
7700
            }
7701
            case 0x2: /* SRHADD, URHADD */
7702
            {
7703
                static NeonGenTwoOpFn * const fns[3][2] = {
7704
                    { gen_helper_neon_rhadd_s8, gen_helper_neon_rhadd_u8 },
7705
                    { gen_helper_neon_rhadd_s16, gen_helper_neon_rhadd_u16 },
7706
                    { gen_helper_neon_rhadd_s32, gen_helper_neon_rhadd_u32 },
7707
                };
7708
                genfn = fns[size][u];
7709
                break;
7710
            }
7711
            case 0x4: /* SHSUB, UHSUB */
7712
            {
7713
                static NeonGenTwoOpFn * const fns[3][2] = {
7714
                    { gen_helper_neon_hsub_s8, gen_helper_neon_hsub_u8 },
7715
                    { gen_helper_neon_hsub_s16, gen_helper_neon_hsub_u16 },
7716
                    { gen_helper_neon_hsub_s32, gen_helper_neon_hsub_u32 },
7717
                };
7718
                genfn = fns[size][u];
7719
                break;
7720
            }
7721
            case 0x5: /* SQSUB, UQSUB */
7722
            {
7723
                static NeonGenTwoOpEnvFn * const fns[3][2] = {
7724
                    { gen_helper_neon_qsub_s8, gen_helper_neon_qsub_u8 },
7725
                    { gen_helper_neon_qsub_s16, gen_helper_neon_qsub_u16 },
7726
                    { gen_helper_neon_qsub_s32, gen_helper_neon_qsub_u32 },
7727
                };
7728
                genenvfn = fns[size][u];
7729
                break;
7730
            }
7731
            case 0x6: /* CMGT, CMHI */
7732
            {
7733
                static NeonGenTwoOpFn * const fns[3][2] = {
7734
                    { gen_helper_neon_cgt_s8, gen_helper_neon_cgt_u8 },
7735
                    { gen_helper_neon_cgt_s16, gen_helper_neon_cgt_u16 },
7736
                    { gen_helper_neon_cgt_s32, gen_helper_neon_cgt_u32 },
7737
                };
7738
                genfn = fns[size][u];
7739
                break;
7740
            }
7741
            case 0x7: /* CMGE, CMHS */
7742
            {
7743
                static NeonGenTwoOpFn * const fns[3][2] = {
7744
                    { gen_helper_neon_cge_s8, gen_helper_neon_cge_u8 },
7745
                    { gen_helper_neon_cge_s16, gen_helper_neon_cge_u16 },
7746
                    { gen_helper_neon_cge_s32, gen_helper_neon_cge_u32 },
7747
                };
7748
                genfn = fns[size][u];
7749
                break;
7750
            }
7751
            case 0x8: /* SSHL, USHL */
7752
            {
7753
                static NeonGenTwoOpFn * const fns[3][2] = {
7754
                    { gen_helper_neon_shl_s8, gen_helper_neon_shl_u8 },
7755
                    { gen_helper_neon_shl_s16, gen_helper_neon_shl_u16 },
7756
                    { gen_helper_neon_shl_s32, gen_helper_neon_shl_u32 },
7757
                };
7758
                genfn = fns[size][u];
7759
                break;
7760
            }
7761
            case 0x9: /* SQSHL, UQSHL */
7762
            {
7763
                static NeonGenTwoOpEnvFn * const fns[3][2] = {
7764
                    { gen_helper_neon_qshl_s8, gen_helper_neon_qshl_u8 },
7765
                    { gen_helper_neon_qshl_s16, gen_helper_neon_qshl_u16 },
7766
                    { gen_helper_neon_qshl_s32, gen_helper_neon_qshl_u32 },
7767
                };
7768
                genenvfn = fns[size][u];
7769
                break;
7770
            }
7771
            case 0xa: /* SRSHL, URSHL */
7772
            {
7773
                static NeonGenTwoOpFn * const fns[3][2] = {
7774
                    { gen_helper_neon_rshl_s8, gen_helper_neon_rshl_u8 },
7775
                    { gen_helper_neon_rshl_s16, gen_helper_neon_rshl_u16 },
7776
                    { gen_helper_neon_rshl_s32, gen_helper_neon_rshl_u32 },
7777
                };
7778
                genfn = fns[size][u];
7779
                break;
7780
            }
7781
            case 0xb: /* SQRSHL, UQRSHL */
7782
            {
7783
                static NeonGenTwoOpEnvFn * const fns[3][2] = {
7784
                    { gen_helper_neon_qrshl_s8, gen_helper_neon_qrshl_u8 },
7785
                    { gen_helper_neon_qrshl_s16, gen_helper_neon_qrshl_u16 },
7786
                    { gen_helper_neon_qrshl_s32, gen_helper_neon_qrshl_u32 },
7787
                };
7788
                genenvfn = fns[size][u];
7789
                break;
7790
            }
7791
            case 0xc: /* SMAX, UMAX */
7792
            {
7793
                static NeonGenTwoOpFn * const fns[3][2] = {
7794
                    { gen_helper_neon_max_s8, gen_helper_neon_max_u8 },
7795
                    { gen_helper_neon_max_s16, gen_helper_neon_max_u16 },
7796
                    { gen_max_s32, gen_max_u32 },
7797
                };
7798
                genfn = fns[size][u];
7799
                break;
7800
            }
7801

    
7802
            case 0xd: /* SMIN, UMIN */
7803
            {
7804
                static NeonGenTwoOpFn * const fns[3][2] = {
7805
                    { gen_helper_neon_min_s8, gen_helper_neon_min_u8 },
7806
                    { gen_helper_neon_min_s16, gen_helper_neon_min_u16 },
7807
                    { gen_min_s32, gen_min_u32 },
7808
                };
7809
                genfn = fns[size][u];
7810
                break;
7811
            }
7812
            case 0xe: /* SABD, UABD */
7813
            case 0xf: /* SABA, UABA */
7814
            {
7815
                static NeonGenTwoOpFn * const fns[3][2] = {
7816
                    { gen_helper_neon_abd_s8, gen_helper_neon_abd_u8 },
7817
                    { gen_helper_neon_abd_s16, gen_helper_neon_abd_u16 },
7818
                    { gen_helper_neon_abd_s32, gen_helper_neon_abd_u32 },
7819
                };
7820
                genfn = fns[size][u];
7821
                break;
7822
            }
7823
            case 0x10: /* ADD, SUB */
7824
            {
7825
                static NeonGenTwoOpFn * const fns[3][2] = {
7826
                    { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 },
7827
                    { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 },
7828
                    { tcg_gen_add_i32, tcg_gen_sub_i32 },
7829
                };
7830
                genfn = fns[size][u];
7831
                break;
7832
            }
7833
            case 0x11: /* CMTST, CMEQ */
7834
            {
7835
                static NeonGenTwoOpFn * const fns[3][2] = {
7836
                    { gen_helper_neon_tst_u8, gen_helper_neon_ceq_u8 },
7837
                    { gen_helper_neon_tst_u16, gen_helper_neon_ceq_u16 },
7838
                    { gen_helper_neon_tst_u32, gen_helper_neon_ceq_u32 },
7839
                };
7840
                genfn = fns[size][u];
7841
                break;
7842
            }
7843
            case 0x13: /* MUL, PMUL */
7844
                if (u) {
7845
                    /* PMUL */
7846
                    assert(size == 0);
7847
                    genfn = gen_helper_neon_mul_p8;
7848
                    break;
7849
                }
7850
                /* fall through : MUL */
7851
            case 0x12: /* MLA, MLS */
7852
            {
7853
                static NeonGenTwoOpFn * const fns[3] = {
7854
                    gen_helper_neon_mul_u8,
7855
                    gen_helper_neon_mul_u16,
7856
                    tcg_gen_mul_i32,
7857
                };
7858
                genfn = fns[size];
7859
                break;
7860
            }
7861
            case 0x16: /* SQDMULH, SQRDMULH */
7862
            {
7863
                static NeonGenTwoOpEnvFn * const fns[2][2] = {
7864
                    { gen_helper_neon_qdmulh_s16, gen_helper_neon_qrdmulh_s16 },
7865
                    { gen_helper_neon_qdmulh_s32, gen_helper_neon_qrdmulh_s32 },
7866
                };
7867
                assert(size == 1 || size == 2);
7868
                genenvfn = fns[size - 1][u];
7869
                break;
7870
            }
7871
            default:
7872
                g_assert_not_reached();
7873
            }
7874

    
7875
            if (genenvfn) {
7876
                genenvfn(tcg_res, cpu_env, tcg_op1, tcg_op2);
7877
            } else {
7878
                genfn(tcg_res, tcg_op1, tcg_op2);
7879
            }
7880

    
7881
            if (opcode == 0xf || opcode == 0x12) {
7882
                /* SABA, UABA, MLA, MLS: accumulating ops */
7883
                static NeonGenTwoOpFn * const fns[3][2] = {
7884
                    { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 },
7885
                    { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 },
7886
                    { tcg_gen_add_i32, tcg_gen_sub_i32 },
7887
                };
7888
                bool is_sub = (opcode == 0x12 && u); /* MLS */
7889

    
7890
                genfn = fns[size][is_sub];
7891
                read_vec_element_i32(s, tcg_op1, rd, pass, MO_32);
7892
                genfn(tcg_res, tcg_res, tcg_op1);
7893
            }
7894

    
7895
            write_vec_element_i32(s, tcg_res, rd, pass, MO_32);
7896

    
7897
            tcg_temp_free_i32(tcg_res);
7898
            tcg_temp_free_i32(tcg_op1);
7899
            tcg_temp_free_i32(tcg_op2);
7900
        }
7901
    }
7902

    
7903
    if (!is_q) {
7904
        clear_vec_high(s, rd);
7905
    }
7906
}
7907

    
7908
/* C3.6.16 AdvSIMD three same
7909
 *  31  30  29  28       24 23  22  21 20  16 15    11  10 9    5 4    0
7910
 * +---+---+---+-----------+------+---+------+--------+---+------+------+
7911
 * | 0 | Q | U | 0 1 1 1 0 | size | 1 |  Rm  | opcode | 1 |  Rn  |  Rd  |
7912
 * +---+---+---+-----------+------+---+------+--------+---+------+------+
7913
 */
7914
static void disas_simd_three_reg_same(DisasContext *s, uint32_t insn)
7915
{
7916
    int opcode = extract32(insn, 11, 5);
7917

    
7918
    switch (opcode) {
7919
    case 0x3: /* logic ops */
7920
        disas_simd_3same_logic(s, insn);
7921
        break;
7922
    case 0x17: /* ADDP */
7923
    case 0x14: /* SMAXP, UMAXP */
7924
    case 0x15: /* SMINP, UMINP */
7925
    {
7926
        /* Pairwise operations */
7927
        int is_q = extract32(insn, 30, 1);
7928
        int u = extract32(insn, 29, 1);
7929
        int size = extract32(insn, 22, 2);
7930
        int rm = extract32(insn, 16, 5);
7931
        int rn = extract32(insn, 5, 5);
7932
        int rd = extract32(insn, 0, 5);
7933
        if (opcode == 0x17) {
7934
            if (u || (size == 3 && !is_q)) {
7935
                unallocated_encoding(s);
7936
                return;
7937
            }
7938
        } else {
7939
            if (size == 3) {
7940
                unallocated_encoding(s);
7941
                return;
7942
            }
7943
        }
7944
        handle_simd_3same_pair(s, is_q, u, opcode, size, rn, rm, rd);
7945
        break;
7946
    }
7947
    case 0x18 ... 0x31:
7948
        /* floating point ops, sz[1] and U are part of opcode */
7949
        disas_simd_3same_float(s, insn);
7950
        break;
7951
    default:
7952
        disas_simd_3same_int(s, insn);
7953
        break;
7954
    }
7955
}
7956

    
7957
static void handle_2misc_narrow(DisasContext *s, int opcode, bool u, bool is_q,
7958
                                int size, int rn, int rd)
7959
{
7960
    /* Handle 2-reg-misc ops which are narrowing (so each 2*size element
7961
     * in the source becomes a size element in the destination).
7962
     */
7963
    int pass;
7964
    TCGv_i32 tcg_res[2];
7965
    int destelt = is_q ? 2 : 0;
7966

    
7967
    for (pass = 0; pass < 2; pass++) {
7968
        TCGv_i64 tcg_op = tcg_temp_new_i64();
7969
        NeonGenNarrowFn *genfn = NULL;
7970
        NeonGenNarrowEnvFn *genenvfn = NULL;
7971

    
7972
        read_vec_element(s, tcg_op, rn, pass, MO_64);
7973
        tcg_res[pass] = tcg_temp_new_i32();
7974

    
7975
        switch (opcode) {
7976
        case 0x12: /* XTN, SQXTUN */
7977
        {
7978
            static NeonGenNarrowFn * const xtnfns[3] = {
7979
                gen_helper_neon_narrow_u8,
7980
                gen_helper_neon_narrow_u16,
7981
                tcg_gen_trunc_i64_i32,
7982
            };
7983
            static NeonGenNarrowEnvFn * const sqxtunfns[3] = {
7984
                gen_helper_neon_unarrow_sat8,
7985
                gen_helper_neon_unarrow_sat16,
7986
                gen_helper_neon_unarrow_sat32,
7987
            };
7988
            if (u) {
7989
                genenvfn = sqxtunfns[size];
7990
            } else {
7991
                genfn = xtnfns[size];
7992
            }
7993
            break;
7994
        }
7995
        case 0x14: /* SQXTN, UQXTN */
7996
        {
7997
            static NeonGenNarrowEnvFn * const fns[3][2] = {
7998
                { gen_helper_neon_narrow_sat_s8,
7999
                  gen_helper_neon_narrow_sat_u8 },
8000
                { gen_helper_neon_narrow_sat_s16,
8001
                  gen_helper_neon_narrow_sat_u16 },
8002
                { gen_helper_neon_narrow_sat_s32,
8003
                  gen_helper_neon_narrow_sat_u32 },
8004
            };
8005
            genenvfn = fns[size][u];
8006
            break;
8007
        }
8008
        default:
8009
            g_assert_not_reached();
8010
        }
8011

    
8012
        if (genfn) {
8013
            genfn(tcg_res[pass], tcg_op);
8014
        } else {
8015
            genenvfn(tcg_res[pass], cpu_env, tcg_op);
8016
        }
8017

    
8018
        tcg_temp_free_i64(tcg_op);
8019
    }
8020

    
8021
    for (pass = 0; pass < 2; pass++) {
8022
        write_vec_element_i32(s, tcg_res[pass], rd, destelt + pass, MO_32);
8023
        tcg_temp_free_i32(tcg_res[pass]);
8024
    }
8025
    if (!is_q) {
8026
        clear_vec_high(s, rd);
8027
    }
8028
}
8029

    
8030
static void handle_rev(DisasContext *s, int opcode, bool u,
8031
                       bool is_q, int size, int rn, int rd)
8032
{
8033
    int op = (opcode << 1) | u;
8034
    int opsz = op + size;
8035
    int grp_size = 3 - opsz;
8036
    int dsize = is_q ? 128 : 64;
8037
    int i;
8038

    
8039
    if (opsz >= 3) {
8040
        unallocated_encoding(s);
8041
        return;
8042
    }
8043

    
8044
    if (size == 0) {
8045
        /* Special case bytes, use bswap op on each group of elements */
8046
        int groups = dsize / (8 << grp_size);
8047

    
8048
        for (i = 0; i < groups; i++) {
8049
            TCGv_i64 tcg_tmp = tcg_temp_new_i64();
8050

    
8051
            read_vec_element(s, tcg_tmp, rn, i, grp_size);
8052
            switch (grp_size) {
8053
            case MO_16:
8054
                tcg_gen_bswap16_i64(tcg_tmp, tcg_tmp);
8055
                break;
8056
            case MO_32:
8057
                tcg_gen_bswap32_i64(tcg_tmp, tcg_tmp);
8058
                break;
8059
            case MO_64:
8060
                tcg_gen_bswap64_i64(tcg_tmp, tcg_tmp);
8061
                break;
8062
            default:
8063
                g_assert_not_reached();
8064
            }
8065
            write_vec_element(s, tcg_tmp, rd, i, grp_size);
8066
            tcg_temp_free_i64(tcg_tmp);
8067
        }
8068
        if (!is_q) {
8069
            clear_vec_high(s, rd);
8070
        }
8071
    } else {
8072
        int revmask = (1 << grp_size) - 1;
8073
        int esize = 8 << size;
8074
        int elements = dsize / esize;
8075
        TCGv_i64 tcg_rn = tcg_temp_new_i64();
8076
        TCGv_i64 tcg_rd = tcg_const_i64(0);
8077
        TCGv_i64 tcg_rd_hi = tcg_const_i64(0);
8078

    
8079
        for (i = 0; i < elements; i++) {
8080
            int e_rev = (i & 0xf) ^ revmask;
8081
            int off = e_rev * esize;
8082
            read_vec_element(s, tcg_rn, rn, i, size);
8083
            if (off >= 64) {
8084
                tcg_gen_deposit_i64(tcg_rd_hi, tcg_rd_hi,
8085
                                    tcg_rn, off - 64, esize);
8086
            } else {
8087
                tcg_gen_deposit_i64(tcg_rd, tcg_rd, tcg_rn, off, esize);
8088
            }
8089
        }
8090
        write_vec_element(s, tcg_rd, rd, 0, MO_64);
8091
        write_vec_element(s, tcg_rd_hi, rd, 1, MO_64);
8092

    
8093
        tcg_temp_free_i64(tcg_rd_hi);
8094
        tcg_temp_free_i64(tcg_rd);
8095
        tcg_temp_free_i64(tcg_rn);
8096
    }
8097
}
8098

    
8099
/* C3.6.17 AdvSIMD two reg misc
8100
 *   31  30  29 28       24 23  22 21       17 16    12 11 10 9    5 4    0
8101
 * +---+---+---+-----------+------+-----------+--------+-----+------+------+
8102
 * | 0 | Q | U | 0 1 1 1 0 | size | 1 0 0 0 0 | opcode | 1 0 |  Rn  |  Rd  |
8103
 * +---+---+---+-----------+------+-----------+--------+-----+------+------+
8104
 */
8105
static void disas_simd_two_reg_misc(DisasContext *s, uint32_t insn)
8106
{
8107
    int size = extract32(insn, 22, 2);
8108
    int opcode = extract32(insn, 12, 5);
8109
    bool u = extract32(insn, 29, 1);
8110
    bool is_q = extract32(insn, 30, 1);
8111
    int rn = extract32(insn, 5, 5);
8112
    int rd = extract32(insn, 0, 5);
8113

    
8114
    switch (opcode) {
8115
    case 0x0: /* REV64, REV32 */
8116
    case 0x1: /* REV16 */
8117
        handle_rev(s, opcode, u, is_q, size, rn, rd);
8118
        return;
8119
    case 0x5: /* CNT, NOT, RBIT */
8120
        if (u && size == 0) {
8121
            /* NOT: adjust size so we can use the 64-bits-at-a-time loop. */
8122
            size = 3;
8123
            break;
8124
        } else if (u && size == 1) {
8125
            /* RBIT */
8126
            break;
8127
        } else if (!u && size == 0) {
8128
            /* CNT */
8129
            break;
8130
        }
8131
        unallocated_encoding(s);
8132
        return;
8133
    case 0x12: /* XTN, XTN2, SQXTUN, SQXTUN2 */
8134
    case 0x14: /* SQXTN, SQXTN2, UQXTN, UQXTN2 */
8135
        if (size == 3) {
8136
            unallocated_encoding(s);
8137
            return;
8138
        }
8139
        handle_2misc_narrow(s, opcode, u, is_q, size, rn, rd);
8140
        return;
8141
    case 0x2: /* SADDLP, UADDLP */
8142
    case 0x4: /* CLS, CLZ */
8143
    case 0x6: /* SADALP, UADALP */
8144
        if (size == 3) {
8145
            unallocated_encoding(s);
8146
            return;
8147
        }
8148
        unsupported_encoding(s, insn);
8149
        return;
8150
    case 0x13: /* SHLL, SHLL2 */
8151
        if (u == 0 || size == 3) {
8152
            unallocated_encoding(s);
8153
            return;
8154
        }
8155
        unsupported_encoding(s, insn);
8156
        return;
8157
    case 0xa: /* CMLT */
8158
        if (u == 1) {
8159
            unallocated_encoding(s);
8160
            return;
8161
        }
8162
        /* fall through */
8163
    case 0x8: /* CMGT, CMGE */
8164
    case 0x9: /* CMEQ, CMLE */
8165
    case 0xb: /* ABS, NEG */
8166
        if (size == 3 && !is_q) {
8167
            unallocated_encoding(s);
8168
            return;
8169
        }
8170
        break;
8171
    case 0x3: /* SUQADD, USQADD */
8172
    case 0x7: /* SQABS, SQNEG */
8173
        if (size == 3 && !is_q) {
8174
            unallocated_encoding(s);
8175
            return;
8176
        }
8177
        unsupported_encoding(s, insn);
8178
        return;
8179
    case 0xc ... 0xf:
8180
    case 0x16 ... 0x1d:
8181
    case 0x1f:
8182
    {
8183
        /* Floating point: U, size[1] and opcode indicate operation;
8184
         * size[0] indicates single or double precision.
8185
         */
8186
        opcode |= (extract32(size, 1, 1) << 5) | (u << 6);
8187
        size = extract32(size, 0, 1) ? 3 : 2;
8188
        switch (opcode) {
8189
        case 0x2f: /* FABS */
8190
        case 0x6f: /* FNEG */
8191
            if (size == 3 && !is_q) {
8192
                unallocated_encoding(s);
8193
                return;
8194
            }
8195
            break;
8196
        case 0x2c: /* FCMGT (zero) */
8197
        case 0x2d: /* FCMEQ (zero) */
8198
        case 0x2e: /* FCMLT (zero) */
8199
        case 0x6c: /* FCMGE (zero) */
8200
        case 0x6d: /* FCMLE (zero) */
8201
            if (size == 3 && !is_q) {
8202
                unallocated_encoding(s);
8203
                return;
8204
            }
8205
            handle_2misc_fcmp_zero(s, opcode, false, u, is_q, size, rn, rd);
8206
            return;
8207
        case 0x16: /* FCVTN, FCVTN2 */
8208
        case 0x17: /* FCVTL, FCVTL2 */
8209
        case 0x18: /* FRINTN */
8210
        case 0x19: /* FRINTM */
8211
        case 0x1a: /* FCVTNS */
8212
        case 0x1b: /* FCVTMS */
8213
        case 0x1c: /* FCVTAS */
8214
        case 0x1d: /* SCVTF */
8215
        case 0x38: /* FRINTP */
8216
        case 0x39: /* FRINTZ */
8217
        case 0x3a: /* FCVTPS */
8218
        case 0x3b: /* FCVTZS */
8219
        case 0x3c: /* URECPE */
8220
        case 0x3d: /* FRECPE */
8221
        case 0x56: /* FCVTXN, FCVTXN2 */
8222
        case 0x58: /* FRINTA */
8223
        case 0x59: /* FRINTX */
8224
        case 0x5a: /* FCVTNU */
8225
        case 0x5b: /* FCVTMU */
8226
        case 0x5c: /* FCVTAU */
8227
        case 0x5d: /* UCVTF */
8228
        case 0x79: /* FRINTI */
8229
        case 0x7a: /* FCVTPU */
8230
        case 0x7b: /* FCVTZU */
8231
        case 0x7c: /* URSQRTE */
8232
        case 0x7d: /* FRSQRTE */
8233
        case 0x7f: /* FSQRT */
8234
            unsupported_encoding(s, insn);
8235
            return;
8236
        default:
8237
            unallocated_encoding(s);
8238
            return;
8239
        }
8240
        break;
8241
    }
8242
    default:
8243
        unallocated_encoding(s);
8244
        return;
8245
    }
8246

    
8247
    if (size == 3) {
8248
        /* All 64-bit element operations can be shared with scalar 2misc */
8249
        int pass;
8250

    
8251
        for (pass = 0; pass < (is_q ? 2 : 1); pass++) {
8252
            TCGv_i64 tcg_op = tcg_temp_new_i64();
8253
            TCGv_i64 tcg_res = tcg_temp_new_i64();
8254

    
8255
            read_vec_element(s, tcg_op, rn, pass, MO_64);
8256

    
8257
            handle_2misc_64(s, opcode, u, tcg_res, tcg_op);
8258

    
8259
            write_vec_element(s, tcg_res, rd, pass, MO_64);
8260

    
8261
            tcg_temp_free_i64(tcg_res);
8262
            tcg_temp_free_i64(tcg_op);
8263
        }
8264
    } else {
8265
        int pass;
8266

    
8267
        for (pass = 0; pass < (is_q ? 4 : 2); pass++) {
8268
            TCGv_i32 tcg_op = tcg_temp_new_i32();
8269
            TCGv_i32 tcg_res = tcg_temp_new_i32();
8270
            TCGCond cond;
8271

    
8272
            read_vec_element_i32(s, tcg_op, rn, pass, MO_32);
8273

    
8274
            if (size == 2) {
8275
                /* Special cases for 32 bit elements */
8276
                switch (opcode) {
8277
                case 0xa: /* CMLT */
8278
                    /* 32 bit integer comparison against zero, result is
8279
                     * test ? (2^32 - 1) : 0. We implement via setcond(test)
8280
                     * and inverting.
8281
                     */
8282
                    cond = TCG_COND_LT;
8283
                do_cmop:
8284
                    tcg_gen_setcondi_i32(cond, tcg_res, tcg_op, 0);
8285
                    tcg_gen_neg_i32(tcg_res, tcg_res);
8286
                    break;
8287
                case 0x8: /* CMGT, CMGE */
8288
                    cond = u ? TCG_COND_GE : TCG_COND_GT;
8289
                    goto do_cmop;
8290
                case 0x9: /* CMEQ, CMLE */
8291
                    cond = u ? TCG_COND_LE : TCG_COND_EQ;
8292
                    goto do_cmop;
8293
                case 0xb: /* ABS, NEG */
8294
                    if (u) {
8295
                        tcg_gen_neg_i32(tcg_res, tcg_op);
8296
                    } else {
8297
                        TCGv_i32 tcg_zero = tcg_const_i32(0);
8298
                        tcg_gen_neg_i32(tcg_res, tcg_op);
8299
                        tcg_gen_movcond_i32(TCG_COND_GT, tcg_res, tcg_op,
8300
                                            tcg_zero, tcg_op, tcg_res);
8301
                        tcg_temp_free_i32(tcg_zero);
8302
                    }
8303
                    break;
8304
                case 0x2f: /* FABS */
8305
                    gen_helper_vfp_abss(tcg_res, tcg_op);
8306
                    break;
8307
                case 0x6f: /* FNEG */
8308
                    gen_helper_vfp_negs(tcg_res, tcg_op);
8309
                    break;
8310
                default:
8311
                    g_assert_not_reached();
8312
                }
8313
            } else {
8314
                /* Use helpers for 8 and 16 bit elements */
8315
                switch (opcode) {
8316
                case 0x5: /* CNT, RBIT */
8317
                    /* For these two insns size is part of the opcode specifier
8318
                     * (handled earlier); they always operate on byte elements.
8319
                     */
8320
                    if (u) {
8321
                        gen_helper_neon_rbit_u8(tcg_res, tcg_op);
8322
                    } else {
8323
                        gen_helper_neon_cnt_u8(tcg_res, tcg_op);
8324
                    }
8325
                    break;
8326
                case 0x8: /* CMGT, CMGE */
8327
                case 0x9: /* CMEQ, CMLE */
8328
                case 0xa: /* CMLT */
8329
                {
8330
                    static NeonGenTwoOpFn * const fns[3][2] = {
8331
                        { gen_helper_neon_cgt_s8, gen_helper_neon_cgt_s16 },
8332
                        { gen_helper_neon_cge_s8, gen_helper_neon_cge_s16 },
8333
                        { gen_helper_neon_ceq_u8, gen_helper_neon_ceq_u16 },
8334
                    };
8335
                    NeonGenTwoOpFn *genfn;
8336
                    int comp;
8337
                    bool reverse;
8338
                    TCGv_i32 tcg_zero = tcg_const_i32(0);
8339

    
8340
                    /* comp = index into [CMGT, CMGE, CMEQ, CMLE, CMLT] */
8341
                    comp = (opcode - 0x8) * 2 + u;
8342
                    /* ...but LE, LT are implemented as reverse GE, GT */
8343
                    reverse = (comp > 2);
8344
                    if (reverse) {
8345
                        comp = 4 - comp;
8346
                    }
8347
                    genfn = fns[comp][size];
8348
                    if (reverse) {
8349
                        genfn(tcg_res, tcg_zero, tcg_op);
8350
                    } else {
8351
                        genfn(tcg_res, tcg_op, tcg_zero);
8352
                    }
8353
                    tcg_temp_free_i32(tcg_zero);
8354
                    break;
8355
                }
8356
                case 0xb: /* ABS, NEG */
8357
                    if (u) {
8358
                        TCGv_i32 tcg_zero = tcg_const_i32(0);
8359
                        if (size) {
8360
                            gen_helper_neon_sub_u16(tcg_res, tcg_zero, tcg_op);
8361
                        } else {
8362
                            gen_helper_neon_sub_u8(tcg_res, tcg_zero, tcg_op);
8363
                        }
8364
                        tcg_temp_free_i32(tcg_zero);
8365
                    } else {
8366
                        if (size) {
8367
                            gen_helper_neon_abs_s16(tcg_res, tcg_op);
8368
                        } else {
8369
                            gen_helper_neon_abs_s8(tcg_res, tcg_op);
8370
                        }
8371
                    }
8372
                    break;
8373
                default:
8374
                    g_assert_not_reached();
8375
                }
8376
            }
8377

    
8378
            write_vec_element_i32(s, tcg_res, rd, pass, MO_32);
8379

    
8380
            tcg_temp_free_i32(tcg_res);
8381
            tcg_temp_free_i32(tcg_op);
8382
        }
8383
    }
8384
    if (!is_q) {
8385
        clear_vec_high(s, rd);
8386
    }
8387
}
8388

    
8389
/* C3.6.13 AdvSIMD scalar x indexed element
8390
 *  31 30  29 28       24 23  22 21  20  19  16 15 12  11  10 9    5 4    0
8391
 * +-----+---+-----------+------+---+---+------+-----+---+---+------+------+
8392
 * | 0 1 | U | 1 1 1 1 1 | size | L | M |  Rm  | opc | H | 0 |  Rn  |  Rd  |
8393
 * +-----+---+-----------+------+---+---+------+-----+---+---+------+------+
8394
 * C3.6.18 AdvSIMD vector x indexed element
8395
 *   31  30  29 28       24 23  22 21  20  19  16 15 12  11  10 9    5 4    0
8396
 * +---+---+---+-----------+------+---+---+------+-----+---+---+------+------+
8397
 * | 0 | Q | U | 0 1 1 1 1 | size | L | M |  Rm  | opc | H | 0 |  Rn  |  Rd  |
8398
 * +---+---+---+-----------+------+---+---+------+-----+---+---+------+------+
8399
 */
8400
static void disas_simd_indexed(DisasContext *s, uint32_t insn)
8401
{
8402
    /* This encoding has two kinds of instruction:
8403
     *  normal, where we perform elt x idxelt => elt for each
8404
     *     element in the vector
8405
     *  long, where we perform elt x idxelt and generate a result of
8406
     *     double the width of the input element
8407
     * The long ops have a 'part' specifier (ie come in INSN, INSN2 pairs).
8408
     */
8409
    bool is_scalar = extract32(insn, 28, 1);
8410
    bool is_q = extract32(insn, 30, 1);
8411
    bool u = extract32(insn, 29, 1);
8412
    int size = extract32(insn, 22, 2);
8413
    int l = extract32(insn, 21, 1);
8414
    int m = extract32(insn, 20, 1);
8415
    /* Note that the Rm field here is only 4 bits, not 5 as it usually is */
8416
    int rm = extract32(insn, 16, 4);
8417
    int opcode = extract32(insn, 12, 4);
8418
    int h = extract32(insn, 11, 1);
8419
    int rn = extract32(insn, 5, 5);
8420
    int rd = extract32(insn, 0, 5);
8421
    bool is_long = false;
8422
    bool is_fp = false;
8423
    int index;
8424
    TCGv_ptr fpst;
8425

    
8426
    switch (opcode) {
8427
    case 0x0: /* MLA */
8428
    case 0x4: /* MLS */
8429
        if (!u || is_scalar) {
8430
            unallocated_encoding(s);
8431
            return;
8432
        }
8433
        break;
8434
    case 0x2: /* SMLAL, SMLAL2, UMLAL, UMLAL2 */
8435
    case 0x6: /* SMLSL, SMLSL2, UMLSL, UMLSL2 */
8436
    case 0xa: /* SMULL, SMULL2, UMULL, UMULL2 */
8437
        if (is_scalar) {
8438
            unallocated_encoding(s);
8439
            return;
8440
        }
8441
        is_long = true;
8442
        break;
8443
    case 0x3: /* SQDMLAL, SQDMLAL2 */
8444
    case 0x7: /* SQDMLSL, SQDMLSL2 */
8445
    case 0xb: /* SQDMULL, SQDMULL2 */
8446
        is_long = true;
8447
        /* fall through */
8448
    case 0xc: /* SQDMULH */
8449
    case 0xd: /* SQRDMULH */
8450
        if (u) {
8451
            unallocated_encoding(s);
8452
            return;
8453
        }
8454
        break;
8455
    case 0x8: /* MUL */
8456
        if (u || is_scalar) {
8457
            unallocated_encoding(s);
8458
            return;
8459
        }
8460
        break;
8461
    case 0x1: /* FMLA */
8462
    case 0x5: /* FMLS */
8463
        if (u) {
8464
            unallocated_encoding(s);
8465
            return;
8466
        }
8467
        /* fall through */
8468
    case 0x9: /* FMUL, FMULX */
8469
        if (!extract32(size, 1, 1)) {
8470
            unallocated_encoding(s);
8471
            return;
8472
        }
8473
        is_fp = true;
8474
        break;
8475
    default:
8476
        unallocated_encoding(s);
8477
        return;
8478
    }
8479

    
8480
    if (is_fp) {
8481
        /* low bit of size indicates single/double */
8482
        size = extract32(size, 0, 1) ? 3 : 2;
8483
        if (size == 2) {
8484
            index = h << 1 | l;
8485
        } else {
8486
            if (l || !is_q) {
8487
                unallocated_encoding(s);
8488
                return;
8489
            }
8490
            index = h;
8491
        }
8492
        rm |= (m << 4);
8493
    } else {
8494
        switch (size) {
8495
        case 1:
8496
            index = h << 2 | l << 1 | m;
8497
            break;
8498
        case 2:
8499
            index = h << 1 | l;
8500
            rm |= (m << 4);
8501
            break;
8502
        default:
8503
            unallocated_encoding(s);
8504
            return;
8505
        }
8506
    }
8507

    
8508
    if (is_fp) {
8509
        fpst = get_fpstatus_ptr();
8510
    } else {
8511
        TCGV_UNUSED_PTR(fpst);
8512
    }
8513

    
8514
    if (size == 3) {
8515
        TCGv_i64 tcg_idx = tcg_temp_new_i64();
8516
        int pass;
8517

    
8518
        assert(is_fp && is_q && !is_long);
8519

    
8520
        read_vec_element(s, tcg_idx, rm, index, MO_64);
8521

    
8522
        for (pass = 0; pass < (is_scalar ? 1 : 2); pass++) {
8523
            TCGv_i64 tcg_op = tcg_temp_new_i64();
8524
            TCGv_i64 tcg_res = tcg_temp_new_i64();
8525

    
8526
            read_vec_element(s, tcg_op, rn, pass, MO_64);
8527

    
8528
            switch (opcode) {
8529
            case 0x5: /* FMLS */
8530
                /* As usual for ARM, separate negation for fused multiply-add */
8531
                gen_helper_vfp_negd(tcg_op, tcg_op);
8532
                /* fall through */
8533
            case 0x1: /* FMLA */
8534
                read_vec_element(s, tcg_res, rd, pass, MO_64);
8535
                gen_helper_vfp_muladdd(tcg_res, tcg_op, tcg_idx, tcg_res, fpst);
8536
                break;
8537
            case 0x9: /* FMUL, FMULX */
8538
                if (u) {
8539
                    gen_helper_vfp_mulxd(tcg_res, tcg_op, tcg_idx, fpst);
8540
                } else {
8541
                    gen_helper_vfp_muld(tcg_res, tcg_op, tcg_idx, fpst);
8542
                }
8543
                break;
8544
            default:
8545
                g_assert_not_reached();
8546
            }
8547

    
8548
            write_vec_element(s, tcg_res, rd, pass, MO_64);
8549
            tcg_temp_free_i64(tcg_op);
8550
            tcg_temp_free_i64(tcg_res);
8551
        }
8552

    
8553
        if (is_scalar) {
8554
            clear_vec_high(s, rd);
8555
        }
8556

    
8557
        tcg_temp_free_i64(tcg_idx);
8558
    } else if (!is_long) {
8559
        /* 32 bit floating point, or 16 or 32 bit integer.
8560
         * For the 16 bit scalar case we use the usual Neon helpers and
8561
         * rely on the fact that 0 op 0 == 0 with no side effects.
8562
         */
8563
        TCGv_i32 tcg_idx = tcg_temp_new_i32();
8564
        int pass, maxpasses;
8565

    
8566
        if (is_scalar) {
8567
            maxpasses = 1;
8568
        } else {
8569
            maxpasses = is_q ? 4 : 2;
8570
        }
8571

    
8572
        read_vec_element_i32(s, tcg_idx, rm, index, size);
8573

    
8574
        if (size == 1 && !is_scalar) {
8575
            /* The simplest way to handle the 16x16 indexed ops is to duplicate
8576
             * the index into both halves of the 32 bit tcg_idx and then use
8577
             * the usual Neon helpers.
8578
             */
8579
            tcg_gen_deposit_i32(tcg_idx, tcg_idx, tcg_idx, 16, 16);
8580
        }
8581

    
8582
        for (pass = 0; pass < maxpasses; pass++) {
8583
            TCGv_i32 tcg_op = tcg_temp_new_i32();
8584
            TCGv_i32 tcg_res = tcg_temp_new_i32();
8585

    
8586
            read_vec_element_i32(s, tcg_op, rn, pass, is_scalar ? size : MO_32);
8587

    
8588
            switch (opcode) {
8589
            case 0x0: /* MLA */
8590
            case 0x4: /* MLS */
8591
            case 0x8: /* MUL */
8592
            {
8593
                static NeonGenTwoOpFn * const fns[2][2] = {
8594
                    { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 },
8595
                    { tcg_gen_add_i32, tcg_gen_sub_i32 },
8596
                };
8597
                NeonGenTwoOpFn *genfn;
8598
                bool is_sub = opcode == 0x4;
8599

    
8600
                if (size == 1) {
8601
                    gen_helper_neon_mul_u16(tcg_res, tcg_op, tcg_idx);
8602
                } else {
8603
                    tcg_gen_mul_i32(tcg_res, tcg_op, tcg_idx);
8604
                }
8605
                if (opcode == 0x8) {
8606
                    break;
8607
                }
8608
                read_vec_element_i32(s, tcg_op, rd, pass, MO_32);
8609
                genfn = fns[size - 1][is_sub];
8610
                genfn(tcg_res, tcg_op, tcg_res);
8611
                break;
8612
            }
8613
            case 0x5: /* FMLS */
8614
                /* As usual for ARM, separate negation for fused multiply-add */
8615
                gen_helper_vfp_negs(tcg_op, tcg_op);
8616
                /* fall through */
8617
            case 0x1: /* FMLA */
8618
                read_vec_element_i32(s, tcg_res, rd, pass, MO_32);
8619
                gen_helper_vfp_muladds(tcg_res, tcg_op, tcg_idx, tcg_res, fpst);
8620
                break;
8621
            case 0x9: /* FMUL, FMULX */
8622
                if (u) {
8623
                    gen_helper_vfp_mulxs(tcg_res, tcg_op, tcg_idx, fpst);
8624
                } else {
8625
                    gen_helper_vfp_muls(tcg_res, tcg_op, tcg_idx, fpst);
8626
                }
8627
                break;
8628
            case 0xc: /* SQDMULH */
8629
                if (size == 1) {
8630
                    gen_helper_neon_qdmulh_s16(tcg_res, cpu_env,
8631
                                               tcg_op, tcg_idx);
8632
                } else {
8633
                    gen_helper_neon_qdmulh_s32(tcg_res, cpu_env,
8634
                                               tcg_op, tcg_idx);
8635
                }
8636
                break;
8637
            case 0xd: /* SQRDMULH */
8638
                if (size == 1) {
8639
                    gen_helper_neon_qrdmulh_s16(tcg_res, cpu_env,
8640
                                                tcg_op, tcg_idx);
8641
                } else {
8642
                    gen_helper_neon_qrdmulh_s32(tcg_res, cpu_env,
8643
                                                tcg_op, tcg_idx);
8644
                }
8645
                break;
8646
            default:
8647
                g_assert_not_reached();
8648
            }
8649

    
8650
            if (is_scalar) {
8651
                write_fp_sreg(s, rd, tcg_res);
8652
            } else {
8653
                write_vec_element_i32(s, tcg_res, rd, pass, MO_32);
8654
            }
8655

    
8656
            tcg_temp_free_i32(tcg_op);
8657
            tcg_temp_free_i32(tcg_res);
8658
        }
8659

    
8660
        tcg_temp_free_i32(tcg_idx);
8661

    
8662
        if (!is_q) {
8663
            clear_vec_high(s, rd);
8664
        }
8665
    } else {
8666
        /* long ops: 16x16->32 or 32x32->64 */
8667
        TCGv_i64 tcg_res[2];
8668
        int pass;
8669
        bool satop = extract32(opcode, 0, 1);
8670
        TCGMemOp memop = MO_32;
8671

    
8672
        if (satop || !u) {
8673
            memop |= MO_SIGN;
8674
        }
8675

    
8676
        if (size == 2) {
8677
            TCGv_i64 tcg_idx = tcg_temp_new_i64();
8678

    
8679
            read_vec_element(s, tcg_idx, rm, index, memop);
8680

    
8681
            for (pass = 0; pass < (is_scalar ? 1 : 2); pass++) {
8682
                TCGv_i64 tcg_op = tcg_temp_new_i64();
8683
                TCGv_i64 tcg_passres;
8684
                int passelt;
8685

    
8686
                if (is_scalar) {
8687
                    passelt = 0;
8688
                } else {
8689
                    passelt = pass + (is_q * 2);
8690
                }
8691

    
8692
                read_vec_element(s, tcg_op, rn, passelt, memop);
8693

    
8694
                tcg_res[pass] = tcg_temp_new_i64();
8695

    
8696
                if (opcode == 0xa || opcode == 0xb) {
8697
                    /* Non-accumulating ops */
8698
                    tcg_passres = tcg_res[pass];
8699
                } else {
8700
                    tcg_passres = tcg_temp_new_i64();
8701
                }
8702

    
8703
                tcg_gen_mul_i64(tcg_passres, tcg_op, tcg_idx);
8704
                tcg_temp_free_i64(tcg_op);
8705

    
8706
                if (satop) {
8707
                    /* saturating, doubling */
8708
                    gen_helper_neon_addl_saturate_s64(tcg_passres, cpu_env,
8709
                                                      tcg_passres, tcg_passres);
8710
                }
8711

    
8712
                if (opcode == 0xa || opcode == 0xb) {
8713
                    continue;
8714
                }
8715

    
8716
                /* Accumulating op: handle accumulate step */
8717
                read_vec_element(s, tcg_res[pass], rd, pass, MO_64);
8718

    
8719
                switch (opcode) {
8720
                case 0x2: /* SMLAL, SMLAL2, UMLAL, UMLAL2 */
8721
                    tcg_gen_add_i64(tcg_res[pass], tcg_res[pass], tcg_passres);
8722
                    break;
8723
                case 0x6: /* SMLSL, SMLSL2, UMLSL, UMLSL2 */
8724
                    tcg_gen_sub_i64(tcg_res[pass], tcg_res[pass], tcg_passres);
8725
                    break;
8726
                case 0x7: /* SQDMLSL, SQDMLSL2 */
8727
                    tcg_gen_neg_i64(tcg_passres, tcg_passres);
8728
                    /* fall through */
8729
                case 0x3: /* SQDMLAL, SQDMLAL2 */
8730
                    gen_helper_neon_addl_saturate_s64(tcg_res[pass], cpu_env,
8731
                                                      tcg_res[pass],
8732
                                                      tcg_passres);
8733
                    break;
8734
                default:
8735
                    g_assert_not_reached();
8736
                }
8737
                tcg_temp_free_i64(tcg_passres);
8738
            }
8739
            tcg_temp_free_i64(tcg_idx);
8740

    
8741
            if (is_scalar) {
8742
                clear_vec_high(s, rd);
8743
            }
8744
        } else {
8745
            TCGv_i32 tcg_idx = tcg_temp_new_i32();
8746

    
8747
            assert(size == 1);
8748
            read_vec_element_i32(s, tcg_idx, rm, index, size);
8749

    
8750
            if (!is_scalar) {
8751
                /* The simplest way to handle the 16x16 indexed ops is to
8752
                 * duplicate the index into both halves of the 32 bit tcg_idx
8753
                 * and then use the usual Neon helpers.
8754
                 */
8755
                tcg_gen_deposit_i32(tcg_idx, tcg_idx, tcg_idx, 16, 16);
8756
            }
8757

    
8758
            for (pass = 0; pass < (is_scalar ? 1 : 2); pass++) {
8759
                TCGv_i32 tcg_op = tcg_temp_new_i32();
8760
                TCGv_i64 tcg_passres;
8761

    
8762
                if (is_scalar) {
8763
                    read_vec_element_i32(s, tcg_op, rn, pass, size);
8764
                } else {
8765
                    read_vec_element_i32(s, tcg_op, rn,
8766
                                         pass + (is_q * 2), MO_32);
8767
                }
8768

    
8769
                tcg_res[pass] = tcg_temp_new_i64();
8770

    
8771
                if (opcode == 0xa || opcode == 0xb) {
8772
                    /* Non-accumulating ops */
8773
                    tcg_passres = tcg_res[pass];
8774
                } else {
8775
                    tcg_passres = tcg_temp_new_i64();
8776
                }
8777

    
8778
                if (memop & MO_SIGN) {
8779
                    gen_helper_neon_mull_s16(tcg_passres, tcg_op, tcg_idx);
8780
                } else {
8781
                    gen_helper_neon_mull_u16(tcg_passres, tcg_op, tcg_idx);
8782
                }
8783
                if (satop) {
8784
                    gen_helper_neon_addl_saturate_s32(tcg_passres, cpu_env,
8785
                                                      tcg_passres, tcg_passres);
8786
                }
8787
                tcg_temp_free_i32(tcg_op);
8788

    
8789
                if (opcode == 0xa || opcode == 0xb) {
8790
                    continue;
8791
                }
8792

    
8793
                /* Accumulating op: handle accumulate step */
8794
                read_vec_element(s, tcg_res[pass], rd, pass, MO_64);
8795

    
8796
                switch (opcode) {
8797
                case 0x2: /* SMLAL, SMLAL2, UMLAL, UMLAL2 */
8798
                    gen_helper_neon_addl_u32(tcg_res[pass], tcg_res[pass],
8799
                                             tcg_passres);
8800
                    break;
8801
                case 0x6: /* SMLSL, SMLSL2, UMLSL, UMLSL2 */
8802
                    gen_helper_neon_subl_u32(tcg_res[pass], tcg_res[pass],
8803
                                             tcg_passres);
8804
                    break;
8805
                case 0x7: /* SQDMLSL, SQDMLSL2 */
8806
                    gen_helper_neon_negl_u32(tcg_passres, tcg_passres);
8807
                    /* fall through */
8808
                case 0x3: /* SQDMLAL, SQDMLAL2 */
8809
                    gen_helper_neon_addl_saturate_s32(tcg_res[pass], cpu_env,
8810
                                                      tcg_res[pass],
8811
                                                      tcg_passres);
8812
                    break;
8813
                default:
8814
                    g_assert_not_reached();
8815
                }
8816
                tcg_temp_free_i64(tcg_passres);
8817
            }
8818
            tcg_temp_free_i32(tcg_idx);
8819

    
8820
            if (is_scalar) {
8821
                tcg_gen_ext32u_i64(tcg_res[0], tcg_res[0]);
8822
            }
8823
        }
8824

    
8825
        if (is_scalar) {
8826
            tcg_res[1] = tcg_const_i64(0);
8827
        }
8828

    
8829
        for (pass = 0; pass < 2; pass++) {
8830
            write_vec_element(s, tcg_res[pass], rd, pass, MO_64);
8831
            tcg_temp_free_i64(tcg_res[pass]);
8832
        }
8833
    }
8834

    
8835
    if (!TCGV_IS_UNUSED_PTR(fpst)) {
8836
        tcg_temp_free_ptr(fpst);
8837
    }
8838
}
8839

    
8840
/* C3.6.19 Crypto AES
8841
 *  31             24 23  22 21       17 16    12 11 10 9    5 4    0
8842
 * +-----------------+------+-----------+--------+-----+------+------+
8843
 * | 0 1 0 0 1 1 1 0 | size | 1 0 1 0 0 | opcode | 1 0 |  Rn  |  Rd  |
8844
 * +-----------------+------+-----------+--------+-----+------+------+
8845
 */
8846
static void disas_crypto_aes(DisasContext *s, uint32_t insn)
8847
{
8848
    unsupported_encoding(s, insn);
8849
}
8850

    
8851
/* C3.6.20 Crypto three-reg SHA
8852
 *  31             24 23  22  21 20  16  15 14    12 11 10 9    5 4    0
8853
 * +-----------------+------+---+------+---+--------+-----+------+------+
8854
 * | 0 1 0 1 1 1 1 0 | size | 0 |  Rm  | 0 | opcode | 0 0 |  Rn  |  Rd  |
8855
 * +-----------------+------+---+------+---+--------+-----+------+------+
8856
 */
8857
static void disas_crypto_three_reg_sha(DisasContext *s, uint32_t insn)
8858
{
8859
    unsupported_encoding(s, insn);
8860
}
8861

    
8862
/* C3.6.21 Crypto two-reg SHA
8863
 *  31             24 23  22 21       17 16    12 11 10 9    5 4    0
8864
 * +-----------------+------+-----------+--------+-----+------+------+
8865
 * | 0 1 0 1 1 1 1 0 | size | 1 0 1 0 0 | opcode | 1 0 |  Rn  |  Rd  |
8866
 * +-----------------+------+-----------+--------+-----+------+------+
8867
 */
8868
static void disas_crypto_two_reg_sha(DisasContext *s, uint32_t insn)
8869
{
8870
    unsupported_encoding(s, insn);
8871
}
8872

    
8873
/* C3.6 Data processing - SIMD, inc Crypto
8874
 *
8875
 * As the decode gets a little complex we are using a table based
8876
 * approach for this part of the decode.
8877
 */
8878
static const AArch64DecodeTable data_proc_simd[] = {
8879
    /* pattern  ,  mask     ,  fn                        */
8880
    { 0x0e200400, 0x9f200400, disas_simd_three_reg_same },
8881
    { 0x0e200000, 0x9f200c00, disas_simd_three_reg_diff },
8882
    { 0x0e200800, 0x9f3e0c00, disas_simd_two_reg_misc },
8883
    { 0x0e300800, 0x9f3e0c00, disas_simd_across_lanes },
8884
    { 0x0e000400, 0x9fe08400, disas_simd_copy },
8885
    { 0x0f000000, 0x9f000400, disas_simd_indexed }, /* vector indexed */
8886
    /* simd_mod_imm decode is a subset of simd_shift_imm, so must precede it */
8887
    { 0x0f000400, 0x9ff80400, disas_simd_mod_imm },
8888
    { 0x0f000400, 0x9f800400, disas_simd_shift_imm },
8889
    { 0x0e000000, 0xbf208c00, disas_simd_tb },
8890
    { 0x0e000800, 0xbf208c00, disas_simd_zip_trn },
8891
    { 0x2e000000, 0xbf208400, disas_simd_ext },
8892
    { 0x5e200400, 0xdf200400, disas_simd_scalar_three_reg_same },
8893
    { 0x5e200000, 0xdf200c00, disas_simd_scalar_three_reg_diff },
8894
    { 0x5e200800, 0xdf3e0c00, disas_simd_scalar_two_reg_misc },
8895
    { 0x5e300800, 0xdf3e0c00, disas_simd_scalar_pairwise },
8896
    { 0x5e000400, 0xdfe08400, disas_simd_scalar_copy },
8897
    { 0x5f000000, 0xdf000400, disas_simd_indexed }, /* scalar indexed */
8898
    { 0x5f000400, 0xdf800400, disas_simd_scalar_shift_imm },
8899
    { 0x4e280800, 0xff3e0c00, disas_crypto_aes },
8900
    { 0x5e000000, 0xff208c00, disas_crypto_three_reg_sha },
8901
    { 0x5e280800, 0xff3e0c00, disas_crypto_two_reg_sha },
8902
    { 0x00000000, 0x00000000, NULL }
8903
};
8904

    
8905
static void disas_data_proc_simd(DisasContext *s, uint32_t insn)
8906
{
8907
    /* Note that this is called with all non-FP cases from
8908
     * table C3-6 so it must UNDEF for entries not specifically
8909
     * allocated to instructions in that table.
8910
     */
8911
    AArch64DecodeFn *fn = lookup_disas_fn(&data_proc_simd[0], insn);
8912
    if (fn) {
8913
        fn(s, insn);
8914
    } else {
8915
        unallocated_encoding(s);
8916
    }
8917
}
8918

    
8919
/* C3.6 Data processing - SIMD and floating point */
8920
static void disas_data_proc_simd_fp(DisasContext *s, uint32_t insn)
8921
{
8922
    if (extract32(insn, 28, 1) == 1 && extract32(insn, 30, 1) == 0) {
8923
        disas_data_proc_fp(s, insn);
8924
    } else {
8925
        /* SIMD, including crypto */
8926
        disas_data_proc_simd(s, insn);
8927
    }
8928
}
8929

    
8930
/* C3.1 A64 instruction index by encoding */
8931
static void disas_a64_insn(CPUARMState *env, DisasContext *s)
8932
{
8933
    uint32_t insn;
8934

    
8935
    insn = arm_ldl_code(env, s->pc, s->bswap_code);
8936
    s->insn = insn;
8937
    s->pc += 4;
8938

    
8939
    switch (extract32(insn, 25, 4)) {
8940
    case 0x0: case 0x1: case 0x2: case 0x3: /* UNALLOCATED */
8941
        unallocated_encoding(s);
8942
        break;
8943
    case 0x8: case 0x9: /* Data processing - immediate */
8944
        disas_data_proc_imm(s, insn);
8945
        break;
8946
    case 0xa: case 0xb: /* Branch, exception generation and system insns */
8947
        disas_b_exc_sys(s, insn);
8948
        break;
8949
    case 0x4:
8950
    case 0x6:
8951
    case 0xc:
8952
    case 0xe:      /* Loads and stores */
8953
        disas_ldst(s, insn);
8954
        break;
8955
    case 0x5:
8956
    case 0xd:      /* Data processing - register */
8957
        disas_data_proc_reg(s, insn);
8958
        break;
8959
    case 0x7:
8960
    case 0xf:      /* Data processing - SIMD and floating point */
8961
        disas_data_proc_simd_fp(s, insn);
8962
        break;
8963
    default:
8964
        assert(FALSE); /* all 15 cases should be handled above */
8965
        break;
8966
    }
8967

    
8968
    /* if we allocated any temporaries, free them here */
8969
    free_tmp_a64(s);
8970
}
8971

    
8972
void gen_intermediate_code_internal_a64(ARMCPU *cpu,
8973
                                        TranslationBlock *tb,
8974
                                        bool search_pc)
8975
{
8976
    CPUState *cs = CPU(cpu);
8977
    CPUARMState *env = &cpu->env;
8978
    DisasContext dc1, *dc = &dc1;
8979
    CPUBreakpoint *bp;
8980
    uint16_t *gen_opc_end;
8981
    int j, lj;
8982
    target_ulong pc_start;
8983
    target_ulong next_page_start;
8984
    int num_insns;
8985
    int max_insns;
8986

    
8987
    pc_start = tb->pc;
8988

    
8989
    dc->tb = tb;
8990

    
8991
    gen_opc_end = tcg_ctx.gen_opc_buf + OPC_MAX_SIZE;
8992

    
8993
    dc->is_jmp = DISAS_NEXT;
8994
    dc->pc = pc_start;
8995
    dc->singlestep_enabled = cs->singlestep_enabled;
8996
    dc->condjmp = 0;
8997

    
8998
    dc->aarch64 = 1;
8999
    dc->thumb = 0;
9000
    dc->bswap_code = 0;
9001
    dc->condexec_mask = 0;
9002
    dc->condexec_cond = 0;
9003
#if !defined(CONFIG_USER_ONLY)
9004
    dc->user = 0;
9005
#endif
9006
    dc->vfp_enabled = 0;
9007
    dc->vec_len = 0;
9008
    dc->vec_stride = 0;
9009
    dc->cp_regs = cpu->cp_regs;
9010
    dc->current_pl = arm_current_pl(env);
9011

    
9012
    init_tmp_a64_array(dc);
9013

    
9014
    next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
9015
    lj = -1;
9016
    num_insns = 0;
9017
    max_insns = tb->cflags & CF_COUNT_MASK;
9018
    if (max_insns == 0) {
9019
        max_insns = CF_COUNT_MASK;
9020
    }
9021

    
9022
    gen_tb_start();
9023

    
9024
    tcg_clear_temp_count();
9025

    
9026
    do {
9027
        if (unlikely(!QTAILQ_EMPTY(&env->breakpoints))) {
9028
            QTAILQ_FOREACH(bp, &env->breakpoints, entry) {
9029
                if (bp->pc == dc->pc) {
9030
                    gen_exception_insn(dc, 0, EXCP_DEBUG);
9031
                    /* Advance PC so that clearing the breakpoint will
9032
                       invalidate this TB.  */
9033
                    dc->pc += 2;
9034
                    goto done_generating;
9035
                }
9036
            }
9037
        }
9038

    
9039
        if (search_pc) {
9040
            j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
9041
            if (lj < j) {
9042
                lj++;
9043
                while (lj < j) {
9044
                    tcg_ctx.gen_opc_instr_start[lj++] = 0;
9045
                }
9046
            }
9047
            tcg_ctx.gen_opc_pc[lj] = dc->pc;
9048
            tcg_ctx.gen_opc_instr_start[lj] = 1;
9049
            tcg_ctx.gen_opc_icount[lj] = num_insns;
9050
        }
9051

    
9052
        if (num_insns + 1 == max_insns && (tb->cflags & CF_LAST_IO)) {
9053
            gen_io_start();
9054
        }
9055

    
9056
        if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP | CPU_LOG_TB_OP_OPT))) {
9057
            tcg_gen_debug_insn_start(dc->pc);
9058
        }
9059

    
9060
        disas_a64_insn(env, dc);
9061

    
9062
        if (tcg_check_temp_count()) {
9063
            fprintf(stderr, "TCG temporary leak before "TARGET_FMT_lx"\n",
9064
                    dc->pc);
9065
        }
9066

    
9067
        /* Translation stops when a conditional branch is encountered.
9068
         * Otherwise the subsequent code could get translated several times.
9069
         * Also stop translation when a page boundary is reached.  This
9070
         * ensures prefetch aborts occur at the right place.
9071
         */
9072
        num_insns++;
9073
    } while (!dc->is_jmp && tcg_ctx.gen_opc_ptr < gen_opc_end &&
9074
             !cs->singlestep_enabled &&
9075
             !singlestep &&
9076
             dc->pc < next_page_start &&
9077
             num_insns < max_insns);
9078

    
9079
    if (tb->cflags & CF_LAST_IO) {
9080
        gen_io_end();
9081
    }
9082

    
9083
    if (unlikely(cs->singlestep_enabled) && dc->is_jmp != DISAS_EXC) {
9084
        /* Note that this means single stepping WFI doesn't halt the CPU.
9085
         * For conditional branch insns this is harmless unreachable code as
9086
         * gen_goto_tb() has already handled emitting the debug exception
9087
         * (and thus a tb-jump is not possible when singlestepping).
9088
         */
9089
        assert(dc->is_jmp != DISAS_TB_JUMP);
9090
        if (dc->is_jmp != DISAS_JUMP) {
9091
            gen_a64_set_pc_im(dc->pc);
9092
        }
9093
        gen_exception(EXCP_DEBUG);
9094
    } else {
9095
        switch (dc->is_jmp) {
9096
        case DISAS_NEXT:
9097
            gen_goto_tb(dc, 1, dc->pc);
9098
            break;
9099
        default:
9100
        case DISAS_UPDATE:
9101
            gen_a64_set_pc_im(dc->pc);
9102
            /* fall through */
9103
        case DISAS_JUMP:
9104
            /* indicate that the hash table must be used to find the next TB */
9105
            tcg_gen_exit_tb(0);
9106
            break;
9107
        case DISAS_TB_JUMP:
9108
        case DISAS_EXC:
9109
        case DISAS_SWI:
9110
            break;
9111
        case DISAS_WFI:
9112
            /* This is a special case because we don't want to just halt the CPU
9113
             * if trying to debug across a WFI.
9114
             */
9115
            gen_helper_wfi(cpu_env);
9116
            break;
9117
        }
9118
    }
9119

    
9120
done_generating:
9121
    gen_tb_end(tb, num_insns);
9122
    *tcg_ctx.gen_opc_ptr = INDEX_op_end;
9123

    
9124
#ifdef DEBUG_DISAS
9125
    if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
9126
        qemu_log("----------------\n");
9127
        qemu_log("IN: %s\n", lookup_symbol(pc_start));
9128
        log_target_disas(env, pc_start, dc->pc - pc_start,
9129
                         4 | (dc->bswap_code << 1));
9130
        qemu_log("\n");
9131
    }
9132
#endif
9133
    if (search_pc) {
9134
        j = tcg_ctx.gen_opc_ptr - tcg_ctx.gen_opc_buf;
9135
        lj++;
9136
        while (lj <= j) {
9137
            tcg_ctx.gen_opc_instr_start[lj++] = 0;
9138
        }
9139
    } else {
9140
        tb->size = dc->pc - pc_start;
9141
        tb->icount = num_insns;
9142
    }
9143
}