summaryrefslogtreecommitdiff
path: root/deps/v8/src/compiler/backend/arm/instruction-codes-arm.h
blob: d4e0c2c4578b6d6b326121ca62236c1a65549a1d (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
// Copyright 2014 the V8 project authors. All rights reserved.
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.

#ifndef V8_COMPILER_BACKEND_ARM_INSTRUCTION_CODES_ARM_H_
#define V8_COMPILER_BACKEND_ARM_INSTRUCTION_CODES_ARM_H_

namespace v8 {
namespace internal {
namespace compiler {

// ARM-specific opcodes that specify which assembly sequence to emit.
// Most opcodes specify a single instruction.

// Opcodes that support a MemoryAccessMode.
#define TARGET_ARCH_OPCODE_WITH_MEMORY_ACCESS_MODE_LIST(V)  // None.

#define TARGET_ARCH_OPCODE_LIST(V)                   \
  TARGET_ARCH_OPCODE_WITH_MEMORY_ACCESS_MODE_LIST(V) \
  V(ArmAdd)                                          \
  V(ArmAnd)                                          \
  V(ArmBic)                                          \
  V(ArmClz)                                          \
  V(ArmCmp)                                          \
  V(ArmCmn)                                          \
  V(ArmTst)                                          \
  V(ArmTeq)                                          \
  V(ArmOrr)                                          \
  V(ArmEor)                                          \
  V(ArmSub)                                          \
  V(ArmRsb)                                          \
  V(ArmMul)                                          \
  V(ArmMla)                                          \
  V(ArmMls)                                          \
  V(ArmSmull)                                        \
  V(ArmSmmul)                                        \
  V(ArmSmmla)                                        \
  V(ArmUmull)                                        \
  V(ArmSdiv)                                         \
  V(ArmUdiv)                                         \
  V(ArmMov)                                          \
  V(ArmMvn)                                          \
  V(ArmBfc)                                          \
  V(ArmUbfx)                                         \
  V(ArmSbfx)                                         \
  V(ArmSxtb)                                         \
  V(ArmSxth)                                         \
  V(ArmSxtab)                                        \
  V(ArmSxtah)                                        \
  V(ArmUxtb)                                         \
  V(ArmUxth)                                         \
  V(ArmUxtab)                                        \
  V(ArmRbit)                                         \
  V(ArmRev)                                          \
  V(ArmUxtah)                                        \
  V(ArmAddPair)                                      \
  V(ArmSubPair)                                      \
  V(ArmMulPair)                                      \
  V(ArmLslPair)                                      \
  V(ArmLsrPair)                                      \
  V(ArmAsrPair)                                      \
  V(ArmVcmpF32)                                      \
  V(ArmVaddF32)                                      \
  V(ArmVsubF32)                                      \
  V(ArmVmulF32)                                      \
  V(ArmVmlaF32)                                      \
  V(ArmVmlsF32)                                      \
  V(ArmVdivF32)                                      \
  V(ArmVabsF32)                                      \
  V(ArmVnegF32)                                      \
  V(ArmVsqrtF32)                                     \
  V(ArmVcmpF64)                                      \
  V(ArmVaddF64)                                      \
  V(ArmVsubF64)                                      \
  V(ArmVmulF64)                                      \
  V(ArmVmlaF64)                                      \
  V(ArmVmlsF64)                                      \
  V(ArmVdivF64)                                      \
  V(ArmVmodF64)                                      \
  V(ArmVabsF64)                                      \
  V(ArmVnegF64)                                      \
  V(ArmVsqrtF64)                                     \
  V(ArmVmullLow)                                     \
  V(ArmVmullHigh)                                    \
  V(ArmVrintmF32)                                    \
  V(ArmVrintmF64)                                    \
  V(ArmVrintpF32)                                    \
  V(ArmVrintpF64)                                    \
  V(ArmVrintzF32)                                    \
  V(ArmVrintzF64)                                    \
  V(ArmVrintaF64)                                    \
  V(ArmVrintnF32)                                    \
  V(ArmVrintnF64)                                    \
  V(ArmVcvtF32F64)                                   \
  V(ArmVcvtF64F32)                                   \
  V(ArmVcvtF32S32)                                   \
  V(ArmVcvtF32U32)                                   \
  V(ArmVcvtF64S32)                                   \
  V(ArmVcvtF64U32)                                   \
  V(ArmVcvtS32F32)                                   \
  V(ArmVcvtU32F32)                                   \
  V(ArmVcvtS32F64)                                   \
  V(ArmVcvtU32F64)                                   \
  V(ArmVmovU32F32)                                   \
  V(ArmVmovF32U32)                                   \
  V(ArmVmovLowU32F64)                                \
  V(ArmVmovLowF64U32)                                \
  V(ArmVmovHighU32F64)                               \
  V(ArmVmovHighF64U32)                               \
  V(ArmVmovF64U32U32)                                \
  V(ArmVmovU32U32F64)                                \
  V(ArmVldrF32)                                      \
  V(ArmVstrF32)                                      \
  V(ArmVldrF64)                                      \
  V(ArmVld1F64)                                      \
  V(ArmVstrF64)                                      \
  V(ArmVst1F64)                                      \
  V(ArmVld1S128)                                     \
  V(ArmVst1S128)                                     \
  V(ArmVcnt)                                         \
  V(ArmVpadal)                                       \
  V(ArmVpaddl)                                       \
  V(ArmFloat32Max)                                   \
  V(ArmFloat64Max)                                   \
  V(ArmFloat32Min)                                   \
  V(ArmFloat64Min)                                   \
  V(ArmFloat64SilenceNaN)                            \
  V(ArmLdrb)                                         \
  V(ArmLdrsb)                                        \
  V(ArmStrb)                                         \
  V(ArmLdrh)                                         \
  V(ArmLdrsh)                                        \
  V(ArmStrh)                                         \
  V(ArmLdr)                                          \
  V(ArmStr)                                          \
  V(ArmPush)                                         \
  V(ArmPoke)                                         \
  V(ArmPeek)                                         \
  V(ArmDmbIsh)                                       \
  V(ArmDsbIsb)                                       \
  V(ArmF64x2Splat)                                   \
  V(ArmF64x2ExtractLane)                             \
  V(ArmF64x2ReplaceLane)                             \
  V(ArmF64x2Abs)                                     \
  V(ArmF64x2Neg)                                     \
  V(ArmF64x2Sqrt)                                    \
  V(ArmF64x2Add)                                     \
  V(ArmF64x2Sub)                                     \
  V(ArmF64x2Mul)                                     \
  V(ArmF64x2Div)                                     \
  V(ArmF64x2Min)                                     \
  V(ArmF64x2Max)                                     \
  V(ArmF64x2Eq)                                      \
  V(ArmF64x2Ne)                                      \
  V(ArmF64x2Lt)                                      \
  V(ArmF64x2Le)                                      \
  V(ArmF64x2Pmin)                                    \
  V(ArmF64x2Pmax)                                    \
  V(ArmF64x2Ceil)                                    \
  V(ArmF64x2Floor)                                   \
  V(ArmF64x2Trunc)                                   \
  V(ArmF64x2NearestInt)                              \
  V(ArmF64x2ConvertLowI32x4S)                        \
  V(ArmF64x2ConvertLowI32x4U)                        \
  V(ArmF64x2PromoteLowF32x4)                         \
  V(ArmF32x4Splat)                                   \
  V(ArmF32x4ExtractLane)                             \
  V(ArmF32x4ReplaceLane)                             \
  V(ArmF32x4SConvertI32x4)                           \
  V(ArmF32x4UConvertI32x4)                           \
  V(ArmF32x4Abs)                                     \
  V(ArmF32x4Neg)                                     \
  V(ArmF32x4Sqrt)                                    \
  V(ArmF32x4RecipApprox)                             \
  V(ArmF32x4RecipSqrtApprox)                         \
  V(ArmF32x4Add)                                     \
  V(ArmF32x4Sub)                                     \
  V(ArmF32x4Mul)                                     \
  V(ArmF32x4Div)                                     \
  V(ArmF32x4Min)                                     \
  V(ArmF32x4Max)                                     \
  V(ArmF32x4Eq)                                      \
  V(ArmF32x4Ne)                                      \
  V(ArmF32x4Lt)                                      \
  V(ArmF32x4Le)                                      \
  V(ArmF32x4Pmin)                                    \
  V(ArmF32x4Pmax)                                    \
  V(ArmF32x4DemoteF64x2Zero)                         \
  V(ArmI64x2SplatI32Pair)                            \
  V(ArmI64x2ReplaceLaneI32Pair)                      \
  V(ArmI64x2Abs)                                     \
  V(ArmI64x2Neg)                                     \
  V(ArmI64x2Shl)                                     \
  V(ArmI64x2ShrS)                                    \
  V(ArmI64x2Add)                                     \
  V(ArmI64x2Sub)                                     \
  V(ArmI64x2Mul)                                     \
  V(ArmI64x2ShrU)                                    \
  V(ArmI64x2BitMask)                                 \
  V(ArmI64x2Eq)                                      \
  V(ArmI64x2Ne)                                      \
  V(ArmI64x2GtS)                                     \
  V(ArmI64x2GeS)                                     \
  V(ArmI64x2SConvertI32x4Low)                        \
  V(ArmI64x2SConvertI32x4High)                       \
  V(ArmI64x2UConvertI32x4Low)                        \
  V(ArmI64x2UConvertI32x4High)                       \
  V(ArmI32x4Splat)                                   \
  V(ArmI32x4ExtractLane)                             \
  V(ArmI32x4ReplaceLane)                             \
  V(ArmI32x4SConvertF32x4)                           \
  V(ArmI32x4SConvertI16x8Low)                        \
  V(ArmI32x4SConvertI16x8High)                       \
  V(ArmI32x4Neg)                                     \
  V(ArmI32x4Shl)                                     \
  V(ArmI32x4ShrS)                                    \
  V(ArmI32x4Add)                                     \
  V(ArmI32x4Sub)                                     \
  V(ArmI32x4Mul)                                     \
  V(ArmI32x4MinS)                                    \
  V(ArmI32x4MaxS)                                    \
  V(ArmI32x4Eq)                                      \
  V(ArmI32x4Ne)                                      \
  V(ArmI32x4GtS)                                     \
  V(ArmI32x4GeS)                                     \
  V(ArmI32x4UConvertF32x4)                           \
  V(ArmI32x4UConvertI16x8Low)                        \
  V(ArmI32x4UConvertI16x8High)                       \
  V(ArmI32x4ShrU)                                    \
  V(ArmI32x4MinU)                                    \
  V(ArmI32x4MaxU)                                    \
  V(ArmI32x4GtU)                                     \
  V(ArmI32x4GeU)                                     \
  V(ArmI32x4Abs)                                     \
  V(ArmI32x4BitMask)                                 \
  V(ArmI32x4DotI16x8S)                               \
  V(ArmI32x4TruncSatF64x2SZero)                      \
  V(ArmI32x4TruncSatF64x2UZero)                      \
  V(ArmI16x8Splat)                                   \
  V(ArmI16x8ExtractLaneS)                            \
  V(ArmI16x8ReplaceLane)                             \
  V(ArmI16x8SConvertI8x16Low)                        \
  V(ArmI16x8SConvertI8x16High)                       \
  V(ArmI16x8Neg)                                     \
  V(ArmI16x8Shl)                                     \
  V(ArmI16x8ShrS)                                    \
  V(ArmI16x8SConvertI32x4)                           \
  V(ArmI16x8Add)                                     \
  V(ArmI16x8AddSatS)                                 \
  V(ArmI16x8Sub)                                     \
  V(ArmI16x8SubSatS)                                 \
  V(ArmI16x8Mul)                                     \
  V(ArmI16x8MinS)                                    \
  V(ArmI16x8MaxS)                                    \
  V(ArmI16x8Eq)                                      \
  V(ArmI16x8Ne)                                      \
  V(ArmI16x8GtS)                                     \
  V(ArmI16x8GeS)                                     \
  V(ArmI16x8ExtractLaneU)                            \
  V(ArmI16x8UConvertI8x16Low)                        \
  V(ArmI16x8UConvertI8x16High)                       \
  V(ArmI16x8ShrU)                                    \
  V(ArmI16x8UConvertI32x4)                           \
  V(ArmI16x8AddSatU)                                 \
  V(ArmI16x8SubSatU)                                 \
  V(ArmI16x8MinU)                                    \
  V(ArmI16x8MaxU)                                    \
  V(ArmI16x8GtU)                                     \
  V(ArmI16x8GeU)                                     \
  V(ArmI16x8RoundingAverageU)                        \
  V(ArmI16x8Abs)                                     \
  V(ArmI16x8BitMask)                                 \
  V(ArmI16x8Q15MulRSatS)                             \
  V(ArmI8x16Splat)                                   \
  V(ArmI8x16ExtractLaneS)                            \
  V(ArmI8x16ReplaceLane)                             \
  V(ArmI8x16Neg)                                     \
  V(ArmI8x16Shl)                                     \
  V(ArmI8x16ShrS)                                    \
  V(ArmI8x16SConvertI16x8)                           \
  V(ArmI8x16Add)                                     \
  V(ArmI8x16AddSatS)                                 \
  V(ArmI8x16Sub)                                     \
  V(ArmI8x16SubSatS)                                 \
  V(ArmI8x16MinS)                                    \
  V(ArmI8x16MaxS)                                    \
  V(ArmI8x16Eq)                                      \
  V(ArmI8x16Ne)                                      \
  V(ArmI8x16GtS)                                     \
  V(ArmI8x16GeS)                                     \
  V(ArmI8x16ExtractLaneU)                            \
  V(ArmI8x16ShrU)                                    \
  V(ArmI8x16UConvertI16x8)                           \
  V(ArmI8x16AddSatU)                                 \
  V(ArmI8x16SubSatU)                                 \
  V(ArmI8x16MinU)                                    \
  V(ArmI8x16MaxU)                                    \
  V(ArmI8x16GtU)                                     \
  V(ArmI8x16GeU)                                     \
  V(ArmI8x16RoundingAverageU)                        \
  V(ArmI8x16Abs)                                     \
  V(ArmI8x16BitMask)                                 \
  V(ArmS128Const)                                    \
  V(ArmS128Zero)                                     \
  V(ArmS128AllOnes)                                  \
  V(ArmS128Dup)                                      \
  V(ArmS128And)                                      \
  V(ArmS128Or)                                       \
  V(ArmS128Xor)                                      \
  V(ArmS128Not)                                      \
  V(ArmS128Select)                                   \
  V(ArmS128AndNot)                                   \
  V(ArmS32x4ZipLeft)                                 \
  V(ArmS32x4ZipRight)                                \
  V(ArmS32x4UnzipLeft)                               \
  V(ArmS32x4UnzipRight)                              \
  V(ArmS32x4TransposeLeft)                           \
  V(ArmS32x4TransposeRight)                          \
  V(ArmS32x4Shuffle)                                 \
  V(ArmS16x8ZipLeft)                                 \
  V(ArmS16x8ZipRight)                                \
  V(ArmS16x8UnzipLeft)                               \
  V(ArmS16x8UnzipRight)                              \
  V(ArmS16x8TransposeLeft)                           \
  V(ArmS16x8TransposeRight)                          \
  V(ArmS8x16ZipLeft)                                 \
  V(ArmS8x16ZipRight)                                \
  V(ArmS8x16UnzipLeft)                               \
  V(ArmS8x16UnzipRight)                              \
  V(ArmS8x16TransposeLeft)                           \
  V(ArmS8x16TransposeRight)                          \
  V(ArmS8x16Concat)                                  \
  V(ArmI8x16Swizzle)                                 \
  V(ArmI8x16Shuffle)                                 \
  V(ArmS32x2Reverse)                                 \
  V(ArmS16x4Reverse)                                 \
  V(ArmS16x2Reverse)                                 \
  V(ArmS8x8Reverse)                                  \
  V(ArmS8x4Reverse)                                  \
  V(ArmS8x2Reverse)                                  \
  V(ArmI64x2AllTrue)                                 \
  V(ArmI32x4AllTrue)                                 \
  V(ArmI16x8AllTrue)                                 \
  V(ArmV128AnyTrue)                                  \
  V(ArmI8x16AllTrue)                                 \
  V(ArmS128Load8Splat)                               \
  V(ArmS128Load16Splat)                              \
  V(ArmS128Load32Splat)                              \
  V(ArmS128Load64Splat)                              \
  V(ArmS128Load8x8S)                                 \
  V(ArmS128Load8x8U)                                 \
  V(ArmS128Load16x4S)                                \
  V(ArmS128Load16x4U)                                \
  V(ArmS128Load32x2S)                                \
  V(ArmS128Load32x2U)                                \
  V(ArmS128Load32Zero)                               \
  V(ArmS128Load64Zero)                               \
  V(ArmS128LoadLaneLow)                              \
  V(ArmS128LoadLaneHigh)                             \
  V(ArmS128StoreLaneLow)                             \
  V(ArmS128StoreLaneHigh)                            \
  V(ArmWord32AtomicPairLoad)                         \
  V(ArmWord32AtomicPairStore)                        \
  V(ArmWord32AtomicPairAdd)                          \
  V(ArmWord32AtomicPairSub)                          \
  V(ArmWord32AtomicPairAnd)                          \
  V(ArmWord32AtomicPairOr)                           \
  V(ArmWord32AtomicPairXor)                          \
  V(ArmWord32AtomicPairExchange)                     \
  V(ArmWord32AtomicPairCompareExchange)

// Addressing modes represent the "shape" of inputs to an instruction.
// Many instructions support multiple addressing modes. Addressing modes
// are encoded into the InstructionCode of the instruction and tell the
// code generator after register allocation which assembler method to call.
#define TARGET_ADDRESSING_MODE_LIST(V)  \
  V(Offset_RI)        /* [%r0 + K] */   \
  V(Offset_RR)        /* [%r0 + %r1] */ \
  V(Operand2_I)       /* K */           \
  V(Operand2_R)       /* %r0 */         \
  V(Operand2_R_ASR_I) /* %r0 ASR K */   \
  V(Operand2_R_LSL_I) /* %r0 LSL K */   \
  V(Operand2_R_LSR_I) /* %r0 LSR K */   \
  V(Operand2_R_ROR_I) /* %r0 ROR K */   \
  V(Operand2_R_ASR_R) /* %r0 ASR %r1 */ \
  V(Operand2_R_LSL_R) /* %r0 LSL %r1 */ \
  V(Operand2_R_LSR_R) /* %r0 LSR %r1 */ \
  V(Operand2_R_ROR_R) /* %r0 ROR %r1 */ \
  V(Root)             /* [%rr + K] */

}  // namespace compiler
}  // namespace internal
}  // namespace v8

#endif  // V8_COMPILER_BACKEND_ARM_INSTRUCTION_CODES_ARM_H_