|  | // Copyright 2014 the V8 project authors. All rights reserved. | 
|  | // Use of this source code is governed by a BSD-style license that can be | 
|  | // found in the LICENSE file. | 
|  |  | 
|  | #ifndef V8_COMPILER_ARM64_INSTRUCTION_CODES_ARM64_H_ | 
|  | #define V8_COMPILER_ARM64_INSTRUCTION_CODES_ARM64_H_ | 
|  |  | 
|  | namespace v8 { | 
|  | namespace internal { | 
|  | namespace compiler { | 
|  |  | 
|  | // ARM64-specific opcodes that specify which assembly sequence to emit. | 
|  | // Most opcodes specify a single instruction. | 
|  | #define TARGET_ARCH_OPCODE_LIST(V)          \ | 
|  | V(Arm64Add)                               \ | 
|  | V(Arm64Add32)                             \ | 
|  | V(Arm64And)                               \ | 
|  | V(Arm64And32)                             \ | 
|  | V(Arm64Bic)                               \ | 
|  | V(Arm64Bic32)                             \ | 
|  | V(Arm64Clz)                               \ | 
|  | V(Arm64Clz32)                             \ | 
|  | V(Arm64Cmp)                               \ | 
|  | V(Arm64Cmp32)                             \ | 
|  | V(Arm64Cmn)                               \ | 
|  | V(Arm64Cmn32)                             \ | 
|  | V(Arm64Tst)                               \ | 
|  | V(Arm64Tst32)                             \ | 
|  | V(Arm64Or)                                \ | 
|  | V(Arm64Or32)                              \ | 
|  | V(Arm64Orn)                               \ | 
|  | V(Arm64Orn32)                             \ | 
|  | V(Arm64Eor)                               \ | 
|  | V(Arm64Eor32)                             \ | 
|  | V(Arm64Eon)                               \ | 
|  | V(Arm64Eon32)                             \ | 
|  | V(Arm64Sub)                               \ | 
|  | V(Arm64Sub32)                             \ | 
|  | V(Arm64Mul)                               \ | 
|  | V(Arm64Mul32)                             \ | 
|  | V(Arm64Smull)                             \ | 
|  | V(Arm64Umull)                             \ | 
|  | V(Arm64Madd)                              \ | 
|  | V(Arm64Madd32)                            \ | 
|  | V(Arm64Msub)                              \ | 
|  | V(Arm64Msub32)                            \ | 
|  | V(Arm64Mneg)                              \ | 
|  | V(Arm64Mneg32)                            \ | 
|  | V(Arm64Idiv)                              \ | 
|  | V(Arm64Idiv32)                            \ | 
|  | V(Arm64Udiv)                              \ | 
|  | V(Arm64Udiv32)                            \ | 
|  | V(Arm64Imod)                              \ | 
|  | V(Arm64Imod32)                            \ | 
|  | V(Arm64Umod)                              \ | 
|  | V(Arm64Umod32)                            \ | 
|  | V(Arm64Not)                               \ | 
|  | V(Arm64Not32)                             \ | 
|  | V(Arm64Lsl)                               \ | 
|  | V(Arm64Lsl32)                             \ | 
|  | V(Arm64Lsr)                               \ | 
|  | V(Arm64Lsr32)                             \ | 
|  | V(Arm64Asr)                               \ | 
|  | V(Arm64Asr32)                             \ | 
|  | V(Arm64Ror)                               \ | 
|  | V(Arm64Ror32)                             \ | 
|  | V(Arm64Mov32)                             \ | 
|  | V(Arm64Sxtb32)                            \ | 
|  | V(Arm64Sxth32)                            \ | 
|  | V(Arm64Sxtb)                              \ | 
|  | V(Arm64Sxth)                              \ | 
|  | V(Arm64Sxtw)                              \ | 
|  | V(Arm64Sbfx32)                            \ | 
|  | V(Arm64Ubfx)                              \ | 
|  | V(Arm64Ubfx32)                            \ | 
|  | V(Arm64Ubfiz32)                           \ | 
|  | V(Arm64Bfi)                               \ | 
|  | V(Arm64Rbit)                              \ | 
|  | V(Arm64Rbit32)                            \ | 
|  | V(Arm64TestAndBranch32)                   \ | 
|  | V(Arm64TestAndBranch)                     \ | 
|  | V(Arm64CompareAndBranch32)                \ | 
|  | V(Arm64CompareAndBranch)                  \ | 
|  | V(Arm64Claim)                             \ | 
|  | V(Arm64Poke)                              \ | 
|  | V(Arm64PokePair)                          \ | 
|  | V(Arm64Peek)                              \ | 
|  | V(Arm64Float32Cmp)                        \ | 
|  | V(Arm64Float32Add)                        \ | 
|  | V(Arm64Float32Sub)                        \ | 
|  | V(Arm64Float32Mul)                        \ | 
|  | V(Arm64Float32Div)                        \ | 
|  | V(Arm64Float32Abs)                        \ | 
|  | V(Arm64Float32Neg)                        \ | 
|  | V(Arm64Float32Sqrt)                       \ | 
|  | V(Arm64Float32RoundDown)                  \ | 
|  | V(Arm64Float32Max)                        \ | 
|  | V(Arm64Float32Min)                        \ | 
|  | V(Arm64Float64Cmp)                        \ | 
|  | V(Arm64Float64Add)                        \ | 
|  | V(Arm64Float64Sub)                        \ | 
|  | V(Arm64Float64Mul)                        \ | 
|  | V(Arm64Float64Div)                        \ | 
|  | V(Arm64Float64Mod)                        \ | 
|  | V(Arm64Float64Max)                        \ | 
|  | V(Arm64Float64Min)                        \ | 
|  | V(Arm64Float64Abs)                        \ | 
|  | V(Arm64Float64Neg)                        \ | 
|  | V(Arm64Float64Sqrt)                       \ | 
|  | V(Arm64Float64RoundDown)                  \ | 
|  | V(Arm64Float32RoundUp)                    \ | 
|  | V(Arm64Float64RoundUp)                    \ | 
|  | V(Arm64Float64RoundTiesAway)              \ | 
|  | V(Arm64Float32RoundTruncate)              \ | 
|  | V(Arm64Float64RoundTruncate)              \ | 
|  | V(Arm64Float32RoundTiesEven)              \ | 
|  | V(Arm64Float64RoundTiesEven)              \ | 
|  | V(Arm64Float64SilenceNaN)                 \ | 
|  | V(Arm64Float32ToFloat64)                  \ | 
|  | V(Arm64Float64ToFloat32)                  \ | 
|  | V(Arm64Float32ToInt32)                    \ | 
|  | V(Arm64Float64ToInt32)                    \ | 
|  | V(Arm64Float32ToUint32)                   \ | 
|  | V(Arm64Float64ToUint32)                   \ | 
|  | V(Arm64Float32ToInt64)                    \ | 
|  | V(Arm64Float64ToInt64)                    \ | 
|  | V(Arm64Float32ToUint64)                   \ | 
|  | V(Arm64Float64ToUint64)                   \ | 
|  | V(Arm64Int32ToFloat32)                    \ | 
|  | V(Arm64Int32ToFloat64)                    \ | 
|  | V(Arm64Int64ToFloat32)                    \ | 
|  | V(Arm64Int64ToFloat64)                    \ | 
|  | V(Arm64Uint32ToFloat32)                   \ | 
|  | V(Arm64Uint32ToFloat64)                   \ | 
|  | V(Arm64Uint64ToFloat32)                   \ | 
|  | V(Arm64Uint64ToFloat64)                   \ | 
|  | V(Arm64Float64ExtractLowWord32)           \ | 
|  | V(Arm64Float64ExtractHighWord32)          \ | 
|  | V(Arm64Float64InsertLowWord32)            \ | 
|  | V(Arm64Float64InsertHighWord32)           \ | 
|  | V(Arm64Float64MoveU64)                    \ | 
|  | V(Arm64U64MoveFloat64)                    \ | 
|  | V(Arm64LdrS)                              \ | 
|  | V(Arm64StrS)                              \ | 
|  | V(Arm64LdrD)                              \ | 
|  | V(Arm64StrD)                              \ | 
|  | V(Arm64LdrQ)                              \ | 
|  | V(Arm64StrQ)                              \ | 
|  | V(Arm64Ldrb)                              \ | 
|  | V(Arm64Ldrsb)                             \ | 
|  | V(Arm64Strb)                              \ | 
|  | V(Arm64Ldrh)                              \ | 
|  | V(Arm64Ldrsh)                             \ | 
|  | V(Arm64Strh)                              \ | 
|  | V(Arm64Ldrsw)                             \ | 
|  | V(Arm64LdrW)                              \ | 
|  | V(Arm64StrW)                              \ | 
|  | V(Arm64Ldr)                               \ | 
|  | V(Arm64Str)                               \ | 
|  | V(Arm64DsbIsb)                            \ | 
|  | V(Arm64F32x4Splat)                        \ | 
|  | V(Arm64F32x4ExtractLane)                  \ | 
|  | V(Arm64F32x4ReplaceLane)                  \ | 
|  | V(Arm64F32x4SConvertI32x4)                \ | 
|  | V(Arm64F32x4UConvertI32x4)                \ | 
|  | V(Arm64F32x4Abs)                          \ | 
|  | V(Arm64F32x4Neg)                          \ | 
|  | V(Arm64F32x4RecipApprox)                  \ | 
|  | V(Arm64F32x4RecipSqrtApprox)              \ | 
|  | V(Arm64F32x4Add)                          \ | 
|  | V(Arm64F32x4AddHoriz)                     \ | 
|  | V(Arm64F32x4Sub)                          \ | 
|  | V(Arm64F32x4Mul)                          \ | 
|  | V(Arm64F32x4Min)                          \ | 
|  | V(Arm64F32x4Max)                          \ | 
|  | V(Arm64F32x4Eq)                           \ | 
|  | V(Arm64F32x4Ne)                           \ | 
|  | V(Arm64F32x4Lt)                           \ | 
|  | V(Arm64F32x4Le)                           \ | 
|  | V(Arm64I32x4Splat)                        \ | 
|  | V(Arm64I32x4ExtractLane)                  \ | 
|  | V(Arm64I32x4ReplaceLane)                  \ | 
|  | V(Arm64I32x4SConvertF32x4)                \ | 
|  | V(Arm64I32x4SConvertI16x8Low)             \ | 
|  | V(Arm64I32x4SConvertI16x8High)            \ | 
|  | V(Arm64I32x4Neg)                          \ | 
|  | V(Arm64I32x4Shl)                          \ | 
|  | V(Arm64I32x4ShrS)                         \ | 
|  | V(Arm64I32x4Add)                          \ | 
|  | V(Arm64I32x4AddHoriz)                     \ | 
|  | V(Arm64I32x4Sub)                          \ | 
|  | V(Arm64I32x4Mul)                          \ | 
|  | V(Arm64I32x4MinS)                         \ | 
|  | V(Arm64I32x4MaxS)                         \ | 
|  | V(Arm64I32x4Eq)                           \ | 
|  | V(Arm64I32x4Ne)                           \ | 
|  | V(Arm64I32x4GtS)                          \ | 
|  | V(Arm64I32x4GeS)                          \ | 
|  | V(Arm64I32x4UConvertF32x4)                \ | 
|  | V(Arm64I32x4UConvertI16x8Low)             \ | 
|  | V(Arm64I32x4UConvertI16x8High)            \ | 
|  | V(Arm64I32x4ShrU)                         \ | 
|  | V(Arm64I32x4MinU)                         \ | 
|  | V(Arm64I32x4MaxU)                         \ | 
|  | V(Arm64I32x4GtU)                          \ | 
|  | V(Arm64I32x4GeU)                          \ | 
|  | V(Arm64I16x8Splat)                        \ | 
|  | V(Arm64I16x8ExtractLane)                  \ | 
|  | V(Arm64I16x8ReplaceLane)                  \ | 
|  | V(Arm64I16x8SConvertI8x16Low)             \ | 
|  | V(Arm64I16x8SConvertI8x16High)            \ | 
|  | V(Arm64I16x8Neg)                          \ | 
|  | V(Arm64I16x8Shl)                          \ | 
|  | V(Arm64I16x8ShrS)                         \ | 
|  | V(Arm64I16x8SConvertI32x4)                \ | 
|  | V(Arm64I16x8Add)                          \ | 
|  | V(Arm64I16x8AddSaturateS)                 \ | 
|  | V(Arm64I16x8AddHoriz)                     \ | 
|  | V(Arm64I16x8Sub)                          \ | 
|  | V(Arm64I16x8SubSaturateS)                 \ | 
|  | V(Arm64I16x8Mul)                          \ | 
|  | V(Arm64I16x8MinS)                         \ | 
|  | V(Arm64I16x8MaxS)                         \ | 
|  | V(Arm64I16x8Eq)                           \ | 
|  | V(Arm64I16x8Ne)                           \ | 
|  | V(Arm64I16x8GtS)                          \ | 
|  | V(Arm64I16x8GeS)                          \ | 
|  | V(Arm64I16x8UConvertI8x16Low)             \ | 
|  | V(Arm64I16x8UConvertI8x16High)            \ | 
|  | V(Arm64I16x8ShrU)                         \ | 
|  | V(Arm64I16x8UConvertI32x4)                \ | 
|  | V(Arm64I16x8AddSaturateU)                 \ | 
|  | V(Arm64I16x8SubSaturateU)                 \ | 
|  | V(Arm64I16x8MinU)                         \ | 
|  | V(Arm64I16x8MaxU)                         \ | 
|  | V(Arm64I16x8GtU)                          \ | 
|  | V(Arm64I16x8GeU)                          \ | 
|  | V(Arm64I8x16Splat)                        \ | 
|  | V(Arm64I8x16ExtractLane)                  \ | 
|  | V(Arm64I8x16ReplaceLane)                  \ | 
|  | V(Arm64I8x16Neg)                          \ | 
|  | V(Arm64I8x16Shl)                          \ | 
|  | V(Arm64I8x16ShrS)                         \ | 
|  | V(Arm64I8x16SConvertI16x8)                \ | 
|  | V(Arm64I8x16Add)                          \ | 
|  | V(Arm64I8x16AddSaturateS)                 \ | 
|  | V(Arm64I8x16Sub)                          \ | 
|  | V(Arm64I8x16SubSaturateS)                 \ | 
|  | V(Arm64I8x16Mul)                          \ | 
|  | V(Arm64I8x16MinS)                         \ | 
|  | V(Arm64I8x16MaxS)                         \ | 
|  | V(Arm64I8x16Eq)                           \ | 
|  | V(Arm64I8x16Ne)                           \ | 
|  | V(Arm64I8x16GtS)                          \ | 
|  | V(Arm64I8x16GeS)                          \ | 
|  | V(Arm64I8x16ShrU)                         \ | 
|  | V(Arm64I8x16UConvertI16x8)                \ | 
|  | V(Arm64I8x16AddSaturateU)                 \ | 
|  | V(Arm64I8x16SubSaturateU)                 \ | 
|  | V(Arm64I8x16MinU)                         \ | 
|  | V(Arm64I8x16MaxU)                         \ | 
|  | V(Arm64I8x16GtU)                          \ | 
|  | V(Arm64I8x16GeU)                          \ | 
|  | V(Arm64S128Zero)                          \ | 
|  | V(Arm64S128Dup)                           \ | 
|  | V(Arm64S128And)                           \ | 
|  | V(Arm64S128Or)                            \ | 
|  | V(Arm64S128Xor)                           \ | 
|  | V(Arm64S128Not)                           \ | 
|  | V(Arm64S128Select)                        \ | 
|  | V(Arm64S32x4ZipLeft)                      \ | 
|  | V(Arm64S32x4ZipRight)                     \ | 
|  | V(Arm64S32x4UnzipLeft)                    \ | 
|  | V(Arm64S32x4UnzipRight)                   \ | 
|  | V(Arm64S32x4TransposeLeft)                \ | 
|  | V(Arm64S32x4TransposeRight)               \ | 
|  | V(Arm64S32x4Shuffle)                      \ | 
|  | V(Arm64S16x8ZipLeft)                      \ | 
|  | V(Arm64S16x8ZipRight)                     \ | 
|  | V(Arm64S16x8UnzipLeft)                    \ | 
|  | V(Arm64S16x8UnzipRight)                   \ | 
|  | V(Arm64S16x8TransposeLeft)                \ | 
|  | V(Arm64S16x8TransposeRight)               \ | 
|  | V(Arm64S8x16ZipLeft)                      \ | 
|  | V(Arm64S8x16ZipRight)                     \ | 
|  | V(Arm64S8x16UnzipLeft)                    \ | 
|  | V(Arm64S8x16UnzipRight)                   \ | 
|  | V(Arm64S8x16TransposeLeft)                \ | 
|  | V(Arm64S8x16TransposeRight)               \ | 
|  | V(Arm64S8x16Concat)                       \ | 
|  | V(Arm64S8x16Shuffle)                      \ | 
|  | V(Arm64S32x2Reverse)                      \ | 
|  | V(Arm64S16x4Reverse)                      \ | 
|  | V(Arm64S16x2Reverse)                      \ | 
|  | V(Arm64S8x8Reverse)                       \ | 
|  | V(Arm64S8x4Reverse)                       \ | 
|  | V(Arm64S8x2Reverse)                       \ | 
|  | V(Arm64S1x4AnyTrue)                       \ | 
|  | V(Arm64S1x4AllTrue)                       \ | 
|  | V(Arm64S1x8AnyTrue)                       \ | 
|  | V(Arm64S1x8AllTrue)                       \ | 
|  | V(Arm64S1x16AnyTrue)                      \ | 
|  | V(Arm64S1x16AllTrue)                      \ | 
|  | V(Arm64Word64AtomicLoadUint8)             \ | 
|  | V(Arm64Word64AtomicLoadUint16)            \ | 
|  | V(Arm64Word64AtomicLoadUint32)            \ | 
|  | V(Arm64Word64AtomicLoadUint64)            \ | 
|  | V(Arm64Word64AtomicStoreWord8)            \ | 
|  | V(Arm64Word64AtomicStoreWord16)           \ | 
|  | V(Arm64Word64AtomicStoreWord32)           \ | 
|  | V(Arm64Word64AtomicStoreWord64)           \ | 
|  | V(Arm64Word64AtomicAddUint8)              \ | 
|  | V(Arm64Word64AtomicAddUint16)             \ | 
|  | V(Arm64Word64AtomicAddUint32)             \ | 
|  | V(Arm64Word64AtomicAddUint64)             \ | 
|  | V(Arm64Word64AtomicSubUint8)              \ | 
|  | V(Arm64Word64AtomicSubUint16)             \ | 
|  | V(Arm64Word64AtomicSubUint32)             \ | 
|  | V(Arm64Word64AtomicSubUint64)             \ | 
|  | V(Arm64Word64AtomicAndUint8)              \ | 
|  | V(Arm64Word64AtomicAndUint16)             \ | 
|  | V(Arm64Word64AtomicAndUint32)             \ | 
|  | V(Arm64Word64AtomicAndUint64)             \ | 
|  | V(Arm64Word64AtomicOrUint8)               \ | 
|  | V(Arm64Word64AtomicOrUint16)              \ | 
|  | V(Arm64Word64AtomicOrUint32)              \ | 
|  | V(Arm64Word64AtomicOrUint64)              \ | 
|  | V(Arm64Word64AtomicXorUint8)              \ | 
|  | V(Arm64Word64AtomicXorUint16)             \ | 
|  | V(Arm64Word64AtomicXorUint32)             \ | 
|  | V(Arm64Word64AtomicXorUint64)             \ | 
|  | V(Arm64Word64AtomicExchangeUint8)         \ | 
|  | V(Arm64Word64AtomicExchangeUint16)        \ | 
|  | V(Arm64Word64AtomicExchangeUint32)        \ | 
|  | V(Arm64Word64AtomicExchangeUint64)        \ | 
|  | V(Arm64Word64AtomicCompareExchangeUint8)  \ | 
|  | V(Arm64Word64AtomicCompareExchangeUint16) \ | 
|  | V(Arm64Word64AtomicCompareExchangeUint32) \ | 
|  | V(Arm64Word64AtomicCompareExchangeUint64) | 
|  |  | 
|  | // Addressing modes represent the "shape" of inputs to an instruction. | 
|  | // Many instructions support multiple addressing modes. Addressing modes | 
|  | // are encoded into the InstructionCode of the instruction and tell the | 
|  | // code generator after register allocation which assembler method to call. | 
|  | // | 
|  | // We use the following local notation for addressing modes: | 
|  | // | 
|  | // R = register | 
|  | // O = register or stack slot | 
|  | // D = double register | 
|  | // I = immediate (handle, external, int32) | 
|  | // MRI = [register + immediate] | 
|  | // MRR = [register + register] | 
|  | #define TARGET_ADDRESSING_MODE_LIST(V)                          \ | 
|  | V(MRI)              /* [%r0 + K] */                           \ | 
|  | V(MRR)              /* [%r0 + %r1] */                         \ | 
|  | V(Operand2_R_LSL_I) /* %r0 LSL K */                           \ | 
|  | V(Operand2_R_LSR_I) /* %r0 LSR K */                           \ | 
|  | V(Operand2_R_ASR_I) /* %r0 ASR K */                           \ | 
|  | V(Operand2_R_ROR_I) /* %r0 ROR K */                           \ | 
|  | V(Operand2_R_UXTB)  /* %r0 UXTB (unsigned extend byte) */     \ | 
|  | V(Operand2_R_UXTH)  /* %r0 UXTH (unsigned extend halfword) */ \ | 
|  | V(Operand2_R_SXTB)  /* %r0 SXTB (signed extend byte) */       \ | 
|  | V(Operand2_R_SXTH)  /* %r0 SXTH (signed extend halfword) */   \ | 
|  | V(Operand2_R_SXTW)  /* %r0 SXTW (signed extend word) */       \ | 
|  | V(Root)             /* [%rr + K] */ | 
|  |  | 
|  | }  // namespace compiler | 
|  | }  // namespace internal | 
|  | }  // namespace v8 | 
|  |  | 
|  | #endif  // V8_COMPILER_ARM64_INSTRUCTION_CODES_ARM64_H_ |