diff options
Diffstat (limited to 'chromium/v8/src/compiler/backend/ppc/instruction-codes-ppc.h')
-rw-r--r-- | chromium/v8/src/compiler/backend/ppc/instruction-codes-ppc.h | 816 |
1 files changed, 405 insertions, 411 deletions
diff --git a/chromium/v8/src/compiler/backend/ppc/instruction-codes-ppc.h b/chromium/v8/src/compiler/backend/ppc/instruction-codes-ppc.h index 64f532a52b0..4f9003257f6 100644 --- a/chromium/v8/src/compiler/backend/ppc/instruction-codes-ppc.h +++ b/chromium/v8/src/compiler/backend/ppc/instruction-codes-ppc.h @@ -11,417 +11,411 @@ namespace compiler { // PPC-specific opcodes that specify which assembly sequence to emit. // Most opcodes specify a single instruction. -#define TARGET_ARCH_OPCODE_LIST(V) \ - V(PPC_Peek) \ - V(PPC_Sync) \ - V(PPC_And) \ - V(PPC_AndComplement) \ - V(PPC_Or) \ - V(PPC_OrComplement) \ - V(PPC_Xor) \ - V(PPC_ShiftLeft32) \ - V(PPC_ShiftLeft64) \ - V(PPC_ShiftLeftPair) \ - V(PPC_ShiftRight32) \ - V(PPC_ShiftRight64) \ - V(PPC_ShiftRightPair) \ - V(PPC_ShiftRightAlg32) \ - V(PPC_ShiftRightAlg64) \ - V(PPC_ShiftRightAlgPair) \ - V(PPC_RotRight32) \ - V(PPC_RotRight64) \ - V(PPC_Not) \ - V(PPC_RotLeftAndMask32) \ - V(PPC_RotLeftAndClear64) \ - V(PPC_RotLeftAndClearLeft64) \ - V(PPC_RotLeftAndClearRight64) \ - V(PPC_Add32) \ - V(PPC_Add64) \ - V(PPC_AddWithOverflow32) \ - V(PPC_AddPair) \ - V(PPC_AddDouble) \ - V(PPC_Sub) \ - V(PPC_SubWithOverflow32) \ - V(PPC_SubPair) \ - V(PPC_SubDouble) \ - V(PPC_Mul32) \ - V(PPC_Mul32WithHigh32) \ - V(PPC_Mul64) \ - V(PPC_MulHigh32) \ - V(PPC_MulHighU32) \ - V(PPC_MulPair) \ - V(PPC_MulDouble) \ - V(PPC_Div32) \ - V(PPC_Div64) \ - V(PPC_DivU32) \ - V(PPC_DivU64) \ - V(PPC_DivDouble) \ - V(PPC_Mod32) \ - V(PPC_Mod64) \ - V(PPC_ModU32) \ - V(PPC_ModU64) \ - V(PPC_ModDouble) \ - V(PPC_Neg) \ - V(PPC_NegDouble) \ - V(PPC_SqrtDouble) \ - V(PPC_FloorDouble) \ - V(PPC_CeilDouble) \ - V(PPC_TruncateDouble) \ - V(PPC_RoundDouble) \ - V(PPC_MaxDouble) \ - V(PPC_MinDouble) \ - V(PPC_AbsDouble) \ - V(PPC_Cntlz32) \ - V(PPC_Cntlz64) \ - V(PPC_Popcnt32) \ - V(PPC_Popcnt64) \ - V(PPC_Cmp32) \ - V(PPC_Cmp64) \ - V(PPC_CmpDouble) \ - V(PPC_Tst32) \ - V(PPC_Tst64) \ - V(PPC_Push) \ - V(PPC_PushFrame) \ - V(PPC_StoreToStackSlot) \ - V(PPC_ExtendSignWord8) \ - V(PPC_ExtendSignWord16) \ - V(PPC_ExtendSignWord32) \ - V(PPC_Uint32ToUint64) \ - V(PPC_Int64ToInt32) \ - V(PPC_Int64ToFloat32) \ - V(PPC_Int64ToDouble) \ - V(PPC_Uint64ToFloat32) \ - V(PPC_Uint64ToDouble) \ - V(PPC_Int32ToFloat32) \ - V(PPC_Int32ToDouble) \ - V(PPC_Uint32ToFloat32) \ - V(PPC_Float32ToInt32) \ - V(PPC_Float32ToUint32) \ - V(PPC_Uint32ToDouble) \ - V(PPC_Float32ToDouble) \ - V(PPC_Float64SilenceNaN) \ - V(PPC_DoubleToInt32) \ - V(PPC_DoubleToUint32) \ - V(PPC_DoubleToInt64) \ - V(PPC_DoubleToUint64) \ - V(PPC_DoubleToFloat32) \ - V(PPC_DoubleExtractLowWord32) \ - V(PPC_DoubleExtractHighWord32) \ - V(PPC_DoubleInsertLowWord32) \ - V(PPC_DoubleInsertHighWord32) \ - V(PPC_DoubleConstruct) \ - V(PPC_BitcastInt32ToFloat32) \ - V(PPC_BitcastFloat32ToInt32) \ - V(PPC_BitcastInt64ToDouble) \ - V(PPC_BitcastDoubleToInt64) \ - V(PPC_LoadWordS8) \ - V(PPC_LoadWordU8) \ - V(PPC_LoadWordS16) \ - V(PPC_LoadWordU16) \ - V(PPC_LoadWordS32) \ - V(PPC_LoadWordU32) \ - V(PPC_LoadByteRev32) \ - V(PPC_LoadWord64) \ - V(PPC_LoadByteRev64) \ - V(PPC_LoadFloat32) \ - V(PPC_LoadDouble) \ - V(PPC_LoadSimd128) \ - V(PPC_LoadReverseSimd128RR) \ - V(PPC_StoreWord8) \ - V(PPC_StoreWord16) \ - V(PPC_StoreWord32) \ - V(PPC_StoreByteRev32) \ - V(PPC_StoreWord64) \ - V(PPC_StoreByteRev64) \ - V(PPC_StoreFloat32) \ - V(PPC_StoreDouble) \ - V(PPC_StoreSimd128) \ - V(PPC_ByteRev32) \ - V(PPC_ByteRev64) \ - V(PPC_CompressSigned) \ - V(PPC_CompressPointer) \ - V(PPC_CompressAny) \ - V(PPC_AtomicStoreUint8) \ - V(PPC_AtomicStoreUint16) \ - V(PPC_AtomicStoreWord32) \ - V(PPC_AtomicStoreWord64) \ - V(PPC_AtomicLoadUint8) \ - V(PPC_AtomicLoadUint16) \ - V(PPC_AtomicLoadWord32) \ - V(PPC_AtomicLoadWord64) \ - V(PPC_AtomicExchangeUint8) \ - V(PPC_AtomicExchangeUint16) \ - V(PPC_AtomicExchangeWord32) \ - V(PPC_AtomicExchangeWord64) \ - V(PPC_AtomicCompareExchangeUint8) \ - V(PPC_AtomicCompareExchangeUint16) \ - V(PPC_AtomicCompareExchangeWord32) \ - V(PPC_AtomicCompareExchangeWord64) \ - V(PPC_AtomicAddUint8) \ - V(PPC_AtomicAddUint16) \ - V(PPC_AtomicAddUint32) \ - V(PPC_AtomicAddUint64) \ - V(PPC_AtomicAddInt8) \ - V(PPC_AtomicAddInt16) \ - V(PPC_AtomicAddInt32) \ - V(PPC_AtomicAddInt64) \ - V(PPC_AtomicSubUint8) \ - V(PPC_AtomicSubUint16) \ - V(PPC_AtomicSubUint32) \ - V(PPC_AtomicSubUint64) \ - V(PPC_AtomicSubInt8) \ - V(PPC_AtomicSubInt16) \ - V(PPC_AtomicSubInt32) \ - V(PPC_AtomicSubInt64) \ - V(PPC_AtomicAndUint8) \ - V(PPC_AtomicAndUint16) \ - V(PPC_AtomicAndUint32) \ - V(PPC_AtomicAndUint64) \ - V(PPC_AtomicAndInt8) \ - V(PPC_AtomicAndInt16) \ - V(PPC_AtomicAndInt32) \ - V(PPC_AtomicAndInt64) \ - V(PPC_AtomicOrUint8) \ - V(PPC_AtomicOrUint16) \ - V(PPC_AtomicOrUint32) \ - V(PPC_AtomicOrUint64) \ - V(PPC_AtomicOrInt8) \ - V(PPC_AtomicOrInt16) \ - V(PPC_AtomicOrInt32) \ - V(PPC_AtomicOrInt64) \ - V(PPC_AtomicXorUint8) \ - V(PPC_AtomicXorUint16) \ - V(PPC_AtomicXorUint32) \ - V(PPC_AtomicXorUint64) \ - V(PPC_AtomicXorInt8) \ - V(PPC_AtomicXorInt16) \ - V(PPC_AtomicXorInt32) \ - V(PPC_AtomicXorInt64) \ - V(PPC_F64x2Splat) \ - V(PPC_F64x2ExtractLane) \ - V(PPC_F64x2ReplaceLane) \ - V(PPC_F64x2Add) \ - V(PPC_F64x2Sub) \ - V(PPC_F64x2Mul) \ - V(PPC_F64x2Eq) \ - V(PPC_F64x2Ne) \ - V(PPC_F64x2Le) \ - V(PPC_F64x2Lt) \ - V(PPC_F64x2Abs) \ - V(PPC_F64x2Neg) \ - V(PPC_F64x2Sqrt) \ - V(PPC_F64x2Qfma) \ - V(PPC_F64x2Qfms) \ - V(PPC_F64x2Div) \ - V(PPC_F64x2Min) \ - V(PPC_F64x2Max) \ - V(PPC_F64x2Ceil) \ - V(PPC_F64x2Floor) \ - V(PPC_F64x2Trunc) \ - V(PPC_F64x2Pmin) \ - V(PPC_F64x2Pmax) \ - V(PPC_F64x2ConvertLowI32x4S) \ - V(PPC_F64x2ConvertLowI32x4U) \ - V(PPC_F64x2PromoteLowF32x4) \ - V(PPC_F32x4Splat) \ - V(PPC_F32x4ExtractLane) \ - V(PPC_F32x4ReplaceLane) \ - V(PPC_F32x4Add) \ - V(PPC_F32x4Sub) \ - V(PPC_F32x4Mul) \ - V(PPC_F32x4Eq) \ - V(PPC_F32x4Ne) \ - V(PPC_F32x4Lt) \ - V(PPC_F32x4Le) \ - V(PPC_F32x4Abs) \ - V(PPC_F32x4Neg) \ - V(PPC_F32x4RecipApprox) \ - V(PPC_F32x4RecipSqrtApprox) \ - V(PPC_F32x4Sqrt) \ - V(PPC_F32x4SConvertI32x4) \ - V(PPC_F32x4UConvertI32x4) \ - V(PPC_F32x4Div) \ - V(PPC_F32x4Min) \ - V(PPC_F32x4Max) \ - V(PPC_F32x4Ceil) \ - V(PPC_F32x4Floor) \ - V(PPC_F32x4Trunc) \ - V(PPC_F32x4Pmin) \ - V(PPC_F32x4Pmax) \ - V(PPC_F32x4Qfma) \ - V(PPC_F32x4Qfms) \ - V(PPC_F32x4DemoteF64x2Zero) \ - V(PPC_I64x2Splat) \ - V(PPC_I64x2ExtractLane) \ - V(PPC_I64x2ReplaceLane) \ - V(PPC_I64x2Add) \ - V(PPC_I64x2Sub) \ - V(PPC_I64x2Mul) \ - V(PPC_I64x2Eq) \ - V(PPC_I64x2Ne) \ - V(PPC_I64x2GtS) \ - V(PPC_I64x2GeS) \ - V(PPC_I64x2Shl) \ - V(PPC_I64x2ShrS) \ - V(PPC_I64x2ShrU) \ - V(PPC_I64x2Neg) \ - V(PPC_I64x2BitMask) \ - V(PPC_I64x2SConvertI32x4Low) \ - V(PPC_I64x2SConvertI32x4High) \ - V(PPC_I64x2UConvertI32x4Low) \ - V(PPC_I64x2UConvertI32x4High) \ - V(PPC_I64x2ExtMulLowI32x4S) \ - V(PPC_I64x2ExtMulHighI32x4S) \ - V(PPC_I64x2ExtMulLowI32x4U) \ - V(PPC_I64x2ExtMulHighI32x4U) \ - V(PPC_I64x2Abs) \ - V(PPC_I32x4Splat) \ - V(PPC_I32x4ExtractLane) \ - V(PPC_I32x4ReplaceLane) \ - V(PPC_I32x4Add) \ - V(PPC_I32x4Sub) \ - V(PPC_I32x4Mul) \ - V(PPC_I32x4MinS) \ - V(PPC_I32x4MinU) \ - V(PPC_I32x4MaxS) \ - V(PPC_I32x4MaxU) \ - V(PPC_I32x4Eq) \ - V(PPC_I32x4Ne) \ - V(PPC_I32x4GtS) \ - V(PPC_I32x4GeS) \ - V(PPC_I32x4GtU) \ - V(PPC_I32x4GeU) \ - V(PPC_I32x4Shl) \ - V(PPC_I32x4ShrS) \ - V(PPC_I32x4ShrU) \ - V(PPC_I32x4Neg) \ - V(PPC_I32x4Abs) \ - V(PPC_I32x4SConvertF32x4) \ - V(PPC_I32x4UConvertF32x4) \ - V(PPC_I32x4SConvertI16x8Low) \ - V(PPC_I32x4SConvertI16x8High) \ - V(PPC_I32x4UConvertI16x8Low) \ - V(PPC_I32x4UConvertI16x8High) \ - V(PPC_I32x4BitMask) \ - V(PPC_I32x4DotI16x8S) \ - V(PPC_I32x4ExtAddPairwiseI16x8S) \ - V(PPC_I32x4ExtAddPairwiseI16x8U) \ - V(PPC_I32x4ExtMulLowI16x8S) \ - V(PPC_I32x4ExtMulHighI16x8S) \ - V(PPC_I32x4ExtMulLowI16x8U) \ - V(PPC_I32x4ExtMulHighI16x8U) \ - V(PPC_I32x4TruncSatF64x2SZero) \ - V(PPC_I32x4TruncSatF64x2UZero) \ - V(PPC_I16x8Splat) \ - V(PPC_I16x8ExtractLaneU) \ - V(PPC_I16x8ExtractLaneS) \ - V(PPC_I16x8ReplaceLane) \ - V(PPC_I16x8Add) \ - V(PPC_I16x8Sub) \ - V(PPC_I16x8Mul) \ - V(PPC_I16x8MinS) \ - V(PPC_I16x8MinU) \ - V(PPC_I16x8MaxS) \ - V(PPC_I16x8MaxU) \ - V(PPC_I16x8Eq) \ - V(PPC_I16x8Ne) \ - V(PPC_I16x8GtS) \ - V(PPC_I16x8GeS) \ - V(PPC_I16x8GtU) \ - V(PPC_I16x8GeU) \ - V(PPC_I16x8Shl) \ - V(PPC_I16x8ShrS) \ - V(PPC_I16x8ShrU) \ - V(PPC_I16x8Neg) \ - V(PPC_I16x8Abs) \ - V(PPC_I16x8SConvertI32x4) \ - V(PPC_I16x8UConvertI32x4) \ - V(PPC_I16x8SConvertI8x16Low) \ - V(PPC_I16x8SConvertI8x16High) \ - V(PPC_I16x8UConvertI8x16Low) \ - V(PPC_I16x8UConvertI8x16High) \ - V(PPC_I16x8AddSatS) \ - V(PPC_I16x8SubSatS) \ - V(PPC_I16x8AddSatU) \ - V(PPC_I16x8SubSatU) \ - V(PPC_I16x8RoundingAverageU) \ - V(PPC_I16x8BitMask) \ - V(PPC_I16x8ExtAddPairwiseI8x16S) \ - V(PPC_I16x8ExtAddPairwiseI8x16U) \ - V(PPC_I16x8Q15MulRSatS) \ - V(PPC_I16x8ExtMulLowI8x16S) \ - V(PPC_I16x8ExtMulHighI8x16S) \ - V(PPC_I16x8ExtMulLowI8x16U) \ - V(PPC_I16x8ExtMulHighI8x16U) \ - V(PPC_I8x16Splat) \ - V(PPC_I8x16ExtractLaneU) \ - V(PPC_I8x16ExtractLaneS) \ - V(PPC_I8x16ReplaceLane) \ - V(PPC_I8x16Add) \ - V(PPC_I8x16Sub) \ - V(PPC_I8x16MinS) \ - V(PPC_I8x16MinU) \ - V(PPC_I8x16MaxS) \ - V(PPC_I8x16MaxU) \ - V(PPC_I8x16Eq) \ - V(PPC_I8x16Ne) \ - V(PPC_I8x16GtS) \ - V(PPC_I8x16GeS) \ - V(PPC_I8x16GtU) \ - V(PPC_I8x16GeU) \ - V(PPC_I8x16Shl) \ - V(PPC_I8x16ShrS) \ - V(PPC_I8x16ShrU) \ - V(PPC_I8x16Neg) \ - V(PPC_I8x16Abs) \ - V(PPC_I8x16SConvertI16x8) \ - V(PPC_I8x16UConvertI16x8) \ - V(PPC_I8x16AddSatS) \ - V(PPC_I8x16SubSatS) \ - V(PPC_I8x16AddSatU) \ - V(PPC_I8x16SubSatU) \ - V(PPC_I8x16RoundingAverageU) \ - V(PPC_I8x16Shuffle) \ - V(PPC_I8x16Swizzle) \ - V(PPC_I8x16BitMask) \ - V(PPC_I8x16Popcnt) \ - V(PPC_I64x2AllTrue) \ - V(PPC_I32x4AllTrue) \ - V(PPC_I16x8AllTrue) \ - V(PPC_I8x16AllTrue) \ - V(PPC_V128AnyTrue) \ - V(PPC_S128And) \ - V(PPC_S128Or) \ - V(PPC_S128Xor) \ - V(PPC_S128Const) \ - V(PPC_S128Zero) \ - V(PPC_S128AllOnes) \ - V(PPC_S128Not) \ - V(PPC_S128Select) \ - V(PPC_S128AndNot) \ - V(PPC_S128Load8Splat) \ - V(PPC_S128Load16Splat) \ - V(PPC_S128Load32Splat) \ - V(PPC_S128Load64Splat) \ - V(PPC_S128Load8x8S) \ - V(PPC_S128Load8x8U) \ - V(PPC_S128Load16x4S) \ - V(PPC_S128Load16x4U) \ - V(PPC_S128Load32x2S) \ - V(PPC_S128Load32x2U) \ - V(PPC_S128Load32Zero) \ - V(PPC_S128Load64Zero) \ - V(PPC_S128Load8Lane) \ - V(PPC_S128Load16Lane) \ - V(PPC_S128Load32Lane) \ - V(PPC_S128Load64Lane) \ - V(PPC_S128Store8Lane) \ - V(PPC_S128Store16Lane) \ - V(PPC_S128Store32Lane) \ - V(PPC_S128Store64Lane) \ - V(PPC_StoreCompressTagged) \ - V(PPC_LoadDecompressTaggedSigned) \ - V(PPC_LoadDecompressTaggedPointer) \ + +// Opcodes that support a MemoryAccessMode. +#define TARGET_ARCH_OPCODE_WITH_MEMORY_ACCESS_MODE_LIST(V) // None. + +#define TARGET_ARCH_OPCODE_LIST(V) \ + TARGET_ARCH_OPCODE_WITH_MEMORY_ACCESS_MODE_LIST(V) \ + V(PPC_Peek) \ + V(PPC_Sync) \ + V(PPC_And) \ + V(PPC_AndComplement) \ + V(PPC_Or) \ + V(PPC_OrComplement) \ + V(PPC_Xor) \ + V(PPC_ShiftLeft32) \ + V(PPC_ShiftLeft64) \ + V(PPC_ShiftLeftPair) \ + V(PPC_ShiftRight32) \ + V(PPC_ShiftRight64) \ + V(PPC_ShiftRightPair) \ + V(PPC_ShiftRightAlg32) \ + V(PPC_ShiftRightAlg64) \ + V(PPC_ShiftRightAlgPair) \ + V(PPC_RotRight32) \ + V(PPC_RotRight64) \ + V(PPC_Not) \ + V(PPC_RotLeftAndMask32) \ + V(PPC_RotLeftAndClear64) \ + V(PPC_RotLeftAndClearLeft64) \ + V(PPC_RotLeftAndClearRight64) \ + V(PPC_Add32) \ + V(PPC_Add64) \ + V(PPC_AddWithOverflow32) \ + V(PPC_AddPair) \ + V(PPC_AddDouble) \ + V(PPC_Sub) \ + V(PPC_SubWithOverflow32) \ + V(PPC_SubPair) \ + V(PPC_SubDouble) \ + V(PPC_Mul32) \ + V(PPC_Mul32WithHigh32) \ + V(PPC_Mul64) \ + V(PPC_MulHigh32) \ + V(PPC_MulHighU32) \ + V(PPC_MulPair) \ + V(PPC_MulDouble) \ + V(PPC_Div32) \ + V(PPC_Div64) \ + V(PPC_DivU32) \ + V(PPC_DivU64) \ + V(PPC_DivDouble) \ + V(PPC_Mod32) \ + V(PPC_Mod64) \ + V(PPC_ModU32) \ + V(PPC_ModU64) \ + V(PPC_ModDouble) \ + V(PPC_Neg) \ + V(PPC_NegDouble) \ + V(PPC_SqrtDouble) \ + V(PPC_FloorDouble) \ + V(PPC_CeilDouble) \ + V(PPC_TruncateDouble) \ + V(PPC_RoundDouble) \ + V(PPC_MaxDouble) \ + V(PPC_MinDouble) \ + V(PPC_AbsDouble) \ + V(PPC_Cntlz32) \ + V(PPC_Cntlz64) \ + V(PPC_Popcnt32) \ + V(PPC_Popcnt64) \ + V(PPC_Cmp32) \ + V(PPC_Cmp64) \ + V(PPC_CmpDouble) \ + V(PPC_Tst32) \ + V(PPC_Tst64) \ + V(PPC_Push) \ + V(PPC_PushFrame) \ + V(PPC_StoreToStackSlot) \ + V(PPC_ExtendSignWord8) \ + V(PPC_ExtendSignWord16) \ + V(PPC_ExtendSignWord32) \ + V(PPC_Uint32ToUint64) \ + V(PPC_Int64ToInt32) \ + V(PPC_Int64ToFloat32) \ + V(PPC_Int64ToDouble) \ + V(PPC_Uint64ToFloat32) \ + V(PPC_Uint64ToDouble) \ + V(PPC_Int32ToFloat32) \ + V(PPC_Int32ToDouble) \ + V(PPC_Uint32ToFloat32) \ + V(PPC_Float32ToInt32) \ + V(PPC_Float32ToUint32) \ + V(PPC_Uint32ToDouble) \ + V(PPC_Float32ToDouble) \ + V(PPC_Float64SilenceNaN) \ + V(PPC_DoubleToInt32) \ + V(PPC_DoubleToUint32) \ + V(PPC_DoubleToInt64) \ + V(PPC_DoubleToUint64) \ + V(PPC_DoubleToFloat32) \ + V(PPC_DoubleExtractLowWord32) \ + V(PPC_DoubleExtractHighWord32) \ + V(PPC_DoubleInsertLowWord32) \ + V(PPC_DoubleInsertHighWord32) \ + V(PPC_DoubleConstruct) \ + V(PPC_BitcastInt32ToFloat32) \ + V(PPC_BitcastFloat32ToInt32) \ + V(PPC_BitcastInt64ToDouble) \ + V(PPC_BitcastDoubleToInt64) \ + V(PPC_LoadWordS8) \ + V(PPC_LoadWordU8) \ + V(PPC_LoadWordS16) \ + V(PPC_LoadWordU16) \ + V(PPC_LoadWordS32) \ + V(PPC_LoadWordU32) \ + V(PPC_LoadByteRev32) \ + V(PPC_LoadWord64) \ + V(PPC_LoadByteRev64) \ + V(PPC_LoadFloat32) \ + V(PPC_LoadDouble) \ + V(PPC_LoadSimd128) \ + V(PPC_LoadReverseSimd128RR) \ + V(PPC_StoreWord8) \ + V(PPC_StoreWord16) \ + V(PPC_StoreWord32) \ + V(PPC_StoreByteRev32) \ + V(PPC_StoreWord64) \ + V(PPC_StoreByteRev64) \ + V(PPC_StoreFloat32) \ + V(PPC_StoreDouble) \ + V(PPC_StoreSimd128) \ + V(PPC_ByteRev32) \ + V(PPC_ByteRev64) \ + V(PPC_AtomicExchangeUint8) \ + V(PPC_AtomicExchangeUint16) \ + V(PPC_AtomicExchangeWord32) \ + V(PPC_AtomicExchangeWord64) \ + V(PPC_AtomicCompareExchangeUint8) \ + V(PPC_AtomicCompareExchangeUint16) \ + V(PPC_AtomicCompareExchangeWord32) \ + V(PPC_AtomicCompareExchangeWord64) \ + V(PPC_AtomicAddUint8) \ + V(PPC_AtomicAddUint16) \ + V(PPC_AtomicAddUint32) \ + V(PPC_AtomicAddUint64) \ + V(PPC_AtomicAddInt8) \ + V(PPC_AtomicAddInt16) \ + V(PPC_AtomicAddInt32) \ + V(PPC_AtomicAddInt64) \ + V(PPC_AtomicSubUint8) \ + V(PPC_AtomicSubUint16) \ + V(PPC_AtomicSubUint32) \ + V(PPC_AtomicSubUint64) \ + V(PPC_AtomicSubInt8) \ + V(PPC_AtomicSubInt16) \ + V(PPC_AtomicSubInt32) \ + V(PPC_AtomicSubInt64) \ + V(PPC_AtomicAndUint8) \ + V(PPC_AtomicAndUint16) \ + V(PPC_AtomicAndUint32) \ + V(PPC_AtomicAndUint64) \ + V(PPC_AtomicAndInt8) \ + V(PPC_AtomicAndInt16) \ + V(PPC_AtomicAndInt32) \ + V(PPC_AtomicAndInt64) \ + V(PPC_AtomicOrUint8) \ + V(PPC_AtomicOrUint16) \ + V(PPC_AtomicOrUint32) \ + V(PPC_AtomicOrUint64) \ + V(PPC_AtomicOrInt8) \ + V(PPC_AtomicOrInt16) \ + V(PPC_AtomicOrInt32) \ + V(PPC_AtomicOrInt64) \ + V(PPC_AtomicXorUint8) \ + V(PPC_AtomicXorUint16) \ + V(PPC_AtomicXorUint32) \ + V(PPC_AtomicXorUint64) \ + V(PPC_AtomicXorInt8) \ + V(PPC_AtomicXorInt16) \ + V(PPC_AtomicXorInt32) \ + V(PPC_AtomicXorInt64) \ + V(PPC_F64x2Splat) \ + V(PPC_F64x2ExtractLane) \ + V(PPC_F64x2ReplaceLane) \ + V(PPC_F64x2Add) \ + V(PPC_F64x2Sub) \ + V(PPC_F64x2Mul) \ + V(PPC_F64x2Eq) \ + V(PPC_F64x2Ne) \ + V(PPC_F64x2Le) \ + V(PPC_F64x2Lt) \ + V(PPC_F64x2Abs) \ + V(PPC_F64x2Neg) \ + V(PPC_F64x2Sqrt) \ + V(PPC_F64x2Qfma) \ + V(PPC_F64x2Qfms) \ + V(PPC_F64x2Div) \ + V(PPC_F64x2Min) \ + V(PPC_F64x2Max) \ + V(PPC_F64x2Ceil) \ + V(PPC_F64x2Floor) \ + V(PPC_F64x2Trunc) \ + V(PPC_F64x2Pmin) \ + V(PPC_F64x2Pmax) \ + V(PPC_F64x2ConvertLowI32x4S) \ + V(PPC_F64x2ConvertLowI32x4U) \ + V(PPC_F64x2PromoteLowF32x4) \ + V(PPC_F32x4Splat) \ + V(PPC_F32x4ExtractLane) \ + V(PPC_F32x4ReplaceLane) \ + V(PPC_F32x4Add) \ + V(PPC_F32x4Sub) \ + V(PPC_F32x4Mul) \ + V(PPC_F32x4Eq) \ + V(PPC_F32x4Ne) \ + V(PPC_F32x4Lt) \ + V(PPC_F32x4Le) \ + V(PPC_F32x4Abs) \ + V(PPC_F32x4Neg) \ + V(PPC_F32x4RecipApprox) \ + V(PPC_F32x4RecipSqrtApprox) \ + V(PPC_F32x4Sqrt) \ + V(PPC_F32x4SConvertI32x4) \ + V(PPC_F32x4UConvertI32x4) \ + V(PPC_F32x4Div) \ + V(PPC_F32x4Min) \ + V(PPC_F32x4Max) \ + V(PPC_F32x4Ceil) \ + V(PPC_F32x4Floor) \ + V(PPC_F32x4Trunc) \ + V(PPC_F32x4Pmin) \ + V(PPC_F32x4Pmax) \ + V(PPC_F32x4Qfma) \ + V(PPC_F32x4Qfms) \ + V(PPC_F32x4DemoteF64x2Zero) \ + V(PPC_I64x2Splat) \ + V(PPC_I64x2ExtractLane) \ + V(PPC_I64x2ReplaceLane) \ + V(PPC_I64x2Add) \ + V(PPC_I64x2Sub) \ + V(PPC_I64x2Mul) \ + V(PPC_I64x2Eq) \ + V(PPC_I64x2Ne) \ + V(PPC_I64x2GtS) \ + V(PPC_I64x2GeS) \ + V(PPC_I64x2Shl) \ + V(PPC_I64x2ShrS) \ + V(PPC_I64x2ShrU) \ + V(PPC_I64x2Neg) \ + V(PPC_I64x2BitMask) \ + V(PPC_I64x2SConvertI32x4Low) \ + V(PPC_I64x2SConvertI32x4High) \ + V(PPC_I64x2UConvertI32x4Low) \ + V(PPC_I64x2UConvertI32x4High) \ + V(PPC_I64x2ExtMulLowI32x4S) \ + V(PPC_I64x2ExtMulHighI32x4S) \ + V(PPC_I64x2ExtMulLowI32x4U) \ + V(PPC_I64x2ExtMulHighI32x4U) \ + V(PPC_I64x2Abs) \ + V(PPC_I32x4Splat) \ + V(PPC_I32x4ExtractLane) \ + V(PPC_I32x4ReplaceLane) \ + V(PPC_I32x4Add) \ + V(PPC_I32x4Sub) \ + V(PPC_I32x4Mul) \ + V(PPC_I32x4MinS) \ + V(PPC_I32x4MinU) \ + V(PPC_I32x4MaxS) \ + V(PPC_I32x4MaxU) \ + V(PPC_I32x4Eq) \ + V(PPC_I32x4Ne) \ + V(PPC_I32x4GtS) \ + V(PPC_I32x4GeS) \ + V(PPC_I32x4GtU) \ + V(PPC_I32x4GeU) \ + V(PPC_I32x4Shl) \ + V(PPC_I32x4ShrS) \ + V(PPC_I32x4ShrU) \ + V(PPC_I32x4Neg) \ + V(PPC_I32x4Abs) \ + V(PPC_I32x4SConvertF32x4) \ + V(PPC_I32x4UConvertF32x4) \ + V(PPC_I32x4SConvertI16x8Low) \ + V(PPC_I32x4SConvertI16x8High) \ + V(PPC_I32x4UConvertI16x8Low) \ + V(PPC_I32x4UConvertI16x8High) \ + V(PPC_I32x4BitMask) \ + V(PPC_I32x4DotI16x8S) \ + V(PPC_I32x4ExtAddPairwiseI16x8S) \ + V(PPC_I32x4ExtAddPairwiseI16x8U) \ + V(PPC_I32x4ExtMulLowI16x8S) \ + V(PPC_I32x4ExtMulHighI16x8S) \ + V(PPC_I32x4ExtMulLowI16x8U) \ + V(PPC_I32x4ExtMulHighI16x8U) \ + V(PPC_I32x4TruncSatF64x2SZero) \ + V(PPC_I32x4TruncSatF64x2UZero) \ + V(PPC_I16x8Splat) \ + V(PPC_I16x8ExtractLaneU) \ + V(PPC_I16x8ExtractLaneS) \ + V(PPC_I16x8ReplaceLane) \ + V(PPC_I16x8Add) \ + V(PPC_I16x8Sub) \ + V(PPC_I16x8Mul) \ + V(PPC_I16x8MinS) \ + V(PPC_I16x8MinU) \ + V(PPC_I16x8MaxS) \ + V(PPC_I16x8MaxU) \ + V(PPC_I16x8Eq) \ + V(PPC_I16x8Ne) \ + V(PPC_I16x8GtS) \ + V(PPC_I16x8GeS) \ + V(PPC_I16x8GtU) \ + V(PPC_I16x8GeU) \ + V(PPC_I16x8Shl) \ + V(PPC_I16x8ShrS) \ + V(PPC_I16x8ShrU) \ + V(PPC_I16x8Neg) \ + V(PPC_I16x8Abs) \ + V(PPC_I16x8SConvertI32x4) \ + V(PPC_I16x8UConvertI32x4) \ + V(PPC_I16x8SConvertI8x16Low) \ + V(PPC_I16x8SConvertI8x16High) \ + V(PPC_I16x8UConvertI8x16Low) \ + V(PPC_I16x8UConvertI8x16High) \ + V(PPC_I16x8AddSatS) \ + V(PPC_I16x8SubSatS) \ + V(PPC_I16x8AddSatU) \ + V(PPC_I16x8SubSatU) \ + V(PPC_I16x8RoundingAverageU) \ + V(PPC_I16x8BitMask) \ + V(PPC_I16x8ExtAddPairwiseI8x16S) \ + V(PPC_I16x8ExtAddPairwiseI8x16U) \ + V(PPC_I16x8Q15MulRSatS) \ + V(PPC_I16x8ExtMulLowI8x16S) \ + V(PPC_I16x8ExtMulHighI8x16S) \ + V(PPC_I16x8ExtMulLowI8x16U) \ + V(PPC_I16x8ExtMulHighI8x16U) \ + V(PPC_I8x16Splat) \ + V(PPC_I8x16ExtractLaneU) \ + V(PPC_I8x16ExtractLaneS) \ + V(PPC_I8x16ReplaceLane) \ + V(PPC_I8x16Add) \ + V(PPC_I8x16Sub) \ + V(PPC_I8x16MinS) \ + V(PPC_I8x16MinU) \ + V(PPC_I8x16MaxS) \ + V(PPC_I8x16MaxU) \ + V(PPC_I8x16Eq) \ + V(PPC_I8x16Ne) \ + V(PPC_I8x16GtS) \ + V(PPC_I8x16GeS) \ + V(PPC_I8x16GtU) \ + V(PPC_I8x16GeU) \ + V(PPC_I8x16Shl) \ + V(PPC_I8x16ShrS) \ + V(PPC_I8x16ShrU) \ + V(PPC_I8x16Neg) \ + V(PPC_I8x16Abs) \ + V(PPC_I8x16SConvertI16x8) \ + V(PPC_I8x16UConvertI16x8) \ + V(PPC_I8x16AddSatS) \ + V(PPC_I8x16SubSatS) \ + V(PPC_I8x16AddSatU) \ + V(PPC_I8x16SubSatU) \ + V(PPC_I8x16RoundingAverageU) \ + V(PPC_I8x16Shuffle) \ + V(PPC_I8x16Swizzle) \ + V(PPC_I8x16BitMask) \ + V(PPC_I8x16Popcnt) \ + V(PPC_I64x2AllTrue) \ + V(PPC_I32x4AllTrue) \ + V(PPC_I16x8AllTrue) \ + V(PPC_I8x16AllTrue) \ + V(PPC_V128AnyTrue) \ + V(PPC_S128And) \ + V(PPC_S128Or) \ + V(PPC_S128Xor) \ + V(PPC_S128Const) \ + V(PPC_S128Zero) \ + V(PPC_S128AllOnes) \ + V(PPC_S128Not) \ + V(PPC_S128Select) \ + V(PPC_S128AndNot) \ + V(PPC_S128Load8Splat) \ + V(PPC_S128Load16Splat) \ + V(PPC_S128Load32Splat) \ + V(PPC_S128Load64Splat) \ + V(PPC_S128Load8x8S) \ + V(PPC_S128Load8x8U) \ + V(PPC_S128Load16x4S) \ + V(PPC_S128Load16x4U) \ + V(PPC_S128Load32x2S) \ + V(PPC_S128Load32x2U) \ + V(PPC_S128Load32Zero) \ + V(PPC_S128Load64Zero) \ + V(PPC_S128Load8Lane) \ + V(PPC_S128Load16Lane) \ + V(PPC_S128Load32Lane) \ + V(PPC_S128Load64Lane) \ + V(PPC_S128Store8Lane) \ + V(PPC_S128Store16Lane) \ + V(PPC_S128Store32Lane) \ + V(PPC_S128Store64Lane) \ + V(PPC_StoreCompressTagged) \ + V(PPC_LoadDecompressTaggedSigned) \ + V(PPC_LoadDecompressTaggedPointer) \ V(PPC_LoadDecompressAnyTagged) // Addressing modes represent the "shape" of inputs to an instruction. |