AdvSimd Classe

Définition

Important

Cette API n’est pas conforme CLS.

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

public ref class AdvSimd abstract : System::Runtime::Intrinsics::Arm::ArmBase
[System.CLSCompliant(false)]
public abstract class AdvSimd : System.Runtime.Intrinsics.Arm.ArmBase
[<System.CLSCompliant(false)>]
type AdvSimd = class
    inherit ArmBase
Public MustInherit Class AdvSimd
Inherits ArmBase
Héritage
AdvSimd
Dérivé
Attributs

Propriétés

IsSupported

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

Méthodes

Abs(Vector128<Int16>)

int16x8_t vabsq_s16 (int16x8_t a)

A32 : VABS.S16 Qd, Qm

A64 : ABS Vd.8H, Vn.8H

Abs(Vector128<Int32>)

int32x4_t vabsq_s32 (int32x4_t a)

A32 : VABS.S32 Qd, Qm

A64 : ABS Vd.4S, Vn.4S

Abs(Vector128<SByte>)

int8x16_t vabsq_s8 (int8x16_t a)

A32 : VABS.S8 Qd, Qm

A64 : ABS Vd.16B, Vn.16B

Abs(Vector128<Single>)

float32x4_t vabsq_f32 (float32x4_t a)

A32 : VABS.F32 Qd, Qm

A64 : FABS Vd.4S, Vn.4S

Abs(Vector64<Int16>)

int16x4_t vabs_s16 (int16x4_t a)

A32 : VABS.S16 Dd, Dm

A64 : ABS Vd.4H, Vn.4H

Abs(Vector64<Int32>)

int32x2_t vabs_s32 (int32x2_t a)

A32 : VABS.S32 Dd, Dm

A64 : ABS Vd.2S, Vn.2S

Abs(Vector64<SByte>)

int8x8_t vabs_s8 (int8x8_t a)

A32 : VABS.S8 Dd, Dm

A64 : ABS Vd.8B, Vn.8B

Abs(Vector64<Single>)

float32x2_t vabs_f32 (float32x2_t a)

A32 : VABS.F32 Dd, Dm

A64 : FABS Vd.2S, Vn.2S

AbsoluteCompareGreaterThan(Vector128<Single>, Vector128<Single>)

uint32x4_t vcagtq_f32 (float32x4_t a, float32x4_t b)

A32 : VACGT.F32 Qd, Qn, Qm

A64 : FACGT Vd.4S, Vn.4S, Vm.4S

AbsoluteCompareGreaterThan(Vector64<Single>, Vector64<Single>)

uint32x2_t vcagt_f32 (float32x2_t a, float32x2_t b)

A32 : VACGT.F32 Dd, Dn, Dm

A64 : FACGT Vd.2S, Vn.2S, Vm.2S

AbsoluteCompareGreaterThanOrEqual(Vector128<Single>, Vector128<Single>)

uint32x4_t vcageq_f32 (float32x4_t a, float32x4_t b)

A32 : VACGE.F32 Qd, Qn, Qm

A64 : FACGE Vd.4S, Vn.4S, Vm.4S

AbsoluteCompareGreaterThanOrEqual(Vector64<Single>, Vector64<Single>)

uint32x2_t vcage_f32 (float32x2_t a, float32x2_t b)

A32 : VACGE.F32 Dd, Dn, Dm

A64 : FACGE Vd.2S, Vn.2S, Vm.2S

AbsoluteCompareLessThan(Vector128<Single>, Vector128<Single>)

uint32x4_t vcaltq_f32 (float32x4_t a, float32x4_t b)

A32 : VACLT.F32 Qd, Qn, Qm

A64 : FACGT Vd.4S, Vn.4S, Vm.4S

AbsoluteCompareLessThan(Vector64<Single>, Vector64<Single>)

uint32x2_t vcalt_f32 (float32x2_t a, float32x2_t b)

A32 : VACLT.F32 Dd, Dn, Dm

A64 : FACGT Vd.2S, Vn.2S, Vm.2S

AbsoluteCompareLessThanOrEqual(Vector128<Single>, Vector128<Single>)

uint32x4_t vcaleq_f32 (float32x4_t a, float32x4_t b)

A32 : VACLE.F32 Qd, Qn, Qm

A64 : FACGE Vd.4S, Vn.4S, Vm.4S

AbsoluteCompareLessThanOrEqual(Vector64<Single>, Vector64<Single>)

uint32x2_t vcale_f32 (float32x2_t a, float32x2_t b)

A32 : VACLE.F32 Dd, Dn, Dm

A64 : FACGE Vd.2S, Vn.2S, Vm.2S

AbsoluteDifference(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vabdq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VABD.U8 Qd, Qn, Qm

A64 : UABD Vd.16B, Vn.16B, Vm.16B

AbsoluteDifference(Vector128<Int16>, Vector128<Int16>)

int16x8_t vabdq_s16 (int16x8_t a, int16x8_t b)

A32 : VABD.S16 Qd, Qn, Qm

A64 : SABD Vd.8H, Vn.8H, Vm.8H

AbsoluteDifference(Vector128<Int32>, Vector128<Int32>)

int32x4_t vabdq_s32 (int32x4_t a, int32x4_t b)

A32 : VABD.S32 Qd, Qn, Qm

A64 : SABD Vd.4S, Vn.4S, Vm.4S

AbsoluteDifference(Vector128<SByte>, Vector128<SByte>)

int8x16_t vabdq_s8 (int8x16_t a, int8x16_t b)

A32 : VABD.S8 Qd, Qn, Qm

A64 : SABD Vd.16B, Vn.16B, Vm.16B

AbsoluteDifference(Vector128<Single>, Vector128<Single>)

float32x4_t vabdq_f32 (float32x4_t a, float32x4_t b)

A32 : VABD.F32 Qd, Qn, Qm

A64 : FABD Vd.4S, Vn.4S, Vm.4S

AbsoluteDifference(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vabdq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VABD.U16 Qd, Qn, Qm

A64 : UABD Vd.8H, Vn.8H, Vm.8H

AbsoluteDifference(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vabdq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VABD.U32 Qd, Qn, Qm

A64 : UABD Vd.4S, Vn.4S, Vm.4S

AbsoluteDifference(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vabd_u8 (uint8x8_t a, uint8x8_t b)

A32 : VABD.U8 Dd, Dn, Dm

A64 : UABD Vd.8B, Vn.8B, Vm.8B

AbsoluteDifference(Vector64<Int16>, Vector64<Int16>)

int16x4_t vabd_s16 (int16x4_t a, int16x4_t b)

A32 : VABD.S16 Dd, Dn, Dm

A64 : SABD Vd.4H, Vn.4H, Vm.4H

AbsoluteDifference(Vector64<Int32>, Vector64<Int32>)

int32x2_t vabd_s32 (int32x2_t a, int32x2_t b)

A32 : VABD.S32 Dd, Dn, Dm

A64 : SABD Vd.2S, Vn.2S, Vm.2S

AbsoluteDifference(Vector64<SByte>, Vector64<SByte>)

int8x8_t vabd_s8 (int8x8_t a, int8x8_t b)

A32 : VABD.S8 Dd, Dn, Dm

A64 : SABD Vd.8B, Vn.8B, Vm.8B

AbsoluteDifference(Vector64<Single>, Vector64<Single>)

float32x2_t vabd_f32 (float32x2_t a, float32x2_t b)

A32 : VABD.F32 Dd, Dn, Dm

A64 : FABD Vd.2S, Vn.2S, Vm.2S

AbsoluteDifference(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vabd_u16 (uint16x4_t a, uint16x4_t b)

A32 : VABD.U16 Dd, Dn, Dm

A64 : UABD Vd.4H, Vn.4H, Vm.4H

AbsoluteDifference(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vabd_u32 (uint32x2_t a, uint32x2_t b)

A32 : VABD.U32 Dd, Dn, Dm

A64 : UABD Vd.2S, Vn.2S, Vm.2S

AbsoluteDifferenceAdd(Vector128<Byte>, Vector128<Byte>, Vector128<Byte>)

uint8x16_t vabaq_u8 (uint8x16_t a, uint8x16_t b, uint8x16_t c)

A32 : VABA.U8 Qd, Qn, Qm

A64 : UABA Vd.16B, Vn.16B, Vm.16B

AbsoluteDifferenceAdd(Vector128<Int16>, Vector128<Int16>, Vector128<Int16>)

int16x8_t vabaq_s16 (int16x8_t a, int16x8_t b, int16x8_t c)

A32 : VABA.S16 Qd, Qn, Qm

A64 : SABA Vd.8H, Vn.8H, Vm.8H

AbsoluteDifferenceAdd(Vector128<Int32>, Vector128<Int32>, Vector128<Int32>)

int32x4_t vabaq_s32 (int32x4_t a, int32x4_t b, int32x4_t c)

A32 : VABA.S32 Qd, Qn, Qm

A64 : SABA Vd.4S, Vn.4S, Vm.4S

AbsoluteDifferenceAdd(Vector128<SByte>, Vector128<SByte>, Vector128<SByte>)

int8x16_t vabaq_s8 (int8x16_t a, int8x16_t b, int8x16_t c)

A32 : VABA.S8 Qd, Qn, Qm

A64 : SABA Vd.16B, Vn.16B, Vm.16B

AbsoluteDifferenceAdd(Vector128<UInt16>, Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vabaq_u16 (uint16x8_t a, uint16x8_t b, uint16x8_t c)

A32 : VABA.U16 Qd, Qn, Qm

A64 : UABA Vd.8H, Vn.8H, Vm.8H

AbsoluteDifferenceAdd(Vector128<UInt32>, Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vabaq_u32 (uint32x4_t a, uint32x4_t b, uint32x4_t c)

A32 : VABA.U32 Qd, Qn, Qm

A64 : UABA Vd.4S, Vn.4S, Vm.4S

AbsoluteDifferenceAdd(Vector64<Byte>, Vector64<Byte>, Vector64<Byte>)

uint8x8_t vaba_u8 (uint8x8_t a, uint8x8_t b, uint8x8_t c)

A32 : VABA.U8 Dd, Dn, Dm

A64 : UABA Vd.8B, Vn.8B, Vm.8B

AbsoluteDifferenceAdd(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>)

int16x4_t vaba_s16 (int16x4_t a, int16x4_t b, int16x4_t c)

A32 : VABA.S16 Dd, Dn, Dm

A64 : SABA Vd.4H, Vn.4H, Vm.4H

AbsoluteDifferenceAdd(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>)

int32x2_t vaba_s32 (int32x2_t a, int32x2_t b, int32x2_t c)

A32 : VABA.S32 Dd, Dn, Dm

A64 : SABA Vd.2S, Vn.2S, Vm.2S

AbsoluteDifferenceAdd(Vector64<SByte>, Vector64<SByte>, Vector64<SByte>)

int8x8_t vaba_s8 (int8x8_t a, int8x8_t b, int8x8_t c)

A32 : VABA.S8 Dd, Dn, Dm

A64 : SABA Vd.8B, Vn.8B, Vm.8B

AbsoluteDifferenceAdd(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vaba_u16 (uint16x4_t a, uint16x4_t b, uint16x4_t c)

A32 : VABA.U16 Dd, Dn, Dm

A64 : UABA Vd.4H, Vn.4H, Vm.4H

AbsoluteDifferenceAdd(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vaba_u32 (uint32x2_t a, uint32x2_t b, uint32x2_t c)

A32 : VABA.U32 Dd, Dn, Dm

A64 : UABA Vd.2S, Vn.2S, Vm.2S

AbsoluteDifferenceWideningLower(Vector64<Byte>, Vector64<Byte>)

uint16x8_t vabdl_u8 (uint8x8_t a, uint8x8_t b)

A32 : VABDL.U8 Qd, Dn, Dm

A64 : UABDL Vd.8H, Vn.8B, Vm.8B

AbsoluteDifferenceWideningLower(Vector64<Int16>, Vector64<Int16>)

int32x4_t vabdl_s16 (int16x4_t a, int16x4_t b)

A32 : VABDL.S16 Qd, Dn, Dm

A64 : SABDL Vd.4S, Vn.4H, Vm.4H

AbsoluteDifferenceWideningLower(Vector64<Int32>, Vector64<Int32>)

int64x2_t vabdl_s32 (int32x2_t a, int32x2_t b)

A32 : VABDL.S32 Qd, Dn, Dm

A64 : SABDL Vd.2D, Vn.2S, Vm.2S

AbsoluteDifferenceWideningLower(Vector64<SByte>, Vector64<SByte>)

int16x8_t vabdl_s8 (int8x8_t a, int8x8_t b)

A32 : VABDL.S8 Qd, Dn, Dm

A64 : SABDL Vd.8H, Vn.8B, Vm.8B

AbsoluteDifferenceWideningLower(Vector64<UInt16>, Vector64<UInt16>)

uint32x4_t vabdl_u16 (uint16x4_t a, uint16x4_t b)

A32 : VABDL.U16 Qd, Dn, Dm

A64 : UABDL Vd.4S, Vn.4H, Vm.4H

AbsoluteDifferenceWideningLower(Vector64<UInt32>, Vector64<UInt32>)

uint64x2_t vabdl_u32 (uint32x2_t a, uint32x2_t b)

A32 : VABDL.U32 Qd, Dn, Dm

A64 : UABDL Vd.2D, Vn.2S, Vm.2S

AbsoluteDifferenceWideningLowerAndAdd(Vector128<Int16>, Vector64<SByte>, Vector64<SByte>)

int16x8_t vabal_s8 (int16x8_t a, int8x8_t b, int8x8_t c)

A32 : VABAL.S8 Qd, Dn, Dm

A64 : SABAL Vd.8H, Vn.8B, Vm.8B

AbsoluteDifferenceWideningLowerAndAdd(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>)

int32x4_t vabal_s16 (int32x4_t a, int16x4_t b, int16x4_t c)

A32 : VABAL.S16 Qd, Dn, Dm

A64 : SABAL Vd.4S, Vn.4H, Vm.4H

AbsoluteDifferenceWideningLowerAndAdd(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>)

int64x2_t vabal_s32 (int64x2_t a, int32x2_t b, int32x2_t c)

A32 : VABAL.S32 Qd, Dn, Dm

A64 : SABAL Vd.2D, Vn.2S, Vm.2S

AbsoluteDifferenceWideningLowerAndAdd(Vector128<UInt16>, Vector64<Byte>, Vector64<Byte>)

uint16x8_t vabal_u8 (uint16x8_t a, uint8x8_t b, uint8x8_t c)

A32 : VABAL.U8 Qd, Dn, Dm

A64 : UABAL Vd.8H, Vn.8B, Vm.8B

AbsoluteDifferenceWideningLowerAndAdd(Vector128<UInt32>, Vector64<UInt16>, Vector64<UInt16>)

uint32x4_t vabal_u16 (uint32x4_t a, uint16x4_t b, uint16x4_t c)

A32 : VABAL.U16 Qd, Dn, Dm

A64 : UABAL Vd.4S, Vn.4H, Vm.4H

AbsoluteDifferenceWideningLowerAndAdd(Vector128<UInt64>, Vector64<UInt32>, Vector64<UInt32>)

uint64x2_t vabal_u32 (uint64x2_t a, uint32x2_t b, uint32x2_t c)

A32 : VABAL.U32 Qd, Dn, Dm

A64 : UABAL Vd.2D, Vn.2S, Vm.2S

AbsoluteDifferenceWideningUpper(Vector128<Byte>, Vector128<Byte>)

uint16x8_t vabdl_high_u8 (uint8x16_t a, uint8x16_t b)

A32 : VABDL.U8 Qd, Dn+1, Dm+1

A64 : UABDL2 Vd.8H, Vn.16B, Vm.16B

AbsoluteDifferenceWideningUpper(Vector128<Int16>, Vector128<Int16>)

int32x4_t vabdl_high_s16 (int16x8_t a, int16x8_t b)

A32 : VABDL.S16 Qd, Dn+1, Dm+1

A64 : SABDL2 Vd.4S, Vn.8H, Vm.8H

AbsoluteDifferenceWideningUpper(Vector128<Int32>, Vector128<Int32>)

int64x2_t vabdl_high_s32 (int32x4_t a, int32x4_t b)

A32 : VABDL.S32 Qd, Dn+1, Dm+1

A64 : SABDL2 Vd.2D, Vn.4S, Vm.4S

AbsoluteDifferenceWideningUpper(Vector128<SByte>, Vector128<SByte>)

int16x8_t vabdl_high_s8 (int8x16_t a, int8x16_t b)

A32 : VABDL.S8 Qd, Dn+1, Dm+1

A64 : SABDL2 Vd.8H, Vn.16B, Vm.16B

AbsoluteDifferenceWideningUpper(Vector128<UInt16>, Vector128<UInt16>)

uint32x4_t vabdl_high_u16 (uint16x8_t a, uint16x8_t b)

A32 : VABDL.U16 Qd, Dn+1, Dm+1

A64 : UABDL2 Vd.4S, Vn.8H, Vm.8H

AbsoluteDifferenceWideningUpper(Vector128<UInt32>, Vector128<UInt32>)

uint64x2_t vabdl_high_u32 (uint32x4_t a, uint32x4_t b)

A32 : VABDL.U32 Qd, Dn+1, Dm+1

A64 : UABDL2 Vd.2D, Vn.4S, Vm.4S

AbsoluteDifferenceWideningUpperAndAdd(Vector128<Int16>, Vector128<SByte>, Vector128<SByte>)

int16x8_t vabal_high_s8 (int16x8_t a, int8x16_t b, int8x16_t c)

A32 : VABAL.S8 Qd, Dn+1, Dm+1

A64 : SABAL2 Vd.8H, Vn.16B, Vm.16B

AbsoluteDifferenceWideningUpperAndAdd(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>)

int32x4_t vabal_high_s16 (int32x4_t a, int16x8_t b, int16x8_t c)

A32 : VABAL.S16 Qd, Dn+1, Dm+1

A64 : SABAL2 Vd.4S, Vn.8H, Vm.8H

AbsoluteDifferenceWideningUpperAndAdd(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>)

int64x2_t vabal_high_s32 (int64x2_t a, int32x4_t b, int32x4_t c)

A32 : VABAL.S32 Qd, Dn+1, Dm+1

A64 : SABAL2 Vd.2D, Vn.4S, Vm.4S

AbsoluteDifferenceWideningUpperAndAdd(Vector128<UInt16>, Vector128<Byte>, Vector128<Byte>)

uint16x8_t vabal_high_u8 (uint16x8_t a, uint8x16_t b, uint8x16_t c)

A32 : VABAL.U8 Qd, Dn+1, Dm+1

A64 : UABAL2 Vd.8H, Vn.16B, Vm.16B

AbsoluteDifferenceWideningUpperAndAdd(Vector128<UInt32>, Vector128<UInt16>, Vector128<UInt16>)

uint32x4_t vabal_high_u16 (uint32x4_t a, uint16x8_t b, uint16x8_t c)

A32 : VABAL.U16 Qd, Dn+1, Dm+1

A64 : UABAL2 Vd.4S, Vn.8H, Vm.8H

AbsoluteDifferenceWideningUpperAndAdd(Vector128<UInt64>, Vector128<UInt32>, Vector128<UInt32>)

uint64x2_t vabal_high_u32 (uint64x2_t a, uint32x4_t b, uint32x4_t c)

A32 : VABAL.U32 Qd, Dn+1, Dm+1

A64 : UABAL2 Vd.2D, Vn.4S, Vm.4S

AbsSaturate(Vector128<Int16>)

int16x8_t vqabsq_s16 (int16x8_t a)

A32 : VQABS.S16 Qd, Qm

A64 : SQABS Vd.8H, Vn.8H

AbsSaturate(Vector128<Int32>)

int32x4_t vqabsq_s32 (int32x4_t a)

A32 : VQABS.S32 Qd, Qm

A64 : SQABS Vd.4S, Vn.4S

AbsSaturate(Vector128<SByte>)

int8x16_t vqabsq_s8 (int8x16_t a)

A32 : VQABS.S8 Qd, Qm

A64 : SQABS Vd.16B, Vn.16B

AbsSaturate(Vector64<Int16>)

int16x4_t vqabs_s16 (int16x4_t a)

A32 : VQABS.S16 Dd, Dm

A64 : SQABS Vd.4H, Vn.4H

AbsSaturate(Vector64<Int32>)

int32x2_t vqabs_s32 (int32x2_t a)

A32 : VQABS.S32 Dd, Dm

A64 : SQABS Vd.2S, Vn.2S

AbsSaturate(Vector64<SByte>)

int8x8_t vqabs_s8 (int8x8_t a)

A32 : VQABS.S8 Dd, Dm

A64 : SQABS Vd.8B, Vn.8B

AbsScalar(Vector64<Double>)

float64x1_t vabs_f64 (float64x1_t a)

A32 : VABS.F64 Dd, Dm

A64 : FABS Dd, Dn

AbsScalar(Vector64<Single>)

float32_t vabss_f32 (float32_t a)

A32 : VABS.F32 Sd, Sm

A64 : FABS Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Add(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vaddq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VADD.I8 Qd, Qn, Qm

A64 : ADD Vd.16B, Vn.16B, Vm.16B

Add(Vector128<Int16>, Vector128<Int16>)

int16x8_t vaddq_s16 (int16x8_t a, int16x8_t b)

A32 : VADD.I16 Qd, Qn, Qm

A64 : ADD Vd.8H, Vn.8H, Vm.8H

Add(Vector128<Int32>, Vector128<Int32>)

int32x4_t vaddq_s32 (int32x4_t a, int32x4_t b)

A32 : VADD.I32 Qd, Qn, Qm

A64 : ADD Vd.4S, Vn.4S, Vm.4S

Add(Vector128<Int64>, Vector128<Int64>)

int64x2_t vaddq_s64 (int64x2_t a, int64x2_t b)

A32 : VADD.I64 Qd, Qn, Qm

A64 : ADD Vd.2D, Vn.2D, Vm.2D

Add(Vector128<SByte>, Vector128<SByte>)

int8x16_t vaddq_s8 (int8x16_t a, int8x16_t b)

A32 : VADD.I8 Qd, Qn, Qm

A64 : ADD Vd.16B, Vn.16B, Vm.16B

Add(Vector128<Single>, Vector128<Single>)

float32x4_t vaddq_f32 (float32x4_t a, float32x4_t b)

A32 : VADD.F32 Qd, Qn, Qm

A64 : FADD Vd.4S, Vn.4S, Vm.4S

Add(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vaddq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VADD.I16 Qd, Qn, Qm

A64 : ADD Vd.8H, Vn.8H, Vm.8H

Add(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vaddq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VADD.I32 Qd, Qn, Qm

A64 : ADD Vd.4S, Vn.4S, Vm.4S

Add(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vaddq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VADD.I64 Qd, Qn, Qm

A64 : ADD Vd.2D, Vn.2D, Vm.2D

Add(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vadd_u8 (uint8x8_t a, uint8x8_t b)

A32 : VADD.I8 Dd, Dn, Dm

A64 : ADD Vd.8B, Vn.8B, Vm.8B

Add(Vector64<Int16>, Vector64<Int16>)

int16x4_t vadd_s16 (int16x4_t a, int16x4_t b)

A32 : VADD.I16 Dd, Dn, Dm

A64 : ADD Vd.4H, Vn.4H, Vm.4H

Add(Vector64<Int32>, Vector64<Int32>)

int32x2_t vadd_s32 (int32x2_t a, int32x2_t b)

A32 : VADD.I32 Dd, Dn, Dm

A64 : ADD Vd.2S, Vn.2S, Vm.2S

Add(Vector64<SByte>, Vector64<SByte>)

int8x8_t vadd_s8 (int8x8_t a, int8x8_t b)

A32 : VADD.I8 Dd, Dn, Dm

A64 : ADD Vd.8B, Vn.8B, Vm.8B

Add(Vector64<Single>, Vector64<Single>)

float32x2_t vadd_f32 (float32x2_t a, float32x2_t b)

A32 : VADD.F32 Dd, Dn, Dm

A64 : FADD Vd.2S, Vn.2S, Vm.2S

Add(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vadd_u16 (uint16x4_t a, uint16x4_t b)

A32 : VADD.I16 Dd, Dn, Dm

A64 : ADD Vd.4H, Vn.4H, Vm.4H

Add(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vadd_u32 (uint32x2_t a, uint32x2_t b)

A32 : VADD.I32 Dd, Dn, Dm

A64 : ADD Vd.2S, Vn.2S, Vm.2S

AddHighNarrowingLower(Vector128<Int16>, Vector128<Int16>)

int8x8_t vaddhn_s16 (int16x8_t a, int16x8_t b)

A32 : VADDHN.I16 Dd, Qn, Qm

A64 : ADDHN Vd.8B, Vn.8H, Vm.8H

AddHighNarrowingLower(Vector128<Int32>, Vector128<Int32>)

int16x4_t vaddhn_s32 (int32x4_t a, int32x4_t b)

A32 : VADDHN.I32 Dd, Qn, Qm

A64 : ADDHN Vd.4H, Vn.4S, Vm.4S

AddHighNarrowingLower(Vector128<Int64>, Vector128<Int64>)

int32x2_t vaddhn_s64 (int64x2_t a, int64x2_t b)

A32 : VADDHN.I64 Dd, Qn, Qm

A64 : ADDHN Vd.2S, Vn.2D, Vm.2D

AddHighNarrowingLower(Vector128<UInt16>, Vector128<UInt16>)

uint8x8_t vaddhn_u16 (uint16x8_t a, uint16x8_t b)

A32 : VADDHN.I16 Dd, Qn, Qm

A64 : ADDHN Vd.8B, Vn.8H, Vm.8H

AddHighNarrowingLower(Vector128<UInt32>, Vector128<UInt32>)

uint16x4_t vaddhn_u32 (uint32x4_t a, uint32x4_t b)

A32 : VADDHN.I32 Dd, Qn, Qm

A64 : ADDHN Vd.4H, Vn.4S, Vm.4S

AddHighNarrowingLower(Vector128<UInt64>, Vector128<UInt64>)

uint32x2_t vaddhn_u64 (uint64x2_t a, uint64x2_t b)

A32 : VADDHN.I64 Dd, Qn, Qm

A64 : ADDHN Vd.2S, Vn.2D, Vm.2D

AddHighNarrowingUpper(Vector64<Byte>, Vector128<UInt16>, Vector128<UInt16>)

uint8x16_t vaddhn_high_u16 (uint8x8_t r, uint16x8_t a, uint16x8_t b)

A32 : VADDHN.I16 Dd+1, Qn, Qm

A64 : ADDHN2 Vd.16B, Vn.8H, Vm.8H

AddHighNarrowingUpper(Vector64<Int16>, Vector128<Int32>, Vector128<Int32>)

int16x8_t vaddhn_high_s32 (int16x4_t r, int32x4_t a, int32x4_t b)

A32 : VADDHN.I32 Dd+1, Qn, Qm

A64 : ADDHN2 Vd.8H, Vn.4S, Vm.4S

AddHighNarrowingUpper(Vector64<Int32>, Vector128<Int64>, Vector128<Int64>)

int32x4_t vaddhn_high_s64 (int32x2_t r, int64x2_t a, int64x2_t b)

A32 : VADDHN.I64 Dd+1, Qn, Qm

A64 : ADDHN2 Vd.4S, Vn.2D, Vm.2D

AddHighNarrowingUpper(Vector64<SByte>, Vector128<Int16>, Vector128<Int16>)

int8x16_t vaddhn_high_s16 (int8x8_t r, int16x8_t a, int16x8_t b)

A32 : VADDHN.I16 Dd+1, Qn, Qm

A64 : ADDHN2 Vd.16B, Vn.8H, Vm.8H

AddHighNarrowingUpper(Vector64<UInt16>, Vector128<UInt32>, Vector128<UInt32>)

uint16x8_t vaddhn_high_u32 (uint16x4_t r, uint32x4_t a, uint32x4_t b)

A32 : VADDHN.I32 Dd+1, Qn, Qm

A64 : ADDHN2 Vd.8H, Vn.4S, Vm.4S

AddHighNarrowingUpper(Vector64<UInt32>, Vector128<UInt64>, Vector128<UInt64>)

uint32x4_t vaddhn_high_u64 (uint32x2_t r, uint64x2_t a, uint64x2_t b)

A32 : VADDHN.I64 Dd+1, Qn, Qm

A64 : ADDHN2 Vd.4S, Vn.2D, Vm.2D

AddPairwise(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vpadd_u8 (uint8x8_t a, uint8x8_t b)

A32 : VPADD.I8 Dd, Dn, Dm

A64 : ADDP Vd.8B, Vn.8B, Vm.8B

AddPairwise(Vector64<Int16>, Vector64<Int16>)

int16x4_t vpadd_s16 (int16x4_t a, int16x4_t b)

A32 : VPADD.I16 Dd, Dn, Dm

A64 : ADDP Vd.4H, Vn.4H, Vm.4H

AddPairwise(Vector64<Int32>, Vector64<Int32>)

int32x2_t vpadd_s32 (int32x2_t a, int32x2_t b)

A32 : VPADD.I32 Dd, Dn, Dm

A64 : ADDP Vd.2S, Vn.2S, Vm.2S

AddPairwise(Vector64<SByte>, Vector64<SByte>)

int8x8_t vpadd_s8 (int8x8_t a, int8x8_t b)

A32 : VPADD.I8 Dd, Dn, Dm

A64 : ADDP Vd.8B, Vn.8B, Vm.8B

AddPairwise(Vector64<Single>, Vector64<Single>)

float32x2_t vpadd_f32 (float32x2_t a, float32x2_t b)

A32 : VPADD.F32 Dd, Dn, Dm

A64 : FADDP Vd.2S, Vn.2S, Vm.2S

AddPairwise(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vpadd_u16 (uint16x4_t a, uint16x4_t b)

A32 : VPADD.I16 Dd, Dn, Dm

A64 : ADDP Vd.4H, Vn.4H, Vm.4H

AddPairwise(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vpadd_u32 (uint32x2_t a, uint32x2_t b)

A32 : VPADD.I32 Dd, Dn, Dm

A64 : ADDP Vd.2S, Vn.2S, Vm.2S

AddPairwiseWidening(Vector128<Byte>)

uint16x8_t vpaddlq_u8 (uint8x16_t a)

A32 : VPADDL.U8 Qd, Qm

A64 : UADDLP Vd.8H, Vn.16B

AddPairwiseWidening(Vector128<Int16>)

int32x4_t vpaddlq_s16 (int16x8_t a)

A32 : VPADDL.S16 Qd, Qm

A64 : SADDLP Vd.4S, Vn.8H

AddPairwiseWidening(Vector128<Int32>)

int64x2_t vpaddlq_s32 (int32x4_t a)

A32 : VPADDL.S32 Qd, Qm

A64 : SADDLP Vd.2D, Vn.4S

AddPairwiseWidening(Vector128<SByte>)

int16x8_t vpaddlq_s8 (int8x16_t a)

A32 : VPADDL.S8 Qd, Qm

A64 : SADDLP Vd.8H, Vn.16B

AddPairwiseWidening(Vector128<UInt16>)

uint32x4_t vpaddlq_u16 (uint16x8_t a)

A32 : VPADDL.U16 Qd, Qm

A64 : UADDLP Vd.4S, Vn.8H

AddPairwiseWidening(Vector128<UInt32>)

uint64x2_t vpaddlq_u32 (uint32x4_t a)

A32 : VPADDL.U32 Qd, Qm

A64 : UADDLP Vd.2D, Vn.4S

AddPairwiseWidening(Vector64<Byte>)

uint16x4_t vpaddl_u8 (uint8x8_t a)

A32 : VPADDL.U8 Dd, Dm

A64 : UADDLP Vd.4H, Vn.8B

AddPairwiseWidening(Vector64<Int16>)

int32x2_t vpaddl_s16 (int16x4_t a)

A32 : VPADDL.S16 Dd, Dm

A64 : SADDLP Vd.2S, Vn.4H

AddPairwiseWidening(Vector64<SByte>)

int16x4_t vpaddl_s8 (int8x8_t a)

A32 : VPADDL.S8 Dd, Dm

A64 : SADDLP Vd.4H, Vn.8B

AddPairwiseWidening(Vector64<UInt16>)

uint32x2_t vpaddl_u16 (uint16x4_t a)

A32 : VPADDL.U16 Dd, Dm

A64 : UADDLP Vd.2S, Vn.4H

AddPairwiseWideningAndAdd(Vector128<Int16>, Vector128<SByte>)

int16x8_t vpadalq_s8 (int16x8_t a, int8x16_t b)

A32 : VPADAL.S8 Qd, Qm

A64 : SADALP Vd.8H, Vn.16B

AddPairwiseWideningAndAdd(Vector128<Int32>, Vector128<Int16>)

int32x4_t vpadalq_s16 (int32x4_t a, int16x8_t b)

A32 : VPADAL.S16 Qd, Qm

A64 : SADALP Vd.4S, Vn.8H

AddPairwiseWideningAndAdd(Vector128<Int64>, Vector128<Int32>)

int64x2_t vpadalq_s32 (int64x2_t a, int32x4_t b)

A32 : VPADAL.S32 Qd, Qm

A64 : SADALP Vd.2D, Vn.4S

AddPairwiseWideningAndAdd(Vector128<UInt16>, Vector128<Byte>)

uint16x8_t vpadalq_u8 (uint16x8_t a, uint8x16_t b)

A32 : VPADAL.U8 Qd, Qm

A64 : UADALP Vd.8H, Vn.16B

AddPairwiseWideningAndAdd(Vector128<UInt32>, Vector128<UInt16>)

uint32x4_t vpadalq_u16 (uint32x4_t a, uint16x8_t b)

A32 : VPADAL.U16 Qd, Qm

A64 : UADALP Vd.4S, Vn.8H

AddPairwiseWideningAndAdd(Vector128<UInt64>, Vector128<UInt32>)

uint64x2_t vpadalq_u32 (uint64x2_t a, uint32x4_t b)

A32 : VPADAL.U32 Qd, Qm

A64 : UADALP Vd.2D, Vn.4S

AddPairwiseWideningAndAdd(Vector64<Int16>, Vector64<SByte>)

int16x4_t vpadal_s8 (int16x4_t a, int8x8_t b)

A32 : VPADAL.S8 Dd, Dm

A64 : SADALP Vd.4H, Vn.8B

AddPairwiseWideningAndAdd(Vector64<Int32>, Vector64<Int16>)

int32x2_t vpadal_s16 (int32x2_t a, int16x4_t b)

A32 : VPADAL.S16 Dd, Dm

A64 : SADALP Vd.2S, Vn.4H

AddPairwiseWideningAndAdd(Vector64<UInt16>, Vector64<Byte>)

uint16x4_t vpadal_u8 (uint16x4_t a, uint8x8_t b)

A32 : VPADAL.U8 Dd, Dm

A64 : UADALP Vd.4H, Vn.8B

AddPairwiseWideningAndAdd(Vector64<UInt32>, Vector64<UInt16>)

uint32x2_t vpadal_u16 (uint32x2_t a, uint16x4_t b)

A32 : VPADAL.U16 Dd, Dm

A64 : UADALP Vd.2S, Vn.4H

AddPairwiseWideningAndAddScalar(Vector64<Int64>, Vector64<Int32>)

int64x1_t vpadal_s32 (int64x1_t a, int32x2_t b)

A32 : VPADAL.S32 Dd, Dm

A64 : SADALP Vd.1D, Vn.2S

AddPairwiseWideningAndAddScalar(Vector64<UInt64>, Vector64<UInt32>)

uint64x1_t vpadal_u32 (uint64x1_t a, uint32x2_t b)

A32 : VPADAL.U32 Dd, Dm

A64 : UADALP Vd.1D, Vn.2S

AddPairwiseWideningScalar(Vector64<Int32>)

int64x1_t vpaddl_s32 (int32x2_t a)

A32 : VPADDL.S32 Dd, Dm

A64 : SADDLP Dd, Vn.2S

AddPairwiseWideningScalar(Vector64<UInt32>)

uint64x1_t vpaddl_u32 (uint32x2_t a)

A32 : VPADDL.U32 Dd, Dm

A64 : UADDLP Dd, Vn.2S

AddRoundedHighNarrowingLower(Vector128<Int16>, Vector128<Int16>)

int8x8_t vraddhn_s16 (int16x8_t a, int16x8_t b)

A32 : VRADDHN.I16 Dd, Qn, Qm

A64 : RADDHN Vd.8B, Vn.8H, Vm.8H

AddRoundedHighNarrowingLower(Vector128<Int32>, Vector128<Int32>)

int16x4_t vraddhn_s32 (int32x4_t a, int32x4_t b)

A32 : VRADDHN.I32 Dd, Qn, Qm

A64 : RADDHN Vd.4H, Vn.4S, Vm.4S

AddRoundedHighNarrowingLower(Vector128<Int64>, Vector128<Int64>)

int32x2_t vraddhn_s64 (int64x2_t a, int64x2_t b)

A32 : VRADDHN.I64 Dd, Qn, Qm

A64 : RADDHN Vd.2S, Vn.2D, Vm.2D

AddRoundedHighNarrowingLower(Vector128<UInt16>, Vector128<UInt16>)

uint8x8_t vraddhn_u16 (uint16x8_t a, uint16x8_t b)

A32 : VRADDHN.I16 Dd, Qn, Qm

A64 : RADDHN Vd.8B, Vn.8H, Vm.8H

AddRoundedHighNarrowingLower(Vector128<UInt32>, Vector128<UInt32>)

uint16x4_t vraddhn_u32 (uint32x4_t a, uint32x4_t b)

A32 : VRADDHN.I32 Dd, Qn, Qm

A64 : RADDHN Vd.4H, Vn.4S, Vm.4S

AddRoundedHighNarrowingLower(Vector128<UInt64>, Vector128<UInt64>)

uint32x2_t vraddhn_u64 (uint64x2_t a, uint64x2_t b)

A32 : VRADDHN.I64 Dd, Qn, Qm

A64 : RADDHN Vd.2S, Vn.2D, Vm.2D

AddRoundedHighNarrowingUpper(Vector64<Byte>, Vector128<UInt16>, Vector128<UInt16>)

uint8x16_t vraddhn_high_u16 (uint8x8_t r, uint16x8_t a, uint16x8_t b)

A32 : VRADDHN.I16 Dd+1, Qn, Qm

A64 : RADDHN2 Vd.16B, Vn.8H, Vm.8H

AddRoundedHighNarrowingUpper(Vector64<Int16>, Vector128<Int32>, Vector128<Int32>)

int16x8_t vraddhn_high_s32 (int16x4_t r, int32x4_t a, int32x4_t b)

A32 : VRADDHN.I32 Dd+1, Qn, Qm

A64 : RADDHN2 Vd.8H, Vn.4S, Vm.4S

AddRoundedHighNarrowingUpper(Vector64<Int32>, Vector128<Int64>, Vector128<Int64>)

int32x4_t vraddhn_high_s64 (int32x2_t r, int64x2_t a, int64x2_t b)

A32 : VRADDHN.I64 Dd+1, Qn, Qm

A64 : RADDHN2 Vd.4S, Vn.2D, Vm.2D

AddRoundedHighNarrowingUpper(Vector64<SByte>, Vector128<Int16>, Vector128<Int16>)

int8x16_t vraddhn_high_s16 (int8x8_t r, int16x8_t a, int16x8_t b)

A32 : VRADDHN.I16 Dd+1, Qn, Qm

A64 : RADDHN2 Vd.16B, Vn.8H, Vm.8H

AddRoundedHighNarrowingUpper(Vector64<UInt16>, Vector128<UInt32>, Vector128<UInt32>)

uint16x8_t vraddhn_high_u32 (uint16x4_t r, uint32x4_t a, uint32x4_t b)

A32 : VRADDHN.I32 Dd+1, Qn, Qm

A64 : RADDHN2 Vd.8H, Vn.4S, Vm.4S

AddRoundedHighNarrowingUpper(Vector64<UInt32>, Vector128<UInt64>, Vector128<UInt64>)

uint32x4_t vraddhn_high_u64 (uint32x2_t r, uint64x2_t a, uint64x2_t b)

A32 : VRADDHN.I64 Dd+1, Qn, Qm

A64 : RADDHN2 Vd.4S, Vn.2D, Vm.2D

AddSaturate(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vqaddq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VQADD.U8 Qd, Qn, Qm

A64 : UQADD Vd.16B, Vn.16B, Vm.16B

AddSaturate(Vector128<Int16>, Vector128<Int16>)

int16x8_t vqaddq_s16 (int16x8_t a, int16x8_t b)

A32 : VQADD.S16 Qd, Qn, Qm

A64 : SQADD Vd.8H, Vn.8H, Vm.8H

AddSaturate(Vector128<Int32>, Vector128<Int32>)

int32x4_t vqaddq_s32 (int32x4_t a, int32x4_t b)

A32 : VQADD.S32 Qd, Qn, Qm

A64 : SQADD Vd.4S, Vn.4S, Vm.4S

AddSaturate(Vector128<Int64>, Vector128<Int64>)

int64x2_t vqaddq_s64 (int64x2_t a, int64x2_t b)

A32 : VQADD.S64 Qd, Qn, Qm

A64 : SQADD Vd.2D, Vn.2D, Vm.2D

AddSaturate(Vector128<SByte>, Vector128<SByte>)

int8x16_t vqaddq_s8 (int8x16_t a, int8x16_t b)

A32 : VQADD.S8 Qd, Qn, Qm

A64 : SQADD Vd.16B, Vn.16B, Vm.16B

AddSaturate(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vqaddq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VQADD.U16 Qd, Qn, Qm

A64 : UQADD Vd.8H, Vn.8H, Vm.8H

AddSaturate(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vqaddq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VQADD.U32 Qd, Qn, Qm

A64 : UQADD Vd.4S, Vn.4S, Vm.4S

AddSaturate(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vqaddq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VQADD.U64 Qd, Qn, Qm

A64 : UQADD Vd.2D, Vn.2D, Vm.2D

AddSaturate(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vqadd_u8 (uint8x8_t a, uint8x8_t b)

A32 : VQADD.U8 Dd, Dn, Dm

A64 : UQADD Vd.8B, Vn.8B, Vm.8B

AddSaturate(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqadd_s16 (int16x4_t a, int16x4_t b)

A32 : VQADD.S16 Dd, Dn, Dm

A64 : SQADD Vd.4H, Vn.4H, Vm.4H

AddSaturate(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqadd_s32 (int32x2_t a, int32x2_t b)

A32 : VQADD.S32 Dd, Dn, Dm

A64 : SQADD Vd.2S, Vn.2S, Vm.2S

AddSaturate(Vector64<SByte>, Vector64<SByte>)

int8x8_t vqadd_s8 (int8x8_t a, int8x8_t b)

A32 : VQADD.S8 Dd, Dn, Dm

A64 : SQADD Vd.8B, Vn.8B, Vm.8B

AddSaturate(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vqadd_u16 (uint16x4_t a, uint16x4_t b)

A32 : VQADD.U16 Dd, Dn, Dm

A64 : UQADD Vd.4H, Vn.4H, Vm.4H

AddSaturate(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vqadd_u32 (uint32x2_t a, uint32x2_t b)

A32 : VQADD.U32 Dd, Dn, Dm

A64 : UQADD Vd.2S, Vn.2S, Vm.2S

AddSaturateScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vqadd_s64 (int64x1_t a, int64x1_t b)

A32 : VQADD.S64 Dd, Dn, Dm

A64 : SQADD Dd, Dn, Dm

AddSaturateScalar(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vqadd_u64 (uint64x1_t a, uint64x1_t b)

A32 : VQADD.U64 Dd, Dn, Dm

A64 : UQADD Dd, Dn, Dm

AddScalar(Vector64<Double>, Vector64<Double>)

float64x1_t vadd_f64 (float64x1_t a, float64x1_t b)

A32 : VADD.F64 Dd, Dn, Dm

A64 : FADD Dd, Dn, Dm

AddScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vadd_s64 (int64x1_t a, int64x1_t b)

A32 : VADD.I64 Dd, Dn, Dm

A64 : ADD Dd, Dn, Dm

AddScalar(Vector64<Single>, Vector64<Single>)

float32_t vadds_f32 (float32_t a, float32_t b)

A32 : VADD.F32 Sd, Sn, Sm

A64 : FADD Sd, Sn, Sm La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

AddScalar(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vadd_u64 (uint64x1_t a, uint64x1_t b)

A32 : VADD.I64 Dd, Dn, Dm

A64 : ADD Dd, Dn, Dm

AddWideningLower(Vector128<Int16>, Vector64<SByte>)

int16x8_t vaddw_s8 (int16x8_t a, int8x8_t b)

A32 : VADDW.S8 Qd, Qn, Dm

A64 : SADDW Vd.8H, Vn.8H, Vm.8B

AddWideningLower(Vector128<Int32>, Vector64<Int16>)

int32x4_t vaddw_s16 (int32x4_t a, int16x4_t b)

A32 : VADDW.S16 Qd, Qn, Dm

A64 : SADDW Vd.4S, Vn.4S, Vm.4H

AddWideningLower(Vector128<Int64>, Vector64<Int32>)

int64x2_t vaddw_s32 (int64x2_t a, int32x2_t b)

A32 : VADDW.S32 Qd, Qn, Dm

A64 : SADDW Vd.2D, Vn.2D, Vm.2S

AddWideningLower(Vector128<UInt16>, Vector64<Byte>)

uint16x8_t vaddw_u8 (uint16x8_t a, uint8x8_t b)

A32 : VADDW.U8 Qd, Qn, Dm

A64 : UADDW Vd.8H, Vn.8H, Vm.8B

AddWideningLower(Vector128<UInt32>, Vector64<UInt16>)

uint32x4_t vaddw_u16 (uint32x4_t a, uint16x4_t b)

A32 : VADDW.U16 Qd, Qn, Dm

A64 : UADDW Vd.4S, Vn.4S, Vm.4H

AddWideningLower(Vector128<UInt64>, Vector64<UInt32>)

uint64x2_t vaddw_u32 (uint64x2_t a, uint32x2_t b)

A32 : VADDW.U32 Qd, Qn, Dm

A64 : UADDW Vd.2D, Vn.2D, Vm.2S

AddWideningLower(Vector64<Byte>, Vector64<Byte>)

uint16x8_t vaddl_u8 (uint8x8_t a, uint8x8_t b)

A32 : VADDL.U8 Qd, Dn, Dm

A64 : UADDL Vd.8H, Vn.8B, Vm.8B

AddWideningLower(Vector64<Int16>, Vector64<Int16>)

int32x4_t vaddl_s16 (int16x4_t a, int16x4_t b)

A32 : VADDL.S16 Qd, Dn, Dm

A64 : SADDL Vd.4S, Vn.4H, Vm.4H

AddWideningLower(Vector64<Int32>, Vector64<Int32>)

int64x2_t vaddl_s32 (int32x2_t a, int32x2_t b)

A32 : VADDL.S32 Qd, Dn, Dm

A64 : SADDL Vd.2D, Vn.2S, Vm.2S

AddWideningLower(Vector64<SByte>, Vector64<SByte>)

int16x8_t vaddl_s8 (int8x8_t a, int8x8_t b)

A32 : VADDL.S8 Qd, Dn, Dm

A64 : SADDL Vd.8H, Vn.8B, Vm.8B

AddWideningLower(Vector64<UInt16>, Vector64<UInt16>)

uint32x4_t vaddl_u16 (uint16x4_t a, uint16x4_t b)

A32 : VADDL.U16 Qd, Dn, Dm

A64 : UADDL Vd.4S, Vn.4H, Vm.4H

AddWideningLower(Vector64<UInt32>, Vector64<UInt32>)

uint64x2_t vaddl_u32 (uint32x2_t a, uint32x2_t b)

A32 : VADDL.U32 Qd, Dn, Dm

A64 : UADDL Vd.2D, Vn.2S, Vm.2S

AddWideningUpper(Vector128<Byte>, Vector128<Byte>)

uint16x8_t vaddl_high_u8 (uint8x16_t a, uint8x16_t b)

A32 : VADDL.U8 Qd, Dn+1, Dm+1

A64 : UADDL2 Vd.8H, Vn.16B, Vm.16B

AddWideningUpper(Vector128<Int16>, Vector128<Int16>)

int32x4_t vaddl_high_s16 (int16x8_t a, int16x8_t b)

A32 : VADDL.S16 Qd, Dn+1, Dm+1

A64 : SADDL2 Vd.4S, Vn.8H, Vm.8H

AddWideningUpper(Vector128<Int16>, Vector128<SByte>)

int16x8_t vaddw_high_s8 (int16x8_t a, int8x16_t b)

A32 : VADDW.S8 Qd, Qn, Dm+1

A64 : SADDW2 Vd.8H, Vn.8H, Vm.16B

AddWideningUpper(Vector128<Int32>, Vector128<Int16>)

int32x4_t vaddw_high_s16 (int32x4_t a, int16x8_t b)

A32 : VADDW.S16 Qd, Qn, Dm+1

A64 : SADDW2 Vd.4S, Vn.4S, Vm.8H

AddWideningUpper(Vector128<Int32>, Vector128<Int32>)

int64x2_t vaddl_high_s32 (int32x4_t a, int32x4_t b)

A32 : VADDL.S32 Qd, Dn+1, Dm+1

A64 : SADDL2 Vd.2D, Vn.4S, Vm.4S

AddWideningUpper(Vector128<Int64>, Vector128<Int32>)

int64x2_t vaddw_high_s32 (int64x2_t a, int32x4_t b)

A32 : VADDW.S32 Qd, Qn, Dm+1

A64 : SADDW2 Vd.2D, Vn.2D, Vm.4S

AddWideningUpper(Vector128<SByte>, Vector128<SByte>)

int16x8_t vaddl_high_s8 (int8x16_t a, int8x16_t b)

A32 : VADDL.S8 Qd, Dn+1, Dm+1

A64 : SADDL2 Vd.8H, Vn.16B, Vm.16B

AddWideningUpper(Vector128<UInt16>, Vector128<Byte>)

uint16x8_t vaddw_high_u8 (uint16x8_t a, uint8x16_t b)

A32 : VADDW.U8 Qd, Qn, Dm+1

A64 : UADDW2 Vd.8H, Vn.8H, Vm.16B

AddWideningUpper(Vector128<UInt16>, Vector128<UInt16>)

uint32x4_t vaddl_high_u16 (uint16x8_t a, uint16x8_t b)

A32 : VADDL.U16 Qd, Dn+1, Dm+1

A64 : UADDL2 Vd.4S, Vn.8H, Vm.8H

AddWideningUpper(Vector128<UInt32>, Vector128<UInt16>)

uint32x4_t vaddw_high_u16 (uint32x4_t a, uint16x8_t b)

A32 : VADDW.U16 Qd, Qn, Dm+1

A64 : UADDW2 Vd.4S, Vn.4S, Vm.8H

AddWideningUpper(Vector128<UInt32>, Vector128<UInt32>)

uint64x2_t vaddl_high_u32 (uint32x4_t a, uint32x4_t b)

A32 : VADDL.U32 Qd, Dn+1, Dm+1

A64 : UADDL2 Vd.2D, Vn.4S, Vm.4S

AddWideningUpper(Vector128<UInt64>, Vector128<UInt32>)

uint64x2_t vaddw_high_u32 (uint64x2_t a, uint32x4_t b)

A32 : VADDW.U32 Qd, Qn, Dm+1

A64 : UADDW2 Vd.2D, Vn.2D, Vm.4S

And(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vandq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector128<Double>, Vector128<Double>)

float64x2_t vandq_f64 (float64x2_t a, float64x2_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

And(Vector128<Int16>, Vector128<Int16>)

int16x8_t vandq_s16 (int16x8_t a, int16x8_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector128<Int32>, Vector128<Int32>)

int32x4_t vandq_s32 (int32x4_t a, int32x4_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector128<Int64>, Vector128<Int64>)

int64x2_t vandq_s64 (int64x2_t a, int64x2_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector128<SByte>, Vector128<SByte>)

int8x16_t vandq_s8 (int8x16_t a, int8x16_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector128<Single>, Vector128<Single>)

float32x4_t vandq_f32 (float32x4_t a, float32x4_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

And(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vandq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vandq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vandq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VAND Qd, Qn, Qm

A64 : AND Vd.16B, Vn.16B, Vm.16B

And(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vand_u8 (uint8x8_t a, uint8x8_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

And(Vector64<Double>, Vector64<Double>)

float64x1_t vand_f64 (float64x1_t a, float64x1_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

And(Vector64<Int16>, Vector64<Int16>)

int16x4_t vand_s16 (int16x4_t a, int16x4_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

And(Vector64<Int32>, Vector64<Int32>)

int32x2_t vand_s32 (int32x2_t a, int32x2_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

And(Vector64<Int64>, Vector64<Int64>)

int64x1_t vand_s64 (int64x1_t a, int64x1_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

And(Vector64<SByte>, Vector64<SByte>)

int8x8_t vand_s8 (int8x8_t a, int8x8_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

And(Vector64<Single>, Vector64<Single>)

float32x2_t vand_f32 (float32x2_t a, float32x2_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

And(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vand_u16 (uint16x4_t a, uint16x4_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

And(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vand_u32 (uint32x2_t a, uint32x2_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

And(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vand_u64 (uint64x1_t a, uint64x1_t b)

A32 : VAND Dd, Dn, Dm

A64 : AND Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vbicq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector128<Double>, Vector128<Double>)

float64x2_t vbicq_f64 (float64x2_t a, float64x2_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

BitwiseClear(Vector128<Int16>, Vector128<Int16>)

int16x8_t vbicq_s16 (int16x8_t a, int16x8_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector128<Int32>, Vector128<Int32>)

int32x4_t vbicq_s32 (int32x4_t a, int32x4_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector128<Int64>, Vector128<Int64>)

int64x2_t vbicq_s64 (int64x2_t a, int64x2_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector128<SByte>, Vector128<SByte>)

int8x16_t vbicq_s8 (int8x16_t a, int8x16_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector128<Single>, Vector128<Single>)

float32x4_t vbicq_f32 (float32x4_t a, float32x4_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

BitwiseClear(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vbicq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vbicq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vbicq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VBIC Qd, Qn, Qm

A64 : BIC Vd.16B, Vn.16B, Vm.16B

BitwiseClear(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vbic_u8 (uint8x8_t a, uint8x8_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector64<Double>, Vector64<Double>)

float64x1_t vbic_f64 (float64x1_t a, float64x1_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

BitwiseClear(Vector64<Int16>, Vector64<Int16>)

int16x4_t vbic_s16 (int16x4_t a, int16x4_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector64<Int32>, Vector64<Int32>)

int32x2_t vbic_s32 (int32x2_t a, int32x2_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector64<Int64>, Vector64<Int64>)

int64x1_t vbic_s64 (int64x1_t a, int64x1_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector64<SByte>, Vector64<SByte>)

int8x8_t vbic_s8 (int8x8_t a, int8x8_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector64<Single>, Vector64<Single>)

float32x2_t vbic_f32 (float32x2_t a, float32x2_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

BitwiseClear(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vbic_u16 (uint16x4_t a, uint16x4_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vbic_u32 (uint32x2_t a, uint32x2_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseClear(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vbic_u64 (uint64x1_t a, uint64x1_t b)

A32 : VBIC Dd, Dn, Dm

A64 : BIC Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector128<Byte>, Vector128<Byte>, Vector128<Byte>)

uint8x16_t vbslq_u8 (uint8x16_t a, uint8x16_t b, uint8x16_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<Double>, Vector128<Double>, Vector128<Double>)

float64x2_t vbslq_f64 (uint64x2_t a, float64x2_t b, float64x2_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<Int16>, Vector128<Int16>, Vector128<Int16>)

int16x8_t vbslq_s16 (uint16x8_t a, int16x8_t b, int16x8_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<Int32>, Vector128<Int32>, Vector128<Int32>)

int32x4_t vbslq_s32 (uint32x4_t a, int32x4_t b, int32x4_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<Int64>, Vector128<Int64>, Vector128<Int64>)

int64x2_t vbslq_s64 (uint64x2_t a, int64x2_t b, int64x2_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<SByte>, Vector128<SByte>, Vector128<SByte>)

int8x16_t vbslq_s8 (uint8x16_t a, int8x16_t b, int8x16_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<Single>, Vector128<Single>, Vector128<Single>)

float32x4_t vbslq_f32 (uint32x4_t a, float32x4_t b, float32x4_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<UInt16>, Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vbslq_u16 (uint16x8_t a, uint16x8_t b, uint16x8_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<UInt32>, Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vbslq_u32 (uint32x4_t a, uint32x4_t b, uint32x4_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector128<UInt64>, Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vbslq_u64 (uint64x2_t a, uint64x2_t b, uint64x2_t c)

A32 : VBSL Qd, Qn, Qm

A64 : BSL Vd.16B, Vn.16B, Vm.16B

BitwiseSelect(Vector64<Byte>, Vector64<Byte>, Vector64<Byte>)

uint8x8_t vbsl_u8 (uint8x8_t a, uint8x8_t b, uint8x8_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<Double>, Vector64<Double>, Vector64<Double>)

float64x1_t vbsl_f64 (uint64x1_t a, float64x1_t b, float64x1_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>)

int16x4_t vbsl_s16 (uint16x4_t a, int16x4_t b, int16x4_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>)

int32x2_t vbsl_s32 (uint32x2_t a, int32x2_t b, int32x2_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<Int64>, Vector64<Int64>, Vector64<Int64>)

int64x1_t vbsl_s64 (uint64x1_t a, int64x1_t b, int64x1_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<SByte>, Vector64<SByte>, Vector64<SByte>)

int8x8_t vbsl_s8 (uint8x8_t a, int8x8_t b, int8x8_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<Single>, Vector64<Single>, Vector64<Single>)

float32x2_t vbsl_f32 (uint32x2_t a, float32x2_t b, float32x2_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vbsl_u16 (uint16x4_t a, uint16x4_t b, uint16x4_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vbsl_u32 (uint32x2_t a, uint32x2_t b, uint32x2_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

BitwiseSelect(Vector64<UInt64>, Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vbsl_u64 (uint64x1_t a, uint64x1_t b, uint64x1_t c)

A32 : VBSL Dd, Dn, Dm

A64 : BSL Vd.8B, Vn.8B, Vm.8B

Ceiling(Vector128<Single>)

float32x4_t vrndpq_f32 (float32x4_t a)

A32 : VRINTP.F32 Qd, Qm

A64 : FRINTP Vd.4S, Vn.4S

Ceiling(Vector64<Single>)

float32x2_t vrndp_f32 (float32x2_t a)

A32 : VRINTP.F32 Dd, Dm

A64 : FRINTP Vd.2S, Vn.2S

CeilingScalar(Vector64<Double>)

float64x1_t vrndp_f64 (float64x1_t a)

A32 : VRINTP.F64 Dd, Dm

A64 : FRINTP Dd, Dn

CeilingScalar(Vector64<Single>)

float32_t vrndps_f32 (float32_t a)

A32 : VRINTP.F32 Sd, Sm

A64 : FRINTP Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

CompareEqual(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vceqq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VCEQ.I8 Qd, Qn, Qm

A64 : CMEQ Vd.16B, Vn.16B, Vm.16B

CompareEqual(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vceqq_s16 (int16x8_t a, int16x8_t b)

A32 : VCEQ.I16 Qd, Qn, Qm

A64 : CMEQ Vd.8H, Vn.8H, Vm.8H

CompareEqual(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vceqq_s32 (int32x4_t a, int32x4_t b)

A32 : VCEQ.I32 Qd, Qn, Qm

A64 : CMEQ Vd.4S, Vn.4S, Vm.4S

CompareEqual(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vceqq_s8 (int8x16_t a, int8x16_t b)

A32 : VCEQ.I8 Qd, Qn, Qm

A64 : CMEQ Vd.16B, Vn.16B, Vm.16B

CompareEqual(Vector128<Single>, Vector128<Single>)

uint32x4_t vceqq_f32 (float32x4_t a, float32x4_t b)

A32 : VCEQ.F32 Qd, Qn, Qm

A64 : FCMEQ Vd.4S, Vn.4S, Vm.4S

CompareEqual(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vceqq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VCEQ.I16 Qd, Qn, Qm

A64 : CMEQ Vd.8H, Vn.8H, Vm.8H

CompareEqual(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vceqq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VCEQ.I32 Qd, Qn, Qm

A64 : CMEQ Vd.4S, Vn.4S, Vm.4S

CompareEqual(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vceq_u8 (uint8x8_t a, uint8x8_t b)

A32 : VCEQ.I8 Dd, Dn, Dm

A64 : CMEQ Vd.8B, Vn.8B, Vm.8B

CompareEqual(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vceq_s16 (int16x4_t a, int16x4_t b)

A32 : VCEQ.I16 Dd, Dn, Dm

A64 : CMEQ Vd.4H, Vn.4H, Vm.4H

CompareEqual(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vceq_s32 (int32x2_t a, int32x2_t b)

A32 : VCEQ.I32 Dd, Dn, Dm

A64 : CMEQ Vd.2S, Vn.2S, Vm.2S

CompareEqual(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vceq_s8 (int8x8_t a, int8x8_t b)

A32 : VCEQ.I8 Dd, Dn, Dm

A64 : CMEQ Vd.8B, Vn.8B, Vm.8B

CompareEqual(Vector64<Single>, Vector64<Single>)

uint32x2_t vceq_f32 (float32x2_t a, float32x2_t b)

A32 : VCEQ.F32 Dd, Dn, Dm

A64 : FCMEQ Vd.2S, Vn.2S, Vm.2S

CompareEqual(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vceq_u16 (uint16x4_t a, uint16x4_t b)

A32 : VCEQ.I16 Dd, Dn, Dm

A64 : CMEQ Vd.4H, Vn.4H, Vm.4H

CompareEqual(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vceq_u32 (uint32x2_t a, uint32x2_t b)

A32 : VCEQ.I32 Dd, Dn, Dm

A64 : CMEQ Vd.2S, Vn.2S, Vm.2S

CompareGreaterThan(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vcgtq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VCGT.U8 Qd, Qn, Qm

A64 : CMHI Vd.16B, Vn.16B, Vm.16B

CompareGreaterThan(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vcgtq_s16 (int16x8_t a, int16x8_t b)

A32 : VCGT.S16 Qd, Qn, Qm

A64 : CMGT Vd.8H, Vn.8H, Vm.8H

CompareGreaterThan(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vcgtq_s32 (int32x4_t a, int32x4_t b)

A32 : VCGT.S32 Qd, Qn, Qm

A64 : CMGT Vd.4S, Vn.4S, Vm.4S

CompareGreaterThan(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vcgtq_s8 (int8x16_t a, int8x16_t b)

A32 : VCGT.S8 Qd, Qn, Qm

A64 : CMGT Vd.16B, Vn.16B, Vm.16B

CompareGreaterThan(Vector128<Single>, Vector128<Single>)

uint32x4_t vcgtq_f32 (float32x4_t a, float32x4_t b)

A32 : VCGT.F32 Qd, Qn, Qm

A64 : FCMGT Vd.4S, Vn.4S, Vm.4S

CompareGreaterThan(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vcgtq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VCGT.U16 Qd, Qn, Qm

A64 : CMHI Vd.8H, Vn.8H, Vm.8H

CompareGreaterThan(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vcgtq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VCGT.U32 Qd, Qn, Qm

A64 : CMHI Vd.4S, Vn.4S, Vm.4S

CompareGreaterThan(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vcgt_u8 (uint8x8_t a, uint8x8_t b)

A32 : VCGT.U8 Dd, Dn, Dm

A64 : CMHI Vd.8B, Vn.8B, Vm.8B

CompareGreaterThan(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vcgt_s16 (int16x4_t a, int16x4_t b)

A32 : VCGT.S16 Dd, Dn, Dm

A64 : CMGT Vd.4H, Vn.4H, Vm.4H

CompareGreaterThan(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vcgt_s32 (int32x2_t a, int32x2_t b)

A32 : VCGT.S32 Dd, Dn, Dm

A64 : CMGT Vd.2S, Vn.2S, Vm.2S

CompareGreaterThan(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vcgt_s8 (int8x8_t a, int8x8_t b)

A32 : VCGT.S8 Dd, Dn, Dm

A64 : CMGT Vd.8B, Vn.8B, Vm.8B

CompareGreaterThan(Vector64<Single>, Vector64<Single>)

uint32x2_t vcgt_f32 (float32x2_t a, float32x2_t b)

A32 : VCGT.F32 Dd, Dn, Dm

A64 : FCMGT Vd.2S, Vn.2S, Vm.2S

CompareGreaterThan(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vcgt_u16 (uint16x4_t a, uint16x4_t b)

A32 : VCGT.U16 Dd, Dn, Dm

A64 : CMHI Vd.4H, Vn.4H, Vm.4H

CompareGreaterThan(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vcgt_u32 (uint32x2_t a, uint32x2_t b)

A32 : VCGT.U32 Dd, Dn, Dm

A64 : CMHI Vd.2S, Vn.2S, Vm.2S

CompareGreaterThanOrEqual(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vcgeq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VCGE.U8 Qd, Qn, Qm

A64 : CMHS Vd.16B, Vn.16B, Vm.16B

CompareGreaterThanOrEqual(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vcgeq_s16 (int16x8_t a, int16x8_t b)

A32 : VCGE.S16 Qd, Qn, Qm

A64 : CMGE Vd.8H, Vn.8H, Vm.8H

CompareGreaterThanOrEqual(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vcgeq_s32 (int32x4_t a, int32x4_t b)

A32 : VCGE.S32 Qd, Qn, Qm

A64 : CMGE Vd.4S, Vn.4S, Vm.4S

CompareGreaterThanOrEqual(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vcgeq_s8 (int8x16_t a, int8x16_t b)

A32 : VCGE.S8 Qd, Qn, Qm

A64 : CMGE Vd.16B, Vn.16B, Vm.16B

CompareGreaterThanOrEqual(Vector128<Single>, Vector128<Single>)

uint32x4_t vcgeq_f32 (float32x4_t a, float32x4_t b)

A32 : VCGE.F32 Qd, Qn, Qm

A64 : FCMGE Vd.4S, Vn.4S, Vm.4S

CompareGreaterThanOrEqual(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vcgeq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VCGE.U16 Qd, Qn, Qm

A64 : CMHS Vd.8H, Vn.8H, Vm.8H

CompareGreaterThanOrEqual(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vcgeq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VCGE.U32 Qd, Qn, Qm

A64 : CMHS Vd.4S, Vn.4S, Vm.4S

CompareGreaterThanOrEqual(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vcge_u8 (uint8x8_t a, uint8x8_t b)

A32 : VCGE.U8 Dd, Dn, Dm

A64 : CMHS Vd.8B, Vn.8B, Vm.8B

CompareGreaterThanOrEqual(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vcge_s16 (int16x4_t a, int16x4_t b)

A32 : VCGE.S16 Dd, Dn, Dm

A64 : CMGE Vd.4H, Vn.4H, Vm.4H

CompareGreaterThanOrEqual(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vcge_s32 (int32x2_t a, int32x2_t b)

A32 : VCGE.S32 Dd, Dn, Dm

A64 : CMGE Vd.2S, Vn.2S, Vm.2S

CompareGreaterThanOrEqual(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vcge_s8 (int8x8_t a, int8x8_t b)

A32 : VCGE.S8 Dd, Dn, Dm

A64 : CMGE Vd.8B, Vn.8B, Vm.8B

CompareGreaterThanOrEqual(Vector64<Single>, Vector64<Single>)

uint32x2_t vcge_f32 (float32x2_t a, float32x2_t b)

A32 : VCGE.F32 Dd, Dn, Dm

A64 : FCMGE Vd.2S, Vn.2S, Vm.2S

CompareGreaterThanOrEqual(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vcge_u16 (uint16x4_t a, uint16x4_t b)

A32 : VCGE.U16 Dd, Dn, Dm

A64 : CMHS Vd.4H, Vn.4H, Vm.4H

CompareGreaterThanOrEqual(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vcge_u32 (uint32x2_t a, uint32x2_t b)

A32 : VCGE.U32 Dd, Dn, Dm

A64 : CMHS Vd.2S, Vn.2S, Vm.2S

CompareLessThan(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vcltq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VCLT.U8 Qd, Qn, Qm

A64 : CMHI Vd.16B, Vn.16B, Vm.16B

CompareLessThan(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vcltq_s16 (int16x8_t a, int16x8_t b)

A32 : VCLT.S16 Qd, Qn, Qm

A64 : CMGT Vd.8H, Vn.8H, Vm.8H

CompareLessThan(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vcltq_s32 (int32x4_t a, int32x4_t b)

A32 : VCLT.S32 Qd, Qn, Qm

A64 : CMGT Vd.4S, Vn.4S, Vm.4S

CompareLessThan(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vcltq_s8 (int8x16_t a, int8x16_t b)

A32 : VCLT.S8 Qd, Qn, Qm

A64 : CMGT Vd.16B, Vn.16B, Vm.16B

CompareLessThan(Vector128<Single>, Vector128<Single>)

uint32x4_t vcltq_f32 (float32x4_t a, float32x4_t b)

A32 : VCLT.F32 Qd, Qn, Qm

A64 : FCMGT Vd.4S, Vn.4S, Vm.4S

CompareLessThan(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vcltq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VCLT.U16 Qd, Qn, Qm

A64 : CMHI Vd.8H, Vn.8H, Vm.8H

CompareLessThan(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vcltq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VCLT.U32 Qd, Qn, Qm

A64 : CMHI Vd.4S, Vn.4S, Vm.4S

CompareLessThan(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vclt_u8 (uint8x8_t a, uint8x8_t b)

A32 : VCLT.U8 Dd, Dn, Dm

A64 : CMHI Vd.8B, Vn.8B, Vm.8B

CompareLessThan(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vclt_s16 (int16x4_t a, int16x4_t b)

A32 : VCLT.S16 Dd, Dn, Dm

A64 : CMGT Vd.4H, Vn.4H, Vm.4H

CompareLessThan(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vclt_s32 (int32x2_t a, int32x2_t b)

A32 : VCLT.S32 Dd, Dn, Dm

A64 : CMGT Vd.2S, Vn.2S, Vm.2S

CompareLessThan(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vclt_s8 (int8x8_t a, int8x8_t b)

A32 : VCLT.S8 Dd, Dn, Dm

A64 : CMGT Vd.8B, Vn.8B, Vm.8B

CompareLessThan(Vector64<Single>, Vector64<Single>)

uint32x2_t vclt_f32 (float32x2_t a, float32x2_t b)

A32 : VCLT.F32 Dd, Dn, Dm

A64 : FCMGT Vd.2S, Vn.2S, Vm.2S

CompareLessThan(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vclt_u16 (uint16x4_t a, uint16x4_t b)

A32 : VCLT.U16 Dd, Dn, Dm

A64 : CMHI Vd.4H, Vn.4H, Vm.4H

CompareLessThan(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vclt_u32 (uint32x2_t a, uint32x2_t b)

A32 : VCLT.U32 Dd, Dn, Dm

A64 : CMHI Vd.2S, Vn.2S, Vm.2S

CompareLessThanOrEqual(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vcleq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VCLE.U8 Qd, Qn, Qm

A64 : CMHS Vd.16B, Vn.16B, Vm.16B

CompareLessThanOrEqual(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vcleq_s16 (int16x8_t a, int16x8_t b)

A32 : VCLE.S16 Qd, Qn, Qm

A64 : CMGE Vd.8H, Vn.8H, Vm.8H

CompareLessThanOrEqual(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vcleq_s32 (int32x4_t a, int32x4_t b)

A32 : VCLE.S32 Qd, Qn, Qm

A64 : CMGE Vd.4S, Vn.4S, Vm.4S

CompareLessThanOrEqual(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vcleq_s8 (int8x16_t a, int8x16_t b)

A32 : VCLE.S8 Qd, Qn, Qm

A64 : CMGE Vd.16B, Vn.16B, Vm.16B

CompareLessThanOrEqual(Vector128<Single>, Vector128<Single>)

uint32x4_t vcleq_f32 (float32x4_t a, float32x4_t b)

A32 : VCLE.F32 Qd, Qn, Qm

A64 : FCMGE Vd.4S, Vn.4S, Vm.4S

CompareLessThanOrEqual(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vcleq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VCLE.U16 Qd, Qn, Qm

A64 : CMHS Vd.8H, Vn.8H, Vm.8H

CompareLessThanOrEqual(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vcleq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VCLE.U32 Qd, Qn, Qm

A64 : CMHS Vd.4S, Vn.4S, Vm.4S

CompareLessThanOrEqual(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vcle_u8 (uint8x8_t a, uint8x8_t b)

A32 : VCLE.U8 Dd, Dn, Dm

A64 : CMHS Vd.8B, Vn.8B, Vm.8B

CompareLessThanOrEqual(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vcle_s16 (int16x4_t a, int16x4_t b)

A32 : VCLE.S16 Dd, Dn, Dm

A64 : CMGE Vd.4H, Vn.4H, Vm.4H

CompareLessThanOrEqual(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vcle_s32 (int32x2_t a, int32x2_t b)

A32 : VCLE.S32 Dd, Dn, Dm

A64 : CMGE Vd.2S, Vn.2S, Vm.2S

CompareLessThanOrEqual(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vcle_s8 (int8x8_t a, int8x8_t b)

A32 : VCLE.S8 Dd, Dn, Dm

A64 : CMGE Vd.8B, Vn.8B, Vm.8B

CompareLessThanOrEqual(Vector64<Single>, Vector64<Single>)

uint32x2_t vcle_f32 (float32x2_t a, float32x2_t b)

A32 : VCLE.F32 Dd, Dn, Dm

A64 : FCMGE Vd.2S, Vn.2S, Vm.2S

CompareLessThanOrEqual(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vcle_u16 (uint16x4_t a, uint16x4_t b)

A32 : VCLE.U16 Dd, Dn, Dm

A64 : CMHS Vd.4H, Vn.4H, Vm.4H

CompareLessThanOrEqual(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vcle_u32 (uint32x2_t a, uint32x2_t b)

A32 : VCLE.U32 Dd, Dn, Dm

A64 : CMHS Vd.2S, Vn.2S, Vm.2S

CompareTest(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vtstq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VTST.8 Qd, Qn, Qm

A64 : CMTST Vd.16B, Vn.16B, Vm.16B

CompareTest(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vtstq_s16 (int16x8_t a, int16x8_t b)

A32 : VTST.16 Qd, Qn, Qm

A64 : CMTST Vd.8H, Vn.8H, Vm.8H

CompareTest(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vtstq_s32 (int32x4_t a, int32x4_t b)

A32 : VTST.32 Qd, Qn, Qm

A64 : CMTST Vd.4S, Vn.4S, Vm.4S

CompareTest(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vtstq_s8 (int8x16_t a, int8x16_t b)

A32 : VTST.8 Qd, Qn, Qm

A64 : CMTST Vd.16B, Vn.16B, Vm.16B

CompareTest(Vector128<Single>, Vector128<Single>)

uint32x4_t vtstq_f32 (float32x4_t a, float32x4_t b)

A32 : VTST.32 Qd, Qn, Qm

A64 : CMTST Vd.4S, Vn.4S, Vm.4S La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

CompareTest(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vtstq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VTST.16 Qd, Qn, Qm

A64 : CMTST Vd.8H, Vn.8H, Vm.8H

CompareTest(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vtstq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VTST.32 Qd, Qn, Qm

A64 : CMTST Vd.4S, Vn.4S, Vm.4S

CompareTest(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vtst_u8 (uint8x8_t a, uint8x8_t b)

A32 : VTST.8 Dd, Dn, Dm

A64 : CMTST Vd.8B, Vn.8B, Vm.8B

CompareTest(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vtst_s16 (int16x4_t a, int16x4_t b)

A32 : VTST.16 Dd, Dn, Dm

A64 : CMTST Vd.4H, Vn.4H, Vm.4H

CompareTest(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vtst_s32 (int32x2_t a, int32x2_t b)

A32 : VTST.32 Dd, Dn, Dm

A64 : CMTST Vd.2S, Vn.2S, Vm.2S

CompareTest(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vtst_s8 (int8x8_t a, int8x8_t b)

A32 : VTST.8 Dd, Dn, Dm

A64 : CMTST Vd.8B, Vn.8B, Vm.8B

CompareTest(Vector64<Single>, Vector64<Single>)

uint32x2_t vtst_f32 (float32x2_t a, float32x2_t b)

A32 : VTST.32 Dd, Dn, Dm

A64 : CMTST Vd.2S, Vn.2S, Vm.2S La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

CompareTest(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vtst_u16 (uint16x4_t a, uint16x4_t b)

A32 : VTST.16 Dd, Dn, Dm

A64 : CMTST Vd.4H, Vn.4H, Vm.4H

CompareTest(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vtst_u32 (uint32x2_t a, uint32x2_t b)

A32 : VTST.32 Dd, Dn, Dm

A64 : CMTST Vd.2S, Vn.2S, Vm.2S

ConvertToInt32RoundAwayFromZero(Vector128<Single>)

int32x4_t vcvtaq_s32_f32 (float32x4_t a)

A32 : VCVTA.S32.F32 Qd, Qm

A64 : FCVTAS Vd.4S, Vn.4S

ConvertToInt32RoundAwayFromZero(Vector64<Single>)

int32x2_t vcvta_s32_f32 (float32x2_t a)

A32 : VCVTA.S32.F32 Dd, Dm

A64 : FCVTAS Vd.2S, Vn.2S

ConvertToInt32RoundAwayFromZeroScalar(Vector64<Single>)

int32_t vcvtas_s32_f32 (float32_t a)

A32 : VCVTA.S32.F32 Sd, Sm

A64 : FCVTAS Sd, Sn

ConvertToInt32RoundToEven(Vector128<Single>)

int32x4_t vcvtnq_s32_f32 (float32x4_t a)

A32 : VCVTN.S32.F32 Qd, Qm

A64 : FCVTNS Vd.4S, Vn.4S

ConvertToInt32RoundToEven(Vector64<Single>)

int32x2_t vcvtn_s32_f32 (float32x2_t a)

A32 : VCVTN.S32.F32 Dd, Dm

A64 : FCVTNS Vd.2S, Vn.2S

ConvertToInt32RoundToEvenScalar(Vector64<Single>)

int32_t vcvtns_s32_f32 (float32_t a)

A32 : VCVTN.S32.F32 Sd, Sm

A64 : FCVTNS Sd, Sn

ConvertToInt32RoundToNegativeInfinity(Vector128<Single>)

int32x4_t vcvtmq_s32_f32 (float32x4_t a)

A32 : VCVTM.S32.F32 Qd, Qm

A64 : FCVTMS Vd.4S, Vn.4S

ConvertToInt32RoundToNegativeInfinity(Vector64<Single>)

int32x2_t vcvtm_s32_f32 (float32x2_t a)

A32 : VCVTM.S32.F32 Dd, Dm

A64 : FCVTMS Vd.2S, Vn.2S

ConvertToInt32RoundToNegativeInfinityScalar(Vector64<Single>)

int32_t vcvtms_s32_f32 (float32_t a)

A32 : VCVTM.S32.F32 Sd, Sm

A64 : FCVTMS Sd, Sn

ConvertToInt32RoundToPositiveInfinity(Vector128<Single>)

int32x4_t vcvtpq_s32_f32 (float32x4_t a)

A32 : VCVTP.S32.F32 Qd, Qm

A64 : FCVTPS Vd.4S, Vn.4S

ConvertToInt32RoundToPositiveInfinity(Vector64<Single>)

int32x2_t vcvtp_s32_f32 (float32x2_t a)

A32 : VCVTP.S32.F32 Dd, Dm

A64 : FCVTPS Vd.2S, Vn.2S

ConvertToInt32RoundToPositiveInfinityScalar(Vector64<Single>)

int32_t vcvtps_s32_f32 (float32_t a)

A32 : VCVTP.S32.F32 Sd, Sm

A64 : FCVTPS Sd, Sn

ConvertToInt32RoundToZero(Vector128<Single>)

int32x4_t vcvtq_s32_f32 (float32x4_t a)

A32 : VCVT.S32.F32 Qd, Qm

A64 : FCVTZS Vd.4S, Vn.4S

ConvertToInt32RoundToZero(Vector64<Single>)

int32x2_t vcvt_s32_f32 (float32x2_t a)

A32 : VCVT.S32.F32 Dd, Dm

A64 : FCVTZS Vd.2S, Vn.2S

ConvertToInt32RoundToZeroScalar(Vector64<Single>)

int32_t vcvts_s32_f32 (float32_t a)

A32 : VCVT.S32.F32 Sd, Sm

A64 : FCVTZS Sd, Sn

ConvertToSingle(Vector128<Int32>)

float32x4_t vcvtq_f32_s32 (int32x4_t a)

A32 : VCVT.F32.S32 Qd, Qm

A64 : SCVTF Vd.4S, Vn.4S

ConvertToSingle(Vector128<UInt32>)

float32x4_t vcvtq_f32_u32 (uint32x4_t a)

A32 : VCVT.F32.U32 Qd, Qm

A64 : UCVTF Vd.4S, Vn.4S

ConvertToSingle(Vector64<Int32>)

float32x2_t vcvt_f32_s32 (int32x2_t a)

A32 : VCVT.F32.S32 Dd, Dm

A64 : SCVTF Vd.2S, Vn.2S

ConvertToSingle(Vector64<UInt32>)

float32x2_t vcvt_f32_u32 (uint32x2_t a)

A32 : VCVT.F32.U32 Dd, Dm

A64 : UCVTF Vd.2S, Vn.2S

ConvertToSingleScalar(Vector64<Int32>)

float32_t vcvts_f32_s32 (int32_t a)

A32 : VCVT.F32.S32 Sd, Sm

A64 : SCVTF Sd, Sn

ConvertToSingleScalar(Vector64<UInt32>)

float32_t vcvts_f32_u32 (uint32_t a)

A32 : VCVT.F32.U32 Sd, Sm

A64 : UCVTF Sd, Sn

ConvertToUInt32RoundAwayFromZero(Vector128<Single>)

uint32x4_t vcvtaq_u32_f32 (float32x4_t a)

A32 : VCVTA.U32.F32 Qd, Qm

A64 : FCVTAU Vd.4S, Vn.4S

ConvertToUInt32RoundAwayFromZero(Vector64<Single>)

uint32x2_t vcvta_u32_f32 (float32x2_t a)

A32 : VCVTA.U32.F32 Dd, Dm

A64 : FCVTAU Vd.2S, Vn.2S

ConvertToUInt32RoundAwayFromZeroScalar(Vector64<Single>)

uint32_t vcvtas_u32_f32 (float32_t a)

A32 : VCVTA.U32.F32 Sd, Sm

A64 : FCVTAU Sd, Sn

ConvertToUInt32RoundToEven(Vector128<Single>)

uint32x4_t vcvtnq_u32_f32 (float32x4_t a)

A32 : VCVTN.U32.F32 Qd, Qm

A64 : FCVTNU Vd.4S, Vn.4S

ConvertToUInt32RoundToEven(Vector64<Single>)

uint32x2_t vcvtn_u32_f32 (float32x2_t a)

A32 : VCVTN.U32.F32 Dd, Dm

A64 : FCVTNU Vd.2S, Vn.2S

ConvertToUInt32RoundToEvenScalar(Vector64<Single>)

uint32_t vcvtns_u32_f32 (float32_t a)

A32 : VCVTN.U32.F32 Sd, Sm

A64 : FCVTNU Sd, Sn

ConvertToUInt32RoundToNegativeInfinity(Vector128<Single>)

uint32x4_t vcvtmq_u32_f32 (float32x4_t a)

A32 : VCVTM.U32.F32 Qd, Qm

A64 : FCVTMU Vd.4S, Vn.4S

ConvertToUInt32RoundToNegativeInfinity(Vector64<Single>)

uint32x2_t vcvtm_u32_f32 (float32x2_t a)

A32 : VCVTM.U32.F32 Dd, Dm

A64 : FCVTMU Vd.2S, Vn.2S

ConvertToUInt32RoundToNegativeInfinityScalar(Vector64<Single>)

uint32_t vcvtms_u32_f32 (float32_t a)

A32 : VCVTM.U32.F32 Sd, Sm

A64 : FCVTMU Sd, Sn

ConvertToUInt32RoundToPositiveInfinity(Vector128<Single>)

uint32x4_t vcvtpq_u32_f32 (float32x4_t a)

A32 : VCVTP.U32.F32 Qd, Qm

A64 : FCVTPU Vd.4S, Vn.4S

ConvertToUInt32RoundToPositiveInfinity(Vector64<Single>)

uint32x2_t vcvtp_u32_f32 (float32x2_t a)

A32 : VCVTP.U32.F32 Dd, Dm

A64 : FCVTPU Vd.2S, Vn.2S

ConvertToUInt32RoundToPositiveInfinityScalar(Vector64<Single>)

uint32_t vcvtps_u32_f32 (float32_t a)

A32 : VCVTP.U32.F32 Sd, Sm

A64 : FCVTPU Sd, Sn

ConvertToUInt32RoundToZero(Vector128<Single>)

uint32x4_t vcvtq_u32_f32 (float32x4_t a)

A32 : VCVT.U32.F32 Qd, Qm

A64 : FCVTZU Vd.4S, Vn.4S

ConvertToUInt32RoundToZero(Vector64<Single>)

uint32x2_t vcvt_u32_f32 (float32x2_t a)

A32 : VCVT.U32.F32 Dd, Dm

A64 : FCVTZU Vd.2S, Vn.2S

ConvertToUInt32RoundToZeroScalar(Vector64<Single>)

uint32_t vcvts_u32_f32 (float32_t a)

A32 : VCVT.U32.F32 Sd, Sm

A64 : FCVTZU Sd, Sn

DivideScalar(Vector64<Double>, Vector64<Double>)

float64x1_t vdiv_f64 (float64x1_t a, float64x1_t b)

A32 : VDIV.F64 Dd, Dn, Dm

A64 : FDIV Dd, Dn, Dm

DivideScalar(Vector64<Single>, Vector64<Single>)

float32_t vdivs_f32 (float32_t a, float32_t b)

A32 : VDIV.F32 Sd, Sn, Sm

A64 : FDIV Sd, Sn, Sm La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

DuplicateSelectedScalarToVector128(Vector128<Byte>, Byte)

uint8x16_t vdupq_lane_u8 (uint8x16_t vec, const int lane)

A32 : VDUP.8 Qd, Dm[index]

A64 : DUP Vd.16B, Vn.B[index]

DuplicateSelectedScalarToVector128(Vector128<Int16>, Byte)

int16x8_t vdupq_lane_s16 (int16x8_t vec, const int lane)

A32 : VDUP.16 Qd, Dm[index]

A64 : DUP Vd.8H, Vn.H[index]

DuplicateSelectedScalarToVector128(Vector128<Int32>, Byte)

int32x4_t vdupq_lane_s32 (int32x4_t vec, const int lane)

A32 : VDUP.32 Qd, Dm[index]

A64 : DUP Vd.4S, Vn.S[index]

DuplicateSelectedScalarToVector128(Vector128<SByte>, Byte)

int8x16_t vdupq_lane_s8 (int8x16_t vec, const int lane)

A32 : VDUP.8 Qd, Dm[index]

A64 : DUP Vd.16B, Vn.B[index]

DuplicateSelectedScalarToVector128(Vector128<Single>, Byte)

float32x4_t vdupq_lane_f32 (float32x4_t vec, const int lane)

A32 : VDUP.32 Qd, Dm[index]

A64 : DUP Vd.4S, Vn.S[index]

DuplicateSelectedScalarToVector128(Vector128<UInt16>, Byte)

uint16x8_t vdupq_lane_u16 (uint16x8_t vec, const int lane)

A32 : VDUP.16 Qd, Dm[index]

A64 : DUP Vd.8H, Vn.H[index]

DuplicateSelectedScalarToVector128(Vector128<UInt32>, Byte)

uint32x4_t vdupq_lane_u32 (uint32x4_t vec, const int lane)

A32 : VDUP.32 Qd, Dm[index]

A64 : DUP Vd.4S, Vn.S[index]

DuplicateSelectedScalarToVector128(Vector64<Byte>, Byte)

uint8x16_t vdupq_lane_u8 (uint8x8_t vec, const int lane)

A32 : VDUP.8 Qd, Dm[index]

A64 : DUP Vd.16B, Vn.B[index]

DuplicateSelectedScalarToVector128(Vector64<Int16>, Byte)

int16x8_t vdupq_lane_s16 (int16x4_t vec, const int lane)

A32 : VDUP.16 Qd, Dm[index]

A64 : DUP Vd.8H, Vn.H[index]

DuplicateSelectedScalarToVector128(Vector64<Int32>, Byte)

int32x4_t vdupq_lane_s32 (int32x2_t vec, const int lane)

A32 : VDUP.32 Qd, Dm[index]

A64 : DUP Vd.4S, Vn.S[index]

DuplicateSelectedScalarToVector128(Vector64<SByte>, Byte)

int8x16_t vdupq_lane_s8 (int8x8_t vec, const int lane)

A32 : VDUP.8 Qd, Dm[index]

A64 : DUP Vd.16B, Vn.B[index]

DuplicateSelectedScalarToVector128(Vector64<Single>, Byte)

float32x4_t vdupq_lane_f32 (float32x2_t vec, const int lane)

A32 : VDUP.32 Qd, Dm[index]

A64 : DUP Vd.4S, Vn.S[index]

DuplicateSelectedScalarToVector128(Vector64<UInt16>, Byte)

uint16x8_t vdupq_lane_u16 (uint16x4_t vec, const int lane)

A32 : VDUP.16 Qd, Dm[index]

A64 : DUP Vd.8H, Vn.H[index]

DuplicateSelectedScalarToVector128(Vector64<UInt32>, Byte)

uint32x4_t vdupq_lane_u32 (uint32x2_t vec, const int lane)

A32 : VDUP.32 Qd, Dm[index]

A64 : DUP Vd.4S, Vn.S[index]

DuplicateSelectedScalarToVector64(Vector128<Byte>, Byte)

uint8x8_t vdup_laneq_u8 (uint8x16_t vec, const int lane)

A32 : VDUP.8 Dd, Dm[index]

A64 : DUP Vd.8B, Vn.B[index]

DuplicateSelectedScalarToVector64(Vector128<Int16>, Byte)

int16x4_t vdup_laneq_s16 (int16x8_t vec, const int lane)

A32 : VDUP.16 Dd, Dm[index]

A64 : DUP Vd.4H, Vn.H[index]

DuplicateSelectedScalarToVector64(Vector128<Int32>, Byte)

int32x2_t vdup_laneq_s32 (int32x4_t vec, const int lane)

A32 : VDUP.32 Dd, Dm[index]

A64 : DUP Vd.2S, Vn.S[index]

DuplicateSelectedScalarToVector64(Vector128<SByte>, Byte)

int8x8_t vdup_laneq_s8 (int8x16_t vec, const int lane)

A32 : VDUP.8 Dd, Dm[index]

A64 : DUP Vd.8B, Vn.B[index]

DuplicateSelectedScalarToVector64(Vector128<Single>, Byte)

float32x2_t vdup_laneq_f32 (float32x4_t vec, const int lane)

A32 : VDUP.32 Dd, Dm[index]

A64 : DUP Vd.2S, Vn.S[index]

DuplicateSelectedScalarToVector64(Vector128<UInt16>, Byte)

uint16x4_t vdup_laneq_u16 (uint16x8_t vec, const int lane)

A32 : VDUP.16 Dd, Dm[index]

A64 : DUP Vd.4H, Vn.H[index]

DuplicateSelectedScalarToVector64(Vector128<UInt32>, Byte)

uint32x2_t vdup_laneq_u32 (uint32x4_t vec, const int lane)

A32 : VDUP.32 Dd, Dm[index]

A64 : DUP Vd.2S, Vn.S[index]

DuplicateSelectedScalarToVector64(Vector64<Byte>, Byte)

uint8x8_t vdup_lane_u8 (uint8x8_t vec, const int lane)

A32 : VDUP.8 Dd, Dm[index]

A64 : DUP Vd.8B, Vn.B[index]

DuplicateSelectedScalarToVector64(Vector64<Int16>, Byte)

int16x4_t vdup_lane_s16 (int16x4_t vec, const int lane)

A32 : VDUP.16 Dd, Dm[index]

A64 : DUP Vd.4H, Vn.H[index]

DuplicateSelectedScalarToVector64(Vector64<Int32>, Byte)

int32x2_t vdup_lane_s32 (int32x2_t vec, const int lane)

A32 : VDUP.32 Dd, Dm[index]

A64 : DUP Vd.2S, Vn.S[index]

DuplicateSelectedScalarToVector64(Vector64<SByte>, Byte)

int8x8_t vdup_lane_s8 (int8x8_t vec, const int lane)

A32 : VDUP.8 Dd, Dm[index]

A64 : DUP Vd.8B, Vn.B[index]

DuplicateSelectedScalarToVector64(Vector64<Single>, Byte)

float32x2_t vdup_lane_f32 (float32x2_t vec, const int lane)

A32 : VDUP.32 Dd, Dm[index]

A64 : DUP Vd.2S, Vn.S[index]

DuplicateSelectedScalarToVector64(Vector64<UInt16>, Byte)

uint16x4_t vdup_lane_u16 (uint16x4_t vec, const int lane)

A32 : VDUP.16 Dd, Dm[index]

A64 : DUP Vd.4H, Vn.H[index]

DuplicateSelectedScalarToVector64(Vector64<UInt32>, Byte)

uint32x2_t vdup_lane_u32 (uint32x2_t vec, const int lane)

A32 : VDUP.32 Dd, Dm[index]

A64 : DUP Vd.2S, Vn.S[index]

DuplicateToVector128(Byte)

uint8x16_t vdupq_n_u8 (uint8_t value)

A32 : VDUP.8 Qd, Rt

A64 : DUP Vd.16B, Rn

DuplicateToVector128(Int16)

int16x8_t vdupq_n_s16 (int16_t value)

A32 : VDUP.16 Qd, Rt

A64 : DUP Vd.8H, Rn

DuplicateToVector128(Int32)

int32x4_t vdupq_n_s32 (int32_t value)

A32 : VDUP.32 Qd, Rt

A64 : DUP Vd.4S, Rn

DuplicateToVector128(SByte)

int8x16_t vdupq_n_s8 (int8_t value)

A32 : VDUP.8 Qd, Rt

A64 : DUP Vd.16B, Rn

DuplicateToVector128(Single)

float32x4_t vdupq_n_f32 (float32_t value)

A32 : VDUP Qd, Dm[0]

A64 : DUP Vd.4S, Vn.S[0]

DuplicateToVector128(UInt16)

uint16x8_t vdupq_n_u16 (uint16_t value)

A32 : VDUP.16 Qd, Rt

A64 : DUP Vd.8H, Rn

DuplicateToVector128(UInt32)

uint32x4_t vdupq_n_u32 (uint32_t value)

A32 : VDUP.32 Qd, Rt

A64 : DUP Vd.4S, Rn

DuplicateToVector64(Byte)

uint8x8_t vdup_n_u8 (uint8_t value)

A32 : VDUP.8 Dd, Rt

A64 : DUP Vd.8B, Rn

DuplicateToVector64(Int16)

int16x4_t vdup_n_s16 (int16_t value)

A32 : VDUP.16 Dd, Rt

A64 : DUP Vd.4H, Rn

DuplicateToVector64(Int32)

int32x2_t vdup_n_s32 (int32_t value)

A32 : VDUP.32 Dd, Rt

A64 : DUP Vd.2S, Rn

DuplicateToVector64(SByte)

int8x8_t vdup_n_s8 (int8_t value)

A32 : VDUP.8 Dd, Rt

A64 : DUP Vd.8B, Rn

DuplicateToVector64(Single)

float32x2_t vdup_n_f32 (float32_t value)

A32 : VDUP Dd, Dm[0]

A64 : DUP Vd.2S, Vn.S[0]

DuplicateToVector64(UInt16)

uint16x4_t vdup_n_u16 (uint16_t value)

A32 : VDUP.16 Dd, Rt

A64 : DUP Vd.4H, Rn

DuplicateToVector64(UInt32)

uint32x2_t vdup_n_u32 (uint32_t value)

A32 : VDUP.32 Dd, Rt

A64 : DUP Vd.2S, Rn

Equals(Object)

Détermine si l'objet spécifié est égal à l'objet actuel.

(Hérité de Object)
Extract(Vector128<Byte>, Byte)

uint8_t vgetq_lane_u8 (uint8x16_t v, const int lane)

A32 : VMOV.U8 Rt, Dn[lane]

A64 : UMOV Wd, Vn.B[lane]

Extract(Vector128<Double>, Byte)

float64_t vgetq_lane_f64 (float64x2_t v, const int lane)

A32 : VMOV.F64 Dd, Dm

A64 : DUP Dd, Vn.D[lane]

Extract(Vector128<Int16>, Byte)

int16_t vgetq_lane_s16 (int16x8_t v, const int lane)

A32 : VMOV.S16 Rt, Dn[lane]

A64 : SMOV Wd, Vn.H[lane]

Extract(Vector128<Int32>, Byte)

int32_t vgetq_lane_s32 (int32x4_t v, const int lane)

A32 : VMOV.32 Rt, Dn[lane]

A64 : SMOV Wd, Vn.S[lane]

Extract(Vector128<Int64>, Byte)

int64_t vgetq_lane_s64 (int64x2_t v, const int lane)

A32 : VMOV Rt, Rt2, Dm

A64 : UMOV Xd, Vn.D[lane]

Extract(Vector128<SByte>, Byte)

int8_t vgetq_lane_s8 (int8x16_t v, const int lane)

A32 : VMOV.S8 Rt, Dn[lane]

A64 : SMOV Wd, Vn.B[lane]

Extract(Vector128<Single>, Byte)

float32_t vgetq_lane_f32 (float32x4_t v, const int lane)

A32 : VMOV.F32 Sd, Sm

A64 : DUP Sd, Vn.S[lane]

Extract(Vector128<UInt16>, Byte)

uint16_t vgetq_lane_u16 (uint16x8_t v, const int lane)

A32 : VMOV.U16 Rt, Dn[lane]

A64 : UMOV Wd, Vn.H[lane]

Extract(Vector128<UInt32>, Byte)

uint32_t vgetq_lane_u32 (uint32x4_t v, const int lane)

A32 : VMOV.32 Rt, Dn[lane]

A64 : UMOV Wd, Vn.S[lane]

Extract(Vector128<UInt64>, Byte)

uint64_t vgetq_lane_u64 (uint64x2_t v, const int lane)

A32 : VMOV Rt, Rt2, Dm

A64 : UMOV Xd, Vn.D[lane]

Extract(Vector64<Byte>, Byte)

uint8_t vget_lane_u8 (uint8x8_t v, const int lane)

A32 : VMOV.U8 Rt, Dn[lane]

A64 : UMOV Wd, Vn.B[lane]

Extract(Vector64<Int16>, Byte)

int16_t vget_lane_s16 (int16x4_t v, const int lane)

A32 : VMOV.S16 Rt, Dn[lane]

A64 : SMOV Wd, Vn.H[lane]

Extract(Vector64<Int32>, Byte)

int32_t vget_lane_s32 (int32x2_t v, const int lane)

A32 : VMOV.32 Rt, Dn[lane]

A64 : SMOV Wd, Vn.S[lane]

Extract(Vector64<SByte>, Byte)

int8_t vget_lane_s8 (int8x8_t v, const int lane)

A32 : VMOV.S8 Rt, Dn[lane]

A64 : SMOV Wd, Vn.B[lane]

Extract(Vector64<Single>, Byte)

float32_t vget_lane_f32 (float32x2_t v, const int lane)

A32 : VMOV.F32 Sd, Sm

A64 : DUP Sd, Vn.S[lane]

Extract(Vector64<UInt16>, Byte)

uint16_t vget_lane_u16 (uint16x4_t v, const int lane)

A32 : VMOV.U16 Rt, Dn[lane]

A64 : UMOV Wd, Vn.H[lane]

Extract(Vector64<UInt32>, Byte)

uint32_t vget_lane_u32 (uint32x2_t v, const int lane)

A32 : VMOV.32 Rt, Dn[lane]

A64 : UMOV Wd, Vn.S[lane]

ExtractNarrowingLower(Vector128<Int16>)

int8x8_t vmovn_s16 (int16x8_t a)

A32 : VMOVN.I16 Dd, Qm

A64 : XTN Vd.8B, Vn.8H

ExtractNarrowingLower(Vector128<Int32>)

int16x4_t vmovn_s32 (int32x4_t a)

A32 : VMOVN.I32 Dd, Qm

A64 : XTN Vd.4H, Vn.4S

ExtractNarrowingLower(Vector128<Int64>)

int32x2_t vmovn_s64 (int64x2_t a)

A32 : VMOVN.I64 Dd, Qm

A64 : XTN Vd.2S, Vn.2D

ExtractNarrowingLower(Vector128<UInt16>)

uint8x8_t vmovn_u16 (uint16x8_t a)

A32 : VMOVN.I16 Dd, Qm

A64 : XTN Vd.8B, Vn.8H

ExtractNarrowingLower(Vector128<UInt32>)

uint16x4_t vmovn_u32 (uint32x4_t a)

A32 : VMOVN.I32 Dd, Qm

A64 : XTN Vd.4H, Vn.4S

ExtractNarrowingLower(Vector128<UInt64>)

uint32x2_t vmovn_u64 (uint64x2_t a)

A32 : VMOVN.I64 Dd, Qm

A64 : XTN Vd.2S, Vn.2D

ExtractNarrowingSaturateLower(Vector128<Int16>)

int8x8_t vqmovn_s16 (int16x8_t a)

A32 : VQMOVN. S16 Dd, Qm

A64 : SQXTN Vd.8B, Vn.8H

ExtractNarrowingSaturateLower(Vector128<Int32>)

int16x4_t vqmovn_s32 (int32x4_t a)

A32 : VQMOVN. S32 Dd, Qm

A64 : SQXTN Vd.4H, Vn.4S

ExtractNarrowingSaturateLower(Vector128<Int64>)

int32x2_t vqmovn_s64 (int64x2_t a)

A32 : VQMOVN. S64 Dd, Qm

A64 : SQXTN Vd.2S, Vn.2D

ExtractNarrowingSaturateLower(Vector128<UInt16>)

uint8x8_t vqmovn_u16 (uint16x8_t a)

A32 : VQMOVN. U16 Dd, Qm

A64 : UQXTN Vd.8B, Vn.8H

ExtractNarrowingSaturateLower(Vector128<UInt32>)

uint16x4_t vqmovn_u32 (uint32x4_t a)

A32 : VQMOVN. U32 Dd, Qm

A64 : UQXTN Vd.4H, Vn.4S

ExtractNarrowingSaturateLower(Vector128<UInt64>)

uint32x2_t vqmovn_u64 (uint64x2_t a)

A32 : VQMOVN. U64 Dd, Qm

A64 : UQXTN Vd.2S, Vn.2D

ExtractNarrowingSaturateUnsignedLower(Vector128<Int16>)

uint8x8_t vqmovun_s16 (int16x8_t a)

A32 : VQMOVUN. S16 Dd, Qm

A64 : SQXTUN Vd.8B, Vn.8H

ExtractNarrowingSaturateUnsignedLower(Vector128<Int32>)

uint16x4_t vqmovun_s32 (int32x4_t a)

A32 : VQMOVUN. S32 Dd, Qm

A64 : SQXTUN Vd.4H, Vn.4S

ExtractNarrowingSaturateUnsignedLower(Vector128<Int64>)

uint32x2_t vqmovun_s64 (int64x2_t a)

A32 : VQMOVUN. S64 Dd, Qm

A64 : SQXTUN Vd.2S, Vn.2D

ExtractNarrowingSaturateUnsignedUpper(Vector64<Byte>, Vector128<Int16>)

uint8x16_t vqmovun_high_s16 (uint8x8_t r, int16x8_t a)

A32 : VQMOVUN. S16 Dd+1, Qm

A64 : SQXTUN2 Vd.16B, Vn.8H

ExtractNarrowingSaturateUnsignedUpper(Vector64<UInt16>, Vector128<Int32>)

uint16x8_t vqmovun_high_s32 (uint16x4_t r, int32x4_t a)

A32 : VQMOVUN. S32 Dd+1, Qm

A64 : SQXTUN2 Vd.8H, Vn.4S

ExtractNarrowingSaturateUnsignedUpper(Vector64<UInt32>, Vector128<Int64>)

uint32x4_t vqmovun_high_s64 (uint32x2_t r, int64x2_t a)

A32 : VQMOVUN. S64 Dd+1, Qm

A64 : SQXTUN2 Vd.4S, Vn.2D

ExtractNarrowingSaturateUpper(Vector64<Byte>, Vector128<UInt16>)

uint8x16_t vqmovn_high_u16 (uint8x8_t r, uint16x8_t a)

A32 : VQMOVN. U16 Dd+1, Qm

A64 : UQXTN2 Vd.16B, Vn.8H

ExtractNarrowingSaturateUpper(Vector64<Int16>, Vector128<Int32>)

int16x8_t vqmovn_high_s32 (int16x4_t r, int32x4_t a)

A32 : VQMOVN. S32 Dd+1, Qm

A64 : SQXTN2 Vd.8H, Vn.4S

ExtractNarrowingSaturateUpper(Vector64<Int32>, Vector128<Int64>)

int32x4_t vqmovn_high_s64 (int32x2_t r, int64x2_t a)

A32 : VQMOVN. S64 Dd+1, Qm

A64 : SQXTN2 Vd.4S, Vn.2D

ExtractNarrowingSaturateUpper(Vector64<SByte>, Vector128<Int16>)

int8x16_t vqmovn_high_s16 (int8x8_t r, int16x8_t a)

A32 : VQMOVN. S16 Dd+1, Qm

A64 : SQXTN2 Vd.16B, Vn.8H

ExtractNarrowingSaturateUpper(Vector64<UInt16>, Vector128<UInt32>)

uint16x8_t vqmovn_high_u32 (uint16x4_t r, uint32x4_t a)

A32 : VQMOVN. U32 Dd+1, Qm

A64 : UQXTN2 Vd.8H, Vn.4S

ExtractNarrowingSaturateUpper(Vector64<UInt32>, Vector128<UInt64>)

uint32x4_t vqmovn_high_u64 (uint32x2_t r, uint64x2_t a)

A32 : VQMOVN. U64 Dd+1, Qm

A64 : UQXTN2 Vd.4S, Vn.2D

ExtractNarrowingUpper(Vector64<Byte>, Vector128<UInt16>)

uint8x16_t vmovn_high_u16 (uint8x8_t r, uint16x8_t a)

A32 : VMOVN.I16 Dd+1, Qm

A64 : XTN2 Vd.16B, Vn.8H

ExtractNarrowingUpper(Vector64<Int16>, Vector128<Int32>)

int16x8_t vmovn_high_s32 (int16x4_t r, int32x4_t a)

A32 : VMOVN.I32 Dd+1, Qm

A64 : XTN2 Vd.8H, Vn.4S

ExtractNarrowingUpper(Vector64<Int32>, Vector128<Int64>)

int32x4_t vmovn_high_s64 (int32x2_t r, int64x2_t a)

A32 : VMOVN.I64 Dd+1, Qm

A64 : XTN2 Vd.4S, Vn.2D

ExtractNarrowingUpper(Vector64<SByte>, Vector128<Int16>)

int8x16_t vmovn_high_s16 (int8x8_t r, int16x8_t a)

A32 : VMOVN.I16 Dd+1, Qm

A64 : XTN2 Vd.16B, Vn.8H

ExtractNarrowingUpper(Vector64<UInt16>, Vector128<UInt32>)

uint16x8_t vmovn_high_u32 (uint16x4_t r, uint32x4_t a)

A32 : VMOVN.I32 Dd+1, Qm

A64 : XTN2 Vd.8H, Vn.4S

ExtractNarrowingUpper(Vector64<UInt32>, Vector128<UInt64>)

uint32x4_t vmovn_high_u64 (uint32x2_t r, uint64x2_t a)

A32 : VMOVN.I64 Dd+1, Qm

A64 : XTN2 Vd.4S, Vn.2D

ExtractVector128(Vector128<Byte>, Vector128<Byte>, Byte)

uint8x16_t vextq_s8 (uint8x16_t a, uint8x16_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #n

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #n

ExtractVector128(Vector128<Double>, Vector128<Double>, Byte)

float64x2_t vextq_f64 (float64x2_t a, float64x2_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*8)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*8)

ExtractVector128(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vextq_s16 (int16x8_t a, int16x8_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*2)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*2)

ExtractVector128(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vextq_s32 (int32x4_t a, int32x4_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*4)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*4)

ExtractVector128(Vector128<Int64>, Vector128<Int64>, Byte)

int64x2_t vextq_s64 (int64x2_t a, int64x2_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*8)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*8)

ExtractVector128(Vector128<SByte>, Vector128<SByte>, Byte)

int8x16_t vextq_s8 (int8x16_t a, int8x16_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #n

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #n

ExtractVector128(Vector128<Single>, Vector128<Single>, Byte)

float32x4_t vextq_f32 (float32x4_t a, float32x4_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*4)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*4)

ExtractVector128(Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vextq_s16 (uint16x8_t a, uint16x8_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*2)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*2)

ExtractVector128(Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vextq_s32 (uint32x4_t a, uint32x4_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*4)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*4)

ExtractVector128(Vector128<UInt64>, Vector128<UInt64>, Byte)

uint64x2_t vextq_s64 (uint64x2_t a, uint64x2_t b, const int n)

A32 : VEXT.8 Qd, Qn, Qm, #(n*8)

A64 : EXT Vd.16B, Vn.16B, Vm.16B, #(n*8)

ExtractVector64(Vector64<Byte>, Vector64<Byte>, Byte)

uint8x8_t vext_s8 (uint8x8_t a, uint8x8_t b, const int n)

A32 : VEXT.8 Dd, Dn, Dm, #n

A64 : EXT Vd.8B, Vn.8B, Vm.8B, #n

ExtractVector64(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vext_s16 (int16x4_t a, int16x4_t b, const int n)

A32 : VEXT.8 Dd, Dn, Dm, #(n*2)

A64 : EXT Vd.8B, Vn.8B, Vm.8B, #(n*2)

ExtractVector64(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vext_s32 (int32x2_t a, int32x2_t b, const int n)

A32 : VEXT.8 Dd, Dn, Dm, #(n*4)

A64 : EXT Vd.8B, Vn.8B, Vm.8B, #(n*4)

ExtractVector64(Vector64<SByte>, Vector64<SByte>, Byte)

int8x8_t vext_s8 (int8x8_t a, int8x8_t b, const int n)

A32 : VEXT.8 Dd, Dn, Dm, #n

A64 : EXT Vd.8B, Vn.8B, Vm.8B, #n

ExtractVector64(Vector64<Single>, Vector64<Single>, Byte)

float32x2_t vext_f32 (float32x2_t a, float32x2_t b, const int n)

A32 : VEXT.8 Dd, Dn, Dm, #(n*4)

A64 : EXT Vd.8B, Vn.8B, Vm.8B, #(n*4)

ExtractVector64(Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vext_s16 (uint16x4_t a, uint16x4_t b, const int n)

A32 : VEXT.8 Dd, Dn, Dm, #(n*2)

A64 : EXT Vd.8B, Vn.8B, Vm.8B, #(n*2)

ExtractVector64(Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vext_s32 (uint32x2_t a, uint32x2_t b, const int n)

A32 : VEXT.8 Dd, Dn, Dm, #(n*4)

A64 : EXT Vd.8B, Vn.8B, Vm.8B, #(n*4)

Floor(Vector128<Single>)

float32x4_t vrndmq_f32 (float32x4_t a)

A32 : VRINTM.F32 Qd, Qm

A64 : FRINTM Vd.4S, Vn.4S

Floor(Vector64<Single>)

float32x2_t vrndm_f32 (float32x2_t a)

A32 : VRINTM.F32 Dd, Dm

A64 : FRINTM Vd.2S, Vn.2S

FloorScalar(Vector64<Double>)

float64x1_t vrndm_f64 (float64x1_t a)

A32 : VRINTM.F64 Dd, Dm

A64 : FRINTM Dd, Dn

FloorScalar(Vector64<Single>)

float32_t vrndms_f32 (float32_t a)

A32 : VRINTM.F32 Sd, Sm

A64 : FRINTM Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

FusedAddHalving(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vhaddq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VHADD.U8 Qd, Qn, Qm

A64 : UHADD Vd.16B, Vn.16B, Vm.16B

FusedAddHalving(Vector128<Int16>, Vector128<Int16>)

int16x8_t vhaddq_s16 (int16x8_t a, int16x8_t b)

A32 : VHADD.S16 Qd, Qn, Qm

A64 : SHADD Vd.8H, Vn.8H, Vm.8H

FusedAddHalving(Vector128<Int32>, Vector128<Int32>)

int32x4_t vhaddq_s32 (int32x4_t a, int32x4_t b)

A32 : VHADD.S32 Qd, Qn, Qm

A64 : SHADD Vd.4S, Vn.4S, Vm.4S

FusedAddHalving(Vector128<SByte>, Vector128<SByte>)

int8x16_t vhaddq_s8 (int8x16_t a, int8x16_t b)

A32 : VHADD.S8 Qd, Qn, Qm

A64 : SHADD Vd.16B, Vn.16B, Vm.16B

FusedAddHalving(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vhaddq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VHADD.U16 Qd, Qn, Qm

A64 : UHADD Vd.8H, Vn.8H, Vm.8H

FusedAddHalving(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vhaddq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VHADD.U32 Qd, Qn, Qm

A64 : UHADD Vd.4S, Vn.4S, Vm.4S

FusedAddHalving(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vhadd_u8 (uint8x8_t a, uint8x8_t b)

A32 : VHADD.U8 Dd, Dn, Dm

A64 : UHADD Vd.8B, Vn.8B, Vm.8B

FusedAddHalving(Vector64<Int16>, Vector64<Int16>)

int16x4_t vhadd_s16 (int16x4_t a, int16x4_t b)

A32 : VHADD.S16 Dd, Dn, Dm

A64 : SHADD Vd.4H, Vn.4H, Vm.4H

FusedAddHalving(Vector64<Int32>, Vector64<Int32>)

int32x2_t vhadd_s32 (int32x2_t a, int32x2_t b)

A32 : VHADD.S32 Dd, Dn, Dm

A64 : SHADD Vd.2S, Vn.2S, Vm.2S

FusedAddHalving(Vector64<SByte>, Vector64<SByte>)

int8x8_t vhadd_s8 (int8x8_t a, int8x8_t b)

A32 : VHADD.S8 Dd, Dn, Dm

A64 : SHADD Vd.8B, Vn.8B, Vm.8B

FusedAddHalving(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vhadd_u16 (uint16x4_t a, uint16x4_t b)

A32 : VHADD.U16 Dd, Dn, Dm

A64 : UHADD Vd.4H, Vn.4H, Vm.4H

FusedAddHalving(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vhadd_u32 (uint32x2_t a, uint32x2_t b)

A32 : VHADD.U32 Dd, Dn, Dm

A64 : UHADD Vd.2S, Vn.2S, Vm.2S

FusedAddRoundedHalving(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vrhaddq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VRHADD.U8 Qd, Qn, Qm

A64 : URHADD Vd.16B, Vn.16B, Vm.16B

FusedAddRoundedHalving(Vector128<Int16>, Vector128<Int16>)

int16x8_t vrhaddq_s16 (int16x8_t a, int16x8_t b)

A32 : VRHADD.S16 Qd, Qn, Qm

A64 : SRHADD Vd.8H, Vn.8H, Vm.8H

FusedAddRoundedHalving(Vector128<Int32>, Vector128<Int32>)

int32x4_t vrhaddq_s32 (int32x4_t a, int32x4_t b)

A32 : VRHADD.S32 Qd, Qn, Qm

A64 : SRHADD Vd.4S, Vn.4S, Vm.4S

FusedAddRoundedHalving(Vector128<SByte>, Vector128<SByte>)

int8x16_t vrhaddq_s8 (int8x16_t a, int8x16_t b)

A32 : VRHADD.S8 Qd, Qn, Qm

A64 : SRHADD Vd.16B, Vn.16B, Vm.16B

FusedAddRoundedHalving(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vrhaddq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VRHADD.U16 Qd, Qn, Qm

A64 : URHADD Vd.8H, Vn.8H, Vm.8H

FusedAddRoundedHalving(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vrhaddq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VRHADD.U32 Qd, Qn, Qm

A64 : URHADD Vd.4S, Vn.4S, Vm.4S

FusedAddRoundedHalving(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vrhadd_u8 (uint8x8_t a, uint8x8_t b)

A32 : VRHADD.U8 Dd, Dn, Dm

A64 : URHADD Vd.8B, Vn.8B, Vm.8B

FusedAddRoundedHalving(Vector64<Int16>, Vector64<Int16>)

int16x4_t vrhadd_s16 (int16x4_t a, int16x4_t b)

A32 : VRHADD.S16 Dd, Dn, Dm

A64 : SRHADD Vd.4H, Vn.4H, Vm.4H

FusedAddRoundedHalving(Vector64<Int32>, Vector64<Int32>)

int32x2_t vrhadd_s32 (int32x2_t a, int32x2_t b)

A32 : VRHADD.S32 Dd, Dn, Dm

A64 : SRHADD Vd.2S, Vn.2S, Vm.2S

FusedAddRoundedHalving(Vector64<SByte>, Vector64<SByte>)

int8x8_t vrhadd_s8 (int8x8_t a, int8x8_t b)

A32 : VRHADD.S8 Dd, Dn, Dm

A64 : SRHADD Vd.8B, Vn.8B, Vm.8B

FusedAddRoundedHalving(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vrhadd_u16 (uint16x4_t a, uint16x4_t b)

A32 : VRHADD.U16 Dd, Dn, Dm

A64 : URHADD Vd.4H, Vn.4H, Vm.4H

FusedAddRoundedHalving(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vrhadd_u32 (uint32x2_t a, uint32x2_t b)

A32 : VRHADD.U32 Dd, Dn, Dm

A64 : URHADD Vd.2S, Vn.2S, Vm.2S

FusedMultiplyAdd(Vector128<Single>, Vector128<Single>, Vector128<Single>)

float32x4_t vfmaq_f32 (float32x4_t a, float32x4_t b, float32x4_t c)

A32 : VFMA.F32 Qd, Qn, Qm

A64 : FMLA Vd.4S, Vn.4S, Vm.4S

FusedMultiplyAdd(Vector64<Single>, Vector64<Single>, Vector64<Single>)

float32x2_t vfma_f32 (float32x2_t a, float32x2_t b, float32x2_t c)

A32 : VFMA.F32 Dd, Dn, Dm

A64 : FMLA Vd.2S, Vn.2S, Vm.2S

FusedMultiplyAddNegatedScalar(Vector64<Double>, Vector64<Double>, Vector64<Double>)

float64x1_t vfnma_f64 (float64x1_t a, float64x1_t b, float64x1_t c)

A32 : VFNMA.F64 Dd, Dn, Dm

A64 : FNMADD Dd, Dn, Dm, Da La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

FusedMultiplyAddNegatedScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>)

float32_t vfnmas_f32 (float32_t a, float32_t b, float32_t c)

A32 : VFNMA.F32 Sd, Sn, Sm

A64 : FNMADD Sd, Sn, Sm, Sa La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

FusedMultiplyAddScalar(Vector64<Double>, Vector64<Double>, Vector64<Double>)

float64x1_t vfma_f64 (float64x1_t a, float64x1_t b, float64x1_t c)

A32 : VFMA.F64 Dd, Dn, Dm

A64 : FMADD Dd, Dn, Dm, Da

FusedMultiplyAddScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>)

float32_t vfmas_f32 (float32_t a, float32_t b, float32_t c)

A32 : VFMA.F32 Sd, Sn, Sm

A64 : FMADD Sd, Sn, Sm, Sa La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

FusedMultiplySubtract(Vector128<Single>, Vector128<Single>, Vector128<Single>)

float32x4_t vfmsq_f32 (float32x4_t a, float32x4_t b, float32x4_t c)

A32 : VFMS.F32 Qd, Qn, Qm

A64 : FMLS Vd.4S, Vn.4S, Vm.4S

FusedMultiplySubtract(Vector64<Single>, Vector64<Single>, Vector64<Single>)

float32x2_t vfms_f32 (float32x2_t a, float32x2_t b, float32x2_t c)

A32 : VFMS.F32 Dd, Dn, Dm

A64 : FMLS Vd.2S, Vn.2S, Vm.2S

FusedMultiplySubtractNegatedScalar(Vector64<Double>, Vector64<Double>, Vector64<Double>)

float64x1_t vfnms_f64 (float64x1_t a, float64x1_t b, float64x1_t c)

A32 : VFNMS.F64 Dd, Dn, Dm

A64 : FNMSUB Dd, Dn, Dm, Da La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

FusedMultiplySubtractNegatedScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>)

float32_t vfnmss_f32 (float32_t a, float32_t b, float32_t c)

A32 : VFNMS.F32 Sd, Sn, Sm

A64 : FNMSUB Sd, Sn, Sm, Sa La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

FusedMultiplySubtractScalar(Vector64<Double>, Vector64<Double>, Vector64<Double>)

float64x1_t vfms_f64 (float64x1_t a, float64x1_t b, float64x1_t c)

A32 : VFMS.F64 Dd, Dn, Dm

A64 : FMSUB Dd, Dn, Dm, Da

FusedMultiplySubtractScalar(Vector64<Single>, Vector64<Single>, Vector64<Single>)

float32_t vfmss_f32 (float32_t a, float32_t b, float32_t c)

A32 : VFMS.F32 Sd, Sn, Sm

A64 : FMSUB Sd, Sn, Sm, Sa La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

FusedSubtractHalving(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vhsubq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VHSUB.U8 Qd, Qn, Qm

A64 : UHSUB Vd.16B, Vn.16B, Vm.16B

FusedSubtractHalving(Vector128<Int16>, Vector128<Int16>)

int16x8_t vhsubq_s16 (int16x8_t a, int16x8_t b)

A32 : VHSUB.S16 Qd, Qn, Qm

A64 : SHSUB Vd.8H, Vn.8H, Vm.8H

FusedSubtractHalving(Vector128<Int32>, Vector128<Int32>)

int32x4_t vhsubq_s32 (int32x4_t a, int32x4_t b)

A32 : VHSUB.S32 Qd, Qn, Qm

A64 : SHSUB Vd.4S, Vn.4S, Vm.4S

FusedSubtractHalving(Vector128<SByte>, Vector128<SByte>)

int8x16_t vhsubq_s8 (int8x16_t a, int8x16_t b)

A32 : VHSUB.S8 Qd, Qn, Qm

A64 : SHSUB Vd.16B, Vn.16B, Vm.16B

FusedSubtractHalving(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vhsubq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VHSUB.U16 Qd, Qn, Qm

A64 : UHSUB Vd.8H, Vn.8H, Vm.8H

FusedSubtractHalving(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vhsubq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VHSUB.U32 Qd, Qn, Qm

A64 : UHSUB Vd.4S, Vn.4S, Vm.4S

FusedSubtractHalving(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vhsub_u8 (uint8x8_t a, uint8x8_t b)

A32 : VHSUB.U8 Dd, Dn, Dm

A64 : UHSUB Vd.8B, Vn.8B, Vm.8B

FusedSubtractHalving(Vector64<Int16>, Vector64<Int16>)

int16x4_t vhsub_s16 (int16x4_t a, int16x4_t b)

A32 : VHSUB.S16 Dd, Dn, Dm

A64 : SHSUB Vd.4H, Vn.4H, Vm.4H

FusedSubtractHalving(Vector64<Int32>, Vector64<Int32>)

int32x2_t vhsub_s32 (int32x2_t a, int32x2_t b)

A32 : VHSUB.S32 Dd, Dn, Dm

A64 : SHSUB Vd.2S, Vn.2S, Vm.2S

FusedSubtractHalving(Vector64<SByte>, Vector64<SByte>)

int8x8_t vhsub_s8 (int8x8_t a, int8x8_t b)

A32 : VHSUB.S8 Dd, Dn, Dm

A64 : SHSUB Vd.8B, Vn.8B, Vm.8B

FusedSubtractHalving(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vhsub_u16 (uint16x4_t a, uint16x4_t b)

A32 : VHSUB.U16 Dd, Dn, Dm

A64 : UHSUB Vd.4H, Vn.4H, Vm.4H

FusedSubtractHalving(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vhsub_u32 (uint32x2_t a, uint32x2_t b)

A32 : VHSUB.U32 Dd, Dn, Dm

A64 : UHSUB Vd.2S, Vn.2S, Vm.2S

GetHashCode()

Fait office de fonction de hachage par défaut.

(Hérité de Object)
GetType()

Obtient le Type de l'instance actuelle.

(Hérité de Object)
Insert(Vector128<Byte>, Byte, Byte)

uint8x16_t vsetq_lane_u8 (uint8_t a, uint8x16_t v, const int lane)

A32 : VMOV.8 Dd[lane], Rt

A64 : INS Vd.B[lane], Wn

Insert(Vector128<Double>, Byte, Double)

float64x2_t vsetq_lane_f64 (float64_t a, float64x2_t v, const int lane)

A32 : VMOV.F64 Dd, Dm

A64 : INS Vd.D[lane], Vn.D[0]

Insert(Vector128<Int16>, Byte, Int16)

int16x8_t vsetq_lane_s16 (int16_t a, int16x8_t v, const int lane)

A32 : VMOV.16 Dd[lane], Rt

A64 : INS Vd.H[lane], Wn

Insert(Vector128<Int32>, Byte, Int32)

int32x4_t vsetq_lane_s32 (int32_t a, int32x4_t v, const int lane)

A32 : VMOV.32 Dd[lane], Rt

A64 : INS Vd.S[lane], Wn

Insert(Vector128<Int64>, Byte, Int64)

int64x2_t vsetq_lane_s64 (int64_t a, int64x2_t v, const int lane)

A32 : VMOV.64 Dd, Rt, Rt2

A64 : INS Vd.D[lane], Xn

Insert(Vector128<SByte>, Byte, SByte)

int8x16_t vsetq_lane_s8 (int8_t a, int8x16_t v, const int lane)

A32 : VMOV.8 Dd[lane], Rt

A64 : INS Vd.B[lane], Wn

Insert(Vector128<Single>, Byte, Single)

float32x4_t vsetq_lane_f32 (float32_t a, float32x4_t v, const int lane)

A32 : VMOV.F32 Sd, Sm

A64 : INS Vd.S[lane], Vn.S[0]

Insert(Vector128<UInt16>, Byte, UInt16)

uint16x8_t vsetq_lane_u16 (uint16_t a, uint16x8_t v, const int lane)

A32 : VMOV.16 Dd[lane], Rt

A64 : INS Vd.H[lane], Wn

Insert(Vector128<UInt32>, Byte, UInt32)

uint32x4_t vsetq_lane_u32 (uint32_t a, uint32x4_t v, const int lane)

A32 : VMOV.32 Dd[lane], Rt

A64 : INS Vd.S[lane], Wn

Insert(Vector128<UInt64>, Byte, UInt64)

uint64x2_t vsetq_lane_u64 (uint64_t a, uint64x2_t v, const int lane)

A32 : VMOV.64 Dd, Rt, Rt2

A64 : INS Vd.D[lane], Xn

Insert(Vector64<Byte>, Byte, Byte)

uint8x8_t vset_lane_u8 (uint8_t a, uint8x8_t v, const int lane)

A32 : VMOV.8 Dd[lane], Rt

A64 : INS Vd.B[lane], Wn

Insert(Vector64<Int16>, Byte, Int16)

int16x4_t vset_lane_s16 (int16_t a, int16x4_t v, const int lane)

A32 : VMOV.16 Dd[lane], Rt

A64 : INS Vd.H[lane], Wn

Insert(Vector64<Int32>, Byte, Int32)

int32x2_t vset_lane_s32 (int32_t a, int32x2_t v, const int lane)

A32 : VMOV.32 Dd[lane], Rt

A64 : INS Vd.S[lane], Wn

Insert(Vector64<SByte>, Byte, SByte)

int8x8_t vset_lane_s8 (int8_t a, int8x8_t v, const int lane)

A32 : VMOV.8 Dd[lane], Rt

A64 : INS Vd.B[lane], Wn

Insert(Vector64<Single>, Byte, Single)

float32x2_t vset_lane_f32 (float32_t a, float32x2_t v, const int lane)

A32 : VMOV.F32 Sd, Sm

A64 : INS Vd.S[lane], Vn.S[0]

Insert(Vector64<UInt16>, Byte, UInt16)

uint16x4_t vset_lane_u16 (uint16_t a, uint16x4_t v, const int lane)

A32 : VMOV.16 Dd[lane], Rt

A64 : INS Vd.H[lane], Wn

Insert(Vector64<UInt32>, Byte, UInt32)

uint32x2_t vset_lane_u32 (uint32_t a, uint32x2_t v, const int lane)

A32 : VMOV.32 Dd[lane], Rt

A64 : INS Vd.S[lane], Wn

InsertScalar(Vector128<Double>, Byte, Vector64<Double>)

float64x2_t vcopyq_lane_f64 (float64x2_t a, const int lane1, float64x1_t b, const int lane2)

A32 : VMOV.F64 Dd, Dm

A64 : INS Vd.D[lane1], Vn.D[0]

InsertScalar(Vector128<Int64>, Byte, Vector64<Int64>)

int64x2_t vcopyq_lane_s64 (int64x2_t a, const int lane1, int64x1_t b, const int lane2)

A32 : VMOV Dd, Dm

A64 : INS Vd.D[lane1], Vn.D[0]

InsertScalar(Vector128<UInt64>, Byte, Vector64<UInt64>)

uint64x2_t vcopyq_lane_u64 (uint64x2_t a, const int lane1, uint64x1_t b, const int lane2)

A32 : VMOV Dd, Dm

A64 : INS Vd.D[lane1], Vn.D[0]

LeadingSignCount(Vector128<Int16>)

int16x8_t vclsq_s16 (int16x8_t a)

A32 : VCLS.S16 Qd, Qm

A64 : CLS Vd.8H, Vn.8H

LeadingSignCount(Vector128<Int32>)

int32x4_t vclsq_s32 (int32x4_t a)

A32 : VCLS.S32 Qd, Qm

A64 : CLS Vd.4S, Vn.4S

LeadingSignCount(Vector128<SByte>)

int8x16_t vclsq_s8 (int8x16_t a)

A32 : VCLS.S8 Qd, Qm

A64 : CLS Vd.16B, Vn.16B

LeadingSignCount(Vector64<Int16>)

int16x4_t vcls_s16 (int16x4_t a)

A32 : VCLS.S16 Dd, Dm

A64 : CLS Vd.4H, Vn.4H

LeadingSignCount(Vector64<Int32>)

int32x2_t vcls_s32 (int32x2_t a)

A32 : VCLS.S32 Dd, Dm

A64 : CLS Vd.2S, Vn.2S

LeadingSignCount(Vector64<SByte>)

int8x8_t vcls_s8 (int8x8_t a)

A32 : VCLS.S8 Dd, Dm

A64 : CLS Vd.8B, Vn.8B

LeadingZeroCount(Vector128<Byte>)

uint8x16_t vclzq_u8 (uint8x16_t a)

A32 : VCLZ.I8 Qd, Qm

A64 : CLZ Vd.16B, Vn.16B

LeadingZeroCount(Vector128<Int16>)

int16x8_t vclzq_s16 (int16x8_t a)

A32 : VCLZ.I16 Qd, Qm

A64 : CLZ Vd.8H, Vn.8H

LeadingZeroCount(Vector128<Int32>)

int32x4_t vclzq_s32 (int32x4_t a)

A32 : VCLZ.I32 Qd, Qm

A64 : CLZ Vd.4S, Vn.4S

LeadingZeroCount(Vector128<SByte>)

int8x16_t vclzq_s8 (int8x16_t a)

A32 : VCLZ.I8 Qd, Qm

A64 : CLZ Vd.16B, Vn.16B

LeadingZeroCount(Vector128<UInt16>)

uint16x8_t vclzq_u16 (uint16x8_t a)

A32 : VCLZ.I16 Qd, Qm

A64 : CLZ Vd.8H, Vn.8H

LeadingZeroCount(Vector128<UInt32>)

uint32x4_t vclzq_u32 (uint32x4_t a)

A32 : VCLZ.I32 Qd, Qm

A64 : CLZ Vd.4S, Vn.4S

LeadingZeroCount(Vector64<Byte>)

uint8x8_t vclz_u8 (uint8x8_t a)

A32 : VCLZ.I8 Dd, Dm

A64 : CLZ Vd.8B, Vn.8B

LeadingZeroCount(Vector64<Int16>)

int16x4_t vclz_s16 (int16x4_t a)

A32 : VCLZ.I16 Dd, Dm

A64 : CLZ Vd.4H, Vn.4H

LeadingZeroCount(Vector64<Int32>)

int32x2_t vclz_s32 (int32x2_t a)

A32 : VCLZ.I32 Dd, Dm

A64 : CLZ Vd.2S, Vn.2S

LeadingZeroCount(Vector64<SByte>)

int8x8_t vclz_s8 (int8x8_t a)

A32 : VCLZ.I8 Dd, Dm

A64 : CLZ Vd.8B, Vn.8B

LeadingZeroCount(Vector64<UInt16>)

uint16x4_t vclz_u16 (uint16x4_t a)

A32 : VCLZ.I16 Dd, Dm

A64 : CLZ Vd.4H, Vn.4H

LeadingZeroCount(Vector64<UInt32>)

uint32x2_t vclz_u32 (uint32x2_t a)

A32 : VCLZ.I32 Dd, Dm

A64 : CLZ Vd.2S, Vn.2S

LoadAndInsertScalar(ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>, Byte, Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>, Byte, Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Byte>,Vector64<Byte>>, Byte, Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>, Byte, Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>, Byte, Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Int16>,Vector64<Int16>>, Byte, Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>, Byte, Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>, Byte, Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Int32>,Vector64<Int32>>, Byte, Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>, Byte, SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>, Byte, SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<SByte>,Vector64<SByte>>, Byte, SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>,Vector64<Single>>, Byte, Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>>, Byte, Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<Single>,Vector64<Single>>, Byte, Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>, Byte, UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>, Byte, UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<UInt16>,Vector64<UInt16>>, Byte, UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>, Byte, UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>, Byte, UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(ValueTuple<Vector64<UInt32>,Vector64<UInt32>>, Byte, UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndInsertScalar(Vector128<Byte>, Byte, Byte*)

uint8x16_t vld1q_lane_u8 (uint8_t const * ptr, uint8x16_t src, const int lane)

A32 : VLD1.8 { Dd[index] }, [Rn]

A64 : LD1 { Vt.B }[index], [Xn]

LoadAndInsertScalar(Vector128<Double>, Byte, Double*)

float64x2_t vld1q_lane_f64 (float64_t const * ptr, float64x2_t src, const int lane)

A32 : VLDR.64 Dd, [Rn]

A64 : LD1 { Vt.D }[index], [Xn]

LoadAndInsertScalar(Vector128<Int16>, Byte, Int16*)

int16x8_t vld1q_lane_s16 (int16_t const * ptr, int16x8_t src, const int lane)

A32 : VLD1.16 { Dd[index] }, [Rn]

A64 : LD1 { Vt.H }[index], [Xn]

LoadAndInsertScalar(Vector128<Int32>, Byte, Int32*)

int32x4_t vld1q_lane_s32 (int32_t const * ptr, int32x4_t src, const int lane)

A32 : VLD1.32 { Dd[index] }, [Rn]

A64 : LD1 { Vt.S }[index], [Xn]

LoadAndInsertScalar(Vector128<Int64>, Byte, Int64*)

int64x2_t vld1q_lane_s64 (int64_t const * ptr, int64x2_t src, const int lane)

A32 : VLDR.64 Dd, [Rn]

A64 : LD1 { Vt.D }[index], [Xn]

LoadAndInsertScalar(Vector128<SByte>, Byte, SByte*)

int8x16_t vld1q_lane_s8 (int8_t const * ptr, int8x16_t src, const int lane)

A32 : VLD1.8 { Dd[index] }, [Rn]

A64 : LD1 { Vt.B }[index], [Xn]

LoadAndInsertScalar(Vector128<Single>, Byte, Single*)

float32x4_t vld1q_lane_f32 (float32_t const * ptr, float32x4_t src, const int lane)

A32 : VLD1.32 { Dd[index] }, [Rn]

A64 : LD1 { Vt.S }[index], [Xn]

LoadAndInsertScalar(Vector128<UInt16>, Byte, UInt16*)

uint16x8_t vld1q_lane_u16 (uint16_t const * ptr, uint16x8_t src, const int lane)

A32 : VLD1.16 { Dd[index] }, [Rn]

A64 : LD1 { Vt.H }[index], [Xn]

LoadAndInsertScalar(Vector128<UInt32>, Byte, UInt32*)

uint32x4_t vld1q_lane_u32 (uint32_t const * ptr, uint32x4_t src, const int lane)

A32 : VLD1.32 { Dd[index] }, [Rn]

A64 : LD1 { Vt.S }[index], [Xn]

LoadAndInsertScalar(Vector128<UInt64>, Byte, UInt64*)

uint64x2_t vld1q_lane_u64 (uint64_t const * ptr, uint64x2_t src, const int lane)

A32 : VLDR.64 Dd, [Rn]

A64 : LD1 { Vt.D }[index], [Xn]

LoadAndInsertScalar(Vector64<Byte>, Byte, Byte*)

uint8x8_t vld1_lane_u8 (uint8_t const * ptr, uint8x8_t src, const int lane)

A32 : VLD1.8 { Dd[index] }, [Rn]

A64 : LD1 { Vt.B }[index], [Xn]

LoadAndInsertScalar(Vector64<Int16>, Byte, Int16*)

int16x4_t vld1_lane_s16 (int16_t const * ptr, int16x4_t src, const int lane)

A32 : VLD1.16 { Dd[index] }, [Rn]

A64 : LD1 { Vt.H }[index], [Xn]

LoadAndInsertScalar(Vector64<Int32>, Byte, Int32*)

int32x2_t vld1_lane_s32 (int32_t const * ptr, int32x2_t src, const int lane)

A32 : VLD1.32 { Dd[index] }, [Rn]

A64 : LD1 { Vt.S }[index], [Xn]

LoadAndInsertScalar(Vector64<SByte>, Byte, SByte*)

int8x8_t vld1_lane_s8 (int8_t const * ptr, int8x8_t src, const int lane)

A32 : VLD1.8 { Dd[index] }, [Rn]

A64 : LD1 { Vt.B }[index], [Xn]

LoadAndInsertScalar(Vector64<Single>, Byte, Single*)

float32x2_t vld1_lane_f32 (float32_t const * ptr, float32x2_t src, const int lane)

A32 : VLD1.32 { Dd[index] }, [Rn]

A64 : LD1 { Vt.S }[index], [Xn]

LoadAndInsertScalar(Vector64<UInt16>, Byte, UInt16*)

uint16x4_t vld1_lane_u16 (uint16_t const * ptr, uint16x4_t src, const int lane)

A32 : VLD1.16 { Dd[index] }, [Rn]

A64 : LD1 { Vt.H }[index], [Xn]

LoadAndInsertScalar(Vector64<UInt32>, Byte, UInt32*)

uint32x2_t vld1_lane_u32 (uint32_t const * ptr, uint32x2_t src, const int lane)

A32 : VLD1.32 { Dd[index] }, [Rn]

A64 : LD1 { Vt.S }[index], [Xn]

LoadAndReplicateToVector128(Byte*)

uint8x16_t vld1q_dup_u8 (uint8_t const * ptr)

A32 : VLD1.8 { Dd[], Dd+1[] }, [Rn]

A64 : LD1R { Vt.16B }, [Xn]

LoadAndReplicateToVector128(Int16*)

int16x8_t vld1q_dup_s16 (int16_t const * ptr)

A32 : VLD1.16 { Dd[], Dd+1[] }, [Rn]

A64 : LD1R { Vt.8H }, [Xn]

LoadAndReplicateToVector128(Int32*)

int32x4_t vld1q_dup_s32 (int32_t const * ptr)

A32 : VLD1.32 { Dd[], Dd+1[] }, [Rn]

A64 : LD1R { Vt.4S }, [Xn]

LoadAndReplicateToVector128(SByte*)

int8x16_t vld1q_dup_s8 (int8_t const * ptr)

A32 : VLD1.8 { Dd[], Dd+1[] }, [Rn]

A64 : LD1R { Vt.16B }, [Xn]

LoadAndReplicateToVector128(Single*)

float32x4_t vld1q_dup_f32 (float32_t const * ptr)

A32 : VLD1.32 { Dd[], Dd+1[] }, [Rn]

A64 : LD1R { Vt.4S }, [Xn]

LoadAndReplicateToVector128(UInt16*)

uint16x8_t vld1q_dup_u16 (uint16_t const * ptr)

A32 : VLD1.16 { Dd[], Dd+1[] }, [Rn]

A64 : LD1R { Vt.8H }, [Xn]

LoadAndReplicateToVector128(UInt32*)

uint32x4_t vld1q_dup_u32 (uint32_t const * ptr)

A32 : VLD1.32 { Dd[], Dd+1[] }, [Rn]

A64 : LD1R { Vt.4S }, [Xn]

LoadAndReplicateToVector64(Byte*)

uint8x8_t vld1_dup_u8 (uint8_t const * ptr)

A32 : VLD1.8 { Dd[] }, [Rn]

A64 : LD1R { Vt.8B }, [Xn]

LoadAndReplicateToVector64(Int16*)

int16x4_t vld1_dup_s16 (int16_t const * ptr)

A32 : VLD1.16 { Dd[] }, [Rn]

A64 : LD1R { Vt.4H }, [Xn]

LoadAndReplicateToVector64(Int32*)

int32x2_t vld1_dup_s32 (int32_t const * ptr)

A32 : VLD1.32 { Dd[] }, [Rn]

A64 : LD1R { Vt.2S }, [Xn]

LoadAndReplicateToVector64(SByte*)

int8x8_t vld1_dup_s8 (int8_t const * ptr)

A32 : VLD1.8 { Dd[] }, [Rn]

A64 : LD1R { Vt.8B }, [Xn]

LoadAndReplicateToVector64(Single*)

float32x2_t vld1_dup_f32 (float32_t const * ptr)

A32 : VLD1.32 { Dd[] }, [Rn]

A64 : LD1R { Vt.2S }, [Xn]

LoadAndReplicateToVector64(UInt16*)

uint16x4_t vld1_dup_u16 (uint16_t const * ptr)

A32 : VLD1.16 { Dd[] }, [Rn]

A64 : LD1R { Vt.4H }, [Xn]

LoadAndReplicateToVector64(UInt32*)

uint32x2_t vld1_dup_u32 (uint32_t const * ptr)

A32 : VLD1.32 { Dd[] }, [Rn]

A64 : LD1R { Vt.2S }, [Xn]

LoadAndReplicateToVector64x2(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x2(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x2(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x2(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x2(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x2(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x2(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x3(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x3(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x3(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x3(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x3(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x3(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x3(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x4(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x4(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x4(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x4(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x4(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x4(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadAndReplicateToVector64x4(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector128(Byte*)

uint8x16_t vld1q_u8 (uint8_t const * ptr)

A32 : VLD1.8 Dd, Dd+1, [Rn]

A64 : LD1 Vt.16B, [Xn]

LoadVector128(Double*)

float64x2_t vld1q_f64 (float64_t const * ptr)

A32 : VLD1.64 Dd, Dd+1, [Rn]

A64 : LD1 Vt.2D, [Xn]

LoadVector128(Int16*)

int16x8_t vld1q_s16 (int16_t const * ptr)

A32 : VLD1.16 Dd, Dd+1, [Rn]

A64 : LD1 Vt.8H, [Xn]

LoadVector128(Int32*)

int32x4_t vld1q_s32 (int32_t const * ptr)

A32 : VLD1.32 Dd, Dd+1, [Rn]

A64 : LD1 Vt.4S, [Xn]

LoadVector128(Int64*)

int64x2_t vld1q_s64 (int64_t const * ptr)

A32 : VLD1.64 Dd, Dd+1, [Rn]

A64 : LD1 Vt.2D, [Xn]

LoadVector128(SByte*)

int8x16_t vld1q_s8 (int8_t const * ptr)

A32 : VLD1.8 Dd, Dd+1, [Rn]

A64 : LD1 Vt.16B, [Xn]

LoadVector128(Single*)

float32x4_t vld1q_f32 (float32_t const * ptr)

A32 : VLD1.32 Dd, Dd+1, [Rn]

A64 : LD1 Vt.4S, [Xn]

LoadVector128(UInt16*)

uint16x8_t vld1q_s16 (uint16_t const * ptr)

A32 : VLD1.16 Dd, Dd+1, [Rn]

A64 : LD1 Vt.8H, [Xn]

LoadVector128(UInt32*)

uint32x4_t vld1q_s32 (uint32_t const * ptr)

A32 : VLD1.32 Dd, Dd+1, [Rn]

A64 : LD1 Vt.4S, [Xn]

LoadVector128(UInt64*)

uint64x2_t vld1q_u64 (uint64_t const * ptr)

A32 : VLD1.64 Dd, Dd+1, [Rn]

A64 : LD1 Vt.2D, [Xn]

LoadVector64(Byte*)

uint8x8_t vld1_u8 (uint8_t const * ptr)

A32 : VLD1.8 Dd, [Rn]

A64 : LD1 Vt.8B, [Xn]

LoadVector64(Double*)

float64x1_t vld1_f64 (float64_t const * ptr)

A32 : VLD1.64 Dd, [Rn]

A64 : LD1 Vt.1D, [Xn]

LoadVector64(Int16*)

int16x4_t vld1_s16 (int16_t const * ptr)

A32 : VLD1.16 Dd, [Rn]

A64 : LD1 Vt.4H, [Xn]

LoadVector64(Int32*)

int32x2_t vld1_s32 (int32_t const * ptr)

A32 : VLD1.32 Dd, [Rn]

A64 : LD1 Vt.2S, [Xn]

LoadVector64(Int64*)

int64x1_t vld1_s64 (int64_t const * ptr)

A32 : VLD1.64 Dd, [Rn]

A64 : LD1 Vt.1D, [Xn]

LoadVector64(SByte*)

int8x8_t vld1_s8 (int8_t const * ptr)

A32 : VLD1.8 Dd, [Rn]

A64 : LD1 Vt.8B, [Xn]

LoadVector64(Single*)

float32x2_t vld1_f32 (float32_t const * ptr)

A32 : VLD1.32 Dd, [Rn]

A64 : LD1 Vt.2S, [Xn]

LoadVector64(UInt16*)

uint16x4_t vld1_u16 (uint16_t const * ptr)

A32 : VLD1.16 Dd, [Rn]

A64 : LD1 Vt.4H, [Xn]

LoadVector64(UInt32*)

uint32x2_t vld1_u32 (uint32_t const * ptr)

A32 : VLD1.32 Dd, [Rn]

A64 : LD1 Vt.2S, [Xn]

LoadVector64(UInt64*)

uint64x1_t vld1_u64 (uint64_t const * ptr)

A32 : VLD1.64 Dd, [Rn]

A64 : LD1 Vt.1D, [Xn]

LoadVector64x2(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2AndUnzip(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2AndUnzip(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2AndUnzip(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2AndUnzip(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2AndUnzip(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2AndUnzip(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x2AndUnzip(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3AndUnzip(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3AndUnzip(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3AndUnzip(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3AndUnzip(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3AndUnzip(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3AndUnzip(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x3AndUnzip(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4AndUnzip(Byte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4AndUnzip(Int16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4AndUnzip(Int32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4AndUnzip(SByte*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4AndUnzip(Single*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4AndUnzip(UInt16*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

LoadVector64x4AndUnzip(UInt32*)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

Max(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vmaxq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VMAX.U8 Qd, Qn, Qm

A64 : UMAX Vd.16B, Vn.16B, Vm.16B

Max(Vector128<Int16>, Vector128<Int16>)

int16x8_t vmaxq_s16 (int16x8_t a, int16x8_t b)

A32 : VMAX.S16 Qd, Qn, Qm

A64 : SMAX Vd.8H, Vn.8H, Vm.8H

Max(Vector128<Int32>, Vector128<Int32>)

int32x4_t vmaxq_s32 (int32x4_t a, int32x4_t b)

A32 : VMAX.S32 Qd, Qn, Qm

A64 : SMAX Vd.4S, Vn.4S, Vm.4S

Max(Vector128<SByte>, Vector128<SByte>)

int8x16_t vmaxq_s8 (int8x16_t a, int8x16_t b)

A32 : VMAX.S8 Qd, Qn, Qm

A64 : SMAX Vd.16B, Vn.16B, Vm.16B

Max(Vector128<Single>, Vector128<Single>)

float32x4_t vmaxq_f32 (float32x4_t a, float32x4_t b)

A32 : VMAX.F32 Qd, Qn, Qm

A64 : FMAX Vd.4S, Vn.4S, Vm.4S

Max(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vmaxq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VMAX.U16 Qd, Qn, Qm

A64 : UMAX Vd.8H, Vn.8H, Vm.8H

Max(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vmaxq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VMAX.U32 Qd, Qn, Qm

A64 : UMAX Vd.4S, Vn.4S, Vm.4S

Max(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vmax_u8 (uint8x8_t a, uint8x8_t b)

A32 : VMAX.U8 Dd, Dn, Dm

A64 : UMAX Vd.8B, Vn.8B, Vm.8B

Max(Vector64<Int16>, Vector64<Int16>)

int16x4_t vmax_s16 (int16x4_t a, int16x4_t b)

A32 : VMAX.S16 Dd, Dn, Dm

A64 : SMAX Vd.4H, Vn.4H, Vm.4H

Max(Vector64<Int32>, Vector64<Int32>)

int32x2_t vmax_s32 (int32x2_t a, int32x2_t b)

A32 : VMAX.S32 Dd, Dn, Dm

A64 : SMAX Vd.2S, Vn.2S, Vm.2S

Max(Vector64<SByte>, Vector64<SByte>)

int8x8_t vmax_s8 (int8x8_t a, int8x8_t b)

A32 : VMAX.S8 Dd, Dn, Dm

A64 : SMAX Vd.8B, Vn.8B, Vm.8B

Max(Vector64<Single>, Vector64<Single>)

float32x2_t vmax_f32 (float32x2_t a, float32x2_t b)

A32 : VMAX.F32 Dd, Dn, Dm

A64 : FMAX Vd.2S, Vn.2S, Vm.2S

Max(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmax_u16 (uint16x4_t a, uint16x4_t b)

A32 : VMAX.U16 Dd, Dn, Dm

A64 : UMAX Vd.4H, Vn.4H, Vm.4H

Max(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmax_u32 (uint32x2_t a, uint32x2_t b)

A32 : VMAX.U32 Dd, Dn, Dm

A64 : UMAX Vd.2S, Vn.2S, Vm.2S

MaxNumber(Vector128<Single>, Vector128<Single>)

float32x4_t vmaxnmq_f32 (float32x4_t a, float32x4_t b)

A32 : VMAXNM.F32 Qd, Qn, Qm

A64 : FMAXNM Vd.4S, Vn.4S, Vm.4S

MaxNumber(Vector64<Single>, Vector64<Single>)

float32x2_t vmaxnm_f32 (float32x2_t a, float32x2_t b)

A32 : VMAXNM.F32 Dd, Dn, Dm

A64 : FMAXNM Vd.2S, Vn.2S, Vm.2S

MaxNumberScalar(Vector64<Double>, Vector64<Double>)

float64x1_t vmaxnm_f64 (float64x1_t a, float64x1_t b)

A32 : VMAXNM.F64 Dd, Dn, Dm

A64 : FMAXNM Dd, Dn, Dm

MaxNumberScalar(Vector64<Single>, Vector64<Single>)

float32_t vmaxnms_f32 (float32_t a, float32_t b)

A32 : VMAXNM.F32 Sd, Sn, Sm

A64 : FMAXNM Sd, Sn, Sm

MaxPairwise(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vpmax_u8 (uint8x8_t a, uint8x8_t b)

A32 : VPMAX.U8 Dd, Dn, Dm

A64 : UMAXP Vd.8B, Vn.8B, Vm.8B

MaxPairwise(Vector64<Int16>, Vector64<Int16>)

int16x4_t vpmax_s16 (int16x4_t a, int16x4_t b)

A32 : VPMAX.S16 Dd, Dn, Dm

A64 : SMAXP Vd.4H, Vn.4H, Vm.4H

MaxPairwise(Vector64<Int32>, Vector64<Int32>)

int32x2_t vpmax_s32 (int32x2_t a, int32x2_t b)

A32 : VPMAX.S32 Dd, Dn, Dm

A64 : SMAXP Vd.2S, Vn.2S, Vm.2S

MaxPairwise(Vector64<SByte>, Vector64<SByte>)

int8x8_t vpmax_s8 (int8x8_t a, int8x8_t b)

A32 : VPMAX.S8 Dd, Dn, Dm

A64 : SMAXP Vd.8B, Vn.8B, Vm.8B

MaxPairwise(Vector64<Single>, Vector64<Single>)

float32x2_t vpmax_f32 (float32x2_t a, float32x2_t b)

A32 : VPMAX.F32 Dd, Dn, Dm

A64 : FMAXP Vd.2S, Vn.2S, Vm.2S

MaxPairwise(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vpmax_u16 (uint16x4_t a, uint16x4_t b)

A32 : VPMAX.U16 Dd, Dn, Dm

A64 : UMAXP Vd.4H, Vn.4H, Vm.4H

MaxPairwise(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vpmax_u32 (uint32x2_t a, uint32x2_t b)

A32 : VPMAX.U32 Dd, Dn, Dm

A64 : UMAXP Vd.2S, Vn.2S, Vm.2S

MemberwiseClone()

Crée une copie superficielle du Object actuel.

(Hérité de Object)
Min(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vminq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VMIN.U8 Qd, Qn, Qm

A64 : UMIN Vd.16B, Vn.16B, Vm.16B

Min(Vector128<Int16>, Vector128<Int16>)

int16x8_t vminq_s16 (int16x8_t a, int16x8_t b)

A32 : VMIN.S16 Qd, Qn, Qm

A64 : SMIN Vd.8H, Vn.8H, Vm.8H

Min(Vector128<Int32>, Vector128<Int32>)

int32x4_t vminq_s32 (int32x4_t a, int32x4_t b)

A32 : VMIN.S32 Qd, Qn, Qm

A64 : SMIN Vd.4S, Vn.4S, Vm.4S

Min(Vector128<SByte>, Vector128<SByte>)

int8x16_t vminq_s8 (int8x16_t a, int8x16_t b)

A32 : VMIN.S8 Qd, Qn, Qm

A64 : SMIN Vd.16B, Vn.16B, Vm.16B

Min(Vector128<Single>, Vector128<Single>)

float32x4_t vminq_f32 (float32x4_t a, float32x4_t b)

A32 : VMIN.F32 Qd, Qn, Qm

A64 : FMIN Vd.4S, Vn.4S, Vm.4S

Min(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vminq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VMIN.U16 Qd, Qn, Qm

A64 : UMIN Vd.8H, Vn.8H, Vm.8H

Min(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vminq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VMIN.U32 Qd, Qn, Qm

A64 : UMIN Vd.4S, Vn.4S, Vm.4S

Min(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vmin_u8 (uint8x8_t a, uint8x8_t b)

A32 : VMIN.U8 Dd, Dn, Dm

A64 : UMIN Vd.8B, Vn.8B, Vm.8B

Min(Vector64<Int16>, Vector64<Int16>)

int16x4_t vmin_s16 (int16x4_t a, int16x4_t b)

A32 : VMIN.S16 Dd, Dn, Dm

A64 : SMIN Vd.4H, Vn.4H, Vm.4H

Min(Vector64<Int32>, Vector64<Int32>)

int32x2_t vmin_s32 (int32x2_t a, int32x2_t b)

A32 : VMIN.S32 Dd, Dn, Dm

A64 : SMIN Vd.2S, Vn.2S, Vm.2S

Min(Vector64<SByte>, Vector64<SByte>)

int8x8_t vmin_s8 (int8x8_t a, int8x8_t b)

A32 : VMIN.S8 Dd, Dn, Dm

A64 : SMIN Vd.8B, Vn.8B, Vm.8B

Min(Vector64<Single>, Vector64<Single>)

float32x2_t vmin_f32 (float32x2_t a, float32x2_t b)

A32 : VMIN.F32 Dd, Dn, Dm

A64 : FMIN Vd.2S, Vn.2S, Vm.2S

Min(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmin_u16 (uint16x4_t a, uint16x4_t b)

A32 : VMIN.U16 Dd, Dn, Dm

A64 : UMIN Vd.4H, Vn.4H, Vm.4H

Min(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmin_u32 (uint32x2_t a, uint32x2_t b)

A32 : VMIN.U32 Dd, Dn, Dm

A64 : UMIN Vd.2S, Vn.2S, Vm.2S

MinNumber(Vector128<Single>, Vector128<Single>)

float32x4_t vminnmq_f32 (float32x4_t a, float32x4_t b)

A32 : VMINNM.F32 Qd, Qn, Qm

A64 : FMINNM Vd.4S, Vn.4S, Vm.4S

MinNumber(Vector64<Single>, Vector64<Single>)

float32x2_t vminnm_f32 (float32x2_t a, float32x2_t b)

A32 : VMINNM.F32 Dd, Dn, Dm

A64 : FMINNM Vd.2S, Vn.2S, Vm.2S

MinNumberScalar(Vector64<Double>, Vector64<Double>)

float64x1_t vminnm_f64 (float64x1_t a, float64x1_t b)

A32 : VMINNM.F64 Dd, Dn, Dm

A64 : FMINNM Dd, Dn, Dm

MinNumberScalar(Vector64<Single>, Vector64<Single>)

float32_t vminnms_f32 (float32_t a, float32_t b)

A32 : VMINNM.F32 Sd, Sn, Sm

A64 : FMINNM Sd, Sn, Sm

MinPairwise(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vpmin_u8 (uint8x8_t a, uint8x8_t b)

A32 : VPMIN.U8 Dd, Dn, Dm

A64 : UMINP Vd.8B, Vn.8B, Vm.8B

MinPairwise(Vector64<Int16>, Vector64<Int16>)

int16x4_t vpmin_s16 (int16x4_t a, int16x4_t b)

A32 : VPMIN.S16 Dd, Dn, Dm

A64 : SMINP Vd.4H, Vn.4H, Vm.4H

MinPairwise(Vector64<Int32>, Vector64<Int32>)

int32x2_t vpmin_s32 (int32x2_t a, int32x2_t b)

A32 : VPMIN.S32 Dd, Dn, Dm

A64 : SMINP Vd.2S, Vn.2S, Vm.2S

MinPairwise(Vector64<SByte>, Vector64<SByte>)

int8x8_t vpmin_s8 (int8x8_t a, int8x8_t b)

A32 : VPMIN.S8 Dd, Dn, Dm

A64 : SMINP Vd.8B, Vn.8B, Vm.8B

MinPairwise(Vector64<Single>, Vector64<Single>)

float32x2_t vpmin_f32 (float32x2_t a, float32x2_t b)

A32 : VPMIN.F32 Dd, Dn, Dm

A64 : FMINP Vd.2S, Vn.2S, Vm.2S

MinPairwise(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vpmin_u16 (uint16x4_t a, uint16x4_t b)

A32 : VPMIN.U16 Dd, Dn, Dm

A64 : UMINP Vd.4H, Vn.4H, Vm.4H

MinPairwise(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vpmin_u32 (uint32x2_t a, uint32x2_t b)

A32 : VPMIN.U32 Dd, Dn, Dm

A64 : UMINP Vd.2S, Vn.2S, Vm.2S

Multiply(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vmulq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VMUL.I8 Qd, Qn, Qm

A64 : MUL Vd.16B, Vn.16B, Vm.16B

Multiply(Vector128<Int16>, Vector128<Int16>)

int16x8_t vmulq_s16 (int16x8_t a, int16x8_t b)

A32 : VMUL.I16 Qd, Qn, Qm

A64 : MUL Vd.8H, Vn.8H, Vm.8H

Multiply(Vector128<Int32>, Vector128<Int32>)

int32x4_t vmulq_s32 (int32x4_t a, int32x4_t b)

A32 : VMUL.I32 Qd, Qn, Qm

A64 : MUL Vd.4S, Vn.4S, Vm.4S

Multiply(Vector128<SByte>, Vector128<SByte>)

int8x16_t vmulq_s8 (int8x16_t a, int8x16_t b)

A32 : VMUL.I8 Qd, Qn, Qm

A64 : MUL Vd.16B, Vn.16B, Vm.16B

Multiply(Vector128<Single>, Vector128<Single>)

float32x4_t vmulq_f32 (float32x4_t a, float32x4_t b)

A32 : VMUL.F32 Qd, Qn, Qm

A64 : FMUL Vd.4S, Vn.4S, Vm.4S

Multiply(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vmulq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VMUL.I16 Qd, Qn, Qm

A64 : MUL Vd.8H, Vn.8H, Vm.8H

Multiply(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vmulq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VMUL.I32 Qd, Qn, Qm

A64 : MUL Vd.4S, Vn.4S, Vm.4S

Multiply(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vmul_u8 (uint8x8_t a, uint8x8_t b)

A32 : VMUL.I8 Dd, Dn, Dm

A64 : MUL Vd.8B, Vn.8B, Vm.8B

Multiply(Vector64<Int16>, Vector64<Int16>)

int16x4_t vmul_s16 (int16x4_t a, int16x4_t b)

A32 : VMUL.I16 Dd, Dn, Dm

A64 : MUL Vd.4H, Vn.4H, Vm.4H

Multiply(Vector64<Int32>, Vector64<Int32>)

int32x2_t vmul_s32 (int32x2_t a, int32x2_t b)

A32 : VMUL.I32 Dd, Dn, Dm

A64 : MUL Vd.2S, Vn.2S, Vm.2S

Multiply(Vector64<SByte>, Vector64<SByte>)

int8x8_t vmul_s8 (int8x8_t a, int8x8_t b)

A32 : VMUL.I8 Dd, Dn, Dm

A64 : MUL Vd.8B, Vn.8B, Vm.8B

Multiply(Vector64<Single>, Vector64<Single>)

float32x2_t vmul_f32 (float32x2_t a, float32x2_t b)

A32 : VMUL.F32 Dd, Dn, Dm

A64 : FMUL Vd.2S, Vn.2S, Vm.2S

Multiply(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmul_u16 (uint16x4_t a, uint16x4_t b)

A32 : VMUL.I16 Dd, Dn, Dm

A64 : MUL Vd.4H, Vn.4H, Vm.4H

Multiply(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmul_u32 (uint32x2_t a, uint32x2_t b)

A32 : VMUL.I32 Dd, Dn, Dm

A64 : MUL Vd.2S, Vn.2S, Vm.2S

MultiplyAdd(Vector128<Byte>, Vector128<Byte>, Vector128<Byte>)

uint8x16_t vmlaq_u8 (uint8x16_t a, uint8x16_t b, uint8x16_t c)

A32 : VMLA.I8 Qd, Qn, Qm

A64 : MLA Vd.16B, Vn.16B, Vm.16B

MultiplyAdd(Vector128<Int16>, Vector128<Int16>, Vector128<Int16>)

int16x8_t vmlaq_s16 (int16x8_t a, int16x8_t b, int16x8_t c)

A32 : VMLA.I16 Qd, Qn, Qm

A64 : MLA Vd.8H, Vn.8H, Vm.8H

MultiplyAdd(Vector128<Int32>, Vector128<Int32>, Vector128<Int32>)

int32x4_t vmlaq_s32 (int32x4_t a, int32x4_t b, int32x4_t c)

A32 : VMLA.I32 Qd, Qn, Qm

A64 : MLA Vd.4S, Vn.4S, Vm.4S

MultiplyAdd(Vector128<SByte>, Vector128<SByte>, Vector128<SByte>)

int8x16_t vmlaq_s8 (int8x16_t a, int8x16_t b, int8x16_t c)

A32 : VMLA.I8 Qd, Qn, Qm

A64 : MLA Vd.16B, Vn.16B, Vm.16B

MultiplyAdd(Vector128<UInt16>, Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vmlaq_u16 (uint16x8_t a, uint16x8_t b, uint16x8_t c)

A32 : VMLA.I16 Qd, Qn, Qm

A64 : MLA Vd.8H, Vn.8H, Vm.8H

MultiplyAdd(Vector128<UInt32>, Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vmlaq_u32 (uint32x4_t a, uint32x4_t b, uint32x4_t c)

A32 : VMLA.I32 Qd, Qn, Qm

A64 : MLA Vd.4S, Vn.4S, Vm.4S

MultiplyAdd(Vector64<Byte>, Vector64<Byte>, Vector64<Byte>)

uint8x8_t vmla_u8 (uint8x8_t a, uint8x8_t b, uint8x8_t c)

A32 : VMLA.I8 Dd, Dn, Dm

A64 : MLA Vd.8B, Vn.8B, Vm.8B

MultiplyAdd(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>)

int16x4_t vmla_s16 (int16x4_t a, int16x4_t b, int16x4_t c)

A32 : VMLA.I16 Dd, Dn, Dm

A64 : MLA Vd.4H, Vn.4H, Vm.4H

MultiplyAdd(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>)

int32x2_t vmla_s32 (int32x2_t a, int32x2_t b, int32x2_t c)

A32 : VMLA.I32 Dd, Dn, Dm

A64 : MLA Vd.2S, Vn.2S, Vm.2S

MultiplyAdd(Vector64<SByte>, Vector64<SByte>, Vector64<SByte>)

int8x8_t vmla_s8 (int8x8_t a, int8x8_t b, int8x8_t c)

A32 : VMLA.I8 Dd, Dn, Dm

A64 : MLA Vd.8B, Vn.8B, Vm.8B

MultiplyAdd(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmla_u16 (uint16x4_t a, uint16x4_t b, uint16x4_t c)

A32 : VMLA.I16 Dd, Dn, Dm

A64 : MLA Vd.4H, Vn.4H, Vm.4H

MultiplyAdd(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmla_u32 (uint32x2_t a, uint32x2_t b, uint32x2_t c)

A32 : VMLA.I32 Dd, Dn, Dm

A64 : MLA Vd.2S, Vn.2S, Vm.2S

MultiplyAddByScalar(Vector128<Int16>, Vector128<Int16>, Vector64<Int16>)

int16x8_t vmlaq_n_s16 (int16x8_t a, int16x8_t b, int16_t c)

A32 : VMLA.I16 Qd, Qn, Dm[0]

A64 : MLA Vd.8H, Vn.8H, Vm.H[0]

MultiplyAddByScalar(Vector128<Int32>, Vector128<Int32>, Vector64<Int32>)

int32x4_t vmlaq_n_s32 (int32x4_t a, int32x4_t b, int32_t c)

A32 : VMLA.I32 Qd, Qn, Dm[0]

A64 : MLA Vd.4S, Vn.4S, Vm.S[0]

MultiplyAddByScalar(Vector128<UInt16>, Vector128<UInt16>, Vector64<UInt16>)

uint16x8_t vmlaq_n_u16 (uint16x8_t a, uint16x8_t b, uint16_t c)

A32 : VMLA.I16 Qd, Qn, Dm[0]

A64 : MLA Vd.8H, Vn.8H, Vm.H[0]

MultiplyAddByScalar(Vector128<UInt32>, Vector128<UInt32>, Vector64<UInt32>)

uint32x4_t vmlaq_n_u32 (uint32x4_t a, uint32x4_t b, uint32_t c)

A32 : VMLA.I32 Qd, Qn, Dm[0]

A64 : MLA Vd.4S, Vn.4S, Vm.S[0]

MultiplyAddByScalar(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>)

int16x4_t vmla_n_s16 (int16x4_t a, int16x4_t b, int16_t c)

A32 : VMLA.I16 Dd, Dn, Dm[0]

A64 : MLA Vd.4H, Vn.4H, Vm.H[0]

MultiplyAddByScalar(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>)

int32x2_t vmla_n_s32 (int32x2_t a, int32x2_t b, int32_t c)

A32 : VMLA.I32 Dd, Dn, Dm[0]

A64 : MLA Vd.2S, Vn.2S, Vm.S[0]

MultiplyAddByScalar(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmla_n_u16 (uint16x4_t a, uint16x4_t b, uint16_t c)

A32 : VMLA.I16 Dd, Dn, Dm[0]

A64 : MLA Vd.4H, Vn.4H, Vm.H[0]

MultiplyAddByScalar(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmla_n_u32 (uint32x2_t a, uint32x2_t b, uint32_t c)

A32 : VMLA.I32 Dd, Dn, Dm[0]

A64 : MLA Vd.2S, Vn.2S, Vm.S[0]

MultiplyAddBySelectedScalar(Vector128<Int16>, Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vmlaq_laneq_s16 (int16x8_t a, int16x8_t b, int16x8_t v, const int lane)

A32 : VMLA.I16 Qd, Qn, Dm[lane]

A64 : MLA Vd.8H, Vn.8H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector128<Int16>, Vector128<Int16>, Vector64<Int16>, Byte)

int16x8_t vmlaq_lane_s16 (int16x8_t a, int16x8_t b, int16x4_t v, const int lane)

A32 : VMLA.I16 Qd, Qn, Dm[lane]

A64 : MLA Vd.8H, Vn.8H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector128<Int32>, Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vmlaq_laneq_s32 (int32x4_t a, int32x4_t b, int32x4_t v, const int lane)

A32 : VMLA.I32 Qd, Qn, Dm[lane]

A64 : MLA Vd.4S, Vn.4S, Vm.S[lane]

MultiplyAddBySelectedScalar(Vector128<Int32>, Vector128<Int32>, Vector64<Int32>, Byte)

int32x4_t vmlaq_lane_s32 (int32x4_t a, int32x4_t b, int32x2_t v, const int lane)

A32 : VMLA.I32 Qd, Qn, Dm[lane]

A64 : MLA Vd.4S, Vn.4S, Vm.S[lane]

MultiplyAddBySelectedScalar(Vector128<UInt16>, Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vmlaq_laneq_u16 (uint16x8_t a, uint16x8_t b, uint16x8_t v, const int lane)

A32 : VMLA.I16 Qd, Qn, Dm[lane]

A64 : MLA Vd.8H, Vn.8H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector128<UInt16>, Vector128<UInt16>, Vector64<UInt16>, Byte)

uint16x8_t vmlaq_lane_u16 (uint16x8_t a, uint16x8_t b, uint16x4_t v, const int lane)

A32 : VMLA.I16 Qd, Qn, Dm[lane]

A64 : MLA Vd.8H, Vn.8H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector128<UInt32>, Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vmlaq_laneq_u32 (uint32x4_t a, uint32x4_t b, uint32x4_t v, const int lane)

A32 : VMLA.I32 Qd, Qn, Dm[lane]

A64 : MLA Vd.4S, Vn.4S, Vm.S[lane]

MultiplyAddBySelectedScalar(Vector128<UInt32>, Vector128<UInt32>, Vector64<UInt32>, Byte)

uint32x4_t vmlaq_lane_u32 (uint32x4_t a, uint32x4_t b, uint32x2_t v, const int lane)

A32 : VMLA.I32 Qd, Qn, Dm[lane]

A64 : MLA Vd.4S, Vn.4S, Vm.S[lane]

MultiplyAddBySelectedScalar(Vector64<Int16>, Vector64<Int16>, Vector128<Int16>, Byte)

int16x4_t vmla_laneq_s16 (int16x4_t a, int16x4_t b, int16x8_t v, const int lane)

A32 : VMLA.I16 Dd, Dn, Dm[lane]

A64 : MLA Vd.4H, Vn.4H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vmla_lane_s16 (int16x4_t a, int16x4_t b, int16x4_t v, const int lane)

A32 : VMLA.I16 Dd, Dn, Dm[lane]

A64 : MLA Vd.4H, Vn.4H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector64<Int32>, Vector64<Int32>, Vector128<Int32>, Byte)

int32x2_t vmla_laneq_s32 (int32x2_t a, int32x2_t b, int32x4_t v, const int lane)

A32 : VMLA.I32 Dd, Dn, Dm[lane]

A64 : MLA Vd.2S, Vn.2S, Vm.S[lane]

MultiplyAddBySelectedScalar(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vmla_lane_s32 (int32x2_t a, int32x2_t b, int32x2_t v, const int lane)

A32 : VMLA.I32 Dd, Dn, Dm[lane]

A64 : MLA Vd.2S, Vn.2S, Vm.S[lane]

MultiplyAddBySelectedScalar(Vector64<UInt16>, Vector64<UInt16>, Vector128<UInt16>, Byte)

uint16x4_t vmla_laneq_u16 (uint16x4_t a, uint16x4_t b, uint16x8_t v, const int lane)

A32 : VMLA.I16 Dd, Dn, Dm[lane]

A64 : MLA Vd.4H, Vn.4H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vmla_lane_u16 (uint16x4_t a, uint16x4_t b, uint16x4_t v, const int lane)

A32 : VMLA.I16 Dd, Dn, Dm[lane]

A64 : MLA Vd.4H, Vn.4H, Vm.H[lane]

MultiplyAddBySelectedScalar(Vector64<UInt32>, Vector64<UInt32>, Vector128<UInt32>, Byte)

uint32x2_t vmla_laneq_u32 (uint32x2_t a, uint32x2_t b, uint32x4_t v, const int lane)

A32 : VMLA.I32 Dd, Dn, Dm[lane]

A64 : MLA Vd.2S, Vn.2S, Vm.S[lane]

MultiplyAddBySelectedScalar(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vmla_lane_u32 (uint32x2_t a, uint32x2_t b, uint32x2_t v, const int lane)

A32 : VMLA.I32 Dd, Dn, Dm[lane]

A64 : MLA Vd.2S, Vn.2S, Vm.S[lane]

MultiplyByScalar(Vector128<Int16>, Vector64<Int16>)

int16x8_t vmulq_n_s16 (int16x8_t a, int16_t b)

A32 : VMUL.I16 Qd, Qn, Dm[0]

A64 : MUL Vd.8H, Vn.8H, Vm.H[0]

MultiplyByScalar(Vector128<Int32>, Vector64<Int32>)

int32x4_t vmulq_n_s32 (int32x4_t a, int32_t b)

A32 : VMUL.I32 Qd, Qn, Dm[0]

A64 : MUL Vd.4S, Vn.4S, Vm.S[0]

MultiplyByScalar(Vector128<Single>, Vector64<Single>)

float32x4_t vmulq_n_f32 (float32x4_t a, float32_t b)

A32 : VMUL.F32 Qd, Qn, Dm[0]

A64 : FMUL Vd.4S, Vn.4S, Vm.S[0]

MultiplyByScalar(Vector128<UInt16>, Vector64<UInt16>)

uint16x8_t vmulq_n_u16 (uint16x8_t a, uint16_t b)

A32 : VMUL.I16 Qd, Qn, Dm[0]

A64 : MUL Vd.8H, Vn.8H, Vm.H[0]

MultiplyByScalar(Vector128<UInt32>, Vector64<UInt32>)

uint32x4_t vmulq_n_u32 (uint32x4_t a, uint32_t b)

A32 : VMUL.I32 Qd, Qn, Dm[0]

A64 : MUL Vd.4S, Vn.4S, Vm.S[0]

MultiplyByScalar(Vector64<Int16>, Vector64<Int16>)

int16x4_t vmul_n_s16 (int16x4_t a, int16_t b)

A32 : VMUL.I16 Dd, Dn, Dm[0]

A64 : MUL Vd.4H, Vn.4H, Vm.H[0]

MultiplyByScalar(Vector64<Int32>, Vector64<Int32>)

int32x2_t vmul_n_s32 (int32x2_t a, int32_t b)

A32 : VMUL.I32 Dd, Dn, Dm[0]

A64 : MUL Vd.2S, Vn.2S, Vm.S[0]

MultiplyByScalar(Vector64<Single>, Vector64<Single>)

float32x2_t vmul_n_f32 (float32x2_t a, float32_t b)

A32 : VMUL.F32 Dd, Dn, Dm[0]

A64 : FMUL Vd.2S, Vn.2S, Vm.S[0]

MultiplyByScalar(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmul_n_u16 (uint16x4_t a, uint16_t b)

A32 : VMUL.I16 Dd, Dn, Dm[0]

A64 : MUL Vd.4H, Vn.4H, Vm.H[0]

MultiplyByScalar(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmul_n_u32 (uint32x2_t a, uint32_t b)

A32 : VMUL.I32 Dd, Dn, Dm[0]

A64 : MUL Vd.2S, Vn.2S, Vm.S[0]

MultiplyBySelectedScalar(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vmulq_laneq_s16 (int16x8_t a, int16x8_t v, const int lane)

A32 : VMUL.I16 Qd, Qn, Dm[lane]

A64 : MUL Vd.8H, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalar(Vector128<Int16>, Vector64<Int16>, Byte)

int16x8_t vmulq_lane_s16 (int16x8_t a, int16x4_t v, const int lane)

A32 : VMUL.I16 Qd, Qn, Dm[lane]

A64 : MUL Vd.8H, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalar(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vmulq_laneq_s32 (int32x4_t a, int32x4_t v, const int lane)

A32 : VMUL.I32 Qd, Qn, Dm[lane]

A64 : MUL Vd.4S, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalar(Vector128<Int32>, Vector64<Int32>, Byte)

int32x4_t vmulq_lane_s32 (int32x4_t a, int32x2_t v, const int lane)

A32 : VMUL.I32 Qd, Qn, Dm[lane]

A64 : MUL Vd.4S, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalar(Vector128<Single>, Vector128<Single>, Byte)

float32x4_t vmulq_laneq_f32 (float32x4_t a, float32x4_t v, const int lane)

A32 : VMUL.F32 Qd, Qn, Dm[lane]

A64 : FMUL Vd.4S, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalar(Vector128<Single>, Vector64<Single>, Byte)

float32x4_t vmulq_lane_f32 (float32x4_t a, float32x2_t v, const int lane)

A32 : VMUL.F32 Qd, Qn, Dm[lane]

A64 : FMUL Vd.4S, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalar(Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vmulq_laneq_u16 (uint16x8_t a, uint16x8_t v, const int lane)

A32 : VMUL.I16 Qd, Qn, Dm[lane]

A64 : MUL Vd.8H, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalar(Vector128<UInt16>, Vector64<UInt16>, Byte)

uint16x8_t vmulq_lane_u16 (uint16x8_t a, uint16x4_t v, const int lane)

A32 : VMUL.I16 Qd, Qn, Dm[lane]

A64 : MUL Vd.8H, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalar(Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vmulq_laneq_u32 (uint32x4_t a, uint32x4_t v, const int lane)

A32 : VMUL.I32 Qd, Qn, Dm[lane]

A64 : MUL Vd.4S, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalar(Vector128<UInt32>, Vector64<UInt32>, Byte)

uint32x4_t vmulq_lane_u32 (uint32x4_t a, uint32x2_t v, const int lane)

A32 : VMUL.I32 Qd, Qn, Dm[lane]

A64 : MUL Vd.4S, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalar(Vector64<Int16>, Vector128<Int16>, Byte)

int16x4_t vmul_laneq_s16 (int16x4_t a, int16x8_t v, const int lane)

A32 : VMUL.I16 Dd, Dn, Dm[lane]

A64 : MUL Vd.4H, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalar(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vmul_lane_s16 (int16x4_t a, int16x4_t v, const int lane)

A32 : VMUL.I16 Dd, Dn, Dm[lane]

A64 : MUL Vd.4H, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalar(Vector64<Int32>, Vector128<Int32>, Byte)

int32x2_t vmul_laneq_s32 (int32x2_t a, int32x4_t v, const int lane)

A32 : VMUL.I32 Dd, Dn, Dm[lane]

A64 : MUL Vd.2S, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalar(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vmul_lane_s32 (int32x2_t a, int32x2_t v, const int lane)

A32 : VMUL.I32 Dd, Dn, Dm[lane]

A64 : MUL Vd.2S, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalar(Vector64<Single>, Vector128<Single>, Byte)

float32x2_t vmul_laneq_f32 (float32x2_t a, float32x4_t v, const int lane)

A32 : VMUL.F32 Dd, Dn, Dm[lane]

A64 : FMUL Vd.2S, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalar(Vector64<Single>, Vector64<Single>, Byte)

float32x2_t vmul_lane_f32 (float32x2_t a, float32x2_t v, const int lane)

A32 : VMUL.F32 Dd, Dn, Dm[lane]

A64 : FMUL Vd.2S, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalar(Vector64<UInt16>, Vector128<UInt16>, Byte)

uint16x4_t vmul_laneq_u16 (uint16x4_t a, uint16x8_t v, const int lane)

A32 : VMUL.I16 Dd, Dn, Dm[lane]

A64 : MUL Vd.4H, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalar(Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vmul_lane_u16 (uint16x4_t a, uint16x4_t v, const int lane)

A32 : VMUL.I16 Dd, Dn, Dm[lane]

A64 : MUL Vd.4H, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalar(Vector64<UInt32>, Vector128<UInt32>, Byte)

uint32x2_t vmul_laneq_u32 (uint32x2_t a, uint32x4_t v, const int lane)

A32 : VMUL.I32 Dd, Dn, Dm[lane]

A64 : MUL Vd.2S, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalar(Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vmul_lane_u32 (uint32x2_t a, uint32x2_t v, const int lane)

A32 : VMUL.I32 Dd, Dn, Dm[lane]

A64 : MUL Vd.2S, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLower(Vector64<Int16>, Vector128<Int16>, Byte)

int32x4_t vmull_laneq_s16 (int16x4_t a, int16x8_t v, const int lane)

A32 : VMULL.S16 Qd, Dn, Dm[lane]

A64 : SMULL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLower(Vector64<Int16>, Vector64<Int16>, Byte)

int32x4_t vmull_lane_s16 (int16x4_t a, int16x4_t v, const int lane)

A32 : VMULL.S16 Qd, Dn, Dm[lane]

A64 : SMULL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLower(Vector64<Int32>, Vector128<Int32>, Byte)

int64x2_t vmull_laneq_s32 (int32x2_t a, int32x4_t v, const int lane)

A32 : VMULL.S32 Qd, Dn, Dm[lane]

A64 : SMULL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLower(Vector64<Int32>, Vector64<Int32>, Byte)

int64x2_t vmull_lane_s32 (int32x2_t a, int32x2_t v, const int lane)

A32 : VMULL.S32 Qd, Dn, Dm[lane]

A64 : SMULL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLower(Vector64<UInt16>, Vector128<UInt16>, Byte)

uint32x4_t vmull_laneq_u16 (uint16x4_t a, uint16x8_t v, const int lane)

A32 : VMULL.U16 Qd, Dn, Dm[lane]

A64 : UMULL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLower(Vector64<UInt16>, Vector64<UInt16>, Byte)

uint32x4_t vmull_lane_u16 (uint16x4_t a, uint16x4_t v, const int lane)

A32 : VMULL.U16 Qd, Dn, Dm[lane]

A64 : UMULL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLower(Vector64<UInt32>, Vector128<UInt32>, Byte)

uint64x2_t vmull_laneq_u32 (uint32x2_t a, uint32x4_t v, const int lane)

A32 : VMULL.U32 Qd, Dn, Dm[lane]

A64 : UMULL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLower(Vector64<UInt32>, Vector64<UInt32>, Byte)

uint64x2_t vmull_lane_u32 (uint32x2_t a, uint32x2_t v, const int lane)

A32 : VMULL.U32 Qd, Dn, Dm[lane]

A64 : UMULL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<Int32>, Vector64<Int16>, Vector128<Int16>, Byte)

int32x4_t vmlal_laneq_s16 (int32x4_t a, int16x4_t b, int16x8_t v, const int lane)

A32 : VMLAL.S16 Qd, Dn, Dm[lane]

A64 : SMLAL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>, Byte)

int32x4_t vmlal_lane_s16 (int32x4_t a, int16x4_t b, int16x4_t v, const int lane)

A32 : VMLAL.S16 Qd, Dn, Dm[lane]

A64 : SMLAL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<Int64>, Vector64<Int32>, Vector128<Int32>, Byte)

int64x2_t vmlal_laneq_s32 (int64x2_t a, int32x2_t b, int32x4_t v, const int lane)

A32 : VMLAL.S32 Qd, Dn, Dm[lane]

A64 : SMLAL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>, Byte)

int64x2_t vmlal_lane_s32 (int64x2_t a, int32x2_t b, int32x2_t v, const int lane)

A32 : VMLAL.S32 Qd, Dn, Dm[lane]

A64 : SMLAL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<UInt32>, Vector64<UInt16>, Vector128<UInt16>, Byte)

uint32x4_t vmlal_laneq_u16 (uint32x4_t a, uint16x4_t b, uint16x8_t v, const int lane)

A32 : VMLAL.U16 Qd, Dn, Dm[lane]

A64 : UMLAL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<UInt32>, Vector64<UInt16>, Vector64<UInt16>, Byte)

uint32x4_t vmlal_lane_u16 (uint32x4_t a, uint16x4_t b, uint16x4_t v, const int lane)

A32 : VMLAL.U16 Qd, Dn, Dm[lane]

A64 : UMLAL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<UInt64>, Vector64<UInt32>, Vector128<UInt32>, Byte)

uint64x2_t vmlal_laneq_u32 (uint64x2_t a, uint32x2_t b, uint32x4_t v, const int lane)

A32 : VMLAL.U32 Qd, Dn, Dm[lane]

A64 : UMLAL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndAdd(Vector128<UInt64>, Vector64<UInt32>, Vector64<UInt32>, Byte)

uint64x2_t vmlal_lane_u32 (uint64x2_t a, uint32x2_t b, uint32x2_t v, const int lane)

A32 : VMLAL.U32 Qd, Dn, Dm[lane]

A64 : UMLAL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<Int32>, Vector64<Int16>, Vector128<Int16>, Byte)

int32x4_t vmlsl_laneq_s16 (int32x4_t a, int16x4_t b, int16x8_t v, const int lane)

A32 : VMLSL.S16 Qd, Dn, Dm[lane]

A64 : SMLSL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>, Byte)

int32x4_t vmlsl_lane_s16 (int32x4_t a, int16x4_t b, int16x4_t v, const int lane)

A32 : VMLSL.S16 Qd, Dn, Dm[lane]

A64 : SMLSL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<Int64>, Vector64<Int32>, Vector128<Int32>, Byte)

int64x2_t vmlsl_laneq_s32 (int64x2_t a, int32x2_t b, int32x4_t v, const int lane)

A32 : VMLSL.S32 Qd, Dn, Dm[lane]

A64 : SMLSL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>, Byte)

int64x2_t vmlsl_lane_s32 (int64x2_t a, int32x2_t b, int32x2_t v, const int lane)

A32 : VMLSL.S32 Qd, Dn, Dm[lane]

A64 : SMLSL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<UInt32>, Vector64<UInt16>, Vector128<UInt16>, Byte)

uint32x4_t vmlsl_laneq_u16 (uint32x4_t a, uint16x4_t b, uint16x8_t v, const int lane)

A32 : VMLSL.U16 Qd, Dn, Dm[lane]

A64 : UMLSL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<UInt32>, Vector64<UInt16>, Vector64<UInt16>, Byte)

uint32x4_t vmlsl_lane_u16 (uint32x4_t a, uint16x4_t b, uint16x4_t v, const int lane)

A32 : VMLSL.U16 Qd, Dn, Dm[lane]

A64 : UMLSL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<UInt64>, Vector64<UInt32>, Vector128<UInt32>, Byte)

uint64x2_t vmlsl_laneq_u32 (uint64x2_t a, uint32x2_t b, uint32x4_t v, const int lane)

A32 : VMLSL.U32 Qd, Dn, Dm[lane]

A64 : UMLSL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningLowerAndSubtract(Vector128<UInt64>, Vector64<UInt32>, Vector64<UInt32>, Byte)

uint64x2_t vmlsl_lane_u32 (uint64x2_t a, uint32x2_t b, uint32x2_t v, const int lane)

A32 : VMLSL.U32 Qd, Dn, Dm[lane]

A64 : UMLSL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<Int16>, Vector128<Int16>, Byte)

int32x4_t vmull_high_laneq_s16 (int16x8_t a, int16x8_t v, const int lane)

A32 : VMULL.S16 Qd, Dn+1, Dm[lane]

A64 : SMULL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<Int16>, Vector64<Int16>, Byte)

int32x4_t vmull_high_lane_s16 (int16x8_t a, int16x4_t v, const int lane)

A32 : VMULL.S16 Qd, Dn+1, Dm[lane]

A64 : SMULL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<Int32>, Vector128<Int32>, Byte)

int64x2_t vmull_high_laneq_s32 (int32x4_t a, int32x4_t v, const int lane)

A32 : VMULL.S32 Qd, Dn+1, Dm[lane]

A64 : SMULL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<Int32>, Vector64<Int32>, Byte)

int64x2_t vmull_high_lane_s32 (int32x4_t a, int32x2_t v, const int lane)

A32 : VMULL.S32 Qd, Dn+1, Dm[lane]

A64 : SMULL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<UInt16>, Vector128<UInt16>, Byte)

uint32x4_t vmull_high_laneq_u16 (uint16x8_t a, uint16x8_t v, const int lane)

A32 : VMULL.U16 Qd, Dn+1, Dm[lane]

A64 : UMULL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<UInt16>, Vector64<UInt16>, Byte)

uint32x4_t vmull_high_lane_u16 (uint16x8_t a, uint16x4_t v, const int lane)

A32 : VMULL.U16 Qd, Dn+1, Dm[lane]

A64 : UMULL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<UInt32>, Vector128<UInt32>, Byte)

uint64x2_t vmull_high_laneq_u32 (uint32x4_t a, uint32x4_t v, const int lane)

A32 : VMULL.U32 Qd, Dn+1, Dm[lane]

A64 : UMULL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpper(Vector128<UInt32>, Vector64<UInt32>, Byte)

uint64x2_t vmull_high_lane_u32 (uint32x4_t a, uint32x2_t v, const int lane)

A32 : VMULL.U32 Qd, Dn+1, Dm[lane]

A64 : UMULL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>, Byte)

int32x4_t vmlal_high_laneq_s16 (int32x4_t a, int16x8_t b, int16x8_t v, const int lane)

A32 : VMLAL.S16 Qd, Dn+1, Dm[lane]

A64 : SMLAL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<Int32>, Vector128<Int16>, Vector64<Int16>, Byte)

int32x4_t vmlal_high_lane_s16 (int32x4_t a, int16x8_t b, int16x4_t v, const int lane)

A32 : VMLAL.S16 Qd, Dn+1, Dm[lane]

A64 : SMLAL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>, Byte)

int64x2_t vmlal_high_laneq_s32 (int64x2_t a, int32x4_t b, int32x4_t v, const int lane)

A32 : VMLAL.S32 Qd, Dn+1, Dm[lane]

A64 : SMLAL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<Int64>, Vector128<Int32>, Vector64<Int32>, Byte)

int64x2_t vmlal_high_lane_s32 (int64x2_t a, int32x4_t b, int32x2_t v, const int lane)

A32 : VMLAL.S32 Qd, Dn+1, Dm[lane]

A64 : SMLAL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<UInt32>, Vector128<UInt16>, Vector128<UInt16>, Byte)

uint32x4_t vmlal_high_laneq_u16 (uint32x4_t a, uint16x8_t b, uint16x8_t v, const int lane)

A32 : VMLAL.U16 Qd, Dn+1, Dm[lane]

A64 : UMLAL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<UInt32>, Vector128<UInt16>, Vector64<UInt16>, Byte)

uint32x4_t vmlal_high_lane_u16 (uint32x4_t a, uint16x8_t b, uint16x4_t v, const int lane)

A32 : VMLAL.U16 Qd, Dn+1, Dm[lane]

A64 : UMLAL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<UInt64>, Vector128<UInt32>, Vector128<UInt32>, Byte)

uint64x2_t vmlal_high_laneq_u32 (uint64x2_t a, uint32x4_t b, uint32x4_t v, const int lane)

A32 : VMLAL.U32 Qd, Dn+1, Dm[lane]

A64 : UMLAL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndAdd(Vector128<UInt64>, Vector128<UInt32>, Vector64<UInt32>, Byte)

uint64x2_t vmlal_high_lane_u32 (uint64x2_t a, uint32x4_t b, uint32x2_t v, const int lane)

A32 : VMLAL.U32 Qd, Dn+1, Dm[lane]

A64 : UMLAL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>, Byte)

int32x4_t vmlsl_high_laneq_s16 (int32x4_t a, int16x8_t b, int16x8_t v, const int lane)

A32 : VMLSL.S16 Qd, Dn+1, Dm[lane]

A64 : SMLSL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<Int32>, Vector128<Int16>, Vector64<Int16>, Byte)

int32x4_t vmlsl_high_lane_s16 (int32x4_t a, int16x8_t b, int16x4_t v, const int lane)

A32 : VMLSL.S16 Qd, Dn+1, Dm[lane]

A64 : SMLSL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>, Byte)

int64x2_t vmlsl_high_laneq_s32 (int64x2_t a, int32x4_t b, int32x4_t v, const int lane)

A32 : VMLSL.S32 Qd, Dn+1, Dm[lane]

A64 : SMLSL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<Int64>, Vector128<Int32>, Vector64<Int32>, Byte)

int64x2_t vmlsl_high_lane_s32 (int64x2_t a, int32x4_t b, int32x2_t v, const int lane)

A32 : VMLSL.S32 Qd, Dn+1, Dm[lane]

A64 : SMLSL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<UInt32>, Vector128<UInt16>, Vector128<UInt16>, Byte)

uint32x4_t vmlsl_high_laneq_u16 (uint32x4_t a, uint16x8_t b, uint16x8_t v, const int lane)

A32 : VMLSL.U16 Qd, Dn+1, Dm[lane]

A64 : UMLSL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<UInt32>, Vector128<UInt16>, Vector64<UInt16>, Byte)

uint32x4_t vmlsl_high_lane_u16 (uint32x4_t a, uint16x8_t b, uint16x4_t v, const int lane)

A32 : VMLSL.U16 Qd, Dn+1, Dm[lane]

A64 : UMLSL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<UInt64>, Vector128<UInt32>, Vector128<UInt32>, Byte)

uint64x2_t vmlsl_high_laneq_u32 (uint64x2_t a, uint32x4_t b, uint32x4_t v, const int lane)

A32 : VMLSL.U32 Qd, Dn+1, Dm[lane]

A64 : UMLSL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyBySelectedScalarWideningUpperAndSubtract(Vector128<UInt64>, Vector128<UInt32>, Vector64<UInt32>, Byte)

uint64x2_t vmlsl_high_lane_u32 (uint64x2_t a, uint32x4_t b, uint32x2_t v, const int lane)

A32 : VMLSL.U32 Qd, Dn+1, Dm[lane]

A64 : UMLSL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyDoublingByScalarSaturateHigh(Vector128<Int16>, Vector64<Int16>)

int16x8_t vqdmulhq_n_s16 (int16x8_t a, int16_t b)

A32 : VQDMULH. S16 Qd, Qn, Dm[0]

A64 : SQDMULH Vd.8H, Vn.8H, Vm.H[0]

MultiplyDoublingByScalarSaturateHigh(Vector128<Int32>, Vector64<Int32>)

int32x4_t vqdmulhq_n_s32 (int32x4_t a, int32_t b)

A32 : VQDMULH. S32 Qd, Qn, Dm[0]

A64 : SQDMULH Vd.4S, Vn.4S, Vm.S[0]

MultiplyDoublingByScalarSaturateHigh(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqdmulh_n_s16 (int16x4_t a, int16_t b)

A32 : VQDMULH. S16 Dd, Dn, Dm[0]

A64 : SQDMULH Vd.4H, Vn.4H, Vm.H[0]

MultiplyDoublingByScalarSaturateHigh(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqdmulh_n_s32 (int32x2_t a, int32_t b)

A32 : VQDMULH. S32 Dd, Dn, Dm[0]

A64 : SQDMULH Vd.2S, Vn.2S, Vm.S[0]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vqdmulhq_laneq_s16 (int16x8_t a, int16x8_t v, const int lane)

A32 : VQDMULH. S16 Qd, Qn, Dm[lane]

A64 : SQDMULH Vd.8H, Vn.8H, Vm.H[lane]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector128<Int16>, Vector64<Int16>, Byte)

int16x8_t vqdmulhq_lane_s16 (int16x8_t a, int16x4_t v, const int lane)

A32 : VQDMULH. S16 Qd, Qn, Dm[lane]

A64 : SQDMULH Vd.8H, Vn.8H, Vm.H[lane]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vqdmulhq_laneq_s32 (int32x4_t a, int32x4_t v, const int lane)

A32 : VQDMULH. S32 Qd, Qn, Dm[lane]

A64 : SQDMULH Vd.4S, Vn.4S, Vm.S[lane]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector128<Int32>, Vector64<Int32>, Byte)

int32x4_t vqdmulhq_lane_s32 (int32x4_t a, int32x2_t v, const int lane)

A32 : VQDMULH. S32 Qd, Qn, Dm[lane]

A64 : SQDMULH Vd.4S, Vn.4S, Vm.S[lane]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector64<Int16>, Vector128<Int16>, Byte)

int16x4_t vqdmulh_laneq_s16 (int16x4_t a, int16x8_t v, const int lane)

A32 : VQDMULH. S16 Dd, Dn, Dm[lane]

A64 : SQDMULH Vd.4H, Vn.4H, Vm.H[lane]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vqdmulh_lane_s16 (int16x4_t a, int16x4_t v, const int lane)

A32 : VQDMULH. S16 Dd, Dn, Dm[lane]

A64 : SQDMULH Vd.4H, Vn.4H, Vm.H[lane]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector64<Int32>, Vector128<Int32>, Byte)

int32x2_t vqdmulh_laneq_s32 (int32x2_t a, int32x4_t v, const int lane)

A32 : VQDMULH. S32 Dd, Dn, Dm[lane]

A64 : SQDMULH Vd.2S, Vn.2S, Vm.S[lane]

MultiplyDoublingBySelectedScalarSaturateHigh(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vqdmulh_lane_s32 (int32x2_t a, int32x2_t v, const int lane)

A32 : VQDMULH. S32 Dd, Dn, Dm[lane]

A64 : SQDMULH Vd.2S, Vn.2S, Vm.S[lane]

MultiplyDoublingSaturateHigh(Vector128<Int16>, Vector128<Int16>)

int16x8_t vqdmulhq_s16 (int16x8_t a, int16x8_t b)

A32 : VQDMULH. S16 Qd, Qn, Qm

A64 : SQDMULH Vd.8H, Vn.8H, Vm.8H

MultiplyDoublingSaturateHigh(Vector128<Int32>, Vector128<Int32>)

int32x4_t vqdmulhq_s32 (int32x4_t a, int32x4_t b)

A32 : VQDMULH. S32 Qd, Qn, Qm

A64 : SQDMULH Vd.4S, Vn.4S, Vm.4S

MultiplyDoublingSaturateHigh(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqdmulh_s16 (int16x4_t a, int16x4_t b)

A32 : VQDMULH. S16 Dd, Dn, Dm

A64 : SQDMULH Vd.4H, Vn.4H, Vm.4H

MultiplyDoublingSaturateHigh(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqdmulh_s32 (int32x2_t a, int32x2_t b)

A32 : VQDMULH. S32 Dd, Dn, Dm

A64 : SQDMULH Vd.2S, Vn.2S, Vm.2S

MultiplyDoublingWideningLowerAndAddSaturate(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>)

int32x4_t vqdmlal_s16 (int32x4_t a, int16x4_t b, int16x4_t c)

A32 : VQDMLAL. S16 Qd, Dn, Dm

A64 : SQDMLAL Vd.4S, Vn.4H, Vm.4H

MultiplyDoublingWideningLowerAndAddSaturate(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>)

int64x2_t vqdmlal_s32 (int64x2_t a, int32x2_t b, int32x2_t c)

A32 : VQDMLAL. S32 Qd, Dn, Dm

A64 : SQDMLAL Vd.2D, Vn.2S, Vm.2S

MultiplyDoublingWideningLowerAndSubtractSaturate(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>)

int32x4_t vqdmlsl_s16 (int32x4_t a, int16x4_t b, int16x4_t c)

A32 : VQDMLSL. S16 Qd, Dn, Dm

A64 : SQDMLSL Vd.4S, Vn.4H, Vm.4H

MultiplyDoublingWideningLowerAndSubtractSaturate(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>)

int64x2_t vqdmlsl_s32 (int64x2_t a, int32x2_t b, int32x2_t c)

A32 : VQDMLSL. S32 Qd, Dn, Dm

A64 : SQDMLSL Vd.2D, Vn.2S, Vm.2S

MultiplyDoublingWideningLowerByScalarAndAddSaturate(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>)

int32x4_t vqdmlal_n_s16 (int32x4_t a, int16x4_t b, int16_t c)

A32 : VQDMLAL. S16 Qd, Dn, Dm[0]

A64 : SQDMLAL Vd.4S, Vn.4H, Vm.H[0]

MultiplyDoublingWideningLowerByScalarAndAddSaturate(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>)

int64x2_t vqdmlal_n_s32 (int64x2_t a, int32x2_t b, int32_t c)

A32 : VQDMLAL. S32 Qd, Dn, Dm[0]

A64 : SQDMLAL Vd.2D, Vn.2S, Vm.S[0]

MultiplyDoublingWideningLowerByScalarAndSubtractSaturate(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>)

int32x4_t vqdmlsl_n_s16 (int32x4_t a, int16x4_t b, int16_t c)

A32 : VQDMLSL. S16 Qd, Dn, Dm[0]

A64 : SQDMLSL Vd.4S, Vn.4H, Vm.H[0]

MultiplyDoublingWideningLowerByScalarAndSubtractSaturate(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>)

int64x2_t vqdmlsl_n_s32 (int64x2_t a, int32x2_t b, int32_t c)

A32 : VQDMLSL. S32 Qd, Dn, Dm[0]

A64 : SQDMLSL Vd.2D, Vn.2S, Vm.S[0]

MultiplyDoublingWideningLowerBySelectedScalarAndAddSaturate(Vector128<Int32>, Vector64<Int16>, Vector128<Int16>, Byte)

int32x4_t vqdmlal_laneq_s16 (int32x4_t a, int16x4_t b, int16x8_t v, const int lane)

A32 : VQDMLAL. S16 Qd, Dn, Dm[lane]

A64 : SQDMLAL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyDoublingWideningLowerBySelectedScalarAndAddSaturate(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>, Byte)

int32x4_t vqdmlal_lane_s16 (int32x4_t a, int16x4_t b, int16x4_t v, const int lane)

A32 : VQDMLAL. S16 Qd, Dn, Dm[lane]

A64 : SQDMLAL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyDoublingWideningLowerBySelectedScalarAndAddSaturate(Vector128<Int64>, Vector64<Int32>, Vector128<Int32>, Byte)

int64x2_t vqdmlal_laneq_s32 (int64x2_t a, int32x2_t b, int32x4_t v, const int lane)

A32 : VQDMLAL. S32 Qd, Dn, Dm[lane]

A64 : SQDMLAL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyDoublingWideningLowerBySelectedScalarAndAddSaturate(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>, Byte)

int64x2_t vqdmlal_lane_s32 (int64x2_t a, int32x2_t b, int32x2_t v, const int lane)

A32 : VQDMLAL. S32 Qd, Dn, Dm[lane]

A64 : SQDMLAL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyDoublingWideningLowerBySelectedScalarAndSubtractSaturate(Vector128<Int32>, Vector64<Int16>, Vector128<Int16>, Byte)

int32x4_t vqdmlsl_laneq_s16 (int32x4_t a, int16x4_t b, int16x8_t v, const int lane)

A32 : VQDMLSL. S16 Qd, Dn, Dm[lane]

A64 : SQDMLSL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyDoublingWideningLowerBySelectedScalarAndSubtractSaturate(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>, Byte)

int32x4_t vqdmlsl_lane_s16 (int32x4_t a, int16x4_t b, int16x4_t v, const int lane)

A32 : VQDMLSL. S16 Qd, Dn, Dm[lane]

A64 : SQDMLSL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyDoublingWideningLowerBySelectedScalarAndSubtractSaturate(Vector128<Int64>, Vector64<Int32>, Vector128<Int32>, Byte)

int64x2_t vqdmlsl_laneq_s32 (int64x2_t a, int32x2_t b, int32x4_t v, const int lane)

A32 : VQDMLSL. S32 Qd, Dn, Dm[lane]

A64 : SQDMLSL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyDoublingWideningLowerBySelectedScalarAndSubtractSaturate(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>, Byte)

int64x2_t vqdmlsl_lane_s32 (int64x2_t a, int32x2_t b, int32x2_t v, const int lane)

A32 : VQDMLSL. S32 Qd, Dn, Dm[lane]

A64 : SQDMLSL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyDoublingWideningSaturateLower(Vector64<Int16>, Vector64<Int16>)

int32x4_t vqdmull_s16 (int16x4_t a, int16x4_t b)

A32 : VQDMULL. S16 Qd, Dn, Dm

A64 : SQDMULL Vd.4S, Vn.4H, Vm.4H

MultiplyDoublingWideningSaturateLower(Vector64<Int32>, Vector64<Int32>)

int64x2_t vqdmull_s32 (int32x2_t a, int32x2_t b)

A32 : VQDMULL. S32 Qd, Dn, Dm

A64 : SQDMULL Vd.2D, Vn.2S, Vm.2S

MultiplyDoublingWideningSaturateLowerByScalar(Vector64<Int16>, Vector64<Int16>)

int32x4_t vqdmull_n_s16 (int16x4_t a, int16_t b)

A32 : VQDMULL. S16 Qd, Dn, Dm[0]

A64 : SQDMULL Vd.4S, Vn.4H, Vm.H[0]

MultiplyDoublingWideningSaturateLowerByScalar(Vector64<Int32>, Vector64<Int32>)

int64x2_t vqdmull_n_s32 (int32x2_t a, int32_t b)

A32 : VQDMULL. S32 Qd, Dn, Dm[0]

A64 : SQDMULL Vd.2D, Vn.2S, Vm.S[0]

MultiplyDoublingWideningSaturateLowerBySelectedScalar(Vector64<Int16>, Vector128<Int16>, Byte)

int32x4_t vqdmull_laneq_s16 (int16x4_t a, int16x8_t v, const int lane)

A32 : VQDMULL. S16 Qd, Dn, Dm[lane]

A64 : SQDMULL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyDoublingWideningSaturateLowerBySelectedScalar(Vector64<Int16>, Vector64<Int16>, Byte)

int32x4_t vqdmull_lane_s16 (int16x4_t a, int16x4_t v, const int lane)

A32 : VQDMULL. S16 Qd, Dn, Dm[lane]

A64 : SQDMULL Vd.4S, Vn.4H, Vm.H[lane]

MultiplyDoublingWideningSaturateLowerBySelectedScalar(Vector64<Int32>, Vector128<Int32>, Byte)

int64x2_t vqdmull_laneq_s32 (int32x2_t a, int32x4_t v, const int lane)

A32 : VQDMULL. S32 Qd, Dn, Dm[lane]

A64 : SQDMULL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyDoublingWideningSaturateLowerBySelectedScalar(Vector64<Int32>, Vector64<Int32>, Byte)

int64x2_t vqdmull_lane_s32 (int32x2_t a, int32x2_t v, const int lane)

A32 : VQDMULL. S32 Qd, Dn, Dm[lane]

A64 : SQDMULL Vd.2D, Vn.2S, Vm.S[lane]

MultiplyDoublingWideningSaturateUpper(Vector128<Int16>, Vector128<Int16>)

int32x4_t vqdmull_high_s16 (int16x8_t a, int16x8_t b)

A32 : VQDMULL. S16 Qd, Dn+1, Dm+1

A64 : SQDMULL2 Vd.4S, Vn.8H, Vm.8H

MultiplyDoublingWideningSaturateUpper(Vector128<Int32>, Vector128<Int32>)

int64x2_t vqdmull_high_s32 (int32x4_t a, int32x4_t b)

A32 : VQDMULL. S32 Qd, Dn+1, Dm+1

A64 : SQDMULL2 Vd.2D, Vn.4S, Vm.4S

MultiplyDoublingWideningSaturateUpperByScalar(Vector128<Int16>, Vector64<Int16>)

int32x4_t vqdmull_high_n_s16 (int16x8_t a, int16_t b)

A32 : VQDMULL. S16 Qd, Dn+1, Dm[0]

A64 : SQDMULL2 Vd.4S, Vn.8H, Vm.H[0]

MultiplyDoublingWideningSaturateUpperByScalar(Vector128<Int32>, Vector64<Int32>)

int64x2_t vqdmull_high_n_s32 (int32x4_t a, int32_t b)

A32 : VQDMULL. S32 Qd, Dn+1, Dm[0]

A64 : SQDMULL2 Vd.2D, Vn.4S, Vm.S[0]

MultiplyDoublingWideningSaturateUpperBySelectedScalar(Vector128<Int16>, Vector128<Int16>, Byte)

int32x4_t vqdmull_high_laneq_s16 (int16x8_t a, int16x8_t v, const int lane)

A32 : VQDMULL. S16 Qd, Dn+1, Dm[lane]

A64 : SQDMULL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyDoublingWideningSaturateUpperBySelectedScalar(Vector128<Int16>, Vector64<Int16>, Byte)

int32x4_t vqdmull_high_lane_s16 (int16x8_t a, int16x4_t v, const int lane)

A32 : VQDMULL. S16 Qd, Dn+1, Dm[lane]

A64 : SQDMULL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyDoublingWideningSaturateUpperBySelectedScalar(Vector128<Int32>, Vector128<Int32>, Byte)

int64x2_t vqdmull_high_laneq_s32 (int32x4_t a, int32x4_t v, const int lane)

A32 : VQDMULL. S32 Qd, Dn+1, Dm[lane]

A64 : SQDMULL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyDoublingWideningSaturateUpperBySelectedScalar(Vector128<Int32>, Vector64<Int32>, Byte)

int64x2_t vqdmull_high_lane_s32 (int32x4_t a, int32x2_t v, const int lane)

A32 : VQDMULL. S32 Qd, Dn+1, Dm[lane]

A64 : SQDMULL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyDoublingWideningUpperAndAddSaturate(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>)

int32x4_t vqdmlal_high_s16 (int32x4_t a, int16x8_t b, int16x8_t c)

A32 : VQDMLAL. S16 Qd, Dn+1, Dm+1

A64 : SQDMLAL2 Vd.4S, Vn.8H, Vm.8H

MultiplyDoublingWideningUpperAndAddSaturate(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>)

int64x2_t vqdmlal_high_s32 (int64x2_t a, int32x4_t b, int32x4_t c)

A32 : VQDMLAL. S32 Qd, Dn+1, Dm+1

A64 : SQDMLAL2 Vd.2D, Vn.4S, Vm.4S

MultiplyDoublingWideningUpperAndSubtractSaturate(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>)

int32x4_t vqdmlsl_high_s16 (int32x4_t a, int16x8_t b, int16x8_t c)

A32 : VQDMLSL. S16 Qd, Dn+1, Dm+1

A64 : SQDMLSL2 Vd.4S, Vn.8H, Vm.8H

MultiplyDoublingWideningUpperAndSubtractSaturate(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>)

int64x2_t vqdmlsl_high_s32 (int64x2_t a, int32x4_t b, int32x4_t c)

A32 : VQDMLSL. S32 Qd, Dn+1, Dm+1

A64 : SQDMLSL2 Vd.2D, Vn.4S, Vm.4S

MultiplyDoublingWideningUpperByScalarAndAddSaturate(Vector128<Int32>, Vector128<Int16>, Vector64<Int16>)

int32x4_t vqdmlal_high_n_s16 (int32x4_t a, int16x8_t b, int16_t c)

A32 : VQDMLAL. S16 Qd, Dn+1, Dm[0]

A64 : SQDMLAL2 Vd.4S, Vn.8H, Vm.H[0]

MultiplyDoublingWideningUpperByScalarAndAddSaturate(Vector128<Int64>, Vector128<Int32>, Vector64<Int32>)

int64x2_t vqdmlal_high_n_s32 (int64x2_t a, int32x4_t b, int32_t c)

A32 : VQDMLAL. S32 Qd, Dn+1, Dm[0]

A64 : SQDMLAL2 Vd.2D, Vn.4S, Vm.S[0]

MultiplyDoublingWideningUpperByScalarAndSubtractSaturate(Vector128<Int32>, Vector128<Int16>, Vector64<Int16>)

int32x4_t vqdmlsl_high_n_s16 (int32x4_t a, int16x8_t b, int16_t c)

A32 : VQDMLSL. S16 Qd, Dn+1, Dm[0]

A64 : SQDMLSL2 Vd.4S, Vn.8H, Vm.H[0]

MultiplyDoublingWideningUpperByScalarAndSubtractSaturate(Vector128<Int64>, Vector128<Int32>, Vector64<Int32>)

int64x2_t vqdmlsl_high_n_s32 (int64x2_t a, int32x4_t b, int32_t c)

A32 : VQDMLSL. S32 Qd, Dn+1, Dm[0]

A64 : SQDMLSL2 Vd.2D, Vn.4S, Vm.S[0]

MultiplyDoublingWideningUpperBySelectedScalarAndAddSaturate(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>, Byte)

int32x4_t vqdmlal_high_laneq_s16 (int32x4_t a, int16x8_t b, int16x8_t v, const int lane)

A32 : VQDMLAL. S16 Qd, Dn+1, Dm[lane]

A64 : SQDMLAL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyDoublingWideningUpperBySelectedScalarAndAddSaturate(Vector128<Int32>, Vector128<Int16>, Vector64<Int16>, Byte)

int32x4_t vqdmlal_high_lane_s16 (int32x4_t a, int16x8_t b, int16x4_t v, const int lane)

A32 : VQDMLAL. S16 Qd, Dn+1, Dm[lane]

A64 : SQDMLAL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyDoublingWideningUpperBySelectedScalarAndAddSaturate(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>, Byte)

int64x2_t vqdmlal_high_laneq_s32 (int64x2_t a, int32x4_t b, int32x4_t v, const int lane)

A32 : VQDMLAL. S32 Qd, Dn+1, Dm[lane]

A64 : SQDMLAL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyDoublingWideningUpperBySelectedScalarAndAddSaturate(Vector128<Int64>, Vector128<Int32>, Vector64<Int32>, Byte)

int64x2_t vqdmlal_high_lane_s32 (int64x2_t a, int32x4_t b, int32x2_t v, const int lane)

A32 : VQDMLAL. S32 Qd, Dn+1, Dm[lane]

A64 : SQDMLAL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyDoublingWideningUpperBySelectedScalarAndSubtractSaturate(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>, Byte)

int32x4_t vqdmlsl_high_laneq_s16 (int32x4_t a, int16x8_t b, int16x8_t v, const int lane)

A32 : VQDMLSL. S16 Qd, Dn+1, Dm[lane]

A64 : SQDMLSL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyDoublingWideningUpperBySelectedScalarAndSubtractSaturate(Vector128<Int32>, Vector128<Int16>, Vector64<Int16>, Byte)

int32x4_t vqdmlsl_high_lane_s16 (int32x4_t a, int16x8_t b, int16x4_t v, const int lane)

A32 : VQDMLSL. S16 Qd, Dn+1, Dm[lane]

A64 : SQDMLSL2 Vd.4S, Vn.8H, Vm.H[lane]

MultiplyDoublingWideningUpperBySelectedScalarAndSubtractSaturate(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>, Byte)

int64x2_t vqdmlsl_high_laneq_s32 (int64x2_t a, int32x4_t b, int32x4_t v, const int lane)

A32 : VQDMLSL. S32 Qd, Dn+1, Dm[lane]

A64 : SQDMLSL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyDoublingWideningUpperBySelectedScalarAndSubtractSaturate(Vector128<Int64>, Vector128<Int32>, Vector64<Int32>, Byte)

int64x2_t vqdmlsl_high_lane_s32 (int64x2_t a, int32x4_t b, int32x2_t v, const int lane)

A32 : VQDMLSL. S32 Qd, Dn+1, Dm[lane]

A64 : SQDMLSL2 Vd.2D, Vn.4S, Vm.S[lane]

MultiplyRoundedDoublingByScalarSaturateHigh(Vector128<Int16>, Vector64<Int16>)

int16x8_t vqrdmulhq_n_s16 (int16x8_t a, int16_t b)

A32 : VQRDMULH. S16 Qd, Qn, Dm[0]

A64 : SQRDMULH Vd.8H, Vn.8H, Vm.H[0]

MultiplyRoundedDoublingByScalarSaturateHigh(Vector128<Int32>, Vector64<Int32>)

int32x4_t vqrdmulhq_n_s32 (int32x4_t a, int32_t b)

A32 : VQRDMULH. S32 Qd, Qn, Dm[0]

A64 : SQRDMULH Vd.4S, Vn.4S, Vm.S[0]

MultiplyRoundedDoublingByScalarSaturateHigh(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqrdmulh_n_s16 (int16x4_t a, int16_t b)

A32 : VQRDMULH. S16 Dd, Dn, Dm[0]

A64 : SQRDMULH Vd.4H, Vn.4H, Vm.H[0]

MultiplyRoundedDoublingByScalarSaturateHigh(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqrdmulh_n_s32 (int32x2_t a, int32_t b)

A32 : VQRDMULH. S32 Dd, Dn, Dm[0]

A64 : SQRDMULH Vd.2S, Vn.2S, Vm.S[0]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vqrdmulhq_laneq_s16 (int16x8_t a, int16x8_t v, const int lane)

A32 : VQRDMULH. S16 Qd, Qn, Dm[lane]

A64 : SQRDMULH Vd.8H, Vn.8H, Vm.H[lane]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector128<Int16>, Vector64<Int16>, Byte)

int16x8_t vqrdmulhq_lane_s16 (int16x8_t a, int16x4_t v, const int lane)

A32 : VQRDMULH. S16 Qd, Qn, Dm[lane]

A64 : SQRDMULH Vd.8H, Vn.8H, Vm.H[lane]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vqrdmulhq_laneq_s32 (int32x4_t a, int32x4_t v, const int lane)

A32 : VQRDMULH. S32 Qd, Qn, Dm[lane]

A64 : SQRDMULH Vd.4S, Vn.4S, Vm.S[lane]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector128<Int32>, Vector64<Int32>, Byte)

int32x4_t vqrdmulhq_lane_s32 (int32x4_t a, int32x2_t v, const int lane)

A32 : VQRDMULH. S32 Qd, Qn, Dm[lane]

A64 : SQRDMULH Vd.4S, Vn.4S, Vm.S[lane]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector64<Int16>, Vector128<Int16>, Byte)

int16x4_t vqrdmulh_laneq_s16 (int16x4_t a, int16x8_t v, const int lane)

A32 : VQRDMULH. S16 Dd, Dn, Dm[lane]

A64 : SQRDMULH Vd.4H, Vn.4H, Vm.H[lane]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vqrdmulh_lane_s16 (int16x4_t a, int16x4_t v, const int lane)

A32 : VQRDMULH. S16 Dd, Dn, Dm[lane]

A64 : SQRDMULH Vd.4H, Vn.4H, Vm.H[lane]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector64<Int32>, Vector128<Int32>, Byte)

int32x2_t vqrdmulh_laneq_s32 (int32x2_t a, int32x4_t v, const int lane)

A32 : VQRDMULH. S32 Dd, Dn, Dm[lane]

A64 : SQRDMULH Vd.2S, Vn.2S, Vm.S[lane]

MultiplyRoundedDoublingBySelectedScalarSaturateHigh(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vqrdmulh_lane_s32 (int32x2_t a, int32x2_t v, const int lane)

A32 : VQRDMULH. S32 Dd, Dn, Dm[lane]

A64 : SQRDMULH Vd.2S, Vn.2S, Vm.S[lane]

MultiplyRoundedDoublingSaturateHigh(Vector128<Int16>, Vector128<Int16>)

int16x8_t vqrdmulhq_s16 (int16x8_t a, int16x8_t b)

A32 : VQRDMULH. S16 Qd, Qn, Qm

A64 : SQRDMULH Vd.8H, Vn.8H, Vm.8H

MultiplyRoundedDoublingSaturateHigh(Vector128<Int32>, Vector128<Int32>)

int32x4_t vqrdmulhq_s32 (int32x4_t a, int32x4_t b)

A32 : VQRDMULH. S32 Qd, Qn, Qm

A64 : SQRDMULH Vd.4S, Vn.4S, Vm.4S

MultiplyRoundedDoublingSaturateHigh(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqrdmulh_s16 (int16x4_t a, int16x4_t b)

A32 : VQRDMULH. S16 Dd, Dn, Dm

A64 : SQRDMULH Vd.4H, Vn.4H, Vm.4H

MultiplyRoundedDoublingSaturateHigh(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqrdmulh_s32 (int32x2_t a, int32x2_t b)

A32 : VQRDMULH. S32 Dd, Dn, Dm

A64 : SQRDMULH Vd.2S, Vn.2S, Vm.2S

MultiplyScalar(Vector64<Double>, Vector64<Double>)

float64x1_t vmul_f64 (float64x1_t a, float64x1_t b)

A32 : VMUL.F64 Dd, Dn, Dm

A64 : FMUL Dd, Dn, Dm

MultiplyScalar(Vector64<Single>, Vector64<Single>)

float32_t vmuls_f32 (float32_t a, float32_t b)

A32 : VMUL.F32 Sd, Sn, Sm

A64 : FMUL Sd, Sn, Sm La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

MultiplyScalarBySelectedScalar(Vector64<Single>, Vector128<Single>, Byte)

float32_t vmuls_laneq_f32 (float32_t a, float32x4_t v, const int lane)

A32 : VMUL.F32 Sd, Sn, Dm[lane]

A64 : FMUL Sd, Sn, Vm.S[lane]

MultiplyScalarBySelectedScalar(Vector64<Single>, Vector64<Single>, Byte)

float32_t vmuls_lane_f32 (float32_t a, float32x2_t v, const int lane)

A32 : VMUL.F32 Sd, Sn, Dm[lane]

A64 : FMUL Sd, Sn, Vm.S[lane]

MultiplySubtract(Vector128<Byte>, Vector128<Byte>, Vector128<Byte>)

uint8x16_t vmlsq_u8 (uint8x16_t a, uint8x16_t b, uint8x16_t c)

A32 : VMLS.I8 Qd, Qn, Qm

A64 : MLS Vd.16B, Vn.16B, Vm.16B

MultiplySubtract(Vector128<Int16>, Vector128<Int16>, Vector128<Int16>)

int16x8_t vmlsq_s16 (int16x8_t a, int16x8_t b, int16x8_t c)

A32 : VMLS.I16 Qd, Qn, Qm

A64 : MLS Vd.8H, Vn.8H, Vm.8H

MultiplySubtract(Vector128<Int32>, Vector128<Int32>, Vector128<Int32>)

int32x4_t vmlsq_s32 (int32x4_t a, int32x4_t b, int32x4_t c)

A32 : VMLS.I32 Qd, Qn, Qm

A64 : MLS Vd.4S, Vn.4S, Vm.4S

MultiplySubtract(Vector128<SByte>, Vector128<SByte>, Vector128<SByte>)

int8x16_t vmlsq_s8 (int8x16_t a, int8x16_t b, int8x16_t c)

A32 : VMLS.I8 Qd, Qn, Qm

A64 : MLS Vd.16B, Vn.16B, Vm.16B

MultiplySubtract(Vector128<UInt16>, Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vmlsq_u16 (uint16x8_t a, uint16x8_t b, uint16x8_t c)

A32 : VMLS.I16 Qd, Qn, Qm

A64 : MLS Vd.8H, Vn.8H, Vm.8H

MultiplySubtract(Vector128<UInt32>, Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vmlsq_u32 (uint32x4_t a, uint32x4_t b, uint32x4_t c)

A32 : VMLS.I32 Qd, Qn, Qm

A64 : MLS Vd.4S, Vn.4S, Vm.4S

MultiplySubtract(Vector64<Byte>, Vector64<Byte>, Vector64<Byte>)

uint8x8_t vmls_u8 (uint8x8_t a, uint8x8_t b, uint8x8_t c)

A32 : VMLS.I8 Dd, Dn, Dm

A64 : MLS Vd.8B, Vn.8B, Vm.8B

MultiplySubtract(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>)

int16x4_t vmls_s16 (int16x4_t a, int16x4_t b, int16x4_t c)

A32 : VMLS.I16 Dd, Dn, Dm

A64 : MLS Vd.4H, Vn.4H, Vm.4H

MultiplySubtract(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>)

int32x2_t vmls_s32 (int32x2_t a, int32x2_t b, int32x2_t c)

A32 : VMLS.I32 Dd, Dn, Dm

A64 : MLS Vd.2S, Vn.2S, Vm.2S

MultiplySubtract(Vector64<SByte>, Vector64<SByte>, Vector64<SByte>)

int8x8_t vmls_s8 (int8x8_t a, int8x8_t b, int8x8_t c)

A32 : VMLS.I8 Dd, Dn, Dm

A64 : MLS Vd.8B, Vn.8B, Vm.8B

MultiplySubtract(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmls_u16 (uint16x4_t a, uint16x4_t b, uint16x4_t c)

A32 : VMLS.I16 Dd, Dn, Dm

A64 : MLS Vd.4H, Vn.4H, Vm.4H

MultiplySubtract(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmls_u32 (uint32x2_t a, uint32x2_t b, uint32x2_t c)

A32 : VMLS.I32 Dd, Dn, Dm

A64 : MLS Vd.2S, Vn.2S, Vm.2S

MultiplySubtractByScalar(Vector128<Int16>, Vector128<Int16>, Vector64<Int16>)

int16x8_t vmlsq_n_s16 (int16x8_t a, int16x8_t b, int16_t c)

A32 : VMLS.I16 Qd, Qn, Dm[0]

A64 : MLS Vd.8H, Vn.8H, Vm.H[0]

MultiplySubtractByScalar(Vector128<Int32>, Vector128<Int32>, Vector64<Int32>)

int32x4_t vmlsq_n_s32 (int32x4_t a, int32x4_t b, int32_t c)

A32 : VMLS.I32 Qd, Qn, Dm[0]

A64 : MLS Vd.4S, Vn.4S, Vm.S[0]

MultiplySubtractByScalar(Vector128<UInt16>, Vector128<UInt16>, Vector64<UInt16>)

uint16x8_t vmlsq_n_u16 (uint16x8_t a, uint16x8_t b, uint16_t c)

A32 : VMLS.I16 Qd, Qn, Dm[0]

A64 : MLS Vd.8H, Vn.8H, Vm.H[0]

MultiplySubtractByScalar(Vector128<UInt32>, Vector128<UInt32>, Vector64<UInt32>)

uint32x4_t vmlsq_n_u32 (uint32x4_t a, uint32x4_t b, uint32_t c)

A32 : VMLS.I32 Qd, Qn, Dm[0]

A64 : MLS Vd.4S, Vn.4S, Vm.S[0]

MultiplySubtractByScalar(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>)

int16x4_t vmls_n_s16 (int16x4_t a, int16x4_t b, int16_t c)

A32 : VMLS.I16 Dd, Dn, Dm[0]

A64 : MLS Vd.4H, Vn.4H, Vm.H[0]

MultiplySubtractByScalar(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>)

int32x2_t vmls_n_s32 (int32x2_t a, int32x2_t b, int32_t c)

A32 : VMLS.I32 Dd, Dn, Dm[0]

A64 : MLS Vd.2S, Vn.2S, Vm.S[0]

MultiplySubtractByScalar(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vmls_n_u16 (uint16x4_t a, uint16x4_t b, uint16_t c)

A32 : VMLS.I16 Dd, Dn, Dm[0]

A64 : MLS Vd.4H, Vn.4H, Vm.H[0]

MultiplySubtractByScalar(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vmls_n_u32 (uint32x2_t a, uint32x2_t b, uint32_t c)

A32 : VMLS.I32 Dd, Dn, Dm[0]

A64 : MLS Vd.2S, Vn.2S, Vm.S[0]

MultiplySubtractBySelectedScalar(Vector128<Int16>, Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vmlsq_laneq_s16 (int16x8_t a, int16x8_t b, int16x8_t v, const int lane)

A32 : VMLS.I16 Qd, Qn, Dm[lane]

A64 : MLS Vd.8H, Vn.8H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector128<Int16>, Vector128<Int16>, Vector64<Int16>, Byte)

int16x8_t vmlsq_lane_s16 (int16x8_t a, int16x8_t b, int16x4_t v, const int lane)

A32 : VMLS.I16 Qd, Qn, Dm[lane]

A64 : MLS Vd.8H, Vn.8H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector128<Int32>, Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vmlsq_laneq_s32 (int32x4_t a, int32x4_t b, int32x4_t v, const int lane)

A32 : VMLS.I32 Qd, Qn, Dm[lane]

A64 : MLS Vd.4S, Vn.4S, Vm.S[lane]

MultiplySubtractBySelectedScalar(Vector128<Int32>, Vector128<Int32>, Vector64<Int32>, Byte)

int32x4_t vmlsq_lane_s32 (int32x4_t a, int32x4_t b, int32x2_t v, const int lane)

A32 : VMLS.I32 Qd, Qn, Dm[lane]

A64 : MLS Vd.4S, Vn.4S, Vm.S[lane]

MultiplySubtractBySelectedScalar(Vector128<UInt16>, Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vmlsq_laneq_u16 (uint16x8_t a, uint16x8_t b, uint16x8_t v, const int lane)

A32 : VMLS.I16 Qd, Qn, Dm[lane]

A64 : MLS Vd.8H, Vn.8H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector128<UInt16>, Vector128<UInt16>, Vector64<UInt16>, Byte)

uint16x8_t vmlsq_lane_u16 (uint16x8_t a, uint16x8_t b, uint16x4_t v, const int lane)

A32 : VMLS.I16 Qd, Qn, Dm[lane]

A64 : MLS Vd.8H, Vn.8H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector128<UInt32>, Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vmlsq_laneq_u32 (uint32x4_t a, uint32x4_t b, uint32x4_t v, const int lane)

A32 : VMLS.I32 Qd, Qn, Dm[lane]

A64 : MLS Vd.4S, Vn.4S, Vm.S[lane]

MultiplySubtractBySelectedScalar(Vector128<UInt32>, Vector128<UInt32>, Vector64<UInt32>, Byte)

uint32x4_t vmlsq_lane_u32 (uint32x4_t a, uint32x4_t b, uint32x2_t v, const int lane)

A32 : VMLS.I32 Qd, Qn, Dm[lane]

A64 : MLS Vd.4S, Vn.4S, Vm.S[lane]

MultiplySubtractBySelectedScalar(Vector64<Int16>, Vector64<Int16>, Vector128<Int16>, Byte)

int16x4_t vmls_laneq_s16 (int16x4_t a, int16x4_t b, int16x8_t v, const int lane)

A32 : VMLS.I16 Dd, Dn, Dm[lane]

A64 : MLS Vd.4H, Vn.4H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector64<Int16>, Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vmls_lane_s16 (int16x4_t a, int16x4_t b, int16x4_t v, const int lane)

A32 : VMLS.I16 Dd, Dn, Dm[lane]

A64 : MLS Vd.4H, Vn.4H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector64<Int32>, Vector64<Int32>, Vector128<Int32>, Byte)

int32x2_t vmls_laneq_s32 (int32x2_t a, int32x2_t b, int32x4_t v, const int lane)

A32 : VMLS.I32 Dd, Dn, Dm[lane]

A64 : MLS Vd.2S, Vn.2S, Vm.S[lane]

MultiplySubtractBySelectedScalar(Vector64<Int32>, Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vmls_lane_s32 (int32x2_t a, int32x2_t b, int32x2_t v, const int lane)

A32 : VMLS.I32 Dd, Dn, Dm[lane]

A64 : MLS Vd.2S, Vn.2S, Vm.S[lane]

MultiplySubtractBySelectedScalar(Vector64<UInt16>, Vector64<UInt16>, Vector128<UInt16>, Byte)

uint16x4_t vmls_laneq_u16 (uint16x4_t a, uint16x4_t b, uint16x8_t v, const int lane)

A32 : VMLS.I16 Dd, Dn, Dm[lane]

A64 : MLS Vd.4H, Vn.4H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector64<UInt16>, Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vmls_lane_u16 (uint16x4_t a, uint16x4_t b, uint16x4_t v, const int lane)

A32 : VMLS.I16 Dd, Dn, Dm[lane]

A64 : MLS Vd.4H, Vn.4H, Vm.H[lane]

MultiplySubtractBySelectedScalar(Vector64<UInt32>, Vector64<UInt32>, Vector128<UInt32>, Byte)

uint32x2_t vmls_laneq_u32 (uint32x2_t a, uint32x2_t b, uint32x4_t v, const int lane)

A32 : VMLS.I32 Dd, Dn, Dm[lane]

A64 : MLS Vd.2S, Vn.2S, Vm.S[lane]

MultiplySubtractBySelectedScalar(Vector64<UInt32>, Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vmls_lane_u32 (uint32x2_t a, uint32x2_t b, uint32x2_t v, const int lane)

A32 : VMLS.I32 Dd, Dn, Dm[lane]

A64 : MLS Vd.2S, Vn.2S, Vm.S[lane]

MultiplyWideningLower(Vector64<Byte>, Vector64<Byte>)

uint16x8_t vmull_u8 (uint8x8_t a, uint8x8_t b)

A32 : VMULL.U8 Qd, Dn, Dm

A64 : UMULL Vd.8H, Vn.8B, Vm.8B

MultiplyWideningLower(Vector64<Int16>, Vector64<Int16>)

int32x4_t vmull_s16 (int16x4_t a, int16x4_t b)

A32 : VMULL.S16 Qd, Dn, Dm

A64 : SMULL Vd.4S, Vn.4H, Vm.4H

MultiplyWideningLower(Vector64<Int32>, Vector64<Int32>)

int64x2_t vmull_s32 (int32x2_t a, int32x2_t b)

A32 : VMULL.S32 Qd, Dn, Dm

A64 : SMULL Vd.2D, Vn.2S, Vm.2S

MultiplyWideningLower(Vector64<SByte>, Vector64<SByte>)

int16x8_t vmull_s8 (int8x8_t a, int8x8_t b)

A32 : VMULL.S8 Qd, Dn, Dm

A64 : SMULL Vd.8H, Vn.8B, Vm.8B

MultiplyWideningLower(Vector64<UInt16>, Vector64<UInt16>)

uint32x4_t vmull_u16 (uint16x4_t a, uint16x4_t b)

A32 : VMULL.U16 Qd, Dn, Dm

A64 : UMULL Vd.4S, Vn.4H, Vm.4H

MultiplyWideningLower(Vector64<UInt32>, Vector64<UInt32>)

uint64x2_t vmull_u32 (uint32x2_t a, uint32x2_t b)

A32 : VMULL.U32 Qd, Dn, Dm

A64 : UMULL Vd.2D, Vn.2S, Vm.2S

MultiplyWideningLowerAndAdd(Vector128<Int16>, Vector64<SByte>, Vector64<SByte>)

int16x8_t vmlal_s8 (int16x8_t a, int8x8_t b, int8x8_t c)

A32 : VMLAL.S8 Qd, Dn, Dm

A64 : SMLAL Vd.8H, Vn.8B, Vm.8B

MultiplyWideningLowerAndAdd(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>)

int32x4_t vmlal_s16 (int32x4_t a, int16x4_t b, int16x4_t c)

A32 : VMLAL.S16 Qd, Dn, Dm

A64 : SMLAL Vd.4S, Vn.4H, Vm.4H

MultiplyWideningLowerAndAdd(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>)

int64x2_t vmlal_s32 (int64x2_t a, int32x2_t b, int32x2_t c)

A32 : VMLAL.S32 Qd, Dn, Dm

A64 : SMLAL Vd.2D, Vn.2S, Vm.2S

MultiplyWideningLowerAndAdd(Vector128<UInt16>, Vector64<Byte>, Vector64<Byte>)

uint16x8_t vmlal_u8 (uint16x8_t a, uint8x8_t b, uint8x8_t c)

A32 : VMLAL.U8 Qd, Dn, Dm

A64 : UMLAL Vd.8H, Vn.8B, Vm.8B

MultiplyWideningLowerAndAdd(Vector128<UInt32>, Vector64<UInt16>, Vector64<UInt16>)

uint32x4_t vmlal_u16 (uint32x4_t a, uint16x4_t b, uint16x4_t c)

A32 : VMLAL.U16 Qd, Dn, Dm

A64 : UMLAL Vd.4S, Vn.4H, Vm.4H

MultiplyWideningLowerAndAdd(Vector128<UInt64>, Vector64<UInt32>, Vector64<UInt32>)

uint64x2_t vmlal_u32 (uint64x2_t a, uint32x2_t b, uint32x2_t c)

A32 : VMLAL.U32 Qd, Dn, Dm

A64 : UMLAL Vd.2D, Vn.2S, Vm.2S

MultiplyWideningLowerAndSubtract(Vector128<Int16>, Vector64<SByte>, Vector64<SByte>)

int16x8_t vmlsl_s8 (int16x8_t a, int8x8_t b, int8x8_t c)

A32 : VMLSL.S8 Qd, Dn, Dm

A64 : SMLSL Vd.8H, Vn.8B, Vm.8B

MultiplyWideningLowerAndSubtract(Vector128<Int32>, Vector64<Int16>, Vector64<Int16>)

int32x4_t vmlsl_s16 (int32x4_t a, int16x4_t b, int16x4_t c)

A32 : VMLSL.S16 Qd, Dn, Dm

A64 : SMLSL Vd.4S, Vn.4H, Vm.4H

MultiplyWideningLowerAndSubtract(Vector128<Int64>, Vector64<Int32>, Vector64<Int32>)

int64x2_t vmlsl_s32 (int64x2_t a, int32x2_t b, int32x2_t c)

A32 : VMLSL.S32 Qd, Dn, Dm

A64 : SMLSL Vd.2D, Vn.2S, Vm.2S

MultiplyWideningLowerAndSubtract(Vector128<UInt16>, Vector64<Byte>, Vector64<Byte>)

uint16x8_t vmlsl_u8 (uint16x8_t a, uint8x8_t b, uint8x8_t c)

A32 : VMLSL.U8 Qd, Dn, Dm

A64 : UMLSL Vd.8H, Vn.8B, Vm.8B

MultiplyWideningLowerAndSubtract(Vector128<UInt32>, Vector64<UInt16>, Vector64<UInt16>)

uint32x4_t vmlsl_u16 (uint32x4_t a, uint16x4_t b, uint16x4_t c)

A32 : VMLSL.U16 Qd, Dn, Dm

A64 : UMLSL Vd.4S, Vn.4H, Vm.4H

MultiplyWideningLowerAndSubtract(Vector128<UInt64>, Vector64<UInt32>, Vector64<UInt32>)

uint64x2_t vmlsl_u32 (uint64x2_t a, uint32x2_t b, uint32x2_t c)

A32 : VMLSL.U32 Qd, Dn, Dm

A64 : UMLSL Vd.2D, Vn.2S, Vm.2S

MultiplyWideningUpper(Vector128<Byte>, Vector128<Byte>)

uint16x8_t vmull_high_u8 (uint8x16_t a, uint8x16_t b)

A32 : VMULL.U8 Qd, Dn+1, Dm+1

A64 : UMULL2 Vd.8H, Vn.16B, Vm.16B

MultiplyWideningUpper(Vector128<Int16>, Vector128<Int16>)

int32x4_t vmull_high_s16 (int16x8_t a, int16x8_t b)

A32 : VMULL.S16 Qd, Dn+1, Dm+1

A64 : SMULL2 Vd.4S, Vn.8H, Vm.8H

MultiplyWideningUpper(Vector128<Int32>, Vector128<Int32>)

int64x2_t vmull_high_s32 (int32x4_t a, int32x4_t b)

A32 : VMULL.S32 Qd, Dn+1, Dm+1

A64 : SMULL2 Vd.2D, Vn.4S, Vm.4S

MultiplyWideningUpper(Vector128<SByte>, Vector128<SByte>)

int16x8_t vmull_high_s8 (int8x16_t a, int8x16_t b)

A32 : VMULL.S8 Qd, Dn+1, Dm+1

A64 : SMULL2 Vd.8H, Vn.16B, Vm.16B

MultiplyWideningUpper(Vector128<UInt16>, Vector128<UInt16>)

uint32x4_t vmull_high_u16 (uint16x8_t a, uint16x8_t b)

A32 : VMULL.U16 Qd, Dn+1, Dm+1

A64 : UMULL2 Vd.4S, Vn.8H, Vm.8H

MultiplyWideningUpper(Vector128<UInt32>, Vector128<UInt32>)

uint64x2_t vmull_high_u32 (uint32x4_t a, uint32x4_t b)

A32 : VMULL.U32 Qd, Dn+1, Dm+1

A64 : UMULL2 Vd.2D, Vn.4S, Vm.4S

MultiplyWideningUpperAndAdd(Vector128<Int16>, Vector128<SByte>, Vector128<SByte>)

int16x8_t vmlal_high_s8 (int16x8_t a, int8x16_t b, int8x16_t c)

A32 : VMLAL.S8 Qd, Dn+1, Dm+1

A64 : SMLAL2 Vd.8H, Vn.16B, Vm.16B

MultiplyWideningUpperAndAdd(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>)

int32x4_t vmlal_high_s16 (int32x4_t a, int16x8_t b, int16x8_t c)

A32 : VMLAL.S16 Qd, Dn+1, Dm+1

A64 : SMLAL2 Vd.4S, Vn.8H, Vm.8H

MultiplyWideningUpperAndAdd(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>)

int64x2_t vmlal_high_s32 (int64x2_t a, int32x4_t b, int32x4_t c)

A32 : VMLAL.S32 Qd, Dn+1, Dm+1

A64 : SMLAL2 Vd.2D, Vn.4S, Vm.4S

MultiplyWideningUpperAndAdd(Vector128<UInt16>, Vector128<Byte>, Vector128<Byte>)

uint16x8_t vmlal_high_u8 (uint16x8_t a, uint8x16_t b, uint8x16_t c)

A32 : VMLAL.U8 Qd, Dn+1, Dm+1

A64 : UMLAL2 Vd.8H, Vn.16B, Vm.16B

MultiplyWideningUpperAndAdd(Vector128<UInt32>, Vector128<UInt16>, Vector128<UInt16>)

uint32x4_t vmlal_high_u16 (uint32x4_t a, uint16x8_t b, uint16x8_t c)

A32 : VMLAL.U16 Qd, Dn+1, Dm+1

A64 : UMLAL2 Vd.4S, Vn.8H, Vm.8H

MultiplyWideningUpperAndAdd(Vector128<UInt64>, Vector128<UInt32>, Vector128<UInt32>)

uint64x2_t vmlal_high_u32 (uint64x2_t a, uint32x4_t b, uint32x4_t c)

A32 : VMLAL.U32 Qd, Dn+1, Dm+1

A64 : UMLAL2 Vd.2D, Vn.4S, Vm.4S

MultiplyWideningUpperAndSubtract(Vector128<Int16>, Vector128<SByte>, Vector128<SByte>)

int16x8_t vmlsl_high_s8 (int16x8_t a, int8x16_t b, int8x16_t c)

A32 : VMLSL.S8 Qd, Dn+1, Dm+1

A64 : SMLSL2 Vd.8H, Vn.16B, Vm.16B

MultiplyWideningUpperAndSubtract(Vector128<Int32>, Vector128<Int16>, Vector128<Int16>)

int32x4_t vmlsl_high_s16 (int32x4_t a, int16x8_t b, int16x8_t c)

A32 : VMLSL.S16 Qd, Dn+1, Dm+1

A64 : SMLSL2 Vd.4S, Vn.8H, Vm.8H

MultiplyWideningUpperAndSubtract(Vector128<Int64>, Vector128<Int32>, Vector128<Int32>)

int64x2_t vmlsl_high_s32 (int64x2_t a, int32x4_t b, int32x4_t c)

A32 : VMLSL.S32 Qd, Dn+1, Dm+1

A64 : SMLSL2 Vd.2D, Vn.4S, Vm.4S

MultiplyWideningUpperAndSubtract(Vector128<UInt16>, Vector128<Byte>, Vector128<Byte>)

uint16x8_t vmlsl_high_u8 (uint16x8_t a, uint8x16_t b, uint8x16_t c)

A32 : VMLSL.U8 Qd, Dn+1, Dm+1

A64 : UMLSL2 Vd.8H, Vn.16B, Vm.16B

MultiplyWideningUpperAndSubtract(Vector128<UInt32>, Vector128<UInt16>, Vector128<UInt16>)

uint32x4_t vmlsl_high_u16 (uint32x4_t a, uint16x8_t b, uint16x8_t c)

A32 : VMLSL.U16 Qd, Dn+1, Dm+1

A64 : UMLSL2 Vd.4S, Vn.8H, Vm.8H

MultiplyWideningUpperAndSubtract(Vector128<UInt64>, Vector128<UInt32>, Vector128<UInt32>)

uint64x2_t vmlsl_high_u32 (uint64x2_t a, uint32x4_t b, uint32x4_t c)

A32 : VMLSL.U32 Qd, Dn+1, Dm+1

A64 : UMLSL2 Vd.2D, Vn.4S, Vm.4S

Negate(Vector128<Int16>)

int16x8_t vnegq_s16 (int16x8_t a)

A32 : VNEG.S16 Qd, Qm

A64 : NEG Vd.8H, Vn.8H

Negate(Vector128<Int32>)

int32x4_t vnegq_s32 (int32x4_t a)

A32 : VNEG.S32 Qd, Qm

A64 : NEG Vd.4S, Vn.4S

Negate(Vector128<SByte>)

int8x16_t vnegq_s8 (int8x16_t a)

A32 : VNEG.S8 Qd, Qm

A64 : NEG Vd.16B, Vn.16B

Negate(Vector128<Single>)

float32x4_t vnegq_f32 (float32x4_t a)

A32 : VNEG.F32 Qd, Qm

A64 : FNEG Vd.4S, Vn.4S

Negate(Vector64<Int16>)

int16x4_t vneg_s16 (int16x4_t a)

A32 : VNEG.S16 Dd, Dm

A64 : NEG Vd.4H, Vn.4H

Negate(Vector64<Int32>)

int32x2_t vneg_s32 (int32x2_t a)

A32 : VNEG.S32 Dd, Dm

A64 : NEG Vd.2S, Vn.2S

Negate(Vector64<SByte>)

int8x8_t vneg_s8 (int8x8_t a)

A32 : VNEG.S8 Dd, Dm

A64 : NEG Vd.8B, Vn.8B

Negate(Vector64<Single>)

float32x2_t vneg_f32 (float32x2_t a)

A32 : VNEG.F32 Dd, Dm

A64 : FNEG Vd.2S, Vn.2S

NegateSaturate(Vector128<Int16>)

int16x8_t vqnegq_s16 (int16x8_t a)

A32 : VQNEG.S16 Qd, Qm

A64 : SQNEG Vd.8H, Vn.8H

NegateSaturate(Vector128<Int32>)

int32x4_t vqnegq_s32 (int32x4_t a)

A32 : VQNEG.S32 Qd, Qm

A64 : SQNEG Vd.4S, Vn.4S

NegateSaturate(Vector128<SByte>)

int8x16_t vqnegq_s8 (int8x16_t a)

A32 : VQNEG.S8 Qd, Qm

A64 : SQNEG Vd.16B, Vn.16B

NegateSaturate(Vector64<Int16>)

int16x4_t vqneg_s16 (int16x4_t a)

A32 : VQNEG.S16 Dd, Dm

A64 : SQNEG Vd.4H, Vn.4H

NegateSaturate(Vector64<Int32>)

int32x2_t vqneg_s32 (int32x2_t a)

A32 : VQNEG.S32 Dd, Dm

A64 : SQNEG Vd.2S, Vn.2S

NegateSaturate(Vector64<SByte>)

int8x8_t vqneg_s8 (int8x8_t a)

A32 : VQNEG.S8 Dd, Dm

A64 : SQNEG Vd.8B, Vn.8B

NegateScalar(Vector64<Double>)

float64x1_t vneg_f64 (float64x1_t a)

A32 : VNEG.F64 Dd, Dm

A64 : FNEG Dd, Dn

NegateScalar(Vector64<Single>)

float32_t vnegs_f32 (float32_t a)

A32 : VNEG.F32 Sd, Sm

A64 : FNEG Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Not(Vector128<Byte>)

uint8x16_t vmvnq_u8 (uint8x16_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector128<Double>)

float64x2_t vmvnq_f64 (float64x2_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Not(Vector128<Int16>)

int16x8_t vmvnq_s16 (int16x8_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector128<Int32>)

int32x4_t vmvnq_s32 (int32x4_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector128<Int64>)

int64x2_t vmvnq_s64 (int64x2_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector128<SByte>)

int8x16_t vmvnq_s8 (int8x16_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector128<Single>)

float32x4_t vmvnq_f32 (float32x4_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Not(Vector128<UInt16>)

uint16x8_t vmvnq_u16 (uint16x8_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector128<UInt32>)

uint32x4_t vmvnq_u32 (uint32x4_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector128<UInt64>)

uint64x2_t vmvnq_u64 (uint64x2_t a)

A32 : VMVN Qd, Qm

A64 : MVN Vd.16B, Vn.16B

Not(Vector64<Byte>)

uint8x8_t vmvn_u8 (uint8x8_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Not(Vector64<Double>)

float64x1_t vmvn_f64 (float64x1_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Not(Vector64<Int16>)

int16x4_t vmvn_s16 (int16x4_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Not(Vector64<Int32>)

int32x2_t vmvn_s32 (int32x2_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Not(Vector64<Int64>)

int64x1_t vmvn_s64 (int64x1_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Not(Vector64<SByte>)

int8x8_t vmvn_s8 (int8x8_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Not(Vector64<Single>)

float32x2_t vmvn_f32 (float32x2_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Not(Vector64<UInt16>)

uint16x4_t vmvn_u16 (uint16x4_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Not(Vector64<UInt32>)

uint32x2_t vmvn_u32 (uint32x2_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Not(Vector64<UInt64>)

uint64x1_t vmvn_u64 (uint64x1_t a)

A32 : VMVN Dd, Dm

A64 : MVN Vd.8B, Vn.8B

Or(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vorrq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector128<Double>, Vector128<Double>)

float64x2_t vorrq_f64 (float64x2_t a, float64x2_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Or(Vector128<Int16>, Vector128<Int16>)

int16x8_t vorrq_s16 (int16x8_t a, int16x8_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector128<Int32>, Vector128<Int32>)

int32x4_t vorrq_s32 (int32x4_t a, int32x4_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector128<Int64>, Vector128<Int64>)

int64x2_t vorrq_s64 (int64x2_t a, int64x2_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector128<SByte>, Vector128<SByte>)

int8x16_t vorrq_s8 (int8x16_t a, int8x16_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector128<Single>, Vector128<Single>)

float32x4_t vorrq_f32 (float32x4_t a, float32x4_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Or(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vorrq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vorrq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vorrq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VORR Qd, Qn, Qm

A64 : ORR Vd.16B, Vn.16B, Vm.16B

Or(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vorr_u8 (uint8x8_t a, uint8x8_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

Or(Vector64<Double>, Vector64<Double>)

float64x1_t vorr_f64 (float64x1_t a, float64x1_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Or(Vector64<Int16>, Vector64<Int16>)

int16x4_t vorr_s16 (int16x4_t a, int16x4_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

Or(Vector64<Int32>, Vector64<Int32>)

int32x2_t vorr_s32 (int32x2_t a, int32x2_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

Or(Vector64<Int64>, Vector64<Int64>)

int64x1_t vorr_s64 (int64x1_t a, int64x1_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

Or(Vector64<SByte>, Vector64<SByte>)

int8x8_t vorr_s8 (int8x8_t a, int8x8_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

Or(Vector64<Single>, Vector64<Single>)

float32x2_t vorr_f32 (float32x2_t a, float32x2_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Or(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vorr_u16 (uint16x4_t a, uint16x4_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

Or(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vorr_u32 (uint32x2_t a, uint32x2_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

Or(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vorr_u64 (uint64x1_t a, uint64x1_t b)

A32 : VORR Dd, Dn, Dm

A64 : ORR Vd.8B, Vn.8B, Vm.8B

OrNot(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vornq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector128<Double>, Vector128<Double>)

float64x2_t vornq_f64 (float64x2_t a, float64x2_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

OrNot(Vector128<Int16>, Vector128<Int16>)

int16x8_t vornq_s16 (int16x8_t a, int16x8_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector128<Int32>, Vector128<Int32>)

int32x4_t vornq_s32 (int32x4_t a, int32x4_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector128<Int64>, Vector128<Int64>)

int64x2_t vornq_s64 (int64x2_t a, int64x2_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector128<SByte>, Vector128<SByte>)

int8x16_t vornq_s8 (int8x16_t a, int8x16_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector128<Single>, Vector128<Single>)

float32x4_t vornq_f32 (float32x4_t a, float32x4_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

OrNot(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vornq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vornq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vornq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VORN Qd, Qn, Qm

A64 : ORN Vd.16B, Vn.16B, Vm.16B

OrNot(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vorn_u8 (uint8x8_t a, uint8x8_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

OrNot(Vector64<Double>, Vector64<Double>)

float64x1_t vorn_f64 (float64x1_t a, float64x1_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

OrNot(Vector64<Int16>, Vector64<Int16>)

int16x4_t vorn_s16 (int16x4_t a, int16x4_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

OrNot(Vector64<Int32>, Vector64<Int32>)

int32x2_t vorn_s32 (int32x2_t a, int32x2_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

OrNot(Vector64<Int64>, Vector64<Int64>)

int64x1_t vorn_s64 (int64x1_t a, int64x1_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

OrNot(Vector64<SByte>, Vector64<SByte>)

int8x8_t vorn_s8 (int8x8_t a, int8x8_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

OrNot(Vector64<Single>, Vector64<Single>)

float32x2_t vorn_f32 (float32x2_t a, float32x2_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

OrNot(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vorn_u16 (uint16x4_t a, uint16x4_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

OrNot(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vorn_u32 (uint32x2_t a, uint32x2_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

OrNot(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vorn_u64 (uint64x1_t a, uint64x1_t b)

A32 : VORN Dd, Dn, Dm

A64 : ORN Vd.8B, Vn.8B, Vm.8B

PolynomialMultiply(Vector128<Byte>, Vector128<Byte>)

poly8x16_t vmulq_p8 (poly8x16_t a, poly8x16_t b)

A32 : VMUL.P8 Qd, Qn, Qm

A64 : PMUL Vd.16B, Vn.16B, Vm.16B

PolynomialMultiply(Vector128<SByte>, Vector128<SByte>)

poly8x16_t vmulq_p8 (poly8x16_t a, poly8x16_t b)

A32 : VMUL.P8 Qd, Qn, Qm

A64 : PMUL Vd.16B, Vn.16B, Vm.16B

PolynomialMultiply(Vector64<Byte>, Vector64<Byte>)

poly8x8_t vmul_p8 (poly8x8_t a, poly8x8_t b)

A32 : VMUL.P8 Dd, Dn, Dm

A64 : PMUL Vd.8B, Vn.8B, Vm.8B

PolynomialMultiply(Vector64<SByte>, Vector64<SByte>)

poly8x8_t vmul_p8 (poly8x8_t a, poly8x8_t b)

A32 : VMUL.P8 Dd, Dn, Dm

A64 : PMUL Vd.8B, Vn.8B, Vm.8B

PolynomialMultiplyWideningLower(Vector64<Byte>, Vector64<Byte>)

poly16x8_t vmull_p8 (poly8x8_t a, poly8x8_t b)

A32 : VMULL.P8 Qd, Dn, Dm

A64 : PMULL Vd.16B, Vn.8B, Vm.8B

PolynomialMultiplyWideningLower(Vector64<SByte>, Vector64<SByte>)

poly16x8_t vmull_p8 (poly8x8_t a, poly8x8_t b)

A32 : VMULL.P8 Qd, Dn, Dm

A64 : PMULL Vd.16B, Vn.8B, Vm.8B

PolynomialMultiplyWideningUpper(Vector128<Byte>, Vector128<Byte>)

poly16x8_t vmull_high_p8 (poly8x16_t a, poly8x16_t b)

A32 : VMULL.P8 Qd, Dn+1, Dm+1

A64 : PMULL2 Vd.16B, Vn.16B, Vm.16B

PolynomialMultiplyWideningUpper(Vector128<SByte>, Vector128<SByte>)

poly16x8_t vmull_high_p8 (poly8x16_t a, poly8x16_t b)

A32 : VMULL.P8 Qd, Dn+1, Dm+1

A64 : PMULL2 Vd.16B, Vn.16B, Vm.16B

PopCount(Vector128<Byte>)

uint8x16_t vcntq_u8 (uint8x16_t a)

A32 : VCNT.I8 Qd, Qm

A64 : CNT Vd.16B, Vn.16B

PopCount(Vector128<SByte>)

int8x16_t vcntq_s8 (int8x16_t a)

A32 : VCNT.I8 Qd, Qm

A64 : CNT Vd.16B, Vn.16B

PopCount(Vector64<Byte>)

uint8x8_t vcnt_u8 (uint8x8_t a)

A32 : VCNT.I8 Dd, Dm

A64 : CNT Vd.8B, Vn.8B

PopCount(Vector64<SByte>)

int8x8_t vcnt_s8 (int8x8_t a)

A32 : VCNT.I8 Dd, Dm

A64 : CNT Vd.8B, Vn.8B

ReciprocalEstimate(Vector128<Single>)

float32x4_t vrecpeq_f32 (float32x4_t a)

A32 : VRECPE.F32 Qd, Qm

A64 : FRECPE Vd.4S, Vn.4S

ReciprocalEstimate(Vector128<UInt32>)

uint32x4_t vrecpeq_u32 (uint32x4_t a)

A32 : VRECPE.U32 Qd, Qm

A64 : URECPE Vd.4S, Vn.4S

ReciprocalEstimate(Vector64<Single>)

float32x2_t vrecpe_f32 (float32x2_t a)

A32 : VRECPE.F32 Dd, Dm

A64 : FRECPE Vd.2S, Vn.2S

ReciprocalEstimate(Vector64<UInt32>)

uint32x2_t vrecpe_u32 (uint32x2_t a)

A32 : VRECPE.U32 Dd, Dm

A64 : URECPE Vd.2S, Vn.2S

ReciprocalSquareRootEstimate(Vector128<Single>)

float32x4_t vrsqrteq_f32 (float32x4_t a)

A32 : VRSQRTE.F32 Qd, Qm

A64 : FRSQRTE Vd.4S, Vn.4S

ReciprocalSquareRootEstimate(Vector128<UInt32>)

uint32x4_t vrsqrteq_u32 (uint32x4_t a)

A32 : VRSQRTE.U32 Qd, Qm

A64 : URSQRTE Vd.4S, Vn.4S

ReciprocalSquareRootEstimate(Vector64<Single>)

float32x2_t vrsqrte_f32 (float32x2_t a)

A32 : VRSQRTE.F32 Dd, Dm

A64 : FRSQRTE Vd.2S, Vn.2S

ReciprocalSquareRootEstimate(Vector64<UInt32>)

uint32x2_t vrsqrte_u32 (uint32x2_t a)

A32 : VRSQRTE.U32 Dd, Dm

A64 : URSQRTE Vd.2S, Vn.2S

ReciprocalSquareRootStep(Vector128<Single>, Vector128<Single>)

float32x4_t vrsqrtsq_f32 (float32x4_t a, float32x4_t b)

A32 : VRSQRTS.F32 Qd, Qn, Qm

A64 : FRSQRTS Vd.4S, Vn.4S, Vm.4S

ReciprocalSquareRootStep(Vector64<Single>, Vector64<Single>)

float32x2_t vrsqrts_f32 (float32x2_t a, float32x2_t b)

A32 : VRSQRTS.F32 Dd, Dn, Dm

A64 : FRSQRTS Vd.2S, Vn.2S, Vm.2S

ReciprocalStep(Vector128<Single>, Vector128<Single>)

float32x4_t vrecpsq_f32 (float32x4_t a, float32x4_t b)

A32 : VRECPS.F32 Qd, Qn, Qm

A64 : FRECPS Vd.4S, Vn.4S, Vm.4S

ReciprocalStep(Vector64<Single>, Vector64<Single>)

float32x2_t vrecps_f32 (float32x2_t a, float32x2_t b)

A32 : VRECPS.F32 Dd, Dn, Dm

A64 : FRECPS Vd.2S, Vn.2S, Vm.2S

ReverseElement16(Vector128<Int32>)

int16x8_t vrev32q_s16 (int16x8_t vec)

A32 : VREV32.16 Qd, Qm

A64 : REV32 Vd.8H, Vn.8H

ReverseElement16(Vector128<Int64>)

int16x8_t vrev64q_s16 (int16x8_t vec)

A32 : VREV64.16 Qd, Qm

A64 : REV64 Vd.8H, Vn.8H

ReverseElement16(Vector128<UInt32>)

uint16x8_t vrev32q_u16 (uint16x8_t vec)

A32 : VREV32.16 Qd, Qm

A64 : REV32 Vd.8H, Vn.8H

ReverseElement16(Vector128<UInt64>)

uint16x8_t vrev64q_u16 (uint16x8_t vec)

A32 : VREV64.16 Qd, Qm

A64 : REV64 Vd.8H, Vn.8H

ReverseElement16(Vector64<Int32>)

int16x4_t vrev32_s16 (int16x4_t vec)

A32 : VREV32.16 Dd, Dm

A64 : REV32 Vd.4H, Vn.4H

ReverseElement16(Vector64<Int64>)

int16x4_t vrev64_s16 (int16x4_t vec)

A32 : VREV64.16 Dd, Dm

A64 : REV64 Vd.4H, Vn.4H

ReverseElement16(Vector64<UInt32>)

uint16x4_t vrev32_u16 (uint16x4_t vec)

A32 : VREV32.16 Dd, Dm

A64 : REV32 Vd.4H, Vn.4H

ReverseElement16(Vector64<UInt64>)

uint16x4_t vrev64_u16 (uint16x4_t vec)

A32 : VREV64.16 Dd, Dm

A64 : REV64 Vd.4H, Vn.4H

ReverseElement32(Vector128<Int64>)

int32x4_t vrev64q_s32 (int32x4_t vec)

A32 : VREV64.32 Qd, Qm

A64 : REV64 Vd.4S, Vn.4S

ReverseElement32(Vector128<UInt64>)

uint32x4_t vrev64q_u32 (uint32x4_t vec)

A32 : VREV64.32 Qd, Qm

A64 : REV64 Vd.4S, Vn.4S

ReverseElement32(Vector64<Int64>)

int32x2_t vrev64_s32 (int32x2_t vec)

A32 : VREV64.32 Dd, Dm

A64 : REV64 Vd.2S, Vn.2S

ReverseElement32(Vector64<UInt64>)

uint32x2_t vrev64_u32 (uint32x2_t vec)

A32 : VREV64.32 Dd, Dm

A64 : REV64 Vd.2S, Vn.2S

ReverseElement8(Vector128<Int16>)

int8x16_t vrev16q_s8 (int8x16_t vec)

A32 : VREV16.8 Qd, Qm

A64 : REV16 Vd.16B, Vn.16B

ReverseElement8(Vector128<Int32>)

int8x16_t vrev32q_s8 (int8x16_t vec)

A32 : VREV32.8 Qd, Qm

A64 : REV32 Vd.16B, Vn.16B

ReverseElement8(Vector128<Int64>)

int8x16_t vrev64q_s8 (int8x16_t vec)

A32 : VREV64.8 Qd, Qm

A64 : REV64 Vd.16B, Vn.16B

ReverseElement8(Vector128<UInt16>)

uint8x16_t vrev16q_u8 (uint8x16_t vec)

A32 : VREV16.8 Qd, Qm

A64 : REV16 Vd.16B, Vn.16B

ReverseElement8(Vector128<UInt32>)

uint8x16_t vrev32q_u8 (uint8x16_t vec)

A32 : VREV32.8 Qd, Qm

A64 : REV32 Vd.16B, Vn.16B

ReverseElement8(Vector128<UInt64>)

uint8x16_t vrev64q_u8 (uint8x16_t vec)

A32 : VREV64.8 Qd, Qm

A64 : REV64 Vd.16B, Vn.16B

ReverseElement8(Vector64<Int16>)

int8x8_t vrev16_s8 (int8x8_t vec)

A32 : VREV16.8 Dd, Dm

A64 : REV16 Vd.8B, Vn.8B

ReverseElement8(Vector64<Int32>)

int8x8_t vrev32_s8 (int8x8_t vec)

A32 : VREV32.8 Dd, Dm

A64 : REV32 Vd.8B, Vn.8B

ReverseElement8(Vector64<Int64>)

int8x8_t vrev64_s8 (int8x8_t vec)

A32 : VREV64.8 Dd, Dm

A64 : REV64 Vd.8B, Vn.8B

ReverseElement8(Vector64<UInt16>)

uint8x8_t vrev16_u8 (uint8x8_t vec)

A32 : VREV16.8 Dd, Dm

A64 : REV16 Vd.8B, Vn.8B

ReverseElement8(Vector64<UInt32>)

uint8x8_t vrev32_u8 (uint8x8_t vec)

A32 : VREV32.8 Dd, Dm

A64 : REV32 Vd.8B, Vn.8B

ReverseElement8(Vector64<UInt64>)

uint8x8_t vrev64_u8 (uint8x8_t vec)

A32 : VREV64.8 Dd, Dm

A64 : REV64 Vd.8B, Vn.8B

RoundAwayFromZero(Vector128<Single>)

float32x4_t vrndaq_f32 (float32x4_t a)

A32 : VRINTA.F32 Qd, Qm

A64 : FRINTA Vd.4S, Vn.4S

RoundAwayFromZero(Vector64<Single>)

float32x2_t vrnda_f32 (float32x2_t a)

A32 : VRINTA.F32 Dd, Dm

A64 : FRINTA Vd.2S, Vn.2S

RoundAwayFromZeroScalar(Vector64<Double>)

float64x1_t vrnda_f64 (float64x1_t a)

A32 : VRINTA.F64 Dd, Dm

A64 : FRINTA Dd, Dn

RoundAwayFromZeroScalar(Vector64<Single>)

float32_t vrndas_f32 (float32_t a)

A32 : VRINTA.F32 Sd, Sm

A64 : FRINTA Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

RoundToNearest(Vector128<Single>)

float32x4_t vrndnq_f32 (float32x4_t a)

A32 : VRINTN.F32 Qd, Qm

A64 : FRINTN Vd.4S, Vn.4S

RoundToNearest(Vector64<Single>)

float32x2_t vrndn_f32 (float32x2_t a)

A32 : VRINTN.F32 Dd, Dm

A64 : FRINTN Vd.2S, Vn.2S

RoundToNearestScalar(Vector64<Double>)

float64x1_t vrndn_f64 (float64x1_t a)

A32 : VRINTN.F64 Dd, Dm

A64 : FRINTN Dd, Dn

RoundToNearestScalar(Vector64<Single>)

float32_t vrndns_f32 (float32_t a)

A32 : VRINTN.F32 Sd, Sm

A64 : FRINTN Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

RoundToNegativeInfinity(Vector128<Single>)

float32x4_t vrndmq_f32 (float32x4_t a)

A32 : VRINTM.F32 Qd, Qm

A64 : FRINTM Vd.4S, Vn.4S

RoundToNegativeInfinity(Vector64<Single>)

float32x2_t vrndm_f32 (float32x2_t a)

A32 : VRINTM.F32 Dd, Dm

A64 : FRINTM Vd.2S, Vn.2S

RoundToNegativeInfinityScalar(Vector64<Double>)

float64x1_t vrndm_f64 (float64x1_t a)

A32 : VRINTM.F64 Dd, Dm

A64 : FRINTM Dd, Dn

RoundToNegativeInfinityScalar(Vector64<Single>)

float32_t vrndms_f32 (float32_t a)

A32 : VRINTM.F32 Sd, Sm

A64 : FRINTM Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

RoundToPositiveInfinity(Vector128<Single>)

float32x4_t vrndpq_f32 (float32x4_t a)

A32 : VRINTP.F32 Qd, Qm

A64 : FRINTP Vd.4S, Vn.4S

RoundToPositiveInfinity(Vector64<Single>)

float32x2_t vrndp_f32 (float32x2_t a)

A32 : VRINTP.F32 Dd, Dm

A64 : FRINTP Vd.2S, Vn.2S

RoundToPositiveInfinityScalar(Vector64<Double>)

float64x1_t vrndp_f64 (float64x1_t a)

A32 : VRINTP.F64 Dd, Dm

A64 : FRINTP Dd, Dn

RoundToPositiveInfinityScalar(Vector64<Single>)

float32_t vrndps_f32 (float32_t a)

A32 : VRINTP.F32 Sd, Sm

A64 : FRINTP Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

RoundToZero(Vector128<Single>)

float32x4_t vrndq_f32 (float32x4_t a)

A32 : VRINTZ.F32 Qd, Qm

A64 : FRINTZ Vd.4S, Vn.4S

RoundToZero(Vector64<Single>)

float32x2_t vrnd_f32 (float32x2_t a)

A32 : VRINTZ.F32 Dd, Dm

A64 : FRINTZ Vd.2S, Vn.2S

RoundToZeroScalar(Vector64<Double>)

float64x1_t vrnd_f64 (float64x1_t a)

A32 : VRINTZ.F64 Dd, Dm

A64 : FRINTZ Dd, Dn

RoundToZeroScalar(Vector64<Single>)

float32_t vrnds_f32 (float32_t a)

A32 : VRINTZ.F32 Sd, Sm

A64 : FRINTZ Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

ShiftArithmetic(Vector128<Int16>, Vector128<Int16>)

int16x8_t vshlq_s16 (int16x8_t a, int16x8_t b)

A32 : VSHL.S16 Qd, Qn, Qm

A64 : SSHL Vd.8H, Vn.8H, Vm.8H

ShiftArithmetic(Vector128<Int32>, Vector128<Int32>)

int32x4_t vshlq_s32 (int32x4_t a, int32x4_t b)

A32 : VSHL.S32 Qd, Qn, Qm

A64 : SSHL Vd.4S, Vn.4S, Vm.4S

ShiftArithmetic(Vector128<Int64>, Vector128<Int64>)

int64x2_t vshlq_s64 (int64x2_t a, int64x2_t b)

A32 : VSHL.S64 Qd, Qn, Qm

A64 : SSHL Vd.2D, Vn.2D, Vm.2D

ShiftArithmetic(Vector128<SByte>, Vector128<SByte>)

int8x16_t vshlq_s8 (int8x16_t a, int8x16_t b)

A32 : VSHL.S8 Qd, Qn, Qm

A64 : SSHL Vd.16B, Vn.16B, Vm.16B

ShiftArithmetic(Vector64<Int16>, Vector64<Int16>)

int16x4_t vshl_s16 (int16x4_t a, int16x4_t b)

A32 : VSHL.S16 Dd, Dn, Dm

A64 : SSHL Vd.4H, Vn.4H, Vm.4H

ShiftArithmetic(Vector64<Int32>, Vector64<Int32>)

int32x2_t vshl_s32 (int32x2_t a, int32x2_t b)

A32 : VSHL.S32 Dd, Dn, Dm

A64 : SSHL Vd.2S, Vn.2S, Vm.2S

ShiftArithmetic(Vector64<SByte>, Vector64<SByte>)

int8x8_t vshl_s8 (int8x8_t a, int8x8_t b)

A32 : VSHL.S8 Dd, Dn, Dm

A64 : SSHL Vd.8B, Vn.8B, Vm.8B

ShiftArithmeticRounded(Vector128<Int16>, Vector128<Int16>)

int16x8_t vrshlq_s16 (int16x8_t a, int16x8_t b)

A32 : VRSHL.S16 Qd, Qn, Qm

A64 : SRSHL Vd.8H, Vn.8H, Vm.8H

ShiftArithmeticRounded(Vector128<Int32>, Vector128<Int32>)

int32x4_t vrshlq_s32 (int32x4_t a, int32x4_t b)

A32 : VRSHL.S32 Qd, Qn, Qm

A64 : SRSHL Vd.4S, Vn.4S, Vm.4S

ShiftArithmeticRounded(Vector128<Int64>, Vector128<Int64>)

int64x2_t vrshlq_s64 (int64x2_t a, int64x2_t b)

A32 : VRSHL.S64 Qd, Qn, Qm

A64 : SRSHL Vd.2D, Vn.2D, Vm.2D

ShiftArithmeticRounded(Vector128<SByte>, Vector128<SByte>)

int8x16_t vrshlq_s8 (int8x16_t a, int8x16_t b)

A32 : VRSHL.S8 Qd, Qn, Qm

A64 : SRSHL Vd.16B, Vn.16B, Vm.16B

ShiftArithmeticRounded(Vector64<Int16>, Vector64<Int16>)

int16x4_t vrshl_s16 (int16x4_t a, int16x4_t b)

A32 : VRSHL.S16 Dd, Dn, Dm

A64 : SRSHL Vd.4H, Vn.4H, Vm.4H

ShiftArithmeticRounded(Vector64<Int32>, Vector64<Int32>)

int32x2_t vrshl_s32 (int32x2_t a, int32x2_t b)

A32 : VRSHL.S32 Dd, Dn, Dm

A64 : SRSHL Vd.2S, Vn.2S, Vm.2S

ShiftArithmeticRounded(Vector64<SByte>, Vector64<SByte>)

int8x8_t vrshl_s8 (int8x8_t a, int8x8_t b)

A32 : VRSHL.S8 Dd, Dn, Dm

A64 : SRSHL Vd.8B, Vn.8B, Vm.8B

ShiftArithmeticRoundedSaturate(Vector128<Int16>, Vector128<Int16>)

int16x8_t vqrshlq_s16 (int16x8_t a, int16x8_t b)

A32 : VQRSHL.S16 Qd, Qn, Qm

A64 : SQRSHL Vd.8H, Vn.8H, Vm.8H

ShiftArithmeticRoundedSaturate(Vector128<Int32>, Vector128<Int32>)

int32x4_t vqrshlq_s32 (int32x4_t a, int32x4_t b)

A32 : VQRSHL.S32 Qd, Qn, Qm

A64 : SQRSHL Vd.4S, Vn.4S, Vm.4S

ShiftArithmeticRoundedSaturate(Vector128<Int64>, Vector128<Int64>)

int64x2_t vqrshlq_s64 (int64x2_t a, int64x2_t b)

A32 : VQRSHL.S64 Qd, Qn, Qm

A64 : SQRSHL Vd.2D, Vn.2D, Vm.2D

ShiftArithmeticRoundedSaturate(Vector128<SByte>, Vector128<SByte>)

int8x16_t vqrshlq_s8 (int8x16_t a, int8x16_t b)

A32 : VQRSHL.S8 Qd, Qn, Qm

A64 : SQRSHL Vd.16B, Vn.16B, Vm.16B

ShiftArithmeticRoundedSaturate(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqrshl_s16 (int16x4_t a, int16x4_t b)

A32 : VQRSHL.S16 Dd, Dn, Dm

A64 : SQRSHL Vd.4H, Vn.4H, Vm.4H

ShiftArithmeticRoundedSaturate(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqrshl_s32 (int32x2_t a, int32x2_t b)

A32 : VQRSHL.S32 Dd, Dn, Dm

A64 : SQRSHL Vd.2S, Vn.2S, Vm.2S

ShiftArithmeticRoundedSaturate(Vector64<SByte>, Vector64<SByte>)

int8x8_t vqrshl_s8 (int8x8_t a, int8x8_t b)

A32 : VQRSHL.S8 Dd, Dn, Dm

A64 : SQRSHL Vd.8B, Vn.8B, Vm.8B

ShiftArithmeticRoundedSaturateScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vqrshl_s64 (int64x1_t a, int64x1_t b)

A32 : VQRSHL.S64 Dd, Dn, Dm

A64 : SQRSHL Dd, Dn, Dm

ShiftArithmeticRoundedScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vrshl_s64 (int64x1_t a, int64x1_t b)

A32 : VRSHL.S64 Dd, Dn, Dm

A64 : SRSHL Dd, Dn, Dm

ShiftArithmeticSaturate(Vector128<Int16>, Vector128<Int16>)

int16x8_t vqshlq_s16 (int16x8_t a, int16x8_t b)

A32 : VQSHL.S16 Qd, Qn, Qm

A64 : SQSHL Vd.8H, Vn.8H, Vm.8H

ShiftArithmeticSaturate(Vector128<Int32>, Vector128<Int32>)

int32x4_t vqshlq_s32 (int32x4_t a, int32x4_t b)

A32 : VQSHL.S32 Qd, Qn, Qm

A64 : SQSHL Vd.4S, Vn.4S, Vm.4S

ShiftArithmeticSaturate(Vector128<Int64>, Vector128<Int64>)

int64x2_t vqshlq_s64 (int64x2_t a, int64x2_t b)

A32 : VQSHL.S64 Qd, Qn, Qm

A64 : SQSHL Vd.2D, Vn.2D, Vm.2D

ShiftArithmeticSaturate(Vector128<SByte>, Vector128<SByte>)

int8x16_t vqshlq_s8 (int8x16_t a, int8x16_t b)

A32 : VQSHL.S8 Qd, Qn, Qm

A64 : SQSHL Vd.16B, Vn.16B, Vm.16B

ShiftArithmeticSaturate(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqshl_s16 (int16x4_t a, int16x4_t b)

A32 : VQSHL.S16 Dd, Dn, Dm

A64 : SQSHL Vd.4H, Vn.4H, Vm.4H

ShiftArithmeticSaturate(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqshl_s32 (int32x2_t a, int32x2_t b)

A32 : VQSHL.S32 Dd, Dn, Dm

A64 : SQSHL Vd.2S, Vn.2S, Vm.2S

ShiftArithmeticSaturate(Vector64<SByte>, Vector64<SByte>)

int8x8_t vqshl_s8 (int8x8_t a, int8x8_t b)

A32 : VQSHL.S8 Dd, Dn, Dm

A64 : SQSHL Vd.8B, Vn.8B, Vm.8B

ShiftArithmeticSaturateScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vqshl_s64 (int64x1_t a, int64x1_t b)

A32 : VQSHL.S64 Dd, Dn, Dm

A64 : SQSHL Dd, Dn, Dm

ShiftArithmeticScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vshl_s64 (int64x1_t a, int64x1_t b)

A32 : VSHL.S64 Dd, Dn, Dm

A64 : SSHL Dd, Dn, Dm

ShiftLeftAndInsert(Vector128<Byte>, Vector128<Byte>, Byte)

uint8x16_t vsliq_n_u8(uint8x16_t a, uint8x16_t b, __builtin_constant_p(n))

A32 : VSLI.8 Qd, Qm, #n

A64 : SLI Vd.16B, Vn.16B, #n

ShiftLeftAndInsert(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vsliq_n_s16(int16x8_t a, int16x8_t b, __builtin_constant_p(n))

A32 : VSLI.16 Qd, Qm, #n

A64 : SLI Vd.8H, Vn.8H, #n

ShiftLeftAndInsert(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vsliq_n_s32(int32x4_t a, int32x4_t b, __builtin_constant_p(n))

A32 : VSLI.32 Qd, Qm, #n

A64 : SLI Vd.4S, Vn.4S, #n

ShiftLeftAndInsert(Vector128<Int64>, Vector128<Int64>, Byte)

int64x2_t vsliq_n_s64(int64x2_t a, int64x2_t b, __builtin_constant_p(n))

A32 : VSLI.64 Qd, Qm, #n

A64 : SLI Vd.2D, Vn.2D, #n

ShiftLeftAndInsert(Vector128<SByte>, Vector128<SByte>, Byte)

int8x16_t vsliq_n_s8(int8x16_t a, int8x16_t b, __builtin_constant_p(n))

A32 : VSLI.8 Qd, Qm, #n

A64 : SLI Vd.16B, Vn.16B, #n

ShiftLeftAndInsert(Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vsliq_n_u16(uint16x8_t a, uint16x8_t b, __builtin_constant_p(n))

A32 : VSLI.16 Qd, Qm, #n

A64 : SLI Vd.8H, Vn.8H, #n

ShiftLeftAndInsert(Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vsliq_n_u32(uint32x4_t a, uint32x4_t b, __builtin_constant_p(n))

A32 : VSLI.32 Qd, Qm, #n

A64 : SLI Vd.4S, Vn.4S, #n

ShiftLeftAndInsert(Vector128<UInt64>, Vector128<UInt64>, Byte)

uint64x2_t vsliq_n_u64(uint64x2_t a, uint64x2_t b, __builtin_constant_p(n))

A32 : VSLI.64 Qd, Qm, #n

A64 : SLI Vd.2D, Vn.2D, #n

ShiftLeftAndInsert(Vector64<Byte>, Vector64<Byte>, Byte)

uint8x8_t vsli_n_u8(uint8x8_t a, uint8x8_t b, __builtin_constant_p(n))

A32 : VSLI.8 Dd, Dm, #n

A64 : SLI Vd.8B, Vn.8B, #n

ShiftLeftAndInsert(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vsli_n_s16(int16x4_t a, int16x4_t b, __builtin_constant_p(n))

A32 : VSLI.16 Dd, Dm, #n

A64 : SLI Vd.4H, Vn.4H, #n

ShiftLeftAndInsert(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vsli_n_s32(int32x2_t a, int32x2_t b, __builtin_constant_p(n))

A32 : VSLI.32 Dd, Dm, #n

A64 : SLI Vd.2S, Vn.2S, #n

ShiftLeftAndInsert(Vector64<SByte>, Vector64<SByte>, Byte)

int8x8_t vsli_n_s8(int8x8_t a, int8x8_t b, __builtin_constant_p(n))

A32 : VSLI.8 Dd, Dm, #n

A64 : SLI Vd.8B, Vn.8B, #n

ShiftLeftAndInsert(Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vsli_n_u16(uint16x4_t a, uint16x4_t b, __builtin_constant_p(n))

A32 : VSLI.16 Dd, Dm, #n

A64 : SLI Vd.4H, Vn.4H, #n

ShiftLeftAndInsert(Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vsli_n_u32(uint32x2_t a, uint32x2_t b, __builtin_constant_p(n))

A32 : VSLI.32 Dd, Dm, #n

A64 : SLI Vd.2S, Vn.2S, #n

ShiftLeftAndInsertScalar(Vector64<Int64>, Vector64<Int64>, Byte)

int64_t vslid_n_s64(int64_t a, int64_t b, __builtin_constant_p(n))

A32 : VSLI.64 Dd, Dm, #n

A64 : SLI Dd, Dn, #n

ShiftLeftAndInsertScalar(Vector64<UInt64>, Vector64<UInt64>, Byte)

uint64_t vslid_n_u64(uint64_t a, uint64_t b, __builtin_constant_p(n))

A32 : VSLI.64 Dd, Dm, #n

A64 : SLI Dd, Dn, #n

ShiftLeftLogical(Vector128<Byte>, Byte)

uint8x16_t vshlq_n_u8 (uint8x16_t a, const int n)

A32 : VSHL.I8 Qd, Qm, #n

A64 : SHL Vd.16B, Vn.16B, #n

ShiftLeftLogical(Vector128<Int16>, Byte)

int16x8_t vshlq_n_s16 (int16x8_t a, const int n)

A32 : VSHL.I16 Qd, Qm, #n

A64 : SHL Vd.8H, Vn.8H, #n

ShiftLeftLogical(Vector128<Int64>, Byte)

int64x2_t vshlq_n_s64 (int64x2_t a, const int n)

A32 : VSHL.I64 Qd, Qm, #n

A64 : SHL Vd.2D, Vn.2D, #n

ShiftLeftLogical(Vector128<SByte>, Byte)

int8x16_t vshlq_n_s8 (int8x16_t a, const int n)

A32 : VSHL.I8 Qd, Qm, #n

A64 : SHL Vd.16B, Vn.16B, #n

ShiftLeftLogical(Vector128<UInt16>, Byte)

uint16x8_t vshlq_n_u16 (uint16x8_t a, const int n)

A32 : VSHL.I16 Qd, Qm, #n

A64 : SHL Vd.8H, Vn.8H, #n

ShiftLeftLogical(Vector128<UInt32>, Byte)

uint32x4_t vshlq_n_u32 (uint32x4_t a, const int n)

A32 : VSHL.I32 Qd, Qm, #n

A64 : SHL Vd.4S, Vn.4S, #n

ShiftLeftLogical(Vector128<UInt64>, Byte)

uint64x2_t vshlq_n_u64 (uint64x2_t a, const int n)

A32 : VSHL.I64 Qd, Qm, #n

A64 : SHL Vd.2D, Vn.2D, #n

ShiftLeftLogical(Vector64<Byte>, Byte)

uint8x8_t vshl_n_u8 (uint8x8_t a, const int n)

A32 : VSHL.I8 Dd, Dm, #n

A64 : SHL Vd.8B, Vn.8B, #n

ShiftLeftLogical(Vector64<Int16>, Byte)

int16x4_t vshl_n_s16 (int16x4_t a, const int n)

A32 : VSHL.I16 Dd, Dm, #n

A64 : SHL Vd.4H, Vn.4H, #n

ShiftLeftLogical(Vector64<Int32>, Byte)

int32x2_t vshl_n_s32 (int32x2_t a, const int n)

A32 : VSHL.I32 Dd, Dm, #n

A64 : SHL Vd.2S, Vn.2S, #n

ShiftLeftLogical(Vector64<SByte>, Byte)

int8x8_t vshl_n_s8 (int8x8_t a, const int n)

A32 : VSHL.I8 Dd, Dm, #n

A64 : SHL Vd.8B, Vn.8B, #n

ShiftLeftLogical(Vector64<UInt16>, Byte)

uint16x4_t vshl_n_u16 (uint16x4_t a, const int n)

A32 : VSHL.I16 Dd, Dm, #n

A64 : SHL Vd.4H, Vn.4H, #n

ShiftLeftLogical(Vector64<UInt32>, Byte)

uint32x2_t vshl_n_u32 (uint32x2_t a, const int n)

A32 : VSHL.I32 Dd, Dm, #n

A64 : SHL Vd.2S, Vn.2S, #n

ShiftLeftLogicalSaturate(Vector128<Byte>, Byte)

uint8x16_t vqshlq_n_u8 (uint8x16_t a, const int n)

A32 : VQSHL.U8 Qd, Qm, #n

A64 : UQSHL Vd.16B, Vn.16B, #n

ShiftLeftLogicalSaturate(Vector128<Int16>, Byte)

int16x8_t vqshlq_n_s16 (int16x8_t a, const int n)

A32 : VQSHL.S16 Qd, Qm, #n

A64 : SQSHL Vd.8H, Vn.8H, #n

ShiftLeftLogicalSaturate(Vector128<Int32>, Byte)

int32x4_t vqshlq_n_s32 (int32x4_t a, const int n)

A32 : VQSHL.S32 Qd, Qm, #n

A64 : SQSHL Vd.4S, Vn.4S, #n

ShiftLeftLogicalSaturate(Vector128<Int64>, Byte)

int64x2_t vqshlq_n_s64 (int64x2_t a, const int n)

A32 : VQSHL.S64 Qd, Qm, #n

A64 : SQSHL Vd.2D, Vn.2D, #n

ShiftLeftLogicalSaturate(Vector128<SByte>, Byte)

int8x16_t vqshlq_n_s8 (int8x16_t a, const int n)

A32 : VQSHL.S8 Qd, Qm, #n

A64 : SQSHL Vd.16B, Vn.16B, #n

ShiftLeftLogicalSaturate(Vector128<UInt16>, Byte)

uint16x8_t vqshlq_n_u16 (uint16x8_t a, const int n)

A32 : VQSHL.U16 Qd, Qm, #n

A64 : UQSHL Vd.8H, Vn.8H, #n

ShiftLeftLogicalSaturate(Vector128<UInt32>, Byte)

uint32x4_t vqshlq_n_u32 (uint32x4_t a, const int n)

A32 : VQSHL.U32 Qd, Qm, #n

A64 : UQSHL Vd.4S, Vn.4S, #n

ShiftLeftLogicalSaturate(Vector128<UInt64>, Byte)

uint64x2_t vqshlq_n_u64 (uint64x2_t a, const int n)

A32 : VQSHL.U64 Qd, Qm, #n

A64 : UQSHL Vd.2D, Vn.2D, #n

ShiftLeftLogicalSaturate(Vector64<Byte>, Byte)

uint8x8_t vqshl_n_u8 (uint8x8_t a, const int n)

A32 : VQSHL.U8 Dd, Dm, #n

A64 : UQSHL Vd.8B, Vn.8B, #n

ShiftLeftLogicalSaturate(Vector64<Int16>, Byte)

int16x4_t vqshl_n_s16 (int16x4_t a, const int n)

A32 : VQSHL.S16 Dd, Dm, #n

A64 : SQSHL Vd.4H, Vn.4H, #n

ShiftLeftLogicalSaturate(Vector64<Int32>, Byte)

int32x2_t vqshl_n_s32 (int32x2_t a, const int n)

A32 : VQSHL.S32 Dd, Dm, #n

A64 : SQSHL Vd.2S, Vn.2S, #n

ShiftLeftLogicalSaturate(Vector64<SByte>, Byte)

int8x8_t vqshl_n_s8 (int8x8_t a, const int n)

A32 : VQSHL.S8 Dd, Dm, #n

A64 : SQSHL Vd.8B, Vn.8B, #n

ShiftLeftLogicalSaturate(Vector64<UInt16>, Byte)

uint16x4_t vqshl_n_u16 (uint16x4_t a, const int n)

A32 : VQSHL.U16 Dd, Dm, #n

A64 : UQSHL Vd.4H, Vn.4H, #n

ShiftLeftLogicalSaturate(Vector64<UInt32>, Byte)

uint32x2_t vqshl_n_u32 (uint32x2_t a, const int n)

A32 : VQSHL.U32 Dd, Dm, #n

A64 : UQSHL Vd.2S, Vn.2S, #n

ShiftLeftLogicalSaturateScalar(Vector64<Int64>, Byte)

int64x1_t vqshl_n_s64 (int64x1_t a, const int n)

A32 : VQSHL.S64 Dd, Dm, #n

A64 : SQSHL Dd, Dn, #n

ShiftLeftLogicalSaturateScalar(Vector64<UInt64>, Byte)

uint64x1_t vqshl_n_u64 (uint64x1_t a, const int n)

A32 : VQSHL.U64 Dd, Dm, #n

A64 : UQSHL Dd, Dn, #n

ShiftLeftLogicalSaturateUnsigned(Vector128<Int16>, Byte)

uint16x8_t vqshluq_n_s16 (int16x8_t a, const int n)

A32 : VQSHLU.S16 Qd, Qm, #n

A64 : SQSHLU Vd.8H, Vn.8H, #n

ShiftLeftLogicalSaturateUnsigned(Vector128<Int32>, Byte)

uint32x4_t vqshluq_n_s32 (int32x4_t a, const int n)

A32 : VQSHLU.S32 Qd, Qm, #n

A64 : SQSHLU Vd.4S, Vn.4S, #n

ShiftLeftLogicalSaturateUnsigned(Vector128<Int64>, Byte)

uint64x2_t vqshluq_n_s64 (int64x2_t a, const int n)

A32 : VQSHLU.S64 Qd, Qm, #n

A64 : SQSHLU Vd.2D, Vn.2D, #n

ShiftLeftLogicalSaturateUnsigned(Vector128<SByte>, Byte)

uint8x16_t vqshluq_n_s8 (int8x16_t a, const int n)

A32 : VQSHLU.S8 Qd, Qm, #n

A64 : SQSHLU Vd.16B, Vn.16B, #n

ShiftLeftLogicalSaturateUnsigned(Vector64<Int16>, Byte)

uint16x4_t vqshlu_n_s16 (int16x4_t a, const int n)

A32 : VQSHLU.S16 Dd, Dm, #n

A64 : SQSHLU Vd.4H, Vn.4H, #n

ShiftLeftLogicalSaturateUnsigned(Vector64<Int32>, Byte)

uint32x2_t vqshlu_n_s32 (int32x2_t a, const int n)

A32 : VQSHLU.S32 Dd, Dm, #n

A64 : SQSHLU Vd.2S, Vn.2S, #n

ShiftLeftLogicalSaturateUnsigned(Vector64<SByte>, Byte)

uint8x8_t vqshlu_n_s8 (int8x8_t a, const int n)

A32 : VQSHLU.S8 Dd, Dm, #n

A64 : SQSHLU Vd.8B, Vn.8B, #n

ShiftLeftLogicalSaturateUnsignedScalar(Vector64<Int64>, Byte)

uint64x1_t vqshlu_n_s64 (int64x1_t a, const int n)

A32 : VQSHLU.S64 Dd, Dm, #n

A64 : SQSHLU Dd, Dn, #n

ShiftLeftLogicalScalar(Vector64<Int64>, Byte)

int64x1_t vshl_n_s64 (int64x1_t a, const int n)

A32 : VSHL.I64 Dd, Dm, #n

A64 : SHL Dd, Dn, #n

ShiftLeftLogicalScalar(Vector64<UInt64>, Byte)

uint64x1_t vshl_n_u64 (uint64x1_t a, const int n)

A32 : VSHL.I64 Dd, Dm, #n

A64 : SHL Dd, Dn, #n

ShiftLeftLogicalWideningLower(Vector64<Byte>, Byte)

uint16x8_t vshll_n_u8 (uint8x8_t a, const int n)

A32 : VSHLL.U8 Qd, Dm, #n

A64 : USHLL Vd.8H, Vn.8B, #n

ShiftLeftLogicalWideningLower(Vector64<Int16>, Byte)

int32x4_t vshll_n_s16 (int16x4_t a, const int n)

A32 : VSHLL.S16 Qd, Dm, #n

A64 : SSHLL Vd.4S, Vn.4H, #n

ShiftLeftLogicalWideningLower(Vector64<Int32>, Byte)

int64x2_t vshll_n_s32 (int32x2_t a, const int n)

A32 : VSHLL.S32 Qd, Dm, #n

A64 : SSHLL Vd.2D, Vn.2S, #n

ShiftLeftLogicalWideningLower(Vector64<SByte>, Byte)

int16x8_t vshll_n_s8 (int8x8_t a, const int n)

A32 : VSHLL.S8 Qd, Dm, #n

A64 : SSHLL Vd.8H, Vn.8B, #n

ShiftLeftLogicalWideningLower(Vector64<UInt16>, Byte)

uint32x4_t vshll_n_u16 (uint16x4_t a, const int n)

A32 : VSHLL.U16 Qd, Dm, #n

A64 : USHLL Vd.4S, Vn.4H, #n

ShiftLeftLogicalWideningLower(Vector64<UInt32>, Byte)

uint64x2_t vshll_n_u32 (uint32x2_t a, const int n)

A32 : VSHLL.U32 Qd, Dm, #n

A64 : USHLL Vd.2D, Vn.2S, #n

ShiftLeftLogicalWideningUpper(Vector128<Byte>, Byte)

uint16x8_t vshll_high_n_u8 (uint8x16_t a, const int n)

A32 : VSHLL.U8 Qd, Dm+1, #n

A64 : USHLL2 Vd.8H, Vn.16B, #n

ShiftLeftLogicalWideningUpper(Vector128<Int16>, Byte)

int32x4_t vshll_high_n_s16 (int16x8_t a, const int n)

A32 : VSHLL.S16 Qd, Dm+1, #n

A64 : SSHLL2 Vd.4S, Vn.8H, #n

ShiftLeftLogicalWideningUpper(Vector128<Int32>, Byte)

int64x2_t vshll_high_n_s32 (int32x4_t a, const int n)

A32 : VSHLL.S32 Qd, Dm+1, #n

A64 : SSHLL2 Vd.2D, Vn.4S, #n

ShiftLeftLogicalWideningUpper(Vector128<SByte>, Byte)

int16x8_t vshll_high_n_s8 (int8x16_t a, const int n)

A32 : VSHLL.S8 Qd, Dm+1, #n

A64 : SSHLL2 Vd.8H, Vn.16B, #n

ShiftLeftLogicalWideningUpper(Vector128<UInt16>, Byte)

uint32x4_t vshll_high_n_u16 (uint16x8_t a, const int n)

A32 : VSHLL.U16 Qd, Dm+1, #n

A64 : USHLL2 Vd.4S, Vn.8H, #n

ShiftLeftLogicalWideningUpper(Vector128<UInt32>, Byte)

uint64x2_t vshll_high_n_u32 (uint32x4_t a, const int n)

A32 : VSHLL.U32 Qd, Dm+1, #n

A64 : USHLL2 Vd.2D, Vn.4S, #n

ShiftLogical(Vector128<Byte>, Vector128<SByte>)

uint8x16_t vshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VSHL.U8 Qd, Qn, Qm

A64 : USHL Vd.16B, Vn.16B, Vm.16B

ShiftLogical(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VSHL.U16 Qd, Qn, Qm

A64 : USHL Vd.8H, Vn.8H, Vm.8H

ShiftLogical(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VSHL.U32 Qd, Qn, Qm

A64 : USHL Vd.4S, Vn.4S, Vm.4S

ShiftLogical(Vector128<Int64>, Vector128<Int64>)

uint64x2_t vshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VSHL.U64 Qd, Qn, Qm

A64 : USHL Vd.2D, Vn.2D, Vm.2D

ShiftLogical(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VSHL.U8 Qd, Qn, Qm

A64 : USHL Vd.16B, Vn.16B, Vm.16B

ShiftLogical(Vector128<UInt16>, Vector128<Int16>)

uint16x8_t vshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VSHL.U16 Qd, Qn, Qm

A64 : USHL Vd.8H, Vn.8H, Vm.8H

ShiftLogical(Vector128<UInt32>, Vector128<Int32>)

uint32x4_t vshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VSHL.U32 Qd, Qn, Qm

A64 : USHL Vd.4S, Vn.4S, Vm.4S

ShiftLogical(Vector128<UInt64>, Vector128<Int64>)

uint64x2_t vshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VSHL.U64 Qd, Qn, Qm

A64 : USHL Vd.2D, Vn.2D, Vm.2D

ShiftLogical(Vector64<Byte>, Vector64<SByte>)

uint8x8_t vshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VSHL.U8 Dd, Dn, Dm

A64 : USHL Vd.8B, Vn.8B, Vm.8B

ShiftLogical(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VSHL.U16 Dd, Dn, Dm

A64 : USHL Vd.4H, Vn.4H, Vm.4H

ShiftLogical(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VSHL.U32 Dd, Dn, Dm

A64 : USHL Vd.2S, Vn.2S, Vm.2S

ShiftLogical(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VSHL.U8 Dd, Dn, Dm

A64 : USHL Vd.8B, Vn.8B, Vm.8B

ShiftLogical(Vector64<UInt16>, Vector64<Int16>)

uint16x4_t vshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VSHL.U16 Dd, Dn, Dm

A64 : USHL Vd.4H, Vn.4H, Vm.4H

ShiftLogical(Vector64<UInt32>, Vector64<Int32>)

uint32x2_t vshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VSHL.U32 Dd, Dn, Dm

A64 : USHL Vd.2S, Vn.2S, Vm.2S

ShiftLogicalRounded(Vector128<Byte>, Vector128<SByte>)

uint8x16_t vrshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VRSHL.U8 Qd, Qn, Qm

A64 : URSHL Vd.16B, Vn.16B, Vm.16B

ShiftLogicalRounded(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vrshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VRSHL.U16 Qd, Qn, Qm

A64 : URSHL Vd.8H, Vn.8H, Vm.8H

ShiftLogicalRounded(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vrshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VRSHL.U32 Qd, Qn, Qm

A64 : URSHL Vd.4S, Vn.4S, Vm.4S

ShiftLogicalRounded(Vector128<Int64>, Vector128<Int64>)

uint64x2_t vrshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VRSHL.U64 Qd, Qn, Qm

A64 : URSHL Vd.2D, Vn.2D, Vm.2D

ShiftLogicalRounded(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vrshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VRSHL.U8 Qd, Qn, Qm

A64 : URSHL Vd.16B, Vn.16B, Vm.16B

ShiftLogicalRounded(Vector128<UInt16>, Vector128<Int16>)

uint16x8_t vrshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VRSHL.U16 Qd, Qn, Qm

A64 : URSHL Vd.8H, Vn.8H, Vm.8H

ShiftLogicalRounded(Vector128<UInt32>, Vector128<Int32>)

uint32x4_t vrshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VRSHL.U32 Qd, Qn, Qm

A64 : URSHL Vd.4S, Vn.4S, Vm.4S

ShiftLogicalRounded(Vector128<UInt64>, Vector128<Int64>)

uint64x2_t vrshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VRSHL.U64 Qd, Qn, Qm

A64 : URSHL Vd.2D, Vn.2D, Vm.2D

ShiftLogicalRounded(Vector64<Byte>, Vector64<SByte>)

uint8x8_t vrshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VRSHL.U8 Dd, Dn, Dm

A64 : URSHL Vd.8B, Vn.8B, Vm.8B

ShiftLogicalRounded(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vrshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VRSHL.U16 Dd, Dn, Dm

A64 : URSHL Vd.4H, Vn.4H, Vm.4H

ShiftLogicalRounded(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vrshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VRSHL.U32 Dd, Dn, Dm

A64 : URSHL Vd.2S, Vn.2S, Vm.2S

ShiftLogicalRounded(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vrshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VRSHL.U8 Dd, Dn, Dm

A64 : URSHL Vd.8B, Vn.8B, Vm.8B

ShiftLogicalRounded(Vector64<UInt16>, Vector64<Int16>)

uint16x4_t vrshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VRSHL.U16 Dd, Dn, Dm

A64 : URSHL Vd.4H, Vn.4H, Vm.4H

ShiftLogicalRounded(Vector64<UInt32>, Vector64<Int32>)

uint32x2_t vrshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VRSHL.U32 Dd, Dn, Dm

A64 : URSHL Vd.2S, Vn.2S, Vm.2S

ShiftLogicalRoundedSaturate(Vector128<Byte>, Vector128<SByte>)

uint8x16_t vqrshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VQRSHL.U8 Qd, Qn, Qm

A64 : UQRSHL Vd.16B, Vn.16B, Vm.16B

ShiftLogicalRoundedSaturate(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vqrshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VQRSHL.U16 Qd, Qn, Qm

A64 : UQRSHL Vd.8H, Vn.8H, Vm.8H

ShiftLogicalRoundedSaturate(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vqrshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VQRSHL.U32 Qd, Qn, Qm

A64 : UQRSHL Vd.4S, Vn.4S, Vm.4S

ShiftLogicalRoundedSaturate(Vector128<Int64>, Vector128<Int64>)

uint64x2_t vqrshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VQRSHL.U64 Qd, Qn, Qm

A64 : UQRSHL Vd.2D, Vn.2D, Vm.2D

ShiftLogicalRoundedSaturate(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vqrshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VQRSHL.U8 Qd, Qn, Qm

A64 : UQRSHL Vd.16B, Vn.16B, Vm.16B

ShiftLogicalRoundedSaturate(Vector128<UInt16>, Vector128<Int16>)

uint16x8_t vqrshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VQRSHL.U16 Qd, Qn, Qm

A64 : UQRSHL Vd.8H, Vn.8H, Vm.8H

ShiftLogicalRoundedSaturate(Vector128<UInt32>, Vector128<Int32>)

uint32x4_t vqrshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VQRSHL.U32 Qd, Qn, Qm

A64 : UQRSHL Vd.4S, Vn.4S, Vm.4S

ShiftLogicalRoundedSaturate(Vector128<UInt64>, Vector128<Int64>)

uint64x2_t vqrshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VQRSHL.U64 Qd, Qn, Qm

A64 : UQRSHL Vd.2D, Vn.2D, Vm.2D

ShiftLogicalRoundedSaturate(Vector64<Byte>, Vector64<SByte>)

uint8x8_t vqrshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VQRSHL.U8 Dd, Dn, Dm

A64 : UQRSHL Vd.8B, Vn.8B, Vm.8B

ShiftLogicalRoundedSaturate(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vqrshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VQRSHL.U16 Dd, Dn, Dm

A64 : UQRSHL Vd.4H, Vn.4H, Vm.4H

ShiftLogicalRoundedSaturate(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vqrshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VQRSHL.U32 Dd, Dn, Dm

A64 : UQRSHL Vd.2S, Vn.2S, Vm.2S

ShiftLogicalRoundedSaturate(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vqrshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VQRSHL.U8 Dd, Dn, Dm

A64 : UQRSHL Vd.8B, Vn.8B, Vm.8B

ShiftLogicalRoundedSaturate(Vector64<UInt16>, Vector64<Int16>)

uint16x4_t vqrshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VQRSHL.U16 Dd, Dn, Dm

A64 : UQRSHL Vd.4H, Vn.4H, Vm.4H

ShiftLogicalRoundedSaturate(Vector64<UInt32>, Vector64<Int32>)

uint32x2_t vqrshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VQRSHL.U32 Dd, Dn, Dm

A64 : UQRSHL Vd.2S, Vn.2S, Vm.2S

ShiftLogicalRoundedSaturateScalar(Vector64<Int64>, Vector64<Int64>)

uint64x1_t vqrshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VQRSHL.U64 Dd, Dn, Dm

A64 : UQRSHL Dd, Dn, Dm

ShiftLogicalRoundedSaturateScalar(Vector64<UInt64>, Vector64<Int64>)

uint64x1_t vqrshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VQRSHL.U64 Dd, Dn, Dm

A64 : UQRSHL Dd, Dn, Dm

ShiftLogicalRoundedScalar(Vector64<Int64>, Vector64<Int64>)

uint64x1_t vrshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VRSHL.U64 Dd, Dn, Dm

A64 : URSHL Dd, Dn, Dm

ShiftLogicalRoundedScalar(Vector64<UInt64>, Vector64<Int64>)

uint64x1_t vrshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VRSHL.U64 Dd, Dn, Dm

A64 : URSHL Dd, Dn, Dm

ShiftLogicalSaturate(Vector128<Byte>, Vector128<SByte>)

uint8x16_t vqshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VQSHL.U8 Qd, Qn, Qm

A64 : UQSHL Vd.16B, Vn.16B, Vm.16B

ShiftLogicalSaturate(Vector128<Int16>, Vector128<Int16>)

uint16x8_t vqshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VQSHL.U16 Qd, Qn, Qm

A64 : UQSHL Vd.8H, Vn.8H, Vm.8H

ShiftLogicalSaturate(Vector128<Int32>, Vector128<Int32>)

uint32x4_t vqshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VQSHL.U32 Qd, Qn, Qm

A64 : UQSHL Vd.4S, Vn.4S, Vm.4S

ShiftLogicalSaturate(Vector128<Int64>, Vector128<Int64>)

uint64x2_t vqshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VQSHL.U64 Qd, Qn, Qm

A64 : UQSHL Vd.2D, Vn.2D, Vm.2D

ShiftLogicalSaturate(Vector128<SByte>, Vector128<SByte>)

uint8x16_t vqshlq_u8 (uint8x16_t a, int8x16_t b)

A32 : VQSHL.U8 Qd, Qn, Qm

A64 : UQSHL Vd.16B, Vn.16B, Vm.16B

ShiftLogicalSaturate(Vector128<UInt16>, Vector128<Int16>)

uint16x8_t vqshlq_u16 (uint16x8_t a, int16x8_t b)

A32 : VQSHL.U16 Qd, Qn, Qm

A64 : UQSHL Vd.8H, Vn.8H, Vm.8H

ShiftLogicalSaturate(Vector128<UInt32>, Vector128<Int32>)

uint32x4_t vqshlq_u32 (uint32x4_t a, int32x4_t b)

A32 : VQSHL.U32 Qd, Qn, Qm

A64 : UQSHL Vd.4S, Vn.4S, Vm.4S

ShiftLogicalSaturate(Vector128<UInt64>, Vector128<Int64>)

uint64x2_t vqshlq_u64 (uint64x2_t a, int64x2_t b)

A32 : VQSHL.U64 Qd, Qn, Qm

A64 : UQSHL Vd.2D, Vn.2D, Vm.2D

ShiftLogicalSaturate(Vector64<Byte>, Vector64<SByte>)

uint8x8_t vqshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VQSHL.U8 Dd, Dn, Dm

A64 : UQSHL Vd.8B, Vn.8B, Vm.8B

ShiftLogicalSaturate(Vector64<Int16>, Vector64<Int16>)

uint16x4_t vqshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VQSHL.U16 Dd, Dn, Dm

A64 : UQSHL Vd.4H, Vn.4H, Vm.4H

ShiftLogicalSaturate(Vector64<Int32>, Vector64<Int32>)

uint32x2_t vqshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VQSHL.U32 Dd, Dn, Dm

A64 : UQSHL Vd.2S, Vn.2S, Vm.2S

ShiftLogicalSaturate(Vector64<SByte>, Vector64<SByte>)

uint8x8_t vqshl_u8 (uint8x8_t a, int8x8_t b)

A32 : VQSHL.U8 Dd, Dn, Dm

A64 : UQSHL Vd.8B, Vn.8B, Vm.8B

ShiftLogicalSaturate(Vector64<UInt16>, Vector64<Int16>)

uint16x4_t vqshl_u16 (uint16x4_t a, int16x4_t b)

A32 : VQSHL.U16 Dd, Dn, Dm

A64 : UQSHL Vd.4H, Vn.4H, Vm.4H

ShiftLogicalSaturate(Vector64<UInt32>, Vector64<Int32>)

uint32x2_t vqshl_u32 (uint32x2_t a, int32x2_t b)

A32 : VQSHL.U32 Dd, Dn, Dm

A64 : UQSHL Vd.2S, Vn.2S, Vm.2S

ShiftLogicalSaturateScalar(Vector64<Int64>, Vector64<Int64>)

uint64x1_t vqshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VQSHL.U64 Dd, Dn, Dm

A64 : UQSHL Dd, Dn, Dm

ShiftLogicalSaturateScalar(Vector64<UInt64>, Vector64<Int64>)

uint64x1_t vqshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VQSHL.U64 Dd, Dn, Dm

A64 : UQSHL Dd, Dn, Dm

ShiftLogicalScalar(Vector64<Int64>, Vector64<Int64>)

uint64x1_t vshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VSHL.U64 Dd, Dn, Dm

A64 : USHL Dd, Dn, Dm

ShiftLogicalScalar(Vector64<UInt64>, Vector64<Int64>)

uint64x1_t vshl_u64 (uint64x1_t a, int64x1_t b)

A32 : VSHL.U64 Dd, Dn, Dm

A64 : USHL Dd, Dn, Dm

ShiftRightAndInsert(Vector128<Byte>, Vector128<Byte>, Byte)

uint8x16_t vsriq_n_u8(uint8x16_t a, uint8x16_t b, __builtin_constant_p(n))

A32 : VSRI.8 Qd, Qm, #n

A64 : SRI Vd.16B, Vn.16B, #n

ShiftRightAndInsert(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vsriq_n_s16(int16x8_t a, int16x8_t b, __builtin_constant_p(n))

A32 : VSRI.16 Qd, Qm, #n

A64 : SRI Vd.8H, Vn.8H, #n

ShiftRightAndInsert(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vsriq_n_s32(int32x4_t a, int32x4_t b, __builtin_constant_p(n))

A32 : VSRI.32 Qd, Qm, #n

A64 : SRI Vd.4S, Vn.4S, #n

ShiftRightAndInsert(Vector128<Int64>, Vector128<Int64>, Byte)

int64x2_t vsriq_n_s64(int64x2_t a, int64x2_t b, __builtin_constant_p(n))

A32 : VSRI.64 Qd, Qm, #n

A64 : SRI Vd.2D, Vn.2D, #n

ShiftRightAndInsert(Vector128<SByte>, Vector128<SByte>, Byte)

int8x16_t vsriq_n_s8(int8x16_t a, int8x16_t b, __builtin_constant_p(n))

A32 : VSRI.8 Qd, Qm, #n

A64 : SRI Vd.16B, Vn.16B, #n

ShiftRightAndInsert(Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vsriq_n_u16(uint16x8_t a, uint16x8_t b, __builtin_constant_p(n))

A32 : VSRI.16 Qd, Qm, #n

A64 : SRI Vd.8H, Vn.8H, #n

ShiftRightAndInsert(Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vsriq_n_u32(uint32x4_t a, uint32x4_t b, __builtin_constant_p(n))

A32 : VSRI.32 Qd, Qm, #n

A64 : SRI Vd.4S, Vn.4S, #n

ShiftRightAndInsert(Vector128<UInt64>, Vector128<UInt64>, Byte)

uint64x2_t vsriq_n_u64(uint64x2_t a, uint64x2_t b, __builtin_constant_p(n))

A32 : VSRI.64 Qd, Qm, #n

A64 : SRI Vd.2D, Vn.2D, #n

ShiftRightAndInsert(Vector64<Byte>, Vector64<Byte>, Byte)

uint8x8_t vsri_n_u8(uint8x8_t a, uint8x8_t b, __builtin_constant_p(n))

A32 : VSRI.8 Dd, Dm, #n

A64 : SRI Vd.8B, Vn.8B, #n

ShiftRightAndInsert(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vsri_n_s16(int16x4_t a, int16x4_t b, __builtin_constant_p(n))

A32 : VSRI.16 Dd, Dm, #n

A64 : SRI Vd.4H, Vn.4H, #n

ShiftRightAndInsert(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vsri_n_s32(int32x2_t a, int32x2_t b, __builtin_constant_p(n))

A32 : VSRI.32 Dd, Dm, #n

A64 : SRI Vd.2S, Vn.2S, #n

ShiftRightAndInsert(Vector64<SByte>, Vector64<SByte>, Byte)

int8x8_t vsri_n_s8(int8x8_t a, int8x8_t b, __builtin_constant_p(n))

A32 : VSRI.8 Dd, Dm, #n

A64 : SRI Vd.8B, Vn.8B, #n

ShiftRightAndInsert(Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vsri_n_u16(uint16x4_t a, uint16x4_t b, __builtin_constant_p(n))

A32 : VSRI.16 Dd, Dm, #n

A64 : SRI Vd.4H, Vn.4H, #n

ShiftRightAndInsert(Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vsri_n_u32(uint32x2_t a, uint32x2_t b, __builtin_constant_p(n))

A32 : VSRI.32 Dd, Dm, #n

A64 : SRI Vd.2S, Vn.2S, #n

ShiftRightAndInsertScalar(Vector64<Int64>, Vector64<Int64>, Byte)

int64_t vsrid_n_s64(int64_t a, int64_t b, __builtin_constant_p(n))

A32 : VSRI.64 Dd, Dm, #n

A64 : SRI Dd, Dn, #n

ShiftRightAndInsertScalar(Vector64<UInt64>, Vector64<UInt64>, Byte)

uint64_t vsrid_n_u64(uint64_t a, uint64_t b, __builtin_constant_p(n))

A32 : VSRI.64 Dd, Dm, #n

A64 : SRI Dd, Dn, #n

ShiftRightArithmetic(Vector128<Int16>, Byte)

int16x8_t vshrq_n_s16 (int16x8_t a, const int n)

A32 : VSHR.S16 Qd, Qm, #n

A64 : SSHR Vd.8H, Vn.8H, #n

ShiftRightArithmetic(Vector128<Int32>, Byte)

int32x4_t vshrq_n_s32 (int32x4_t a, const int n)

A32 : VSHR.S32 Qd, Qm, #n

A64 : SSHR Vd.4S, Vn.4S, #n

ShiftRightArithmetic(Vector128<Int64>, Byte)

int64x2_t vshrq_n_s64 (int64x2_t a, const int n)

A32 : VSHR.S64 Qd, Qm, #n

A64 : SSHR Vd.2D, Vn.2D, #n

ShiftRightArithmetic(Vector128<SByte>, Byte)

int8x16_t vshrq_n_s8 (int8x16_t a, const int n)

A32 : VSHR.S8 Qd, Qm, #n

A64 : SSHR Vd.16B, Vn.16B, #n

ShiftRightArithmetic(Vector64<Int16>, Byte)

int16x4_t vshr_n_s16 (int16x4_t a, const int n)

A32 : VSHR.S16 Dd, Dm, #n

A64 : SSHR Vd.4H, Vn.4H, #n

ShiftRightArithmetic(Vector64<Int32>, Byte)

int32x2_t vshr_n_s32 (int32x2_t a, const int n)

A32 : VSHR.S32 Dd, Dm, #n

A64 : SSHR Vd.2S, Vn.2S, #n

ShiftRightArithmetic(Vector64<SByte>, Byte)

int8x8_t vshr_n_s8 (int8x8_t a, const int n)

A32 : VSHR.S8 Dd, Dm, #n

A64 : SSHR Vd.8B, Vn.8B, #n

ShiftRightArithmeticAdd(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vsraq_n_s16 (int16x8_t a, int16x8_t b, const int n)

A32 : VSRA.S16 Qd, Qm, #n

A64 : SSRA Vd.8H, Vn.8H, #n

ShiftRightArithmeticAdd(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vsraq_n_s32 (int32x4_t a, int32x4_t b, const int n)

A32 : VSRA.S32 Qd, Qm, #n

A64 : SSRA Vd.4S, Vn.4S, #n

ShiftRightArithmeticAdd(Vector128<Int64>, Vector128<Int64>, Byte)

int64x2_t vsraq_n_s64 (int64x2_t a, int64x2_t b, const int n)

A32 : VSRA.S64 Qd, Qm, #n

A64 : SSRA Vd.2D, Vn.2D, #n

ShiftRightArithmeticAdd(Vector128<SByte>, Vector128<SByte>, Byte)

int8x16_t vsraq_n_s8 (int8x16_t a, int8x16_t b, const int n)

A32 : VSRA.S8 Qd, Qm, #n

A64 : SSRA Vd.16B, Vn.16B, #n

ShiftRightArithmeticAdd(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vsra_n_s16 (int16x4_t a, int16x4_t b, const int n)

A32 : VSRA.S16 Dd, Dm, #n

A64 : SSRA Vd.4H, Vn.4H, #n

ShiftRightArithmeticAdd(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vsra_n_s32 (int32x2_t a, int32x2_t b, const int n)

A32 : VSRA.S32 Dd, Dm, #n

A64 : SSRA Vd.2S, Vn.2S, #n

ShiftRightArithmeticAdd(Vector64<SByte>, Vector64<SByte>, Byte)

int8x8_t vsra_n_s8 (int8x8_t a, int8x8_t b, const int n)

A32 : VSRA.S8 Dd, Dm, #n

A64 : SSRA Vd.8B, Vn.8B, #n

ShiftRightArithmeticAddScalar(Vector64<Int64>, Vector64<Int64>, Byte)

int64x1_t vsra_n_s64 (int64x1_t a, int64x1_t b, const int n)

A32 : VSRA.S64 Dd, Dm, #n

A64 : SSRA Dd, Dn, #n

ShiftRightArithmeticNarrowingSaturateLower(Vector128<Int16>, Byte)

int8x8_t vqshrn_n_s16 (int16x8_t a, const int n)

A32 : VQSHRN.S16 Dd, Qm, #n

A64 : SQSHRN Vd.8B, Vn.8H, #n

ShiftRightArithmeticNarrowingSaturateLower(Vector128<Int32>, Byte)

int16x4_t vqshrn_n_s32 (int32x4_t a, const int n)

A32 : VQSHRN.S32 Dd, Qm, #n

A64 : SQSHRN Vd.4H, Vn.4S, #n

ShiftRightArithmeticNarrowingSaturateLower(Vector128<Int64>, Byte)

int32x2_t vqshrn_n_s64 (int64x2_t a, const int n)

A32 : VQSHRN.S64 Dd, Qm, #n

A64 : SQSHRN Vd.2S, Vn.2D, #n

ShiftRightArithmeticNarrowingSaturateUnsignedLower(Vector128<Int16>, Byte)

uint8x8_t vqshrun_n_s16 (int16x8_t a, const int n)

A32 : VQSHRUN.S16 Dd, Qm, #n

A64 : SQSHRUN Vd.8B, Vn.8H, #n

ShiftRightArithmeticNarrowingSaturateUnsignedLower(Vector128<Int32>, Byte)

uint16x4_t vqshrun_n_s32 (int32x4_t a, const int n)

A32 : VQSHRUN.S32 Dd, Qm, #n

A64 : SQSHRUN Vd.4H, Vn.4S, #n

ShiftRightArithmeticNarrowingSaturateUnsignedLower(Vector128<Int64>, Byte)

uint32x2_t vqshrun_n_s64 (int64x2_t a, const int n)

A32 : VQSHRUN.S64 Dd, Qm, #n

A64 : SQSHRUN Vd.2S, Vn.2D, #n

ShiftRightArithmeticNarrowingSaturateUnsignedUpper(Vector64<Byte>, Vector128<Int16>, Byte)

uint8x16_t vqshrun_high_n_s16 (uint8x8_t r, int16x8_t a, const int n)

A32 : VQSHRUN.S16 Dd+1, Dn, #n

A64 : SQSHRUN2 Vd.16B, Vn.8H, #n

ShiftRightArithmeticNarrowingSaturateUnsignedUpper(Vector64<UInt16>, Vector128<Int32>, Byte)

uint16x8_t vqshrun_high_n_s32 (uint16x4_t r, int32x4_t a, const int n)

A32 : VQSHRUN.S32 Dd+1, Dn, #n

A64 : SQSHRUN2 Vd.8H, Vn.4S, #n

ShiftRightArithmeticNarrowingSaturateUnsignedUpper(Vector64<UInt32>, Vector128<Int64>, Byte)

uint32x4_t vqshrun_high_n_s64 (uint32x2_t r, int64x2_t a, const int n)

A32 : VQSHRUN.S64 Dd+1, Dn, #n

A64 : SQSHRUN2 Vd.4S, Vn.2D, #n

ShiftRightArithmeticNarrowingSaturateUpper(Vector64<Int16>, Vector128<Int32>, Byte)

int16x8_t vqshrn_high_n_s32 (int16x4_t r, int32x4_t a, const int n)

A32 : VQSHRN.S32 Dd+1, Qm, #n

A64 : SQSHRN2 Vd.8H, Vn.4S, #n

ShiftRightArithmeticNarrowingSaturateUpper(Vector64<Int32>, Vector128<Int64>, Byte)

int32x4_t vqshrn_high_n_s64 (int32x2_t r, int64x2_t a, const int n)

A32 : VQSHRN.S64 Dd+1, Qm, #n

A64 : SQSHRN2 Vd.4S, Vn.2D, #n

ShiftRightArithmeticNarrowingSaturateUpper(Vector64<SByte>, Vector128<Int16>, Byte)

int8x16_t vqshrn_high_n_s16 (int8x8_t r, int16x8_t a, const int n)

A32 : VQSHRN.S16 Dd+1, Qm, #n

A64 : SQSHRN2 Vd.16B, Vn.8H, #n

ShiftRightArithmeticRounded(Vector128<Int16>, Byte)

int16x8_t vrshrq_n_s16 (int16x8_t a, const int n)

A32 : VRSHR.S16 Qd, Qm, #n

A64 : SRSHR Vd.8H, Vn.8H, #n

ShiftRightArithmeticRounded(Vector128<Int32>, Byte)

int32x4_t vrshrq_n_s32 (int32x4_t a, const int n)

A32 : VRSHR.S32 Qd, Qm, #n

A64 : SRSHR Vd.4S, Vn.4S, #n

ShiftRightArithmeticRounded(Vector128<Int64>, Byte)

int64x2_t vrshrq_n_s64 (int64x2_t a, const int n)

A32 : VRSHR.S64 Qd, Qm, #n

A64 : SRSHR Vd.2D, Vn.2D, #n

ShiftRightArithmeticRounded(Vector128<SByte>, Byte)

int8x16_t vrshrq_n_s8 (int8x16_t a, const int n)

A32 : VRSHR.S8 Qd, Qm, #n

A64 : SRSHR Vd.16B, Vn.16B, #n

ShiftRightArithmeticRounded(Vector64<Int16>, Byte)

int16x4_t vrshr_n_s16 (int16x4_t a, const int n)

A32 : VRSHR.S16 Dd, Dm, #n

A64 : SRSHR Vd.4H, Vn.4H, #n

ShiftRightArithmeticRounded(Vector64<Int32>, Byte)

int32x2_t vrshr_n_s32 (int32x2_t a, const int n)

A32 : VRSHR.S32 Dd, Dm, #n

A64 : SRSHR Vd.2S, Vn.2S, #n

ShiftRightArithmeticRounded(Vector64<SByte>, Byte)

int8x8_t vrshr_n_s8 (int8x8_t a, const int n)

A32 : VRSHR.S8 Dd, Dm, #n

A64 : SRSHR Vd.8B, Vn.8B, #n

ShiftRightArithmeticRoundedAdd(Vector128<Int16>, Vector128<Int16>, Byte)

int16x8_t vrsraq_n_s16 (int16x8_t a, int16x8_t b, const int n)

A32 : VRSRA.S16 Qd, Qm, #n

A64 : SRSRA Vd.8H, Vn.8H, #n

ShiftRightArithmeticRoundedAdd(Vector128<Int32>, Vector128<Int32>, Byte)

int32x4_t vrsraq_n_s32 (int32x4_t a, int32x4_t b, const int n)

A32 : VRSRA.S32 Qd, Qm, #n

A64 : SRSRA Vd.4S, Vn.4S, #n

ShiftRightArithmeticRoundedAdd(Vector128<Int64>, Vector128<Int64>, Byte)

int64x2_t vrsraq_n_s64 (int64x2_t a, int64x2_t b, const int n)

A32 : VRSRA.S64 Qd, Qm, #n

A64 : SRSRA Vd.2D, Vn.2D, #n

ShiftRightArithmeticRoundedAdd(Vector128<SByte>, Vector128<SByte>, Byte)

int8x16_t vrsraq_n_s8 (int8x16_t a, int8x16_t b, const int n)

A32 : VRSRA.S8 Qd, Qm, #n

A64 : SRSRA Vd.16B, Vn.16B, #n

ShiftRightArithmeticRoundedAdd(Vector64<Int16>, Vector64<Int16>, Byte)

int16x4_t vrsra_n_s16 (int16x4_t a, int16x4_t b, const int n)

A32 : VRSRA.S16 Dd, Dm, #n

A64 : SRSRA Vd.4H, Vn.4H, #n

ShiftRightArithmeticRoundedAdd(Vector64<Int32>, Vector64<Int32>, Byte)

int32x2_t vrsra_n_s32 (int32x2_t a, int32x2_t b, const int n)

A32 : VRSRA.S32 Dd, Dm, #n

A64 : SRSRA Vd.2S, Vn.2S, #n

ShiftRightArithmeticRoundedAdd(Vector64<SByte>, Vector64<SByte>, Byte)

int8x8_t vrsra_n_s8 (int8x8_t a, int8x8_t b, const int n)

A32 : VRSRA.S8 Dd, Dm, #n

A64 : SRSRA Vd.8B, Vn.8B, #n

ShiftRightArithmeticRoundedAddScalar(Vector64<Int64>, Vector64<Int64>, Byte)

int64x1_t vrsra_n_s64 (int64x1_t a, int64x1_t b, const int n)

A32 : VRSRA.S64 Dd, Dm, #n

A64 : SRSRA Dd, Dn, #n

ShiftRightArithmeticRoundedNarrowingSaturateLower(Vector128<Int16>, Byte)

int8x8_t vqrshrn_n_s16 (int16x8_t a, const int n)

A32 : VQRSHRN.S16 Dd, Qm, #n

A64 : SQRSHRN Vd.8B, Vn.8H, #n

ShiftRightArithmeticRoundedNarrowingSaturateLower(Vector128<Int32>, Byte)

int16x4_t vqrshrn_n_s32 (int32x4_t a, const int n)

A32 : VQRSHRN.S32 Dd, Qm, #n

A64 : SQRSHRN Vd.4H, Vn.4S, #n

ShiftRightArithmeticRoundedNarrowingSaturateLower(Vector128<Int64>, Byte)

int32x2_t vqrshrn_n_s64 (int64x2_t a, const int n)

A32 : VQRSHRN.S64 Dd, Qm, #n

A64 : SQRSHRN Vd.2S, Vn.2D, #n

ShiftRightArithmeticRoundedNarrowingSaturateUnsignedLower(Vector128<Int16>, Byte)

uint8x8_t vqrshrun_n_s16 (int16x8_t a, const int n)

A32 : VQRSHRUN.S16 Dd, Qm, #n

A64 : SQRSHRUN Vd.8B, Vn.8H, #n

ShiftRightArithmeticRoundedNarrowingSaturateUnsignedLower(Vector128<Int32>, Byte)

uint16x4_t vqrshrun_n_s32 (int32x4_t a, const int n)

A32 : VQRSHRUN.S32 Dd, Qm, #n

A64 : SQRSHRUN Vd.4H, Vn.4S, #n

ShiftRightArithmeticRoundedNarrowingSaturateUnsignedLower(Vector128<Int64>, Byte)

uint32x2_t vqrshrun_n_s64 (int64x2_t a, const int n)

A32 : VQRSHRUN.S64 Dd, Qm, #n

A64 : SQRSHRUN Vd.2S, Vn.2D, #n

ShiftRightArithmeticRoundedNarrowingSaturateUnsignedUpper(Vector64<Byte>, Vector128<Int16>, Byte)

uint8x16_t vqrshrun_high_n_s16 (uint8x8_t r, int16x8_t a, const int n)

A32 : VQRSHRUN.S16 Dd+1, Dn, #n

A64 : SQRSHRUN2 Vd.16B, Vn.8H, #n

ShiftRightArithmeticRoundedNarrowingSaturateUnsignedUpper(Vector64<UInt16>, Vector128<Int32>, Byte)

uint16x8_t vqrshrun_high_n_s32 (uint16x4_t r, int32x4_t a, const int n)

A32 : VQRSHRUN.S32 Dd+1, Dn, #n

A64 : SQRSHRUN2 Vd.8H, Vn.4S, #n

ShiftRightArithmeticRoundedNarrowingSaturateUnsignedUpper(Vector64<UInt32>, Vector128<Int64>, Byte)

uint32x4_t vqrshrun_high_n_s64 (uint32x2_t r, int64x2_t a, const int n)

A32 : VQRSHRUN.S64 Dd+1, Dn, #n

A64 : SQRSHRUN2 Vd.4S, Vn.2D, #n

ShiftRightArithmeticRoundedNarrowingSaturateUpper(Vector64<Int16>, Vector128<Int32>, Byte)

int16x8_t vqrshrn_high_n_s32 (int16x4_t r, int32x4_t a, const int n)

A32 : VQRSHRN.S32 Dd+1, Dn, #n

A64 : SQRSHRN2 Vd.8H, Vn.4S, #n

ShiftRightArithmeticRoundedNarrowingSaturateUpper(Vector64<Int32>, Vector128<Int64>, Byte)

int32x4_t vqrshrn_high_n_s64 (int32x2_t r, int64x2_t a, const int n)

A32 : VQRSHRN.S64 Dd+1, Dn, #n

A64 : SQRSHRN2 Vd.4S, Vn.2D, #n

ShiftRightArithmeticRoundedNarrowingSaturateUpper(Vector64<SByte>, Vector128<Int16>, Byte)

int8x16_t vqrshrn_high_n_s16 (int8x8_t r, int16x8_t a, const int n)

A32 : VQRSHRN.S16 Dd+1, Dn, #n

A64 : SQRSHRN2 Vd.16B, Vn.8H, #n

ShiftRightArithmeticRoundedScalar(Vector64<Int64>, Byte)

int64x1_t vrshr_n_s64 (int64x1_t a, const int n)

A32 : VRSHR.S64 Dd, Dm, #n

A64 : SRSHR Dd, Dn, #n

ShiftRightArithmeticScalar(Vector64<Int64>, Byte)

int64x1_t vshr_n_s64 (int64x1_t a, const int n)

A32 : VSHR.S64 Dd, Dm, #n

A64 : SSHR Dd, Dn, #n

ShiftRightLogical(Vector128<Byte>, Byte)

uint8x16_t vshrq_n_u8 (uint8x16_t a, const int n)

A32 : VSHR.U8 Qd, Qm, #n

A64 : USHR Vd.16B, Vn.16B, #n

ShiftRightLogical(Vector128<Int16>, Byte)

uint16x8_t vshrq_n_u16 (uint16x8_t a, const int n)

A32 : VSHR.U16 Qd, Qm, #n

A64 : USHR Vd.8H, Vn.8H, #n

ShiftRightLogical(Vector128<Int32>, Byte)

uint32x4_t vshrq_n_u32 (uint32x4_t a, const int n)

A32 : VSHR.U32 Qd, Qm, #n

A64 : USHR Vd.4S, Vn.4S, #n

ShiftRightLogical(Vector128<Int64>, Byte)

uint64x2_t vshrq_n_u64 (uint64x2_t a, const int n)

A32 : VSHR.U64 Qd, Qm, #n

A64 : USHR Vd.2D, Vn.2D, #n

ShiftRightLogical(Vector128<SByte>, Byte)

uint8x16_t vshrq_n_u8 (uint8x16_t a, const int n)

A32 : VSHR.U8 Qd, Qm, #n

A64 : USHR Vd.16B, Vn.16B, #n

ShiftRightLogical(Vector128<UInt16>, Byte)

uint16x8_t vshrq_n_u16 (uint16x8_t a, const int n)

A32 : VSHR.U16 Qd, Qm, #n

A64 : USHR Vd.8H, Vn.8H, #n

ShiftRightLogical(Vector128<UInt32>, Byte)

uint32x4_t vshrq_n_u32 (uint32x4_t a, const int n)

A32 : VSHR.U32 Qd, Qm, #n

A64 : USHR Vd.4S, Vn.4S, #n

ShiftRightLogical(Vector128<UInt64>, Byte)

uint64x2_t vshrq_n_u64 (uint64x2_t a, const int n)

A32 : VSHR.U64 Qd, Qm, #n

A64 : USHR Vd.2D, Vn.2D, #n

ShiftRightLogical(Vector64<Byte>, Byte)

uint8x8_t vshr_n_u8 (uint8x8_t a, const int n)

A32 : VSHR.U8 Dd, Dm, #n

A64 : USHR Vd.8B, Vn.8B, #n

ShiftRightLogical(Vector64<Int16>, Byte)

uint16x4_t vshr_n_u16 (uint16x4_t a, const int n)

A32 : VSHR.U16 Dd, Dm, #n

A64 : USHR Vd.4H, Vn.4H, #n

ShiftRightLogical(Vector64<Int32>, Byte)

uint32x2_t vshr_n_u32 (uint32x2_t a, const int n)

A32 : VSHR.U32 Dd, Dm, #n

A64 : USHR Vd.2S, Vn.2S, #n

ShiftRightLogical(Vector64<SByte>, Byte)

uint8x8_t vshr_n_u8 (uint8x8_t a, const int n)

A32 : VSHR.U8 Dd, Dm, #n

A64 : USHR Vd.8B, Vn.8B, #n

ShiftRightLogical(Vector64<UInt16>, Byte)

uint16x4_t vshr_n_u16 (uint16x4_t a, const int n)

A32 : VSHR.U16 Dd, Dm, #n

A64 : USHR Vd.4H, Vn.4H, #n

ShiftRightLogical(Vector64<UInt32>, Byte)

uint32x2_t vshr_n_u32 (uint32x2_t a, const int n)

A32 : VSHR.U32 Dd, Dm, #n

A64 : USHR Vd.2S, Vn.2S, #n

ShiftRightLogicalAdd(Vector128<Byte>, Vector128<Byte>, Byte)

uint8x16_t vsraq_n_u8 (uint8x16_t a, uint8x16_t b, const int n)

A32 : VSRA.U8 Qd, Qm, #n

A64 : USRA Vd.16B, Vn.16B, #n

ShiftRightLogicalAdd(Vector128<Int16>, Vector128<Int16>, Byte)

uint16x8_t vsraq_n_u16 (uint16x8_t a, uint16x8_t b, const int n)

A32 : VSRA.U16 Qd, Qm, #n

A64 : USRA Vd.8H, Vn.8H, #n

ShiftRightLogicalAdd(Vector128<Int32>, Vector128<Int32>, Byte)

uint32x4_t vsraq_n_u32 (uint32x4_t a, uint32x4_t b, const int n)

A32 : VSRA.U32 Qd, Qm, #n

A64 : USRA Vd.4S, Vn.4S, #n

ShiftRightLogicalAdd(Vector128<Int64>, Vector128<Int64>, Byte)

uint64x2_t vsraq_n_u64 (uint64x2_t a, uint64x2_t b, const int n)

A32 : VSRA.U64 Qd, Qm, #n

A64 : USRA Vd.2D, Vn.2D, #n

ShiftRightLogicalAdd(Vector128<SByte>, Vector128<SByte>, Byte)

uint8x16_t vsraq_n_u8 (uint8x16_t a, uint8x16_t b, const int n)

A32 : VSRA.U8 Qd, Qm, #n

A64 : USRA Vd.16B, Vn.16B, #n

ShiftRightLogicalAdd(Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vsraq_n_u16 (uint16x8_t a, uint16x8_t b, const int n)

A32 : VSRA.U16 Qd, Qm, #n

A64 : USRA Vd.8H, Vn.8H, #n

ShiftRightLogicalAdd(Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vsraq_n_u32 (uint32x4_t a, uint32x4_t b, const int n)

A32 : VSRA.U32 Qd, Qm, #n

A64 : USRA Vd.4S, Vn.4S, #n

ShiftRightLogicalAdd(Vector128<UInt64>, Vector128<UInt64>, Byte)

uint64x2_t vsraq_n_u64 (uint64x2_t a, uint64x2_t b, const int n)

A32 : VSRA.U64 Qd, Qm, #n

A64 : USRA Vd.2D, Vn.2D, #n

ShiftRightLogicalAdd(Vector64<Byte>, Vector64<Byte>, Byte)

uint8x8_t vsra_n_u8 (uint8x8_t a, uint8x8_t b, const int n)

A32 : VSRA.U8 Dd, Dm, #n

A64 : USRA Vd.8B, Vn.8B, #n

ShiftRightLogicalAdd(Vector64<Int16>, Vector64<Int16>, Byte)

uint16x4_t vsra_n_u16 (uint16x4_t a, uint16x4_t b, const int n)

A32 : VSRA.U16 Dd, Dm, #n

A64 : USRA Vd.4H, Vn.4H, #n

ShiftRightLogicalAdd(Vector64<Int32>, Vector64<Int32>, Byte)

uint32x2_t vsra_n_u32 (uint32x2_t a, uint32x2_t b, const int n)

A32 : VSRA.U32 Dd, Dm, #n

A64 : USRA Vd.2S, Vn.2S, #n

ShiftRightLogicalAdd(Vector64<SByte>, Vector64<SByte>, Byte)

uint8x8_t vsra_n_u8 (uint8x8_t a, uint8x8_t b, const int n)

A32 : VSRA.U8 Dd, Dm, #n

A64 : USRA Vd.8B, Vn.8B, #n

ShiftRightLogicalAdd(Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vsra_n_u16 (uint16x4_t a, uint16x4_t b, const int n)

A32 : VSRA.U16 Dd, Dm, #n

A64 : USRA Vd.4H, Vn.4H, #n

ShiftRightLogicalAdd(Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vsra_n_u32 (uint32x2_t a, uint32x2_t b, const int n)

A32 : VSRA.U32 Dd, Dm, #n

A64 : USRA Vd.2S, Vn.2S, #n

ShiftRightLogicalAddScalar(Vector64<Int64>, Vector64<Int64>, Byte)

uint64x1_t vsra_n_u64 (uint64x1_t a, uint64x1_t b, const int n)

A32 : VSRA.U64 Dd, Dm, #n

A64 : USRA Dd, Dn, #n

ShiftRightLogicalAddScalar(Vector64<UInt64>, Vector64<UInt64>, Byte)

uint64x1_t vsra_n_u64 (uint64x1_t a, uint64x1_t b, const int n)

A32 : VSRA.U64 Dd, Dm, #n

A64 : USRA Dd, Dn, #n

ShiftRightLogicalNarrowingLower(Vector128<Int16>, Byte)

int8x8_t vshrn_n_s16 (int16x8_t a, const int n)

A32 : VSHRN.I16 Dd, Qm, #n

A64 : SHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalNarrowingLower(Vector128<Int32>, Byte)

int16x4_t vshrn_n_s32 (int32x4_t a, const int n)

A32 : VSHRN.I32 Dd, Qm, #n

A64 : SHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalNarrowingLower(Vector128<Int64>, Byte)

int32x2_t vshrn_n_s64 (int64x2_t a, const int n)

A32 : VSHRN.I64 Dd, Qm, #n

A64 : SHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalNarrowingLower(Vector128<UInt16>, Byte)

uint8x8_t vshrn_n_u16 (uint16x8_t a, const int n)

A32 : VSHRN.I16 Dd, Qm, #n

A64 : SHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalNarrowingLower(Vector128<UInt32>, Byte)

uint16x4_t vshrn_n_u32 (uint32x4_t a, const int n)

A32 : VSHRN.I32 Dd, Qm, #n

A64 : SHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalNarrowingLower(Vector128<UInt64>, Byte)

uint32x2_t vshrn_n_u64 (uint64x2_t a, const int n)

A32 : VSHRN.I64 Dd, Qm, #n

A64 : SHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalNarrowingSaturateLower(Vector128<Int16>, Byte)

uint8x8_t vqshrn_n_u16 (uint16x8_t a, const int n)

A32 : VQSHRN.U16 Dd, Qm, #n

A64 : UQSHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalNarrowingSaturateLower(Vector128<Int32>, Byte)

uint16x4_t vqshrn_n_u32 (uint32x4_t a, const int n)

A32 : VQSHRN.U32 Dd, Qm, #n

A64 : UQSHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalNarrowingSaturateLower(Vector128<Int64>, Byte)

uint32x2_t vqshrn_n_u64 (uint64x2_t a, const int n)

A32 : VQSHRN.U64 Dd, Qm, #n

A64 : UQSHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalNarrowingSaturateLower(Vector128<UInt16>, Byte)

uint8x8_t vqshrn_n_u16 (uint16x8_t a, const int n)

A32 : VQSHRN.U16 Dd, Qm, #n

A64 : UQSHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalNarrowingSaturateLower(Vector128<UInt32>, Byte)

uint16x4_t vqshrn_n_u32 (uint32x4_t a, const int n)

A32 : VQSHRN.U32 Dd, Qm, #n

A64 : UQSHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalNarrowingSaturateLower(Vector128<UInt64>, Byte)

uint32x2_t vqshrn_n_u64 (uint64x2_t a, const int n)

A32 : VQSHRN.U64 Dd, Qm, #n

A64 : UQSHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalNarrowingSaturateUpper(Vector64<Byte>, Vector128<UInt16>, Byte)

uint8x16_t vqshrn_high_n_u16 (uint8x8_t r, uint16x8_t a, const int n)

A32 : VQSHRN.U16 Dd+1, Qm, #n

A64 : UQSHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalNarrowingSaturateUpper(Vector64<Int16>, Vector128<Int32>, Byte)

uint16x8_t vqshrn_high_n_u32 (uint16x4_t r, uint32x4_t a, const int n)

A32 : VQSHRN.U32 Dd+1, Qm, #n

A64 : UQSHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalNarrowingSaturateUpper(Vector64<Int32>, Vector128<Int64>, Byte)

uint32x4_t vqshrn_high_n_u64 (uint32x2_t r, uint64x2_t a, const int n)

A32 : VQSHRN.U64 Dd+1, Qm, #n

A64 : UQSHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalNarrowingSaturateUpper(Vector64<SByte>, Vector128<Int16>, Byte)

uint8x16_t vqshrn_high_n_u16 (uint8x8_t r, uint16x8_t a, const int n)

A32 : VQSHRN.U16 Dd+1, Qm, #n

A64 : UQSHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalNarrowingSaturateUpper(Vector64<UInt16>, Vector128<UInt32>, Byte)

uint16x8_t vqshrn_high_n_u32 (uint16x4_t r, uint32x4_t a, const int n)

A32 : VQSHRN.U32 Dd+1, Qm, #n

A64 : UQSHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalNarrowingSaturateUpper(Vector64<UInt32>, Vector128<UInt64>, Byte)

uint32x4_t vqshrn_high_n_u64 (uint32x2_t r, uint64x2_t a, const int n)

A32 : VQSHRN.U64 Dd+1, Qm, #n

A64 : UQSHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalNarrowingUpper(Vector64<Byte>, Vector128<UInt16>, Byte)

uint8x16_t vshrn_high_n_u16 (uint8x8_t r, uint16x8_t a, const int n)

A32 : VSHRN.I16 Dd+1, Qm, #n

A64 : SHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalNarrowingUpper(Vector64<Int16>, Vector128<Int32>, Byte)

int16x8_t vshrn_high_n_s32 (int16x4_t r, int32x4_t a, const int n)

A32 : VSHRN.I32 Dd+1, Qm, #n

A64 : SHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalNarrowingUpper(Vector64<Int32>, Vector128<Int64>, Byte)

int32x4_t vshrn_high_n_s64 (int32x2_t r, int64x2_t a, const int n)

A32 : VSHRN.I64 Dd+1, Qm, #n

A64 : SHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalNarrowingUpper(Vector64<SByte>, Vector128<Int16>, Byte)

int8x16_t vshrn_high_n_s16 (int8x8_t r, int16x8_t a, const int n)

A32 : VSHRN.I16 Dd+1, Qm, #n

A64 : SHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalNarrowingUpper(Vector64<UInt16>, Vector128<UInt32>, Byte)

uint16x8_t vshrn_high_n_u32 (uint16x4_t r, uint32x4_t a, const int n)

A32 : VSHRN.I32 Dd+1, Qm, #n

A64 : SHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalNarrowingUpper(Vector64<UInt32>, Vector128<UInt64>, Byte)

uint32x4_t vshrn_high_n_u64 (uint32x2_t r, uint64x2_t a, const int n)

A32 : VSHRN.I64 Dd+1, Qm, #n

A64 : SHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalRounded(Vector128<Byte>, Byte)

uint8x16_t vrshrq_n_u8 (uint8x16_t a, const int n)

A32 : VRSHR.U8 Qd, Qm, #n

A64 : URSHR Vd.16B, Vn.16B, #n

ShiftRightLogicalRounded(Vector128<Int16>, Byte)

uint16x8_t vrshrq_n_u16 (uint16x8_t a, const int n)

A32 : VRSHR.U16 Qd, Qm, #n

A64 : URSHR Vd.8H, Vn.8H, #n

ShiftRightLogicalRounded(Vector128<Int32>, Byte)

uint32x4_t vrshrq_n_u32 (uint32x4_t a, const int n)

A32 : VRSHR.U32 Qd, Qm, #n

A64 : URSHR Vd.4S, Vn.4S, #n

ShiftRightLogicalRounded(Vector128<Int64>, Byte)

uint64x2_t vrshrq_n_u64 (uint64x2_t a, const int n)

A32 : VRSHR.U64 Qd, Qm, #n

A64 : URSHR Vd.2D, Vn.2D, #n

ShiftRightLogicalRounded(Vector128<SByte>, Byte)

uint8x16_t vrshrq_n_u8 (uint8x16_t a, const int n)

A32 : VRSHR.U8 Qd, Qm, #n

A64 : URSHR Vd.16B, Vn.16B, #n

ShiftRightLogicalRounded(Vector128<UInt16>, Byte)

uint16x8_t vrshrq_n_u16 (uint16x8_t a, const int n)

A32 : VRSHR.U16 Qd, Qm, #n

A64 : URSHR Vd.8H, Vn.8H, #n

ShiftRightLogicalRounded(Vector128<UInt32>, Byte)

uint32x4_t vrshrq_n_u32 (uint32x4_t a, const int n)

A32 : VRSHR.U32 Qd, Qm, #n

A64 : URSHR Vd.4S, Vn.4S, #n

ShiftRightLogicalRounded(Vector128<UInt64>, Byte)

uint64x2_t vrshrq_n_u64 (uint64x2_t a, const int n)

A32 : VRSHR.U64 Qd, Qm, #n

A64 : URSHR Vd.2D, Vn.2D, #n

ShiftRightLogicalRounded(Vector64<Byte>, Byte)

uint8x8_t vrshr_n_u8 (uint8x8_t a, const int n)

A32 : VRSHR.U8 Dd, Dm, #n

A64 : URSHR Vd.8B, Vn.8B, #n

ShiftRightLogicalRounded(Vector64<Int16>, Byte)

uint16x4_t vrshr_n_u16 (uint16x4_t a, const int n)

A32 : VRSHR.U16 Dd, Dm, #n

A64 : URSHR Vd.4H, Vn.4H, #n

ShiftRightLogicalRounded(Vector64<Int32>, Byte)

uint32x2_t vrshr_n_u32 (uint32x2_t a, const int n)

A32 : VRSHR.U32 Dd, Dm, #n

A64 : URSHR Vd.2S, Vn.2S, #n

ShiftRightLogicalRounded(Vector64<SByte>, Byte)

uint8x8_t vrshr_n_u8 (uint8x8_t a, const int n)

A32 : VRSHR.U8 Dd, Dm, #n

A64 : URSHR Vd.8B, Vn.8B, #n

ShiftRightLogicalRounded(Vector64<UInt16>, Byte)

uint16x4_t vrshr_n_u16 (uint16x4_t a, const int n)

A32 : VRSHR.U16 Dd, Dm, #n

A64 : URSHR Vd.4H, Vn.4H, #n

ShiftRightLogicalRounded(Vector64<UInt32>, Byte)

uint32x2_t vrshr_n_u32 (uint32x2_t a, const int n)

A32 : VRSHR.U32 Dd, Dm, #n

A64 : URSHR Vd.2S, Vn.2S, #n

ShiftRightLogicalRoundedAdd(Vector128<Byte>, Vector128<Byte>, Byte)

uint8x16_t vrsraq_n_u8 (uint8x16_t a, uint8x16_t b, const int n)

A32 : VRSRA.U8 Qd, Qm, #n

A64 : URSRA Vd.16B, Vn.16B, #n

ShiftRightLogicalRoundedAdd(Vector128<Int16>, Vector128<Int16>, Byte)

uint16x8_t vrsraq_n_u16 (uint16x8_t a, uint16x8_t b, const int n)

A32 : VRSRA.U16 Qd, Qm, #n

A64 : URSRA Vd.8H, Vn.8H, #n

ShiftRightLogicalRoundedAdd(Vector128<Int32>, Vector128<Int32>, Byte)

uint32x4_t vrsraq_n_u32 (uint32x4_t a, uint32x4_t b, const int n)

A32 : VRSRA.U32 Qd, Qm, #n

A64 : URSRA Vd.4S, Vn.4S, #n

ShiftRightLogicalRoundedAdd(Vector128<Int64>, Vector128<Int64>, Byte)

uint64x2_t vrsraq_n_u64 (uint64x2_t a, uint64x2_t b, const int n)

A32 : VRSRA.U64 Qd, Qm, #n

A64 : URSRA Vd.2D, Vn.2D, #n

ShiftRightLogicalRoundedAdd(Vector128<SByte>, Vector128<SByte>, Byte)

uint8x16_t vrsraq_n_u8 (uint8x16_t a, uint8x16_t b, const int n)

A32 : VRSRA.U8 Qd, Qm, #n

A64 : URSRA Vd.16B, Vn.16B, #n

ShiftRightLogicalRoundedAdd(Vector128<UInt16>, Vector128<UInt16>, Byte)

uint16x8_t vrsraq_n_u16 (uint16x8_t a, uint16x8_t b, const int n)

A32 : VRSRA.U16 Qd, Qm, #n

A64 : URSRA Vd.8H, Vn.8H, #n

ShiftRightLogicalRoundedAdd(Vector128<UInt32>, Vector128<UInt32>, Byte)

uint32x4_t vrsraq_n_u32 (uint32x4_t a, uint32x4_t b, const int n)

A32 : VRSRA.U32 Qd, Qm, #n

A64 : URSRA Vd.4S, Vn.4S, #n

ShiftRightLogicalRoundedAdd(Vector128<UInt64>, Vector128<UInt64>, Byte)

uint64x2_t vrsraq_n_u64 (uint64x2_t a, uint64x2_t b, const int n)

A32 : VRSRA.U64 Qd, Qm, #n

A64 : URSRA Vd.2D, Vn.2D, #n

ShiftRightLogicalRoundedAdd(Vector64<Byte>, Vector64<Byte>, Byte)

uint8x8_t vrsra_n_u8 (uint8x8_t a, uint8x8_t b, const int n)

A32 : VRSRA.U8 Dd, Dm, #n

A64 : URSRA Vd.8B, Vn.8B, #n

ShiftRightLogicalRoundedAdd(Vector64<Int16>, Vector64<Int16>, Byte)

uint16x4_t vrsra_n_u16 (uint16x4_t a, uint16x4_t b, const int n)

A32 : VRSRA.U16 Dd, Dm, #n

A64 : URSRA Vd.4H, Vn.4H, #n

ShiftRightLogicalRoundedAdd(Vector64<Int32>, Vector64<Int32>, Byte)

uint32x2_t vrsra_n_u32 (uint32x2_t a, uint32x2_t b, const int n)

A32 : VRSRA.U32 Dd, Dm, #n

A64 : URSRA Vd.2S, Vn.2S, #n

ShiftRightLogicalRoundedAdd(Vector64<SByte>, Vector64<SByte>, Byte)

uint8x8_t vrsra_n_u8 (uint8x8_t a, uint8x8_t b, const int n)

A32 : VRSRA.U8 Dd, Dm, #n

A64 : URSRA Vd.8B, Vn.8B, #n

ShiftRightLogicalRoundedAdd(Vector64<UInt16>, Vector64<UInt16>, Byte)

uint16x4_t vrsra_n_u16 (uint16x4_t a, uint16x4_t b, const int n)

A32 : VRSRA.U16 Dd, Dm, #n

A64 : URSRA Vd.4H, Vn.4H, #n

ShiftRightLogicalRoundedAdd(Vector64<UInt32>, Vector64<UInt32>, Byte)

uint32x2_t vrsra_n_u32 (uint32x2_t a, uint32x2_t b, const int n)

A32 : VRSRA.U32 Dd, Dm, #n

A64 : URSRA Vd.2S, Vn.2S, #n

ShiftRightLogicalRoundedAddScalar(Vector64<Int64>, Vector64<Int64>, Byte)

uint64x1_t vrsra_n_u64 (uint64x1_t a, uint64x1_t b, const int n)

A32 : VRSRA.U64 Dd, Dm, #n

A64 : URSRA Dd, Dn, #n

ShiftRightLogicalRoundedAddScalar(Vector64<UInt64>, Vector64<UInt64>, Byte)

uint64x1_t vrsra_n_u64 (uint64x1_t a, uint64x1_t b, const int n)

A32 : VRSRA.U64 Dd, Dm, #n

A64 : URSRA Dd, Dn, #n

ShiftRightLogicalRoundedNarrowingLower(Vector128<Int16>, Byte)

int8x8_t vrshrn_n_s16 (int16x8_t a, const int n)

A32 : VRSHRN.I16 Dd, Qm, #n

A64 : RSHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingLower(Vector128<Int32>, Byte)

int16x4_t vrshrn_n_s32 (int32x4_t a, const int n)

A32 : VRSHRN.I32 Dd, Qm, #n

A64 : RSHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingLower(Vector128<Int64>, Byte)

int32x2_t vrshrn_n_s64 (int64x2_t a, const int n)

A32 : VRSHRN.I64 Dd, Qm, #n

A64 : RSHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalRoundedNarrowingLower(Vector128<UInt16>, Byte)

uint8x8_t vrshrn_n_u16 (uint16x8_t a, const int n)

A32 : VRSHRN.I16 Dd, Qm, #n

A64 : RSHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingLower(Vector128<UInt32>, Byte)

uint16x4_t vrshrn_n_u32 (uint32x4_t a, const int n)

A32 : VRSHRN.I32 Dd, Qm, #n

A64 : RSHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingLower(Vector128<UInt64>, Byte)

uint32x2_t vrshrn_n_u64 (uint64x2_t a, const int n)

A32 : VRSHRN.I64 Dd, Qm, #n

A64 : RSHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalRoundedNarrowingSaturateLower(Vector128<Int16>, Byte)

uint8x8_t vqrshrn_n_u16 (uint16x8_t a, const int n)

A32 : VQRSHRN.U16 Dd, Qm, #n

A64 : UQRSHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingSaturateLower(Vector128<Int32>, Byte)

uint16x4_t vqrshrn_n_u32 (uint32x4_t a, const int n)

A32 : VQRSHRN.U32 Dd, Qm, #n

A64 : UQRSHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingSaturateLower(Vector128<Int64>, Byte)

uint32x2_t vqrshrn_n_u64 (uint64x2_t a, const int n)

A32 : VQRSHRN.U64 Dd, Qm, #n

A64 : UQRSHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalRoundedNarrowingSaturateLower(Vector128<UInt16>, Byte)

uint8x8_t vqrshrn_n_u16 (uint16x8_t a, const int n)

A32 : VQRSHRN.U16 Dd, Qm, #n

A64 : UQRSHRN Vd.8B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingSaturateLower(Vector128<UInt32>, Byte)

uint16x4_t vqrshrn_n_u32 (uint32x4_t a, const int n)

A32 : VQRSHRN.U32 Dd, Qm, #n

A64 : UQRSHRN Vd.4H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingSaturateLower(Vector128<UInt64>, Byte)

uint32x2_t vqrshrn_n_u64 (uint64x2_t a, const int n)

A32 : VQRSHRN.U64 Dd, Qm, #n

A64 : UQRSHRN Vd.2S, Vn.2D, #n

ShiftRightLogicalRoundedNarrowingSaturateUpper(Vector64<Byte>, Vector128<UInt16>, Byte)

uint8x16_t vqrshrn_high_n_u16 (uint8x8_t r, uint16x8_t a, const int n)

A32 : VQRSHRN.U16 Dd+1, Dn, #n

A64 : UQRSHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingSaturateUpper(Vector64<Int16>, Vector128<Int32>, Byte)

uint16x8_t vqrshrn_high_n_u32 (uint16x4_t r, uint32x4_t a, const int n)

A32 : VQRSHRN.U32 Dd+1, Dn, #n

A64 : UQRSHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingSaturateUpper(Vector64<Int32>, Vector128<Int64>, Byte)

uint32x4_t vqrshrn_high_n_u64 (uint32x2_t r, uint64x2_t a, const int n)

A32 : VQRSHRN.U64 Dd+1, Dn, #n

A64 : UQRSHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalRoundedNarrowingSaturateUpper(Vector64<SByte>, Vector128<Int16>, Byte)

uint8x16_t vqrshrn_high_n_u16 (uint8x8_t r, uint16x8_t a, const int n)

A32 : VQRSHRN.U16 Dd+1, Dn, #n

A64 : UQRSHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingSaturateUpper(Vector64<UInt16>, Vector128<UInt32>, Byte)

uint16x8_t vqrshrn_high_n_u32 (uint16x4_t r, uint32x4_t a, const int n)

A32 : VQRSHRN.U32 Dd+1, Dn, #n

A64 : UQRSHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingSaturateUpper(Vector64<UInt32>, Vector128<UInt64>, Byte)

uint32x4_t vqrshrn_high_n_u64 (uint32x2_t r, uint64x2_t a, const int n)

A32 : VQRSHRN.U64 Dd+1, Dn, #n

A64 : UQRSHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalRoundedNarrowingUpper(Vector64<Byte>, Vector128<UInt16>, Byte)

uint8x16_t vrshrn_high_n_u16 (uint8x8_t r, uint16x8_t a, const int n)

A32 : VRSHRN.I16 Dd+1, Qm, #n

A64 : RSHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingUpper(Vector64<Int16>, Vector128<Int32>, Byte)

int16x8_t vrshrn_high_n_s32 (int16x4_t r, int32x4_t a, const int n)

A32 : VRSHRN.I32 Dd+1, Qm, #n

A64 : RSHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingUpper(Vector64<Int32>, Vector128<Int64>, Byte)

int32x4_t vrshrn_high_n_s64 (int32x2_t r, int64x2_t a, const int n)

A32 : VRSHRN.I64 Dd+1, Qm, #n

A64 : RSHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalRoundedNarrowingUpper(Vector64<SByte>, Vector128<Int16>, Byte)

int8x16_t vrshrn_high_n_s16 (int8x8_t r, int16x8_t a, const int n)

A32 : VRSHRN.I16 Dd+1, Qm, #n

A64 : RSHRN2 Vd.16B, Vn.8H, #n

ShiftRightLogicalRoundedNarrowingUpper(Vector64<UInt16>, Vector128<UInt32>, Byte)

uint16x8_t vrshrn_high_n_u32 (uint16x4_t r, uint32x4_t a, const int n)

A32 : VRSHRN.I32 Dd+1, Qm, #n

A64 : RSHRN2 Vd.8H, Vn.4S, #n

ShiftRightLogicalRoundedNarrowingUpper(Vector64<UInt32>, Vector128<UInt64>, Byte)

uint32x4_t vrshrn_high_n_u64 (uint32x2_t r, uint64x2_t a, const int n)

A32 : VRSHRN.I64 Dd+1, Qm, #n

A64 : RSHRN2 Vd.4S, Vn.2D, #n

ShiftRightLogicalRoundedScalar(Vector64<Int64>, Byte)

uint64x1_t vrshr_n_u64 (uint64x1_t a, const int n)

A32 : VRSHR.U64 Dd, Dm, #n

A64 : URSHR Dd, Dn, #n

ShiftRightLogicalRoundedScalar(Vector64<UInt64>, Byte)

uint64x1_t vrshr_n_u64 (uint64x1_t a, const int n)

A32 : VRSHR.U64 Dd, Dm, #n

A64 : URSHR Dd, Dn, #n

ShiftRightLogicalScalar(Vector64<Int64>, Byte)

uint64x1_t vshr_n_u64 (uint64x1_t a, const int n)

A32 : VSHR.U64 Dd, Dm, #n

A64 : USHR Dd, Dn, #n

ShiftRightLogicalScalar(Vector64<UInt64>, Byte)

uint64x1_t vshr_n_u64 (uint64x1_t a, const int n)

A32 : VSHR.U64 Dd, Dm, #n

A64 : USHR Dd, Dn, #n

SignExtendWideningLower(Vector64<Int16>)

int32x4_t vmovl_s16 (int16x4_t a)

A32 : VMOVL.S16 Qd, Dm

A64 : SXTL Vd.4S, Vn.4H

SignExtendWideningLower(Vector64<Int32>)

int64x2_t vmovl_s32 (int32x2_t a)

A32 : VMOVL.S32 Qd, Dm

A64 : SXTL Vd.2D, Vn.2S

SignExtendWideningLower(Vector64<SByte>)

int16x8_t vmovl_s8 (int8x8_t a)

A32 : VMOVL.S8 Qd, Dm

A64 : SXTL Vd.8H, Vn.8B

SignExtendWideningUpper(Vector128<Int16>)

int32x4_t vmovl_high_s16 (int16x8_t a)

A32 : VMOVL.S16 Qd, Dm+1

A64 : SXTL2 Vd.4S, Vn.8H

SignExtendWideningUpper(Vector128<Int32>)

int64x2_t vmovl_high_s32 (int32x4_t a)

A32 : VMOVL.S32 Qd, Dm+1

A64 : SXTL2 Vd.2D, Vn.4S

SignExtendWideningUpper(Vector128<SByte>)

int16x8_t vmovl_high_s8 (int8x16_t a)

A32 : VMOVL.S8 Qd, Dm+1

A64 : SXTL2 Vd.8H, Vn.16B

SqrtScalar(Vector64<Double>)

float64x1_t vsqrt_f64 (float64x1_t a)

A32 : VSQRT.F64 Dd, Dm

A64 : FSQRT Dd, Dn

SqrtScalar(Vector64<Single>)

float32_t vsqrts_f32 (float32_t a)

A32 : VSQRT.F32 Sd, Sm

A64 : FSQRT Sd, Sn La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Store(Byte*, Vector128<Byte>)

void vst1q_u8 (uint8_t * ptr, uint8x16_t val)

A32 : VST1.8 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.16B }, [Xn]

Store(Byte*, Vector64<Byte>)

void vst1_u8 (uint8_t * ptr, uint8x8_t val)

A32 : VST1.8 { Dd }, [Rn]

A64 : ST1 { Vt.8B }, [Xn]

Store(Double*, Vector128<Double>)

void vst1q_f64 (float64_t * ptr, float64x2_t val)

A32 : VST1.64 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.2D }, [Xn]

Store(Double*, Vector64<Double>)

void vst1_f64 (float64_t * ptr, float64x1_t val)

A32 : VST1.64 { Dd }, [Rn]

A64 : ST1 { Vt.1D }, [Xn]

Store(Int16*, Vector128<Int16>)

void vst1q_s16 (int16_t * ptr, int16x8_t val)

A32 : VST1.16 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.8H }, [Xn]

Store(Int16*, Vector64<Int16>)

void vst1_s16 (int16_t * ptr, int16x4_t val)

A32 : VST1.16 { Dd }, [Rn]

A64 : ST1 {Vt.4H }, [Xn]

Store(Int32*, Vector128<Int32>)

void vst1q_s32 (int32_t * ptr, int32x4_t val)

A32 : VST1.32 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.4S }, [Xn]

Store(Int32*, Vector64<Int32>)

void vst1_s32 (int32_t * ptr, int32x2_t val)

A32 : VST1.32 { Dd }, [Rn]

A64 : ST1 { Vt.2S }, [Xn]

Store(Int64*, Vector128<Int64>)

void vst1q_s64 (int64_t * ptr, int64x2_t val)

A32 : VST1.64 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.2D }, [Xn]

Store(Int64*, Vector64<Int64>)

void vst1_s64 (int64_t * ptr, int64x1_t val)

A32 : VST1.64 { Dd }, [Rn]

A64 : ST1 { Vt.1D }, [Xn]

Store(SByte*, Vector128<SByte>)

void vst1q_s8 (int8_t * ptr, int8x16_t val)

A32 : VST1.8 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.16B }, [Xn]

Store(SByte*, Vector64<SByte>)

void vst1_s8 (int8_t * ptr, int8x8_t val)

A32 : VST1.8 { Dd }, [Rn]

A64 : ST1 { Vt.8B }, [Xn]

Store(Single*, Vector128<Single>)

void vst1q_f32 (float32_t * ptr, float32x4_t val)

A32 : VST1.32 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.4S }, [Xn]

Store(Single*, Vector64<Single>)

void vst1_f32 (float32_t * ptr, float32x2_t val)

A32 : VST1.32 { Dd }, [Rn]

A64 : ST1 { Vt.2S }, [Xn]

Store(UInt16*, Vector128<UInt16>)

void vst1q_u16 (uint16_t * ptr, uint16x8_t val)

A32 : VST1.16 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.8H }, [Xn]

Store(UInt16*, Vector64<UInt16>)

void vst1_u16 (uint16_t * ptr, uint16x4_t val)

A32 : VST1.16 { Dd }, [Rn]

A64 : ST1 { Vt.4H }, [Xn]

Store(UInt32*, Vector128<UInt32>)

void vst1q_u32 (uint32_t * ptr, uint32x4_t val)

A32 : VST1.32 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.4S }, [Xn]

Store(UInt32*, Vector64<UInt32>)

void vst1_u32 (uint32_t * ptr, uint32x2_t val)

A32 : VST1.32 { Dd }, [Rn]

A64 : ST1 { Vt.2S }, [Xn]

Store(UInt64*, Vector128<UInt64>)

void vst1q_u64 (uint64_t * ptr, uint64x2_t val)

A32 : VST1.64 { Dd, Dd+1 }, [Rn]

A64 : ST1 { Vt.2D }, [Xn]

Store(UInt64*, Vector64<UInt64>)

void vst1_u64 (uint64_t * ptr, uint64x1_t val)

A32 : VST1.64 { Dd }, [Rn]

A64 : ST1 { Vt.1D }, [Xn]

StoreSelectedScalar(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Byte*, Vector128<Byte>, Byte)

void vst1q_lane_u8 (uint8_t * ptr, uint8x16_t val, const int lane)

A32 : VST1.8 { Dd[index] }, [Rn]

A64 : ST1 { Vt.B }[index], [Xn]

StoreSelectedScalar(Byte*, Vector64<Byte>, Byte)

void vst1_lane_u8 (uint8_t * ptr, uint8x8_t val, const int lane)

A32 : VST1.8 { Dd[index] }, [Rn]

A64 : ST1 { Vt.B }[index], [Xn]

StoreSelectedScalar(Double*, Vector128<Double>, Byte)

void vst1q_lane_f64 (float64_t * ptr, float64x2_t val, const int lane)

A32 : VSTR.64 Dd, [Rn]

A64 : ST1 { Vt.D }[index], [Xn]

StoreSelectedScalar(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Int16*, Vector128<Int16>, Byte)

void vst1q_lane_s16 (int16_t * ptr, int16x8_t val, const int lane)

A32 : VST1.16 { Dd[index] }, [Rn]

A64 : ST1 { Vt.H }[index], [Xn]

StoreSelectedScalar(Int16*, Vector64<Int16>, Byte)

void vst1_lane_s16 (int16_t * ptr, int16x4_t val, const int lane)

A32 : VST1.16 { Dd[index] }, [Rn]

A64 : ST1 { Vt.H }[index], [Xn]

StoreSelectedScalar(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Int32*, Vector128<Int32>, Byte)

void vst1q_lane_s32 (int32_t * ptr, int32x4_t val, const int lane)

A32 : VST1.32 { Dd[index] }, [Rn]

A64 : ST1 { Vt.S }[index], [Xn]

StoreSelectedScalar(Int32*, Vector64<Int32>, Byte)

void vst1_lane_s32 (int32_t * ptr, int32x2_t val, const int lane)

A32 : VST1.32 { Dd[index] }, [Rn]

A64 : ST1 { Vt.S }[index], [Xn]

StoreSelectedScalar(Int64*, Vector128<Int64>, Byte)

void vst1q_lane_s64 (int64_t * ptr, int64x2_t val, const int lane)

A32 : VSTR.64 Dd, [Rn]

A64 : ST1 { Vt.D }[index], [Xn]

StoreSelectedScalar(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(SByte*, Vector128<SByte>, Byte)

void vst1q_lane_s8 (int8_t * ptr, int8x16_t val, const int lane)

A32 : VST1.8 { Dd[index] }, [Rn]

A64 : ST1 { Vt.B }[index], [Xn]

StoreSelectedScalar(SByte*, Vector64<SByte>, Byte)

void vst1_lane_s8 (int8_t * ptr, int8x8_t val, const int lane)

A32 : VST1.8 { Dd[index] }, [Rn]

A64 : ST1 { Vt.B }[index], [Xn]

StoreSelectedScalar(Single*, ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>,Vector64<Single>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Single*, ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Single*, ValueTuple<Vector64<Single>,Vector64<Single>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(Single*, Vector128<Single>, Byte)

void vst1q_lane_f32 (float32_t * ptr, float32x4_t val, const int lane)

A32 : VST1.32 { Dd[index] }, [Rn]

A64 : ST1 { Vt.S }[index], [Xn]

StoreSelectedScalar(Single*, Vector64<Single>, Byte)

void vst1_lane_f32 (float32_t * ptr, float32x2_t val, const int lane)

A32 : VST1.32 { Dd[index] }, [Rn]

A64 : ST1 { Vt.S }[index], [Xn]

StoreSelectedScalar(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(UInt16*, Vector128<UInt16>, Byte)

void vst1q_lane_u16 (uint16_t * ptr, uint16x8_t val, const int lane)

A32 : VST1.16 { Dd[index] }, [Rn]

A64 : ST1 { Vt.H }[index], [Xn]

StoreSelectedScalar(UInt16*, Vector64<UInt16>, Byte)

void vst1_lane_u16 (uint16_t * ptr, uint16x4_t val, const int lane)

A32 : VST1.16 { Dd[index] }, [Rn]

A64 : ST1 { Vt.H }[index], [Xn]

StoreSelectedScalar(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>>, Byte)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreSelectedScalar(UInt32*, Vector128<UInt32>, Byte)

void vst1q_lane_u32 (uint32_t * ptr, uint32x4_t val, const int lane)

A32 : VST1.32 { Dd[index] }, [Rn]

A64 : ST1 { Vt.S }[index], [Xn]

StoreSelectedScalar(UInt32*, Vector64<UInt32>, Byte)

void vst1_lane_u32 (uint32_t * ptr, uint32x2_t val, const int lane)

A32 : VST1.32 { Dd[index] }, [Rn]

A64 : ST1 { Vt.S }[index], [Xn]

StoreSelectedScalar(UInt64*, Vector128<UInt64>, Byte)

void vst1q_lane_u64 (uint64_t * ptr, uint64x2_t val, const int lane)

A32 : VSTR.64 Dd, [Rn]

A64 : ST1 { Vt.D }[index], [Xn]

StoreVector64x2(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2(Single*, ValueTuple<Vector64<Single>,Vector64<Single>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2AndZip(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2AndZip(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2AndZip(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2AndZip(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2AndZip(Single*, ValueTuple<Vector64<Single>,Vector64<Single>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2AndZip(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x2AndZip(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3(Single*, ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3AndZip(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3AndZip(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3AndZip(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3AndZip(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3AndZip(Single*, ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3AndZip(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x3AndZip(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4(Single*, ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>,Vector64<Single>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4AndZip(Byte*, ValueTuple<Vector64<Byte>,Vector64<Byte>,Vector64<Byte>,Vector64<Byte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4AndZip(Int16*, ValueTuple<Vector64<Int16>,Vector64<Int16>,Vector64<Int16>,Vector64<Int16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4AndZip(Int32*, ValueTuple<Vector64<Int32>,Vector64<Int32>,Vector64<Int32>,Vector64<Int32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4AndZip(SByte*, ValueTuple<Vector64<SByte>,Vector64<SByte>,Vector64<SByte>,Vector64<SByte>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4AndZip(Single*, ValueTuple<Vector64<Single>,Vector64<Single>,Vector64<Single>,Vector64<Single>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4AndZip(UInt16*, ValueTuple<Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>,Vector64<UInt16>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

StoreVector64x4AndZip(UInt32*, ValueTuple<Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>,Vector64<UInt32>>)

Cette classe fournit un accès aux instructions matérielles ARM AdvSIMD via des intrinsèques.

Subtract(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vsubq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VSUB.I8 Qd, Qn, Qm

A64 : SUB Vd.16B, Vn.16B, Vm.16B

Subtract(Vector128<Int16>, Vector128<Int16>)

int16x8_t vsubq_s16 (int16x8_t a, int16x8_t b)

A32 : VSUB.I16 Qd, Qn, Qm

A64 : SUB Vd.8H, Vn.8H, Vm.8H

Subtract(Vector128<Int32>, Vector128<Int32>)

int32x4_t vsubq_s32 (int32x4_t a, int32x4_t b)

A32 : VSUB.I32 Qd, Qn, Qm

A64 : SUB Vd.4S, Vn.4S, Vm.4S

Subtract(Vector128<Int64>, Vector128<Int64>)

int64x2_t vsubq_s64 (int64x2_t a, int64x2_t b)

A32 : VSUB.I64 Qd, Qn, Qm

A64 : SUB Vd.2D, Vn.2D, Vm.2D

Subtract(Vector128<SByte>, Vector128<SByte>)

int8x16_t vsubq_s8 (int8x16_t a, int8x16_t b)

A32 : VSUB.I8 Qd, Qn, Qm

A64 : SUB Vd.16B, Vn.16B, Vm.16B

Subtract(Vector128<Single>, Vector128<Single>)

float32x4_t vsubq_f32 (float32x4_t a, float32x4_t b)

A32 : VSUB.F32 Qd, Qn, Qm

A64 : FSUB Vd.4S, Vn.4S, Vm.4S

Subtract(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vsubq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VSUB.I16 Qd, Qn, Qm

A64 : SUB Vd.8H, Vn.8H, Vm.8H

Subtract(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vsubq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VSUB.I32 Qd, Qn, Qm

A64 : SUB Vd.4S, Vn.4S, Vm.4S

Subtract(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vsubq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VSUB.I64 Qd, Qn, Qm

A64 : SUB Vd.2D, Vn.2D, Vm.2D

Subtract(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vsub_u8 (uint8x8_t a, uint8x8_t b)

A32 : VSUB.I8 Dd, Dn, Dm

A64 : SUB Vd.8B, Vn.8B, Vm.8B

Subtract(Vector64<Int16>, Vector64<Int16>)

int16x4_t vsub_s16 (int16x4_t a, int16x4_t b)

A32 : VSUB.I16 Dd, Dn, Dm

A64 : SUB Vd.4H, Vn.4H, Vm.4H

Subtract(Vector64<Int32>, Vector64<Int32>)

int32x2_t vsub_s32 (int32x2_t a, int32x2_t b)

A32 : VSUB.I32 Dd, Dn, Dm

A64 : SUB Vd.2S, Vn.2S, Vm.2S

Subtract(Vector64<SByte>, Vector64<SByte>)

int8x8_t vsub_s8 (int8x8_t a, int8x8_t b)

A32 : VSUB.I8 Dd, Dn, Dm

A64 : SUB Vd.8B, Vn.8B, Vm.8B

Subtract(Vector64<Single>, Vector64<Single>)

float32x2_t vsub_f32 (float32x2_t a, float32x2_t b)

A32 : VSUB.F32 Dd, Dn, Dm

A64 : FSUB Vd.2S, Vn.2S, Vm.2S

Subtract(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vsub_u16 (uint16x4_t a, uint16x4_t b)

A32 : VSUB.I16 Dd, Dn, Dm

A64 : SUB Vd.4H, Vn.4H, Vm.4H

Subtract(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vsub_u32 (uint32x2_t a, uint32x2_t b)

A32 : VSUB.I32 Dd, Dn, Dm

A64 : SUB Vd.2S, Vn.2S, Vm.2S

SubtractHighNarrowingLower(Vector128<Int16>, Vector128<Int16>)

int8x8_t vsubhn_s16 (int16x8_t a, int16x8_t b)

A32 : VSUBHN.I16 Dd, Qn, Qm

A64 : SUBHN Vd.8B, Vn.8H, Vm.8H

SubtractHighNarrowingLower(Vector128<Int32>, Vector128<Int32>)

int16x4_t vsubhn_s32 (int32x4_t a, int32x4_t b)

A32 : VSUBHN.I32 Dd, Qn, Qm

A64 : SUBHN Vd.4H, Vn.4S, Vm.4S

SubtractHighNarrowingLower(Vector128<Int64>, Vector128<Int64>)

int32x2_t vsubhn_s64 (int64x2_t a, int64x2_t b)

A32 : VSUBHN.I64 Dd, Qn, Qm

A64 : SUBHN Vd.2S, Vn.2D, Vm.2D

SubtractHighNarrowingLower(Vector128<UInt16>, Vector128<UInt16>)

uint8x8_t vsubhn_u16 (uint16x8_t a, uint16x8_t b)

A32 : VSUBHN.I16 Dd, Qn, Qm

A64 : SUBHN Vd.8B, Vn.8H, Vm.8H

SubtractHighNarrowingLower(Vector128<UInt32>, Vector128<UInt32>)

uint16x4_t vsubhn_u32 (uint32x4_t a, uint32x4_t b)

A32 : VSUBHN.I32 Dd, Qn, Qm

A64 : SUBHN Vd.4H, Vn.4S, Vm.4S

SubtractHighNarrowingLower(Vector128<UInt64>, Vector128<UInt64>)

uint32x2_t vsubhn_u64 (uint64x2_t a, uint64x2_t b)

A32 : VSUBHN.I64 Dd, Qn, Qm

A64 : SUBHN Vd.2S, Vn.2D, Vm.2D

SubtractHighNarrowingUpper(Vector64<Byte>, Vector128<UInt16>, Vector128<UInt16>)

uint8x16_t vsubhn_high_u16 (uint8x8_t r, uint16x8_t a, uint16x8_t b)

A32 : VSUBHN.I16 Dd+1, Qn, Qm

A64 : SUBHN2 Vd.16B, Vn.8H, Vm.8H

SubtractHighNarrowingUpper(Vector64<Int16>, Vector128<Int32>, Vector128<Int32>)

int16x8_t vsubhn_high_s32 (int16x4_t r, int32x4_t a, int32x4_t b)

A32 : VSUBHN.I32 Dd+1, Qn, Qm

A64 : SUBHN2 Vd.8H, Vn.4S, Vm.4S

SubtractHighNarrowingUpper(Vector64<Int32>, Vector128<Int64>, Vector128<Int64>)

int32x4_t vsubhn_high_s64 (int32x2_t r, int64x2_t a, int64x2_t b)

A32 : VSUBHN.I64 Dd+1, Qn, Qm

A64 : SUBHN2 Vd.4S, Vn.2D, Vm.2D

SubtractHighNarrowingUpper(Vector64<SByte>, Vector128<Int16>, Vector128<Int16>)

int8x16_t vsubhn_high_s16 (int8x8_t r, int16x8_t a, int16x8_t b)

A32 : VSUBHN.I16 Dd+1, Qn, Qm

A64 : SUBHN2 Vd.16B, Vn.8H, Vm.8H

SubtractHighNarrowingUpper(Vector64<UInt16>, Vector128<UInt32>, Vector128<UInt32>)

uint16x8_t vsubhn_high_u32 (uint16x4_t r, uint32x4_t a, uint32x4_t b)

A32 : VSUBHN.I32 Dd+1, Qn, Qm

A64 : SUBHN2 Vd.8H, Vn.4S, Vm.4S

SubtractHighNarrowingUpper(Vector64<UInt32>, Vector128<UInt64>, Vector128<UInt64>)

uint32x4_t vsubhn_high_u64 (uint32x2_t r, uint64x2_t a, uint64x2_t b)

A32 : VSUBHN.I64 Dd+1, Qn, Qm

A64 : SUBHN2 Vd.4S, Vn.2D, Vm.2D

SubtractRoundedHighNarrowingLower(Vector128<Int16>, Vector128<Int16>)

int8x8_t vrsubhn_s16 (int16x8_t a, int16x8_t b)

A32 : VRSUBHN.I16 Dd, Qn, Qm

A64 : RSUBHN Vd.8B, Vn.8H, Vm.8H

SubtractRoundedHighNarrowingLower(Vector128<Int32>, Vector128<Int32>)

int16x4_t vrsubhn_s32 (int32x4_t a, int32x4_t b)

A32 : VRSUBHN.I32 Dd, Qn, Qm

A64 : RSUBHN Vd.4H, Vn.4S, Vm.4S

SubtractRoundedHighNarrowingLower(Vector128<Int64>, Vector128<Int64>)

int32x2_t vrsubhn_s64 (int64x2_t a, int64x2_t b)

A32 : VRSUBHN.I64 Dd, Qn, Qm

A64 : RSUBHN Vd.2S, Vn.2D, Vm.2D

SubtractRoundedHighNarrowingLower(Vector128<UInt16>, Vector128<UInt16>)

uint8x8_t vrsubhn_u16 (uint16x8_t a, uint16x8_t b)

A32 : VRSUBHN.I16 Dd, Qn, Qm

A64 : RSUBHN Vd.8B, Vn.8H, Vm.8H

SubtractRoundedHighNarrowingLower(Vector128<UInt32>, Vector128<UInt32>)

uint16x4_t vrsubhn_u32 (uint32x4_t a, uint32x4_t b)

A32 : VRSUBHN.I32 Dd, Qn, Qm

A64 : RSUBHN Vd.4H, Vn.4S, Vm.4S

SubtractRoundedHighNarrowingLower(Vector128<UInt64>, Vector128<UInt64>)

uint32x2_t vrsubhn_u64 (uint64x2_t a, uint64x2_t b)

A32 : VRSUBHN.I64 Dd, Qn, Qm

A64 : RSUBHN Vd.2S, Vn.2D, Vm.2D

SubtractRoundedHighNarrowingUpper(Vector64<Byte>, Vector128<UInt16>, Vector128<UInt16>)

uint8x16_t vrsubhn_high_u16 (uint8x8_t r, uint16x8_t a, uint16x8_t b)

A32 : VRSUBHN.I16 Dd+1, Qn, Qm

A64 : RSUBHN2 Vd.16B, Vn.8H, Vm.8H

SubtractRoundedHighNarrowingUpper(Vector64<Int16>, Vector128<Int32>, Vector128<Int32>)

int16x8_t vrsubhn_high_s32 (int16x4_t r, int32x4_t a, int32x4_t b)

A32 : VRSUBHN.I32 Dd+1, Qn, Qm

A64 : RSUBHN2 Vd.8H, Vn.4S, Vm.4S

SubtractRoundedHighNarrowingUpper(Vector64<Int32>, Vector128<Int64>, Vector128<Int64>)

int32x4_t vrsubhn_high_s64 (int32x2_t r, int64x2_t a, int64x2_t b)

A32 : VRSUBHN.I64 Dd+1, Qn, Qm

A64 : RSUBHN2 Vd.4S, Vn.2D, Vm.2D

SubtractRoundedHighNarrowingUpper(Vector64<SByte>, Vector128<Int16>, Vector128<Int16>)

int8x16_t vrsubhn_high_s16 (int8x8_t r, int16x8_t a, int16x8_t b)

A32 : VRSUBHN.I16 Dd+1, Qn, Qm

A64 : RSUBHN2 Vd.16B, Vn.8H, Vm.8H

SubtractRoundedHighNarrowingUpper(Vector64<UInt16>, Vector128<UInt32>, Vector128<UInt32>)

uint16x8_t vrsubhn_high_u32 (uint16x4_t r, uint32x4_t a, uint32x4_t b)

A32 : VRSUBHN.I32 Dd+1, Qn, Qm

A64 : RSUBHN2 Vd.8H, Vn.4S, Vm.4S

SubtractRoundedHighNarrowingUpper(Vector64<UInt32>, Vector128<UInt64>, Vector128<UInt64>)

uint32x4_t vrsubhn_high_u64 (uint32x2_t r, uint64x2_t a, uint64x2_t b)

A32 : VRSUBHN.I64 Dd+1, Qn, Qm

A64 : RSUBHN2 Vd.4S, Vn.2D, Vm.2D

SubtractSaturate(Vector128<Byte>, Vector128<Byte>)

uint8x16_t vqsubq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VQSUB.U8 Qd, Qn, Qm

A64 : UQSUB Vd.16B, Vn.16B, Vm.16B

SubtractSaturate(Vector128<Int16>, Vector128<Int16>)

int16x8_t vqsubq_s16 (int16x8_t a, int16x8_t b)

A32 : VQSUB.S16 Qd, Qn, Qm

A64 : SQSUB Vd.8H, Vn.8H, Vm.8H

SubtractSaturate(Vector128<Int32>, Vector128<Int32>)

int32x4_t vqsubq_s32 (int32x4_t a, int32x4_t b)

A32 : VQSUB.S32 Qd, Qn, Qm

A64 : SQSUB Vd.4S, Vn.4S, Vm.4S

SubtractSaturate(Vector128<Int64>, Vector128<Int64>)

int64x2_t vqsubq_s64 (int64x2_t a, int64x2_t b)

A32 : VQSUB.S64 Qd, Qn, Qm

A64 : SQSUB Vd.2D, Vn.2D, Vm.2D

SubtractSaturate(Vector128<SByte>, Vector128<SByte>)

int8x16_t vqsubq_s8 (int8x16_t a, int8x16_t b)

A32 : VQSUB.S8 Qd, Qn, Qm

A64 : SQSUB Vd.16B, Vn.16B, Vm.16B

SubtractSaturate(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t vqsubq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VQSUB.U16 Qd, Qn, Qm

A64 : UQSUB Vd.8H, Vn.8H, Vm.8H

SubtractSaturate(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t vqsubq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VQSUB.U32 Qd, Qn, Qm

A64 : UQSUB Vd.4S, Vn.4S, Vm.4S

SubtractSaturate(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t vqsubq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VQSUB.U64 Qd, Qn, Qm

A64 : UQSUB Vd.2D, Vn.2D, Vm.2D

SubtractSaturate(Vector64<Byte>, Vector64<Byte>)

uint8x8_t vqsub_u8 (uint8x8_t a, uint8x8_t b)

A32 : VQSUB.U8 Dd, Dn, Dm

A64 : UQSUB Vd.8B, Vn.8B, Vm.8B

SubtractSaturate(Vector64<Int16>, Vector64<Int16>)

int16x4_t vqsub_s16 (int16x4_t a, int16x4_t b)

A32 : VQSUB.S16 Dd, Dn, Dm

A64 : SQSUB Vd.4H, Vn.4H, Vm.4H

SubtractSaturate(Vector64<Int32>, Vector64<Int32>)

int32x2_t vqsub_s32 (int32x2_t a, int32x2_t b)

A32 : VQSUB.S32 Dd, Dn, Dm

A64 : SQSUB Vd.2S, Vn.2S, Vm.2S

SubtractSaturate(Vector64<SByte>, Vector64<SByte>)

int8x8_t vqsub_s8 (int8x8_t a, int8x8_t b)

A32 : VQSUB.S8 Dd, Dn, Dm

A64 : SQSUB Vd.8B, Vn.8B, Vm.8B

SubtractSaturate(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t vqsub_u16 (uint16x4_t a, uint16x4_t b)

A32 : VQSUB.U16 Dd, Dn, Dm

A64 : UQSUB Vd.4H, Vn.4H, Vm.4H

SubtractSaturate(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t vqsub_u32 (uint32x2_t a, uint32x2_t b)

A32 : VQSUB.U32 Dd, Dn, Dm

A64 : UQSUB Vd.2S, Vn.2S, Vm.2S

SubtractSaturateScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vqsub_s64 (int64x1_t a, int64x1_t b)

A32 : VQSUB.S64 Dd, Dn, Dm

A64 : SQSUB Dd, Dn, Dm

SubtractSaturateScalar(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vqsub_u64 (uint64x1_t a, uint64x1_t b)

A32 : VQSUB.U64 Dd, Dn, Dm

A64 : UQSUB Dd, Dn, Dm

SubtractScalar(Vector64<Double>, Vector64<Double>)

float64x1_t vsub_f64 (float64x1_t a, float64x1_t b)

A32 : VSUB.F64 Dd, Dn, Dm

A64 : FSUB Dd, Dn, Dm

SubtractScalar(Vector64<Int64>, Vector64<Int64>)

int64x1_t vsub_s64 (int64x1_t a, int64x1_t b)

A32 : VSUB.I64 Dd, Dn, Dm

A64 : SUB Dd, Dn, Dm

SubtractScalar(Vector64<Single>, Vector64<Single>)

float32_t vsubs_f32 (float32_t a, float32_t b)

A32 : VSUB.F32 Sd, Sn, Sm

A64 : FSUB Sd, Sn, Sm La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

SubtractScalar(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t vsub_u64 (uint64x1_t a, uint64x1_t b)

A32 : VSUB.I64 Dd, Dn, Dm

A64 : SUB Dd, Dn, Dm

SubtractWideningLower(Vector128<Int16>, Vector64<SByte>)

int16x8_t vsubw_s8 (int16x8_t a, int8x8_t b)

A32 : VSUBW.S8 Qd, Qn, Dm

A64 : SSUBW Vd.8H, Vn.8H, Vm.8B

SubtractWideningLower(Vector128<Int32>, Vector64<Int16>)

int32x4_t vsubw_s16 (int32x4_t a, int16x4_t b)

A32 : VSUBW.S16 Qd, Qn, Dm

A64 : SSUBW Vd.4S, Vn.4S, Vm.4H

SubtractWideningLower(Vector128<Int64>, Vector64<Int32>)

int64x2_t vsubw_s32 (int64x2_t a, int32x2_t b)

A32 : VSUBW.S32 Qd, Qn, Dm

A64 : SSUBW Vd.2D, Vn.2D, Vm.2S

SubtractWideningLower(Vector128<UInt16>, Vector64<Byte>)

uint16x8_t vsubw_u8 (uint16x8_t a, uint8x8_t b)

A32 : VSUBW.U8 Qd, Qn, Dm

A64 : USUBW Vd.8H, Vn.8H, Vm.8B

SubtractWideningLower(Vector128<UInt32>, Vector64<UInt16>)

uint32x4_t vsubw_u16 (uint32x4_t a, uint16x4_t b)

A32 : VSUBW.U16 Qd, Qn, Dm

A64 : USUBW Vd.4S, Vn.4S, Vm.4H

SubtractWideningLower(Vector128<UInt64>, Vector64<UInt32>)

uint64x2_t vsubw_u32 (uint64x2_t a, uint32x2_t b)

A32 : VSUBW.U32 Qd, Qn, Dm

A64 : USUBW Vd.2D, Vn.2D, Vm.2S

SubtractWideningLower(Vector64<Byte>, Vector64<Byte>)

uint16x8_t vsubl_u8 (uint8x8_t a, uint8x8_t b)

A32 : VSUBL.U8 Qd, Dn, Dm

A64 : USUBL Vd.8H, Vn.8B, Vm.8B

SubtractWideningLower(Vector64<Int16>, Vector64<Int16>)

int32x4_t vsubl_s16 (int16x4_t a, int16x4_t b)

A32 : VSUBL.S16 Qd, Dn, Dm

A64 : SSUBL Vd.4S, Vn.4H, Vm.4H

SubtractWideningLower(Vector64<Int32>, Vector64<Int32>)

int64x2_t vsubl_s32 (int32x2_t a, int32x2_t b)

A32 : VSUBL.S32 Qd, Dn, Dm

A64 : SSUBL Vd.2D, Vn.2S, Vm.2S

SubtractWideningLower(Vector64<SByte>, Vector64<SByte>)

int16x8_t vsubl_s8 (int8x8_t a, int8x8_t b)

A32 : VSUBL.S8 Qd, Dn, Dm

A64 : SSUBL Vd.8H, Vn.8B, Vm.8B

SubtractWideningLower(Vector64<UInt16>, Vector64<UInt16>)

uint32x4_t vsubl_u16 (uint16x4_t a, uint16x4_t b)

A32 : VSUBL.U16 Qd, Dn, Dm

A64 : USUBL Vd.4S, Vn.4H, Vm.4H

SubtractWideningLower(Vector64<UInt32>, Vector64<UInt32>)

uint64x2_t vsubl_u32 (uint32x2_t a, uint32x2_t b)

A32 : VSUBL.U32 Qd, Dn, Dm

A64 : USUBL Vd.2D, Vn.2S, Vm.2S

SubtractWideningUpper(Vector128<Byte>, Vector128<Byte>)

uint16x8_t vsubl_high_u8 (uint8x16_t a, uint8x16_t b)

A32 : VSUBL.U8 Qd, Dn+1, Dm+1

A64 : USUBL2 Vd.8H, Vn.16B, Vm.16B

SubtractWideningUpper(Vector128<Int16>, Vector128<Int16>)

int32x4_t vsubl_high_s16 (int16x8_t a, int16x8_t b)

A32 : VSUBL.S16 Qd, Dn+1, Dm+1

A64 : SSUBL2 Vd.4S, Vn.8H, Vm.8H

SubtractWideningUpper(Vector128<Int16>, Vector128<SByte>)

int16x8_t vsubw_high_s8 (int16x8_t a, int8x16_t b)

A32 : VSUBW.S8 Qd, Qn, Dm+1

A64 : SSUBW2 Vd.8H, Vn.8H, Vm.16B

SubtractWideningUpper(Vector128<Int32>, Vector128<Int16>)

int32x4_t vsubw_high_s16 (int32x4_t a, int16x8_t b)

A32 : VSUBW.S16 Qd, Qn, Dm+1

A64 : SSUBW2 Vd.4S, Vn.4S, Vm.8H

SubtractWideningUpper(Vector128<Int32>, Vector128<Int32>)

int64x2_t vsubl_high_s32 (int32x4_t a, int32x4_t b)

A32 : VSUBL.S32 Qd, Dn+1, Dm+1

A64 : SSUBL2 Vd.2D, Vn.4S, Vm.4S

SubtractWideningUpper(Vector128<Int64>, Vector128<Int32>)

int64x2_t vsubw_high_s32 (int64x2_t a, int32x4_t b)

A32 : VSUBW.S32 Qd, Qn, Dm+1

A64 : SSUBW2 Vd.2D, Vn.2D, Vm.4S

SubtractWideningUpper(Vector128<SByte>, Vector128<SByte>)

int16x8_t vsubl_high_s8 (int8x16_t a, int8x16_t b)

A32 : VSUBL.S8 Qd, Dn+1, Dm+1

A64 : SSUBL2 Vd.8H, Vn.16B, Vm.16B

SubtractWideningUpper(Vector128<UInt16>, Vector128<Byte>)

uint16x8_t vsubw_high_u8 (uint16x8_t a, uint8x16_t b)

A32 : VSUBW.U8 Qd, Qn, Dm+1

A64 : USUBW2 Vd.8H, Vn.8H, Vm.16B

SubtractWideningUpper(Vector128<UInt16>, Vector128<UInt16>)

uint32x4_t vsubl_high_u16 (uint16x8_t a, uint16x8_t b)

A32 : VSUBL.U16 Qd, Dn+1, Dm+1

A64 : USUBL2 Vd.4S, Vn.8H, Vm.8H

SubtractWideningUpper(Vector128<UInt32>, Vector128<UInt16>)

uint32x4_t vsubw_high_u16 (uint32x4_t a, uint16x8_t b)

A32 : VSUBW.U16 Qd, Qn, Dm+1

A64 : USUBW2 Vd.4S, Vn.4S, Vm.8H

SubtractWideningUpper(Vector128<UInt32>, Vector128<UInt32>)

uint64x2_t vsubl_high_u32 (uint32x4_t a, uint32x4_t b)

A32 : VSUBL.U32 Qd, Dn+1, Dm+1

A64 : USUBL2 Vd.2D, Vn.4S, Vm.4S

SubtractWideningUpper(Vector128<UInt64>, Vector128<UInt32>)

uint64x2_t vsubw_high_u32 (uint64x2_t a, uint32x4_t b)

A32 : VSUBW.U32 Qd, Qn, Dm+1

A64 : USUBW2 Vd.2D, Vn.2D, Vm.4S

ToString()

Retourne une chaîne qui représente l'objet actuel.

(Hérité de Object)
VectorTableLookup(ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector64<Byte>)

uint8x8_t vqtbl4q_u8(uint8x16x4_t t, uint8x8_t idx)

A64 : TBL Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.8B

VectorTableLookup(ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector64<Byte>)

uint8x8_t vqtbl3q_u8(uint8x16x3_t t, uint8x8_t idx)

A64 : TBL Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.8B

VectorTableLookup(ValueTuple<Vector128<Byte>,Vector128<Byte>>, Vector64<Byte>)

uint8x8_t vqtbl2q_u8(uint8x16x2_t t, uint8x8_t idx)

A64 : TBL Vd.8B, {Vn.16B, Vn+1.16B}, Vm.8B

VectorTableLookup(ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector64<SByte>)

int8x8_t vqtbl4q_u8(int8x16x4_t t, uint8x8_t idx)

A64 : TBL Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.8B

VectorTableLookup(ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector64<SByte>)

int8x8_t vqtbl3q_u8(int8x16x3_t t, uint8x8_t idx)

A64 : TBL Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.8B

VectorTableLookup(ValueTuple<Vector128<SByte>,Vector128<SByte>>, Vector64<SByte>)

int8x8_t vqtbl2q_u8(int8x16x2_t t, uint8x8_t idx)

A64 : TBL Vd.8B, {Vn.16B, Vn+1.16B}, Vm.8B

VectorTableLookup(Vector128<Byte>, Vector64<Byte>)

uint8x8_t vqvtbl1_u8(uint8x16_t t, uint8x8_t idx)

A32 : VTBL Dd, {Dn, Dn+1}, Dm

A64 : TBL Vd.8B, {Vn.16B}, Vm.8B

VectorTableLookup(Vector128<SByte>, Vector64<SByte>)

int8x8_t vqvtbl1_s8(int8x16_t t, uint8x8_t idx)

A32 : VTBL Dd, {Dn, Dn+1}, Dm

A64 : TBL Vd.8B, {Vn.16B}, Vm.8B

VectorTableLookupExtension(Vector64<Byte>, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector64<Byte>)

uint8x8_t vqtbx4q_u8(uint8x16x4_t t, uint8x8_t idx)

A64 : TBX Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.8B

VectorTableLookupExtension(Vector64<Byte>, ValueTuple<Vector128<Byte>,Vector128<Byte>,Vector128<Byte>>, Vector64<Byte>)

uint8x8_t vqtbx3q_u8(uint8x16x3_t t, uint8x8_t idx)

A64 : TBX Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.8B

VectorTableLookupExtension(Vector64<Byte>, ValueTuple<Vector128<Byte>,Vector128<Byte>>, Vector64<Byte>)

uint8x8_t vqtbx2q_u8(uint8x16x2_t t, uint8x8_t idx)

A64 : TBX Vd.8B, {Vn.16B, Vn+1.16B}, Vm.8B

VectorTableLookupExtension(Vector64<Byte>, Vector128<Byte>, Vector64<Byte>)

uint8x8_t vqvtbx1_u8(uint8x8_t r, uint8x16_t t, uint8x8_t idx)

A32 : VTBX Dd, {Dn, Dn+1}, Dm

A64 : TBX Vd.8B, {Vn.16B}, Vm.8B

VectorTableLookupExtension(Vector64<SByte>, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector64<SByte>)

int8x8_t vqtbx4q_u8(int8x16x4_t t, uint8x8_t idx)

A64 : TBX Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B, Vn+3.16B}, Vm.8B

VectorTableLookupExtension(Vector64<SByte>, ValueTuple<Vector128<SByte>,Vector128<SByte>,Vector128<SByte>>, Vector64<SByte>)

int8x8_t vqtbx3q_u8(int8x16x3_t t, uint8x8_t idx)

A64 : TBX Vd.8B, {Vn.16B, Vn+1.16B, Vn+2.16B}, Vm.8B

VectorTableLookupExtension(Vector64<SByte>, ValueTuple<Vector128<SByte>,Vector128<SByte>>, Vector64<SByte>)

int8x8_t vqtbx2q_u8(int8x16x2_t t, uint8x8_t idx)

A64 : TBX Vd.8B, {Vn.16B, Vn+1.16B}, Vm.8B

VectorTableLookupExtension(Vector64<SByte>, Vector128<SByte>, Vector64<SByte>)

int8x8_t vqvtbx1_s8(int8x8_t r, int8x16_t t, uint8x8_t idx)

A32 : VTBX Dd, {Dn, Dn+1}, Dm

A64 : TBX Vd.8B, {Vn.16B}, Vm.8B

Xor(Vector128<Byte>, Vector128<Byte>)

uint8x16_t veorq_u8 (uint8x16_t a, uint8x16_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector128<Double>, Vector128<Double>)

float64x2_t veorq_f64 (float64x2_t a, float64x2_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Xor(Vector128<Int16>, Vector128<Int16>)

int16x8_t veorq_s16 (int16x8_t a, int16x8_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector128<Int32>, Vector128<Int32>)

int32x4_t veorq_s32 (int32x4_t a, int32x4_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector128<Int64>, Vector128<Int64>)

int64x2_t veorq_s64 (int64x2_t a, int64x2_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector128<SByte>, Vector128<SByte>)

int8x16_t veorq_s8 (int8x16_t a, int8x16_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector128<Single>, Vector128<Single>)

float32x4_t veorq_f32 (float32x4_t a, float32x4_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Xor(Vector128<UInt16>, Vector128<UInt16>)

uint16x8_t veorq_u16 (uint16x8_t a, uint16x8_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector128<UInt32>, Vector128<UInt32>)

uint32x4_t veorq_u32 (uint32x4_t a, uint32x4_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector128<UInt64>, Vector128<UInt64>)

uint64x2_t veorq_u64 (uint64x2_t a, uint64x2_t b)

A32 : VEOR Qd, Qn, Qm

A64 : EOR Vd.16B, Vn.16B, Vm.16B

Xor(Vector64<Byte>, Vector64<Byte>)

uint8x8_t veor_u8 (uint8x8_t a, uint8x8_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

Xor(Vector64<Double>, Vector64<Double>)

float64x1_t veor_f64 (float64x1_t a, float64x1_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Xor(Vector64<Int16>, Vector64<Int16>)

int16x4_t veor_s16 (int16x4_t a, int16x4_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

Xor(Vector64<Int32>, Vector64<Int32>)

int32x2_t veor_s32 (int32x2_t a, int32x2_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

Xor(Vector64<Int64>, Vector64<Int64>)

int64x1_t veor_s64 (int64x1_t a, int64x1_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

Xor(Vector64<SByte>, Vector64<SByte>)

int8x8_t veor_s8 (int8x8_t a, int8x8_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

Xor(Vector64<Single>, Vector64<Single>)

float32x2_t veor_f32 (float32x2_t a, float32x2_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B La signature native ci-dessus n’existe pas. Nous fournissons cette surcharge supplémentaire pour la cohérence avec les autres API scalaires.

Xor(Vector64<UInt16>, Vector64<UInt16>)

uint16x4_t veor_u16 (uint16x4_t a, uint16x4_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

Xor(Vector64<UInt32>, Vector64<UInt32>)

uint32x2_t veor_u32 (uint32x2_t a, uint32x2_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

Xor(Vector64<UInt64>, Vector64<UInt64>)

uint64x1_t veor_u64 (uint64x1_t a, uint64x1_t b)

A32 : VEOR Dd, Dn, Dm

A64 : EOR Vd.8B, Vn.8B, Vm.8B

ZeroExtendWideningLower(Vector64<Byte>)

uint16x8_t vmovl_u8 (uint8x8_t a)

A32 : VMOVL.U8 Qd, Dm

A64 : UXTL Vd.8H, Vn.8B

ZeroExtendWideningLower(Vector64<Int16>)

uint32x4_t vmovl_u16 (uint16x4_t a)

A32 : VMOVL.U16 Qd, Dm

A64 : UXTL Vd.4S, Vn.4H

ZeroExtendWideningLower(Vector64<Int32>)

uint64x2_t vmovl_u32 (uint32x2_t a)

A32 : VMOVL.U32 Qd, Dm

A64 : UXTL Vd.2D, Vn.2S

ZeroExtendWideningLower(Vector64<SByte>)

uint16x8_t vmovl_u8 (uint8x8_t a)

A32 : VMOVL.U8 Qd, Dm

A64 : UXTL Vd.8H, Vn.8B

ZeroExtendWideningLower(Vector64<UInt16>)

uint32x4_t vmovl_u16 (uint16x4_t a)

A32 : VMOVL.U16 Qd, Dm

A64 : UXTL Vd.4S, Vn.4H

ZeroExtendWideningLower(Vector64<UInt32>)

uint64x2_t vmovl_u32 (uint32x2_t a)

A32 : VMOVL.U32 Qd, Dm

A64 : UXTL Vd.2D, Vn.2S

ZeroExtendWideningUpper(Vector128<Byte>)

uint16x8_t vmovl_high_u8 (uint8x16_t a)

A32 : VMOVL.U8 Qd, Dm+1

A64 : UXTL2 Vd.8H, Vn.16B

ZeroExtendWideningUpper(Vector128<Int16>)

uint32x4_t vmovl_high_u16 (uint16x8_t a)

A32 : VMOVL.U16 Qd, Dm+1

A64 : UXTL2 Vd.4S, Vn.8H

ZeroExtendWideningUpper(Vector128<Int32>)

uint64x2_t vmovl_high_u32 (uint32x4_t a)

A32 : VMOVL.U32 Qd, Dm+1

A64 : UXTL2 Vd.2D, Vn.4S

ZeroExtendWideningUpper(Vector128<SByte>)

uint16x8_t vmovl_high_u8 (uint8x16_t a)

A32 : VMOVL.U8 Qd, Dm+1

A64 : UXTL2 Vd.8H, Vn.16B

ZeroExtendWideningUpper(Vector128<UInt16>)

uint32x4_t vmovl_high_u16 (uint16x8_t a)

A32 : VMOVL.U16 Qd, Dm+1

A64 : UXTL2 Vd.4S, Vn.8H

ZeroExtendWideningUpper(Vector128<UInt32>)

uint64x2_t vmovl_high_u32 (uint32x4_t a)

A32 : VMOVL.U32 Qd, Dm+1

A64 : UXTL2 Vd.2D, Vn.4S

S’applique à