|  | @ RUN: llvm-mc -triple armv8 -mattr=+fp-armv8 -show-encoding < %s | FileCheck %s | 
|  |  | 
|  | @ VCVT{B,T} | 
|  |  | 
|  | vcvtt.f64.f16 d3, s1 | 
|  | @ CHECK: vcvtt.f64.f16 d3, s1      @ encoding: [0xe0,0x3b,0xb2,0xee] | 
|  | vcvtt.f16.f64 s5, d12 | 
|  | @ CHECK: vcvtt.f16.f64 s5, d12     @ encoding: [0xcc,0x2b,0xf3,0xee] | 
|  |  | 
|  | vcvtb.f64.f16 d3, s1 | 
|  | @ CHECK: vcvtb.f64.f16 d3, s1     @ encoding: [0x60,0x3b,0xb2,0xee] | 
|  | vcvtb.f16.f64 s4, d1 | 
|  | @ CHECK: vcvtb.f16.f64 s4, d1     @ encoding: [0x41,0x2b,0xb3,0xee] | 
|  |  | 
|  | vcvttge.f64.f16 d3, s1 | 
|  | @ CHECK: vcvttge.f64.f16 d3, s1      @ encoding: [0xe0,0x3b,0xb2,0xae] | 
|  | vcvttgt.f16.f64 s5, d12 | 
|  | @ CHECK: vcvttgt.f16.f64 s5, d12     @ encoding: [0xcc,0x2b,0xf3,0xce] | 
|  |  | 
|  | vcvtbeq.f64.f16 d3, s1 | 
|  | @ CHECK: vcvtbeq.f64.f16 d3, s1     @ encoding: [0x60,0x3b,0xb2,0x0e] | 
|  | vcvtblt.f16.f64 s4, d1 | 
|  | @ CHECK: vcvtblt.f16.f64 s4, d1     @ encoding: [0x41,0x2b,0xb3,0xbe] | 
|  |  | 
|  |  | 
|  | @ VCVT{A,N,P,M} | 
|  |  | 
|  | vcvta.s32.f32 s2, s3 | 
|  | @ CHECK: vcvta.s32.f32 s2, s3     @ encoding: [0xe1,0x1a,0xbc,0xfe] | 
|  | vcvta.s32.f64 s2, d3 | 
|  | @ CHECK: vcvta.s32.f64 s2, d3     @ encoding: [0xc3,0x1b,0xbc,0xfe] | 
|  | vcvtn.s32.f32 s6, s23 | 
|  | @ CHECK: vcvtn.s32.f32 s6, s23     @ encoding: [0xeb,0x3a,0xbd,0xfe] | 
|  | vcvtn.s32.f64 s6, d23 | 
|  | @ CHECK: vcvtn.s32.f64 s6, d23     @ encoding: [0xe7,0x3b,0xbd,0xfe] | 
|  | vcvtp.s32.f32 s0, s4 | 
|  | @ CHECK: vcvtp.s32.f32 s0, s4     @ encoding: [0xc2,0x0a,0xbe,0xfe] | 
|  | vcvtp.s32.f64 s0, d4 | 
|  | @ CHECK: vcvtp.s32.f64 s0, d4     @ encoding: [0xc4,0x0b,0xbe,0xfe] | 
|  | vcvtm.s32.f32 s17, s8 | 
|  | @ CHECK: vcvtm.s32.f32 s17, s8     @ encoding: [0xc4,0x8a,0xff,0xfe] | 
|  | vcvtm.s32.f64 s17, d8 | 
|  | @ CHECK: vcvtm.s32.f64 s17, d8     @ encoding: [0xc8,0x8b,0xff,0xfe] | 
|  |  | 
|  | vcvta.u32.f32 s2, s3 | 
|  | @ CHECK: vcvta.u32.f32 s2, s3     @ encoding: [0x61,0x1a,0xbc,0xfe] | 
|  | vcvta.u32.f64 s2, d3 | 
|  | @ CHECK: vcvta.u32.f64 s2, d3     @ encoding: [0x43,0x1b,0xbc,0xfe] | 
|  | vcvtn.u32.f32 s6, s23 | 
|  | @ CHECK: vcvtn.u32.f32 s6, s23     @ encoding: [0x6b,0x3a,0xbd,0xfe] | 
|  | vcvtn.u32.f64 s6, d23 | 
|  | @ CHECK: vcvtn.u32.f64 s6, d23     @ encoding: [0x67,0x3b,0xbd,0xfe] | 
|  | vcvtp.u32.f32 s0, s4 | 
|  | @ CHECK: vcvtp.u32.f32 s0, s4     @ encoding: [0x42,0x0a,0xbe,0xfe] | 
|  | vcvtp.u32.f64 s0, d4 | 
|  | @ CHECK: vcvtp.u32.f64 s0, d4     @ encoding: [0x44,0x0b,0xbe,0xfe] | 
|  | vcvtm.u32.f32 s17, s8 | 
|  | @ CHECK: vcvtm.u32.f32 s17, s8     @ encoding: [0x44,0x8a,0xff,0xfe] | 
|  | vcvtm.u32.f64 s17, d8 | 
|  | @ CHECK: vcvtm.u32.f64 s17, d8     @ encoding: [0x48,0x8b,0xff,0xfe] | 
|  |  | 
|  |  | 
|  | @ VSEL | 
|  | vselge.f32 s4, s1, s23 | 
|  | @ CHECK: vselge.f32 s4, s1, s23    @ encoding: [0xab,0x2a,0x20,0xfe] | 
|  | vselge.f64 d30, d31, d23 | 
|  | @ CHECK: vselge.f64 d30, d31, d23  @ encoding: [0xa7,0xeb,0x6f,0xfe] | 
|  | vselgt.f32 s0, s1, s0 | 
|  | @ CHECK: vselgt.f32 s0, s1, s0    @ encoding: [0x80,0x0a,0x30,0xfe] | 
|  | vselgt.f64 d5, d10, d20 | 
|  | @ CHECK: vselgt.f64 d5, d10, d20  @ encoding: [0x24,0x5b,0x3a,0xfe] | 
|  | vseleq.f32 s30, s28, s23 | 
|  | @ CHECK: vseleq.f32 s30, s28, s23 @ encoding: [0x2b,0xfa,0x0e,0xfe] | 
|  | vseleq.f64 d2, d4, d8 | 
|  | @ CHECK: vseleq.f64 d2, d4, d8    @ encoding: [0x08,0x2b,0x04,0xfe] | 
|  | vselvs.f32 s21, s16, s14 | 
|  | @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe] | 
|  | vselvs.f64 d0, d1, d31 | 
|  | @ CHECK: vselvs.f64 d0, d1, d31   @ encoding: [0x2f,0x0b,0x11,0xfe] | 
|  |  | 
|  |  | 
|  | @ VMAXNM / VMINNM | 
|  | vmaxnm.f32 s5, s12, s0 | 
|  | @ CHECK: vmaxnm.f32 s5, s12, s0    @ encoding: [0x00,0x2a,0xc6,0xfe] | 
|  | vmaxnm.f64 d5, d22, d30 | 
|  | @ CHECK: vmaxnm.f64 d5, d22, d30   @ encoding: [0xae,0x5b,0x86,0xfe] | 
|  | vminnm.f32 s0, s0, s12 | 
|  | @ CHECK: vminnm.f32 s0, s0, s12    @ encoding: [0x46,0x0a,0x80,0xfe] | 
|  | vminnm.f64 d4, d6, d9 | 
|  | @ CHECK: vminnm.f64 d4, d6, d9     @ encoding: [0x49,0x4b,0x86,0xfe] | 
|  |  | 
|  | @ VRINT{Z,R,X} | 
|  |  | 
|  | vrintzge.f64 d3, d12 | 
|  | @ CHECK: vrintzge.f64 d3, d12   @ encoding: [0xcc,0x3b,0xb6,0xae] | 
|  | vrintz.f32 s3, s24 | 
|  | @ CHECK: vrintz.f32 s3, s24     @ encoding: [0xcc,0x1a,0xf6,0xee] | 
|  | vrintrlt.f64 d5, d0 | 
|  | @ CHECK: vrintrlt.f64 d5, d0    @ encoding: [0x40,0x5b,0xb6,0xbe] | 
|  | vrintr.f32 s0, s9 | 
|  | @ CHECK: vrintr.f32 s0, s9      @ encoding: [0x64,0x0a,0xb6,0xee] | 
|  | vrintxeq.f64 d28, d30 | 
|  | @ CHECK: vrintxeq.f64 d28, d30  @ encoding: [0x6e,0xcb,0xf7,0x0e] | 
|  | vrintxvs.f32 s10, s14 | 
|  | @ CHECK: vrintxvs.f32 s10, s14  @ encoding: [0x47,0x5a,0xb7,0x6e] | 
|  |  | 
|  | @ VRINT{A,N,P,M} | 
|  |  | 
|  | vrinta.f64 d3, d4 | 
|  | @ CHECK: vrinta.f64 d3, d4     @ encoding: [0x44,0x3b,0xb8,0xfe] | 
|  | vrinta.f32 s12, s1 | 
|  | @ CHECK: vrinta.f32 s12, s1    @ encoding: [0x60,0x6a,0xb8,0xfe] | 
|  | vrintn.f64 d3, d4 | 
|  | @ CHECK: vrintn.f64 d3, d4     @ encoding: [0x44,0x3b,0xb9,0xfe] | 
|  | vrintn.f32 s12, s1 | 
|  | @ CHECK: vrintn.f32 s12, s1    @ encoding: [0x60,0x6a,0xb9,0xfe] | 
|  | vrintp.f64 d3, d4 | 
|  | @ CHECK: vrintp.f64 d3, d4     @ encoding: [0x44,0x3b,0xba,0xfe] | 
|  | vrintp.f32 s12, s1 | 
|  | @ CHECK: vrintp.f32 s12, s1    @ encoding: [0x60,0x6a,0xba,0xfe] | 
|  | vrintm.f64 d3, d4 | 
|  | @ CHECK: vrintm.f64 d3, d4     @ encoding: [0x44,0x3b,0xbb,0xfe] | 
|  | vrintm.f32 s12, s1 | 
|  | @ CHECK: vrintm.f32 s12, s1    @ encoding: [0x60,0x6a,0xbb,0xfe] | 
|  |  | 
|  | @ MVFR2 | 
|  |  | 
|  | vmrs sp, mvfr2 | 
|  | @ CHECK: vmrs sp, mvfr2        @ encoding: [0x10,0xda,0xf5,0xee] |