| // RUN: %clang_cc1 -triple arm64-apple-darwin -target-feature +neon \ |
| // RUN: -fallow-half-arguments-and-returns -disable-O0-optnone -emit-llvm -o - %s \ |
| // RUN: | opt -S -mem2reg | FileCheck %s |
| |
| #include <arm_neon.h> |
| |
| // CHECK-LABEL: define i8 @test_vget_lane_u8(<8 x i8> %a) #0 { |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <8 x i8> %a, i32 7 |
| // CHECK: ret i8 [[VGET_LANE]] |
| uint8_t test_vget_lane_u8(uint8x8_t a) { |
| return vget_lane_u8(a, 7); |
| } |
| |
| // CHECK-LABEL: define i16 @test_vget_lane_u16(<4 x i16> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <4 x i16> [[TMP1]], i32 3 |
| // CHECK: ret i16 [[VGET_LANE]] |
| uint16_t test_vget_lane_u16(uint16x4_t a) { |
| return vget_lane_u16(a, 3); |
| } |
| |
| // CHECK-LABEL: define i32 @test_vget_lane_u32(<2 x i32> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i32> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x i32> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <2 x i32> [[TMP1]], i32 1 |
| // CHECK: ret i32 [[VGET_LANE]] |
| uint32_t test_vget_lane_u32(uint32x2_t a) { |
| return vget_lane_u32(a, 1); |
| } |
| |
| // CHECK-LABEL: define i8 @test_vget_lane_s8(<8 x i8> %a) #0 { |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <8 x i8> %a, i32 7 |
| // CHECK: ret i8 [[VGET_LANE]] |
| int8_t test_vget_lane_s8(int8x8_t a) { |
| return vget_lane_s8(a, 7); |
| } |
| |
| // CHECK-LABEL: define i16 @test_vget_lane_s16(<4 x i16> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <4 x i16> [[TMP1]], i32 3 |
| // CHECK: ret i16 [[VGET_LANE]] |
| int16_t test_vget_lane_s16(int16x4_t a) { |
| return vget_lane_s16(a, 3); |
| } |
| |
| // CHECK-LABEL: define i32 @test_vget_lane_s32(<2 x i32> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i32> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x i32> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <2 x i32> [[TMP1]], i32 1 |
| // CHECK: ret i32 [[VGET_LANE]] |
| int32_t test_vget_lane_s32(int32x2_t a) { |
| return vget_lane_s32(a, 1); |
| } |
| |
| // CHECK-LABEL: define i8 @test_vget_lane_p8(<8 x i8> %a) #0 { |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <8 x i8> %a, i32 7 |
| // CHECK: ret i8 [[VGET_LANE]] |
| poly8_t test_vget_lane_p8(poly8x8_t a) { |
| return vget_lane_p8(a, 7); |
| } |
| |
| // CHECK-LABEL: define i16 @test_vget_lane_p16(<4 x i16> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <4 x i16> [[TMP1]], i32 3 |
| // CHECK: ret i16 [[VGET_LANE]] |
| poly16_t test_vget_lane_p16(poly16x4_t a) { |
| return vget_lane_p16(a, 3); |
| } |
| |
| // CHECK-LABEL: define float @test_vget_lane_f32(<2 x float> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x float> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x float> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <2 x float> [[TMP1]], i32 1 |
| // CHECK: ret float [[VGET_LANE]] |
| float32_t test_vget_lane_f32(float32x2_t a) { |
| return vget_lane_f32(a, 1); |
| } |
| |
| // CHECK-LABEL: define float @test_vget_lane_f16(<4 x half> %a) #0 { |
| // CHECK: [[__REINT_242:%.*]] = alloca <4 x half>, align 8 |
| // CHECK: [[__REINT1_242:%.*]] = alloca i16, align 2 |
| // CHECK: store <4 x half> %a, <4 x half>* [[__REINT_242]], align 8 |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x half>* [[__REINT_242]] to <4 x i16>* |
| // CHECK: [[TMP1:%.*]] = load <4 x i16>, <4 x i16>* [[TMP0]], align 8 |
| // CHECK: [[TMP2:%.*]] = bitcast <4 x i16> [[TMP1]] to <8 x i8> |
| // CHECK: [[TMP3:%.*]] = bitcast <8 x i8> [[TMP2]] to <4 x i16> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <4 x i16> [[TMP3]], i32 1 |
| // CHECK: store i16 [[VGET_LANE]], i16* [[__REINT1_242]], align 2 |
| // CHECK: [[TMP4:%.*]] = bitcast i16* [[__REINT1_242]] to half* |
| // CHECK: [[TMP5:%.*]] = load half, half* [[TMP4]], align 2 |
| // CHECK: [[CONV:%.*]] = fpext half [[TMP5]] to float |
| // CHECK: ret float [[CONV]] |
| float32_t test_vget_lane_f16(float16x4_t a) { |
| return vget_lane_f16(a, 1); |
| } |
| |
| // CHECK-LABEL: define i8 @test_vgetq_lane_u8(<16 x i8> %a) #0 { |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <16 x i8> %a, i32 15 |
| // CHECK: ret i8 [[VGETQ_LANE]] |
| uint8_t test_vgetq_lane_u8(uint8x16_t a) { |
| return vgetq_lane_u8(a, 15); |
| } |
| |
| // CHECK-LABEL: define i16 @test_vgetq_lane_u16(<8 x i16> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <8 x i16> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <8 x i16> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <8 x i16> [[TMP1]], i32 7 |
| // CHECK: ret i16 [[VGETQ_LANE]] |
| uint16_t test_vgetq_lane_u16(uint16x8_t a) { |
| return vgetq_lane_u16(a, 7); |
| } |
| |
| // CHECK-LABEL: define i32 @test_vgetq_lane_u32(<4 x i32> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i32> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <4 x i32> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <4 x i32> [[TMP1]], i32 3 |
| // CHECK: ret i32 [[VGETQ_LANE]] |
| uint32_t test_vgetq_lane_u32(uint32x4_t a) { |
| return vgetq_lane_u32(a, 3); |
| } |
| |
| // CHECK-LABEL: define i8 @test_vgetq_lane_s8(<16 x i8> %a) #0 { |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <16 x i8> %a, i32 15 |
| // CHECK: ret i8 [[VGETQ_LANE]] |
| int8_t test_vgetq_lane_s8(int8x16_t a) { |
| return vgetq_lane_s8(a, 15); |
| } |
| |
| // CHECK-LABEL: define i16 @test_vgetq_lane_s16(<8 x i16> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <8 x i16> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <8 x i16> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <8 x i16> [[TMP1]], i32 7 |
| // CHECK: ret i16 [[VGETQ_LANE]] |
| int16_t test_vgetq_lane_s16(int16x8_t a) { |
| return vgetq_lane_s16(a, 7); |
| } |
| |
| // CHECK-LABEL: define i32 @test_vgetq_lane_s32(<4 x i32> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i32> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <4 x i32> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <4 x i32> [[TMP1]], i32 3 |
| // CHECK: ret i32 [[VGETQ_LANE]] |
| int32_t test_vgetq_lane_s32(int32x4_t a) { |
| return vgetq_lane_s32(a, 3); |
| } |
| |
| // CHECK-LABEL: define i8 @test_vgetq_lane_p8(<16 x i8> %a) #0 { |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <16 x i8> %a, i32 15 |
| // CHECK: ret i8 [[VGETQ_LANE]] |
| poly8_t test_vgetq_lane_p8(poly8x16_t a) { |
| return vgetq_lane_p8(a, 15); |
| } |
| |
| // CHECK-LABEL: define i16 @test_vgetq_lane_p16(<8 x i16> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <8 x i16> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <8 x i16> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <8 x i16> [[TMP1]], i32 7 |
| // CHECK: ret i16 [[VGETQ_LANE]] |
| poly16_t test_vgetq_lane_p16(poly16x8_t a) { |
| return vgetq_lane_p16(a, 7); |
| } |
| |
| // CHECK-LABEL: define float @test_vgetq_lane_f32(<4 x float> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x float> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <4 x float> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <4 x float> [[TMP1]], i32 3 |
| // CHECK: ret float [[VGETQ_LANE]] |
| float32_t test_vgetq_lane_f32(float32x4_t a) { |
| return vgetq_lane_f32(a, 3); |
| } |
| |
| // CHECK-LABEL: define float @test_vgetq_lane_f16(<8 x half> %a) #0 { |
| // CHECK: [[__REINT_244:%.*]] = alloca <8 x half>, align 16 |
| // CHECK: [[__REINT1_244:%.*]] = alloca i16, align 2 |
| // CHECK: store <8 x half> %a, <8 x half>* [[__REINT_244]], align 16 |
| // CHECK: [[TMP0:%.*]] = bitcast <8 x half>* [[__REINT_244]] to <8 x i16>* |
| // CHECK: [[TMP1:%.*]] = load <8 x i16>, <8 x i16>* [[TMP0]], align 16 |
| // CHECK: [[TMP2:%.*]] = bitcast <8 x i16> [[TMP1]] to <16 x i8> |
| // CHECK: [[TMP3:%.*]] = bitcast <16 x i8> [[TMP2]] to <8 x i16> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <8 x i16> [[TMP3]], i32 3 |
| // CHECK: store i16 [[VGETQ_LANE]], i16* [[__REINT1_244]], align 2 |
| // CHECK: [[TMP4:%.*]] = bitcast i16* [[__REINT1_244]] to half* |
| // CHECK: [[TMP5:%.*]] = load half, half* [[TMP4]], align 2 |
| // CHECK: [[CONV:%.*]] = fpext half [[TMP5]] to float |
| // CHECK: ret float [[CONV]] |
| float32_t test_vgetq_lane_f16(float16x8_t a) { |
| return vgetq_lane_f16(a, 3); |
| } |
| |
| // CHECK-LABEL: define i64 @test_vget_lane_s64(<1 x i64> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <1 x i64> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <1 x i64> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <1 x i64> [[TMP1]], i32 0 |
| // CHECK: ret i64 [[VGET_LANE]] |
| int64_t test_vget_lane_s64(int64x1_t a) { |
| return vget_lane_s64(a, 0); |
| } |
| |
| // CHECK-LABEL: define i64 @test_vget_lane_u64(<1 x i64> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <1 x i64> %a to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <1 x i64> |
| // CHECK: [[VGET_LANE:%.*]] = extractelement <1 x i64> [[TMP1]], i32 0 |
| // CHECK: ret i64 [[VGET_LANE]] |
| uint64_t test_vget_lane_u64(uint64x1_t a) { |
| return vget_lane_u64(a, 0); |
| } |
| |
| // CHECK-LABEL: define i64 @test_vgetq_lane_s64(<2 x i64> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i64> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <2 x i64> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <2 x i64> [[TMP1]], i32 1 |
| // CHECK: ret i64 [[VGETQ_LANE]] |
| int64_t test_vgetq_lane_s64(int64x2_t a) { |
| return vgetq_lane_s64(a, 1); |
| } |
| |
| // CHECK-LABEL: define i64 @test_vgetq_lane_u64(<2 x i64> %a) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i64> %a to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <2 x i64> |
| // CHECK: [[VGETQ_LANE:%.*]] = extractelement <2 x i64> [[TMP1]], i32 1 |
| // CHECK: ret i64 [[VGETQ_LANE]] |
| uint64_t test_vgetq_lane_u64(uint64x2_t a) { |
| return vgetq_lane_u64(a, 1); |
| } |
| |
| |
| // CHECK-LABEL: define <8 x i8> @test_vset_lane_u8(i8 %a, <8 x i8> %b) #0 { |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <8 x i8> %b, i8 %a, i32 7 |
| // CHECK: ret <8 x i8> [[VSET_LANE]] |
| uint8x8_t test_vset_lane_u8(uint8_t a, uint8x8_t b) { |
| return vset_lane_u8(a, b, 7); |
| } |
| |
| // CHECK-LABEL: define <4 x i16> @test_vset_lane_u16(i16 %a, <4 x i16> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <4 x i16> [[TMP1]], i16 %a, i32 3 |
| // CHECK: ret <4 x i16> [[VSET_LANE]] |
| uint16x4_t test_vset_lane_u16(uint16_t a, uint16x4_t b) { |
| return vset_lane_u16(a, b, 3); |
| } |
| |
| // CHECK-LABEL: define <2 x i32> @test_vset_lane_u32(i32 %a, <2 x i32> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i32> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x i32> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <2 x i32> [[TMP1]], i32 %a, i32 1 |
| // CHECK: ret <2 x i32> [[VSET_LANE]] |
| uint32x2_t test_vset_lane_u32(uint32_t a, uint32x2_t b) { |
| return vset_lane_u32(a, b, 1); |
| } |
| |
| // CHECK-LABEL: define <8 x i8> @test_vset_lane_s8(i8 %a, <8 x i8> %b) #0 { |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <8 x i8> %b, i8 %a, i32 7 |
| // CHECK: ret <8 x i8> [[VSET_LANE]] |
| int8x8_t test_vset_lane_s8(int8_t a, int8x8_t b) { |
| return vset_lane_s8(a, b, 7); |
| } |
| |
| // CHECK-LABEL: define <4 x i16> @test_vset_lane_s16(i16 %a, <4 x i16> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <4 x i16> [[TMP1]], i16 %a, i32 3 |
| // CHECK: ret <4 x i16> [[VSET_LANE]] |
| int16x4_t test_vset_lane_s16(int16_t a, int16x4_t b) { |
| return vset_lane_s16(a, b, 3); |
| } |
| |
| // CHECK-LABEL: define <2 x i32> @test_vset_lane_s32(i32 %a, <2 x i32> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i32> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x i32> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <2 x i32> [[TMP1]], i32 %a, i32 1 |
| // CHECK: ret <2 x i32> [[VSET_LANE]] |
| int32x2_t test_vset_lane_s32(int32_t a, int32x2_t b) { |
| return vset_lane_s32(a, b, 1); |
| } |
| |
| // CHECK-LABEL: define <8 x i8> @test_vset_lane_p8(i8 %a, <8 x i8> %b) #0 { |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <8 x i8> %b, i8 %a, i32 7 |
| // CHECK: ret <8 x i8> [[VSET_LANE]] |
| poly8x8_t test_vset_lane_p8(poly8_t a, poly8x8_t b) { |
| return vset_lane_p8(a, b, 7); |
| } |
| |
| // CHECK-LABEL: define <4 x i16> @test_vset_lane_p16(i16 %a, <4 x i16> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i16> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <4 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <4 x i16> [[TMP1]], i16 %a, i32 3 |
| // CHECK: ret <4 x i16> [[VSET_LANE]] |
| poly16x4_t test_vset_lane_p16(poly16_t a, poly16x4_t b) { |
| return vset_lane_p16(a, b, 3); |
| } |
| |
| // CHECK-LABEL: define <2 x float> @test_vset_lane_f32(float %a, <2 x float> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x float> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <2 x float> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <2 x float> [[TMP1]], float %a, i32 1 |
| // CHECK: ret <2 x float> [[VSET_LANE]] |
| float32x2_t test_vset_lane_f32(float32_t a, float32x2_t b) { |
| return vset_lane_f32(a, b, 1); |
| } |
| |
| // CHECK-LABEL: define <4 x half> @test_vset_lane_f16(half* %a, <4 x half> %b) #0 { |
| // CHECK: [[__REINT_246:%.*]] = alloca half, align 2 |
| // CHECK: [[__REINT1_246:%.*]] = alloca <4 x half>, align 8 |
| // CHECK: [[__REINT2_246:%.*]] = alloca <4 x i16>, align 8 |
| // CHECK: [[TMP0:%.*]] = load half, half* %a, align 2 |
| // CHECK: store half [[TMP0]], half* [[__REINT_246]], align 2 |
| // CHECK: store <4 x half> %b, <4 x half>* [[__REINT1_246]], align 8 |
| // CHECK: [[TMP1:%.*]] = bitcast half* [[__REINT_246]] to i16* |
| // CHECK: [[TMP2:%.*]] = load i16, i16* [[TMP1]], align 2 |
| // CHECK: [[TMP3:%.*]] = bitcast <4 x half>* [[__REINT1_246]] to <4 x i16>* |
| // CHECK: [[TMP4:%.*]] = load <4 x i16>, <4 x i16>* [[TMP3]], align 8 |
| // CHECK: [[TMP5:%.*]] = bitcast <4 x i16> [[TMP4]] to <8 x i8> |
| // CHECK: [[TMP6:%.*]] = bitcast <8 x i8> [[TMP5]] to <4 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <4 x i16> [[TMP6]], i16 [[TMP2]], i32 3 |
| // CHECK: store <4 x i16> [[VSET_LANE]], <4 x i16>* [[__REINT2_246]], align 8 |
| // CHECK: [[TMP7:%.*]] = bitcast <4 x i16>* [[__REINT2_246]] to <4 x half>* |
| // CHECK: [[TMP8:%.*]] = load <4 x half>, <4 x half>* [[TMP7]], align 8 |
| // CHECK: ret <4 x half> [[TMP8]] |
| float16x4_t test_vset_lane_f16(float16_t *a, float16x4_t b) { |
| return vset_lane_f16(*a, b, 3); |
| } |
| |
| // CHECK-LABEL: define <16 x i8> @test_vsetq_lane_u8(i8 %a, <16 x i8> %b) #0 { |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <16 x i8> %b, i8 %a, i32 15 |
| // CHECK: ret <16 x i8> [[VSET_LANE]] |
| uint8x16_t test_vsetq_lane_u8(uint8_t a, uint8x16_t b) { |
| return vsetq_lane_u8(a, b, 15); |
| } |
| |
| // CHECK-LABEL: define <8 x i16> @test_vsetq_lane_u16(i16 %a, <8 x i16> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <8 x i16> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <8 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <8 x i16> [[TMP1]], i16 %a, i32 7 |
| // CHECK: ret <8 x i16> [[VSET_LANE]] |
| uint16x8_t test_vsetq_lane_u16(uint16_t a, uint16x8_t b) { |
| return vsetq_lane_u16(a, b, 7); |
| } |
| |
| // CHECK-LABEL: define <4 x i32> @test_vsetq_lane_u32(i32 %a, <4 x i32> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i32> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <4 x i32> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <4 x i32> [[TMP1]], i32 %a, i32 3 |
| // CHECK: ret <4 x i32> [[VSET_LANE]] |
| uint32x4_t test_vsetq_lane_u32(uint32_t a, uint32x4_t b) { |
| return vsetq_lane_u32(a, b, 3); |
| } |
| |
| // CHECK-LABEL: define <16 x i8> @test_vsetq_lane_s8(i8 %a, <16 x i8> %b) #0 { |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <16 x i8> %b, i8 %a, i32 15 |
| // CHECK: ret <16 x i8> [[VSET_LANE]] |
| int8x16_t test_vsetq_lane_s8(int8_t a, int8x16_t b) { |
| return vsetq_lane_s8(a, b, 15); |
| } |
| |
| // CHECK-LABEL: define <8 x i16> @test_vsetq_lane_s16(i16 %a, <8 x i16> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <8 x i16> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <8 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <8 x i16> [[TMP1]], i16 %a, i32 7 |
| // CHECK: ret <8 x i16> [[VSET_LANE]] |
| int16x8_t test_vsetq_lane_s16(int16_t a, int16x8_t b) { |
| return vsetq_lane_s16(a, b, 7); |
| } |
| |
| // CHECK-LABEL: define <4 x i32> @test_vsetq_lane_s32(i32 %a, <4 x i32> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x i32> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <4 x i32> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <4 x i32> [[TMP1]], i32 %a, i32 3 |
| // CHECK: ret <4 x i32> [[VSET_LANE]] |
| int32x4_t test_vsetq_lane_s32(int32_t a, int32x4_t b) { |
| return vsetq_lane_s32(a, b, 3); |
| } |
| |
| // CHECK-LABEL: define <16 x i8> @test_vsetq_lane_p8(i8 %a, <16 x i8> %b) #0 { |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <16 x i8> %b, i8 %a, i32 15 |
| // CHECK: ret <16 x i8> [[VSET_LANE]] |
| poly8x16_t test_vsetq_lane_p8(poly8_t a, poly8x16_t b) { |
| return vsetq_lane_p8(a, b, 15); |
| } |
| |
| // CHECK-LABEL: define <8 x i16> @test_vsetq_lane_p16(i16 %a, <8 x i16> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <8 x i16> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <8 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <8 x i16> [[TMP1]], i16 %a, i32 7 |
| // CHECK: ret <8 x i16> [[VSET_LANE]] |
| poly16x8_t test_vsetq_lane_p16(poly16_t a, poly16x8_t b) { |
| return vsetq_lane_p16(a, b, 7); |
| } |
| |
| // CHECK-LABEL: define <4 x float> @test_vsetq_lane_f32(float %a, <4 x float> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <4 x float> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <4 x float> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <4 x float> [[TMP1]], float %a, i32 3 |
| // CHECK: ret <4 x float> [[VSET_LANE]] |
| float32x4_t test_vsetq_lane_f32(float32_t a, float32x4_t b) { |
| return vsetq_lane_f32(a, b, 3); |
| } |
| |
| // CHECK-LABEL: define <8 x half> @test_vsetq_lane_f16(half* %a, <8 x half> %b) #0 { |
| // CHECK: [[__REINT_248:%.*]] = alloca half, align 2 |
| // CHECK: [[__REINT1_248:%.*]] = alloca <8 x half>, align 16 |
| // CHECK: [[__REINT2_248:%.*]] = alloca <8 x i16>, align 16 |
| // CHECK: [[TMP0:%.*]] = load half, half* %a, align 2 |
| // CHECK: store half [[TMP0]], half* [[__REINT_248]], align 2 |
| // CHECK: store <8 x half> %b, <8 x half>* [[__REINT1_248]], align 16 |
| // CHECK: [[TMP1:%.*]] = bitcast half* [[__REINT_248]] to i16* |
| // CHECK: [[TMP2:%.*]] = load i16, i16* [[TMP1]], align 2 |
| // CHECK: [[TMP3:%.*]] = bitcast <8 x half>* [[__REINT1_248]] to <8 x i16>* |
| // CHECK: [[TMP4:%.*]] = load <8 x i16>, <8 x i16>* [[TMP3]], align 16 |
| // CHECK: [[TMP5:%.*]] = bitcast <8 x i16> [[TMP4]] to <16 x i8> |
| // CHECK: [[TMP6:%.*]] = bitcast <16 x i8> [[TMP5]] to <8 x i16> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <8 x i16> [[TMP6]], i16 [[TMP2]], i32 7 |
| // CHECK: store <8 x i16> [[VSET_LANE]], <8 x i16>* [[__REINT2_248]], align 16 |
| // CHECK: [[TMP7:%.*]] = bitcast <8 x i16>* [[__REINT2_248]] to <8 x half>* |
| // CHECK: [[TMP8:%.*]] = load <8 x half>, <8 x half>* [[TMP7]], align 16 |
| // CHECK: ret <8 x half> [[TMP8]] |
| float16x8_t test_vsetq_lane_f16(float16_t *a, float16x8_t b) { |
| return vsetq_lane_f16(*a, b, 7); |
| } |
| |
| // CHECK-LABEL: define <1 x i64> @test_vset_lane_s64(i64 %a, <1 x i64> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <1 x i64> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <1 x i64> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <1 x i64> [[TMP1]], i64 %a, i32 0 |
| // CHECK: ret <1 x i64> [[VSET_LANE]] |
| int64x1_t test_vset_lane_s64(int64_t a, int64x1_t b) { |
| return vset_lane_s64(a, b, 0); |
| } |
| |
| // CHECK-LABEL: define <1 x i64> @test_vset_lane_u64(i64 %a, <1 x i64> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <1 x i64> %b to <8 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <8 x i8> [[TMP0]] to <1 x i64> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <1 x i64> [[TMP1]], i64 %a, i32 0 |
| // CHECK: ret <1 x i64> [[VSET_LANE]] |
| uint64x1_t test_vset_lane_u64(uint64_t a, uint64x1_t b) { |
| return vset_lane_u64(a, b, 0); |
| } |
| |
| // CHECK-LABEL: define <2 x i64> @test_vsetq_lane_s64(i64 %a, <2 x i64> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i64> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <2 x i64> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <2 x i64> [[TMP1]], i64 %a, i32 1 |
| // CHECK: ret <2 x i64> [[VSET_LANE]] |
| int64x2_t test_vsetq_lane_s64(int64_t a, int64x2_t b) { |
| return vsetq_lane_s64(a, b, 1); |
| } |
| |
| // CHECK-LABEL: define <2 x i64> @test_vsetq_lane_u64(i64 %a, <2 x i64> %b) #0 { |
| // CHECK: [[TMP0:%.*]] = bitcast <2 x i64> %b to <16 x i8> |
| // CHECK: [[TMP1:%.*]] = bitcast <16 x i8> [[TMP0]] to <2 x i64> |
| // CHECK: [[VSET_LANE:%.*]] = insertelement <2 x i64> [[TMP1]], i64 %a, i32 1 |
| // CHECK: ret <2 x i64> [[VSET_LANE]] |
| uint64x2_t test_vsetq_lane_u64(uint64_t a, uint64x2_t b) { |
| return vsetq_lane_u64(a, b, 1); |
| } |