mirror of
https://github.com/Gericom/teak-llvm.git
synced 2025-06-19 11:35:51 -04:00

Summary: The vector pattern `(a + b + 1) / 2` was previously selected to an avgr_u instruction regardless of nuw flags, but this is incorrect in the case where either addition may have an unsigned wrap. This CL changes the existing pattern to require both adds to have nuw flags and adds builtin functions and intrinsics for the avgr_u instructions because the corrected pattern is not representable in C. Reviewers: aheejin Subscribers: dschuff, sbc100, jgravelle-google, hiraditya, sunfish, cfe-commits, llvm-commits Tags: #clang, #llvm Differential Revision: https://reviews.llvm.org/D71648
615 lines
21 KiB
C
615 lines
21 KiB
C
// RUN: %clang_cc1 -triple wasm32-unknown-unknown -target-feature +unimplemented-simd128 -target-feature +nontrapping-fptoint -target-feature +exception-handling -target-feature +bulk-memory -flax-vector-conversions=none -O3 -emit-llvm -o - %s | FileCheck %s -check-prefixes WEBASSEMBLY,WEBASSEMBLY32
|
|
// RUN: %clang_cc1 -triple wasm64-unknown-unknown -target-feature +unimplemented-simd128 -target-feature +nontrapping-fptoint -target-feature +exception-handling -target-feature +bulk-memory -flax-vector-conversions=none -O3 -emit-llvm -o - %s | FileCheck %s -check-prefixes WEBASSEMBLY,WEBASSEMBLY64
|
|
// RUN: not %clang_cc1 -triple wasm64-unknown-unknown -target-feature +nontrapping-fptoint -target-feature +exception-handling -target-feature +bulk-memory -flax-vector-conversions=none -O3 -emit-llvm -o - %s 2>&1 | FileCheck %s -check-prefixes MISSING-SIMD
|
|
|
|
// SIMD convenience types
|
|
typedef char i8x16 __attribute((vector_size(16)));
|
|
typedef short i16x8 __attribute((vector_size(16)));
|
|
typedef int i32x4 __attribute((vector_size(16)));
|
|
typedef long long i64x2 __attribute((vector_size(16)));
|
|
typedef unsigned char u8x16 __attribute((vector_size(16)));
|
|
typedef unsigned short u16x8 __attribute((vector_size(16)));
|
|
typedef unsigned int u32x4 __attribute((vector_size(16)));
|
|
typedef unsigned long long u64x2 __attribute((vector_size(16)));
|
|
typedef float f32x4 __attribute((vector_size(16)));
|
|
typedef double f64x2 __attribute((vector_size(16)));
|
|
|
|
__SIZE_TYPE__ memory_size(void) {
|
|
return __builtin_wasm_memory_size(0);
|
|
// WEBASSEMBLY32: call {{i.*}} @llvm.wasm.memory.size.i32(i32 0)
|
|
// WEBASSEMBLY64: call {{i.*}} @llvm.wasm.memory.size.i64(i32 0)
|
|
}
|
|
|
|
__SIZE_TYPE__ memory_grow(__SIZE_TYPE__ delta) {
|
|
return __builtin_wasm_memory_grow(0, delta);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.memory.grow.i32(i32 0, i32 %{{.*}})
|
|
// WEBASSEMBLY64: call i64 @llvm.wasm.memory.grow.i64(i32 0, i64 %{{.*}})
|
|
}
|
|
|
|
void memory_init(void *dest, int offset, int size) {
|
|
__builtin_wasm_memory_init(3, 0, dest, offset, size);
|
|
// WEBASSEMBLY32: call void @llvm.wasm.memory.init(i32 3, i32 0, i8* %{{.*}}, i32 %{{.*}}, i32 %{{.*}})
|
|
// WEBASSEMBLY64: call void @llvm.wasm.memory.init(i32 3, i32 0, i8* %{{.*}}, i32 %{{.*}}, i32 %{{.*}})
|
|
}
|
|
|
|
void data_drop() {
|
|
__builtin_wasm_data_drop(3);
|
|
// WEBASSEMBLY32: call void @llvm.wasm.data.drop(i32 3)
|
|
// WEBASSEMBLY64: call void @llvm.wasm.data.drop(i32 3)
|
|
}
|
|
|
|
__SIZE_TYPE__ tls_size() {
|
|
return __builtin_wasm_tls_size();
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.tls.size.i32()
|
|
// WEBASSEMBLY64: call i64 @llvm.wasm.tls.size.i64()
|
|
}
|
|
|
|
__SIZE_TYPE__ tls_align() {
|
|
return __builtin_wasm_tls_align();
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.tls.align.i32()
|
|
// WEBASSEMBLY64: call i64 @llvm.wasm.tls.align.i64()
|
|
}
|
|
|
|
void *tls_base() {
|
|
return __builtin_wasm_tls_base();
|
|
// WEBASSEMBLY: call i8* @llvm.wasm.tls.base()
|
|
}
|
|
|
|
void throw(void *obj) {
|
|
return __builtin_wasm_throw(0, obj);
|
|
// WEBASSEMBLY32: call void @llvm.wasm.throw(i32 0, i8* %{{.*}})
|
|
// WEBASSEMBLY64: call void @llvm.wasm.throw(i32 0, i8* %{{.*}})
|
|
}
|
|
|
|
void rethrow_in_catch(void) {
|
|
return __builtin_wasm_rethrow_in_catch();
|
|
// WEBASSEMBLY32: call void @llvm.wasm.rethrow.in.catch()
|
|
// WEBASSEMBLY64: call void @llvm.wasm.rethrow.in.catch()
|
|
}
|
|
|
|
int atomic_wait_i32(int *addr, int expected, long long timeout) {
|
|
return __builtin_wasm_atomic_wait_i32(addr, expected, timeout);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.atomic.wait.i32(i32* %{{.*}}, i32 %{{.*}}, i64 %{{.*}})
|
|
// WEBASSEMBLY64: call i32 @llvm.wasm.atomic.wait.i32(i32* %{{.*}}, i32 %{{.*}}, i64 %{{.*}})
|
|
}
|
|
|
|
int atomic_wait_i64(long long *addr, long long expected, long long timeout) {
|
|
return __builtin_wasm_atomic_wait_i64(addr, expected, timeout);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.atomic.wait.i64(i64* %{{.*}}, i64 %{{.*}}, i64 %{{.*}})
|
|
// WEBASSEMBLY64: call i32 @llvm.wasm.atomic.wait.i64(i64* %{{.*}}, i64 %{{.*}}, i64 %{{.*}})
|
|
}
|
|
|
|
unsigned int atomic_notify(int *addr, unsigned int count) {
|
|
return __builtin_wasm_atomic_notify(addr, count);
|
|
// WEBASSEMBLY32: call i32 @llvm.wasm.atomic.notify(i32* %{{.*}}, i32 %{{.*}})
|
|
// WEBASSEMBLY64: call i32 @llvm.wasm.atomic.notify(i32* %{{.*}}, i32 %{{.*}})
|
|
}
|
|
|
|
int trunc_s_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_s_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.signed.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_u_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_u_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.unsigned.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_s_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_s_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.signed.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_u_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_u_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.unsigned.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_s_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_s_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.signed.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_u_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_u_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.unsigned.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_s_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_s_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.signed.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_u_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_u_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.unsigned.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_s_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_s_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.saturate.signed.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_u_i32_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_u_i32_f32(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.saturate.unsigned.i32.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_s_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_s_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.saturate.signed.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int trunc_saturate_u_i32_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_u_i32_f64(f);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.trunc.saturate.unsigned.i32.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_s_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_s_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.saturate.signed.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_u_i64_f32(float f) {
|
|
return __builtin_wasm_trunc_saturate_u_i64_f32(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.saturate.unsigned.i64.f32(float %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_s_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_s_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.saturate.signed.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long trunc_saturate_u_i64_f64(double f) {
|
|
return __builtin_wasm_trunc_saturate_u_i64_f64(f);
|
|
// WEBASSEMBLY: call i64 @llvm.wasm.trunc.saturate.unsigned.i64.f64(double %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
float min_f32(float x, float y) {
|
|
return __builtin_wasm_min_f32(x, y);
|
|
// WEBASSEMBLY: call float @llvm.minimum.f32(float %x, float %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
float max_f32(float x, float y) {
|
|
return __builtin_wasm_max_f32(x, y);
|
|
// WEBASSEMBLY: call float @llvm.maximum.f32(float %x, float %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
double min_f64(double x, double y) {
|
|
return __builtin_wasm_min_f64(x, y);
|
|
// WEBASSEMBLY: call double @llvm.minimum.f64(double %x, double %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
double max_f64(double x, double y) {
|
|
return __builtin_wasm_max_f64(x, y);
|
|
// WEBASSEMBLY: call double @llvm.maximum.f64(double %x, double %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int extract_lane_s_i8x16(i8x16 v) {
|
|
return __builtin_wasm_extract_lane_s_i8x16(v, 13);
|
|
// MISSING-SIMD: error: '__builtin_wasm_extract_lane_s_i8x16' needs target feature simd128
|
|
// WEBASSEMBLY: extractelement <16 x i8> %v, i32 13
|
|
// WEBASSEMBLY-NEXT: sext
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int extract_lane_u_i8x16(i8x16 v) {
|
|
return __builtin_wasm_extract_lane_u_i8x16(v, 13);
|
|
// WEBASSEMBLY: extractelement <16 x i8> %v, i32 13
|
|
// WEBASSEMBLY-NEXT: zext
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int extract_lane_s_i16x8(i16x8 v) {
|
|
return __builtin_wasm_extract_lane_s_i16x8(v, 7);
|
|
// WEBASSEMBLY: extractelement <8 x i16> %v, i32 7
|
|
// WEBASSEMBLY-NEXT: sext
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int extract_lane_u_i16x8(i16x8 v) {
|
|
return __builtin_wasm_extract_lane_u_i16x8(v, 7);
|
|
// WEBASSEMBLY: extractelement <8 x i16> %v, i32 7
|
|
// WEBASSEMBLY-NEXT: zext
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int extract_lane_i32x4(i32x4 v) {
|
|
return __builtin_wasm_extract_lane_i32x4(v, 3);
|
|
// WEBASSEMBLY: extractelement <4 x i32> %v, i32 3
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
long long extract_lane_i64x2(i64x2 v) {
|
|
return __builtin_wasm_extract_lane_i64x2(v, 1);
|
|
// WEBASSEMBLY: extractelement <2 x i64> %v, i32 1
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
float extract_lane_f32x4(f32x4 v) {
|
|
return __builtin_wasm_extract_lane_f32x4(v, 3);
|
|
// WEBASSEMBLY: extractelement <4 x float> %v, i32 3
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
double extract_lane_f64x2(f64x2 v) {
|
|
return __builtin_wasm_extract_lane_f64x2(v, 1);
|
|
// WEBASSEMBLY: extractelement <2 x double> %v, i32 1
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 replace_lane_i8x16(i8x16 v, int x) {
|
|
return __builtin_wasm_replace_lane_i8x16(v, 13, x);
|
|
// WEBASSEMBLY: trunc i32 %x to i8
|
|
// WEBASSEMBLY-NEXT: insertelement <16 x i8> %v, i8 %{{.*}}, i32 13
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 replace_lane_i16x8(i16x8 v, int x) {
|
|
return __builtin_wasm_replace_lane_i16x8(v, 7, x);
|
|
// WEBASSEMBLY: trunc i32 %x to i16
|
|
// WEBASSEMBLY-NEXT: insertelement <8 x i16> %v, i16 %{{.*}}, i32 7
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 replace_lane_i32x4(i32x4 v, int x) {
|
|
return __builtin_wasm_replace_lane_i32x4(v, 3, x);
|
|
// WEBASSEMBLY: insertelement <4 x i32> %v, i32 %x, i32 3
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i64x2 replace_lane_i64x2(i64x2 v, long long x) {
|
|
return __builtin_wasm_replace_lane_i64x2(v, 1, x);
|
|
// WEBASSEMBLY: insertelement <2 x i64> %v, i64 %x, i32 1
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 replace_lane_f32x4(f32x4 v, float x) {
|
|
return __builtin_wasm_replace_lane_f32x4(v, 3, x);
|
|
// WEBASSEMBLY: insertelement <4 x float> %v, float %x, i32 3
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 replace_lane_f64x2(f64x2 v, double x) {
|
|
return __builtin_wasm_replace_lane_f64x2(v, 1, x);
|
|
// WEBASSEMBLY: insertelement <2 x double> %v, double %x, i32 1
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 add_saturate_s_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_add_saturate_s_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.sadd.sat.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 add_saturate_u_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_add_saturate_u_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.uadd.sat.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 add_saturate_s_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_add_saturate_s_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.sadd.sat.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 add_saturate_u_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_add_saturate_u_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.uadd.sat.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 sub_saturate_s_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_sub_saturate_s_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.sub.saturate.signed.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 sub_saturate_u_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_sub_saturate_u_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.sub.saturate.unsigned.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 sub_saturate_s_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_sub_saturate_s_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.sub.saturate.signed.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 sub_saturate_u_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_sub_saturate_u_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.sub.saturate.unsigned.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 avgr_u_i8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_avgr_u_i8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.avgr.unsigned.v16i8(
|
|
// WEBASSEMBLY-SAME: <16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i16x8 avgr_u_i16x8(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_avgr_u_i16x8(x, y);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.avgr.unsigned.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 dot_i16x8_s(i16x8 x, i16x8 y) {
|
|
return __builtin_wasm_dot_s_i32x4_i16x8(x, y);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.dot(<8 x i16> %x, <8 x i16> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 bitselect(i32x4 x, i32x4 y, i32x4 c) {
|
|
return __builtin_wasm_bitselect(x, y, c);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.bitselect.v4i32(
|
|
// WEBASSEMBLY-SAME: <4 x i32> %x, <4 x i32> %y, <4 x i32> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
int any_true_i8x16(i8x16 x) {
|
|
return __builtin_wasm_any_true_i8x16(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.anytrue.v16i8(<16 x i8> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int any_true_i16x8(i16x8 x) {
|
|
return __builtin_wasm_any_true_i16x8(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.anytrue.v8i16(<8 x i16> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int any_true_i32x4(i32x4 x) {
|
|
return __builtin_wasm_any_true_i32x4(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.anytrue.v4i32(<4 x i32> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int any_true_i64x2(i64x2 x) {
|
|
return __builtin_wasm_any_true_i64x2(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.anytrue.v2i64(<2 x i64> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i8x16(i8x16 x) {
|
|
return __builtin_wasm_all_true_i8x16(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v16i8(<16 x i8> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i16x8(i16x8 x) {
|
|
return __builtin_wasm_all_true_i16x8(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v8i16(<8 x i16> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i32x4(i32x4 x) {
|
|
return __builtin_wasm_all_true_i32x4(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v4i32(<4 x i32> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
int all_true_i64x2(i64x2 x) {
|
|
return __builtin_wasm_all_true_i64x2(x);
|
|
// WEBASSEMBLY: call i32 @llvm.wasm.alltrue.v2i64(<2 x i64> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 abs_f32x4(f32x4 x) {
|
|
return __builtin_wasm_abs_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.fabs.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 abs_f64x2(f64x2 x) {
|
|
return __builtin_wasm_abs_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.fabs.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 min_f32x4(f32x4 x, f32x4 y) {
|
|
return __builtin_wasm_min_f32x4(x, y);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.minimum.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %x, <4 x float> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 max_f32x4(f32x4 x, f32x4 y) {
|
|
return __builtin_wasm_max_f32x4(x, y);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.maximum.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %x, <4 x float> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 min_f64x2(f64x2 x, f64x2 y) {
|
|
return __builtin_wasm_min_f64x2(x, y);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.minimum.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %x, <2 x double> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 max_f64x2(f64x2 x, f64x2 y) {
|
|
return __builtin_wasm_max_f64x2(x, y);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.maximum.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %x, <2 x double> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 sqrt_f32x4(f32x4 x) {
|
|
return __builtin_wasm_sqrt_f32x4(x);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.sqrt.v4f32(<4 x float> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f64x2 sqrt_f64x2(f64x2 x) {
|
|
return __builtin_wasm_sqrt_f64x2(x);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.sqrt.v2f64(<2 x double> %x)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
f32x4 qfma_f32x4(f32x4 a, f32x4 b, f32x4 c) {
|
|
return __builtin_wasm_qfma_f32x4(a, b, c);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.qfma.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %a, <4 x float> %b, <4 x float> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f32x4 qfms_f32x4(f32x4 a, f32x4 b, f32x4 c) {
|
|
return __builtin_wasm_qfms_f32x4(a, b, c);
|
|
// WEBASSEMBLY: call <4 x float> @llvm.wasm.qfms.v4f32(
|
|
// WEBASSEMBLY-SAME: <4 x float> %a, <4 x float> %b, <4 x float> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 qfma_f64x2(f64x2 a, f64x2 b, f64x2 c) {
|
|
return __builtin_wasm_qfma_f64x2(a, b, c);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.qfma.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %a, <2 x double> %b, <2 x double> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
f64x2 qfms_f64x2(f64x2 a, f64x2 b, f64x2 c) {
|
|
return __builtin_wasm_qfms_f64x2(a, b, c);
|
|
// WEBASSEMBLY: call <2 x double> @llvm.wasm.qfms.v2f64(
|
|
// WEBASSEMBLY-SAME: <2 x double> %a, <2 x double> %b, <2 x double> %c)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 trunc_saturate_s_i32x4_f32x4(f32x4 f) {
|
|
return __builtin_wasm_trunc_saturate_s_i32x4_f32x4(f);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.trunc.saturate.signed.v4i32.v4f32(<4 x float> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i32x4 trunc_saturate_u_i32x4_f32x4(f32x4 f) {
|
|
return __builtin_wasm_trunc_saturate_u_i32x4_f32x4(f);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.trunc.saturate.unsigned.v4i32.v4f32(<4 x float> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i64x2 trunc_saturate_s_i64x2_f64x2(f64x2 f) {
|
|
return __builtin_wasm_trunc_saturate_s_i64x2_f64x2(f);
|
|
// WEBASSEMBLY: call <2 x i64> @llvm.wasm.trunc.saturate.signed.v2i64.v2f64(<2 x double> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i64x2 trunc_saturate_u_i64x2_f64x2(f64x2 f) {
|
|
return __builtin_wasm_trunc_saturate_u_i64x2_f64x2(f);
|
|
// WEBASSEMBLY: call <2 x i64> @llvm.wasm.trunc.saturate.unsigned.v2i64.v2f64(<2 x double> %f)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|
|
|
|
i8x16 narrow_s_i8x16_i16x8(i16x8 low, i16x8 high) {
|
|
return __builtin_wasm_narrow_s_i8x16_i16x8(low, high);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.narrow.signed.v16i8.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %low, <8 x i16> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i8x16 narrow_u_i8x16_i16x8(i16x8 low, i16x8 high) {
|
|
return __builtin_wasm_narrow_u_i8x16_i16x8(low, high);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.narrow.unsigned.v16i8.v8i16(
|
|
// WEBASSEMBLY-SAME: <8 x i16> %low, <8 x i16> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i16x8 narrow_s_i16x8_i32x4(i32x4 low, i32x4 high) {
|
|
return __builtin_wasm_narrow_s_i16x8_i32x4(low, high);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.narrow.signed.v8i16.v4i32(
|
|
// WEBASSEMBLY-SAME: <4 x i32> %low, <4 x i32> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i16x8 narrow_u_i16x8_i32x4(i32x4 low, i32x4 high) {
|
|
return __builtin_wasm_narrow_u_i16x8_i32x4(low, high);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.narrow.unsigned.v8i16.v4i32(
|
|
// WEBASSEMBLY-SAME: <4 x i32> %low, <4 x i32> %high)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i16x8 widen_low_s_i16x8_i8x16(i8x16 v) {
|
|
return __builtin_wasm_widen_low_s_i16x8_i8x16(v);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.widen.low.signed.v8i16.v16i8(<16 x i8> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i16x8 widen_high_s_i16x8_i8x16(i8x16 v) {
|
|
return __builtin_wasm_widen_high_s_i16x8_i8x16(v);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.widen.high.signed.v8i16.v16i8(<16 x i8> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i16x8 widen_low_u_i16x8_i8x16(i8x16 v) {
|
|
return __builtin_wasm_widen_low_u_i16x8_i8x16(v);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.widen.low.unsigned.v8i16.v16i8(<16 x i8> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i16x8 widen_high_u_i16x8_i8x16(i8x16 v) {
|
|
return __builtin_wasm_widen_high_u_i16x8_i8x16(v);
|
|
// WEBASSEMBLY: call <8 x i16> @llvm.wasm.widen.high.unsigned.v8i16.v16i8(<16 x i8> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i32x4 widen_low_s_i32x4_i16x8(i16x8 v) {
|
|
return __builtin_wasm_widen_low_s_i32x4_i16x8(v);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.widen.low.signed.v4i32.v8i16(<8 x i16> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i32x4 widen_high_s_i32x4_i16x8(i16x8 v) {
|
|
return __builtin_wasm_widen_high_s_i32x4_i16x8(v);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.widen.high.signed.v4i32.v8i16(<8 x i16> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i32x4 widen_low_u_i32x4_i16x8(i16x8 v) {
|
|
return __builtin_wasm_widen_low_u_i32x4_i16x8(v);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.widen.low.unsigned.v4i32.v8i16(<8 x i16> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i32x4 widen_high_u_i32x4_i16x8(i16x8 v) {
|
|
return __builtin_wasm_widen_high_u_i32x4_i16x8(v);
|
|
// WEBASSEMBLY: call <4 x i32> @llvm.wasm.widen.high.unsigned.v4i32.v8i16(<8 x i16> %v)
|
|
// WEBASSEMBLY: ret
|
|
}
|
|
|
|
i8x16 swizzle_v8x16(i8x16 x, i8x16 y) {
|
|
return __builtin_wasm_swizzle_v8x16(x, y);
|
|
// WEBASSEMBLY: call <16 x i8> @llvm.wasm.swizzle(<16 x i8> %x, <16 x i8> %y)
|
|
// WEBASSEMBLY-NEXT: ret
|
|
}
|