diff options
Diffstat (limited to 'test/CodeGen/bmi-builtins.c')
| -rw-r--r-- | test/CodeGen/bmi-builtins.c | 87 | 
1 files changed, 86 insertions, 1 deletions
diff --git a/test/CodeGen/bmi-builtins.c b/test/CodeGen/bmi-builtins.c index 2e1ba12d9b3ca..92332e3a1265d 100644 --- a/test/CodeGen/bmi-builtins.c +++ b/test/CodeGen/bmi-builtins.c @@ -5,6 +5,14 @@  #include <x86intrin.h> +// The double underscore intrinsics are for compatibility with  +// AMD's BMI interface. The single underscore intrinsics +// are for compatibility with Intel's BMI interface. +// Apart from the underscores, the interfaces are identical +// except in one case: although the 'bextr' register-form  +// instruction is identical in hardware, the AMD and Intel  +// intrinsics are different!  +  unsigned short test__tzcnt_u16(unsigned short __X) {    // CHECK: @llvm.cttz.i16    return __tzcnt_u16(__X); @@ -39,7 +47,7 @@ unsigned int test__blsr_u32(unsigned int __X) {    return __blsr_u32(__X);  } -unsigned int test_tzcnt_u32(unsigned int __X) { +unsigned int test__tzcnt_u32(unsigned int __X) {    // CHECK: @llvm.cttz.i32    return __tzcnt_u32(__X);  } @@ -77,3 +85,80 @@ unsigned long long test__tzcnt_u64(unsigned long long __X) {    // CHECK: @llvm.cttz.i64    return __tzcnt_u64(__X);  } + +// Intel intrinsics + +unsigned short test_tzcnt_u16(unsigned short __X) { +  // CHECK: @llvm.cttz.i16 +  return _tzcnt_u16(__X); +} + +unsigned int test_andn_u32(unsigned int __X, unsigned int __Y) { +  // CHECK: [[DEST:%.*]] = xor i32 %{{.*}}, -1 +  // CHECK-NEXT: %{{.*}} = and i32 %{{.*}}, [[DEST]] +  return _andn_u32(__X, __Y); +} + +unsigned int test_bextr_u32(unsigned int __X, unsigned int __Y,  +                            unsigned int __Z) { +  // CHECK: @llvm.x86.bmi.bextr.32 +  return _bextr_u32(__X, __Y, __Z); +} + +unsigned int test_blsi_u32(unsigned int __X) { +  // CHECK: [[DEST:%.*]] = sub i32 0, [[SRC:%.*]] +  // CHECK-NEXT: %{{.*}} = and i32 [[SRC]], [[DEST]] +  return _blsi_u32(__X); +} + +unsigned int test_blsmsk_u32(unsigned int __X) { +  // CHECK: [[DEST:%.*]] = add i32 [[SRC:%.*]], -1 +  // CHECK-NEXT: %{{.*}} = xor i32 [[DEST]], [[SRC]] +  return _blsmsk_u32(__X); +} + +unsigned int test_blsr_u32(unsigned int __X) { +  // CHECK: [[DEST:%.*]] = add i32 [[SRC:%.*]], -1 +  // CHECK-NEXT: %{{.*}} = and i32 [[DEST]], [[SRC]] +  return _blsr_u32(__X); +} + +unsigned int test_tzcnt_u32(unsigned int __X) { +  // CHECK: @llvm.cttz.i32 +  return _tzcnt_u32(__X); +} + +unsigned long long test_andn_u64(unsigned long __X, unsigned long __Y) { +  // CHECK: [[DEST:%.*]] = xor i64 %{{.*}}, -1 +  // CHECK-NEXT: %{{.*}} = and i64 %{{.*}}, [[DEST]] +  return _andn_u64(__X, __Y); +} + +unsigned long long test_bextr_u64(unsigned long __X, unsigned int __Y,  +                                  unsigned int __Z) { +  // CHECK: @llvm.x86.bmi.bextr.64 +  return _bextr_u64(__X, __Y, __Z); +} + +unsigned long long test_blsi_u64(unsigned long long __X) { +  // CHECK: [[DEST:%.*]] = sub i64 0, [[SRC:%.*]] +  // CHECK-NEXT: %{{.*}} = and i64 [[SRC]], [[DEST]] +  return _blsi_u64(__X); +} + +unsigned long long test_blsmsk_u64(unsigned long long __X) { +  // CHECK: [[DEST:%.*]] = add i64 [[SRC:%.*]], -1 +  // CHECK-NEXT: %{{.*}} = xor i64 [[DEST]], [[SRC]] +  return _blsmsk_u64(__X); +} + +unsigned long long test_blsr_u64(unsigned long long __X) { +  // CHECK: [[DEST:%.*]] = add i64 [[SRC:%.*]], -1 +  // CHECK-NEXT: %{{.*}} = and i64 [[DEST]], [[SRC]] +  return _blsr_u64(__X); +} + +unsigned long long test_tzcnt_u64(unsigned long long __X) { +  // CHECK: @llvm.cttz.i64 +  return _tzcnt_u64(__X); +}  | 
