1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
|
// RUN: %clang_cc1 %s -O3 -triple=x86_64-apple-darwin -target-feature +bmi -emit-llvm -o - | FileCheck %s
// Don't include mm_malloc.h, it's system specific.
#define __MM_MALLOC_H
#include <x86intrin.h>
// The double underscore intrinsics are for compatibility with
// AMD's BMI interface. The single underscore intrinsics
// are for compatibility with Intel's BMI interface.
// Apart from the underscores, the interfaces are identical
// except in one case: although the 'bextr' register-form
// instruction is identical in hardware, the AMD and Intel
// intrinsics are different!
unsigned short test__tzcnt_u16(unsigned short __X) {
// CHECK: @llvm.cttz.i16
return __tzcnt_u16(__X);
}
unsigned int test__andn_u32(unsigned int __X, unsigned int __Y) {
// CHECK: [[DEST:%.*]] = xor i32 %{{.*}}, -1
// CHECK-NEXT: %{{.*}} = and i32 %{{.*}}, [[DEST]]
return __andn_u32(__X, __Y);
}
unsigned int test__bextr_u32(unsigned int __X, unsigned int __Y) {
// CHECK: @llvm.x86.bmi.bextr.32
return __bextr_u32(__X, __Y);
}
unsigned int test__blsi_u32(unsigned int __X) {
// CHECK: [[DEST:%.*]] = sub i32 0, [[SRC:%.*]]
// CHECK-NEXT: %{{.*}} = and i32 [[SRC]], [[DEST]]
return __blsi_u32(__X);
}
unsigned int test__blsmsk_u32(unsigned int __X) {
// CHECK: [[DEST:%.*]] = add i32 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = xor i32 [[DEST]], [[SRC]]
return __blsmsk_u32(__X);
}
unsigned int test__blsr_u32(unsigned int __X) {
// CHECK: [[DEST:%.*]] = add i32 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = and i32 [[DEST]], [[SRC]]
return __blsr_u32(__X);
}
unsigned int test__tzcnt_u32(unsigned int __X) {
// CHECK: @llvm.cttz.i32
return __tzcnt_u32(__X);
}
unsigned long long test__andn_u64(unsigned long __X, unsigned long __Y) {
// CHECK: [[DEST:%.*]] = xor i64 %{{.*}}, -1
// CHECK-NEXT: %{{.*}} = and i64 %{{.*}}, [[DEST]]
return __andn_u64(__X, __Y);
}
unsigned long long test__bextr_u64(unsigned long __X, unsigned long __Y) {
// CHECK: @llvm.x86.bmi.bextr.64
return __bextr_u64(__X, __Y);
}
unsigned long long test__blsi_u64(unsigned long long __X) {
// CHECK: [[DEST:%.*]] = sub i64 0, [[SRC:%.*]]
// CHECK-NEXT: %{{.*}} = and i64 [[SRC]], [[DEST]]
return __blsi_u64(__X);
}
unsigned long long test__blsmsk_u64(unsigned long long __X) {
// CHECK: [[DEST:%.*]] = add i64 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = xor i64 [[DEST]], [[SRC]]
return __blsmsk_u64(__X);
}
unsigned long long test__blsr_u64(unsigned long long __X) {
// CHECK: [[DEST:%.*]] = add i64 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = and i64 [[DEST]], [[SRC]]
return __blsr_u64(__X);
}
unsigned long long test__tzcnt_u64(unsigned long long __X) {
// CHECK: @llvm.cttz.i64
return __tzcnt_u64(__X);
}
// Intel intrinsics
unsigned short test_tzcnt_u16(unsigned short __X) {
// CHECK: @llvm.cttz.i16
return _tzcnt_u16(__X);
}
unsigned int test_andn_u32(unsigned int __X, unsigned int __Y) {
// CHECK: [[DEST:%.*]] = xor i32 %{{.*}}, -1
// CHECK-NEXT: %{{.*}} = and i32 %{{.*}}, [[DEST]]
return _andn_u32(__X, __Y);
}
unsigned int test_bextr_u32(unsigned int __X, unsigned int __Y,
unsigned int __Z) {
// CHECK: @llvm.x86.bmi.bextr.32
return _bextr_u32(__X, __Y, __Z);
}
unsigned int test_blsi_u32(unsigned int __X) {
// CHECK: [[DEST:%.*]] = sub i32 0, [[SRC:%.*]]
// CHECK-NEXT: %{{.*}} = and i32 [[SRC]], [[DEST]]
return _blsi_u32(__X);
}
unsigned int test_blsmsk_u32(unsigned int __X) {
// CHECK: [[DEST:%.*]] = add i32 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = xor i32 [[DEST]], [[SRC]]
return _blsmsk_u32(__X);
}
unsigned int test_blsr_u32(unsigned int __X) {
// CHECK: [[DEST:%.*]] = add i32 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = and i32 [[DEST]], [[SRC]]
return _blsr_u32(__X);
}
unsigned int test_tzcnt_u32(unsigned int __X) {
// CHECK: @llvm.cttz.i32
return _tzcnt_u32(__X);
}
unsigned long long test_andn_u64(unsigned long __X, unsigned long __Y) {
// CHECK: [[DEST:%.*]] = xor i64 %{{.*}}, -1
// CHECK-NEXT: %{{.*}} = and i64 %{{.*}}, [[DEST]]
return _andn_u64(__X, __Y);
}
unsigned long long test_bextr_u64(unsigned long __X, unsigned int __Y,
unsigned int __Z) {
// CHECK: @llvm.x86.bmi.bextr.64
return _bextr_u64(__X, __Y, __Z);
}
unsigned long long test_blsi_u64(unsigned long long __X) {
// CHECK: [[DEST:%.*]] = sub i64 0, [[SRC:%.*]]
// CHECK-NEXT: %{{.*}} = and i64 [[SRC]], [[DEST]]
return _blsi_u64(__X);
}
unsigned long long test_blsmsk_u64(unsigned long long __X) {
// CHECK: [[DEST:%.*]] = add i64 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = xor i64 [[DEST]], [[SRC]]
return _blsmsk_u64(__X);
}
unsigned long long test_blsr_u64(unsigned long long __X) {
// CHECK: [[DEST:%.*]] = add i64 [[SRC:%.*]], -1
// CHECK-NEXT: %{{.*}} = and i64 [[DEST]], [[SRC]]
return _blsr_u64(__X);
}
unsigned long long test_tzcnt_u64(unsigned long long __X) {
// CHECK: @llvm.cttz.i64
return _tzcnt_u64(__X);
}
|