root / lab4 / .minix-src / include / clang-3.6 / bmiintrin.h @ 14
History | View | Annotate | Download (4.79 KB)
1 |
/*===---- bmiintrin.h - BMI intrinsics -------------------------------------===
|
---|---|
2 |
*
|
3 |
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
4 |
* of this software and associated documentation files (the "Software"), to deal
|
5 |
* in the Software without restriction, including without limitation the rights
|
6 |
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
7 |
* copies of the Software, and to permit persons to whom the Software is
|
8 |
* furnished to do so, subject to the following conditions:
|
9 |
*
|
10 |
* The above copyright notice and this permission notice shall be included in
|
11 |
* all copies or substantial portions of the Software.
|
12 |
*
|
13 |
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
14 |
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
15 |
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
16 |
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
17 |
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
18 |
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
19 |
* THE SOFTWARE.
|
20 |
*
|
21 |
*===-----------------------------------------------------------------------===
|
22 |
*/
|
23 |
|
24 |
#if !defined __X86INTRIN_H && !defined __IMMINTRIN_H
|
25 |
#error "Never use <bmiintrin.h> directly; include <x86intrin.h> instead." |
26 |
#endif
|
27 |
|
28 |
#ifndef __BMI__
|
29 |
# error "BMI instruction set not enabled" |
30 |
#endif /* __BMI__ */ |
31 |
|
32 |
#ifndef __BMIINTRIN_H
|
33 |
#define __BMIINTRIN_H
|
34 |
|
35 |
#define _tzcnt_u16(a) (__tzcnt_u16((a)))
|
36 |
#define _andn_u32(a, b) (__andn_u32((a), (b)))
|
37 |
/* _bextr_u32 != __bextr_u32 */
|
38 |
#define _blsi_u32(a) (__blsi_u32((a)))
|
39 |
#define _blsmsk_u32(a) (__blsmsk_u32((a)))
|
40 |
#define _blsr_u32(a) (__blsr_u32((a)))
|
41 |
#define _tzcnt_u32(a) (__tzcnt_u32((a)))
|
42 |
|
43 |
static __inline__ unsigned short __attribute__((__always_inline__, __nodebug__)) |
44 |
__tzcnt_u16(unsigned short __X) |
45 |
{ |
46 |
return __X ? __builtin_ctzs(__X) : 16; |
47 |
} |
48 |
|
49 |
static __inline__ unsigned int __attribute__((__always_inline__, __nodebug__)) |
50 |
__andn_u32(unsigned int __X, unsigned int __Y) |
51 |
{ |
52 |
return ~__X & __Y;
|
53 |
} |
54 |
|
55 |
/* AMD-specified, double-leading-underscore version of BEXTR */
|
56 |
static __inline__ unsigned int __attribute__((__always_inline__, __nodebug__)) |
57 |
__bextr_u32(unsigned int __X, unsigned int __Y) |
58 |
{ |
59 |
return __builtin_ia32_bextr_u32(__X, __Y);
|
60 |
} |
61 |
|
62 |
/* Intel-specified, single-leading-underscore version of BEXTR */
|
63 |
static __inline__ unsigned int __attribute__((__always_inline__, __nodebug__)) |
64 |
_bextr_u32(unsigned int __X, unsigned int __Y, unsigned int __Z) |
65 |
{ |
66 |
return __builtin_ia32_bextr_u32 (__X, ((__Y & 0xff) | ((__Z & 0xff) << 8))); |
67 |
} |
68 |
|
69 |
static __inline__ unsigned int __attribute__((__always_inline__, __nodebug__)) |
70 |
__blsi_u32(unsigned int __X) |
71 |
{ |
72 |
return __X & -__X;
|
73 |
} |
74 |
|
75 |
static __inline__ unsigned int __attribute__((__always_inline__, __nodebug__)) |
76 |
__blsmsk_u32(unsigned int __X) |
77 |
{ |
78 |
return __X ^ (__X - 1); |
79 |
} |
80 |
|
81 |
static __inline__ unsigned int __attribute__((__always_inline__, __nodebug__)) |
82 |
__blsr_u32(unsigned int __X) |
83 |
{ |
84 |
return __X & (__X - 1); |
85 |
} |
86 |
|
87 |
static __inline__ unsigned int __attribute__((__always_inline__, __nodebug__)) |
88 |
__tzcnt_u32(unsigned int __X) |
89 |
{ |
90 |
return __X ? __builtin_ctz(__X) : 32; |
91 |
} |
92 |
|
93 |
#ifdef __x86_64__
|
94 |
|
95 |
#define _andn_u64(a, b) (__andn_u64((a), (b)))
|
96 |
/* _bextr_u64 != __bextr_u64 */
|
97 |
#define _blsi_u64(a) (__blsi_u64((a)))
|
98 |
#define _blsmsk_u64(a) (__blsmsk_u64((a)))
|
99 |
#define _blsr_u64(a) (__blsr_u64((a)))
|
100 |
#define _tzcnt_u64(a) (__tzcnt_u64((a)))
|
101 |
|
102 |
static __inline__ unsigned long long __attribute__((__always_inline__, __nodebug__)) |
103 |
__andn_u64 (unsigned long long __X, unsigned long long __Y) |
104 |
{ |
105 |
return ~__X & __Y;
|
106 |
} |
107 |
|
108 |
/* AMD-specified, double-leading-underscore version of BEXTR */
|
109 |
static __inline__ unsigned long long __attribute__((__always_inline__, __nodebug__)) |
110 |
__bextr_u64(unsigned long long __X, unsigned long long __Y) |
111 |
{ |
112 |
return __builtin_ia32_bextr_u64(__X, __Y);
|
113 |
} |
114 |
|
115 |
/* Intel-specified, single-leading-underscore version of BEXTR */
|
116 |
static __inline__ unsigned long long __attribute__((__always_inline__, __nodebug__)) |
117 |
_bextr_u64(unsigned long long __X, unsigned int __Y, unsigned int __Z) |
118 |
{ |
119 |
return __builtin_ia32_bextr_u64 (__X, ((__Y & 0xff) | ((__Z & 0xff) << 8))); |
120 |
} |
121 |
|
122 |
static __inline__ unsigned long long __attribute__((__always_inline__, __nodebug__)) |
123 |
__blsi_u64(unsigned long long __X) |
124 |
{ |
125 |
return __X & -__X;
|
126 |
} |
127 |
|
128 |
static __inline__ unsigned long long __attribute__((__always_inline__, __nodebug__)) |
129 |
__blsmsk_u64(unsigned long long __X) |
130 |
{ |
131 |
return __X ^ (__X - 1); |
132 |
} |
133 |
|
134 |
static __inline__ unsigned long long __attribute__((__always_inline__, __nodebug__)) |
135 |
__blsr_u64(unsigned long long __X) |
136 |
{ |
137 |
return __X & (__X - 1); |
138 |
} |
139 |
|
140 |
static __inline__ unsigned long long __attribute__((__always_inline__, __nodebug__)) |
141 |
__tzcnt_u64(unsigned long long __X) |
142 |
{ |
143 |
return __X ? __builtin_ctzll(__X) : 64; |
144 |
} |
145 |
|
146 |
#endif /* __x86_64__ */ |
147 |
|
148 |
#endif /* __BMIINTRIN_H */ |