Print this page
6263 add missing cc clobbers to intel atomic inlines
Split |
Close |
Expand all |
Collapse all |
--- old/usr/src/uts/intel/asm/atomic.h
+++ new/usr/src/uts/intel/asm/atomic.h
1 1 /*
2 2 * CDDL HEADER START
3 3 *
4 4 * The contents of this file are subject to the terms of the
5 5 * Common Development and Distribution License, Version 1.0 only
6 6 * (the "License"). You may not use this file except in compliance
7 7 * with the License.
8 8 *
9 9 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10 10 * or http://www.opensolaris.org/os/licensing.
11 11 * See the License for the specific language governing permissions
12 12 * and limitations under the License.
13 13 *
14 14 * When distributing Covered Code, include this CDDL HEADER in each
15 15 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
↓ open down ↓ |
15 lines elided |
↑ open up ↑ |
16 16 * If applicable, add the following below this CDDL HEADER, with the
17 17 * fields enclosed by brackets "[]" replaced with your own identifying
18 18 * information: Portions Copyright [yyyy] [name of copyright owner]
19 19 *
20 20 * CDDL HEADER END
21 21 */
22 22 /*
23 23 * Copyright 2005 Sun Microsystems, Inc. All rights reserved.
24 24 * Use is subject to license terms.
25 25 * Copyright 2014 Nexenta Systems, Inc. All rights reserved.
26 + * Copyright 2015 Josef 'Jeff' Sipek <jeffpc@josefsipek.net>
26 27 */
27 28
28 29 #ifndef _ASM_ATOMIC_H
29 30 #define _ASM_ATOMIC_H
30 31
31 32 #include <sys/ccompile.h>
32 33 #include <sys/types.h>
33 34
34 35 #ifdef __cplusplus
35 36 extern "C" {
36 37 #endif
37 38
38 39 #if !defined(__lint) && defined(__GNUC__)
39 40
40 41 /* BEGIN CSTYLED */
41 42 /*
42 43 * This file contains a number of static inline functions implementing
43 44 * various atomic variable functions. Note that these are *not* all of the
44 45 * atomic_* functions as defined in usr/src/uts/common/sys/atomic.h. All
45 46 * possible atomic_* functions are implemented in usr/src/common/atomic in
46 47 * pure assembly. In the absence of an identically named function in this
47 48 * header file, any use of the function will result in the compiler emitting
48 49 * a function call as usual. On the other hand, if an identically named
49 50 * function exists in this header as a static inline, the compiler will
50 51 * inline its contents and the linker never sees the symbol reference. We
51 52 * use this to avoid implementing some of the more complex and less used
52 53 * functions and instead falling back to function calls. Note that in some
53 54 * cases (e.g., atomic_inc_64) we implement a static inline only on AMD64
54 55 * but not i386.
55 56 */
56 57
57 58 /*
58 59 * Instruction suffixes for various operand sizes (assuming AMD64)
59 60 */
60 61 #define SUF_8 "b"
61 62 #define SUF_16 "w"
62 63 #define SUF_32 "l"
63 64 #define SUF_64 "q"
64 65
65 66 #if defined(__amd64)
66 67 #define SUF_LONG SUF_64
67 68 #define SUF_PTR SUF_64
68 69 #define __ATOMIC_OP64(...) __ATOMIC_OPXX(__VA_ARGS__)
69 70 #elif defined(__i386)
70 71 #define SUF_LONG SUF_32
71 72 #define SUF_PTR SUF_32
72 73 #define __ATOMIC_OP64(...)
73 74 #else
74 75 #error "port me"
↓ open down ↓ |
39 lines elided |
↑ open up ↑ |
75 76 #endif
76 77
77 78 #if defined(__amd64) || defined(__i386)
78 79
79 80 #define __ATOMIC_OPXX(fxn, type, op) \
80 81 extern __GNU_INLINE void \
81 82 fxn(volatile type *target) \
82 83 { \
83 84 __asm__ __volatile__( \
84 85 "lock; " op " %0" \
85 - : "+m" (*target)); \
86 + : "+m" (*target) \
87 + : /* no inputs */ \
88 + : "cc"); \
86 89 }
87 90
88 91 __ATOMIC_OPXX(atomic_inc_8, uint8_t, "inc" SUF_8)
89 92 __ATOMIC_OPXX(atomic_inc_16, uint16_t, "inc" SUF_16)
90 93 __ATOMIC_OPXX(atomic_inc_32, uint32_t, "inc" SUF_32)
91 94 __ATOMIC_OP64(atomic_inc_64, uint64_t, "inc" SUF_64)
92 95 __ATOMIC_OPXX(atomic_inc_uchar, uchar_t, "inc" SUF_8)
93 96 __ATOMIC_OPXX(atomic_inc_ushort, ushort_t, "inc" SUF_16)
94 97 __ATOMIC_OPXX(atomic_inc_uint, uint_t, "inc" SUF_32)
95 98 __ATOMIC_OPXX(atomic_inc_ulong, ulong_t, "inc" SUF_LONG)
96 99
97 100 __ATOMIC_OPXX(atomic_dec_8, uint8_t, "dec" SUF_8)
98 101 __ATOMIC_OPXX(atomic_dec_16, uint16_t, "dec" SUF_16)
99 102 __ATOMIC_OPXX(atomic_dec_32, uint32_t, "dec" SUF_32)
100 103 __ATOMIC_OP64(atomic_dec_64, uint64_t, "dec" SUF_64)
101 104 __ATOMIC_OPXX(atomic_dec_uchar, uchar_t, "dec" SUF_8)
102 105 __ATOMIC_OPXX(atomic_dec_ushort, ushort_t, "dec" SUF_16)
103 106 __ATOMIC_OPXX(atomic_dec_uint, uint_t, "dec" SUF_32)
104 107 __ATOMIC_OPXX(atomic_dec_ulong, ulong_t, "dec" SUF_LONG)
↓ open down ↓ |
9 lines elided |
↑ open up ↑ |
105 108
106 109 #undef __ATOMIC_OPXX
107 110
108 111 #define __ATOMIC_OPXX(fxn, type1, type2, op, reg) \
109 112 extern __GNU_INLINE void \
110 113 fxn(volatile type1 *target, type2 delta) \
111 114 { \
112 115 __asm__ __volatile__( \
113 116 "lock; " op " %1,%0" \
114 117 : "+m" (*target) \
115 - : "i" reg (delta)); \
118 + : "i" reg (delta) \
119 + : "cc"); \
116 120 }
117 121
118 122 __ATOMIC_OPXX(atomic_add_8, uint8_t, int8_t, "add" SUF_8, "q")
119 123 __ATOMIC_OPXX(atomic_add_16, uint16_t, int16_t, "add" SUF_16, "r")
120 124 __ATOMIC_OPXX(atomic_add_32, uint32_t, int32_t, "add" SUF_32, "r")
121 125 __ATOMIC_OP64(atomic_add_64, uint64_t, int64_t, "add" SUF_64, "r")
122 126 __ATOMIC_OPXX(atomic_add_char, uchar_t, signed char, "add" SUF_8, "q")
123 127 __ATOMIC_OPXX(atomic_add_short, ushort_t, short, "add" SUF_16, "r")
124 128 __ATOMIC_OPXX(atomic_add_int, uint_t, int, "add" SUF_32, "r")
125 129 __ATOMIC_OPXX(atomic_add_long, ulong_t, long, "add" SUF_LONG, "r")
126 130
127 131 /*
128 132 * We don't use the above macro here because atomic_add_ptr has an
129 133 * inconsistent type. The first argument should really be a 'volatile void
↓ open down ↓ |
4 lines elided |
↑ open up ↑ |
130 134 * **'.
131 135 */
132 136 extern __GNU_INLINE void
133 137 atomic_add_ptr(volatile void *target, ssize_t delta)
134 138 {
135 139 volatile void **tmp = (volatile void **)target;
136 140
137 141 __asm__ __volatile__(
138 142 "lock; add" SUF_PTR " %1,%0"
139 143 : "+m" (*tmp)
140 - : "ir" (delta));
144 + : "ir" (delta)
145 + : "cc");
141 146 }
142 147
143 148 __ATOMIC_OPXX(atomic_or_8, uint8_t, uint8_t, "or" SUF_8, "q")
144 149 __ATOMIC_OPXX(atomic_or_16, uint16_t, uint16_t, "or" SUF_16, "r")
145 150 __ATOMIC_OPXX(atomic_or_32, uint32_t, uint32_t, "or" SUF_32, "r")
146 151 __ATOMIC_OP64(atomic_or_64, uint64_t, uint64_t, "or" SUF_64, "r")
147 152 __ATOMIC_OPXX(atomic_or_uchar, uchar_t, uchar_t, "or" SUF_8, "q")
148 153 __ATOMIC_OPXX(atomic_or_ushort, ushort_t, ushort_t, "or" SUF_16, "r")
149 154 __ATOMIC_OPXX(atomic_or_uint, uint_t, uint_t, "or" SUF_32, "r")
150 155 __ATOMIC_OPXX(atomic_or_ulong, ulong_t, ulong_t, "or" SUF_LONG, "r")
151 156
152 157 __ATOMIC_OPXX(atomic_and_8, uint8_t, uint8_t, "and" SUF_8, "q")
153 158 __ATOMIC_OPXX(atomic_and_16, uint16_t, uint16_t, "and" SUF_16, "r")
154 159 __ATOMIC_OPXX(atomic_and_32, uint32_t, uint32_t, "and" SUF_32, "r")
155 160 __ATOMIC_OP64(atomic_and_64, uint64_t, uint64_t, "and" SUF_64, "r")
156 161 __ATOMIC_OPXX(atomic_and_uchar, uchar_t, uchar_t, "and" SUF_8, "q")
157 162 __ATOMIC_OPXX(atomic_and_ushort, ushort_t, ushort_t, "and" SUF_16, "r")
158 163 __ATOMIC_OPXX(atomic_and_uint, uint_t, uint_t, "and" SUF_32, "r")
159 164 __ATOMIC_OPXX(atomic_and_ulong, ulong_t, ulong_t, "and" SUF_LONG, "r")
160 165
161 166 #undef __ATOMIC_OPXX
162 167
163 168 #define __ATOMIC_OPXX(fxn, type, op, reg) \
164 169 extern __GNU_INLINE type \
165 170 fxn(volatile type *target, type cmp, type new) \
166 171 { \
167 172 type ret; \
168 173 __asm__ __volatile__( \
169 174 "lock; " op " %2,%0" \
170 175 : "+m" (*target), "=a" (ret) \
171 176 : reg (new), "1" (cmp) \
172 177 : "cc"); \
173 178 return (ret); \
174 179 }
175 180
176 181 __ATOMIC_OPXX(atomic_cas_8, uint8_t, "cmpxchg" SUF_8, "q")
177 182 __ATOMIC_OPXX(atomic_cas_16, uint16_t, "cmpxchg" SUF_16, "r")
178 183 __ATOMIC_OPXX(atomic_cas_32, uint32_t, "cmpxchg" SUF_32, "r")
179 184 __ATOMIC_OP64(atomic_cas_64, uint64_t, "cmpxchg" SUF_64, "r")
180 185 __ATOMIC_OPXX(atomic_cas_uchar, uchar_t, "cmpxchg" SUF_8, "q")
181 186 __ATOMIC_OPXX(atomic_cas_ushort, ushort_t, "cmpxchg" SUF_16, "r")
182 187 __ATOMIC_OPXX(atomic_cas_uint, uint_t, "cmpxchg" SUF_32, "r")
183 188 __ATOMIC_OPXX(atomic_cas_ulong, ulong_t, "cmpxchg" SUF_LONG, "r")
184 189
185 190 #undef __ATOMIC_OPXX
186 191
187 192 /*
188 193 * We don't use the above macro here because atomic_cas_ptr has an
189 194 * inconsistent type. The first argument should really be a 'volatile void
190 195 * **'.
191 196 */
192 197 extern __GNU_INLINE void *
193 198 atomic_cas_ptr(volatile void *target, void *cmp, void *new)
194 199 {
195 200 volatile void **tmp = (volatile void **)target;
196 201 void *ret;
197 202
198 203 __asm__ __volatile__(
199 204 "lock; cmpxchg" SUF_PTR " %2,%0"
200 205 : "+m" (*tmp), "=a" (ret)
201 206 : "r" (new), "1" (cmp)
202 207 : "cc");
203 208
204 209 return (ret);
205 210 }
206 211
207 212 #define __ATOMIC_OPXX(fxn, type, op, reg) \
208 213 extern __GNU_INLINE type \
209 214 fxn(volatile type *target, type val) \
210 215 { \
211 216 __asm__ __volatile__( \
212 217 op " %1,%0" \
213 218 : "+m" (*target), "+" reg (val)); \
214 219 return (val); \
215 220 }
216 221
217 222 __ATOMIC_OPXX(atomic_swap_8, uint8_t, "xchg" SUF_8, "q")
218 223 __ATOMIC_OPXX(atomic_swap_16, uint16_t, "xchg" SUF_16, "r")
219 224 __ATOMIC_OPXX(atomic_swap_32, uint32_t, "xchg" SUF_32, "r")
220 225 __ATOMIC_OP64(atomic_swap_64, uint64_t, "xchg" SUF_64, "r")
221 226 __ATOMIC_OPXX(atomic_swap_uchar, uchar_t, "xchg" SUF_8, "q")
222 227 __ATOMIC_OPXX(atomic_swap_ushort, ushort_t, "xchg" SUF_16, "r")
223 228 __ATOMIC_OPXX(atomic_swap_uint, uint_t, "xchg" SUF_32, "r")
224 229 __ATOMIC_OPXX(atomic_swap_ulong, ulong_t, "xchg" SUF_LONG, "r")
225 230
226 231 #undef __ATOMIC_OPXX
227 232
228 233 /*
229 234 * We don't use the above macro here because atomic_swap_ptr has an
230 235 * inconsistent type. The first argument should really be a 'volatile void
231 236 * **'.
232 237 */
233 238 extern __GNU_INLINE void *
234 239 atomic_swap_ptr(volatile void *target, void *val)
235 240 {
236 241 volatile void **tmp = (volatile void **)target;
237 242
238 243 __asm__ __volatile__(
239 244 "xchg" SUF_PTR " %1,%0"
240 245 : "+m" (*tmp), "+r" (val));
241 246
242 247 return (val);
243 248 }
244 249
245 250 #else
246 251 #error "port me"
247 252 #endif
248 253
249 254 #undef SUF_8
250 255 #undef SUF_16
251 256 #undef SUF_32
252 257 #undef SUF_64
253 258 #undef SUF_LONG
254 259 #undef SUF_PTR
255 260
256 261 #undef __ATOMIC_OP64
257 262
258 263 /* END CSTYLED */
259 264
260 265 #endif /* !__lint && __GNUC__ */
261 266
262 267 #ifdef __cplusplus
263 268 }
264 269 #endif
265 270
266 271 #endif /* _ASM_ATOMIC_H */
↓ open down ↓ |
116 lines elided |
↑ open up ↑ |
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX