root / lab4 / .minix-src / include / sys / mutex.h @ 13
History | View | Annotate | Download (6.33 KB)
1 |
/* $NetBSD: mutex.h,v 1.20 2010/02/08 09:54:27 skrll Exp $ */
|
---|---|
2 |
|
3 |
/*-
|
4 |
* Copyright (c) 2002, 2006, 2007, 2008, 2009 The NetBSD Foundation, Inc.
|
5 |
* All rights reserved.
|
6 |
*
|
7 |
* This code is derived from software contributed to The NetBSD Foundation
|
8 |
* by Jason R. Thorpe and Andrew Doran.
|
9 |
*
|
10 |
* Redistribution and use in source and binary forms, with or without
|
11 |
* modification, are permitted provided that the following conditions
|
12 |
* are met:
|
13 |
* 1. Redistributions of source code must retain the above copyright
|
14 |
* notice, this list of conditions and the following disclaimer.
|
15 |
* 2. Redistributions in binary form must reproduce the above copyright
|
16 |
* notice, this list of conditions and the following disclaimer in the
|
17 |
* documentation and/or other materials provided with the distribution.
|
18 |
*
|
19 |
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
|
20 |
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
|
21 |
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
22 |
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
|
23 |
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
24 |
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
25 |
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
26 |
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
27 |
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
28 |
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
29 |
* POSSIBILITY OF SUCH DAMAGE.
|
30 |
*/
|
31 |
|
32 |
#ifndef _SYS_MUTEX_H_
|
33 |
#define _SYS_MUTEX_H_
|
34 |
|
35 |
/*
|
36 |
* There are 2 types of mutexes:
|
37 |
*
|
38 |
* * Adaptive -- If the lock is already held, the thread attempting
|
39 |
* to acquire the lock determines if the thread that holds it is
|
40 |
* currently running. If so, it spins, else it sleeps.
|
41 |
*
|
42 |
* * Spin -- If the lock is already held, the thread attempting to
|
43 |
* acquire the lock spins. The IPL will be raised on entry.
|
44 |
*
|
45 |
* Machine dependent code must provide the following:
|
46 |
*
|
47 |
* struct mutex
|
48 |
* The actual mutex structure. This structure is mostly
|
49 |
* opaque to machine-independent code; most access are done
|
50 |
* through macros. However, machine-independent code must
|
51 |
* be able to access the following members:
|
52 |
*
|
53 |
* uintptr_t mtx_owner
|
54 |
* ipl_cookie_t mtx_ipl
|
55 |
* __cpu_simple_lock_t mtx_lock
|
56 |
*
|
57 |
* If an architecture can be considered 'simple' (no interlock required in
|
58 |
* the MP case, or no MP) it need only define __HAVE_SIMPLE_MUTEXES and
|
59 |
* provide the following:
|
60 |
*
|
61 |
* struct mutex
|
62 |
*
|
63 |
* [additionally:]
|
64 |
* volatile integer mtx_id
|
65 |
*
|
66 |
* MUTEX_RECEIVE(mtx)
|
67 |
* Post a load fence after acquiring the mutex, if necessary.
|
68 |
*
|
69 |
* MUTEX_GIVE(mtx)
|
70 |
* Post a load/store fence after releasing the mutex, if
|
71 |
* necessary.
|
72 |
*
|
73 |
* MUTEX_CAS(ptr, old, new)
|
74 |
* Perform an atomic "compare and swap" operation and
|
75 |
* evaluate to true or false according to the success
|
76 |
*
|
77 |
* Otherwise, the following must be defined:
|
78 |
*
|
79 |
* MUTEX_INITIALIZE_SPIN(mtx, dodebug, minipl)
|
80 |
* Initialize a spin mutex.
|
81 |
*
|
82 |
* MUTEX_INITIALIZE_ADAPTIVE(mtx, dodebug)
|
83 |
* Initialize an adaptive mutex.
|
84 |
*
|
85 |
* MUTEX_DESTROY(mtx)
|
86 |
* Tear down a mutex.
|
87 |
*
|
88 |
* MUTEX_ADAPTIVE_P(mtx)
|
89 |
* Evaluates to true if the mutex is an adaptive mutex.
|
90 |
*
|
91 |
* MUTEX_SPIN_P(mtx)
|
92 |
* Evaluates to true if the mutex is a spin mutex.
|
93 |
*
|
94 |
* MUTEX_OWNER(owner)
|
95 |
* Returns the owner of the adaptive mutex (LWP address).
|
96 |
*
|
97 |
* MUTEX_OWNED(owner)
|
98 |
* Returns non-zero if an adaptive mutex is currently
|
99 |
* held by an LWP.
|
100 |
*
|
101 |
* MUTEX_HAS_WAITERS(mtx)
|
102 |
* Returns true if the mutex has waiters.
|
103 |
*
|
104 |
* MUTEX_SET_WAITERS(mtx)
|
105 |
* Mark the mutex has having waiters.
|
106 |
*
|
107 |
* MUTEX_ACQUIRE(mtx, owner)
|
108 |
* Try to acquire an adaptive mutex such that:
|
109 |
* if (lock held OR waiters)
|
110 |
* return 0;
|
111 |
* else
|
112 |
* return 1;
|
113 |
* Must be MP/interrupt atomic.
|
114 |
*
|
115 |
* MUTEX_RELEASE(mtx)
|
116 |
* Release the lock and clear the "has waiters" indication.
|
117 |
* Must be interrupt atomic, need not be MP safe.
|
118 |
*
|
119 |
* MUTEX_DEBUG_P(mtx)
|
120 |
* Evaluates to true if the mutex is initialized with
|
121 |
* dodebug==true. Only used in the LOCKDEBUG case.
|
122 |
*
|
123 |
* Machine dependent code may optionally provide stubs for the following
|
124 |
* functions to implement the easy (unlocked / no waiters) cases. If
|
125 |
* these stubs are provided, __HAVE_MUTEX_STUBS should be defined.
|
126 |
*
|
127 |
* mutex_enter()
|
128 |
* mutex_exit()
|
129 |
*
|
130 |
* Two additional stubs may be implemented that handle only the spinlock
|
131 |
* case, primarily for the scheduler. __HAVE_SPIN_MUTEX_STUBS should be
|
132 |
* defined if these are provided:
|
133 |
*
|
134 |
* mutex_spin_enter()
|
135 |
* mutex_spin_exit()
|
136 |
*/
|
137 |
|
138 |
#if defined(_KERNEL_OPT)
|
139 |
#include "opt_lockdebug.h" |
140 |
#endif
|
141 |
|
142 |
#if !defined(_KERNEL)
|
143 |
#include <sys/types.h> |
144 |
#include <sys/inttypes.h> |
145 |
#endif
|
146 |
|
147 |
typedef enum kmutex_type_t { |
148 |
MUTEX_SPIN = 0, /* To get a spin mutex at IPL_NONE */ |
149 |
MUTEX_ADAPTIVE = 1, /* For porting code written for Solaris */ |
150 |
MUTEX_DEFAULT = 2, /* The only native, endorsed type */ |
151 |
MUTEX_DRIVER = 3, /* For porting code written for Solaris */ |
152 |
MUTEX_NODEBUG = 4 /* Disables LOCKDEBUG; use with care */ |
153 |
} kmutex_type_t; |
154 |
|
155 |
typedef struct kmutex kmutex_t; |
156 |
|
157 |
#if defined(__MUTEX_PRIVATE)
|
158 |
|
159 |
#define MUTEX_THREAD ((uintptr_t)-16L) |
160 |
|
161 |
#define MUTEX_BIT_SPIN 0x01 |
162 |
#define MUTEX_BIT_WAITERS 0x02 |
163 |
|
164 |
#if defined(LOCKDEBUG)
|
165 |
#define MUTEX_BIT_NODEBUG 0x04 /* LOCKDEBUG disabled */ |
166 |
#else
|
167 |
#define MUTEX_BIT_NODEBUG 0x00 /* do nothing */ |
168 |
#endif /* LOCKDEBUG */ |
169 |
|
170 |
#define MUTEX_SPIN_IPL(mtx) ((mtx)->mtx_ipl)
|
171 |
#define MUTEX_SPIN_OLDSPL(ci) ((ci)->ci_mtx_oldspl)
|
172 |
|
173 |
void mutex_vector_enter(kmutex_t *);
|
174 |
void mutex_vector_exit(kmutex_t *);
|
175 |
void mutex_spin_retry(kmutex_t *);
|
176 |
void mutex_wakeup(kmutex_t *);
|
177 |
|
178 |
#endif /* __MUTEX_PRIVATE */ |
179 |
|
180 |
#ifdef _KERNEL
|
181 |
#include <sys/intr.h> |
182 |
#endif
|
183 |
|
184 |
#include <machine/mutex.h> |
185 |
|
186 |
/*
|
187 |
* Return true if no spin mutexes are held by the current CPU.
|
188 |
*/
|
189 |
#ifndef MUTEX_NO_SPIN_ACTIVE_P
|
190 |
#define MUTEX_NO_SPIN_ACTIVE_P(ci) ((ci)->ci_mtx_count == 0) |
191 |
#endif
|
192 |
|
193 |
#ifdef _KERNEL
|
194 |
|
195 |
void mutex_init(kmutex_t *, kmutex_type_t, int); |
196 |
void mutex_destroy(kmutex_t *);
|
197 |
|
198 |
void mutex_enter(kmutex_t *);
|
199 |
void mutex_exit(kmutex_t *);
|
200 |
|
201 |
void mutex_spin_enter(kmutex_t *);
|
202 |
void mutex_spin_exit(kmutex_t *);
|
203 |
|
204 |
int mutex_tryenter(kmutex_t *);
|
205 |
|
206 |
int mutex_owned(kmutex_t *);
|
207 |
lwp_t *mutex_owner(kmutex_t *); |
208 |
|
209 |
void mutex_obj_init(void); |
210 |
kmutex_t *mutex_obj_alloc(kmutex_type_t, int);
|
211 |
void mutex_obj_hold(kmutex_t *);
|
212 |
bool mutex_obj_free(kmutex_t *);
|
213 |
|
214 |
#endif /* _KERNEL */ |
215 |
|
216 |
#endif /* _SYS_MUTEX_H_ */ |