Rev 6336 | Only display areas with differences | Regard whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 6336 | Rev 6936 | ||
---|---|---|---|
1 | /* |
1 | /* |
2 | * Read-Copy Update mechanism for mutual exclusion, the Bloatwatch edition. |
2 | * Read-Copy Update mechanism for mutual exclusion, the Bloatwatch edition. |
3 | * |
3 | * |
4 | * This program is free software; you can redistribute it and/or modify |
4 | * This program is free software; you can redistribute it and/or modify |
5 | * it under the terms of the GNU General Public License as published by |
5 | * it under the terms of the GNU General Public License as published by |
6 | * the Free Software Foundation; either version 2 of the License, or |
6 | * the Free Software Foundation; either version 2 of the License, or |
7 | * (at your option) any later version. |
7 | * (at your option) any later version. |
8 | * |
8 | * |
9 | * This program is distributed in the hope that it will be useful, |
9 | * This program is distributed in the hope that it will be useful, |
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
12 | * GNU General Public License for more details. |
12 | * GNU General Public License for more details. |
13 | * |
13 | * |
14 | * You should have received a copy of the GNU General Public License |
14 | * You should have received a copy of the GNU General Public License |
15 | * along with this program; if not, you can access it online at |
15 | * along with this program; if not, you can access it online at |
16 | * http://www.gnu.org/licenses/gpl-2.0.html. |
16 | * http://www.gnu.org/licenses/gpl-2.0.html. |
17 | * |
17 | * |
18 | * Copyright IBM Corporation, 2008 |
18 | * Copyright IBM Corporation, 2008 |
19 | * |
19 | * |
20 | * Author: Paul E. McKenney |
20 | * Author: Paul E. McKenney |
21 | * |
21 | * |
22 | * For detailed explanation of Read-Copy Update mechanism see - |
22 | * For detailed explanation of Read-Copy Update mechanism see - |
23 | * Documentation/RCU |
23 | * Documentation/RCU |
24 | */ |
24 | */ |
25 | #ifndef __LINUX_TINY_H |
25 | #ifndef __LINUX_TINY_H |
26 | #define __LINUX_TINY_H |
26 | #define __LINUX_TINY_H |
27 | 27 | ||
28 | #include |
28 | #include |
29 | 29 | ||
30 | static inline unsigned long get_state_synchronize_rcu(void) |
30 | static inline unsigned long get_state_synchronize_rcu(void) |
31 | { |
31 | { |
32 | return 0; |
32 | return 0; |
33 | } |
33 | } |
34 | 34 | ||
35 | static inline void cond_synchronize_rcu(unsigned long oldstate) |
35 | static inline void cond_synchronize_rcu(unsigned long oldstate) |
36 | { |
36 | { |
37 | might_sleep(); |
37 | might_sleep(); |
38 | } |
38 | } |
39 | 39 | ||
40 | static inline unsigned long get_state_synchronize_sched(void) |
40 | static inline unsigned long get_state_synchronize_sched(void) |
41 | { |
41 | { |
42 | return 0; |
42 | return 0; |
43 | } |
43 | } |
44 | static inline void rcu_barrier_bh(void) |
44 | static inline void rcu_barrier_bh(void) |
45 | { |
45 | { |
46 | wait_rcu_gp(call_rcu_bh); |
46 | wait_rcu_gp(call_rcu_bh); |
47 | } |
47 | } |
48 | 48 | ||
49 | static inline void rcu_barrier_sched(void) |
49 | static inline void rcu_barrier_sched(void) |
50 | { |
50 | { |
51 | wait_rcu_gp(call_rcu_sched); |
51 | wait_rcu_gp(call_rcu_sched); |
52 | } |
52 | } |
53 | 53 | ||
54 | static inline void synchronize_rcu_expedited(void) |
54 | static inline void synchronize_rcu_expedited(void) |
55 | { |
55 | { |
56 | synchronize_sched(); /* Only one CPU, so pretty fast anyway!!! */ |
56 | synchronize_sched(); /* Only one CPU, so pretty fast anyway!!! */ |
57 | } |
57 | } |
58 | 58 | ||
59 | static inline void rcu_barrier(void) |
59 | static inline void rcu_barrier(void) |
60 | { |
60 | { |
61 | rcu_barrier_sched(); /* Only one CPU, so only one list of callbacks! */ |
61 | rcu_barrier_sched(); /* Only one CPU, so only one list of callbacks! */ |
62 | } |
62 | } |
63 | 63 | ||
64 | static inline void synchronize_rcu_bh(void) |
64 | static inline void synchronize_rcu_bh(void) |
65 | { |
65 | { |
66 | synchronize_sched(); |
66 | synchronize_sched(); |
67 | } |
67 | } |
68 | 68 | ||
69 | static inline void synchronize_rcu_bh_expedited(void) |
69 | static inline void synchronize_rcu_bh_expedited(void) |
70 | { |
70 | { |
71 | synchronize_sched(); |
71 | synchronize_sched(); |
72 | } |
72 | } |
73 | 73 | ||
74 | static inline void synchronize_sched_expedited(void) |
74 | static inline void synchronize_sched_expedited(void) |
75 | { |
75 | { |
76 | synchronize_sched(); |
76 | synchronize_sched(); |
77 | } |
77 | } |
78 | 78 | ||
79 | static inline void kfree_call_rcu(struct rcu_head *head, |
79 | static inline void kfree_call_rcu(struct rcu_head *head, |
80 | void (*func)(struct rcu_head *rcu)) |
80 | void (*func)(struct rcu_head *rcu)) |
81 | { |
81 | { |
82 | call_rcu(head, func); |
82 | call_rcu(head, func); |
83 | } |
83 | } |
84 | 84 | ||
85 | static inline void rcu_note_context_switch(void) |
85 | static inline void rcu_note_context_switch(void) |
86 | { |
86 | { |
87 | rcu_sched_qs(); |
87 | rcu_sched_qs(); |
88 | } |
88 | } |
89 | 89 | ||
90 | /* |
90 | /* |
91 | * Take advantage of the fact that there is only one CPU, which |
91 | * Take advantage of the fact that there is only one CPU, which |
92 | * allows us to ignore virtualization-based context switches. |
92 | * allows us to ignore virtualization-based context switches. |
93 | */ |
93 | */ |
94 | static inline void rcu_virt_note_context_switch(int cpu) |
94 | static inline void rcu_virt_note_context_switch(int cpu) |
95 | { |
95 | { |
96 | } |
96 | } |
97 | 97 | ||
98 | /* |
98 | /* |
99 | * Return the number of grace periods started. |
99 | * Return the number of grace periods started. |
100 | */ |
100 | */ |
101 | static inline unsigned long rcu_batches_started(void) |
101 | static inline unsigned long rcu_batches_started(void) |
102 | { |
102 | { |
103 | return 0; |
103 | return 0; |
104 | } |
104 | } |
105 | 105 | ||
106 | /* |
106 | /* |
107 | * Return the number of bottom-half grace periods started. |
107 | * Return the number of bottom-half grace periods started. |
108 | */ |
108 | */ |
109 | static inline unsigned long rcu_batches_started_bh(void) |
109 | static inline unsigned long rcu_batches_started_bh(void) |
110 | { |
110 | { |
111 | return 0; |
111 | return 0; |
112 | } |
112 | } |
113 | 113 | ||
114 | /* |
114 | /* |
115 | * Return the number of sched grace periods started. |
115 | * Return the number of sched grace periods started. |
116 | */ |
116 | */ |
117 | static inline unsigned long rcu_batches_started_sched(void) |
117 | static inline unsigned long rcu_batches_started_sched(void) |
118 | { |
118 | { |
119 | return 0; |
119 | return 0; |
120 | } |
120 | } |
121 | 121 | ||
122 | /* |
122 | /* |
123 | * Return the number of grace periods completed. |
123 | * Return the number of grace periods completed. |
124 | */ |
124 | */ |
125 | static inline unsigned long rcu_batches_completed(void) |
125 | static inline unsigned long rcu_batches_completed(void) |
126 | { |
126 | { |
127 | return 0; |
127 | return 0; |
128 | } |
128 | } |
129 | 129 | ||
130 | /* |
130 | /* |
131 | * Return the number of bottom-half grace periods completed. |
131 | * Return the number of bottom-half grace periods completed. |
132 | */ |
132 | */ |
133 | static inline unsigned long rcu_batches_completed_bh(void) |
133 | static inline unsigned long rcu_batches_completed_bh(void) |
134 | { |
134 | { |
135 | return 0; |
135 | return 0; |
136 | } |
136 | } |
137 | 137 | ||
138 | /* |
138 | /* |
139 | * Return the number of sched grace periods completed. |
139 | * Return the number of sched grace periods completed. |
140 | */ |
140 | */ |
141 | static inline unsigned long rcu_batches_completed_sched(void) |
141 | static inline unsigned long rcu_batches_completed_sched(void) |
142 | { |
142 | { |
143 | return 0; |
143 | return 0; |
144 | } |
144 | } |
145 | 145 | ||
146 | static inline void rcu_force_quiescent_state(void) |
146 | static inline void rcu_force_quiescent_state(void) |
147 | { |
147 | { |
148 | } |
148 | } |
149 | 149 | ||
150 | static inline void rcu_bh_force_quiescent_state(void) |
150 | static inline void rcu_bh_force_quiescent_state(void) |
151 | { |
151 | { |
152 | } |
152 | } |
153 | 153 | ||
154 | static inline void rcu_sched_force_quiescent_state(void) |
154 | static inline void rcu_sched_force_quiescent_state(void) |
155 | { |
155 | { |
156 | } |
156 | } |
157 | 157 | ||
158 | static inline void show_rcu_gp_kthreads(void) |
158 | static inline void show_rcu_gp_kthreads(void) |
159 | { |
159 | { |
160 | } |
160 | } |
161 | 161 | ||
162 | static inline void rcu_cpu_stall_reset(void) |
162 | static inline void rcu_cpu_stall_reset(void) |
163 | { |
163 | { |
164 | } |
164 | } |
165 | 165 | ||
166 | static inline void rcu_idle_enter(void) |
166 | static inline void rcu_idle_enter(void) |
167 | { |
167 | { |
168 | } |
168 | } |
169 | 169 | ||
170 | static inline void rcu_idle_exit(void) |
170 | static inline void rcu_idle_exit(void) |
171 | { |
171 | { |
172 | } |
172 | } |
173 | 173 | ||
174 | static inline void rcu_irq_enter(void) |
174 | static inline void rcu_irq_enter(void) |
175 | { |
175 | { |
176 | } |
176 | } |
- | 177 | ||
- | 178 | static inline void rcu_irq_exit_irqson(void) |
|
- | 179 | { |
|
- | 180 | } |
|
- | 181 | ||
- | 182 | static inline void rcu_irq_enter_irqson(void) |
|
- | 183 | { |
|
- | 184 | } |
|
177 | 185 | ||
178 | static inline void rcu_irq_exit(void) |
186 | static inline void rcu_irq_exit(void) |
179 | { |
187 | { |
180 | } |
188 | } |
181 | 189 | ||
182 | static inline void exit_rcu(void) |
190 | static inline void exit_rcu(void) |
183 | { |
191 | { |
184 | } |
192 | } |
185 | 193 | ||
186 | #ifdef CONFIG_DEBUG_LOCK_ALLOC |
194 | #ifdef CONFIG_DEBUG_LOCK_ALLOC |
187 | extern int rcu_scheduler_active __read_mostly; |
195 | extern int rcu_scheduler_active __read_mostly; |
188 | void rcu_scheduler_starting(void); |
196 | void rcu_scheduler_starting(void); |
189 | #else /* #ifdef CONFIG_DEBUG_LOCK_ALLOC */ |
197 | #else /* #ifdef CONFIG_DEBUG_LOCK_ALLOC */ |
190 | static inline void rcu_scheduler_starting(void) |
198 | static inline void rcu_scheduler_starting(void) |
191 | { |
199 | { |
192 | } |
200 | } |
193 | #endif /* #else #ifdef CONFIG_DEBUG_LOCK_ALLOC */ |
201 | #endif /* #else #ifdef CONFIG_DEBUG_LOCK_ALLOC */ |
194 | 202 | ||
195 | #if defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) |
203 | #if defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) |
196 | 204 | ||
197 | static inline bool rcu_is_watching(void) |
205 | static inline bool rcu_is_watching(void) |
198 | { |
206 | { |
199 | return __rcu_is_watching(); |
207 | return __rcu_is_watching(); |
200 | } |
208 | } |
201 | 209 | ||
202 | #else /* defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) */ |
210 | #else /* defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) */ |
203 | 211 | ||
204 | static inline bool rcu_is_watching(void) |
212 | static inline bool rcu_is_watching(void) |
205 | { |
213 | { |
206 | return true; |
214 | return true; |
207 | } |
215 | } |
208 | 216 | ||
209 | #endif /* #else defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) */ |
217 | #endif /* #else defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) */ |
210 | 218 | ||
211 | static inline void rcu_all_qs(void) |
219 | static inline void rcu_all_qs(void) |
212 | { |
220 | { |
213 | } |
221 | } |
214 | 222 | ||
215 | #endif /* __LINUX_RCUTINY_H */ |
223 | #endif /* __LINUX_RCUTINY_H */ |