Rev 6336 | Details | Compare with Previous | Last modification | View Log | RSS feed
Rev | Author | Line No. | Line |
---|---|---|---|
5270 | serge | 1 | /* |
2 | * Read-Copy Update mechanism for mutual exclusion, the Bloatwatch edition. |
||
3 | * |
||
4 | * This program is free software; you can redistribute it and/or modify |
||
5 | * it under the terms of the GNU General Public License as published by |
||
6 | * the Free Software Foundation; either version 2 of the License, or |
||
7 | * (at your option) any later version. |
||
8 | * |
||
9 | * This program is distributed in the hope that it will be useful, |
||
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
||
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
||
12 | * GNU General Public License for more details. |
||
13 | * |
||
14 | * You should have received a copy of the GNU General Public License |
||
15 | * along with this program; if not, you can access it online at |
||
16 | * http://www.gnu.org/licenses/gpl-2.0.html. |
||
17 | * |
||
18 | * Copyright IBM Corporation, 2008 |
||
19 | * |
||
20 | * Author: Paul E. McKenney |
||
21 | * |
||
22 | * For detailed explanation of Read-Copy Update mechanism see - |
||
23 | * Documentation/RCU |
||
24 | */ |
||
25 | #ifndef __LINUX_TINY_H |
||
26 | #define __LINUX_TINY_H |
||
27 | |||
28 | #include |
||
29 | |||
30 | static inline unsigned long get_state_synchronize_rcu(void) |
||
31 | { |
||
32 | return 0; |
||
33 | } |
||
34 | |||
35 | static inline void cond_synchronize_rcu(unsigned long oldstate) |
||
36 | { |
||
37 | might_sleep(); |
||
38 | } |
||
39 | |||
6336 | serge | 40 | static inline unsigned long get_state_synchronize_sched(void) |
41 | { |
||
42 | return 0; |
||
43 | } |
||
5270 | serge | 44 | static inline void rcu_barrier_bh(void) |
45 | { |
||
46 | wait_rcu_gp(call_rcu_bh); |
||
47 | } |
||
48 | |||
49 | static inline void rcu_barrier_sched(void) |
||
50 | { |
||
51 | wait_rcu_gp(call_rcu_sched); |
||
52 | } |
||
53 | |||
54 | static inline void synchronize_rcu_expedited(void) |
||
55 | { |
||
56 | synchronize_sched(); /* Only one CPU, so pretty fast anyway!!! */ |
||
57 | } |
||
58 | |||
59 | static inline void rcu_barrier(void) |
||
60 | { |
||
61 | rcu_barrier_sched(); /* Only one CPU, so only one list of callbacks! */ |
||
62 | } |
||
63 | |||
64 | static inline void synchronize_rcu_bh(void) |
||
65 | { |
||
66 | synchronize_sched(); |
||
67 | } |
||
68 | |||
69 | static inline void synchronize_rcu_bh_expedited(void) |
||
70 | { |
||
71 | synchronize_sched(); |
||
72 | } |
||
73 | |||
74 | static inline void synchronize_sched_expedited(void) |
||
75 | { |
||
76 | synchronize_sched(); |
||
77 | } |
||
78 | |||
79 | static inline void kfree_call_rcu(struct rcu_head *head, |
||
80 | void (*func)(struct rcu_head *rcu)) |
||
81 | { |
||
82 | call_rcu(head, func); |
||
83 | } |
||
84 | |||
85 | static inline void rcu_note_context_switch(void) |
||
86 | { |
||
87 | rcu_sched_qs(); |
||
88 | } |
||
89 | |||
90 | /* |
||
91 | * Take advantage of the fact that there is only one CPU, which |
||
92 | * allows us to ignore virtualization-based context switches. |
||
93 | */ |
||
94 | static inline void rcu_virt_note_context_switch(int cpu) |
||
95 | { |
||
96 | } |
||
97 | |||
98 | /* |
||
6082 | serge | 99 | * Return the number of grace periods started. |
5270 | serge | 100 | */ |
6082 | serge | 101 | static inline unsigned long rcu_batches_started(void) |
5270 | serge | 102 | { |
103 | return 0; |
||
104 | } |
||
105 | |||
106 | /* |
||
6082 | serge | 107 | * Return the number of bottom-half grace periods started. |
5270 | serge | 108 | */ |
6082 | serge | 109 | static inline unsigned long rcu_batches_started_bh(void) |
5270 | serge | 110 | { |
111 | return 0; |
||
112 | } |
||
113 | |||
6082 | serge | 114 | /* |
115 | * Return the number of sched grace periods started. |
||
116 | */ |
||
117 | static inline unsigned long rcu_batches_started_sched(void) |
||
118 | { |
||
119 | return 0; |
||
120 | } |
||
121 | |||
122 | /* |
||
123 | * Return the number of grace periods completed. |
||
124 | */ |
||
125 | static inline unsigned long rcu_batches_completed(void) |
||
126 | { |
||
127 | return 0; |
||
128 | } |
||
129 | |||
130 | /* |
||
131 | * Return the number of bottom-half grace periods completed. |
||
132 | */ |
||
133 | static inline unsigned long rcu_batches_completed_bh(void) |
||
134 | { |
||
135 | return 0; |
||
136 | } |
||
137 | |||
138 | /* |
||
139 | * Return the number of sched grace periods completed. |
||
140 | */ |
||
141 | static inline unsigned long rcu_batches_completed_sched(void) |
||
142 | { |
||
143 | return 0; |
||
144 | } |
||
145 | |||
5270 | serge | 146 | static inline void rcu_force_quiescent_state(void) |
147 | { |
||
148 | } |
||
149 | |||
150 | static inline void rcu_bh_force_quiescent_state(void) |
||
151 | { |
||
152 | } |
||
153 | |||
154 | static inline void rcu_sched_force_quiescent_state(void) |
||
155 | { |
||
156 | } |
||
157 | |||
158 | static inline void show_rcu_gp_kthreads(void) |
||
159 | { |
||
160 | } |
||
161 | |||
162 | static inline void rcu_cpu_stall_reset(void) |
||
163 | { |
||
164 | } |
||
165 | |||
6082 | serge | 166 | static inline void rcu_idle_enter(void) |
167 | { |
||
168 | } |
||
169 | |||
170 | static inline void rcu_idle_exit(void) |
||
171 | { |
||
172 | } |
||
173 | |||
174 | static inline void rcu_irq_enter(void) |
||
175 | { |
||
176 | } |
||
177 | |||
6936 | serge | 178 | static inline void rcu_irq_exit_irqson(void) |
179 | { |
||
180 | } |
||
181 | |||
182 | static inline void rcu_irq_enter_irqson(void) |
||
183 | { |
||
184 | } |
||
185 | |||
6082 | serge | 186 | static inline void rcu_irq_exit(void) |
187 | { |
||
188 | } |
||
189 | |||
5270 | serge | 190 | static inline void exit_rcu(void) |
191 | { |
||
192 | } |
||
193 | |||
194 | #ifdef CONFIG_DEBUG_LOCK_ALLOC |
||
195 | extern int rcu_scheduler_active __read_mostly; |
||
196 | void rcu_scheduler_starting(void); |
||
197 | #else /* #ifdef CONFIG_DEBUG_LOCK_ALLOC */ |
||
198 | static inline void rcu_scheduler_starting(void) |
||
199 | { |
||
200 | } |
||
201 | #endif /* #else #ifdef CONFIG_DEBUG_LOCK_ALLOC */ |
||
202 | |||
203 | #if defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) |
||
204 | |||
205 | static inline bool rcu_is_watching(void) |
||
206 | { |
||
207 | return __rcu_is_watching(); |
||
208 | } |
||
209 | |||
210 | #else /* defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) */ |
||
211 | |||
212 | static inline bool rcu_is_watching(void) |
||
213 | { |
||
214 | return true; |
||
215 | } |
||
216 | |||
217 | #endif /* #else defined(CONFIG_DEBUG_LOCK_ALLOC) || defined(CONFIG_RCU_TRACE) */ |
||
218 | |||
6082 | serge | 219 | static inline void rcu_all_qs(void) |
220 | { |
||
221 | } |
||
222 | |||
5270 | serge | 223 | #endif /* __LINUX_RCUTINY_H */ |