Rev 1498 | Go to most recent revision | Details | Compare with Previous | Last modification | View Log | RSS feed
Rev | Author | Line No. | Line |
---|---|---|---|
1498 | serge | 1 | /****************************************************************************** |
2 | * |
||
3 | * Name: acmacros.h - C macros for the entire subsystem. |
||
4 | * |
||
5 | *****************************************************************************/ |
||
6 | |||
7 | /****************************************************************************** |
||
8 | * |
||
9 | * 1. Copyright Notice |
||
10 | * |
||
2216 | Serge | 11 | * Some or all of this work - Copyright (c) 1999 - 2011, Intel Corp. |
1498 | serge | 12 | * All rights reserved. |
13 | * |
||
14 | * 2. License |
||
15 | * |
||
16 | * 2.1. This is your license from Intel Corp. under its intellectual property |
||
17 | * rights. You may have additional license terms from the party that provided |
||
18 | * you this software, covering your right to use that party's intellectual |
||
19 | * property rights. |
||
20 | * |
||
21 | * 2.2. Intel grants, free of charge, to any person ("Licensee") obtaining a |
||
22 | * copy of the source code appearing in this file ("Covered Code") an |
||
23 | * irrevocable, perpetual, worldwide license under Intel's copyrights in the |
||
24 | * base code distributed originally by Intel ("Original Intel Code") to copy, |
||
25 | * make derivatives, distribute, use and display any portion of the Covered |
||
26 | * Code in any form, with the right to sublicense such rights; and |
||
27 | * |
||
28 | * 2.3. Intel grants Licensee a non-exclusive and non-transferable patent |
||
29 | * license (with the right to sublicense), under only those claims of Intel |
||
30 | * patents that are infringed by the Original Intel Code, to make, use, sell, |
||
31 | * offer to sell, and import the Covered Code and derivative works thereof |
||
32 | * solely to the minimum extent necessary to exercise the above copyright |
||
33 | * license, and in no event shall the patent license extend to any additions |
||
34 | * to or modifications of the Original Intel Code. No other license or right |
||
35 | * is granted directly or by implication, estoppel or otherwise; |
||
36 | * |
||
37 | * The above copyright and patent license is granted only if the following |
||
38 | * conditions are met: |
||
39 | * |
||
40 | * 3. Conditions |
||
41 | * |
||
42 | * 3.1. Redistribution of Source with Rights to Further Distribute Source. |
||
43 | * Redistribution of source code of any substantial portion of the Covered |
||
44 | * Code or modification with rights to further distribute source must include |
||
45 | * the above Copyright Notice, the above License, this list of Conditions, |
||
46 | * and the following Disclaimer and Export Compliance provision. In addition, |
||
47 | * Licensee must cause all Covered Code to which Licensee contributes to |
||
48 | * contain a file documenting the changes Licensee made to create that Covered |
||
49 | * Code and the date of any change. Licensee must include in that file the |
||
50 | * documentation of any changes made by any predecessor Licensee. Licensee |
||
51 | * must include a prominent statement that the modification is derived, |
||
52 | * directly or indirectly, from Original Intel Code. |
||
53 | * |
||
54 | * 3.2. Redistribution of Source with no Rights to Further Distribute Source. |
||
55 | * Redistribution of source code of any substantial portion of the Covered |
||
56 | * Code or modification without rights to further distribute source must |
||
57 | * include the following Disclaimer and Export Compliance provision in the |
||
58 | * documentation and/or other materials provided with distribution. In |
||
59 | * addition, Licensee may not authorize further sublicense of source of any |
||
60 | * portion of the Covered Code, and must include terms to the effect that the |
||
61 | * license from Licensee to its licensee is limited to the intellectual |
||
62 | * property embodied in the software Licensee provides to its licensee, and |
||
63 | * not to intellectual property embodied in modifications its licensee may |
||
64 | * make. |
||
65 | * |
||
66 | * 3.3. Redistribution of Executable. Redistribution in executable form of any |
||
67 | * substantial portion of the Covered Code or modification must reproduce the |
||
68 | * above Copyright Notice, and the following Disclaimer and Export Compliance |
||
69 | * provision in the documentation and/or other materials provided with the |
||
70 | * distribution. |
||
71 | * |
||
72 | * 3.4. Intel retains all right, title, and interest in and to the Original |
||
73 | * Intel Code. |
||
74 | * |
||
75 | * 3.5. Neither the name Intel nor any other trademark owned or controlled by |
||
76 | * Intel shall be used in advertising or otherwise to promote the sale, use or |
||
77 | * other dealings in products derived from or relating to the Covered Code |
||
78 | * without prior written authorization from Intel. |
||
79 | * |
||
80 | * 4. Disclaimer and Export Compliance |
||
81 | * |
||
82 | * 4.1. INTEL MAKES NO WARRANTY OF ANY KIND REGARDING ANY SOFTWARE PROVIDED |
||
83 | * HERE. ANY SOFTWARE ORIGINATING FROM INTEL OR DERIVED FROM INTEL SOFTWARE |
||
84 | * IS PROVIDED "AS IS," AND INTEL WILL NOT PROVIDE ANY SUPPORT, ASSISTANCE, |
||
85 | * INSTALLATION, TRAINING OR OTHER SERVICES. INTEL WILL NOT PROVIDE ANY |
||
86 | * UPDATES, ENHANCEMENTS OR EXTENSIONS. INTEL SPECIFICALLY DISCLAIMS ANY |
||
87 | * IMPLIED WARRANTIES OF MERCHANTABILITY, NONINFRINGEMENT AND FITNESS FOR A |
||
88 | * PARTICULAR PURPOSE. |
||
89 | * |
||
90 | * 4.2. IN NO EVENT SHALL INTEL HAVE ANY LIABILITY TO LICENSEE, ITS LICENSEES |
||
91 | * OR ANY OTHER THIRD PARTY, FOR ANY LOST PROFITS, LOST DATA, LOSS OF USE OR |
||
92 | * COSTS OF PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES, OR FOR ANY INDIRECT, |
||
93 | * SPECIAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THIS AGREEMENT, UNDER ANY |
||
94 | * CAUSE OF ACTION OR THEORY OF LIABILITY, AND IRRESPECTIVE OF WHETHER INTEL |
||
95 | * HAS ADVANCE NOTICE OF THE POSSIBILITY OF SUCH DAMAGES. THESE LIMITATIONS |
||
96 | * SHALL APPLY NOTWITHSTANDING THE FAILURE OF THE ESSENTIAL PURPOSE OF ANY |
||
97 | * LIMITED REMEDY. |
||
98 | * |
||
99 | * 4.3. Licensee shall not export, either directly or indirectly, any of this |
||
100 | * software or system incorporating such software without first obtaining any |
||
101 | * required license or other approval from the U. S. Department of Commerce or |
||
102 | * any other agency or department of the United States Government. In the |
||
103 | * event Licensee exports any such software from the United States or |
||
104 | * re-exports any such software from a foreign destination, Licensee shall |
||
105 | * ensure that the distribution and export/re-export of the software is in |
||
106 | * compliance with all laws, regulations, orders, or other restrictions of the |
||
107 | * U.S. Export Administration Regulations. Licensee agrees that neither it nor |
||
108 | * any of its subsidiaries will export/re-export any technical data, process, |
||
109 | * software, or service, directly or indirectly, to any country for which the |
||
110 | * United States government or any agency thereof requires an export license, |
||
111 | * other governmental approval, or letter of assurance, without first obtaining |
||
112 | * such license, approval or letter. |
||
113 | * |
||
114 | *****************************************************************************/ |
||
115 | |||
116 | #ifndef __ACMACROS_H__ |
||
117 | #define __ACMACROS_H__ |
||
118 | |||
119 | |||
120 | /* |
||
121 | * Extract data using a pointer. Any more than a byte and we |
||
122 | * get into potential aligment issues -- see the STORE macros below. |
||
123 | * Use with care. |
||
124 | */ |
||
125 | #define ACPI_GET8(ptr) *ACPI_CAST_PTR (UINT8, ptr) |
||
126 | #define ACPI_GET16(ptr) *ACPI_CAST_PTR (UINT16, ptr) |
||
127 | #define ACPI_GET32(ptr) *ACPI_CAST_PTR (UINT32, ptr) |
||
128 | #define ACPI_GET64(ptr) *ACPI_CAST_PTR (UINT64, ptr) |
||
129 | #define ACPI_SET8(ptr) *ACPI_CAST_PTR (UINT8, ptr) |
||
130 | #define ACPI_SET16(ptr) *ACPI_CAST_PTR (UINT16, ptr) |
||
131 | #define ACPI_SET32(ptr) *ACPI_CAST_PTR (UINT32, ptr) |
||
132 | #define ACPI_SET64(ptr) *ACPI_CAST_PTR (UINT64, ptr) |
||
133 | |||
134 | /* |
||
135 | * printf() format helpers |
||
136 | */ |
||
137 | |||
138 | /* Split 64-bit integer into two 32-bit values. Use with %8.8X%8.8X */ |
||
139 | |||
140 | #define ACPI_FORMAT_UINT64(i) ACPI_HIDWORD(i), ACPI_LODWORD(i) |
||
141 | |||
142 | #if ACPI_MACHINE_WIDTH == 64 |
||
143 | #define ACPI_FORMAT_NATIVE_UINT(i) ACPI_FORMAT_UINT64(i) |
||
144 | #else |
||
145 | #define ACPI_FORMAT_NATIVE_UINT(i) 0, (i) |
||
146 | #endif |
||
147 | |||
148 | |||
149 | /* |
||
150 | * Macros for moving data around to/from buffers that are possibly unaligned. |
||
151 | * If the hardware supports the transfer of unaligned data, just do the store. |
||
152 | * Otherwise, we have to move one byte at a time. |
||
153 | */ |
||
154 | #ifdef ACPI_BIG_ENDIAN |
||
155 | /* |
||
156 | * Macros for big-endian machines |
||
157 | */ |
||
158 | |||
159 | /* These macros reverse the bytes during the move, converting little-endian to big endian */ |
||
160 | |||
161 | /* Big Endian <== Little Endian */ |
||
162 | /* Hi...Lo Lo...Hi */ |
||
163 | /* 16-bit source, 16/32/64 destination */ |
||
164 | |||
165 | #define ACPI_MOVE_16_TO_16(d, s) {(( UINT8 *)(void *)(d))[0] = ((UINT8 *)(void *)(s))[1];\ |
||
166 | (( UINT8 *)(void *)(d))[1] = ((UINT8 *)(void *)(s))[0];} |
||
167 | |||
168 | #define ACPI_MOVE_16_TO_32(d, s) {(*(UINT32 *)(void *)(d))=0;\ |
||
169 | ((UINT8 *)(void *)(d))[2] = ((UINT8 *)(void *)(s))[1];\ |
||
170 | ((UINT8 *)(void *)(d))[3] = ((UINT8 *)(void *)(s))[0];} |
||
171 | |||
172 | #define ACPI_MOVE_16_TO_64(d, s) {(*(UINT64 *)(void *)(d))=0;\ |
||
173 | ((UINT8 *)(void *)(d))[6] = ((UINT8 *)(void *)(s))[1];\ |
||
174 | ((UINT8 *)(void *)(d))[7] = ((UINT8 *)(void *)(s))[0];} |
||
175 | |||
176 | /* 32-bit source, 16/32/64 destination */ |
||
177 | |||
178 | #define ACPI_MOVE_32_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */ |
||
179 | |||
180 | #define ACPI_MOVE_32_TO_32(d, s) {(( UINT8 *)(void *)(d))[0] = ((UINT8 *)(void *)(s))[3];\ |
||
181 | (( UINT8 *)(void *)(d))[1] = ((UINT8 *)(void *)(s))[2];\ |
||
182 | (( UINT8 *)(void *)(d))[2] = ((UINT8 *)(void *)(s))[1];\ |
||
183 | (( UINT8 *)(void *)(d))[3] = ((UINT8 *)(void *)(s))[0];} |
||
184 | |||
185 | #define ACPI_MOVE_32_TO_64(d, s) {(*(UINT64 *)(void *)(d))=0;\ |
||
186 | ((UINT8 *)(void *)(d))[4] = ((UINT8 *)(void *)(s))[3];\ |
||
187 | ((UINT8 *)(void *)(d))[5] = ((UINT8 *)(void *)(s))[2];\ |
||
188 | ((UINT8 *)(void *)(d))[6] = ((UINT8 *)(void *)(s))[1];\ |
||
189 | ((UINT8 *)(void *)(d))[7] = ((UINT8 *)(void *)(s))[0];} |
||
190 | |||
191 | /* 64-bit source, 16/32/64 destination */ |
||
192 | |||
193 | #define ACPI_MOVE_64_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */ |
||
194 | |||
195 | #define ACPI_MOVE_64_TO_32(d, s) ACPI_MOVE_32_TO_32(d, s) /* Truncate to 32 */ |
||
196 | |||
197 | #define ACPI_MOVE_64_TO_64(d, s) {(( UINT8 *)(void *)(d))[0] = ((UINT8 *)(void *)(s))[7];\ |
||
198 | (( UINT8 *)(void *)(d))[1] = ((UINT8 *)(void *)(s))[6];\ |
||
199 | (( UINT8 *)(void *)(d))[2] = ((UINT8 *)(void *)(s))[5];\ |
||
200 | (( UINT8 *)(void *)(d))[3] = ((UINT8 *)(void *)(s))[4];\ |
||
201 | (( UINT8 *)(void *)(d))[4] = ((UINT8 *)(void *)(s))[3];\ |
||
202 | (( UINT8 *)(void *)(d))[5] = ((UINT8 *)(void *)(s))[2];\ |
||
203 | (( UINT8 *)(void *)(d))[6] = ((UINT8 *)(void *)(s))[1];\ |
||
204 | (( UINT8 *)(void *)(d))[7] = ((UINT8 *)(void *)(s))[0];} |
||
205 | #else |
||
206 | /* |
||
207 | * Macros for little-endian machines |
||
208 | */ |
||
209 | |||
210 | #ifndef ACPI_MISALIGNMENT_NOT_SUPPORTED |
||
211 | |||
212 | /* The hardware supports unaligned transfers, just do the little-endian move */ |
||
213 | |||
214 | /* 16-bit source, 16/32/64 destination */ |
||
215 | |||
216 | #define ACPI_MOVE_16_TO_16(d, s) *(UINT16 *)(void *)(d) = *(UINT16 *)(void *)(s) |
||
217 | #define ACPI_MOVE_16_TO_32(d, s) *(UINT32 *)(void *)(d) = *(UINT16 *)(void *)(s) |
||
218 | #define ACPI_MOVE_16_TO_64(d, s) *(UINT64 *)(void *)(d) = *(UINT16 *)(void *)(s) |
||
219 | |||
220 | /* 32-bit source, 16/32/64 destination */ |
||
221 | |||
222 | #define ACPI_MOVE_32_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */ |
||
223 | #define ACPI_MOVE_32_TO_32(d, s) *(UINT32 *)(void *)(d) = *(UINT32 *)(void *)(s) |
||
224 | #define ACPI_MOVE_32_TO_64(d, s) *(UINT64 *)(void *)(d) = *(UINT32 *)(void *)(s) |
||
225 | |||
226 | /* 64-bit source, 16/32/64 destination */ |
||
227 | |||
228 | #define ACPI_MOVE_64_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */ |
||
229 | #define ACPI_MOVE_64_TO_32(d, s) ACPI_MOVE_32_TO_32(d, s) /* Truncate to 32 */ |
||
230 | #define ACPI_MOVE_64_TO_64(d, s) *(UINT64 *)(void *)(d) = *(UINT64 *)(void *)(s) |
||
231 | |||
232 | #else |
||
233 | /* |
||
234 | * The hardware does not support unaligned transfers. We must move the |
||
235 | * data one byte at a time. These macros work whether the source or |
||
236 | * the destination (or both) is/are unaligned. (Little-endian move) |
||
237 | */ |
||
238 | |||
239 | /* 16-bit source, 16/32/64 destination */ |
||
240 | |||
241 | #define ACPI_MOVE_16_TO_16(d, s) {(( UINT8 *)(void *)(d))[0] = ((UINT8 *)(void *)(s))[0];\ |
||
242 | (( UINT8 *)(void *)(d))[1] = ((UINT8 *)(void *)(s))[1];} |
||
243 | |||
244 | #define ACPI_MOVE_16_TO_32(d, s) {(*(UINT32 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d, s);} |
||
245 | #define ACPI_MOVE_16_TO_64(d, s) {(*(UINT64 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d, s);} |
||
246 | |||
247 | /* 32-bit source, 16/32/64 destination */ |
||
248 | |||
249 | #define ACPI_MOVE_32_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */ |
||
250 | |||
251 | #define ACPI_MOVE_32_TO_32(d, s) {(( UINT8 *)(void *)(d))[0] = ((UINT8 *)(void *)(s))[0];\ |
||
252 | (( UINT8 *)(void *)(d))[1] = ((UINT8 *)(void *)(s))[1];\ |
||
253 | (( UINT8 *)(void *)(d))[2] = ((UINT8 *)(void *)(s))[2];\ |
||
254 | (( UINT8 *)(void *)(d))[3] = ((UINT8 *)(void *)(s))[3];} |
||
255 | |||
256 | #define ACPI_MOVE_32_TO_64(d, s) {(*(UINT64 *)(void *)(d)) = 0; ACPI_MOVE_32_TO_32(d, s);} |
||
257 | |||
258 | /* 64-bit source, 16/32/64 destination */ |
||
259 | |||
260 | #define ACPI_MOVE_64_TO_16(d, s) ACPI_MOVE_16_TO_16(d, s) /* Truncate to 16 */ |
||
261 | #define ACPI_MOVE_64_TO_32(d, s) ACPI_MOVE_32_TO_32(d, s) /* Truncate to 32 */ |
||
262 | #define ACPI_MOVE_64_TO_64(d, s) {(( UINT8 *)(void *)(d))[0] = ((UINT8 *)(void *)(s))[0];\ |
||
263 | (( UINT8 *)(void *)(d))[1] = ((UINT8 *)(void *)(s))[1];\ |
||
264 | (( UINT8 *)(void *)(d))[2] = ((UINT8 *)(void *)(s))[2];\ |
||
265 | (( UINT8 *)(void *)(d))[3] = ((UINT8 *)(void *)(s))[3];\ |
||
266 | (( UINT8 *)(void *)(d))[4] = ((UINT8 *)(void *)(s))[4];\ |
||
267 | (( UINT8 *)(void *)(d))[5] = ((UINT8 *)(void *)(s))[5];\ |
||
268 | (( UINT8 *)(void *)(d))[6] = ((UINT8 *)(void *)(s))[6];\ |
||
269 | (( UINT8 *)(void *)(d))[7] = ((UINT8 *)(void *)(s))[7];} |
||
270 | #endif |
||
271 | #endif |
||
272 | |||
273 | |||
274 | /* |
||
275 | * Fast power-of-two math macros for non-optimized compilers |
||
276 | */ |
||
277 | #define _ACPI_DIV(value, PowerOf2) ((UINT32) ((value) >> (PowerOf2))) |
||
278 | #define _ACPI_MUL(value, PowerOf2) ((UINT32) ((value) << (PowerOf2))) |
||
279 | #define _ACPI_MOD(value, Divisor) ((UINT32) ((value) & ((Divisor) -1))) |
||
280 | |||
281 | #define ACPI_DIV_2(a) _ACPI_DIV(a, 1) |
||
282 | #define ACPI_MUL_2(a) _ACPI_MUL(a, 1) |
||
283 | #define ACPI_MOD_2(a) _ACPI_MOD(a, 2) |
||
284 | |||
285 | #define ACPI_DIV_4(a) _ACPI_DIV(a, 2) |
||
286 | #define ACPI_MUL_4(a) _ACPI_MUL(a, 2) |
||
287 | #define ACPI_MOD_4(a) _ACPI_MOD(a, 4) |
||
288 | |||
289 | #define ACPI_DIV_8(a) _ACPI_DIV(a, 3) |
||
290 | #define ACPI_MUL_8(a) _ACPI_MUL(a, 3) |
||
291 | #define ACPI_MOD_8(a) _ACPI_MOD(a, 8) |
||
292 | |||
293 | #define ACPI_DIV_16(a) _ACPI_DIV(a, 4) |
||
294 | #define ACPI_MUL_16(a) _ACPI_MUL(a, 4) |
||
295 | #define ACPI_MOD_16(a) _ACPI_MOD(a, 16) |
||
296 | |||
297 | #define ACPI_DIV_32(a) _ACPI_DIV(a, 5) |
||
298 | #define ACPI_MUL_32(a) _ACPI_MUL(a, 5) |
||
299 | #define ACPI_MOD_32(a) _ACPI_MOD(a, 32) |
||
300 | |||
301 | /* |
||
302 | * Rounding macros (Power of two boundaries only) |
||
303 | */ |
||
304 | #define ACPI_ROUND_DOWN(value, boundary) (((ACPI_SIZE)(value)) & \ |
||
305 | (~(((ACPI_SIZE) boundary)-1))) |
||
306 | |||
307 | #define ACPI_ROUND_UP(value, boundary) ((((ACPI_SIZE)(value)) + \ |
||
308 | (((ACPI_SIZE) boundary)-1)) & \ |
||
309 | (~(((ACPI_SIZE) boundary)-1))) |
||
310 | |||
311 | /* Note: sizeof(ACPI_SIZE) evaluates to either 4 or 8 (32- vs 64-bit mode) */ |
||
312 | |||
313 | #define ACPI_ROUND_DOWN_TO_32BIT(a) ACPI_ROUND_DOWN(a, 4) |
||
314 | #define ACPI_ROUND_DOWN_TO_64BIT(a) ACPI_ROUND_DOWN(a, 8) |
||
315 | #define ACPI_ROUND_DOWN_TO_NATIVE_WORD(a) ACPI_ROUND_DOWN(a, sizeof(ACPI_SIZE)) |
||
316 | |||
317 | #define ACPI_ROUND_UP_TO_32BIT(a) ACPI_ROUND_UP(a, 4) |
||
318 | #define ACPI_ROUND_UP_TO_64BIT(a) ACPI_ROUND_UP(a, 8) |
||
319 | #define ACPI_ROUND_UP_TO_NATIVE_WORD(a) ACPI_ROUND_UP(a, sizeof(ACPI_SIZE)) |
||
320 | |||
321 | #define ACPI_ROUND_BITS_UP_TO_BYTES(a) ACPI_DIV_8((a) + 7) |
||
322 | #define ACPI_ROUND_BITS_DOWN_TO_BYTES(a) ACPI_DIV_8((a)) |
||
323 | |||
324 | #define ACPI_ROUND_UP_TO_1K(a) (((a) + 1023) >> 10) |
||
325 | |||
326 | /* Generic (non-power-of-two) rounding */ |
||
327 | |||
328 | #define ACPI_ROUND_UP_TO(value, boundary) (((value) + ((boundary)-1)) / (boundary)) |
||
329 | |||
330 | #define ACPI_IS_MISALIGNED(value) (((ACPI_SIZE) value) & (sizeof(ACPI_SIZE)-1)) |
||
331 | |||
332 | /* |
||
333 | * Bitmask creation |
||
334 | * Bit positions start at zero. |
||
335 | * MASK_BITS_ABOVE creates a mask starting AT the position and above |
||
336 | * MASK_BITS_BELOW creates a mask starting one bit BELOW the position |
||
337 | */ |
||
338 | #define ACPI_MASK_BITS_ABOVE(position) (~((ACPI_UINT64_MAX) << ((UINT32) (position)))) |
||
339 | #define ACPI_MASK_BITS_BELOW(position) ((ACPI_UINT64_MAX) << ((UINT32) (position))) |
||
340 | |||
341 | /* Bitfields within ACPI registers */ |
||
342 | |||
343 | #define ACPI_REGISTER_PREPARE_BITS(Val, Pos, Mask) ((Val << Pos) & Mask) |
||
344 | #define ACPI_REGISTER_INSERT_VALUE(Reg, Pos, Mask, Val) Reg = (Reg & (~(Mask))) | ACPI_REGISTER_PREPARE_BITS(Val, Pos, Mask) |
||
345 | |||
346 | #define ACPI_INSERT_BITS(Target, Mask, Source) Target = ((Target & (~(Mask))) | (Source & Mask)) |
||
347 | |||
348 | /* |
||
349 | * An ACPI_NAMESPACE_NODE can appear in some contexts |
||
350 | * where a pointer to an ACPI_OPERAND_OBJECT can also |
||
351 | * appear. This macro is used to distinguish them. |
||
352 | * |
||
353 | * The "Descriptor" field is the first field in both structures. |
||
354 | */ |
||
355 | #define ACPI_GET_DESCRIPTOR_TYPE(d) (((ACPI_DESCRIPTOR *)(void *)(d))->Common.DescriptorType) |
||
356 | #define ACPI_SET_DESCRIPTOR_TYPE(d, t) (((ACPI_DESCRIPTOR *)(void *)(d))->Common.DescriptorType = t) |
||
357 | |||
358 | /* |
||
359 | * Macros for the master AML opcode table |
||
360 | */ |
||
361 | #if defined (ACPI_DISASSEMBLER) || defined (ACPI_DEBUG_OUTPUT) |
||
362 | #define ACPI_OP(Name, PArgs, IArgs, ObjType, Class, Type, Flags) \ |
||
363 | {Name, (UINT32)(PArgs), (UINT32)(IArgs), (UINT32)(Flags), ObjType, Class, Type} |
||
364 | #else |
||
365 | #define ACPI_OP(Name, PArgs, IArgs, ObjType, Class, Type, Flags) \ |
||
366 | {(UINT32)(PArgs), (UINT32)(IArgs), (UINT32)(Flags), ObjType, Class, Type} |
||
367 | #endif |
||
368 | |||
369 | #define ARG_TYPE_WIDTH 5 |
||
370 | #define ARG_1(x) ((UINT32)(x)) |
||
371 | #define ARG_2(x) ((UINT32)(x) << (1 * ARG_TYPE_WIDTH)) |
||
372 | #define ARG_3(x) ((UINT32)(x) << (2 * ARG_TYPE_WIDTH)) |
||
373 | #define ARG_4(x) ((UINT32)(x) << (3 * ARG_TYPE_WIDTH)) |
||
374 | #define ARG_5(x) ((UINT32)(x) << (4 * ARG_TYPE_WIDTH)) |
||
375 | #define ARG_6(x) ((UINT32)(x) << (5 * ARG_TYPE_WIDTH)) |
||
376 | |||
377 | #define ARGI_LIST1(a) (ARG_1(a)) |
||
378 | #define ARGI_LIST2(a, b) (ARG_1(b)|ARG_2(a)) |
||
379 | #define ARGI_LIST3(a, b, c) (ARG_1(c)|ARG_2(b)|ARG_3(a)) |
||
380 | #define ARGI_LIST4(a, b, c, d) (ARG_1(d)|ARG_2(c)|ARG_3(b)|ARG_4(a)) |
||
381 | #define ARGI_LIST5(a, b, c, d, e) (ARG_1(e)|ARG_2(d)|ARG_3(c)|ARG_4(b)|ARG_5(a)) |
||
382 | #define ARGI_LIST6(a, b, c, d, e, f) (ARG_1(f)|ARG_2(e)|ARG_3(d)|ARG_4(c)|ARG_5(b)|ARG_6(a)) |
||
383 | |||
384 | #define ARGP_LIST1(a) (ARG_1(a)) |
||
385 | #define ARGP_LIST2(a, b) (ARG_1(a)|ARG_2(b)) |
||
386 | #define ARGP_LIST3(a, b, c) (ARG_1(a)|ARG_2(b)|ARG_3(c)) |
||
387 | #define ARGP_LIST4(a, b, c, d) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)) |
||
388 | #define ARGP_LIST5(a, b, c, d, e) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e)) |
||
389 | #define ARGP_LIST6(a, b, c, d, e, f) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e)|ARG_6(f)) |
||
390 | |||
391 | #define GET_CURRENT_ARG_TYPE(List) (List & ((UINT32) 0x1F)) |
||
392 | #define INCREMENT_ARG_LIST(List) (List >>= ((UINT32) ARG_TYPE_WIDTH)) |
||
393 | |||
394 | /* |
||
395 | * Ascii error messages can be configured out |
||
396 | */ |
||
397 | #ifndef ACPI_NO_ERROR_MESSAGES |
||
398 | /* |
||
399 | * Error reporting. Callers module and line number are inserted by AE_INFO, |
||
400 | * the plist contains a set of parens to allow variable-length lists. |
||
401 | * These macros are used for both the debug and non-debug versions of the code. |
||
402 | */ |
||
2216 | Serge | 403 | #define ACPI_ERROR_NAMESPACE(s, e) AcpiUtNamespaceError (AE_INFO, s, e); |
404 | #define ACPI_ERROR_METHOD(s, n, p, e) AcpiUtMethodError (AE_INFO, s, n, p, e); |
||
1498 | serge | 405 | #define ACPI_WARN_PREDEFINED(plist) AcpiUtPredefinedWarning plist |
406 | #define ACPI_INFO_PREDEFINED(plist) AcpiUtPredefinedInfo plist |
||
407 | |||
408 | #else |
||
409 | |||
410 | /* No error messages */ |
||
411 | |||
412 | #define ACPI_ERROR_NAMESPACE(s, e) |
||
413 | #define ACPI_ERROR_METHOD(s, n, p, e) |
||
414 | #define ACPI_WARN_PREDEFINED(plist) |
||
415 | #define ACPI_INFO_PREDEFINED(plist) |
||
416 | |||
417 | #endif /* ACPI_NO_ERROR_MESSAGES */ |
||
418 | |||
419 | /* |
||
420 | * Debug macros that are conditionally compiled |
||
421 | */ |
||
422 | #ifdef ACPI_DEBUG_OUTPUT |
||
423 | /* |
||
424 | * Function entry tracing |
||
425 | */ |
||
426 | #define ACPI_FUNCTION_TRACE(a) ACPI_FUNCTION_NAME(a) \ |
||
427 | AcpiUtTrace(ACPI_DEBUG_PARAMETERS) |
||
428 | #define ACPI_FUNCTION_TRACE_PTR(a, b) ACPI_FUNCTION_NAME(a) \ |
||
429 | AcpiUtTracePtr(ACPI_DEBUG_PARAMETERS, (void *)b) |
||
430 | #define ACPI_FUNCTION_TRACE_U32(a, b) ACPI_FUNCTION_NAME(a) \ |
||
431 | AcpiUtTraceU32(ACPI_DEBUG_PARAMETERS, (UINT32)b) |
||
432 | #define ACPI_FUNCTION_TRACE_STR(a, b) ACPI_FUNCTION_NAME(a) \ |
||
433 | AcpiUtTraceStr(ACPI_DEBUG_PARAMETERS, (char *)b) |
||
434 | |||
435 | #define ACPI_FUNCTION_ENTRY() AcpiUtTrackStackPtr() |
||
436 | |||
437 | /* |
||
438 | * Function exit tracing. |
||
439 | * WARNING: These macros include a return statement. This is usually considered |
||
440 | * bad form, but having a separate exit macro is very ugly and difficult to maintain. |
||
441 | * One of the FUNCTION_TRACE macros above must be used in conjunction with these macros |
||
442 | * so that "_AcpiFunctionName" is defined. |
||
443 | * |
||
444 | * Note: the DO_WHILE0 macro is used to prevent some compilers from complaining |
||
445 | * about these constructs. |
||
446 | */ |
||
447 | #ifdef ACPI_USE_DO_WHILE_0 |
||
448 | #define ACPI_DO_WHILE0(a) do a while(0) |
||
449 | #else |
||
450 | #define ACPI_DO_WHILE0(a) a |
||
451 | #endif |
||
452 | |||
453 | #define return_VOID ACPI_DO_WHILE0 ({ \ |
||
454 | AcpiUtExit (ACPI_DEBUG_PARAMETERS); \ |
||
455 | return;}) |
||
456 | /* |
||
457 | * There are two versions of most of the return macros. The default version is |
||
458 | * safer, since it avoids side-effects by guaranteeing that the argument will |
||
459 | * not be evaluated twice. |
||
460 | * |
||
461 | * A less-safe version of the macros is provided for optional use if the |
||
462 | * compiler uses excessive CPU stack (for example, this may happen in the |
||
463 | * debug case if code optimzation is disabled.) |
||
464 | */ |
||
465 | #ifndef ACPI_SIMPLE_RETURN_MACROS |
||
466 | |||
467 | #define return_ACPI_STATUS(s) ACPI_DO_WHILE0 ({ \ |
||
468 | register ACPI_STATUS _s = (s); \ |
||
469 | AcpiUtStatusExit (ACPI_DEBUG_PARAMETERS, _s); \ |
||
470 | return (_s); }) |
||
471 | #define return_PTR(s) ACPI_DO_WHILE0 ({ \ |
||
472 | register void *_s = (void *) (s); \ |
||
473 | AcpiUtPtrExit (ACPI_DEBUG_PARAMETERS, (UINT8 *) _s); \ |
||
474 | return (_s); }) |
||
475 | #define return_VALUE(s) ACPI_DO_WHILE0 ({ \ |
||
476 | register UINT64 _s = (s); \ |
||
477 | AcpiUtValueExit (ACPI_DEBUG_PARAMETERS, _s); \ |
||
478 | return (_s); }) |
||
479 | #define return_UINT8(s) ACPI_DO_WHILE0 ({ \ |
||
480 | register UINT8 _s = (UINT8) (s); \ |
||
481 | AcpiUtValueExit (ACPI_DEBUG_PARAMETERS, (UINT64) _s); \ |
||
482 | return (_s); }) |
||
483 | #define return_UINT32(s) ACPI_DO_WHILE0 ({ \ |
||
484 | register UINT32 _s = (UINT32) (s); \ |
||
485 | AcpiUtValueExit (ACPI_DEBUG_PARAMETERS, (UINT64) _s); \ |
||
486 | return (_s); }) |
||
487 | #else /* Use original less-safe macros */ |
||
488 | |||
489 | #define return_ACPI_STATUS(s) ACPI_DO_WHILE0 ({ \ |
||
490 | AcpiUtStatusExit (ACPI_DEBUG_PARAMETERS, (s)); \ |
||
491 | return((s)); }) |
||
492 | #define return_PTR(s) ACPI_DO_WHILE0 ({ \ |
||
493 | AcpiUtPtrExit (ACPI_DEBUG_PARAMETERS, (UINT8 *) (s)); \ |
||
494 | return((s)); }) |
||
495 | #define return_VALUE(s) ACPI_DO_WHILE0 ({ \ |
||
496 | AcpiUtValueExit (ACPI_DEBUG_PARAMETERS, (UINT64) (s)); \ |
||
497 | return((s)); }) |
||
498 | #define return_UINT8(s) return_VALUE(s) |
||
499 | #define return_UINT32(s) return_VALUE(s) |
||
500 | |||
501 | #endif /* ACPI_SIMPLE_RETURN_MACROS */ |
||
502 | |||
503 | |||
504 | /* Conditional execution */ |
||
505 | |||
506 | #define ACPI_DEBUG_EXEC(a) a |
||
507 | #define ACPI_DEBUG_ONLY_MEMBERS(a) a; |
||
508 | #define _VERBOSE_STRUCTURES |
||
509 | |||
510 | |||
511 | /* Various object display routines for debug */ |
||
512 | |||
513 | #define ACPI_DUMP_STACK_ENTRY(a) AcpiExDumpOperand((a), 0) |
||
514 | #define ACPI_DUMP_OPERANDS(a, b ,c) AcpiExDumpOperands(a, b, c) |
||
515 | #define ACPI_DUMP_ENTRY(a, b) AcpiNsDumpEntry (a, b) |
||
516 | #define ACPI_DUMP_PATHNAME(a, b, c, d) AcpiNsDumpPathname(a, b, c, d) |
||
517 | #define ACPI_DUMP_BUFFER(a, b) AcpiUtDumpBuffer((UINT8 *) a, b, DB_BYTE_DISPLAY, _COMPONENT) |
||
518 | |||
519 | #else |
||
520 | /* |
||
521 | * This is the non-debug case -- make everything go away, |
||
522 | * leaving no executable debug code! |
||
523 | */ |
||
524 | #define ACPI_DEBUG_EXEC(a) |
||
525 | #define ACPI_DEBUG_ONLY_MEMBERS(a) |
||
526 | #define ACPI_FUNCTION_TRACE(a) |
||
527 | #define ACPI_FUNCTION_TRACE_PTR(a, b) |
||
528 | #define ACPI_FUNCTION_TRACE_U32(a, b) |
||
529 | #define ACPI_FUNCTION_TRACE_STR(a, b) |
||
530 | #define ACPI_FUNCTION_EXIT |
||
531 | #define ACPI_FUNCTION_STATUS_EXIT(s) |
||
532 | #define ACPI_FUNCTION_VALUE_EXIT(s) |
||
533 | #define ACPI_FUNCTION_ENTRY() |
||
534 | #define ACPI_DUMP_STACK_ENTRY(a) |
||
535 | #define ACPI_DUMP_OPERANDS(a, b, c) |
||
536 | #define ACPI_DUMP_ENTRY(a, b) |
||
537 | #define ACPI_DUMP_TABLES(a, b) |
||
538 | #define ACPI_DUMP_PATHNAME(a, b, c, d) |
||
539 | #define ACPI_DUMP_BUFFER(a, b) |
||
540 | #define ACPI_DEBUG_PRINT(pl) |
||
541 | #define ACPI_DEBUG_PRINT_RAW(pl) |
||
542 | |||
543 | #define return_VOID return |
||
544 | #define return_ACPI_STATUS(s) return(s) |
||
545 | #define return_VALUE(s) return(s) |
||
546 | #define return_UINT8(s) return(s) |
||
547 | #define return_UINT32(s) return(s) |
||
548 | #define return_PTR(s) return(s) |
||
549 | |||
550 | #endif /* ACPI_DEBUG_OUTPUT */ |
||
551 | |||
552 | /* |
||
553 | * Some code only gets executed when the debugger is built in. |
||
554 | * Note that this is entirely independent of whether the |
||
555 | * DEBUG_PRINT stuff (set by ACPI_DEBUG_OUTPUT) is on, or not. |
||
556 | */ |
||
557 | #ifdef ACPI_DEBUGGER |
||
558 | #define ACPI_DEBUGGER_EXEC(a) a |
||
559 | #else |
||
560 | #define ACPI_DEBUGGER_EXEC(a) |
||
561 | #endif |
||
562 | |||
563 | |||
564 | /* |
||
565 | * Memory allocation tracking (DEBUG ONLY) |
||
566 | */ |
||
567 | #define ACPI_MEM_PARAMETERS _COMPONENT, _AcpiModuleName, __LINE__ |
||
568 | |||
569 | #ifndef ACPI_DBG_TRACK_ALLOCATIONS |
||
570 | |||
571 | /* Memory allocation */ |
||
572 | |||
573 | #define ACPI_ALLOCATE(a) AcpiUtAllocate((ACPI_SIZE) (a), ACPI_MEM_PARAMETERS) |
||
574 | #define ACPI_ALLOCATE_ZEROED(a) AcpiUtAllocateZeroed((ACPI_SIZE) (a), ACPI_MEM_PARAMETERS) |
||
575 | #define ACPI_FREE(a) AcpiOsFree(a) |
||
576 | #define ACPI_MEM_TRACKING(a) |
||
577 | |||
578 | #else |
||
579 | |||
580 | /* Memory allocation */ |
||
581 | |||
582 | #define ACPI_ALLOCATE(a) AcpiUtAllocateAndTrack((ACPI_SIZE) (a), ACPI_MEM_PARAMETERS) |
||
583 | #define ACPI_ALLOCATE_ZEROED(a) AcpiUtAllocateZeroedAndTrack((ACPI_SIZE) (a), ACPI_MEM_PARAMETERS) |
||
584 | #define ACPI_FREE(a) AcpiUtFreeAndTrack(a, ACPI_MEM_PARAMETERS) |
||
585 | #define ACPI_MEM_TRACKING(a) a |
||
586 | |||
587 | #endif /* ACPI_DBG_TRACK_ALLOCATIONS */ |
||
588 | |||
589 | |||
590 | /* |
||
591 | * Macros used for ACPICA utilities only |
||
592 | */ |
||
593 | |||
594 | /* Generate a UUID */ |
||
595 | |||
596 | #define ACPI_INIT_UUID(a, b, c, d0, d1, d2, d3, d4, d5, d6, d7) \ |
||
597 | (a) & 0xFF, ((a) >> 8) & 0xFF, ((a) >> 16) & 0xFF, ((a) >> 24) & 0xFF, \ |
||
598 | (b) & 0xFF, ((b) >> 8) & 0xFF, \ |
||
599 | (c) & 0xFF, ((c) >> 8) & 0xFF, \ |
||
600 | (d0), (d1), (d2), (d3), (d4), (d5), (d6), (d7) |
||
601 | |||
602 | #define ACPI_IS_OCTAL_DIGIT(d) (((char)(d) >= '0') && ((char)(d) <= '7')) |
||
603 | |||
604 | |||
605 | #endif /* ACMACROS_H */=>><>><>><>><>><>><>><>><>><>==> |