Subversion Repositories Kolibri OS

Rev

Go to most recent revision | Details | Last modification | View Log | RSS feed

Rev Author Line No. Line
1901 serge 1
/*
2
 * Mesa 3-D graphics library
3
 * Version:  6.1
4
 *
5
 * Copyright (C) 1999-2004  Brian Paul   All Rights Reserved.
6
 *
7
 * Permission is hereby granted, free of charge, to any person obtaining a
8
 * copy of this software and associated documentation files (the "Software"),
9
 * to deal in the Software without restriction, including without limitation
10
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
11
 * and/or sell copies of the Software, and to permit persons to whom the
12
 * Software is furnished to do so, subject to the following conditions:
13
 *
14
 * The above copyright notice and this permission notice shall be included
15
 * in all copies or substantial portions of the Software.
16
 *
17
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
18
 * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
20
 * BRIAN PAUL BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN
21
 * AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
22
 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
23
 *
24
 * Authors:
25
 *    Gareth Hughes
26
 */
27
 
28
#ifndef __M_DEBUG_UTIL_H__
29
#define __M_DEBUG_UTIL_H__
30
 
31
 
32
#ifdef DEBUG_MATH  /* This code only used for debugging */
33
 
34
 
35
/* Comment this out to deactivate the cycle counter.
36
 * NOTE: it works only on CPUs which know the 'rdtsc' command (586 or higher)
37
 * (hope, you don't try to debug Mesa on a 386 ;)
38
 */
39
#if defined(__GNUC__) && \
40
    ((defined(__i386__) && defined(USE_X86_ASM)) || \
41
     (defined(__sparc__) && defined(USE_SPARC_ASM)))
42
#define  RUN_DEBUG_BENCHMARK
43
#endif
44
 
45
#define TEST_COUNT		128	/* size of the tested vector array   */
46
 
47
#define REQUIRED_PRECISION	10	/* allow 4 bits to miss              */
48
#define MAX_PRECISION		24	/* max. precision possible           */
49
 
50
 
51
#ifdef  RUN_DEBUG_BENCHMARK
52
/* Overhead of profiling counter in cycles.  Automatically adjusted to
53
 * your machine at run time - counter initialization should give very
54
 * consistent results.
55
 */
56
extern long counter_overhead;
57
 
58
/* This is the value of the environment variable MESA_PROFILE, and is
59
 * used to determine if we should benchmark the functions as well as
60
 * verify their correctness.
61
 */
62
extern char *mesa_profile;
63
 
64
/* Modify the number of tests if you like.
65
 * We take the minimum of all results, because every error should be
66
 * positive (time used by other processes, task switches etc).
67
 * It is assumed that all calculations are done in the cache.
68
 */
69
 
70
#if defined(__i386__)
71
 
72
#if 1 /* PPro, PII, PIII version */
73
 
74
/* Profiling on the P6 architecture requires a little more work, due to
75
 * the internal out-of-order execution.  We must perform a serializing
76
 * 'cpuid' instruction before and after the 'rdtsc' instructions to make
77
 * sure no other uops are executed when we sample the timestamp counter.
78
 */
79
#define  INIT_COUNTER()							\
80
   do {									\
81
      int cycle_i;							\
82
      counter_overhead = LONG_MAX;					\
83
      for ( cycle_i = 0 ; cycle_i < 8 ; cycle_i++ ) {			\
84
	 long cycle_tmp1 = 0, cycle_tmp2 = 0;				\
85
	 __asm__ __volatile__ ( "push %%ebx       \n"			\
86
				"xor %%eax, %%eax \n"			\
87
				"cpuid            \n"			\
88
				"rdtsc            \n"			\
89
				"mov %%eax, %0    \n"			\
90
				"xor %%eax, %%eax \n"			\
91
				"cpuid            \n"			\
92
				"pop %%ebx        \n"			\
93
				"push %%ebx       \n"			\
94
				"xor %%eax, %%eax \n"			\
95
				"cpuid            \n"			\
96
				"rdtsc            \n"			\
97
				"mov %%eax, %1    \n"			\
98
				"xor %%eax, %%eax \n"			\
99
				"cpuid            \n"			\
100
				"pop %%ebx        \n"			\
101
				: "=m" (cycle_tmp1), "=m" (cycle_tmp2)	\
102
				: : "eax", "ecx", "edx" );		\
103
	 if ( counter_overhead > (cycle_tmp2 - cycle_tmp1) ) {		\
104
	    counter_overhead = cycle_tmp2 - cycle_tmp1;			\
105
	 }								\
106
      }									\
107
   } while (0)
108
 
109
#define  BEGIN_RACE(x)							\
110
   x = LONG_MAX;							\
111
   for ( cycle_i = 0 ; cycle_i < 10 ; cycle_i++ ) {			\
112
      long cycle_tmp1 = 0, cycle_tmp2 = 0;				\
113
      __asm__ __volatile__ ( "push %%ebx       \n"			\
114
			     "xor %%eax, %%eax \n"			\
115
			     "cpuid            \n"			\
116
			     "rdtsc            \n"			\
117
			     "mov %%eax, %0    \n"			\
118
			     "xor %%eax, %%eax \n"			\
119
			     "cpuid            \n"			\
120
			     "pop %%ebx        \n"			\
121
			     : "=m" (cycle_tmp1)			\
122
			     : : "eax", "ecx", "edx" );
123
 
124
#define END_RACE(x)							\
125
      __asm__ __volatile__ ( "push %%ebx       \n"			\
126
			     "xor %%eax, %%eax \n"			\
127
			     "cpuid            \n"			\
128
			     "rdtsc            \n"			\
129
			     "mov %%eax, %0    \n"			\
130
			     "xor %%eax, %%eax \n"			\
131
			     "cpuid            \n"			\
132
			     "pop %%ebx        \n"			\
133
			     : "=m" (cycle_tmp2)			\
134
			     : : "eax", "ecx", "edx" );			\
135
      if ( x > (cycle_tmp2 - cycle_tmp1) ) {				\
136
	 x = cycle_tmp2 - cycle_tmp1;					\
137
      }									\
138
   }									\
139
   x -= counter_overhead;
140
 
141
#else /* PPlain, PMMX version */
142
 
143
/* To ensure accurate results, we stall the pipelines with the
144
 * non-pairable 'cdq' instruction.  This ensures all the code being
145
 * profiled is complete when the 'rdtsc' instruction executes.
146
 */
147
#define  INIT_COUNTER(x)						\
148
   do {									\
149
      int cycle_i;							\
150
      x = LONG_MAX;							\
151
      for ( cycle_i = 0 ; cycle_i < 32 ; cycle_i++ ) {			\
152
	 long cycle_tmp1, cycle_tmp2, dummy;				\
153
	 __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp1) );		\
154
	 __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp2) );		\
155
	 __asm__ ( "cdq" );						\
156
	 __asm__ ( "cdq" );						\
157
	 __asm__ ( "rdtsc" : "=a" (cycle_tmp1), "=d" (dummy) );		\
158
	 __asm__ ( "cdq" );						\
159
	 __asm__ ( "cdq" );						\
160
	 __asm__ ( "rdtsc" : "=a" (cycle_tmp2), "=d" (dummy) );		\
161
	 if ( x > (cycle_tmp2 - cycle_tmp1) )				\
162
	    x = cycle_tmp2 - cycle_tmp1;				\
163
      }									\
164
   } while (0)
165
 
166
#define  BEGIN_RACE(x)							\
167
   x = LONG_MAX;							\
168
   for ( cycle_i = 0 ; cycle_i < 16 ; cycle_i++ ) {			\
169
      long cycle_tmp1, cycle_tmp2, dummy;				\
170
      __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp1) );			\
171
      __asm__ ( "mov %%eax, %0" : "=a" (cycle_tmp2) );			\
172
      __asm__ ( "cdq" );						\
173
      __asm__ ( "cdq" );						\
174
      __asm__ ( "rdtsc" : "=a" (cycle_tmp1), "=d" (dummy) );
175
 
176
 
177
#define END_RACE(x)							\
178
      __asm__ ( "cdq" );						\
179
      __asm__ ( "cdq" );						\
180
      __asm__ ( "rdtsc" : "=a" (cycle_tmp2), "=d" (dummy) );		\
181
      if ( x > (cycle_tmp2 - cycle_tmp1) )				\
182
	 x = cycle_tmp2 - cycle_tmp1;					\
183
   }									\
184
   x -= counter_overhead;
185
 
186
#endif
187
 
188
#elif defined(__x86_64__)
189
 
190
#define rdtscll(val) do { \
191
     unsigned int a,d; \
192
     __asm__ volatile("rdtsc" : "=a" (a), "=d" (d)); \
193
     (val) = ((unsigned long)a) | (((unsigned long)d)<<32); \
194
} while(0)
195
 
196
/* Copied from i386 PIII version */
197
#define  INIT_COUNTER()							\
198
   do {									\
199
      int cycle_i;							\
200
      counter_overhead = LONG_MAX;					\
201
      for ( cycle_i = 0 ; cycle_i < 16 ; cycle_i++ ) {			\
202
	 unsigned long cycle_tmp1, cycle_tmp2;        			\
203
	 rdtscll(cycle_tmp1);						\
204
	 rdtscll(cycle_tmp2);						\
205
	 if ( counter_overhead > (cycle_tmp2 - cycle_tmp1) ) {		\
206
	    counter_overhead = cycle_tmp2 - cycle_tmp1;			\
207
	 }								\
208
      }									\
209
   } while (0)
210
 
211
 
212
#define  BEGIN_RACE(x)							\
213
   x = LONG_MAX;							\
214
   for ( cycle_i = 0 ; cycle_i < 10 ; cycle_i++ ) {			\
215
      unsigned long cycle_tmp1, cycle_tmp2;				\
216
      rdtscll(cycle_tmp1);						\
217
 
218
#define END_RACE(x)							\
219
      rdtscll(cycle_tmp2);						\
220
      if ( x > (cycle_tmp2 - cycle_tmp1) ) {				\
221
	 x = cycle_tmp2 - cycle_tmp1;					\
222
      }									\
223
   }									\
224
   x -= counter_overhead;
225
 
226
#elif defined(__sparc__)
227
 
228
#define  INIT_COUNTER()	\
229
	 do { counter_overhead = 5; } while(0)
230
 
231
#define  BEGIN_RACE(x)                                                        \
232
x = LONG_MAX;                                                                 \
233
for (cycle_i = 0; cycle_i <10; cycle_i++) {                                   \
234
   register long cycle_tmp1 __asm__("l0");				      \
235
   register long cycle_tmp2 __asm__("l1");				      \
236
   /* rd %tick, %l0 */							      \
237
   __asm__ __volatile__ (".word 0xa1410000" : "=r" (cycle_tmp1));  /*  save timestamp   */
238
 
239
#define END_RACE(x)                                                           \
240
   /* rd %tick, %l1 */							      \
241
   __asm__ __volatile__ (".word 0xa3410000" : "=r" (cycle_tmp2));	      \
242
   if (x > (cycle_tmp2-cycle_tmp1)) x = cycle_tmp2 - cycle_tmp1;              \
243
}                                                                             \
244
x -= counter_overhead;
245
 
246
#else
247
#error Your processor is not supported for RUN_XFORM_BENCHMARK
248
#endif
249
 
250
#else
251
 
252
#define BEGIN_RACE(x)
253
#define END_RACE(x)
254
 
255
#endif
256
 
257
 
258
/* =============================================================
259
 * Helper functions
260
 */
261
 
262
static GLfloat rnd( void )
263
{
264
   GLfloat f = (GLfloat)rand() / (GLfloat)RAND_MAX;
265
   GLfloat gran = (GLfloat)(1 << 13);
266
 
267
   f = (GLfloat)(GLint)(f * gran) / gran;
268
 
269
   return f * 2.0 - 1.0;
270
}
271
 
272
static int significand_match( GLfloat a, GLfloat b )
273
{
274
   GLfloat d = a - b;
275
   int a_ex, b_ex, d_ex;
276
 
277
   if ( d == 0.0F ) {
278
      return MAX_PRECISION;   /* Exact match */
279
   }
280
 
281
   if ( a == 0.0F || b == 0.0F ) {
282
      /* It would probably be better to check if the
283
       * non-zero number is denormalized and return
284
       * the index of the highest set bit here.
285
       */
286
      return 0;
287
   }
288
 
289
   FREXPF( a, &a_ex );
290
   FREXPF( b, &b_ex );
291
   FREXPF( d, &d_ex );
292
 
293
   if ( a_ex < b_ex ) {
294
      return a_ex - d_ex;
295
   } else {
296
      return b_ex - d_ex;
297
   }
298
}
299
 
300
enum { NIL = 0, ONE = 1, NEG = -1, VAR = 2 };
301
 
302
/* Ensure our arrays are correctly aligned.
303
 */
304
#if defined(__GNUC__)
305
#  define ALIGN16(type, array)	type array __attribute__ ((aligned (16)))
306
#elif defined(_MSC_VER)
307
#  define ALIGN16(type, array)	type array __declspec(align(16)) /* GH: Does this work? */
308
#elif defined(__WATCOMC__)
309
#  define ALIGN16(type, array)	                    /* Watcom does not support this */
310
#elif defined(__xlC__)
311
#  define ALIGN16(type, array)       type __align (16) array
312
#else
313
#  warning "ALIGN16 will not 16-byte align!\n"
314
#  define ALIGN16
315
#endif
316
 
317
 
318
#endif /* DEBUG_MATH */
319
 
320
#endif /* __M_DEBUG_UTIL_H__ */