Subversion Repositories Kolibri OS

Rev

Rev 1117 | Rev 1120 | Go to most recent revision | Show entire file | Regard whitespace | Details | Blame | Last modification | View Log | RSS feed

Rev 1117 Rev 1119
Line 45... Line 45...
45
 * Some of these functions might be used by newer ASICs.
45
 * Some of these functions might be used by newer ASICs.
46
 */
46
 */
47
void r520_gpu_init(struct radeon_device *rdev);
47
void r520_gpu_init(struct radeon_device *rdev);
48
int r520_mc_wait_for_idle(struct radeon_device *rdev);
48
int r520_mc_wait_for_idle(struct radeon_device *rdev);
Line 49... Line -...
49
 
-
 
50
#if 0
49
 
51
/*
50
/*
52
 * MC
51
 * MC
53
 */
52
 */
54
int r520_mc_init(struct radeon_device *rdev)
53
int r520_mc_init(struct radeon_device *rdev)
55
{
54
{
56
	uint32_t tmp;
55
	uint32_t tmp;
Line -... Line 56...
-
 
56
	int r;
-
 
57
 
57
	int r;
58
    dbgprintf("%s\n",__FUNCTION__);
58
 
59
 
59
	if (r100_debugfs_rbbm_init(rdev)) {
60
//   if (r100_debugfs_rbbm_init(rdev)) {
60
		DRM_ERROR("Failed to register debugfs file for RBBM !\n");
61
//       DRM_ERROR("Failed to register debugfs file for RBBM !\n");
61
	}
62
//   }
62
	if (rv515_debugfs_pipes_info_init(rdev)) {
63
//   if (rv515_debugfs_pipes_info_init(rdev)) {
63
		DRM_ERROR("Failed to register debugfs file for pipes !\n");
64
//       DRM_ERROR("Failed to register debugfs file for pipes !\n");
64
	}
65
//   }
65
	if (rv515_debugfs_ga_info_init(rdev)) {
66
//   if (rv515_debugfs_ga_info_init(rdev)) {
Line 66... Line 67...
66
		DRM_ERROR("Failed to register debugfs file for pipes !\n");
67
//       DRM_ERROR("Failed to register debugfs file for pipes !\n");
67
	}
68
//   }
Line 68... Line 69...
68
 
69
 
Line 111... Line 112...
111
	} else {
112
	} else {
112
		WREG32_MC(R520_MC_AGP_LOCATION, 0x0FFFFFFF);
113
		WREG32_MC(R520_MC_AGP_LOCATION, 0x0FFFFFFF);
113
		WREG32_MC(R520_MC_AGP_BASE, 0);
114
		WREG32_MC(R520_MC_AGP_BASE, 0);
114
		WREG32_MC(R520_MC_AGP_BASE_2, 0);
115
		WREG32_MC(R520_MC_AGP_BASE_2, 0);
115
	}
116
	}
-
 
117
 
-
 
118
    dbgprintf("done: %s\n",__FUNCTION__);
-
 
119
 
116
	return 0;
120
	return 0;
117
}
121
}
Line 118... Line 122...
118
 
122
 
119
void r520_mc_fini(struct radeon_device *rdev)
123
void r520_mc_fini(struct radeon_device *rdev)
120
{
124
{
121
	rv370_pcie_gart_disable(rdev);
125
	rv370_pcie_gart_disable(rdev);
122
	radeon_gart_table_vram_free(rdev);
126
	radeon_gart_table_vram_free(rdev);
123
	radeon_gart_fini(rdev);
127
	radeon_gart_fini(rdev);
Line 124... Line -...
124
}
-
 
Line 125... Line 128...
125
 
128
}
126
#endif
129
 
127
 
130
 
128
/*
131
/*
129
 * Global GPU functions
132
 * Global GPU functions
130
 */
133
 */
131
void r520_errata(struct radeon_device *rdev)
134
void r520_errata(struct radeon_device *rdev)
Line 132... Line -...
132
{
-
 
133
	rdev->pll_errata = 0;
135
{
134
}
136
	rdev->pll_errata = 0;
135
 
137
}
136
#if 0
138
 
Line 151... Line 153...
151
}
153
}
Line 152... Line 154...
152
 
154
 
153
void r520_gpu_init(struct radeon_device *rdev)
155
void r520_gpu_init(struct radeon_device *rdev)
154
{
156
{
-
 
157
	unsigned pipe_select_current, gb_pipe_select, tmp;
Line 155... Line 158...
155
	unsigned pipe_select_current, gb_pipe_select, tmp;
158
    dbgprintf("%s\n\r",__FUNCTION__);
156
 
159
 
157
	r100_hdp_reset(rdev);
160
	r100_hdp_reset(rdev);
158
	rs600_disable_vga(rdev);
161
	rs600_disable_vga(rdev);
Line 192... Line 195...
192
		printk(KERN_WARNING "Failed to wait MC idle while "
195
		printk(KERN_WARNING "Failed to wait MC idle while "
193
		       "programming pipes. Bad things might happen.\n");
196
		       "programming pipes. Bad things might happen.\n");
194
	}
197
	}
195
}
198
}
Line 196... Line -...
196
 
-
 
197
#endif
-
 
Line 198... Line 199...
198
 
199
 
199
 
200
 
200
/*
201
/*
201
 * VRAM info
202
 * VRAM info
202
 */
203
 */
203
static void r520_vram_get_type(struct radeon_device *rdev)
204
static void r520_vram_get_type(struct radeon_device *rdev)
-
 
205
{
Line 204... Line 206...
204
{
206
	uint32_t tmp;
205
	uint32_t tmp;
207
    dbgprintf("%s\n\r",__FUNCTION__);
206
 
208
 
207
	rdev->mc.vram_width = 128;
209
	rdev->mc.vram_width = 128;
Line 235... Line 237...
235
 
237
 
236
	rdev->mc.aper_base = drm_get_resource_start(rdev->ddev, 0);
238
	rdev->mc.aper_base = drm_get_resource_start(rdev->ddev, 0);
237
	rdev->mc.aper_size = drm_get_resource_len(rdev->ddev, 0);
239
	rdev->mc.aper_size = drm_get_resource_len(rdev->ddev, 0);
Line -... Line 240...
-
 
240
}
-
 
241
 
-
 
242
/*
-
 
243
 * Global GPU functions
-
 
244
 */
-
 
245
void rs600_disable_vga(struct radeon_device *rdev)
-
 
246
{
-
 
247
    unsigned tmp;
-
 
248
    dbgprintf("%s\n\r",__FUNCTION__);
-
 
249
 
-
 
250
    WREG32(0x330, 0);
-
 
251
    WREG32(0x338, 0);
-
 
252
    tmp = RREG32(0x300);
-
 
253
    tmp &= ~(3 << 16);
-
 
254
    WREG32(0x300, tmp);
-
 
255
    WREG32(0x308, (1 << 8));
-
 
256
    WREG32(0x310, rdev->mc.vram_location);
-
 
257
    WREG32(0x594, 0);
-
 
258
}
-
 
259
 
-
 
260
 
-
 
261
void r420_pipes_init(struct radeon_device *rdev)
-
 
262
{
-
 
263
    unsigned tmp;
-
 
264
    unsigned gb_pipe_select;
-
 
265
    unsigned num_pipes;
-
 
266
 
-
 
267
    dbgprintf("%s\n\r",__FUNCTION__);
-
 
268
 
-
 
269
    /* GA_ENHANCE workaround TCL deadlock issue */
-
 
270
    WREG32(0x4274, (1 << 0) | (1 << 1) | (1 << 2) | (1 << 3));
-
 
271
    /* get max number of pipes */
-
 
272
    gb_pipe_select = RREG32(0x402C);
-
 
273
    num_pipes = ((gb_pipe_select >> 12) & 3) + 1;
-
 
274
    rdev->num_gb_pipes = num_pipes;
-
 
275
    tmp = 0;
-
 
276
    switch (num_pipes) {
-
 
277
    default:
-
 
278
        /* force to 1 pipe */
-
 
279
        num_pipes = 1;
-
 
280
    case 1:
-
 
281
        tmp = (0 << 1);
-
 
282
        break;
-
 
283
    case 2:
-
 
284
        tmp = (3 << 1);
-
 
285
        break;
-
 
286
    case 3:
-
 
287
        tmp = (6 << 1);
-
 
288
        break;
-
 
289
    case 4:
-
 
290
        tmp = (7 << 1);
-
 
291
        break;
-
 
292
    }
-
 
293
    WREG32(0x42C8, (1 << num_pipes) - 1);
-
 
294
    /* Sub pixel 1/12 so we can have 4K rendering according to doc */
-
 
295
    tmp |= (1 << 4) | (1 << 0);
-
 
296
    WREG32(0x4018, tmp);
-
 
297
    if (r100_gui_wait_for_idle(rdev)) {
-
 
298
        printk(KERN_WARNING "Failed to wait GUI idle while "
-
 
299
               "programming pipes. Bad things might happen.\n");
-
 
300
    }
-
 
301
 
-
 
302
    tmp = RREG32(0x170C);
-
 
303
    WREG32(0x170C, tmp | (1 << 31));
-
 
304
 
-
 
305
    WREG32(R300_RB2D_DSTCACHE_MODE,
-
 
306
           RREG32(R300_RB2D_DSTCACHE_MODE) |
-
 
307
           R300_DC_AUTOFLUSH_ENABLE |
-
 
308
           R300_DC_DC_DISABLE_IGNORE_PE);
-
 
309
 
-
 
310
    if (r100_gui_wait_for_idle(rdev)) {
-
 
311
        printk(KERN_WARNING "Failed to wait GUI idle while "
-
 
312
               "programming pipes. Bad things might happen.\n");
-
 
313
    }
-
 
314
    DRM_INFO("radeon: %d pipes initialized.\n", rdev->num_gb_pipes);
-
 
315
}
-
 
316
 
-
 
317
void rv370_pcie_gart_disable(struct radeon_device *rdev)
-
 
318
{
-
 
319
    uint32_t tmp;
-
 
320
    dbgprintf("%s\n\r",__FUNCTION__);
-
 
321
 
-
 
322
    tmp = RREG32_PCIE(RADEON_PCIE_TX_GART_CNTL);
-
 
323
    tmp |= RADEON_PCIE_TX_GART_UNMAPPED_ACCESS_DISCARD;
-
 
324
    WREG32_PCIE(RADEON_PCIE_TX_GART_CNTL, tmp & ~RADEON_PCIE_TX_GART_EN);
-
 
325
    if (rdev->gart.table.vram.robj) {
-
 
326
//        radeon_object_kunmap(rdev->gart.table.vram.robj);
-
 
327
//        radeon_object_unpin(rdev->gart.table.vram.robj);
-
 
328
    }
-
 
329
}
-
 
330
 
-
 
331
void radeon_gart_table_vram_free(struct radeon_device *rdev)
-
 
332
{
-
 
333
    if (rdev->gart.table.vram.robj == NULL) {
-
 
334
        return;
-
 
335
    }
-
 
336
//    radeon_object_kunmap(rdev->gart.table.vram.robj);
-
 
337
//    radeon_object_unpin(rdev->gart.table.vram.robj);
-
 
338
//    radeon_object_unref(&rdev->gart.table.vram.robj);
-
 
339
}
-
 
340
 
-
 
341
/*
-
 
342
 * Common gart functions.
-
 
343
 */
-
 
344
void radeon_gart_unbind(struct radeon_device *rdev, unsigned offset,
-
 
345
            int pages)
-
 
346
{
-
 
347
    unsigned t;
-
 
348
    unsigned p;
-
 
349
    int i, j;
-
 
350
    dbgprintf("%s\n\r",__FUNCTION__);
-
 
351
 
-
 
352
    if (!rdev->gart.ready) {
-
 
353
        dbgprintf("trying to unbind memory to unitialized GART !\n");
-
 
354
        return;
-
 
355
    }
-
 
356
    t = offset / 4096;
-
 
357
    p = t / (PAGE_SIZE / 4096);
-
 
358
    for (i = 0; i < pages; i++, p++) {
-
 
359
        if (rdev->gart.pages[p]) {
-
 
360
//            pci_unmap_page(rdev->pdev, rdev->gart.pages_addr[p],
-
 
361
//                       PAGE_SIZE, PCI_DMA_BIDIRECTIONAL);
-
 
362
            rdev->gart.pages[p] = NULL;
-
 
363
            rdev->gart.pages_addr[p] = 0;
-
 
364
            for (j = 0; j < (PAGE_SIZE / 4096); j++, t++) {
-
 
365
                radeon_gart_set_page(rdev, t, 0);
-
 
366
            }
-
 
367
        }
-
 
368
    }
-
 
369
    mb();
-
 
370
    radeon_gart_tlb_flush(rdev);
-
 
371
}
-
 
372
 
-
 
373
 
-
 
374
 
-
 
375
void radeon_gart_fini(struct radeon_device *rdev)
-
 
376
{
-
 
377
    if (rdev->gart.pages && rdev->gart.pages_addr && rdev->gart.ready) {
-
 
378
        /* unbind pages */
-
 
379
        radeon_gart_unbind(rdev, 0, rdev->gart.num_cpu_pages);
-
 
380
    }
-
 
381
    rdev->gart.ready = false;
-
 
382
//    kfree(rdev->gart.pages);
-
 
383
//    kfree(rdev->gart.pages_addr);
-
 
384
    rdev->gart.pages = NULL;
-
 
385
    rdev->gart.pages_addr = NULL;
-
 
386
}
-
 
387
 
-
 
388
 
-
 
389
 
-
 
390
int radeon_agp_init(struct radeon_device *rdev)
-
 
391
{
-
 
392
 
-
 
393
    dbgprintf("%s\n\r",__FUNCTION__);
-
 
394
 
-
 
395
#if __OS_HAS_AGP
-
 
396
    struct radeon_agpmode_quirk *p = radeon_agpmode_quirk_list;
-
 
397
    struct drm_agp_mode mode;
-
 
398
    struct drm_agp_info info;
-
 
399
    uint32_t agp_status;
-
 
400
    int default_mode;
-
 
401
    bool is_v3;
-
 
402
    int ret;
-
 
403
 
-
 
404
    /* Acquire AGP. */
-
 
405
    if (!rdev->ddev->agp->acquired) {
-
 
406
        ret = drm_agp_acquire(rdev->ddev);
-
 
407
        if (ret) {
-
 
408
            DRM_ERROR("Unable to acquire AGP: %d\n", ret);
-
 
409
            return ret;
-
 
410
        }
-
 
411
    }
-
 
412
 
-
 
413
    ret = drm_agp_info(rdev->ddev, &info);
-
 
414
    if (ret) {
-
 
415
        DRM_ERROR("Unable to get AGP info: %d\n", ret);
-
 
416
        return ret;
-
 
417
    }
-
 
418
    mode.mode = info.mode;
-
 
419
    agp_status = (RREG32(RADEON_AGP_STATUS) | RADEON_AGPv3_MODE) & mode.mode;
-
 
420
    is_v3 = !!(agp_status & RADEON_AGPv3_MODE);
-
 
421
 
-
 
422
    if (is_v3) {
-
 
423
        default_mode = (agp_status & RADEON_AGPv3_8X_MODE) ? 8 : 4;
-
 
424
    } else {
-
 
425
        if (agp_status & RADEON_AGP_4X_MODE) {
-
 
426
            default_mode = 4;
-
 
427
        } else if (agp_status & RADEON_AGP_2X_MODE) {
-
 
428
            default_mode = 2;
-
 
429
        } else {
-
 
430
            default_mode = 1;
-
 
431
        }
-
 
432
    }
-
 
433
 
-
 
434
    /* Apply AGPMode Quirks */
-
 
435
    while (p && p->chip_device != 0) {
-
 
436
        if (info.id_vendor == p->hostbridge_vendor &&
-
 
437
            info.id_device == p->hostbridge_device &&
-
 
438
            rdev->pdev->vendor == p->chip_vendor &&
-
 
439
            rdev->pdev->device == p->chip_device &&
-
 
440
            rdev->pdev->subsystem_vendor == p->subsys_vendor &&
-
 
441
            rdev->pdev->subsystem_device == p->subsys_device) {
-
 
442
            default_mode = p->default_mode;
-
 
443
        }
-
 
444
        ++p;
-
 
445
    }
-
 
446
 
-
 
447
    if (radeon_agpmode > 0) {
-
 
448
        if ((radeon_agpmode < (is_v3 ? 4 : 1)) ||
-
 
449
            (radeon_agpmode > (is_v3 ? 8 : 4)) ||
-
 
450
            (radeon_agpmode & (radeon_agpmode - 1))) {
-
 
451
            DRM_ERROR("Illegal AGP Mode: %d (valid %s), leaving at %d\n",
-
 
452
                  radeon_agpmode, is_v3 ? "4, 8" : "1, 2, 4",
-
 
453
                  default_mode);
-
 
454
            radeon_agpmode = default_mode;
-
 
455
        } else {
-
 
456
            DRM_INFO("AGP mode requested: %d\n", radeon_agpmode);
-
 
457
        }
-
 
458
    } else {
-
 
459
        radeon_agpmode = default_mode;
-
 
460
    }
-
 
461
 
-
 
462
    mode.mode &= ~RADEON_AGP_MODE_MASK;
-
 
463
    if (is_v3) {
-
 
464
        switch (radeon_agpmode) {
-
 
465
        case 8:
-
 
466
            mode.mode |= RADEON_AGPv3_8X_MODE;
-
 
467
            break;
-
 
468
        case 4:
-
 
469
        default:
-
 
470
            mode.mode |= RADEON_AGPv3_4X_MODE;
-
 
471
            break;
-
 
472
        }
-
 
473
    } else {
-
 
474
        switch (radeon_agpmode) {
-
 
475
        case 4:
-
 
476
            mode.mode |= RADEON_AGP_4X_MODE;
-
 
477
            break;
-
 
478
        case 2:
-
 
479
            mode.mode |= RADEON_AGP_2X_MODE;
-
 
480
            break;
-
 
481
        case 1:
-
 
482
        default:
-
 
483
            mode.mode |= RADEON_AGP_1X_MODE;
-
 
484
            break;
-
 
485
        }
-
 
486
    }
-
 
487
 
-
 
488
    mode.mode &= ~RADEON_AGP_FW_MODE; /* disable fw */
-
 
489
    ret = drm_agp_enable(rdev->ddev, mode);
-
 
490
    if (ret) {
-
 
491
        DRM_ERROR("Unable to enable AGP (mode = 0x%lx)\n", mode.mode);
-
 
492
        return ret;
-
 
493
    }
-
 
494
 
-
 
495
    rdev->mc.agp_base = rdev->ddev->agp->agp_info.aper_base;
-
 
496
    rdev->mc.gtt_size = rdev->ddev->agp->agp_info.aper_size << 20;
-
 
497
 
-
 
498
    /* workaround some hw issues */
-
 
499
    if (rdev->family < CHIP_R200) {
-
 
500
        WREG32(RADEON_AGP_CNTL, RREG32(RADEON_AGP_CNTL) | 0x000e0000);
-
 
501
    }
-
 
502
    return 0;
-
 
503
#else
-
 
504
    return 0;
-
 
505
#endif
-
 
506
}
-
 
507
 
-
 
508
 
-
 
509
void rs600_mc_disable_clients(struct radeon_device *rdev)
-
 
510
{
-
 
511
    unsigned tmp;
-
 
512
    dbgprintf("%s\n",__FUNCTION__);
-
 
513
 
-
 
514
    if (r100_gui_wait_for_idle(rdev)) {
-
 
515
        printk(KERN_WARNING "Failed to wait GUI idle while "
-
 
516
               "programming pipes. Bad things might happen.\n");
-
 
517
    }
-
 
518
 
-
 
519
    tmp = RREG32(AVIVO_D1VGA_CONTROL);
-
 
520
    WREG32(AVIVO_D1VGA_CONTROL, tmp & ~AVIVO_DVGA_CONTROL_MODE_ENABLE);
-
 
521
    tmp = RREG32(AVIVO_D2VGA_CONTROL);
-
 
522
    WREG32(AVIVO_D2VGA_CONTROL, tmp & ~AVIVO_DVGA_CONTROL_MODE_ENABLE);
-
 
523
 
-
 
524
    tmp = RREG32(AVIVO_D1CRTC_CONTROL);
-
 
525
    WREG32(AVIVO_D1CRTC_CONTROL, tmp & ~AVIVO_CRTC_EN);
-
 
526
    tmp = RREG32(AVIVO_D2CRTC_CONTROL);
-
 
527
    WREG32(AVIVO_D2CRTC_CONTROL, tmp & ~AVIVO_CRTC_EN);
-
 
528
 
-
 
529
    /* make sure all previous write got through */
-
 
530
    tmp = RREG32(AVIVO_D2CRTC_CONTROL);
-
 
531
 
-
 
532
    mdelay(1);
-
 
533
 
-
 
534
    dbgprintf("done\n");
-
 
535
 
-
 
536
}
-
 
537
 
-
 
538
int rv370_pcie_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr)
-
 
539
{
-
 
540
    void __iomem *ptr = (void *)rdev->gart.table.vram.ptr;
-
 
541
 
-
 
542
    if (i < 0 || i > rdev->gart.num_gpu_pages) {
-
 
543
        return -EINVAL;
-
 
544
    }
-
 
545
    addr = (((u32_t)addr) >> 8) | ((upper_32_bits(addr) & 0xff) << 4) | 0xC;
-
 
546
    writel(cpu_to_le32(addr), ((void __iomem *)ptr) + (i * 4));
-
 
547
    return 0;
-
 
548
}
-
 
549
 
-
 
550
 
-
 
551
int radeon_gart_init(struct radeon_device *rdev)
-
 
552
{
-
 
553
 
-
 
554
    dbgprintf("%s\n",__FUNCTION__);
-
 
555
 
-
 
556
    if (rdev->gart.pages) {
-
 
557
        return 0;
-
 
558
    }
-
 
559
    /* We need PAGE_SIZE >= 4096 */
-
 
560
    if (PAGE_SIZE < 4096) {
-
 
561
        DRM_ERROR("Page size is smaller than GPU page size!\n");
-
 
562
        return -EINVAL;
-
 
563
    }
-
 
564
    /* Compute table size */
-
 
565
    rdev->gart.num_cpu_pages = rdev->mc.gtt_size / PAGE_SIZE;
-
 
566
    rdev->gart.num_gpu_pages = rdev->mc.gtt_size / 4096;
-
 
567
    DRM_INFO("GART: num cpu pages %u, num gpu pages %u\n",
-
 
568
         rdev->gart.num_cpu_pages, rdev->gart.num_gpu_pages);
-
 
569
    /* Allocate pages table */
-
 
570
    rdev->gart.pages = kzalloc(sizeof(void *) * rdev->gart.num_cpu_pages,
-
 
571
                   GFP_KERNEL);
-
 
572
    if (rdev->gart.pages == NULL) {
-
 
573
//        radeon_gart_fini(rdev);
-
 
574
        return -ENOMEM;
-
 
575
    }
-
 
576
    rdev->gart.pages_addr = kzalloc(sizeof(u32_t) *
-
 
577
                    rdev->gart.num_cpu_pages, GFP_KERNEL);
-
 
578
    if (rdev->gart.pages_addr == NULL) {
-
 
579
//        radeon_gart_fini(rdev);
-
 
580
        return -ENOMEM;
-
 
581
    }
-
 
582
    return 0;
-
 
583
}
-
 
584
 
-
 
585
int radeon_gart_table_vram_alloc(struct radeon_device *rdev)
-
 
586
{
-
 
587
    uint32_t gpu_addr;
-
 
588
    int r;
-
 
589
 
-
 
590
//    if (rdev->gart.table.vram.robj == NULL) {
-
 
591
//        r = radeon_object_create(rdev, NULL,
-
 
592
//                     rdev->gart.table_size,
-
 
593
//                     true,
-
 
594
//                     RADEON_GEM_DOMAIN_VRAM,
-
 
595
//                     false, &rdev->gart.table.vram.robj);
-
 
596
//        if (r) {
-
 
597
//            return r;
-
 
598
//        }
-
 
599
//    }
-
 
600
//    r = radeon_object_pin(rdev->gart.table.vram.robj,
-
 
601
//                  RADEON_GEM_DOMAIN_VRAM, &gpu_addr);
-
 
602
//    if (r) {
-
 
603
//        radeon_object_unref(&rdev->gart.table.vram.robj);
-
 
604
//        return r;
-
 
605
//    }
-
 
606
//    r = radeon_object_kmap(rdev->gart.table.vram.robj,
-
 
607
//                   (void **)&rdev->gart.table.vram.ptr);
-
 
608
//    if (r) {
-
 
609
//        radeon_object_unpin(rdev->gart.table.vram.robj);
-
 
610
//        radeon_object_unref(&rdev->gart.table.vram.robj);
-
 
611
//        DRM_ERROR("radeon: failed to map gart vram table.\n");
-
 
612
//        return r;
-
 
613
//    }
-
 
614
 
-
 
615
    gpu_addr = 0x800000;
-
 
616
 
-
 
617
    u32_t pci_addr = rdev->mc.aper_base + gpu_addr;
-
 
618
 
-
 
619
    rdev->gart.table.vram.ptr = (void*)MapIoMem(pci_addr, rdev->gart.table_size, PG_SW);
-
 
620
 
-
 
621
    rdev->gart.table_addr = gpu_addr;
-
 
622
 
-
 
623
    dbgprintf("alloc gart vram:\n  gpu_base %x pci_base %x lin_addr %x",
-
 
624
               gpu_addr, pci_addr, rdev->gart.table.vram.ptr);
-
 
625
 
-
 
626
    return 0;
-
 
627
}
-
 
628
 
-
 
629
void rv370_pcie_gart_tlb_flush(struct radeon_device *rdev);
-
 
630
 
-
 
631
int rv370_pcie_gart_enable(struct radeon_device *rdev)
-
 
632
{
-
 
633
    uint32_t table_addr;
-
 
634
    uint32_t tmp;
-
 
635
    int r;
-
 
636
 
-
 
637
    dbgprintf("%s\n",__FUNCTION__);
-
 
638
 
-
 
639
    /* Initialize common gart structure */
-
 
640
    r = radeon_gart_init(rdev);
-
 
641
    if (r) {
-
 
642
        return r;
-
 
643
    }
-
 
644
 //   r = rv370_debugfs_pcie_gart_info_init(rdev);
-
 
645
 //   if (r) {
-
 
646
 //       DRM_ERROR("Failed to register debugfs file for PCIE gart !\n");
-
 
647
 //   }
-
 
648
    rdev->gart.table_size = rdev->gart.num_gpu_pages * 4;
-
 
649
    r = radeon_gart_table_vram_alloc(rdev);
-
 
650
    if (r) {
-
 
651
        return r;
-
 
652
    }
-
 
653
    /* discard memory request outside of configured range */
-
 
654
    tmp = RADEON_PCIE_TX_GART_UNMAPPED_ACCESS_DISCARD;
-
 
655
    WREG32_PCIE(RADEON_PCIE_TX_GART_CNTL, tmp);
-
 
656
    WREG32_PCIE(RADEON_PCIE_TX_GART_START_LO, rdev->mc.gtt_location);
-
 
657
    tmp = rdev->mc.gtt_location + rdev->mc.gtt_size - 4096;
-
 
658
    WREG32_PCIE(RADEON_PCIE_TX_GART_END_LO, tmp);
-
 
659
    WREG32_PCIE(RADEON_PCIE_TX_GART_START_HI, 0);
-
 
660
    WREG32_PCIE(RADEON_PCIE_TX_GART_END_HI, 0);
-
 
661
    table_addr = rdev->gart.table_addr;
-
 
662
    WREG32_PCIE(RADEON_PCIE_TX_GART_BASE, table_addr);
-
 
663
    /* FIXME: setup default page */
-
 
664
    WREG32_PCIE(RADEON_PCIE_TX_DISCARD_RD_ADDR_LO, rdev->mc.vram_location);
-
 
665
    WREG32_PCIE(RADEON_PCIE_TX_DISCARD_RD_ADDR_HI, 0);
-
 
666
    /* Clear error */
-
 
667
    WREG32_PCIE(0x18, 0);
-
 
668
    tmp = RREG32_PCIE(RADEON_PCIE_TX_GART_CNTL);
-
 
669
    tmp |= RADEON_PCIE_TX_GART_EN;
-
 
670
    tmp |= RADEON_PCIE_TX_GART_UNMAPPED_ACCESS_DISCARD;
-
 
671
    WREG32_PCIE(RADEON_PCIE_TX_GART_CNTL, tmp);
-
 
672
    rv370_pcie_gart_tlb_flush(rdev);
-
 
673
    DRM_INFO("PCIE GART of %uM enabled (table at 0x%08X).\n",
-
 
674
         rdev->mc.gtt_size >> 20, table_addr);
-
 
675
    rdev->gart.ready = true;
-
 
676
    return 0;
-
 
677
}
-
 
678
 
-
 
679
void rv370_pcie_gart_tlb_flush(struct radeon_device *rdev)
-
 
680
{
-
 
681
    uint32_t tmp;
-
 
682
    int i;
-
 
683
 
-
 
684
    /* Workaround HW bug do flush 2 times */
-
 
685
    for (i = 0; i < 2; i++) {
-
 
686
        tmp = RREG32_PCIE(RADEON_PCIE_TX_GART_CNTL);
-
 
687
        WREG32_PCIE(RADEON_PCIE_TX_GART_CNTL, tmp | RADEON_PCIE_TX_GART_INVALIDATE_TLB);
-
 
688
        (void)RREG32_PCIE(RADEON_PCIE_TX_GART_CNTL);
-
 
689
        WREG32_PCIE(RADEON_PCIE_TX_GART_CNTL, tmp);
-
 
690
        mb();
-
 
691
    }
-
 
692
}
-
 
693
 
-
 
694
int r300_gart_enable(struct radeon_device *rdev)
-
 
695
{
-
 
696
#if __OS_HAS_AGP
-
 
697
    if (rdev->flags & RADEON_IS_AGP) {
-
 
698
        if (rdev->family > CHIP_RV350) {
-
 
699
            rv370_pcie_gart_disable(rdev);
-
 
700
        } else {
-
 
701
            r100_pci_gart_disable(rdev);
-
 
702
        }
-
 
703
        return 0;
-
 
704
    }
-
 
705
#endif
-
 
706
    if (rdev->flags & RADEON_IS_PCIE) {
-
 
707
        rdev->asic->gart_disable = &rv370_pcie_gart_disable;
-
 
708
        rdev->asic->gart_tlb_flush = &rv370_pcie_gart_tlb_flush;
-
 
709
        rdev->asic->gart_set_page = &rv370_pcie_gart_set_page;
-
 
710
        return rv370_pcie_gart_enable(rdev);
-
 
711
    }
-
 
712
 //   return r100_pci_gart_enable(rdev);
-
 
713
}
-
 
714
 
-
 
715
 
-
 
716
 
-
 
717
int radeon_fence_driver_init(struct radeon_device *rdev)
-
 
718
{
-
 
719
    unsigned long irq_flags;
-
 
720
    int r;
-
 
721
 
-
 
722
//    write_lock_irqsave(&rdev->fence_drv.lock, irq_flags);
-
 
723
    r = radeon_scratch_get(rdev, &rdev->fence_drv.scratch_reg);
-
 
724
    if (r) {
-
 
725
        DRM_ERROR("Fence failed to get a scratch register.");
-
 
726
//        write_unlock_irqrestore(&rdev->fence_drv.lock, irq_flags);
-
 
727
        return r;
-
 
728
    }
-
 
729
    WREG32(rdev->fence_drv.scratch_reg, 0);
-
 
730
//    atomic_set(&rdev->fence_drv.seq, 0);
-
 
731
//    INIT_LIST_HEAD(&rdev->fence_drv.created);
-
 
732
//    INIT_LIST_HEAD(&rdev->fence_drv.emited);
-
 
733
//    INIT_LIST_HEAD(&rdev->fence_drv.signaled);
-
 
734
    rdev->fence_drv.count_timeout = 0;
-
 
735
//    init_waitqueue_head(&rdev->fence_drv.queue);
-
 
736
//    write_unlock_irqrestore(&rdev->fence_drv.lock, irq_flags);
-
 
737
//    if (radeon_debugfs_fence_init(rdev)) {
-
 
738
//        DRM_ERROR("Failed to register debugfs file for fence !\n");
-
 
739
//    }
-
 
740
    return 0;
-
 
741
}
-
 
742
 
-
 
743
 
-
 
744
int radeon_gart_bind(struct radeon_device *rdev, unsigned offset,
-
 
745
             int pages, u32_t *pagelist)
-
 
746
{
-
 
747
    unsigned t;
-
 
748
    unsigned p;
-
 
749
    uint64_t page_base;
-
 
750
    int i, j;
-
 
751
 
-
 
752
    dbgprintf("%s\n\r",__FUNCTION__);
-
 
753
 
-
 
754
 
-
 
755
    if (!rdev->gart.ready) {
-
 
756
        DRM_ERROR("trying to bind memory to unitialized GART !\n");
-
 
757
        return -EINVAL;
-
 
758
    }
-
 
759
    t = offset / 4096;
-
 
760
    p = t / (PAGE_SIZE / 4096);
-
 
761
 
-
 
762
    for (i = 0; i < pages; i++, p++) {
-
 
763
        /* we need to support large memory configurations */
-
 
764
        /* assume that unbind have already been call on the range */
-
 
765
 
-
 
766
        rdev->gart.pages_addr[p] = pagelist[i] & ~4095;
-
 
767
 
-
 
768
        //if (pci_dma_mapping_error(rdev->pdev, rdev->gart.pages_addr[p])) {
-
 
769
        //    /* FIXME: failed to map page (return -ENOMEM?) */
-
 
770
        //    radeon_gart_unbind(rdev, offset, pages);
-
 
771
        //    return -ENOMEM;
-
 
772
        //}
-
 
773
        rdev->gart.pages[p] = pagelist[i];
-
 
774
        page_base = (uint32_t)rdev->gart.pages_addr[p];
-
 
775
        for (j = 0; j < (PAGE_SIZE / 4096); j++, t++) {
-
 
776
            radeon_gart_set_page(rdev, t, page_base);
-
 
777
            page_base += 4096;
-
 
778
        }
-
 
779
    }
-
 
780
    mb();
-
 
781
    radeon_gart_tlb_flush(rdev);
-
 
782
 
-
 
783
    dbgprintf("done %s\n",__FUNCTION__);
-
 
784
 
-
 
785
    return 0;
-
 
786
}
-
 
787