Rev 2005 | Rev 3120 | Go to most recent revision | Only display areas with differences | Regard whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 2005 | Rev 2997 | ||
---|---|---|---|
1 | /* |
1 | /* |
2 | * Copyright 2008 Advanced Micro Devices, Inc. |
2 | * Copyright 2008 Advanced Micro Devices, Inc. |
3 | * Copyright 2008 Red Hat Inc. |
3 | * Copyright 2008 Red Hat Inc. |
4 | * Copyright 2009 Jerome Glisse. |
4 | * Copyright 2009 Jerome Glisse. |
5 | * |
5 | * |
6 | * Permission is hereby granted, free of charge, to any person obtaining a |
6 | * Permission is hereby granted, free of charge, to any person obtaining a |
7 | * copy of this software and associated documentation files (the "Software"), |
7 | * copy of this software and associated documentation files (the "Software"), |
8 | * to deal in the Software without restriction, including without limitation |
8 | * to deal in the Software without restriction, including without limitation |
9 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
9 | * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
10 | * and/or sell copies of the Software, and to permit persons to whom the |
10 | * and/or sell copies of the Software, and to permit persons to whom the |
11 | * Software is furnished to do so, subject to the following conditions: |
11 | * Software is furnished to do so, subject to the following conditions: |
12 | * |
12 | * |
13 | * The above copyright notice and this permission notice shall be included in |
13 | * The above copyright notice and this permission notice shall be included in |
14 | * all copies or substantial portions of the Software. |
14 | * all copies or substantial portions of the Software. |
15 | * |
15 | * |
16 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
16 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
17 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
17 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
18 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
18 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
19 | * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR |
19 | * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR |
20 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
20 | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
21 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR |
21 | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR |
22 | * OTHER DEALINGS IN THE SOFTWARE. |
22 | * OTHER DEALINGS IN THE SOFTWARE. |
23 | * |
23 | * |
24 | * Authors: Dave Airlie |
24 | * Authors: Dave Airlie |
25 | * Alex Deucher |
25 | * Alex Deucher |
26 | * Jerome Glisse |
26 | * Jerome Glisse |
27 | */ |
27 | */ |
28 | #include |
28 | #include |
29 | #include |
29 | #include |
30 | #include "drmP.h" |
30 | #include |
31 | #include "radeon_reg.h" |
31 | #include "radeon_reg.h" |
32 | #include "radeon.h" |
32 | #include "radeon.h" |
33 | #include "radeon_asic.h" |
33 | #include "radeon_asic.h" |
34 | #include "atom.h" |
34 | #include "atom.h" |
35 | #include "r100d.h" |
35 | #include "r100d.h" |
36 | #include "r420d.h" |
36 | #include "r420d.h" |
37 | #include "r420_reg_safe.h" |
37 | #include "r420_reg_safe.h" |
38 | 38 | ||
39 | static void r420_set_reg_safe(struct radeon_device *rdev) |
39 | static void r420_set_reg_safe(struct radeon_device *rdev) |
40 | { |
40 | { |
41 | rdev->config.r300.reg_safe_bm = r420_reg_safe_bm; |
41 | rdev->config.r300.reg_safe_bm = r420_reg_safe_bm; |
42 | rdev->config.r300.reg_safe_bm_size = ARRAY_SIZE(r420_reg_safe_bm); |
42 | rdev->config.r300.reg_safe_bm_size = ARRAY_SIZE(r420_reg_safe_bm); |
43 | } |
43 | } |
44 | 44 | ||
45 | void r420_pipes_init(struct radeon_device *rdev) |
45 | void r420_pipes_init(struct radeon_device *rdev) |
46 | { |
46 | { |
47 | unsigned tmp; |
47 | unsigned tmp; |
48 | unsigned gb_pipe_select; |
48 | unsigned gb_pipe_select; |
49 | unsigned num_pipes; |
49 | unsigned num_pipes; |
50 | 50 | ||
51 | /* GA_ENHANCE workaround TCL deadlock issue */ |
51 | /* GA_ENHANCE workaround TCL deadlock issue */ |
52 | WREG32(R300_GA_ENHANCE, R300_GA_DEADLOCK_CNTL | R300_GA_FASTSYNC_CNTL | |
52 | WREG32(R300_GA_ENHANCE, R300_GA_DEADLOCK_CNTL | R300_GA_FASTSYNC_CNTL | |
53 | (1 << 2) | (1 << 3)); |
53 | (1 << 2) | (1 << 3)); |
54 | /* add idle wait as per freedesktop.org bug 24041 */ |
54 | /* add idle wait as per freedesktop.org bug 24041 */ |
55 | if (r100_gui_wait_for_idle(rdev)) { |
55 | if (r100_gui_wait_for_idle(rdev)) { |
56 | printk(KERN_WARNING "Failed to wait GUI idle while " |
56 | printk(KERN_WARNING "Failed to wait GUI idle while " |
57 | "programming pipes. Bad things might happen.\n"); |
57 | "programming pipes. Bad things might happen.\n"); |
58 | } |
58 | } |
59 | /* get max number of pipes */ |
59 | /* get max number of pipes */ |
60 | gb_pipe_select = RREG32(R400_GB_PIPE_SELECT); |
60 | gb_pipe_select = RREG32(R400_GB_PIPE_SELECT); |
61 | num_pipes = ((gb_pipe_select >> 12) & 3) + 1; |
61 | num_pipes = ((gb_pipe_select >> 12) & 3) + 1; |
62 | 62 | ||
63 | /* SE chips have 1 pipe */ |
63 | /* SE chips have 1 pipe */ |
64 | if ((rdev->pdev->device == 0x5e4c) || |
64 | if ((rdev->pdev->device == 0x5e4c) || |
65 | (rdev->pdev->device == 0x5e4f)) |
65 | (rdev->pdev->device == 0x5e4f)) |
66 | num_pipes = 1; |
66 | num_pipes = 1; |
67 | 67 | ||
68 | rdev->num_gb_pipes = num_pipes; |
68 | rdev->num_gb_pipes = num_pipes; |
69 | tmp = 0; |
69 | tmp = 0; |
70 | switch (num_pipes) { |
70 | switch (num_pipes) { |
71 | default: |
71 | default: |
72 | /* force to 1 pipe */ |
72 | /* force to 1 pipe */ |
73 | num_pipes = 1; |
73 | num_pipes = 1; |
74 | case 1: |
74 | case 1: |
75 | tmp = (0 << 1); |
75 | tmp = (0 << 1); |
76 | break; |
76 | break; |
77 | case 2: |
77 | case 2: |
78 | tmp = (3 << 1); |
78 | tmp = (3 << 1); |
79 | break; |
79 | break; |
80 | case 3: |
80 | case 3: |
81 | tmp = (6 << 1); |
81 | tmp = (6 << 1); |
82 | break; |
82 | break; |
83 | case 4: |
83 | case 4: |
84 | tmp = (7 << 1); |
84 | tmp = (7 << 1); |
85 | break; |
85 | break; |
86 | } |
86 | } |
87 | WREG32(R500_SU_REG_DEST, (1 << num_pipes) - 1); |
87 | WREG32(R500_SU_REG_DEST, (1 << num_pipes) - 1); |
88 | /* Sub pixel 1/12 so we can have 4K rendering according to doc */ |
88 | /* Sub pixel 1/12 so we can have 4K rendering according to doc */ |
89 | tmp |= R300_TILE_SIZE_16 | R300_ENABLE_TILING; |
89 | tmp |= R300_TILE_SIZE_16 | R300_ENABLE_TILING; |
90 | WREG32(R300_GB_TILE_CONFIG, tmp); |
90 | WREG32(R300_GB_TILE_CONFIG, tmp); |
91 | if (r100_gui_wait_for_idle(rdev)) { |
91 | if (r100_gui_wait_for_idle(rdev)) { |
92 | printk(KERN_WARNING "Failed to wait GUI idle while " |
92 | printk(KERN_WARNING "Failed to wait GUI idle while " |
93 | "programming pipes. Bad things might happen.\n"); |
93 | "programming pipes. Bad things might happen.\n"); |
94 | } |
94 | } |
95 | 95 | ||
96 | tmp = RREG32(R300_DST_PIPE_CONFIG); |
96 | tmp = RREG32(R300_DST_PIPE_CONFIG); |
97 | WREG32(R300_DST_PIPE_CONFIG, tmp | R300_PIPE_AUTO_CONFIG); |
97 | WREG32(R300_DST_PIPE_CONFIG, tmp | R300_PIPE_AUTO_CONFIG); |
98 | 98 | ||
99 | WREG32(R300_RB2D_DSTCACHE_MODE, |
99 | WREG32(R300_RB2D_DSTCACHE_MODE, |
100 | RREG32(R300_RB2D_DSTCACHE_MODE) | |
100 | RREG32(R300_RB2D_DSTCACHE_MODE) | |
101 | R300_DC_AUTOFLUSH_ENABLE | |
101 | R300_DC_AUTOFLUSH_ENABLE | |
102 | R300_DC_DC_DISABLE_IGNORE_PE); |
102 | R300_DC_DC_DISABLE_IGNORE_PE); |
103 | 103 | ||
104 | if (r100_gui_wait_for_idle(rdev)) { |
104 | if (r100_gui_wait_for_idle(rdev)) { |
105 | printk(KERN_WARNING "Failed to wait GUI idle while " |
105 | printk(KERN_WARNING "Failed to wait GUI idle while " |
106 | "programming pipes. Bad things might happen.\n"); |
106 | "programming pipes. Bad things might happen.\n"); |
107 | } |
107 | } |
108 | 108 | ||
109 | if (rdev->family == CHIP_RV530) { |
109 | if (rdev->family == CHIP_RV530) { |
110 | tmp = RREG32(RV530_GB_PIPE_SELECT2); |
110 | tmp = RREG32(RV530_GB_PIPE_SELECT2); |
111 | if ((tmp & 3) == 3) |
111 | if ((tmp & 3) == 3) |
112 | rdev->num_z_pipes = 2; |
112 | rdev->num_z_pipes = 2; |
113 | else |
113 | else |
114 | rdev->num_z_pipes = 1; |
114 | rdev->num_z_pipes = 1; |
115 | } else |
115 | } else |
116 | rdev->num_z_pipes = 1; |
116 | rdev->num_z_pipes = 1; |
117 | 117 | ||
118 | DRM_INFO("radeon: %d quad pipes, %d z pipes initialized.\n", |
118 | DRM_INFO("radeon: %d quad pipes, %d z pipes initialized.\n", |
119 | rdev->num_gb_pipes, rdev->num_z_pipes); |
119 | rdev->num_gb_pipes, rdev->num_z_pipes); |
120 | } |
120 | } |
121 | 121 | ||
122 | u32 r420_mc_rreg(struct radeon_device *rdev, u32 reg) |
122 | u32 r420_mc_rreg(struct radeon_device *rdev, u32 reg) |
123 | { |
123 | { |
124 | u32 r; |
124 | u32 r; |
125 | 125 | ||
126 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg)); |
126 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg)); |
127 | r = RREG32(R_0001FC_MC_IND_DATA); |
127 | r = RREG32(R_0001FC_MC_IND_DATA); |
128 | return r; |
128 | return r; |
129 | } |
129 | } |
130 | 130 | ||
131 | void r420_mc_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
131 | void r420_mc_wreg(struct radeon_device *rdev, u32 reg, u32 v) |
132 | { |
132 | { |
133 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg) | |
133 | WREG32(R_0001F8_MC_IND_INDEX, S_0001F8_MC_IND_ADDR(reg) | |
134 | S_0001F8_MC_IND_WR_EN(1)); |
134 | S_0001F8_MC_IND_WR_EN(1)); |
135 | WREG32(R_0001FC_MC_IND_DATA, v); |
135 | WREG32(R_0001FC_MC_IND_DATA, v); |
136 | } |
136 | } |
137 | 137 | ||
138 | static void r420_debugfs(struct radeon_device *rdev) |
138 | static void r420_debugfs(struct radeon_device *rdev) |
139 | { |
139 | { |
140 | if (r100_debugfs_rbbm_init(rdev)) { |
140 | if (r100_debugfs_rbbm_init(rdev)) { |
141 | DRM_ERROR("Failed to register debugfs file for RBBM !\n"); |
141 | DRM_ERROR("Failed to register debugfs file for RBBM !\n"); |
142 | } |
142 | } |
143 | if (r420_debugfs_pipes_info_init(rdev)) { |
143 | if (r420_debugfs_pipes_info_init(rdev)) { |
144 | DRM_ERROR("Failed to register debugfs file for pipes !\n"); |
144 | DRM_ERROR("Failed to register debugfs file for pipes !\n"); |
145 | } |
145 | } |
146 | } |
146 | } |
147 | 147 | ||
148 | static void r420_clock_resume(struct radeon_device *rdev) |
148 | static void r420_clock_resume(struct radeon_device *rdev) |
149 | { |
149 | { |
150 | u32 sclk_cntl; |
150 | u32 sclk_cntl; |
151 | 151 | ||
152 | if (radeon_dynclks != -1 && radeon_dynclks) |
152 | if (radeon_dynclks != -1 && radeon_dynclks) |
153 | radeon_atom_set_clock_gating(rdev, 1); |
153 | radeon_atom_set_clock_gating(rdev, 1); |
154 | sclk_cntl = RREG32_PLL(R_00000D_SCLK_CNTL); |
154 | sclk_cntl = RREG32_PLL(R_00000D_SCLK_CNTL); |
155 | sclk_cntl |= S_00000D_FORCE_CP(1) | S_00000D_FORCE_VIP(1); |
155 | sclk_cntl |= S_00000D_FORCE_CP(1) | S_00000D_FORCE_VIP(1); |
156 | if (rdev->family == CHIP_R420) |
156 | if (rdev->family == CHIP_R420) |
157 | sclk_cntl |= S_00000D_FORCE_PX(1) | S_00000D_FORCE_TX(1); |
157 | sclk_cntl |= S_00000D_FORCE_PX(1) | S_00000D_FORCE_TX(1); |
158 | WREG32_PLL(R_00000D_SCLK_CNTL, sclk_cntl); |
158 | WREG32_PLL(R_00000D_SCLK_CNTL, sclk_cntl); |
159 | } |
159 | } |
160 | 160 | ||
161 | static void r420_cp_errata_init(struct radeon_device *rdev) |
161 | static void r420_cp_errata_init(struct radeon_device *rdev) |
162 | { |
162 | { |
- | 163 | struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; |
|
- | 164 | ||
163 | /* RV410 and R420 can lock up if CP DMA to host memory happens |
165 | /* RV410 and R420 can lock up if CP DMA to host memory happens |
164 | * while the 2D engine is busy. |
166 | * while the 2D engine is busy. |
165 | * |
167 | * |
166 | * The proper workaround is to queue a RESYNC at the beginning |
168 | * The proper workaround is to queue a RESYNC at the beginning |
167 | * of the CP init, apparently. |
169 | * of the CP init, apparently. |
168 | */ |
170 | */ |
169 | radeon_scratch_get(rdev, &rdev->config.r300.resync_scratch); |
171 | radeon_scratch_get(rdev, &rdev->config.r300.resync_scratch); |
170 | radeon_ring_lock(rdev, 8); |
172 | radeon_ring_lock(rdev, ring, 8); |
171 | radeon_ring_write(rdev, PACKET0(R300_CP_RESYNC_ADDR, 1)); |
173 | radeon_ring_write(ring, PACKET0(R300_CP_RESYNC_ADDR, 1)); |
172 | radeon_ring_write(rdev, rdev->config.r300.resync_scratch); |
174 | radeon_ring_write(ring, rdev->config.r300.resync_scratch); |
173 | radeon_ring_write(rdev, 0xDEADBEEF); |
175 | radeon_ring_write(ring, 0xDEADBEEF); |
174 | radeon_ring_unlock_commit(rdev); |
176 | radeon_ring_unlock_commit(rdev, ring); |
175 | } |
177 | } |
176 | 178 | ||
177 | static void r420_cp_errata_fini(struct radeon_device *rdev) |
179 | static void r420_cp_errata_fini(struct radeon_device *rdev) |
178 | { |
180 | { |
- | 181 | struct radeon_ring *ring = &rdev->ring[RADEON_RING_TYPE_GFX_INDEX]; |
|
- | 182 | ||
179 | /* Catch the RESYNC we dispatched all the way back, |
183 | /* Catch the RESYNC we dispatched all the way back, |
180 | * at the very beginning of the CP init. |
184 | * at the very beginning of the CP init. |
181 | */ |
185 | */ |
182 | radeon_ring_lock(rdev, 8); |
186 | radeon_ring_lock(rdev, ring, 8); |
183 | radeon_ring_write(rdev, PACKET0(R300_RB3D_DSTCACHE_CTLSTAT, 0)); |
187 | radeon_ring_write(ring, PACKET0(R300_RB3D_DSTCACHE_CTLSTAT, 0)); |
184 | radeon_ring_write(rdev, R300_RB3D_DC_FINISH); |
188 | radeon_ring_write(ring, R300_RB3D_DC_FINISH); |
185 | radeon_ring_unlock_commit(rdev); |
189 | radeon_ring_unlock_commit(rdev, ring); |
186 | radeon_scratch_free(rdev, rdev->config.r300.resync_scratch); |
190 | radeon_scratch_free(rdev, rdev->config.r300.resync_scratch); |
187 | } |
191 | } |
188 | 192 | ||
189 | static int r420_startup(struct radeon_device *rdev) |
193 | static int r420_startup(struct radeon_device *rdev) |
190 | { |
194 | { |
191 | int r; |
195 | int r; |
192 | 196 | ||
193 | /* set common regs */ |
197 | /* set common regs */ |
194 | r100_set_common_regs(rdev); |
198 | r100_set_common_regs(rdev); |
195 | /* program mc */ |
199 | /* program mc */ |
196 | r300_mc_program(rdev); |
200 | r300_mc_program(rdev); |
197 | /* Resume clock */ |
201 | /* Resume clock */ |
198 | r420_clock_resume(rdev); |
202 | r420_clock_resume(rdev); |
199 | /* Initialize GART (initialize after TTM so we can allocate |
203 | /* Initialize GART (initialize after TTM so we can allocate |
200 | * memory through TTM but finalize after TTM) */ |
204 | * memory through TTM but finalize after TTM) */ |
201 | if (rdev->flags & RADEON_IS_PCIE) { |
205 | if (rdev->flags & RADEON_IS_PCIE) { |
202 | r = rv370_pcie_gart_enable(rdev); |
206 | r = rv370_pcie_gart_enable(rdev); |
203 | if (r) |
207 | if (r) |
204 | return r; |
208 | return r; |
205 | } |
209 | } |
206 | if (rdev->flags & RADEON_IS_PCI) { |
210 | if (rdev->flags & RADEON_IS_PCI) { |
207 | r = r100_pci_gart_enable(rdev); |
211 | r = r100_pci_gart_enable(rdev); |
208 | if (r) |
212 | if (r) |
209 | return r; |
213 | return r; |
210 | } |
214 | } |
211 | r420_pipes_init(rdev); |
215 | r420_pipes_init(rdev); |
212 | 216 | ||
213 | /* allocate wb buffer */ |
217 | /* allocate wb buffer */ |
214 | r = radeon_wb_init(rdev); |
218 | r = radeon_wb_init(rdev); |
215 | if (r) |
219 | if (r) |
216 | return r; |
220 | return r; |
217 | 221 | ||
218 | /* Enable IRQ */ |
222 | /* Enable IRQ */ |
219 | r100_irq_set(rdev); |
223 | r100_irq_set(rdev); |
220 | rdev->config.r300.hdp_cntl = RREG32(RADEON_HOST_PATH_CNTL); |
224 | rdev->config.r300.hdp_cntl = RREG32(RADEON_HOST_PATH_CNTL); |
221 | /* 1M ring buffer */ |
225 | /* 1M ring buffer */ |
222 | r = r100_cp_init(rdev, 1024 * 1024); |
226 | r = r100_cp_init(rdev, 1024 * 1024); |
223 | if (r) { |
227 | if (r) { |
224 | dev_err(rdev->dev, "failed initializing CP (%d).\n", r); |
228 | dev_err(rdev->dev, "failed initializing CP (%d).\n", r); |
225 | return r; |
229 | return r; |
226 | } |
230 | } |
227 | r420_cp_errata_init(rdev); |
231 | r420_cp_errata_init(rdev); |
- | 232 | ||
228 | r = r100_ib_init(rdev); |
233 | r = radeon_ib_pool_init(rdev); |
229 | if (r) { |
234 | if (r) { |
230 | dev_err(rdev->dev, "failed initializing IB (%d).\n", r); |
235 | dev_err(rdev->dev, "IB initialization failed (%d).\n", r); |
231 | return r; |
236 | return r; |
232 | } |
237 | } |
- | 238 | ||
233 | return 0; |
239 | return 0; |
234 | } |
240 | } |
235 | - | ||
236 | int r420_resume(struct radeon_device *rdev) |
241 | |
237 | { |
- | |
238 | /* Make sur GART are not working */ |
- | |
239 | if (rdev->flags & RADEON_IS_PCIE) |
- | |
240 | rv370_pcie_gart_disable(rdev); |
- | |
241 | if (rdev->flags & RADEON_IS_PCI) |
- | |
242 | r100_pci_gart_disable(rdev); |
- | |
243 | /* Resume clock before doing reset */ |
- | |
244 | r420_clock_resume(rdev); |
- | |
245 | /* Reset gpu before posting otherwise ATOM will enter infinite loop */ |
- | |
246 | if (radeon_asic_reset(rdev)) { |
- | |
247 | dev_warn(rdev->dev, "GPU reset failed ! (0xE40=0x%08X, 0x7C0=0x%08X)\n", |
- | |
248 | RREG32(R_000E40_RBBM_STATUS), |
- | |
249 | RREG32(R_0007C0_CP_STAT)); |
- | |
250 | } |
- | |
251 | /* check if cards are posted or not */ |
- | |
252 | if (rdev->is_atom_bios) { |
- | |
253 | atom_asic_init(rdev->mode_info.atom_context); |
- | |
254 | } else { |
- | |
255 | radeon_combios_asic_init(rdev->ddev); |
- | |
256 | } |
- | |
257 | /* Resume clock after posting */ |
- | |
258 | r420_clock_resume(rdev); |
- | |
259 | /* Initialize surface registers */ |
- | |
260 | radeon_surface_init(rdev); |
- | |
261 | return r420_startup(rdev); |
242 | |
262 | } |
243 | |
263 | 244 | ||
264 | 245 | ||
265 | 246 | ||
266 | int r420_init(struct radeon_device *rdev) |
247 | int r420_init(struct radeon_device *rdev) |
267 | { |
248 | { |
268 | int r; |
249 | int r; |
269 | 250 | ||
270 | /* Initialize scratch registers */ |
251 | /* Initialize scratch registers */ |
271 | radeon_scratch_init(rdev); |
252 | radeon_scratch_init(rdev); |
272 | /* Initialize surface registers */ |
253 | /* Initialize surface registers */ |
273 | radeon_surface_init(rdev); |
254 | radeon_surface_init(rdev); |
274 | /* TODO: disable VGA need to use VGA request */ |
255 | /* TODO: disable VGA need to use VGA request */ |
275 | /* restore some register to sane defaults */ |
256 | /* restore some register to sane defaults */ |
276 | r100_restore_sanity(rdev); |
257 | r100_restore_sanity(rdev); |
277 | /* BIOS*/ |
258 | /* BIOS*/ |
278 | if (!radeon_get_bios(rdev)) { |
259 | if (!radeon_get_bios(rdev)) { |
279 | if (ASIC_IS_AVIVO(rdev)) |
260 | if (ASIC_IS_AVIVO(rdev)) |
280 | return -EINVAL; |
261 | return -EINVAL; |
281 | } |
262 | } |
282 | if (rdev->is_atom_bios) { |
263 | if (rdev->is_atom_bios) { |
283 | r = radeon_atombios_init(rdev); |
264 | r = radeon_atombios_init(rdev); |
284 | if (r) { |
265 | if (r) { |
285 | return r; |
266 | return r; |
286 | } |
267 | } |
287 | } else { |
268 | } else { |
288 | r = radeon_combios_init(rdev); |
269 | r = radeon_combios_init(rdev); |
289 | if (r) { |
270 | if (r) { |
290 | return r; |
271 | return r; |
291 | } |
272 | } |
292 | } |
273 | } |
293 | /* Reset gpu before posting otherwise ATOM will enter infinite loop */ |
274 | /* Reset gpu before posting otherwise ATOM will enter infinite loop */ |
294 | if (radeon_asic_reset(rdev)) { |
275 | if (radeon_asic_reset(rdev)) { |
295 | dev_warn(rdev->dev, |
276 | dev_warn(rdev->dev, |
296 | "GPU reset failed ! (0xE40=0x%08X, 0x7C0=0x%08X)\n", |
277 | "GPU reset failed ! (0xE40=0x%08X, 0x7C0=0x%08X)\n", |
297 | RREG32(R_000E40_RBBM_STATUS), |
278 | RREG32(R_000E40_RBBM_STATUS), |
298 | RREG32(R_0007C0_CP_STAT)); |
279 | RREG32(R_0007C0_CP_STAT)); |
299 | } |
280 | } |
300 | /* check if cards are posted or not */ |
281 | /* check if cards are posted or not */ |
301 | if (radeon_boot_test_post_card(rdev) == false) |
282 | if (radeon_boot_test_post_card(rdev) == false) |
302 | return -EINVAL; |
283 | return -EINVAL; |
303 | 284 | ||
304 | /* Initialize clocks */ |
285 | /* Initialize clocks */ |
305 | radeon_get_clock_info(rdev->ddev); |
286 | radeon_get_clock_info(rdev->ddev); |
306 | /* initialize AGP */ |
287 | /* initialize AGP */ |
307 | if (rdev->flags & RADEON_IS_AGP) { |
288 | if (rdev->flags & RADEON_IS_AGP) { |
308 | r = radeon_agp_init(rdev); |
289 | r = radeon_agp_init(rdev); |
309 | if (r) { |
290 | if (r) { |
310 | radeon_agp_disable(rdev); |
291 | radeon_agp_disable(rdev); |
311 | } |
292 | } |
312 | } |
293 | } |
313 | /* initialize memory controller */ |
294 | /* initialize memory controller */ |
314 | r300_mc_init(rdev); |
295 | r300_mc_init(rdev); |
315 | r420_debugfs(rdev); |
296 | r420_debugfs(rdev); |
316 | /* Fence driver */ |
297 | /* Fence driver */ |
317 | r = radeon_fence_driver_init(rdev); |
298 | r = radeon_fence_driver_init(rdev); |
318 | if (r) { |
299 | if (r) { |
319 | return r; |
300 | return r; |
320 | } |
301 | } |
321 | r = radeon_irq_kms_init(rdev); |
302 | r = radeon_irq_kms_init(rdev); |
322 | if (r) { |
303 | if (r) { |
323 | return r; |
304 | return r; |
324 | } |
305 | } |
325 | /* Memory manager */ |
306 | /* Memory manager */ |
326 | r = radeon_bo_init(rdev); |
307 | r = radeon_bo_init(rdev); |
327 | if (r) { |
308 | if (r) { |
328 | return r; |
309 | return r; |
329 | } |
310 | } |
330 | if (rdev->family == CHIP_R420) |
311 | if (rdev->family == CHIP_R420) |
331 | r100_enable_bm(rdev); |
312 | r100_enable_bm(rdev); |
332 | 313 | ||
333 | if (rdev->flags & RADEON_IS_PCIE) { |
314 | if (rdev->flags & RADEON_IS_PCIE) { |
334 | r = rv370_pcie_gart_init(rdev); |
315 | r = rv370_pcie_gart_init(rdev); |
335 | if (r) |
316 | if (r) |
336 | return r; |
317 | return r; |
337 | } |
318 | } |
338 | if (rdev->flags & RADEON_IS_PCI) { |
319 | if (rdev->flags & RADEON_IS_PCI) { |
339 | r = r100_pci_gart_init(rdev); |
320 | r = r100_pci_gart_init(rdev); |
340 | if (r) |
321 | if (r) |
341 | return r; |
322 | return r; |
342 | } |
323 | } |
343 | r420_set_reg_safe(rdev); |
324 | r420_set_reg_safe(rdev); |
- | 325 | ||
344 | rdev->accel_working = true; |
326 | rdev->accel_working = true; |
345 | r = r420_startup(rdev); |
327 | r = r420_startup(rdev); |
346 | if (r) { |
328 | if (r) { |
347 | /* Somethings want wront with the accel init stop accel */ |
329 | /* Somethings want wront with the accel init stop accel */ |
348 | dev_err(rdev->dev, "Disabling GPU acceleration\n"); |
330 | dev_err(rdev->dev, "Disabling GPU acceleration\n"); |
349 | if (rdev->flags & RADEON_IS_PCIE) |
331 | if (rdev->flags & RADEON_IS_PCIE) |
350 | rv370_pcie_gart_fini(rdev); |
332 | rv370_pcie_gart_fini(rdev); |
351 | if (rdev->flags & RADEON_IS_PCI) |
333 | if (rdev->flags & RADEON_IS_PCI) |
352 | r100_pci_gart_fini(rdev); |
334 | r100_pci_gart_fini(rdev); |
353 | rdev->accel_working = false; |
335 | rdev->accel_working = false; |
354 | } |
336 | } |
355 | return 0; |
337 | return 0; |
356 | } |
338 | } |
357 | 339 | ||
358 | /* |
340 | /* |
359 | * Debugfs info |
341 | * Debugfs info |
360 | */ |
342 | */ |
361 | #if defined(CONFIG_DEBUG_FS) |
343 | #if defined(CONFIG_DEBUG_FS) |
362 | static int r420_debugfs_pipes_info(struct seq_file *m, void *data) |
344 | static int r420_debugfs_pipes_info(struct seq_file *m, void *data) |
363 | { |
345 | { |
364 | struct drm_info_node *node = (struct drm_info_node *) m->private; |
346 | struct drm_info_node *node = (struct drm_info_node *) m->private; |
365 | struct drm_device *dev = node->minor->dev; |
347 | struct drm_device *dev = node->minor->dev; |
366 | struct radeon_device *rdev = dev->dev_private; |
348 | struct radeon_device *rdev = dev->dev_private; |
367 | uint32_t tmp; |
349 | uint32_t tmp; |
368 | 350 | ||
369 | tmp = RREG32(R400_GB_PIPE_SELECT); |
351 | tmp = RREG32(R400_GB_PIPE_SELECT); |
370 | seq_printf(m, "GB_PIPE_SELECT 0x%08x\n", tmp); |
352 | seq_printf(m, "GB_PIPE_SELECT 0x%08x\n", tmp); |
371 | tmp = RREG32(R300_GB_TILE_CONFIG); |
353 | tmp = RREG32(R300_GB_TILE_CONFIG); |
372 | seq_printf(m, "GB_TILE_CONFIG 0x%08x\n", tmp); |
354 | seq_printf(m, "GB_TILE_CONFIG 0x%08x\n", tmp); |
373 | tmp = RREG32(R300_DST_PIPE_CONFIG); |
355 | tmp = RREG32(R300_DST_PIPE_CONFIG); |
374 | seq_printf(m, "DST_PIPE_CONFIG 0x%08x\n", tmp); |
356 | seq_printf(m, "DST_PIPE_CONFIG 0x%08x\n", tmp); |
375 | return 0; |
357 | return 0; |
376 | } |
358 | } |
377 | 359 | ||
378 | static struct drm_info_list r420_pipes_info_list[] = { |
360 | static struct drm_info_list r420_pipes_info_list[] = { |
379 | {"r420_pipes_info", r420_debugfs_pipes_info, 0, NULL}, |
361 | {"r420_pipes_info", r420_debugfs_pipes_info, 0, NULL}, |
380 | }; |
362 | }; |
381 | #endif |
363 | #endif |
382 | 364 | ||
383 | int r420_debugfs_pipes_info_init(struct radeon_device *rdev) |
365 | int r420_debugfs_pipes_info_init(struct radeon_device *rdev) |
384 | { |
366 | { |
385 | #if defined(CONFIG_DEBUG_FS) |
367 | #if defined(CONFIG_DEBUG_FS) |
386 | return radeon_debugfs_add_files(rdev, r420_pipes_info_list, 1); |
368 | return radeon_debugfs_add_files(rdev, r420_pipes_info_list, 1); |
387 | #else |
369 | #else |
388 | return 0; |
370 | return 0; |
389 | #endif |
371 | #endif |
390 | }><>><>><>><>><>><>><> |
372 | }><>><>><>><>><>><>><> |