Rev 2330 | Rev 2336 | Go to most recent revision | Only display areas with differences | Regard whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 2330 | Rev 2332 | ||
---|---|---|---|
1 | /* i915_drv.h -- Private header for the I915 driver -*- linux-c -*- |
1 | /* i915_drv.h -- Private header for the I915 driver -*- linux-c -*- |
2 | */ |
2 | */ |
3 | /* |
3 | /* |
4 | * |
4 | * |
5 | * Copyright 2003 Tungsten Graphics, Inc., Cedar Park, Texas. |
5 | * Copyright 2003 Tungsten Graphics, Inc., Cedar Park, Texas. |
6 | * All Rights Reserved. |
6 | * All Rights Reserved. |
7 | * |
7 | * |
8 | * Permission is hereby granted, free of charge, to any person obtaining a |
8 | * Permission is hereby granted, free of charge, to any person obtaining a |
9 | * copy of this software and associated documentation files (the |
9 | * copy of this software and associated documentation files (the |
10 | * "Software"), to deal in the Software without restriction, including |
10 | * "Software"), to deal in the Software without restriction, including |
11 | * without limitation the rights to use, copy, modify, merge, publish, |
11 | * without limitation the rights to use, copy, modify, merge, publish, |
12 | * distribute, sub license, and/or sell copies of the Software, and to |
12 | * distribute, sub license, and/or sell copies of the Software, and to |
13 | * permit persons to whom the Software is furnished to do so, subject to |
13 | * permit persons to whom the Software is furnished to do so, subject to |
14 | * the following conditions: |
14 | * the following conditions: |
15 | * |
15 | * |
16 | * The above copyright notice and this permission notice (including the |
16 | * The above copyright notice and this permission notice (including the |
17 | * next paragraph) shall be included in all copies or substantial portions |
17 | * next paragraph) shall be included in all copies or substantial portions |
18 | * of the Software. |
18 | * of the Software. |
19 | * |
19 | * |
20 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS |
20 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS |
21 | * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
21 | * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF |
22 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. |
22 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. |
23 | * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR |
23 | * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR |
24 | * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, |
24 | * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, |
25 | * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE |
25 | * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE |
26 | * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
26 | * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
27 | * |
27 | * |
28 | */ |
28 | */ |
29 | 29 | ||
30 | #ifndef _I915_DRV_H_ |
30 | #ifndef _I915_DRV_H_ |
31 | #define _I915_DRV_H_ |
31 | #define _I915_DRV_H_ |
32 | 32 | ||
33 | #include "i915_reg.h" |
33 | #include "i915_reg.h" |
34 | #include "intel_bios.h" |
34 | #include "intel_bios.h" |
35 | #include "intel_ringbuffer.h" |
35 | #include "intel_ringbuffer.h" |
36 | //#include |
36 | //#include |
37 | #include |
37 | #include |
38 | //#include |
38 | #include |
39 | //#include |
39 | //#include |
40 | 40 | ||
41 | #include |
41 | #include |
42 | 42 | ||
43 | /* General customization: |
43 | /* General customization: |
44 | */ |
44 | */ |
45 | 45 | ||
46 | #define I915_TILING_NONE 0 |
46 | #define I915_TILING_NONE 0 |
47 | 47 | ||
48 | 48 | ||
49 | #define DRIVER_AUTHOR "Tungsten Graphics, Inc." |
49 | #define DRIVER_AUTHOR "Tungsten Graphics, Inc." |
50 | 50 | ||
51 | #define DRIVER_NAME "i915" |
51 | #define DRIVER_NAME "i915" |
52 | #define DRIVER_DESC "Intel Graphics" |
52 | #define DRIVER_DESC "Intel Graphics" |
53 | #define DRIVER_DATE "20080730" |
53 | #define DRIVER_DATE "20080730" |
54 | 54 | ||
55 | enum pipe { |
55 | enum pipe { |
56 | PIPE_A = 0, |
56 | PIPE_A = 0, |
57 | PIPE_B, |
57 | PIPE_B, |
58 | PIPE_C, |
58 | PIPE_C, |
59 | I915_MAX_PIPES |
59 | I915_MAX_PIPES |
60 | }; |
60 | }; |
61 | #define pipe_name(p) ((p) + 'A') |
61 | #define pipe_name(p) ((p) + 'A') |
62 | 62 | ||
63 | enum plane { |
63 | enum plane { |
64 | PLANE_A = 0, |
64 | PLANE_A = 0, |
65 | PLANE_B, |
65 | PLANE_B, |
66 | PLANE_C, |
66 | PLANE_C, |
67 | }; |
67 | }; |
68 | #define plane_name(p) ((p) + 'A') |
68 | #define plane_name(p) ((p) + 'A') |
69 | 69 | ||
70 | #define I915_GEM_GPU_DOMAINS (~(I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT)) |
70 | #define I915_GEM_GPU_DOMAINS (~(I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT)) |
71 | 71 | ||
72 | #define for_each_pipe(p) for ((p) = 0; (p) < dev_priv->num_pipe; (p)++) |
72 | #define for_each_pipe(p) for ((p) = 0; (p) < dev_priv->num_pipe; (p)++) |
73 | 73 | ||
74 | /* Interface history: |
74 | /* Interface history: |
75 | * |
75 | * |
76 | * 1.1: Original. |
76 | * 1.1: Original. |
77 | * 1.2: Add Power Management |
77 | * 1.2: Add Power Management |
78 | * 1.3: Add vblank support |
78 | * 1.3: Add vblank support |
79 | * 1.4: Fix cmdbuffer path, add heap destroy |
79 | * 1.4: Fix cmdbuffer path, add heap destroy |
80 | * 1.5: Add vblank pipe configuration |
80 | * 1.5: Add vblank pipe configuration |
81 | * 1.6: - New ioctl for scheduling buffer swaps on vertical blank |
81 | * 1.6: - New ioctl for scheduling buffer swaps on vertical blank |
82 | * - Support vertical blank on secondary display pipe |
82 | * - Support vertical blank on secondary display pipe |
83 | */ |
83 | */ |
84 | #define DRIVER_MAJOR 1 |
84 | #define DRIVER_MAJOR 1 |
85 | #define DRIVER_MINOR 6 |
85 | #define DRIVER_MINOR 6 |
86 | #define DRIVER_PATCHLEVEL 0 |
86 | #define DRIVER_PATCHLEVEL 0 |
87 | 87 | ||
88 | #define WATCH_COHERENCY 0 |
88 | #define WATCH_COHERENCY 0 |
89 | #define WATCH_LISTS 0 |
89 | #define WATCH_LISTS 0 |
90 | 90 | ||
91 | #define I915_GEM_PHYS_CURSOR_0 1 |
91 | #define I915_GEM_PHYS_CURSOR_0 1 |
92 | #define I915_GEM_PHYS_CURSOR_1 2 |
92 | #define I915_GEM_PHYS_CURSOR_1 2 |
93 | #define I915_GEM_PHYS_OVERLAY_REGS 3 |
93 | #define I915_GEM_PHYS_OVERLAY_REGS 3 |
94 | #define I915_MAX_PHYS_OBJECT (I915_GEM_PHYS_OVERLAY_REGS) |
94 | #define I915_MAX_PHYS_OBJECT (I915_GEM_PHYS_OVERLAY_REGS) |
95 | 95 | ||
96 | struct mem_block { |
96 | struct mem_block { |
97 | struct mem_block *next; |
97 | struct mem_block *next; |
98 | struct mem_block *prev; |
98 | struct mem_block *prev; |
99 | int start; |
99 | int start; |
100 | int size; |
100 | int size; |
101 | struct drm_file *file_priv; /* NULL: free, -1: heap, other: real files */ |
101 | struct drm_file *file_priv; /* NULL: free, -1: heap, other: real files */ |
102 | }; |
102 | }; |
103 | 103 | ||
104 | struct opregion_header; |
104 | struct opregion_header; |
105 | struct opregion_acpi; |
105 | struct opregion_acpi; |
106 | struct opregion_swsci; |
106 | struct opregion_swsci; |
107 | struct opregion_asle; |
107 | struct opregion_asle; |
108 | 108 | ||
109 | struct intel_opregion { |
109 | struct intel_opregion { |
110 | struct opregion_header *header; |
110 | struct opregion_header *header; |
111 | struct opregion_acpi *acpi; |
111 | struct opregion_acpi *acpi; |
112 | struct opregion_swsci *swsci; |
112 | struct opregion_swsci *swsci; |
113 | struct opregion_asle *asle; |
113 | struct opregion_asle *asle; |
114 | void *vbt; |
114 | void *vbt; |
115 | u32 __iomem *lid_state; |
115 | u32 __iomem *lid_state; |
116 | }; |
116 | }; |
117 | #define OPREGION_SIZE (8*1024) |
117 | #define OPREGION_SIZE (8*1024) |
118 | 118 | ||
119 | struct intel_overlay; |
119 | struct intel_overlay; |
120 | struct intel_overlay_error_state; |
120 | struct intel_overlay_error_state; |
121 | 121 | ||
122 | struct drm_i915_master_private { |
122 | struct drm_i915_master_private { |
123 | drm_local_map_t *sarea; |
123 | drm_local_map_t *sarea; |
124 | struct _drm_i915_sarea *sarea_priv; |
124 | struct _drm_i915_sarea *sarea_priv; |
125 | }; |
125 | }; |
126 | #define I915_FENCE_REG_NONE -1 |
126 | #define I915_FENCE_REG_NONE -1 |
127 | 127 | ||
128 | struct drm_i915_fence_reg { |
128 | struct drm_i915_fence_reg { |
129 | struct list_head lru_list; |
129 | struct list_head lru_list; |
130 | struct drm_i915_gem_object *obj; |
130 | struct drm_i915_gem_object *obj; |
131 | uint32_t setup_seqno; |
131 | uint32_t setup_seqno; |
132 | }; |
132 | }; |
133 | 133 | ||
134 | struct sdvo_device_mapping { |
134 | struct sdvo_device_mapping { |
135 | u8 initialized; |
135 | u8 initialized; |
136 | u8 dvo_port; |
136 | u8 dvo_port; |
137 | u8 slave_addr; |
137 | u8 slave_addr; |
138 | u8 dvo_wiring; |
138 | u8 dvo_wiring; |
139 | u8 i2c_pin; |
139 | u8 i2c_pin; |
140 | u8 i2c_speed; |
140 | u8 i2c_speed; |
141 | u8 ddc_pin; |
141 | u8 ddc_pin; |
142 | }; |
142 | }; |
143 | 143 | ||
144 | struct intel_display_error_state; |
144 | struct intel_display_error_state; |
145 | 145 | ||
146 | struct drm_i915_error_state { |
146 | struct drm_i915_error_state { |
147 | u32 eir; |
147 | u32 eir; |
148 | u32 pgtbl_er; |
148 | u32 pgtbl_er; |
149 | u32 pipestat[I915_MAX_PIPES]; |
149 | u32 pipestat[I915_MAX_PIPES]; |
150 | u32 ipeir; |
150 | u32 ipeir; |
151 | u32 ipehr; |
151 | u32 ipehr; |
152 | u32 instdone; |
152 | u32 instdone; |
153 | u32 acthd; |
153 | u32 acthd; |
154 | u32 error; /* gen6+ */ |
154 | u32 error; /* gen6+ */ |
155 | u32 bcs_acthd; /* gen6+ blt engine */ |
155 | u32 bcs_acthd; /* gen6+ blt engine */ |
156 | u32 bcs_ipehr; |
156 | u32 bcs_ipehr; |
157 | u32 bcs_ipeir; |
157 | u32 bcs_ipeir; |
158 | u32 bcs_instdone; |
158 | u32 bcs_instdone; |
159 | u32 bcs_seqno; |
159 | u32 bcs_seqno; |
160 | u32 vcs_acthd; /* gen6+ bsd engine */ |
160 | u32 vcs_acthd; /* gen6+ bsd engine */ |
161 | u32 vcs_ipehr; |
161 | u32 vcs_ipehr; |
162 | u32 vcs_ipeir; |
162 | u32 vcs_ipeir; |
163 | u32 vcs_instdone; |
163 | u32 vcs_instdone; |
164 | u32 vcs_seqno; |
164 | u32 vcs_seqno; |
165 | u32 instpm; |
165 | u32 instpm; |
166 | u32 instps; |
166 | u32 instps; |
167 | u32 instdone1; |
167 | u32 instdone1; |
168 | u32 seqno; |
168 | u32 seqno; |
169 | u64 bbaddr; |
169 | u64 bbaddr; |
170 | u64 fence[16]; |
170 | u64 fence[16]; |
171 | struct timeval time; |
171 | struct timeval time; |
172 | struct drm_i915_error_object { |
172 | struct drm_i915_error_object { |
173 | int page_count; |
173 | int page_count; |
174 | u32 gtt_offset; |
174 | u32 gtt_offset; |
175 | u32 *pages[0]; |
175 | u32 *pages[0]; |
176 | } *ringbuffer[I915_NUM_RINGS], *batchbuffer[I915_NUM_RINGS]; |
176 | } *ringbuffer[I915_NUM_RINGS], *batchbuffer[I915_NUM_RINGS]; |
177 | struct drm_i915_error_buffer { |
177 | struct drm_i915_error_buffer { |
178 | u32 size; |
178 | u32 size; |
179 | u32 name; |
179 | u32 name; |
180 | u32 seqno; |
180 | u32 seqno; |
181 | u32 gtt_offset; |
181 | u32 gtt_offset; |
182 | u32 read_domains; |
182 | u32 read_domains; |
183 | u32 write_domain; |
183 | u32 write_domain; |
184 | s32 fence_reg:5; |
184 | s32 fence_reg:5; |
185 | s32 pinned:2; |
185 | s32 pinned:2; |
186 | u32 tiling:2; |
186 | u32 tiling:2; |
187 | u32 dirty:1; |
187 | u32 dirty:1; |
188 | u32 purgeable:1; |
188 | u32 purgeable:1; |
189 | u32 ring:4; |
189 | u32 ring:4; |
190 | u32 cache_level:2; |
190 | u32 cache_level:2; |
191 | } *active_bo, *pinned_bo; |
191 | } *active_bo, *pinned_bo; |
192 | u32 active_bo_count, pinned_bo_count; |
192 | u32 active_bo_count, pinned_bo_count; |
193 | struct intel_overlay_error_state *overlay; |
193 | struct intel_overlay_error_state *overlay; |
194 | struct intel_display_error_state *display; |
194 | struct intel_display_error_state *display; |
195 | }; |
195 | }; |
196 | 196 | ||
197 | struct drm_i915_display_funcs { |
197 | struct drm_i915_display_funcs { |
198 | void (*dpms)(struct drm_crtc *crtc, int mode); |
198 | void (*dpms)(struct drm_crtc *crtc, int mode); |
199 | bool (*fbc_enabled)(struct drm_device *dev); |
199 | bool (*fbc_enabled)(struct drm_device *dev); |
200 | void (*enable_fbc)(struct drm_crtc *crtc, unsigned long interval); |
200 | void (*enable_fbc)(struct drm_crtc *crtc, unsigned long interval); |
201 | void (*disable_fbc)(struct drm_device *dev); |
201 | void (*disable_fbc)(struct drm_device *dev); |
202 | int (*get_display_clock_speed)(struct drm_device *dev); |
202 | int (*get_display_clock_speed)(struct drm_device *dev); |
203 | int (*get_fifo_size)(struct drm_device *dev, int plane); |
203 | int (*get_fifo_size)(struct drm_device *dev, int plane); |
204 | void (*update_wm)(struct drm_device *dev); |
204 | void (*update_wm)(struct drm_device *dev); |
205 | int (*crtc_mode_set)(struct drm_crtc *crtc, |
205 | int (*crtc_mode_set)(struct drm_crtc *crtc, |
206 | struct drm_display_mode *mode, |
206 | struct drm_display_mode *mode, |
207 | struct drm_display_mode *adjusted_mode, |
207 | struct drm_display_mode *adjusted_mode, |
208 | int x, int y, |
208 | int x, int y, |
209 | struct drm_framebuffer *old_fb); |
209 | struct drm_framebuffer *old_fb); |
210 | void (*fdi_link_train)(struct drm_crtc *crtc); |
210 | void (*fdi_link_train)(struct drm_crtc *crtc); |
211 | void (*init_clock_gating)(struct drm_device *dev); |
211 | void (*init_clock_gating)(struct drm_device *dev); |
212 | void (*init_pch_clock_gating)(struct drm_device *dev); |
212 | void (*init_pch_clock_gating)(struct drm_device *dev); |
213 | int (*queue_flip)(struct drm_device *dev, struct drm_crtc *crtc, |
213 | int (*queue_flip)(struct drm_device *dev, struct drm_crtc *crtc, |
214 | struct drm_framebuffer *fb, |
214 | struct drm_framebuffer *fb, |
215 | struct drm_i915_gem_object *obj); |
215 | struct drm_i915_gem_object *obj); |
216 | int (*update_plane)(struct drm_crtc *crtc, struct drm_framebuffer *fb, |
216 | int (*update_plane)(struct drm_crtc *crtc, struct drm_framebuffer *fb, |
217 | int x, int y); |
217 | int x, int y); |
218 | /* clock updates for mode set */ |
218 | /* clock updates for mode set */ |
219 | /* cursor updates */ |
219 | /* cursor updates */ |
220 | /* render clock increase/decrease */ |
220 | /* render clock increase/decrease */ |
221 | /* display clock increase/decrease */ |
221 | /* display clock increase/decrease */ |
222 | /* pll clock increase/decrease */ |
222 | /* pll clock increase/decrease */ |
223 | }; |
223 | }; |
224 | 224 | ||
225 | struct intel_device_info { |
225 | struct intel_device_info { |
226 | u8 gen; |
226 | u8 gen; |
227 | u8 is_mobile : 1; |
227 | u8 is_mobile : 1; |
228 | u8 is_i85x : 1; |
228 | u8 is_i85x : 1; |
229 | u8 is_i915g : 1; |
229 | u8 is_i915g : 1; |
230 | u8 is_i945gm : 1; |
230 | u8 is_i945gm : 1; |
231 | u8 is_g33 : 1; |
231 | u8 is_g33 : 1; |
232 | u8 need_gfx_hws : 1; |
232 | u8 need_gfx_hws : 1; |
233 | u8 is_g4x : 1; |
233 | u8 is_g4x : 1; |
234 | u8 is_pineview : 1; |
234 | u8 is_pineview : 1; |
235 | u8 is_broadwater : 1; |
235 | u8 is_broadwater : 1; |
236 | u8 is_crestline : 1; |
236 | u8 is_crestline : 1; |
237 | u8 is_ivybridge : 1; |
237 | u8 is_ivybridge : 1; |
238 | u8 has_fbc : 1; |
238 | u8 has_fbc : 1; |
239 | u8 has_pipe_cxsr : 1; |
239 | u8 has_pipe_cxsr : 1; |
240 | u8 has_hotplug : 1; |
240 | u8 has_hotplug : 1; |
241 | u8 cursor_needs_physical : 1; |
241 | u8 cursor_needs_physical : 1; |
242 | u8 has_overlay : 1; |
242 | u8 has_overlay : 1; |
243 | u8 overlay_needs_physical : 1; |
243 | u8 overlay_needs_physical : 1; |
244 | u8 supports_tv : 1; |
244 | u8 supports_tv : 1; |
245 | u8 has_bsd_ring : 1; |
245 | u8 has_bsd_ring : 1; |
246 | u8 has_blt_ring : 1; |
246 | u8 has_blt_ring : 1; |
247 | }; |
247 | }; |
248 | 248 | ||
249 | enum no_fbc_reason { |
249 | enum no_fbc_reason { |
250 | FBC_NO_OUTPUT, /* no outputs enabled to compress */ |
250 | FBC_NO_OUTPUT, /* no outputs enabled to compress */ |
251 | FBC_STOLEN_TOO_SMALL, /* not enough space to hold compressed buffers */ |
251 | FBC_STOLEN_TOO_SMALL, /* not enough space to hold compressed buffers */ |
252 | FBC_UNSUPPORTED_MODE, /* interlace or doublescanned mode */ |
252 | FBC_UNSUPPORTED_MODE, /* interlace or doublescanned mode */ |
253 | FBC_MODE_TOO_LARGE, /* mode too large for compression */ |
253 | FBC_MODE_TOO_LARGE, /* mode too large for compression */ |
254 | FBC_BAD_PLANE, /* fbc not supported on plane */ |
254 | FBC_BAD_PLANE, /* fbc not supported on plane */ |
255 | FBC_NOT_TILED, /* buffer not tiled */ |
255 | FBC_NOT_TILED, /* buffer not tiled */ |
256 | FBC_MULTIPLE_PIPES, /* more than one pipe active */ |
256 | FBC_MULTIPLE_PIPES, /* more than one pipe active */ |
257 | FBC_MODULE_PARAM, |
257 | FBC_MODULE_PARAM, |
258 | }; |
258 | }; |
259 | 259 | ||
260 | enum intel_pch { |
260 | enum intel_pch { |
261 | PCH_IBX, /* Ibexpeak PCH */ |
261 | PCH_IBX, /* Ibexpeak PCH */ |
262 | PCH_CPT, /* Cougarpoint PCH */ |
262 | PCH_CPT, /* Cougarpoint PCH */ |
263 | }; |
263 | }; |
264 | 264 | ||
265 | #define QUIRK_PIPEA_FORCE (1<<0) |
265 | #define QUIRK_PIPEA_FORCE (1<<0) |
266 | #define QUIRK_LVDS_SSC_DISABLE (1<<1) |
266 | #define QUIRK_LVDS_SSC_DISABLE (1<<1) |
267 | 267 | ||
268 | struct intel_fbdev; |
268 | struct intel_fbdev; |
269 | struct intel_fbc_work; |
269 | struct intel_fbc_work; |
270 | 270 | ||
271 | 271 | ||
272 | typedef struct drm_i915_private { |
272 | typedef struct drm_i915_private { |
273 | struct drm_device *dev; |
273 | struct drm_device *dev; |
274 | 274 | ||
275 | const struct intel_device_info *info; |
275 | const struct intel_device_info *info; |
276 | 276 | ||
277 | int has_gem; |
277 | int has_gem; |
278 | int relative_constants_mode; |
278 | int relative_constants_mode; |
279 | 279 | ||
280 | void __iomem *regs; |
280 | void __iomem *regs; |
281 | u32 gt_fifo_count; |
281 | u32 gt_fifo_count; |
282 | 282 | ||
283 | struct intel_gmbus { |
283 | struct intel_gmbus { |
284 | struct i2c_adapter adapter; |
284 | struct i2c_adapter adapter; |
285 | struct i2c_adapter *force_bit; |
285 | struct i2c_adapter *force_bit; |
286 | u32 reg0; |
286 | u32 reg0; |
287 | } *gmbus; |
287 | } *gmbus; |
288 | 288 | ||
289 | struct pci_dev *bridge_dev; |
289 | struct pci_dev *bridge_dev; |
290 | struct intel_ring_buffer ring[I915_NUM_RINGS]; |
290 | struct intel_ring_buffer ring[I915_NUM_RINGS]; |
291 | uint32_t next_seqno; |
291 | uint32_t next_seqno; |
292 | 292 | ||
293 | drm_dma_handle_t *status_page_dmah; |
293 | drm_dma_handle_t *status_page_dmah; |
294 | // uint32_t counter; |
294 | // uint32_t counter; |
295 | // drm_local_map_t hws_map; |
295 | // drm_local_map_t hws_map; |
296 | // struct drm_i915_gem_object *pwrctx; |
296 | struct drm_i915_gem_object *pwrctx; |
297 | // struct drm_i915_gem_object *renderctx; |
297 | struct drm_i915_gem_object *renderctx; |
298 | 298 | ||
299 | // struct resource mch_res; |
299 | // struct resource mch_res; |
300 | 300 | ||
301 | unsigned int cpp; |
301 | unsigned int cpp; |
302 | int back_offset; |
302 | int back_offset; |
303 | int front_offset; |
303 | int front_offset; |
304 | int current_page; |
304 | int current_page; |
305 | int page_flipping; |
305 | int page_flipping; |
306 | 306 | ||
307 | atomic_t irq_received; |
307 | atomic_t irq_received; |
308 | 308 | ||
309 | /* protects the irq masks */ |
309 | /* protects the irq masks */ |
310 | spinlock_t irq_lock; |
310 | spinlock_t irq_lock; |
311 | /** Cached value of IMR to avoid reads in updating the bitfield */ |
311 | /** Cached value of IMR to avoid reads in updating the bitfield */ |
312 | u32 pipestat[2]; |
312 | u32 pipestat[2]; |
313 | u32 irq_mask; |
313 | u32 irq_mask; |
314 | u32 gt_irq_mask; |
314 | u32 gt_irq_mask; |
315 | u32 pch_irq_mask; |
315 | u32 pch_irq_mask; |
316 | 316 | ||
317 | u32 hotplug_supported_mask; |
317 | u32 hotplug_supported_mask; |
318 | // struct work_struct hotplug_work; |
318 | // struct work_struct hotplug_work; |
319 | 319 | ||
320 | int tex_lru_log_granularity; |
320 | int tex_lru_log_granularity; |
321 | int allow_batchbuffer; |
321 | int allow_batchbuffer; |
322 | struct mem_block *agp_heap; |
322 | struct mem_block *agp_heap; |
323 | unsigned int sr01, adpa, ppcr, dvob, dvoc, lvds; |
323 | unsigned int sr01, adpa, ppcr, dvob, dvoc, lvds; |
324 | int vblank_pipe; |
324 | int vblank_pipe; |
325 | int num_pipe; |
325 | int num_pipe; |
326 | 326 | ||
327 | /* For hangcheck timer */ |
327 | /* For hangcheck timer */ |
328 | #define DRM_I915_HANGCHECK_PERIOD 1500 /* in ms */ |
328 | #define DRM_I915_HANGCHECK_PERIOD 1500 /* in ms */ |
329 | struct timer_list hangcheck_timer; |
329 | struct timer_list hangcheck_timer; |
330 | int hangcheck_count; |
330 | int hangcheck_count; |
331 | uint32_t last_acthd; |
331 | uint32_t last_acthd; |
332 | uint32_t last_instdone; |
332 | uint32_t last_instdone; |
333 | uint32_t last_instdone1; |
333 | uint32_t last_instdone1; |
334 | 334 | ||
335 | unsigned long cfb_size; |
335 | unsigned long cfb_size; |
336 | unsigned int cfb_fb; |
336 | unsigned int cfb_fb; |
337 | enum plane cfb_plane; |
337 | enum plane cfb_plane; |
338 | int cfb_y; |
338 | int cfb_y; |
339 | // struct intel_fbc_work *fbc_work; |
339 | // struct intel_fbc_work *fbc_work; |
340 | 340 | ||
341 | struct intel_opregion opregion; |
341 | struct intel_opregion opregion; |
342 | 342 | ||
343 | /* overlay */ |
343 | /* overlay */ |
344 | // struct intel_overlay *overlay; |
344 | // struct intel_overlay *overlay; |
345 | 345 | ||
346 | /* LVDS info */ |
346 | /* LVDS info */ |
347 | int backlight_level; /* restore backlight to this value */ |
347 | int backlight_level; /* restore backlight to this value */ |
348 | bool backlight_enabled; |
348 | bool backlight_enabled; |
349 | struct drm_display_mode *panel_fixed_mode; |
349 | struct drm_display_mode *panel_fixed_mode; |
350 | struct drm_display_mode *lfp_lvds_vbt_mode; /* if any */ |
350 | struct drm_display_mode *lfp_lvds_vbt_mode; /* if any */ |
351 | struct drm_display_mode *sdvo_lvds_vbt_mode; /* if any */ |
351 | struct drm_display_mode *sdvo_lvds_vbt_mode; /* if any */ |
352 | 352 | ||
353 | /* Feature bits from the VBIOS */ |
353 | /* Feature bits from the VBIOS */ |
354 | unsigned int int_tv_support:1; |
354 | unsigned int int_tv_support:1; |
355 | unsigned int lvds_dither:1; |
355 | unsigned int lvds_dither:1; |
356 | unsigned int lvds_vbt:1; |
356 | unsigned int lvds_vbt:1; |
357 | unsigned int int_crt_support:1; |
357 | unsigned int int_crt_support:1; |
358 | unsigned int lvds_use_ssc:1; |
358 | unsigned int lvds_use_ssc:1; |
359 | int lvds_ssc_freq; |
359 | int lvds_ssc_freq; |
360 | struct { |
360 | struct { |
361 | int rate; |
361 | int rate; |
362 | int lanes; |
362 | int lanes; |
363 | int preemphasis; |
363 | int preemphasis; |
364 | int vswing; |
364 | int vswing; |
365 | 365 | ||
366 | bool initialized; |
366 | bool initialized; |
367 | bool support; |
367 | bool support; |
368 | int bpp; |
368 | int bpp; |
369 | struct edp_power_seq pps; |
369 | struct edp_power_seq pps; |
370 | } edp; |
370 | } edp; |
371 | bool no_aux_handshake; |
371 | bool no_aux_handshake; |
372 | 372 | ||
373 | // struct notifier_block lid_notifier; |
373 | // struct notifier_block lid_notifier; |
374 | 374 | ||
375 | int crt_ddc_pin; |
375 | int crt_ddc_pin; |
376 | struct drm_i915_fence_reg fence_regs[16]; /* assume 965 */ |
376 | struct drm_i915_fence_reg fence_regs[16]; /* assume 965 */ |
377 | int fence_reg_start; /* 4 if userland hasn't ioctl'd us yet */ |
377 | int fence_reg_start; /* 4 if userland hasn't ioctl'd us yet */ |
378 | int num_fence_regs; /* 8 on pre-965, 16 otherwise */ |
378 | int num_fence_regs; /* 8 on pre-965, 16 otherwise */ |
379 | 379 | ||
380 | unsigned int fsb_freq, mem_freq, is_ddr3; |
380 | unsigned int fsb_freq, mem_freq, is_ddr3; |
381 | 381 | ||
382 | spinlock_t error_lock; |
382 | spinlock_t error_lock; |
383 | // struct drm_i915_error_state *first_error; |
383 | // struct drm_i915_error_state *first_error; |
384 | // struct work_struct error_work; |
384 | // struct work_struct error_work; |
385 | // struct completion error_completion; |
385 | // struct completion error_completion; |
386 | // struct workqueue_struct *wq; |
386 | // struct workqueue_struct *wq; |
387 | 387 | ||
388 | /* Display functions */ |
388 | /* Display functions */ |
389 | struct drm_i915_display_funcs display; |
389 | struct drm_i915_display_funcs display; |
390 | 390 | ||
391 | /* PCH chipset type */ |
391 | /* PCH chipset type */ |
392 | enum intel_pch pch_type; |
392 | enum intel_pch pch_type; |
393 | 393 | ||
394 | unsigned long quirks; |
394 | unsigned long quirks; |
395 | 395 | ||
396 | /* Register state */ |
396 | /* Register state */ |
397 | bool modeset_on_lid; |
397 | bool modeset_on_lid; |
398 | u8 saveLBB; |
398 | u8 saveLBB; |
399 | u32 saveDSPACNTR; |
399 | u32 saveDSPACNTR; |
400 | u32 saveDSPBCNTR; |
400 | u32 saveDSPBCNTR; |
401 | u32 saveDSPARB; |
401 | u32 saveDSPARB; |
402 | u32 saveHWS; |
402 | u32 saveHWS; |
403 | u32 savePIPEACONF; |
403 | u32 savePIPEACONF; |
404 | u32 savePIPEBCONF; |
404 | u32 savePIPEBCONF; |
405 | u32 savePIPEASRC; |
405 | u32 savePIPEASRC; |
406 | u32 savePIPEBSRC; |
406 | u32 savePIPEBSRC; |
407 | u32 saveFPA0; |
407 | u32 saveFPA0; |
408 | u32 saveFPA1; |
408 | u32 saveFPA1; |
409 | u32 saveDPLL_A; |
409 | u32 saveDPLL_A; |
410 | u32 saveDPLL_A_MD; |
410 | u32 saveDPLL_A_MD; |
411 | u32 saveHTOTAL_A; |
411 | u32 saveHTOTAL_A; |
412 | u32 saveHBLANK_A; |
412 | u32 saveHBLANK_A; |
413 | u32 saveHSYNC_A; |
413 | u32 saveHSYNC_A; |
414 | u32 saveVTOTAL_A; |
414 | u32 saveVTOTAL_A; |
415 | u32 saveVBLANK_A; |
415 | u32 saveVBLANK_A; |
416 | u32 saveVSYNC_A; |
416 | u32 saveVSYNC_A; |
417 | u32 saveBCLRPAT_A; |
417 | u32 saveBCLRPAT_A; |
418 | u32 saveTRANSACONF; |
418 | u32 saveTRANSACONF; |
419 | u32 saveTRANS_HTOTAL_A; |
419 | u32 saveTRANS_HTOTAL_A; |
420 | u32 saveTRANS_HBLANK_A; |
420 | u32 saveTRANS_HBLANK_A; |
421 | u32 saveTRANS_HSYNC_A; |
421 | u32 saveTRANS_HSYNC_A; |
422 | u32 saveTRANS_VTOTAL_A; |
422 | u32 saveTRANS_VTOTAL_A; |
423 | u32 saveTRANS_VBLANK_A; |
423 | u32 saveTRANS_VBLANK_A; |
424 | u32 saveTRANS_VSYNC_A; |
424 | u32 saveTRANS_VSYNC_A; |
425 | u32 savePIPEASTAT; |
425 | u32 savePIPEASTAT; |
426 | u32 saveDSPASTRIDE; |
426 | u32 saveDSPASTRIDE; |
427 | u32 saveDSPASIZE; |
427 | u32 saveDSPASIZE; |
428 | u32 saveDSPAPOS; |
428 | u32 saveDSPAPOS; |
429 | u32 saveDSPAADDR; |
429 | u32 saveDSPAADDR; |
430 | u32 saveDSPASURF; |
430 | u32 saveDSPASURF; |
431 | u32 saveDSPATILEOFF; |
431 | u32 saveDSPATILEOFF; |
432 | u32 savePFIT_PGM_RATIOS; |
432 | u32 savePFIT_PGM_RATIOS; |
433 | u32 saveBLC_HIST_CTL; |
433 | u32 saveBLC_HIST_CTL; |
434 | u32 saveBLC_PWM_CTL; |
434 | u32 saveBLC_PWM_CTL; |
435 | u32 saveBLC_PWM_CTL2; |
435 | u32 saveBLC_PWM_CTL2; |
436 | u32 saveBLC_CPU_PWM_CTL; |
436 | u32 saveBLC_CPU_PWM_CTL; |
437 | u32 saveBLC_CPU_PWM_CTL2; |
437 | u32 saveBLC_CPU_PWM_CTL2; |
438 | u32 saveFPB0; |
438 | u32 saveFPB0; |
439 | u32 saveFPB1; |
439 | u32 saveFPB1; |
440 | u32 saveDPLL_B; |
440 | u32 saveDPLL_B; |
441 | u32 saveDPLL_B_MD; |
441 | u32 saveDPLL_B_MD; |
442 | u32 saveHTOTAL_B; |
442 | u32 saveHTOTAL_B; |
443 | u32 saveHBLANK_B; |
443 | u32 saveHBLANK_B; |
444 | u32 saveHSYNC_B; |
444 | u32 saveHSYNC_B; |
445 | u32 saveVTOTAL_B; |
445 | u32 saveVTOTAL_B; |
446 | u32 saveVBLANK_B; |
446 | u32 saveVBLANK_B; |
447 | u32 saveVSYNC_B; |
447 | u32 saveVSYNC_B; |
448 | u32 saveBCLRPAT_B; |
448 | u32 saveBCLRPAT_B; |
449 | u32 saveTRANSBCONF; |
449 | u32 saveTRANSBCONF; |
450 | u32 saveTRANS_HTOTAL_B; |
450 | u32 saveTRANS_HTOTAL_B; |
451 | u32 saveTRANS_HBLANK_B; |
451 | u32 saveTRANS_HBLANK_B; |
452 | u32 saveTRANS_HSYNC_B; |
452 | u32 saveTRANS_HSYNC_B; |
453 | u32 saveTRANS_VTOTAL_B; |
453 | u32 saveTRANS_VTOTAL_B; |
454 | u32 saveTRANS_VBLANK_B; |
454 | u32 saveTRANS_VBLANK_B; |
455 | u32 saveTRANS_VSYNC_B; |
455 | u32 saveTRANS_VSYNC_B; |
456 | u32 savePIPEBSTAT; |
456 | u32 savePIPEBSTAT; |
457 | u32 saveDSPBSTRIDE; |
457 | u32 saveDSPBSTRIDE; |
458 | u32 saveDSPBSIZE; |
458 | u32 saveDSPBSIZE; |
459 | u32 saveDSPBPOS; |
459 | u32 saveDSPBPOS; |
460 | u32 saveDSPBADDR; |
460 | u32 saveDSPBADDR; |
461 | u32 saveDSPBSURF; |
461 | u32 saveDSPBSURF; |
462 | u32 saveDSPBTILEOFF; |
462 | u32 saveDSPBTILEOFF; |
463 | u32 saveVGA0; |
463 | u32 saveVGA0; |
464 | u32 saveVGA1; |
464 | u32 saveVGA1; |
465 | u32 saveVGA_PD; |
465 | u32 saveVGA_PD; |
466 | u32 saveVGACNTRL; |
466 | u32 saveVGACNTRL; |
467 | u32 saveADPA; |
467 | u32 saveADPA; |
468 | u32 saveLVDS; |
468 | u32 saveLVDS; |
469 | u32 savePP_ON_DELAYS; |
469 | u32 savePP_ON_DELAYS; |
470 | u32 savePP_OFF_DELAYS; |
470 | u32 savePP_OFF_DELAYS; |
471 | u32 saveDVOA; |
471 | u32 saveDVOA; |
472 | u32 saveDVOB; |
472 | u32 saveDVOB; |
473 | u32 saveDVOC; |
473 | u32 saveDVOC; |
474 | u32 savePP_ON; |
474 | u32 savePP_ON; |
475 | u32 savePP_OFF; |
475 | u32 savePP_OFF; |
476 | u32 savePP_CONTROL; |
476 | u32 savePP_CONTROL; |
477 | u32 savePP_DIVISOR; |
477 | u32 savePP_DIVISOR; |
478 | u32 savePFIT_CONTROL; |
478 | u32 savePFIT_CONTROL; |
479 | u32 save_palette_a[256]; |
479 | u32 save_palette_a[256]; |
480 | u32 save_palette_b[256]; |
480 | u32 save_palette_b[256]; |
481 | u32 saveDPFC_CB_BASE; |
481 | u32 saveDPFC_CB_BASE; |
482 | u32 saveFBC_CFB_BASE; |
482 | u32 saveFBC_CFB_BASE; |
483 | u32 saveFBC_LL_BASE; |
483 | u32 saveFBC_LL_BASE; |
484 | u32 saveFBC_CONTROL; |
484 | u32 saveFBC_CONTROL; |
485 | u32 saveFBC_CONTROL2; |
485 | u32 saveFBC_CONTROL2; |
486 | u32 saveIER; |
486 | u32 saveIER; |
487 | u32 saveIIR; |
487 | u32 saveIIR; |
488 | u32 saveIMR; |
488 | u32 saveIMR; |
489 | u32 saveDEIER; |
489 | u32 saveDEIER; |
490 | u32 saveDEIMR; |
490 | u32 saveDEIMR; |
491 | u32 saveGTIER; |
491 | u32 saveGTIER; |
492 | u32 saveGTIMR; |
492 | u32 saveGTIMR; |
493 | u32 saveFDI_RXA_IMR; |
493 | u32 saveFDI_RXA_IMR; |
494 | u32 saveFDI_RXB_IMR; |
494 | u32 saveFDI_RXB_IMR; |
495 | u32 saveCACHE_MODE_0; |
495 | u32 saveCACHE_MODE_0; |
496 | u32 saveMI_ARB_STATE; |
496 | u32 saveMI_ARB_STATE; |
497 | u32 saveSWF0[16]; |
497 | u32 saveSWF0[16]; |
498 | u32 saveSWF1[16]; |
498 | u32 saveSWF1[16]; |
499 | u32 saveSWF2[3]; |
499 | u32 saveSWF2[3]; |
500 | u8 saveMSR; |
500 | u8 saveMSR; |
501 | u8 saveSR[8]; |
501 | u8 saveSR[8]; |
502 | u8 saveGR[25]; |
502 | u8 saveGR[25]; |
503 | u8 saveAR_INDEX; |
503 | u8 saveAR_INDEX; |
504 | u8 saveAR[21]; |
504 | u8 saveAR[21]; |
505 | u8 saveDACMASK; |
505 | u8 saveDACMASK; |
506 | u8 saveCR[37]; |
506 | u8 saveCR[37]; |
507 | uint64_t saveFENCE[16]; |
507 | uint64_t saveFENCE[16]; |
508 | u32 saveCURACNTR; |
508 | u32 saveCURACNTR; |
509 | u32 saveCURAPOS; |
509 | u32 saveCURAPOS; |
510 | u32 saveCURABASE; |
510 | u32 saveCURABASE; |
511 | u32 saveCURBCNTR; |
511 | u32 saveCURBCNTR; |
512 | u32 saveCURBPOS; |
512 | u32 saveCURBPOS; |
513 | u32 saveCURBBASE; |
513 | u32 saveCURBBASE; |
514 | u32 saveCURSIZE; |
514 | u32 saveCURSIZE; |
515 | u32 saveDP_B; |
515 | u32 saveDP_B; |
516 | u32 saveDP_C; |
516 | u32 saveDP_C; |
517 | u32 saveDP_D; |
517 | u32 saveDP_D; |
518 | u32 savePIPEA_GMCH_DATA_M; |
518 | u32 savePIPEA_GMCH_DATA_M; |
519 | u32 savePIPEB_GMCH_DATA_M; |
519 | u32 savePIPEB_GMCH_DATA_M; |
520 | u32 savePIPEA_GMCH_DATA_N; |
520 | u32 savePIPEA_GMCH_DATA_N; |
521 | u32 savePIPEB_GMCH_DATA_N; |
521 | u32 savePIPEB_GMCH_DATA_N; |
522 | u32 savePIPEA_DP_LINK_M; |
522 | u32 savePIPEA_DP_LINK_M; |
523 | u32 savePIPEB_DP_LINK_M; |
523 | u32 savePIPEB_DP_LINK_M; |
524 | u32 savePIPEA_DP_LINK_N; |
524 | u32 savePIPEA_DP_LINK_N; |
525 | u32 savePIPEB_DP_LINK_N; |
525 | u32 savePIPEB_DP_LINK_N; |
526 | u32 saveFDI_RXA_CTL; |
526 | u32 saveFDI_RXA_CTL; |
527 | u32 saveFDI_TXA_CTL; |
527 | u32 saveFDI_TXA_CTL; |
528 | u32 saveFDI_RXB_CTL; |
528 | u32 saveFDI_RXB_CTL; |
529 | u32 saveFDI_TXB_CTL; |
529 | u32 saveFDI_TXB_CTL; |
530 | u32 savePFA_CTL_1; |
530 | u32 savePFA_CTL_1; |
531 | u32 savePFB_CTL_1; |
531 | u32 savePFB_CTL_1; |
532 | u32 savePFA_WIN_SZ; |
532 | u32 savePFA_WIN_SZ; |
533 | u32 savePFB_WIN_SZ; |
533 | u32 savePFB_WIN_SZ; |
534 | u32 savePFA_WIN_POS; |
534 | u32 savePFA_WIN_POS; |
535 | u32 savePFB_WIN_POS; |
535 | u32 savePFB_WIN_POS; |
536 | u32 savePCH_DREF_CONTROL; |
536 | u32 savePCH_DREF_CONTROL; |
537 | u32 saveDISP_ARB_CTL; |
537 | u32 saveDISP_ARB_CTL; |
538 | u32 savePIPEA_DATA_M1; |
538 | u32 savePIPEA_DATA_M1; |
539 | u32 savePIPEA_DATA_N1; |
539 | u32 savePIPEA_DATA_N1; |
540 | u32 savePIPEA_LINK_M1; |
540 | u32 savePIPEA_LINK_M1; |
541 | u32 savePIPEA_LINK_N1; |
541 | u32 savePIPEA_LINK_N1; |
542 | u32 savePIPEB_DATA_M1; |
542 | u32 savePIPEB_DATA_M1; |
543 | u32 savePIPEB_DATA_N1; |
543 | u32 savePIPEB_DATA_N1; |
544 | u32 savePIPEB_LINK_M1; |
544 | u32 savePIPEB_LINK_M1; |
545 | u32 savePIPEB_LINK_N1; |
545 | u32 savePIPEB_LINK_N1; |
546 | u32 saveMCHBAR_RENDER_STANDBY; |
546 | u32 saveMCHBAR_RENDER_STANDBY; |
547 | u32 savePCH_PORT_HOTPLUG; |
547 | u32 savePCH_PORT_HOTPLUG; |
548 | 548 | ||
549 | struct { |
549 | struct { |
550 | /** Bridge to intel-gtt-ko */ |
550 | /** Bridge to intel-gtt-ko */ |
551 | const struct intel_gtt *gtt; |
551 | const struct intel_gtt *gtt; |
552 | /** Memory allocator for GTT stolen memory */ |
552 | /** Memory allocator for GTT stolen memory */ |
553 | struct drm_mm stolen; |
553 | struct drm_mm stolen; |
554 | /** Memory allocator for GTT */ |
554 | /** Memory allocator for GTT */ |
555 | // struct drm_mm gtt_space; |
555 | struct drm_mm gtt_space; |
556 | /** List of all objects in gtt_space. Used to restore gtt |
556 | /** List of all objects in gtt_space. Used to restore gtt |
557 | * mappings on resume */ |
557 | * mappings on resume */ |
558 | struct list_head gtt_list; |
558 | struct list_head gtt_list; |
559 | 559 | ||
560 | /** Usable portion of the GTT for GEM */ |
560 | /** Usable portion of the GTT for GEM */ |
561 | unsigned long gtt_start; |
561 | unsigned long gtt_start; |
562 | unsigned long gtt_mappable_end; |
562 | unsigned long gtt_mappable_end; |
563 | unsigned long gtt_end; |
563 | unsigned long gtt_end; |
564 | 564 | ||
565 | // struct io_mapping *gtt_mapping; |
565 | // struct io_mapping *gtt_mapping; |
566 | int gtt_mtrr; |
566 | int gtt_mtrr; |
567 | 567 | ||
568 | // struct shrinker inactive_shrinker; |
568 | // struct shrinker inactive_shrinker; |
569 | 569 | ||
570 | /** |
570 | /** |
571 | * List of objects currently involved in rendering. |
571 | * List of objects currently involved in rendering. |
572 | * |
572 | * |
573 | * Includes buffers having the contents of their GPU caches |
573 | * Includes buffers having the contents of their GPU caches |
574 | * flushed, not necessarily primitives. last_rendering_seqno |
574 | * flushed, not necessarily primitives. last_rendering_seqno |
575 | * represents when the rendering involved will be completed. |
575 | * represents when the rendering involved will be completed. |
576 | * |
576 | * |
577 | * A reference is held on the buffer while on this list. |
577 | * A reference is held on the buffer while on this list. |
578 | */ |
578 | */ |
579 | struct list_head active_list; |
579 | struct list_head active_list; |
580 | 580 | ||
581 | /** |
581 | /** |
582 | * List of objects which are not in the ringbuffer but which |
582 | * List of objects which are not in the ringbuffer but which |
583 | * still have a write_domain which needs to be flushed before |
583 | * still have a write_domain which needs to be flushed before |
584 | * unbinding. |
584 | * unbinding. |
585 | * |
585 | * |
586 | * last_rendering_seqno is 0 while an object is in this list. |
586 | * last_rendering_seqno is 0 while an object is in this list. |
587 | * |
587 | * |
588 | * A reference is held on the buffer while on this list. |
588 | * A reference is held on the buffer while on this list. |
589 | */ |
589 | */ |
590 | struct list_head flushing_list; |
590 | struct list_head flushing_list; |
591 | 591 | ||
592 | /** |
592 | /** |
593 | * LRU list of objects which are not in the ringbuffer and |
593 | * LRU list of objects which are not in the ringbuffer and |
594 | * are ready to unbind, but are still in the GTT. |
594 | * are ready to unbind, but are still in the GTT. |
595 | * |
595 | * |
596 | * last_rendering_seqno is 0 while an object is in this list. |
596 | * last_rendering_seqno is 0 while an object is in this list. |
597 | * |
597 | * |
598 | * A reference is not held on the buffer while on this list, |
598 | * A reference is not held on the buffer while on this list, |
599 | * as merely being GTT-bound shouldn't prevent its being |
599 | * as merely being GTT-bound shouldn't prevent its being |
600 | * freed, and we'll pull it off the list in the free path. |
600 | * freed, and we'll pull it off the list in the free path. |
601 | */ |
601 | */ |
602 | struct list_head inactive_list; |
602 | struct list_head inactive_list; |
603 | 603 | ||
604 | /** |
604 | /** |
605 | * LRU list of objects which are not in the ringbuffer but |
605 | * LRU list of objects which are not in the ringbuffer but |
606 | * are still pinned in the GTT. |
606 | * are still pinned in the GTT. |
607 | */ |
607 | */ |
608 | struct list_head pinned_list; |
608 | struct list_head pinned_list; |
609 | 609 | ||
610 | /** LRU list of objects with fence regs on them. */ |
610 | /** LRU list of objects with fence regs on them. */ |
611 | struct list_head fence_list; |
611 | struct list_head fence_list; |
612 | 612 | ||
613 | /** |
613 | /** |
614 | * List of objects currently pending being freed. |
614 | * List of objects currently pending being freed. |
615 | * |
615 | * |
616 | * These objects are no longer in use, but due to a signal |
616 | * These objects are no longer in use, but due to a signal |
617 | * we were prevented from freeing them at the appointed time. |
617 | * we were prevented from freeing them at the appointed time. |
618 | */ |
618 | */ |
619 | struct list_head deferred_free_list; |
619 | struct list_head deferred_free_list; |
620 | 620 | ||
621 | /** |
621 | /** |
622 | * We leave the user IRQ off as much as possible, |
622 | * We leave the user IRQ off as much as possible, |
623 | * but this means that requests will finish and never |
623 | * but this means that requests will finish and never |
624 | * be retired once the system goes idle. Set a timer to |
624 | * be retired once the system goes idle. Set a timer to |
625 | * fire periodically while the ring is running. When it |
625 | * fire periodically while the ring is running. When it |
626 | * fires, go retire requests. |
626 | * fires, go retire requests. |
627 | */ |
627 | */ |
628 | // struct delayed_work retire_work; |
628 | // struct delayed_work retire_work; |
629 | 629 | ||
630 | /** |
630 | /** |
631 | * Are we in a non-interruptible section of code like |
631 | * Are we in a non-interruptible section of code like |
632 | * modesetting? |
632 | * modesetting? |
633 | */ |
633 | */ |
634 | bool interruptible; |
634 | bool interruptible; |
635 | 635 | ||
636 | /** |
636 | /** |
637 | * Flag if the X Server, and thus DRM, is not currently in |
637 | * Flag if the X Server, and thus DRM, is not currently in |
638 | * control of the device. |
638 | * control of the device. |
639 | * |
639 | * |
640 | * This is set between LeaveVT and EnterVT. It needs to be |
640 | * This is set between LeaveVT and EnterVT. It needs to be |
641 | * replaced with a semaphore. It also needs to be |
641 | * replaced with a semaphore. It also needs to be |
642 | * transitioned away from for kernel modesetting. |
642 | * transitioned away from for kernel modesetting. |
643 | */ |
643 | */ |
644 | int suspended; |
644 | int suspended; |
645 | 645 | ||
646 | /** |
646 | /** |
647 | * Flag if the hardware appears to be wedged. |
647 | * Flag if the hardware appears to be wedged. |
648 | * |
648 | * |
649 | * This is set when attempts to idle the device timeout. |
649 | * This is set when attempts to idle the device timeout. |
650 | * It prevents command submission from occurring and makes |
650 | * It prevents command submission from occurring and makes |
651 | * every pending request fail |
651 | * every pending request fail |
652 | */ |
652 | */ |
653 | atomic_t wedged; |
653 | atomic_t wedged; |
654 | 654 | ||
655 | /** Bit 6 swizzling required for X tiling */ |
655 | /** Bit 6 swizzling required for X tiling */ |
656 | uint32_t bit_6_swizzle_x; |
656 | uint32_t bit_6_swizzle_x; |
657 | /** Bit 6 swizzling required for Y tiling */ |
657 | /** Bit 6 swizzling required for Y tiling */ |
658 | uint32_t bit_6_swizzle_y; |
658 | uint32_t bit_6_swizzle_y; |
659 | 659 | ||
660 | /* storage for physical objects */ |
660 | /* storage for physical objects */ |
661 | // struct drm_i915_gem_phys_object *phys_objs[I915_MAX_PHYS_OBJECT]; |
661 | // struct drm_i915_gem_phys_object *phys_objs[I915_MAX_PHYS_OBJECT]; |
662 | 662 | ||
663 | /* accounting, useful for userland debugging */ |
663 | /* accounting, useful for userland debugging */ |
664 | size_t gtt_total; |
664 | size_t gtt_total; |
665 | size_t mappable_gtt_total; |
665 | size_t mappable_gtt_total; |
666 | size_t object_memory; |
666 | size_t object_memory; |
667 | u32 object_count; |
667 | u32 object_count; |
668 | } mm; |
668 | } mm; |
669 | struct sdvo_device_mapping sdvo_mappings[2]; |
669 | struct sdvo_device_mapping sdvo_mappings[2]; |
670 | /* indicate whether the LVDS_BORDER should be enabled or not */ |
670 | /* indicate whether the LVDS_BORDER should be enabled or not */ |
671 | unsigned int lvds_border_bits; |
671 | unsigned int lvds_border_bits; |
672 | /* Panel fitter placement and size for Ironlake+ */ |
672 | /* Panel fitter placement and size for Ironlake+ */ |
673 | u32 pch_pf_pos, pch_pf_size; |
673 | u32 pch_pf_pos, pch_pf_size; |
674 | int panel_t3, panel_t12; |
674 | int panel_t3, panel_t12; |
675 | 675 | ||
676 | struct drm_crtc *plane_to_crtc_mapping[2]; |
676 | struct drm_crtc *plane_to_crtc_mapping[2]; |
677 | struct drm_crtc *pipe_to_crtc_mapping[2]; |
677 | struct drm_crtc *pipe_to_crtc_mapping[2]; |
678 | // wait_queue_head_t pending_flip_queue; |
678 | // wait_queue_head_t pending_flip_queue; |
679 | bool flip_pending_is_done; |
679 | bool flip_pending_is_done; |
680 | 680 | ||
681 | /* Reclocking support */ |
681 | /* Reclocking support */ |
682 | bool render_reclock_avail; |
682 | bool render_reclock_avail; |
683 | bool lvds_downclock_avail; |
683 | bool lvds_downclock_avail; |
684 | /* indicates the reduced downclock for LVDS*/ |
684 | /* indicates the reduced downclock for LVDS*/ |
685 | int lvds_downclock; |
685 | int lvds_downclock; |
686 | // struct work_struct idle_work; |
686 | // struct work_struct idle_work; |
687 | struct timer_list idle_timer; |
687 | struct timer_list idle_timer; |
688 | bool busy; |
688 | bool busy; |
689 | u16 orig_clock; |
689 | u16 orig_clock; |
690 | int child_dev_num; |
690 | int child_dev_num; |
691 | struct child_device_config *child_dev; |
691 | struct child_device_config *child_dev; |
692 | struct drm_connector *int_lvds_connector; |
692 | struct drm_connector *int_lvds_connector; |
693 | struct drm_connector *int_edp_connector; |
693 | struct drm_connector *int_edp_connector; |
694 | 694 | ||
695 | bool mchbar_need_disable; |
695 | bool mchbar_need_disable; |
696 | 696 | ||
697 | // struct work_struct rps_work; |
697 | // struct work_struct rps_work; |
698 | spinlock_t rps_lock; |
698 | spinlock_t rps_lock; |
699 | u32 pm_iir; |
699 | u32 pm_iir; |
700 | 700 | ||
701 | u8 cur_delay; |
701 | u8 cur_delay; |
702 | u8 min_delay; |
702 | u8 min_delay; |
703 | u8 max_delay; |
703 | u8 max_delay; |
704 | u8 fmax; |
704 | u8 fmax; |
705 | u8 fstart; |
705 | u8 fstart; |
706 | 706 | ||
707 | u64 last_count1; |
707 | u64 last_count1; |
708 | unsigned long last_time1; |
708 | unsigned long last_time1; |
709 | u64 last_count2; |
709 | u64 last_count2; |
710 | struct timespec last_time2; |
710 | struct timespec last_time2; |
711 | unsigned long gfx_power; |
711 | unsigned long gfx_power; |
712 | int c_m; |
712 | int c_m; |
713 | int r_t; |
713 | int r_t; |
714 | u8 corr; |
714 | u8 corr; |
715 | spinlock_t *mchdev_lock; |
715 | spinlock_t *mchdev_lock; |
716 | 716 | ||
717 | // enum no_fbc_reason no_fbc_reason; |
717 | // enum no_fbc_reason no_fbc_reason; |
718 | 718 | ||
719 | // struct drm_mm_node *compressed_fb; |
719 | // struct drm_mm_node *compressed_fb; |
720 | // struct drm_mm_node *compressed_llb; |
720 | // struct drm_mm_node *compressed_llb; |
721 | 721 | ||
722 | unsigned long last_gpu_reset; |
722 | unsigned long last_gpu_reset; |
723 | 723 | ||
724 | /* list of fbdev register on this device */ |
724 | /* list of fbdev register on this device */ |
725 | // struct intel_fbdev *fbdev; |
725 | struct intel_fbdev *fbdev; |
726 | 726 | ||
727 | // struct backlight_device *backlight; |
727 | // struct backlight_device *backlight; |
728 | 728 | ||
729 | // struct drm_property *broadcast_rgb_property; |
729 | // struct drm_property *broadcast_rgb_property; |
730 | // struct drm_property *force_audio_property; |
730 | // struct drm_property *force_audio_property; |
731 | 731 | ||
732 | atomic_t forcewake_count; |
732 | atomic_t forcewake_count; |
733 | } drm_i915_private_t; |
733 | } drm_i915_private_t; |
734 | 734 | ||
735 | enum i915_cache_level { |
735 | enum i915_cache_level { |
736 | I915_CACHE_NONE, |
736 | I915_CACHE_NONE, |
737 | I915_CACHE_LLC, |
737 | I915_CACHE_LLC, |
738 | I915_CACHE_LLC_MLC, /* gen6+ */ |
738 | I915_CACHE_LLC_MLC, /* gen6+ */ |
739 | }; |
739 | }; |
740 | 740 | ||
741 | struct drm_i915_gem_object { |
741 | struct drm_i915_gem_object { |
742 | struct drm_gem_object base; |
742 | struct drm_gem_object base; |
743 | 743 | ||
744 | /** Current space allocated to this object in the GTT, if any. */ |
744 | /** Current space allocated to this object in the GTT, if any. */ |
745 | struct drm_mm_node *gtt_space; |
745 | struct drm_mm_node *gtt_space; |
746 | struct list_head gtt_list; |
746 | struct list_head gtt_list; |
747 | 747 | ||
748 | /** This object's place on the active/flushing/inactive lists */ |
748 | /** This object's place on the active/flushing/inactive lists */ |
749 | struct list_head ring_list; |
749 | struct list_head ring_list; |
750 | struct list_head mm_list; |
750 | struct list_head mm_list; |
751 | /** This object's place on GPU write list */ |
751 | /** This object's place on GPU write list */ |
752 | struct list_head gpu_write_list; |
752 | struct list_head gpu_write_list; |
753 | /** This object's place in the batchbuffer or on the eviction list */ |
753 | /** This object's place in the batchbuffer or on the eviction list */ |
754 | struct list_head exec_list; |
754 | struct list_head exec_list; |
755 | 755 | ||
756 | /** |
756 | /** |
757 | * This is set if the object is on the active or flushing lists |
757 | * This is set if the object is on the active or flushing lists |
758 | * (has pending rendering), and is not set if it's on inactive (ready |
758 | * (has pending rendering), and is not set if it's on inactive (ready |
759 | * to be unbound). |
759 | * to be unbound). |
760 | */ |
760 | */ |
761 | unsigned int active : 1; |
761 | unsigned int active : 1; |
762 | 762 | ||
763 | /** |
763 | /** |
764 | * This is set if the object has been written to since last bound |
764 | * This is set if the object has been written to since last bound |
765 | * to the GTT |
765 | * to the GTT |
766 | */ |
766 | */ |
767 | unsigned int dirty : 1; |
767 | unsigned int dirty : 1; |
768 | 768 | ||
769 | /** |
769 | /** |
770 | * This is set if the object has been written to since the last |
770 | * This is set if the object has been written to since the last |
771 | * GPU flush. |
771 | * GPU flush. |
772 | */ |
772 | */ |
773 | unsigned int pending_gpu_write : 1; |
773 | unsigned int pending_gpu_write : 1; |
774 | 774 | ||
775 | /** |
775 | /** |
776 | * Fence register bits (if any) for this object. Will be set |
776 | * Fence register bits (if any) for this object. Will be set |
777 | * as needed when mapped into the GTT. |
777 | * as needed when mapped into the GTT. |
778 | * Protected by dev->struct_mutex. |
778 | * Protected by dev->struct_mutex. |
779 | * |
779 | * |
780 | * Size: 4 bits for 16 fences + sign (for FENCE_REG_NONE) |
780 | * Size: 4 bits for 16 fences + sign (for FENCE_REG_NONE) |
781 | */ |
781 | */ |
782 | signed int fence_reg : 5; |
782 | signed int fence_reg : 5; |
783 | 783 | ||
784 | /** |
784 | /** |
785 | * Advice: are the backing pages purgeable? |
785 | * Advice: are the backing pages purgeable? |
786 | */ |
786 | */ |
787 | unsigned int madv : 2; |
787 | unsigned int madv : 2; |
788 | 788 | ||
789 | /** |
789 | /** |
790 | * Current tiling mode for the object. |
790 | * Current tiling mode for the object. |
791 | */ |
791 | */ |
792 | unsigned int tiling_mode : 2; |
792 | unsigned int tiling_mode : 2; |
793 | unsigned int tiling_changed : 1; |
793 | unsigned int tiling_changed : 1; |
794 | 794 | ||
795 | /** How many users have pinned this object in GTT space. The following |
795 | /** How many users have pinned this object in GTT space. The following |
796 | * users can each hold at most one reference: pwrite/pread, pin_ioctl |
796 | * users can each hold at most one reference: pwrite/pread, pin_ioctl |
797 | * (via user_pin_count), execbuffer (objects are not allowed multiple |
797 | * (via user_pin_count), execbuffer (objects are not allowed multiple |
798 | * times for the same batchbuffer), and the framebuffer code. When |
798 | * times for the same batchbuffer), and the framebuffer code. When |
799 | * switching/pageflipping, the framebuffer code has at most two buffers |
799 | * switching/pageflipping, the framebuffer code has at most two buffers |
800 | * pinned per crtc. |
800 | * pinned per crtc. |
801 | * |
801 | * |
802 | * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3 |
802 | * In the worst case this is 1 + 1 + 1 + 2*2 = 7. That would fit into 3 |
803 | * bits with absolutely no headroom. So use 4 bits. */ |
803 | * bits with absolutely no headroom. So use 4 bits. */ |
804 | unsigned int pin_count : 4; |
804 | unsigned int pin_count : 4; |
805 | #define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf |
805 | #define DRM_I915_GEM_OBJECT_MAX_PIN_COUNT 0xf |
806 | 806 | ||
807 | /** |
807 | /** |
808 | * Is the object at the current location in the gtt mappable and |
808 | * Is the object at the current location in the gtt mappable and |
809 | * fenceable? Used to avoid costly recalculations. |
809 | * fenceable? Used to avoid costly recalculations. |
810 | */ |
810 | */ |
811 | unsigned int map_and_fenceable : 1; |
811 | unsigned int map_and_fenceable : 1; |
812 | 812 | ||
813 | /** |
813 | /** |
814 | * Whether the current gtt mapping needs to be mappable (and isn't just |
814 | * Whether the current gtt mapping needs to be mappable (and isn't just |
815 | * mappable by accident). Track pin and fault separate for a more |
815 | * mappable by accident). Track pin and fault separate for a more |
816 | * accurate mappable working set. |
816 | * accurate mappable working set. |
817 | */ |
817 | */ |
818 | unsigned int fault_mappable : 1; |
818 | unsigned int fault_mappable : 1; |
819 | unsigned int pin_mappable : 1; |
819 | unsigned int pin_mappable : 1; |
820 | 820 | ||
821 | /* |
821 | /* |
822 | * Is the GPU currently using a fence to access this buffer, |
822 | * Is the GPU currently using a fence to access this buffer, |
823 | */ |
823 | */ |
824 | unsigned int pending_fenced_gpu_access:1; |
824 | unsigned int pending_fenced_gpu_access:1; |
825 | unsigned int fenced_gpu_access:1; |
825 | unsigned int fenced_gpu_access:1; |
826 | 826 | ||
827 | unsigned int cache_level:2; |
827 | unsigned int cache_level:2; |
828 | 828 | ||
829 | struct page **pages; |
829 | struct page **pages; |
830 | 830 | ||
831 | /** |
831 | /** |
832 | * DMAR support |
832 | * DMAR support |
833 | */ |
833 | */ |
834 | struct scatterlist *sg_list; |
834 | struct scatterlist *sg_list; |
835 | int num_sg; |
835 | int num_sg; |
836 | 836 | ||
837 | /** |
837 | /** |
838 | * Used for performing relocations during execbuffer insertion. |
838 | * Used for performing relocations during execbuffer insertion. |
839 | */ |
839 | */ |
840 | struct hlist_node exec_node; |
840 | struct hlist_node exec_node; |
841 | unsigned long exec_handle; |
841 | unsigned long exec_handle; |
842 | struct drm_i915_gem_exec_object2 *exec_entry; |
842 | struct drm_i915_gem_exec_object2 *exec_entry; |
843 | 843 | ||
844 | /** |
844 | /** |
845 | * Current offset of the object in GTT space. |
845 | * Current offset of the object in GTT space. |
846 | * |
846 | * |
847 | * This is the same as gtt_space->start |
847 | * This is the same as gtt_space->start |
848 | */ |
848 | */ |
849 | uint32_t gtt_offset; |
849 | uint32_t gtt_offset; |
850 | 850 | ||
851 | /** Breadcrumb of last rendering to the buffer. */ |
851 | /** Breadcrumb of last rendering to the buffer. */ |
852 | uint32_t last_rendering_seqno; |
852 | uint32_t last_rendering_seqno; |
853 | struct intel_ring_buffer *ring; |
853 | struct intel_ring_buffer *ring; |
854 | 854 | ||
855 | /** Breadcrumb of last fenced GPU access to the buffer. */ |
855 | /** Breadcrumb of last fenced GPU access to the buffer. */ |
856 | uint32_t last_fenced_seqno; |
856 | uint32_t last_fenced_seqno; |
857 | struct intel_ring_buffer *last_fenced_ring; |
857 | struct intel_ring_buffer *last_fenced_ring; |
858 | 858 | ||
859 | /** Current tiling stride for the object, if it's tiled. */ |
859 | /** Current tiling stride for the object, if it's tiled. */ |
860 | uint32_t stride; |
860 | uint32_t stride; |
861 | 861 | ||
862 | /** Record of address bit 17 of each page at last unbind. */ |
862 | /** Record of address bit 17 of each page at last unbind. */ |
863 | unsigned long *bit_17; |
863 | unsigned long *bit_17; |
864 | 864 | ||
865 | 865 | ||
866 | /** |
866 | /** |
867 | * If present, while GEM_DOMAIN_CPU is in the read domain this array |
867 | * If present, while GEM_DOMAIN_CPU is in the read domain this array |
868 | * flags which individual pages are valid. |
868 | * flags which individual pages are valid. |
869 | */ |
869 | */ |
870 | uint8_t *page_cpu_valid; |
870 | uint8_t *page_cpu_valid; |
871 | 871 | ||
872 | /** User space pin count and filp owning the pin */ |
872 | /** User space pin count and filp owning the pin */ |
873 | uint32_t user_pin_count; |
873 | uint32_t user_pin_count; |
874 | struct drm_file *pin_filp; |
874 | struct drm_file *pin_filp; |
875 | 875 | ||
876 | /** for phy allocated objects */ |
876 | /** for phy allocated objects */ |
877 | struct drm_i915_gem_phys_object *phys_obj; |
877 | struct drm_i915_gem_phys_object *phys_obj; |
878 | 878 | ||
879 | /** |
879 | /** |
880 | * Number of crtcs where this object is currently the fb, but |
880 | * Number of crtcs where this object is currently the fb, but |
881 | * will be page flipped away on the next vblank. When it |
881 | * will be page flipped away on the next vblank. When it |
882 | * reaches 0, dev_priv->pending_flip_queue will be woken up. |
882 | * reaches 0, dev_priv->pending_flip_queue will be woken up. |
883 | */ |
883 | */ |
884 | atomic_t pending_flip; |
884 | atomic_t pending_flip; |
885 | }; |
885 | }; |
886 | 886 | ||
887 | 887 | ||
888 | #define to_intel_bo(x) container_of(x, struct drm_i915_gem_object, base) |
888 | #define to_intel_bo(x) container_of(x, struct drm_i915_gem_object, base) |
889 | 889 | ||
890 | /** |
890 | /** |
891 | * Request queue structure. |
891 | * Request queue structure. |
892 | * |
892 | * |
893 | * The request queue allows us to note sequence numbers that have been emitted |
893 | * The request queue allows us to note sequence numbers that have been emitted |
894 | * and may be associated with active buffers to be retired. |
894 | * and may be associated with active buffers to be retired. |
895 | * |
895 | * |
896 | * By keeping this list, we can avoid having to do questionable |
896 | * By keeping this list, we can avoid having to do questionable |
897 | * sequence-number comparisons on buffer last_rendering_seqnos, and associate |
897 | * sequence-number comparisons on buffer last_rendering_seqnos, and associate |
898 | * an emission time with seqnos for tracking how far ahead of the GPU we are. |
898 | * an emission time with seqnos for tracking how far ahead of the GPU we are. |
899 | */ |
899 | */ |
900 | struct drm_i915_gem_request { |
900 | struct drm_i915_gem_request { |
901 | /** On Which ring this request was generated */ |
901 | /** On Which ring this request was generated */ |
902 | struct intel_ring_buffer *ring; |
902 | struct intel_ring_buffer *ring; |
903 | 903 | ||
904 | /** GEM sequence number associated with this request. */ |
904 | /** GEM sequence number associated with this request. */ |
905 | uint32_t seqno; |
905 | uint32_t seqno; |
906 | 906 | ||
907 | /** Time at which this request was emitted, in jiffies. */ |
907 | /** Time at which this request was emitted, in jiffies. */ |
908 | unsigned long emitted_jiffies; |
908 | unsigned long emitted_jiffies; |
909 | 909 | ||
910 | /** global list entry for this request */ |
910 | /** global list entry for this request */ |
911 | struct list_head list; |
911 | struct list_head list; |
912 | 912 | ||
913 | struct drm_i915_file_private *file_priv; |
913 | struct drm_i915_file_private *file_priv; |
914 | /** file_priv list entry for this request */ |
914 | /** file_priv list entry for this request */ |
915 | struct list_head client_list; |
915 | struct list_head client_list; |
916 | }; |
916 | }; |
917 | 917 | ||
918 | struct drm_i915_file_private { |
918 | struct drm_i915_file_private { |
919 | struct { |
919 | struct { |
920 | // struct spinlock lock; |
920 | // struct spinlock lock; |
921 | struct list_head request_list; |
921 | struct list_head request_list; |
922 | } mm; |
922 | } mm; |
923 | }; |
923 | }; |
924 | 924 | ||
925 | #define INTEL_INFO(dev) (((struct drm_i915_private *) (dev)->dev_private)->info) |
925 | #define INTEL_INFO(dev) (((struct drm_i915_private *) (dev)->dev_private)->info) |
926 | 926 | ||
927 | #define IS_I830(dev) ((dev)->pci_device == 0x3577) |
927 | #define IS_I830(dev) ((dev)->pci_device == 0x3577) |
928 | #define IS_845G(dev) ((dev)->pci_device == 0x2562) |
928 | #define IS_845G(dev) ((dev)->pci_device == 0x2562) |
929 | #define IS_I85X(dev) (INTEL_INFO(dev)->is_i85x) |
929 | #define IS_I85X(dev) (INTEL_INFO(dev)->is_i85x) |
930 | #define IS_I865G(dev) ((dev)->pci_device == 0x2572) |
930 | #define IS_I865G(dev) ((dev)->pci_device == 0x2572) |
931 | #define IS_I915G(dev) (INTEL_INFO(dev)->is_i915g) |
931 | #define IS_I915G(dev) (INTEL_INFO(dev)->is_i915g) |
932 | #define IS_I915GM(dev) ((dev)->pci_device == 0x2592) |
932 | #define IS_I915GM(dev) ((dev)->pci_device == 0x2592) |
933 | #define IS_I945G(dev) ((dev)->pci_device == 0x2772) |
933 | #define IS_I945G(dev) ((dev)->pci_device == 0x2772) |
934 | #define IS_I945GM(dev) (INTEL_INFO(dev)->is_i945gm) |
934 | #define IS_I945GM(dev) (INTEL_INFO(dev)->is_i945gm) |
935 | #define IS_BROADWATER(dev) (INTEL_INFO(dev)->is_broadwater) |
935 | #define IS_BROADWATER(dev) (INTEL_INFO(dev)->is_broadwater) |
936 | #define IS_CRESTLINE(dev) (INTEL_INFO(dev)->is_crestline) |
936 | #define IS_CRESTLINE(dev) (INTEL_INFO(dev)->is_crestline) |
937 | #define IS_GM45(dev) ((dev)->pci_device == 0x2A42) |
937 | #define IS_GM45(dev) ((dev)->pci_device == 0x2A42) |
938 | #define IS_G4X(dev) (INTEL_INFO(dev)->is_g4x) |
938 | #define IS_G4X(dev) (INTEL_INFO(dev)->is_g4x) |
939 | #define IS_PINEVIEW_G(dev) ((dev)->pci_device == 0xa001) |
939 | #define IS_PINEVIEW_G(dev) ((dev)->pci_device == 0xa001) |
940 | #define IS_PINEVIEW_M(dev) ((dev)->pci_device == 0xa011) |
940 | #define IS_PINEVIEW_M(dev) ((dev)->pci_device == 0xa011) |
941 | #define IS_PINEVIEW(dev) (INTEL_INFO(dev)->is_pineview) |
941 | #define IS_PINEVIEW(dev) (INTEL_INFO(dev)->is_pineview) |
942 | #define IS_G33(dev) (INTEL_INFO(dev)->is_g33) |
942 | #define IS_G33(dev) (INTEL_INFO(dev)->is_g33) |
943 | #define IS_IRONLAKE_D(dev) ((dev)->pci_device == 0x0042) |
943 | #define IS_IRONLAKE_D(dev) ((dev)->pci_device == 0x0042) |
944 | #define IS_IRONLAKE_M(dev) ((dev)->pci_device == 0x0046) |
944 | #define IS_IRONLAKE_M(dev) ((dev)->pci_device == 0x0046) |
945 | #define IS_IVYBRIDGE(dev) (INTEL_INFO(dev)->is_ivybridge) |
945 | #define IS_IVYBRIDGE(dev) (INTEL_INFO(dev)->is_ivybridge) |
946 | #define IS_MOBILE(dev) (INTEL_INFO(dev)->is_mobile) |
946 | #define IS_MOBILE(dev) (INTEL_INFO(dev)->is_mobile) |
947 | 947 | ||
948 | /* |
948 | /* |
949 | * The genX designation typically refers to the render engine, so render |
949 | * The genX designation typically refers to the render engine, so render |
950 | * capability related checks should use IS_GEN, while display and other checks |
950 | * capability related checks should use IS_GEN, while display and other checks |
951 | * have their own (e.g. HAS_PCH_SPLIT for ILK+ display, IS_foo for particular |
951 | * have their own (e.g. HAS_PCH_SPLIT for ILK+ display, IS_foo for particular |
952 | * chips, etc.). |
952 | * chips, etc.). |
953 | */ |
953 | */ |
954 | #define IS_GEN2(dev) (INTEL_INFO(dev)->gen == 2) |
954 | #define IS_GEN2(dev) (INTEL_INFO(dev)->gen == 2) |
955 | #define IS_GEN3(dev) (INTEL_INFO(dev)->gen == 3) |
955 | #define IS_GEN3(dev) (INTEL_INFO(dev)->gen == 3) |
956 | #define IS_GEN4(dev) (INTEL_INFO(dev)->gen == 4) |
956 | #define IS_GEN4(dev) (INTEL_INFO(dev)->gen == 4) |
957 | #define IS_GEN5(dev) (INTEL_INFO(dev)->gen == 5) |
957 | #define IS_GEN5(dev) (INTEL_INFO(dev)->gen == 5) |
958 | #define IS_GEN6(dev) (INTEL_INFO(dev)->gen == 6) |
958 | #define IS_GEN6(dev) (INTEL_INFO(dev)->gen == 6) |
959 | #define IS_GEN7(dev) (INTEL_INFO(dev)->gen == 7) |
959 | #define IS_GEN7(dev) (INTEL_INFO(dev)->gen == 7) |
960 | 960 | ||
961 | #define HAS_BSD(dev) (INTEL_INFO(dev)->has_bsd_ring) |
961 | #define HAS_BSD(dev) (INTEL_INFO(dev)->has_bsd_ring) |
962 | #define HAS_BLT(dev) (INTEL_INFO(dev)->has_blt_ring) |
962 | #define HAS_BLT(dev) (INTEL_INFO(dev)->has_blt_ring) |
963 | #define I915_NEED_GFX_HWS(dev) (INTEL_INFO(dev)->need_gfx_hws) |
963 | #define I915_NEED_GFX_HWS(dev) (INTEL_INFO(dev)->need_gfx_hws) |
964 | 964 | ||
965 | #define HAS_OVERLAY(dev) (INTEL_INFO(dev)->has_overlay) |
965 | #define HAS_OVERLAY(dev) (INTEL_INFO(dev)->has_overlay) |
966 | #define OVERLAY_NEEDS_PHYSICAL(dev) (INTEL_INFO(dev)->overlay_needs_physical) |
966 | #define OVERLAY_NEEDS_PHYSICAL(dev) (INTEL_INFO(dev)->overlay_needs_physical) |
967 | 967 | ||
968 | /* With the 945 and later, Y tiling got adjusted so that it was 32 128-byte |
968 | /* With the 945 and later, Y tiling got adjusted so that it was 32 128-byte |
969 | * rows, which changed the alignment requirements and fence programming. |
969 | * rows, which changed the alignment requirements and fence programming. |
970 | */ |
970 | */ |
971 | #define HAS_128_BYTE_Y_TILING(dev) (!IS_GEN2(dev) && !(IS_I915G(dev) || \ |
971 | #define HAS_128_BYTE_Y_TILING(dev) (!IS_GEN2(dev) && !(IS_I915G(dev) || \ |
972 | IS_I915GM(dev))) |
972 | IS_I915GM(dev))) |
973 | #define SUPPORTS_DIGITAL_OUTPUTS(dev) (!IS_GEN2(dev) && !IS_PINEVIEW(dev)) |
973 | #define SUPPORTS_DIGITAL_OUTPUTS(dev) (!IS_GEN2(dev) && !IS_PINEVIEW(dev)) |
974 | #define SUPPORTS_INTEGRATED_HDMI(dev) (IS_G4X(dev) || IS_GEN5(dev)) |
974 | #define SUPPORTS_INTEGRATED_HDMI(dev) (IS_G4X(dev) || IS_GEN5(dev)) |
975 | #define SUPPORTS_INTEGRATED_DP(dev) (IS_G4X(dev) || IS_GEN5(dev)) |
975 | #define SUPPORTS_INTEGRATED_DP(dev) (IS_G4X(dev) || IS_GEN5(dev)) |
976 | #define SUPPORTS_EDP(dev) (IS_IRONLAKE_M(dev)) |
976 | #define SUPPORTS_EDP(dev) (IS_IRONLAKE_M(dev)) |
977 | #define SUPPORTS_TV(dev) (INTEL_INFO(dev)->supports_tv) |
977 | #define SUPPORTS_TV(dev) (INTEL_INFO(dev)->supports_tv) |
978 | #define I915_HAS_HOTPLUG(dev) (INTEL_INFO(dev)->has_hotplug) |
978 | #define I915_HAS_HOTPLUG(dev) (INTEL_INFO(dev)->has_hotplug) |
979 | /* dsparb controlled by hw only */ |
979 | /* dsparb controlled by hw only */ |
980 | #define DSPARB_HWCONTROL(dev) (IS_G4X(dev) || IS_IRONLAKE(dev)) |
980 | #define DSPARB_HWCONTROL(dev) (IS_G4X(dev) || IS_IRONLAKE(dev)) |
981 | 981 | ||
982 | #define HAS_FW_BLC(dev) (INTEL_INFO(dev)->gen > 2) |
982 | #define HAS_FW_BLC(dev) (INTEL_INFO(dev)->gen > 2) |
983 | #define HAS_PIPE_CXSR(dev) (INTEL_INFO(dev)->has_pipe_cxsr) |
983 | #define HAS_PIPE_CXSR(dev) (INTEL_INFO(dev)->has_pipe_cxsr) |
984 | #define I915_HAS_FBC(dev) (INTEL_INFO(dev)->has_fbc) |
984 | #define I915_HAS_FBC(dev) (INTEL_INFO(dev)->has_fbc) |
985 | 985 | ||
986 | #define HAS_PCH_SPLIT(dev) (IS_GEN5(dev) || IS_GEN6(dev) || IS_IVYBRIDGE(dev)) |
986 | #define HAS_PCH_SPLIT(dev) (IS_GEN5(dev) || IS_GEN6(dev) || IS_IVYBRIDGE(dev)) |
987 | #define HAS_PIPE_CONTROL(dev) (INTEL_INFO(dev)->gen >= 5) |
987 | #define HAS_PIPE_CONTROL(dev) (INTEL_INFO(dev)->gen >= 5) |
988 | 988 | ||
989 | #define INTEL_PCH_TYPE(dev) (((struct drm_i915_private *)(dev)->dev_private)->pch_type) |
989 | #define INTEL_PCH_TYPE(dev) (((struct drm_i915_private *)(dev)->dev_private)->pch_type) |
990 | #define HAS_PCH_CPT(dev) (INTEL_PCH_TYPE(dev) == PCH_CPT) |
990 | #define HAS_PCH_CPT(dev) (INTEL_PCH_TYPE(dev) == PCH_CPT) |
991 | #define HAS_PCH_IBX(dev) (INTEL_PCH_TYPE(dev) == PCH_IBX) |
991 | #define HAS_PCH_IBX(dev) (INTEL_PCH_TYPE(dev) == PCH_IBX) |
992 | 992 | ||
993 | //#include "i915_trace.h" |
993 | //#include "i915_trace.h" |
994 | 994 | ||
995 | extern int i915_max_ioctl; |
995 | extern int i915_max_ioctl; |
996 | extern unsigned int i915_fbpercrtc; |
996 | extern unsigned int i915_fbpercrtc; |
997 | extern int i915_panel_ignore_lid; |
997 | extern int i915_panel_ignore_lid; |
998 | extern unsigned int i915_powersave; |
998 | extern unsigned int i915_powersave; |
999 | extern unsigned int i915_semaphores; |
999 | extern unsigned int i915_semaphores; |
1000 | extern unsigned int i915_lvds_downclock; |
1000 | extern unsigned int i915_lvds_downclock; |
1001 | extern unsigned int i915_panel_use_ssc; |
1001 | extern unsigned int i915_panel_use_ssc; |
1002 | extern int i915_vbt_sdvo_panel_type; |
1002 | extern int i915_vbt_sdvo_panel_type; |
1003 | extern unsigned int i915_enable_rc6; |
1003 | extern unsigned int i915_enable_rc6; |
1004 | extern unsigned int i915_enable_fbc; |
1004 | extern unsigned int i915_enable_fbc; |
1005 | extern bool i915_enable_hangcheck; |
1005 | extern bool i915_enable_hangcheck; |
1006 | 1006 | ||
1007 | extern int i915_resume(struct drm_device *dev); |
1007 | extern int i915_resume(struct drm_device *dev); |
1008 | extern int i915_master_create(struct drm_device *dev, struct drm_master *master); |
1008 | extern int i915_master_create(struct drm_device *dev, struct drm_master *master); |
1009 | extern void i915_master_destroy(struct drm_device *dev, struct drm_master *master); |
1009 | extern void i915_master_destroy(struct drm_device *dev, struct drm_master *master); |
1010 | 1010 | ||
1011 | /* i915_dma.c */ |
1011 | /* i915_dma.c */ |
1012 | extern void i915_kernel_lost_context(struct drm_device * dev); |
1012 | extern void i915_kernel_lost_context(struct drm_device * dev); |
1013 | extern int i915_driver_load(struct drm_device *, unsigned long flags); |
1013 | extern int i915_driver_load(struct drm_device *, unsigned long flags); |
1014 | extern int i915_driver_unload(struct drm_device *); |
1014 | extern int i915_driver_unload(struct drm_device *); |
1015 | extern int i915_driver_open(struct drm_device *dev, struct drm_file *file_priv); |
1015 | extern int i915_driver_open(struct drm_device *dev, struct drm_file *file_priv); |
1016 | extern void i915_driver_lastclose(struct drm_device * dev); |
1016 | extern void i915_driver_lastclose(struct drm_device * dev); |
1017 | extern void i915_driver_preclose(struct drm_device *dev, |
1017 | extern void i915_driver_preclose(struct drm_device *dev, |
1018 | struct drm_file *file_priv); |
1018 | struct drm_file *file_priv); |
1019 | extern void i915_driver_postclose(struct drm_device *dev, |
1019 | extern void i915_driver_postclose(struct drm_device *dev, |
1020 | struct drm_file *file_priv); |
1020 | struct drm_file *file_priv); |
1021 | extern int i915_driver_device_is_agp(struct drm_device * dev); |
1021 | extern int i915_driver_device_is_agp(struct drm_device * dev); |
1022 | extern long i915_compat_ioctl(struct file *filp, unsigned int cmd, |
1022 | extern long i915_compat_ioctl(struct file *filp, unsigned int cmd, |
1023 | unsigned long arg); |
1023 | unsigned long arg); |
1024 | extern int i915_emit_box(struct drm_device *dev, |
1024 | extern int i915_emit_box(struct drm_device *dev, |
1025 | struct drm_clip_rect *box, |
1025 | struct drm_clip_rect *box, |
1026 | int DR1, int DR4); |
1026 | int DR1, int DR4); |
1027 | extern int i915_reset(struct drm_device *dev, u8 flags); |
1027 | extern int i915_reset(struct drm_device *dev, u8 flags); |
1028 | extern unsigned long i915_chipset_val(struct drm_i915_private *dev_priv); |
1028 | extern unsigned long i915_chipset_val(struct drm_i915_private *dev_priv); |
1029 | extern unsigned long i915_mch_val(struct drm_i915_private *dev_priv); |
1029 | extern unsigned long i915_mch_val(struct drm_i915_private *dev_priv); |
1030 | extern unsigned long i915_gfx_val(struct drm_i915_private *dev_priv); |
1030 | extern unsigned long i915_gfx_val(struct drm_i915_private *dev_priv); |
1031 | extern void i915_update_gfx_val(struct drm_i915_private *dev_priv); |
1031 | extern void i915_update_gfx_val(struct drm_i915_private *dev_priv); |
1032 | 1032 | ||
1033 | 1033 | ||
1034 | /* i915_irq.c */ |
1034 | /* i915_irq.c */ |
1035 | void i915_hangcheck_elapsed(unsigned long data); |
1035 | void i915_hangcheck_elapsed(unsigned long data); |
1036 | void i915_handle_error(struct drm_device *dev, bool wedged); |
1036 | void i915_handle_error(struct drm_device *dev, bool wedged); |
1037 | extern int i915_irq_emit(struct drm_device *dev, void *data, |
1037 | extern int i915_irq_emit(struct drm_device *dev, void *data, |
1038 | struct drm_file *file_priv); |
1038 | struct drm_file *file_priv); |
1039 | extern int i915_irq_wait(struct drm_device *dev, void *data, |
1039 | extern int i915_irq_wait(struct drm_device *dev, void *data, |
1040 | struct drm_file *file_priv); |
1040 | struct drm_file *file_priv); |
1041 | 1041 | ||
1042 | extern void intel_irq_init(struct drm_device *dev); |
1042 | extern void intel_irq_init(struct drm_device *dev); |
1043 | 1043 | ||
1044 | extern int i915_vblank_pipe_set(struct drm_device *dev, void *data, |
1044 | extern int i915_vblank_pipe_set(struct drm_device *dev, void *data, |
1045 | struct drm_file *file_priv); |
1045 | struct drm_file *file_priv); |
1046 | extern int i915_vblank_pipe_get(struct drm_device *dev, void *data, |
1046 | extern int i915_vblank_pipe_get(struct drm_device *dev, void *data, |
1047 | struct drm_file *file_priv); |
1047 | struct drm_file *file_priv); |
1048 | extern int i915_vblank_swap(struct drm_device *dev, void *data, |
1048 | extern int i915_vblank_swap(struct drm_device *dev, void *data, |
1049 | struct drm_file *file_priv); |
1049 | struct drm_file *file_priv); |
1050 | 1050 | ||
1051 | void |
1051 | void |
1052 | i915_enable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); |
1052 | i915_enable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); |
1053 | 1053 | ||
1054 | void |
1054 | void |
1055 | i915_disable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); |
1055 | i915_disable_pipestat(drm_i915_private_t *dev_priv, int pipe, u32 mask); |
1056 | 1056 | ||
1057 | void intel_enable_asle (struct drm_device *dev); |
1057 | void intel_enable_asle (struct drm_device *dev); |
1058 | 1058 | ||
1059 | #ifdef CONFIG_DEBUG_FS |
1059 | #ifdef CONFIG_DEBUG_FS |
1060 | extern void i915_destroy_error_state(struct drm_device *dev); |
1060 | extern void i915_destroy_error_state(struct drm_device *dev); |
1061 | #else |
1061 | #else |
1062 | #define i915_destroy_error_state(x) |
1062 | #define i915_destroy_error_state(x) |
1063 | #endif |
1063 | #endif |
1064 | 1064 | ||
1065 | 1065 | ||
1066 | /* i915_mem.c */ |
1066 | /* i915_mem.c */ |
1067 | extern int i915_mem_alloc(struct drm_device *dev, void *data, |
1067 | extern int i915_mem_alloc(struct drm_device *dev, void *data, |
1068 | struct drm_file *file_priv); |
1068 | struct drm_file *file_priv); |
1069 | extern int i915_mem_free(struct drm_device *dev, void *data, |
1069 | extern int i915_mem_free(struct drm_device *dev, void *data, |
1070 | struct drm_file *file_priv); |
1070 | struct drm_file *file_priv); |
1071 | extern int i915_mem_init_heap(struct drm_device *dev, void *data, |
1071 | extern int i915_mem_init_heap(struct drm_device *dev, void *data, |
1072 | struct drm_file *file_priv); |
1072 | struct drm_file *file_priv); |
1073 | extern int i915_mem_destroy_heap(struct drm_device *dev, void *data, |
1073 | extern int i915_mem_destroy_heap(struct drm_device *dev, void *data, |
1074 | struct drm_file *file_priv); |
1074 | struct drm_file *file_priv); |
1075 | extern void i915_mem_takedown(struct mem_block **heap); |
1075 | extern void i915_mem_takedown(struct mem_block **heap); |
1076 | extern void i915_mem_release(struct drm_device * dev, |
1076 | extern void i915_mem_release(struct drm_device * dev, |
1077 | struct drm_file *file_priv, struct mem_block *heap); |
1077 | struct drm_file *file_priv, struct mem_block *heap); |
1078 | /* i915_gem.c */ |
1078 | /* i915_gem.c */ |
1079 | int i915_gem_init_ioctl(struct drm_device *dev, void *data, |
1079 | int i915_gem_init_ioctl(struct drm_device *dev, void *data, |
1080 | struct drm_file *file_priv); |
1080 | struct drm_file *file_priv); |
1081 | int i915_gem_create_ioctl(struct drm_device *dev, void *data, |
1081 | int i915_gem_create_ioctl(struct drm_device *dev, void *data, |
1082 | struct drm_file *file_priv); |
1082 | struct drm_file *file_priv); |
1083 | int i915_gem_pread_ioctl(struct drm_device *dev, void *data, |
1083 | int i915_gem_pread_ioctl(struct drm_device *dev, void *data, |
1084 | struct drm_file *file_priv); |
1084 | struct drm_file *file_priv); |
1085 | int i915_gem_pwrite_ioctl(struct drm_device *dev, void *data, |
1085 | int i915_gem_pwrite_ioctl(struct drm_device *dev, void *data, |
1086 | struct drm_file *file_priv); |
1086 | struct drm_file *file_priv); |
1087 | int i915_gem_mmap_ioctl(struct drm_device *dev, void *data, |
1087 | int i915_gem_mmap_ioctl(struct drm_device *dev, void *data, |
1088 | struct drm_file *file_priv); |
1088 | struct drm_file *file_priv); |
1089 | int i915_gem_mmap_gtt_ioctl(struct drm_device *dev, void *data, |
1089 | int i915_gem_mmap_gtt_ioctl(struct drm_device *dev, void *data, |
1090 | struct drm_file *file_priv); |
1090 | struct drm_file *file_priv); |
1091 | int i915_gem_set_domain_ioctl(struct drm_device *dev, void *data, |
1091 | int i915_gem_set_domain_ioctl(struct drm_device *dev, void *data, |
1092 | struct drm_file *file_priv); |
1092 | struct drm_file *file_priv); |
1093 | int i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data, |
1093 | int i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data, |
1094 | struct drm_file *file_priv); |
1094 | struct drm_file *file_priv); |
1095 | int i915_gem_execbuffer(struct drm_device *dev, void *data, |
1095 | int i915_gem_execbuffer(struct drm_device *dev, void *data, |
1096 | struct drm_file *file_priv); |
1096 | struct drm_file *file_priv); |
1097 | int i915_gem_execbuffer2(struct drm_device *dev, void *data, |
1097 | int i915_gem_execbuffer2(struct drm_device *dev, void *data, |
1098 | struct drm_file *file_priv); |
1098 | struct drm_file *file_priv); |
1099 | int i915_gem_pin_ioctl(struct drm_device *dev, void *data, |
1099 | int i915_gem_pin_ioctl(struct drm_device *dev, void *data, |
1100 | struct drm_file *file_priv); |
1100 | struct drm_file *file_priv); |
1101 | int i915_gem_unpin_ioctl(struct drm_device *dev, void *data, |
1101 | int i915_gem_unpin_ioctl(struct drm_device *dev, void *data, |
1102 | struct drm_file *file_priv); |
1102 | struct drm_file *file_priv); |
1103 | int i915_gem_busy_ioctl(struct drm_device *dev, void *data, |
1103 | int i915_gem_busy_ioctl(struct drm_device *dev, void *data, |
1104 | struct drm_file *file_priv); |
1104 | struct drm_file *file_priv); |
1105 | int i915_gem_throttle_ioctl(struct drm_device *dev, void *data, |
1105 | int i915_gem_throttle_ioctl(struct drm_device *dev, void *data, |
1106 | struct drm_file *file_priv); |
1106 | struct drm_file *file_priv); |
1107 | int i915_gem_madvise_ioctl(struct drm_device *dev, void *data, |
1107 | int i915_gem_madvise_ioctl(struct drm_device *dev, void *data, |
1108 | struct drm_file *file_priv); |
1108 | struct drm_file *file_priv); |
1109 | int i915_gem_entervt_ioctl(struct drm_device *dev, void *data, |
1109 | int i915_gem_entervt_ioctl(struct drm_device *dev, void *data, |
1110 | struct drm_file *file_priv); |
1110 | struct drm_file *file_priv); |
1111 | int i915_gem_leavevt_ioctl(struct drm_device *dev, void *data, |
1111 | int i915_gem_leavevt_ioctl(struct drm_device *dev, void *data, |
1112 | struct drm_file *file_priv); |
1112 | struct drm_file *file_priv); |
1113 | int i915_gem_set_tiling(struct drm_device *dev, void *data, |
1113 | int i915_gem_set_tiling(struct drm_device *dev, void *data, |
1114 | struct drm_file *file_priv); |
1114 | struct drm_file *file_priv); |
1115 | int i915_gem_get_tiling(struct drm_device *dev, void *data, |
1115 | int i915_gem_get_tiling(struct drm_device *dev, void *data, |
1116 | struct drm_file *file_priv); |
1116 | struct drm_file *file_priv); |
1117 | int i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, |
1117 | int i915_gem_get_aperture_ioctl(struct drm_device *dev, void *data, |
1118 | struct drm_file *file_priv); |
1118 | struct drm_file *file_priv); |
1119 | void i915_gem_load(struct drm_device *dev); |
1119 | void i915_gem_load(struct drm_device *dev); |
1120 | int i915_gem_init_object(struct drm_gem_object *obj); |
1120 | int i915_gem_init_object(struct drm_gem_object *obj); |
1121 | int __must_check i915_gem_flush_ring(struct intel_ring_buffer *ring, |
1121 | int __must_check i915_gem_flush_ring(struct intel_ring_buffer *ring, |
1122 | uint32_t invalidate_domains, |
1122 | uint32_t invalidate_domains, |
1123 | uint32_t flush_domains); |
1123 | uint32_t flush_domains); |
1124 | struct drm_i915_gem_object *i915_gem_alloc_object(struct drm_device *dev, |
1124 | struct drm_i915_gem_object *i915_gem_alloc_object(struct drm_device *dev, |
1125 | size_t size); |
1125 | size_t size); |
1126 | void i915_gem_free_object(struct drm_gem_object *obj); |
1126 | void i915_gem_free_object(struct drm_gem_object *obj); |
1127 | int __must_check i915_gem_object_pin(struct drm_i915_gem_object *obj, |
1127 | int __must_check i915_gem_object_pin(struct drm_i915_gem_object *obj, |
1128 | uint32_t alignment, |
1128 | uint32_t alignment, |
1129 | bool map_and_fenceable); |
1129 | bool map_and_fenceable); |
1130 | void i915_gem_object_unpin(struct drm_i915_gem_object *obj); |
1130 | void i915_gem_object_unpin(struct drm_i915_gem_object *obj); |
1131 | int __must_check i915_gem_object_unbind(struct drm_i915_gem_object *obj); |
1131 | int __must_check i915_gem_object_unbind(struct drm_i915_gem_object *obj); |
1132 | void i915_gem_release_mmap(struct drm_i915_gem_object *obj); |
1132 | void i915_gem_release_mmap(struct drm_i915_gem_object *obj); |
1133 | void i915_gem_lastclose(struct drm_device *dev); |
1133 | void i915_gem_lastclose(struct drm_device *dev); |
1134 | 1134 | ||
1135 | int __must_check i915_mutex_lock_interruptible(struct drm_device *dev); |
1135 | int __must_check i915_mutex_lock_interruptible(struct drm_device *dev); |
1136 | int __must_check i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj); |
1136 | int __must_check i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj); |
1137 | void i915_gem_object_move_to_active(struct drm_i915_gem_object *obj, |
1137 | void i915_gem_object_move_to_active(struct drm_i915_gem_object *obj, |
1138 | struct intel_ring_buffer *ring, |
1138 | struct intel_ring_buffer *ring, |
1139 | u32 seqno); |
1139 | u32 seqno); |
1140 | 1140 | ||
1141 | int i915_gem_dumb_create(struct drm_file *file_priv, |
1141 | int i915_gem_dumb_create(struct drm_file *file_priv, |
1142 | struct drm_device *dev, |
1142 | struct drm_device *dev, |
1143 | struct drm_mode_create_dumb *args); |
1143 | struct drm_mode_create_dumb *args); |
1144 | int i915_gem_mmap_gtt(struct drm_file *file_priv, struct drm_device *dev, |
1144 | int i915_gem_mmap_gtt(struct drm_file *file_priv, struct drm_device *dev, |
1145 | uint32_t handle, uint64_t *offset); |
1145 | uint32_t handle, uint64_t *offset); |
1146 | int i915_gem_dumb_destroy(struct drm_file *file_priv, struct drm_device *dev, |
1146 | int i915_gem_dumb_destroy(struct drm_file *file_priv, struct drm_device *dev, |
1147 | uint32_t handle); |
1147 | uint32_t handle); |
1148 | /** |
1148 | /** |
1149 | * Returns true if seq1 is later than seq2. |
1149 | * Returns true if seq1 is later than seq2. |
1150 | */ |
1150 | */ |
1151 | //static inline bool |
1151 | //static inline bool |
1152 | //i915_seqno_passed(uint32_t seq1, uint32_t seq2) |
1152 | //i915_seqno_passed(uint32_t seq1, uint32_t seq2) |
1153 | //{ |
1153 | //{ |
1154 | // return (int32_t)(seq1 - seq2) >= 0; |
1154 | // return (int32_t)(seq1 - seq2) >= 0; |
1155 | //} |
1155 | //} |
1156 | 1156 | ||
1157 | //static inline u32 |
1157 | static inline u32 |
1158 | //i915_gem_next_request_seqno(struct intel_ring_buffer *ring) |
1158 | i915_gem_next_request_seqno(struct intel_ring_buffer *ring) |
1159 | //{ |
1159 | { |
1160 | // drm_i915_private_t *dev_priv = ring->dev->dev_private; |
1160 | drm_i915_private_t *dev_priv = ring->dev->dev_private; |
1161 | // return ring->outstanding_lazy_request = dev_priv->next_seqno; |
1161 | return ring->outstanding_lazy_request = dev_priv->next_seqno; |
- | 1162 | } |
|
1162 | //} |
- | |
1163 | 1163 | ||
1164 | /* |
1164 | |
1165 | void i915_gem_retire_requests(struct drm_device *dev); |
1165 | void i915_gem_retire_requests(struct drm_device *dev); |
1166 | void i915_gem_reset(struct drm_device *dev); |
1166 | void i915_gem_reset(struct drm_device *dev); |
1167 | void i915_gem_clflush_object(struct drm_i915_gem_object *obj); |
1167 | void i915_gem_clflush_object(struct drm_i915_gem_object *obj); |
1168 | int __must_check i915_gem_object_set_domain(struct drm_i915_gem_object *obj, |
1168 | int __must_check i915_gem_object_set_domain(struct drm_i915_gem_object *obj, |
1169 | uint32_t read_domains, |
1169 | uint32_t read_domains, |
1170 | uint32_t write_domain); |
1170 | uint32_t write_domain); |
1171 | int __must_check i915_gem_object_finish_gpu(struct drm_i915_gem_object *obj); |
1171 | int __must_check i915_gem_object_finish_gpu(struct drm_i915_gem_object *obj); |
1172 | int __must_check i915_gem_init_ringbuffer(struct drm_device *dev); |
1172 | int __must_check i915_gem_init_ringbuffer(struct drm_device *dev); |
1173 | void i915_gem_cleanup_ringbuffer(struct drm_device *dev); |
1173 | void i915_gem_cleanup_ringbuffer(struct drm_device *dev); |
1174 | void i915_gem_do_init(struct drm_device *dev, |
1174 | void i915_gem_do_init(struct drm_device *dev, |
1175 | unsigned long start, |
1175 | unsigned long start, |
1176 | unsigned long mappable_end, |
1176 | unsigned long mappable_end, |
1177 | unsigned long end); |
1177 | unsigned long end); |
1178 | int __must_check i915_gpu_idle(struct drm_device *dev); |
1178 | int __must_check i915_gpu_idle(struct drm_device *dev); |
1179 | int __must_check i915_gem_idle(struct drm_device *dev); |
1179 | int __must_check i915_gem_idle(struct drm_device *dev); |
1180 | int __must_check i915_add_request(struct intel_ring_buffer *ring, |
1180 | int __must_check i915_add_request(struct intel_ring_buffer *ring, |
1181 | struct drm_file *file, |
1181 | struct drm_file *file, |
1182 | struct drm_i915_gem_request *request); |
1182 | struct drm_i915_gem_request *request); |
1183 | int __must_check i915_wait_request(struct intel_ring_buffer *ring, |
1183 | int __must_check i915_wait_request(struct intel_ring_buffer *ring, |
1184 | uint32_t seqno); |
1184 | uint32_t seqno); |
1185 | int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); |
1185 | int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); |
1186 | int __must_check |
1186 | int __must_check |
1187 | i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, |
1187 | i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, |
1188 | bool write); |
1188 | bool write); |
1189 | int __must_check |
1189 | int __must_check |
1190 | i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj, |
1190 | i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj, |
1191 | u32 alignment, |
1191 | u32 alignment, |
1192 | struct intel_ring_buffer *pipelined); |
1192 | struct intel_ring_buffer *pipelined); |
1193 | int i915_gem_attach_phys_object(struct drm_device *dev, |
1193 | int i915_gem_attach_phys_object(struct drm_device *dev, |
1194 | struct drm_i915_gem_object *obj, |
1194 | struct drm_i915_gem_object *obj, |
1195 | int id, |
1195 | int id, |
1196 | int align); |
1196 | int align); |
1197 | void i915_gem_detach_phys_object(struct drm_device *dev, |
1197 | void i915_gem_detach_phys_object(struct drm_device *dev, |
1198 | struct drm_i915_gem_object *obj); |
1198 | struct drm_i915_gem_object *obj); |
1199 | void i915_gem_free_all_phys_object(struct drm_device *dev); |
1199 | void i915_gem_free_all_phys_object(struct drm_device *dev); |
1200 | void i915_gem_release(struct drm_device *dev, struct drm_file *file); |
1200 | void i915_gem_release(struct drm_device *dev, struct drm_file *file); |
1201 | 1201 | ||
1202 | uint32_t |
1202 | uint32_t |
1203 | i915_gem_get_unfenced_gtt_alignment(struct drm_device *dev, |
1203 | i915_gem_get_unfenced_gtt_alignment(struct drm_device *dev, |
1204 | uint32_t size, |
1204 | uint32_t size, |
1205 | int tiling_mode); |
1205 | int tiling_mode); |
1206 | 1206 | ||
1207 | int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, |
1207 | int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, |
1208 | enum i915_cache_level cache_level); |
1208 | enum i915_cache_level cache_level); |
1209 | */ |
1209 | |
1210 | 1210 | ||
1211 | /* i915_gem_gtt.c */ |
1211 | /* i915_gem_gtt.c */ |
1212 | void i915_gem_restore_gtt_mappings(struct drm_device *dev); |
1212 | void i915_gem_restore_gtt_mappings(struct drm_device *dev); |
1213 | int __must_check i915_gem_gtt_bind_object(struct drm_i915_gem_object *obj); |
1213 | int __must_check i915_gem_gtt_bind_object(struct drm_i915_gem_object *obj); |
1214 | void i915_gem_gtt_rebind_object(struct drm_i915_gem_object *obj, |
1214 | void i915_gem_gtt_rebind_object(struct drm_i915_gem_object *obj, |
1215 | enum i915_cache_level cache_level); |
1215 | enum i915_cache_level cache_level); |
1216 | void i915_gem_gtt_unbind_object(struct drm_i915_gem_object *obj); |
1216 | void i915_gem_gtt_unbind_object(struct drm_i915_gem_object *obj); |
1217 | 1217 | ||
1218 | /* i915_gem_evict.c */ |
1218 | /* i915_gem_evict.c */ |
1219 | int __must_check i915_gem_evict_something(struct drm_device *dev, int min_size, |
1219 | int __must_check i915_gem_evict_something(struct drm_device *dev, int min_size, |
1220 | unsigned alignment, bool mappable); |
1220 | unsigned alignment, bool mappable); |
1221 | int __must_check i915_gem_evict_everything(struct drm_device *dev, |
1221 | int __must_check i915_gem_evict_everything(struct drm_device *dev, |
1222 | bool purgeable_only); |
1222 | bool purgeable_only); |
1223 | int __must_check i915_gem_evict_inactive(struct drm_device *dev, |
1223 | int __must_check i915_gem_evict_inactive(struct drm_device *dev, |
1224 | bool purgeable_only); |
1224 | bool purgeable_only); |
1225 | 1225 | ||
1226 | /* i915_gem_tiling.c */ |
1226 | /* i915_gem_tiling.c */ |
1227 | void i915_gem_detect_bit_6_swizzle(struct drm_device *dev); |
1227 | void i915_gem_detect_bit_6_swizzle(struct drm_device *dev); |
1228 | void i915_gem_object_do_bit_17_swizzle(struct drm_i915_gem_object *obj); |
1228 | void i915_gem_object_do_bit_17_swizzle(struct drm_i915_gem_object *obj); |
1229 | void i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj); |
1229 | void i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj); |
1230 | 1230 | ||
1231 | /* i915_gem_debug.c */ |
1231 | /* i915_gem_debug.c */ |
1232 | void i915_gem_dump_object(struct drm_i915_gem_object *obj, int len, |
1232 | void i915_gem_dump_object(struct drm_i915_gem_object *obj, int len, |
1233 | const char *where, uint32_t mark); |
1233 | const char *where, uint32_t mark); |
1234 | #if WATCH_LISTS |
1234 | #if WATCH_LISTS |
1235 | int i915_verify_lists(struct drm_device *dev); |
1235 | int i915_verify_lists(struct drm_device *dev); |
1236 | #else |
1236 | #else |
1237 | #define i915_verify_lists(dev) 0 |
1237 | #define i915_verify_lists(dev) 0 |
1238 | #endif |
1238 | #endif |
1239 | void i915_gem_object_check_coherency(struct drm_i915_gem_object *obj, |
1239 | void i915_gem_object_check_coherency(struct drm_i915_gem_object *obj, |
1240 | int handle); |
1240 | int handle); |
1241 | void i915_gem_dump_object(struct drm_i915_gem_object *obj, int len, |
1241 | void i915_gem_dump_object(struct drm_i915_gem_object *obj, int len, |
1242 | const char *where, uint32_t mark); |
1242 | const char *where, uint32_t mark); |
1243 | 1243 | ||
1244 | /* i915_debugfs.c */ |
1244 | /* i915_debugfs.c */ |
1245 | int i915_debugfs_init(struct drm_minor *minor); |
1245 | int i915_debugfs_init(struct drm_minor *minor); |
1246 | void i915_debugfs_cleanup(struct drm_minor *minor); |
1246 | void i915_debugfs_cleanup(struct drm_minor *minor); |
1247 | 1247 | ||
1248 | /* i915_suspend.c */ |
1248 | /* i915_suspend.c */ |
1249 | extern int i915_save_state(struct drm_device *dev); |
1249 | extern int i915_save_state(struct drm_device *dev); |
1250 | extern int i915_restore_state(struct drm_device *dev); |
1250 | extern int i915_restore_state(struct drm_device *dev); |
1251 | 1251 | ||
1252 | /* i915_suspend.c */ |
1252 | /* i915_suspend.c */ |
1253 | extern int i915_save_state(struct drm_device *dev); |
1253 | extern int i915_save_state(struct drm_device *dev); |
1254 | extern int i915_restore_state(struct drm_device *dev); |
1254 | extern int i915_restore_state(struct drm_device *dev); |
1255 | 1255 | ||
1256 | /* intel_i2c.c */ |
1256 | /* intel_i2c.c */ |
1257 | extern int intel_setup_gmbus(struct drm_device *dev); |
1257 | extern int intel_setup_gmbus(struct drm_device *dev); |
1258 | extern void intel_teardown_gmbus(struct drm_device *dev); |
1258 | extern void intel_teardown_gmbus(struct drm_device *dev); |
1259 | extern void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed); |
1259 | extern void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed); |
1260 | extern void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit); |
1260 | extern void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit); |
1261 | 1261 | ||
1262 | //extern inline bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter) |
1262 | //extern inline bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter) |
1263 | //{ |
1263 | //{ |
1264 | // return container_of(adapter, struct intel_gmbus, adapter)->force_bit; |
1264 | // return container_of(adapter, struct intel_gmbus, adapter)->force_bit; |
1265 | //} |
1265 | //} |
1266 | 1266 | ||
1267 | extern void intel_i2c_reset(struct drm_device *dev); |
1267 | extern void intel_i2c_reset(struct drm_device *dev); |
1268 | 1268 | ||
1269 | /* intel_opregion.c */ |
1269 | /* intel_opregion.c */ |
1270 | extern int intel_opregion_setup(struct drm_device *dev); |
1270 | extern int intel_opregion_setup(struct drm_device *dev); |
1271 | #ifdef CONFIG_ACPI |
1271 | #ifdef CONFIG_ACPI |
1272 | extern void intel_opregion_init(struct drm_device *dev); |
1272 | extern void intel_opregion_init(struct drm_device *dev); |
1273 | extern void intel_opregion_fini(struct drm_device *dev); |
1273 | extern void intel_opregion_fini(struct drm_device *dev); |
1274 | extern void intel_opregion_asle_intr(struct drm_device *dev); |
1274 | extern void intel_opregion_asle_intr(struct drm_device *dev); |
1275 | extern void intel_opregion_gse_intr(struct drm_device *dev); |
1275 | extern void intel_opregion_gse_intr(struct drm_device *dev); |
1276 | extern void intel_opregion_enable_asle(struct drm_device *dev); |
1276 | extern void intel_opregion_enable_asle(struct drm_device *dev); |
1277 | #else |
1277 | #else |
1278 | static inline void intel_opregion_init(struct drm_device *dev) { return; } |
1278 | static inline void intel_opregion_init(struct drm_device *dev) { return; } |
1279 | static inline void intel_opregion_fini(struct drm_device *dev) { return; } |
1279 | static inline void intel_opregion_fini(struct drm_device *dev) { return; } |
1280 | static inline void intel_opregion_asle_intr(struct drm_device *dev) { return; } |
1280 | static inline void intel_opregion_asle_intr(struct drm_device *dev) { return; } |
1281 | static inline void intel_opregion_gse_intr(struct drm_device *dev) { return; } |
1281 | static inline void intel_opregion_gse_intr(struct drm_device *dev) { return; } |
1282 | static inline void intel_opregion_enable_asle(struct drm_device *dev) { return; } |
1282 | static inline void intel_opregion_enable_asle(struct drm_device *dev) { return; } |
1283 | #endif |
1283 | #endif |
1284 | 1284 | ||
1285 | /* intel_acpi.c */ |
1285 | /* intel_acpi.c */ |
1286 | #ifdef CONFIG_ACPI |
1286 | #ifdef CONFIG_ACPI |
1287 | extern void intel_register_dsm_handler(void); |
1287 | extern void intel_register_dsm_handler(void); |
1288 | extern void intel_unregister_dsm_handler(void); |
1288 | extern void intel_unregister_dsm_handler(void); |
1289 | #else |
1289 | #else |
1290 | static inline void intel_register_dsm_handler(void) { return; } |
1290 | static inline void intel_register_dsm_handler(void) { return; } |
1291 | static inline void intel_unregister_dsm_handler(void) { return; } |
1291 | static inline void intel_unregister_dsm_handler(void) { return; } |
1292 | #endif /* CONFIG_ACPI */ |
1292 | #endif /* CONFIG_ACPI */ |
1293 | 1293 | ||
1294 | /* modesetting */ |
1294 | /* modesetting */ |
1295 | extern void intel_modeset_init(struct drm_device *dev); |
1295 | extern void intel_modeset_init(struct drm_device *dev); |
1296 | extern void intel_modeset_gem_init(struct drm_device *dev); |
1296 | extern void intel_modeset_gem_init(struct drm_device *dev); |
1297 | extern void intel_modeset_cleanup(struct drm_device *dev); |
1297 | extern void intel_modeset_cleanup(struct drm_device *dev); |
1298 | extern int intel_modeset_vga_set_state(struct drm_device *dev, bool state); |
1298 | extern int intel_modeset_vga_set_state(struct drm_device *dev, bool state); |
1299 | extern bool intel_fbc_enabled(struct drm_device *dev); |
1299 | extern bool intel_fbc_enabled(struct drm_device *dev); |
1300 | extern void intel_disable_fbc(struct drm_device *dev); |
1300 | extern void intel_disable_fbc(struct drm_device *dev); |
1301 | extern bool ironlake_set_drps(struct drm_device *dev, u8 val); |
1301 | extern bool ironlake_set_drps(struct drm_device *dev, u8 val); |
1302 | extern void ironlake_enable_rc6(struct drm_device *dev); |
1302 | extern void ironlake_enable_rc6(struct drm_device *dev); |
1303 | extern void gen6_set_rps(struct drm_device *dev, u8 val); |
1303 | extern void gen6_set_rps(struct drm_device *dev, u8 val); |
1304 | extern void intel_detect_pch (struct drm_device *dev); |
1304 | extern void intel_detect_pch (struct drm_device *dev); |
1305 | extern int intel_trans_dp_port_sel (struct drm_crtc *crtc); |
1305 | extern int intel_trans_dp_port_sel (struct drm_crtc *crtc); |
1306 | 1306 | ||
1307 | /* overlay */ |
1307 | /* overlay */ |
1308 | #ifdef CONFIG_DEBUG_FS |
1308 | #ifdef CONFIG_DEBUG_FS |
1309 | extern struct intel_overlay_error_state *intel_overlay_capture_error_state(struct drm_device *dev); |
1309 | extern struct intel_overlay_error_state *intel_overlay_capture_error_state(struct drm_device *dev); |
1310 | extern void intel_overlay_print_error_state(struct seq_file *m, struct intel_overlay_error_state *error); |
1310 | extern void intel_overlay_print_error_state(struct seq_file *m, struct intel_overlay_error_state *error); |
1311 | 1311 | ||
1312 | extern struct intel_display_error_state *intel_display_capture_error_state(struct drm_device *dev); |
1312 | extern struct intel_display_error_state *intel_display_capture_error_state(struct drm_device *dev); |
1313 | extern void intel_display_print_error_state(struct seq_file *m, |
1313 | extern void intel_display_print_error_state(struct seq_file *m, |
1314 | struct drm_device *dev, |
1314 | struct drm_device *dev, |
1315 | struct intel_display_error_state *error); |
1315 | struct intel_display_error_state *error); |
1316 | #endif |
1316 | #endif |
1317 | 1317 | ||
1318 | #define LP_RING(d) (&((struct drm_i915_private *)(d))->ring[RCS]) |
1318 | #define LP_RING(d) (&((struct drm_i915_private *)(d))->ring[RCS]) |
1319 | 1319 | ||
1320 | #define BEGIN_LP_RING(n) \ |
1320 | #define BEGIN_LP_RING(n) \ |
1321 | intel_ring_begin(LP_RING(dev_priv), (n)) |
1321 | intel_ring_begin(LP_RING(dev_priv), (n)) |
1322 | 1322 | ||
1323 | #define OUT_RING(x) \ |
1323 | #define OUT_RING(x) \ |
1324 | intel_ring_emit(LP_RING(dev_priv), x) |
1324 | intel_ring_emit(LP_RING(dev_priv), x) |
1325 | 1325 | ||
1326 | #define ADVANCE_LP_RING() \ |
1326 | #define ADVANCE_LP_RING() \ |
1327 | intel_ring_advance(LP_RING(dev_priv)) |
1327 | intel_ring_advance(LP_RING(dev_priv)) |
1328 | 1328 | ||
1329 | /** |
1329 | /** |
1330 | * Lock test for when it's just for synchronization of ring access. |
1330 | * Lock test for when it's just for synchronization of ring access. |
1331 | * |
1331 | * |
1332 | * In that case, we don't need to do it when GEM is initialized as nobody else |
1332 | * In that case, we don't need to do it when GEM is initialized as nobody else |
1333 | * has access to the ring. |
1333 | * has access to the ring. |
1334 | */ |
1334 | */ |
1335 | #define RING_LOCK_TEST_WITH_RETURN(dev, file) do { \ |
1335 | #define RING_LOCK_TEST_WITH_RETURN(dev, file) do { \ |
1336 | if (LP_RING(dev->dev_private)->obj == NULL) \ |
1336 | if (LP_RING(dev->dev_private)->obj == NULL) \ |
1337 | LOCK_TEST_WITH_RETURN(dev, file); \ |
1337 | LOCK_TEST_WITH_RETURN(dev, file); \ |
1338 | } while (0) |
1338 | } while (0) |
1339 | 1339 | ||
1340 | /* On SNB platform, before reading ring registers forcewake bit |
1340 | /* On SNB platform, before reading ring registers forcewake bit |
1341 | * must be set to prevent GT core from power down and stale values being |
1341 | * must be set to prevent GT core from power down and stale values being |
1342 | * returned. |
1342 | * returned. |
1343 | */ |
1343 | */ |
1344 | void gen6_gt_force_wake_get(struct drm_i915_private *dev_priv); |
1344 | void gen6_gt_force_wake_get(struct drm_i915_private *dev_priv); |
1345 | void gen6_gt_force_wake_put(struct drm_i915_private *dev_priv); |
1345 | void gen6_gt_force_wake_put(struct drm_i915_private *dev_priv); |
1346 | void __gen6_gt_wait_for_fifo(struct drm_i915_private *dev_priv); |
1346 | void __gen6_gt_wait_for_fifo(struct drm_i915_private *dev_priv); |
1347 | 1347 | ||
1348 | /* We give fast paths for the really cool registers */ |
1348 | /* We give fast paths for the really cool registers */ |
1349 | #define NEEDS_FORCE_WAKE(dev_priv, reg) \ |
1349 | #define NEEDS_FORCE_WAKE(dev_priv, reg) \ |
1350 | (((dev_priv)->info->gen >= 6) && \ |
1350 | (((dev_priv)->info->gen >= 6) && \ |
1351 | ((reg) < 0x40000) && \ |
1351 | ((reg) < 0x40000) && \ |
1352 | ((reg) != FORCEWAKE)) |
1352 | ((reg) != FORCEWAKE)) |
1353 | 1353 | ||
1354 | #define __i915_read(x, y) \ |
1354 | #define __i915_read(x, y) \ |
1355 | static inline u##x i915_read##x(struct drm_i915_private *dev_priv, u32 reg) { \ |
1355 | static inline u##x i915_read##x(struct drm_i915_private *dev_priv, u32 reg) { \ |
1356 | u##x val = 0; \ |
1356 | u##x val = 0; \ |
1357 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ |
1357 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ |
1358 | gen6_gt_force_wake_get(dev_priv); \ |
1358 | gen6_gt_force_wake_get(dev_priv); \ |
1359 | val = read##y(dev_priv->regs + reg); \ |
1359 | val = read##y(dev_priv->regs + reg); \ |
1360 | gen6_gt_force_wake_put(dev_priv); \ |
1360 | gen6_gt_force_wake_put(dev_priv); \ |
1361 | } else { \ |
1361 | } else { \ |
1362 | val = read##y(dev_priv->regs + reg); \ |
1362 | val = read##y(dev_priv->regs + reg); \ |
1363 | } \ |
1363 | } \ |
1364 | /* trace_i915_reg_rw(false, reg, val, sizeof(val)); */\ |
1364 | /* trace_i915_reg_rw(false, reg, val, sizeof(val)); */\ |
1365 | return val; \ |
1365 | return val; \ |
1366 | } |
1366 | } |
1367 | 1367 | ||
1368 | __i915_read(8, b) |
1368 | __i915_read(8, b) |
1369 | __i915_read(16, w) |
1369 | __i915_read(16, w) |
1370 | __i915_read(32, l) |
1370 | __i915_read(32, l) |
1371 | __i915_read(64, q) |
1371 | __i915_read(64, q) |
1372 | #undef __i915_read |
1372 | #undef __i915_read |
1373 | 1373 | ||
1374 | #define __i915_write(x, y) \ |
1374 | #define __i915_write(x, y) \ |
1375 | static inline void i915_write##x(struct drm_i915_private *dev_priv, u32 reg, u##x val) { \ |
1375 | static inline void i915_write##x(struct drm_i915_private *dev_priv, u32 reg, u##x val) { \ |
1376 | /* trace_i915_reg_rw(true, reg, val, sizeof(val));*/ \ |
1376 | /* trace_i915_reg_rw(true, reg, val, sizeof(val));*/ \ |
1377 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ |
1377 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ |
1378 | __gen6_gt_wait_for_fifo(dev_priv); \ |
1378 | __gen6_gt_wait_for_fifo(dev_priv); \ |
1379 | } \ |
1379 | } \ |
1380 | write##y(val, dev_priv->regs + reg); \ |
1380 | write##y(val, dev_priv->regs + reg); \ |
1381 | } |
1381 | } |
1382 | __i915_write(8, b) |
1382 | __i915_write(8, b) |
1383 | __i915_write(16, w) |
1383 | __i915_write(16, w) |
1384 | __i915_write(32, l) |
1384 | __i915_write(32, l) |
1385 | __i915_write(64, q) |
1385 | __i915_write(64, q) |
1386 | #undef __i915_write |
1386 | #undef __i915_write |
1387 | 1387 | ||
1388 | #define I915_READ8(reg) i915_read8(dev_priv, (reg)) |
1388 | #define I915_READ8(reg) i915_read8(dev_priv, (reg)) |
1389 | #define I915_WRITE8(reg, val) i915_write8(dev_priv, (reg), (val)) |
1389 | #define I915_WRITE8(reg, val) i915_write8(dev_priv, (reg), (val)) |
1390 | 1390 | ||
1391 | #define I915_READ16(reg) i915_read16(dev_priv, (reg)) |
1391 | #define I915_READ16(reg) i915_read16(dev_priv, (reg)) |
1392 | #define I915_WRITE16(reg, val) i915_write16(dev_priv, (reg), (val)) |
1392 | #define I915_WRITE16(reg, val) i915_write16(dev_priv, (reg), (val)) |
1393 | #define I915_READ16_NOTRACE(reg) readw(dev_priv->regs + (reg)) |
1393 | #define I915_READ16_NOTRACE(reg) readw(dev_priv->regs + (reg)) |
1394 | #define I915_WRITE16_NOTRACE(reg, val) writew(val, dev_priv->regs + (reg)) |
1394 | #define I915_WRITE16_NOTRACE(reg, val) writew(val, dev_priv->regs + (reg)) |
1395 | 1395 | ||
1396 | #define I915_READ(reg) i915_read32(dev_priv, (reg)) |
1396 | #define I915_READ(reg) i915_read32(dev_priv, (reg)) |
1397 | #define I915_WRITE(reg, val) i915_write32(dev_priv, (reg), (val)) |
1397 | #define I915_WRITE(reg, val) i915_write32(dev_priv, (reg), (val)) |
1398 | #define I915_READ_NOTRACE(reg) readl(dev_priv->regs + (reg)) |
1398 | #define I915_READ_NOTRACE(reg) readl(dev_priv->regs + (reg)) |
1399 | #define I915_WRITE_NOTRACE(reg, val) writel(val, dev_priv->regs + (reg)) |
1399 | #define I915_WRITE_NOTRACE(reg, val) writel(val, dev_priv->regs + (reg)) |
1400 | 1400 | ||
1401 | #define I915_WRITE64(reg, val) i915_write64(dev_priv, (reg), (val)) |
1401 | #define I915_WRITE64(reg, val) i915_write64(dev_priv, (reg), (val)) |
1402 | #define I915_READ64(reg) i915_read64(dev_priv, (reg)) |
1402 | #define I915_READ64(reg) i915_read64(dev_priv, (reg)) |
1403 | 1403 | ||
1404 | #define POSTING_READ(reg) (void)I915_READ_NOTRACE(reg) |
1404 | #define POSTING_READ(reg) (void)I915_READ_NOTRACE(reg) |
1405 | #define POSTING_READ16(reg) (void)I915_READ16_NOTRACE(reg) |
1405 | #define POSTING_READ16(reg) (void)I915_READ16_NOTRACE(reg) |
1406 | 1406 | ||
1407 | 1407 | ||
1408 | #endif>1) |
1408 | #endif>1) |
1409 | 1409 | ||
1410 | struct><1) |
1410 | struct><1) |
1411 | 1411 | ||
1412 | struct>0) |
1412 | struct>0) |
1413 | #define><0) |
1413 | #define><0) |
1414 | #define>> |
1414 | #define>> |