Rev 2338 | Rev 2342 | Go to most recent revision | Show entire file | Regard whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 2338 | Rev 2340 | ||
---|---|---|---|
Line 288... | Line 288... | ||
288 | struct pci_dev *bridge_dev; |
288 | struct pci_dev *bridge_dev; |
289 | struct intel_ring_buffer ring[I915_NUM_RINGS]; |
289 | struct intel_ring_buffer ring[I915_NUM_RINGS]; |
290 | uint32_t next_seqno; |
290 | uint32_t next_seqno; |
Line 291... | Line 291... | ||
291 | 291 | ||
292 | drm_dma_handle_t *status_page_dmah; |
292 | drm_dma_handle_t *status_page_dmah; |
293 | // uint32_t counter; |
293 | uint32_t counter; |
294 | // drm_local_map_t hws_map; |
294 | drm_local_map_t hws_map; |
295 | struct drm_i915_gem_object *pwrctx; |
295 | struct drm_i915_gem_object *pwrctx; |
Line 296... | Line 296... | ||
296 | struct drm_i915_gem_object *renderctx; |
296 | struct drm_i915_gem_object *renderctx; |
Line 881... | Line 881... | ||
881 | * reaches 0, dev_priv->pending_flip_queue will be woken up. |
881 | * reaches 0, dev_priv->pending_flip_queue will be woken up. |
882 | */ |
882 | */ |
883 | atomic_t pending_flip; |
883 | atomic_t pending_flip; |
884 | }; |
884 | }; |
Line 885... | Line -... | ||
885 | - | ||
886 | 885 | ||
Line 887... | Line 886... | ||
887 | #define to_intel_bo(x) container_of(x, struct drm_i915_gem_object, base) |
886 | #define to_intel_bo(x) container_of(x, struct drm_i915_gem_object, base) |
888 | 887 | ||
889 | /** |
888 | /** |
Line 1145... | Line 1144... | ||
1145 | int i915_gem_dumb_destroy(struct drm_file *file_priv, struct drm_device *dev, |
1144 | int i915_gem_dumb_destroy(struct drm_file *file_priv, struct drm_device *dev, |
1146 | uint32_t handle); |
1145 | uint32_t handle); |
1147 | /** |
1146 | /** |
1148 | * Returns true if seq1 is later than seq2. |
1147 | * Returns true if seq1 is later than seq2. |
1149 | */ |
1148 | */ |
1150 | //static inline bool |
1149 | static inline bool |
1151 | //i915_seqno_passed(uint32_t seq1, uint32_t seq2) |
1150 | i915_seqno_passed(uint32_t seq1, uint32_t seq2) |
1152 | //{ |
1151 | { |
1153 | // return (int32_t)(seq1 - seq2) >= 0; |
1152 | return (int32_t)(seq1 - seq2) >= 0; |
1154 | //} |
1153 | } |
Line 1155... | Line 1154... | ||
1155 | 1154 | ||
1156 | static inline u32 |
1155 | static inline u32 |
1157 | i915_gem_next_request_seqno(struct intel_ring_buffer *ring) |
1156 | i915_gem_next_request_seqno(struct intel_ring_buffer *ring) |
1158 | { |
1157 | { |
Line 1204... | Line 1203... | ||
1204 | int tiling_mode); |
1203 | int tiling_mode); |
Line 1205... | Line 1204... | ||
1205 | 1204 | ||
1206 | int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, |
1205 | int i915_gem_object_set_cache_level(struct drm_i915_gem_object *obj, |
Line 1207... | Line -... | ||
1207 | enum i915_cache_level cache_level); |
- | |
1208 | 1206 | enum i915_cache_level cache_level); |
|
1209 | 1207 | ||
1210 | /* i915_gem_gtt.c */ |
1208 | /* i915_gem_gtt.c */ |
1211 | void i915_gem_restore_gtt_mappings(struct drm_device *dev); |
1209 | void i915_gem_restore_gtt_mappings(struct drm_device *dev); |
1212 | int __must_check i915_gem_gtt_bind_object(struct drm_i915_gem_object *obj); |
1210 | int __must_check i915_gem_gtt_bind_object(struct drm_i915_gem_object *obj); |