138,7 → 138,6 |
dev_priv->mm.object_memory -= size; |
} |
|
#if 0 |
|
static int |
i915_gem_wait_for_error(struct drm_device *dev) |
150,7 → 149,7 |
|
if (!atomic_read(&dev_priv->mm.wedged)) |
return 0; |
|
#if 0 |
/* |
* Only wait 10 seconds for the gpu reset to complete to avoid hanging |
* userspace. If it takes that long something really bad is going on and |
174,6 → 173,8 |
x->done++; |
spin_unlock_irqrestore(&x->wait.lock, flags); |
} |
#endif |
|
return 0; |
} |
|
185,14 → 186,11 |
if (ret) |
return ret; |
|
ret = mutex_lock_interruptible(&dev->struct_mutex); |
if (ret) |
return ret; |
mutex_lock(&dev->struct_mutex); |
|
WARN_ON(i915_verify_lists(dev)); |
return 0; |
} |
#endif |
|
static inline bool |
i915_gem_object_is_inactive(struct drm_i915_gem_object *obj) |
251,7 → 249,6 |
return 0; |
} |
|
#if 0 |
static int |
i915_gem_create(struct drm_file *file, |
struct drm_device *dev, |
284,6 → 281,7 |
trace_i915_gem_object_create(obj); |
|
*handle_p = handle; |
|
return 0; |
} |
|
319,6 → 317,8 |
args->size, &args->handle); |
} |
|
#if 0 |
|
static int i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_object *obj) |
{ |
drm_i915_private_t *dev_priv = obj->base.dev->dev_private; |
1473,7 → 1473,7 |
* Fail silently without starting the shrinker |
*/ |
for_each_sg(st->sgl, sg, page_count, i) { |
page = AllocPage(); // oh-oh |
page = (struct page *)AllocPage(); // oh-oh |
if ( page == 0 ) |
goto err_pages; |
|
3054,7 → 3054,6 |
obj->pin_mappable = false; |
} |
|
#if 0 |
int |
i915_gem_pin_ioctl(struct drm_device *dev, void *data, |
struct drm_file *file) |
3107,6 → 3106,8 |
return ret; |
} |
|
#if 0 |
|
int |
i915_gem_unpin_ioctl(struct drm_device *dev, void *data, |
struct drm_file *file) |