diff options
| author | Dave Airlie <airlied@linux.ie> | 2007-11-05 19:09:18 +1000 | 
|---|---|---|
| committer | Dave Airlie <airlied@linux.ie> | 2007-11-05 19:09:18 +1000 | 
| commit | 6ee5412da0cc7516472235805482b8632cb374ef (patch) | |
| tree | a3d4f61858d3c8b2b6bfe3b8078badfcccde711d /linux-core | |
| parent | 7ad38907075852e347b5b4432c96d199387f5ce8 (diff) | |
drm/ttm: apply linux kernel coding style to bo_lock/move/object/ttm.c
Diffstat (limited to 'linux-core')
| -rw-r--r-- | linux-core/drm_bo_lock.c | 9 | ||||
| -rw-r--r-- | linux-core/drm_bo_move.c | 59 | ||||
| -rw-r--r-- | linux-core/drm_object.c | 36 | ||||
| -rw-r--r-- | linux-core/drm_ttm.c | 63 | 
4 files changed, 76 insertions, 91 deletions
diff --git a/linux-core/drm_bo_lock.c b/linux-core/drm_bo_lock.c index 46318f6a..f967fb7c 100644 --- a/linux-core/drm_bo_lock.c +++ b/linux-core/drm_bo_lock.c @@ -73,7 +73,6 @@ void drm_bo_read_unlock(struct drm_bo_lock *lock)  	if (atomic_read(&lock->readers) == 0)  		wake_up_interruptible(&lock->queue);  } -  EXPORT_SYMBOL(drm_bo_read_unlock);  int drm_bo_read_lock(struct drm_bo_lock *lock) @@ -95,7 +94,6 @@ int drm_bo_read_lock(struct drm_bo_lock *lock)  	}  	return 0;  } -  EXPORT_SYMBOL(drm_bo_read_lock);  static int __drm_bo_write_unlock(struct drm_bo_lock *lock) @@ -123,9 +121,8 @@ int drm_bo_write_lock(struct drm_bo_lock *lock, struct drm_file *file_priv)  	int ret = 0;  	struct drm_device *dev; -	if (unlikely(atomic_cmpxchg(&lock->write_lock_pending, 0, 1) != 0)) { +	if (unlikely(atomic_cmpxchg(&lock->write_lock_pending, 0, 1) != 0))  		return -EINVAL; -	}  	while (unlikely(atomic_cmpxchg(&lock->readers, 0, -1) != 0)) {  		ret = wait_event_interruptible @@ -149,9 +146,9 @@ int drm_bo_write_lock(struct drm_bo_lock *lock, struct drm_file *file_priv)  	ret = drm_add_user_object(file_priv, &lock->base, 0);  	lock->base.remove = &drm_bo_write_lock_remove;  	lock->base.type = drm_lock_type; -	if (ret) { +	if (ret)  		(void)__drm_bo_write_unlock(lock); -	} +  	mutex_unlock(&dev->struct_mutex);  	return ret; diff --git a/linux-core/drm_bo_move.c b/linux-core/drm_bo_move.c index 9ab28b03..2aba3271 100644 --- a/linux-core/drm_bo_move.c +++ b/linux-core/drm_bo_move.c @@ -35,7 +35,7 @@   * have not been requested to free also pinned regions.   */ -static void drm_bo_free_old_node(struct drm_buffer_object * bo) +static void drm_bo_free_old_node(struct drm_buffer_object *bo)  {  	struct drm_bo_mem_reg *old_mem = &bo->mem; @@ -48,8 +48,8 @@ static void drm_bo_free_old_node(struct drm_buffer_object * bo)  	old_mem->mm_node = NULL;  } -int drm_bo_move_ttm(struct drm_buffer_object * bo, -		    int evict, int no_wait, struct drm_bo_mem_reg * new_mem) +int drm_bo_move_ttm(struct drm_buffer_object *bo, +		    int evict, int no_wait, struct drm_bo_mem_reg *new_mem)  {  	struct drm_ttm *ttm = bo->ttm;  	struct drm_bo_mem_reg *old_mem = &bo->mem; @@ -82,7 +82,6 @@ int drm_bo_move_ttm(struct drm_buffer_object * bo,  	DRM_FLAG_MASKED(save_flags, new_mem->flags, DRM_BO_MASK_MEMTYPE);  	return 0;  } -  EXPORT_SYMBOL(drm_bo_move_ttm);  /** @@ -100,7 +99,7 @@ EXPORT_SYMBOL(drm_bo_move_ttm);   * Call bo->mutex locked.   */ -int drm_mem_reg_ioremap(struct drm_device * dev, struct drm_bo_mem_reg * mem, +int drm_mem_reg_ioremap(struct drm_device *dev, struct drm_bo_mem_reg *mem,  			void **virtual)  {  	struct drm_buffer_manager *bm = &dev->bm; @@ -136,7 +135,7 @@ EXPORT_SYMBOL(drm_mem_reg_ioremap);   * Call bo->mutex locked.   */ -void drm_mem_reg_iounmap(struct drm_device * dev, struct drm_bo_mem_reg * mem, +void drm_mem_reg_iounmap(struct drm_device *dev, struct drm_bo_mem_reg *mem,  			 void *virtual)  {  	struct drm_buffer_manager *bm; @@ -145,9 +144,8 @@ void drm_mem_reg_iounmap(struct drm_device * dev, struct drm_bo_mem_reg * mem,  	bm = &dev->bm;  	man = &bm->man[mem->mem_type]; -	if (virtual && (man->flags & _DRM_FLAG_NEEDS_IOREMAP)) { +	if (virtual && (man->flags & _DRM_FLAG_NEEDS_IOREMAP))  		iounmap(virtual); -	}  }  static int drm_copy_io_page(void *dst, void *src, unsigned long page) @@ -163,7 +161,8 @@ static int drm_copy_io_page(void *dst, void *src, unsigned long page)  	return 0;  } -static int drm_copy_io_ttm_page(struct drm_ttm * ttm, void *src, unsigned long page) +static int drm_copy_io_ttm_page(struct drm_ttm *ttm, void *src, +				unsigned long page)  {  	struct page *d = drm_ttm_get_page(ttm, page);  	void *dst; @@ -181,7 +180,7 @@ static int drm_copy_io_ttm_page(struct drm_ttm * ttm, void *src, unsigned long p  	return 0;  } -static int drm_copy_ttm_io_page(struct drm_ttm * ttm, void *dst, unsigned long page) +static int drm_copy_ttm_io_page(struct drm_ttm *ttm, void *dst, unsigned long page)  {  	struct page *s = drm_ttm_get_page(ttm, page);  	void *src; @@ -199,8 +198,8 @@ static int drm_copy_ttm_io_page(struct drm_ttm * ttm, void *dst, unsigned long p  	return 0;  } -int drm_bo_move_memcpy(struct drm_buffer_object * bo, -		       int evict, int no_wait, struct drm_bo_mem_reg * new_mem) +int drm_bo_move_memcpy(struct drm_buffer_object *bo, +		       int evict, int no_wait, struct drm_bo_mem_reg *new_mem)  {  	struct drm_device *dev = bo->dev;  	struct drm_mem_type_manager *man = &dev->bm.man[new_mem->mem_type]; @@ -251,7 +250,7 @@ int drm_bo_move_memcpy(struct drm_buffer_object * bo,  			goto out1;  	}  	mb(); -      out2: +out2:  	drm_bo_free_old_node(bo);  	*old_mem = *new_mem; @@ -265,13 +264,12 @@ int drm_bo_move_memcpy(struct drm_buffer_object * bo,  		bo->ttm = NULL;  	} -      out1: +out1:  	drm_mem_reg_iounmap(dev, new_mem, new_iomap); -      out: +out:  	drm_mem_reg_iounmap(dev, &old_copy, old_iomap);  	return ret;  } -  EXPORT_SYMBOL(drm_bo_move_memcpy);  /* @@ -280,8 +278,8 @@ EXPORT_SYMBOL(drm_bo_move_memcpy);   * object. Call bo->mutex locked.   */ -int drm_buffer_object_transfer(struct drm_buffer_object * bo, -			       struct drm_buffer_object ** new_obj) +int drm_buffer_object_transfer(struct drm_buffer_object *bo, +			       struct drm_buffer_object **new_obj)  {  	struct drm_buffer_object *fbo;  	struct drm_device *dev = bo->dev; @@ -322,12 +320,10 @@ int drm_buffer_object_transfer(struct drm_buffer_object * bo,   * We cannot restart until it has finished.   */ -int drm_bo_move_accel_cleanup(struct drm_buffer_object * bo, -			      int evict, -			      int no_wait, -			      uint32_t fence_class, -			      uint32_t fence_type, -			      uint32_t fence_flags, struct drm_bo_mem_reg * new_mem) +int drm_bo_move_accel_cleanup(struct drm_buffer_object *bo, +			      int evict, int no_wait, uint32_t fence_class, +			      uint32_t fence_type, uint32_t fence_flags, +			      struct drm_bo_mem_reg *new_mem)  {  	struct drm_device *dev = bo->dev;  	struct drm_mem_type_manager *man = &dev->bm.man[new_mem->mem_type]; @@ -407,7 +403,6 @@ int drm_bo_move_accel_cleanup(struct drm_buffer_object * bo,  	DRM_FLAG_MASKED(save_flags, new_mem->flags, DRM_BO_MASK_MEMTYPE);  	return 0;  } -  EXPORT_SYMBOL(drm_bo_move_accel_cleanup);  int drm_bo_same_page(unsigned long offset, @@ -420,13 +415,11 @@ EXPORT_SYMBOL(drm_bo_same_page);  unsigned long drm_bo_offset_end(unsigned long offset,  				unsigned long end)  { -  	offset = (offset + PAGE_SIZE) & PAGE_MASK;  	return (end < offset) ? end : offset;  }  EXPORT_SYMBOL(drm_bo_offset_end); -  static pgprot_t drm_kernel_io_prot(uint32_t map_type)  {  	pgprot_t tmp = PAGE_KERNEL; @@ -475,8 +468,9 @@ static int drm_bo_ioremap(struct drm_buffer_object *bo, unsigned long bus_base,  	return (!map->virtual) ? -ENOMEM : 0;  } -static int drm_bo_kmap_ttm(struct drm_buffer_object *bo, unsigned long start_page, -			   unsigned long num_pages, struct drm_bo_kmap_obj *map) +static int drm_bo_kmap_ttm(struct drm_buffer_object *bo, +			   unsigned long start_page, unsigned long num_pages, +			   struct drm_bo_kmap_obj *map)  {  	struct drm_device *dev = bo->dev;  	struct drm_bo_mem_reg *mem = &bo->mem; @@ -503,7 +497,7 @@ static int drm_bo_kmap_ttm(struct drm_buffer_object *bo, unsigned long start_pag  		 * Populate the part we're mapping;  		 */ -		for (i = start_page; i< start_page + num_pages; ++i) { +		for (i = start_page; i < start_page + num_pages; ++i) {  			d = drm_ttm_get_page(ttm, i);  			if (!d)  				return -ENOMEM; @@ -530,7 +524,8 @@ static int drm_bo_kmap_ttm(struct drm_buffer_object *bo, unsigned long start_pag   * and caching policy the buffer currently has.   * Mapping multiple pages or buffers that live in io memory is a bit slow and   * consumes vmalloc space. Be restrictive with such mappings. - * Mapping single pages usually returns the logical kernel address, (which is fast) + * Mapping single pages usually returns the logical kernel address, + * (which is fast)   * BUG may use slower temporary mappings for high memory pages or   * uncached / write-combined pages.   * @@ -581,7 +576,7 @@ void drm_bo_kunmap(struct drm_bo_kmap_obj *map)  	if (!map->virtual)  		return; -	switch(map->bo_kmap_type) { +	switch (map->bo_kmap_type) {  	case bo_map_iomap:  		iounmap(map->virtual);  		break; diff --git a/linux-core/drm_object.c b/linux-core/drm_object.c index a6d6c0d7..5cc1c8d9 100644 --- a/linux-core/drm_object.c +++ b/linux-core/drm_object.c @@ -30,7 +30,7 @@  #include "drmP.h" -int drm_add_user_object(struct drm_file * priv, struct drm_user_object * item, +int drm_add_user_object(struct drm_file *priv, struct drm_user_object *item,  			int shareable)  {  	struct drm_device *dev = priv->head->dev; @@ -56,7 +56,7 @@ int drm_add_user_object(struct drm_file * priv, struct drm_user_object * item,  }  EXPORT_SYMBOL(drm_add_user_object); -struct drm_user_object *drm_lookup_user_object(struct drm_file * priv, uint32_t key) +struct drm_user_object *drm_lookup_user_object(struct drm_file *priv, uint32_t key)  {  	struct drm_device *dev = priv->head->dev;  	struct drm_hash_item *hash; @@ -66,9 +66,9 @@ struct drm_user_object *drm_lookup_user_object(struct drm_file * priv, uint32_t  	DRM_ASSERT_LOCKED(&dev->struct_mutex);  	ret = drm_ht_find_item(&dev->object_hash, key, &hash); -	if (ret) { +	if (ret)  		return NULL; -	} +  	item = drm_hash_entry(hash, struct drm_user_object, hash);  	if (priv != item->owner) { @@ -83,7 +83,7 @@ struct drm_user_object *drm_lookup_user_object(struct drm_file * priv, uint32_t  }  EXPORT_SYMBOL(drm_lookup_user_object); -static void drm_deref_user_object(struct drm_file * priv, struct drm_user_object * item) +static void drm_deref_user_object(struct drm_file *priv, struct drm_user_object *item)  {  	struct drm_device *dev = priv->head->dev;  	int ret; @@ -95,7 +95,7 @@ static void drm_deref_user_object(struct drm_file * priv, struct drm_user_object  	}  } -static int drm_object_ref_action(struct drm_file * priv, struct drm_user_object * ro, +static int drm_object_ref_action(struct drm_file *priv, struct drm_user_object *ro,  				 enum drm_ref_type action)  {  	int ret = 0; @@ -114,7 +114,7 @@ static int drm_object_ref_action(struct drm_file * priv, struct drm_user_object  	return ret;  } -int drm_add_ref_object(struct drm_file * priv, struct drm_user_object * referenced_object, +int drm_add_ref_object(struct drm_file *priv, struct drm_user_object *referenced_object,  		       enum drm_ref_type ref_action)  {  	int ret = 0; @@ -167,12 +167,12 @@ int drm_add_ref_object(struct drm_file * priv, struct drm_user_object * referenc  	list_add(&item->list, &priv->refd_objects);  	ret = drm_object_ref_action(priv, referenced_object, ref_action); -      out: +out:  	return ret;  } -struct drm_ref_object *drm_lookup_ref_object(struct drm_file * priv, -					struct drm_user_object * referenced_object, +struct drm_ref_object *drm_lookup_ref_object(struct drm_file *priv, +					struct drm_user_object *referenced_object,  					enum drm_ref_type ref_action)  {  	struct drm_hash_item *hash; @@ -188,8 +188,8 @@ struct drm_ref_object *drm_lookup_ref_object(struct drm_file * priv,  }  EXPORT_SYMBOL(drm_lookup_ref_object); -static void drm_remove_other_references(struct drm_file * priv, -					struct drm_user_object * ro) +static void drm_remove_other_references(struct drm_file *priv, +					struct drm_user_object *ro)  {  	int i;  	struct drm_open_hash *ht; @@ -205,7 +205,7 @@ static void drm_remove_other_references(struct drm_file * priv,  	}  } -void drm_remove_ref_object(struct drm_file * priv, struct drm_ref_object * item) +void drm_remove_ref_object(struct drm_file *priv, struct drm_ref_object *item)  {  	int ret;  	struct drm_user_object *user_object = (struct drm_user_object *) item->hash.key; @@ -235,8 +235,8 @@ void drm_remove_ref_object(struct drm_file * priv, struct drm_ref_object * item)  } -int drm_user_object_ref(struct drm_file * priv, uint32_t user_token, -			enum drm_object_type type, struct drm_user_object ** object) +int drm_user_object_ref(struct drm_file *priv, uint32_t user_token, +			enum drm_object_type type, struct drm_user_object **object)  {  	struct drm_device *dev = priv->head->dev;  	struct drm_user_object *uo; @@ -260,12 +260,12 @@ int drm_user_object_ref(struct drm_file * priv, uint32_t user_token,  	mutex_unlock(&dev->struct_mutex);  	*object = uo;  	return 0; -      out_err: +out_err:  	mutex_unlock(&dev->struct_mutex);  	return ret;  } -int drm_user_object_unref(struct drm_file * priv, uint32_t user_token, +int drm_user_object_unref(struct drm_file *priv, uint32_t user_token,  			  enum drm_object_type type)  {  	struct drm_device *dev = priv->head->dev; @@ -287,7 +287,7 @@ int drm_user_object_unref(struct drm_file * priv, uint32_t user_token,  	drm_remove_ref_object(priv, ro);  	mutex_unlock(&dev->struct_mutex);  	return 0; -      out_err: +out_err:  	mutex_unlock(&dev->struct_mutex);  	return ret;  } diff --git a/linux-core/drm_ttm.c b/linux-core/drm_ttm.c index 4d51f9fe..8a44070d 100644 --- a/linux-core/drm_ttm.c +++ b/linux-core/drm_ttm.c @@ -46,7 +46,7 @@ EXPORT_SYMBOL(drm_ttm_cache_flush);   * Use kmalloc if possible. Otherwise fall back to vmalloc.   */ -static void ttm_alloc_pages(struct drm_ttm * ttm) +static void ttm_alloc_pages(struct drm_ttm *ttm)  {  	unsigned long size = ttm->num_pages * sizeof(*ttm->pages);  	ttm->pages = NULL; @@ -54,20 +54,19 @@ static void ttm_alloc_pages(struct drm_ttm * ttm)  	if (drm_alloc_memctl(size))  		return; -	if (size <= PAGE_SIZE) { +	if (size <= PAGE_SIZE)  		ttm->pages = drm_calloc(1, size, DRM_MEM_TTM); -	} +  	if (!ttm->pages) {  		ttm->pages = vmalloc_user(size);  		if (ttm->pages)  			ttm->page_flags |= DRM_TTM_PAGE_VMALLOC;  	} -	if (!ttm->pages) { +	if (!ttm->pages)  		drm_free_memctl(size); -	}  } -static void ttm_free_pages(struct drm_ttm * ttm) +static void ttm_free_pages(struct drm_ttm *ttm)  {  	unsigned long size = ttm->num_pages * sizeof(*ttm->pages); @@ -85,9 +84,9 @@ static struct page *drm_ttm_alloc_page(void)  {  	struct page *page; -	if (drm_alloc_memctl(PAGE_SIZE)) { +	if (drm_alloc_memctl(PAGE_SIZE))  		return NULL; -	} +  	page = alloc_page(GFP_KERNEL | __GFP_ZERO | GFP_DMA32);  	if (!page) {  		drm_free_memctl(PAGE_SIZE); @@ -106,7 +105,7 @@ static struct page *drm_ttm_alloc_page(void)   * for range of pages in a ttm.   */ -static int drm_set_caching(struct drm_ttm * ttm, int noncached) +static int drm_set_caching(struct drm_ttm *ttm, int noncached)  {  	int i;  	struct page **cur_page; @@ -153,7 +152,7 @@ static void drm_ttm_free_user_pages(struct drm_ttm *ttm)  	dirty = ((ttm->page_flags & DRM_TTM_PAGE_USER_DIRTY) != 0);  	down_read(&mm->mmap_sem); -	for (i=0; i<ttm->num_pages; ++i) { +	for (i = 0; i < ttm->num_pages; ++i) {  		page = ttm->pages[i];  		if (page == NULL)  			continue; @@ -186,14 +185,10 @@ static void drm_ttm_free_alloced_pages(struct drm_ttm *ttm)  #else  			ClearPageReserved(*cur_page);  #endif -			if (page_count(*cur_page) != 1) { -				DRM_ERROR("Erroneous page count. " -					  "Leaking pages.\n"); -			} -			if (page_mapped(*cur_page)) { -				DRM_ERROR("Erroneous map count. " -					  "Leaking page mappings.\n"); -			} +			if (page_count(*cur_page) != 1) +				DRM_ERROR("Erroneous page count. Leaking pages.\n"); +			if (page_mapped(*cur_page)) +				DRM_ERROR("Erroneous map count. Leaking page mappings.\n");  			__free_page(*cur_page);  			drm_free_memctl(PAGE_SIZE);  			--bm->cur_pages; @@ -205,7 +200,7 @@ static void drm_ttm_free_alloced_pages(struct drm_ttm *ttm)   * Free all resources associated with a ttm.   */ -int drm_destroy_ttm(struct drm_ttm * ttm) +int drm_destroy_ttm(struct drm_ttm *ttm)  {  	struct drm_ttm_backend *be; @@ -234,7 +229,7 @@ int drm_destroy_ttm(struct drm_ttm * ttm)  	return 0;  } -struct page *drm_ttm_get_page(struct drm_ttm * ttm, int index) +struct page *drm_ttm_get_page(struct drm_ttm *ttm, int index)  {  	struct page *p;  	struct drm_buffer_manager *bm = &ttm->dev->bm; @@ -283,10 +278,9 @@ int drm_ttm_set_user(struct drm_ttm *ttm,  		return -ENOMEM;  	} -	for (i=0; i<num_pages; ++i) { -		if (ttm->pages[i] == NULL) { +	for (i = 0; i < num_pages; ++i) { +		if (ttm->pages[i] == NULL)  			ttm->pages[i] = ttm->dummy_read_page; -		}  	}  	return 0; @@ -294,7 +288,7 @@ int drm_ttm_set_user(struct drm_ttm *ttm, -int drm_ttm_populate(struct drm_ttm * ttm) +int drm_ttm_populate(struct drm_ttm *ttm)  {  	struct page *page;  	unsigned long i; @@ -318,7 +312,7 @@ int drm_ttm_populate(struct drm_ttm * ttm)   * Initialize a ttm.   */ -struct drm_ttm *drm_ttm_init(struct drm_device * dev, unsigned long size) +struct drm_ttm *drm_ttm_init(struct drm_device *dev, unsigned long size)  {  	struct drm_bo_driver *bo_driver = dev->driver->bo_driver;  	struct drm_ttm *ttm; @@ -362,7 +356,7 @@ struct drm_ttm *drm_ttm_init(struct drm_device * dev, unsigned long size)   * Unbind a ttm region from the aperture.   */ -void drm_ttm_evict(struct drm_ttm * ttm) +void drm_ttm_evict(struct drm_ttm *ttm)  {  	struct drm_ttm_backend *be = ttm->be;  	int ret; @@ -375,19 +369,18 @@ void drm_ttm_evict(struct drm_ttm * ttm)  	ttm->state = ttm_evicted;  } -void drm_ttm_fixup_caching(struct drm_ttm * ttm) +void drm_ttm_fixup_caching(struct drm_ttm *ttm)  {  	if (ttm->state == ttm_evicted) {  		struct drm_ttm_backend *be = ttm->be; -		if (be->func->needs_ub_cache_adjust(be)) { +		if (be->func->needs_ub_cache_adjust(be))  			drm_set_caching(ttm, 0); -		}  		ttm->state = ttm_unbound;  	}  } -void drm_ttm_unbind(struct drm_ttm * ttm) +void drm_ttm_unbind(struct drm_ttm *ttm)  {  	if (ttm->state == ttm_bound)  		drm_ttm_evict(ttm); @@ -395,7 +388,7 @@ void drm_ttm_unbind(struct drm_ttm * ttm)  	drm_ttm_fixup_caching(ttm);  } -int drm_bind_ttm(struct drm_ttm * ttm, struct drm_bo_mem_reg *bo_mem) +int drm_bind_ttm(struct drm_ttm *ttm, struct drm_bo_mem_reg *bo_mem)  {  	struct drm_bo_driver *bo_driver = ttm->dev->driver->bo_driver;  	int ret = 0; @@ -412,13 +405,14 @@ int drm_bind_ttm(struct drm_ttm * ttm, struct drm_bo_mem_reg *bo_mem)  	if (ret)  		return ret; -	if (ttm->state == ttm_unbound && !(bo_mem->flags & DRM_BO_FLAG_CACHED)) { +	if (ttm->state == ttm_unbound && !(bo_mem->flags & DRM_BO_FLAG_CACHED))  		drm_set_caching(ttm, DRM_TTM_PAGE_UNCACHED); -	} else if ((bo_mem->flags & DRM_BO_FLAG_CACHED_MAPPED) && +	else if ((bo_mem->flags & DRM_BO_FLAG_CACHED_MAPPED) &&  		   bo_driver->ttm_cache_flush)  		bo_driver->ttm_cache_flush(ttm); -	if ((ret = be->func->bind(be, bo_mem))) { +	ret = be->func->bind(be, bo_mem); +	if (ret) {  		ttm->state = ttm_evicted;  		DRM_ERROR("Couldn't bind backend.\n");  		return ret; @@ -429,5 +423,4 @@ int drm_bind_ttm(struct drm_ttm * ttm, struct drm_bo_mem_reg *bo_mem)  		ttm->page_flags |= DRM_TTM_PAGE_USER_DIRTY;  	return 0;  } -  EXPORT_SYMBOL(drm_bind_ttm);  | 
