summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorThomas Hellstrom <thomas-at-tungstengraphics-dot-com>2007-02-06 15:56:43 +0100
committerThomas Hellstrom <thomas-at-tungstengraphics-dot-com>2007-02-06 15:56:43 +0100
commit40ce53dfde11f84d7bf8db5db93fb73715b2e96e (patch)
tree0acd792653752cdc579c56cf8a22bf53a5d6d253
parent609e3b037526021d20c7cc18b7fed1152206dc68 (diff)
Implement a drm_mem_reg_t substructure in the buffer object type.
-rw-r--r--linux-core/drmP.h8
-rw-r--r--linux-core/drm_bo.c173
2 files changed, 89 insertions, 92 deletions
diff --git a/linux-core/drmP.h b/linux-core/drmP.h
index 5834c9dc..a8f5e3e2 100644
--- a/linux-core/drmP.h
+++ b/linux-core/drmP.h
@@ -1022,13 +1022,9 @@ typedef struct drm_buffer_object{
unsigned long buffer_start;
drm_bo_type_t type;
unsigned long offset;
- uint32_t page_alignment;
atomic_t mapped;
- uint32_t flags;
- uint32_t mask;
- uint32_t mem_type;
+ drm_bo_mem_reg_t mem;
- drm_mm_node_t *mm_node;
struct list_head lru;
struct list_head ddestroy;
@@ -1042,11 +1038,9 @@ typedef struct drm_buffer_object{
/* For vm */
drm_map_list_t map_list;
- drm_mm_node_t *node;
uint32_t memory_type;
drm_ttm_t *ttm;
unsigned long bus_offset;
- unsigned long num_pages;
uint32_t vm_flags;
void *iomap;
diff --git a/linux-core/drm_bo.c b/linux-core/drm_bo.c
index fa659d04..8a6b49dc 100644
--- a/linux-core/drm_bo.c
+++ b/linux-core/drm_bo.c
@@ -78,24 +78,24 @@ static void drm_bo_add_to_lru(drm_buffer_object_t * bo,
struct list_head *list;
drm_mem_type_manager_t *man;
- bo->mem_type = 0;
+ bo->mem.mem_type = 0;
- switch(bo->flags & DRM_BO_MASK_MEM) {
+ switch(bo->mem.flags & DRM_BO_MASK_MEM) {
case DRM_BO_FLAG_MEM_TT:
- bo->mem_type = DRM_BO_MEM_TT;
+ bo->mem.mem_type = DRM_BO_MEM_TT;
break;
case DRM_BO_FLAG_MEM_VRAM:
- bo->mem_type = DRM_BO_MEM_VRAM;
+ bo->mem.mem_type = DRM_BO_MEM_VRAM;
break;
case DRM_BO_FLAG_MEM_LOCAL:
- bo->mem_type = DRM_BO_MEM_LOCAL;
+ bo->mem.mem_type = DRM_BO_MEM_LOCAL;
break;
default:
BUG_ON(1);
}
- man = &bm->man[bo->mem_type];
- list = (bo->flags & (DRM_BO_FLAG_NO_EVICT | DRM_BO_FLAG_NO_MOVE)) ?
+ man = &bm->man[bo->mem.mem_type];
+ list = (bo->mem.flags & (DRM_BO_FLAG_NO_EVICT | DRM_BO_FLAG_NO_MOVE)) ?
&man->pinned : &man->lru;
list_add_tail(&bo->lru, list);
return;
@@ -111,7 +111,7 @@ static int drm_move_tt_to_local(drm_buffer_object_t * bo, int evict,
drm_device_t *dev = bo->dev;
int ret = 0;
- if (bo->mm_node) {
+ if (bo->mem.mm_node) {
#ifdef DRM_ODD_MM_COMPAT
mutex_lock(&dev->struct_mutex);
ret = drm_bo_lock_kmm(bo);
@@ -133,16 +133,16 @@ static int drm_move_tt_to_local(drm_buffer_object_t * bo, int evict,
else
drm_ttm_unbind(bo->ttm);
- bo->mem_type = DRM_BO_MEM_LOCAL;
- if (!(bo->flags & DRM_BO_FLAG_NO_MOVE) || force_no_move) {
- drm_mm_put_block(bo->mm_node);
- bo->mm_node = NULL;
+ bo->mem.mem_type = DRM_BO_MEM_LOCAL;
+ if (!(bo->mem.flags & DRM_BO_FLAG_NO_MOVE) || force_no_move) {
+ drm_mm_put_block(bo->mem.mm_node);
+ bo->mem.mm_node = NULL;
}
mutex_unlock(&dev->struct_mutex);
}
- bo->flags &= ~DRM_BO_FLAG_MEM_TT;
- bo->flags |= DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED;
+ bo->mem.flags &= ~DRM_BO_FLAG_MEM_TT;
+ bo->mem.flags |= DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED;
return 0;
}
@@ -235,9 +235,9 @@ static void drm_bo_cleanup_refs(drm_buffer_object_t *bo, int remove_all)
if (!bo->fence) {
list_del_init(&bo->lru);
- if (bo->mm_node) {
- drm_mm_put_block(bo->mm_node);
- bo->mm_node = NULL;
+ if (bo->mem.mm_node) {
+ drm_mm_put_block(bo->mem.mm_node);
+ bo->mem.mm_node = NULL;
}
list_del_init(&bo->ddestroy);
mutex_unlock(&bo->mutex);
@@ -269,7 +269,7 @@ static void drm_bo_destroy_locked(drm_buffer_object_t *bo)
drm_device_t *dev = bo->dev;
drm_buffer_manager_t *bm = &dev->bm;
- if (list_empty(&bo->lru) && bo->mm_node == NULL && atomic_read(&bo->usage) == 0) {
+ if (list_empty(&bo->lru) && bo->mem.mm_node == NULL && atomic_read(&bo->usage) == 0) {
BUG_ON(bo->fence != NULL);
#ifdef DRM_ODD_MM_COMPAT
@@ -507,7 +507,7 @@ static int drm_bo_evict(drm_buffer_object_t * bo, unsigned mem_type,
if (bo->priv_flags & _DRM_BO_FLAG_UNFENCED)
goto out;
- if (!(bo->flags & drm_bo_type_flags(mem_type)))
+ if (!(bo->mem.flags & drm_bo_type_flags(mem_type)))
goto out;
ret = drm_bo_wait(bo, 0, 0, no_wait);
@@ -518,9 +518,7 @@ static int drm_bo_evict(drm_buffer_object_t * bo, unsigned mem_type,
goto out;
}
- evict_mem.num_pages = bo->num_pages;
- evict_mem.page_alignment = bo->page_alignment;
- evict_mem.size = evict_mem.num_pages << PAGE_SHIFT;
+ evict_mem = bo->mem;
evict_mem.mask = dev->driver->bo_driver->evict_flags(dev, mem_type);
ret = drm_bo_mem_space(dev, &evict_mem, no_wait);
@@ -585,7 +583,7 @@ static int drm_bo_mem_force_space(drm_device_t *dev,
atomic_inc(&entry->usage);
mutex_unlock(&dev->struct_mutex);
mutex_lock(&entry->mutex);
- BUG_ON(entry->flags & (DRM_BO_FLAG_NO_MOVE | DRM_BO_FLAG_NO_EVICT));
+ BUG_ON(entry->mem.flags & (DRM_BO_FLAG_NO_MOVE | DRM_BO_FLAG_NO_EVICT));
ret = drm_bo_evict(entry, mem_type, no_wait, 0);
mutex_unlock(&entry->mutex);
@@ -694,10 +692,10 @@ static int drm_move_local_to_tt(drm_buffer_object_t * bo,
drm_device_t *dev = bo->dev;
int ret = 0;
- bo->mm_node = mem->mm_node;
+ bo->mem.mm_node = mem->mm_node;
DRM_DEBUG("Flipping in to AGP 0x%08lx 0x%08lx\n",
- bo->mm_node->start, bo->mm_node->size);
+ bo->mem.mm_node->start, bo->mem.mm_node->size);
#ifdef DRM_ODD_MM_COMPAT
mutex_lock(&dev->struct_mutex);
@@ -708,8 +706,8 @@ static int drm_move_local_to_tt(drm_buffer_object_t * bo,
}
#endif
drm_bo_unmap_virtual(bo);
- ret = drm_bind_ttm(bo->ttm, bo->flags & DRM_BO_FLAG_BIND_CACHED,
- bo->mm_node->start);
+ ret = drm_bind_ttm(bo->ttm, bo->mem.flags & DRM_BO_FLAG_BIND_CACHED,
+ bo->mem.mm_node->start);
if (ret) {
#ifdef DRM_ODD_MM_COMPAT
@@ -719,12 +717,12 @@ static int drm_move_local_to_tt(drm_buffer_object_t * bo,
goto out_put_unlock;
}
- if (!(bo->flags & DRM_BO_FLAG_BIND_CACHED))
- bo->flags &= DRM_BO_FLAG_CACHED;
- bo->flags &= ~DRM_BO_MASK_MEM;
- bo->flags |= DRM_BO_FLAG_MEM_TT;
- bo->mem_type = DRM_BO_MEM_TT;
- bo->offset = bo->mm_node->start << PAGE_SHIFT;
+ if (!(bo->mem.flags & DRM_BO_FLAG_BIND_CACHED))
+ bo->mem.flags &= DRM_BO_FLAG_CACHED;
+ bo->mem.flags &= ~DRM_BO_MASK_MEM;
+ bo->mem.flags |= DRM_BO_FLAG_MEM_TT;
+ bo->mem.mem_type = DRM_BO_MEM_TT;
+ bo->offset = bo->mem.mm_node->start << PAGE_SHIFT;
#ifdef DRM_ODD_MM_COMPAT
ret = drm_bo_remap_bound(bo);
@@ -736,7 +734,7 @@ static int drm_move_local_to_tt(drm_buffer_object_t * bo,
#endif
if (bo->priv_flags & _DRM_BO_FLAG_EVICTED) {
- ret = dev->driver->bo_driver->invalidate_caches(dev, bo->flags);
+ ret = dev->driver->bo_driver->invalidate_caches(dev, bo->mem.flags);
if (ret)
DRM_ERROR("Could not flush read caches\n");
}
@@ -746,8 +744,8 @@ static int drm_move_local_to_tt(drm_buffer_object_t * bo,
out_put_unlock:
mutex_lock(&dev->struct_mutex);
- drm_mm_put_block(bo->mm_node);
- bo->mm_node = NULL;
+ drm_mm_put_block(bo->mem.mm_node);
+ bo->mem.mm_node = NULL;
mutex_unlock(&dev->struct_mutex);
return ret;
}
@@ -948,7 +946,7 @@ static int drm_bo_read_cached(drm_buffer_object_t * bo)
int ret = 0;
BUG_ON(bo->priv_flags & _DRM_BO_FLAG_UNFENCED);
- if (bo->mm_node)
+ if (bo->mem.mm_node)
ret = drm_bo_evict(bo, DRM_BO_MEM_TT, 1, 0);
return ret;
}
@@ -1039,15 +1037,15 @@ static void drm_bo_fill_rep_arg(drm_buffer_object_t * bo,
drm_bo_arg_reply_t * rep)
{
rep->handle = bo->base.hash.key;
- rep->flags = bo->flags;
- rep->size = bo->num_pages * PAGE_SIZE;
+ rep->flags = bo->mem.flags;
+ rep->size = bo->mem.num_pages * PAGE_SIZE;
rep->offset = bo->offset;
rep->arg_handle = bo->map_list.user_token;
- rep->mask = bo->mask;
+ rep->mask = bo->mem.mask;
rep->buffer_start = bo->buffer_start;
rep->fence_flags = bo->fence_type;
rep->rep_flags = 0;
- rep->page_alignment = bo->page_alignment;
+ rep->page_alignment = bo->mem.page_alignment;
if ((bo->priv_flags & _DRM_BO_FLAG_UNFENCED) || drm_bo_quick_busy(bo)) {
DRM_FLAG_MASKED(rep->rep_flags, DRM_BO_REP_BUSY,
@@ -1105,14 +1103,14 @@ static int drm_buffer_object_map(drm_file_t * priv, uint32_t handle,
}
if ((map_flags & DRM_BO_FLAG_READ) &&
- (bo->flags & DRM_BO_FLAG_READ_CACHED) &&
- (!(bo->flags & DRM_BO_FLAG_CACHED))) {
+ (bo->mem.flags & DRM_BO_FLAG_READ_CACHED) &&
+ (!(bo->mem.flags & DRM_BO_FLAG_CACHED))) {
drm_bo_read_cached(bo);
}
break;
} else if ((map_flags & DRM_BO_FLAG_READ) &&
- (bo->flags & DRM_BO_FLAG_READ_CACHED) &&
- (!(bo->flags & DRM_BO_FLAG_CACHED))) {
+ (bo->mem.flags & DRM_BO_FLAG_READ_CACHED) &&
+ (!(bo->mem.flags & DRM_BO_FLAG_CACHED))) {
/*
* We are already mapped with different flags.
@@ -1228,18 +1226,23 @@ static int drm_bo_move_buffer(drm_buffer_object_t * bo, uint32_t new_mem_flags,
return ret;
- mem.num_pages = bo->num_pages;
+ mem.num_pages = bo->mem.num_pages;
mem.size = mem.num_pages << PAGE_SHIFT;
mem.mask = new_mem_flags;
- mem.page_alignment = bo->page_alignment;
+ mem.page_alignment = bo->mem.page_alignment;
mutex_lock(&bm->evict_mutex);
mutex_lock(&dev->struct_mutex);
list_del(&bo->lru);
list_add_tail(&bo->lru,&bm->unfenced);
- DRM_FLAG_MASKED(bo->priv_flags, _DRM_BO_FLAG_UNFENCED, _DRM_BO_FLAG_UNFENCED);
+ DRM_FLAG_MASKED(bo->priv_flags, _DRM_BO_FLAG_UNFENCED,
+ _DRM_BO_FLAG_UNFENCED);
mutex_unlock(&dev->struct_mutex);
+ /*
+ * Determine where to move the buffer.
+ */
+
ret = drm_bo_mem_space(dev, &mem, no_wait);
mutex_unlock(&bm->evict_mutex);
@@ -1272,7 +1275,7 @@ static int drm_buffer_object_validate(drm_buffer_object_t * bo,
{
drm_device_t *dev = bo->dev;
drm_buffer_manager_t *bm = &dev->bm;
- uint32_t flag_diff = (new_flags ^ bo->flags);
+ uint32_t flag_diff = (new_flags ^ bo->mem.flags);
drm_bo_driver_t *driver = dev->driver->bo_driver;
int ret;
@@ -1282,7 +1285,7 @@ static int drm_buffer_object_validate(drm_buffer_object_t * bo,
return -EINVAL;
}
- DRM_DEBUG("New flags 0x%08x, Old flags 0x%08x\n", new_flags, bo->flags);
+ DRM_DEBUG("New flags 0x%08x, Old flags 0x%08x\n", new_flags, bo->mem.flags);
ret = driver->fence_type(new_flags, &bo->fence_class, &bo->fence_type);
if (ret) {
DRM_ERROR("Driver did not support given buffer permissions\n");
@@ -1294,8 +1297,8 @@ static int drm_buffer_object_validate(drm_buffer_object_t * bo,
*/
if ((flag_diff & DRM_BO_FLAG_BIND_CACHED) &&
- !(bo->flags & DRM_BO_FLAG_MEM_LOCAL)) {
- if (bo->flags & (DRM_BO_FLAG_NO_EVICT | DRM_BO_FLAG_NO_MOVE)) {
+ !(bo->mem.flags & DRM_BO_FLAG_MEM_LOCAL)) {
+ if (bo->mem.flags & (DRM_BO_FLAG_NO_EVICT | DRM_BO_FLAG_NO_MOVE)) {
DRM_ERROR("Cannot change caching policy of "
"pinned buffer.\n");
return -EINVAL;
@@ -1307,8 +1310,8 @@ static int drm_buffer_object_validate(drm_buffer_object_t * bo,
return ret;
}
}
- DRM_MASK_VAL(bo->flags, DRM_BO_FLAG_BIND_CACHED, new_flags);
- flag_diff = (new_flags ^ bo->flags);
+ DRM_MASK_VAL(bo->mem.flags, DRM_BO_FLAG_BIND_CACHED, new_flags);
+ flag_diff = (new_flags ^ bo->mem.flags);
/*
* Check whether we dropped no_move policy, and in that case,
@@ -1318,9 +1321,9 @@ static int drm_buffer_object_validate(drm_buffer_object_t * bo,
if ((flag_diff & DRM_BO_FLAG_NO_MOVE) &&
!(new_flags & DRM_BO_FLAG_NO_MOVE)) {
mutex_lock(&dev->struct_mutex);
- if (bo->mm_node) {
- drm_mm_put_block(bo->mm_node);
- bo->mm_node = NULL;
+ if (bo->mem.mm_node) {
+ drm_mm_put_block(bo->mem.mm_node);
+ bo->mem.mm_node = NULL;
}
mutex_unlock(&dev->struct_mutex);
}
@@ -1359,7 +1362,7 @@ static int drm_buffer_object_validate(drm_buffer_object_t * bo,
mutex_unlock(&dev->struct_mutex);
}
- bo->flags = new_flags;
+ bo->mem.flags = new_flags;
return 0;
}
@@ -1384,9 +1387,9 @@ static int drm_bo_handle_validate(drm_file_t * priv, uint32_t handle,
if (ret)
goto out;
- ret = drm_bo_new_flags(dev, bo->flags,
- (flags & mask) | (bo->mask & ~mask), hint,
- 0, &new_flags, &bo->mask);
+ ret = drm_bo_new_flags(dev, bo->mem.flags,
+ (flags & mask) | (bo->mem.mask & ~mask), hint,
+ 0, &new_flags, &bo->mem.mask);
if (ret)
goto out;
@@ -1469,7 +1472,7 @@ static int drm_bo_add_ttm(drm_file_t * priv, drm_buffer_object_t * bo)
mutex_unlock(&dev->struct_mutex);
if (ret)
break;
- bo->ttm = drm_ttm_init(dev, bo->num_pages << PAGE_SHIFT);
+ bo->ttm = drm_ttm_init(dev, bo->mem.num_pages << PAGE_SHIFT);
if (!bo->ttm)
ret = -ENOMEM;
break;
@@ -1511,12 +1514,12 @@ int drm_buffer_object_transfer(drm_buffer_object_t *bo,
INIT_LIST_HEAD(&fbo->lru);
list_splice_init(&bo->lru, &fbo->lru);
- bo->mm_node = NULL;
+ bo->mem.mm_node = NULL;
bo->ttm = NULL;
bo->fence = NULL;
- bo->flags = 0;
+ bo->mem.flags = 0;
- fbo->mm_node->private = (void *)fbo;
+ fbo->mem.mm_node->private = (void *)fbo;
atomic_set(&fbo->usage, 1);
atomic_inc(&bm->count);
mutex_unlock(&dev->struct_mutex);
@@ -1572,9 +1575,9 @@ int drm_buffer_object_create(drm_file_t * priv,
#endif
bo->dev = dev;
bo->type = type;
- bo->num_pages = num_pages;
- bo->mm_node = NULL;
- bo->page_alignment = page_alignment;
+ bo->mem.num_pages = num_pages;
+ bo->mem.mm_node = NULL;
+ bo->mem.page_alignment = page_alignment;
if (bo->type == drm_bo_type_fake) {
bo->offset = buffer_start;
bo->buffer_start = 0;
@@ -1582,10 +1585,10 @@ int drm_buffer_object_create(drm_file_t * priv,
bo->buffer_start = buffer_start;
}
bo->priv_flags = 0;
- bo->flags = DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED;
+ bo->mem.flags = DRM_BO_FLAG_MEM_LOCAL | DRM_BO_FLAG_CACHED;
atomic_inc(&bm->count);
- ret = drm_bo_new_flags(dev, bo->flags, mask, hint,
- 1, &new_flags, &bo->mask);
+ ret = drm_bo_new_flags(dev, bo->mem.flags, mask, hint,
+ 1, &new_flags, &bo->mem.mask);
if (ret)
goto out_err;
ret = drm_bo_add_ttm(priv, bo);
@@ -1800,7 +1803,7 @@ static int drm_bo_force_list_clean(drm_device_t * dev,
drm_bo_usage_deref_locked(entry);
goto retry;
}
- if (entry->mm_node) {
+ if (entry->mem.mm_node) {
clean = 0;
/*
@@ -1836,14 +1839,14 @@ static int drm_bo_force_list_clean(drm_device_t * dev,
0);
if (force_no_move) {
- DRM_MASK_VAL(entry->flags, DRM_BO_FLAG_NO_MOVE,
+ DRM_MASK_VAL(entry->mem.flags, DRM_BO_FLAG_NO_MOVE,
0);
}
- if (entry->flags & DRM_BO_FLAG_NO_EVICT) {
+ if (entry->mem.flags & DRM_BO_FLAG_NO_EVICT) {
DRM_ERROR("A DRM_BO_NO_EVICT buffer present at "
"cleanup. Removing flag and evicting.\n");
- entry->flags &= ~DRM_BO_FLAG_NO_EVICT;
- entry->mask &= ~DRM_BO_FLAG_NO_EVICT;
+ entry->mem.flags &= ~DRM_BO_FLAG_NO_EVICT;
+ entry->mem.mask &= ~DRM_BO_FLAG_NO_EVICT;
}
ret = drm_bo_evict(entry, mem_type, 1, force_no_move);
@@ -2181,7 +2184,7 @@ int drm_bo_pci_offset(const drm_buffer_object_t *bo,
{
drm_device_t *dev = bo->dev;
drm_buffer_manager_t *bm = &dev->bm;
- drm_mem_type_manager_t *man = &bm->man[bo->mem_type];
+ drm_mem_type_manager_t *man = &bm->man[bo->mem.mem_type];
*bus_size = 0;
if (bo->type != drm_bo_type_dc)
@@ -2208,8 +2211,8 @@ int drm_bo_pci_offset(const drm_buffer_object_t *bo,
*bus_base = man->io_offset;
}
- *bus_offset = bo->mm_node->start << PAGE_SHIFT;
- *bus_size = bo->num_pages << PAGE_SHIFT;
+ *bus_offset = bo->mem.mm_node->start << PAGE_SHIFT;
+ *bus_size = bo->mem.num_pages << PAGE_SHIFT;
return 0;
}
@@ -2232,7 +2235,7 @@ int drm_bo_ioremap(drm_buffer_object_t *bo)
{
drm_device_t *dev = bo->dev;
drm_buffer_manager_t *bm = &dev->bm;
- drm_mem_type_manager_t *man = &bm->man[bo->mem_type];
+ drm_mem_type_manager_t *man = &bm->man[bo->mem.mem_type];
unsigned long bus_offset;
unsigned long bus_size;
unsigned long bus_base;
@@ -2271,7 +2274,7 @@ void drm_bo_iounmap(drm_buffer_object_t *bo)
bm = &dev->bm;
- man = &bm->man[bo->mem_type];
+ man = &bm->man[bo->mem.mem_type];
if (bo->iomap && (man->flags & _DRM_FLAG_NEEDS_IOREMAP))
iounmap(bo->iomap);
@@ -2291,7 +2294,7 @@ void drm_bo_unmap_virtual(drm_buffer_object_t *bo)
{
drm_device_t *dev = bo->dev;
loff_t offset = ((loff_t) bo->map_list.hash.key) << PAGE_SHIFT;
- loff_t holelen = ((loff_t) bo->num_pages) << PAGE_SHIFT;
+ loff_t holelen = ((loff_t) bo->mem.num_pages) << PAGE_SHIFT;
unmap_mapping_range(dev->dev_mapping, offset, holelen, 1);
}
@@ -2335,12 +2338,12 @@ static int drm_bo_setup_vm_locked(drm_buffer_object_t *bo)
map->offset = 0;
map->type = _DRM_TTM;
map->flags = _DRM_REMOVABLE;
- map->size = bo->num_pages * PAGE_SIZE;
+ map->size = bo->mem.num_pages * PAGE_SIZE;
atomic_inc(&bo->usage);
map->handle = (void *) bo;
list->file_offset_node = drm_mm_search_free(&dev->offset_manager,
- bo->num_pages, 0, 0);
+ bo->mem.num_pages, 0, 0);
if (!list->file_offset_node) {
drm_bo_takedown_vm_locked(bo);
@@ -2348,7 +2351,7 @@ static int drm_bo_setup_vm_locked(drm_buffer_object_t *bo)
}
list->file_offset_node = drm_mm_get_block(list->file_offset_node,
- bo->num_pages, 0);
+ bo->mem.num_pages, 0);
list->hash.key = list->file_offset_node->start;
if (drm_ht_insert_item(&dev->map_hash, &list->hash)) {