summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorEric Anholt <eric@anholt.net>2008-05-09 14:34:20 -0700
committerEric Anholt <eric@anholt.net>2008-05-09 17:38:32 -0700
commitc5c59eab809604e4d0d4d1dc71fc11186d0220f8 (patch)
treeaa6fd6723baf5e9d1187938e58b25fcd3c0e05d8
parentf56f2acb5a3f34ad6916ff315d3d2058bd4b8f9c (diff)
GEM: Separate the LRU into execution list and LRU list.
Now, the LRU list has objects that are completely done rendering and ready to kick out, while the execution list has things with active rendering, which have associated cookies and reference counts on them.
-rw-r--r--linux-core/i915_gem.c130
-rw-r--r--shared-core/i915_dma.c2
-rw-r--r--shared-core/i915_drv.h9
3 files changed, 106 insertions, 35 deletions
diff --git a/linux-core/i915_gem.c b/linux-core/i915_gem.c
index caead91f..ec5a9872 100644
--- a/linux-core/i915_gem.c
+++ b/linux-core/i915_gem.c
@@ -150,6 +150,7 @@ static int
i915_gem_object_wait_rendering(struct drm_gem_object *obj)
{
struct drm_device *dev = obj->dev;
+ drm_i915_private_t *dev_priv = dev->dev_private;
struct drm_i915_gem_object *obj_priv = obj->driver_private;
int ret;
@@ -163,9 +164,19 @@ i915_gem_object_wait_rendering(struct drm_gem_object *obj)
#endif
i915_gem_flush(dev, 0, obj->write_domain);
obj->write_domain = 0;
+
+ /* Add a reference since we're gaining a cookie. */
if (obj_priv->last_rendering_cookie == 0)
drm_gem_object_reference(obj);
+ /* Move from whatever list we were on to the tail of execution.
+ */
+ list_move_tail(&obj_priv->gtt_lru_entry,
+ &dev_priv->mm.execution_list);
obj_priv->last_rendering_cookie = i915_emit_irq(dev);
+ BUG_ON(obj_priv->last_rendering_cookie == 0);
+#if WATCH_LRU
+ DRM_INFO("%s: flush moves to exec list %p\n", __func__, obj);
+#endif
}
/* If there is rendering queued on the buffer being evicted, wait for
* it.
@@ -178,12 +189,19 @@ i915_gem_object_wait_rendering(struct drm_gem_object *obj)
ret = i915_wait_irq(dev, obj_priv->last_rendering_cookie);
if (ret != 0)
return ret;
+
/* Clear it now that we know it's passed. */
obj_priv->last_rendering_cookie = 0;
-
- /*
- * The cookie held a reference to the object,
- * release that now
+ /* We were on the execution list since we had a cookie.
+ * Move to the tail of the LRU list now since we're done.
+ */
+ list_move_tail(&obj_priv->gtt_lru_entry,
+ &dev_priv->mm.gtt_lru);
+#if WATCH_LRU
+ DRM_INFO("%s: wait moves to lru list %p\n", __func__, obj);
+#endif
+ /* The cookie held a reference to the object, release that
+ * now
*/
drm_gem_object_unreference(obj);
}
@@ -206,7 +224,11 @@ i915_gem_object_unbind(struct drm_gem_object *obj)
if (obj_priv->gtt_space == NULL)
return;
- i915_gem_object_wait_rendering(obj);
+ /* Ignore the return value of wait_rendering. If we're here but
+ * a wait_rendering hasn't completed, we're in the freeing process,
+ * and we want the buffer to go away even if the command queue is hung.
+ */
+ (void)i915_gem_object_wait_rendering(obj);
if (obj_priv->agp_mem != NULL) {
drm_unbind_agp(obj_priv->agp_mem);
@@ -218,8 +240,17 @@ i915_gem_object_unbind(struct drm_gem_object *obj)
drm_memrange_put_block(obj_priv->gtt_space);
obj_priv->gtt_space = NULL;
- if (!list_empty(&obj_priv->gtt_lru_entry))
+
+ /* Remove ourselves from the LRU list if present. */
+ if (!list_empty(&obj_priv->gtt_lru_entry)) {
list_del_init(&obj_priv->gtt_lru_entry);
+ if (obj_priv->last_rendering_cookie) {
+ DRM_ERROR("Failed to wait on buffer when unbinding, "
+ "continued anyway.\n");
+ obj_priv->last_rendering_cookie = 0;
+ drm_gem_object_unreference(obj);
+ }
+ }
}
#if WATCH_BUF | WATCH_EXEC
@@ -274,6 +305,13 @@ i915_dump_lru(struct drm_device *dev, const char *where)
drm_i915_private_t *dev_priv = dev->dev_private;
struct drm_i915_gem_object *obj_priv;
+ DRM_INFO("GTT execution list %s {\n", where);
+ list_for_each_entry(obj_priv, &dev_priv->mm.execution_list,
+ gtt_lru_entry)
+ {
+ DRM_INFO(" %p: %08x\n", obj_priv,
+ obj_priv->last_rendering_cookie);
+ }
DRM_INFO("GTT LRU %s {\n", where);
list_for_each_entry(obj_priv, &dev_priv->mm.gtt_lru, gtt_lru_entry) {
DRM_INFO(" %p: %08x\n", obj_priv,
@@ -292,11 +330,22 @@ i915_gem_evict_something(struct drm_device *dev)
int ret;
/* Find the LRU buffer. */
- BUG_ON(list_empty(&dev_priv->mm.gtt_lru));
- obj_priv = list_first_entry(&dev_priv->mm.gtt_lru,
- struct drm_i915_gem_object,
- gtt_lru_entry);
+ if (!list_empty(&dev_priv->mm.gtt_lru)) {
+ obj_priv = list_first_entry(&dev_priv->mm.gtt_lru,
+ struct drm_i915_gem_object,
+ gtt_lru_entry);
+ } else if (!list_empty(&dev_priv->mm.execution_list)) {
+ /* If there's nothing unused and ready, grab the LRU
+ * from the currently executing list.
+ */
+ obj_priv = list_first_entry(&dev_priv->mm.execution_list,
+ struct drm_i915_gem_object,
+ gtt_lru_entry);
+ } else {
+ return -ENOMEM;
+ }
obj = obj_priv->obj;
+ drm_gem_object_reference(obj);
#if WATCH_LRU
DRM_INFO("%s: evicting %p\n", __func__, obj);
#endif
@@ -317,6 +366,7 @@ i915_gem_evict_something(struct drm_device *dev)
#if WATCH_LRU
DRM_INFO("%s: evicted %p\n", __func__, obj);
#endif
+ drm_gem_object_unreference(obj);
return 0;
}
@@ -360,7 +410,8 @@ i915_gem_object_bind_to_gtt(struct drm_gem_object *obj, unsigned alignment)
#if WATCH_LRU
DRM_INFO("%s: GTT full, evicting something\n", __func__);
#endif
- if (list_empty(&dev_priv->mm.gtt_lru)) {
+ if (list_empty(&dev_priv->mm.gtt_lru) &&
+ list_empty(&dev_priv->mm.execution_list)) {
DRM_ERROR("GTT full, but LRU list empty\n");
return -ENOMEM;
}
@@ -529,7 +580,6 @@ i915_gem_reloc_and_validate_object(struct drm_gem_object *obj,
struct drm_i915_gem_validate_entry *entry)
{
struct drm_device *dev = obj->dev;
- drm_i915_private_t *dev_priv = dev->dev_private;
struct drm_i915_gem_relocation_entry reloc;
struct drm_i915_gem_relocation_entry __user *relocs;
struct drm_i915_gem_object *obj_priv = obj->driver_private;
@@ -546,19 +596,6 @@ i915_gem_reloc_and_validate_object(struct drm_gem_object *obj,
entry->buffer_offset = obj_priv->gtt_offset;
- if (obj_priv->pin_count == 0) {
- /* Move our buffer to the head of the LRU. */
- if (list_empty(&obj_priv->gtt_lru_entry))
- list_add_tail(&obj_priv->gtt_lru_entry,
- &dev_priv->mm.gtt_lru);
- else
- list_move_tail(&obj_priv->gtt_lru_entry,
- &dev_priv->mm.gtt_lru);
-#if WATCH_LRU && 0
- i915_dump_lru(dev, __func__);
-#endif
- }
-
relocs = (struct drm_i915_gem_relocation_entry __user *)
(uintptr_t) entry->relocs_ptr;
/* Apply the relocations, using the GTT aperture to avoid cache
@@ -770,6 +807,7 @@ int
i915_gem_execbuffer(struct drm_device *dev, void *data,
struct drm_file *file_priv)
{
+ drm_i915_private_t *dev_priv = dev->dev_private;
struct drm_i915_gem_execbuffer *args = data;
struct drm_i915_gem_validate_entry *validate_list;
struct drm_gem_object **object_list;
@@ -894,7 +932,17 @@ i915_gem_execbuffer(struct drm_device *dev, void *data,
if (obj_priv->last_rendering_cookie == 0)
drm_gem_object_reference(obj);
obj_priv->last_rendering_cookie = cookie;
+ BUG_ON(obj_priv->last_rendering_cookie == 0);
+ /* Move our buffer to the tail of the execution list. */
+ list_move_tail(&obj_priv->gtt_lru_entry,
+ &dev_priv->mm.execution_list);
+#if WATCH_LRU
+ DRM_INFO("%s: move to exec list %p\n", __func__, obj);
+#endif
}
+#if WATCH_LRU && 0
+ i915_dump_lru(dev, __func__);
+#endif
/* Copy the new buffer offsets back to the user's validate list. */
ret = copy_to_user((struct drm_i915_relocation_entry __user *)
@@ -911,14 +959,6 @@ err:
drm_gem_object_unreference(object_list[i]);
}
-#if 1
- /* XXX kludge for now as we don't clean the exec ring yet */
- if (object_list != NULL) {
- for (i = 0; i < args->buffer_count; i++)
- i915_gem_object_wait_rendering(object_list[i]);
- }
-#endif
-
drm_free(object_list, sizeof(*object_list) * args->buffer_count,
DRM_MEM_DRIVER);
drm_free(validate_list, sizeof(*validate_list) * args->buffer_count,
@@ -1014,3 +1054,27 @@ i915_gem_set_domain_ioctl(struct drm_gem_object *obj,
i915_gem_dev_set_domain(obj->dev);
return 0;
}
+
+void
+i915_gem_lastclose(struct drm_device *dev)
+{
+ drm_i915_private_t *dev_priv = dev->dev_private;
+
+ /* Assume that the chip has been idled at this point. Just pull them
+ * off the execution list and unref them. Since this is the last
+ * close, this is also the last ref and they'll go away.
+ */
+
+ while (!list_empty(&dev_priv->mm.execution_list)) {
+ struct drm_i915_gem_object *obj_priv;
+
+ obj_priv = list_first_entry(&dev_priv->mm.execution_list,
+ struct drm_i915_gem_object,
+ gtt_lru_entry);
+
+ list_del_init(&obj_priv->gtt_lru_entry);
+ obj_priv->last_rendering_cookie = 0;
+ obj_priv->obj->write_domain = 0;
+ drm_gem_object_unreference(obj_priv->obj);
+ }
+}
diff --git a/shared-core/i915_dma.c b/shared-core/i915_dma.c
index 0601b89c..24f361e0 100644
--- a/shared-core/i915_dma.c
+++ b/shared-core/i915_dma.c
@@ -1052,6 +1052,7 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags)
_DRM_KERNEL | _DRM_DRIVER, &dev_priv->mmio_map);
INIT_LIST_HEAD(&dev_priv->mm.gtt_lru);
+ INIT_LIST_HEAD(&dev_priv->mm.execution_list);
#ifdef __linux__
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,25)
@@ -1093,6 +1094,7 @@ void i915_driver_lastclose(struct drm_device * dev)
dev_priv->val_bufs = NULL;
}
#endif
+ i915_gem_lastclose(dev);
if (drm_getsarea(dev) && dev_priv->sarea_priv)
i915_do_cleanup_pageflip(dev);
diff --git a/shared-core/i915_drv.h b/shared-core/i915_drv.h
index 96257ab6..413fca89 100644
--- a/shared-core/i915_drv.h
+++ b/shared-core/i915_drv.h
@@ -244,7 +244,12 @@ typedef struct drm_i915_private {
struct {
struct drm_memrange gtt_space;
- /** LRU List of unpinned objects in the GTT. */
+ /**
+ * List of objects currently involved in rendering from the
+ * ringbuffer.
+ */
+ struct list_head execution_list;
+ /** LRU List of non-executing objects still in the GTT. */
struct list_head gtt_lru;
} mm;
} drm_i915_private_t;
@@ -386,7 +391,7 @@ void i915_gem_free_object(struct drm_gem_object *obj);
int i915_gem_set_domain_ioctl (struct drm_gem_object *obj,
uint32_t read_domains,
uint32_t write_domain);
-
+void i915_gem_lastclose(struct drm_device *dev);
#endif
#ifdef __linux__