summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--libdrm/xf86drm.c12
-rw-r--r--linux-core/drm_bo.c56
-rw-r--r--linux-core/drm_fence.c77
-rw-r--r--linux-core/i915_buffer.c5
4 files changed, 97 insertions, 53 deletions
diff --git a/libdrm/xf86drm.c b/libdrm/xf86drm.c
index 2ea6656f..3a23eae2 100644
--- a/libdrm/xf86drm.c
+++ b/libdrm/xf86drm.c
@@ -2995,11 +2995,19 @@ int drmAddValidateItem(drmBOList *list, drmBO *buf, unsigned flags,
if (!memFlags) {
drmMsg("Incompatible memory location requests "
"on validate list.\n");
+ drmMsg("Previous flag: 0x%08lx, mask: 0x%08lx\n",
+ cur->arg0, cur->arg1);
+ drmMsg("Current flag: 0x%08lx, mask: 0x%08lx\n",
+ flags, mask);
return -EINVAL;
}
- if ((cur->arg1 | mask) & ~DRM_BO_MASK_MEM & (cur->arg0 ^ flags)) {
+ if (mask & cur->arg1 & ~DRM_BO_MASK_MEM & (cur->arg0 ^ flags)) {
drmMsg("Incompatible buffer flag requests "
- " on validate list.\n");
+ "on validate list.\n");
+ drmMsg("Previous flag: 0x%08lx, mask: 0x%08lx\n",
+ cur->arg0, cur->arg1);
+ drmMsg("Current flag: 0x%08lx, mask: 0x%08lx\n",
+ flags, mask);
return -EINVAL;
}
cur->arg1 |= mask;
diff --git a/linux-core/drm_bo.c b/linux-core/drm_bo.c
index 3a9c2313..abad398f 100644
--- a/linux-core/drm_bo.c
+++ b/linux-core/drm_bo.c
@@ -91,15 +91,16 @@ static void drm_bo_destroy_locked(drm_device_t * dev, drm_buffer_object_t * bo)
drm_buffer_manager_t *bm = &dev->bm;
- DRM_FLAG_MASKED(bo->priv_flags, 0, _DRM_BO_FLAG_UNFENCED);
+ DRM_FLAG_MASKED(bo->priv_flags, 0, _DRM_BO_FLAG_UNFENCED);
if (bo->fence) {
if (!drm_fence_object_signaled(bo->fence, bo->fence_flags)) {
drm_fence_object_flush(dev, bo->fence, bo->fence_flags);
list_add_tail(&bo->ddestroy, &bm->ddestroy);
- schedule_delayed_work(&bm->wq,
- ((DRM_HZ/100) < 1) ? 1 : DRM_HZ/100);
+ schedule_delayed_work(&bm->wq,
+ ((DRM_HZ / 100) <
+ 1) ? 1 : DRM_HZ / 100);
return;
} else {
drm_fence_usage_deref_locked(dev, bo->fence);
@@ -168,9 +169,10 @@ static void drm_bo_delayed_workqueue(void *data)
DRM_ERROR("Delayed delete Worker\n");
#endif
drm_bo_delayed_delete(dev);
- mutex_lock(&dev->struct_mutex);
+ mutex_lock(&dev->struct_mutex);
if (!list_empty(&bm->ddestroy)) {
- schedule_delayed_work(&bm->wq, ((DRM_HZ/100) < 1) ? 1 : DRM_HZ/100);
+ schedule_delayed_work(&bm->wq,
+ ((DRM_HZ / 100) < 1) ? 1 : DRM_HZ / 100);
}
mutex_unlock(&dev->struct_mutex);
}
@@ -205,9 +207,9 @@ void drm_bo_usage_deref_unlocked(drm_device_t * dev, drm_buffer_object_t * bo)
*/
int drm_fence_buffer_objects(drm_file_t * priv,
- struct list_head *list,
- drm_fence_object_t *fence,
- drm_fence_object_t **used_fence)
+ struct list_head *list,
+ drm_fence_object_t * fence,
+ drm_fence_object_t ** used_fence)
{
drm_device_t *dev = priv->head->dev;
drm_buffer_manager_t *bm = &dev->bm;
@@ -226,6 +228,12 @@ int drm_fence_buffer_objects(drm_file_t * priv,
list_for_each_entry(entry, list, head) {
BUG_ON(!(entry->priv_flags & _DRM_BO_FLAG_UNFENCED));
fence_flags |= entry->fence_flags;
+ if (entry->fence_class != 0) {
+ DRM_ERROR("Fence class %d is not implemented yet.\n",
+ entry->fence_class);
+ ret = -EINVAL;
+ goto out;
+ }
count++;
}
@@ -238,7 +246,7 @@ int drm_fence_buffer_objects(drm_file_t * priv,
/*
* Transfer to a local list before we release the dev->struct_mutex;
* This is so we don't get any new unfenced objects while fencing
- * these.
+ * the ones we already have..
*/
list_add_tail(&f_list, list);
@@ -256,7 +264,7 @@ int drm_fence_buffer_objects(drm_file_t * priv,
ret = drm_fence_object_create(dev, fence_flags, 1, &fence);
mutex_lock(&dev->struct_mutex);
if (ret)
- goto out;
+ goto out;
}
count = 0;
@@ -286,7 +294,7 @@ int drm_fence_buffer_objects(drm_file_t * priv,
list_add_tail(&entry->head, &bm->other);
} else {
#ifdef BODEBUG
- DRM_ERROR("Huh? Fenced object on unfenced list\n");
+ DRM_ERROR("Huh? Fenced object on unfenced list\n");
#endif
}
mutex_unlock(&entry->mutex);
@@ -302,8 +310,8 @@ int drm_fence_buffer_objects(drm_file_t * priv,
*used_fence = fence;
return ret;
}
-EXPORT_SYMBOL(drm_fence_buffer_objects);
+EXPORT_SYMBOL(drm_fence_buffer_objects);
/*
* Call bo->mutex locked.
@@ -822,7 +830,7 @@ static int drm_buffer_object_map(drm_file_t * priv, uint32_t handle,
while (1) {
if (atomic_inc_and_test(&bo->mapped)) {
- if (no_wait && drm_bo_busy(bo)) {
+ if (no_wait && drm_bo_busy(bo)) {
atomic_dec(&bo->mapped);
ret = -EBUSY;
goto out;
@@ -1109,7 +1117,7 @@ static int drm_bo_handle_wait(drm_file_t * priv, uint32_t handle,
drm_buffer_object_t *bo;
int no_wait = hint & DRM_BO_HINT_DONT_BLOCK;
int ret;
-
+
bo = drm_lookup_buffer_object(priv, handle, 1);
if (!bo) {
return -EINVAL;
@@ -1118,13 +1126,12 @@ static int drm_bo_handle_wait(drm_file_t * priv, uint32_t handle,
ret = drm_bo_wait_unfenced(bo, no_wait, 0);
if (ret)
goto out;
- ret = drm_bo_wait(bo, hint & DRM_BO_HINT_WAIT_LAZY,
- 0, no_wait);
+ ret = drm_bo_wait(bo, hint & DRM_BO_HINT_WAIT_LAZY, 0, no_wait);
if (ret)
goto out;
drm_bo_fill_rep_arg(bo, rep);
-out:
+ out:
mutex_unlock(&bo->mutex);
drm_bo_usage_deref_unlocked(bo->dev, bo);
return 0;
@@ -1204,11 +1211,10 @@ int drm_buffer_object_create(drm_file_t * priv,
int ret = 0;
uint32_t new_flags;
unsigned long num_pages;
-
+
drm_bo_delayed_delete(dev);
- if ((buffer_start & ~PAGE_MASK) &&
- (type != drm_bo_type_fake)) {
+ if ((buffer_start & ~PAGE_MASK) && (type != drm_bo_type_fake)) {
DRM_ERROR("Invalid buffer object start.\n");
return -EINVAL;
}
@@ -1350,8 +1356,7 @@ int drm_bo_ioctl(DRM_IOCTL_ARGS)
case drm_bo_map:
rep.ret = drm_buffer_object_map(priv, req->handle,
req->mask,
- req->hint,
- &rep);
+ req->hint, &rep);
break;
case drm_bo_destroy:
mutex_lock(&dev->struct_mutex);
@@ -1404,7 +1409,7 @@ int drm_bo_ioctl(DRM_IOCTL_ARGS)
rep.ret = drm_bo_handle_info(priv, req->handle, &rep);
break;
case drm_bo_wait_idle:
- rep.ret = drm_bo_handle_wait(priv, req->handle,
+ rep.ret = drm_bo_handle_wait(priv, req->handle,
req->hint, &rep);
break;
case drm_bo_ref_fence:
@@ -1443,7 +1448,7 @@ static void drm_bo_force_clean(drm_device_t * dev)
int ret = 0;
l = bm->ddestroy.next;
- while(l != &bm->ddestroy) {
+ while (l != &bm->ddestroy) {
entry = list_entry(l, drm_buffer_object_t, ddestroy);
list_del(l);
if (entry->fence) {
@@ -1584,8 +1589,7 @@ int drm_mm_init_ioctl(DRM_IOCTL_ARGS)
DRM_ERROR("Memory manager not clean. "
"Delaying takedown\n");
}
- DRM_DEBUG("We have %ld still locked pages\n",
- bm->cur_pages);
+ DRM_DEBUG("We have %ld still locked pages\n", bm->cur_pages);
break;
default:
DRM_ERROR("Function not implemented yet\n");
diff --git a/linux-core/drm_fence.c b/linux-core/drm_fence.c
index df5db702..4ee5d2d0 100644
--- a/linux-core/drm_fence.c
+++ b/linux-core/drm_fence.c
@@ -112,7 +112,7 @@ void drm_fence_handler(drm_device_t * dev, uint32_t sequence, uint32_t type)
if ((fence->signaled | relevant) != fence->signaled) {
fence->signaled |= relevant;
#ifdef BODEBUG
- DRM_ERROR("Fence 0x%08lx signaled 0x%08x\n",
+ DRM_ERROR("Fence 0x%08lx signaled 0x%08x\n",
fence->base.hash.key, fence->signaled);
#endif
fence->submitted_flush |= relevant;
@@ -135,7 +135,7 @@ void drm_fence_handler(drm_device_t * dev, uint32_t sequence, uint32_t type)
if (!(fence->type & ~fence->signaled)) {
#ifdef BODEBUG
- DRM_ERROR("Fence completely signaled 0x%08lx\n",
+ DRM_ERROR("Fence completely signaled 0x%08lx\n",
fence->base.hash.key);
#endif
fence_list = &fence->ring;
@@ -275,7 +275,7 @@ int drm_fence_object_signaled(drm_fence_object_t * fence, uint32_t type)
*/
int drm_fence_object_flush(drm_device_t * dev,
- drm_fence_object_t * fence, uint32_t type)
+ drm_fence_object_t * fence, uint32_t type)
{
drm_fence_manager_t *fm = &dev->fm;
drm_fence_driver_t *driver = dev->driver->fence_driver;
@@ -341,6 +341,7 @@ int drm_fence_object_wait(drm_device_t * dev, drm_fence_object_t * fence,
drm_fence_driver_t *driver = dev->driver->fence_driver;
int ret = 0;
unsigned long _end;
+ int signaled;
if (mask & ~fence->type) {
DRM_ERROR("Wait trying to extend fence type\n");
@@ -353,25 +354,51 @@ int drm_fence_object_wait(drm_device_t * dev, drm_fence_object_t * fence,
_end = jiffies + 3 * DRM_HZ;
drm_fence_object_flush(dev, fence, mask);
+
if (lazy && driver->lazy_capable) {
+
do {
DRM_WAIT_ON(ret, fm->fence_queue, 3 * DRM_HZ,
fence_signaled(dev, fence, mask, 1));
if (time_after_eq(jiffies, _end))
break;
- } while (ret == -EINTR && ignore_signals);
+ } while (ret == -EINTR && ignore_signals);
if (time_after_eq(jiffies, _end) && (ret != 0))
ret = -EBUSY;
- if (ret)
+ if (ret)
return ((ret == -EINTR) ? -EAGAIN : ret);
- } else {
- int signaled;
+
+ } else if ((fence->class == 0) && (mask & DRM_FENCE_EXE) &&
+ driver->lazy_capable) {
+
+ /*
+ * We use IRQ wait for EXE fence if available to gain
+ * CPU in some cases.
+ */
+
do {
- signaled = fence_signaled(dev, fence, mask, 1);
- } while (!signaled && !time_after_eq(jiffies, _end));
- if (!signaled)
- return -EBUSY;
+ DRM_WAIT_ON(ret, fm->fence_queue, 3 * DRM_HZ,
+ fence_signaled(dev, fence, DRM_FENCE_EXE,
+ 1));
+ if (time_after_eq(jiffies, _end))
+ break;
+ } while (ret == -EINTR && ignore_signals);
+ if (time_after_eq(jiffies, _end) && (ret != 0))
+ ret = -EBUSY;
+ if (ret)
+ return ((ret == -EINTR) ? -EAGAIN : ret);
}
+
+ if (fence_signaled(dev, fence, mask, 0))
+ return 0;
+
+ do {
+ signaled = fence_signaled(dev, fence, mask, 1);
+ } while (!signaled && !time_after_eq(jiffies, _end));
+
+ if (!signaled)
+ return -EBUSY;
+
return 0;
}
@@ -413,7 +440,7 @@ int drm_fence_object_init(drm_device_t * dev, uint32_t type, int emit,
write_lock_irqsave(&fm->lock, flags);
INIT_LIST_HEAD(&fence->ring);
- fence->class = 0;
+ fence->class = 0;
fence->type = type;
fence->flush_mask = 0;
fence->submitted_flush = 0;
@@ -428,7 +455,7 @@ int drm_fence_object_init(drm_device_t * dev, uint32_t type, int emit,
EXPORT_SYMBOL(drm_fence_object_init);
-int drm_fence_add_user_object(drm_file_t *priv, drm_fence_object_t *fence,
+int drm_fence_add_user_object(drm_file_t * priv, drm_fence_object_t * fence,
int shareable)
{
drm_device_t *dev = priv->head->dev;
@@ -446,10 +473,11 @@ int drm_fence_add_user_object(drm_file_t *priv, drm_fence_object_t *fence,
#endif
return 0;
}
+
EXPORT_SYMBOL(drm_fence_add_user_object);
-int drm_fence_object_create(drm_device_t *dev, uint32_t type,
- int emit, drm_fence_object_t **c_fence)
+int drm_fence_object_create(drm_device_t * dev, uint32_t type,
+ int emit, drm_fence_object_t ** c_fence)
{
drm_fence_object_t *fence;
int ret;
@@ -462,11 +490,11 @@ int drm_fence_object_create(drm_device_t *dev, uint32_t type,
drm_fence_usage_deref_unlocked(dev, fence);
return ret;
}
- *c_fence = fence;
+ *c_fence = fence;
return 0;
}
-EXPORT_SYMBOL(drm_fence_object_create);
+EXPORT_SYMBOL(drm_fence_object_create);
void drm_fence_manager_init(drm_device_t * dev)
{
@@ -535,8 +563,8 @@ int drm_fence_ioctl(DRM_IOCTL_ARGS)
&fence);
if (ret)
return ret;
- ret = drm_fence_add_user_object(priv, fence,
- arg.flags &
+ ret = drm_fence_add_user_object(priv, fence,
+ arg.flags &
DRM_FENCE_FLAG_SHAREABLE);
if (ret) {
drm_fence_usage_deref_unlocked(dev, fence);
@@ -588,8 +616,7 @@ int drm_fence_ioctl(DRM_IOCTL_ARGS)
ret =
drm_fence_object_wait(dev, fence,
arg.flags & DRM_FENCE_FLAG_WAIT_LAZY,
- 0,
- arg.type);
+ 0, arg.type);
break;
case drm_fence_emit:
LOCK_TEST_WITH_RETURN(dev, filp);
@@ -605,16 +632,16 @@ int drm_fence_ioctl(DRM_IOCTL_ARGS)
}
LOCK_TEST_WITH_RETURN(dev, filp);
ret = drm_fence_buffer_objects(priv, NULL, NULL, &fence);
- if (ret)
+ if (ret)
return ret;
- ret = drm_fence_add_user_object(priv, fence,
- arg.flags &
+ ret = drm_fence_add_user_object(priv, fence,
+ arg.flags &
DRM_FENCE_FLAG_SHAREABLE);
if (ret)
return ret;
atomic_inc(&fence->usage);
arg.handle = fence->base.hash.key;
- break;
+ break;
default:
return -EINVAL;
}
diff --git a/linux-core/i915_buffer.c b/linux-core/i915_buffer.c
index 598f8772..9e8ae4a9 100644
--- a/linux-core/i915_buffer.c
+++ b/linux-core/i915_buffer.c
@@ -53,6 +53,10 @@ int i915_fence_types(uint32_t buffer_flags, uint32_t *class, uint32_t *type)
int i915_invalidate_caches(drm_device_t *dev, uint32_t flags)
{
+ /*
+ * FIXME: Only emit once per batchbuffer submission.
+ */
+
uint32_t flush_cmd = MI_NO_WRITE_FLUSH;
if (flags & DRM_BO_FLAG_READ)
@@ -60,5 +64,6 @@ int i915_invalidate_caches(drm_device_t *dev, uint32_t flags)
if (flags & DRM_BO_FLAG_EXE)
flush_cmd |= MI_EXE_FLUSH;
+
return i915_emit_mi_flush(dev, flush_cmd);
}