summaryrefslogtreecommitdiff
path: root/shared-core
diff options
context:
space:
mode:
Diffstat (limited to 'shared-core')
-rw-r--r--shared-core/i915_dma.c282
-rw-r--r--shared-core/nouveau_dma.c29
-rw-r--r--shared-core/nouveau_drm.h6
-rw-r--r--shared-core/nouveau_fifo.c5
-rw-r--r--shared-core/radeon_state.c14
5 files changed, 229 insertions, 107 deletions
diff --git a/shared-core/i915_dma.c b/shared-core/i915_dma.c
index 727aeee0..eee24dd1 100644
--- a/shared-core/i915_dma.c
+++ b/shared-core/i915_dma.c
@@ -769,7 +769,10 @@ struct i915_relocatee_info {
struct drm_i915_validate_buffer {
struct drm_buffer_object *buffer;
+ struct drm_bo_info_rep rep;
int presumed_offset_correct;
+ void __user *data;
+ int ret;
};
static void i915_dereference_buffers_locked(struct drm_i915_validate_buffer *buffers,
@@ -790,6 +793,10 @@ int i915_apply_reloc(struct drm_file *file_priv, int num_buffers,
int ret, i;
int buf_index = -1;
+ /*
+ * FIXME: O(relocs * buffers) complexity.
+ */
+
for (i = 0; i <= num_buffers; i++)
if (buffers[i].buffer)
if (reloc[2] == buffers[i].buffer->base.hash.key)
@@ -811,21 +818,14 @@ int i915_apply_reloc(struct drm_file *file_priv, int num_buffers,
if (!relocatee->data_page ||
!drm_bo_same_page(relocatee->offset, new_cmd_offset)) {
drm_bo_kunmap(&relocatee->kmap);
+ relocatee->data_page = NULL;
relocatee->offset = new_cmd_offset;
- mutex_lock (&relocatee->buf->mutex);
- ret = drm_bo_wait (relocatee->buf, 0, 0, FALSE);
- mutex_unlock (&relocatee->buf->mutex);
- if (ret) {
- DRM_ERROR("Could not wait for buffer to apply relocs\n %08lx", new_cmd_offset);
- return ret;
- }
ret = drm_bo_kmap(relocatee->buf, new_cmd_offset >> PAGE_SHIFT,
1, &relocatee->kmap);
if (ret) {
DRM_ERROR("Could not map command buffer to apply relocs\n %08lx", new_cmd_offset);
return ret;
}
-
relocatee->data_page = drm_bmo_virtual(&relocatee->kmap,
&relocatee->is_iomem);
relocatee->page_offset = (relocatee->offset & PAGE_MASK);
@@ -844,7 +844,11 @@ int i915_apply_reloc(struct drm_file *file_priv, int num_buffers,
reloc[0], reloc[1], buf_index, relocatee->data_page[index], val);
}
}
- relocatee->data_page[index] = val;
+
+ if (relocatee->is_iomem)
+ iowrite32(val, relocatee->data_page + index);
+ else
+ relocatee->data_page[index] = val;
return 0;
}
@@ -912,11 +916,13 @@ int i915_process_relocs(struct drm_file *file_priv,
}
out:
-
if (reloc_buf)
kfree(reloc_buf);
- drm_bo_kunmap(&relocatee->kmap);
- relocatee->data_page = NULL;
+
+ if (relocatee->data_page) {
+ drm_bo_kunmap(&relocatee->kmap);
+ relocatee->data_page = NULL;
+ }
return ret;
}
@@ -956,22 +962,60 @@ static int i915_exec_reloc(struct drm_file *file_priv, drm_handle_t buf_handle,
goto out_err;
}
+ mutex_lock (&relocatee.buf->mutex);
+ ret = drm_bo_wait (relocatee.buf, 0, 0, FALSE);
+ if (ret)
+ goto out_err1;
+
while (reloc_user_ptr) {
ret = i915_process_relocs(file_priv, buf_handle, &reloc_user_ptr, &relocatee, buffers, buf_count);
if (ret) {
DRM_ERROR("process relocs failed\n");
- break;
+ goto out_err1;
}
}
- mutex_lock(&dev->struct_mutex);
- drm_bo_usage_deref_locked(&relocatee.buf);
- mutex_unlock(&dev->struct_mutex);
-
+out_err1:
+ mutex_unlock (&relocatee.buf->mutex);
+ drm_bo_usage_deref_unlocked(&relocatee.buf);
out_err:
return ret;
}
+static int i915_check_presumed(struct drm_i915_op_arg *arg,
+ struct drm_buffer_object *bo,
+ uint32_t __user *data,
+ int *presumed_ok)
+{
+ struct drm_bo_op_req *req = &arg->d.req;
+ uint32_t hint_offset;
+ uint32_t hint = req->bo_req.hint;
+
+ *presumed_ok = 0;
+
+ if (!(hint & DRM_BO_HINT_PRESUMED_OFFSET))
+ return 0;
+ if (bo->offset == req->bo_req.presumed_offset) {
+ *presumed_ok = 1;
+ return 0;
+ }
+
+ /*
+ * We need to turn off the HINT_PRESUMED_OFFSET for this buffer in
+ * the user-space IOCTL argument list, since the buffer has moved,
+ * we're about to apply relocations and we might subsequently
+ * hit an -EAGAIN. In that case the argument list will be reused by
+ * user-space, but the presumed offset is no longer valid.
+ *
+ * Needless to say, this is a bit ugly.
+ */
+
+ hint_offset = (uint32_t *)&req->bo_req.hint - (uint32_t *)arg;
+ hint &= ~DRM_BO_HINT_PRESUMED_OFFSET;
+ return __put_user(hint, data + hint_offset);
+}
+
+
/*
* Validate, add fence and relocate a block of bos from a userspace list
*/
@@ -982,13 +1026,11 @@ int i915_validate_buffer_list(struct drm_file *file_priv,
{
struct drm_i915_op_arg arg;
struct drm_bo_op_req *req = &arg.d.req;
- struct drm_bo_arg_rep rep;
- unsigned long next = 0;
int ret = 0;
unsigned buf_count = 0;
- struct drm_device *dev = file_priv->minor->dev;
uint32_t buf_handle;
uint32_t __user *reloc_user_ptr;
+ struct drm_i915_validate_buffer *item = buffers;
do {
if (buf_count >= *num_buffers) {
@@ -996,31 +1038,26 @@ int i915_validate_buffer_list(struct drm_file *file_priv,
ret = -EINVAL;
goto out_err;
}
+ item = buffers + buf_count;
+ item->buffer = NULL;
+ item->presumed_offset_correct = 0;
buffers[buf_count].buffer = NULL;
- buffers[buf_count].presumed_offset_correct = 0;
if (copy_from_user(&arg, (void __user *)(unsigned long)data, sizeof(arg))) {
ret = -EFAULT;
goto out_err;
}
- if (arg.handled) {
- data = arg.next;
- mutex_lock(&dev->struct_mutex);
- buffers[buf_count].buffer = drm_lookup_buffer_object(file_priv, req->arg_handle, 1);
- mutex_unlock(&dev->struct_mutex);
- buf_count++;
- continue;
- }
-
- rep.ret = 0;
+ ret = 0;
if (req->op != drm_bo_validate) {
DRM_ERROR
("Buffer object operation wasn't \"validate\".\n");
- rep.ret = -EINVAL;
+ ret = -EINVAL;
goto out_err;
}
+ item->ret = 0;
+ item->data = (void __user *) (unsigned long) data;
buf_handle = req->bo_req.handle;
reloc_user_ptr = (uint32_t *)(unsigned long)arg.reloc_ptr;
@@ -1032,48 +1069,146 @@ int i915_validate_buffer_list(struct drm_file *file_priv,
DRM_MEMORYBARRIER();
}
- rep.ret = drm_bo_handle_validate(file_priv, req->bo_req.handle,
- req->bo_req.flags, req->bo_req.mask,
- req->bo_req.hint,
- req->bo_req.fence_class, 0,
- &rep.bo_info,
- &buffers[buf_count].buffer);
+ ret = drm_bo_handle_validate(file_priv, req->bo_req.handle,
+ req->bo_req.flags, req->bo_req.mask,
+ req->bo_req.hint,
+ req->bo_req.fence_class, 0,
+ &item->rep,
+ &item->buffer);
+
+ if (ret) {
+ DRM_ERROR("error on handle validate %d\n", ret);
+ goto out_err;
+ }
+
+ buf_count++;
- if (rep.ret) {
- DRM_ERROR("error on handle validate %d\n", rep.ret);
+ ret = i915_check_presumed(&arg, item->buffer,
+ (uint32_t __user *)
+ (unsigned long) data,
+ &item->presumed_offset_correct);
+ if (ret)
goto out_err;
+
+ data = arg.next;
+ } while (data != 0);
+out_err:
+ *num_buffers = buf_count;
+ item->ret = (ret != -EAGAIN) ? ret : 0;
+ return ret;
+}
+
+
+/*
+ * Remove all buffers from the unfenced list.
+ * If the execbuffer operation was aborted, for example due to a signal,
+ * this also make sure that buffers retain their original state and
+ * fence pointers.
+ * Copy back buffer information to user-space unless we were interrupted
+ * by a signal. In which case the IOCTL must be rerun.
+ */
+
+static int i915_handle_copyback(struct drm_device *dev,
+ struct drm_i915_validate_buffer *buffers,
+ unsigned int num_buffers, int ret)
+{
+ int err = ret;
+ int i;
+ struct drm_i915_op_arg arg;
+
+ if (ret)
+ drm_putback_buffer_objects(dev);
+
+ if (ret != -EAGAIN) {
+ for (i = 0; i < num_buffers; ++i) {
+ arg.handled = 1;
+ arg.d.rep.ret = buffers->ret;
+ arg.d.rep.bo_info = buffers->rep;
+ if (__copy_to_user(buffers->data, &arg, sizeof(arg)))
+ err = -EFAULT;
+ buffers++;
}
+ }
+
+ return err;
+}
+
+/*
+ * Create a fence object, and if that fails, pretend that everything is
+ * OK and just idle the GPU.
+ */
+
+void i915_fence_or_sync(struct drm_file *file_priv,
+ uint32_t fence_flags,
+ struct drm_fence_arg *fence_arg,
+ struct drm_fence_object **fence_p)
+{
+ struct drm_device *dev = file_priv->minor->dev;
+ int ret;
+ struct drm_fence_object *fence;
+
+ ret = drm_fence_buffer_objects(dev, NULL, fence_flags,
+ NULL, &fence);
+
+ if (ret) {
+
/*
- * If the user provided a presumed offset hint, check whether
- * the buffer is in the same place, if so, relocations relative to
- * this buffer need not be performed
+ * Fence creation failed.
+ * Fall back to synchronous operation and idle the engine.
*/
- if ((req->bo_req.hint & DRM_BO_HINT_PRESUMED_OFFSET) &&
- buffers[buf_count].buffer->offset == req->bo_req.presumed_offset) {
- buffers[buf_count].presumed_offset_correct = 1;
- }
- next = arg.next;
- arg.handled = 1;
- arg.d.rep = rep;
+ (void) i915_emit_mi_flush(dev, MI_READ_FLUSH);
+ (void) i915_quiescent(dev);
- if (copy_to_user((void __user *)(unsigned long)data, &arg, sizeof(arg)))
- return -EFAULT;
+ if (!(fence_flags & DRM_FENCE_FLAG_NO_USER)) {
- data = next;
- buf_count++;
+ /*
+ * Communicate to user-space that
+ * fence creation has failed and that
+ * the engine is idle.
+ */
- } while (next != 0);
- *num_buffers = buf_count;
- return 0;
-out_err:
- mutex_lock(&dev->struct_mutex);
- i915_dereference_buffers_locked(buffers, buf_count);
- mutex_unlock(&dev->struct_mutex);
- *num_buffers = 0;
- return (ret) ? ret : rep.ret;
+ fence_arg->handle = ~0;
+ fence_arg->error = ret;
+ }
+
+ drm_putback_buffer_objects(dev);
+ if (fence_p)
+ *fence_p = NULL;
+ return;
+ }
+
+ if (!(fence_flags & DRM_FENCE_FLAG_NO_USER)) {
+
+ ret = drm_fence_add_user_object(file_priv, fence,
+ fence_flags &
+ DRM_FENCE_FLAG_SHAREABLE);
+ if (!ret)
+ drm_fence_fill_arg(fence, fence_arg);
+ else {
+ /*
+ * Fence user object creation failed.
+ * We must idle the engine here as well, as user-
+ * space expects a fence object to wait on. Since we
+ * have a fence object we wait for it to signal
+ * to indicate engine "sufficiently" idle.
+ */
+
+ (void) drm_fence_object_wait(fence, 0, 1,
+ fence->type);
+ drm_fence_usage_deref_unlocked(&fence);
+ fence_arg->handle = ~0;
+ fence_arg->error = ret;
+ }
+ }
+
+ if (fence_p)
+ *fence_p = fence;
+ else if (fence)
+ drm_fence_usage_deref_unlocked(&fence);
}
+
static int i915_execbuffer(struct drm_device *dev, void *data,
struct drm_file *file_priv)
{
@@ -1087,7 +1222,6 @@ static int i915_execbuffer(struct drm_device *dev, void *data,
int num_buffers;
int ret;
struct drm_i915_validate_buffer *buffers;
- struct drm_fence_object *fence;
if (!dev_priv->allow_batchbuffer) {
DRM_ERROR("Batchbuffer ioctl disabled\n");
@@ -1132,7 +1266,7 @@ static int i915_execbuffer(struct drm_device *dev, void *data,
ret = i915_validate_buffer_list(file_priv, 0, exec_buf->ops_list,
buffers, &num_buffers);
if (ret)
- goto out_free;
+ goto out_err0;
/* make sure all previous memory operations have passed */
DRM_MEMORYBARRIER();
@@ -1151,30 +1285,16 @@ static int i915_execbuffer(struct drm_device *dev, void *data,
if (sarea_priv)
sarea_priv->last_dispatch = READ_BREADCRUMB(dev_priv);
- /* fence */
- ret = drm_fence_buffer_objects(dev, NULL, fence_arg->flags,
- NULL, &fence);
- if (ret)
- goto out_err0;
+ i915_fence_or_sync(file_priv, fence_arg->flags, fence_arg, NULL);
- if (!(fence_arg->flags & DRM_FENCE_FLAG_NO_USER)) {
- ret = drm_fence_add_user_object(file_priv, fence, fence_arg->flags & DRM_FENCE_FLAG_SHAREABLE);
- if (!ret) {
- fence_arg->handle = fence->base.hash.key;
- fence_arg->fence_class = fence->fence_class;
- fence_arg->type = fence->type;
- fence_arg->signaled = fence->signaled_types;
- }
- }
- drm_fence_usage_deref_unlocked(&fence);
out_err0:
/* handle errors */
+ ret = i915_handle_copyback(dev, buffers, num_buffers, ret);
mutex_lock(&dev->struct_mutex);
i915_dereference_buffers_locked(buffers, num_buffers);
mutex_unlock(&dev->struct_mutex);
-out_free:
drm_free(buffers, (exec_buf->num_buffers * sizeof(struct drm_buffer_object *)), DRM_MEM_DRIVER);
mutex_unlock(&dev_priv->cmdbuf_mutex);
diff --git a/shared-core/nouveau_dma.c b/shared-core/nouveau_dma.c
index dff786d4..e519dc4e 100644
--- a/shared-core/nouveau_dma.c
+++ b/shared-core/nouveau_dma.c
@@ -29,11 +29,6 @@
#include "nouveau_drv.h"
#include "nouveau_dma.h"
-/* FIXME : should go into a nouveau_drm.h define ?
- * (it's shared between DRI & DDX & DRM)
- */
-#define SKIPS 8
-
int
nouveau_dma_channel_init(struct drm_device *dev)
{
@@ -76,10 +71,10 @@ nouveau_dma_channel_init(struct drm_device *dev)
dchan->cur = dchan->put;
dchan->free = dchan->max - dchan->cur;
- /* Insert NOPS for SKIPS */
- dchan->free -= SKIPS;
- dchan->push_free = SKIPS;
- for (i=0; i<SKIPS; i++)
+ /* Insert NOPS for NOUVEAU_DMA_SKIPS */
+ dchan->free -= NOUVEAU_DMA_SKIPS;
+ dchan->push_free = NOUVEAU_DMA_SKIPS;
+ for (i=0; i < NOUVEAU_DMA_SKIPS; i++)
OUT_RING(0);
/* NV_MEMORY_TO_MEMORY_FORMAT requires a notifier */
@@ -131,8 +126,6 @@ nouveau_dma_channel_takedown(struct drm_device *dev)
}
}
-#define RING_SKIPS 8
-
#define READ_GET() ((NV_READ(dchan->chan->get) - \
dchan->chan->pushbuf_base) >> 2)
#define WRITE_PUT(val) do { \
@@ -156,19 +149,19 @@ nouveau_dma_wait(struct drm_device *dev, int size)
if (dchan->free < size) {
dchan->push_free = 1;
OUT_RING(0x20000000|dchan->chan->pushbuf_base);
- if (get <= RING_SKIPS) {
+ if (get <= NOUVEAU_DMA_SKIPS) {
/*corner case - will be idle*/
- if (dchan->put <= RING_SKIPS)
- WRITE_PUT(RING_SKIPS + 1);
+ if (dchan->put <= NOUVEAU_DMA_SKIPS)
+ WRITE_PUT(NOUVEAU_DMA_SKIPS + 1);
do {
get = READ_GET();
- } while (get <= RING_SKIPS);
+ } while (get <= NOUVEAU_DMA_SKIPS);
}
- WRITE_PUT(RING_SKIPS);
- dchan->cur = dchan->put = RING_SKIPS;
- dchan->free = get - (RING_SKIPS + 1);
+ WRITE_PUT(NOUVEAU_DMA_SKIPS);
+ dchan->cur = dchan->put = NOUVEAU_DMA_SKIPS;
+ dchan->free = get - (NOUVEAU_DMA_SKIPS + 1);
}
} else {
dchan->free = get - dchan->cur - 1;
diff --git a/shared-core/nouveau_drm.h b/shared-core/nouveau_drm.h
index 3bf40089..5f07fcbc 100644
--- a/shared-core/nouveau_drm.h
+++ b/shared-core/nouveau_drm.h
@@ -70,6 +70,12 @@ struct drm_nouveau_gpuobj_free {
uint32_t handle;
};
+/* This is needed to avoid a race condition.
+ * Otherwise you may be writing in the fetch area.
+ * Is this large enough, as it's only 32 bytes, and the maximum fetch size is 256 bytes?
+ */
+#define NOUVEAU_DMA_SKIPS 8
+
#define NOUVEAU_MEM_FB 0x00000001
#define NOUVEAU_MEM_AGP 0x00000002
#define NOUVEAU_MEM_FB_ACCEPTABLE 0x00000004
diff --git a/shared-core/nouveau_fifo.c b/shared-core/nouveau_fifo.c
index 0daf9ac4..056cb6d0 100644
--- a/shared-core/nouveau_fifo.c
+++ b/shared-core/nouveau_fifo.c
@@ -45,11 +45,6 @@ int nouveau_fifo_ctx_size(struct drm_device *dev)
* functions doing the actual work
***********************************/
-/* voir nv_xaa.c : NVResetGraphics
- * mémoire mappée par nv_driver.c : NVMapMem
- * voir nv_driver.c : NVPreInit
- */
-
static int nouveau_fifo_instmem_configure(struct drm_device *dev)
{
struct drm_nouveau_private *dev_priv = dev->dev_private;
diff --git a/shared-core/radeon_state.c b/shared-core/radeon_state.c
index 289fddc7..6ff952e1 100644
--- a/shared-core/radeon_state.c
+++ b/shared-core/radeon_state.c
@@ -1662,7 +1662,7 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
u32 height;
int i;
u32 texpitch, microtile;
- u32 offset;
+ u32 offset, byte_offset;
RING_LOCALS;
if (radeon_check_and_fixup_offset(dev_priv, file_priv, &tex->offset)) {
@@ -1727,6 +1727,13 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
} else
microtile = 0;
+ /* this might fail for zero-sized uploads - are those illegal? */
+ if (!radeon_check_offset(dev_priv, tex->offset + image->height *
+ blit_width - 1)) {
+ DRM_ERROR("Invalid final destination offset\n");
+ return -EINVAL;
+ }
+
DRM_DEBUG("tex=%dx%d blit=%d\n", tex_width, tex->height, blit_width);
do {
@@ -1840,6 +1847,7 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
}
#undef RADEON_COPY_MT
+ byte_offset = (image->y & ~2047) * blit_width;
buf->file_priv = file_priv;
buf->used = size;
offset = dev_priv->gart_buffers_offset + buf->offset;
@@ -1854,9 +1862,9 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
RADEON_DP_SRC_SOURCE_MEMORY |
RADEON_GMC_CLR_CMP_CNTL_DIS | RADEON_GMC_WR_MSK_DIS);
OUT_RING((spitch << 22) | (offset >> 10));
- OUT_RING((texpitch << 22) | (tex->offset >> 10));
+ OUT_RING((texpitch << 22) | ((tex->offset >> 10) + (byte_offset >> 10)));
OUT_RING(0);
- OUT_RING((image->x << 16) | image->y);
+ OUT_RING((image->x << 16) | (image->y % 2048));
OUT_RING((image->width << 16) | height);
RADEON_WAIT_UNTIL_2D_IDLE();
ADVANCE_RING();