summaryrefslogtreecommitdiff
path: root/libdrm
diff options
context:
space:
mode:
Diffstat (limited to 'libdrm')
-rw-r--r--libdrm/xf86drm.c58
1 files changed, 31 insertions, 27 deletions
diff --git a/libdrm/xf86drm.c b/libdrm/xf86drm.c
index a7d4beb4..31d1c164 100644
--- a/libdrm/xf86drm.c
+++ b/libdrm/xf86drm.c
@@ -2252,6 +2252,7 @@ int drmFenceCreate(int fd, unsigned flags, int class,unsigned type,
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.type = type;
arg.class = class;
arg.op = drm_fence_create;
@@ -2275,6 +2276,7 @@ int drmFenceBuffers(int fd, unsigned flags, drmFence *fence)
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.flags = flags;
arg.op = drm_fence_buffers;
if (ioctl(fd, DRM_IOCTL_FENCE, &arg))
@@ -2291,6 +2293,7 @@ int drmFenceDestroy(int fd, const drmFence *fence)
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.handle = fence->handle;
arg.op = drm_fence_destroy;
if (ioctl(fd, DRM_IOCTL_FENCE, &arg))
@@ -2302,6 +2305,7 @@ int drmFenceReference(int fd, unsigned handle, drmFence *fence)
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.handle = handle;
arg.op = drm_fence_reference;
if (ioctl(fd, DRM_IOCTL_FENCE, &arg))
@@ -2318,6 +2322,7 @@ int drmFenceUnreference(int fd, const drmFence *fence)
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.handle = fence->handle;
arg.op = drm_fence_unreference;
if (ioctl(fd, DRM_IOCTL_FENCE, &arg))
@@ -2329,6 +2334,7 @@ int drmFenceFlush(int fd, drmFence *fence, unsigned flush_type)
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.handle = fence->handle;
arg.type = flush_type;
arg.op = drm_fence_flush;
@@ -2344,6 +2350,7 @@ int drmFenceUpdate(int fd, drmFence *fence)
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.handle = fence->handle;
arg.op = drm_fence_signaled;
if (ioctl(fd, DRM_IOCTL_FENCE, &arg))
@@ -2384,6 +2391,7 @@ int drmFenceEmit(int fd, unsigned flags, drmFence *fence, unsigned emit_type)
{
drm_fence_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.flags = flags;
arg.handle = fence->handle;
arg.type = emit_type;
@@ -2417,6 +2425,7 @@ int drmFenceWait(int fd, unsigned flags, drmFence *fence, unsigned flush_type)
}
}
+ memset(&arg, 0, sizeof(arg));
arg.handle = fence->handle;
arg.type = flush_type;
arg.flags = flags;
@@ -2438,6 +2447,7 @@ int drmTTMCreate(int fd, drmTTM *ttm, unsigned long size, unsigned flags)
{
drm_ttm_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.op = drm_ttm_create;
arg.flags = flags;
arg.size = size;
@@ -2458,6 +2468,7 @@ int drmTTMDestroy(int fd, const drmTTM *ttm)
{
drm_ttm_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.op = drm_ttm_destroy;
arg.handle = ttm->handle;
if (ioctl(fd, DRM_IOCTL_TTM, &arg))
@@ -2470,6 +2481,7 @@ int drmTTMReference(int fd, unsigned handle, drmTTM *ttm)
{
drm_ttm_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.handle = handle;
arg.op = drm_ttm_reference;
if (ioctl(fd, DRM_IOCTL_TTM, &arg))
@@ -2485,6 +2497,7 @@ int drmTTMUnreference(int fd, const drmTTM *ttm)
{
drm_ttm_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.op = drm_ttm_destroy;
arg.handle = ttm->handle;
if (ioctl(fd, DRM_IOCTL_TTM, &arg))
@@ -2660,8 +2673,9 @@ int drmBOCreate(int fd, drmTTM *ttm, unsigned long start, unsigned long size,
drm_bo_arg_t arg;
drm_bo_arg_request_t *req = &arg.d.req;
drm_bo_arg_reply_t *rep = &arg.d.rep;
-
- arg.handled = 0;
+
+ memset(buf, 0, sizeof(*buf));
+ memset(&arg, 0, sizeof(arg));
req->mask = mask;
req->hint = hint;
req->size = size;
@@ -2690,7 +2704,6 @@ int drmBOCreate(int fd, drmTTM *ttm, unsigned long start, unsigned long size,
return -EINVAL;
}
req->op = drm_bo_create;
- arg.next = 0;
if (ioctl(fd, DRM_IOCTL_BUFOBJ, &arg))
return -errno;
@@ -2721,10 +2734,9 @@ int drmBODestroy(int fd, drmBO *buf)
buf->virtual = NULL;
}
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->op = drm_bo_destroy;
- arg.next = 0;
if (ioctl(fd, DRM_IOCTL_BUFOBJ, &arg))
return -errno;
@@ -2746,10 +2758,9 @@ int drmBOReference(int fd, unsigned handle, drmBO *buf)
drm_bo_arg_request_t *req = &arg.d.req;
drm_bo_arg_reply_t *rep = &arg.d.rep;
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = handle;
req->op = drm_bo_reference;
- arg.next = 0;
if (ioctl(fd, DRM_IOCTL_BUFOBJ, &arg))
return -errno;
@@ -2782,10 +2793,9 @@ int drmBOUnReference(int fd, drmBO *buf)
buf->virtual = NULL;
}
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->op = drm_bo_unreference;
- arg.next = 0;
if (ioctl(fd, DRM_IOCTL_BUFOBJ, &arg))
return -errno;
@@ -2832,12 +2842,11 @@ int drmBOMap(int fd, drmBO *buf, unsigned mapFlags, unsigned mapHint,
#endif
}
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->mask = mapFlags;
req->hint = mapHint;
req->op = drm_bo_map;
- arg.next = 0;
/*
* May hang if the buffer object is busy.
@@ -2870,10 +2879,9 @@ int drmBOUnmap(int fd, drmBO *buf)
drm_bo_arg_reply_t *rep = &arg.d.rep;
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->op = drm_bo_unmap;
- arg.next = 0;
if (ioctl(fd, DRM_IOCTL_BUFOBJ, &arg)) {
return -errno;
@@ -2894,20 +2902,17 @@ int drmBOValidate(int fd, drmBO *buf, unsigned flags, unsigned mask,
drm_bo_arg_reply_t *rep = &arg.d.rep;
int ret = 0;
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->mask = flags;
req->hint = hint;
req->arg_handle = mask; /* Encode mask in the arg_handle field :/ */
req->op = drm_bo_validate;
- arg.next = 0;
-
do{
ret = ioctl(fd, DRM_IOCTL_BUFOBJ, &arg);
} while (ret && errno == EAGAIN);
-
if (ret)
return ret;
if (!arg.handled)
@@ -2927,12 +2932,11 @@ int drmBOFence(int fd, drmBO *buf, unsigned flags, unsigned fenceHandle)
drm_bo_arg_reply_t *rep = &arg.d.rep;
int ret = 0;
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->mask = flags;
req->arg_handle = fenceHandle;
req->op = drm_bo_validate;
- arg.next = 0;
ret = ioctl(fd, DRM_IOCTL_BUFOBJ, &arg);
@@ -2952,10 +2956,9 @@ int drmBOInfo(int fd, drmBO *buf)
drm_bo_arg_reply_t *rep = &arg.d.rep;
int ret = 0;
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->op = drm_bo_info;
- arg.next = 0;
ret = ioctl(fd, DRM_IOCTL_BUFOBJ, &arg);
@@ -2978,11 +2981,10 @@ int drmBOWaitIdle(int fd, drmBO *buf, unsigned hint)
if ((buf->flags & DRM_BO_FLAG_SHAREABLE) ||
(buf->replyFlags & DRM_BO_REP_BUSY)) {
- arg.handled = 0;
+ memset(&arg, 0, sizeof(arg));
req->handle = buf->handle;
req->op = drm_bo_wait_idle;
req->hint = hint;
- arg.next = 0;
do {
ret = ioctl(fd, DRM_IOCTL_BUFOBJ, &arg);
@@ -3098,9 +3100,8 @@ int drmBOValidateList(int fd, drmBOList *list)
if (prevNext)
*prevNext = (unsigned long) arg;
- arg->next = 0;
+ memset(arg, 0, sizeof(*arg));
prevNext = &arg->next;
- arg->handled = 0;
req->handle = node->buf->handle;
req->op = drm_bo_validate;
req->mask = node->arg0;
@@ -3166,9 +3167,8 @@ int drmBOFenceList(int fd, drmBOList *list, unsigned fenceHandle)
if (prevNext)
*prevNext = (unsigned long) arg;
- arg->next = 0;
+ memset(arg, 0, sizeof(*arg));
prevNext = &arg->next;
- arg->handled = 0;
req->handle = node->buf->handle;
req->op = drm_bo_fence;
req->mask = node->arg0;
@@ -3205,6 +3205,7 @@ int drmMMInit(int fd, unsigned long vramPOffset, unsigned long vramPSize,
{
drm_mm_init_arg_t arg;
+ memset(&arg, 0, sizeof(arg));
arg.req.op = mm_init;
arg.req.vr_p_offset = vramPOffset;
arg.req.vr_p_size = vramPSize;
@@ -3221,6 +3222,9 @@ int drmMMInit(int fd, unsigned long vramPOffset, unsigned long vramPSize,
int drmMMTakedown(int fd)
{
drm_mm_init_arg_t arg;
+
+
+ memset(&arg, 0, sizeof(arg));
arg.req.op = mm_takedown;
if (ioctl(fd, DRM_IOCTL_MM_INIT, &arg))
id='n536' href='#n536'>536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602
/* via_dma.c -- DMA support for the VIA Unichrome/Pro
 */
/**************************************************************************
 * 
 * Copyright 2003 Tungsten Graphics, Inc., Cedar Park, Texas.
 * All Rights Reserved.
 *
 * Copyright 2004 Digeo, Inc., Palo Alto, CA, U.S.A.
 * All Rights Reserved.
 * 
 **************************************************************************/

#include "via.h"
#include "drmP.h"
#include "drm.h"
#include "via_drm.h"
#include "via_drv.h"

#define VIA_2D_CMD 0xF0000000

static void via_cmdbuf_start(drm_via_private_t * dev_priv);
static void via_cmdbuf_pause(drm_via_private_t * dev_priv);
static void via_cmdbuf_reset(drm_via_private_t * dev_priv);
static void via_cmdbuf_rewind(drm_via_private_t * dev_priv);
static int  via_wait_idle(drm_via_private_t * dev_priv);

static inline int
via_cmdbuf_wait(drm_via_private_t * dev_priv, unsigned int size)
{
	uint32_t agp_base = dev_priv->dma_offset + (uint32_t) dev_priv->agpAddr;
	uint32_t cur_addr, hw_addr, next_addr;
	volatile uint32_t * hw_addr_ptr;
	uint32_t count;
	hw_addr_ptr = dev_priv->hw_addr_ptr;
	cur_addr = agp_base + dev_priv->dma_low;
	/* At high resolution (i.e. 1280x1024) and with high workload within
	 * a short commmand stream, the following test will fail. It may be
	 * that the engine is too busy to update hw_addr. Therefore, add
	 * a large 64KB window between buffer head and tail.
	 */
	next_addr = cur_addr + size + 64 * 1024;
	count = 1000000; /* How long is this? */
	do {
		hw_addr = *hw_addr_ptr;
		if (count-- == 0) {
			DRM_ERROR("via_cmdbuf_wait timed out hw %x dma_low %x\n",
					hw_addr, dev_priv->dma_low);
			return -1;
		}
	} while ((cur_addr < hw_addr) && (next_addr >= hw_addr));
	return 0;
}

/*
 * Checks whether buffer head has reach the end. Rewind the ring buffer
 * when necessary.
 *
 * Returns virtual pointer to ring buffer.
 */
static inline uint32_t *
via_check_dma(drm_via_private_t * dev_priv, unsigned int size)
{
	if ((dev_priv->dma_low + size + 0x400) > dev_priv->dma_high) {
		via_cmdbuf_rewind(dev_priv);
	}
	if (via_cmdbuf_wait(dev_priv, size) != 0) {
		return NULL;
	}

	return (uint32_t*)(dev_priv->dma_ptr + dev_priv->dma_low);
}

int via_dma_cleanup(drm_device_t *dev)
{
	if (dev->dev_private) {
		drm_via_private_t *dev_priv = 
				(drm_via_private_t *) dev->dev_private;

		if (dev_priv->ring.virtual_start) {
			via_cmdbuf_reset(dev_priv);

			drm_core_ioremapfree( &dev_priv->ring.map, dev);
			dev_priv->ring.virtual_start = NULL;
		}
	}

	return 0;
}


static int via_initialize(drm_device_t *dev, 
			       drm_via_private_t *dev_priv,
			       drm_via_dma_init_t *init)
{
	if (!dev_priv || !dev_priv->mmio) {
		DRM_ERROR("via_dma_init called before via_map_init\n");
		return DRM_ERR(EFAULT);
	}

	if (dev_priv->ring.virtual_start != NULL) {
		DRM_ERROR("%s called again without calling cleanup\n",
				__FUNCTION__);
		return DRM_ERR(EFAULT);
	}

	dev_priv->ring.map.offset = dev->agp->base + init->offset;
	dev_priv->ring.map.size = init->size;
	dev_priv->ring.map.type = 0;
	dev_priv->ring.map.flags = 0;
	dev_priv->ring.map.mtrr = 0;

	drm_core_ioremap( &dev_priv->ring.map, dev );

	if (dev_priv->ring.map.handle == NULL) {
		via_dma_cleanup(dev);
		DRM_ERROR("can not ioremap virtual address for"
				" ring buffer\n");
		return DRM_ERR(ENOMEM);
	}

	dev_priv->ring.virtual_start = dev_priv->ring.map.handle;

	dev_priv->dma_ptr = dev_priv->ring.virtual_start;
	dev_priv->dma_low = 0;
	dev_priv->dma_high = init->size;
	dev_priv->dma_offset = init->offset;
	dev_priv->last_pause_ptr = NULL;
	dev_priv->hw_addr_ptr = dev_priv->mmio->handle + init->reg_pause_addr;

	via_cmdbuf_start(dev_priv);

	return 0;
}


int via_dma_init( DRM_IOCTL_ARGS )
{
	DRM_DEVICE;
	drm_via_private_t *dev_priv = (drm_via_private_t *)dev->dev_private;
	drm_via_dma_init_t init;
	int retcode = 0;

	DRM_COPY_FROM_USER_IOCTL(init, (drm_via_dma_init_t *)data, sizeof(init));

	switch(init.func) {
	case VIA_INIT_DMA:
		retcode = via_initialize(dev, dev_priv, &init);
		break;
	case VIA_CLEANUP_DMA:
		retcode = via_dma_cleanup(dev);
		break;
	default:
		retcode = DRM_ERR(EINVAL);
		break;
	}

	return retcode;
}


static int via_dispatch_cmdbuffer(drm_device_t *dev, 
				   drm_via_cmdbuffer_t *cmd )
{
	drm_via_private_t *dev_priv = dev->dev_private;
	uint32_t * vb;
	vb = via_check_dma(dev_priv, cmd->size);
	if (vb == NULL) {
		return DRM_ERR(EAGAIN);
	}
	if (DRM_COPY_FROM_USER(vb, cmd->buf, cmd->size)) {
		return DRM_ERR(EFAULT);
	}
	dev_priv->dma_low += cmd->size;
	via_cmdbuf_pause(dev_priv);

	return 0;
}


static int via_quiescent(drm_device_t *dev)
{
	drm_via_private_t *dev_priv = dev->dev_private;

	if (!via_wait_idle(dev_priv)) {
		return DRM_ERR(EAGAIN);
	}
	return 0;
}


int via_flush_ioctl( DRM_IOCTL_ARGS )
{
	DRM_DEVICE;

	if(!_DRM_LOCK_IS_HELD(dev->lock.hw_lock->lock)) {
		DRM_ERROR("via_flush_ioctl called without lock held\n");
		return DRM_ERR(EINVAL);
	}

	return via_quiescent(dev); 
}


int via_cmdbuffer( DRM_IOCTL_ARGS )
{
	DRM_DEVICE;
	drm_via_cmdbuffer_t cmdbuf;
	int ret;

	DRM_COPY_FROM_USER_IOCTL( cmdbuf, (drm_via_cmdbuffer_t *)data, 
			sizeof(cmdbuf) );

	DRM_DEBUG("via cmdbuffer, buf %p size %lu\n", cmdbuf.buf, cmdbuf.size);

	if(!_DRM_LOCK_IS_HELD(dev->lock.hw_lock->lock)) {
		DRM_ERROR("via_cmdbuffer called without lock held\n");
		return DRM_ERR(EINVAL);
	}

	ret = via_dispatch_cmdbuffer( dev, &cmdbuf );
	if (ret) {
		return ret;
	}

	return 0;
}

static int via_parse_pci_cmdbuffer( drm_device_t *dev, const char *buf, 
				    unsigned int size )
{
	drm_via_private_t *dev_priv = dev->dev_private;
	uint32_t offset, value;
	const uint32_t *regbuf = (uint32_t *)buf;
	unsigned int i;

	size >>=3 ;
	for (i=0; i<size; ++i) {
		offset = *regbuf; 
		regbuf += 2;
		if ((offset > ((0x7FF >> 2) | VIA_2D_CMD)) && 
		    (offset < ((0xC00 >> 2) | VIA_2D_CMD)) ) {
			DRM_DEBUG("Attempt to access Burst Command Area.\n");
			return DRM_ERR( EINVAL );
		} else if (offset > ((0xDFF >> 2) | VIA_2D_CMD)) {
			DRM_DEBUG("Attempt to access DMA or VGA registers.\n");
			return DRM_ERR( EINVAL );
		}
	}
			
	regbuf = (uint32_t *)buf;
	for ( i=0; i<size; ++i ) {
		offset = (*regbuf++ & ~VIA_2D_CMD) << 2;
		value = *regbuf++;
		VIA_WRITE( offset, value );
	}
	return 0;
}
		
static int via_dispatch_pci_cmdbuffer(drm_device_t *dev, 
				      drm_via_cmdbuffer_t *cmd )
{
	drm_via_private_t *dev_priv = dev->dev_private;
	char *hugebuf;
	int ret;

	/*
	 * We must be able to parse the buffer all at a time, so as
	 * to return an error on an invalid operation without doing
	 * anything.
	 * Small buffers must, on the other hand be handled fast.
	 */

	if ( cmd->size > VIA_MAX_PCI_SIZE ) {
		return DRM_ERR( ENOMEM );
	} else if ( cmd->size > VIA_PREALLOCATED_PCI_SIZE ) {
	  if (NULL == (hugebuf = (char *) kmalloc( cmd-> size, GFP_KERNEL )))
			return DRM_ERR( ENOMEM );
		if (DRM_COPY_FROM_USER( hugebuf, cmd->buf, cmd->size ))
			return DRM_ERR(EFAULT);
		ret = via_parse_pci_cmdbuffer( dev, hugebuf, cmd->size );
		kfree( hugebuf );
	} else {
		if (DRM_COPY_FROM_USER( dev_priv->pci_buf, cmd->buf, cmd->size ))
			return DRM_ERR(EFAULT);
		ret = via_parse_pci_cmdbuffer( dev, dev_priv->pci_buf, cmd->size );
	}
	return ret;
}



int via_pci_cmdbuffer( DRM_IOCTL_ARGS )
{
	DRM_DEVICE;
	drm_via_cmdbuffer_t cmdbuf;
	int ret;

	DRM_COPY_FROM_USER_IOCTL( cmdbuf, (drm_via_cmdbuffer_t *)data, 
			sizeof(cmdbuf) );

	DRM_DEBUG("via_pci_cmdbuffer, buf %p size %lu\n", cmdbuf.buf, cmdbuf.size);

	if(!_DRM_LOCK_IS_HELD(dev->lock.hw_lock->lock)) {
		DRM_ERROR("via_pci_cmdbuffer called without lock held\n");
		return DRM_ERR(EINVAL);
	}

	ret = via_dispatch_pci_cmdbuffer( dev, &cmdbuf );
	if (ret) {
		return ret;
	}

	return 0;
}




/************************************************************************/
#include "via_3d_reg.h"

#define CMDBUF_ALIGNMENT_SIZE   (0x100)
#define CMDBUF_ALIGNMENT_MASK   (0xff)

/* defines for VIA 3D registers */
#define VIA_REG_STATUS          0x400
#define VIA_REG_TRANSET         0x43C
#define VIA_REG_TRANSPACE       0x440
                                                                                                                                                                                  
/* VIA_REG_STATUS(0x400): Engine Status */
#define VIA_CMD_RGTR_BUSY       0x00000080  /* Command Regulator is busy */
#define VIA_2D_ENG_BUSY         0x00000001  /* 2D Engine is busy */
#define VIA_3D_ENG_BUSY         0x00000002  /* 3D Engine is busy */
#define VIA_VR_QUEUE_BUSY       0x00020000 /* Virtual Queue is busy */


#define SetReg2DAGP(nReg, nData) { \
	*((uint32_t *)(vb)) = ((nReg) >> 2) | 0xF0000000; \
	*((uint32_t *)(vb) + 1) = (nData); \
	vb = ((uint32_t *)vb) + 2; \
	dev_priv->dma_low +=8; \
}

static uint32_t via_swap_count = 0;

static inline uint32_t *
via_align_buffer(drm_via_private_t * dev_priv, uint32_t * vb, int qw_count)
{
	for ( ; qw_count > 0; --qw_count) {
		*vb++ = (0xcc000000 | (dev_priv->dma_low & 0xffffff));
		*vb++ = (0xdd400000 | via_swap_count);
		dev_priv->dma_low += 8;
	}
	via_swap_count = (via_swap_count + 1) & 0xffff;
	return vb;
}

/*
 * This function is used internally by ring buffer mangement code.
 *
 * Returns virtual pointer to ring buffer.
 */
static inline uint32_t * via_get_dma(drm_via_private_t * dev_priv)
{
	return (uint32_t*)(dev_priv->dma_ptr + dev_priv->dma_low);
}


static int via_wait_idle(drm_via_private_t * dev_priv)
{
	int count = 10000000;
	while (count-- && (VIA_READ(VIA_REG_STATUS) &
		(VIA_CMD_RGTR_BUSY | VIA_2D_ENG_BUSY | VIA_3D_ENG_BUSY)));
	return count;
}

static inline void
via_dummy_bitblt(drm_via_private_t * dev_priv)
{
	uint32_t * vb = via_get_dma(dev_priv);
	/* GEDST*/
	SetReg2DAGP(0x0C, (0 | (0 << 16)));
	/* GEWD*/
	SetReg2DAGP(0x10, 0 | (0 << 16));
	/* BITBLT*/
	SetReg2DAGP(0x0, 0x1 | 0x2000 | 0xAA000000);
}

static void via_cmdbuf_start(drm_via_private_t * dev_priv)
{
	uint32_t agp_base;
	uint32_t pause_addr, pause_addr_lo, pause_addr_hi;
	uint32_t start_addr, start_addr_lo;
	uint32_t end_addr, end_addr_lo;
	uint32_t qw_pad_count;
	uint32_t command;
	uint32_t * vb;

	dev_priv->dma_low = 0;
	vb = via_get_dma(dev_priv);

	agp_base = dev_priv->dma_offset + (uint32_t) dev_priv->agpAddr;
	start_addr = agp_base;
	end_addr = agp_base + dev_priv->dma_high;

	start_addr_lo = ((HC_SubA_HAGPBstL << 24) | (start_addr & 0xFFFFFF));
	end_addr_lo = ((HC_SubA_HAGPBendL << 24) | (end_addr & 0xFFFFFF));
	command = ((HC_SubA_HAGPCMNT << 24) | (start_addr >> 24) |
			((end_addr & 0xff000000) >> 16));

	*vb++ = HC_HEADER2 | ((VIA_REG_TRANSET>>2)<<12) |
			(VIA_REG_TRANSPACE>>2);
	*vb++ = (HC_ParaType_PreCR<<16);
	dev_priv->dma_low += 8;

	qw_pad_count = (CMDBUF_ALIGNMENT_SIZE>>3) -
			((dev_priv->dma_low & CMDBUF_ALIGNMENT_MASK) >> 3);

	pause_addr = agp_base + dev_priv->dma_low - 8 + (qw_pad_count<<3);
	pause_addr_lo = ((HC_SubA_HAGPBpL<<24) |
			HC_HAGPBpID_PAUSE |
			(pause_addr & 0xffffff));
	pause_addr_hi = ((HC_SubA_HAGPBpH<<24) | (pause_addr >> 24));

	vb = via_align_buffer(dev_priv, vb, qw_pad_count-1);

	*vb++ = pause_addr_hi;
	*vb++ = pause_addr_lo;
	dev_priv->dma_low += 8;
	dev_priv->last_pause_ptr = vb-1;

	VIA_WRITE(VIA_REG_TRANSET, (HC_ParaType_PreCR << 16));
	VIA_WRITE(VIA_REG_TRANSPACE, command);
	VIA_WRITE(VIA_REG_TRANSPACE, start_addr_lo);
	VIA_WRITE(VIA_REG_TRANSPACE, end_addr_lo);

	VIA_WRITE(VIA_REG_TRANSPACE, pause_addr_hi);
	VIA_WRITE(VIA_REG_TRANSPACE, pause_addr_lo);

	VIA_WRITE(VIA_REG_TRANSPACE, command | HC_HAGPCMNT_MASK);
}

static void via_cmdbuf_jump(drm_via_private_t * dev_priv)
{
	uint32_t agp_base;
	uint32_t pause_addr, pause_addr_lo, pause_addr_hi;
	uint32_t start_addr;
	uint32_t end_addr, end_addr_lo;
	uint32_t * vb;
	uint32_t qw_pad_count;
	uint32_t command;
	uint32_t jump_addr, jump_addr_lo, jump_addr_hi;

	/* Seems like Unichrome has bug that when the PAUSE register is
	 * set in the AGP command stream immediately after a PCI write to
	 * the same register, the command regulator goes into a looping
	 * state. Prepending a BitBLT command to stall the command
	 * regulator for a moment seems to solve the problem.
	 */
	via_cmdbuf_wait(dev_priv, 48);
	via_dummy_bitblt(dev_priv);

	via_cmdbuf_wait(dev_priv, 2*CMDBUF_ALIGNMENT_SIZE);

	/* At end of buffer, rewind with a JUMP command. */
	vb = via_get_dma(dev_priv);

	*vb++ = HC_HEADER2 | ((VIA_REG_TRANSET>>2)<<12) |
			(VIA_REG_TRANSPACE>>2);
	*vb++ = (HC_ParaType_PreCR<<16);
	dev_priv->dma_low += 8;

	qw_pad_count = (CMDBUF_ALIGNMENT_SIZE>>3) -
			((dev_priv->dma_low & CMDBUF_ALIGNMENT_MASK) >> 3);

	agp_base = dev_priv->dma_offset + (uint32_t) dev_priv->agpAddr;
	start_addr = agp_base;
	end_addr = agp_base + dev_priv->dma_low - 8 + (qw_pad_count<<3);

	jump_addr = end_addr;
	jump_addr_lo = ((HC_SubA_HAGPBpL<<24) | HC_HAGPBpID_JUMP |
			(jump_addr & 0xffffff));
	jump_addr_hi = ((HC_SubA_HAGPBpH<<24) | (jump_addr >> 24));

	end_addr_lo = ((HC_SubA_HAGPBendL << 24) | (end_addr & 0xFFFFFF));
	command = ((HC_SubA_HAGPCMNT << 24) | (start_addr >> 24) |
			((end_addr & 0xff000000) >> 16));

	*vb++ = command;
	*vb++ = end_addr_lo;
	dev_priv->dma_low += 8;

	vb = via_align_buffer(dev_priv, vb, qw_pad_count-1);


	/* Now at beginning of buffer, make sure engine will pause here. */
	dev_priv->dma_low = 0;
	if (via_cmdbuf_wait(dev_priv, CMDBUF_ALIGNMENT_SIZE) != 0) {
		DRM_ERROR("via_cmdbuf_jump failed\n");
	}
	vb = via_get_dma(dev_priv);

	end_addr = agp_base + dev_priv->dma_high;

	end_addr_lo = ((HC_SubA_HAGPBendL << 24) | (end_addr & 0xFFFFFF));
	command = ((HC_SubA_HAGPCMNT << 24) | (start_addr >> 24) |
			((end_addr & 0xff000000) >> 16));

	qw_pad_count = (CMDBUF_ALIGNMENT_SIZE>>3) -
			((dev_priv->dma_low & CMDBUF_ALIGNMENT_MASK) >> 3);

	pause_addr = agp_base + dev_priv->dma_low - 8 + (qw_pad_count<<3);
	pause_addr_lo = ((HC_SubA_HAGPBpL<<24) | HC_HAGPBpID_PAUSE |
			(pause_addr & 0xffffff));
	pause_addr_hi = ((HC_SubA_HAGPBpH<<24) | (pause_addr >> 24));

	*vb++ = HC_HEADER2 | ((VIA_REG_TRANSET>>2)<<12) |
			(VIA_REG_TRANSPACE>>2);
	*vb++ = (HC_ParaType_PreCR<<16);
	dev_priv->dma_low += 8;

	*vb++ = pause_addr_hi;
	*vb++ = pause_addr_lo;
	dev_priv->dma_low += 8;

	*vb++ = command;
	*vb++ = end_addr_lo;
	dev_priv->dma_low += 8;

	vb = via_align_buffer(dev_priv, vb, qw_pad_count - 4);

	*vb++ = pause_addr_hi;
	*vb++ = pause_addr_lo;
	dev_priv->dma_low += 8;

	*dev_priv->last_pause_ptr = jump_addr_lo;
	dev_priv->last_pause_ptr = vb-1;

	if (VIA_READ(0x41c) & 0x80000000) {
		VIA_WRITE(VIA_REG_TRANSET, (HC_ParaType_PreCR << 16));
		VIA_WRITE(VIA_REG_TRANSPACE, jump_addr_hi);
		VIA_WRITE(VIA_REG_TRANSPACE, jump_addr_lo);
	}
}

static void via_cmdbuf_rewind(drm_via_private_t * dev_priv)
{
	via_cmdbuf_pause(dev_priv);
	via_cmdbuf_jump(dev_priv);
}

static void via_cmdbuf_flush(drm_via_private_t * dev_priv, uint32_t cmd_type)
{
	uint32_t agp_base;
	uint32_t pause_addr, pause_addr_lo, pause_addr_hi;
	uint32_t * vb;
	uint32_t qw_pad_count;

	via_cmdbuf_wait(dev_priv, 0x200);

	vb = via_get_dma(dev_priv);
	*vb++ = HC_HEADER2 | ((VIA_REG_TRANSET>>2)<<12) |
			(VIA_REG_TRANSPACE>>2);
	*vb++ = (HC_ParaType_PreCR<<16);
	dev_priv->dma_low += 8;

	agp_base = dev_priv->dma_offset + (uint32_t) dev_priv->agpAddr;
	qw_pad_count = (CMDBUF_ALIGNMENT_SIZE>>3) -
			((dev_priv->dma_low & CMDBUF_ALIGNMENT_MASK) >> 3);

	pause_addr = agp_base + dev_priv->dma_low - 8 + (qw_pad_count<<3);
	pause_addr_lo = ((HC_SubA_HAGPBpL<<24) | cmd_type |
			(pause_addr & 0xffffff));
	pause_addr_hi = ((HC_SubA_HAGPBpH<<24) | (pause_addr >> 24));

	vb = via_align_buffer(dev_priv, vb, qw_pad_count-1);

	*vb++ = pause_addr_hi;
	*vb++ = pause_addr_lo;
	dev_priv->dma_low += 8;
	*dev_priv->last_pause_ptr = pause_addr_lo;
	dev_priv->last_pause_ptr = vb-1;

	if (VIA_READ(0x41c) & 0x80000000) {
		VIA_WRITE(VIA_REG_TRANSET, (HC_ParaType_PreCR << 16));
		VIA_WRITE(VIA_REG_TRANSPACE, pause_addr_hi);
		VIA_WRITE(VIA_REG_TRANSPACE, pause_addr_lo);
	}
}

static void via_cmdbuf_pause(drm_via_private_t * dev_priv)
{
	via_cmdbuf_flush(dev_priv, HC_HAGPBpID_PAUSE);
}

static void via_cmdbuf_reset(drm_via_private_t * dev_priv)
{
	via_cmdbuf_flush(dev_priv, HC_HAGPBpID_STOP);
	via_wait_idle(dev_priv);
}

/************************************************************************/