Merge branch 'master' of git+ssh://git.freedesktop.org/git/mesa/drm into modesetting-101

Conflicts:

	shared-core/i915_dma.c
main
Alan Hourihane 2008-02-26 15:42:41 +00:00
commit 1e66322633
6 changed files with 232 additions and 109 deletions

View File

@ -248,6 +248,7 @@ static void i915_restore_vga(struct drm_device *dev)
dev_priv->saveGR[0x18]); dev_priv->saveGR[0x18]);
/* Attribute controller registers */ /* Attribute controller registers */
inb(st01); /* switch back to index mode */
for (i = 0; i < 20; i++) for (i = 0; i < 20; i++)
i915_write_ar(st01, i, dev_priv->saveAR[i], 0); i915_write_ar(st01, i, dev_priv->saveAR[i], 0);
inb(st01); /* switch back to index mode */ inb(st01); /* switch back to index mode */

View File

@ -769,7 +769,10 @@ struct i915_relocatee_info {
struct drm_i915_validate_buffer { struct drm_i915_validate_buffer {
struct drm_buffer_object *buffer; struct drm_buffer_object *buffer;
struct drm_bo_info_rep rep;
int presumed_offset_correct; int presumed_offset_correct;
void __user *data;
int ret;
}; };
static void i915_dereference_buffers_locked(struct drm_i915_validate_buffer *buffers, static void i915_dereference_buffers_locked(struct drm_i915_validate_buffer *buffers,
@ -790,6 +793,10 @@ int i915_apply_reloc(struct drm_file *file_priv, int num_buffers,
int ret, i; int ret, i;
int buf_index = -1; int buf_index = -1;
/*
* FIXME: O(relocs * buffers) complexity.
*/
for (i = 0; i <= num_buffers; i++) for (i = 0; i <= num_buffers; i++)
if (buffers[i].buffer) if (buffers[i].buffer)
if (reloc[2] == buffers[i].buffer->base.hash.key) if (reloc[2] == buffers[i].buffer->base.hash.key)
@ -811,21 +818,14 @@ int i915_apply_reloc(struct drm_file *file_priv, int num_buffers,
if (!relocatee->data_page || if (!relocatee->data_page ||
!drm_bo_same_page(relocatee->offset, new_cmd_offset)) { !drm_bo_same_page(relocatee->offset, new_cmd_offset)) {
drm_bo_kunmap(&relocatee->kmap); drm_bo_kunmap(&relocatee->kmap);
relocatee->data_page = NULL;
relocatee->offset = new_cmd_offset; relocatee->offset = new_cmd_offset;
mutex_lock (&relocatee->buf->mutex);
ret = drm_bo_wait (relocatee->buf, 0, 0, FALSE);
mutex_unlock (&relocatee->buf->mutex);
if (ret) {
DRM_ERROR("Could not wait for buffer to apply relocs\n %08lx", new_cmd_offset);
return ret;
}
ret = drm_bo_kmap(relocatee->buf, new_cmd_offset >> PAGE_SHIFT, ret = drm_bo_kmap(relocatee->buf, new_cmd_offset >> PAGE_SHIFT,
1, &relocatee->kmap); 1, &relocatee->kmap);
if (ret) { if (ret) {
DRM_ERROR("Could not map command buffer to apply relocs\n %08lx", new_cmd_offset); DRM_ERROR("Could not map command buffer to apply relocs\n %08lx", new_cmd_offset);
return ret; return ret;
} }
relocatee->data_page = drm_bmo_virtual(&relocatee->kmap, relocatee->data_page = drm_bmo_virtual(&relocatee->kmap,
&relocatee->is_iomem); &relocatee->is_iomem);
relocatee->page_offset = (relocatee->offset & PAGE_MASK); relocatee->page_offset = (relocatee->offset & PAGE_MASK);
@ -844,6 +844,10 @@ int i915_apply_reloc(struct drm_file *file_priv, int num_buffers,
reloc[0], reloc[1], buf_index, relocatee->data_page[index], val); reloc[0], reloc[1], buf_index, relocatee->data_page[index], val);
} }
} }
if (relocatee->is_iomem)
iowrite32(val, relocatee->data_page + index);
else
relocatee->data_page[index] = val; relocatee->data_page[index] = val;
return 0; return 0;
} }
@ -912,11 +916,13 @@ int i915_process_relocs(struct drm_file *file_priv,
} }
out: out:
if (reloc_buf) if (reloc_buf)
kfree(reloc_buf); kfree(reloc_buf);
if (relocatee->data_page) {
drm_bo_kunmap(&relocatee->kmap); drm_bo_kunmap(&relocatee->kmap);
relocatee->data_page = NULL; relocatee->data_page = NULL;
}
return ret; return ret;
} }
@ -956,22 +962,60 @@ static int i915_exec_reloc(struct drm_file *file_priv, drm_handle_t buf_handle,
goto out_err; goto out_err;
} }
mutex_lock (&relocatee.buf->mutex);
ret = drm_bo_wait (relocatee.buf, 0, 0, FALSE);
if (ret)
goto out_err1;
while (reloc_user_ptr) { while (reloc_user_ptr) {
ret = i915_process_relocs(file_priv, buf_handle, &reloc_user_ptr, &relocatee, buffers, buf_count); ret = i915_process_relocs(file_priv, buf_handle, &reloc_user_ptr, &relocatee, buffers, buf_count);
if (ret) { if (ret) {
DRM_ERROR("process relocs failed\n"); DRM_ERROR("process relocs failed\n");
break; goto out_err1;
} }
} }
mutex_lock(&dev->struct_mutex); out_err1:
drm_bo_usage_deref_locked(&relocatee.buf); mutex_unlock (&relocatee.buf->mutex);
mutex_unlock(&dev->struct_mutex); drm_bo_usage_deref_unlocked(&relocatee.buf);
out_err: out_err:
return ret; return ret;
} }
static int i915_check_presumed(struct drm_i915_op_arg *arg,
struct drm_buffer_object *bo,
uint32_t __user *data,
int *presumed_ok)
{
struct drm_bo_op_req *req = &arg->d.req;
uint32_t hint_offset;
uint32_t hint = req->bo_req.hint;
*presumed_ok = 0;
if (!(hint & DRM_BO_HINT_PRESUMED_OFFSET))
return 0;
if (bo->offset == req->bo_req.presumed_offset) {
*presumed_ok = 1;
return 0;
}
/*
* We need to turn off the HINT_PRESUMED_OFFSET for this buffer in
* the user-space IOCTL argument list, since the buffer has moved,
* we're about to apply relocations and we might subsequently
* hit an -EAGAIN. In that case the argument list will be reused by
* user-space, but the presumed offset is no longer valid.
*
* Needless to say, this is a bit ugly.
*/
hint_offset = (uint32_t *)&req->bo_req.hint - (uint32_t *)arg;
hint &= ~DRM_BO_HINT_PRESUMED_OFFSET;
return __put_user(hint, data + hint_offset);
}
/* /*
* Validate, add fence and relocate a block of bos from a userspace list * Validate, add fence and relocate a block of bos from a userspace list
*/ */
@ -982,13 +1026,11 @@ int i915_validate_buffer_list(struct drm_file *file_priv,
{ {
struct drm_i915_op_arg arg; struct drm_i915_op_arg arg;
struct drm_bo_op_req *req = &arg.d.req; struct drm_bo_op_req *req = &arg.d.req;
struct drm_bo_arg_rep rep;
unsigned long next = 0;
int ret = 0; int ret = 0;
unsigned buf_count = 0; unsigned buf_count = 0;
struct drm_device *dev = file_priv->minor->dev;
uint32_t buf_handle; uint32_t buf_handle;
uint32_t __user *reloc_user_ptr; uint32_t __user *reloc_user_ptr;
struct drm_i915_validate_buffer *item = buffers;
do { do {
if (buf_count >= *num_buffers) { if (buf_count >= *num_buffers) {
@ -996,31 +1038,26 @@ int i915_validate_buffer_list(struct drm_file *file_priv,
ret = -EINVAL; ret = -EINVAL;
goto out_err; goto out_err;
} }
item = buffers + buf_count;
item->buffer = NULL;
item->presumed_offset_correct = 0;
buffers[buf_count].buffer = NULL; buffers[buf_count].buffer = NULL;
buffers[buf_count].presumed_offset_correct = 0;
if (copy_from_user(&arg, (void __user *)(unsigned long)data, sizeof(arg))) { if (copy_from_user(&arg, (void __user *)(unsigned long)data, sizeof(arg))) {
ret = -EFAULT; ret = -EFAULT;
goto out_err; goto out_err;
} }
if (arg.handled) { ret = 0;
data = arg.next;
mutex_lock(&dev->struct_mutex);
buffers[buf_count].buffer = drm_lookup_buffer_object(file_priv, req->arg_handle, 1);
mutex_unlock(&dev->struct_mutex);
buf_count++;
continue;
}
rep.ret = 0;
if (req->op != drm_bo_validate) { if (req->op != drm_bo_validate) {
DRM_ERROR DRM_ERROR
("Buffer object operation wasn't \"validate\".\n"); ("Buffer object operation wasn't \"validate\".\n");
rep.ret = -EINVAL; ret = -EINVAL;
goto out_err; goto out_err;
} }
item->ret = 0;
item->data = (void __user *) (unsigned long) data;
buf_handle = req->bo_req.handle; buf_handle = req->bo_req.handle;
reloc_user_ptr = (uint32_t *)(unsigned long)arg.reloc_ptr; reloc_user_ptr = (uint32_t *)(unsigned long)arg.reloc_ptr;
@ -1032,48 +1069,146 @@ int i915_validate_buffer_list(struct drm_file *file_priv,
DRM_MEMORYBARRIER(); DRM_MEMORYBARRIER();
} }
rep.ret = drm_bo_handle_validate(file_priv, req->bo_req.handle, ret = drm_bo_handle_validate(file_priv, req->bo_req.handle,
req->bo_req.flags, req->bo_req.mask, req->bo_req.flags, req->bo_req.mask,
req->bo_req.hint, req->bo_req.hint,
req->bo_req.fence_class, 0, req->bo_req.fence_class, 0,
&rep.bo_info, &item->rep,
&buffers[buf_count].buffer); &item->buffer);
if (rep.ret) { if (ret) {
DRM_ERROR("error on handle validate %d\n", rep.ret); DRM_ERROR("error on handle validate %d\n", ret);
goto out_err; goto out_err;
} }
/*
* If the user provided a presumed offset hint, check whether
* the buffer is in the same place, if so, relocations relative to
* this buffer need not be performed
*/
if ((req->bo_req.hint & DRM_BO_HINT_PRESUMED_OFFSET) &&
buffers[buf_count].buffer->offset == req->bo_req.presumed_offset) {
buffers[buf_count].presumed_offset_correct = 1;
}
next = arg.next;
arg.handled = 1;
arg.d.rep = rep;
if (copy_to_user((void __user *)(unsigned long)data, &arg, sizeof(arg)))
return -EFAULT;
data = next;
buf_count++; buf_count++;
} while (next != 0); ret = i915_check_presumed(&arg, item->buffer,
*num_buffers = buf_count; (uint32_t __user *)
return 0; (unsigned long) data,
&item->presumed_offset_correct);
if (ret)
goto out_err;
data = arg.next;
} while (data != 0);
out_err: out_err:
mutex_lock(&dev->struct_mutex); *num_buffers = buf_count;
i915_dereference_buffers_locked(buffers, buf_count); item->ret = (ret != -EAGAIN) ? ret : 0;
mutex_unlock(&dev->struct_mutex); return ret;
*num_buffers = 0;
return (ret) ? ret : rep.ret;
} }
/*
* Remove all buffers from the unfenced list.
* If the execbuffer operation was aborted, for example due to a signal,
* this also make sure that buffers retain their original state and
* fence pointers.
* Copy back buffer information to user-space unless we were interrupted
* by a signal. In which case the IOCTL must be rerun.
*/
static int i915_handle_copyback(struct drm_device *dev,
struct drm_i915_validate_buffer *buffers,
unsigned int num_buffers, int ret)
{
int err = ret;
int i;
struct drm_i915_op_arg arg;
if (ret)
drm_putback_buffer_objects(dev);
if (ret != -EAGAIN) {
for (i = 0; i < num_buffers; ++i) {
arg.handled = 1;
arg.d.rep.ret = buffers->ret;
arg.d.rep.bo_info = buffers->rep;
if (__copy_to_user(buffers->data, &arg, sizeof(arg)))
err = -EFAULT;
buffers++;
}
}
return err;
}
/*
* Create a fence object, and if that fails, pretend that everything is
* OK and just idle the GPU.
*/
void i915_fence_or_sync(struct drm_file *file_priv,
uint32_t fence_flags,
struct drm_fence_arg *fence_arg,
struct drm_fence_object **fence_p)
{
struct drm_device *dev = file_priv->minor->dev;
int ret;
struct drm_fence_object *fence;
ret = drm_fence_buffer_objects(dev, NULL, fence_flags,
NULL, &fence);
if (ret) {
/*
* Fence creation failed.
* Fall back to synchronous operation and idle the engine.
*/
(void) i915_emit_mi_flush(dev, MI_READ_FLUSH);
(void) i915_quiescent(dev);
if (!(fence_flags & DRM_FENCE_FLAG_NO_USER)) {
/*
* Communicate to user-space that
* fence creation has failed and that
* the engine is idle.
*/
fence_arg->handle = ~0;
fence_arg->error = ret;
}
drm_putback_buffer_objects(dev);
if (fence_p)
*fence_p = NULL;
return;
}
if (!(fence_flags & DRM_FENCE_FLAG_NO_USER)) {
ret = drm_fence_add_user_object(file_priv, fence,
fence_flags &
DRM_FENCE_FLAG_SHAREABLE);
if (!ret)
drm_fence_fill_arg(fence, fence_arg);
else {
/*
* Fence user object creation failed.
* We must idle the engine here as well, as user-
* space expects a fence object to wait on. Since we
* have a fence object we wait for it to signal
* to indicate engine "sufficiently" idle.
*/
(void) drm_fence_object_wait(fence, 0, 1,
fence->type);
drm_fence_usage_deref_unlocked(&fence);
fence_arg->handle = ~0;
fence_arg->error = ret;
}
}
if (fence_p)
*fence_p = fence;
else if (fence)
drm_fence_usage_deref_unlocked(&fence);
}
static int i915_execbuffer(struct drm_device *dev, void *data, static int i915_execbuffer(struct drm_device *dev, void *data,
struct drm_file *file_priv) struct drm_file *file_priv)
{ {
@ -1087,7 +1222,6 @@ static int i915_execbuffer(struct drm_device *dev, void *data,
int num_buffers; int num_buffers;
int ret; int ret;
struct drm_i915_validate_buffer *buffers; struct drm_i915_validate_buffer *buffers;
struct drm_fence_object *fence;
if (!dev_priv->allow_batchbuffer) { if (!dev_priv->allow_batchbuffer) {
DRM_ERROR("Batchbuffer ioctl disabled\n"); DRM_ERROR("Batchbuffer ioctl disabled\n");
@ -1132,7 +1266,7 @@ static int i915_execbuffer(struct drm_device *dev, void *data,
ret = i915_validate_buffer_list(file_priv, 0, exec_buf->ops_list, ret = i915_validate_buffer_list(file_priv, 0, exec_buf->ops_list,
buffers, &num_buffers); buffers, &num_buffers);
if (ret) if (ret)
goto out_free; goto out_err0;
/* make sure all previous memory operations have passed */ /* make sure all previous memory operations have passed */
DRM_MEMORYBARRIER(); DRM_MEMORYBARRIER();
@ -1151,30 +1285,16 @@ static int i915_execbuffer(struct drm_device *dev, void *data,
if (sarea_priv) if (sarea_priv)
sarea_priv->last_dispatch = READ_BREADCRUMB(dev_priv); sarea_priv->last_dispatch = READ_BREADCRUMB(dev_priv);
/* fence */ i915_fence_or_sync(file_priv, fence_arg->flags, fence_arg, NULL);
ret = drm_fence_buffer_objects(dev, NULL, fence_arg->flags,
NULL, &fence);
if (ret)
goto out_err0;
if (!(fence_arg->flags & DRM_FENCE_FLAG_NO_USER)) {
ret = drm_fence_add_user_object(file_priv, fence, fence_arg->flags & DRM_FENCE_FLAG_SHAREABLE);
if (!ret) {
fence_arg->handle = fence->base.hash.key;
fence_arg->fence_class = fence->fence_class;
fence_arg->type = fence->type;
fence_arg->signaled = fence->signaled_types;
}
}
drm_fence_usage_deref_unlocked(&fence);
out_err0: out_err0:
/* handle errors */ /* handle errors */
ret = i915_handle_copyback(dev, buffers, num_buffers, ret);
mutex_lock(&dev->struct_mutex); mutex_lock(&dev->struct_mutex);
i915_dereference_buffers_locked(buffers, num_buffers); i915_dereference_buffers_locked(buffers, num_buffers);
mutex_unlock(&dev->struct_mutex); mutex_unlock(&dev->struct_mutex);
out_free:
drm_free(buffers, (exec_buf->num_buffers * sizeof(struct drm_buffer_object *)), DRM_MEM_DRIVER); drm_free(buffers, (exec_buf->num_buffers * sizeof(struct drm_buffer_object *)), DRM_MEM_DRIVER);
mutex_unlock(&dev_priv->cmdbuf_mutex); mutex_unlock(&dev_priv->cmdbuf_mutex);

View File

@ -29,11 +29,6 @@
#include "nouveau_drv.h" #include "nouveau_drv.h"
#include "nouveau_dma.h" #include "nouveau_dma.h"
/* FIXME : should go into a nouveau_drm.h define ?
* (it's shared between DRI & DDX & DRM)
*/
#define SKIPS 8
int int
nouveau_dma_channel_init(struct drm_device *dev) nouveau_dma_channel_init(struct drm_device *dev)
{ {
@ -76,10 +71,10 @@ nouveau_dma_channel_init(struct drm_device *dev)
dchan->cur = dchan->put; dchan->cur = dchan->put;
dchan->free = dchan->max - dchan->cur; dchan->free = dchan->max - dchan->cur;
/* Insert NOPS for SKIPS */ /* Insert NOPS for NOUVEAU_DMA_SKIPS */
dchan->free -= SKIPS; dchan->free -= NOUVEAU_DMA_SKIPS;
dchan->push_free = SKIPS; dchan->push_free = NOUVEAU_DMA_SKIPS;
for (i=0; i<SKIPS; i++) for (i=0; i < NOUVEAU_DMA_SKIPS; i++)
OUT_RING(0); OUT_RING(0);
/* NV_MEMORY_TO_MEMORY_FORMAT requires a notifier */ /* NV_MEMORY_TO_MEMORY_FORMAT requires a notifier */
@ -131,8 +126,6 @@ nouveau_dma_channel_takedown(struct drm_device *dev)
} }
} }
#define RING_SKIPS 8
#define READ_GET() ((NV_READ(dchan->chan->get) - \ #define READ_GET() ((NV_READ(dchan->chan->get) - \
dchan->chan->pushbuf_base) >> 2) dchan->chan->pushbuf_base) >> 2)
#define WRITE_PUT(val) do { \ #define WRITE_PUT(val) do { \
@ -156,19 +149,19 @@ nouveau_dma_wait(struct drm_device *dev, int size)
if (dchan->free < size) { if (dchan->free < size) {
dchan->push_free = 1; dchan->push_free = 1;
OUT_RING(0x20000000|dchan->chan->pushbuf_base); OUT_RING(0x20000000|dchan->chan->pushbuf_base);
if (get <= RING_SKIPS) { if (get <= NOUVEAU_DMA_SKIPS) {
/*corner case - will be idle*/ /*corner case - will be idle*/
if (dchan->put <= RING_SKIPS) if (dchan->put <= NOUVEAU_DMA_SKIPS)
WRITE_PUT(RING_SKIPS + 1); WRITE_PUT(NOUVEAU_DMA_SKIPS + 1);
do { do {
get = READ_GET(); get = READ_GET();
} while (get <= RING_SKIPS); } while (get <= NOUVEAU_DMA_SKIPS);
} }
WRITE_PUT(RING_SKIPS); WRITE_PUT(NOUVEAU_DMA_SKIPS);
dchan->cur = dchan->put = RING_SKIPS; dchan->cur = dchan->put = NOUVEAU_DMA_SKIPS;
dchan->free = get - (RING_SKIPS + 1); dchan->free = get - (NOUVEAU_DMA_SKIPS + 1);
} }
} else { } else {
dchan->free = get - dchan->cur - 1; dchan->free = get - dchan->cur - 1;

View File

@ -70,6 +70,12 @@ struct drm_nouveau_gpuobj_free {
uint32_t handle; uint32_t handle;
}; };
/* This is needed to avoid a race condition.
* Otherwise you may be writing in the fetch area.
* Is this large enough, as it's only 32 bytes, and the maximum fetch size is 256 bytes?
*/
#define NOUVEAU_DMA_SKIPS 8
#define NOUVEAU_MEM_FB 0x00000001 #define NOUVEAU_MEM_FB 0x00000001
#define NOUVEAU_MEM_AGP 0x00000002 #define NOUVEAU_MEM_AGP 0x00000002
#define NOUVEAU_MEM_FB_ACCEPTABLE 0x00000004 #define NOUVEAU_MEM_FB_ACCEPTABLE 0x00000004

View File

@ -45,11 +45,6 @@ int nouveau_fifo_ctx_size(struct drm_device *dev)
* functions doing the actual work * functions doing the actual work
***********************************/ ***********************************/
/* voir nv_xaa.c : NVResetGraphics
* mémoire mappée par nv_driver.c : NVMapMem
* voir nv_driver.c : NVPreInit
*/
static int nouveau_fifo_instmem_configure(struct drm_device *dev) static int nouveau_fifo_instmem_configure(struct drm_device *dev)
{ {
struct drm_nouveau_private *dev_priv = dev->dev_private; struct drm_nouveau_private *dev_priv = dev->dev_private;

View File

@ -1662,7 +1662,7 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
u32 height; u32 height;
int i; int i;
u32 texpitch, microtile; u32 texpitch, microtile;
u32 offset; u32 offset, byte_offset;
RING_LOCALS; RING_LOCALS;
if (radeon_check_and_fixup_offset(dev_priv, file_priv, &tex->offset)) { if (radeon_check_and_fixup_offset(dev_priv, file_priv, &tex->offset)) {
@ -1727,6 +1727,13 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
} else } else
microtile = 0; microtile = 0;
/* this might fail for zero-sized uploads - are those illegal? */
if (!radeon_check_offset(dev_priv, tex->offset + image->height *
blit_width - 1)) {
DRM_ERROR("Invalid final destination offset\n");
return -EINVAL;
}
DRM_DEBUG("tex=%dx%d blit=%d\n", tex_width, tex->height, blit_width); DRM_DEBUG("tex=%dx%d blit=%d\n", tex_width, tex->height, blit_width);
do { do {
@ -1840,6 +1847,7 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
} }
#undef RADEON_COPY_MT #undef RADEON_COPY_MT
byte_offset = (image->y & ~2047) * blit_width;
buf->file_priv = file_priv; buf->file_priv = file_priv;
buf->used = size; buf->used = size;
offset = dev_priv->gart_buffers_offset + buf->offset; offset = dev_priv->gart_buffers_offset + buf->offset;
@ -1854,9 +1862,9 @@ static int radeon_cp_dispatch_texture(struct drm_device * dev,
RADEON_DP_SRC_SOURCE_MEMORY | RADEON_DP_SRC_SOURCE_MEMORY |
RADEON_GMC_CLR_CMP_CNTL_DIS | RADEON_GMC_WR_MSK_DIS); RADEON_GMC_CLR_CMP_CNTL_DIS | RADEON_GMC_WR_MSK_DIS);
OUT_RING((spitch << 22) | (offset >> 10)); OUT_RING((spitch << 22) | (offset >> 10));
OUT_RING((texpitch << 22) | (tex->offset >> 10)); OUT_RING((texpitch << 22) | ((tex->offset >> 10) + (byte_offset >> 10)));
OUT_RING(0); OUT_RING(0);
OUT_RING((image->x << 16) | image->y); OUT_RING((image->x << 16) | (image->y % 2048));
OUT_RING((image->width << 16) | height); OUT_RING((image->width << 16) | height);
RADEON_WAIT_UNTIL_2D_IDLE(); RADEON_WAIT_UNTIL_2D_IDLE();
ADVANCE_RING(); ADVANCE_RING();