2007-02-07 04:52:23 -07:00
|
|
|
|
/**************************************************************************
|
2007-11-04 19:42:22 -07:00
|
|
|
|
*
|
2007-02-08 16:07:29 -07:00
|
|
|
|
* Copyright (c) 2007 Tungsten Graphics, Inc., Cedar Park, TX., USA
|
2007-02-07 04:52:23 -07:00
|
|
|
|
* All Rights Reserved.
|
2007-11-04 19:42:22 -07:00
|
|
|
|
*
|
2007-02-07 04:52:23 -07:00
|
|
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
|
|
|
* copy of this software and associated documentation files (the
|
|
|
|
|
* "Software"), to deal in the Software without restriction, including
|
|
|
|
|
* without limitation the rights to use, copy, modify, merge, publish,
|
|
|
|
|
* distribute, sub license, and/or sell copies of the Software, and to
|
|
|
|
|
* permit persons to whom the Software is furnished to do so, subject to
|
|
|
|
|
* the following conditions:
|
2007-11-04 19:42:22 -07:00
|
|
|
|
*
|
2007-02-07 04:52:23 -07:00
|
|
|
|
* The above copyright notice and this permission notice (including the
|
|
|
|
|
* next paragraph) shall be included in all copies or substantial portions
|
|
|
|
|
* of the Software.
|
|
|
|
|
*
|
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
|
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
|
|
|
|
|
* THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
|
2007-11-04 19:42:22 -07:00
|
|
|
|
* DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
|
|
|
|
|
* OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
|
2007-02-07 04:52:23 -07:00
|
|
|
|
* USE OR OTHER DEALINGS IN THE SOFTWARE.
|
|
|
|
|
*
|
|
|
|
|
**************************************************************************/
|
|
|
|
|
/*
|
|
|
|
|
* Authors: Thomas Hellstr<EFBFBD>m <thomas-at-tungstengraphics-dot-com>
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
#include "drmP.h"
|
|
|
|
|
|
2007-02-12 09:47:57 -07:00
|
|
|
|
/**
|
|
|
|
|
* Free the old memory node unless it's a pinned region and we
|
|
|
|
|
* have not been requested to free also pinned regions.
|
|
|
|
|
*/
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
static void drm_bo_free_old_node(struct drm_buffer_object *bo)
|
2007-02-12 09:47:57 -07:00
|
|
|
|
{
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_bo_mem_reg *old_mem = &bo->mem;
|
2007-02-12 09:47:57 -07:00
|
|
|
|
|
2007-02-12 12:34:50 -07:00
|
|
|
|
if (old_mem->mm_node && (old_mem->mm_node != bo->pinned_node)) {
|
2007-02-12 09:47:57 -07:00
|
|
|
|
mutex_lock(&bo->dev->struct_mutex);
|
|
|
|
|
drm_mm_put_block(old_mem->mm_node);
|
|
|
|
|
old_mem->mm_node = NULL;
|
|
|
|
|
mutex_unlock(&bo->dev->struct_mutex);
|
|
|
|
|
}
|
|
|
|
|
old_mem->mm_node = NULL;
|
|
|
|
|
}
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
int drm_bo_move_ttm(struct drm_buffer_object *bo,
|
|
|
|
|
int evict, int no_wait, struct drm_bo_mem_reg *new_mem)
|
2007-02-07 04:52:23 -07:00
|
|
|
|
{
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_ttm *ttm = bo->ttm;
|
|
|
|
|
struct drm_bo_mem_reg *old_mem = &bo->mem;
|
2007-10-26 02:25:57 -06:00
|
|
|
|
uint64_t save_flags = old_mem->flags;
|
|
|
|
|
uint64_t save_mask = old_mem->mask;
|
2007-02-07 04:52:23 -07:00
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
|
|
if (old_mem->mem_type == DRM_BO_MEM_TT) {
|
|
|
|
|
if (evict)
|
|
|
|
|
drm_ttm_evict(ttm);
|
|
|
|
|
else
|
|
|
|
|
drm_ttm_unbind(ttm);
|
|
|
|
|
|
2007-02-12 09:47:57 -07:00
|
|
|
|
drm_bo_free_old_node(bo);
|
2007-02-12 12:34:50 -07:00
|
|
|
|
DRM_FLAG_MASKED(old_mem->flags,
|
|
|
|
|
DRM_BO_FLAG_CACHED | DRM_BO_FLAG_MAPPABLE |
|
|
|
|
|
DRM_BO_FLAG_MEM_LOCAL, DRM_BO_MASK_MEMTYPE);
|
2007-02-07 09:25:13 -07:00
|
|
|
|
old_mem->mem_type = DRM_BO_MEM_LOCAL;
|
|
|
|
|
save_flags = old_mem->flags;
|
2007-02-12 12:34:50 -07:00
|
|
|
|
}
|
2007-02-07 09:25:13 -07:00
|
|
|
|
if (new_mem->mem_type != DRM_BO_MEM_LOCAL) {
|
2007-09-22 05:34:33 -06:00
|
|
|
|
ret = drm_bind_ttm(ttm, new_mem);
|
2007-02-07 04:52:23 -07:00
|
|
|
|
if (ret)
|
|
|
|
|
return ret;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
*old_mem = *new_mem;
|
|
|
|
|
new_mem->mm_node = NULL;
|
|
|
|
|
old_mem->mask = save_mask;
|
2007-02-07 09:25:13 -07:00
|
|
|
|
DRM_FLAG_MASKED(save_flags, new_mem->flags, DRM_BO_MASK_MEMTYPE);
|
2007-02-07 04:52:23 -07:00
|
|
|
|
return 0;
|
|
|
|
|
}
|
2007-02-12 12:34:50 -07:00
|
|
|
|
EXPORT_SYMBOL(drm_bo_move_ttm);
|
2007-02-08 05:29:08 -07:00
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
|
* \c Return a kernel virtual address to the buffer object PCI memory.
|
|
|
|
|
*
|
|
|
|
|
* \param bo The buffer object.
|
|
|
|
|
* \return Failure indication.
|
2007-11-04 19:42:22 -07:00
|
|
|
|
*
|
2007-02-08 05:29:08 -07:00
|
|
|
|
* Returns -EINVAL if the buffer object is currently not mappable.
|
|
|
|
|
* Returns -ENOMEM if the ioremap operation failed.
|
|
|
|
|
* Otherwise returns zero.
|
2007-11-04 19:42:22 -07:00
|
|
|
|
*
|
2007-02-08 05:29:08 -07:00
|
|
|
|
* After a successfull call, bo->iomap contains the virtual address, or NULL
|
2007-11-04 19:42:22 -07:00
|
|
|
|
* if the buffer object content is not accessible through PCI space.
|
2007-02-08 05:29:08 -07:00
|
|
|
|
* Call bo->mutex locked.
|
|
|
|
|
*/
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
int drm_mem_reg_ioremap(struct drm_device *dev, struct drm_bo_mem_reg *mem,
|
2007-02-12 12:34:50 -07:00
|
|
|
|
void **virtual)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
{
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_buffer_manager *bm = &dev->bm;
|
|
|
|
|
struct drm_mem_type_manager *man = &bm->man[mem->mem_type];
|
2007-02-08 05:29:08 -07:00
|
|
|
|
unsigned long bus_offset;
|
|
|
|
|
unsigned long bus_size;
|
|
|
|
|
unsigned long bus_base;
|
|
|
|
|
int ret;
|
|
|
|
|
void *addr;
|
|
|
|
|
|
|
|
|
|
*virtual = NULL;
|
|
|
|
|
ret = drm_bo_pci_offset(dev, mem, &bus_base, &bus_offset, &bus_size);
|
2007-02-12 12:34:50 -07:00
|
|
|
|
if (ret || bus_size == 0)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
|
|
if (!(man->flags & _DRM_FLAG_NEEDS_IOREMAP))
|
2007-02-12 12:34:50 -07:00
|
|
|
|
addr = (void *)(((u8 *) man->io_addr) + bus_offset);
|
2007-02-08 05:29:08 -07:00
|
|
|
|
else {
|
|
|
|
|
addr = ioremap_nocache(bus_base + bus_offset, bus_size);
|
|
|
|
|
if (!addr)
|
|
|
|
|
return -ENOMEM;
|
|
|
|
|
}
|
|
|
|
|
*virtual = addr;
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
2007-08-06 06:11:18 -06:00
|
|
|
|
EXPORT_SYMBOL(drm_mem_reg_ioremap);
|
2007-02-08 05:29:08 -07:00
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
|
* \c Unmap mapping obtained using drm_bo_ioremap
|
|
|
|
|
*
|
|
|
|
|
* \param bo The buffer object.
|
|
|
|
|
*
|
|
|
|
|
* Call bo->mutex locked.
|
|
|
|
|
*/
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
void drm_mem_reg_iounmap(struct drm_device *dev, struct drm_bo_mem_reg *mem,
|
2007-02-08 05:29:08 -07:00
|
|
|
|
void *virtual)
|
|
|
|
|
{
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_buffer_manager *bm;
|
|
|
|
|
struct drm_mem_type_manager *man;
|
2007-02-08 05:29:08 -07:00
|
|
|
|
|
|
|
|
|
bm = &dev->bm;
|
|
|
|
|
man = &bm->man[mem->mem_type];
|
2007-02-12 12:34:50 -07:00
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
if (virtual && (man->flags & _DRM_FLAG_NEEDS_IOREMAP))
|
2007-02-08 05:29:08 -07:00
|
|
|
|
iounmap(virtual);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int drm_copy_io_page(void *dst, void *src, unsigned long page)
|
|
|
|
|
{
|
2007-02-12 12:34:50 -07:00
|
|
|
|
uint32_t *dstP =
|
|
|
|
|
(uint32_t *) ((unsigned long)dst + (page << PAGE_SHIFT));
|
|
|
|
|
uint32_t *srcP =
|
|
|
|
|
(uint32_t *) ((unsigned long)src + (page << PAGE_SHIFT));
|
2007-02-08 05:29:08 -07:00
|
|
|
|
|
|
|
|
|
int i;
|
2007-02-12 12:34:50 -07:00
|
|
|
|
for (i = 0; i < PAGE_SIZE / sizeof(uint32_t); ++i)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
iowrite32(ioread32(srcP++), dstP++);
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
static int drm_copy_io_ttm_page(struct drm_ttm *ttm, void *src,
|
|
|
|
|
unsigned long page)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
{
|
|
|
|
|
struct page *d = drm_ttm_get_page(ttm, page);
|
|
|
|
|
void *dst;
|
|
|
|
|
|
|
|
|
|
if (!d)
|
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
2007-02-12 12:34:50 -07:00
|
|
|
|
src = (void *)((unsigned long)src + (page << PAGE_SHIFT));
|
2007-02-08 05:29:08 -07:00
|
|
|
|
dst = kmap(d);
|
|
|
|
|
if (!dst)
|
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
|
|
memcpy_fromio(dst, src, PAGE_SIZE);
|
2007-02-10 04:06:36 -07:00
|
|
|
|
kunmap(d);
|
2007-02-08 05:29:08 -07:00
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
static int drm_copy_ttm_io_page(struct drm_ttm *ttm, void *dst, unsigned long page)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
{
|
|
|
|
|
struct page *s = drm_ttm_get_page(ttm, page);
|
|
|
|
|
void *src;
|
|
|
|
|
|
|
|
|
|
if (!s)
|
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
2007-02-12 12:34:50 -07:00
|
|
|
|
dst = (void *)((unsigned long)dst + (page << PAGE_SHIFT));
|
2007-02-08 05:29:08 -07:00
|
|
|
|
src = kmap(s);
|
|
|
|
|
if (!src)
|
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
|
|
memcpy_toio(dst, src, PAGE_SIZE);
|
2007-02-10 04:06:36 -07:00
|
|
|
|
kunmap(s);
|
2007-02-08 05:29:08 -07:00
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
int drm_bo_move_memcpy(struct drm_buffer_object *bo,
|
|
|
|
|
int evict, int no_wait, struct drm_bo_mem_reg *new_mem)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
{
|
2007-07-15 20:32:51 -06:00
|
|
|
|
struct drm_device *dev = bo->dev;
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_mem_type_manager *man = &dev->bm.man[new_mem->mem_type];
|
|
|
|
|
struct drm_ttm *ttm = bo->ttm;
|
|
|
|
|
struct drm_bo_mem_reg *old_mem = &bo->mem;
|
|
|
|
|
struct drm_bo_mem_reg old_copy = *old_mem;
|
2007-02-08 05:29:08 -07:00
|
|
|
|
void *old_iomap;
|
|
|
|
|
void *new_iomap;
|
|
|
|
|
int ret;
|
2007-10-26 02:25:57 -06:00
|
|
|
|
uint64_t save_flags = old_mem->flags;
|
|
|
|
|
uint64_t save_mask = old_mem->mask;
|
2007-02-08 05:29:08 -07:00
|
|
|
|
unsigned long i;
|
|
|
|
|
unsigned long page;
|
|
|
|
|
unsigned long add = 0;
|
|
|
|
|
int dir;
|
2007-02-12 12:34:50 -07:00
|
|
|
|
|
2007-02-08 05:29:08 -07:00
|
|
|
|
ret = drm_mem_reg_ioremap(dev, old_mem, &old_iomap);
|
|
|
|
|
if (ret)
|
|
|
|
|
return ret;
|
|
|
|
|
ret = drm_mem_reg_ioremap(dev, new_mem, &new_iomap);
|
2007-02-12 12:34:50 -07:00
|
|
|
|
if (ret)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
|
|
if (old_iomap == NULL && new_iomap == NULL)
|
|
|
|
|
goto out2;
|
2007-02-08 11:06:39 -07:00
|
|
|
|
if (old_iomap == NULL && ttm == NULL)
|
|
|
|
|
goto out2;
|
2007-02-12 12:34:50 -07:00
|
|
|
|
|
2007-02-08 05:29:08 -07:00
|
|
|
|
add = 0;
|
|
|
|
|
dir = 1;
|
|
|
|
|
|
2007-02-12 12:34:50 -07:00
|
|
|
|
if ((old_mem->mem_type == new_mem->mem_type) &&
|
|
|
|
|
(new_mem->mm_node->start <
|
|
|
|
|
old_mem->mm_node->start + old_mem->mm_node->size)) {
|
2007-02-08 05:29:08 -07:00
|
|
|
|
dir = -1;
|
|
|
|
|
add = new_mem->num_pages - 1;
|
|
|
|
|
}
|
|
|
|
|
|
2007-02-12 12:34:50 -07:00
|
|
|
|
for (i = 0; i < new_mem->num_pages; ++i) {
|
|
|
|
|
page = i * dir + add;
|
|
|
|
|
if (old_iomap == NULL)
|
2007-02-08 05:29:08 -07:00
|
|
|
|
ret = drm_copy_ttm_io_page(ttm, new_iomap, page);
|
|
|
|
|
else if (new_iomap == NULL)
|
|
|
|
|
ret = drm_copy_io_ttm_page(ttm, old_iomap, page);
|
2007-02-12 12:34:50 -07:00
|
|
|
|
else
|
2007-02-08 05:29:08 -07:00
|
|
|
|
ret = drm_copy_io_page(new_iomap, old_iomap, page);
|
|
|
|
|
if (ret)
|
|
|
|
|
goto out1;
|
|
|
|
|
}
|
2007-02-08 08:21:38 -07:00
|
|
|
|
mb();
|
2007-11-05 02:09:18 -07:00
|
|
|
|
out2:
|
2007-02-12 09:47:57 -07:00
|
|
|
|
drm_bo_free_old_node(bo);
|
|
|
|
|
|
2007-02-08 05:29:08 -07:00
|
|
|
|
*old_mem = *new_mem;
|
|
|
|
|
new_mem->mm_node = NULL;
|
|
|
|
|
old_mem->mask = save_mask;
|
|
|
|
|
DRM_FLAG_MASKED(save_flags, new_mem->flags, DRM_BO_MASK_MEMTYPE);
|
2007-02-08 08:21:38 -07:00
|
|
|
|
|
|
|
|
|
if ((man->flags & _DRM_FLAG_MEMTYPE_FIXED) && (ttm != NULL)) {
|
|
|
|
|
drm_ttm_unbind(ttm);
|
|
|
|
|
drm_destroy_ttm(ttm);
|
|
|
|
|
bo->ttm = NULL;
|
|
|
|
|
}
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
out1:
|
2007-02-08 08:21:38 -07:00
|
|
|
|
drm_mem_reg_iounmap(dev, new_mem, new_iomap);
|
2007-11-05 02:09:18 -07:00
|
|
|
|
out:
|
2007-02-08 08:21:38 -07:00
|
|
|
|
drm_mem_reg_iounmap(dev, &old_copy, old_iomap);
|
2007-02-08 05:29:08 -07:00
|
|
|
|
return ret;
|
|
|
|
|
}
|
2007-02-08 13:28:33 -07:00
|
|
|
|
EXPORT_SYMBOL(drm_bo_move_memcpy);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Transfer a buffer object's memory and LRU status to a newly
|
|
|
|
|
* created object. User-space references remains with the old
|
|
|
|
|
* object. Call bo->mutex locked.
|
|
|
|
|
*/
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
int drm_buffer_object_transfer(struct drm_buffer_object *bo,
|
|
|
|
|
struct drm_buffer_object **new_obj)
|
2007-02-08 13:28:33 -07:00
|
|
|
|
{
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_buffer_object *fbo;
|
2007-07-15 20:32:51 -06:00
|
|
|
|
struct drm_device *dev = bo->dev;
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_buffer_manager *bm = &dev->bm;
|
2007-02-08 13:28:33 -07:00
|
|
|
|
|
2007-02-10 04:06:36 -07:00
|
|
|
|
fbo = drm_ctl_calloc(1, sizeof(*fbo), DRM_MEM_BUFOBJ);
|
2007-02-08 13:28:33 -07:00
|
|
|
|
if (!fbo)
|
|
|
|
|
return -ENOMEM;
|
2007-02-12 12:34:50 -07:00
|
|
|
|
|
2007-02-08 13:28:33 -07:00
|
|
|
|
*fbo = *bo;
|
|
|
|
|
mutex_init(&fbo->mutex);
|
|
|
|
|
mutex_lock(&fbo->mutex);
|
|
|
|
|
mutex_lock(&dev->struct_mutex);
|
|
|
|
|
|
2007-02-10 04:06:36 -07:00
|
|
|
|
DRM_INIT_WAITQUEUE(&bo->event_queue);
|
2007-02-08 13:28:33 -07:00
|
|
|
|
INIT_LIST_HEAD(&fbo->ddestroy);
|
|
|
|
|
INIT_LIST_HEAD(&fbo->lru);
|
2007-02-12 09:47:57 -07:00
|
|
|
|
INIT_LIST_HEAD(&fbo->pinned_lru);
|
2007-02-10 04:06:36 -07:00
|
|
|
|
#ifdef DRM_ODD_MM_COMPAT
|
|
|
|
|
INIT_LIST_HEAD(&fbo->vma_list);
|
|
|
|
|
INIT_LIST_HEAD(&fbo->p_mm_list);
|
|
|
|
|
#endif
|
2007-02-08 13:28:33 -07:00
|
|
|
|
|
2007-11-19 06:37:21 -07:00
|
|
|
|
fbo->fence = drm_fence_reference_locked(bo->fence);
|
2007-02-12 09:47:57 -07:00
|
|
|
|
fbo->pinned_node = NULL;
|
2007-02-08 13:28:33 -07:00
|
|
|
|
fbo->mem.mm_node->private = (void *)fbo;
|
|
|
|
|
atomic_set(&fbo->usage, 1);
|
|
|
|
|
atomic_inc(&bm->count);
|
|
|
|
|
mutex_unlock(&dev->struct_mutex);
|
|
|
|
|
mutex_unlock(&fbo->mutex);
|
|
|
|
|
|
|
|
|
|
*new_obj = fbo;
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
2007-02-10 04:06:36 -07:00
|
|
|
|
/*
|
|
|
|
|
* Since move is underway, we need to block signals in this function.
|
|
|
|
|
* We cannot restart until it has finished.
|
|
|
|
|
*/
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
int drm_bo_move_accel_cleanup(struct drm_buffer_object *bo,
|
|
|
|
|
int evict, int no_wait, uint32_t fence_class,
|
|
|
|
|
uint32_t fence_type, uint32_t fence_flags,
|
|
|
|
|
struct drm_bo_mem_reg *new_mem)
|
2007-02-08 13:28:33 -07:00
|
|
|
|
{
|
2007-07-15 20:32:51 -06:00
|
|
|
|
struct drm_device *dev = bo->dev;
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_mem_type_manager *man = &dev->bm.man[new_mem->mem_type];
|
|
|
|
|
struct drm_bo_mem_reg *old_mem = &bo->mem;
|
2007-02-08 13:28:33 -07:00
|
|
|
|
int ret;
|
2007-10-26 02:25:57 -06:00
|
|
|
|
uint64_t save_flags = old_mem->flags;
|
|
|
|
|
uint64_t save_mask = old_mem->mask;
|
2007-07-15 21:37:02 -06:00
|
|
|
|
struct drm_buffer_object *old_obj;
|
2007-02-12 12:34:50 -07:00
|
|
|
|
|
2007-02-08 13:28:33 -07:00
|
|
|
|
if (bo->fence)
|
2007-06-29 04:50:12 -06:00
|
|
|
|
drm_fence_usage_deref_unlocked(&bo->fence);
|
2007-02-15 04:10:33 -07:00
|
|
|
|
ret = drm_fence_object_create(dev, fence_class, fence_type,
|
2007-02-08 13:28:33 -07:00
|
|
|
|
fence_flags | DRM_FENCE_FLAG_EMIT,
|
|
|
|
|
&bo->fence);
|
2007-09-22 05:34:33 -06:00
|
|
|
|
bo->fence_type = fence_type;
|
2007-02-08 13:28:33 -07:00
|
|
|
|
if (ret)
|
|
|
|
|
return ret;
|
|
|
|
|
|
2007-02-10 04:06:36 -07:00
|
|
|
|
#ifdef DRM_ODD_MM_COMPAT
|
|
|
|
|
/*
|
|
|
|
|
* In this mode, we don't allow pipelining a copy blit,
|
2007-11-04 19:42:22 -07:00
|
|
|
|
* since the buffer will be accessible from user space
|
2007-02-10 04:06:36 -07:00
|
|
|
|
* the moment we return and rebuild the page tables.
|
|
|
|
|
*
|
|
|
|
|
* With normal vm operation, page tables are rebuilt
|
2007-11-04 19:42:22 -07:00
|
|
|
|
* on demand using fault(), which waits for buffer idle.
|
2007-02-10 04:06:36 -07:00
|
|
|
|
*/
|
|
|
|
|
if (1)
|
|
|
|
|
#else
|
2007-02-12 12:34:50 -07:00
|
|
|
|
if (evict || ((bo->mem.mm_node == bo->pinned_node) &&
|
2007-02-12 12:32:03 -07:00
|
|
|
|
bo->mem.mm_node != NULL))
|
2007-02-10 04:06:36 -07:00
|
|
|
|
#endif
|
|
|
|
|
{
|
2007-02-08 13:28:33 -07:00
|
|
|
|
ret = drm_bo_wait(bo, 0, 1, 0);
|
2007-02-12 12:34:50 -07:00
|
|
|
|
if (ret)
|
2007-02-08 13:28:33 -07:00
|
|
|
|
return ret;
|
2007-02-12 09:47:57 -07:00
|
|
|
|
|
|
|
|
|
drm_bo_free_old_node(bo);
|
|
|
|
|
|
2007-02-12 12:34:50 -07:00
|
|
|
|
if ((man->flags & _DRM_FLAG_MEMTYPE_FIXED) && (bo->ttm != NULL)) {
|
2007-02-08 13:28:33 -07:00
|
|
|
|
drm_ttm_unbind(bo->ttm);
|
|
|
|
|
drm_destroy_ttm(bo->ttm);
|
|
|
|
|
bo->ttm = NULL;
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
|
|
|
|
|
/* This should help pipeline ordinary buffer moves.
|
|
|
|
|
*
|
|
|
|
|
* Hang old buffer memory on a new buffer object,
|
2007-02-08 16:02:02 -07:00
|
|
|
|
* and leave it to be released when the GPU
|
2007-02-08 13:28:33 -07:00
|
|
|
|
* operation has completed.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
ret = drm_buffer_object_transfer(bo, &old_obj);
|
2007-02-10 04:06:36 -07:00
|
|
|
|
|
2007-02-08 13:28:33 -07:00
|
|
|
|
if (ret)
|
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
|
|
if (!(man->flags & _DRM_FLAG_MEMTYPE_FIXED))
|
|
|
|
|
old_obj->ttm = NULL;
|
|
|
|
|
else
|
|
|
|
|
bo->ttm = NULL;
|
|
|
|
|
|
|
|
|
|
mutex_lock(&dev->struct_mutex);
|
2007-02-12 09:47:57 -07:00
|
|
|
|
list_del_init(&old_obj->lru);
|
2007-02-10 04:06:36 -07:00
|
|
|
|
DRM_FLAG_MASKED(bo->priv_flags, 0, _DRM_BO_FLAG_UNFENCED);
|
2007-02-12 12:32:03 -07:00
|
|
|
|
drm_bo_add_to_lru(old_obj);
|
2007-02-12 12:34:50 -07:00
|
|
|
|
|
2007-06-29 04:50:12 -06:00
|
|
|
|
drm_bo_usage_deref_locked(&old_obj);
|
2007-02-08 13:28:33 -07:00
|
|
|
|
mutex_unlock(&dev->struct_mutex);
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
*old_mem = *new_mem;
|
|
|
|
|
new_mem->mm_node = NULL;
|
|
|
|
|
old_mem->mask = save_mask;
|
|
|
|
|
DRM_FLAG_MASKED(save_flags, new_mem->flags, DRM_BO_MASK_MEMTYPE);
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
2007-02-12 12:34:50 -07:00
|
|
|
|
EXPORT_SYMBOL(drm_bo_move_accel_cleanup);
|
2007-09-22 05:34:33 -06:00
|
|
|
|
|
|
|
|
|
int drm_bo_same_page(unsigned long offset,
|
|
|
|
|
unsigned long offset2)
|
|
|
|
|
{
|
|
|
|
|
return (offset & PAGE_MASK) == (offset2 & PAGE_MASK);
|
|
|
|
|
}
|
|
|
|
|
EXPORT_SYMBOL(drm_bo_same_page);
|
|
|
|
|
|
|
|
|
|
unsigned long drm_bo_offset_end(unsigned long offset,
|
|
|
|
|
unsigned long end)
|
|
|
|
|
{
|
|
|
|
|
offset = (offset + PAGE_SIZE) & PAGE_MASK;
|
|
|
|
|
return (end < offset) ? end : offset;
|
|
|
|
|
}
|
|
|
|
|
EXPORT_SYMBOL(drm_bo_offset_end);
|
|
|
|
|
|
|
|
|
|
static pgprot_t drm_kernel_io_prot(uint32_t map_type)
|
|
|
|
|
{
|
|
|
|
|
pgprot_t tmp = PAGE_KERNEL;
|
|
|
|
|
|
|
|
|
|
#if defined(__i386__) || defined(__x86_64__)
|
|
|
|
|
#ifdef USE_PAT_WC
|
|
|
|
|
#warning using pat
|
|
|
|
|
if (drm_use_pat() && map_type == _DRM_TTM) {
|
|
|
|
|
pgprot_val(tmp) |= _PAGE_PAT;
|
|
|
|
|
return tmp;
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
if (boot_cpu_data.x86 > 3 && map_type != _DRM_AGP) {
|
|
|
|
|
pgprot_val(tmp) |= _PAGE_PCD;
|
|
|
|
|
pgprot_val(tmp) &= ~_PAGE_PWT;
|
|
|
|
|
}
|
|
|
|
|
#elif defined(__powerpc__)
|
|
|
|
|
pgprot_val(tmp) |= _PAGE_NO_CACHE;
|
|
|
|
|
if (map_type == _DRM_REGISTERS)
|
|
|
|
|
pgprot_val(tmp) |= _PAGE_GUARDED;
|
|
|
|
|
#endif
|
|
|
|
|
#if defined(__ia64__)
|
|
|
|
|
if (map_type == _DRM_TTM)
|
|
|
|
|
tmp = pgprot_writecombine(tmp);
|
|
|
|
|
else
|
|
|
|
|
tmp = pgprot_noncached(tmp);
|
|
|
|
|
#endif
|
|
|
|
|
return tmp;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int drm_bo_ioremap(struct drm_buffer_object *bo, unsigned long bus_base,
|
|
|
|
|
unsigned long bus_offset, unsigned long bus_size,
|
|
|
|
|
struct drm_bo_kmap_obj *map)
|
|
|
|
|
{
|
|
|
|
|
struct drm_device *dev = bo->dev;
|
|
|
|
|
struct drm_bo_mem_reg *mem = &bo->mem;
|
|
|
|
|
struct drm_mem_type_manager *man = &dev->bm.man[mem->mem_type];
|
|
|
|
|
|
|
|
|
|
if (!(man->flags & _DRM_FLAG_NEEDS_IOREMAP)) {
|
|
|
|
|
map->bo_kmap_type = bo_map_premapped;
|
|
|
|
|
map->virtual = (void *)(((u8 *) man->io_addr) + bus_offset);
|
|
|
|
|
} else {
|
|
|
|
|
map->bo_kmap_type = bo_map_iomap;
|
|
|
|
|
map->virtual = ioremap_nocache(bus_base + bus_offset, bus_size);
|
|
|
|
|
}
|
|
|
|
|
return (!map->virtual) ? -ENOMEM : 0;
|
|
|
|
|
}
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
static int drm_bo_kmap_ttm(struct drm_buffer_object *bo,
|
|
|
|
|
unsigned long start_page, unsigned long num_pages,
|
|
|
|
|
struct drm_bo_kmap_obj *map)
|
2007-09-22 05:34:33 -06:00
|
|
|
|
{
|
|
|
|
|
struct drm_device *dev = bo->dev;
|
|
|
|
|
struct drm_bo_mem_reg *mem = &bo->mem;
|
|
|
|
|
struct drm_mem_type_manager *man = &dev->bm.man[mem->mem_type];
|
|
|
|
|
pgprot_t prot;
|
|
|
|
|
struct drm_ttm *ttm = bo->ttm;
|
|
|
|
|
struct page *d;
|
|
|
|
|
int i;
|
|
|
|
|
|
|
|
|
|
BUG_ON(!ttm);
|
|
|
|
|
|
|
|
|
|
if (num_pages == 1 && (mem->flags & DRM_BO_FLAG_CACHED)) {
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* We're mapping a single page, and the desired
|
|
|
|
|
* page protection is consistent with the bo.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
map->bo_kmap_type = bo_map_kmap;
|
|
|
|
|
map->page = drm_ttm_get_page(ttm, start_page);
|
|
|
|
|
map->virtual = kmap(map->page);
|
|
|
|
|
} else {
|
2007-10-04 00:13:22 -06:00
|
|
|
|
/*
|
|
|
|
|
* Populate the part we're mapping;
|
|
|
|
|
*/
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
for (i = start_page; i < start_page + num_pages; ++i) {
|
2007-10-04 00:13:22 -06:00
|
|
|
|
d = drm_ttm_get_page(ttm, i);
|
|
|
|
|
if (!d)
|
|
|
|
|
return -ENOMEM;
|
|
|
|
|
}
|
2007-09-22 05:34:33 -06:00
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* We need to use vmap to get the desired page protection
|
|
|
|
|
* or to make the buffer object look contigous.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
prot = (mem->flags & DRM_BO_FLAG_CACHED) ?
|
|
|
|
|
PAGE_KERNEL :
|
|
|
|
|
drm_kernel_io_prot(man->drm_bus_maptype);
|
|
|
|
|
map->bo_kmap_type = bo_map_vmap;
|
|
|
|
|
map->virtual = vmap(ttm->pages + start_page,
|
|
|
|
|
num_pages, 0, prot);
|
|
|
|
|
}
|
|
|
|
|
return (!map->virtual) ? -ENOMEM : 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* This function is to be used for kernel mapping of buffer objects.
|
|
|
|
|
* It chooses the appropriate mapping method depending on the memory type
|
|
|
|
|
* and caching policy the buffer currently has.
|
|
|
|
|
* Mapping multiple pages or buffers that live in io memory is a bit slow and
|
|
|
|
|
* consumes vmalloc space. Be restrictive with such mappings.
|
2007-11-05 02:09:18 -07:00
|
|
|
|
* Mapping single pages usually returns the logical kernel address,
|
|
|
|
|
* (which is fast)
|
2007-09-22 05:34:33 -06:00
|
|
|
|
* BUG may use slower temporary mappings for high memory pages or
|
|
|
|
|
* uncached / write-combined pages.
|
|
|
|
|
*
|
|
|
|
|
* The function fills in a drm_bo_kmap_obj which can be used to return the
|
|
|
|
|
* kernel virtual address of the buffer.
|
|
|
|
|
*
|
|
|
|
|
* Code servicing a non-priviliged user request is only allowed to map one
|
|
|
|
|
* page at a time. We might need to implement a better scheme to stop such
|
|
|
|
|
* processes from consuming all vmalloc space.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
int drm_bo_kmap(struct drm_buffer_object *bo, unsigned long start_page,
|
|
|
|
|
unsigned long num_pages, struct drm_bo_kmap_obj *map)
|
|
|
|
|
{
|
|
|
|
|
int ret;
|
|
|
|
|
unsigned long bus_base;
|
|
|
|
|
unsigned long bus_offset;
|
|
|
|
|
unsigned long bus_size;
|
|
|
|
|
|
|
|
|
|
map->virtual = NULL;
|
|
|
|
|
|
|
|
|
|
if (num_pages > bo->num_pages)
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
if (start_page > bo->num_pages)
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
#if 0
|
|
|
|
|
if (num_pages > 1 && !DRM_SUSER(DRM_CURPROC))
|
|
|
|
|
return -EPERM;
|
|
|
|
|
#endif
|
|
|
|
|
ret = drm_bo_pci_offset(bo->dev, &bo->mem, &bus_base,
|
|
|
|
|
&bus_offset, &bus_size);
|
|
|
|
|
|
|
|
|
|
if (ret)
|
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
|
|
if (bus_size == 0) {
|
|
|
|
|
return drm_bo_kmap_ttm(bo, start_page, num_pages, map);
|
|
|
|
|
} else {
|
|
|
|
|
bus_offset += start_page << PAGE_SHIFT;
|
|
|
|
|
bus_size = num_pages << PAGE_SHIFT;
|
|
|
|
|
return drm_bo_ioremap(bo, bus_base, bus_offset, bus_size, map);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
EXPORT_SYMBOL(drm_bo_kmap);
|
|
|
|
|
|
|
|
|
|
void drm_bo_kunmap(struct drm_bo_kmap_obj *map)
|
|
|
|
|
{
|
|
|
|
|
if (!map->virtual)
|
|
|
|
|
return;
|
|
|
|
|
|
2007-11-05 02:09:18 -07:00
|
|
|
|
switch (map->bo_kmap_type) {
|
2007-09-22 05:34:33 -06:00
|
|
|
|
case bo_map_iomap:
|
|
|
|
|
iounmap(map->virtual);
|
|
|
|
|
break;
|
|
|
|
|
case bo_map_vmap:
|
|
|
|
|
vunmap(map->virtual);
|
|
|
|
|
break;
|
|
|
|
|
case bo_map_kmap:
|
|
|
|
|
kunmap(map->page);
|
|
|
|
|
break;
|
|
|
|
|
case bo_map_premapped:
|
|
|
|
|
break;
|
|
|
|
|
default:
|
|
|
|
|
BUG();
|
|
|
|
|
}
|
|
|
|
|
map->virtual = NULL;
|
|
|
|
|
map->page = NULL;
|
|
|
|
|
}
|
|
|
|
|
EXPORT_SYMBOL(drm_bo_kunmap);
|