amdgpu: clean up non list code path for vamgr v2
v2: Add missing "goto out" Signed-off-by: Chunming Zhou <david1.zhou@amd.com> Reviewed-by: Christian König <christian.koenig@amd.com> Tested-by: Michel Dänzer <michel.daenzer@amd.com>main
parent
fa35b51f63
commit
69f9faeee6
|
@ -53,8 +53,6 @@ struct amdgpu_bo_va_hole {
|
||||||
};
|
};
|
||||||
|
|
||||||
struct amdgpu_bo_va_mgr {
|
struct amdgpu_bo_va_mgr {
|
||||||
/* the start virtual address */
|
|
||||||
uint64_t va_offset;
|
|
||||||
uint64_t va_max;
|
uint64_t va_max;
|
||||||
struct list_head va_holes;
|
struct list_head va_holes;
|
||||||
pthread_mutex_t bo_va_mutex;
|
pthread_mutex_t bo_va_mutex;
|
||||||
|
|
|
@ -48,12 +48,19 @@ int amdgpu_va_range_query(amdgpu_device_handle dev,
|
||||||
drm_private void amdgpu_vamgr_init(struct amdgpu_bo_va_mgr *mgr, uint64_t start,
|
drm_private void amdgpu_vamgr_init(struct amdgpu_bo_va_mgr *mgr, uint64_t start,
|
||||||
uint64_t max, uint64_t alignment)
|
uint64_t max, uint64_t alignment)
|
||||||
{
|
{
|
||||||
mgr->va_offset = start;
|
struct amdgpu_bo_va_hole *n;
|
||||||
|
|
||||||
mgr->va_max = max;
|
mgr->va_max = max;
|
||||||
mgr->va_alignment = alignment;
|
mgr->va_alignment = alignment;
|
||||||
|
|
||||||
list_inithead(&mgr->va_holes);
|
list_inithead(&mgr->va_holes);
|
||||||
pthread_mutex_init(&mgr->bo_va_mutex, NULL);
|
pthread_mutex_init(&mgr->bo_va_mutex, NULL);
|
||||||
|
pthread_mutex_lock(&mgr->bo_va_mutex);
|
||||||
|
n = calloc(1, sizeof(struct amdgpu_bo_va_hole));
|
||||||
|
n->size = mgr->va_max;
|
||||||
|
n->offset = start;
|
||||||
|
list_add(&n->list, &mgr->va_holes);
|
||||||
|
pthread_mutex_unlock(&mgr->bo_va_mutex);
|
||||||
}
|
}
|
||||||
|
|
||||||
drm_private void amdgpu_vamgr_deinit(struct amdgpu_bo_va_mgr *mgr)
|
drm_private void amdgpu_vamgr_deinit(struct amdgpu_bo_va_mgr *mgr)
|
||||||
|
@ -122,41 +129,14 @@ amdgpu_vamgr_find_va(struct amdgpu_bo_va_mgr *mgr, uint64_t size,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (base_required) {
|
|
||||||
if (base_required < mgr->va_offset) {
|
|
||||||
pthread_mutex_unlock(&mgr->bo_va_mutex);
|
|
||||||
return AMDGPU_INVALID_VA_ADDRESS;
|
|
||||||
}
|
|
||||||
offset = mgr->va_offset;
|
|
||||||
waste = base_required - mgr->va_offset;
|
|
||||||
} else {
|
|
||||||
offset = mgr->va_offset;
|
|
||||||
waste = offset % alignment;
|
|
||||||
waste = waste ? alignment - waste : 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (offset + waste + size > mgr->va_max) {
|
|
||||||
pthread_mutex_unlock(&mgr->bo_va_mutex);
|
|
||||||
return AMDGPU_INVALID_VA_ADDRESS;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (waste) {
|
|
||||||
n = calloc(1, sizeof(struct amdgpu_bo_va_hole));
|
|
||||||
n->size = waste;
|
|
||||||
n->offset = offset;
|
|
||||||
list_add(&n->list, &mgr->va_holes);
|
|
||||||
}
|
|
||||||
|
|
||||||
offset += waste;
|
|
||||||
mgr->va_offset += size + waste;
|
|
||||||
pthread_mutex_unlock(&mgr->bo_va_mutex);
|
pthread_mutex_unlock(&mgr->bo_va_mutex);
|
||||||
return offset;
|
return AMDGPU_INVALID_VA_ADDRESS;
|
||||||
}
|
}
|
||||||
|
|
||||||
static drm_private void
|
static drm_private void
|
||||||
amdgpu_vamgr_free_va(struct amdgpu_bo_va_mgr *mgr, uint64_t va, uint64_t size)
|
amdgpu_vamgr_free_va(struct amdgpu_bo_va_mgr *mgr, uint64_t va, uint64_t size)
|
||||||
{
|
{
|
||||||
struct amdgpu_bo_va_hole *hole;
|
struct amdgpu_bo_va_hole *hole, *next;
|
||||||
|
|
||||||
if (va == AMDGPU_INVALID_VA_ADDRESS)
|
if (va == AMDGPU_INVALID_VA_ADDRESS)
|
||||||
return;
|
return;
|
||||||
|
@ -164,61 +144,47 @@ amdgpu_vamgr_free_va(struct amdgpu_bo_va_mgr *mgr, uint64_t va, uint64_t size)
|
||||||
size = ALIGN(size, mgr->va_alignment);
|
size = ALIGN(size, mgr->va_alignment);
|
||||||
|
|
||||||
pthread_mutex_lock(&mgr->bo_va_mutex);
|
pthread_mutex_lock(&mgr->bo_va_mutex);
|
||||||
if ((va + size) == mgr->va_offset) {
|
hole = container_of(&mgr->va_holes, hole, list);
|
||||||
mgr->va_offset = va;
|
LIST_FOR_EACH_ENTRY(next, &mgr->va_holes, list) {
|
||||||
/* Delete uppermost hole if it reaches the new top */
|
if (next->offset < va)
|
||||||
if (!LIST_IS_EMPTY(&mgr->va_holes)) {
|
break;
|
||||||
hole = container_of(mgr->va_holes.next, hole, list);
|
hole = next;
|
||||||
if ((hole->offset + hole->size) == va) {
|
}
|
||||||
mgr->va_offset = hole->offset;
|
|
||||||
|
if (&hole->list != &mgr->va_holes) {
|
||||||
|
/* Grow upper hole if it's adjacent */
|
||||||
|
if (hole->offset == (va + size)) {
|
||||||
|
hole->offset = va;
|
||||||
|
hole->size += size;
|
||||||
|
/* Merge lower hole if it's adjacent */
|
||||||
|
if (next != hole &&
|
||||||
|
&next->list != &mgr->va_holes &&
|
||||||
|
(next->offset + next->size) == va) {
|
||||||
|
next->size += hole->size;
|
||||||
list_del(&hole->list);
|
list_del(&hole->list);
|
||||||
free(hole);
|
free(hole);
|
||||||
}
|
}
|
||||||
}
|
|
||||||
} else {
|
|
||||||
struct amdgpu_bo_va_hole *next;
|
|
||||||
|
|
||||||
hole = container_of(&mgr->va_holes, hole, list);
|
|
||||||
LIST_FOR_EACH_ENTRY(next, &mgr->va_holes, list) {
|
|
||||||
if (next->offset < va)
|
|
||||||
break;
|
|
||||||
hole = next;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (&hole->list != &mgr->va_holes) {
|
|
||||||
/* Grow upper hole if it's adjacent */
|
|
||||||
if (hole->offset == (va + size)) {
|
|
||||||
hole->offset = va;
|
|
||||||
hole->size += size;
|
|
||||||
/* Merge lower hole if it's adjacent */
|
|
||||||
if (next != hole &&
|
|
||||||
&next->list != &mgr->va_holes &&
|
|
||||||
(next->offset + next->size) == va) {
|
|
||||||
next->size += hole->size;
|
|
||||||
list_del(&hole->list);
|
|
||||||
free(hole);
|
|
||||||
}
|
|
||||||
goto out;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Grow lower hole if it's adjacent */
|
|
||||||
if (next != hole && &next->list != &mgr->va_holes &&
|
|
||||||
(next->offset + next->size) == va) {
|
|
||||||
next->size += size;
|
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* FIXME on allocation failure we just lose virtual address space
|
|
||||||
* maybe print a warning
|
|
||||||
*/
|
|
||||||
next = calloc(1, sizeof(struct amdgpu_bo_va_hole));
|
|
||||||
if (next) {
|
|
||||||
next->size = size;
|
|
||||||
next->offset = va;
|
|
||||||
list_add(&next->list, &hole->list);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Grow lower hole if it's adjacent */
|
||||||
|
if (next != hole && &next->list != &mgr->va_holes &&
|
||||||
|
(next->offset + next->size) == va) {
|
||||||
|
next->size += size;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* FIXME on allocation failure we just lose virtual address space
|
||||||
|
* maybe print a warning
|
||||||
|
*/
|
||||||
|
next = calloc(1, sizeof(struct amdgpu_bo_va_hole));
|
||||||
|
if (next) {
|
||||||
|
next->size = size;
|
||||||
|
next->offset = va;
|
||||||
|
list_add(&next->list, &hole->list);
|
||||||
|
}
|
||||||
|
|
||||||
out:
|
out:
|
||||||
pthread_mutex_unlock(&mgr->bo_va_mutex);
|
pthread_mutex_unlock(&mgr->bo_va_mutex);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue