drm/ttm: use gem vma_node
Drop vma_node from ttm_buffer_object, use the gem struct (base.vma_node) instead. Signed-off-by: Gerd Hoffmann <kraxel@redhat.com> Reviewed-by: Christian König <christian.koenig@amd.com> Link: http://patchwork.freedesktop.org/patch/msgid/20190805140119.7337-9-kraxel@redhat.com
This commit is contained in:
@@ -672,7 +672,7 @@ static void ttm_bo_release(struct kref *kref)
|
||||
struct ttm_bo_device *bdev = bo->bdev;
|
||||
struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type];
|
||||
|
||||
drm_vma_offset_remove(&bdev->vma_manager, &bo->vma_node);
|
||||
drm_vma_offset_remove(&bdev->vma_manager, &bo->base.vma_node);
|
||||
ttm_mem_io_lock(man, false);
|
||||
ttm_mem_io_free_vm(bo);
|
||||
ttm_mem_io_unlock(man);
|
||||
@@ -1343,9 +1343,9 @@ int ttm_bo_init_reserved(struct ttm_bo_device *bdev,
|
||||
* struct elements we want use regardless.
|
||||
*/
|
||||
reservation_object_init(&bo->base._resv);
|
||||
drm_vma_node_reset(&bo->base.vma_node);
|
||||
}
|
||||
atomic_inc(&bo->bdev->glob->bo_count);
|
||||
drm_vma_node_reset(&bo->vma_node);
|
||||
|
||||
/*
|
||||
* For ttm_bo_type_device buffers, allocate
|
||||
@@ -1353,7 +1353,7 @@ int ttm_bo_init_reserved(struct ttm_bo_device *bdev,
|
||||
*/
|
||||
if (bo->type == ttm_bo_type_device ||
|
||||
bo->type == ttm_bo_type_sg)
|
||||
ret = drm_vma_offset_add(&bdev->vma_manager, &bo->vma_node,
|
||||
ret = drm_vma_offset_add(&bdev->vma_manager, &bo->base.vma_node,
|
||||
bo->mem.num_pages);
|
||||
|
||||
/* passed reservation objects should already be locked,
|
||||
@@ -1781,7 +1781,7 @@ void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo)
|
||||
{
|
||||
struct ttm_bo_device *bdev = bo->bdev;
|
||||
|
||||
drm_vma_node_unmap(&bo->vma_node, bdev->dev_mapping);
|
||||
drm_vma_node_unmap(&bo->base.vma_node, bdev->dev_mapping);
|
||||
ttm_mem_io_free_vm(bo);
|
||||
}
|
||||
|
||||
|
||||
@@ -510,7 +510,7 @@ static int ttm_buffer_object_transfer(struct ttm_buffer_object *bo,
|
||||
INIT_LIST_HEAD(&fbo->base.io_reserve_lru);
|
||||
mutex_init(&fbo->base.wu_mutex);
|
||||
fbo->base.moving = NULL;
|
||||
drm_vma_node_reset(&fbo->base.vma_node);
|
||||
drm_vma_node_reset(&fbo->base.base.vma_node);
|
||||
atomic_set(&fbo->base.cpu_writers, 0);
|
||||
|
||||
kref_init(&fbo->base.list_kref);
|
||||
|
||||
@@ -211,9 +211,9 @@ static vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf)
|
||||
}
|
||||
|
||||
page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) +
|
||||
vma->vm_pgoff - drm_vma_node_start(&bo->vma_node);
|
||||
vma->vm_pgoff - drm_vma_node_start(&bo->base.vma_node);
|
||||
page_last = vma_pages(vma) + vma->vm_pgoff -
|
||||
drm_vma_node_start(&bo->vma_node);
|
||||
drm_vma_node_start(&bo->base.vma_node);
|
||||
|
||||
if (unlikely(page_offset >= bo->num_pages)) {
|
||||
ret = VM_FAULT_SIGBUS;
|
||||
@@ -267,7 +267,7 @@ static vm_fault_t ttm_bo_vm_fault(struct vm_fault *vmf)
|
||||
} else if (unlikely(!page)) {
|
||||
break;
|
||||
}
|
||||
page->index = drm_vma_node_start(&bo->vma_node) +
|
||||
page->index = drm_vma_node_start(&bo->base.vma_node) +
|
||||
page_offset;
|
||||
pfn = page_to_pfn(page);
|
||||
}
|
||||
@@ -413,7 +413,8 @@ static struct ttm_buffer_object *ttm_bo_vm_lookup(struct ttm_bo_device *bdev,
|
||||
|
||||
node = drm_vma_offset_lookup_locked(&bdev->vma_manager, offset, pages);
|
||||
if (likely(node)) {
|
||||
bo = container_of(node, struct ttm_buffer_object, vma_node);
|
||||
bo = container_of(node, struct ttm_buffer_object,
|
||||
base.vma_node);
|
||||
bo = ttm_bo_get_unless_zero(bo);
|
||||
}
|
||||
|
||||
|
||||
Reference in New Issue
Block a user