[RFC PATCH 6/8] drm/virtio: rename virtio_gpu_object_array to virtio_gpu_gem_array
Gurchetan Singh
gurchetansingh at chromium.org
Thu Feb 27 00:25:59 UTC 2020
The plan is to use this array with VRAM objects too.
Signed-off-by: Gurchetan Singh <gurchetansingh at chromium.org>
---
drivers/gpu/drm/virtio/virtgpu_drv.h | 36 ++++----
drivers/gpu/drm/virtio/virtgpu_gem.c | 116 ++++++++++++------------
drivers/gpu/drm/virtio/virtgpu_ioctl.c | 32 +++----
drivers/gpu/drm/virtio/virtgpu_object.c | 20 ++--
drivers/gpu/drm/virtio/virtgpu_plane.c | 22 ++---
drivers/gpu/drm/virtio/virtgpu_vq.c | 62 +++++++------
6 files changed, 145 insertions(+), 143 deletions(-)
diff --git a/drivers/gpu/drm/virtio/virtgpu_drv.h b/drivers/gpu/drm/virtio/virtgpu_drv.h
index 0e99487f2105..a1888a20d311 100644
--- a/drivers/gpu/drm/virtio/virtgpu_drv.h
+++ b/drivers/gpu/drm/virtio/virtgpu_drv.h
@@ -79,7 +79,7 @@ struct virtio_gpu_object {
#define gem_to_virtio_gpu_obj(gobj) \
container_of((gobj), struct virtio_gpu_object, base.base)
-struct virtio_gpu_object_array {
+struct virtio_gpu_gem_array {
struct ww_acquire_ctx ticket;
struct list_head next;
u32 nents, total;
@@ -118,7 +118,7 @@ struct virtio_gpu_vbuffer {
virtio_gpu_resp_cb resp_cb;
void *resp_cb_data;
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
struct list_head list;
};
@@ -244,18 +244,18 @@ int virtio_gpu_mode_dumb_mmap(struct drm_file *file_priv,
struct drm_device *dev,
uint32_t handle, uint64_t *offset_p);
-struct virtio_gpu_object_array *virtio_gpu_array_alloc(u32 nents);
-struct virtio_gpu_object_array*
+struct virtio_gpu_gem_array *virtio_gpu_array_alloc(u32 nents);
+struct virtio_gpu_gem_array*
virtio_gpu_array_from_handles(struct drm_file *drm_file, u32 *handles, u32 nents);
-void virtio_gpu_array_add_obj(struct virtio_gpu_object_array *objs,
+void virtio_gpu_array_add_obj(struct virtio_gpu_gem_array *array,
struct drm_gem_object *obj);
-int virtio_gpu_array_lock_resv(struct virtio_gpu_object_array *objs);
-void virtio_gpu_array_unlock_resv(struct virtio_gpu_object_array *objs);
-void virtio_gpu_array_add_fence(struct virtio_gpu_object_array *objs,
+int virtio_gpu_array_lock_resv(struct virtio_gpu_gem_array *array);
+void virtio_gpu_array_unlock_resv(struct virtio_gpu_gem_array *array);
+void virtio_gpu_array_add_fence(struct virtio_gpu_gem_array *array,
struct dma_fence *fence);
-void virtio_gpu_array_put_free(struct virtio_gpu_object_array *objs);
+void virtio_gpu_array_put_free(struct virtio_gpu_gem_array *array);
void virtio_gpu_array_put_free_delayed(struct virtio_gpu_device *vgdev,
- struct virtio_gpu_object_array *objs);
+ struct virtio_gpu_gem_array *array);
void virtio_gpu_array_put_free_work(struct work_struct *work);
/* virtio vg */
@@ -264,7 +264,7 @@ void virtio_gpu_free_vbufs(struct virtio_gpu_device *vgdev);
void virtio_gpu_cmd_create_resource(struct virtio_gpu_device *vgdev,
struct virtio_gpu_object *bo,
struct virtio_gpu_object_params *params,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence);
void virtio_gpu_cmd_unref_resource(struct virtio_gpu_device *vgdev,
struct drm_gem_object *obj);
@@ -272,7 +272,7 @@ void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev,
uint64_t offset,
uint32_t width, uint32_t height,
uint32_t x, uint32_t y,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence);
void virtio_gpu_cmd_resource_flush(struct virtio_gpu_device *vgdev,
uint32_t resource_id,
@@ -302,32 +302,32 @@ void virtio_gpu_cmd_context_destroy(struct virtio_gpu_device *vgdev,
uint32_t id);
void virtio_gpu_cmd_context_attach_resource(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
- struct virtio_gpu_object_array *objs);
+ struct virtio_gpu_gem_array *array);
void virtio_gpu_cmd_context_detach_resource(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
- struct virtio_gpu_object_array *objs);
+ struct virtio_gpu_gem_array *array);
void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
void *data, uint32_t data_size,
uint32_t ctx_id,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence);
void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
uint64_t offset, uint32_t level,
struct drm_virtgpu_3d_box *box,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence);
void virtio_gpu_cmd_transfer_to_host_3d(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
uint64_t offset, uint32_t level,
struct drm_virtgpu_3d_box *box,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence);
void
virtio_gpu_cmd_resource_create_3d(struct virtio_gpu_device *vgdev,
struct virtio_gpu_object *bo,
struct virtio_gpu_object_params *params,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence);
void virtio_gpu_ctrl_ack(struct virtqueue *vq);
void virtio_gpu_cursor_ack(struct virtqueue *vq);
diff --git a/drivers/gpu/drm/virtio/virtgpu_gem.c b/drivers/gpu/drm/virtio/virtgpu_gem.c
index 0d6152c99a27..53181fe2afe0 100644
--- a/drivers/gpu/drm/virtio/virtgpu_gem.c
+++ b/drivers/gpu/drm/virtio/virtgpu_gem.c
@@ -111,18 +111,18 @@ int virtio_gpu_gem_object_open(struct drm_gem_object *obj,
{
struct virtio_gpu_device *vgdev = obj->dev->dev_private;
struct virtio_gpu_fpriv *vfpriv = file->driver_priv;
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
if (!vgdev->has_virgl_3d)
return 0;
- objs = virtio_gpu_array_alloc(1);
- if (!objs)
+ array = virtio_gpu_array_alloc(1);
+ if (!array)
return -ENOMEM;
- virtio_gpu_array_add_obj(objs, obj);
+ virtio_gpu_array_add_obj(array, obj);
virtio_gpu_cmd_context_attach_resource(vgdev, vfpriv->ctx_id,
- objs);
+ array);
virtio_gpu_notify(vgdev);
return 0;
}
@@ -132,119 +132,119 @@ void virtio_gpu_gem_object_close(struct drm_gem_object *obj,
{
struct virtio_gpu_device *vgdev = obj->dev->dev_private;
struct virtio_gpu_fpriv *vfpriv = file->driver_priv;
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
if (!vgdev->has_virgl_3d)
return;
- objs = virtio_gpu_array_alloc(1);
- if (!objs)
+ array = virtio_gpu_array_alloc(1);
+ if (!array)
return;
- virtio_gpu_array_add_obj(objs, obj);
+ virtio_gpu_array_add_obj(array, obj);
virtio_gpu_cmd_context_detach_resource(vgdev, vfpriv->ctx_id,
- objs);
+ array);
virtio_gpu_notify(vgdev);
}
-struct virtio_gpu_object_array *virtio_gpu_array_alloc(u32 nents)
+struct virtio_gpu_gem_array *virtio_gpu_array_alloc(u32 nents)
{
- struct virtio_gpu_object_array *objs;
- size_t size = sizeof(*objs) + sizeof(objs->objs[0]) * nents;
+ struct virtio_gpu_gem_array *array;
+ size_t size = sizeof(*array) + sizeof(array->objs[0]) * nents;
- objs = kmalloc(size, GFP_KERNEL);
- if (!objs)
+ array = kmalloc(size, GFP_KERNEL);
+ if (!array)
return NULL;
- objs->nents = 0;
- objs->total = nents;
- return objs;
+ array->nents = 0;
+ array->total = nents;
+ return array;
}
-static void virtio_gpu_array_free(struct virtio_gpu_object_array *objs)
+static void virtio_gpu_array_free(struct virtio_gpu_gem_array *array)
{
- kfree(objs);
+ kfree(array);
}
-struct virtio_gpu_object_array*
+struct virtio_gpu_gem_array*
virtio_gpu_array_from_handles(struct drm_file *drm_file, u32 *handles, u32 nents)
{
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
u32 i;
- objs = virtio_gpu_array_alloc(nents);
- if (!objs)
+ array = virtio_gpu_array_alloc(nents);
+ if (!array)
return NULL;
for (i = 0; i < nents; i++) {
- objs->objs[i] = drm_gem_object_lookup(drm_file, handles[i]);
- if (!objs->objs[i]) {
- objs->nents = i;
- virtio_gpu_array_put_free(objs);
+ array->objs[i] = drm_gem_object_lookup(drm_file, handles[i]);
+ if (!array->objs[i]) {
+ array->nents = i;
+ virtio_gpu_array_put_free(array);
return NULL;
}
}
- objs->nents = i;
- return objs;
+ array->nents = i;
+ return array;
}
-void virtio_gpu_array_add_obj(struct virtio_gpu_object_array *objs,
+void virtio_gpu_array_add_obj(struct virtio_gpu_gem_array *array,
struct drm_gem_object *obj)
{
- if (WARN_ON_ONCE(objs->nents == objs->total))
+ if (WARN_ON_ONCE(array->nents == array->total))
return;
drm_gem_object_get(obj);
- objs->objs[objs->nents] = obj;
- objs->nents++;
+ array->objs[array->nents] = obj;
+ array->nents++;
}
-int virtio_gpu_array_lock_resv(struct virtio_gpu_object_array *objs)
+int virtio_gpu_array_lock_resv(struct virtio_gpu_gem_array *array)
{
int ret;
- if (objs->nents == 1) {
- ret = dma_resv_lock_interruptible(objs->objs[0]->resv, NULL);
+ if (array->nents == 1) {
+ ret = dma_resv_lock_interruptible(array->objs[0]->resv, NULL);
} else {
- ret = drm_gem_lock_reservations(objs->objs, objs->nents,
- &objs->ticket);
+ ret = drm_gem_lock_reservations(array->objs, array->nents,
+ &array->ticket);
}
return ret;
}
-void virtio_gpu_array_unlock_resv(struct virtio_gpu_object_array *objs)
+void virtio_gpu_array_unlock_resv(struct virtio_gpu_gem_array *array)
{
- if (objs->nents == 1) {
- dma_resv_unlock(objs->objs[0]->resv);
+ if (array->nents == 1) {
+ dma_resv_unlock(array->objs[0]->resv);
} else {
- drm_gem_unlock_reservations(objs->objs, objs->nents,
- &objs->ticket);
+ drm_gem_unlock_reservations(array->objs, array->nents,
+ &array->ticket);
}
}
-void virtio_gpu_array_add_fence(struct virtio_gpu_object_array *objs,
+void virtio_gpu_array_add_fence(struct virtio_gpu_gem_array *array,
struct dma_fence *fence)
{
int i;
- for (i = 0; i < objs->nents; i++)
- dma_resv_add_excl_fence(objs->objs[i]->resv, fence);
+ for (i = 0; i < array->nents; i++)
+ dma_resv_add_excl_fence(array->objs[i]->resv, fence);
}
-void virtio_gpu_array_put_free(struct virtio_gpu_object_array *objs)
+void virtio_gpu_array_put_free(struct virtio_gpu_gem_array *array)
{
u32 i;
- for (i = 0; i < objs->nents; i++)
- drm_gem_object_put_unlocked(objs->objs[i]);
- virtio_gpu_array_free(objs);
+ for (i = 0; i < array->nents; i++)
+ drm_gem_object_put_unlocked(array->objs[i]);
+ virtio_gpu_array_free(array);
}
void virtio_gpu_array_put_free_delayed(struct virtio_gpu_device *vgdev,
- struct virtio_gpu_object_array *objs)
+ struct virtio_gpu_gem_array *array)
{
spin_lock(&vgdev->obj_free_lock);
- list_add_tail(&objs->next, &vgdev->obj_free_list);
+ list_add_tail(&array->next, &vgdev->obj_free_list);
spin_unlock(&vgdev->obj_free_lock);
schedule_work(&vgdev->obj_free_work);
}
@@ -253,15 +253,15 @@ void virtio_gpu_array_put_free_work(struct work_struct *work)
{
struct virtio_gpu_device *vgdev =
container_of(work, struct virtio_gpu_device, obj_free_work);
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
spin_lock(&vgdev->obj_free_lock);
while (!list_empty(&vgdev->obj_free_list)) {
- objs = list_first_entry(&vgdev->obj_free_list,
- struct virtio_gpu_object_array, next);
- list_del(&objs->next);
+ array = list_first_entry(&vgdev->obj_free_list,
+ struct virtio_gpu_gem_array, next);
+ list_del(&array->next);
spin_unlock(&vgdev->obj_free_lock);
- virtio_gpu_array_put_free(objs);
+ virtio_gpu_array_put_free(array);
spin_lock(&vgdev->obj_free_lock);
}
spin_unlock(&vgdev->obj_free_lock);
diff --git a/drivers/gpu/drm/virtio/virtgpu_ioctl.c b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
index 336cc9143205..9a5bb000ccf2 100644
--- a/drivers/gpu/drm/virtio/virtgpu_ioctl.c
+++ b/drivers/gpu/drm/virtio/virtgpu_ioctl.c
@@ -81,7 +81,7 @@ static int virtio_gpu_execbuffer_ioctl(struct drm_device *dev, void *data,
int ret;
uint32_t *bo_handles = NULL;
void __user *user_bo_handles = NULL;
- struct virtio_gpu_object_array *buflist = NULL;
+ struct virtio_gpu_gem_array *buflist = NULL;
struct sync_file *sync_file;
int in_fence_fd = exbuf->fence_fd;
int out_fence_fd = -1;
@@ -312,7 +312,7 @@ static int virtio_gpu_transfer_from_host_ioctl(struct drm_device *dev,
struct virtio_gpu_device *vgdev = dev->dev_private;
struct virtio_gpu_fpriv *vfpriv = file->driver_priv;
struct drm_virtgpu_3d_transfer_from_host *args = data;
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
struct virtio_gpu_fence *fence;
int ret;
u32 offset = args->offset;
@@ -321,11 +321,11 @@ static int virtio_gpu_transfer_from_host_ioctl(struct drm_device *dev,
return -ENOSYS;
virtio_gpu_create_context(dev, file);
- objs = virtio_gpu_array_from_handles(file, &args->bo_handle, 1);
- if (objs == NULL)
+ array = virtio_gpu_array_from_handles(file, &args->bo_handle, 1);
+ if (array == NULL)
return -ENOENT;
- ret = virtio_gpu_array_lock_resv(objs);
+ ret = virtio_gpu_array_lock_resv(array);
if (ret != 0)
goto err_put_free;
@@ -336,15 +336,15 @@ static int virtio_gpu_transfer_from_host_ioctl(struct drm_device *dev,
}
virtio_gpu_cmd_transfer_from_host_3d
(vgdev, vfpriv->ctx_id, offset, args->level,
- &args->box, objs, fence);
+ &args->box, array, fence);
dma_fence_put(&fence->f);
virtio_gpu_notify(vgdev);
return 0;
err_unlock:
- virtio_gpu_array_unlock_resv(objs);
+ virtio_gpu_array_unlock_resv(array);
err_put_free:
- virtio_gpu_array_put_free(objs);
+ virtio_gpu_array_put_free(array);
return ret;
}
@@ -354,23 +354,23 @@ static int virtio_gpu_transfer_to_host_ioctl(struct drm_device *dev, void *data,
struct virtio_gpu_device *vgdev = dev->dev_private;
struct virtio_gpu_fpriv *vfpriv = file->driver_priv;
struct drm_virtgpu_3d_transfer_to_host *args = data;
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
struct virtio_gpu_fence *fence;
int ret;
u32 offset = args->offset;
- objs = virtio_gpu_array_from_handles(file, &args->bo_handle, 1);
- if (objs == NULL)
+ array = virtio_gpu_array_from_handles(file, &args->bo_handle, 1);
+ if (array == NULL)
return -ENOENT;
if (!vgdev->has_virgl_3d) {
virtio_gpu_cmd_transfer_to_host_2d
(vgdev, offset,
args->box.w, args->box.h, args->box.x, args->box.y,
- objs, NULL);
+ array, NULL);
} else {
virtio_gpu_create_context(dev, file);
- ret = virtio_gpu_array_lock_resv(objs);
+ ret = virtio_gpu_array_lock_resv(array);
if (ret != 0)
goto err_put_free;
@@ -382,16 +382,16 @@ static int virtio_gpu_transfer_to_host_ioctl(struct drm_device *dev, void *data,
virtio_gpu_cmd_transfer_to_host_3d
(vgdev,
vfpriv ? vfpriv->ctx_id : 0, offset,
- args->level, &args->box, objs, fence);
+ args->level, &args->box, array, fence);
dma_fence_put(&fence->f);
}
virtio_gpu_notify(vgdev);
return 0;
err_unlock:
- virtio_gpu_array_unlock_resv(objs);
+ virtio_gpu_array_unlock_resv(array);
err_put_free:
- virtio_gpu_array_put_free(objs);
+ virtio_gpu_array_put_free(array);
return ret;
}
diff --git a/drivers/gpu/drm/virtio/virtgpu_object.c b/drivers/gpu/drm/virtio/virtgpu_object.c
index 84df573e13de..bc8b5a59f364 100644
--- a/drivers/gpu/drm/virtio/virtgpu_object.c
+++ b/drivers/gpu/drm/virtio/virtgpu_object.c
@@ -190,7 +190,7 @@ int virtio_gpu_object_create(struct virtio_gpu_device *vgdev,
struct virtio_gpu_object **bo_ptr,
struct virtio_gpu_fence *fence)
{
- struct virtio_gpu_object_array *objs = NULL;
+ struct virtio_gpu_gem_array *array = NULL;
struct drm_gem_shmem_object *shmem_obj;
struct virtio_gpu_object *bo;
struct virtio_gpu_mem_entry *ents;
@@ -213,22 +213,22 @@ int virtio_gpu_object_create(struct virtio_gpu_device *vgdev,
if (fence) {
ret = -ENOMEM;
- objs = virtio_gpu_array_alloc(1);
- if (!objs)
+ array = virtio_gpu_array_alloc(1);
+ if (!array)
goto err_put_id;
- virtio_gpu_array_add_obj(objs, &bo->base.base);
+ virtio_gpu_array_add_obj(array, &bo->base.base);
- ret = virtio_gpu_array_lock_resv(objs);
+ ret = virtio_gpu_array_lock_resv(array);
if (ret != 0)
- goto err_put_objs;
+ goto err_put_array;
}
if (params->virgl) {
virtio_gpu_cmd_resource_create_3d(vgdev, bo, params,
- objs, fence);
+ array, fence);
} else {
virtio_gpu_cmd_create_resource(vgdev, bo, params,
- objs, fence);
+ array, fence);
}
ret = virtio_gpu_object_shmem_init(vgdev, bo, &ents, &nents);
@@ -247,8 +247,8 @@ int virtio_gpu_object_create(struct virtio_gpu_device *vgdev,
*bo_ptr = bo;
return 0;
-err_put_objs:
- virtio_gpu_array_put_free(objs);
+err_put_array:
+ virtio_gpu_array_put_free(array);
err_put_id:
virtio_gpu_resource_id_put(vgdev, bo->hw_res_handle);
err_free_gem:
diff --git a/drivers/gpu/drm/virtio/virtgpu_plane.c b/drivers/gpu/drm/virtio/virtgpu_plane.c
index 52d24179bcec..fcff5d7a4cee 100644
--- a/drivers/gpu/drm/virtio/virtgpu_plane.c
+++ b/drivers/gpu/drm/virtio/virtgpu_plane.c
@@ -109,7 +109,7 @@ static void virtio_gpu_update_dumb_bo(struct virtio_gpu_device *vgdev,
{
struct virtio_gpu_object *bo =
gem_to_virtio_gpu_obj(state->fb->obj[0]);
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
uint32_t w = rect->x2 - rect->x1;
uint32_t h = rect->y2 - rect->y1;
uint32_t x = rect->x1;
@@ -117,13 +117,13 @@ static void virtio_gpu_update_dumb_bo(struct virtio_gpu_device *vgdev,
uint32_t off = x * state->fb->format->cpp[0] +
y * state->fb->pitches[0];
- objs = virtio_gpu_array_alloc(1);
- if (!objs)
+ array = virtio_gpu_array_alloc(1);
+ if (!array)
return;
- virtio_gpu_array_add_obj(objs, &bo->base.base);
+ virtio_gpu_array_add_obj(array, &bo->base.base);
virtio_gpu_cmd_transfer_to_host_2d(vgdev, off, w, h, x, y,
- objs, NULL);
+ array, NULL);
}
static void virtio_gpu_primary_plane_update(struct drm_plane *plane,
@@ -252,18 +252,18 @@ static void virtio_gpu_cursor_plane_update(struct drm_plane *plane,
if (bo && bo->dumb && (plane->state->fb != old_state->fb)) {
/* new cursor -- update & wait */
- struct virtio_gpu_object_array *objs;
+ struct virtio_gpu_gem_array *array;
- objs = virtio_gpu_array_alloc(1);
- if (!objs)
+ array = virtio_gpu_array_alloc(1);
+ if (!array)
return;
- virtio_gpu_array_add_obj(objs, vgfb->base.obj[0]);
- virtio_gpu_array_lock_resv(objs);
+ virtio_gpu_array_add_obj(array, vgfb->base.obj[0]);
+ virtio_gpu_array_lock_resv(array);
virtio_gpu_cmd_transfer_to_host_2d
(vgdev, 0,
plane->state->crtc_w,
plane->state->crtc_h,
- 0, 0, objs, vgfb->fence);
+ 0, 0, array, vgfb->fence);
virtio_gpu_notify(vgdev);
dma_fence_wait(&vgfb->fence->f, true);
dma_fence_put(&vgfb->fence->f);
diff --git a/drivers/gpu/drm/virtio/virtgpu_vq.c b/drivers/gpu/drm/virtio/virtgpu_vq.c
index 14e64c20eda4..961371566724 100644
--- a/drivers/gpu/drm/virtio/virtgpu_vq.c
+++ b/drivers/gpu/drm/virtio/virtgpu_vq.c
@@ -250,8 +250,8 @@ void virtio_gpu_dequeue_ctrl_func(struct work_struct *work)
virtio_gpu_fence_event_process(vgdev, fence_id);
list_for_each_entry_safe(entry, tmp, &reclaim_list, list) {
- if (entry->objs)
- virtio_gpu_array_put_free_delayed(vgdev, entry->objs);
+ if (entry->array)
+ virtio_gpu_array_put_free_delayed(vgdev, entry->array);
list_del(&entry->list);
free_vbuf(vgdev, entry);
}
@@ -332,8 +332,8 @@ static void virtio_gpu_queue_ctrl_sgs(struct virtio_gpu_device *vgdev,
int ret, idx;
if (!drm_dev_enter(vgdev->ddev, &idx)) {
- if (fence && vbuf->objs)
- virtio_gpu_array_unlock_resv(vbuf->objs);
+ if (fence && vbuf->array)
+ virtio_gpu_array_unlock_resv(vbuf->array);
free_vbuf(vgdev, vbuf);
return;
}
@@ -357,9 +357,9 @@ static void virtio_gpu_queue_ctrl_sgs(struct virtio_gpu_device *vgdev,
if (fence) {
virtio_gpu_fence_emit(vgdev, virtio_gpu_vbuf_ctrl_hdr(vbuf),
fence);
- if (vbuf->objs) {
- virtio_gpu_array_add_fence(vbuf->objs, &fence->f);
- virtio_gpu_array_unlock_resv(vbuf->objs);
+ if (vbuf->array) {
+ virtio_gpu_array_add_fence(vbuf->array, &fence->f);
+ virtio_gpu_array_unlock_resv(vbuf->array);
}
}
@@ -381,6 +381,7 @@ static void virtio_gpu_queue_fenced_ctrl_buffer(struct virtio_gpu_device *vgdev,
{
struct scatterlist *sgs[3], vcmd, vout, vresp;
struct sg_table *sgt = NULL;
+ struct virtio_gpu_gem_array *array = NULL;
int elemcnt = 0, outcnt = 0, incnt = 0;
/* set up vcmd */
@@ -396,8 +397,9 @@ static void virtio_gpu_queue_fenced_ctrl_buffer(struct virtio_gpu_device *vgdev,
sgt = vmalloc_to_sgt(vbuf->data_buf, vbuf->data_size,
&sg_ents);
if (!sgt) {
- if (fence && vbuf->objs)
- virtio_gpu_array_unlock_resv(vbuf->objs);
+ array = vbuf->array;
+ if (fence && array)
+ virtio_gpu_array_unlock_resv(array);
return;
}
@@ -498,7 +500,7 @@ static void virtio_gpu_queue_cursor(struct virtio_gpu_device *vgdev,
void virtio_gpu_cmd_create_resource(struct virtio_gpu_device *vgdev,
struct virtio_gpu_object *bo,
struct virtio_gpu_object_params *params,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence)
{
struct virtio_gpu_resource_create_2d *cmd_p;
@@ -506,7 +508,7 @@ void virtio_gpu_cmd_create_resource(struct virtio_gpu_device *vgdev,
cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
memset(cmd_p, 0, sizeof(*cmd_p));
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_RESOURCE_CREATE_2D);
cmd_p->resource_id = cpu_to_le32(bo->hw_res_handle);
@@ -598,10 +600,10 @@ void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev,
uint64_t offset,
uint32_t width, uint32_t height,
uint32_t x, uint32_t y,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence)
{
- struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(objs->objs[0]);
+ struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(array->objs[0]);
struct virtio_gpu_transfer_to_host_2d *cmd_p;
struct virtio_gpu_vbuffer *vbuf;
bool use_dma_api = !virtio_has_iommu_quirk(vgdev->vdev);
@@ -613,7 +615,7 @@ void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev,
cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
memset(cmd_p, 0, sizeof(*cmd_p));
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_TRANSFER_TO_HOST_2D);
cmd_p->resource_id = cpu_to_le32(bo->hw_res_handle);
@@ -942,15 +944,15 @@ void virtio_gpu_cmd_context_destroy(struct virtio_gpu_device *vgdev,
void virtio_gpu_cmd_context_attach_resource(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
- struct virtio_gpu_object_array *objs)
+ struct virtio_gpu_gem_array *array)
{
- uint32_t handle = virtio_gpu_get_handle(objs->objs[0]);
+ uint32_t handle = virtio_gpu_get_handle(array->objs[0]);
struct virtio_gpu_ctx_resource *cmd_p;
struct virtio_gpu_vbuffer *vbuf;
cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
memset(cmd_p, 0, sizeof(*cmd_p));
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_CTX_ATTACH_RESOURCE);
cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
@@ -960,15 +962,15 @@ void virtio_gpu_cmd_context_attach_resource(struct virtio_gpu_device *vgdev,
void virtio_gpu_cmd_context_detach_resource(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
- struct virtio_gpu_object_array *objs)
+ struct virtio_gpu_gem_array *array)
{
- uint32_t handle = virtio_gpu_get_handle(objs->objs[0]);
+ uint32_t handle = virtio_gpu_get_handle(array->objs[0]);
struct virtio_gpu_ctx_resource *cmd_p;
struct virtio_gpu_vbuffer *vbuf;
cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
memset(cmd_p, 0, sizeof(*cmd_p));
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_CTX_DETACH_RESOURCE);
cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
@@ -980,7 +982,7 @@ void
virtio_gpu_cmd_resource_create_3d(struct virtio_gpu_device *vgdev,
struct virtio_gpu_object *bo,
struct virtio_gpu_object_params *params,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence)
{
struct virtio_gpu_resource_create_3d *cmd_p;
@@ -988,7 +990,7 @@ virtio_gpu_cmd_resource_create_3d(struct virtio_gpu_device *vgdev,
cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
memset(cmd_p, 0, sizeof(*cmd_p));
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_RESOURCE_CREATE_3D);
cmd_p->resource_id = cpu_to_le32(bo->hw_res_handle);
@@ -1013,10 +1015,10 @@ void virtio_gpu_cmd_transfer_to_host_3d(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
uint64_t offset, uint32_t level,
struct drm_virtgpu_3d_box *box,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence)
{
- struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(objs->objs[0]);
+ struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(array->objs[0]);
struct virtio_gpu_transfer_host_3d *cmd_p;
struct virtio_gpu_vbuffer *vbuf;
bool use_dma_api = !virtio_has_iommu_quirk(vgdev->vdev);
@@ -1029,7 +1031,7 @@ void virtio_gpu_cmd_transfer_to_host_3d(struct virtio_gpu_device *vgdev,
cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
memset(cmd_p, 0, sizeof(*cmd_p));
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_TRANSFER_TO_HOST_3D);
cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
@@ -1045,17 +1047,17 @@ void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
uint32_t ctx_id,
uint64_t offset, uint32_t level,
struct drm_virtgpu_3d_box *box,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence)
{
- struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(objs->objs[0]);
+ struct virtio_gpu_object *bo = gem_to_virtio_gpu_obj(array->objs[0]);
struct virtio_gpu_transfer_host_3d *cmd_p;
struct virtio_gpu_vbuffer *vbuf;
cmd_p = virtio_gpu_alloc_cmd(vgdev, &vbuf, sizeof(*cmd_p));
memset(cmd_p, 0, sizeof(*cmd_p));
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_TRANSFER_FROM_HOST_3D);
cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
@@ -1070,7 +1072,7 @@ void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
void *data, uint32_t data_size,
uint32_t ctx_id,
- struct virtio_gpu_object_array *objs,
+ struct virtio_gpu_gem_array *array,
struct virtio_gpu_fence *fence)
{
struct virtio_gpu_cmd_submit *cmd_p;
@@ -1081,7 +1083,7 @@ void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
vbuf->data_buf = data;
vbuf->data_size = data_size;
- vbuf->objs = objs;
+ vbuf->array = array;
cmd_p->hdr.type = cpu_to_le32(VIRTIO_GPU_CMD_SUBMIT_3D);
cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id);
--
2.25.1.481.gfbce0eb801-goog
More information about the dri-devel
mailing list