Mesa (main): r600: Fix userspace pointer support for evergreen compute.
GitLab Mirror
gitlab-mirror at kemper.freedesktop.org
Thu May 12 08:14:34 UTC 2022
Module: Mesa
Branch: main
Commit: 6cc42bc0e77c7fb920246543fb29fec0f998aa62
URL: http://cgit.freedesktop.org/mesa/mesa/commit/?id=6cc42bc0e77c7fb920246543fb29fec0f998aa62
Author: Jason Volk <jason at zemos.net>
Date: Wed Mar 9 12:34:08 2022 -0800
r600: Fix userspace pointer support for evergreen compute.
Resources returned by r600_buffer_from_user_memory() are not compatible
with the evergreen compute memory pool, though they're added to it anyway.
This results in a segfault reproducible from Clover when the user passes
CL_MEM_USE_HOST_PTR.
This patch allows user_ptr resources to participate in the compute global
memory pool as intended. The result appears to finally allow for zero-copy
DMA out of userspace for anonymous pages.
Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/16114>
---
src/gallium/drivers/r600/compute_memory_pool.c | 44 +++++++++++---------------
src/gallium/drivers/r600/compute_memory_pool.h | 6 ++++
src/gallium/drivers/r600/evergreen_compute.c | 10 ++++--
src/gallium/drivers/r600/r600_buffer_common.c | 17 ++++++++--
4 files changed, 47 insertions(+), 30 deletions(-)
diff --git a/src/gallium/drivers/r600/compute_memory_pool.c b/src/gallium/drivers/r600/compute_memory_pool.c
index 03b0cc524c1..d0f6979dc66 100644
--- a/src/gallium/drivers/r600/compute_memory_pool.c
+++ b/src/gallium/drivers/r600/compute_memory_pool.c
@@ -350,7 +350,7 @@ static int compute_memory_promote_item(struct compute_memory_pool *pool,
* In this case, we need to keep the temporary buffer 'alive'
* because it is possible to keep a map active for reading
* while a kernel (that reads from it) executes */
- if (!(item->status & ITEM_MAPPED_FOR_READING)) {
+ if (!(item->status & ITEM_MAPPED_FOR_READING) && !is_item_user_ptr(item)) {
pool->screen->b.b.resource_destroy(screen, src);
item->real_buffer = NULL;
}
@@ -494,6 +494,22 @@ static void compute_memory_move_item(struct compute_memory_pool *pool,
item->start_in_dw = new_start_in_dw;
}
+/**
+ * Frees one item for compute_memory_free()
+ */
+static void compute_memory_free_item(struct pipe_screen *screen,
+ struct compute_memory_item *item)
+{
+ struct pipe_resource *res = (struct pipe_resource *)item->real_buffer;
+
+ list_del(&item->link);
+
+ if (res && !is_item_user_ptr(item))
+ screen->resource_destroy(screen, res);
+
+ free(item);
+}
+
/**
* Frees the memory associated to the item with id \a id from the pool.
* \param id The id of the item to be freed.
@@ -502,45 +518,23 @@ void compute_memory_free(struct compute_memory_pool* pool, int64_t id)
{
struct compute_memory_item *item, *next;
struct pipe_screen *screen = (struct pipe_screen *)pool->screen;
- struct pipe_resource *res;
COMPUTE_DBG(pool->screen, "* compute_memory_free() id + %"PRIi64" \n", id);
LIST_FOR_EACH_ENTRY_SAFE(item, next, pool->item_list, link) {
-
if (item->id == id) {
-
if (item->link.next != pool->item_list) {
pool->status |= POOL_FRAGMENTED;
}
- list_del(&item->link);
-
- if (item->real_buffer) {
- res = (struct pipe_resource *)item->real_buffer;
- pool->screen->b.b.resource_destroy(
- screen, res);
- }
-
- free(item);
-
+ compute_memory_free_item(screen, item);
return;
}
}
LIST_FOR_EACH_ENTRY_SAFE(item, next, pool->unallocated_list, link) {
-
if (item->id == id) {
- list_del(&item->link);
-
- if (item->real_buffer) {
- res = (struct pipe_resource *)item->real_buffer;
- pool->screen->b.b.resource_destroy(
- screen, res);
- }
-
- free(item);
-
+ compute_memory_free_item(screen, item);
return;
}
}
diff --git a/src/gallium/drivers/r600/compute_memory_pool.h b/src/gallium/drivers/r600/compute_memory_pool.h
index 3b90976276c..7ab1ad73b08 100644
--- a/src/gallium/drivers/r600/compute_memory_pool.h
+++ b/src/gallium/drivers/r600/compute_memory_pool.h
@@ -82,6 +82,12 @@ static inline int is_item_in_pool(struct compute_memory_item *item)
return item->start_in_dw != -1;
}
+static inline int is_item_user_ptr(struct compute_memory_item *item)
+{
+ assert(item->real_buffer);
+ return item->real_buffer->b.is_user_ptr;
+}
+
struct compute_memory_pool* compute_memory_pool_new(struct r600_screen *rscreen);
void compute_memory_pool_delete(struct compute_memory_pool* pool);
diff --git a/src/gallium/drivers/r600/evergreen_compute.c b/src/gallium/drivers/r600/evergreen_compute.c
index f7eb788bfb4..4d10e2b604e 100644
--- a/src/gallium/drivers/r600/evergreen_compute.c
+++ b/src/gallium/drivers/r600/evergreen_compute.c
@@ -1277,6 +1277,9 @@ void *r600_compute_global_transfer_map(struct pipe_context *ctx,
assert(box->y == 0);
assert(box->z == 0);
+ if (buffer->base.b.is_user_ptr)
+ return NULL;
+
///TODO: do it better, mapping is not possible if the pool is too big
return pipe_buffer_map_range(ctx, dst,
offset, box->width, usage, ptransfer);
@@ -1311,9 +1314,12 @@ void r600_compute_global_buffer_destroy(struct pipe_screen *screen,
rscreen = (struct r600_screen*)screen;
compute_memory_free(rscreen->global_pool, buffer->chunk->id);
-
buffer->chunk = NULL;
- free(res);
+
+ if (buffer->base.b.is_user_ptr)
+ r600_buffer_destroy(screen, res);
+ else
+ free(res);
}
struct pipe_resource *r600_compute_global_buffer_create(struct pipe_screen *screen,
diff --git a/src/gallium/drivers/r600/r600_buffer_common.c b/src/gallium/drivers/r600/r600_buffer_common.c
index a47faafa928..d1e48c161cd 100644
--- a/src/gallium/drivers/r600/r600_buffer_common.c
+++ b/src/gallium/drivers/r600/r600_buffer_common.c
@@ -26,6 +26,7 @@
#include "r600_cs.h"
#include "evergreen_compute.h"
+#include "compute_memory_pool.h"
#include "util/u_memory.h"
#include "util/u_upload_mgr.h"
#include <inttypes.h>
@@ -347,7 +348,8 @@ void *r600_buffer_transfer_map(struct pipe_context *ctx,
uint8_t *data;
if (r600_resource(resource)->compute_global_bo) {
- return r600_compute_global_transfer_map(ctx, resource, level, usage, box, ptransfer);
+ if ((data = r600_compute_global_transfer_map(ctx, resource, level, usage, box, ptransfer)))
+ return data;
}
assert(box->x + box->width <= resource->width0);
@@ -524,8 +526,9 @@ void r600_buffer_transfer_unmap(struct pipe_context *ctx,
{
struct r600_common_context *rctx = (struct r600_common_context*)ctx;
struct r600_transfer *rtransfer = (struct r600_transfer*)transfer;
+ struct r600_resource *rtransferr = r600_resource(transfer->resource);
- if (r600_resource(transfer->resource)->compute_global_bo) {
+ if (rtransferr->compute_global_bo && !rtransferr->b.is_user_ptr) {
r600_compute_global_transfer_unmap(ctx, transfer);
return;
}
@@ -636,7 +639,15 @@ r600_buffer_from_user_memory(struct pipe_screen *screen,
{
struct r600_common_screen *rscreen = (struct r600_common_screen*)screen;
struct radeon_winsys *ws = rscreen->ws;
- struct r600_resource *rbuffer = r600_alloc_buffer_struct(screen, templ);
+ struct r600_resource *rbuffer;
+
+ if ((templ->bind & PIPE_BIND_GLOBAL) &&
+ (templ->bind & PIPE_BIND_COMPUTE_RESOURCE)) {
+ rbuffer = r600_resource(r600_compute_global_buffer_create(screen, templ));
+ ((struct r600_resource_global *)rbuffer)->chunk->real_buffer = rbuffer;
+ } else {
+ rbuffer = r600_alloc_buffer_struct(screen, templ);
+ }
rbuffer->domains = RADEON_DOMAIN_GTT;
rbuffer->flags = 0;
More information about the mesa-commit
mailing list