[Mesa-dev] [PATCH v3 1/2] r600g/compute: Add documentation to compute_memory_pool

Tom Stellard tom at stellard.net
Mon Jul 28 08:22:00 PDT 2014


On Sun, Jul 27, 2014 at 01:56:15PM +0200, Bruno Jiménez wrote:
> v2: Rebased on top of master

I've pushed both of these patches, thanks!

-Tom

> ---
>  src/gallium/drivers/r600/compute_memory_pool.c | 59 +++++++++++++++++++++-----
>  src/gallium/drivers/r600/compute_memory_pool.h | 58 ++++++++++++++++---------
>  2 files changed, 86 insertions(+), 31 deletions(-)
> 
> diff --git a/src/gallium/drivers/r600/compute_memory_pool.c b/src/gallium/drivers/r600/compute_memory_pool.c
> index d53587f..928618c 100644
> --- a/src/gallium/drivers/r600/compute_memory_pool.c
> +++ b/src/gallium/drivers/r600/compute_memory_pool.c
> @@ -44,7 +44,7 @@
>  
>  #define ITEM_ALIGNMENT 1024
>  /**
> - * Creates a new pool
> + * Creates a new pool.
>   */
>  struct compute_memory_pool* compute_memory_pool_new(
>  	struct r600_screen * rscreen)
> @@ -66,6 +66,12 @@ struct compute_memory_pool* compute_memory_pool_new(
>  	return pool;
>  }
>  
> +/**
> + * Initializes the pool with a size of \a initial_size_in_dw.
> + * \param pool			The pool to be initialized.
> + * \param initial_size_in_dw	The initial size.
> + * \see compute_memory_grow_defrag_pool
> + */
>  static void compute_memory_pool_init(struct compute_memory_pool * pool,
>  	unsigned initial_size_in_dw)
>  {
> @@ -79,7 +85,7 @@ static void compute_memory_pool_init(struct compute_memory_pool * pool,
>  }
>  
>  /**
> - * Frees all stuff in the pool and the pool struct itself too
> + * Frees all stuff in the pool and the pool struct itself too.
>   */
>  void compute_memory_pool_delete(struct compute_memory_pool* pool)
>  {
> @@ -94,7 +100,9 @@ void compute_memory_pool_delete(struct compute_memory_pool* pool)
>  
>  /**
>   * Searches for an empty space in the pool, return with the pointer to the
> - * allocatable space in the pool, returns -1 on failure.
> + * allocatable space in the pool.
> + * \param size_in_dw	The size of the space we are looking for.
> + * \return -1 on failure
>   */
>  int64_t compute_memory_prealloc_chunk(
>  	struct compute_memory_pool* pool,
> @@ -126,6 +134,8 @@ int64_t compute_memory_prealloc_chunk(
>  
>  /**
>   *  Search for the chunk where we can link our new chunk after it.
> + *  \param start_in_dw	The position of the item we want to add to the pool.
> + *  \return The item that is just before the passed position
>   */
>  struct list_head *compute_memory_postalloc_chunk(
>  	struct compute_memory_pool* pool,
> @@ -166,8 +176,9 @@ struct list_head *compute_memory_postalloc_chunk(
>  }
>  
>  /**
> - * Reallocates pool, conserves data.
> - * @returns -1 if it fails, 0 otherwise
> + * Reallocates and defragments the pool, conserves data.
> + * \returns -1 if it fails, 0 otherwise
> + * \see compute_memory_finalize_pending
>   */
>  int compute_memory_grow_defrag_pool(struct compute_memory_pool *pool,
>  	struct pipe_context *pipe, int new_size_in_dw)
> @@ -234,6 +245,8 @@ int compute_memory_grow_defrag_pool(struct compute_memory_pool *pool,
>  
>  /**
>   * Copy pool from device to host, or host to device.
> + * \param device_to_host 1 for device->host, 0 for host->device
> + * \see compute_memory_grow_defrag_pool
>   */
>  void compute_memory_shadow(struct compute_memory_pool* pool,
>  	struct pipe_context * pipe, int device_to_host)
> @@ -251,8 +264,10 @@ void compute_memory_shadow(struct compute_memory_pool* pool,
>  }
>  
>  /**
> - * Allocates pending allocations in the pool
> - * @returns -1 if it fails, 0 otherwise
> + * Moves all the items marked for promotion from the \a unallocated_list
> + * to the \a item_list.
> + * \return -1 if it fails, 0 otherwise
> + * \see evergreen_set_global_binding
>   */
>  int compute_memory_finalize_pending(struct compute_memory_pool* pool,
>  	struct pipe_context * pipe)
> @@ -323,6 +338,9 @@ int compute_memory_finalize_pending(struct compute_memory_pool* pool,
>  /**
>   * Defragments the pool, so that there's no gap between items.
>   * \param pool	The pool to be defragmented
> + * \param src	The origin resource
> + * \param dst	The destination resource
> + * \see compute_memory_grow_defrag_pool and compute_memory_finalize_pending
>   */
>  void compute_memory_defrag(struct compute_memory_pool *pool,
>  	struct pipe_resource *src, struct pipe_resource *dst,
> @@ -348,6 +366,12 @@ void compute_memory_defrag(struct compute_memory_pool *pool,
>  	pool->status &= ~POOL_FRAGMENTED;
>  }
>  
> +/**
> + * Moves an item from the \a unallocated_list to the \a item_list.
> + * \param item	The item that will be promoted.
> + * \return -1 if it fails, 0 otherwise
> + * \see compute_memory_finalize_pending
> + */
>  int compute_memory_promote_item(struct compute_memory_pool *pool,
>  		struct compute_memory_item *item, struct pipe_context *pipe,
>  		int64_t start_in_dw)
> @@ -390,6 +414,11 @@ int compute_memory_promote_item(struct compute_memory_pool *pool,
>  	return 0;
>  }
>  
> +/**
> + * Moves an item from the \a item_list to the \a unallocated_list.
> + * \param item	The item that will be demoted
> + * \see r600_compute_global_transfer_map
> + */
>  void compute_memory_demote_item(struct compute_memory_pool *pool,
>  	struct compute_memory_item *item, struct pipe_context *pipe)
>  {
> @@ -434,7 +463,7 @@ void compute_memory_demote_item(struct compute_memory_pool *pool,
>   * resource \a dst at \a new_start_in_dw
>   *
>   * This function assumes two things:
> - * 1) The item is \b only moved forward
> + * 1) The item is \b only moved forward, unless src is different from dst
>   * 2) The item \b won't change it's position inside the \a item_list
>   *
>   * \param item			The item that will be moved
> @@ -516,6 +545,10 @@ void compute_memory_move_item(struct compute_memory_pool *pool,
>  	item->start_in_dw = new_start_in_dw;
>  }
>  
> +/**
> + * Frees the memory asociated to the item with id \a id from the pool.
> + * \param id	The id of the item to be freed.
> + */
>  void compute_memory_free(struct compute_memory_pool* pool, int64_t id)
>  {
>  	struct compute_memory_item *item, *next;
> @@ -570,7 +603,11 @@ void compute_memory_free(struct compute_memory_pool* pool, int64_t id)
>  }
>  
>  /**
> - * Creates pending allocations
> + * Creates pending allocations for new items, these items are
> + * placed in the unallocated_list.
> + * \param size_in_dw	The size, in double words, of the new item.
> + * \return The new item
> + * \see r600_compute_global_buffer_create
>   */
>  struct compute_memory_item* compute_memory_alloc(
>  	struct compute_memory_pool* pool,
> @@ -601,7 +638,9 @@ struct compute_memory_item* compute_memory_alloc(
>  }
>  
>  /**
> - * Transfer data host<->device, offset and size is in bytes
> + * Transfer data host<->device, offset and size is in bytes.
> + * \param device_to_host 1 for device->host, 0 for host->device.
> + * \see compute_memory_shadow
>   */
>  void compute_memory_transfer(
>  	struct compute_memory_pool* pool,
> diff --git a/src/gallium/drivers/r600/compute_memory_pool.h b/src/gallium/drivers/r600/compute_memory_pool.h
> index c7eb237..161ddd5 100644
> --- a/src/gallium/drivers/r600/compute_memory_pool.h
> +++ b/src/gallium/drivers/r600/compute_memory_pool.h
> @@ -38,13 +38,17 @@ struct compute_memory_pool;
>  
>  struct compute_memory_item
>  {
> -	int64_t id; ///ID of the memory chunk
> +	int64_t id;		/**< ID of the memory chunk */
>  
> -	uint32_t status; ///Will track the status of the item
> +	uint32_t status;	/**< Will track the status of the item */
>  
> -	int64_t start_in_dw; ///Start pointer in dwords relative in the pool bo
> -	int64_t size_in_dw; ///Size of the chunk in dwords
> +	/** Start pointer in dwords relative in the pool bo. If an item
> +	 * is unallocated, then this value must be -1 to indicate this. */
> +	int64_t start_in_dw;
> +	int64_t size_in_dw;	/**< Size of the chunk in dwords */
>  
> +	/** Intermediate buffer asociated with an item. It is used mainly for mapping
> +	 * items against it. They are listed in the pool's unallocated list */
>  	struct r600_resource *real_buffer;
>  
>  	struct compute_memory_pool* pool;
> @@ -54,18 +58,22 @@ struct compute_memory_item
>  
>  struct compute_memory_pool
>  {
> -	int64_t next_id; ///For generating unique IDs for memory chunks
> -	int64_t size_in_dw; ///Size of the pool in dwords
> +	int64_t next_id;	/**< For generating unique IDs for memory chunks */
> +	int64_t size_in_dw;	/**< Size of the pool in dwords */
>  
> -	struct r600_resource *bo; ///The pool buffer object resource
> +	struct r600_resource *bo;	/**< The pool buffer object resource */
>  	struct r600_screen *screen;
>  
> -	uint32_t *shadow; ///host copy of the pool, used for defragmentation
> +	uint32_t *shadow;	/**< host copy of the pool, used for growing the pool */
>  
>  	uint32_t status;	/**< Status of the pool */
>  
> -	struct list_head *item_list; ///Allocated memory chunks in the buffer,they must be ordered by "start_in_dw"
> -	struct list_head *unallocated_list; ///Unallocated memory chunks
> +	/** Allocated memory items in the pool, they must be ordered by "start_in_dw" */
> +	struct list_head *item_list;
> +
> +	/** Unallocated memory items, this list contains all the items that aren't
> +	 * yet in the pool */
> +	struct list_head *unallocated_list;
>  };
>  
>  
> @@ -74,18 +82,21 @@ static inline int is_item_in_pool(struct compute_memory_item *item)
>  	return item->start_in_dw != -1;
>  }
>  
> -struct compute_memory_pool* compute_memory_pool_new(struct r600_screen *rscreen); ///Creates a new pool
> -void compute_memory_pool_delete(struct compute_memory_pool* pool); ///Frees all stuff in the pool and the pool struct itself too
> +struct compute_memory_pool* compute_memory_pool_new(struct r600_screen *rscreen);
> +
> +void compute_memory_pool_delete(struct compute_memory_pool* pool);
>  
> -int64_t compute_memory_prealloc_chunk(struct compute_memory_pool* pool, int64_t size_in_dw); ///searches for an empty space in the pool, return with the pointer to the allocatable space in the pool, returns -1 on failure
> +int64_t compute_memory_prealloc_chunk(struct compute_memory_pool* pool,
> +	int64_t size_in_dw);
>  
> -struct list_head *compute_memory_postalloc_chunk(struct compute_memory_pool* pool, int64_t start_in_dw); ///search for the chunk where we can link our new chunk after it
> +struct list_head *compute_memory_postalloc_chunk(struct compute_memory_pool* pool,
> +	int64_t start_in_dw);
>  
> -int compute_memory_grow_defrag_pool(struct compute_memory_pool* pool, struct pipe_context * pipe,
> -	int new_size_in_dw);
> +int compute_memory_grow_defrag_pool(struct compute_memory_pool* pool,
> +	struct pipe_context *pipe, int new_size_in_dw);
>  
>  void compute_memory_shadow(struct compute_memory_pool* pool,
> -	struct pipe_context * pipe, int device_to_host);
> +	struct pipe_context *pipe, int device_to_host);
>  
>  int compute_memory_finalize_pending(struct compute_memory_pool* pool,
>  	struct pipe_context * pipe);
> @@ -95,8 +106,8 @@ void compute_memory_defrag(struct compute_memory_pool *pool,
>  	struct pipe_context *pipe);
>  
>  int compute_memory_promote_item(struct compute_memory_pool *pool,
> -		struct compute_memory_item *item, struct pipe_context *pipe,
> -		int64_t start_in_dw);
> +	struct compute_memory_item *item, struct pipe_context *pipe,
> +	int64_t allocated);
>  
>  void compute_memory_demote_item(struct compute_memory_pool *pool,
>  	struct compute_memory_item *item, struct pipe_context *pipe);
> @@ -107,13 +118,18 @@ void compute_memory_move_item(struct compute_memory_pool *pool,
>  	struct pipe_context *pipe);
>  
>  void compute_memory_free(struct compute_memory_pool* pool, int64_t id);
> -struct compute_memory_item* compute_memory_alloc(struct compute_memory_pool* pool, int64_t size_in_dw); ///Creates pending allocations
> +
> +struct compute_memory_item* compute_memory_alloc(struct compute_memory_pool* pool,
> +	int64_t size_in_dw);
>  
>  void compute_memory_transfer(struct compute_memory_pool* pool,
>  	struct pipe_context * pipe, int device_to_host,
>  	struct compute_memory_item* chunk, void* data,
>  	int offset_in_chunk, int size);
>  
> -void compute_memory_transfer_direct(struct compute_memory_pool* pool, int chunk_to_data, struct compute_memory_item* chunk, struct r600_resource* data, int offset_in_chunk, int offset_in_data, int size); ///Transfer data between chunk<->data, it is for VRAM<->GART transfers
> +void compute_memory_transfer_direct(struct compute_memory_pool* pool,
> +	int chunk_to_data, struct compute_memory_item* chunk,
> +	struct r600_resource* data, int offset_in_chunk,
> +	int offset_in_data, int size);
>  
>  #endif
> -- 
> 2.0.3
> 
> _______________________________________________
> mesa-dev mailing list
> mesa-dev at lists.freedesktop.org
> http://lists.freedesktop.org/mailman/listinfo/mesa-dev


More information about the mesa-dev mailing list