[PATCH] dma-buf: A collection of typo and documentation fixes
Christian König
christian.koenig at amd.com
Thu Nov 24 07:03:09 UTC 2022
Am 23.11.22 um 20:35 schrieb T.J. Mercier:
> I've been collecting these typo fixes for a while and it feels like
> time to send them in.
>
> Signed-off-by: T.J. Mercier <tjmercier at google.com>
Acked-by: Christian König <christian.koenig at amd.com>
> ---
> drivers/dma-buf/dma-buf.c | 14 +++++++-------
> include/linux/dma-buf.h | 6 +++---
> 2 files changed, 10 insertions(+), 10 deletions(-)
>
> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
> index dd0f83ee505b..614ccd208af4 100644
> --- a/drivers/dma-buf/dma-buf.c
> +++ b/drivers/dma-buf/dma-buf.c
> @@ -1141,7 +1141,7 @@ EXPORT_SYMBOL_NS_GPL(dma_buf_unmap_attachment, DMA_BUF);
> *
> * @dmabuf: [in] buffer which is moving
> *
> - * Informs all attachmenst that they need to destroy and recreated all their
> + * Informs all attachments that they need to destroy and recreate all their
> * mappings.
> */
> void dma_buf_move_notify(struct dma_buf *dmabuf)
> @@ -1159,11 +1159,11 @@ EXPORT_SYMBOL_NS_GPL(dma_buf_move_notify, DMA_BUF);
> /**
> * DOC: cpu access
> *
> - * There are mutliple reasons for supporting CPU access to a dma buffer object:
> + * There are multiple reasons for supporting CPU access to a dma buffer object:
> *
> * - Fallback operations in the kernel, for example when a device is connected
> * over USB and the kernel needs to shuffle the data around first before
> - * sending it away. Cache coherency is handled by braketing any transactions
> + * sending it away. Cache coherency is handled by bracketing any transactions
> * with calls to dma_buf_begin_cpu_access() and dma_buf_end_cpu_access()
> * access.
> *
> @@ -1190,7 +1190,7 @@ EXPORT_SYMBOL_NS_GPL(dma_buf_move_notify, DMA_BUF);
> * replace ION buffers mmap support was needed.
> *
> * There is no special interfaces, userspace simply calls mmap on the dma-buf
> - * fd. But like for CPU access there's a need to braket the actual access,
> + * fd. But like for CPU access there's a need to bracket the actual access,
> * which is handled by the ioctl (DMA_BUF_IOCTL_SYNC). Note that
> * DMA_BUF_IOCTL_SYNC can fail with -EAGAIN or -EINTR, in which case it must
> * be restarted.
> @@ -1264,10 +1264,10 @@ static int __dma_buf_begin_cpu_access(struct dma_buf *dmabuf,
> * preparations. Coherency is only guaranteed in the specified range for the
> * specified access direction.
> * @dmabuf: [in] buffer to prepare cpu access for.
> - * @direction: [in] length of range for cpu access.
> + * @direction: [in] direction of access.
> *
> * After the cpu access is complete the caller should call
> - * dma_buf_end_cpu_access(). Only when cpu access is braketed by both calls is
> + * dma_buf_end_cpu_access(). Only when cpu access is bracketed by both calls is
> * it guaranteed to be coherent with other DMA access.
> *
> * This function will also wait for any DMA transactions tracked through
> @@ -1307,7 +1307,7 @@ EXPORT_SYMBOL_NS_GPL(dma_buf_begin_cpu_access, DMA_BUF);
> * actions. Coherency is only guaranteed in the specified range for the
> * specified access direction.
> * @dmabuf: [in] buffer to complete cpu access for.
> - * @direction: [in] length of range for cpu access.
> + * @direction: [in] direction of access.
> *
> * This terminates CPU access started with dma_buf_begin_cpu_access().
> *
> diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
> index 71731796c8c3..1d61a4f6db35 100644
> --- a/include/linux/dma-buf.h
> +++ b/include/linux/dma-buf.h
> @@ -330,7 +330,7 @@ struct dma_buf {
> * @lock:
> *
> * Used internally to serialize list manipulation, attach/detach and
> - * vmap/unmap. Note that in many cases this is superseeded by
> + * vmap/unmap. Note that in many cases this is superseded by
> * dma_resv_lock() on @resv.
> */
> struct mutex lock;
> @@ -365,7 +365,7 @@ struct dma_buf {
> */
> const char *name;
>
> - /** @name_lock: Spinlock to protect name acces for read access. */
> + /** @name_lock: Spinlock to protect name access for read access. */
> spinlock_t name_lock;
>
> /**
> @@ -402,7 +402,7 @@ struct dma_buf {
> * anything the userspace API considers write access.
> *
> * - Drivers may just always add a write fence, since that only
> - * causes unecessarily synchronization, but no correctness issues.
> + * causes unnecessary synchronization, but no correctness issues.
> *
> * - Some drivers only expose a synchronous userspace API with no
> * pipelining across drivers. These do not set any fences for their
More information about the dri-devel
mailing list