[Intel-xe] [PATCH 4/7] drm/xe: Cleanup CODE_INDENT style issues
Matthew Brost
matthew.brost at intel.com
Fri Jul 14 14:13:16 UTC 2023
On Thu, Jul 13, 2023 at 03:06:08PM +0000, Francois Dugast wrote:
> Remove all existing style issues of type CODE_INDENT reported
> by checkpatch.
>
> Signed-off-by: Francois Dugast <francois.dugast at intel.com>
Reviewed-by: Matthew Brost <matthew.brost at intel.com>
> ---
> drivers/gpu/drm/xe/display/xe_fb_pin.c | 4 ++--
> drivers/gpu/drm/xe/xe_engine.c | 4 ++--
> drivers/gpu/drm/xe/xe_guc_fwif.h | 12 ++++++------
> drivers/gpu/drm/xe/xe_guc_submit.c | 14 +++++++-------
> drivers/gpu/drm/xe/xe_hw_engine.c | 2 +-
> drivers/gpu/drm/xe/xe_uc.c | 10 +++++-----
> 6 files changed, 23 insertions(+), 23 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/display/xe_fb_pin.c b/drivers/gpu/drm/xe/display/xe_fb_pin.c
> index 83c9245dd362..2865e989c30c 100644
> --- a/drivers/gpu/drm/xe/display/xe_fb_pin.c
> +++ b/drivers/gpu/drm/xe/display/xe_fb_pin.c
> @@ -275,13 +275,13 @@ void intel_unpin_fb_vma(struct i915_vma *vma, unsigned long flags)
> int intel_plane_pin_fb(struct intel_plane_state *plane_state)
> {
> struct drm_framebuffer *fb = plane_state->hw.fb;
> - struct xe_bo *bo = intel_fb_obj(fb);
> + struct xe_bo *bo = intel_fb_obj(fb);
> struct i915_vma *vma;
>
> /* We reject creating !SCANOUT fb's, so this is weird.. */
> drm_WARN_ON(bo->ttm.base.dev, !(bo->flags & XE_BO_SCANOUT_BIT));
>
> - vma = __xe_pin_fb_vma(to_intel_framebuffer(fb), &plane_state->view.gtt);
> + vma = __xe_pin_fb_vma(to_intel_framebuffer(fb), &plane_state->view.gtt);
> if (IS_ERR(vma))
> return PTR_ERR(vma);
>
> diff --git a/drivers/gpu/drm/xe/xe_engine.c b/drivers/gpu/drm/xe/xe_engine.c
> index 3831c5f82773..957b084629d6 100644
> --- a/drivers/gpu/drm/xe/xe_engine.c
> +++ b/drivers/gpu/drm/xe/xe_engine.c
> @@ -173,7 +173,7 @@ enum drm_sched_priority
> xe_engine_device_get_max_priority(struct xe_device *xe)
> {
> return capable(CAP_SYS_NICE) ? DRM_SCHED_PRIORITY_HIGH :
> - DRM_SCHED_PRIORITY_NORMAL;
> + DRM_SCHED_PRIORITY_NORMAL;
> }
>
> static int engine_set_priority(struct xe_device *xe, struct xe_engine *e,
> @@ -540,7 +540,7 @@ int xe_engine_create_ioctl(struct drm_device *dev, void *data,
> return -EFAULT;
>
> if (XE_IOCTL_ERR(xe, eci[0].gt_id >= xe->info.tile_count))
> - return -EINVAL;
> + return -EINVAL;
>
> if (eci[0].engine_class == DRM_XE_ENGINE_CLASS_VM_BIND) {
> for_each_gt(gt, xe, id) {
> diff --git a/drivers/gpu/drm/xe/xe_guc_fwif.h b/drivers/gpu/drm/xe/xe_guc_fwif.h
> index 27d132ce2087..e215e8b2c17a 100644
> --- a/drivers/gpu/drm/xe/xe_guc_fwif.h
> +++ b/drivers/gpu/drm/xe/xe_guc_fwif.h
> @@ -64,19 +64,19 @@ struct guc_ctxt_registration_info {
>
> /* 32-bit KLV structure as used by policy updates and others */
> struct guc_klv_generic_dw_t {
> - u32 kl;
> - u32 value;
> + u32 kl;
> + u32 value;
> } __packed;
>
> /* Format of the UPDATE_CONTEXT_POLICIES H2G data packet */
> struct guc_update_engine_policy_header {
> - u32 action;
> - u32 guc_id;
> + u32 action;
> + u32 guc_id;
> } __packed;
>
> struct guc_update_engine_policy {
> - struct guc_update_engine_policy_header header;
> - struct guc_klv_generic_dw_t klv[GUC_CONTEXT_POLICIES_KLV_NUM_IDS];
> + struct guc_update_engine_policy_header header;
> + struct guc_klv_generic_dw_t klv[GUC_CONTEXT_POLICIES_KLV_NUM_IDS];
> } __packed;
>
> /* GUC_CTL_* - Parameters for loading the GuC */
> diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c b/drivers/gpu/drm/xe/xe_guc_submit.c
> index 86c445903560..efbc4b13e9e1 100644
> --- a/drivers/gpu/drm/xe/xe_guc_submit.c
> +++ b/drivers/gpu/drm/xe/xe_guc_submit.c
> @@ -352,17 +352,17 @@ static const int drm_sched_prio_to_guc[] = {
>
> static void init_policies(struct xe_guc *guc, struct xe_engine *e)
> {
> - struct engine_policy policy;
> + struct engine_policy policy;
> enum drm_sched_priority prio = e->entity->priority;
> u32 timeslice_us = e->sched_props.timeslice_us;
> u32 preempt_timeout_us = e->sched_props.preempt_timeout_us;
>
> XE_BUG_ON(!engine_registered(e));
>
> - __guc_engine_policy_start_klv(&policy, e->guc->id);
> - __guc_engine_policy_add_priority(&policy, drm_sched_prio_to_guc[prio]);
> - __guc_engine_policy_add_execution_quantum(&policy, timeslice_us);
> - __guc_engine_policy_add_preemption_timeout(&policy, preempt_timeout_us);
> + __guc_engine_policy_start_klv(&policy, e->guc->id);
> + __guc_engine_policy_add_priority(&policy, drm_sched_prio_to_guc[prio]);
> + __guc_engine_policy_add_execution_quantum(&policy, timeslice_us);
> + __guc_engine_policy_add_preemption_timeout(&policy, preempt_timeout_us);
>
> xe_guc_ct_send(&guc->ct, (u32 *)&policy.h2g,
> __guc_engine_policy_action_size(&policy), 0, 0);
> @@ -372,8 +372,8 @@ static void set_min_preemption_timeout(struct xe_guc *guc, struct xe_engine *e)
> {
> struct engine_policy policy;
>
> - __guc_engine_policy_start_klv(&policy, e->guc->id);
> - __guc_engine_policy_add_preemption_timeout(&policy, 1);
> + __guc_engine_policy_start_klv(&policy, e->guc->id);
> + __guc_engine_policy_add_preemption_timeout(&policy, 1);
>
> xe_guc_ct_send(&guc->ct, (u32 *)&policy.h2g,
> __guc_engine_policy_action_size(&policy), 0, 0);
> diff --git a/drivers/gpu/drm/xe/xe_hw_engine.c b/drivers/gpu/drm/xe/xe_hw_engine.c
> index b7b02c96e998..1af5cccd1142 100644
> --- a/drivers/gpu/drm/xe/xe_hw_engine.c
> +++ b/drivers/gpu/drm/xe/xe_hw_engine.c
> @@ -312,7 +312,7 @@ hw_engine_setup_default_state(struct xe_hw_engine *hwe)
> /* TODO: missing handling of HAS_L3_CCS_READ platforms */
> const u8 mocs_read_idx = gt->mocs.uc_index;
> u32 ring_cmd_cctl_val = REG_FIELD_PREP(CMD_CCTL_WRITE_OVERRIDE_MASK, mocs_write_idx) |
> - REG_FIELD_PREP(CMD_CCTL_READ_OVERRIDE_MASK, mocs_read_idx);
> + REG_FIELD_PREP(CMD_CCTL_READ_OVERRIDE_MASK, mocs_read_idx);
> struct xe_rtp_process_ctx ctx = XE_RTP_PROCESS_CTX_INITIALIZER(hwe);
> const struct xe_rtp_entry_sr engine_entries[] = {
> /*
> diff --git a/drivers/gpu/drm/xe/xe_uc.c b/drivers/gpu/drm/xe/xe_uc.c
> index 70eabf567156..e244d27b55d5 100644
> --- a/drivers/gpu/drm/xe/xe_uc.c
> +++ b/drivers/gpu/drm/xe/xe_uc.c
> @@ -201,14 +201,14 @@ int xe_uc_start(struct xe_uc *uc)
>
> static void uc_reset_wait(struct xe_uc *uc)
> {
> - int ret;
> + int ret;
>
> again:
> - xe_guc_reset_wait(&uc->guc);
> + xe_guc_reset_wait(&uc->guc);
>
> - ret = xe_uc_reset_prepare(uc);
> - if (ret)
> - goto again;
> + ret = xe_uc_reset_prepare(uc);
> + if (ret)
> + goto again;
> }
>
> int xe_uc_suspend(struct xe_uc *uc)
> --
> 2.34.1
>
More information about the Intel-xe
mailing list