[PATCH v5 6/6] drm/sprd: add Unisoc's drm mipi dsi&dphy driver

Maxime Ripard maxime at cerno.tech
Fri Apr 30 09:35:03 UTC 2021


Hi,

On Sun, Apr 25, 2021 at 08:36:07PM +0800, Kevin Tang wrote:
> Adds dsi host controller support for the Unisoc's display subsystem.
> Adds dsi phy support for the Unisoc's display subsystem.
> Only MIPI DSI Displays supported, DP/TV/HMDI will be support
> in the feature.
> 
> v1:
>   - Remove dphy and dsi graph binding, merge the dphy driver into the dsi.
> 
> v2:
>   - Use drm_xxx to replace all DRM_XXX.
>   - Use kzalloc to replace devm_kzalloc for sprd_dsi structure init.
> 
> v4:
>   - Use drmm_helpers to allocate encoder.
>   - Move allocate encoder and connector to bind function.
> 
> v5:
>   - Drop the dsi ip file prefix.
>   - Fix the checkpatch warnings.
>   - Add Signed-off-by for dsi&dphy patch.
>   - Use the mode_flags of mipi_dsi_device to setup crtc DPI and EDPI
>     mode.
> 
> Cc: Orson Zhai <orsonzhai at gmail.com>
> Cc: Chunyan Zhang <zhang.lyra at gmail.com>
> Signed-off-by: Kevin Tang <kevin.tang at unisoc.com>

Output from checkpatch:
total: 0 errors, 3 warnings, 100 checks, 4207 lines checked

> ---
>  drivers/gpu/drm/sprd/Kconfig         |    1 +
>  drivers/gpu/drm/sprd/Makefile        |    6 +-
>  drivers/gpu/drm/sprd/dsi_ctrl.c      |  794 ++++++++++++++
>  drivers/gpu/drm/sprd/dsi_ctrl.h      | 1475 ++++++++++++++++++++++++++
>  drivers/gpu/drm/sprd/dsi_ctrl_ppi.c  |  157 +++
>  drivers/gpu/drm/sprd/dsi_ctrl_ppi.h  |   26 +
>  drivers/gpu/drm/sprd/megacores_pll.c |  317 ++++++
>  drivers/gpu/drm/sprd/megacores_pll.h |  146 +++
>  drivers/gpu/drm/sprd/sprd_dpu.c      |   17 +
>  drivers/gpu/drm/sprd/sprd_drm.c      |    1 +
>  drivers/gpu/drm/sprd/sprd_drm.h      |    1 +
>  drivers/gpu/drm/sprd/sprd_dsi.c      | 1124 ++++++++++++++++++++
>  drivers/gpu/drm/sprd/sprd_dsi.h      |  107 ++
>  13 files changed, 4171 insertions(+), 1 deletion(-)
>  create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl.c
>  create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl.h
>  create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl_ppi.c
>  create mode 100644 drivers/gpu/drm/sprd/dsi_ctrl_ppi.h
>  create mode 100644 drivers/gpu/drm/sprd/megacores_pll.c
>  create mode 100644 drivers/gpu/drm/sprd/megacores_pll.h
>  create mode 100644 drivers/gpu/drm/sprd/sprd_dsi.c
>  create mode 100644 drivers/gpu/drm/sprd/sprd_dsi.h
> 
> diff --git a/drivers/gpu/drm/sprd/Kconfig b/drivers/gpu/drm/sprd/Kconfig
> index 37762c333..3edeaeca0 100644
> --- a/drivers/gpu/drm/sprd/Kconfig
> +++ b/drivers/gpu/drm/sprd/Kconfig
> @@ -5,6 +5,7 @@ config DRM_SPRD
>  	select DRM_GEM_CMA_HELPER
>  	select DRM_KMS_CMA_HELPER
>  	select DRM_KMS_HELPER
> +	select DRM_MIPI_DSI
>  	select VIDEOMODE_HELPERS
>  	help
>  	  Choose this option if you have a Unisoc chipset.
> diff --git a/drivers/gpu/drm/sprd/Makefile b/drivers/gpu/drm/sprd/Makefile
> index ab12b95e6..d49f4977b 100644
> --- a/drivers/gpu/drm/sprd/Makefile
> +++ b/drivers/gpu/drm/sprd/Makefile
> @@ -1,4 +1,8 @@
>  # SPDX-License-Identifier: GPL-2.0
>  
>  obj-y := sprd_drm.o \
> -	sprd_dpu.o
> +	sprd_dpu.o \
> +	sprd_dsi.o \
> +	dw_dsi_ctrl.o \
> +	dw_dsi_ctrl_ppi.o \
> +	megacores_pll.o
> diff --git a/drivers/gpu/drm/sprd/dsi_ctrl.c b/drivers/gpu/drm/sprd/dsi_ctrl.c
> new file mode 100644
> index 000000000..7eccf9654
> --- /dev/null
> +++ b/drivers/gpu/drm/sprd/dsi_ctrl.c
> @@ -0,0 +1,794 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020 Unisoc Inc.
> + */
> +
> +#include <linux/io.h>
> +#include <linux/init.h>
> +#include <linux/delay.h>
> +#include <linux/module.h>
> +
> +#include "dsi_ctrl.h"
> +
> +/*
> + * Modify power status of DSI Host core
> + */
> +void dsi_power_enable(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(enable, &reg->SOFT_RESET);
> +}
> +/*
> + * Enable/disable DPI video mode
> + */
> +void dsi_video_mode(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(0, &reg->DSI_MODE_CFG);
> +}
> +/*
> + * Enable command mode (Generic interface)
> + */
> +void dsi_cmd_mode(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(1, &reg->DSI_MODE_CFG);
> +}
> +
> +bool dsi_is_cmd_mode(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	return readl(&reg->DSI_MODE_CFG);
> +}
> +/*
> + * Configure the read back virtual channel for the generic interface
> + */
> +void dsi_rx_vcid(struct dsi_context *ctx, u8 vc)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x1C virtual_channel_id;
> +
> +	virtual_channel_id.val = readl(&reg->VIRTUAL_CHANNEL_ID);
> +	virtual_channel_id.bits.gen_rx_vcid = vc;
> +
> +	writel(virtual_channel_id.val, &reg->VIRTUAL_CHANNEL_ID);
> +}
> +/*
> + * Write the DPI video virtual channel destination
> + */
> +void dsi_video_vcid(struct dsi_context *ctx, u8 vc)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x1C virtual_channel_id;
> +
> +	virtual_channel_id.val = readl(&reg->VIRTUAL_CHANNEL_ID);
> +	virtual_channel_id.bits.video_pkt_vcid = vc;
> +
> +	writel(virtual_channel_id.val, &reg->VIRTUAL_CHANNEL_ID);
> +}
> +/*
> + * Set DPI video mode type (burst/non-burst - with sync pulses or events)
> + */
> +void dsi_dpi_video_burst_mode(struct dsi_context *ctx, int mode)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x38 vid_mode_cfg;
> +
> +	vid_mode_cfg.val = readl(&reg->VID_MODE_CFG);
> +	vid_mode_cfg.bits.vid_mode_type = mode;
> +
> +	writel(vid_mode_cfg.val, &reg->VID_MODE_CFG);
> +}
> +/*
> + * Set DPI video color coding
> + */
> +void dsi_dpi_color_coding(struct dsi_context *ctx, int coding)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x20 dpi_video_format;
> +
> +	dpi_video_format.val = readl(&reg->DPI_VIDEO_FORMAT);
> +	dpi_video_format.bits.dpi_video_mode_format = coding;
> +
> +	writel(dpi_video_format.val, &reg->DPI_VIDEO_FORMAT);
> +}
> +/*
> + * Configure the Horizontal Line time
> + * param "byte_cycle" taken to transmit the total of the horizontal line
> + */
> +void dsi_dpi_hline_time(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x2C video_line_time;
> +
> +	video_line_time.val = readl(&reg->VIDEO_LINE_TIME);
> +	video_line_time.bits.video_line_time = byte_cycle;
> +
> +	writel(video_line_time.val, &reg->VIDEO_LINE_TIME);
> +}
> +/*
> + * Configure the Horizontal back porch time
> + * param "byte_cycle" taken to transmit the horizontal back porch
> + */
> +void dsi_dpi_hbp_time(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x28 video_line_hblk_time;
> +
> +	video_line_hblk_time.val = readl(&reg->VIDEO_LINE_HBLK_TIME);
> +	video_line_hblk_time.bits.video_line_hbp_time = byte_cycle;
> +
> +	writel(video_line_hblk_time.val, &reg->VIDEO_LINE_HBLK_TIME);
> +}
> +/*
> + * Configure the Horizontal sync time,
> + * param "byte_cycle" taken to transmit the horizontal sync
> + */
> +void dsi_dpi_hsync_time(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x28 video_line_hblk_time;
> +
> +	video_line_hblk_time.val = readl(&reg->VIDEO_LINE_HBLK_TIME);
> +	video_line_hblk_time.bits.video_line_hsa_time = byte_cycle;
> +
> +	writel(video_line_hblk_time.val, &reg->VIDEO_LINE_HBLK_TIME);
> +}
> +/*
> + * Configure the vertical active lines of the video stream
> + */
> +void dsi_dpi_vact(struct dsi_context *ctx, u16 lines)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x34 video_active_lines;
> +
> +	video_active_lines.val = readl(&reg->VIDEO_VACTIVE_LINES);
> +	video_active_lines.bits.vactive_lines = lines;
> +
> +	writel(video_active_lines.val, &reg->VIDEO_VACTIVE_LINES);
> +}
> +
> +void dsi_dpi_vfp(struct dsi_context *ctx, u16 lines)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x30 video_vblk_lines;
> +
> +	video_vblk_lines.val = readl(&reg->VIDEO_VBLK_LINES);
> +	video_vblk_lines.bits.vfp_lines = lines;
> +
> +	writel(video_vblk_lines.val, &reg->VIDEO_VBLK_LINES);
> +}
> +
> +void dsi_dpi_vbp(struct dsi_context *ctx, u16 lines)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x30 video_vblk_lines;
> +
> +	video_vblk_lines.val = readl(&reg->VIDEO_VBLK_LINES);
> +	video_vblk_lines.bits.vbp_lines = lines;
> +
> +	writel(video_vblk_lines.val, &reg->VIDEO_VBLK_LINES);
> +}
> +
> +void dsi_dpi_vsync(struct dsi_context *ctx, u16 lines)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x30 video_vblk_lines;
> +
> +	video_vblk_lines.val = readl(&reg->VIDEO_VBLK_LINES);
> +	video_vblk_lines.bits.vsa_lines = lines;
> +
> +	writel(video_vblk_lines.val, &reg->VIDEO_VBLK_LINES);
> +}
> +
> +void dsi_dpi_hporch_lp_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x38 vid_mode_cfg;
> +
> +	vid_mode_cfg.val = readl(&reg->VID_MODE_CFG);
> +
> +	vid_mode_cfg.bits.lp_hfp_en = enable;
> +	vid_mode_cfg.bits.lp_hbp_en = enable;
> +
> +	writel(vid_mode_cfg.val, &reg->VID_MODE_CFG);
> +}
> +/*
> + * Enable return to low power mode inside vertical active lines periods when
> + * timing allows
> + */
> +void dsi_dpi_vporch_lp_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x38 vid_mode_cfg;
> +
> +	vid_mode_cfg.val = readl(&reg->VID_MODE_CFG);
> +
> +	vid_mode_cfg.bits.lp_vact_en = enable;
> +	vid_mode_cfg.bits.lp_vfp_en = enable;
> +	vid_mode_cfg.bits.lp_vbp_en = enable;
> +	vid_mode_cfg.bits.lp_vsa_en = enable;
> +
> +	writel(vid_mode_cfg.val, &reg->VID_MODE_CFG);
> +}
> +/*
> + * Enable FRAME BTA ACK
> + */
> +void dsi_dpi_frame_ack_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x38 vid_mode_cfg;
> +
> +	vid_mode_cfg.val = readl(&reg->VID_MODE_CFG);
> +	vid_mode_cfg.bits.frame_bta_ack_en = enable;
> +
> +	writel(vid_mode_cfg.val, &reg->VID_MODE_CFG);
> +}
> +/*
> + * Write no of chunks to core - taken into consideration only when multi packet
> + * is enabled
> + */
> +void dsi_dpi_chunk_num(struct dsi_context *ctx, u16 num)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x24 video_pkt_config;
> +
> +	video_pkt_config.val = readl(&reg->VIDEO_PKT_CONFIG);
> +	video_pkt_config.bits.video_line_chunk_num = num;
> +
> +	writel(video_pkt_config.val, &reg->VIDEO_PKT_CONFIG);
> +}
> +/*
> + * Write the null packet size - will only be taken into account when null
> + * packets are enabled.
> + */
> +void dsi_dpi_null_packet_size(struct dsi_context *ctx, u16 size)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xC0 video_nullpkt_size;
> +
> +	video_nullpkt_size.val = readl(&reg->VIDEO_NULLPKT_SIZE);
> +	video_nullpkt_size.bits.video_nullpkt_size = size;
> +
> +	writel(video_nullpkt_size.val, &reg->VIDEO_NULLPKT_SIZE);
> +}
> +/*
> + * Write video packet size. obligatory for sending video
> + */
> +void dsi_dpi_video_packet_size(struct dsi_context *ctx, u16 size)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x24 video_pkt_config;
> +
> +	video_pkt_config.val = readl(&reg->VIDEO_PKT_CONFIG);
> +	video_pkt_config.bits.video_pkt_size = size;
> +
> +	writel(video_pkt_config.val, &reg->VIDEO_PKT_CONFIG);
> +}
> +/*
> + * Specifiy the size of the packet memory write start/continue
> + */
> +void dsi_edpi_max_pkt_size(struct dsi_context *ctx, u16 size)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xC4 dcs_wm_pkt_size;
> +
> +	dcs_wm_pkt_size.val = readl(&reg->DCS_WM_PKT_SIZE);
> +	dcs_wm_pkt_size.bits.dcs_wm_pkt_size = size;
> +
> +	writel(dcs_wm_pkt_size.val, &reg->DCS_WM_PKT_SIZE);
> +}
> +/*
> + * Enable tear effect acknowledge
> + */
> +void dsi_tear_effect_ack_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x68 cmd_mode_cfg;
> +
> +	cmd_mode_cfg.val = readl(&reg->CMD_MODE_CFG);
> +	cmd_mode_cfg.bits.tear_fx_en = enable;
> +
> +	writel(cmd_mode_cfg.val, &reg->CMD_MODE_CFG);
> +}
> +/*
> + * Set DCS command packet transmission to transmission type
> + */
> +void dsi_cmd_mode_lp_cmd_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x68 cmd_mode_cfg;
> +
> +	cmd_mode_cfg.val = readl(&reg->CMD_MODE_CFG);
> +
> +	cmd_mode_cfg.bits.gen_sw_0p_tx = enable;
> +	cmd_mode_cfg.bits.gen_sw_1p_tx = enable;
> +	cmd_mode_cfg.bits.gen_sw_2p_tx = enable;
> +	cmd_mode_cfg.bits.gen_lw_tx = enable;
> +	cmd_mode_cfg.bits.dcs_sw_0p_tx = enable;
> +	cmd_mode_cfg.bits.dcs_sw_1p_tx = enable;
> +	cmd_mode_cfg.bits.dcs_lw_tx = enable;
> +	cmd_mode_cfg.bits.max_rd_pkt_size = enable;
> +
> +	cmd_mode_cfg.bits.gen_sr_0p_tx = enable;
> +	cmd_mode_cfg.bits.gen_sr_1p_tx = enable;
> +	cmd_mode_cfg.bits.gen_sr_2p_tx = enable;
> +	cmd_mode_cfg.bits.dcs_sr_0p_tx = enable;
> +
> +	writel(cmd_mode_cfg.val, &reg->CMD_MODE_CFG);
> +}
> +/*
> + * Set DCS read command packet transmission to transmission type
> + */
> +void dsi_video_mode_lp_cmd_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x38 vid_mode_cfg;
> +
> +	vid_mode_cfg.val = readl(&reg->VID_MODE_CFG);
> +	vid_mode_cfg.bits.lp_cmd_en = enable;
> +
> +	writel(vid_mode_cfg.val, &reg->VID_MODE_CFG);
> +}
> +
> +/*
> + * Write command header in the generic interface (which also sends DCS commands) as a subset
> + */
> +void dsi_set_packet_header(struct dsi_context *ctx,
> +				   u8 vc,
> +				   u8 type,
> +				   u8 wc_lsb,
> +				   u8 wc_msb)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x6C gen_hdr;
> +
> +	gen_hdr.bits.gen_dt = type;
> +	gen_hdr.bits.gen_vc = vc;
> +	gen_hdr.bits.gen_wc_lsbyte = wc_lsb;
> +	gen_hdr.bits.gen_wc_msbyte = wc_msb;
> +
> +	writel(gen_hdr.val, &reg->GEN_HDR);
> +}
> +/*
> + * Write the payload of the long packet commands
> + */
> +void dsi_set_packet_payload(struct dsi_context *ctx, u32 payload)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(payload, &reg->GEN_PLD_DATA);
> +}
> +/*
> + * Read the payload of the long packet commands
> + */
> +u32 dsi_get_rx_payload(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	return readl(&reg->GEN_PLD_DATA);
> +}
> +
> +/*
> + * Enable Bus Turn-around request
> + */
> +void dsi_bta_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(enable, &reg->TA_EN);
> +}
> +/*
> + * Enable EOTp reception
> + */
> +void dsi_eotp_rx_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xBC eotp_en;
> +
> +	eotp_en.val = readl(&reg->EOTP_EN);
> +	eotp_en.bits.rx_eotp_en = enable;
> +
> +	writel(eotp_en.val, &reg->EOTP_EN);
> +}
> +/*
> + * Enable EOTp transmission
> + */
> +void dsi_eotp_tx_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xBC eotp_en;
> +
> +	eotp_en.val = readl(&reg->EOTP_EN);
> +	eotp_en.bits.tx_eotp_en = enable;
> +
> +	writel(eotp_en.val, &reg->EOTP_EN);
> +}
> +/*
> + * Enable ECC reception, error correction and reporting
> + */
> +void dsi_ecc_rx_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xB4 rx_pkt_check_config;
> +
> +	rx_pkt_check_config.val = readl(&reg->RX_PKT_CHECK_CONFIG);
> +	rx_pkt_check_config.bits.rx_pkt_ecc_en = enable;
> +
> +	writel(rx_pkt_check_config.val, &reg->RX_PKT_CHECK_CONFIG);
> +}
> +/*
> + * Enable CRC reception, error reporting
> + */
> +void dsi_crc_rx_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xB4 rx_pkt_check_config;
> +
> +	rx_pkt_check_config.val = readl(&reg->RX_PKT_CHECK_CONFIG);
> +	rx_pkt_check_config.bits.rx_pkt_crc_en = enable;
> +
> +	writel(rx_pkt_check_config.val, &reg->RX_PKT_CHECK_CONFIG);
> +}
> +/*
> + * Get status of read command
> + */
> +bool dsi_is_bta_returned(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x98 cmd_mode_status;
> +
> +	cmd_mode_status.val = readl(&reg->CMD_MODE_STATUS);
> +
> +	return cmd_mode_status.bits.gen_cmd_rdcmd_done;
> +}
> +/*
> + * Get the FULL status of generic read payload fifo
> + */
> +bool dsi_is_rx_payload_fifo_full(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x98 cmd_mode_status;
> +
> +	cmd_mode_status.val = readl(&reg->CMD_MODE_STATUS);
> +
> +	return cmd_mode_status.bits.gen_cmd_rdata_fifo_full;
> +}
> +/*
> + * Get the EMPTY status of generic read payload fifo
> + */
> +bool dsi_is_rx_payload_fifo_empty(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x98 cmd_mode_status;
> +
> +	cmd_mode_status.val = readl(&reg->CMD_MODE_STATUS);
> +
> +	return cmd_mode_status.bits.gen_cmd_rdata_fifo_empty;
> +}
> +/*
> + * Get the FULL status of generic write payload fifo
> + */
> +bool dsi_is_tx_payload_fifo_full(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x98 cmd_mode_status;
> +
> +	cmd_mode_status.val = readl(&reg->CMD_MODE_STATUS);
> +
> +	return cmd_mode_status.bits.gen_cmd_wdata_fifo_full;
> +}
> +/*
> + * Get the EMPTY status of generic write payload fifo
> + */
> +bool dsi_is_tx_payload_fifo_empty(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x98 cmd_mode_status;
> +
> +	cmd_mode_status.val = readl(&reg->CMD_MODE_STATUS);
> +
> +	return cmd_mode_status.bits.gen_cmd_wdata_fifo_empty;
> +}
> +/*
> + * Get the EMPTY status of generic command fifo
> + */
> +bool dsi_is_tx_cmd_fifo_empty(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x98 cmd_mode_status;
> +
> +	cmd_mode_status.val = readl(&reg->CMD_MODE_STATUS);
> +
> +	return cmd_mode_status.bits.gen_cmd_cmd_fifo_empty;
> +}
> +/*
> + * DPI interface signal delay config
> + * param byte_cycle period for waiting after controller receiving HSYNC from
> + * DPI interface to start read pixel data from memory.
> + */
> +void dsi_dpi_sig_delay(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xD0 video_sig_delay_config;
> +
> +	video_sig_delay_config.val = readl(&reg->VIDEO_SIG_DELAY_CONFIG);
> +	video_sig_delay_config.bits.video_sig_delay = byte_cycle;
> +
> +	writel(video_sig_delay_config.val, &reg->VIDEO_SIG_DELAY_CONFIG);
> +}
> +/*
> + * Configure how many cycles of byte clock would the PHY module take
> + * to switch data lane from high speed to low power
> + */
> +void dsi_datalane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xAC phy_datalane_time_config;
> +
> +	phy_datalane_time_config.val = readl(&reg->PHY_DATALANE_TIME_CONFIG);
> +	phy_datalane_time_config.bits.phy_datalane_hs_to_lp_time = byte_cycle;
> +
> +	writel(phy_datalane_time_config.val, &reg->PHY_DATALANE_TIME_CONFIG);
> +}
> +/*
> + * Configure how many cycles of byte clock would the PHY module take
> + * to switch the data lane from to low power high speed
> + */
> +void dsi_datalane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xAC phy_datalane_time_config;
> +
> +	phy_datalane_time_config.val = readl(&reg->PHY_DATALANE_TIME_CONFIG);
> +	phy_datalane_time_config.bits.phy_datalane_lp_to_hs_time = byte_cycle;
> +
> +	writel(phy_datalane_time_config.val, &reg->PHY_DATALANE_TIME_CONFIG);
> +}
> +/*
> + * Configure how many cycles of byte clock would the PHY module take
> + * to switch clock lane from high speed to low power
> + */
> +void dsi_clklane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xA8 phy_clklane_time_config;
> +
> +	phy_clklane_time_config.val = readl(&reg->PHY_CLKLANE_TIME_CONFIG);
> +	phy_clklane_time_config.bits.phy_clklane_hs_to_lp_time = byte_cycle;
> +
> +	writel(phy_clklane_time_config.val, &reg->PHY_CLKLANE_TIME_CONFIG);
> +}
> +/*
> + * Configure how many cycles of byte clock would the PHY module take
> + * to switch clock lane from to low power high speed
> + */
> +void dsi_clklane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xA8 phy_clklane_time_config;
> +
> +	phy_clklane_time_config.val = readl(&reg->PHY_CLKLANE_TIME_CONFIG);
> +	phy_clklane_time_config.bits.phy_clklane_lp_to_hs_time = byte_cycle;
> +
> +	writel(phy_clklane_time_config.val, &reg->PHY_CLKLANE_TIME_CONFIG);
> +}
> +/*
> + * Configure how many cycles of byte clock would the PHY module take
> + * to turn the bus around to start receiving
> + */
> +void dsi_max_read_time(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(byte_cycle, &reg->MAX_READ_TIME);
> +}
> +/*
> + * Enable the automatic mechanism to stop providing clock in the clock
> + * lane when time allows
> + */
> +void dsi_nc_clk_en(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x74 phy_clk_lane_lp_ctrl;
> +
> +	phy_clk_lane_lp_ctrl.val = readl(&reg->PHY_CLK_LANE_LP_CTRL);
> +	phy_clk_lane_lp_ctrl.bits.auto_clklane_ctrl_en = enable;
> +
> +	writel(phy_clk_lane_lp_ctrl.val, &reg->PHY_CLK_LANE_LP_CTRL);
> +}
> +/*
> + * Write transmission escape timeout
> + * a safe guard so that the state machine would reset if transmission
> + * takes too long
> + */
> +void dsi_tx_escape_division(struct dsi_context *ctx, u8 div)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(div, &reg->TX_ESC_CLK_CONFIG);
> +}
> +/*
> + * Configure timeout divisions (so they would have more clock ticks)
> + * div no of hs cycles before transiting back to LP in
> + *  (lane_clk / div)
> + */
> +void dsi_timeout_clock_division(struct dsi_context *ctx, u8 div)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(div, &reg->TIMEOUT_CNT_CLK_CONFIG);
> +}
> +/*
> + * Configure the Low power receive time out
> + */
> +void dsi_lp_rx_timeout(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(byte_cycle, &reg->LRX_H_TO_CONFIG);
> +}
> +/*
> + * Configure a high speed transmission time out
> + */
> +void dsi_hs_tx_timeout(struct dsi_context *ctx, u16 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(byte_cycle, &reg->HTX_TO_CONFIG);
> +}
> +/*
> + * Get the error 0 interrupt register status
> + */
> +u32 dsi_int0_status(struct dsi_context *ctx)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x08 protocol_int_sts;
> +
> +	protocol_int_sts.val = readl(&reg->PROTOCOL_INT_STS);
> +	writel(protocol_int_sts.val, &reg->PROTOCOL_INT_CLR);
> +
> +	if (protocol_int_sts.bits.dphy_errors_0)
> +		drm_err(dsi->drm, "dphy_err: escape entry error\n");
> +
> +	if (protocol_int_sts.bits.dphy_errors_1)
> +		drm_err(dsi->drm, "dphy_err: lp data transmission sync error\n");
> +
> +	if (protocol_int_sts.bits.dphy_errors_2)
> +		drm_err(dsi->drm, "dphy_err: control error\n");
> +
> +	if (protocol_int_sts.bits.dphy_errors_3)
> +		drm_err(dsi->drm, "dphy_err: LP0 contention error\n");
> +
> +	if (protocol_int_sts.bits.dphy_errors_4)
> +		drm_err(dsi->drm, "dphy_err: LP1 contention error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_0)
> +		drm_err(dsi->drm, "ack_err: SoT error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_1)
> +		drm_err(dsi->drm, "ack_err: SoT Sync error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_2)
> +		drm_err(dsi->drm, "ack_err: EoT Sync error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_3)
> +		drm_err(dsi->drm, "ack_err: Escape Mode Entry Command error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_4)
> +		drm_err(dsi->drm, "ack_err: LP Transmit Sync error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_5)
> +		drm_err(dsi->drm, "ack_err: Peripheral Timeout error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_6)
> +		drm_err(dsi->drm, "ack_err: False Control error\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_7)
> +		drm_err(dsi->drm, "ack_err: reserved (specific to device)\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_8)
> +		drm_err(dsi->drm, "ack_err: ECC error, single-bit (corrected)\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_9)
> +		drm_err(dsi->drm, "ack_err: ECC error, multi-bit (not corrected)\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_10)
> +		drm_err(dsi->drm, "ack_err: checksum error (long packet only)\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_11)
> +		drm_err(dsi->drm, "ack_err: not recognized DSI data type\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_12)
> +		drm_err(dsi->drm, "ack_err: DSI VC ID Invalid\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_13)
> +		drm_err(dsi->drm, "ack_err: invalid transmission length\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_14)
> +		drm_err(dsi->drm, "ack_err: reserved (specific to device)\n");
> +
> +	if (protocol_int_sts.bits.ack_with_err_15)
> +		drm_err(dsi->drm, "ack_err: DSI protocol violation\n");
> +
> +	return 0;
> +}
> +/*
> + * Get the error 1 interrupt register status
> + */
> +u32 dsi_int1_status(struct dsi_context *ctx)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x10 internal_int_sts;
> +	u32 status = 0;
> +
> +	internal_int_sts.val = readl(&reg->INTERNAL_INT_STS);
> +	writel(internal_int_sts.val, &reg->INTERNAL_INT_CLR);
> +
> +	if (internal_int_sts.bits.receive_pkt_size_err)
> +		drm_err(dsi->drm, "receive packet size error\n");
> +
> +	if (internal_int_sts.bits.eotp_not_receive_err)
> +		drm_err(dsi->drm, "EoTp packet is not received\n");
> +
> +	if (internal_int_sts.bits.gen_cmd_cmd_fifo_wr_err)
> +		drm_err(dsi->drm, "cmd header-fifo is full\n");
> +
> +	if (internal_int_sts.bits.gen_cmd_rdata_fifo_rd_err)
> +		drm_err(dsi->drm, "cmd read-payload-fifo is empty\n");
> +
> +	if (internal_int_sts.bits.gen_cmd_rdata_fifo_wr_err)
> +		drm_err(dsi->drm, "cmd read-payload-fifo is full\n");
> +
> +	if (internal_int_sts.bits.gen_cmd_wdata_fifo_wr_err)
> +		drm_err(dsi->drm, "cmd write-payload-fifo is full\n");
> +
> +	if (internal_int_sts.bits.gen_cmd_wdata_fifo_rd_err)
> +		drm_err(dsi->drm, "cmd write-payload-fifo is empty\n");
> +
> +	if (internal_int_sts.bits.dpi_pix_fifo_wr_err) {
> +		drm_err(dsi->drm, "DPI pixel-fifo is full\n");
> +		status |= DSI_INT_STS_NEED_SOFT_RESET;
> +	}
> +
> +	if (internal_int_sts.bits.ecc_single_err)
> +		drm_err(dsi->drm, "ECC single error in a received packet\n");
> +
> +	if (internal_int_sts.bits.ecc_multi_err)
> +		drm_err(dsi->drm, "ECC multiple error in a received packet\n");
> +
> +	if (internal_int_sts.bits.crc_err)
> +		drm_err(dsi->drm, "CRC error in the received packet payload\n");
> +
> +	if (internal_int_sts.bits.hs_tx_timeout)
> +		drm_err(dsi->drm, "high-speed transmission timeout\n");
> +
> +	if (internal_int_sts.bits.lp_rx_timeout)
> +		drm_err(dsi->drm, "low-power reception timeout\n");
> +
> +	return status;
> +}
> +/*
> + * Configure MASK (hiding) of interrupts coming from error 0 source
> + */
> +void dsi_int0_mask(struct dsi_context *ctx, u32 mask)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(mask, &reg->MASK_PROTOCOL_INT);
> +}
> +/*
> + * Configure MASK (hiding) of interrupts coming from error 1 source
> + */
> +void dsi_int1_mask(struct dsi_context *ctx, u32 mask)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(mask, &reg->MASK_INTERNAL_INT);
> +}
> diff --git a/drivers/gpu/drm/sprd/dsi_ctrl.h b/drivers/gpu/drm/sprd/dsi_ctrl.h
> new file mode 100644
> index 000000000..88069227f
> --- /dev/null
> +++ b/drivers/gpu/drm/sprd/dsi_ctrl.h
> @@ -0,0 +1,1475 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020 Unisoc Inc.
> + */
> +
> +#ifndef _DSI_CTRL_H_
> +#define _DSI_CTRL_H_
> +
> +#include <asm/types.h>
> +
> +#include "sprd_dsi.h"
> +
> +struct dsi_reg {
> +	union _0x00 {
> +		u32 val;
> +		struct _DSI_VERSION {
> +		u32 dsi_version: 16;
> +		u32 reserved: 16;
> +		} bits;
> +	} DSI_VERSION;

Using unions and structures to define the register is really frowned
upon in favor of defines, like you rightfully did in the crtc driver.

> +	union _0x04 {
> +		u32 val;
> +		struct _SOFT_RESET {
> +		/*
> +		 * This bit configures the core either to work normal or to
> +		 * reset. It's default value is 0. After the core configur-
> +		 * ation, to enable the mipi_dsi_host, set this register to 1.
> +		 * 1: power up     0: reset core
> +		 */
> +		u32 dsi_soft_reset: 1;
> +
> +		u32 reserved: 31;
> +		} bits;
> +	} SOFT_RESET;
> +
> +	union _0x08 {
> +		u32 val;
> +		struct _PROTOCOL_INT_STS {
> +		/* ErrEsc escape entry error from Lane 0 */
> +		u32 dphy_errors_0: 1;
> +
> +		/* ErrSyncEsc low-power data transmission synchronization
> +		 * error from Lane 0
> +		 */
> +		u32 dphy_errors_1: 1;
> +
> +		/* ErrControl error from Lane 0 */
> +		u32 dphy_errors_2: 1;
> +
> +		/* ErrContentionLP0 LP0 contention error from Lane 0 */
> +		u32 dphy_errors_3: 1;
> +
> +		/* ErrContentionLP1 LP1 contention error from Lane 0 */
> +		u32 dphy_errors_4: 1;
> +
> +		/* debug mode protocol errors */
> +		u32 protocol_debug_err: 11;
> +
> +		/* SoT error from the Acknowledge error report */
> +		u32 ack_with_err_0: 1;
> +
> +		/* SoT Sync error from the Acknowledge error report */
> +		u32 ack_with_err_1: 1;
> +
> +		/* EoT Sync error from the Acknowledge error report */
> +		u32 ack_with_err_2: 1;
> +
> +		/* Escape Mode Entry Command error from the Acknowledge
> +		 * error report
> +		 */
> +		u32 ack_with_err_3: 1;
> +
> +		/* LP Transmit Sync error from the Acknowledge error report */
> +		u32 ack_with_err_4: 1;
> +
> +		/* Peripheral Timeout error from the Acknowledge error report */
> +		u32 ack_with_err_5: 1;
> +
> +		/* False Control error from the Acknowledge error report */
> +		u32 ack_with_err_6: 1;
> +
> +		/* reserved (specific to device) from the Acknowledge error
> +		 * report
> +		 */
> +		u32 ack_with_err_7: 1;
> +
> +		/* ECC error, single-bit (detected and corrected) from the
> +		 * Acknowledge error report
> +		 */
> +		u32 ack_with_err_8: 1;
> +
> +		/* ECC error, multi-bit (detected, not corrected) from the
> +		 * Acknowledge error report
> +		 */
> +		u32 ack_with_err_9: 1;
> +
> +		/* checksum error (long packet only) from the Acknowledge
> +		 * error report
> +		 */
> +		u32 ack_with_err_10: 1;
> +
> +		/* not recognized DSI data type from the Acknowledge error
> +		 * report
> +		 */
> +		u32 ack_with_err_11: 1;
> +
> +		/* DSI VC ID Invalid from the Acknowledge error report */
> +		u32 ack_with_err_12: 1;
> +
> +		/* invalid transmission length from the Acknowledge error
> +		 * report
> +		 */
> +		u32 ack_with_err_13: 1;
> +
> +		/* reserved (specific to device) from the Acknowledge error
> +		 * report
> +		 */
> +		u32 ack_with_err_14: 1;
> +
> +		/* DSI protocol violation from the Acknowledge error report */
> +		u32 ack_with_err_15: 1;
> +
> +		} bits;
> +	} PROTOCOL_INT_STS;
> +
> +	union _0x0C {
> +		u32 val;
> +		struct _MASK_PROTOCOL_INT {
> +		u32 mask_dphy_errors_0: 1;
> +		u32 mask_dphy_errors_1: 1;
> +		u32 mask_dphy_errors_2: 1;
> +		u32 mask_dphy_errors_3: 1;
> +		u32 mask_dphy_errors_4: 1;
> +		u32 mask_protocol_debug_err: 11;
> +		u32 mask_ack_with_err_0: 1;
> +		u32 mask_ack_with_err_1: 1;
> +		u32 mask_ack_with_err_2: 1;
> +		u32 mask_ack_with_err_3: 1;
> +		u32 mask_ack_with_err_4: 1;
> +		u32 mask_ack_with_err_5: 1;
> +		u32 mask_ack_with_err_6: 1;
> +		u32 mask_ack_with_err_7: 1;
> +		u32 mask_ack_with_err_8: 1;
> +		u32 mask_ack_with_err_9: 1;
> +		u32 mask_ack_with_err_10: 1;
> +		u32 mask_ack_with_err_11: 1;
> +		u32 mask_ack_with_err_12: 1;
> +		u32 mask_ack_with_err_13: 1;
> +		u32 mask_ack_with_err_14: 1;
> +		u32 mask_ack_with_err_15: 1;
> +		} bits;
> +	} MASK_PROTOCOL_INT;
> +
> +	union _0x10 {
> +		u32 val;
> +		struct _INTERNAL_INT_STS {
> +		/* This bit indicates that the packet size error is detected
> +		 * during the packet reception.
> +		 */
> +		u32 receive_pkt_size_err: 1;
> +
> +		/* This bit indicates that the EoTp packet is not received at
> +		 * the end of the incoming peripheral transmission
> +		 */
> +		u32 eotp_not_receive_err: 1;
> +
> +		/* This bit indicates that the system tried to write a command
> +		 * through the Generic interface and the FIFO is full. There-
> +		 * fore, the command is not written.
> +		 */
> +		u32 gen_cmd_cmd_fifo_wr_err: 1;
> +
> +		/* This bit indicates that during a DCS read data, the payload
> +		 * FIFO becomes	empty and the data sent to the interface is
> +		 * corrupted.
> +		 */
> +		u32 gen_cmd_rdata_fifo_rd_err: 1;
> +
> +		/* This bit indicates that during a generic interface packet
> +		 * read back, the payload FIFO becomes full and the received
> +		 * data is corrupted.
> +		 */
> +		u32 gen_cmd_rdata_fifo_wr_err: 1;
> +
> +		/* This bit indicates that the system tried to write a payload
> +		 * data through the Generic interface and the FIFO is full.
> +		 * Therefore, the payload is not written.
> +		 */
> +		u32 gen_cmd_wdata_fifo_wr_err: 1;
> +
> +		/* This bit indicates that during a Generic interface packet
> +		 * build, the payload FIFO becomes empty and corrupt data is
> +		 * sent.
> +		 */
> +		u32 gen_cmd_wdata_fifo_rd_err: 1;
> +
> +		/* This bit indicates that during a DPI pixel line storage,
> +		 * the payload FIFO becomes full and the data stored is
> +		 * corrupted.
> +		 */
> +		u32 dpi_pix_fifo_wr_err: 1;
> +
> +		/* internal debug error	*/
> +		u32 internal_debug_err: 19;
> +
> +		/* This bit indicates that the ECC single error is detected
> +		 * and corrected in a received packet.
> +		 */
> +		u32 ecc_single_err: 1;
> +
> +		/* This bit indicates that the ECC multiple error is detected
> +		 * in a received packet.
> +		 */
> +		u32 ecc_multi_err: 1;
> +
> +		/* This bit indicates that the CRC error is detected in the
> +		 * received packet payload.
> +		 */
> +		u32 crc_err: 1;
> +
> +		/* This bit indicates that the high-speed transmission timeout
> +		 * counter reached the end and contention is detected.
> +		 */
> +		u32 hs_tx_timeout: 1;
> +
> +		/* This bit indicates that the low-power reception timeout
> +		 * counter reached the end and contention is detected.
> +		 */
> +		u32 lp_rx_timeout: 1;
> +
> +		} bits;
> +	} INTERNAL_INT_STS;
> +
> +	union _0x14 {
> +		u32 val;
> +		struct _MASK_INTERNAL_INT {
> +		u32 mask_receive_pkt_size_err: 1;
> +		u32 mask_eopt_not_receive_err: 1;
> +		u32 mask_gen_cmd_cmd_fifo_wr_err: 1;
> +		u32 mask_gen_cmd_rdata_fifo_rd_err: 1;
> +		u32 mask_gen_cmd_rdata_fifo_wr_err: 1;
> +		u32 mask_gen_cmd_wdata_fifo_wr_err: 1;
> +		u32 mask_gen_cmd_wdata_fifo_rd_err: 1;
> +		u32 mask_dpi_pix_fifo_wr_err: 1;
> +		u32 mask_internal_debug_err: 19;
> +		u32 mask_ecc_single_err: 1;
> +		u32 mask_ecc_multi_err: 1;
> +		u32 mask_crc_err: 1;
> +		u32 mask_hs_tx_timeout: 1;
> +		u32 mask_lp_rx_timeout: 1;
> +		} bits;
> +	} MASK_INTERNAL_INT;
> +
> +	union _0x18 {
> +		u32 val;
> +		struct _DSI_MODE_CFG {
> +		/* This bit configures the operation mode
> +		 * 0: Video mode ;   1: Command mode
> +		 */
> +		u32 cmd_video_mode: 1;
> +
> +		u32 reserved: 31;
> +
> +		} bits;
> +	} DSI_MODE_CFG;
> +
> +	union _0x1C {
> +		u32 val;
> +		struct _VIRTUAL_CHANNEL_ID {
> +		/* This field indicates the Generic interface read-back
> +		 * virtual channel identification
> +		 */
> +		u32 gen_rx_vcid: 2;
> +
> +		/* This field configures the DPI virtual channel id that
> +		 * is indexed to the VIDEO mode packets
> +		 */
> +		u32 video_pkt_vcid: 2;
> +
> +		u32 reserved: 28;
> +
> +		} bits;
> +	} VIRTUAL_CHANNEL_ID;
> +
> +	union _0x20 {
> +		u32 val;
> +		struct _DPI_VIDEO_FORMAT {
> +		/*
> +		 * This field configures the DPI color coding as follows:
> +		 * 0000: 16-bit configuration 1
> +		 * 0001: 16-bit configuration 2
> +		 * 0010: 16-bit configuration 3
> +		 * 0011: 18-bit configuration 1
> +		 * 0100: 18-bit configuration 2
> +		 * 0101: 24-bit
> +		 * 0110: 20-bit YCbCr 4:2:2 loosely packed
> +		 * 0111: 24-bit YCbCr 4:2:2
> +		 * 1000: 16-bit YCbCr 4:2:2
> +		 * 1001: 30-bit
> +		 * 1010: 36-bit
> +		 * 1011: 12-bit YCbCr 4:2:0
> +		 * 1100: Compression Display Stream
> +		 * 1101-1111: 12-bit YCbCr 4:2:0
> +		 */
> +		u32 dpi_video_mode_format: 6;
> +
> +		/* When set to 1, this bit activates loosely packed
> +		 * variant to 18-bit configurations
> +		 */
> +		u32 loosely18_en: 1;
> +
> +		u32 reserved: 25;
> +
> +		} bits;
> +	} DPI_VIDEO_FORMAT;
> +
> +	union _0x24 {
> +		u32 val;
> +		struct _VIDEO_PKT_CONFIG {
> +		/*
> +		 * This field configures the number of pixels in a single
> +		 * video packet. For 18-bit not loosely packed data types,
> +		 * this number must be a multiple of 4. For YCbCr data
> +		 * types, it must be a multiple of 2, as described in the
> +		 * DSI specification.
> +		 */
> +		u32 video_pkt_size: 16;
> +
> +		/*
> +		 * This register configures the number of chunks to be
> +		 * transmitted during a Line period (a chunk consists of
> +		 * a video packet and a null packet). If set to 0 or 1,
> +		 * the video line is transmitted in a single packet. If
> +		 * set to 1, the packet is part of a chunk, so a null packet
> +		 * follows it if vid_null_size > 0. Otherwise, multiple chunks
> +		 * are used to transmit each video line.
> +		 */
> +		u32 video_line_chunk_num: 16;
> +
> +		} bits;
> +	} VIDEO_PKT_CONFIG;
> +
> +	union _0x28 {
> +		u32 val;
> +		struct _VIDEO_LINE_HBLK_TIME {
> +		/* This field configures the Horizontal Back Porch period
> +		 * in lane byte clock cycles
> +		 */
> +		u32 video_line_hbp_time: 16;
> +
> +		/* This field configures the Horizontal Synchronism Active
> +		 * period in lane byte clock cycles
> +		 */
> +		u32 video_line_hsa_time: 16;
> +
> +		} bits;
> +	} VIDEO_LINE_HBLK_TIME;
> +
> +	union _0x2C {
> +		u32 val;
> +		struct _VIDEO_LINE_TIME {
> +		/* This field configures the size of the total line time
> +		 * (HSA+HBP+HACT+HFP) counted in lane byte clock cycles
> +		 */
> +		u32 video_line_time: 16;
> +
> +		u32 reserved: 16;
> +
> +		} bits;
> +	} VIDEO_LINE_TIME;
> +
> +	union _0x30 {
> +		u32 val;
> +		struct _VIDEO_VBLK_LINES {
> +		/* This field configures the Vertical Front Porch period
> +		 * measured in number of horizontal lines
> +		 */
> +		u32 vfp_lines: 10;
> +
> +		/* This field configures the Vertical Back Porch period
> +		 * measured in number of horizontal lines
> +		 */
> +		u32 vbp_lines: 10;
> +
> +		/* This field configures the Vertical Synchronism Active
> +		 * period measured in number of horizontal lines
> +		 */
> +		u32 vsa_lines: 10;
> +
> +		u32 reserved: 2;
> +
> +		} bits;
> +	} VIDEO_VBLK_LINES;
> +
> +	union _0x34 {
> +		u32 val;
> +		struct _VIDEO_VACTIVE_LINES {
> +		/* This field configures the Vertical Active period measured
> +		 * in number of horizontal lines
> +		 */
> +		u32 vactive_lines: 14;
> +
> +		u32 reserved: 18;
> +
> +		} bits;
> +	} VIDEO_VACTIVE_LINES;
> +
> +	union _0x38 {
> +		u32 val;
> +		struct _VID_MODE_CFG {
> +		/*
> +		 * This field indicates the video mode transmission type as
> +		 * follows:
> +		 * 00: Non-burst with sync pulses
> +		 * 01: Non-burst with sync events
> +		 * 10 and 11: Burst mode
> +		 */
> +		u32 vid_mode_type: 2;
> +
> +		u32 reserved_0: 6;
> +
> +		/* When set to 1, this bit enables the return to low-power
> +		 * inside the VSA period when timing allows.
> +		 */
> +		u32 lp_vsa_en: 1;
> +
> +		/* When set to 1, this bit enables the return to low-power
> +		 * inside the VBP period when timing allows.
> +		 */
> +		u32 lp_vbp_en: 1;
> +
> +		/* When set to 1, this bit enables the return to low-power
> +		 * inside the VFP period when timing allows.
> +		 */
> +		u32 lp_vfp_en: 1;
> +
> +		/* When set to 1, this bit enables the return to low-power
> +		 * inside the VACT period when timing allows.
> +		 */
> +		u32 lp_vact_en: 1;
> +
> +		/* When set to 1, this bit enables the return to low-power
> +		 * inside the HBP period when timing allows.
> +		 */
> +		u32 lp_hbp_en: 1;
> +
> +		/* When set to 1, this bit enables the return to low-power
> +		 * inside the HFP period when timing allows.
> +		 */
> +		u32 lp_hfp_en: 1;
> +
> +		/* When set to 1, this bit enables the request for an ack-
> +		 * nowledge response at the end of a frame.
> +		 */
> +		u32 frame_bta_ack_en: 1;
> +
> +		/* When set to 1, this bit enables the command transmission
> +		 * only in low-power mode.
> +		 */
> +		u32 lp_cmd_en: 1;
> +
> +		u32 reserved_1: 16;
> +
> +		} bits;
> +	} VID_MODE_CFG;
> +
> +	union _0x3C {
> +		u32 val;
> +		struct _SDF_MODE_CONFIG {
> +		/*
> +		 * This field defines the 3D mode on/off & display orientation:
> +		 * 00: 3D mode off (2D mode on)
> +		 * 01: 3D mode on, portrait orientation
> +		 * 10: 3D mode on, landscape orientation
> +		 * 11: Reserved
> +		 */
> +		u32 rf_3d_mode: 2;
> +
> +		/*
> +		 * This field defines the 3D image format:
> +		 * 00: Line (alternating lines of left and right data)
> +		 * 01: Frame (alternating frames of left and right data)
> +		 * 10: Pixel (alternating pixels of left and right data)
> +		 * 11: Reserved
> +		 */
> +		u32 rf_3d_format: 2;
> +
> +		/*
> +		 * This field defines whether there is a second VSYNC pulse
> +		 * between Left and Right Images, when 3D Image Format is
> +		 * Frame-based:
> +		 * 0: No sync pulses between left and right data
> +		 * 1: Sync pulse (HSYNC, VSYNC, blanking) between left and
> +		 *    right data
> +		 */
> +		u32 second_vsync_en: 1;
> +
> +		/*
> +		 * This bit defines the left or right order:
> +		 * 0: Left eye data is sent first, and then the right eye data
> +		 *    is sent.
> +		 * 1: Right eye data is sent first, and then the left eye data
> +		 *    is sent.
> +		 */
> +		u32 left_right_order: 1;
> +
> +		u32 reserved_0: 2;
> +
> +		/*
> +		 * When set, causes the next VSS packet to include 3D control
> +		 * payload in every VSS packet.
> +		 */
> +		u32 rf_3d_payload_en: 1;
> +
> +		u32 reserved_1: 23;
> +
> +		} bits;
> +	} SDF_MODE_CONFIG;
> +
> +	union _0x40 {
> +		u32 val;
> +		struct _TIMEOUT_CNT_CLK_CONFIG {
> +		/*
> +		 * This field indicates the division factor for the Time Out
> +		 * clock used as the timing unit in the configuration of HS to
> +		 * LP and LP to HS transition error.
> +		 */
> +		u32 timeout_cnt_clk_config: 16;
> +
> +		u32 reserved: 16;
> +
> +		} bits;
> +	} TIMEOUT_CNT_CLK_CONFIG;
> +
> +	union _0x44 {
> +		u32 val;
> +		struct _HTX_TO_CONFIG {
> +		/*
> +		 * This field configures the timeout counter that triggers
> +		 * a high speed transmission timeout contention detection
> +		 * (measured in TO_CLK_DIVISION cycles).
> +		 *
> +		 * If using the non-burst mode and there is no sufficient
> +		 * time to switch from HS to LP and back in the period which
> +		 * is from one line data finishing to the next line sync
> +		 * start, the DSI link returns the LP state once per frame,
> +		 * then you should configure the TO_CLK_DIVISION and
> +		 * hstx_to_cnt to be in accordance with:
> +		 * hstx_to_cnt * lanebyteclkperiod * TO_CLK_DIVISION >= the
> +		 * time of one FRAME data transmission * (1 + 10%)
> +		 *
> +		 * In burst mode, RGB pixel packets are time-compressed,
> +		 * leaving more time during a scan line. Therefore, if in
> +		 * burst mode and there is sufficient time to switch from HS
> +		 * to LP and back in the period of time from one line data
> +		 * finishing to the next line sync start, the DSI link can
> +		 * return LP mode and back in this time interval to save power.
> +		 * For this, configure the TO_CLK_DIVISION and hstx_to_cnt
> +		 * to be in accordance with:
> +		 * hstx_to_cnt * lanebyteclkperiod * TO_CLK_DIVISION >= the
> +		 * time of one LINE data transmission * (1 + 10%)
> +		 */
> +		u32 htx_to_cnt_limit: 32;
> +		} bits;
> +	} HTX_TO_CONFIG;
> +
> +	union _0x48 {
> +		u32 val;
> +		struct _LRX_H_TO_CONFIG {
> +		/*
> +		 * This field configures the timeout counter that triggers
> +		 * a low-power reception timeout contention detection (measured
> +		 * in TO_CLK_DIVISION cycles).
> +		 */
> +		u32 lrx_h_to_cnt_limit: 32;
> +		} bits;
> +	} LRX_H_TO_CONFIG;
> +
> +	union _0x4C {
> +		u32 val;
> +		struct _RD_PRESP_TO_CONFIG {
> +		/*
> +		 * This field sets a period for which the DWC_mipi_dsi_host
> +		 * keeps the link still, after sending a low-power read oper-
> +		 * ation. This period is measured in cycles of lanebyteclk.
> +		 * The counting starts when the D-PHY enters the Stop state
> +		 * and causes no interrupts.
> +		 */
> +		u32 lprd_presp_to_cnt_limit: 16;
> +
> +		/*
> +		 * This field sets a period for which the DWC_mipi_dsi_host
> +		 * keeps the link still, after sending a high-speed read oper-
> +		 * ation. This period is measured in cycles of lanebyteclk.
> +		 * The counting starts when the D-PHY enters the Stop state
> +		 * and causes no interrupts.
> +		 */
> +		u32 hsrd_presp_to_cnt_limit: 16;
> +
> +		} bits;
> +	} RD_PRESP_TO_CONFIG;
> +
> +	union _0x50 {
> +		u32 val;
> +		struct _HSWR_PRESP_TO_CONFIG {
> +		/*
> +		 * This field sets a period for which the DWC_mipi_dsi_host
> +		 * keeps the link inactive after sending a high-speed write
> +		 * operation. This period is measured in cycles of lanebyteclk.
> +		 * The counting starts when the D-PHY enters the Stop state
> +		 * and causes no interrupts.
> +		 */
> +		u32 hswr_presp_to_cnt_limit: 16;
> +
> +		u32 reserved_0: 8;
> +
> +		/*
> +		 * When set to 1, this bit ensures that the peripheral response
> +		 * timeout caused by hs_wr_to_cnt is used only once per eDPI
> +		 * frame, when both the following conditions are met:
> +		 * dpivsync_edpiwms has risen and fallen.
> +		 * Packets originated from eDPI have been transmitted and its
> +		 * FIFO is empty again In this scenario no non-eDPI requests
> +		 * are sent to the D-PHY, even if there is traffic from generic
> +		 * or DBI ready to be sent, making it return to stop state.
> +		 * When it does so, PRESP_TO counter is activated and only when
> +		 * it finishes does the controller send any other traffic that
> +		 * is ready.
> +		 */
> +		u32 hswr_presp_to_mode: 1;
> +
> +		u32 reserved_1: 7;
> +
> +		} bits;
> +	} HSWR_PRESP_TO_CONFIG;
> +
> +	union _0x54 {
> +		u32 val;
> +		struct _LPWR_PRESP_TO_CONFIG {
> +		/*
> +		 * This field sets a period for which the DWC_mipi_dsi_host
> +		 * keeps the link still, after sending a low-power write oper-
> +		 * ation. This period is measured in cycles of lanebyteclk.
> +		 * The counting starts when the D-PHY enters the Stop state
> +		 * and causes no interrupts.
> +		 */
> +		u32 lpwr_presp_to_cnt_limit: 16;
> +
> +		u32 reserved: 16;
> +
> +		} bits;
> +	} LPWR_PRESP_TO_CONFIG;
> +
> +	union _0x58 {
> +		u32 val;
> +		struct _BTA_PRESP_TO_CONFIG {
> +		/*
> +		 * This field sets a period for which the DWC_mipi_dsi_host
> +		 * keeps the link still, after completing a Bus Turn-Around.
> +		 * This period is measured in cycles of lanebyteclk. The
> +		 * counting starts when the D-PHY enters the Stop state and
> +		 * causes no interrupts.
> +		 */
> +		u32 bta_presp_to_cnt_limit: 16;
> +
> +		u32 reserved: 16;
> +
> +		} bits;
> +	} BTA_PRESP_TO_CONFIG;
> +
> +	union _0x5C {
> +		u32 val;
> +		struct _TX_ESC_CLK_CONFIG {
> +		/*
> +		 * This field indicates the division factor for the TX Escape
> +		 * clock source (lanebyteclk). The values 0 and 1 stop the
> +		 * TX_ESC clock generation.
> +		 */
> +		u32 tx_esc_clk_config: 16;
> +
> +		u32 reserved: 16;
> +
> +		} bits;
> +	} TX_ESC_CLK_CONFIG;
> +
> +	union _0x60 {
> +		u32 val;
> +		struct _VACT_CMD_TRANS_LIMIT {
> +		/*
> +		 * This field is used for the transmission of commands in
> +		 * low-power mode. It defines the size, in bytes, of the
> +		 * largest packet that can fit in a line during the VACT
> +		 * region.
> +		 */
> +		u32 vact_cmd_trans_limit: 8;
> +
> +		u32 reserved: 24;
> +
> +		} bits;
> +	} VACT_CMD_TRANS_LIMIT;
> +
> +	union _0x64 {
> +		u32 val;
> +		struct _VBLK_CMD_TRANS_LIMIT {
> +		/*
> +		 * This field is used for the transmission of commands in
> +		 * low-power mode. It defines the size, in bytes, of the
> +		 * largest packet that can fit in a line during the VSA, VBP,
> +		 * and VFP regions.
> +		 */
> +		u32 vblk_cmd_trans_limit: 8;
> +
> +		u32 reserved: 24;
> +
> +		} bits;
> +	} VBLK_CMD_TRANS_LIMIT;
> +
> +	union _0x68 {
> +		u32 val;
> +		struct _CMD_MODE_CFG {
> +		/*
> +		 * When set to 1, this bit enables the tearing effect
> +		 * acknowledge request.
> +		 */
> +		u32 tear_fx_en: 1;
> +
> +		/*
> +		 * When set to 1, this bit enables the acknowledge request
> +		 * after each packet transmission.
> +		 */
> +		u32 ack_rqst_en: 1;
> +
> +		u32 reserved_0: 3;
> +
> +		u32 pps_tx: 1;
> +		u32 exq_tx: 1;
> +		u32 cmc_tx: 1;
> +
> +		/*
> +		 * This bit configures the Generic short write packet with
> +		 * zero parameter command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 gen_sw_0p_tx: 1;
> +
> +		/*
> +		 * This bit configures the Generic short write packet with
> +		 * one parameter command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 gen_sw_1p_tx: 1;
> +
> +		/*
> +		 * This bit configures the Generic short write packet with
> +		 * two parameters command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 gen_sw_2p_tx: 1;
> +
> +		/*
> +		 * This bit configures the Generic short read packet with
> +		 * zero parameter command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 gen_sr_0p_tx: 1;
> +
> +		/*
> +		 * This bit configures the Generic short read packet with
> +		 * one parameter command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 gen_sr_1p_tx: 1;
> +
> +		/*
> +		 * This bit configures the Generic short read packet with
> +		 * two parameters command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 gen_sr_2p_tx: 1;
> +
> +		/*
> +		 * This bit configures the Generic long write packet command
> +		 * transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 gen_lw_tx: 1;
> +
> +		u32 reserved_1: 1;
> +
> +		/*
> +		 * This bit configures the DCS short write packet with zero
> +		 * parameter command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 dcs_sw_0p_tx: 1;
> +
> +		/*
> +		 * This bit configures the DCS short write packet with one
> +		 * parameter command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 dcs_sw_1p_tx: 1;
> +
> +		/*
> +		 * This bit configures the DCS short read packet with zero
> +		 * parameter command transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 dcs_sr_0p_tx: 1;
> +
> +		/*
> +		 * This bit configures the DCS long write packet command
> +		 * transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 dcs_lw_tx: 1;
> +
> +		u32 reserved_2: 4;
> +
> +		/*
> +		 * This bit configures the maximum read packet size command
> +		 * transmission type:
> +		 * 0: High-speed 1: Low-power
> +		 */
> +		u32 max_rd_pkt_size: 1;
> +
> +		u32 reserved_3: 7;
> +
> +		} bits;
> +	} CMD_MODE_CFG;
> +
> +	union _0x6C {
> +		u32 val;
> +		struct _GEN_HDR {
> +		/*
> +		 * This field configures the packet data type of the header
> +		 * packet.
> +		 */
> +		u32 gen_dt: 6;
> +
> +		/*
> +		 * This field configures the virtual channel id of the header
> +		 * packet.
> +		 */
> +		u32 gen_vc: 2;
> +
> +		/*
> +		 * This field configures the least significant byte of the
> +		 * header packet's Word count for long packets or data 0 for
> +		 * short packets.
> +		 */
> +		u32 gen_wc_lsbyte: 8;
> +
> +		/*
> +		 * This field configures the most significant byte of the
> +		 * header packet's word count for long packets or data 1 for
> +		 * short packets.
> +		 */
> +		u32 gen_wc_msbyte: 8;
> +
> +		u32 reserved: 8;
> +
> +		} bits;
> +	} GEN_HDR;
> +
> +	union _0x70 {
> +		u32 val;
> +		struct _GEN_PLD_DATA {
> +		/* This field indicates byte 1 of the packet payload. */
> +		u32 gen_pld_b1: 8;
> +
> +		/* This field indicates byte 2 of the packet payload. */
> +		u32 gen_pld_b2: 8;
> +
> +		/* This field indicates byte 3 of the packet payload. */
> +		u32 gen_pld_b3: 8;
> +
> +		/* This field indicates byte 4 of the packet payload. */
> +		u32 gen_pld_b4: 8;
> +
> +		} bits;
> +	} GEN_PLD_DATA;
> +
> +	union _0x74 {
> +		u32 val;
> +		struct _PHY_CLK_LANE_LP_CTRL {
> +		/* This bit controls the D-PHY PPI txrequestclkhs signal */
> +		u32 phy_clklane_tx_req_hs: 1;
> +
> +		/* This bit enables the automatic mechanism to stop providing
> +		 * clock in the clock lane when time allows.
> +		 */
> +		u32 auto_clklane_ctrl_en: 1;
> +
> +		u32 reserved: 30;
> +		} bits;
> +	} PHY_CLK_LANE_LP_CTRL;
> +
> +	union _0x78 {
> +		u32 val;
> +		struct _PHY_INTERFACE_CTRL {
> +		/* When set to 0, this bit places the D-PHY macro in power-
> +		 * down state.
> +		 */
> +		u32 rf_phy_shutdown: 1;
> +
> +		/* When set to 0, this bit places the digital section of the
> +		 * D-PHY in the reset state.
> +		 */
> +		u32 rf_phy_reset_n: 1;
> +
> +		/* When set to 1, this bit enables the D-PHY Clock Lane
> +		 * module.
> +		 */
> +		u32 rf_phy_clk_en: 1;
> +
> +		/* When the D-PHY is in ULPS, this bit enables the D-PHY PLL. */
> +		u32 rf_phy_force_pll: 1;
> +
> +		/* ULPS mode Request on clock lane */
> +		u32 rf_phy_clk_txrequlps: 1;
> +
> +		/* ULPS mode Exit on clock lane */
> +		u32 rf_phy_clk_txexitulps: 1;
> +
> +		/* ULPS mode Request on all active data lanes */
> +		u32 rf_phy_data_txrequlps: 1;
> +
> +		/* ULPS mode Exit on all active data lanes */
> +		u32 rf_phy_data_txexitulps: 1;
> +
> +		u32 reserved: 24;
> +		} bits;
> +	} PHY_INTERFACE_CTRL;
> +
> +	union _0x7C {
> +		u32 val;
> +		struct _PHY_TX_TRIGGERS {
> +		/* This field controls the trigger transmissions. */
> +		u32 phy_tx_triggers: 4;
> +
> +		u32 reserved: 28;
> +		} bits;
> +	} PHY_TX_TRIGGERS;
> +
> +	union _0x80 {
> +		u32 val;
> +		struct _DESKEW_START {
> +		u32 deskew_start: 1;
> +		u32 reserved: 31;
> +		} bits;
> +	} DESKEW_START;
> +
> +	union _0x84 {
> +		u32 val;
> +		struct _DESKEW_MODE {
> +		u32 deskew_mode: 2;
> +		u32 reserved: 30;
> +		} bits;
> +	} DESKEW_MODE;
> +
> +	union _0x88 {
> +		u32 val;
> +		struct _DESKEW_TIME {
> +		u32 deskew_time: 32;
> +		} bits;
> +	} DESKEW_TIME;
> +
> +	union _0x8C {
> +		u32 val;
> +		struct _DESKEW_PERIOD {
> +		u32 deskew_period: 32;
> +		} bits;
> +	} DESKEW_PERIOD;
> +
> +	union _0x90 {
> +		u32 val;
> +		struct _DESKEW_BUSY {
> +		u32 deskew_busy: 1;
> +		u32 reserved: 31;
> +		} bits;
> +	} DESKEW_BUSY;
> +
> +	union _0x94 {
> +		u32 val;
> +		struct _DESKEW_LANE_MASK {
> +		u32 deskew_lane0_mask: 1;
> +		u32 deskew_lane1_mask: 1;
> +		u32 deskew_lane2_mask: 1;
> +		u32 deskew_lane3_mask: 1;
> +		u32 reserved: 28;
> +		} bits;
> +	} DESKEW_LANE_MASK;
> +
> +	union _0x98 {
> +		u32 val;
> +		struct _CMD_MODE_STATUS {
> +		/*
> +		 * This bit is set when a read command is issued and cleared
> +		 * when the entire response is stored in the FIFO.
> +		 * Value after reset: 0x0
> +		 *
> +		 * NOTE:
> +		 * For mipi-dsi-r1p0 IP, this bit is set immediately when
> +		 *     the read cmd is set to the GEN_HDR register.
> +		 *
> +		 * For dsi-ctrl-r1p0 IP, this bit is set only after the read
> +		 *     cmd was actually sent out from the controller.
> +		 */
> +		u32 gen_cmd_rdcmd_ongoing: 1;
> +
> +		/*
> +		 * This bit indicates the empty status of the generic read
> +		 * payload FIFO.
> +		 * Value after reset: 0x1
> +		 */
> +		u32 gen_cmd_rdata_fifo_empty: 1;
> +
> +		/*
> +		 * This bit indicates the full status of the generic read
> +		 * payload FIFO.
> +		 * Value after reset: 0x0
> +		 */
> +		u32 gen_cmd_rdata_fifo_full: 1;
> +
> +		/*
> +		 * This bit indicates the empty status of the generic write
> +		 * payload FIFO.
> +		 * Value after reset: 0x1
> +		 */
> +		u32 gen_cmd_wdata_fifo_empty: 1;
> +
> +		/*
> +		 * This bit indicates the full status of the generic write
> +		 * payload FIFO.
> +		 * Value after reset: 0x0
> +		 */
> +		u32 gen_cmd_wdata_fifo_full: 1;
> +
> +		/*
> +		 * This bit indicates the empty status of the generic
> +		 * command FIFO.
> +		 * Value after reset: 0x1
> +		 */
> +		u32 gen_cmd_cmd_fifo_empty: 1;
> +
> +		/*
> +		 * This bit indicates the full status of the generic
> +		 * command FIFO.
> +		 * Value after reset: 0x0
> +		 */
> +		u32 gen_cmd_cmd_fifo_full: 1;
> +
> +		/*
> +		 * This bit is set when the entire response of read is
> +		 * stored in the rx payload FIFO. And it will be cleared
> +		 * automaticlly after read this bit each time.
> +		 * Value after reset: 0x0
> +		 *
> +		 * NOTE: this bit is just supported for dsi-ctrl-r1p0 IP
> +		 */
> +		u32 gen_cmd_rdcmd_done: 1;
> +
> +		u32 reserved : 24;
> +
> +		} bits;
> +	} CMD_MODE_STATUS;
> +
> +	union _0x9C {
> +		u32 val;
> +		struct _PHY_STATUS {
> +		/* the status of phydirection D-PHY signal */
> +		u32 phy_direction: 1;
> +
> +		/* the status of phylock D-PHY signal */
> +		u32 phy_lock: 1;
> +
> +		/* the status of rxulpsesc0lane D-PHY signal */
> +		u32 phy_rxulpsesc0lane: 1;
> +
> +		/* the status of phystopstateclklane D-PHY signal */
> +		u32 phy_stopstateclklane: 1;
> +
> +		/* the status of phystopstate0lane D-PHY signal */
> +		u32 phy_stopstate0lane: 1;
> +
> +		/* the status of phystopstate1lane D-PHY signal */
> +		u32 phy_stopstate1lane: 1;
> +
> +		/* the status of phystopstate2lane D-PHY signal */
> +		u32 phy_stopstate2lane: 1;
> +
> +		/* the status of phystopstate3lane D-PHY signal */
> +		u32 phy_stopstate3lane: 1;
> +
> +		/* the status of phyulpsactivenotclk D-PHY signal */
> +		u32 phy_ulpsactivenotclk: 1;
> +
> +		/* the status of ulpsactivenot0lane D-PHY signal */
> +		u32 phy_ulpsactivenot0lane: 1;
> +
> +		/* the status of ulpsactivenot1lane D-PHY signal */
> +		u32 phy_ulpsactivenot1lane: 1;
> +
> +		/* the status of ulpsactivenot2lane D-PHY signal */
> +		u32 phy_ulpsactivenot2lane: 1;
> +
> +		/* the status of ulpsactivenot3lane D-PHY signal */
> +		u32 phy_ulpsactivenot3lane: 1;
> +
> +		u32 reserved: 19;
> +
> +		} bits;
> +	} PHY_STATUS;
> +
> +	union _0xA0 {
> +		u32 val;
> +		struct _PHY_MIN_STOP_TIME {
> +		/* This field configures the minimum wait period to request
> +		 * a high-speed transmission after the Stop state.
> +		 */
> +		u32 phy_min_stop_time: 8;
> +
> +		u32 reserved: 24;
> +		} bits;
> +	} PHY_MIN_STOP_TIME;
> +
> +	union _0xA4 {
> +		u32 val;
> +		struct _PHY_LANE_NUM_CONFIG {
> +		/*
> +		 * This field configures the number of active data lanes:
> +		 * 00: One data lane (lane 0)
> +		 * 01: Two data lanes (lanes 0 and 1)
> +		 * 10: Three data lanes (lanes 0, 1, and 2)
> +		 * 11: Four data lanes (lanes 0, 1, 2, and 3)
> +		 */
> +		u32 phy_lane_num: 2;
> +
> +		u32 reserved: 30;
> +
> +		} bits;
> +	} PHY_LANE_NUM_CONFIG;
> +
> +	union _0xA8 {
> +		u32 val;
> +		struct _PHY_CLKLANE_TIME_CONFIG {
> +		/*
> +		 * This field configures the maximum time that the D-PHY
> +		 * clock lane takes to go from low-power to high-speed
> +		 * transmission measured in lane byte clock cycles.
> +		 */
> +		u32 phy_clklane_lp_to_hs_time: 16;
> +
> +		/*
> +		 * This field configures the maximum time that the D-PHY
> +		 * clock lane takes to go from high-speed to low-power
> +		 * transmission measured in lane byte clock cycles.
> +		 */
> +		u32 phy_clklane_hs_to_lp_time: 16;
> +
> +		} bits;
> +	} PHY_CLKLANE_TIME_CONFIG;
> +
> +	union _0xAC {
> +		u32 val;
> +		struct _PHY_DATALANE_TIME_CONFIG {
> +		/*
> +		 * This field configures the maximum time that the D-PHY data
> +		 * lanes take to go from low-power to high-speed transmission
> +		 * measured in lane byte clock cycles.
> +		 */
> +		u32 phy_datalane_lp_to_hs_time: 16;
> +
> +		/*
> +		 * This field configures the maximum time that the D-PHY data
> +		 * lanes take to go from high-speed to low-power transmission
> +		 * measured in lane byte clock cycles.
> +		 */
> +		u32 phy_datalane_hs_to_lp_time: 16;
> +
> +		} bits;
> +	} PHY_DATALANE_TIME_CONFIG;
> +
> +	union _0xB0 {
> +		u32 val;
> +		struct _MAX_READ_TIME {
> +		/*
> +		 * This field configures the maximum time required to perform
> +		 * a read command in lane byte clock cycles. This register can
> +		 * only be modified when no read command is in progress.
> +		 */
> +		u32 max_rd_time: 16;
> +
> +		u32 reserved: 16;
> +
> +		} bits;
> +	} MAX_READ_TIME;
> +
> +	union _0xB4 {
> +		u32 val;
> +		struct _RX_PKT_CHECK_CONFIG {
> +		/* When set to 1, this bit enables the ECC reception, error
> +		 * correction, and reporting.
> +		 */
> +		u32 rx_pkt_ecc_en: 1;
> +
> +		/* When set to 1, this bit enables the CRC reception and error
> +		 * reporting.
> +		 */
> +		u32 rx_pkt_crc_en: 1;
> +
> +		u32 reserved: 30;
> +
> +		} bits;
> +	} RX_PKT_CHECK_CONFIG;
> +
> +	union _0xB8 {
> +		u32 val;
> +		struct _TA_EN {
> +		/* When set to 1, this bit enables the Bus Turn-Around (BTA)
> +		 * request.
> +		 */
> +		u32 ta_en: 1;
> +
> +		u32 reserved: 31;
> +
> +		} bits;
> +	} TA_EN;
> +
> +	union _0xBC {
> +		u32 val;
> +		struct _EOTP_EN {
> +		/* When set to 1, this bit enables the EoTp transmission */
> +		u32 tx_eotp_en: 1;
> +
> +		/* When set to 1, this bit enables the EoTp reception. */
> +		u32 rx_eotp_en: 1;
> +
> +		u32 reserved: 30;
> +
> +		} bits;
> +	} EOTP_EN;
> +
> +	union _0xC0 {
> +		u32 val;
> +		struct _VIDEO_NULLPKT_SIZE {
> +		/*
> +		 * This register configures the number of bytes inside a null
> +		 * packet. Setting it to 0 disables the null packets.
> +		 */
> +		u32 video_nullpkt_size: 13;
> +
> +		u32 reserved: 19;
> +
> +		} bits;
> +	} VIDEO_NULLPKT_SIZE;
> +
> +	union _0xC4 {
> +		u32 val;
> +		struct _DCS_WM_PKT_SIZE {
> +		/*
> +		 * This field configures the maximum allowed size for an eDPI
> +		 * write memory command, measured in pixels. Automatic parti-
> +		 * tioning of data obtained from eDPI is permanently enabled.
> +		 */
> +		u32 dcs_wm_pkt_size: 16;
> +
> +		u32 reserved: 16;
> +		} bits;
> +	} DCS_WM_PKT_SIZE;
> +
> +	union _0xC8 {
> +		u32 val;
> +		struct _PROTOCOL_INT_CLR {
> +		u32 clr_dphy_errors_0: 1;
> +		u32 clr_dphy_errors_1: 1;
> +		u32 clr_dphy_errors_2: 1;
> +		u32 clr_dphy_errors_3: 1;
> +		u32 clr_dphy_errors_4: 1;
> +		u32 clr_protocol_debug_err: 11;
> +		u32 clr_ack_with_err_0: 1;
> +		u32 clr_ack_with_err_1: 1;
> +		u32 clr_ack_with_err_2: 1;
> +		u32 clr_ack_with_err_3: 1;
> +		u32 clr_ack_with_err_4: 1;
> +		u32 clr_ack_with_err_5: 1;
> +		u32 clr_ack_with_err_6: 1;
> +		u32 clr_ack_with_err_7: 1;
> +		u32 clr_ack_with_err_8: 1;
> +		u32 clr_ack_with_err_9: 1;
> +		u32 clr_ack_with_err_10: 1;
> +		u32 clr_ack_with_err_11: 1;
> +		u32 clr_ack_with_err_12: 1;
> +		u32 clr_ack_with_err_13: 1;
> +		u32 clr_ack_with_err_14: 1;
> +		u32 clr_ack_with_err_15: 1;
> +		} bits;
> +	} PROTOCOL_INT_CLR;
> +
> +	union _0xCC {
> +		u32 val;
> +		struct _INTERNAL_INT_CLR {
> +		u32 clr_receive_pkt_size_err: 1;
> +		u32 clr_eopt_not_receive_err: 1;
> +		u32 clr_gen_cmd_cmd_fifo_wr_err: 1;
> +		u32 clr_gen_cmd_rdata_fifo_rd_err: 1;
> +		u32 clr_gen_cmd_rdata_fifo_wr_err: 1;
> +		u32 clr_gen_cmd_wdata_fifo_wr_err: 1;
> +		u32 clr_gen_cmd_wdata_fifo_rd_err: 1;
> +		u32 clr_dpi_pix_fifo_wr_err: 1;
> +		u32 clr_internal_debug_err: 19;
> +		u32 clr_ecc_single_err: 1;
> +		u32 clr_ecc_multi_err: 1;
> +		u32 clr_crc_err: 1;
> +		u32 clr_hs_tx_timeout: 1;
> +		u32 clr_lp_rx_timeout: 1;
> +		} bits;
> +	} INTERNAL_INT_CLR;
> +
> +	union _0xD0 {
> +		u32 val;
> +		struct _VIDEO_SIG_DELAY_CONFIG {
> +
> +		/*
> +		 * DPI interface signal delay to be used in clk lanebyte
> +		 * domain for control logic to read video data from pixel
> +		 * memory in mannal mode, measured in clk_lanebyte cycles
> +		 */
> +		u32 video_sig_delay: 24;
> +
> +		/*
> +		 * 1'b1: mannal mode
> +		 *       dsi controller will use video_sig_delay value as
> +		 *       the delay for the packet handle logic to read video
> +		 *       data from pixel memory.
> +		 *
> +		 * 1'b0: auto mode
> +		 *       dsi controller will auto calculate the delay for
> +		 *       the packet handle logic to read video data from
> +		 *       pixel memory.
> +		 */
> +		u32 video_sig_delay_mode: 1;
> +
> +		u32 reserved: 7;
> +		} bits;
> +	} VIDEO_SIG_DELAY_CONFIG;
> +
> +	u32 reservedD4_EC[7];
> +
> +	union _0xF0 {
> +		u32 val;
> +		struct _PHY_TST_CTRL0 {
> +		/* PHY test interface clear (active high) */
> +		u32 phy_testclr: 1;
> +
> +		/* This bit is used to clock the TESTDIN bus into the D-PHY */
> +		u32 phy_testclk: 1;
> +
> +		u32 reserved: 30;
> +		} bits;
> +	} PHY_TST_CTRL0;
> +
> +	union _0xF4 {
> +		u32 val;
> +		struct _PHY_TST_CTRL1 {
> +		/* PHY test interface input 8-bit data bus for internal
> +		 * register programming and test functionalities access.
> +		 */
> +		u32 phy_testdin: 8;
> +
> +		/* PHY output 8-bit data bus for read-back and internal
> +		 * probing functionalities.
> +		 */
> +		u32 phy_testdout: 8;
> +
> +		/*
> +		 * PHY test interface operation selector:
> +		 * 1: The address write operation is set on the falling edge
> +		 *    of the testclk signal.
> +		 * 0: The data write operation is set on the rising edge of
> +		 *    the testclk signal.
> +		 */
> +		u32 phy_testen: 1;
> +
> +		u32 reserved: 15;
> +		} bits;
> +	} PHY_TST_CTRL1;
> +
> +	u32 reservedF8_1FC[66];
> +
> +	union _0x200 {
> +		u32 val;
> +		struct _INT_PLL_STS {
> +		u32 int_pll_sts: 1;
> +		u32 reserved: 31;
> +		} bits;
> +	} INT_PLL_STS;
> +
> +	union _0x204 {
> +		u32 val;
> +		struct _INT_PLL_MSK {
> +		u32 int_pll_msk: 1;
> +		u32 reserved: 31;
> +		} bits;
> +	} INT_PLL_MSK;
> +
> +	union _0x208 {
> +		u32 val;
> +		struct _INT_PLL_CLR {
> +		u32 int_pll_clr: 1;
> +		u32 reserved: 31;
> +		} bits;
> +	} INT_PLL_CLR;
> +
> +};
> +
> +void dsi_power_enable(struct dsi_context *ctx, int enable);
> +void dsi_video_mode(struct dsi_context *ctx);
> +void dsi_cmd_mode(struct dsi_context *ctx);
> +bool dsi_is_cmd_mode(struct dsi_context *ctx);
> +void dsi_rx_vcid(struct dsi_context *ctx, u8 vc);
> +void dsi_video_vcid(struct dsi_context *ctx, u8 vc);
> +void dsi_dpi_video_burst_mode(struct dsi_context *ctx, int mode);
> +void dsi_dpi_color_coding(struct dsi_context *ctx, int coding);
> +void dsi_dpi_sig_delay(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_dpi_hline_time(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_dpi_hsync_time(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_dpi_hbp_time(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_dpi_vact(struct dsi_context *ctx, u16 lines);
> +void dsi_dpi_vfp(struct dsi_context *ctx, u16 lines);
> +void dsi_dpi_vbp(struct dsi_context *ctx, u16 lines);
> +void dsi_dpi_vsync(struct dsi_context *ctx, u16 lines);
> +void dsi_dpi_hporch_lp_en(struct dsi_context *ctx, int enable);
> +void dsi_dpi_vporch_lp_en(struct dsi_context *ctx, int enable);
> +void dsi_dpi_frame_ack_en(struct dsi_context *ctx, int enable);
> +void dsi_dpi_chunk_num(struct dsi_context *ctx, u16 no);
> +void dsi_dpi_null_packet_size(struct dsi_context *ctx, u16 size);
> +void dsi_dpi_video_packet_size(struct dsi_context *ctx, u16 size);
> +void dsi_edpi_max_pkt_size(struct dsi_context *ctx, u16 size);
> +void dsi_tear_effect_ack_en(struct dsi_context *ctx, int enable);
> +void dsi_cmd_mode_lp_cmd_en(struct dsi_context *ctx, int enable);
> +void dsi_video_mode_lp_cmd_en(struct dsi_context *ctx, int enable);
> +void dsi_set_packet_header(struct dsi_context *ctx, u8 vc, u8 type,
> +						u8 wc_lsb, u8 wc_msb);
> +void dsi_set_packet_payload(struct dsi_context *ctx, u32 payload);
> +u32 dsi_get_rx_payload(struct dsi_context *ctx);
> +void dsi_bta_en(struct dsi_context *ctx, int enable);
> +void dsi_eotp_rx_en(struct dsi_context *ctx, int enable);
> +void dsi_eotp_tx_en(struct dsi_context *ctx, int enable);
> +void dsi_ecc_rx_en(struct dsi_context *ctx, int enable);
> +void dsi_crc_rx_en(struct dsi_context *ctx, int enable);
> +bool dsi_is_bta_returned(struct dsi_context *ctx);
> +bool dsi_is_rx_payload_fifo_full(struct dsi_context *ctx);
> +bool dsi_is_rx_payload_fifo_empty(struct dsi_context *ctx);
> +bool dsi_is_tx_payload_fifo_full(struct dsi_context *ctx);
> +bool dsi_is_tx_payload_fifo_empty(struct dsi_context *ctx);
> +bool dsi_is_tx_cmd_fifo_empty(struct dsi_context *ctx);
> +void dsi_datalane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_datalane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_clklane_hs2lp_config(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_clklane_lp2hs_config(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_max_read_time(struct dsi_context *ctx, u16 byte_cycle);
> +void dsi_nc_clk_en(struct dsi_context *ctx, int enable);
> +void dsi_tx_escape_division(struct dsi_context *ctx, u8 div);
> +void dsi_timeout_clock_division(struct dsi_context *ctx, u8 div);
> +void dsi_lp_rx_timeout(struct dsi_context *ctx, u16 count);
> +void dsi_hs_tx_timeout(struct dsi_context *ctx, u16 count);
> +u32 dsi_int0_status(struct dsi_context *ctx);
> +u32 dsi_int1_status(struct dsi_context *ctx);
> +void dsi_int0_mask(struct dsi_context *ctx, u32 mask);
> +void dsi_int1_mask(struct dsi_context *ctx, u32 mask);
> +
> +#endif /* _DSI_CTRL_H_ */
> diff --git a/drivers/gpu/drm/sprd/dsi_ctrl_ppi.c b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.c
> new file mode 100644
> index 000000000..6b3259ae5
> --- /dev/null
> +++ b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.c
> @@ -0,0 +1,157 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020 Unisoc Inc.
> + */
> +
> +#include <linux/io.h>
> +#include <linux/init.h>
> +#include <linux/module.h>
> +
> +#include "dsi_ctrl.h"
> +#include "dsi_ctrl_ppi.h"
> +
> +/*
> + * Reset D-PHY module
> + */
> +void dsi_phy_rstz(struct dsi_context *ctx, int level)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x78 phy_interface_ctrl;
> +
> +	phy_interface_ctrl.val = readl(&reg->PHY_INTERFACE_CTRL);
> +	phy_interface_ctrl.bits.rf_phy_reset_n = level;
> +
> +	writel(phy_interface_ctrl.val, &reg->PHY_INTERFACE_CTRL);
> +}
> +
> +/*
> + * Power up/down D-PHY module
> + */
> +void dsi_phy_shutdownz(struct dsi_context *ctx, int level)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x78 phy_interface_ctrl;
> +
> +	phy_interface_ctrl.val = readl(&reg->PHY_INTERFACE_CTRL);
> +	phy_interface_ctrl.bits.rf_phy_shutdown = level;
> +
> +	writel(phy_interface_ctrl.val, &reg->PHY_INTERFACE_CTRL);
> +}
> +
> +/*
> + * Configure minimum wait period for HS transmission request after a stop state
> + */
> +void dsi_phy_stop_wait_time(struct dsi_context *ctx, u8 byte_cycle)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(byte_cycle, &reg->PHY_MIN_STOP_TIME);
> +}
> +
> +/*
> + * Set number of active lanes
> + */
> +void dsi_phy_datalane_en(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +
> +	writel(ctx->lanes - 1, &reg->PHY_LANE_NUM_CONFIG);
> +}
> +
> +/*
> + * Enable clock lane module
> + */
> +void dsi_phy_clklane_en(struct dsi_context *ctx, int en)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x78 phy_interface_ctrl;
> +
> +	phy_interface_ctrl.val = readl(&reg->PHY_INTERFACE_CTRL);
> +	phy_interface_ctrl.bits.rf_phy_clk_en = en;
> +
> +	writel(phy_interface_ctrl.val, &reg->PHY_INTERFACE_CTRL);
> +}
> +
> +/*
> + * Request the PHY module to start transmission of high speed clock.
> + * This causes the clock lane to start transmitting DDR clock on the
> + * lane interconnect.
> + */
> +void dsi_phy_clk_hs_rqst(struct dsi_context *ctx, int enable)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x74 phy_clk_lane_lp_ctrl;
> +
> +	phy_clk_lane_lp_ctrl.val = readl(&reg->PHY_CLK_LANE_LP_CTRL);
> +	phy_clk_lane_lp_ctrl.bits.auto_clklane_ctrl_en = 0;
> +	phy_clk_lane_lp_ctrl.bits.phy_clklane_tx_req_hs = enable;
> +
> +	writel(phy_clk_lane_lp_ctrl.val, &reg->PHY_CLK_LANE_LP_CTRL);
> +}
> +
> +/*
> + * Get D-PHY PPI status
> + */
> +u8 dsi_phy_is_pll_locked(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0x9C phy_status;
> +
> +	phy_status.val = readl(&reg->PHY_STATUS);
> +
> +	return phy_status.bits.phy_lock;
> +}
> +
> +void dsi_phy_test_clk(struct dsi_context *ctx, u8 value)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xF0 phy_tst_ctrl0;
> +
> +	phy_tst_ctrl0.val = readl(&reg->PHY_TST_CTRL0);
> +	phy_tst_ctrl0.bits.phy_testclk = value;
> +
> +	writel(phy_tst_ctrl0.val, &reg->PHY_TST_CTRL0);
> +}
> +
> +void dsi_phy_test_clr(struct dsi_context *ctx, u8 value)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xF0 phy_tst_ctrl0;
> +
> +	phy_tst_ctrl0.val = readl(&reg->PHY_TST_CTRL0);
> +	phy_tst_ctrl0.bits.phy_testclr = value;
> +
> +	writel(phy_tst_ctrl0.val, &reg->PHY_TST_CTRL0);
> +}
> +
> +void dsi_phy_test_en(struct dsi_context *ctx, u8 value)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xF4 phy_tst_ctrl1;
> +
> +	phy_tst_ctrl1.val = readl(&reg->PHY_TST_CTRL1);
> +	phy_tst_ctrl1.bits.phy_testen = value;
> +
> +	writel(phy_tst_ctrl1.val, &reg->PHY_TST_CTRL1);
> +}
> +
> +u8 dsi_phy_test_dout(struct dsi_context *ctx)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xF4 phy_tst_ctrl1;
> +
> +	phy_tst_ctrl1.val = readl(&reg->PHY_TST_CTRL1);
> +
> +	return phy_tst_ctrl1.bits.phy_testdout;
> +}
> +
> +void dsi_phy_test_din(struct dsi_context *ctx, u8 data)
> +{
> +	struct dsi_reg *reg = (struct dsi_reg *)ctx->base;
> +	union _0xF4 phy_tst_ctrl1;
> +
> +	phy_tst_ctrl1.val = readl(&reg->PHY_TST_CTRL1);
> +	phy_tst_ctrl1.bits.phy_testdin = data;
> +
> +	writel(phy_tst_ctrl1.val, &reg->PHY_TST_CTRL1);
> +}
> diff --git a/drivers/gpu/drm/sprd/dsi_ctrl_ppi.h b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.h
> new file mode 100644
> index 000000000..6fb3fad89
> --- /dev/null
> +++ b/drivers/gpu/drm/sprd/dsi_ctrl_ppi.h
> @@ -0,0 +1,26 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020 Unisoc Inc.
> + */
> +
> +#ifndef _DSI_CTRL_PPI_H_
> +#define _DSI_CTRL_PPI_H_
> +
> +#include "sprd_dsi.h"
> +
> +void dsi_phy_rstz(struct dsi_context *ctx, int level);
> +void dsi_phy_shutdownz(struct dsi_context *ctx, int level);
> +void dsi_phy_force_pll(struct dsi_context *ctx, int force);
> +void dsi_phy_stop_wait_time(struct dsi_context *ctx, u8 byte_clk);
> +void dsi_phy_datalane_en(struct dsi_context *ctx);
> +void dsi_phy_clklane_en(struct dsi_context *ctx, int en);
> +void dsi_phy_clk_hs_rqst(struct dsi_context *ctx, int en);
> +u8 dsi_phy_is_pll_locked(struct dsi_context *ctx);
> +void dsi_phy_test_clk(struct dsi_context *ctx, u8 level);
> +void dsi_phy_test_clr(struct dsi_context *ctx, u8 level);
> +void dsi_phy_test_en(struct dsi_context *ctx, u8 level);
> +u8 dsi_phy_test_dout(struct dsi_context *ctx);
> +void dsi_phy_test_din(struct dsi_context *ctx, u8 data);
> +void dsi_phy_bist_en(struct dsi_context *ctx, int en);
> +
> +#endif /* _DSI_CTRL_PPI_H_ */
> diff --git a/drivers/gpu/drm/sprd/megacores_pll.c b/drivers/gpu/drm/sprd/megacores_pll.c
> new file mode 100644
> index 000000000..03c1f0f00
> --- /dev/null
> +++ b/drivers/gpu/drm/sprd/megacores_pll.c
> @@ -0,0 +1,317 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020 Unisoc Inc.
> + */
> +
> +#include <asm/div64.h>
> +#include <linux/delay.h>
> +#include <linux/init.h>
> +#include <linux/kernel.h>
> +#include <linux/regmap.h>
> +#include <linux/string.h>
> +
> +#include "megacores_pll.h"
> +
> +#define L						0
> +#define H						1
> +#define CLK						0
> +#define DATA					1
> +#define INFINITY				0xffffffff
> +#define MIN_OUTPUT_FREQ			(100)
> +
> +#define AVERAGE(a, b) (min(a, b) + abs((b) - (a)) / 2)
> +
> +/* sharkle */
> +#define VCO_BAND_LOW	750
> +#define VCO_BAND_MID	1100
> +#define VCO_BAND_HIGH	1500
> +#define PHY_REF_CLK	26000
> +
> +static int dphy_calc_pll_param(struct dphy_pll *pll)
> +{
> +	const u32 khz = 1000;
> +	const u32 mhz = 1000000;
> +	const unsigned long long factor = 100;
> +	unsigned long long tmp;
> +	int i;
> +
> +	pll->potential_fvco = pll->freq / khz;
> +	pll->ref_clk = PHY_REF_CLK / khz;
> +
> +	for (i = 0; i < 4; ++i) {
> +		if (pll->potential_fvco >= VCO_BAND_LOW &&
> +			pll->potential_fvco <= VCO_BAND_HIGH) {
> +			pll->fvco = pll->potential_fvco;
> +			pll->out_sel = BIT(i);
> +			break;
> +		}
> +		pll->potential_fvco <<= 1;
> +	}
> +	if (pll->fvco == 0)
> +		return -EINVAL;
> +
> +	if (pll->fvco >= VCO_BAND_LOW && pll->fvco <= VCO_BAND_MID) {
> +		/* vco band control */
> +		pll->vco_band = 0x0;
> +		/* low pass filter control */
> +		pll->lpf_sel = 1;
> +	} else if (pll->fvco > VCO_BAND_MID && pll->fvco <= VCO_BAND_HIGH) {
> +		pll->vco_band = 0x1;
> +		pll->lpf_sel = 0;
> +	} else
> +		return -EINVAL;
> +
> +	pll->nint = pll->fvco / pll->ref_clk;
> +	tmp = pll->fvco * factor * mhz;
> +	do_div(tmp, pll->ref_clk);
> +	tmp = tmp - pll->nint * factor * mhz;
> +	tmp *= BIT(20);
> +	do_div(tmp, 100000000);
> +	pll->kint = (u32)tmp;
> +	pll->refin = 3; /* pre-divider bypass */
> +	pll->sdm_en = true; /* use fraction N PLL */
> +	pll->fdk_s = 0x1; /* fraction */
> +	pll->cp_s = 0x0;
> +	pll->det_delay = 0x1;
> +
> +	return 0;
> +}
> +
> +static void dphy_set_pll_reg(struct dphy_pll *pll, struct regmap *regmap)
> +{
> +	struct pll_reg *reg = &pll->reg;
> +	u8 *val;
> +	int i;
> +
> +	u8 reg_addr[] = {
> +		0x03, 0x04, 0x06, 0x08, 0x09,
> +		0x0a, 0x0b, 0x0e, 0x0f
> +	};
> +
> +	reg->_03.bits.prbs_bist = 1;
> +	reg->_03.bits.en_lp_treot = true;
> +	reg->_03.bits.lpf_sel = pll->lpf_sel;
> +	reg->_03.bits.txfifo_bypass = 0;
> +	reg->_04.bits.div = pll->div;
> +	reg->_04.bits.masterof8lane = 1;
> +	reg->_04.bits.cp_s = pll->cp_s;
> +	reg->_04.bits.fdk_s = pll->fdk_s;
> +	reg->_06.bits.nint = pll->nint;
> +	reg->_08.bits.vco_band = pll->vco_band;
> +	reg->_08.bits.sdm_en = pll->sdm_en;
> +	reg->_08.bits.refin = pll->refin;
> +	reg->_09.bits.kint_h = pll->kint >> 12;
> +	reg->_0a.bits.kint_m = (pll->kint >> 4) & 0xff;
> +	reg->_0b.bits.out_sel = pll->out_sel;
> +	reg->_0b.bits.kint_l = pll->kint & 0xf;
> +	reg->_0e.bits.pll_pu_byp = 0;
> +	reg->_0e.bits.pll_pu = 0;
> +	reg->_0e.bits.stopstate_sel = 1;
> +	reg->_0f.bits.det_delay = pll->det_delay;
> +
> +	val = (u8 *)®
> +
> +	for (i = 0; i < sizeof(reg_addr); ++i) {
> +		regmap_write(regmap, reg_addr[i], val[i]);
> +		DRM_DEBUG("%02x: %02x\n", reg_addr[i], val[i]);
> +	}
> +}
> +
> +int dphy_pll_config(struct dsi_context *ctx)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	struct regmap *regmap = ctx->regmap;
> +	struct dphy_pll *pll = ctx->pll;
> +	int ret;
> +
> +	pll->freq = ctx->byte_clk * 8;
> +
> +	/* FREQ = 26M * (NINT + KINT / 2^20) / out_sel */
> +	ret = dphy_calc_pll_param(pll);
> +	if (ret) {
> +		drm_err(dsi->drm, "failed to calculate dphy pll parameters\n");
> +		return ret;
> +	}
> +	dphy_set_pll_reg(pll, regmap);
> +
> +	return 0;
> +}
> +
> +static void dphy_set_timing_reg(struct regmap *regmap, int type, u8 val[])
> +{
> +	switch (type) {
> +	case REQUEST_TIME:
> +		regmap_write(regmap, 0x31, val[CLK]);
> +		regmap_write(regmap, 0x41, val[DATA]);
> +		regmap_write(regmap, 0x51, val[DATA]);
> +		regmap_write(regmap, 0x61, val[DATA]);
> +		regmap_write(regmap, 0x71, val[DATA]);
> +
> +		regmap_write(regmap, 0x90, val[CLK]);
> +		regmap_write(regmap, 0xa0, val[DATA]);
> +		regmap_write(regmap, 0xb0, val[DATA]);
> +		regmap_write(regmap, 0xc0, val[DATA]);
> +		regmap_write(regmap, 0xd0, val[DATA]);
> +		break;
> +	case PREPARE_TIME:
> +		regmap_write(regmap, 0x32, val[CLK]);
> +		regmap_write(regmap, 0x42, val[DATA]);
> +		regmap_write(regmap, 0x52, val[DATA]);
> +		regmap_write(regmap, 0x62, val[DATA]);
> +		regmap_write(regmap, 0x72, val[DATA]);
> +
> +		regmap_write(regmap, 0x91, val[CLK]);
> +		regmap_write(regmap, 0xa1, val[DATA]);
> +		regmap_write(regmap, 0xb1, val[DATA]);
> +		regmap_write(regmap, 0xc1, val[DATA]);
> +		regmap_write(regmap, 0xd1, val[DATA]);
> +		break;
> +	case ZERO_TIME:
> +		regmap_write(regmap, 0x33, val[CLK]);
> +		regmap_write(regmap, 0x43, val[DATA]);
> +		regmap_write(regmap, 0x53, val[DATA]);
> +		regmap_write(regmap, 0x63, val[DATA]);
> +		regmap_write(regmap, 0x73, val[DATA]);
> +
> +		regmap_write(regmap, 0x92, val[CLK]);
> +		regmap_write(regmap, 0xa2, val[DATA]);
> +		regmap_write(regmap, 0xb2, val[DATA]);
> +		regmap_write(regmap, 0xc2, val[DATA]);
> +		regmap_write(regmap, 0xd2, val[DATA]);
> +		break;
> +	case TRAIL_TIME:
> +		regmap_write(regmap, 0x34, val[CLK]);
> +		regmap_write(regmap, 0x44, val[DATA]);
> +		regmap_write(regmap, 0x54, val[DATA]);
> +		regmap_write(regmap, 0x64, val[DATA]);
> +		regmap_write(regmap, 0x74, val[DATA]);
> +
> +		regmap_write(regmap, 0x93, val[CLK]);
> +		regmap_write(regmap, 0xa3, val[DATA]);
> +		regmap_write(regmap, 0xb3, val[DATA]);
> +		regmap_write(regmap, 0xc3, val[DATA]);
> +		regmap_write(regmap, 0xd3, val[DATA]);
> +		break;
> +	case EXIT_TIME:
> +		regmap_write(regmap, 0x36, val[CLK]);
> +		regmap_write(regmap, 0x46, val[DATA]);
> +		regmap_write(regmap, 0x56, val[DATA]);
> +		regmap_write(regmap, 0x66, val[DATA]);
> +		regmap_write(regmap, 0x76, val[DATA]);
> +
> +		regmap_write(regmap, 0x95, val[CLK]);
> +		regmap_write(regmap, 0xA5, val[DATA]);
> +		regmap_write(regmap, 0xB5, val[DATA]);
> +		regmap_write(regmap, 0xc5, val[DATA]);
> +		regmap_write(regmap, 0xd5, val[DATA]);
> +		break;
> +	case CLKPOST_TIME:
> +		regmap_write(regmap, 0x35, val[CLK]);
> +		regmap_write(regmap, 0x94, val[CLK]);
> +		break;
> +
> +	/* the following just use default value */
> +	case SETTLE_TIME:
> +	case TA_GET:
> +	case TA_GO:
> +	case TA_SURE:
> +		break;
> +	default:
> +		break;
> +	}
> +}
> +
> +void dphy_timing_config(struct dsi_context *ctx)
> +{
> +	struct regmap *regmap = ctx->regmap;
> +	struct dphy_pll *pll = ctx->pll;
> +	const u32 factor = 2;
> +	const u32 scale = 100;
> +	u32 t_ui, t_byteck, t_half_byteck;
> +	u32 range[2], constant;
> +	u8 val[2];
> +	u32 tmp = 0;
> +
> +	/* t_ui: 1 ui, byteck: 8 ui, half byteck: 4 ui */
> +	t_ui = 1000 * scale / (pll->freq / 1000);
> +	t_byteck = t_ui << 3;
> +	t_half_byteck = t_ui << 2;
> +	constant = t_ui << 1;
> +
> +	/* REQUEST_TIME: HS T-LPX: LP-01
> +	 * For T-LPX, mipi spec defined min value is 50ns,
> +	 * but maybe it shouldn't be too small, because BTA,
> +	 * LP-10, LP-00, LP-01, all of this is related to T-LPX.
> +	 */
> +	range[L] = 50 * scale;
> +	range[H] = INFINITY;
> +	val[CLK] = DIV_ROUND_UP(range[L] * (factor << 1), t_byteck) - 2;
> +	val[DATA] = val[CLK];
> +	dphy_set_timing_reg(regmap, REQUEST_TIME, val);
> +
> +	/* PREPARE_TIME: HS sequence: LP-00 */
> +	range[L] = 38 * scale;
> +	range[H] = 95 * scale;
> +	tmp = AVERAGE(range[L], range[H]);
> +	val[CLK] = DIV_ROUND_UP(AVERAGE(range[L], range[H]),
> +			t_half_byteck) - 1;
> +	range[L] = 40 * scale + 4 * t_ui;
> +	range[H] = 85 * scale + 6 * t_ui;
> +	tmp |= AVERAGE(range[L], range[H]) << 16;
> +	val[DATA] = DIV_ROUND_UP(AVERAGE(range[L], range[H]),
> +			t_half_byteck) - 1;
> +	dphy_set_timing_reg(regmap, PREPARE_TIME, val);
> +
> +	/* ZERO_TIME: HS-ZERO */
> +	range[L] = 300 * scale;
> +	range[H] = INFINITY;
> +	val[CLK] = DIV_ROUND_UP(range[L] * factor + (tmp & 0xffff)
> +			- 525 * t_byteck / 100, t_byteck) - 2;
> +	range[L] = 145 * scale + 10 * t_ui;
> +	val[DATA] = DIV_ROUND_UP(range[L] * factor
> +			+ ((tmp >> 16) & 0xffff) - 525 * t_byteck / 100,
> +			t_byteck) - 2;
> +	dphy_set_timing_reg(regmap, ZERO_TIME, val);
> +
> +	/* TRAIL_TIME: HS-TRAIL */
> +	range[L] = 60 * scale;
> +	range[H] = INFINITY;
> +	val[CLK] = DIV_ROUND_UP(range[L] * factor - constant, t_half_byteck);
> +	range[L] = max(8 * t_ui, 60 * scale + 4 * t_ui);
> +	val[DATA] = DIV_ROUND_UP(range[L] * 3 / 2 - constant, t_half_byteck) - 2;
> +	dphy_set_timing_reg(regmap, TRAIL_TIME, val);
> +
> +	/* EXIT_TIME: */
> +	range[L] = 100 * scale;
> +	range[H] = INFINITY;
> +	val[CLK] = DIV_ROUND_UP(range[L] * factor, t_byteck) - 2;
> +	val[DATA] = val[CLK];
> +	dphy_set_timing_reg(regmap, EXIT_TIME, val);
> +
> +	/* CLKPOST_TIME: */
> +	range[L] = 60 * scale + 52 * t_ui;
> +	range[H] = INFINITY;
> +	val[CLK] = DIV_ROUND_UP(range[L] * factor, t_byteck) - 2;
> +	val[DATA] = val[CLK];
> +	dphy_set_timing_reg(regmap, CLKPOST_TIME, val);
> +
> +	/* SETTLE_TIME:
> +	 * This time is used for receiver. So for transmitter,
> +	 * it can be ignored.
> +	 */
> +
> +	/* TA_GO:
> +	 * transmitter drives bridge state(LP-00) before releasing control,
> +	 * reg 0x1f default value: 0x04, which is good.
> +	 */
> +
> +	/* TA_SURE:
> +	 * After LP-10 state and before bridge state(LP-00),
> +	 * reg 0x20 default value: 0x01, which is good.
> +	 */
> +
> +	/* TA_GET:
> +	 * receiver drives Bridge state(LP-00) before releasing control
> +	 * reg 0x21 default value: 0x03, which is good.
> +	 */
> +}
> diff --git a/drivers/gpu/drm/sprd/megacores_pll.h b/drivers/gpu/drm/sprd/megacores_pll.h
> new file mode 100644
> index 000000000..bf20aae65
> --- /dev/null
> +++ b/drivers/gpu/drm/sprd/megacores_pll.h
> @@ -0,0 +1,146 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (C) 2020 Unisoc Inc.
> + */
> +
> +#ifndef _MEGACORES_PLL_H_
> +#define _MEGACORES_PLL_H_
> +
> +#include "sprd_dsi.h"
> +
> +enum PLL_TIMING {
> +	NONE,
> +	REQUEST_TIME,
> +	PREPARE_TIME,
> +	SETTLE_TIME,
> +	ZERO_TIME,
> +	TRAIL_TIME,
> +	EXIT_TIME,
> +	CLKPOST_TIME,
> +	TA_GET,
> +	TA_GO,
> +	TA_SURE,
> +	TA_WAIT,
> +};
> +
> +struct pll_reg {
> +	union __reg_03__ {
> +		struct __03 {
> +			u8 prbs_bist: 1;
> +			u8 en_lp_treot: 1;
> +			u8 lpf_sel: 4;
> +			u8 txfifo_bypass: 1;
> +			u8 freq_hopping: 1;
> +		} bits;
> +		u8 val;
> +	} _03;
> +	union __reg_04__ {
> +		struct __04 {
> +			u8 div: 3;
> +			u8 masterof8lane: 1;
> +			u8 hop_trig: 1;
> +			u8 cp_s: 2;
> +			u8 fdk_s: 1;
> +		} bits;
> +		u8 val;
> +	} _04;
> +	union __reg_06__ {
> +		struct __06 {
> +			u8 nint: 7;
> +			u8 mod_en: 1;
> +		} bits;
> +		u8 val;
> +	} _06;
> +	union __reg_07__ {
> +		struct __07 {
> +			u8 kdelta_h: 8;
> +		} bits;
> +		u8 val;
> +	} _07;
> +	union __reg_08__ {
> +		struct __08 {
> +			u8 vco_band: 1;
> +			u8 sdm_en: 1;
> +			u8 refin: 2;
> +			u8 kdelta_l: 4;
> +		} bits;
> +		u8 val;
> +	} _08;
> +	union __reg_09__ {
> +		struct __09 {
> +			u8 kint_h: 8;
> +		} bits;
> +		u8 val;
> +	} _09;
> +	union __reg_0a__ {
> +		struct __0a {
> +			u8 kint_m: 8;
> +		} bits;
> +		u8 val;
> +	} _0a;
> +	union __reg_0b__ {
> +		struct __0b {
> +			u8 out_sel: 4;
> +			u8 kint_l: 4;
> +		} bits;
> +		u8 val;
> +	} _0b;
> +	union __reg_0c__ {
> +		struct __0c {
> +			u8 kstep_h: 8;
> +		} bits;
> +		u8 val;
> +	} _0c;
> +	union __reg_0d__ {
> +		struct __0d {
> +			u8 kstep_m: 8;
> +		} bits;
> +		u8 val;
> +	} _0d;
> +	union __reg_0e__ {
> +		struct __0e {
> +			u8 pll_pu_byp: 1;
> +			u8 pll_pu: 1;
> +			u8 hsbist_len: 2;
> +			u8 stopstate_sel: 1;
> +			u8 kstep_l: 3;
> +		} bits;
> +		u8 val;
> +	} _0e;
> +	union __reg_0f__ {
> +		struct __0f {
> +			u8 det_delay:2;
> +			u8 kdelta: 4;
> +			u8 ldo0p4:2;
> +		} bits;
> +		u8 val;
> +	} _0f;
> +};
> +
> +struct dphy_pll {
> +	u8 refin; /* Pre-divider control signal */
> +	u8 cp_s; /* 00: SDM_EN=1, 10: SDM_EN=0 */
> +	u8 fdk_s; /* PLL mode control: integer or fraction */
> +	u8 sdm_en;
> +	u8 div;
> +	u8 int_n; /* integer N PLL */
> +	u32 ref_clk; /* dphy reference clock, unit: MHz */
> +	u32 freq; /* panel config, unit: KHz */
> +	u32 fvco;
> +	u32 potential_fvco;
> +	u32 nint; /* sigma delta modulator NINT control */
> +	u32 kint; /* sigma delta modulator KINT control */
> +	u8 lpf_sel; /* low pass filter control */
> +	u8 out_sel; /* post divider control */
> +	u8 vco_band; /* vco range */
> +	u8 det_delay;
> +
> +	struct pll_reg reg;
> +};
> +
> +struct dsi_context;
> +
> +int dphy_pll_config(struct dsi_context *ctx);
> +void dphy_timing_config(struct dsi_context *ctx);
> +
> +#endif /* _MEGACORES_PLL_H_ */
> diff --git a/drivers/gpu/drm/sprd/sprd_dpu.c b/drivers/gpu/drm/sprd/sprd_dpu.c
> index e74c3dbb3..10f6038e6 100644
> --- a/drivers/gpu/drm/sprd/sprd_dpu.c
> +++ b/drivers/gpu/drm/sprd/sprd_dpu.c
> @@ -25,6 +25,7 @@
>  
>  #include "sprd_drm.h"
>  #include "sprd_dpu.h"
> +#include "sprd_dsi.h"
>  
>  /* Global control registers */
>  #define REG_DPU_CTRL	0x04
> @@ -687,9 +688,25 @@ static void sprd_crtc_mode_set_nofb(struct drm_crtc *crtc)
>  {
>  	struct sprd_dpu *dpu = to_sprd_crtc(crtc);
>  	struct drm_display_mode *mode = &crtc->state->adjusted_mode;
> +	struct drm_encoder *encoder;
> +	struct mipi_dsi_device *slave;
> +	struct sprd_dsi *dsi;
>  
>  	if (mode->type & DRM_MODE_TYPE_PREFERRED)
>  		drm_display_mode_to_videomode(mode, &dpu->ctx.vm);
> +
> +	drm_for_each_encoder(encoder, crtc->dev) {
> +		if (encoder->crtc != crtc)
> +			continue;
> +
> +		dsi = encoder_to_dsi(encoder);
> +		slave = dsi->slave;
> +
> +		if (slave->mode_flags & MIPI_DSI_MODE_VIDEO)
> +			dpu->ctx.if_type = SPRD_DPU_IF_DPI;
> +		else
> +			dpu->ctx.if_type = SPRD_DPU_IF_EDPI;
> +	}

This would make more sense to store in a custom crtc_state and to fill
at atomic_check time. That way you can properly fail if you don't find
the encoder, or if it's not a DSI encoder, etc.

>  }
>  
>  static void sprd_crtc_atomic_enable(struct drm_crtc *crtc,
> diff --git a/drivers/gpu/drm/sprd/sprd_drm.c b/drivers/gpu/drm/sprd/sprd_drm.c
> index 549b71278..68cf0933c 100644
> --- a/drivers/gpu/drm/sprd/sprd_drm.c
> +++ b/drivers/gpu/drm/sprd/sprd_drm.c
> @@ -183,6 +183,7 @@ static struct platform_driver sprd_drm_driver = {
>  static struct platform_driver *sprd_drm_drivers[]  = {
>  	&sprd_drm_driver,
>  	&sprd_dpu_driver,
> +	&sprd_dsi_driver,
>  };
>  
>  static int __init sprd_drm_init(void)
> diff --git a/drivers/gpu/drm/sprd/sprd_drm.h b/drivers/gpu/drm/sprd/sprd_drm.h
> index 85d4a8b9f..95d1b972f 100644
> --- a/drivers/gpu/drm/sprd/sprd_drm.h
> +++ b/drivers/gpu/drm/sprd/sprd_drm.h
> @@ -14,5 +14,6 @@ struct sprd_drm {
>  };
>  
>  extern struct platform_driver sprd_dpu_driver;
> +extern struct platform_driver sprd_dsi_driver;
>  
>  #endif /* _SPRD_DRM_H_ */
> diff --git a/drivers/gpu/drm/sprd/sprd_dsi.c b/drivers/gpu/drm/sprd/sprd_dsi.c
> new file mode 100644
> index 000000000..f26a37e50
> --- /dev/null
> +++ b/drivers/gpu/drm/sprd/sprd_dsi.c
> @@ -0,0 +1,1124 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (C) 2020 Unisoc Inc.
> + */
> +
> +#include <linux/component.h>
> +#include <linux/module.h>
> +#include <linux/of_address.h>
> +#include <linux/of_device.h>
> +#include <linux/of_irq.h>
> +#include <linux/of_graph.h>
> +#include <video/mipi_display.h>
> +
> +#include <drm/drm_atomic_helper.h>
> +#include <drm/drm_crtc_helper.h>
> +#include <drm/drm_of.h>
> +#include <drm/drm_probe_helper.h>
> +
> +#include "sprd_drm.h"
> +#include "sprd_dpu.h"
> +#include "sprd_dsi.h"
> +#include "dsi_ctrl.h"
> +#include "dsi_ctrl_ppi.h"
> +
> +#define host_to_dsi(host) \
> +	container_of(host, struct sprd_dsi, host)
> +#define connector_to_dsi(connector) \
> +	container_of(connector, struct sprd_dsi, connector)
> +
> +static int regmap_tst_io_write(void *context, u32 reg, u32 val)
> +{
> +	struct sprd_dsi *dsi = context;
> +	struct dsi_context *ctx = &dsi->ctx;
> +
> +	if (val > 0xff || reg > 0xff)
> +		return -EINVAL;
> +
> +	drm_dbg(dsi->drm, "reg = 0x%02x, val = 0x%02x\n", reg, val);
> +
> +	dsi_phy_test_en(ctx, 1);
> +	dsi_phy_test_din(ctx, reg);
> +	dsi_phy_test_clk(ctx, 1);
> +	dsi_phy_test_clk(ctx, 0);
> +	dsi_phy_test_en(ctx, 0);
> +	dsi_phy_test_din(ctx, val);
> +	dsi_phy_test_clk(ctx, 1);
> +	dsi_phy_test_clk(ctx, 0);
> +
> +	return 0;
> +}
> +
> +static int regmap_tst_io_read(void *context, u32 reg, u32 *val)
> +{
> +	struct sprd_dsi *dsi = context;
> +	struct dsi_context *ctx = &dsi->ctx;
> +	int ret;
> +
> +	if (reg > 0xff)
> +		return -EINVAL;
> +
> +	dsi_phy_test_en(ctx, 1);
> +	dsi_phy_test_din(ctx, reg);
> +	dsi_phy_test_clk(ctx, 1);
> +	dsi_phy_test_clk(ctx, 0);
> +	dsi_phy_test_en(ctx, 0);
> +
> +	udelay(1);
> +
> +	ret = dsi_phy_test_dout(ctx);
> +	if (ret < 0)
> +		return ret;
> +
> +	*val = ret;
> +
> +	drm_dbg(dsi->drm, "reg = 0x%02x, val = 0x%02x\n", reg, *val);
> +	return 0;
> +}
> +
> +static struct regmap_bus regmap_tst_io = {
> +	.reg_write = regmap_tst_io_write,
> +	.reg_read = regmap_tst_io_read,
> +};
> +
> +static const struct regmap_config byte_config = {
> +	.reg_bits = 8,
> +	.val_bits = 8,
> +};
> +
> +static int dphy_wait_pll_locked(struct dsi_context *ctx)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	int i;
> +
> +	for (i = 0; i < 50000; i++) {
> +		if (dsi_phy_is_pll_locked(ctx))
> +			return 0;
> +		udelay(3);
> +	}
> +
> +	drm_err(dsi->drm, "dphy pll can not be locked\n");
> +	return -ETIMEDOUT;
> +}
> +
> +static int dsi_wait_tx_payload_fifo_empty(struct dsi_context *ctx)
> +{
> +	int i;
> +
> +	for (i = 0; i < 5000; i++) {
> +		if (dsi_is_tx_payload_fifo_empty(ctx))
> +			return 0;
> +		udelay(1);
> +	}
> +
> +	return -ETIMEDOUT;
> +}
> +
> +static int dsi_wait_tx_cmd_fifo_empty(struct dsi_context *ctx)
> +{
> +	int i;
> +
> +	for (i = 0; i < 5000; i++) {
> +		if (dsi_is_tx_cmd_fifo_empty(ctx))
> +			return 0;
> +		udelay(1);
> +	}
> +
> +	return -ETIMEDOUT;
> +}
> +
> +static int dsi_wait_rd_resp_completed(struct dsi_context *ctx)
> +{
> +	int i;
> +
> +	for (i = 0; i < 10000; i++) {
> +		if (dsi_is_bta_returned(ctx))
> +			return 0;
> +		udelay(10);
> +	}
> +
> +	return -ETIMEDOUT;
> +}
> +
> +static u16 calc_bytes_per_pixel_x100(int coding)
> +{
> +	u16 Bpp_x100;
> +
> +	switch (coding) {
> +	case COLOR_CODE_16BIT_CONFIG1:
> +	case COLOR_CODE_16BIT_CONFIG2:
> +	case COLOR_CODE_16BIT_CONFIG3:
> +		Bpp_x100 = 200;
> +		break;
> +	case COLOR_CODE_18BIT_CONFIG1:
> +	case COLOR_CODE_18BIT_CONFIG2:
> +		Bpp_x100 = 225;
> +		break;
> +	case COLOR_CODE_24BIT:
> +		Bpp_x100 = 300;
> +		break;
> +	case COLOR_CODE_COMPRESSTION:
> +		Bpp_x100 = 100;
> +		break;
> +	case COLOR_CODE_20BIT_YCC422_LOOSELY:
> +		Bpp_x100 = 250;
> +		break;
> +	case COLOR_CODE_24BIT_YCC422:
> +		Bpp_x100 = 300;
> +		break;
> +	case COLOR_CODE_16BIT_YCC422:
> +		Bpp_x100 = 200;
> +		break;
> +	case COLOR_CODE_30BIT:
> +		Bpp_x100 = 375;
> +		break;
> +	case COLOR_CODE_36BIT:
> +		Bpp_x100 = 450;
> +		break;
> +	case COLOR_CODE_12BIT_YCC420:
> +		Bpp_x100 = 150;
> +		break;
> +	default:
> +		DRM_ERROR("invalid color coding");
> +		Bpp_x100 = 0;
> +		break;
> +	}
> +
> +	return Bpp_x100;
> +}
> +
> +static u8 calc_video_size_step(int coding)
> +{
> +	u8 video_size_step;
> +
> +	switch (coding) {
> +	case COLOR_CODE_16BIT_CONFIG1:
> +	case COLOR_CODE_16BIT_CONFIG2:
> +	case COLOR_CODE_16BIT_CONFIG3:
> +	case COLOR_CODE_18BIT_CONFIG1:
> +	case COLOR_CODE_18BIT_CONFIG2:
> +	case COLOR_CODE_24BIT:
> +	case COLOR_CODE_COMPRESSTION:
> +		return video_size_step = 1;
> +	case COLOR_CODE_20BIT_YCC422_LOOSELY:
> +	case COLOR_CODE_24BIT_YCC422:
> +	case COLOR_CODE_16BIT_YCC422:
> +	case COLOR_CODE_30BIT:
> +	case COLOR_CODE_36BIT:
> +	case COLOR_CODE_12BIT_YCC420:
> +		return video_size_step = 2;
> +	default:
> +		DRM_ERROR("invalid color coding");
> +		return 0;
> +	}
> +}
> +
> +static u16 round_video_size(int coding, u16 video_size)
> +{
> +	switch (coding) {
> +	case COLOR_CODE_16BIT_YCC422:
> +	case COLOR_CODE_24BIT_YCC422:
> +	case COLOR_CODE_20BIT_YCC422_LOOSELY:
> +	case COLOR_CODE_12BIT_YCC420:
> +		/* round up active H pixels to a multiple of 2 */
> +		if ((video_size % 2) != 0)
> +			video_size += 1;
> +		break;
> +	default:
> +		break;
> +	}
> +
> +	return video_size;
> +}
> +
> +#define SPRD_MIPI_DSI_FMT_DSC 0xff
> +static u32 fmt_to_coding(u32 fmt)
> +{
> +	switch (fmt) {
> +	case MIPI_DSI_FMT_RGB565:
> +		return COLOR_CODE_16BIT_CONFIG1;
> +	case MIPI_DSI_FMT_RGB666:
> +	case MIPI_DSI_FMT_RGB666_PACKED:
> +		return COLOR_CODE_18BIT_CONFIG1;
> +	case MIPI_DSI_FMT_RGB888:
> +		return COLOR_CODE_24BIT;
> +	case SPRD_MIPI_DSI_FMT_DSC:
> +		return COLOR_CODE_COMPRESSTION;
> +	default:
> +		DRM_ERROR("Unsupported format (%d)\n", fmt);
> +		return COLOR_CODE_24BIT;
> +	}
> +}
> +
> +#define ns_to_cycle(ns, byte_clk) \
> +	DIV_ROUND_UP((ns) * (byte_clk), 1000000)
> +
> +static void sprd_dsi_init(struct dsi_context *ctx)
> +{
> +	u16 data_hs2lp, data_lp2hs, clk_hs2lp, clk_lp2hs;
> +	u16 max_rd_time;
> +	int div;
> +
> +	dsi_power_enable(ctx, 0);
> +	dsi_int0_mask(ctx, 0xffffffff);
> +	dsi_int1_mask(ctx, 0xffffffff);
> +	dsi_cmd_mode(ctx);
> +	dsi_eotp_rx_en(ctx, 0);
> +	dsi_eotp_tx_en(ctx, 0);
> +	dsi_ecc_rx_en(ctx, 1);
> +	dsi_crc_rx_en(ctx, 1);
> +	dsi_bta_en(ctx, 1);
> +	dsi_video_vcid(ctx, 0);
> +	dsi_rx_vcid(ctx, 0);
> +
> +	div = DIV_ROUND_UP(ctx->byte_clk, ctx->esc_clk);
> +	dsi_tx_escape_division(ctx, div);
> +
> +	max_rd_time = ns_to_cycle(ctx->max_rd_time, ctx->byte_clk);
> +	dsi_max_read_time(ctx, max_rd_time);
> +
> +	data_hs2lp = ns_to_cycle(ctx->data_hs2lp, ctx->byte_clk);
> +	data_lp2hs = ns_to_cycle(ctx->data_lp2hs, ctx->byte_clk);
> +	clk_hs2lp = ns_to_cycle(ctx->clk_hs2lp, ctx->byte_clk);
> +	clk_lp2hs = ns_to_cycle(ctx->clk_lp2hs, ctx->byte_clk);
> +	dsi_datalane_hs2lp_config(ctx, data_hs2lp);
> +	dsi_datalane_lp2hs_config(ctx, data_lp2hs);
> +	dsi_clklane_hs2lp_config(ctx, clk_hs2lp);
> +	dsi_clklane_lp2hs_config(ctx, clk_lp2hs);
> +
> +	dsi_power_enable(ctx, 1);
> +}
> +
> +/*
> + * Free up resources and shutdown host controller and PHY
> + */
> +static void sprd_dsi_fini(struct dsi_context *ctx)
> +{
> +	dsi_int0_mask(ctx, 0xffffffff);
> +	dsi_int1_mask(ctx, 0xffffffff);
> +	dsi_power_enable(ctx, 0);
> +}
> +
> +/*
> + * If not in burst mode, it will compute the video and null packet sizes
> + * according to necessity.
> + * Configure timers for data lanes and/or clock lane to return to LP when
> + * bandwidth is not filled by data.
> + */
> +static int sprd_dsi_dpi_video(struct dsi_context *ctx)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	struct videomode *vm = &ctx->vm;
> +	u16 Bpp_x100;
> +	u16 video_size;
> +	u32 ratio_x1000;
> +	u16 null_pkt_size = 0;
> +	u8 video_size_step;
> +	u32 hs_to;
> +	u32 total_bytes;
> +	u32 bytes_per_chunk;
> +	u32 chunks = 0;
> +	u32 bytes_left = 0;
> +	u32 chunk_overhead;
> +	const u8 pkt_header = 6;
> +	u8 coding;
> +	int div;
> +	u16 hline;
> +
> +	coding = fmt_to_coding(ctx->format);
> +	video_size = round_video_size(coding, vm->hactive);
> +	Bpp_x100 = calc_bytes_per_pixel_x100(coding);
> +	video_size_step = calc_video_size_step(coding);
> +	ratio_x1000 = ctx->byte_clk * 1000 / (vm->pixelclock / 1000);
> +	hline = vm->hactive + vm->hsync_len + vm->hfront_porch +
> +		vm->hback_porch;
> +
> +	dsi_power_enable(ctx, 0);
> +	dsi_dpi_frame_ack_en(ctx, ctx->frame_ack_en);
> +	dsi_dpi_color_coding(ctx, coding);
> +	dsi_dpi_video_burst_mode(ctx, ctx->burst_mode);
> +	dsi_dpi_sig_delay(ctx, 95 * hline * ratio_x1000 / 100000);
> +	dsi_dpi_hline_time(ctx, hline * ratio_x1000 / 1000);
> +	dsi_dpi_hsync_time(ctx, vm->hsync_len * ratio_x1000 / 1000);
> +	dsi_dpi_hbp_time(ctx, vm->hback_porch * ratio_x1000 / 1000);
> +	dsi_dpi_vact(ctx, vm->vactive);
> +	dsi_dpi_vfp(ctx, vm->vfront_porch);
> +	dsi_dpi_vbp(ctx, vm->vback_porch);
> +	dsi_dpi_vsync(ctx, vm->vsync_len);
> +	dsi_dpi_hporch_lp_en(ctx, 1);
> +	dsi_dpi_vporch_lp_en(ctx, 1);
> +
> +	hs_to = (hline * vm->vactive) + (2 * Bpp_x100) / 100;
> +	for (div = 0x80; (div < hs_to) && (div > 2); div--) {
> +		if ((hs_to % div) == 0) {
> +			dsi_timeout_clock_division(ctx, div);
> +			dsi_lp_rx_timeout(ctx, hs_to / div);
> +			dsi_hs_tx_timeout(ctx, hs_to / div);
> +			break;
> +		}
> +	}
> +
> +	if (ctx->burst_mode == VIDEO_BURST_WITH_SYNC_PULSES) {
> +		dsi_dpi_video_packet_size(ctx, video_size);
> +		dsi_dpi_null_packet_size(ctx, 0);
> +		dsi_dpi_chunk_num(ctx, 0);
> +	} else {
> +		/* non burst transmission */
> +		null_pkt_size = 0;
> +
> +		/* bytes to be sent - first as one chunk */
> +		bytes_per_chunk = vm->hactive * Bpp_x100 / 100 + pkt_header;
> +
> +		/* hline total bytes from the DPI interface */
> +		total_bytes = (vm->hactive + vm->hfront_porch) *
> +				ratio_x1000 / ctx->lanes / 1000;
> +
> +		/* check if the pixels actually fit on the DSI link */
> +		if (total_bytes < bytes_per_chunk) {
> +			drm_err(dsi->drm, "current resolution can not be set\n");
> +			return -EINVAL;
> +		}
> +
> +		chunk_overhead = total_bytes - bytes_per_chunk;
> +
> +		/* overhead higher than 1 -> enable multi packets */
> +		if (chunk_overhead > 1) {
> +
> +			/* multi packets */
> +			for (video_size = video_size_step;
> +			     video_size < vm->hactive;
> +			     video_size += video_size_step) {
> +
> +				if (vm->hactive * 1000 / video_size % 1000)
> +					continue;
> +
> +				chunks = vm->hactive / video_size;
> +				bytes_per_chunk = Bpp_x100 * video_size / 100
> +						  + pkt_header;
> +				if (total_bytes >= (bytes_per_chunk * chunks)) {
> +					bytes_left = total_bytes -
> +						     bytes_per_chunk * chunks;
> +					break;
> +				}
> +			}
> +
> +			/* prevent overflow (unsigned - unsigned) */
> +			if (bytes_left > (pkt_header * chunks)) {
> +				null_pkt_size = (bytes_left -
> +						pkt_header * chunks) / chunks;
> +				/* avoid register overflow */
> +				if (null_pkt_size > 1023)
> +					null_pkt_size = 1023;
> +			}
> +
> +		} else {
> +
> +			/* single packet */
> +			chunks = 1;
> +
> +			/* must be a multiple of 4 except 18 loosely */
> +			for (video_size = vm->hactive;
> +			    (video_size % video_size_step) != 0;
> +			     video_size++)
> +				;
> +		}
> +
> +		dsi_dpi_video_packet_size(ctx, video_size);
> +		dsi_dpi_null_packet_size(ctx, null_pkt_size);
> +		dsi_dpi_chunk_num(ctx, chunks);
> +	}
> +
> +	dsi_int0_mask(ctx, ctx->int0_mask);
> +	dsi_int1_mask(ctx, ctx->int1_mask);
> +	dsi_power_enable(ctx, 1);
> +
> +	return 0;
> +}
> +
> +static void sprd_dsi_edpi_video(struct dsi_context *ctx)
> +{
> +	const u32 fifo_depth = 1096;
> +	const u32 word_length = 4;
> +	u32 hactive = ctx->vm.hactive;
> +	u32 Bpp_x100;
> +	u32 max_fifo_len;
> +	u8 coding;
> +
> +	coding = fmt_to_coding(ctx->format);
> +	Bpp_x100 = calc_bytes_per_pixel_x100(coding);
> +	max_fifo_len = word_length * fifo_depth * 100 / Bpp_x100;
> +
> +	dsi_power_enable(ctx, 0);
> +	dsi_dpi_color_coding(ctx, coding);
> +	dsi_tear_effect_ack_en(ctx, ctx->te_ack_en);
> +
> +	if (max_fifo_len > hactive)
> +		dsi_edpi_max_pkt_size(ctx, hactive);
> +	else
> +		dsi_edpi_max_pkt_size(ctx, max_fifo_len);
> +
> +	dsi_int0_mask(ctx, ctx->int0_mask);
> +	dsi_int1_mask(ctx, ctx->int1_mask);
> +	dsi_power_enable(ctx, 1);
> +}
> +
> +/*
> + * Send a packet on the generic interface,
> + * this function has an active delay to wait for the buffer to clear.
> + * The delay is limited to:
> + * (param_length / 4) x DSIH_FIFO_ACTIVE_WAIT x register access time
> + * the controller restricts the sending of.
> + *
> + * This function will not be able to send Null and Blanking packets due to
> + * controller restriction
> + */
> +static int sprd_dsi_wr_pkt(struct dsi_context *ctx, u8 vc, u8 type,
> +			const u8 *param, u16 len)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	u8 wc_lsbyte, wc_msbyte;
> +	u32 payload;
> +	int i, j, ret;
> +
> +	if (vc > 3)
> +		return -EINVAL;
> +
> +
> +	/* 1st: for long packet, must config payload first */
> +	ret = dsi_wait_tx_payload_fifo_empty(ctx);
> +	if (ret) {
> +		drm_err(dsi->drm, "tx payload fifo is not empty\n");
> +		return ret;
> +	}
> +
> +	if (len > 2) {
> +		for (i = 0, j = 0; i < len; i += j) {
> +			payload = 0;
> +			for (j = 0; (j < 4) && ((j + i) < (len)); j++)
> +				payload |= param[i + j] << (j * 8);
> +
> +			dsi_set_packet_payload(ctx, payload);
> +		}
> +		wc_lsbyte = len & 0xff;
> +		wc_msbyte = len >> 8;
> +	} else {
> +		wc_lsbyte = (len > 0) ? param[0] : 0;
> +		wc_msbyte = (len > 1) ? param[1] : 0;
> +	}
> +
> +	/* 2nd: then set packet header */
> +	ret = dsi_wait_tx_cmd_fifo_empty(ctx);
> +	if (ret) {
> +		drm_err(dsi->drm, "tx cmd fifo is not empty\n");
> +		return ret;
> +	}
> +
> +	dsi_set_packet_header(ctx, vc, type, wc_lsbyte, wc_msbyte);
> +
> +	return 0;
> +}
> +
> +/*
> + * Send READ packet to peripheral using the generic interface,
> + * this will force command mode and stop video mode (because of BTA).
> + *
> + * This function has an active delay to wait for the buffer to clear,
> + * the delay is limited to 2 x DSIH_FIFO_ACTIVE_WAIT
> + * (waiting for command buffer, and waiting for receiving)
> + * @note this function will enable BTA
> + */
> +static int sprd_dsi_rd_pkt(struct dsi_context *ctx, u8 vc, u8 type,
> +			u8 msb_byte, u8 lsb_byte,
> +			u8 *buffer, u8 bytes_to_read)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	int i, ret;
> +	int count = 0;
> +	u32 temp;
> +
> +	if (vc > 3)
> +		return -EINVAL;
> +
> +	/* 1st: send read command to peripheral */
> +	if (!dsi_is_tx_cmd_fifo_empty(ctx))
> +		return -EIO;
> +
> +	dsi_set_packet_header(ctx, vc, type, lsb_byte, msb_byte);
> +
> +	/* 2nd: wait peripheral response completed */
> +	ret = dsi_wait_rd_resp_completed(ctx);
> +	if (ret) {
> +		drm_err(dsi->drm, "wait read response time out\n");
> +		return ret;
> +	}
> +
> +	/* 3rd: get data from rx payload fifo */
> +	if (dsi_is_rx_payload_fifo_empty(ctx)) {
> +		drm_err(dsi->drm, "rx payload fifo empty\n");
> +		return -EIO;
> +	}
> +
> +	for (i = 0; i < 100; i++) {
> +		temp = dsi_get_rx_payload(ctx);
> +
> +		if (count < bytes_to_read)
> +			buffer[count++] = temp & 0xff;
> +		if (count < bytes_to_read)
> +			buffer[count++] = (temp >> 8) & 0xff;
> +		if (count < bytes_to_read)
> +			buffer[count++] = (temp >> 16) & 0xff;
> +		if (count < bytes_to_read)
> +			buffer[count++] = (temp >> 24) & 0xff;
> +
> +		if (dsi_is_rx_payload_fifo_empty(ctx))
> +			return count;
> +	}
> +
> +	return 0;
> +}
> +
> +static void sprd_dsi_set_work_mode(struct dsi_context *ctx, u8 mode)
> +{
> +	if (mode == DSI_MODE_CMD)
> +		dsi_cmd_mode(ctx);
> +	else
> +		dsi_video_mode(ctx);
> +}
> +
> +static void sprd_dsi_lp_cmd_enable(struct dsi_context *ctx, bool enable)
> +{
> +	if (dsi_is_cmd_mode(ctx))
> +		dsi_cmd_mode_lp_cmd_en(ctx, enable);
> +	else
> +		dsi_video_mode_lp_cmd_en(ctx, enable);
> +}
> +
> +static void sprd_dsi_state_reset(struct dsi_context *ctx)
> +{
> +	dsi_power_enable(ctx, 0);
> +	udelay(100);
> +	dsi_power_enable(ctx, 1);
> +}
> +
> +static u32 sprd_dsi_int_status(struct dsi_context *ctx, int index)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	u32 status;
> +
> +	if (index == 0)
> +		status = dsi_int0_status(ctx);
> +	else if (index == 1)
> +		status = dsi_int1_status(ctx);
> +	else {
> +		drm_err(dsi->drm, "invalid dsi IRQ index %d\n", index);
> +		status = -EINVAL;
> +	}
> +
> +	return status;
> +}
> +
> +static int sprd_dphy_init(struct dsi_context *ctx)
> +{
> +	struct sprd_dsi *dsi = container_of(ctx, struct sprd_dsi, ctx);
> +	int ret;
> +
> +	dsi_phy_rstz(ctx, 0);
> +	dsi_phy_shutdownz(ctx, 0);
> +	dsi_phy_clklane_en(ctx, 0);
> +
> +	dsi_phy_test_clr(ctx, 0);
> +	dsi_phy_test_clr(ctx, 1);
> +	dsi_phy_test_clr(ctx, 0);
> +
> +	dphy_pll_config(ctx);
> +	dphy_timing_config(ctx);
> +
> +	dsi_phy_shutdownz(ctx, 1);
> +	dsi_phy_rstz(ctx, 1);
> +	dsi_phy_stop_wait_time(ctx, 0x1C);
> +	dsi_phy_clklane_en(ctx, 1);
> +	dsi_phy_datalane_en(ctx);
> +
> +	ret = dphy_wait_pll_locked(ctx);
> +	if (ret) {
> +		drm_err(dsi->drm, "dphy initial failed\n");
> +		return ret;
> +	}
> +
> +	return 0;
> +}
> +
> +static void sprd_dphy_fini(struct dsi_context *ctx)
> +{
> +	dsi_phy_rstz(ctx, 0);
> +	dsi_phy_shutdownz(ctx, 0);
> +	dsi_phy_rstz(ctx, 1);
> +}
> +
> +static void sprd_dsi_encoder_enable(struct drm_encoder *encoder)
> +{
> +	struct sprd_dsi *dsi = encoder_to_dsi(encoder);
> +	struct sprd_dpu *dpu = to_sprd_crtc(encoder->crtc);
> +	struct dsi_context *ctx = &dsi->ctx;
> +
> +	if (ctx->enabled) {
> +		drm_warn(dsi->drm, "dsi is initialized\n");
> +		return;
> +	}
> +
> +	sprd_dsi_init(ctx);
> +	if (ctx->work_mode == DSI_MODE_VIDEO)
> +		sprd_dsi_dpi_video(ctx);
> +	else
> +		sprd_dsi_edpi_video(ctx);
> +
> +	sprd_dphy_init(ctx);
> +
> +	sprd_dsi_lp_cmd_enable(ctx, true);
> +
> +	if (dsi->panel) {
> +		drm_panel_prepare(dsi->panel);
> +		drm_panel_enable(dsi->panel);
> +	}
> +
> +	sprd_dsi_set_work_mode(ctx, ctx->work_mode);
> +	sprd_dsi_state_reset(ctx);
> +
> +	if (ctx->nc_clk_en)
> +		dsi_nc_clk_en(ctx, true);
> +	else {
> +		dsi_phy_clk_hs_rqst(ctx, true);
> +		dphy_wait_pll_locked(ctx);
> +	}
> +
> +	sprd_dpu_run(dpu);
> +
> +	ctx->enabled = true;
> +}
> +
> +static void sprd_dsi_encoder_disable(struct drm_encoder *encoder)
> +{
> +	struct sprd_dsi *dsi = encoder_to_dsi(encoder);
> +	struct sprd_dpu *dpu = to_sprd_crtc(encoder->crtc);
> +	struct dsi_context *ctx = &dsi->ctx;
> +
> +	if (!ctx->enabled) {
> +		drm_warn(dsi->drm, "dsi isn't initialized\n");
> +		return;
> +	}
> +
> +	sprd_dpu_stop(dpu);
> +	sprd_dsi_set_work_mode(ctx, DSI_MODE_CMD);
> +	sprd_dsi_lp_cmd_enable(ctx, true);
> +
> +	if (dsi->panel) {
> +		drm_panel_disable(dsi->panel);
> +		drm_panel_unprepare(dsi->panel);
> +	}
> +
> +	sprd_dphy_fini(ctx);
> +	sprd_dsi_fini(ctx);
> +
> +	ctx->enabled = false;
> +}
> +
> +static void sprd_dsi_encoder_mode_set(struct drm_encoder *encoder,
> +				 struct drm_display_mode *mode,
> +				 struct drm_display_mode *adj_mode)
> +{
> +	struct sprd_dsi *dsi = encoder_to_dsi(encoder);
> +
> +	drm_dbg(dsi->drm, "%s() set mode: %s\n", __func__, dsi->mode->name);
> +}

You can remove that function

> +static int sprd_dsi_encoder_atomic_check(struct drm_encoder *encoder,
> +				    struct drm_crtc_state *crtc_state,
> +				    struct drm_connector_state *conn_state)
> +{
> +	return 0;
> +}

And this one too

> +static const struct drm_encoder_helper_funcs sprd_encoder_helper_funcs = {
> +	.atomic_check	= sprd_dsi_encoder_atomic_check,
> +	.mode_set	= sprd_dsi_encoder_mode_set,
> +	.enable		= sprd_dsi_encoder_enable,
> +	.disable	= sprd_dsi_encoder_disable
> +};
> +
> +static const struct drm_encoder_funcs sprd_encoder_funcs = {
> +	.destroy = drm_encoder_cleanup,
> +};
> +
> +static struct sprd_dsi *sprd_dsi_encoder_init(struct drm_device *drm,
> +			       struct device *dev)
> +{
> +	struct sprd_dsi *dsi;
> +	u32 crtc_mask;
> +
> +	crtc_mask = drm_of_find_possible_crtcs(drm, dev->of_node);
> +	if (!crtc_mask) {
> +		drm_err(drm, "failed to find crtc mask\n");
> +		return ERR_PTR(-EINVAL);
> +	}
> +
> +	drm_dbg(drm, "find possible crtcs: 0x%08x\n", crtc_mask);
> +
> +	dsi = drmm_encoder_alloc(drm, struct sprd_dsi, encoder,
> +			       &sprd_encoder_funcs, DRM_MODE_ENCODER_DSI, NULL);
> +	if (IS_ERR(dsi)) {
> +		drm_err(drm, "failed to init dsi encoder.\n");
> +		return dsi;
> +	}
> +
> +	dsi->encoder.possible_crtcs = crtc_mask;
> +	drm_encoder_helper_add(&dsi->encoder, &sprd_encoder_helper_funcs);
> +
> +	return dsi;
> +}
> +
> +static int sprd_dsi_find_panel(struct sprd_dsi *dsi)
> +{
> +	struct device *dev = dsi->host.dev;
> +	struct device_node *child, *lcds_node;
> +	struct drm_panel *panel;
> +
> +	/* search /lcds child node first */
> +	lcds_node = of_find_node_by_path("/lcds");
> +	for_each_child_of_node(lcds_node, child) {
> +		panel = of_drm_find_panel(child);
> +		if (!IS_ERR(panel)) {
> +			dsi->panel = panel;
> +			return 0;
> +		}
> +	}

That's not part of your binding

> +
> +	/*
> +	 * If /lcds child node search failed, we search
> +	 * the child of dsi host node.
> +	 */
> +	for_each_child_of_node(dev->of_node, child) {
> +		panel = of_drm_find_panel(child);
> +		if (!IS_ERR(panel)) {
> +			dsi->panel = panel;
> +			return 0;
> +		}
> +	}

And you don't need this either. You'll register a mipi_dsi_host, that
will in turn probe all the devices under that bus, and will then call
the .attach callback.

> +	drm_err(dsi->drm, "of_drm_find_panel() failed\n");
> +	return -ENODEV;
> +}
> +
> +static int sprd_dsi_host_attach(struct mipi_dsi_host *host,
> +			   struct mipi_dsi_device *slave)
> +{
> +	struct sprd_dsi *dsi = host_to_dsi(host);
> +	struct dsi_context *ctx = &dsi->ctx;
> +	int ret;
> +
> +	dsi->slave = slave;
> +	ctx->lanes = slave->lanes;
> +	ctx->format = slave->format;
> +	ctx->byte_clk = slave->hs_rate / 8;
> +	ctx->esc_clk = slave->lp_rate;
> +
> +	if (slave->mode_flags & MIPI_DSI_MODE_VIDEO)
> +		ctx->work_mode = DSI_MODE_VIDEO;
> +	else
> +		ctx->work_mode = DSI_MODE_CMD;
> +
> +	if (slave->mode_flags & MIPI_DSI_MODE_VIDEO_BURST)
> +		ctx->burst_mode = VIDEO_BURST_WITH_SYNC_PULSES;
> +	else if (slave->mode_flags & MIPI_DSI_MODE_VIDEO_SYNC_PULSE)
> +		ctx->burst_mode = VIDEO_NON_BURST_WITH_SYNC_PULSES;
> +	else
> +		ctx->burst_mode = VIDEO_NON_BURST_WITH_SYNC_EVENTS;
> +
> +	if (slave->mode_flags & MIPI_DSI_CLOCK_NON_CONTINUOUS)
> +		ctx->nc_clk_en = true;

I'm not sure why you need to duplicate all this, can't you just
dereference the dsi->slave pointer when you need it?

> +	ret = sprd_dsi_find_panel(dsi);
> +	if (ret)
> +		return ret;

The panel is your slave here

> +	return 0;
> +}
> +
> +static int sprd_dsi_host_detach(struct mipi_dsi_host *host,
> +			   struct mipi_dsi_device *slave)
> +{
> +	/* do nothing */
> +	return 0;
> +}
> +
> +static ssize_t sprd_dsi_host_transfer(struct mipi_dsi_host *host,
> +				const struct mipi_dsi_msg *msg)
> +{
> +	struct sprd_dsi *dsi = host_to_dsi(host);
> +	const u8 *tx_buf = msg->tx_buf;
> +
> +	if (msg->rx_buf && msg->rx_len) {
> +		u8 lsb = (msg->tx_len > 0) ? tx_buf[0] : 0;
> +		u8 msb = (msg->tx_len > 1) ? tx_buf[1] : 0;
> +
> +		return sprd_dsi_rd_pkt(&dsi->ctx, msg->channel, msg->type,
> +				msb, lsb, msg->rx_buf, msg->rx_len);
> +	}
> +
> +	if (msg->tx_buf && msg->tx_len)
> +		return sprd_dsi_wr_pkt(&dsi->ctx, msg->channel, msg->type,
> +					tx_buf, msg->tx_len);
> +
> +	return 0;
> +}
> +
> +static const struct mipi_dsi_host_ops sprd_dsi_host_ops = {
> +	.attach = sprd_dsi_host_attach,
> +	.detach = sprd_dsi_host_detach,
> +	.transfer = sprd_dsi_host_transfer,
> +};
> +
> +static int sprd_dsi_host_init(struct sprd_dsi *dsi, struct device *dev)
> +{
> +	int ret;
> +
> +	dsi->host.dev = dev;
> +	dsi->host.ops = &sprd_dsi_host_ops;
> +
> +	ret = mipi_dsi_host_register(&dsi->host);
> +	if (ret)
> +		drm_err(dsi->drm, "failed to register dsi host\n");
> +
> +	return ret;
> +}
> +
> +static int sprd_dsi_connector_get_modes(struct drm_connector *connector)
> +{
> +	struct sprd_dsi *dsi = connector_to_dsi(connector);
> +
> +	return drm_panel_get_modes(dsi->panel, connector);
> +}
> +
> +static enum drm_mode_status
> +sprd_dsi_connector_mode_valid(struct drm_connector *connector,
> +			 struct drm_display_mode *mode)
> +{
> +	struct sprd_dsi *dsi = connector_to_dsi(connector);
> +
> +	drm_dbg(dsi->drm, "%s() mode: "DRM_MODE_FMT"\n", __func__, DRM_MODE_ARG(mode));
> +
> +	if (mode->type & DRM_MODE_TYPE_PREFERRED) {
> +		dsi->mode = mode;
> +		drm_display_mode_to_videomode(dsi->mode, &dsi->ctx.vm);
> +	}

Again, what happens if the mode isn't the preferred one?

You shouldn't be testing this but rather just use the crtc_state
adjusted_mode.

Maxime
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 228 bytes
Desc: not available
URL: <https://lists.freedesktop.org/archives/dri-devel/attachments/20210430/aeee8ea3/attachment-0001.sig>


More information about the dri-devel mailing list