Эх сурвалжийг харах

disp: msm: sde: Update LTM merge mode setting for kailua

Update LTM merge mode setting for kailua since
merge control has its own register.

Change-Id: Ieaacd1e12f410def18a0fce11a77c94832c416f5
Signed-off-by: Renchao Liu <[email protected]>
Renchao Liu 3 жил өмнө
parent
commit
f4730e436e

+ 88 - 24
msm/sde/sde_hw_color_proc_v4.c

@@ -1,5 +1,6 @@
 // SPDX-License-Identifier: GPL-2.0-only
 /*
+ * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
  * Copyright (c) 2017-2021, The Linux Foundation. All rights reserved.
  */
 #include <drm/msm_drm_pp.h>
@@ -316,8 +317,40 @@ void sde_setup_dspp_ltm_hist_bufferv1(struct sde_hw_dspp *ctx, u64 addr)
 			(hs_addr & 0xFFFFFF00));
 }
 
+static void sde_setup_dspp_ltm_hist_ctrl_common(struct sde_hw_dspp *ctx,
+					u64 addr, u32 op_mode,
+					struct sde_ltm_phase_info *phase)
+{
+	u32 offset;
+
+	if (ctx->idx >= DSPP_MAX) {
+		DRM_ERROR("Invalid idx %d\n", ctx->idx);
+		return;
+	}
+
+	if (phase->portrait_en)
+		op_mode |= BIT(2);
+	else
+		op_mode &= ~BIT(2);
+
+	offset = ctx->cap->sblk->ltm.base + 0x8;
+	SDE_REG_WRITE(&ctx->hw, offset, (phase->init_h[ctx->idx] & 0x7FFFFFF));
+	offset += 4;
+	SDE_REG_WRITE(&ctx->hw, offset, (phase->init_v & 0xFFFFFF));
+	offset += 4;
+	SDE_REG_WRITE(&ctx->hw, offset, (phase->inc_h & 0xFFFFFF));
+	offset += 4;
+	SDE_REG_WRITE(&ctx->hw, offset, (phase->inc_v & 0xFFFFFF));
+
+	op_mode |= BIT(0);
+	sde_setup_dspp_ltm_hist_bufferv1(ctx, addr);
+
+	SDE_REG_WRITE(&ctx->hw, ctx->cap->sblk->ltm.base + 0x4,
+			(op_mode & 0x1FFFFFF));
+}
+
 void sde_setup_dspp_ltm_hist_ctrlv1(struct sde_hw_dspp *ctx, void *cfg,
-				    bool enable, u64 addr)
+					bool enable, u64 addr)
 {
 	struct sde_hw_cp_cfg *hw_cfg = cfg;
 	struct sde_ltm_phase_info phase;
@@ -347,38 +380,58 @@ void sde_setup_dspp_ltm_hist_ctrlv1(struct sde_hw_dspp *ctx, void *cfg,
 		return;
 	}
 
-	if (ctx->idx >= DSPP_MAX) {
-		DRM_ERROR("Invalid idx %d\n", ctx->idx);
-		return;
-	}
-
 	memset(&phase, 0, sizeof(phase));
 	sde_ltm_get_phase_info(hw_cfg, &phase);
-
-	if (phase.portrait_en)
-		op_mode |= BIT(2);
-	else
-		op_mode &= ~BIT(2);
-
 	if (phase.merge_en)
 		op_mode |= BIT(16);
 	else
-		op_mode &= ~(BIT(16) | BIT(17));
+		op_mode &= ~LTM_CONFIG_MERGE_MODE_ONLY;
 
-	offset = ctx->cap->sblk->ltm.base + 0x8;
-	SDE_REG_WRITE(&ctx->hw, offset, (phase.init_h[ctx->idx] & 0x7FFFFFF));
-	offset += 4;
-	SDE_REG_WRITE(&ctx->hw, offset, (phase.init_v & 0xFFFFFF));
-	offset += 4;
-	SDE_REG_WRITE(&ctx->hw, offset, (phase.inc_h & 0xFFFFFF));
-	offset += 4;
-	SDE_REG_WRITE(&ctx->hw, offset, (phase.inc_v & 0xFFFFFF));
+	sde_setup_dspp_ltm_hist_ctrl_common(ctx, addr, op_mode, &phase);
+}
 
-	op_mode |= BIT(0);
-	sde_setup_dspp_ltm_hist_bufferv1(ctx, addr);
+void sde_setup_dspp_ltm_hist_ctrlv1_2(struct sde_hw_dspp *ctx, void *cfg,
+					bool enable, u64 addr)
+{
+	struct sde_hw_cp_cfg *hw_cfg = cfg;
+	struct sde_ltm_phase_info phase;
+	u32 op_mode, offset;
+	u32 merge_mode = 0;
 
-	SDE_REG_WRITE(&ctx->hw, ctx->cap->sblk->ltm.base + 0x4,
+	if (!ctx) {
+		DRM_ERROR("invalid parameters ctx %pK\n", ctx);
+		return;
+	}
+
+	if (enable && (!addr || !cfg)) {
+		DRM_ERROR("invalid addr 0x%llx cfg %pK\n", addr, cfg);
+		return;
+	}
+
+	offset = ctx->cap->sblk->ltm.base + 0x4;
+	op_mode = SDE_REG_READ(&ctx->hw, offset);
+	if (!enable) {
+		if (op_mode & BIT(1))
+			op_mode &= ~BIT(0);
+		else
+			op_mode = 0x0;
+
+		SDE_REG_WRITE(&ctx->hw, ctx->cap->sblk->ltm.base + 0x4,
 			(op_mode & 0x1FFFFFF));
+		return;
+	}
+
+	memset(&phase, 0, sizeof(phase));
+	sde_ltm_get_phase_info(hw_cfg, &phase);
+
+	offset = ctx->cap->sblk->ltm.base + 0x18;
+	if (phase.merge_en)
+		merge_mode = BIT(0);
+	else
+		merge_mode = 0x0;
+	SDE_REG_WRITE(&ctx->hw, offset, (merge_mode & 0x3));
+
+	sde_setup_dspp_ltm_hist_ctrl_common(ctx, addr, op_mode, &phase);
 }
 
 void sde_ltm_read_intr_status(struct sde_hw_dspp *ctx, u32 *status)
@@ -399,6 +452,17 @@ void sde_ltm_read_intr_status(struct sde_hw_dspp *ctx, u32 *status)
 	SDE_REG_WRITE(&ctx->hw, ctx->cap->sblk->ltm.base + 0x58, clear);
 }
 
+void sde_ltm_clear_merge_modev1_2(struct sde_hw_dspp *ctx)
+{
+	if (!ctx) {
+		DRM_ERROR("invalid parameters ctx %pK\n", ctx);
+		return;
+	}
+
+	/* clear the merge_mode bit */
+	SDE_REG_WRITE(&ctx->hw, ctx->cap->sblk->ltm.base + 0x18, 0x0);
+}
+
 void sde_ltm_clear_merge_mode(struct sde_hw_dspp *ctx)
 {
 	u32 clear;

+ 18 - 0
msm/sde/sde_hw_color_proc_v4.h

@@ -1,5 +1,6 @@
 /* SPDX-License-Identifier: GPL-2.0-only */
 /*
+ * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
  * Copyright (c) 2017-2019, 2021 The Linux Foundation. All rights reserved.
  */
 #ifndef _SDE_HW_COLOR_PROC_V4_H_
@@ -58,6 +59,17 @@ void sde_setup_dspp_ltm_threshv1(struct sde_hw_dspp *ctx, void *cfg);
  */
 void sde_setup_dspp_ltm_hist_ctrlv1(struct sde_hw_dspp *ctx, void *cfg,
 				    bool enable, u64 addr);
+
+/**
+ * sde_setup_dspp_ltm_hist_ctrlv1_2 - Same as v1 except ltm merge mode setting.
+ * @ctx: dspp ctx pointer
+ * @cfg: pointer to sde_hw_cp_cfg
+ * @enable: feature enable/disable value
+ * @addr: aligned iova address
+ */
+void sde_setup_dspp_ltm_hist_ctrlv1_2(struct sde_hw_dspp *ctx, void *cfg,
+				    bool enable, u64 addr);
+
 /**
  * sde_setup_dspp_ltm_hist_bufferv1 - Function for setting ltm hist buffer v1.
  * @ctx: dspp ctx pointer
@@ -78,6 +90,12 @@ void sde_ltm_read_intr_status(struct sde_hw_dspp *dspp, u32 *status);
  */
 void sde_ltm_clear_merge_mode(struct sde_hw_dspp *dspp);
 
+/**
+ * sde_ltm_clear_merge_modev1_2 - api to clear ltm merge_mode
+ * @dspp: pointer to dspp object
+ */
+void sde_ltm_clear_merge_modev1_2(struct sde_hw_dspp *dspp);
+
 
 /**
  * sde_demura_backlight_cfg - api to set backlight for demura

+ 24 - 8
msm/sde/sde_hw_dspp.c

@@ -1,6 +1,6 @@
 // SPDX-License-Identifier: GPL-2.0-only
 /*
- * Copyright (c) 2021 Qualcomm Innovation Center, Inc. All rights reserved.
+ * Copyright (c) 2021-2022 Qualcomm Innovation Center, Inc. All rights reserved.
  * Copyright (c) 2015-2021, The Linux Foundation. All rights reserved.
  */
 
@@ -231,7 +231,8 @@ static void dspp_ltm(struct sde_hw_dspp *c)
 	int ret = 0;
 
 	if (c->cap->sblk->ltm.version == SDE_COLOR_PROCESS_VER(0x1, 0x0) ||
-		c->cap->sblk->ltm.version == SDE_COLOR_PROCESS_VER(0x1, 0x1)) {
+		c->cap->sblk->ltm.version == SDE_COLOR_PROCESS_VER(0x1, 0x1) ||
+		c->cap->sblk->ltm.version == SDE_COLOR_PROCESS_VER(0x1, 0x2)) {
 		ret = reg_dmav1_init_ltm_op_v6(SDE_LTM_INIT, c->idx);
 		if (!ret)
 			ret = reg_dmav1_init_ltm_op_v6(SDE_LTM_ROI, c->idx);
@@ -239,16 +240,29 @@ static void dspp_ltm(struct sde_hw_dspp *c)
 			ret = reg_dmav1_init_ltm_op_v6(SDE_LTM_VLUT, c->idx);
 
 		if (!ret) {
+			if (c->cap->sblk->ltm.version ==
+				SDE_COLOR_PROCESS_VER(0x1, 0x2)) {
+				c->ops.setup_ltm_vlut =
+					reg_dmav1_setup_ltm_vlutv1_2;
+				c->ops.setup_ltm_hist_ctrl =
+					sde_setup_dspp_ltm_hist_ctrlv1_2;
+				c->ops.clear_ltm_merge_mode =
+					sde_ltm_clear_merge_modev1_2;
+			} else {
+				c->ops.setup_ltm_vlut =
+					reg_dmav1_setup_ltm_vlutv1;
+				c->ops.setup_ltm_hist_ctrl =
+					sde_setup_dspp_ltm_hist_ctrlv1;
+				c->ops.clear_ltm_merge_mode =
+					sde_ltm_clear_merge_mode;
+			}
+
 			c->ops.setup_ltm_init = reg_dmav1_setup_ltm_initv1;
 			c->ops.setup_ltm_roi = reg_dmav1_setup_ltm_roiv1;
-			c->ops.setup_ltm_vlut = reg_dmav1_setup_ltm_vlutv1;
 			c->ops.setup_ltm_thresh = sde_setup_dspp_ltm_threshv1;
-			c->ops.setup_ltm_hist_ctrl =
-				sde_setup_dspp_ltm_hist_ctrlv1;
 			c->ops.setup_ltm_hist_buffer =
 				sde_setup_dspp_ltm_hist_bufferv1;
 			c->ops.ltm_read_intr_status = sde_ltm_read_intr_status;
-			c->ops.clear_ltm_merge_mode = sde_ltm_clear_merge_mode;
 		} else {
 			c->ops.setup_ltm_init = NULL;
 			c->ops.setup_ltm_roi = NULL;
@@ -259,8 +273,10 @@ static void dspp_ltm(struct sde_hw_dspp *c)
 			c->ops.ltm_read_intr_status = NULL;
 			c->ops.clear_ltm_merge_mode = NULL;
 		}
-		if (!ret && c->cap->sblk->ltm.version ==
-			SDE_COLOR_PROCESS_VER(0x1, 0x1))
+		if (!ret && (c->cap->sblk->ltm.version ==
+			SDE_COLOR_PROCESS_VER(0x1, 0x1) ||
+			c->cap->sblk->ltm.version ==
+			SDE_COLOR_PROCESS_VER(0x1, 0x2)))
 			c->ltm_checksum_support = true;
 		else
 			c->ltm_checksum_support = false;

+ 170 - 56
msm/sde/sde_hw_reg_dma_v1_color_proc.c

@@ -1,6 +1,6 @@
 // SPDX-License-Identifier: GPL-2.0-only
 /*
- * Copyright (c) 2021 Qualcomm Innovation Center, Inc. All rights reserved.
+ * Copyright (c) 2021-2022 Qualcomm Innovation Center, Inc. All rights reserved.
  * Copyright (c) 2017-2021, The Linux Foundation. All rights reserved.
  */
 
@@ -3774,7 +3774,7 @@ void reg_dmav1_setup_ltm_roiv1(struct sde_hw_dspp *ctx, void *cfg)
 	}
 }
 
-static void ltm_vlutv1_disable(struct sde_hw_dspp *ctx)
+static void ltm_vlutv1_disable(struct sde_hw_dspp *ctx, u32 clear)
 {
 	enum sde_ltm idx = 0;
 	u32 opmode = 0, offset = 0;
@@ -3792,132 +3792,165 @@ static void ltm_vlutv1_disable(struct sde_hw_dspp *ctx)
 		/* disable VLUT/INIT/ROI */
 		opmode &= REG_DMA_LTM_VLUT_DISABLE_OP_MASK;
 	else
-		opmode &= LTM_CONFIG_MERGE_MODE_ONLY;
+		opmode &= clear;
 	SDE_REG_WRITE(&ctx->hw, offset, opmode);
 }
 
-void reg_dmav1_setup_ltm_vlutv1(struct sde_hw_dspp *ctx, void *cfg)
+static int reg_dmav1_setup_ltm_vlutv1_common(struct sde_hw_dspp *ctx, void *cfg,
+				struct sde_hw_reg_dma_ops *dma_ops,
+				struct sde_reg_dma_setup_ops_cfg *dma_write_cfg,
+				u32 *opmode, enum sde_ltm *dspp_idx)
 {
 	struct drm_msm_ltm_data *payload = NULL;
-	struct sde_reg_dma_setup_ops_cfg dma_write_cfg;
 	struct sde_hw_cp_cfg *hw_cfg = cfg;
-	struct sde_reg_dma_kickoff_cfg kick_off;
-	struct sde_hw_reg_dma_ops *dma_ops;
-	struct sde_ltm_phase_info phase;
-	enum sde_ltm dspp_idx[LTM_MAX] = {0};
 	enum sde_ltm idx = 0;
-	u32 offset, crs = 0, index = 0, len = 0, blk = 0, opmode = 0;
+	u32 offset, crs = 0, index = 0, len = 0, blk = 0;
 	u32 i = 0, num_mixers = 0;
 	int rc = 0;
 
-	rc = reg_dma_ltm_check(ctx, cfg, LTM_VLUT);
-	if (rc)
-		return;
-
-	/* disable case */
-	if (!hw_cfg->payload) {
-		DRM_DEBUG_DRIVER("Disable LTM vlut feature\n");
-		LOG_FEATURE_OFF;
-		ltm_vlutv1_disable(ctx);
-		return;
-	}
-
 	idx = (enum sde_ltm)ctx->idx;
 	num_mixers = hw_cfg->num_of_mixers;
 	rc = reg_dmav1_get_ltm_blk(hw_cfg, idx, &dspp_idx[0], &blk);
 	if (rc) {
 		if (rc != -EALREADY)
 			DRM_ERROR("failed to get the blk info\n");
-		return;
+		return -EINVAL;
 	}
 
 	if (hw_cfg->len != sizeof(struct drm_msm_ltm_data)) {
 		DRM_ERROR("invalid size of payload len %d exp %zd\n",
 				hw_cfg->len, sizeof(struct drm_msm_ltm_data));
-		return;
+		return -EINVAL;
 	}
 
 	offset = ctx->cap->sblk->ltm.base + 0x5c;
 	crs = SDE_REG_READ(&ctx->hw, offset);
 	if (!(crs & BIT(3))) {
 		DRM_ERROR("LTM VLUT buffer is not ready: crs = %d\n", crs);
-		return;
+		return -EINVAL;
 	}
 
-	dma_ops = sde_reg_dma_get_ops();
 	dma_ops->reset_reg_dma_buf(ltm_buf[LTM_VLUT][idx]);
 
-	REG_DMA_INIT_OPS(dma_write_cfg, blk, LTM_VLUT, ltm_buf[LTM_VLUT][idx]);
-	REG_DMA_SETUP_OPS(dma_write_cfg, 0, NULL, 0, HW_BLK_SELECT, 0, 0, 0);
-	rc = dma_ops->setup_payload(&dma_write_cfg);
+	REG_DMA_INIT_OPS(*dma_write_cfg, blk, LTM_VLUT, ltm_buf[LTM_VLUT][idx]);
+	REG_DMA_SETUP_OPS(*dma_write_cfg, 0, NULL, 0, HW_BLK_SELECT, 0, 0, 0);
+	rc = dma_ops->setup_payload(dma_write_cfg);
 	if (rc) {
 		DRM_ERROR("write decode select failed ret %d\n", rc);
-		return;
+		return -EINVAL;
 	}
 
 	/* write VLUT index */
-	REG_DMA_SETUP_OPS(dma_write_cfg, 0x38, &index, sizeof(u32),
+	REG_DMA_SETUP_OPS(*dma_write_cfg, 0x38, &index, sizeof(u32),
 				REG_SINGLE_WRITE, 0, 0, 0);
-	rc = dma_ops->setup_payload(&dma_write_cfg);
+	rc = dma_ops->setup_payload(dma_write_cfg);
 	if (rc) {
 		DRM_ERROR("write VLUT index reg failed ret %d\n", rc);
-		return;
+		return -EINVAL;
 	}
 
 	payload = hw_cfg->payload;
 	len = sizeof(u32) * LTM_DATA_SIZE_0 * LTM_DATA_SIZE_3;
-	REG_DMA_SETUP_OPS(dma_write_cfg, 0x3c, &payload->data[0][0],
+	REG_DMA_SETUP_OPS(*dma_write_cfg, 0x3c, &payload->data[0][0],
 			len, REG_BLK_WRITE_INC, 0, 0, 0);
-	rc = dma_ops->setup_payload(&dma_write_cfg);
+	rc = dma_ops->setup_payload(dma_write_cfg);
 	if (rc) {
 		DRM_ERROR("write VLUT data failed rc %d\n", rc);
-		return;
+		return -EINVAL;
 	}
 
-	sde_ltm_get_phase_info(hw_cfg, &phase);
 	for (i = 0; i < num_mixers; i++) {
 		/* broadcast feature is not supported with REG_SINGLE_MODIFY */
 		/* reset decode select to unicast */
-		dma_write_cfg.blk = ltm_mapping[dspp_idx[i]];
-		REG_DMA_SETUP_OPS(dma_write_cfg, 0, NULL, 0, HW_BLK_SELECT, 0,
+		dma_write_cfg->blk = ltm_mapping[dspp_idx[i]];
+		REG_DMA_SETUP_OPS(*dma_write_cfg, 0, NULL, 0, HW_BLK_SELECT, 0,
 				0, 0);
-		rc = dma_ops->setup_payload(&dma_write_cfg);
+		rc = dma_ops->setup_payload(dma_write_cfg);
 		if (rc) {
 			DRM_ERROR("write decode select failed ret %d\n", rc);
-			return;
+			return -EINVAL;
 		}
 
 		/* set the UPDATE_REQ bit */
 		crs = BIT(0);
-		REG_DMA_SETUP_OPS(dma_write_cfg, 0x5c, &crs, sizeof(u32),
+		REG_DMA_SETUP_OPS(*dma_write_cfg, 0x5c, &crs, sizeof(u32),
 				REG_SINGLE_MODIFY, 0, 0,
 				REG_DMA_LTM_UPDATE_REQ_MASK);
-		rc = dma_ops->setup_payload(&dma_write_cfg);
+		rc = dma_ops->setup_payload(dma_write_cfg);
 		if (rc) {
 			DRM_ERROR("write UPDATE_REQ failed ret %d\n", rc);
-			return;
+			return -EINVAL;
 		}
-
-		opmode = BIT(1);
+		opmode[i] = BIT(1);
 		if (ltm_vlut_ops_mask[dspp_idx[i]] & ltm_unsharp)
-			opmode |= BIT(4);
+			opmode[i] |= BIT(4);
 		if (ltm_vlut_ops_mask[dspp_idx[i]] & ltm_dither)
-			opmode |= BIT(6);
+			opmode[i] |= BIT(6);
 		if (ltm_vlut_ops_mask[dspp_idx[i]] & ltm_roi)
-			opmode |= BIT(24);
+			opmode[i] |= BIT(24);
+		ltm_vlut_ops_mask[dspp_idx[i]] |= ltm_vlut;
+	}
+	return 0;
+}
+
+void reg_dmav1_setup_ltm_vlutv1(struct sde_hw_dspp *ctx, void *cfg)
+{
+	struct sde_reg_dma_setup_ops_cfg dma_write_cfg;
+	struct sde_hw_reg_dma_ops *dma_ops;
+	struct sde_ltm_phase_info phase;
+	struct sde_reg_dma_kickoff_cfg kick_off;
+	u32 *opmode;
+	u32 i = 0, num_mixers = 0;
+	int rc = 0;
+	struct sde_hw_cp_cfg *hw_cfg = cfg;
+	enum sde_ltm dspp_idx[LTM_MAX] = {0};
+	enum sde_ltm idx = 0;
+
+	rc = reg_dma_ltm_check(ctx, cfg, LTM_VLUT);
+	if (rc)
+		return;
+
+	/* disable case */
+	if (!hw_cfg->payload) {
+		DRM_DEBUG_DRIVER("Disable LTM vlut feature\n");
+		LOG_FEATURE_OFF;
+		ltm_vlutv1_disable(ctx, LTM_CONFIG_MERGE_MODE_ONLY);
+		return;
+	}
+
+	idx = (enum sde_ltm)ctx->idx;
+	num_mixers = hw_cfg->num_of_mixers;
+	opmode = kvzalloc((num_mixers * sizeof(u32)), GFP_KERNEL);
+	dma_ops = sde_reg_dma_get_ops();
+
+	rc = reg_dmav1_setup_ltm_vlutv1_common(ctx, cfg, dma_ops,
+					&dma_write_cfg, opmode, dspp_idx);
+	if (rc)
+		goto vlut_exit;
+
+	sde_ltm_get_phase_info(hw_cfg, &phase);
+	for (i = 0; i < num_mixers; i++) {
+		dma_write_cfg.blk = ltm_mapping[dspp_idx[i]];
+		REG_DMA_SETUP_OPS(dma_write_cfg, 0, NULL, 0, HW_BLK_SELECT, 0,
+				0, 0);
+		rc = dma_ops->setup_payload(&dma_write_cfg);
+		if (rc) {
+			DRM_ERROR("write decode select failed ret %d\n", rc);
+			goto vlut_exit;
+		}
+
 		if (phase.merge_en)
-			opmode |= BIT(16);
+			opmode[i] |= BIT(16);
 		else
-			opmode &= ~(BIT(16) | BIT(17));
-		ltm_vlut_ops_mask[dspp_idx[i]] |= ltm_vlut;
+			opmode[i] &= ~LTM_CONFIG_MERGE_MODE_ONLY;
 
-		REG_DMA_SETUP_OPS(dma_write_cfg, 0x4, &opmode, sizeof(u32),
+		REG_DMA_SETUP_OPS(dma_write_cfg, 0x4, &opmode[i], sizeof(u32),
 				REG_SINGLE_MODIFY, 0, 0,
 				REG_DMA_LTM_VLUT_ENABLE_OP_MASK);
 		rc = dma_ops->setup_payload(&dma_write_cfg);
 		if (rc) {
-			DRM_ERROR("write UPDATE_REQ failed ret %d\n", rc);
-			return;
+			DRM_ERROR("write opmode failed ret %d\n", rc);
+			goto vlut_exit;
 		}
 	}
 
@@ -3926,10 +3959,91 @@ void reg_dmav1_setup_ltm_vlutv1(struct sde_hw_dspp *ctx, void *cfg)
 				LTM_VLUT);
 	LOG_FEATURE_ON;
 	rc = dma_ops->kick_off(&kick_off);
-	if (rc) {
+	if (rc)
 		DRM_ERROR("failed to kick off ret %d\n", rc);
+vlut_exit:
+	kvfree(opmode);
+}
+
+void reg_dmav1_setup_ltm_vlutv1_2(struct sde_hw_dspp *ctx, void *cfg)
+{
+	struct sde_reg_dma_setup_ops_cfg dma_write_cfg;
+	struct sde_hw_reg_dma_ops *dma_ops;
+	struct sde_ltm_phase_info phase;
+	struct sde_reg_dma_kickoff_cfg kick_off;
+	u32 merge_mode = 0;
+	u32 *opmode;
+	u32 i = 0, num_mixers = 0;
+	int rc = 0;
+	struct sde_hw_cp_cfg *hw_cfg = cfg;
+	enum sde_ltm dspp_idx[LTM_MAX] = {0};
+	enum sde_ltm idx = 0;
+
+	rc = reg_dma_ltm_check(ctx, cfg, LTM_VLUT);
+	if (rc)
+		return;
+
+	/* disable case */
+	if (!hw_cfg->payload) {
+		DRM_DEBUG_DRIVER("Disable LTM vlut feature\n");
+		LOG_FEATURE_OFF;
+		ltm_vlutv1_disable(ctx, 0x0);
 		return;
 	}
+
+	idx = (enum sde_ltm)ctx->idx;
+	num_mixers = hw_cfg->num_of_mixers;
+	opmode = kvzalloc((num_mixers * sizeof(u32)), GFP_KERNEL);
+	dma_ops = sde_reg_dma_get_ops();
+
+	rc = reg_dmav1_setup_ltm_vlutv1_common(ctx, cfg, dma_ops,
+					&dma_write_cfg, opmode, dspp_idx);
+	if (rc)
+		goto vlut_exit;
+
+	sde_ltm_get_phase_info(hw_cfg, &phase);
+	for (i = 0; i < num_mixers; i++) {
+		dma_write_cfg.blk = ltm_mapping[dspp_idx[i]];
+		REG_DMA_SETUP_OPS(dma_write_cfg, 0, NULL, 0, HW_BLK_SELECT, 0,
+				0, 0);
+		rc = dma_ops->setup_payload(&dma_write_cfg);
+		if (rc) {
+			DRM_ERROR("write decode select failed ret %d\n", rc);
+			goto vlut_exit;
+		}
+
+		if (phase.merge_en)
+			merge_mode = BIT(0);
+		else
+			merge_mode = 0x0;
+		REG_DMA_SETUP_OPS(dma_write_cfg, 0x18, &merge_mode, sizeof(u32),
+				REG_SINGLE_MODIFY, 0, 0,
+				0xFFFFFFFC);
+		rc = dma_ops->setup_payload(&dma_write_cfg);
+		if (rc) {
+			DRM_ERROR("write merge_ctrl failed ret %d\n", rc);
+			goto vlut_exit;
+		}
+
+		REG_DMA_SETUP_OPS(dma_write_cfg, 0x4, &opmode[i], sizeof(u32),
+				REG_SINGLE_MODIFY, 0, 0,
+				REG_DMA_LTM_VLUT_ENABLE_OP_MASK);
+		rc = dma_ops->setup_payload(&dma_write_cfg);
+		if (rc) {
+			DRM_ERROR("write opmode failed ret %d\n", rc);
+			goto vlut_exit;
+		}
+	}
+
+	REG_DMA_SETUP_KICKOFF(kick_off, hw_cfg->ctl, ltm_buf[LTM_VLUT][idx],
+				REG_DMA_WRITE, DMA_CTL_QUEUE0, WRITE_IMMEDIATE,
+				LTM_VLUT);
+	LOG_FEATURE_ON;
+	rc = dma_ops->kick_off(&kick_off);
+	if (rc)
+		DRM_ERROR("failed to kick off ret %d\n", rc);
+vlut_exit:
+	kvfree(opmode);
 }
 
 int reg_dmav2_init_dspp_op_v4(int feature, enum sde_dspp idx)

+ 8 - 0
msm/sde/sde_hw_reg_dma_v1_color_proc.h

@@ -1,5 +1,6 @@
 /* SPDX-License-Identifier: GPL-2.0-only */
 /*
+ * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
  * Copyright (c) 2017-2020, The Linux Foundation. All rights reserved.
  */
 
@@ -251,6 +252,13 @@ void reg_dmav1_setup_ltm_roiv1(struct sde_hw_dspp *ctx, void *cfg);
  */
 void reg_dmav1_setup_ltm_vlutv1(struct sde_hw_dspp *ctx, void *cfg);
 
+/**
+ * reg_dmav1_setup_ltm_vlutv1_2() - Same as v1 except ltm merge mode setting.
+ * @ctx: dspp ctx info
+ * @cfg: pointer to struct sde_hw_cp_cfg
+ */
+void reg_dmav1_setup_ltm_vlutv1_2(struct sde_hw_dspp *ctx, void *cfg);
+
 /**
  * reg_dmav1_setup_rc_datav1() - RC DATA v1 implementation using reg dma v1.
  * @ctx: dspp ctx info