// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2015-2020, The Linux Foundation. All rights reserved. */ #define pr_fmt(fmt) "%s: " fmt, __func__ #include #include #include #include #include #include #include #include #include #include #include #include #include "soc/qcom/secure_buffer.h" #include "sde_rotator_base.h" #include "sde_rotator_util.h" #include "sde_rotator_io_util.h" #include "sde_rotator_smmu.h" #include "sde_rotator_debug.h" #define SMMU_SDE_ROT_SEC "qcom,smmu_sde_rot_sec" #define SMMU_SDE_ROT_UNSEC "qcom,smmu_sde_rot_unsec" struct sde_smmu_domain { char *ctx_name; int domain; }; static inline bool sde_smmu_is_valid_domain_type( struct sde_rot_data_type *mdata, int domain_type) { return true; } static inline bool sde_smmu_is_valid_domain_condition( struct sde_rot_data_type *mdata, int domain_type, bool is_attach) { if (is_attach) { if (test_bit(SDE_CAPS_SEC_ATTACH_DETACH_SMMU, mdata->sde_caps_map) && (mdata->sec_cam_en && domain_type == SDE_IOMMU_DOMAIN_ROT_SECURE)) return false; else return true; } else { if (test_bit(SDE_CAPS_SEC_ATTACH_DETACH_SMMU, mdata->sde_caps_map) && (mdata->sec_cam_en && domain_type == SDE_IOMMU_DOMAIN_ROT_SECURE)) return true; else return false; } } struct sde_smmu_client *sde_smmu_get_cb(u32 domain) { struct sde_rot_data_type *mdata = sde_rot_get_mdata(); if (!sde_smmu_is_valid_domain_type(mdata, domain)) return NULL; return (domain >= SDE_IOMMU_MAX_DOMAIN) ? NULL : &mdata->sde_smmu[domain]; } static int sde_smmu_util_parse_dt_clock(struct platform_device *pdev, struct sde_module_power *mp) { u32 i = 0, rc = 0; const char *clock_name; u32 clock_rate; int num_clk; num_clk = of_property_count_strings(pdev->dev.of_node, "clock-names"); if (num_clk < 0) { SDEROT_DBG("clocks are not defined\n"); num_clk = 0; } mp->num_clk = num_clk; mp->clk_config = devm_kzalloc(&pdev->dev, sizeof(struct sde_clk) * mp->num_clk, GFP_KERNEL); if (num_clk && !mp->clk_config) { rc = -ENOMEM; mp->num_clk = 0; goto clk_err; } for (i = 0; i < mp->num_clk; i++) { of_property_read_string_index(pdev->dev.of_node, "clock-names", i, &clock_name); strlcpy(mp->clk_config[i].clk_name, clock_name, sizeof(mp->clk_config[i].clk_name)); of_property_read_u32_index(pdev->dev.of_node, "clock-rate", i, &clock_rate); mp->clk_config[i].rate = clock_rate; if (!clock_rate) mp->clk_config[i].type = SDE_CLK_AHB; else mp->clk_config[i].type = SDE_CLK_PCLK; } clk_err: return rc; } static int sde_smmu_clk_register(struct platform_device *pdev, struct sde_module_power *mp) { int i, ret; struct clk *clk; ret = sde_smmu_util_parse_dt_clock(pdev, mp); if (ret) { SDEROT_ERR("unable to parse clocks\n"); return -EINVAL; } for (i = 0; i < mp->num_clk; i++) { clk = devm_clk_get(&pdev->dev, mp->clk_config[i].clk_name); if (IS_ERR(clk)) { SDEROT_ERR("unable to get clk: %s\n", mp->clk_config[i].clk_name); return PTR_ERR(clk); } mp->clk_config[i].clk = clk; } return 0; } static int sde_smmu_enable_power(struct sde_smmu_client *sde_smmu, bool enable) { int rc = 0; struct sde_module_power *mp; if (!sde_smmu) return -EINVAL; mp = &sde_smmu->mp; if (!mp->num_vreg && !mp->num_clk) return 0; if (enable) { rc = sde_rot_enable_vreg(mp->vreg_config, mp->num_vreg, true); if (rc) { SDEROT_ERR("vreg enable failed - rc:%d\n", rc); goto end; } sde_update_reg_bus_vote(sde_smmu->reg_bus_clt, VOTE_INDEX_76_MHZ); rc = sde_rot_enable_clk(mp->clk_config, mp->num_clk, true); if (rc) { SDEROT_ERR("clock enable failed - rc:%d\n", rc); sde_update_reg_bus_vote(sde_smmu->reg_bus_clt, VOTE_INDEX_DISABLE); sde_rot_enable_vreg(mp->vreg_config, mp->num_vreg, false); goto end; } } else { sde_rot_enable_clk(mp->clk_config, mp->num_clk, false); sde_update_reg_bus_vote(sde_smmu->reg_bus_clt, VOTE_INDEX_DISABLE); sde_rot_enable_vreg(mp->vreg_config, mp->num_vreg, false); } end: return rc; } /* * sde_smmu_attach() * * Associates each configured VA range with the corresponding smmu context * bank device. Enables the clks as smmu requires voting it before the usage. * And iommu attach is done only once during the initial attach and it is never * detached as smmu v2 uses a feature called 'retention'. */ int sde_smmu_attach(struct sde_rot_data_type *mdata) { struct sde_smmu_client *sde_smmu; int i, rc = 0; for (i = 0; i < SDE_IOMMU_MAX_DOMAIN; i++) { if (!sde_smmu_is_valid_domain_type(mdata, i)) continue; sde_smmu = sde_smmu_get_cb(i); if (sde_smmu && sde_smmu->dev) { rc = sde_smmu_enable_power(sde_smmu, true); if (rc) { SDEROT_ERR( "power enable failed - domain:[%d] rc:%d\n", i, rc); goto err; } if (!sde_smmu->domain_attached && sde_smmu_is_valid_domain_condition(mdata, i, true)) { rc = iommu_attach_device( sde_smmu->rot_domain, sde_smmu->dev); if (rc) { SDEROT_ERR( "iommu attach device failed for domain[%d] with err:%d\n", i, rc); sde_smmu_enable_power(sde_smmu, false); goto err; } sde_smmu->domain_attached = true; SDEROT_DBG("iommu v2 domain[%i] attached\n", i); } } else { SDEROT_DBG( "iommu device not attached for domain[%d]\n", i); } } return 0; err: for (i--; i >= 0; i--) { sde_smmu = sde_smmu_get_cb(i); if (sde_smmu && sde_smmu->dev) { iommu_detach_device(sde_smmu->rot_domain, sde_smmu->dev); sde_smmu_enable_power(sde_smmu, false); sde_smmu->domain_attached = false; } } return rc; } /* * sde_smmu_detach() * * Only disables the clks as it is not required to detach the iommu mapped * VA range from the device in smmu as explained in the sde_smmu_attach */ int sde_smmu_detach(struct sde_rot_data_type *mdata) { struct sde_smmu_client *sde_smmu; int i; for (i = 0; i < SDE_IOMMU_MAX_DOMAIN; i++) { if (!sde_smmu_is_valid_domain_type(mdata, i)) continue; sde_smmu = sde_smmu_get_cb(i); if (sde_smmu && sde_smmu->dev) { if (sde_smmu->domain_attached && sde_smmu_is_valid_domain_condition(mdata, i, false)) { iommu_detach_device(sde_smmu->rot_domain, sde_smmu->dev); SDEROT_DBG("iommu domain[%i] detached\n", i); sde_smmu->domain_attached = false; } else { sde_smmu_enable_power(sde_smmu, false); } } } return 0; } int sde_smmu_get_domain_id(u32 type) { return type; } /* * sde_smmu_dma_buf_attach() * * Same as sde_smmu_dma_buf_attach except that the device is got from * the configured smmu v2 context banks. */ struct dma_buf_attachment *sde_smmu_dma_buf_attach( struct dma_buf *dma_buf, struct device *dev, int domain) { struct sde_smmu_client *sde_smmu = sde_smmu_get_cb(domain); if (!sde_smmu) { SDEROT_ERR("not able to get smmu context\n"); return NULL; } return dma_buf_attach(dma_buf, sde_smmu->dev); } /* * sde_smmu_map_dma_buf() * * Maps existing buffer (by struct scatterlist) into SMMU context bank device. * From which we can take the virtual address and size allocated. * msm_map_dma_buf is depricated with smmu v2 and it uses dma_map_sg instead */ int sde_smmu_map_dma_buf(struct dma_buf *dma_buf, struct sg_table *table, int domain, dma_addr_t *iova, unsigned long *size, int dir) { int rc; struct sde_smmu_client *sde_smmu = sde_smmu_get_cb(domain); unsigned long attrs = 0; if (!sde_smmu) { SDEROT_ERR("not able to get smmu context\n"); return -EINVAL; } rc = dma_map_sg_attrs(sde_smmu->dev, table->sgl, table->nents, dir, attrs); if (!rc) { SDEROT_ERR("dma map sg failed\n"); return -ENOMEM; } *iova = table->sgl->dma_address; *size = table->sgl->dma_length; return 0; } void sde_smmu_unmap_dma_buf(struct sg_table *table, int domain, int dir, struct dma_buf *dma_buf) { struct sde_smmu_client *sde_smmu = sde_smmu_get_cb(domain); if (!sde_smmu) { SDEROT_ERR("not able to get smmu context\n"); return; } dma_unmap_sg(sde_smmu->dev, table->sgl, table->nents, dir); } static DEFINE_MUTEX(sde_smmu_ref_cnt_lock); int sde_smmu_ctrl(int enable) { struct sde_rot_data_type *mdata = sde_rot_get_mdata(); int rc = 0; mutex_lock(&sde_smmu_ref_cnt_lock); SDEROT_EVTLOG(__builtin_return_address(0), enable, mdata->iommu_ref_cnt, mdata->iommu_attached); SDEROT_DBG("%pS: enable:%d ref_cnt:%d attach:%d\n", __builtin_return_address(0), enable, mdata->iommu_ref_cnt, mdata->iommu_attached); if (enable) { if (!mdata->iommu_attached) { rc = sde_smmu_attach(mdata); if (!rc) mdata->iommu_attached = true; } mdata->iommu_ref_cnt++; } else { if (mdata->iommu_ref_cnt) { mdata->iommu_ref_cnt--; if (mdata->iommu_ref_cnt == 0) if (mdata->iommu_attached) { rc = sde_smmu_detach(mdata); if (!rc) mdata->iommu_attached = false; } } else { SDEROT_ERR("unbalanced iommu ref\n"); } } mutex_unlock(&sde_smmu_ref_cnt_lock); if (rc < 0) return rc; else return mdata->iommu_ref_cnt; } int sde_smmu_secure_ctrl(int enable) { struct sde_rot_data_type *mdata = sde_rot_get_mdata(); int rc = 0; mutex_lock(&sde_smmu_ref_cnt_lock); /* * Attach/detach secure context irrespective of ref count, * We come here only when secure camera is disabled */ if (enable) { rc = sde_smmu_attach(mdata); if (!rc) mdata->iommu_attached = true; } else { rc = sde_smmu_detach(mdata); /* * keep iommu_attached equal to true, * so that driver does not attemp to attach * while in secure state */ } mutex_unlock(&sde_smmu_ref_cnt_lock); return rc; } /* * sde_smmu_device_create() * @dev: sde_mdp device * * For smmu, each context bank is a separate child device of sde rot. * Platform devices are created for those smmu related child devices of * sde rot here. This would facilitate probes to happen for these devices in * which the smmu mapping and initialization is handled. */ void sde_smmu_device_create(struct device *dev) { struct device_node *parent, *child; struct sde_rot_data_type *mdata = sde_rot_get_mdata(); parent = dev->of_node; for_each_child_of_node(parent, child) { if (of_device_is_compatible(child, SMMU_SDE_ROT_SEC)) { of_platform_device_create(child, NULL, dev); mdata->sde_smmu [SDE_IOMMU_DOMAIN_ROT_SECURE].domain_attached = true; } else if (of_device_is_compatible(child, SMMU_SDE_ROT_UNSEC)) { of_platform_device_create(child, NULL, dev); mdata->sde_smmu [SDE_IOMMU_DOMAIN_ROT_UNSECURE].domain_attached = true; } } } int sde_smmu_init(struct device *dev) { sde_smmu_device_create(dev); return 0; } static int sde_smmu_fault_handler(struct iommu_domain *domain, struct device *dev, unsigned long iova, int flags, void *token) { struct sde_smmu_client *sde_smmu; int rc = -EINVAL; if (!token) { SDEROT_ERR("Error: token is NULL\n"); return -EINVAL; } sde_smmu = (struct sde_smmu_client *)token; /* trigger rotator dump */ SDEROT_ERR("trigger rotator dump, iova=0x%08lx, flags=0x%x\n", iova, flags); SDEROT_ERR("SMMU device:%s", sde_smmu->dev->kobj.name); /* generate dump, but no panic */ SDEROT_EVTLOG_TOUT_HANDLER("rot", "rot_dbg_bus", "vbif_dbg_bus"); /* * return -ENOSYS to allow smmu driver to dump out useful * debug info. */ return rc; } static struct sde_smmu_domain sde_rot_unsec = { "rot_0", SDE_IOMMU_DOMAIN_ROT_UNSECURE}; static struct sde_smmu_domain sde_rot_sec = { "rot_1", SDE_IOMMU_DOMAIN_ROT_SECURE}; static const struct of_device_id sde_smmu_dt_match[] = { { .compatible = SMMU_SDE_ROT_UNSEC, .data = &sde_rot_unsec}, { .compatible = SMMU_SDE_ROT_SEC, .data = &sde_rot_sec}, {} }; MODULE_DEVICE_TABLE(of, sde_smmu_dt_match); /* * sde_smmu_probe() * @pdev: platform device * * Each smmu context acts as a separate device and the context banks are * configured with a VA range. * Registeres the clks as each context bank has its own clks, for which voting * has to be done everytime before using that context bank. */ int sde_smmu_probe(struct platform_device *pdev) { struct device *dev; struct sde_rot_data_type *mdata = sde_rot_get_mdata(); struct sde_smmu_client *sde_smmu; int rc = 0; struct sde_smmu_domain smmu_domain; const struct of_device_id *match; struct sde_module_power *mp; char name[MAX_CLIENT_NAME_LEN]; u32 sid = 0; if (!mdata) { SDEROT_INFO( "probe failed as mdata is not initializedi, probe defer\n"); return -EPROBE_DEFER; } match = of_match_device(sde_smmu_dt_match, &pdev->dev); if (!match || !match->data) { SDEROT_ERR("probe failed as match data is invalid\n"); return -EINVAL; } smmu_domain = *(struct sde_smmu_domain *) (match->data); if (smmu_domain.domain >= SDE_IOMMU_MAX_DOMAIN) { SDEROT_ERR("no matching device found\n"); return -EINVAL; } if (of_find_property(pdev->dev.of_node, "iommus", NULL)) { dev = &pdev->dev; rc = of_property_read_u32_index(pdev->dev.of_node, "iommus", 1, &sid); if (rc) SDEROT_DBG("SID not defined for domain:%d", smmu_domain.domain); } else { SDEROT_ERR("Invalid SMMU ctx for domain:%d\n", smmu_domain.domain); return -EINVAL; } sde_smmu = &mdata->sde_smmu[smmu_domain.domain]; sde_smmu->domain = smmu_domain.domain; sde_smmu->sid = sid; mp = &sde_smmu->mp; memset(mp, 0, sizeof(struct sde_module_power)); if (of_find_property(pdev->dev.of_node, "gdsc-mdss-supply", NULL)) { mp->vreg_config = devm_kzalloc(&pdev->dev, sizeof(struct sde_vreg), GFP_KERNEL); if (!mp->vreg_config) return -ENOMEM; strlcpy(mp->vreg_config->vreg_name, "gdsc-mdss", sizeof(mp->vreg_config->vreg_name)); mp->num_vreg = 1; } if (mp->vreg_config) { rc = sde_rot_config_vreg(&pdev->dev, mp->vreg_config, mp->num_vreg, true); if (rc) { SDEROT_ERR("vreg config failed rc=%d\n", rc); goto release_vreg; } } rc = sde_smmu_clk_register(pdev, mp); if (rc) { SDEROT_ERR( "smmu clk register failed for domain[%d] with err:%d\n", smmu_domain.domain, rc); goto disable_vreg; } snprintf(name, MAX_CLIENT_NAME_LEN, "smmu:%u", smmu_domain.domain); sde_smmu->reg_bus_clt = sde_reg_bus_vote_client_create(name); if (IS_ERR_OR_NULL(sde_smmu->reg_bus_clt)) { SDEROT_ERR("mdss bus client register failed\n"); rc = PTR_ERR(sde_smmu->reg_bus_clt); sde_smmu->reg_bus_clt = NULL; goto unregister_clk; } rc = sde_smmu_enable_power(sde_smmu, true); if (rc) { SDEROT_ERR("power enable failed - domain:[%d] rc:%d\n", smmu_domain.domain, rc); goto bus_client_destroy; } sde_smmu->dev = &pdev->dev; sde_smmu->rot_domain = iommu_get_domain_for_dev(sde_smmu->dev); if (!sde_smmu->rot_domain) { dev_err(&pdev->dev, "iommu get domain failed\n"); return -EINVAL; } if (!dev->dma_parms) dev->dma_parms = devm_kzalloc(dev, sizeof(*dev->dma_parms), GFP_KERNEL); dma_set_max_seg_size(dev, DMA_BIT_MASK(32)); dma_set_seg_boundary(dev, (unsigned long)DMA_BIT_MASK(64)); iommu_set_fault_handler(sde_smmu->rot_domain, sde_smmu_fault_handler, (void *)sde_smmu); sde_smmu_enable_power(sde_smmu, false); SDEROT_INFO( "iommu v2 domain[%d] mapping and clk register successful!\n", smmu_domain.domain); return 0; bus_client_destroy: sde_reg_bus_vote_client_destroy(sde_smmu->reg_bus_clt); sde_smmu->reg_bus_clt = NULL; unregister_clk: disable_vreg: sde_rot_config_vreg(&pdev->dev, sde_smmu->mp.vreg_config, sde_smmu->mp.num_vreg, false); release_vreg: devm_kfree(&pdev->dev, sde_smmu->mp.vreg_config); sde_smmu->mp.vreg_config = NULL; sde_smmu->mp.num_vreg = 0; return rc; } int sde_smmu_remove(struct platform_device *pdev) { int i; struct sde_smmu_client *sde_smmu; for (i = 0; i < SDE_IOMMU_MAX_DOMAIN; i++) { sde_smmu = sde_smmu_get_cb(i); if (!sde_smmu || !sde_smmu->dev || (sde_smmu->dev != &pdev->dev)) continue; sde_smmu->dev = NULL; sde_smmu->rot_domain = NULL; sde_smmu_enable_power(sde_smmu, false); sde_reg_bus_vote_client_destroy(sde_smmu->reg_bus_clt); sde_smmu->reg_bus_clt = NULL; sde_rot_config_vreg(&pdev->dev, sde_smmu->mp.vreg_config, sde_smmu->mp.num_vreg, false); devm_kfree(&pdev->dev, sde_smmu->mp.vreg_config); sde_smmu->mp.vreg_config = NULL; sde_smmu->mp.num_vreg = 0; } return 0; } static struct platform_driver sde_smmu_driver = { .probe = sde_smmu_probe, .remove = sde_smmu_remove, .shutdown = NULL, .driver = { .name = "sde_smmu", .of_match_table = sde_smmu_dt_match, }, }; static int sde_smmu_register_driver(void) { return platform_driver_register(&sde_smmu_driver); } static int __init sde_smmu_driver_init(void) { int ret; ret = sde_smmu_register_driver(); if (ret) SDEROT_ERR("sde_smmu_register_driver() failed!\n"); return ret; } module_init(sde_smmu_driver_init); static void __exit sde_smmu_driver_cleanup(void) { platform_driver_unregister(&sde_smmu_driver); } module_exit(sde_smmu_driver_cleanup); MODULE_LICENSE("GPL v2"); MODULE_DESCRIPTION("SDE SMMU driver");