Files
android_kernel_samsung_sm86…/smmu-proxy/qti-smmu-proxy-pvm.c
Chris Goldsworthy d71f8c3401 securemsm-kernel: smmu-proxy: Return error on map failure
Return the failure code from the remote VM when smmu_proxy_map()
fails.

Change-Id: I7a7a8a5930275ce13e2bfea38365b878fba534bd
Signed-off-by: Chris Goldsworthy <quic_cgoldswo@quicinc.com>
2023-04-14 15:57:07 -07:00

324 lines
7.5 KiB
C

// SPDX-License-Identifier: GPL-2.0-only
/*
* Copyright (c) 2023 Qualcomm Innovation Center, Inc. All rights reserved.
*/
#include "qti-smmu-proxy-common.h"
#include <linux/qti-smmu-proxy-callbacks.h>
#include <linux/qcom-dma-mapping.h>
#include <linux/of.h>
static void *msgq_hdl;
DEFINE_MUTEX(sender_mutex);
static const struct file_operations smmu_proxy_dev_fops;
int smmu_proxy_unmap(void *data)
{
struct dma_buf *dmabuf;
void *buf;
size_t size;
int ret;
struct smmu_proxy_unmap_req *req;
struct smmu_proxy_unmap_resp *resp;
mutex_lock(&sender_mutex);
buf = kzalloc(GH_MSGQ_MAX_MSG_SIZE_BYTES, GFP_KERNEL);
if (!buf) {
ret = -ENOMEM;
pr_err("%s: Failed to allocate memory!\n", __func__);
goto out;
}
req = buf;
dmabuf = data;
ret = mem_buf_dma_buf_get_memparcel_hdl(dmabuf, &req->hdl);
if (ret) {
pr_err("%s: Failed to get memparcel handle rc: %d\n", __func__, ret);
goto free_buf;
}
req->hdr.msg_type = SMMU_PROXY_UNMAP;
req->hdr.msg_size = sizeof(*req);
ret = gh_msgq_send(msgq_hdl, (void *) req, req->hdr.msg_size, 0);
if (ret < 0) {
pr_err("%s: failed to send message rc: %d\n", __func__, ret);
goto free_buf;
}
/*
* No need to validate size - gh_msgq_recv() ensures that sizeof(*resp) <
* GH_MSGQ_MAX_MSG_SIZE_BYTES
*/
ret = gh_msgq_recv(msgq_hdl, buf, sizeof(*resp), &size, 0);
if (ret < 0) {
pr_err_ratelimited("%s: failed to receive message rc: %d\n", __func__, ret);
goto free_buf;
}
resp = buf;
if (resp->hdr.ret) {
ret = resp->hdr.ret;
pr_err("%s: Unmap call failed on remote VM, rc: %d\n", __func__,
resp->hdr.ret);
}
free_buf:
kfree(buf);
out:
mutex_unlock(&sender_mutex);
return ret;
}
int smmu_proxy_map(struct device *client_dev, struct sg_table *proxy_iova,
struct dma_buf *dmabuf)
{
void *buf;
size_t size;
int ret = 0;
int n_acl_entries, i;
int vmids[2] = { VMID_TVM, VMID_OEMVM };
int perms[2] = { PERM_READ | PERM_WRITE, PERM_READ | PERM_WRITE};
struct csf_version csf_version;
struct mem_buf_lend_kernel_arg arg = {0};
struct smmu_proxy_map_req *req;
struct smmu_proxy_map_resp *resp;
ret = smmu_proxy_get_csf_version(&csf_version);
if (ret) {
return ret;
}
/*
* We enter this function iff the CSF version is 2.5.* . If CSF 2.5.1
* is in use, we set n_acl_entries to two, in order to assign this
* memory to the TVM and OEM VM. If CSF 2.5.0 is in use, we just assign
* it to the TVM.
*/
n_acl_entries = csf_version.min_ver == 1 ? 2 : 1;
mutex_lock(&sender_mutex);
buf = kzalloc(GH_MSGQ_MAX_MSG_SIZE_BYTES, GFP_KERNEL);
if (!buf) {
ret = -ENOMEM;
goto out;
}
if (mem_buf_dma_buf_exclusive_owner(dmabuf)) {
arg.vmids = vmids;
arg.perms = perms;
arg.nr_acl_entries = n_acl_entries;
ret = mem_buf_lend(dmabuf, &arg);
if (ret) {
pr_err("%s: Failed to lend buf rc: %d\n", __func__, ret);
goto free_buf;
}
}
/* Prepare the message */
req = buf;
req->acl_desc.n_acl_entries = n_acl_entries;
for (i = 0; i < n_acl_entries; i++) {
req->acl_desc.acl_entries[i].vmid = vmids[i];
req->acl_desc.acl_entries[i].perms = perms[i];
}
ret = mem_buf_dma_buf_get_memparcel_hdl(dmabuf, &req->hdl);
if (ret) {
pr_err("%s: Failed to get memparcel handle rc: %d\n", __func__, ret);
goto free_buf;
}
ret = of_property_read_u32(client_dev->of_node,
"qti,smmu-proxy-cb-id",
&req->cb_id);
if (ret) {
dev_err(client_dev, "%s: Err reading 'qti,smmu-proxy-cb-id' rc: %d\n",
__func__, ret);
goto free_buf;
}
req->hdr.msg_type = SMMU_PROXY_MAP;
req->hdr.msg_size = offsetof(struct smmu_proxy_map_req,
acl_desc.acl_entries[n_acl_entries]);
ret = gh_msgq_send(msgq_hdl, (void *) req, req->hdr.msg_size, 0);
if (ret < 0) {
pr_err("%s: failed to send message rc: %d\n", __func__, ret);
goto free_buf;
}
/*
* No need to validate size - gh_msgq_recv() ensures that sizeof(*resp) <
* GH_MSGQ_MAX_MSG_SIZE_BYTES
*/
ret = gh_msgq_recv(msgq_hdl, buf, sizeof(*resp), &size, 0);
if (ret < 0) {
pr_err_ratelimited("%s: failed to receive message rc: %d\n", __func__, ret);
goto free_buf;
}
resp = buf;
if (resp->hdr.ret) {
ret = resp->hdr.ret;
pr_err_ratelimited("%s: Map call failed on remote VM, rc: %d\n", __func__,
resp->hdr.ret);
goto free_buf;
}
ret = mem_buf_dma_buf_set_destructor(dmabuf, smmu_proxy_unmap, dmabuf);
if (ret) {
pr_err_ratelimited("%s: Failed to set vmperm destructor, rc: %d\n",
__func__, ret);
goto free_buf;
}
sg_dma_address(proxy_iova->sgl) = resp->iova;
sg_dma_len(proxy_iova->sgl) = resp->mapping_len;
/*
* We set the number of entries to one here, as we only allow the mapping to go
* through on the TVM if the sg_table returned by dma_buf_map_attachment has one
* entry.
*/
proxy_iova->nents = 1;
free_buf:
kfree(buf);
out:
mutex_unlock(&sender_mutex);
return ret;
}
void smmu_proxy_unmap_nop(struct device *client_dev, struct sg_table *table,
struct dma_buf *dmabuf)
{
}
static long smmu_proxy_dev_ioctl(struct file *filp, unsigned int cmd,
unsigned long arg)
{
unsigned int dir = _IOC_DIR(cmd);
union smmu_proxy_ioctl_arg ioctl_arg;
int ret;
if (_IOC_SIZE(cmd) > sizeof(ioctl_arg))
return -EINVAL;
if (copy_from_user(&ioctl_arg, (void __user *)arg, _IOC_SIZE(cmd)))
return -EFAULT;
if (!(dir & _IOC_WRITE))
memset(&ioctl_arg, 0, sizeof(ioctl_arg));
switch (cmd) {
case QTI_SMMU_PROXY_GET_VERSION_IOCTL:
{
struct csf_version *csf_version =
&ioctl_arg.csf_version;
ret = smmu_proxy_get_csf_version(csf_version);
if(ret)
return ret;
break;
}
default:
return -ENOTTY;
}
if (dir & _IOC_READ) {
if (copy_to_user((void __user *)arg, &ioctl_arg,
_IOC_SIZE(cmd)))
return -EFAULT;
}
return 0;
}
static const struct file_operations smmu_proxy_dev_fops = {
.unlocked_ioctl = smmu_proxy_dev_ioctl,
.compat_ioctl = compat_ptr_ioctl,
};
static int sender_probe_handler(struct platform_device *pdev)
{
int ret;
struct csf_version csf_version;
msgq_hdl = gh_msgq_register(GH_MSGQ_LABEL_SMMU_PROXY);
if (IS_ERR(msgq_hdl)) {
ret = PTR_ERR(msgq_hdl);
pr_err("%s: Queue registration failed rc: %d!\n", __func__, PTR_ERR(msgq_hdl));
return ret;
}
ret = smmu_proxy_get_csf_version(&csf_version);
if (ret) {
pr_err("%s: Failed to get CSF version rc: %d\n", __func__, ret);
goto free_msgq;
}
if (csf_version.arch_ver == 2 && csf_version.max_ver == 0) {
ret = qti_smmu_proxy_register_callbacks(NULL, NULL);
} else if (csf_version.arch_ver == 2 && csf_version.max_ver == 5) {
ret = qti_smmu_proxy_register_callbacks(smmu_proxy_map, smmu_proxy_unmap_nop);
} else {
pr_err("%s: Invalid CSF version: %d.%d\n", __func__, csf_version.arch_ver,
csf_version.max_ver);
goto free_msgq;
}
if (ret) {
pr_err("%s: Failed to set SMMU proxy callbacks rc: %d\n", __func__, ret);
goto free_msgq;
}
ret = smmu_proxy_create_dev(&smmu_proxy_dev_fops);
if (ret) {
pr_err("%s: Failed to create character device rc: %d\n", __func__,
ret);
goto set_callbacks_null;
}
return 0;
set_callbacks_null:
qti_smmu_proxy_register_callbacks(NULL, NULL);
free_msgq:
gh_msgq_unregister(msgq_hdl);
return ret;
}
static const struct of_device_id smmu_proxy_match_table[] = {
{.compatible = "smmu-proxy-sender"},
{},
};
static struct platform_driver smmu_proxy_driver = {
.probe = sender_probe_handler,
.driver = {
.name = "qti-smmu-proxy",
.of_match_table = smmu_proxy_match_table,
},
};
int __init init_smmu_proxy_driver(void)
{
return platform_driver_register(&smmu_proxy_driver);
}
module_init(init_smmu_proxy_driver);
MODULE_IMPORT_NS(DMA_BUF);
MODULE_LICENSE("GPL v2");