
The function cam_mem_mgr_release can unmap the buffers when in use. This change prevents unmapping the buffers when in use. CRs-Fixed: 3489559 Change-Id: I2e72e795d39ac15abfa56c19043c419a03686966 Signed-off-by: Shivakumar Malke <quic_smalke@quicinc.com>
573 lines
15 KiB
C
573 lines
15 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (c) 2017-2021, The Linux Foundation. All rights reserved.
|
|
* Copyright (c) 2022-2023 Qualcomm Innovation Center, Inc. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/debugfs.h>
|
|
#include <linux/videodev2.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/uaccess.h>
|
|
#include <media/cam_sync.h>
|
|
#include <media/cam_defs.h>
|
|
#include <media/cam_icp.h>
|
|
#include "cam_node.h"
|
|
#include "cam_context.h"
|
|
#include "cam_context_utils.h"
|
|
#include "cam_icp_context.h"
|
|
#include "cam_req_mgr_util.h"
|
|
#include "cam_mem_mgr.h"
|
|
#include "cam_trace.h"
|
|
#include "cam_debug_util.h"
|
|
#include "cam_packet_util.h"
|
|
#include "cam_req_mgr_dev.h"
|
|
#include "cam_icp_hw_mgr_intf.h"
|
|
|
|
static int cam_icp_context_dump_active_request(void *data, void *args)
|
|
{
|
|
struct cam_context *ctx = (struct cam_context *)data;
|
|
struct cam_ctx_request *req = NULL;
|
|
struct cam_ctx_request *req_temp = NULL;
|
|
struct cam_hw_dump_pf_args *pf_args = (struct cam_hw_dump_pf_args *)args;
|
|
int rc = 0;
|
|
|
|
if (!ctx || !pf_args) {
|
|
CAM_ERR(CAM_ICP, "Invalid ctx %pK or pf args %pK",
|
|
ctx, pf_args);
|
|
return -EINVAL;
|
|
}
|
|
|
|
CAM_INFO(CAM_ICP, "[%s] iommu fault for icp ctx %d state %d",
|
|
ctx->dev_name, ctx->ctx_id, ctx->state);
|
|
|
|
list_for_each_entry_safe(req, req_temp,
|
|
&ctx->active_req_list, list) {
|
|
CAM_INFO(CAM_ICP, "[%s] ctx[%u]: Active req_id: %llu",
|
|
ctx->dev_name, ctx->ctx_id, req->request_id);
|
|
|
|
rc = cam_context_dump_pf_info_to_hw(ctx, pf_args, &req->pf_data);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Failed to dump pf info.ctx->state: %d",
|
|
ctx->dev_name, ctx->ctx_id, ctx->state);
|
|
}
|
|
|
|
/*
|
|
* Faulted ctx found. Since IPE/BPS instances are shared among contexts,
|
|
* faulted ctx is found if and only if the context contains
|
|
* faulted buffer
|
|
*/
|
|
if (pf_args->pf_context_info.ctx_found) {
|
|
/* Send PF notification to UMD if PF found on current CTX */
|
|
rc = cam_context_send_pf_evt(ctx, pf_args);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP,
|
|
"[%s] ctx[%u]: Failed to notify PF event to userspace rc: %d",
|
|
ctx->dev_name, ctx->ctx_id, rc);
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int cam_icp_context_mini_dump(void *priv, void *args)
|
|
{
|
|
int rc;
|
|
struct cam_context *ctx;
|
|
|
|
if (!priv || !args) {
|
|
CAM_ERR(CAM_ICP, "Invalid priv %pK args %pK", priv, args);
|
|
return -EINVAL;
|
|
}
|
|
|
|
ctx = (struct cam_context *)priv;
|
|
rc = cam_context_mini_dump(ctx, args);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Mini Dump failed rc %d",
|
|
ctx->dev_name, ctx->ctx_id, rc);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_acquire_dev_in_available(struct cam_context *ctx,
|
|
struct cam_acquire_dev_cmd *cmd)
|
|
{
|
|
int rc;
|
|
|
|
rc = cam_context_acquire_dev_to_hw(ctx, cmd);
|
|
if (!rc) {
|
|
ctx->state = CAM_CTX_ACQUIRED;
|
|
trace_cam_context_state(ctx->dev_name, ctx);
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_release_dev_in_acquired(struct cam_context *ctx,
|
|
struct cam_release_dev_cmd *cmd)
|
|
{
|
|
int rc;
|
|
|
|
cam_common_release_evt_params(ctx->dev_hdl);
|
|
|
|
rc = cam_context_release_dev_to_hw(ctx, cmd);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Unable to release device",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
|
|
ctx->state = CAM_CTX_AVAILABLE;
|
|
trace_cam_context_state(ctx->dev_name, ctx);
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_start_dev_in_acquired(struct cam_context *ctx,
|
|
struct cam_start_stop_dev_cmd *cmd)
|
|
{
|
|
int rc;
|
|
|
|
rc = cam_context_start_dev_to_hw(ctx, cmd);
|
|
if (!rc) {
|
|
ctx->state = CAM_CTX_READY;
|
|
trace_cam_context_state(ctx->dev_name, ctx);
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_dump_dev_in_ready(
|
|
struct cam_context *ctx,
|
|
struct cam_dump_req_cmd *cmd)
|
|
{
|
|
int rc;
|
|
|
|
rc = cam_context_dump_dev_to_hw(ctx, cmd);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Failed to dump device",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_flush_dev_in_ready(struct cam_context *ctx,
|
|
struct cam_flush_dev_cmd *cmd)
|
|
{
|
|
int rc;
|
|
struct cam_context_utils_flush_args flush_args;
|
|
|
|
flush_args.cmd = cmd;
|
|
flush_args.flush_active_req = false;
|
|
|
|
rc = cam_context_flush_dev_to_hw(ctx, &flush_args);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Failed to flush device",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_config_dev_in_ready(struct cam_context *ctx,
|
|
struct cam_config_dev_cmd *cmd)
|
|
{
|
|
int rc;
|
|
size_t len;
|
|
uintptr_t packet_addr;
|
|
struct cam_packet *packet;
|
|
size_t remain_len = 0;
|
|
|
|
rc = cam_mem_get_cpu_buf((int32_t) cmd->packet_handle,
|
|
&packet_addr, &len);
|
|
if (rc) {
|
|
CAM_ERR(CAM_ICP, "[%s][%d] Can not get packet address",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
rc = -EINVAL;
|
|
return rc;
|
|
}
|
|
|
|
remain_len = len;
|
|
if ((len < sizeof(struct cam_packet)) ||
|
|
(cmd->offset >= (len - sizeof(struct cam_packet)))) {
|
|
CAM_ERR(CAM_CTXT,
|
|
"[%s] ctx[%u]: Invalid offset, len: %zu cmd offset: %llu sizeof packet: %zu",
|
|
ctx->dev_name, ctx->ctx_id,
|
|
len, cmd->offset, sizeof(struct cam_packet));
|
|
goto put_cpu_buf;
|
|
}
|
|
|
|
remain_len -= (size_t)cmd->offset;
|
|
packet = (struct cam_packet *) ((uint8_t *)packet_addr +
|
|
(uint32_t)cmd->offset);
|
|
|
|
rc = cam_packet_util_validate_packet(packet, remain_len);
|
|
if (rc) {
|
|
CAM_ERR(CAM_CTXT, "[%s] ctx[%u]: Invalid packet params, remain length: %zu",
|
|
ctx->dev_name, ctx->ctx_id,
|
|
remain_len);
|
|
goto put_cpu_buf;
|
|
}
|
|
|
|
if (((packet->header.op_code & 0xff) ==
|
|
CAM_ICP_OPCODE_IPE_SETTINGS) ||
|
|
((packet->header.op_code & 0xff) ==
|
|
CAM_ICP_OPCODE_BPS_SETTINGS) ||
|
|
((packet->header.op_code & 0xff) ==
|
|
CAM_ICP_OPCODE_OFE_SETTINGS))
|
|
rc = cam_context_config_dev_to_hw(ctx, cmd);
|
|
else
|
|
rc = cam_context_prepare_dev_to_hw(ctx, cmd);
|
|
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]:Failed to prepare device",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
|
|
put_cpu_buf:
|
|
cam_mem_put_cpu_buf((int32_t) cmd->packet_handle);
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_stop_dev_in_ready(struct cam_context *ctx,
|
|
struct cam_start_stop_dev_cmd *cmd)
|
|
{
|
|
int rc;
|
|
|
|
rc = cam_context_stop_dev_to_hw(ctx);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Failed to stop device",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
|
|
ctx->state = CAM_CTX_ACQUIRED;
|
|
trace_cam_context_state(ctx->dev_name, ctx);
|
|
return rc;
|
|
}
|
|
|
|
static int __cam_icp_release_dev_in_ready(struct cam_context *ctx,
|
|
struct cam_release_dev_cmd *cmd)
|
|
{
|
|
int rc;
|
|
|
|
rc = __cam_icp_stop_dev_in_ready(ctx, NULL);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Failed to stop device",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
|
|
rc = __cam_icp_release_dev_in_acquired(ctx, cmd);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP, "Failed to release device");
|
|
|
|
return rc;
|
|
}
|
|
|
|
static uint32_t cam_icp_context_get_error_code(uint32_t err_type)
|
|
{
|
|
switch (err_type) {
|
|
case CAM_ICP_HW_ERROR_NO_MEM:
|
|
return CAM_REQ_MGR_ICP_NO_MEMORY;
|
|
case CAM_ICP_HW_ERROR_SYSTEM_FAILURE:
|
|
return CAM_REQ_MGR_ICP_SYSTEM_FAILURE;
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static int __cam_icp_notify_v4l2_err_evt(struct cam_context *ctx,
|
|
uint32_t err_type, uint32_t err_code, uint64_t request_id)
|
|
{
|
|
struct cam_req_mgr_message req_msg = {0};
|
|
int rc;
|
|
|
|
req_msg.session_hdl = ctx->session_hdl;
|
|
req_msg.u.err_msg.device_hdl = ctx->dev_hdl;
|
|
req_msg.u.err_msg.error_type = err_type;
|
|
req_msg.u.err_msg.link_hdl = ctx->link_hdl;
|
|
req_msg.u.err_msg.request_id = request_id;
|
|
req_msg.u.err_msg.resource_size = 0x0;
|
|
req_msg.u.err_msg.error_code = err_code;
|
|
|
|
rc = cam_req_mgr_notify_message(&req_msg, V4L_EVENT_CAM_REQ_MGR_ERROR,
|
|
V4L_EVENT_CAM_REQ_MGR_EVENT);
|
|
if (rc)
|
|
CAM_ERR(CAM_ICP,
|
|
"[%s] ctx[%u]: Error in notifying the error time for req id:%lld",
|
|
ctx->dev_name, ctx->ctx_id, request_id);
|
|
|
|
CAM_INFO(CAM_ICP,
|
|
"[%s] ctx[%u]: notifying error to userspace err type: %d, err code: %u, req id: %llu",
|
|
ctx->dev_name, ctx->ctx_id, err_type, err_code, request_id);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int cam_icp_ctx_handle_fatal_error(void *ctx, void *err_evt_data)
|
|
{
|
|
struct cam_icp_hw_error_evt_data *err_evt;
|
|
uint32_t err_code = 0;
|
|
int rc;
|
|
|
|
err_evt = (struct cam_icp_hw_error_evt_data *)err_evt_data;
|
|
err_code = cam_icp_context_get_error_code(err_evt->err_type);
|
|
|
|
rc = __cam_icp_notify_v4l2_err_evt(ctx, CAM_REQ_MGR_ERROR_TYPE_RECOVERY,
|
|
err_code, err_evt->req_id);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int cam_icp_ctx_handle_buf_done_in_ready(void *ctx, void *done_evt_data)
|
|
{
|
|
struct cam_icp_hw_buf_done_evt_data *buf_done;
|
|
|
|
buf_done = (struct cam_icp_hw_buf_done_evt_data *)done_evt_data;
|
|
return cam_context_buf_done_from_hw(ctx, buf_done->buf_done_data, buf_done->evt_id);
|
|
}
|
|
|
|
static int cam_icp_ctx_handle_error_inducement(void *ctx, void *inject_evt_arg)
|
|
{
|
|
return cam_context_apply_evt_injection(ctx, inject_evt_arg);
|
|
}
|
|
|
|
static int __cam_icp_ctx_handle_hw_event(void *ctx,
|
|
uint32_t evt_id, void *evt_data)
|
|
{
|
|
int rc;
|
|
|
|
if (!ctx || !evt_data) {
|
|
CAM_ERR(CAM_ICP, "Invalid ctx %s and event data %s",
|
|
CAM_IS_NULL_TO_STR(ctx), CAM_IS_NULL_TO_STR(evt_data));
|
|
return -EINVAL;
|
|
}
|
|
|
|
switch (evt_id) {
|
|
case CAM_ICP_EVT_ID_BUF_DONE:
|
|
rc = cam_icp_ctx_handle_buf_done_in_ready(ctx, evt_data);
|
|
break;
|
|
case CAM_ICP_EVT_ID_ERROR:
|
|
rc = cam_icp_ctx_handle_fatal_error(ctx, evt_data);
|
|
break;
|
|
case CAM_ICP_EVT_ID_INJECT_EVENT:
|
|
rc = cam_icp_ctx_handle_error_inducement(ctx, evt_data);
|
|
break;
|
|
default:
|
|
CAM_ERR(CAM_ICP, "Invalid event id: %u", evt_id);
|
|
rc = -EINVAL;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int cam_icp_context_validate_event_notify_injection(struct cam_context *ctx,
|
|
struct cam_hw_inject_evt_param *evt_params)
|
|
{
|
|
int rc = 0;
|
|
uint32_t evt_type;
|
|
uint64_t req_id;
|
|
|
|
req_id = evt_params->req_id;
|
|
evt_type = evt_params->u.evt_notify.evt_notify_type;
|
|
|
|
switch (evt_type) {
|
|
case V4L_EVENT_CAM_REQ_MGR_ERROR: {
|
|
struct cam_hw_inject_err_evt_param *err_evt_params =
|
|
&evt_params->u.evt_notify.u.err_evt_params;
|
|
|
|
switch (err_evt_params->err_type) {
|
|
case CAM_REQ_MGR_ERROR_TYPE_RECOVERY:
|
|
case CAM_REQ_MGR_ERROR_TYPE_FULL_RECOVERY:
|
|
break;
|
|
default:
|
|
CAM_ERR(CAM_ICP,
|
|
"[%s] ctx[%u]: Invalid error type: %u for error event injection err code: %u req id: %llu dev hdl: %d",
|
|
ctx->dev_name, ctx->ctx_id, err_evt_params->err_type,
|
|
err_evt_params->err_code, ctx->dev_hdl);
|
|
return -EINVAL;
|
|
}
|
|
|
|
CAM_INFO(CAM_ICP,
|
|
"[%s] ctx[%u]: Inject ERR evt: err code: %u err type: %u req id: %llu dev hdl: %d",
|
|
ctx->dev_name, ctx->ctx_id, err_evt_params->err_code,
|
|
err_evt_params->err_type, req_id, ctx->dev_hdl);
|
|
break;
|
|
}
|
|
case V4L_EVENT_CAM_REQ_MGR_PF_ERROR: {
|
|
struct cam_hw_inject_pf_evt_param *pf_evt_params =
|
|
&evt_params->u.evt_notify.u.pf_evt_params;
|
|
bool non_fatal_en;
|
|
|
|
rc = cam_smmu_is_cb_non_fatal_fault_en(ctx->img_iommu_hdl, &non_fatal_en);
|
|
if (rc) {
|
|
CAM_ERR(CAM_ICP,
|
|
"[%s] ctx[%u]: Fail to query whether device's cb has non-fatal enabled rc: %d",
|
|
ctx->dev_name, ctx->ctx_id, rc);
|
|
return rc;
|
|
}
|
|
|
|
if (!non_fatal_en) {
|
|
CAM_ERR(CAM_ICP,
|
|
"[%s] ctx[%u]: Fail to inject page fault event notification. Page fault is fatal for ICP",
|
|
ctx->dev_name, ctx->ctx_id);
|
|
return -EINVAL;
|
|
}
|
|
|
|
CAM_INFO(CAM_ICP,
|
|
"[%s] ctx[%u]: Inject PF evt: req_id: %llu dev hdl: %d ctx found: %hhu",
|
|
ctx->dev_name, ctx->ctx_id,
|
|
req_id, ctx->dev_hdl, pf_evt_params->ctx_found);
|
|
break;
|
|
}
|
|
default:
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Event notification type not supported: %u",
|
|
ctx->dev_name, ctx->ctx_id, evt_type);
|
|
rc = -EINVAL;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int cam_icp_context_inject_evt(void *context, void *evt_args)
|
|
{
|
|
struct cam_context *ctx = context;
|
|
struct cam_hw_inject_evt_param *evt_params = NULL;
|
|
struct cam_hw_inject_buffer_error_param *buf_err_params = NULL;
|
|
int rc = 0;
|
|
|
|
if (!ctx || !evt_args) {
|
|
CAM_ERR(CAM_ICP,
|
|
"invalid params ctx %s event args %s",
|
|
CAM_IS_NULL_TO_STR(ctx), CAM_IS_NULL_TO_STR(evt_args));
|
|
return -EINVAL;
|
|
}
|
|
|
|
evt_params = (struct cam_hw_inject_evt_param *)evt_args;
|
|
|
|
if (evt_params->inject_id == CAM_COMMON_EVT_INJECT_BUFFER_ERROR_TYPE) {
|
|
buf_err_params = &evt_params->u.buf_err_evt;
|
|
if (buf_err_params->sync_error > CAM_SYNC_ICP_EVENT_START ||
|
|
buf_err_params->sync_error < CAM_SYNC_ICP_EVENT_END) {
|
|
CAM_INFO(CAM_ICP, "[%s] ctx[%u]: Inject buffer sync error %u req id %llu",
|
|
ctx->dev_name, ctx->ctx_id, buf_err_params->sync_error,
|
|
evt_params->req_id);
|
|
} else {
|
|
CAM_ERR(CAM_ICP, "[%s] ctx[%u]: Invalid buffer sync error %u req id %llu",
|
|
ctx->dev_name, ctx->ctx_id, buf_err_params->sync_error,
|
|
evt_params->req_id);
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
rc = cam_icp_context_validate_event_notify_injection(ctx, evt_params);
|
|
if (rc) {
|
|
CAM_ERR(CAM_ICP,
|
|
"[%s] ctx[%u]: Event notification injection failed validation rc: %d",
|
|
ctx->dev_name, ctx->ctx_id, rc);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
if (ctx->hw_mgr_intf->hw_inject_evt)
|
|
ctx->hw_mgr_intf->hw_inject_evt(ctx->ctxt_to_hw_map, evt_args);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static struct cam_ctx_ops
|
|
cam_icp_ctx_state_machine[CAM_CTX_STATE_MAX] = {
|
|
/* Uninit */
|
|
{
|
|
.ioctl_ops = {},
|
|
.crm_ops = {},
|
|
.irq_ops = NULL,
|
|
},
|
|
/* Available */
|
|
{
|
|
.ioctl_ops = {
|
|
.acquire_dev = __cam_icp_acquire_dev_in_available,
|
|
},
|
|
.crm_ops = {},
|
|
.irq_ops = NULL,
|
|
.mini_dump_ops = cam_icp_context_mini_dump,
|
|
},
|
|
/* Acquired */
|
|
{
|
|
.ioctl_ops = {
|
|
.release_dev = __cam_icp_release_dev_in_acquired,
|
|
.start_dev = __cam_icp_start_dev_in_acquired,
|
|
.config_dev = __cam_icp_config_dev_in_ready,
|
|
.flush_dev = __cam_icp_flush_dev_in_ready,
|
|
.dump_dev = __cam_icp_dump_dev_in_ready,
|
|
},
|
|
.crm_ops = {},
|
|
.irq_ops = __cam_icp_ctx_handle_hw_event,
|
|
.pagefault_ops = cam_icp_context_dump_active_request,
|
|
.mini_dump_ops = cam_icp_context_mini_dump,
|
|
.evt_inject_ops = cam_icp_context_inject_evt,
|
|
},
|
|
/* Ready */
|
|
{
|
|
.ioctl_ops = {
|
|
.stop_dev = __cam_icp_stop_dev_in_ready,
|
|
.release_dev = __cam_icp_release_dev_in_ready,
|
|
.config_dev = __cam_icp_config_dev_in_ready,
|
|
.flush_dev = __cam_icp_flush_dev_in_ready,
|
|
.dump_dev = __cam_icp_dump_dev_in_ready,
|
|
},
|
|
.crm_ops = {},
|
|
.irq_ops = __cam_icp_ctx_handle_hw_event,
|
|
.pagefault_ops = cam_icp_context_dump_active_request,
|
|
.mini_dump_ops = cam_icp_context_mini_dump,
|
|
.evt_inject_ops = cam_icp_context_inject_evt,
|
|
},
|
|
/* Flushed */
|
|
{
|
|
.ioctl_ops = {},
|
|
},
|
|
/* Activated */
|
|
{
|
|
.ioctl_ops = {},
|
|
.crm_ops = {},
|
|
.irq_ops = NULL,
|
|
.pagefault_ops = cam_icp_context_dump_active_request,
|
|
.mini_dump_ops = cam_icp_context_mini_dump,
|
|
.evt_inject_ops = cam_icp_context_inject_evt,
|
|
},
|
|
};
|
|
|
|
int cam_icp_context_init(struct cam_icp_context *ctx, struct cam_hw_mgr_intf *hw_intf,
|
|
uint32_t ctx_id, int img_iommu_hdl, const char *icp_dev_name)
|
|
{
|
|
int rc;
|
|
|
|
if ((!ctx) || (!ctx->base) || (!hw_intf) || (!icp_dev_name)) {
|
|
CAM_ERR(CAM_ICP,
|
|
"Invalid params: ctx: %s hw intf: %s dev name: %s",
|
|
CAM_IS_NULL_TO_STR(ctx), CAM_IS_NULL_TO_STR(hw_intf),
|
|
CAM_IS_NULL_TO_STR(icp_dev_name));
|
|
rc = -EINVAL;
|
|
goto err;
|
|
}
|
|
|
|
rc = cam_context_init(ctx->base, icp_dev_name, CAM_ICP, ctx_id,
|
|
NULL, hw_intf, ctx->req_base, CAM_CTX_ICP_REQ_MAX, img_iommu_hdl);
|
|
if (rc) {
|
|
CAM_ERR(CAM_ICP, "[%s] Camera Context Base init failed", icp_dev_name);
|
|
goto err;
|
|
}
|
|
|
|
ctx->base->state_machine = cam_icp_ctx_state_machine;
|
|
ctx->base->ctx_priv = ctx;
|
|
ctx->base->max_hw_update_entries = CAM_CTX_CFG_MAX;
|
|
ctx->base->max_in_map_entries = CAM_CTX_CFG_MAX;
|
|
ctx->base->max_out_map_entries = CAM_CTX_CFG_MAX;
|
|
ctx->ctxt_to_hw_map = NULL;
|
|
|
|
err:
|
|
return rc;
|
|
}
|
|
|
|
int cam_icp_context_deinit(struct cam_icp_context *ctx)
|
|
{
|
|
if ((!ctx) || (!ctx->base)) {
|
|
CAM_ERR(CAM_ICP, "Invalid params: %pK", ctx);
|
|
return -EINVAL;
|
|
}
|
|
|
|
cam_context_deinit(ctx->base);
|
|
memset(ctx, 0, sizeof(*ctx));
|
|
|
|
return 0;
|
|
}
|