
Bring over camera driver changes as of msm-4.19 commit 5a5551a7 (Merge "msm: camera: reqmgr: Fix CRM shift one req issue"). Change-Id: Ic0c2b2d74d1b3470c1c51d98228e312fb13c501a Signed-off-by: Jigarkumar Zala <jzala@codeaurora.org>
943 行
25 KiB
C
943 行
25 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (c) 2017-2019, The Linux Foundation. All rights reserved.
|
|
*/
|
|
|
|
#include <linux/io.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/random.h>
|
|
#include <asm/errno.h>
|
|
#include <linux/timer.h>
|
|
#include <cam_icp.h>
|
|
#include <linux/iopoll.h>
|
|
#include <soc/qcom/socinfo.h>
|
|
|
|
#include "cam_io_util.h"
|
|
#include "hfi_reg.h"
|
|
#include "hfi_sys_defs.h"
|
|
#include "hfi_session_defs.h"
|
|
#include "hfi_intf.h"
|
|
#include "cam_icp_hw_mgr_intf.h"
|
|
#include "cam_debug_util.h"
|
|
#include "cam_soc_util.h"
|
|
|
|
#define HFI_VERSION_INFO_MAJOR_VAL 1
|
|
#define HFI_VERSION_INFO_MINOR_VAL 1
|
|
#define HFI_VERSION_INFO_STEP_VAL 0
|
|
#define HFI_VERSION_INFO_STEP_VAL 0
|
|
#define HFI_VERSION_INFO_MAJOR_BMSK 0xFF000000
|
|
#define HFI_VERSION_INFO_MAJOR_SHFT 24
|
|
#define HFI_VERSION_INFO_MINOR_BMSK 0xFFFF00
|
|
#define HFI_VERSION_INFO_MINOR_SHFT 8
|
|
#define HFI_VERSION_INFO_STEP_BMSK 0xFF
|
|
#define HFI_VERSION_INFO_STEP_SHFT 0
|
|
|
|
#define HFI_MAX_POLL_TRY 5
|
|
|
|
#define HFI_MAX_PC_POLL_TRY 150
|
|
#define HFI_POLL_TRY_SLEEP 1
|
|
|
|
static struct hfi_info *g_hfi;
|
|
unsigned int g_icp_mmu_hdl;
|
|
static DEFINE_MUTEX(hfi_cmd_q_mutex);
|
|
static DEFINE_MUTEX(hfi_msg_q_mutex);
|
|
|
|
void cam_hfi_queue_dump(void)
|
|
{
|
|
struct hfi_qtbl *qtbl;
|
|
struct hfi_qtbl_hdr *qtbl_hdr;
|
|
struct hfi_q_hdr *cmd_q_hdr, *msg_q_hdr;
|
|
struct hfi_mem_info *hfi_mem = NULL;
|
|
uint32_t *read_q, *read_ptr;
|
|
int i;
|
|
|
|
hfi_mem = &g_hfi->map;
|
|
if (!hfi_mem) {
|
|
CAM_ERR(CAM_HFI, "Unable to dump queues hfi memory is NULL");
|
|
return;
|
|
}
|
|
|
|
qtbl = (struct hfi_qtbl *)hfi_mem->qtbl.kva;
|
|
qtbl_hdr = &qtbl->q_tbl_hdr;
|
|
CAM_DBG(CAM_HFI,
|
|
"qtbl: version = %x size = %u num q = %u qhdr_size = %u",
|
|
qtbl_hdr->qtbl_version, qtbl_hdr->qtbl_size,
|
|
qtbl_hdr->qtbl_num_q, qtbl_hdr->qtbl_qhdr_size);
|
|
|
|
cmd_q_hdr = &qtbl->q_hdr[Q_CMD];
|
|
CAM_DBG(CAM_HFI, "cmd: size = %u r_idx = %u w_idx = %u addr = %x",
|
|
cmd_q_hdr->qhdr_q_size, cmd_q_hdr->qhdr_read_idx,
|
|
cmd_q_hdr->qhdr_write_idx, hfi_mem->cmd_q.iova);
|
|
read_q = (uint32_t *)g_hfi->map.cmd_q.kva;
|
|
read_ptr = (uint32_t *)(read_q + 0);
|
|
CAM_DBG(CAM_HFI, "CMD Q START");
|
|
for (i = 0; i < ICP_CMD_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT; i++)
|
|
CAM_DBG(CAM_HFI, "Word: %d Data: 0x%08x ", i, read_ptr[i]);
|
|
|
|
msg_q_hdr = &qtbl->q_hdr[Q_MSG];
|
|
CAM_DBG(CAM_HFI, "msg: size = %u r_idx = %u w_idx = %u addr = %x",
|
|
msg_q_hdr->qhdr_q_size, msg_q_hdr->qhdr_read_idx,
|
|
msg_q_hdr->qhdr_write_idx, hfi_mem->msg_q.iova);
|
|
read_q = (uint32_t *)g_hfi->map.msg_q.kva;
|
|
read_ptr = (uint32_t *)(read_q + 0);
|
|
CAM_DBG(CAM_HFI, "MSG Q START");
|
|
for (i = 0; i < ICP_MSG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT; i++)
|
|
CAM_DBG(CAM_HFI, "Word: %d Data: 0x%08x ", i, read_ptr[i]);
|
|
}
|
|
|
|
int hfi_write_cmd(void *cmd_ptr)
|
|
{
|
|
uint32_t size_in_words, empty_space, new_write_idx, read_idx, temp;
|
|
uint32_t *write_q, *write_ptr;
|
|
struct hfi_qtbl *q_tbl;
|
|
struct hfi_q_hdr *q;
|
|
int rc = 0;
|
|
|
|
if (!cmd_ptr) {
|
|
CAM_ERR(CAM_HFI, "command is null");
|
|
return -EINVAL;
|
|
}
|
|
|
|
mutex_lock(&hfi_cmd_q_mutex);
|
|
if (!g_hfi) {
|
|
CAM_ERR(CAM_HFI, "HFI interface not setup");
|
|
rc = -ENODEV;
|
|
goto err;
|
|
}
|
|
|
|
if (g_hfi->hfi_state != HFI_READY ||
|
|
!g_hfi->cmd_q_state) {
|
|
CAM_ERR(CAM_HFI, "HFI state: %u, cmd q state: %u",
|
|
g_hfi->hfi_state, g_hfi->cmd_q_state);
|
|
rc = -ENODEV;
|
|
goto err;
|
|
}
|
|
|
|
q_tbl = (struct hfi_qtbl *)g_hfi->map.qtbl.kva;
|
|
q = &q_tbl->q_hdr[Q_CMD];
|
|
|
|
write_q = (uint32_t *)g_hfi->map.cmd_q.kva;
|
|
|
|
size_in_words = (*(uint32_t *)cmd_ptr) >> BYTE_WORD_SHIFT;
|
|
if (!size_in_words) {
|
|
CAM_DBG(CAM_HFI, "failed");
|
|
rc = -EINVAL;
|
|
goto err;
|
|
}
|
|
|
|
read_idx = q->qhdr_read_idx;
|
|
empty_space = (q->qhdr_write_idx >= read_idx) ?
|
|
(q->qhdr_q_size - (q->qhdr_write_idx - read_idx)) :
|
|
(read_idx - q->qhdr_write_idx);
|
|
if (empty_space <= size_in_words) {
|
|
CAM_ERR(CAM_HFI, "failed: empty space %u, size_in_words %u",
|
|
empty_space, size_in_words);
|
|
rc = -EIO;
|
|
goto err;
|
|
}
|
|
|
|
new_write_idx = q->qhdr_write_idx + size_in_words;
|
|
write_ptr = (uint32_t *)(write_q + q->qhdr_write_idx);
|
|
|
|
if (new_write_idx < q->qhdr_q_size) {
|
|
memcpy(write_ptr, (uint8_t *)cmd_ptr,
|
|
size_in_words << BYTE_WORD_SHIFT);
|
|
} else {
|
|
new_write_idx -= q->qhdr_q_size;
|
|
temp = (size_in_words - new_write_idx) << BYTE_WORD_SHIFT;
|
|
memcpy(write_ptr, (uint8_t *)cmd_ptr, temp);
|
|
memcpy(write_q, (uint8_t *)cmd_ptr + temp,
|
|
new_write_idx << BYTE_WORD_SHIFT);
|
|
}
|
|
|
|
/*
|
|
* To make sure command data in a command queue before
|
|
* updating write index
|
|
*/
|
|
wmb();
|
|
|
|
q->qhdr_write_idx = new_write_idx;
|
|
|
|
/*
|
|
* Before raising interrupt make sure command data is ready for
|
|
* firmware to process
|
|
*/
|
|
wmb();
|
|
cam_io_w_mb((uint32_t)INTR_ENABLE,
|
|
g_hfi->csr_base + HFI_REG_A5_CSR_HOST2ICPINT);
|
|
err:
|
|
mutex_unlock(&hfi_cmd_q_mutex);
|
|
return rc;
|
|
}
|
|
|
|
int hfi_read_message(uint32_t *pmsg, uint8_t q_id,
|
|
uint32_t *words_read)
|
|
{
|
|
struct hfi_qtbl *q_tbl_ptr;
|
|
struct hfi_q_hdr *q;
|
|
uint32_t new_read_idx, size_in_words, word_diff, temp;
|
|
uint32_t *read_q, *read_ptr, *write_ptr;
|
|
uint32_t size_upper_bound = 0;
|
|
int rc = 0;
|
|
|
|
if (!pmsg) {
|
|
CAM_ERR(CAM_HFI, "Invalid msg");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (q_id > Q_DBG) {
|
|
CAM_ERR(CAM_HFI, "Invalid q :%u", q_id);
|
|
return -EINVAL;
|
|
}
|
|
|
|
mutex_lock(&hfi_msg_q_mutex);
|
|
if (!g_hfi) {
|
|
CAM_ERR(CAM_HFI, "hfi not set up yet");
|
|
rc = -ENODEV;
|
|
goto err;
|
|
}
|
|
|
|
if ((g_hfi->hfi_state != HFI_READY) ||
|
|
!g_hfi->msg_q_state) {
|
|
CAM_ERR(CAM_HFI, "hfi state: %u, msg q state: %u",
|
|
g_hfi->hfi_state, g_hfi->msg_q_state);
|
|
rc = -ENODEV;
|
|
goto err;
|
|
}
|
|
|
|
q_tbl_ptr = (struct hfi_qtbl *)g_hfi->map.qtbl.kva;
|
|
q = &q_tbl_ptr->q_hdr[q_id];
|
|
|
|
if (q->qhdr_read_idx == q->qhdr_write_idx) {
|
|
CAM_DBG(CAM_HFI, "Q not ready, state:%u, r idx:%u, w idx:%u",
|
|
g_hfi->hfi_state, q->qhdr_read_idx, q->qhdr_write_idx);
|
|
rc = -EIO;
|
|
goto err;
|
|
}
|
|
|
|
if (q_id == Q_MSG) {
|
|
read_q = (uint32_t *)g_hfi->map.msg_q.kva;
|
|
size_upper_bound = ICP_HFI_MAX_PKT_SIZE_MSGQ_IN_WORDS;
|
|
} else {
|
|
read_q = (uint32_t *)g_hfi->map.dbg_q.kva;
|
|
size_upper_bound = ICP_HFI_MAX_PKT_SIZE_IN_WORDS;
|
|
}
|
|
|
|
read_ptr = (uint32_t *)(read_q + q->qhdr_read_idx);
|
|
write_ptr = (uint32_t *)(read_q + q->qhdr_write_idx);
|
|
|
|
if (write_ptr > read_ptr)
|
|
size_in_words = write_ptr - read_ptr;
|
|
else {
|
|
word_diff = read_ptr - write_ptr;
|
|
if (q_id == Q_MSG)
|
|
size_in_words = (ICP_MSG_Q_SIZE_IN_BYTES >>
|
|
BYTE_WORD_SHIFT) - word_diff;
|
|
else
|
|
size_in_words = (ICP_DBG_Q_SIZE_IN_BYTES >>
|
|
BYTE_WORD_SHIFT) - word_diff;
|
|
}
|
|
|
|
if ((size_in_words == 0) ||
|
|
(size_in_words > size_upper_bound)) {
|
|
CAM_ERR(CAM_HFI, "invalid HFI message packet size - 0x%08x",
|
|
size_in_words << BYTE_WORD_SHIFT);
|
|
q->qhdr_read_idx = q->qhdr_write_idx;
|
|
rc = -EIO;
|
|
goto err;
|
|
}
|
|
|
|
new_read_idx = q->qhdr_read_idx + size_in_words;
|
|
|
|
if (new_read_idx < q->qhdr_q_size) {
|
|
memcpy(pmsg, read_ptr, size_in_words << BYTE_WORD_SHIFT);
|
|
} else {
|
|
new_read_idx -= q->qhdr_q_size;
|
|
temp = (size_in_words - new_read_idx) << BYTE_WORD_SHIFT;
|
|
memcpy(pmsg, read_ptr, temp);
|
|
memcpy((uint8_t *)pmsg + temp, read_q,
|
|
new_read_idx << BYTE_WORD_SHIFT);
|
|
}
|
|
|
|
q->qhdr_read_idx = new_read_idx;
|
|
*words_read = size_in_words;
|
|
/* Memory Barrier to make sure message
|
|
* queue parameters are updated after read
|
|
*/
|
|
wmb();
|
|
err:
|
|
mutex_unlock(&hfi_msg_q_mutex);
|
|
return rc;
|
|
}
|
|
|
|
int hfi_cmd_ubwc_config(uint32_t *ubwc_cfg)
|
|
{
|
|
uint8_t *prop;
|
|
struct hfi_cmd_prop *dbg_prop;
|
|
uint32_t size = 0;
|
|
|
|
size = sizeof(struct hfi_cmd_prop) +
|
|
sizeof(struct hfi_cmd_ubwc_cfg);
|
|
|
|
CAM_DBG(CAM_HFI,
|
|
"size of ubwc %u, ubwc_cfg [rd-0x%x,wr-0x%x]",
|
|
size, ubwc_cfg[0], ubwc_cfg[1]);
|
|
|
|
prop = kzalloc(size, GFP_KERNEL);
|
|
if (!prop)
|
|
return -ENOMEM;
|
|
|
|
dbg_prop = (struct hfi_cmd_prop *)prop;
|
|
dbg_prop->size = size;
|
|
dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY;
|
|
dbg_prop->num_prop = 1;
|
|
dbg_prop->prop_data[0] = HFI_PROP_SYS_UBWC_CFG;
|
|
dbg_prop->prop_data[1] = ubwc_cfg[0];
|
|
dbg_prop->prop_data[2] = ubwc_cfg[1];
|
|
|
|
hfi_write_cmd(prop);
|
|
kfree(prop);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int hfi_cmd_ubwc_config_ext(uint32_t *ubwc_ipe_cfg,
|
|
uint32_t *ubwc_bps_cfg)
|
|
{
|
|
uint8_t *prop;
|
|
struct hfi_cmd_prop *dbg_prop;
|
|
uint32_t size = 0;
|
|
|
|
size = sizeof(struct hfi_cmd_prop) +
|
|
sizeof(struct hfi_cmd_ubwc_cfg_ext);
|
|
|
|
CAM_DBG(CAM_HFI,
|
|
"size of ubwc %u, ubwc_ipe_cfg[rd-0x%x,wr-0x%x] ubwc_bps_cfg[rd-0x%x,wr-0x%x]",
|
|
size, ubwc_ipe_cfg[0], ubwc_ipe_cfg[1],
|
|
ubwc_bps_cfg[0], ubwc_bps_cfg[1]);
|
|
|
|
prop = kzalloc(size, GFP_KERNEL);
|
|
if (!prop)
|
|
return -ENOMEM;
|
|
|
|
dbg_prop = (struct hfi_cmd_prop *)prop;
|
|
dbg_prop->size = size;
|
|
dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY;
|
|
dbg_prop->num_prop = 1;
|
|
dbg_prop->prop_data[0] = HFI_PROPERTY_SYS_UBWC_CONFIG_EX;
|
|
dbg_prop->prop_data[1] = ubwc_bps_cfg[0];
|
|
dbg_prop->prop_data[2] = ubwc_bps_cfg[1];
|
|
dbg_prop->prop_data[3] = ubwc_ipe_cfg[0];
|
|
dbg_prop->prop_data[4] = ubwc_ipe_cfg[1];
|
|
|
|
hfi_write_cmd(prop);
|
|
kfree(prop);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
int hfi_enable_ipe_bps_pc(bool enable, uint32_t core_info)
|
|
{
|
|
uint8_t *prop;
|
|
struct hfi_cmd_prop *dbg_prop;
|
|
uint32_t size = 0;
|
|
|
|
size = sizeof(struct hfi_cmd_prop) +
|
|
sizeof(struct hfi_ipe_bps_pc);
|
|
|
|
prop = kzalloc(size, GFP_KERNEL);
|
|
if (!prop)
|
|
return -ENOMEM;
|
|
|
|
dbg_prop = (struct hfi_cmd_prop *)prop;
|
|
dbg_prop->size = size;
|
|
dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY;
|
|
dbg_prop->num_prop = 1;
|
|
dbg_prop->prop_data[0] = HFI_PROP_SYS_IPEBPS_PC;
|
|
dbg_prop->prop_data[1] = enable;
|
|
dbg_prop->prop_data[2] = core_info;
|
|
|
|
hfi_write_cmd(prop);
|
|
kfree(prop);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int hfi_set_debug_level(u64 a5_dbg_type, uint32_t lvl)
|
|
{
|
|
uint8_t *prop;
|
|
struct hfi_cmd_prop *dbg_prop;
|
|
uint32_t size = 0, val;
|
|
|
|
val = HFI_DEBUG_MSG_LOW |
|
|
HFI_DEBUG_MSG_MEDIUM |
|
|
HFI_DEBUG_MSG_HIGH |
|
|
HFI_DEBUG_MSG_ERROR |
|
|
HFI_DEBUG_MSG_FATAL |
|
|
HFI_DEBUG_MSG_PERF |
|
|
HFI_DEBUG_CFG_WFI |
|
|
HFI_DEBUG_CFG_ARM9WD;
|
|
|
|
if (lvl > val)
|
|
return -EINVAL;
|
|
|
|
size = sizeof(struct hfi_cmd_prop) +
|
|
sizeof(struct hfi_debug);
|
|
|
|
prop = kzalloc(size, GFP_KERNEL);
|
|
if (!prop)
|
|
return -ENOMEM;
|
|
|
|
dbg_prop = (struct hfi_cmd_prop *)prop;
|
|
dbg_prop->size = size;
|
|
dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY;
|
|
dbg_prop->num_prop = 1;
|
|
dbg_prop->prop_data[0] = HFI_PROP_SYS_DEBUG_CFG;
|
|
dbg_prop->prop_data[1] = lvl;
|
|
dbg_prop->prop_data[2] = a5_dbg_type;
|
|
hfi_write_cmd(prop);
|
|
|
|
kfree(prop);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int hfi_set_fw_dump_level(uint32_t lvl)
|
|
{
|
|
uint8_t *prop = NULL;
|
|
struct hfi_cmd_prop *fw_dump_level_switch_prop = NULL;
|
|
uint32_t size = 0;
|
|
|
|
CAM_DBG(CAM_HFI, "fw dump ENTER");
|
|
|
|
size = sizeof(struct hfi_cmd_prop) + sizeof(lvl);
|
|
prop = kzalloc(size, GFP_KERNEL);
|
|
if (!prop)
|
|
return -ENOMEM;
|
|
|
|
fw_dump_level_switch_prop = (struct hfi_cmd_prop *)prop;
|
|
fw_dump_level_switch_prop->size = size;
|
|
fw_dump_level_switch_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY;
|
|
fw_dump_level_switch_prop->num_prop = 1;
|
|
fw_dump_level_switch_prop->prop_data[0] = HFI_PROP_SYS_FW_DUMP_CFG;
|
|
fw_dump_level_switch_prop->prop_data[1] = lvl;
|
|
|
|
CAM_DBG(CAM_HFI, "prop->size = %d\n"
|
|
"prop->pkt_type = %d\n"
|
|
"prop->num_prop = %d\n"
|
|
"prop->prop_data[0] = %d\n"
|
|
"prop->prop_data[1] = %d\n",
|
|
fw_dump_level_switch_prop->size,
|
|
fw_dump_level_switch_prop->pkt_type,
|
|
fw_dump_level_switch_prop->num_prop,
|
|
fw_dump_level_switch_prop->prop_data[0],
|
|
fw_dump_level_switch_prop->prop_data[1]);
|
|
|
|
hfi_write_cmd(prop);
|
|
kfree(prop);
|
|
return 0;
|
|
}
|
|
|
|
void hfi_send_system_cmd(uint32_t type, uint64_t data, uint32_t size)
|
|
{
|
|
switch (type) {
|
|
case HFI_CMD_SYS_INIT: {
|
|
struct hfi_cmd_sys_init init;
|
|
|
|
memset(&init, 0, sizeof(init));
|
|
|
|
init.size = sizeof(struct hfi_cmd_sys_init);
|
|
init.pkt_type = type;
|
|
hfi_write_cmd(&init);
|
|
}
|
|
break;
|
|
case HFI_CMD_SYS_PC_PREP: {
|
|
struct hfi_cmd_pc_prep prep;
|
|
|
|
prep.size = sizeof(struct hfi_cmd_pc_prep);
|
|
prep.pkt_type = type;
|
|
hfi_write_cmd(&prep);
|
|
}
|
|
break;
|
|
case HFI_CMD_SYS_SET_PROPERTY: {
|
|
struct hfi_cmd_prop prop;
|
|
|
|
if ((uint32_t)data == (uint32_t)HFI_PROP_SYS_DEBUG_CFG) {
|
|
prop.size = sizeof(struct hfi_cmd_prop);
|
|
prop.pkt_type = type;
|
|
prop.num_prop = 1;
|
|
prop.prop_data[0] = HFI_PROP_SYS_DEBUG_CFG;
|
|
hfi_write_cmd(&prop);
|
|
}
|
|
}
|
|
break;
|
|
case HFI_CMD_SYS_GET_PROPERTY:
|
|
break;
|
|
case HFI_CMD_SYS_PING: {
|
|
struct hfi_cmd_ping_pkt ping;
|
|
|
|
ping.size = sizeof(struct hfi_cmd_ping_pkt);
|
|
ping.pkt_type = type;
|
|
ping.user_data = (uint64_t)data;
|
|
hfi_write_cmd(&ping);
|
|
}
|
|
break;
|
|
case HFI_CMD_SYS_RESET: {
|
|
struct hfi_cmd_sys_reset_pkt reset;
|
|
|
|
reset.size = sizeof(struct hfi_cmd_sys_reset_pkt);
|
|
reset.pkt_type = type;
|
|
reset.user_data = (uint64_t)data;
|
|
hfi_write_cmd(&reset);
|
|
}
|
|
break;
|
|
case HFI_CMD_IPEBPS_CREATE_HANDLE: {
|
|
struct hfi_cmd_create_handle handle;
|
|
|
|
handle.size = sizeof(struct hfi_cmd_create_handle);
|
|
handle.pkt_type = type;
|
|
handle.handle_type = (uint32_t)data;
|
|
handle.user_data1 = 0;
|
|
hfi_write_cmd(&handle);
|
|
}
|
|
break;
|
|
case HFI_CMD_IPEBPS_ASYNC_COMMAND_INDIRECT:
|
|
break;
|
|
default:
|
|
CAM_ERR(CAM_HFI, "command not supported :%d", type);
|
|
break;
|
|
}
|
|
}
|
|
|
|
|
|
int hfi_get_hw_caps(void *query_buf)
|
|
{
|
|
int i = 0;
|
|
struct cam_icp_query_cap_cmd *query_cmd = NULL;
|
|
|
|
if (!query_buf) {
|
|
CAM_ERR(CAM_HFI, "query buf is NULL");
|
|
return -EINVAL;
|
|
}
|
|
|
|
query_cmd = (struct cam_icp_query_cap_cmd *)query_buf;
|
|
query_cmd->fw_version.major = 0x12;
|
|
query_cmd->fw_version.minor = 0x12;
|
|
query_cmd->fw_version.revision = 0x12;
|
|
|
|
query_cmd->api_version.major = 0x13;
|
|
query_cmd->api_version.minor = 0x13;
|
|
query_cmd->api_version.revision = 0x13;
|
|
|
|
query_cmd->num_ipe = 2;
|
|
query_cmd->num_bps = 1;
|
|
|
|
for (i = 0; i < CAM_ICP_DEV_TYPE_MAX; i++) {
|
|
query_cmd->dev_ver[i].dev_type = i;
|
|
query_cmd->dev_ver[i].hw_ver.major = 0x34 + i;
|
|
query_cmd->dev_ver[i].hw_ver.minor = 0x34 + i;
|
|
query_cmd->dev_ver[i].hw_ver.incr = 0x34 + i;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void cam_hfi_disable_cpu(void __iomem *icp_base)
|
|
{
|
|
uint32_t data;
|
|
uint32_t val;
|
|
uint32_t try = 0;
|
|
|
|
while (try < HFI_MAX_PC_POLL_TRY) {
|
|
data = cam_io_r_mb(icp_base + HFI_REG_A5_CSR_A5_STATUS);
|
|
CAM_DBG(CAM_HFI, "wfi status = %x\n", (int)data);
|
|
|
|
if (data & ICP_CSR_A5_STATUS_WFI)
|
|
break;
|
|
/* Need to poll here to confirm that FW is going trigger wfi
|
|
* and Host can the proceed. No interrupt is expected from FW
|
|
* at this time.
|
|
*/
|
|
usleep_range(HFI_POLL_TRY_SLEEP * 1000,
|
|
(HFI_POLL_TRY_SLEEP * 1000) + 1000);
|
|
try++;
|
|
}
|
|
|
|
val = cam_io_r(icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
val &= ~(ICP_FLAG_CSR_A5_EN | ICP_FLAG_CSR_WAKE_UP_EN);
|
|
cam_io_w_mb(val, icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
|
|
val = cam_io_r(icp_base + HFI_REG_A5_CSR_NSEC_RESET);
|
|
cam_io_w_mb(val, icp_base + HFI_REG_A5_CSR_NSEC_RESET);
|
|
|
|
cam_io_w_mb((uint32_t)ICP_INIT_REQUEST_RESET,
|
|
icp_base + HFI_REG_HOST_ICP_INIT_REQUEST);
|
|
cam_io_w_mb((uint32_t)INTR_DISABLE,
|
|
icp_base + HFI_REG_A5_CSR_A2HOSTINTEN);
|
|
}
|
|
|
|
void cam_hfi_enable_cpu(void __iomem *icp_base)
|
|
{
|
|
cam_io_w_mb((uint32_t)ICP_FLAG_CSR_A5_EN,
|
|
icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
cam_io_w_mb((uint32_t)0x10, icp_base + HFI_REG_A5_CSR_NSEC_RESET);
|
|
}
|
|
|
|
int cam_hfi_resume(struct hfi_mem_info *hfi_mem,
|
|
void __iomem *icp_base, bool debug)
|
|
{
|
|
int rc = 0;
|
|
uint32_t data;
|
|
uint32_t fw_version, status = 0;
|
|
uint32_t retry_cnt = 0;
|
|
|
|
cam_hfi_enable_cpu(icp_base);
|
|
g_hfi->csr_base = icp_base;
|
|
|
|
if (debug) {
|
|
cam_io_w_mb(ICP_FLAG_A5_CTRL_DBG_EN,
|
|
(icp_base + HFI_REG_A5_CSR_A5_CONTROL));
|
|
|
|
/* Barrier needed as next write should be done after
|
|
* sucessful previous write. Next write enable clock
|
|
* gating
|
|
*/
|
|
wmb();
|
|
|
|
cam_io_w_mb((uint32_t)ICP_FLAG_A5_CTRL_EN,
|
|
icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
|
|
} else {
|
|
cam_io_w_mb((uint32_t)ICP_FLAG_A5_CTRL_EN,
|
|
icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
}
|
|
|
|
while (retry_cnt < HFI_MAX_POLL_TRY) {
|
|
readw_poll_timeout((icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE),
|
|
status, (status == ICP_INIT_RESP_SUCCESS), 100, 10000);
|
|
|
|
CAM_DBG(CAM_HFI, "1: status = %u", status);
|
|
status = cam_io_r_mb(icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE);
|
|
CAM_DBG(CAM_HFI, "2: status = %u", status);
|
|
if (status == ICP_INIT_RESP_SUCCESS)
|
|
break;
|
|
|
|
if (status == ICP_INIT_RESP_FAILED) {
|
|
CAM_ERR(CAM_HFI, "ICP Init Failed. status = %u",
|
|
status);
|
|
fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION);
|
|
CAM_ERR(CAM_HFI, "fw version : [%x]", fw_version);
|
|
return -EINVAL;
|
|
}
|
|
retry_cnt++;
|
|
}
|
|
|
|
if ((retry_cnt == HFI_MAX_POLL_TRY) &&
|
|
(status == ICP_INIT_RESP_RESET)) {
|
|
CAM_ERR(CAM_HFI, "Reached Max retries. status = %u",
|
|
status);
|
|
fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION);
|
|
CAM_ERR(CAM_HFI, "fw version : [%x]", fw_version);
|
|
return -EINVAL;
|
|
}
|
|
|
|
cam_io_w_mb((uint32_t)(INTR_ENABLE|INTR_ENABLE_WD0),
|
|
icp_base + HFI_REG_A5_CSR_A2HOSTINTEN);
|
|
|
|
fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION);
|
|
CAM_DBG(CAM_HFI, "fw version : [%x]", fw_version);
|
|
|
|
data = cam_io_r(icp_base + HFI_REG_A5_CSR_A5_STATUS);
|
|
CAM_DBG(CAM_HFI, "wfi status = %x", (int)data);
|
|
|
|
cam_io_w_mb((uint32_t)hfi_mem->qtbl.iova, icp_base + HFI_REG_QTBL_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->sfr_buf.iova,
|
|
icp_base + HFI_REG_SFR_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->shmem.iova,
|
|
icp_base + HFI_REG_SHARED_MEM_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->shmem.len,
|
|
icp_base + HFI_REG_SHARED_MEM_SIZE);
|
|
cam_io_w_mb((uint32_t)hfi_mem->sec_heap.iova,
|
|
icp_base + HFI_REG_UNCACHED_HEAP_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->sec_heap.len,
|
|
icp_base + HFI_REG_UNCACHED_HEAP_SIZE);
|
|
cam_io_w_mb((uint32_t)hfi_mem->qdss.iova,
|
|
icp_base + HFI_REG_QDSS_IOVA);
|
|
cam_io_w_mb((uint32_t)hfi_mem->qdss.len,
|
|
icp_base + HFI_REG_QDSS_IOVA_SIZE);
|
|
cam_io_w_mb((uint32_t)hfi_mem->io_mem.iova,
|
|
icp_base + HFI_REG_IO_REGION_IOVA);
|
|
cam_io_w_mb((uint32_t)hfi_mem->io_mem.len,
|
|
icp_base + HFI_REG_IO_REGION_SIZE);
|
|
|
|
return rc;
|
|
}
|
|
|
|
int cam_hfi_init(uint8_t event_driven_mode, struct hfi_mem_info *hfi_mem,
|
|
void __iomem *icp_base, bool debug)
|
|
{
|
|
int rc = 0;
|
|
struct hfi_qtbl *qtbl;
|
|
struct hfi_qtbl_hdr *qtbl_hdr;
|
|
struct hfi_q_hdr *cmd_q_hdr, *msg_q_hdr, *dbg_q_hdr;
|
|
uint32_t hw_version, soc_version, fw_version, status = 0;
|
|
uint32_t retry_cnt = 0;
|
|
struct sfr_buf *sfr_buffer;
|
|
|
|
mutex_lock(&hfi_cmd_q_mutex);
|
|
mutex_lock(&hfi_msg_q_mutex);
|
|
|
|
if (!g_hfi) {
|
|
g_hfi = kzalloc(sizeof(struct hfi_info), GFP_KERNEL);
|
|
if (!g_hfi) {
|
|
rc = -ENOMEM;
|
|
goto alloc_fail;
|
|
}
|
|
}
|
|
|
|
if (g_hfi->hfi_state != HFI_DEINIT) {
|
|
CAM_ERR(CAM_HFI, "hfi_init: invalid state");
|
|
return -EINVAL;
|
|
}
|
|
|
|
memcpy(&g_hfi->map, hfi_mem, sizeof(g_hfi->map));
|
|
g_hfi->hfi_state = HFI_DEINIT;
|
|
soc_version = socinfo_get_version();
|
|
if (debug) {
|
|
cam_io_w_mb(
|
|
(uint32_t)(ICP_FLAG_CSR_A5_EN | ICP_FLAG_CSR_WAKE_UP_EN |
|
|
ICP_CSR_EDBGRQ | ICP_CSR_DBGSWENABLE),
|
|
icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
msleep(100);
|
|
cam_io_w_mb((uint32_t)(ICP_FLAG_CSR_A5_EN |
|
|
ICP_FLAG_CSR_WAKE_UP_EN | ICP_CSR_EN_CLKGATE_WFI),
|
|
icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
} else {
|
|
/* Due to hardware bug in V1 ICP clock gating has to be
|
|
* disabled, this is supposed to be fixed in V-2. But enabling
|
|
* the clock gating is causing the firmware hang, hence
|
|
* disabling the clock gating on both V1 and V2 until the
|
|
* hardware team root causes this
|
|
*/
|
|
cam_io_w_mb((uint32_t)ICP_FLAG_CSR_A5_EN |
|
|
ICP_FLAG_CSR_WAKE_UP_EN |
|
|
ICP_CSR_EN_CLKGATE_WFI,
|
|
icp_base + HFI_REG_A5_CSR_A5_CONTROL);
|
|
}
|
|
|
|
qtbl = (struct hfi_qtbl *)hfi_mem->qtbl.kva;
|
|
qtbl_hdr = &qtbl->q_tbl_hdr;
|
|
qtbl_hdr->qtbl_version = 0xFFFFFFFF;
|
|
qtbl_hdr->qtbl_size = sizeof(struct hfi_qtbl);
|
|
qtbl_hdr->qtbl_qhdr0_offset = sizeof(struct hfi_qtbl_hdr);
|
|
qtbl_hdr->qtbl_qhdr_size = sizeof(struct hfi_q_hdr);
|
|
qtbl_hdr->qtbl_num_q = ICP_HFI_NUMBER_OF_QS;
|
|
qtbl_hdr->qtbl_num_active_q = ICP_HFI_NUMBER_OF_QS;
|
|
|
|
/* setup host-to-firmware command queue */
|
|
cmd_q_hdr = &qtbl->q_hdr[Q_CMD];
|
|
cmd_q_hdr->qhdr_status = QHDR_ACTIVE;
|
|
cmd_q_hdr->qhdr_start_addr = hfi_mem->cmd_q.iova;
|
|
cmd_q_hdr->qhdr_q_size = ICP_CMD_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT;
|
|
cmd_q_hdr->qhdr_pkt_size = ICP_HFI_VAR_SIZE_PKT;
|
|
cmd_q_hdr->qhdr_pkt_drop_cnt = RESET;
|
|
cmd_q_hdr->qhdr_read_idx = RESET;
|
|
cmd_q_hdr->qhdr_write_idx = RESET;
|
|
|
|
/* setup firmware-to-Host message queue */
|
|
msg_q_hdr = &qtbl->q_hdr[Q_MSG];
|
|
msg_q_hdr->qhdr_status = QHDR_ACTIVE;
|
|
msg_q_hdr->qhdr_start_addr = hfi_mem->msg_q.iova;
|
|
msg_q_hdr->qhdr_q_size = ICP_MSG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT;
|
|
msg_q_hdr->qhdr_pkt_size = ICP_HFI_VAR_SIZE_PKT;
|
|
msg_q_hdr->qhdr_pkt_drop_cnt = RESET;
|
|
msg_q_hdr->qhdr_read_idx = RESET;
|
|
msg_q_hdr->qhdr_write_idx = RESET;
|
|
|
|
/* setup firmware-to-Host message queue */
|
|
dbg_q_hdr = &qtbl->q_hdr[Q_DBG];
|
|
dbg_q_hdr->qhdr_status = QHDR_ACTIVE;
|
|
dbg_q_hdr->qhdr_start_addr = hfi_mem->dbg_q.iova;
|
|
dbg_q_hdr->qhdr_q_size = ICP_DBG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT;
|
|
dbg_q_hdr->qhdr_pkt_size = ICP_HFI_VAR_SIZE_PKT;
|
|
dbg_q_hdr->qhdr_pkt_drop_cnt = RESET;
|
|
dbg_q_hdr->qhdr_read_idx = RESET;
|
|
dbg_q_hdr->qhdr_write_idx = RESET;
|
|
|
|
sfr_buffer = (struct sfr_buf *)hfi_mem->sfr_buf.kva;
|
|
sfr_buffer->size = ICP_MSG_SFR_SIZE_IN_BYTES;
|
|
|
|
switch (event_driven_mode) {
|
|
case INTR_MODE:
|
|
cmd_q_hdr->qhdr_type = Q_CMD;
|
|
cmd_q_hdr->qhdr_rx_wm = SET;
|
|
cmd_q_hdr->qhdr_tx_wm = SET;
|
|
cmd_q_hdr->qhdr_rx_req = SET;
|
|
cmd_q_hdr->qhdr_tx_req = RESET;
|
|
cmd_q_hdr->qhdr_rx_irq_status = RESET;
|
|
cmd_q_hdr->qhdr_tx_irq_status = RESET;
|
|
|
|
msg_q_hdr->qhdr_type = Q_MSG;
|
|
msg_q_hdr->qhdr_rx_wm = SET;
|
|
msg_q_hdr->qhdr_tx_wm = SET;
|
|
msg_q_hdr->qhdr_rx_req = SET;
|
|
msg_q_hdr->qhdr_tx_req = RESET;
|
|
msg_q_hdr->qhdr_rx_irq_status = RESET;
|
|
msg_q_hdr->qhdr_tx_irq_status = RESET;
|
|
|
|
dbg_q_hdr->qhdr_type = Q_DBG;
|
|
dbg_q_hdr->qhdr_rx_wm = SET;
|
|
dbg_q_hdr->qhdr_tx_wm = SET_WM;
|
|
dbg_q_hdr->qhdr_rx_req = RESET;
|
|
dbg_q_hdr->qhdr_tx_req = RESET;
|
|
dbg_q_hdr->qhdr_rx_irq_status = RESET;
|
|
dbg_q_hdr->qhdr_tx_irq_status = RESET;
|
|
|
|
break;
|
|
|
|
case POLL_MODE:
|
|
cmd_q_hdr->qhdr_type = Q_CMD | TX_EVENT_POLL_MODE_2 |
|
|
RX_EVENT_POLL_MODE_2;
|
|
msg_q_hdr->qhdr_type = Q_MSG | TX_EVENT_POLL_MODE_2 |
|
|
RX_EVENT_POLL_MODE_2;
|
|
dbg_q_hdr->qhdr_type = Q_DBG | TX_EVENT_POLL_MODE_2 |
|
|
RX_EVENT_POLL_MODE_2;
|
|
break;
|
|
|
|
case WM_MODE:
|
|
cmd_q_hdr->qhdr_type = Q_CMD | TX_EVENT_DRIVEN_MODE_2 |
|
|
RX_EVENT_DRIVEN_MODE_2;
|
|
cmd_q_hdr->qhdr_rx_wm = SET;
|
|
cmd_q_hdr->qhdr_tx_wm = SET;
|
|
cmd_q_hdr->qhdr_rx_req = RESET;
|
|
cmd_q_hdr->qhdr_tx_req = SET;
|
|
cmd_q_hdr->qhdr_rx_irq_status = RESET;
|
|
cmd_q_hdr->qhdr_tx_irq_status = RESET;
|
|
|
|
msg_q_hdr->qhdr_type = Q_MSG | TX_EVENT_DRIVEN_MODE_2 |
|
|
RX_EVENT_DRIVEN_MODE_2;
|
|
msg_q_hdr->qhdr_rx_wm = SET;
|
|
msg_q_hdr->qhdr_tx_wm = SET;
|
|
msg_q_hdr->qhdr_rx_req = SET;
|
|
msg_q_hdr->qhdr_tx_req = RESET;
|
|
msg_q_hdr->qhdr_rx_irq_status = RESET;
|
|
msg_q_hdr->qhdr_tx_irq_status = RESET;
|
|
|
|
dbg_q_hdr->qhdr_type = Q_DBG | TX_EVENT_DRIVEN_MODE_2 |
|
|
RX_EVENT_DRIVEN_MODE_2;
|
|
dbg_q_hdr->qhdr_rx_wm = SET;
|
|
dbg_q_hdr->qhdr_tx_wm = SET_WM;
|
|
dbg_q_hdr->qhdr_rx_req = RESET;
|
|
dbg_q_hdr->qhdr_tx_req = RESET;
|
|
dbg_q_hdr->qhdr_rx_irq_status = RESET;
|
|
dbg_q_hdr->qhdr_tx_irq_status = RESET;
|
|
break;
|
|
|
|
default:
|
|
CAM_ERR(CAM_HFI, "Invalid event driven mode :%u",
|
|
event_driven_mode);
|
|
break;
|
|
}
|
|
|
|
cam_io_w_mb((uint32_t)hfi_mem->qtbl.iova,
|
|
icp_base + HFI_REG_QTBL_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->sfr_buf.iova,
|
|
icp_base + HFI_REG_SFR_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->shmem.iova,
|
|
icp_base + HFI_REG_SHARED_MEM_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->shmem.len,
|
|
icp_base + HFI_REG_SHARED_MEM_SIZE);
|
|
cam_io_w_mb((uint32_t)hfi_mem->sec_heap.iova,
|
|
icp_base + HFI_REG_UNCACHED_HEAP_PTR);
|
|
cam_io_w_mb((uint32_t)hfi_mem->sec_heap.len,
|
|
icp_base + HFI_REG_UNCACHED_HEAP_SIZE);
|
|
cam_io_w_mb((uint32_t)ICP_INIT_REQUEST_SET,
|
|
icp_base + HFI_REG_HOST_ICP_INIT_REQUEST);
|
|
cam_io_w_mb((uint32_t)hfi_mem->qdss.iova,
|
|
icp_base + HFI_REG_QDSS_IOVA);
|
|
cam_io_w_mb((uint32_t)hfi_mem->qdss.len,
|
|
icp_base + HFI_REG_QDSS_IOVA_SIZE);
|
|
cam_io_w_mb((uint32_t)hfi_mem->io_mem.iova,
|
|
icp_base + HFI_REG_IO_REGION_IOVA);
|
|
cam_io_w_mb((uint32_t)hfi_mem->io_mem.len,
|
|
icp_base + HFI_REG_IO_REGION_SIZE);
|
|
|
|
hw_version = cam_io_r(icp_base + HFI_REG_A5_HW_VERSION);
|
|
|
|
while (retry_cnt < HFI_MAX_POLL_TRY) {
|
|
readw_poll_timeout((icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE),
|
|
status, (status == ICP_INIT_RESP_SUCCESS), 100, 10000);
|
|
|
|
CAM_DBG(CAM_HFI, "1: status = %u rc = %d", status, rc);
|
|
status = cam_io_r_mb(icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE);
|
|
CAM_DBG(CAM_HFI, "2: status = %u rc = %d", status, rc);
|
|
if (status == ICP_INIT_RESP_SUCCESS)
|
|
break;
|
|
|
|
if (status == ICP_INIT_RESP_FAILED) {
|
|
CAM_ERR(CAM_HFI, "ICP Init Failed. status = %u",
|
|
status);
|
|
fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION);
|
|
CAM_ERR(CAM_HFI, "fw version : [%x]", fw_version);
|
|
goto regions_fail;
|
|
}
|
|
retry_cnt++;
|
|
}
|
|
|
|
if ((retry_cnt == HFI_MAX_POLL_TRY) &&
|
|
(status == ICP_INIT_RESP_RESET)) {
|
|
CAM_ERR(CAM_HFI, "Reached Max retries. status = %u",
|
|
status);
|
|
fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION);
|
|
CAM_ERR(CAM_HFI,
|
|
"hw version : : [%x], fw version : [%x]",
|
|
hw_version, fw_version);
|
|
goto regions_fail;
|
|
}
|
|
|
|
fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION);
|
|
CAM_DBG(CAM_HFI, "hw version : : [%x], fw version : [%x]",
|
|
hw_version, fw_version);
|
|
|
|
g_hfi->csr_base = icp_base;
|
|
g_hfi->hfi_state = HFI_READY;
|
|
g_hfi->cmd_q_state = true;
|
|
g_hfi->msg_q_state = true;
|
|
cam_io_w_mb((uint32_t)(INTR_ENABLE|INTR_ENABLE_WD0),
|
|
icp_base + HFI_REG_A5_CSR_A2HOSTINTEN);
|
|
|
|
mutex_unlock(&hfi_cmd_q_mutex);
|
|
mutex_unlock(&hfi_msg_q_mutex);
|
|
|
|
return rc;
|
|
regions_fail:
|
|
kfree(g_hfi);
|
|
g_hfi = NULL;
|
|
alloc_fail:
|
|
mutex_unlock(&hfi_cmd_q_mutex);
|
|
mutex_unlock(&hfi_msg_q_mutex);
|
|
return rc;
|
|
}
|
|
|
|
void cam_hfi_deinit(void __iomem *icp_base)
|
|
{
|
|
mutex_lock(&hfi_cmd_q_mutex);
|
|
mutex_lock(&hfi_msg_q_mutex);
|
|
|
|
if (!g_hfi) {
|
|
CAM_ERR(CAM_HFI, "hfi path not established yet");
|
|
goto err;
|
|
}
|
|
|
|
g_hfi->cmd_q_state = false;
|
|
g_hfi->msg_q_state = false;
|
|
|
|
kzfree(g_hfi);
|
|
g_hfi = NULL;
|
|
|
|
err:
|
|
mutex_unlock(&hfi_cmd_q_mutex);
|
|
mutex_unlock(&hfi_msg_q_mutex);
|
|
}
|