// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2017-2020, The Linux Foundation. All rights reserved. */ #include #include #include #include #include #include #include #include #include "cam_io_util.h" #include "hfi_reg.h" #include "hfi_sys_defs.h" #include "hfi_session_defs.h" #include "hfi_intf.h" #include "cam_icp_hw_mgr_intf.h" #include "cam_debug_util.h" #define HFI_VERSION_INFO_MAJOR_VAL 1 #define HFI_VERSION_INFO_MINOR_VAL 1 #define HFI_VERSION_INFO_STEP_VAL 0 #define HFI_VERSION_INFO_STEP_VAL 0 #define HFI_VERSION_INFO_MAJOR_BMSK 0xFF000000 #define HFI_VERSION_INFO_MAJOR_SHFT 24 #define HFI_VERSION_INFO_MINOR_BMSK 0xFFFF00 #define HFI_VERSION_INFO_MINOR_SHFT 8 #define HFI_VERSION_INFO_STEP_BMSK 0xFF #define HFI_VERSION_INFO_STEP_SHFT 0 #define HFI_MAX_POLL_TRY 5 #define HFI_MAX_PC_POLL_TRY 150 #define HFI_POLL_TRY_SLEEP 1 static struct hfi_info *g_hfi; unsigned int g_icp_mmu_hdl; static DEFINE_MUTEX(hfi_cmd_q_mutex); static DEFINE_MUTEX(hfi_msg_q_mutex); void cam_hfi_queue_dump(void) { struct hfi_qtbl *qtbl; struct hfi_qtbl_hdr *qtbl_hdr; struct hfi_q_hdr *cmd_q_hdr, *msg_q_hdr; struct hfi_mem_info *hfi_mem = NULL; uint32_t *read_q, *read_ptr; int i; hfi_mem = &g_hfi->map; if (!hfi_mem) { CAM_ERR(CAM_HFI, "Unable to dump queues hfi memory is NULL"); return; } qtbl = (struct hfi_qtbl *)hfi_mem->qtbl.kva; qtbl_hdr = &qtbl->q_tbl_hdr; CAM_DBG(CAM_HFI, "qtbl: version = %x size = %u num q = %u qhdr_size = %u", qtbl_hdr->qtbl_version, qtbl_hdr->qtbl_size, qtbl_hdr->qtbl_num_q, qtbl_hdr->qtbl_qhdr_size); cmd_q_hdr = &qtbl->q_hdr[Q_CMD]; CAM_DBG(CAM_HFI, "cmd: size = %u r_idx = %u w_idx = %u addr = %x", cmd_q_hdr->qhdr_q_size, cmd_q_hdr->qhdr_read_idx, cmd_q_hdr->qhdr_write_idx, hfi_mem->cmd_q.iova); read_q = (uint32_t *)g_hfi->map.cmd_q.kva; read_ptr = (uint32_t *)(read_q + 0); CAM_DBG(CAM_HFI, "CMD Q START"); for (i = 0; i < ICP_CMD_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT; i++) CAM_DBG(CAM_HFI, "Word: %d Data: 0x%08x ", i, read_ptr[i]); msg_q_hdr = &qtbl->q_hdr[Q_MSG]; CAM_DBG(CAM_HFI, "msg: size = %u r_idx = %u w_idx = %u addr = %x", msg_q_hdr->qhdr_q_size, msg_q_hdr->qhdr_read_idx, msg_q_hdr->qhdr_write_idx, hfi_mem->msg_q.iova); read_q = (uint32_t *)g_hfi->map.msg_q.kva; read_ptr = (uint32_t *)(read_q + 0); CAM_DBG(CAM_HFI, "MSG Q START"); for (i = 0; i < ICP_MSG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT; i++) CAM_DBG(CAM_HFI, "Word: %d Data: 0x%08x ", i, read_ptr[i]); } int hfi_write_cmd(void *cmd_ptr) { uint32_t size_in_words, empty_space, new_write_idx, read_idx, temp; uint32_t *write_q, *write_ptr; struct hfi_qtbl *q_tbl; struct hfi_q_hdr *q; int rc = 0; if (!cmd_ptr) { CAM_ERR(CAM_HFI, "command is null"); return -EINVAL; } mutex_lock(&hfi_cmd_q_mutex); if (!g_hfi) { CAM_ERR(CAM_HFI, "HFI interface not setup"); rc = -ENODEV; goto err; } if (g_hfi->hfi_state != HFI_READY || !g_hfi->cmd_q_state) { CAM_ERR(CAM_HFI, "HFI state: %u, cmd q state: %u", g_hfi->hfi_state, g_hfi->cmd_q_state); rc = -ENODEV; goto err; } q_tbl = (struct hfi_qtbl *)g_hfi->map.qtbl.kva; q = &q_tbl->q_hdr[Q_CMD]; write_q = (uint32_t *)g_hfi->map.cmd_q.kva; size_in_words = (*(uint32_t *)cmd_ptr) >> BYTE_WORD_SHIFT; if (!size_in_words) { CAM_DBG(CAM_HFI, "failed"); rc = -EINVAL; goto err; } read_idx = q->qhdr_read_idx; empty_space = (q->qhdr_write_idx >= read_idx) ? (q->qhdr_q_size - (q->qhdr_write_idx - read_idx)) : (read_idx - q->qhdr_write_idx); if (empty_space <= size_in_words) { CAM_ERR(CAM_HFI, "failed: empty space %u, size_in_words %u", empty_space, size_in_words); rc = -EIO; goto err; } new_write_idx = q->qhdr_write_idx + size_in_words; write_ptr = (uint32_t *)(write_q + q->qhdr_write_idx); if (new_write_idx < q->qhdr_q_size) { memcpy(write_ptr, (uint8_t *)cmd_ptr, size_in_words << BYTE_WORD_SHIFT); } else { new_write_idx -= q->qhdr_q_size; temp = (size_in_words - new_write_idx) << BYTE_WORD_SHIFT; memcpy(write_ptr, (uint8_t *)cmd_ptr, temp); memcpy(write_q, (uint8_t *)cmd_ptr + temp, new_write_idx << BYTE_WORD_SHIFT); } /* * To make sure command data in a command queue before * updating write index */ wmb(); q->qhdr_write_idx = new_write_idx; /* * Before raising interrupt make sure command data is ready for * firmware to process */ wmb(); cam_io_w_mb((uint32_t)INTR_ENABLE, g_hfi->csr_base + HFI_REG_A5_CSR_HOST2ICPINT); err: mutex_unlock(&hfi_cmd_q_mutex); return rc; } int hfi_read_message(uint32_t *pmsg, uint8_t q_id, uint32_t *words_read) { struct hfi_qtbl *q_tbl_ptr; struct hfi_q_hdr *q; uint32_t new_read_idx, size_in_words, word_diff, temp; uint32_t *read_q, *read_ptr, *write_ptr; uint32_t size_upper_bound = 0; int rc = 0; if (!pmsg) { CAM_ERR(CAM_HFI, "Invalid msg"); return -EINVAL; } if (q_id > Q_DBG) { CAM_ERR(CAM_HFI, "Invalid q :%u", q_id); return -EINVAL; } mutex_lock(&hfi_msg_q_mutex); if (!g_hfi) { CAM_ERR(CAM_HFI, "hfi not set up yet"); rc = -ENODEV; goto err; } if ((g_hfi->hfi_state != HFI_READY) || !g_hfi->msg_q_state) { CAM_ERR(CAM_HFI, "hfi state: %u, msg q state: %u", g_hfi->hfi_state, g_hfi->msg_q_state); rc = -ENODEV; goto err; } q_tbl_ptr = (struct hfi_qtbl *)g_hfi->map.qtbl.kva; q = &q_tbl_ptr->q_hdr[q_id]; if (q->qhdr_read_idx == q->qhdr_write_idx) { CAM_DBG(CAM_HFI, "Q not ready, state:%u, r idx:%u, w idx:%u", g_hfi->hfi_state, q->qhdr_read_idx, q->qhdr_write_idx); rc = -EIO; goto err; } if (q_id == Q_MSG) { read_q = (uint32_t *)g_hfi->map.msg_q.kva; size_upper_bound = ICP_HFI_MAX_PKT_SIZE_MSGQ_IN_WORDS; } else { read_q = (uint32_t *)g_hfi->map.dbg_q.kva; size_upper_bound = ICP_HFI_MAX_PKT_SIZE_IN_WORDS; } read_ptr = (uint32_t *)(read_q + q->qhdr_read_idx); write_ptr = (uint32_t *)(read_q + q->qhdr_write_idx); if (write_ptr > read_ptr) size_in_words = write_ptr - read_ptr; else { word_diff = read_ptr - write_ptr; if (q_id == Q_MSG) size_in_words = (ICP_MSG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT) - word_diff; else size_in_words = (ICP_DBG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT) - word_diff; } if ((size_in_words == 0) || (size_in_words > size_upper_bound)) { CAM_ERR(CAM_HFI, "invalid HFI message packet size - 0x%08x", size_in_words << BYTE_WORD_SHIFT); q->qhdr_read_idx = q->qhdr_write_idx; rc = -EIO; goto err; } new_read_idx = q->qhdr_read_idx + size_in_words; if (new_read_idx < q->qhdr_q_size) { memcpy(pmsg, read_ptr, size_in_words << BYTE_WORD_SHIFT); } else { new_read_idx -= q->qhdr_q_size; temp = (size_in_words - new_read_idx) << BYTE_WORD_SHIFT; memcpy(pmsg, read_ptr, temp); memcpy((uint8_t *)pmsg + temp, read_q, new_read_idx << BYTE_WORD_SHIFT); } q->qhdr_read_idx = new_read_idx; *words_read = size_in_words; /* Memory Barrier to make sure message * queue parameters are updated after read */ wmb(); err: mutex_unlock(&hfi_msg_q_mutex); return rc; } int hfi_cmd_ubwc_config(uint32_t *ubwc_cfg) { uint8_t *prop; struct hfi_cmd_prop *dbg_prop; uint32_t size = 0; size = sizeof(struct hfi_cmd_prop) + sizeof(struct hfi_cmd_ubwc_cfg); CAM_DBG(CAM_HFI, "size of ubwc %u, ubwc_cfg [rd-0x%x,wr-0x%x]", size, ubwc_cfg[0], ubwc_cfg[1]); prop = kzalloc(size, GFP_KERNEL); if (!prop) return -ENOMEM; dbg_prop = (struct hfi_cmd_prop *)prop; dbg_prop->size = size; dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY; dbg_prop->num_prop = 1; dbg_prop->prop_data[0] = HFI_PROP_SYS_UBWC_CFG; dbg_prop->prop_data[1] = ubwc_cfg[0]; dbg_prop->prop_data[2] = ubwc_cfg[1]; hfi_write_cmd(prop); kfree(prop); return 0; } int hfi_cmd_ubwc_config_ext(uint32_t *ubwc_ipe_cfg, uint32_t *ubwc_bps_cfg) { uint8_t *prop; struct hfi_cmd_prop *dbg_prop; uint32_t size = 0; size = sizeof(struct hfi_cmd_prop) + sizeof(struct hfi_cmd_ubwc_cfg_ext); CAM_DBG(CAM_HFI, "size of ubwc %u, ubwc_ipe_cfg[rd-0x%x,wr-0x%x] ubwc_bps_cfg[rd-0x%x,wr-0x%x]", size, ubwc_ipe_cfg[0], ubwc_ipe_cfg[1], ubwc_bps_cfg[0], ubwc_bps_cfg[1]); prop = kzalloc(size, GFP_KERNEL); if (!prop) return -ENOMEM; dbg_prop = (struct hfi_cmd_prop *)prop; dbg_prop->size = size; dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY; dbg_prop->num_prop = 1; dbg_prop->prop_data[0] = HFI_PROPERTY_SYS_UBWC_CONFIG_EX; dbg_prop->prop_data[1] = ubwc_bps_cfg[0]; dbg_prop->prop_data[2] = ubwc_bps_cfg[1]; dbg_prop->prop_data[3] = ubwc_ipe_cfg[0]; dbg_prop->prop_data[4] = ubwc_ipe_cfg[1]; hfi_write_cmd(prop); kfree(prop); return 0; } int hfi_enable_ipe_bps_pc(bool enable, uint32_t core_info) { uint8_t *prop; struct hfi_cmd_prop *dbg_prop; uint32_t size = 0; size = sizeof(struct hfi_cmd_prop) + sizeof(struct hfi_ipe_bps_pc); prop = kzalloc(size, GFP_KERNEL); if (!prop) return -ENOMEM; dbg_prop = (struct hfi_cmd_prop *)prop; dbg_prop->size = size; dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY; dbg_prop->num_prop = 1; dbg_prop->prop_data[0] = HFI_PROP_SYS_IPEBPS_PC; dbg_prop->prop_data[1] = enable; dbg_prop->prop_data[2] = core_info; hfi_write_cmd(prop); kfree(prop); return 0; } int hfi_set_debug_level(u64 a5_dbg_type, uint32_t lvl) { uint8_t *prop; struct hfi_cmd_prop *dbg_prop; uint32_t size = 0, val; val = HFI_DEBUG_MSG_LOW | HFI_DEBUG_MSG_MEDIUM | HFI_DEBUG_MSG_HIGH | HFI_DEBUG_MSG_ERROR | HFI_DEBUG_MSG_FATAL | HFI_DEBUG_MSG_PERF | HFI_DEBUG_CFG_WFI | HFI_DEBUG_CFG_ARM9WD; if (lvl > val) return -EINVAL; size = sizeof(struct hfi_cmd_prop) + sizeof(struct hfi_debug); prop = kzalloc(size, GFP_KERNEL); if (!prop) return -ENOMEM; dbg_prop = (struct hfi_cmd_prop *)prop; dbg_prop->size = size; dbg_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY; dbg_prop->num_prop = 1; dbg_prop->prop_data[0] = HFI_PROP_SYS_DEBUG_CFG; dbg_prop->prop_data[1] = lvl; dbg_prop->prop_data[2] = a5_dbg_type; hfi_write_cmd(prop); kfree(prop); return 0; } int hfi_set_fw_dump_level(uint32_t lvl) { uint8_t *prop = NULL; struct hfi_cmd_prop *fw_dump_level_switch_prop = NULL; uint32_t size = 0; CAM_DBG(CAM_HFI, "fw dump ENTER"); size = sizeof(struct hfi_cmd_prop) + sizeof(lvl); prop = kzalloc(size, GFP_KERNEL); if (!prop) return -ENOMEM; fw_dump_level_switch_prop = (struct hfi_cmd_prop *)prop; fw_dump_level_switch_prop->size = size; fw_dump_level_switch_prop->pkt_type = HFI_CMD_SYS_SET_PROPERTY; fw_dump_level_switch_prop->num_prop = 1; fw_dump_level_switch_prop->prop_data[0] = HFI_PROP_SYS_FW_DUMP_CFG; fw_dump_level_switch_prop->prop_data[1] = lvl; CAM_DBG(CAM_HFI, "prop->size = %d\n" "prop->pkt_type = %d\n" "prop->num_prop = %d\n" "prop->prop_data[0] = %d\n" "prop->prop_data[1] = %d\n", fw_dump_level_switch_prop->size, fw_dump_level_switch_prop->pkt_type, fw_dump_level_switch_prop->num_prop, fw_dump_level_switch_prop->prop_data[0], fw_dump_level_switch_prop->prop_data[1]); hfi_write_cmd(prop); kfree(prop); return 0; } void hfi_send_system_cmd(uint32_t type, uint64_t data, uint32_t size) { switch (type) { case HFI_CMD_SYS_INIT: { struct hfi_cmd_sys_init init; memset(&init, 0, sizeof(init)); init.size = sizeof(struct hfi_cmd_sys_init); init.pkt_type = type; hfi_write_cmd(&init); } break; case HFI_CMD_SYS_PC_PREP: { struct hfi_cmd_pc_prep prep; prep.size = sizeof(struct hfi_cmd_pc_prep); prep.pkt_type = type; hfi_write_cmd(&prep); } break; case HFI_CMD_SYS_SET_PROPERTY: { struct hfi_cmd_prop prop; if ((uint32_t)data == (uint32_t)HFI_PROP_SYS_DEBUG_CFG) { prop.size = sizeof(struct hfi_cmd_prop); prop.pkt_type = type; prop.num_prop = 1; prop.prop_data[0] = HFI_PROP_SYS_DEBUG_CFG; hfi_write_cmd(&prop); } } break; case HFI_CMD_SYS_GET_PROPERTY: break; case HFI_CMD_SYS_PING: { struct hfi_cmd_ping_pkt ping; ping.size = sizeof(struct hfi_cmd_ping_pkt); ping.pkt_type = type; ping.user_data = (uint64_t)data; hfi_write_cmd(&ping); } break; case HFI_CMD_SYS_RESET: { struct hfi_cmd_sys_reset_pkt reset; reset.size = sizeof(struct hfi_cmd_sys_reset_pkt); reset.pkt_type = type; reset.user_data = (uint64_t)data; hfi_write_cmd(&reset); } break; case HFI_CMD_IPEBPS_CREATE_HANDLE: { struct hfi_cmd_create_handle handle; handle.size = sizeof(struct hfi_cmd_create_handle); handle.pkt_type = type; handle.handle_type = (uint32_t)data; handle.user_data1 = 0; hfi_write_cmd(&handle); } break; case HFI_CMD_IPEBPS_ASYNC_COMMAND_INDIRECT: break; default: CAM_ERR(CAM_HFI, "command not supported :%d", type); break; } } int hfi_get_hw_caps(void *query_buf) { int i = 0; struct cam_icp_query_cap_cmd *query_cmd = NULL; if (!query_buf) { CAM_ERR(CAM_HFI, "query buf is NULL"); return -EINVAL; } query_cmd = (struct cam_icp_query_cap_cmd *)query_buf; query_cmd->fw_version.major = 0x12; query_cmd->fw_version.minor = 0x12; query_cmd->fw_version.revision = 0x12; query_cmd->api_version.major = 0x13; query_cmd->api_version.minor = 0x13; query_cmd->api_version.revision = 0x13; query_cmd->num_ipe = 2; query_cmd->num_bps = 1; for (i = 0; i < CAM_ICP_DEV_TYPE_MAX; i++) { query_cmd->dev_ver[i].dev_type = i; query_cmd->dev_ver[i].hw_ver.major = 0x34 + i; query_cmd->dev_ver[i].hw_ver.minor = 0x34 + i; query_cmd->dev_ver[i].hw_ver.incr = 0x34 + i; } return 0; } void cam_hfi_disable_cpu(void __iomem *icp_base) { uint32_t data; uint32_t val; uint32_t try = 0; while (try < HFI_MAX_PC_POLL_TRY) { data = cam_io_r_mb(icp_base + HFI_REG_A5_CSR_A5_STATUS); CAM_DBG(CAM_HFI, "wfi status = %x\n", (int)data); if (data & ICP_CSR_A5_STATUS_WFI) break; /* Need to poll here to confirm that FW is going trigger wfi * and Host can the proceed. No interrupt is expected from FW * at this time. */ usleep_range(HFI_POLL_TRY_SLEEP * 1000, (HFI_POLL_TRY_SLEEP * 1000) + 1000); try++; } val = cam_io_r(icp_base + HFI_REG_A5_CSR_A5_CONTROL); val &= ~(ICP_FLAG_CSR_A5_EN | ICP_FLAG_CSR_WAKE_UP_EN); cam_io_w_mb(val, icp_base + HFI_REG_A5_CSR_A5_CONTROL); val = cam_io_r(icp_base + HFI_REG_A5_CSR_NSEC_RESET); cam_io_w_mb(val, icp_base + HFI_REG_A5_CSR_NSEC_RESET); cam_io_w_mb((uint32_t)ICP_INIT_REQUEST_RESET, icp_base + HFI_REG_HOST_ICP_INIT_REQUEST); cam_io_w_mb((uint32_t)INTR_DISABLE, icp_base + HFI_REG_A5_CSR_A2HOSTINTEN); } void cam_hfi_enable_cpu(void __iomem *icp_base) { cam_io_w_mb((uint32_t)ICP_FLAG_CSR_A5_EN, icp_base + HFI_REG_A5_CSR_A5_CONTROL); cam_io_w_mb((uint32_t)0x10, icp_base + HFI_REG_A5_CSR_NSEC_RESET); } int cam_hfi_resume(struct hfi_mem_info *hfi_mem, void __iomem *icp_base, bool debug) { int rc = 0; uint32_t data; uint32_t fw_version, status = 0; uint32_t retry_cnt = 0; cam_hfi_enable_cpu(icp_base); g_hfi->csr_base = icp_base; if (debug) { cam_io_w_mb(ICP_FLAG_A5_CTRL_DBG_EN, (icp_base + HFI_REG_A5_CSR_A5_CONTROL)); /* Barrier needed as next write should be done after * sucessful previous write. Next write enable clock * gating */ wmb(); cam_io_w_mb((uint32_t)ICP_FLAG_A5_CTRL_EN, icp_base + HFI_REG_A5_CSR_A5_CONTROL); } else { cam_io_w_mb((uint32_t)ICP_FLAG_A5_CTRL_EN, icp_base + HFI_REG_A5_CSR_A5_CONTROL); } while (retry_cnt < HFI_MAX_POLL_TRY) { readw_poll_timeout((icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE), status, (status == ICP_INIT_RESP_SUCCESS), 100, 10000); CAM_DBG(CAM_HFI, "1: status = %u", status); status = cam_io_r_mb(icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE); CAM_DBG(CAM_HFI, "2: status = %u", status); if (status == ICP_INIT_RESP_SUCCESS) break; if (status == ICP_INIT_RESP_FAILED) { CAM_ERR(CAM_HFI, "ICP Init Failed. status = %u", status); fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION); CAM_ERR(CAM_HFI, "fw version : [%x]", fw_version); return -EINVAL; } retry_cnt++; } if ((retry_cnt == HFI_MAX_POLL_TRY) && (status == ICP_INIT_RESP_RESET)) { CAM_ERR(CAM_HFI, "Reached Max retries. status = %u", status); fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION); CAM_ERR(CAM_HFI, "fw version : [%x]", fw_version); return -EINVAL; } cam_io_w_mb((uint32_t)(INTR_ENABLE|INTR_ENABLE_WD0), icp_base + HFI_REG_A5_CSR_A2HOSTINTEN); fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION); CAM_DBG(CAM_HFI, "fw version : [%x]", fw_version); data = cam_io_r(icp_base + HFI_REG_A5_CSR_A5_STATUS); CAM_DBG(CAM_HFI, "wfi status = %x", (int)data); cam_io_w_mb((uint32_t)hfi_mem->qtbl.iova, icp_base + HFI_REG_QTBL_PTR); cam_io_w_mb((uint32_t)hfi_mem->sfr_buf.iova, icp_base + HFI_REG_SFR_PTR); cam_io_w_mb((uint32_t)hfi_mem->shmem.iova, icp_base + HFI_REG_SHARED_MEM_PTR); cam_io_w_mb((uint32_t)hfi_mem->shmem.len, icp_base + HFI_REG_SHARED_MEM_SIZE); cam_io_w_mb((uint32_t)hfi_mem->sec_heap.iova, icp_base + HFI_REG_UNCACHED_HEAP_PTR); cam_io_w_mb((uint32_t)hfi_mem->sec_heap.len, icp_base + HFI_REG_UNCACHED_HEAP_SIZE); cam_io_w_mb((uint32_t)hfi_mem->qdss.iova, icp_base + HFI_REG_QDSS_IOVA); cam_io_w_mb((uint32_t)hfi_mem->qdss.len, icp_base + HFI_REG_QDSS_IOVA_SIZE); cam_io_w_mb((uint32_t)hfi_mem->io_mem.iova, icp_base + HFI_REG_IO_REGION_IOVA); cam_io_w_mb((uint32_t)hfi_mem->io_mem.len, icp_base + HFI_REG_IO_REGION_SIZE); cam_io_w_mb((uint32_t)hfi_mem->io_mem2.iova, icp_base + HFI_REG_IO2_REGION_IOVA); cam_io_w_mb((uint32_t)hfi_mem->io_mem2.len, icp_base + HFI_REG_IO2_REGION_SIZE); CAM_INFO(CAM_HFI, "Resume IO1 : [0x%x 0x%x] IO2 [0x%x 0x%x]", hfi_mem->io_mem.iova, hfi_mem->io_mem.len, hfi_mem->io_mem2.iova, hfi_mem->io_mem2.len); return rc; } int cam_hfi_init(uint8_t event_driven_mode, struct hfi_mem_info *hfi_mem, void __iomem *icp_base, bool debug) { int rc = 0; struct hfi_qtbl *qtbl; struct hfi_qtbl_hdr *qtbl_hdr; struct hfi_q_hdr *cmd_q_hdr, *msg_q_hdr, *dbg_q_hdr; uint32_t hw_version, fw_version, status = 0; uint32_t retry_cnt = 0; struct sfr_buf *sfr_buffer; mutex_lock(&hfi_cmd_q_mutex); mutex_lock(&hfi_msg_q_mutex); if (!g_hfi) { g_hfi = kzalloc(sizeof(struct hfi_info), GFP_KERNEL); if (!g_hfi) { rc = -ENOMEM; goto alloc_fail; } } if (g_hfi->hfi_state != HFI_DEINIT) { CAM_ERR(CAM_HFI, "hfi_init: invalid state"); return -EINVAL; } memcpy(&g_hfi->map, hfi_mem, sizeof(g_hfi->map)); g_hfi->hfi_state = HFI_DEINIT; if (debug) { cam_io_w_mb( (uint32_t)(ICP_FLAG_CSR_A5_EN | ICP_FLAG_CSR_WAKE_UP_EN | ICP_CSR_EDBGRQ | ICP_CSR_DBGSWENABLE), icp_base + HFI_REG_A5_CSR_A5_CONTROL); msleep(100); cam_io_w_mb((uint32_t)(ICP_FLAG_CSR_A5_EN | ICP_FLAG_CSR_WAKE_UP_EN | ICP_CSR_EN_CLKGATE_WFI), icp_base + HFI_REG_A5_CSR_A5_CONTROL); } else { /* Due to hardware bug in V1 ICP clock gating has to be * disabled, this is supposed to be fixed in V-2. But enabling * the clock gating is causing the firmware hang, hence * disabling the clock gating on both V1 and V2 until the * hardware team root causes this */ cam_io_w_mb((uint32_t)ICP_FLAG_CSR_A5_EN | ICP_FLAG_CSR_WAKE_UP_EN | ICP_CSR_EN_CLKGATE_WFI, icp_base + HFI_REG_A5_CSR_A5_CONTROL); } qtbl = (struct hfi_qtbl *)hfi_mem->qtbl.kva; qtbl_hdr = &qtbl->q_tbl_hdr; qtbl_hdr->qtbl_version = 0xFFFFFFFF; qtbl_hdr->qtbl_size = sizeof(struct hfi_qtbl); qtbl_hdr->qtbl_qhdr0_offset = sizeof(struct hfi_qtbl_hdr); qtbl_hdr->qtbl_qhdr_size = sizeof(struct hfi_q_hdr); qtbl_hdr->qtbl_num_q = ICP_HFI_NUMBER_OF_QS; qtbl_hdr->qtbl_num_active_q = ICP_HFI_NUMBER_OF_QS; /* setup host-to-firmware command queue */ cmd_q_hdr = &qtbl->q_hdr[Q_CMD]; cmd_q_hdr->qhdr_status = QHDR_ACTIVE; cmd_q_hdr->qhdr_start_addr = hfi_mem->cmd_q.iova; cmd_q_hdr->qhdr_q_size = ICP_CMD_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT; cmd_q_hdr->qhdr_pkt_size = ICP_HFI_VAR_SIZE_PKT; cmd_q_hdr->qhdr_pkt_drop_cnt = RESET; cmd_q_hdr->qhdr_read_idx = RESET; cmd_q_hdr->qhdr_write_idx = RESET; /* setup firmware-to-Host message queue */ msg_q_hdr = &qtbl->q_hdr[Q_MSG]; msg_q_hdr->qhdr_status = QHDR_ACTIVE; msg_q_hdr->qhdr_start_addr = hfi_mem->msg_q.iova; msg_q_hdr->qhdr_q_size = ICP_MSG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT; msg_q_hdr->qhdr_pkt_size = ICP_HFI_VAR_SIZE_PKT; msg_q_hdr->qhdr_pkt_drop_cnt = RESET; msg_q_hdr->qhdr_read_idx = RESET; msg_q_hdr->qhdr_write_idx = RESET; /* setup firmware-to-Host message queue */ dbg_q_hdr = &qtbl->q_hdr[Q_DBG]; dbg_q_hdr->qhdr_status = QHDR_ACTIVE; dbg_q_hdr->qhdr_start_addr = hfi_mem->dbg_q.iova; dbg_q_hdr->qhdr_q_size = ICP_DBG_Q_SIZE_IN_BYTES >> BYTE_WORD_SHIFT; dbg_q_hdr->qhdr_pkt_size = ICP_HFI_VAR_SIZE_PKT; dbg_q_hdr->qhdr_pkt_drop_cnt = RESET; dbg_q_hdr->qhdr_read_idx = RESET; dbg_q_hdr->qhdr_write_idx = RESET; sfr_buffer = (struct sfr_buf *)hfi_mem->sfr_buf.kva; sfr_buffer->size = ICP_MSG_SFR_SIZE_IN_BYTES; switch (event_driven_mode) { case INTR_MODE: cmd_q_hdr->qhdr_type = Q_CMD; cmd_q_hdr->qhdr_rx_wm = SET; cmd_q_hdr->qhdr_tx_wm = SET; cmd_q_hdr->qhdr_rx_req = SET; cmd_q_hdr->qhdr_tx_req = RESET; cmd_q_hdr->qhdr_rx_irq_status = RESET; cmd_q_hdr->qhdr_tx_irq_status = RESET; msg_q_hdr->qhdr_type = Q_MSG; msg_q_hdr->qhdr_rx_wm = SET; msg_q_hdr->qhdr_tx_wm = SET; msg_q_hdr->qhdr_rx_req = SET; msg_q_hdr->qhdr_tx_req = RESET; msg_q_hdr->qhdr_rx_irq_status = RESET; msg_q_hdr->qhdr_tx_irq_status = RESET; dbg_q_hdr->qhdr_type = Q_DBG; dbg_q_hdr->qhdr_rx_wm = SET; dbg_q_hdr->qhdr_tx_wm = SET_WM; dbg_q_hdr->qhdr_rx_req = RESET; dbg_q_hdr->qhdr_tx_req = RESET; dbg_q_hdr->qhdr_rx_irq_status = RESET; dbg_q_hdr->qhdr_tx_irq_status = RESET; break; case POLL_MODE: cmd_q_hdr->qhdr_type = Q_CMD | TX_EVENT_POLL_MODE_2 | RX_EVENT_POLL_MODE_2; msg_q_hdr->qhdr_type = Q_MSG | TX_EVENT_POLL_MODE_2 | RX_EVENT_POLL_MODE_2; dbg_q_hdr->qhdr_type = Q_DBG | TX_EVENT_POLL_MODE_2 | RX_EVENT_POLL_MODE_2; break; case WM_MODE: cmd_q_hdr->qhdr_type = Q_CMD | TX_EVENT_DRIVEN_MODE_2 | RX_EVENT_DRIVEN_MODE_2; cmd_q_hdr->qhdr_rx_wm = SET; cmd_q_hdr->qhdr_tx_wm = SET; cmd_q_hdr->qhdr_rx_req = RESET; cmd_q_hdr->qhdr_tx_req = SET; cmd_q_hdr->qhdr_rx_irq_status = RESET; cmd_q_hdr->qhdr_tx_irq_status = RESET; msg_q_hdr->qhdr_type = Q_MSG | TX_EVENT_DRIVEN_MODE_2 | RX_EVENT_DRIVEN_MODE_2; msg_q_hdr->qhdr_rx_wm = SET; msg_q_hdr->qhdr_tx_wm = SET; msg_q_hdr->qhdr_rx_req = SET; msg_q_hdr->qhdr_tx_req = RESET; msg_q_hdr->qhdr_rx_irq_status = RESET; msg_q_hdr->qhdr_tx_irq_status = RESET; dbg_q_hdr->qhdr_type = Q_DBG | TX_EVENT_DRIVEN_MODE_2 | RX_EVENT_DRIVEN_MODE_2; dbg_q_hdr->qhdr_rx_wm = SET; dbg_q_hdr->qhdr_tx_wm = SET_WM; dbg_q_hdr->qhdr_rx_req = RESET; dbg_q_hdr->qhdr_tx_req = RESET; dbg_q_hdr->qhdr_rx_irq_status = RESET; dbg_q_hdr->qhdr_tx_irq_status = RESET; break; default: CAM_ERR(CAM_HFI, "Invalid event driven mode :%u", event_driven_mode); break; } cam_io_w_mb((uint32_t)hfi_mem->qtbl.iova, icp_base + HFI_REG_QTBL_PTR); cam_io_w_mb((uint32_t)hfi_mem->sfr_buf.iova, icp_base + HFI_REG_SFR_PTR); cam_io_w_mb((uint32_t)hfi_mem->shmem.iova, icp_base + HFI_REG_SHARED_MEM_PTR); cam_io_w_mb((uint32_t)hfi_mem->shmem.len, icp_base + HFI_REG_SHARED_MEM_SIZE); cam_io_w_mb((uint32_t)hfi_mem->sec_heap.iova, icp_base + HFI_REG_UNCACHED_HEAP_PTR); cam_io_w_mb((uint32_t)hfi_mem->sec_heap.len, icp_base + HFI_REG_UNCACHED_HEAP_SIZE); cam_io_w_mb((uint32_t)ICP_INIT_REQUEST_SET, icp_base + HFI_REG_HOST_ICP_INIT_REQUEST); cam_io_w_mb((uint32_t)hfi_mem->qdss.iova, icp_base + HFI_REG_QDSS_IOVA); cam_io_w_mb((uint32_t)hfi_mem->qdss.len, icp_base + HFI_REG_QDSS_IOVA_SIZE); cam_io_w_mb((uint32_t)hfi_mem->io_mem.iova, icp_base + HFI_REG_IO_REGION_IOVA); cam_io_w_mb((uint32_t)hfi_mem->io_mem.len, icp_base + HFI_REG_IO_REGION_SIZE); cam_io_w_mb((uint32_t)hfi_mem->io_mem2.iova, icp_base + HFI_REG_IO2_REGION_IOVA); cam_io_w_mb((uint32_t)hfi_mem->io_mem2.len, icp_base + HFI_REG_IO2_REGION_SIZE); CAM_INFO(CAM_HFI, "Init IO1 : [0x%x 0x%x] IO2 [0x%x 0x%x]", hfi_mem->io_mem.iova, hfi_mem->io_mem.len, hfi_mem->io_mem2.iova, hfi_mem->io_mem2.len); hw_version = cam_io_r(icp_base + HFI_REG_A5_HW_VERSION); while (retry_cnt < HFI_MAX_POLL_TRY) { readw_poll_timeout((icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE), status, (status == ICP_INIT_RESP_SUCCESS), 100, 10000); CAM_DBG(CAM_HFI, "1: status = %u rc = %d", status, rc); status = cam_io_r_mb(icp_base + HFI_REG_ICP_HOST_INIT_RESPONSE); CAM_DBG(CAM_HFI, "2: status = %u rc = %d", status, rc); if (status == ICP_INIT_RESP_SUCCESS) break; if (status == ICP_INIT_RESP_FAILED) { CAM_ERR(CAM_HFI, "ICP Init Failed. status = %u", status); fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION); CAM_ERR(CAM_HFI, "fw version : [%x]", fw_version); goto regions_fail; } retry_cnt++; } if ((retry_cnt == HFI_MAX_POLL_TRY) && (status == ICP_INIT_RESP_RESET)) { CAM_ERR(CAM_HFI, "Reached Max retries. status = %u", status); fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION); CAM_ERR(CAM_HFI, "hw version : : [%x], fw version : [%x]", hw_version, fw_version); goto regions_fail; } fw_version = cam_io_r(icp_base + HFI_REG_FW_VERSION); CAM_DBG(CAM_HFI, "hw version : : [%x], fw version : [%x]", hw_version, fw_version); g_hfi->csr_base = icp_base; g_hfi->hfi_state = HFI_READY; g_hfi->cmd_q_state = true; g_hfi->msg_q_state = true; cam_io_w_mb((uint32_t)(INTR_ENABLE|INTR_ENABLE_WD0), icp_base + HFI_REG_A5_CSR_A2HOSTINTEN); mutex_unlock(&hfi_cmd_q_mutex); mutex_unlock(&hfi_msg_q_mutex); return rc; regions_fail: kfree(g_hfi); g_hfi = NULL; alloc_fail: mutex_unlock(&hfi_cmd_q_mutex); mutex_unlock(&hfi_msg_q_mutex); return rc; } void cam_hfi_deinit(void __iomem *icp_base) { mutex_lock(&hfi_cmd_q_mutex); mutex_lock(&hfi_msg_q_mutex); if (!g_hfi) { CAM_ERR(CAM_HFI, "hfi path not established yet"); goto err; } g_hfi->cmd_q_state = false; g_hfi->msg_q_state = false; kzfree(g_hfi); g_hfi = NULL; err: mutex_unlock(&hfi_cmd_q_mutex); mutex_unlock(&hfi_msg_q_mutex); }