// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2021, The Linux Foundation. All rights reserved. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "hfi_packetization.h" #include "msm_cvp_debug.h" #include "cvp_core_hfi.h" #include "cvp_hfi_helper.h" #include "cvp_hfi_io.h" #include "msm_cvp_dsp.h" #include "msm_cvp_clocks.h" #include "cvp_dump.h" #ifdef CVP_MINIDUMP_ENABLED /*Declare and init the head node of the linked list for queue va_md dump*/ static LIST_HEAD(head_node_hfi_queue); /*Declare and init the head node of the linked list for debug struct va_md dump*/ static LIST_HEAD(head_node_dbg_struct); static int eva_struct_list_notif_handler(struct notifier_block *this, unsigned long event, void *ptr); static int eva_hfiq_list_notif_handler(struct notifier_block *this, unsigned long event, void *ptr); static struct notifier_block eva_struct_list_notif_blk = { .notifier_call = eva_struct_list_notif_handler, .priority = INT_MAX-1, }; static struct notifier_block eva_hfiq_list_notif_blk = { .notifier_call = eva_hfiq_list_notif_handler, .priority = INT_MAX, }; struct list_head *dump_array[CVP_MAX_DUMP] = { [CVP_QUEUE_DUMP] = &head_node_hfi_queue, [CVP_DBG_DUMP] = &head_node_dbg_struct, }; int md_eva_dump(const char* name, u64 virt, u64 phys, u64 size) { struct md_region md_entry; if (msm_minidump_enabled()) { dprintk(CVP_INFO, "Minidump is enabled!\n"); strlcpy(md_entry.name, name, sizeof(md_entry.name)); md_entry.virt_addr = (uintptr_t)virt; md_entry.phys_addr = phys; md_entry.size = size; if (msm_minidump_add_region(&md_entry) < 0) { dprintk(CVP_ERR, "Failed to add \"%s\" data in \ Minidump\n", name); return 1; } else { dprintk(CVP_INFO, "add region success for \"%s\" with virt addr:\ 0x%x, phy addr: 0x%x, size: %d", md_entry.name, md_entry.virt_addr, md_entry.phys_addr, md_entry.size); return 0; } } else { dprintk(CVP_ERR, "Minidump is NOT enabled!\n"); return 1; } } void cvp_va_md_register(char* name, void* notf_blk_ptr) { int rc = 0; struct notifier_block* notf_blk = (struct notifier_block*)notf_blk_ptr; rc = qcom_va_md_register(name, notf_blk); if (rc) { dprintk(CVP_ERR, "\"%s\" : qcom_va_md_register failed rc = %d\n", name, rc); } else { dprintk(CVP_INFO, "\"%s\" : eva_queue qcom_va_md_register \ success rc = %d\n", name, rc); } } void cvp_register_va_md_region() { if (qcom_va_md_enabled()) { cvp_va_md_register("eva_queues", &eva_hfiq_list_notif_blk); cvp_va_md_register("dbg_struct", &eva_struct_list_notif_blk); } else { dprintk(CVP_ERR, "VA_Minidump is NOT enabled!\n"); } } void cvp_free_va_md_list(void) { struct eva_va_md_queue *cursor, *temp; list_for_each_entry_safe(cursor, temp, &head_node_hfi_queue, list) { list_del(&cursor->list); kfree(cursor); } list_for_each_entry_safe(cursor, temp, &head_node_dbg_struct, list) { list_del(&cursor->list); kfree(cursor); } } void add_va_node_to_list(enum cvp_dump_type type, void *buff_va, u32 buff_size, const char *region_name, bool copy) { struct list_head *head_node; struct eva_va_md_queue *temp_node = NULL; if (type >= CVP_MAX_DUMP) return; head_node = dump_array[type]; /*Creating Node*/ temp_node = kzalloc(sizeof(struct eva_va_md_queue), GFP_KERNEL); if (!temp_node) { dprintk(CVP_ERR, "Memory allocation failed for list node\n"); return; } INIT_LIST_HEAD(&temp_node->list); temp_node->va_md_buff = buff_va; temp_node->va_md_buff_size = buff_size; strlcpy(temp_node->region_name, region_name, sizeof(temp_node->region_name)); temp_node->copy = copy; list_add_tail(&temp_node->list, head_node); dprintk(CVP_INFO, "\"%s\" added to buffer list, vaddr: %px size: 0x%x\n", temp_node->region_name, temp_node->va_md_buff, temp_node->va_md_buff_size); } void add_hfi_queue_to_va_md_list(void *device) { struct cvp_iface_q_info *iface_q; struct iris_hfi_device *dev; dev = (struct iris_hfi_device*)device; iface_q = &dev->iface_queues[CVP_IFACEQ_CMDQ_IDX]; add_va_node_to_list(CVP_QUEUE_DUMP, iface_q->q_array.align_virtual_addr, iface_q->q_array.mem_size, "eva_cmdq_cpu", false); iface_q = &dev->iface_queues[CVP_IFACEQ_MSGQ_IDX]; add_va_node_to_list(CVP_QUEUE_DUMP, iface_q->q_array.align_virtual_addr, iface_q->q_array.mem_size, "eva_msgq_cpu", false); iface_q = &dev->dsp_iface_queues[CVP_IFACEQ_CMDQ_IDX]; add_va_node_to_list(CVP_QUEUE_DUMP, iface_q->q_array.align_virtual_addr, iface_q->q_array.mem_size, "eva_cmdq_dsp", false); iface_q = &dev->dsp_iface_queues[CVP_IFACEQ_MSGQ_IDX]; add_va_node_to_list(CVP_QUEUE_DUMP, iface_q->q_array.align_virtual_addr, iface_q->q_array.mem_size, "eva_msgq_dsp", false); } void add_queue_header_to_va_md_list(void *device) { struct cvp_iface_q_info *iface_q; struct iris_hfi_device *dev; struct cvp_hfi_queue_header *queue; dev = (struct iris_hfi_device*)device; iface_q = &dev->iface_queues[CVP_IFACEQ_CMDQ_IDX]; queue = (struct cvp_hfi_queue_header *)iface_q->q_hdr; add_va_node_to_list(CVP_DBG_DUMP, queue, sizeof(struct cvp_hfi_queue_header), "cvp_hfi_queue_header-cpucmdQ", false); iface_q = &dev->iface_queues[CVP_IFACEQ_MSGQ_IDX]; queue = (struct cvp_hfi_queue_header *)iface_q->q_hdr; add_va_node_to_list(CVP_DBG_DUMP, queue, sizeof(struct cvp_hfi_queue_header), "cvp_hfi_queue_header-cpumsgQ", false); iface_q = &dev->dsp_iface_queues[CVP_IFACEQ_CMDQ_IDX]; queue = (struct cvp_hfi_queue_header *)iface_q->q_hdr; add_va_node_to_list(CVP_DBG_DUMP, queue, sizeof(struct cvp_hfi_queue_header), "cvp_hfi_queue_header-dspcmdQ", false); iface_q = &dev->dsp_iface_queues[CVP_IFACEQ_MSGQ_IDX]; queue = (struct cvp_hfi_queue_header *)iface_q->q_hdr; add_va_node_to_list(CVP_DBG_DUMP, queue, sizeof(struct cvp_hfi_queue_header), "cvp_hfi_queue_header-dspmsgQ", false); } static int eva_hfiq_list_notif_handler(struct notifier_block *this, unsigned long event, void *ptr) { struct va_md_entry entry; struct eva_va_md_queue *cursor, *temp; int rc = 0; void *temp_data; list_for_each_entry_safe(cursor, temp, &head_node_hfi_queue, list) { entry.vaddr = (unsigned long)cursor->va_md_buff; if (cursor->copy) { dprintk(CVP_INFO, "Copying \"%s\"(%d Bytes)\ to intermediate buffer\n", cursor->region_name, cursor->va_md_buff_size); temp_data = kzalloc(cursor->va_md_buff_size, GFP_KERNEL); if (temp_data) { memcpy(temp_data, cursor->va_md_buff, cursor->va_md_buff_size); entry.vaddr = (unsigned long)temp_data; } } entry.size = cursor->va_md_buff_size; strlcpy(entry.owner, cursor->region_name, sizeof(entry.owner)); entry.cb = NULL; if (msm_cvp_minidump_enable) { rc = qcom_va_md_add_region(&entry); if (rc) dprintk(CVP_ERR, "Add region \"failed\" for \ \"%s\", vaddr: %px size: 0x%x\n", entry.owner, cursor->va_md_buff, entry.size); else dprintk(CVP_INFO, "Add region \"success\" for \ \"%s\", vaddr: %px size: 0x%x\n", entry.owner, cursor->va_md_buff, entry.size); } } return NOTIFY_OK; } static int eva_struct_list_notif_handler(struct notifier_block *this, unsigned long event, void *ptr) { struct va_md_entry entry; struct eva_va_md_queue *cursor, *temp; int rc = 0; void *temp_data; list_for_each_entry_safe(cursor, temp, &head_node_dbg_struct, list) { entry.vaddr = (unsigned long)cursor->va_md_buff; if (cursor->copy) { dprintk(CVP_INFO, "Copying \"%s\"(%d Bytes) to \ intermediate buffer\n", cursor->region_name, cursor->va_md_buff_size); temp_data = kzalloc(cursor->va_md_buff_size, GFP_KERNEL); if (temp_data) { memcpy(temp_data, cursor->va_md_buff, cursor->va_md_buff_size); entry.vaddr = (unsigned long)temp_data; } } entry.size = cursor->va_md_buff_size; strlcpy(entry.owner, cursor->region_name, sizeof(entry.owner)); entry.cb = NULL; if (msm_cvp_minidump_enable) { rc = qcom_va_md_add_region(&entry); if (rc) dprintk(CVP_ERR, "Add region \"failed\" for \ \"%s\", vaddr: %px size: 0x%x\n", entry.owner, cursor->va_md_buff, entry.size); else dprintk(CVP_INFO, "Add region \"success\" for \ \"%s\", vaddr: %px size: 0x%x\n", entry.owner, cursor->va_md_buff, entry.size); } } return NOTIFY_OK; } #endif