12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * Copyright 2018-2020 Broadcom.
- */
- #include <linux/delay.h>
- #include <linux/fs.h>
- #include <linux/hash.h>
- #include <linux/interrupt.h>
- #include <linux/list.h>
- #include <linux/module.h>
- #include <linux/poll.h>
- #include <linux/sizes.h>
- #include <linux/spinlock.h>
- #include <linux/timer.h>
- #include "bcm_vk.h"
- #include "bcm_vk_msg.h"
- #include "bcm_vk_sg.h"
- /* functions to manipulate the transport id in msg block */
- #define BCM_VK_MSG_Q_SHIFT 4
- #define BCM_VK_MSG_Q_MASK 0xF
- #define BCM_VK_MSG_ID_MASK 0xFFF
- #define BCM_VK_DMA_DRAIN_MAX_MS 2000
- /* number x q_size will be the max number of msg processed per loop */
- #define BCM_VK_MSG_PROC_MAX_LOOP 2
- /* module parameter */
- static bool hb_mon = true;
- module_param(hb_mon, bool, 0444);
- MODULE_PARM_DESC(hb_mon, "Monitoring heartbeat continuously.\n");
- static int batch_log = 1;
- module_param(batch_log, int, 0444);
- MODULE_PARM_DESC(batch_log, "Max num of logs per batch operation.\n");
- static bool hb_mon_is_on(void)
- {
- return hb_mon;
- }
- static u32 get_q_num(const struct vk_msg_blk *msg)
- {
- u32 q_num = msg->trans_id & BCM_VK_MSG_Q_MASK;
- if (q_num >= VK_MSGQ_PER_CHAN_MAX)
- q_num = VK_MSGQ_NUM_DEFAULT;
- return q_num;
- }
- static void set_q_num(struct vk_msg_blk *msg, u32 q_num)
- {
- u32 trans_q;
- if (q_num >= VK_MSGQ_PER_CHAN_MAX)
- trans_q = VK_MSGQ_NUM_DEFAULT;
- else
- trans_q = q_num;
- msg->trans_id = (msg->trans_id & ~BCM_VK_MSG_Q_MASK) | trans_q;
- }
- static u32 get_msg_id(const struct vk_msg_blk *msg)
- {
- return ((msg->trans_id >> BCM_VK_MSG_Q_SHIFT) & BCM_VK_MSG_ID_MASK);
- }
- static void set_msg_id(struct vk_msg_blk *msg, u32 val)
- {
- msg->trans_id = (val << BCM_VK_MSG_Q_SHIFT) | get_q_num(msg);
- }
- static u32 msgq_inc(const struct bcm_vk_sync_qinfo *qinfo, u32 idx, u32 inc)
- {
- return ((idx + inc) & qinfo->q_mask);
- }
- static
- struct vk_msg_blk __iomem *msgq_blk_addr(const struct bcm_vk_sync_qinfo *qinfo,
- u32 idx)
- {
- return qinfo->q_start + (VK_MSGQ_BLK_SIZE * idx);
- }
- static u32 msgq_occupied(const struct bcm_vk_msgq __iomem *msgq,
- const struct bcm_vk_sync_qinfo *qinfo)
- {
- u32 wr_idx, rd_idx;
- wr_idx = readl_relaxed(&msgq->wr_idx);
- rd_idx = readl_relaxed(&msgq->rd_idx);
- return ((wr_idx - rd_idx) & qinfo->q_mask);
- }
- static
- u32 msgq_avail_space(const struct bcm_vk_msgq __iomem *msgq,
- const struct bcm_vk_sync_qinfo *qinfo)
- {
- return (qinfo->q_size - msgq_occupied(msgq, qinfo) - 1);
- }
- /* number of retries when enqueue message fails before returning EAGAIN */
- #define BCM_VK_H2VK_ENQ_RETRY 10
- #define BCM_VK_H2VK_ENQ_RETRY_DELAY_MS 50
- bool bcm_vk_drv_access_ok(struct bcm_vk *vk)
- {
- return (!!atomic_read(&vk->msgq_inited));
- }
- void bcm_vk_set_host_alert(struct bcm_vk *vk, u32 bit_mask)
- {
- struct bcm_vk_alert *alert = &vk->host_alert;
- unsigned long flags;
- /* use irqsave version as this maybe called inside timer interrupt */
- spin_lock_irqsave(&vk->host_alert_lock, flags);
- alert->notfs |= bit_mask;
- spin_unlock_irqrestore(&vk->host_alert_lock, flags);
- if (test_and_set_bit(BCM_VK_WQ_NOTF_PEND, vk->wq_offload) == 0)
- queue_work(vk->wq_thread, &vk->wq_work);
- }
- /*
- * Heartbeat related defines
- * The heartbeat from host is a last resort. If stuck condition happens
- * on the card, firmware is supposed to detect it. Therefore, the heartbeat
- * values used will be more relaxed on the driver, which need to be bigger
- * than the watchdog timeout on the card. The watchdog timeout on the card
- * is 20s, with a jitter of 2s => 22s. We use a value of 27s here.
- */
- #define BCM_VK_HB_TIMER_S 3
- #define BCM_VK_HB_TIMER_VALUE (BCM_VK_HB_TIMER_S * HZ)
- #define BCM_VK_HB_LOST_MAX (27 / BCM_VK_HB_TIMER_S)
- static void bcm_vk_hb_poll(struct timer_list *t)
- {
- u32 uptime_s;
- struct bcm_vk_hb_ctrl *hb = container_of(t, struct bcm_vk_hb_ctrl,
- timer);
- struct bcm_vk *vk = container_of(hb, struct bcm_vk, hb_ctrl);
- if (bcm_vk_drv_access_ok(vk) && hb_mon_is_on()) {
- /* read uptime from register and compare */
- uptime_s = vkread32(vk, BAR_0, BAR_OS_UPTIME);
- if (uptime_s == hb->last_uptime)
- hb->lost_cnt++;
- else /* reset to avoid accumulation */
- hb->lost_cnt = 0;
- dev_dbg(&vk->pdev->dev, "Last uptime %d current %d, lost %d\n",
- hb->last_uptime, uptime_s, hb->lost_cnt);
- /*
- * if the interface goes down without any activity, a value
- * of 0xFFFFFFFF will be continuously read, and the detection
- * will be happened eventually.
- */
- hb->last_uptime = uptime_s;
- } else {
- /* reset heart beat lost cnt */
- hb->lost_cnt = 0;
- }
- /* next, check if heartbeat exceeds limit */
- if (hb->lost_cnt > BCM_VK_HB_LOST_MAX) {
- dev_err(&vk->pdev->dev, "Heartbeat Misses %d times, %d s!\n",
- BCM_VK_HB_LOST_MAX,
- BCM_VK_HB_LOST_MAX * BCM_VK_HB_TIMER_S);
- bcm_vk_blk_drv_access(vk);
- bcm_vk_set_host_alert(vk, ERR_LOG_HOST_HB_FAIL);
- }
- /* re-arm timer */
- mod_timer(&hb->timer, jiffies + BCM_VK_HB_TIMER_VALUE);
- }
- void bcm_vk_hb_init(struct bcm_vk *vk)
- {
- struct bcm_vk_hb_ctrl *hb = &vk->hb_ctrl;
- timer_setup(&hb->timer, bcm_vk_hb_poll, 0);
- mod_timer(&hb->timer, jiffies + BCM_VK_HB_TIMER_VALUE);
- }
- void bcm_vk_hb_deinit(struct bcm_vk *vk)
- {
- struct bcm_vk_hb_ctrl *hb = &vk->hb_ctrl;
- del_timer(&hb->timer);
- }
- static void bcm_vk_msgid_bitmap_clear(struct bcm_vk *vk,
- unsigned int start,
- unsigned int nbits)
- {
- spin_lock(&vk->msg_id_lock);
- bitmap_clear(vk->bmap, start, nbits);
- spin_unlock(&vk->msg_id_lock);
- }
- /*
- * allocate a ctx per file struct
- */
- static struct bcm_vk_ctx *bcm_vk_get_ctx(struct bcm_vk *vk, const pid_t pid)
- {
- u32 i;
- struct bcm_vk_ctx *ctx = NULL;
- u32 hash_idx = hash_32(pid, VK_PID_HT_SHIFT_BIT);
- spin_lock(&vk->ctx_lock);
- /* check if it is in reset, if so, don't allow */
- if (vk->reset_pid) {
- dev_err(&vk->pdev->dev,
- "No context allowed during reset by pid %d\n",
- vk->reset_pid);
- goto in_reset_exit;
- }
- for (i = 0; i < ARRAY_SIZE(vk->ctx); i++) {
- if (!vk->ctx[i].in_use) {
- vk->ctx[i].in_use = true;
- ctx = &vk->ctx[i];
- break;
- }
- }
- if (!ctx) {
- dev_err(&vk->pdev->dev, "All context in use\n");
- goto all_in_use_exit;
- }
- /* set the pid and insert it to hash table */
- ctx->pid = pid;
- ctx->hash_idx = hash_idx;
- list_add_tail(&ctx->node, &vk->pid_ht[hash_idx].head);
- /* increase kref */
- kref_get(&vk->kref);
- /* clear counter */
- atomic_set(&ctx->pend_cnt, 0);
- atomic_set(&ctx->dma_cnt, 0);
- init_waitqueue_head(&ctx->rd_wq);
- all_in_use_exit:
- in_reset_exit:
- spin_unlock(&vk->ctx_lock);
- return ctx;
- }
- static u16 bcm_vk_get_msg_id(struct bcm_vk *vk)
- {
- u16 rc = VK_MSG_ID_OVERFLOW;
- u16 test_bit_count = 0;
- spin_lock(&vk->msg_id_lock);
- while (test_bit_count < (VK_MSG_ID_BITMAP_SIZE - 1)) {
- /*
- * first time come in this loop, msg_id will be 0
- * and the first one tested will be 1. We skip
- * VK_SIMPLEX_MSG_ID (0) for one way host2vk
- * communication
- */
- vk->msg_id++;
- if (vk->msg_id == VK_MSG_ID_BITMAP_SIZE)
- vk->msg_id = 1;
- if (test_bit(vk->msg_id, vk->bmap)) {
- test_bit_count++;
- continue;
- }
- rc = vk->msg_id;
- bitmap_set(vk->bmap, vk->msg_id, 1);
- break;
- }
- spin_unlock(&vk->msg_id_lock);
- return rc;
- }
- static int bcm_vk_free_ctx(struct bcm_vk *vk, struct bcm_vk_ctx *ctx)
- {
- u32 idx;
- u32 hash_idx;
- pid_t pid;
- struct bcm_vk_ctx *entry;
- int count = 0;
- if (!ctx) {
- dev_err(&vk->pdev->dev, "NULL context detected\n");
- return -EINVAL;
- }
- idx = ctx->idx;
- pid = ctx->pid;
- spin_lock(&vk->ctx_lock);
- if (!vk->ctx[idx].in_use) {
- dev_err(&vk->pdev->dev, "context[%d] not in use!\n", idx);
- } else {
- vk->ctx[idx].in_use = false;
- vk->ctx[idx].miscdev = NULL;
- /* Remove it from hash list and see if it is the last one. */
- list_del(&ctx->node);
- hash_idx = ctx->hash_idx;
- list_for_each_entry(entry, &vk->pid_ht[hash_idx].head, node) {
- if (entry->pid == pid)
- count++;
- }
- }
- spin_unlock(&vk->ctx_lock);
- return count;
- }
- static void bcm_vk_free_wkent(struct device *dev, struct bcm_vk_wkent *entry)
- {
- int proc_cnt;
- bcm_vk_sg_free(dev, entry->dma, VK_DMA_MAX_ADDRS, &proc_cnt);
- if (proc_cnt)
- atomic_dec(&entry->ctx->dma_cnt);
- kfree(entry->to_h_msg);
- kfree(entry);
- }
- static void bcm_vk_drain_all_pend(struct device *dev,
- struct bcm_vk_msg_chan *chan,
- struct bcm_vk_ctx *ctx)
- {
- u32 num;
- struct bcm_vk_wkent *entry, *tmp;
- struct bcm_vk *vk;
- struct list_head del_q;
- if (ctx)
- vk = container_of(ctx->miscdev, struct bcm_vk, miscdev);
- INIT_LIST_HEAD(&del_q);
- spin_lock(&chan->pendq_lock);
- for (num = 0; num < chan->q_nr; num++) {
- list_for_each_entry_safe(entry, tmp, &chan->pendq[num], node) {
- if ((!ctx) || (entry->ctx->idx == ctx->idx)) {
- list_move_tail(&entry->node, &del_q);
- }
- }
- }
- spin_unlock(&chan->pendq_lock);
- /* batch clean up */
- num = 0;
- list_for_each_entry_safe(entry, tmp, &del_q, node) {
- list_del(&entry->node);
- num++;
- if (ctx) {
- struct vk_msg_blk *msg;
- int bit_set;
- bool responded;
- u32 msg_id;
- /* if it is specific ctx, log for any stuck */
- msg = entry->to_v_msg;
- msg_id = get_msg_id(msg);
- bit_set = test_bit(msg_id, vk->bmap);
- responded = entry->to_h_msg ? true : false;
- if (num <= batch_log)
- dev_info(dev,
- "Drained: fid %u size %u msg 0x%x(seq-%x) ctx 0x%x[fd-%d] args:[0x%x 0x%x] resp %s, bmap %d\n",
- msg->function_id, msg->size,
- msg_id, entry->seq_num,
- msg->context_id, entry->ctx->idx,
- msg->cmd, msg->arg,
- responded ? "T" : "F", bit_set);
- if (responded)
- atomic_dec(&ctx->pend_cnt);
- else if (bit_set)
- bcm_vk_msgid_bitmap_clear(vk, msg_id, 1);
- }
- bcm_vk_free_wkent(dev, entry);
- }
- if (num && ctx)
- dev_info(dev, "Total drained items %d [fd-%d]\n",
- num, ctx->idx);
- }
- void bcm_vk_drain_msg_on_reset(struct bcm_vk *vk)
- {
- bcm_vk_drain_all_pend(&vk->pdev->dev, &vk->to_v_msg_chan, NULL);
- bcm_vk_drain_all_pend(&vk->pdev->dev, &vk->to_h_msg_chan, NULL);
- }
- /*
- * Function to sync up the messages queue info that is provided by BAR1
- */
- int bcm_vk_sync_msgq(struct bcm_vk *vk, bool force_sync)
- {
- struct bcm_vk_msgq __iomem *msgq;
- struct device *dev = &vk->pdev->dev;
- u32 msgq_off;
- u32 num_q;
- struct bcm_vk_msg_chan *chan_list[] = {&vk->to_v_msg_chan,
- &vk->to_h_msg_chan};
- struct bcm_vk_msg_chan *chan;
- int i, j;
- int ret = 0;
- /*
- * If the driver is loaded at startup where vk OS is not up yet,
- * the msgq-info may not be available until a later time. In
- * this case, we skip and the sync function is supposed to be
- * called again.
- */
- if (!bcm_vk_msgq_marker_valid(vk)) {
- dev_info(dev, "BAR1 msgq marker not initialized.\n");
- return -EAGAIN;
- }
- msgq_off = vkread32(vk, BAR_1, VK_BAR1_MSGQ_CTRL_OFF);
- /* each side is always half the total */
- num_q = vkread32(vk, BAR_1, VK_BAR1_MSGQ_NR) / 2;
- if (!num_q || (num_q > VK_MSGQ_PER_CHAN_MAX)) {
- dev_err(dev,
- "Advertised msgq %d error - max %d allowed\n",
- num_q, VK_MSGQ_PER_CHAN_MAX);
- return -EINVAL;
- }
- vk->to_v_msg_chan.q_nr = num_q;
- vk->to_h_msg_chan.q_nr = num_q;
- /* first msgq location */
- msgq = vk->bar[BAR_1] + msgq_off;
- /*
- * if this function is called when it is already inited,
- * something is wrong
- */
- if (bcm_vk_drv_access_ok(vk) && !force_sync) {
- dev_err(dev, "Msgq info already in sync\n");
- return -EPERM;
- }
- for (i = 0; i < ARRAY_SIZE(chan_list); i++) {
- chan = chan_list[i];
- memset(chan->sync_qinfo, 0, sizeof(chan->sync_qinfo));
- for (j = 0; j < num_q; j++) {
- struct bcm_vk_sync_qinfo *qinfo;
- u32 msgq_start;
- u32 msgq_size;
- u32 msgq_nxt;
- u32 msgq_db_offset, q_db_offset;
- chan->msgq[j] = msgq;
- msgq_start = readl_relaxed(&msgq->start);
- msgq_size = readl_relaxed(&msgq->size);
- msgq_nxt = readl_relaxed(&msgq->nxt);
- msgq_db_offset = readl_relaxed(&msgq->db_offset);
- q_db_offset = (msgq_db_offset & ((1 << DB_SHIFT) - 1));
- if (q_db_offset == (~msgq_db_offset >> DB_SHIFT))
- msgq_db_offset = q_db_offset;
- else
- /* fall back to default */
- msgq_db_offset = VK_BAR0_Q_DB_BASE(j);
- dev_info(dev,
- "MsgQ[%d] type %d num %d, @ 0x%x, db_offset 0x%x rd_idx %d wr_idx %d, size %d, nxt 0x%x\n",
- j,
- readw_relaxed(&msgq->type),
- readw_relaxed(&msgq->num),
- msgq_start,
- msgq_db_offset,
- readl_relaxed(&msgq->rd_idx),
- readl_relaxed(&msgq->wr_idx),
- msgq_size,
- msgq_nxt);
- qinfo = &chan->sync_qinfo[j];
- /* formulate and record static info */
- qinfo->q_start = vk->bar[BAR_1] + msgq_start;
- qinfo->q_size = msgq_size;
- /* set low threshold as 50% or 1/2 */
- qinfo->q_low = qinfo->q_size >> 1;
- qinfo->q_mask = qinfo->q_size - 1;
- qinfo->q_db_offset = msgq_db_offset;
- msgq++;
- }
- }
- atomic_set(&vk->msgq_inited, 1);
- return ret;
- }
- static int bcm_vk_msg_chan_init(struct bcm_vk_msg_chan *chan)
- {
- u32 i;
- mutex_init(&chan->msgq_mutex);
- spin_lock_init(&chan->pendq_lock);
- for (i = 0; i < VK_MSGQ_MAX_NR; i++)
- INIT_LIST_HEAD(&chan->pendq[i]);
- return 0;
- }
- static void bcm_vk_append_pendq(struct bcm_vk_msg_chan *chan, u16 q_num,
- struct bcm_vk_wkent *entry)
- {
- struct bcm_vk_ctx *ctx;
- spin_lock(&chan->pendq_lock);
- list_add_tail(&entry->node, &chan->pendq[q_num]);
- if (entry->to_h_msg) {
- ctx = entry->ctx;
- atomic_inc(&ctx->pend_cnt);
- wake_up_interruptible(&ctx->rd_wq);
- }
- spin_unlock(&chan->pendq_lock);
- }
- static u32 bcm_vk_append_ib_sgl(struct bcm_vk *vk,
- struct bcm_vk_wkent *entry,
- struct _vk_data *data,
- unsigned int num_planes)
- {
- unsigned int i;
- unsigned int item_cnt = 0;
- struct device *dev = &vk->pdev->dev;
- struct bcm_vk_msg_chan *chan = &vk->to_v_msg_chan;
- struct vk_msg_blk *msg = &entry->to_v_msg[0];
- struct bcm_vk_msgq __iomem *msgq;
- struct bcm_vk_sync_qinfo *qinfo;
- u32 ib_sgl_size = 0;
- u8 *buf = (u8 *)&entry->to_v_msg[entry->to_v_blks];
- u32 avail;
- u32 q_num;
- /* check if high watermark is hit, and if so, skip */
- q_num = get_q_num(msg);
- msgq = chan->msgq[q_num];
- qinfo = &chan->sync_qinfo[q_num];
- avail = msgq_avail_space(msgq, qinfo);
- if (avail < qinfo->q_low) {
- dev_dbg(dev, "Skip inserting inband SGL, [0x%x/0x%x]\n",
- avail, qinfo->q_size);
- return 0;
- }
- for (i = 0; i < num_planes; i++) {
- if (data[i].address &&
- (ib_sgl_size + data[i].size) <= vk->ib_sgl_size) {
- item_cnt++;
- memcpy(buf, entry->dma[i].sglist, data[i].size);
- ib_sgl_size += data[i].size;
- buf += data[i].size;
- }
- }
- dev_dbg(dev, "Num %u sgl items appended, size 0x%x, room 0x%x\n",
- item_cnt, ib_sgl_size, vk->ib_sgl_size);
- /* round up size */
- ib_sgl_size = (ib_sgl_size + VK_MSGQ_BLK_SIZE - 1)
- >> VK_MSGQ_BLK_SZ_SHIFT;
- return ib_sgl_size;
- }
- void bcm_to_v_q_doorbell(struct bcm_vk *vk, u32 q_num, u32 db_val)
- {
- struct bcm_vk_msg_chan *chan = &vk->to_v_msg_chan;
- struct bcm_vk_sync_qinfo *qinfo = &chan->sync_qinfo[q_num];
- vkwrite32(vk, db_val, BAR_0, qinfo->q_db_offset);
- }
- static int bcm_to_v_msg_enqueue(struct bcm_vk *vk, struct bcm_vk_wkent *entry)
- {
- static u32 seq_num;
- struct bcm_vk_msg_chan *chan = &vk->to_v_msg_chan;
- struct device *dev = &vk->pdev->dev;
- struct vk_msg_blk *src = &entry->to_v_msg[0];
- struct vk_msg_blk __iomem *dst;
- struct bcm_vk_msgq __iomem *msgq;
- struct bcm_vk_sync_qinfo *qinfo;
- u32 q_num = get_q_num(src);
- u32 wr_idx; /* local copy */
- u32 i;
- u32 avail;
- u32 retry;
- if (entry->to_v_blks != src->size + 1) {
- dev_err(dev, "number of blks %d not matching %d MsgId[0x%x]: func %d ctx 0x%x\n",
- entry->to_v_blks,
- src->size + 1,
- get_msg_id(src),
- src->function_id,
- src->context_id);
- return -EMSGSIZE;
- }
- msgq = chan->msgq[q_num];
- qinfo = &chan->sync_qinfo[q_num];
- mutex_lock(&chan->msgq_mutex);
- avail = msgq_avail_space(msgq, qinfo);
- /* if not enough space, return EAGAIN and let app handles it */
- retry = 0;
- while ((avail < entry->to_v_blks) &&
- (retry++ < BCM_VK_H2VK_ENQ_RETRY)) {
- mutex_unlock(&chan->msgq_mutex);
- msleep(BCM_VK_H2VK_ENQ_RETRY_DELAY_MS);
- mutex_lock(&chan->msgq_mutex);
- avail = msgq_avail_space(msgq, qinfo);
- }
- if (retry > BCM_VK_H2VK_ENQ_RETRY) {
- mutex_unlock(&chan->msgq_mutex);
- return -EAGAIN;
- }
- /* at this point, mutex is taken and there is enough space */
- entry->seq_num = seq_num++; /* update debug seq number */
- wr_idx = readl_relaxed(&msgq->wr_idx);
- if (wr_idx >= qinfo->q_size) {
- dev_crit(dev, "Invalid wr_idx 0x%x => max 0x%x!",
- wr_idx, qinfo->q_size);
- bcm_vk_blk_drv_access(vk);
- bcm_vk_set_host_alert(vk, ERR_LOG_HOST_PCIE_DWN);
- goto idx_err;
- }
- dst = msgq_blk_addr(qinfo, wr_idx);
- for (i = 0; i < entry->to_v_blks; i++) {
- memcpy_toio(dst, src, sizeof(*dst));
- src++;
- wr_idx = msgq_inc(qinfo, wr_idx, 1);
- dst = msgq_blk_addr(qinfo, wr_idx);
- }
- /* flush the write pointer */
- writel(wr_idx, &msgq->wr_idx);
- /* log new info for debugging */
- dev_dbg(dev,
- "MsgQ[%d] [Rd Wr] = [%d %d] blks inserted %d - Q = [u-%d a-%d]/%d\n",
- readl_relaxed(&msgq->num),
- readl_relaxed(&msgq->rd_idx),
- wr_idx,
- entry->to_v_blks,
- msgq_occupied(msgq, qinfo),
- msgq_avail_space(msgq, qinfo),
- readl_relaxed(&msgq->size));
- /*
- * press door bell based on queue number. 1 is added to the wr_idx
- * to avoid the value of 0 appearing on the VK side to distinguish
- * from initial value.
- */
- bcm_to_v_q_doorbell(vk, q_num, wr_idx + 1);
- idx_err:
- mutex_unlock(&chan->msgq_mutex);
- return 0;
- }
- int bcm_vk_send_shutdown_msg(struct bcm_vk *vk, u32 shut_type,
- const pid_t pid, const u32 q_num)
- {
- int rc = 0;
- struct bcm_vk_wkent *entry;
- struct device *dev = &vk->pdev->dev;
- /*
- * check if the marker is still good. Sometimes, the PCIe interface may
- * have gone done, and if so and we ship down thing based on broken
- * values, kernel may panic.
- */
- if (!bcm_vk_msgq_marker_valid(vk)) {
- dev_info(dev, "PCIe comm chan - invalid marker (0x%x)!\n",
- vkread32(vk, BAR_1, VK_BAR1_MSGQ_DEF_RDY));
- return -EINVAL;
- }
- entry = kzalloc(struct_size(entry, to_v_msg, 1), GFP_KERNEL);
- if (!entry)
- return -ENOMEM;
- /* fill up necessary data */
- entry->to_v_msg[0].function_id = VK_FID_SHUTDOWN;
- set_q_num(&entry->to_v_msg[0], q_num);
- set_msg_id(&entry->to_v_msg[0], VK_SIMPLEX_MSG_ID);
- entry->to_v_blks = 1; /* always 1 block */
- entry->to_v_msg[0].cmd = shut_type;
- entry->to_v_msg[0].arg = pid;
- rc = bcm_to_v_msg_enqueue(vk, entry);
- if (rc)
- dev_err(dev,
- "Sending shutdown message to q %d for pid %d fails.\n",
- get_q_num(&entry->to_v_msg[0]), pid);
- kfree(entry);
- return rc;
- }
- static int bcm_vk_handle_last_sess(struct bcm_vk *vk, const pid_t pid,
- const u32 q_num)
- {
- int rc = 0;
- struct device *dev = &vk->pdev->dev;
- /*
- * don't send down or do anything if message queue is not initialized
- * and if it is the reset session, clear it.
- */
- if (!bcm_vk_drv_access_ok(vk)) {
- if (vk->reset_pid == pid)
- vk->reset_pid = 0;
- return -EPERM;
- }
- dev_dbg(dev, "No more sessions, shut down pid %d\n", pid);
- /* only need to do it if it is not the reset process */
- if (vk->reset_pid != pid)
- rc = bcm_vk_send_shutdown_msg(vk, VK_SHUTDOWN_PID, pid, q_num);
- else
- /* put reset_pid to 0 if it is exiting last session */
- vk->reset_pid = 0;
- return rc;
- }
- static struct bcm_vk_wkent *bcm_vk_dequeue_pending(struct bcm_vk *vk,
- struct bcm_vk_msg_chan *chan,
- u16 q_num,
- u16 msg_id)
- {
- struct bcm_vk_wkent *entry = NULL, *iter;
- spin_lock(&chan->pendq_lock);
- list_for_each_entry(iter, &chan->pendq[q_num], node) {
- if (get_msg_id(&iter->to_v_msg[0]) == msg_id) {
- list_del(&iter->node);
- entry = iter;
- bcm_vk_msgid_bitmap_clear(vk, msg_id, 1);
- break;
- }
- }
- spin_unlock(&chan->pendq_lock);
- return entry;
- }
- s32 bcm_to_h_msg_dequeue(struct bcm_vk *vk)
- {
- struct device *dev = &vk->pdev->dev;
- struct bcm_vk_msg_chan *chan = &vk->to_h_msg_chan;
- struct vk_msg_blk *data;
- struct vk_msg_blk __iomem *src;
- struct vk_msg_blk *dst;
- struct bcm_vk_msgq __iomem *msgq;
- struct bcm_vk_sync_qinfo *qinfo;
- struct bcm_vk_wkent *entry;
- u32 rd_idx, wr_idx;
- u32 q_num, msg_id, j;
- u32 num_blks;
- s32 total = 0;
- int cnt = 0;
- int msg_processed = 0;
- int max_msg_to_process;
- bool exit_loop;
- /*
- * drain all the messages from the queues, and find its pending
- * entry in the to_v queue, based on msg_id & q_num, and move the
- * entry to the to_h pending queue, waiting for user space
- * program to extract
- */
- mutex_lock(&chan->msgq_mutex);
- for (q_num = 0; q_num < chan->q_nr; q_num++) {
- msgq = chan->msgq[q_num];
- qinfo = &chan->sync_qinfo[q_num];
- max_msg_to_process = BCM_VK_MSG_PROC_MAX_LOOP * qinfo->q_size;
- rd_idx = readl_relaxed(&msgq->rd_idx);
- wr_idx = readl_relaxed(&msgq->wr_idx);
- msg_processed = 0;
- exit_loop = false;
- while ((rd_idx != wr_idx) && !exit_loop) {
- u8 src_size;
- /*
- * Make a local copy and get pointer to src blk
- * The rd_idx is masked before getting the pointer to
- * avoid out of bound access in case the interface goes
- * down. It will end up pointing to the last block in
- * the buffer, but subsequent src->size check would be
- * able to catch this.
- */
- src = msgq_blk_addr(qinfo, rd_idx & qinfo->q_mask);
- src_size = readb(&src->size);
- if ((rd_idx >= qinfo->q_size) ||
- (src_size > (qinfo->q_size - 1))) {
- dev_crit(dev,
- "Invalid rd_idx 0x%x or size 0x%x => max 0x%x!",
- rd_idx, src_size, qinfo->q_size);
- bcm_vk_blk_drv_access(vk);
- bcm_vk_set_host_alert(vk,
- ERR_LOG_HOST_PCIE_DWN);
- goto idx_err;
- }
- num_blks = src_size + 1;
- data = kzalloc(num_blks * VK_MSGQ_BLK_SIZE, GFP_KERNEL);
- if (data) {
- /* copy messages and linearize it */
- dst = data;
- for (j = 0; j < num_blks; j++) {
- memcpy_fromio(dst, src, sizeof(*dst));
- dst++;
- rd_idx = msgq_inc(qinfo, rd_idx, 1);
- src = msgq_blk_addr(qinfo, rd_idx);
- }
- total++;
- } else {
- /*
- * if we could not allocate memory in kernel,
- * that is fatal.
- */
- dev_crit(dev, "Kernel mem allocation failure.\n");
- total = -ENOMEM;
- goto idx_err;
- }
- /* flush rd pointer after a message is dequeued */
- writel(rd_idx, &msgq->rd_idx);
- /* log new info for debugging */
- dev_dbg(dev,
- "MsgQ[%d] [Rd Wr] = [%d %d] blks extracted %d - Q = [u-%d a-%d]/%d\n",
- readl_relaxed(&msgq->num),
- rd_idx,
- wr_idx,
- num_blks,
- msgq_occupied(msgq, qinfo),
- msgq_avail_space(msgq, qinfo),
- readl_relaxed(&msgq->size));
- /*
- * No need to search if it is an autonomous one-way
- * message from driver, as these messages do not bear
- * a to_v pending item. Currently, only the shutdown
- * message falls into this category.
- */
- if (data->function_id == VK_FID_SHUTDOWN) {
- kfree(data);
- continue;
- }
- msg_id = get_msg_id(data);
- /* lookup original message in to_v direction */
- entry = bcm_vk_dequeue_pending(vk,
- &vk->to_v_msg_chan,
- q_num,
- msg_id);
- /*
- * if there is message to does not have prior send,
- * this is the location to add here
- */
- if (entry) {
- entry->to_h_blks = num_blks;
- entry->to_h_msg = data;
- bcm_vk_append_pendq(&vk->to_h_msg_chan,
- q_num, entry);
- } else {
- if (cnt++ < batch_log)
- dev_info(dev,
- "Could not find MsgId[0x%x] for resp func %d bmap %d\n",
- msg_id, data->function_id,
- test_bit(msg_id, vk->bmap));
- kfree(data);
- }
- /* Fetch wr_idx to handle more back-to-back events */
- wr_idx = readl(&msgq->wr_idx);
- /*
- * cap the max so that even we try to handle more back-to-back events,
- * so that it won't hold CPU too long or in case rd/wr idexes are
- * corrupted which triggers infinite looping.
- */
- if (++msg_processed >= max_msg_to_process) {
- dev_warn(dev, "Q[%d] Per loop processing exceeds %d\n",
- q_num, max_msg_to_process);
- exit_loop = true;
- }
- }
- }
- idx_err:
- mutex_unlock(&chan->msgq_mutex);
- dev_dbg(dev, "total %d drained from queues\n", total);
- return total;
- }
- /*
- * init routine for all required data structures
- */
- static int bcm_vk_data_init(struct bcm_vk *vk)
- {
- int i;
- spin_lock_init(&vk->ctx_lock);
- for (i = 0; i < ARRAY_SIZE(vk->ctx); i++) {
- vk->ctx[i].in_use = false;
- vk->ctx[i].idx = i; /* self identity */
- vk->ctx[i].miscdev = NULL;
- }
- spin_lock_init(&vk->msg_id_lock);
- spin_lock_init(&vk->host_alert_lock);
- vk->msg_id = 0;
- /* initialize hash table */
- for (i = 0; i < VK_PID_HT_SZ; i++)
- INIT_LIST_HEAD(&vk->pid_ht[i].head);
- return 0;
- }
- irqreturn_t bcm_vk_msgq_irqhandler(int irq, void *dev_id)
- {
- struct bcm_vk *vk = dev_id;
- if (!bcm_vk_drv_access_ok(vk)) {
- dev_err(&vk->pdev->dev,
- "Interrupt %d received when msgq not inited\n", irq);
- goto skip_schedule_work;
- }
- queue_work(vk->wq_thread, &vk->wq_work);
- skip_schedule_work:
- return IRQ_HANDLED;
- }
- int bcm_vk_open(struct inode *inode, struct file *p_file)
- {
- struct bcm_vk_ctx *ctx;
- struct miscdevice *miscdev = (struct miscdevice *)p_file->private_data;
- struct bcm_vk *vk = container_of(miscdev, struct bcm_vk, miscdev);
- struct device *dev = &vk->pdev->dev;
- int rc = 0;
- /* get a context and set it up for file */
- ctx = bcm_vk_get_ctx(vk, task_tgid_nr(current));
- if (!ctx) {
- dev_err(dev, "Error allocating context\n");
- rc = -ENOMEM;
- } else {
- /*
- * set up context and replace private data with context for
- * other methods to use. Reason for the context is because
- * it is allowed for multiple sessions to open the sysfs, and
- * for each file open, when upper layer query the response,
- * only those that are tied to a specific open should be
- * returned. The context->idx will be used for such binding
- */
- ctx->miscdev = miscdev;
- p_file->private_data = ctx;
- dev_dbg(dev, "ctx_returned with idx %d, pid %d\n",
- ctx->idx, ctx->pid);
- }
- return rc;
- }
- ssize_t bcm_vk_read(struct file *p_file,
- char __user *buf,
- size_t count,
- loff_t *f_pos)
- {
- ssize_t rc = -ENOMSG;
- struct bcm_vk_ctx *ctx = p_file->private_data;
- struct bcm_vk *vk = container_of(ctx->miscdev, struct bcm_vk,
- miscdev);
- struct device *dev = &vk->pdev->dev;
- struct bcm_vk_msg_chan *chan = &vk->to_h_msg_chan;
- struct bcm_vk_wkent *entry = NULL, *iter;
- u32 q_num;
- u32 rsp_length;
- if (!bcm_vk_drv_access_ok(vk))
- return -EPERM;
- dev_dbg(dev, "Buf count %zu\n", count);
- /*
- * search through the pendq on the to_h chan, and return only those
- * that belongs to the same context. Search is always from the high to
- * the low priority queues
- */
- spin_lock(&chan->pendq_lock);
- for (q_num = 0; q_num < chan->q_nr; q_num++) {
- list_for_each_entry(iter, &chan->pendq[q_num], node) {
- if (iter->ctx->idx == ctx->idx) {
- if (count >=
- (iter->to_h_blks * VK_MSGQ_BLK_SIZE)) {
- list_del(&iter->node);
- atomic_dec(&ctx->pend_cnt);
- entry = iter;
- } else {
- /* buffer not big enough */
- rc = -EMSGSIZE;
- }
- goto read_loop_exit;
- }
- }
- }
- read_loop_exit:
- spin_unlock(&chan->pendq_lock);
- if (entry) {
- /* retrieve the passed down msg_id */
- set_msg_id(&entry->to_h_msg[0], entry->usr_msg_id);
- rsp_length = entry->to_h_blks * VK_MSGQ_BLK_SIZE;
- if (copy_to_user(buf, entry->to_h_msg, rsp_length) == 0)
- rc = rsp_length;
- bcm_vk_free_wkent(dev, entry);
- } else if (rc == -EMSGSIZE) {
- struct vk_msg_blk tmp_msg = entry->to_h_msg[0];
- /*
- * in this case, return just the first block, so
- * that app knows what size it is looking for.
- */
- set_msg_id(&tmp_msg, entry->usr_msg_id);
- tmp_msg.size = entry->to_h_blks - 1;
- if (copy_to_user(buf, &tmp_msg, VK_MSGQ_BLK_SIZE) != 0) {
- dev_err(dev, "Error return 1st block in -EMSGSIZE\n");
- rc = -EFAULT;
- }
- }
- return rc;
- }
- ssize_t bcm_vk_write(struct file *p_file,
- const char __user *buf,
- size_t count,
- loff_t *f_pos)
- {
- ssize_t rc;
- struct bcm_vk_ctx *ctx = p_file->private_data;
- struct bcm_vk *vk = container_of(ctx->miscdev, struct bcm_vk,
- miscdev);
- struct bcm_vk_msgq __iomem *msgq;
- struct device *dev = &vk->pdev->dev;
- struct bcm_vk_wkent *entry;
- u32 sgl_extra_blks;
- u32 q_num;
- u32 msg_size;
- u32 msgq_size;
- if (!bcm_vk_drv_access_ok(vk))
- return -EPERM;
- dev_dbg(dev, "Msg count %zu\n", count);
- /* first, do sanity check where count should be multiple of basic blk */
- if (count & (VK_MSGQ_BLK_SIZE - 1)) {
- dev_err(dev, "Failure with size %zu not multiple of %zu\n",
- count, VK_MSGQ_BLK_SIZE);
- rc = -EINVAL;
- goto write_err;
- }
- /* allocate the work entry + buffer for size count and inband sgl */
- entry = kzalloc(sizeof(*entry) + count + vk->ib_sgl_size,
- GFP_KERNEL);
- if (!entry) {
- rc = -ENOMEM;
- goto write_err;
- }
- /* now copy msg from user space, and then formulate the work entry */
- if (copy_from_user(&entry->to_v_msg[0], buf, count)) {
- rc = -EFAULT;
- goto write_free_ent;
- }
- entry->to_v_blks = count >> VK_MSGQ_BLK_SZ_SHIFT;
- entry->ctx = ctx;
- /* do a check on the blk size which could not exceed queue space */
- q_num = get_q_num(&entry->to_v_msg[0]);
- msgq = vk->to_v_msg_chan.msgq[q_num];
- msgq_size = readl_relaxed(&msgq->size);
- if (entry->to_v_blks + (vk->ib_sgl_size >> VK_MSGQ_BLK_SZ_SHIFT)
- > (msgq_size - 1)) {
- dev_err(dev, "Blk size %d exceed max queue size allowed %d\n",
- entry->to_v_blks, msgq_size - 1);
- rc = -EINVAL;
- goto write_free_ent;
- }
- /* Use internal message id */
- entry->usr_msg_id = get_msg_id(&entry->to_v_msg[0]);
- rc = bcm_vk_get_msg_id(vk);
- if (rc == VK_MSG_ID_OVERFLOW) {
- dev_err(dev, "msg_id overflow\n");
- rc = -EOVERFLOW;
- goto write_free_ent;
- }
- set_msg_id(&entry->to_v_msg[0], rc);
- ctx->q_num = q_num;
- dev_dbg(dev,
- "[Q-%d]Message ctx id %d, usr_msg_id 0x%x sent msg_id 0x%x\n",
- ctx->q_num, ctx->idx, entry->usr_msg_id,
- get_msg_id(&entry->to_v_msg[0]));
- if (entry->to_v_msg[0].function_id == VK_FID_TRANS_BUF) {
- /* Convert any pointers to sg list */
- unsigned int num_planes;
- int dir;
- struct _vk_data *data;
- /*
- * check if we are in reset, if so, no buffer transfer is
- * allowed and return error.
- */
- if (vk->reset_pid) {
- dev_dbg(dev, "No Transfer allowed during reset, pid %d.\n",
- ctx->pid);
- rc = -EACCES;
- goto write_free_msgid;
- }
- num_planes = entry->to_v_msg[0].cmd & VK_CMD_PLANES_MASK;
- if ((entry->to_v_msg[0].cmd & VK_CMD_MASK) == VK_CMD_DOWNLOAD)
- dir = DMA_FROM_DEVICE;
- else
- dir = DMA_TO_DEVICE;
- /* Calculate vk_data location */
- /* Go to end of the message */
- msg_size = entry->to_v_msg[0].size;
- if (msg_size > entry->to_v_blks) {
- rc = -EMSGSIZE;
- goto write_free_msgid;
- }
- data = (struct _vk_data *)&entry->to_v_msg[msg_size + 1];
- /* Now back up to the start of the pointers */
- data -= num_planes;
- /* Convert user addresses to DMA SG List */
- rc = bcm_vk_sg_alloc(dev, entry->dma, dir, data, num_planes);
- if (rc)
- goto write_free_msgid;
- atomic_inc(&ctx->dma_cnt);
- /* try to embed inband sgl */
- sgl_extra_blks = bcm_vk_append_ib_sgl(vk, entry, data,
- num_planes);
- entry->to_v_blks += sgl_extra_blks;
- entry->to_v_msg[0].size += sgl_extra_blks;
- } else if (entry->to_v_msg[0].function_id == VK_FID_INIT &&
- entry->to_v_msg[0].context_id == VK_NEW_CTX) {
- /*
- * Init happens in 2 stages, only the first stage contains the
- * pid that needs translating.
- */
- pid_t org_pid, pid;
- /*
- * translate the pid into the unique host space as user
- * may run sessions inside containers or process
- * namespaces.
- */
- #define VK_MSG_PID_MASK 0xffffff00
- #define VK_MSG_PID_SH 8
- org_pid = (entry->to_v_msg[0].arg & VK_MSG_PID_MASK)
- >> VK_MSG_PID_SH;
- pid = task_tgid_nr(current);
- entry->to_v_msg[0].arg =
- (entry->to_v_msg[0].arg & ~VK_MSG_PID_MASK) |
- (pid << VK_MSG_PID_SH);
- if (org_pid != pid)
- dev_dbg(dev, "In PID 0x%x(%d), converted PID 0x%x(%d)\n",
- org_pid, org_pid, pid, pid);
- }
- /*
- * store work entry to pending queue until a response is received.
- * This needs to be done before enqueuing the message
- */
- bcm_vk_append_pendq(&vk->to_v_msg_chan, q_num, entry);
- rc = bcm_to_v_msg_enqueue(vk, entry);
- if (rc) {
- dev_err(dev, "Fail to enqueue msg to to_v queue\n");
- /* remove message from pending list */
- entry = bcm_vk_dequeue_pending
- (vk,
- &vk->to_v_msg_chan,
- q_num,
- get_msg_id(&entry->to_v_msg[0]));
- goto write_free_ent;
- }
- return count;
- write_free_msgid:
- bcm_vk_msgid_bitmap_clear(vk, get_msg_id(&entry->to_v_msg[0]), 1);
- write_free_ent:
- kfree(entry);
- write_err:
- return rc;
- }
- __poll_t bcm_vk_poll(struct file *p_file, struct poll_table_struct *wait)
- {
- __poll_t ret = 0;
- int cnt;
- struct bcm_vk_ctx *ctx = p_file->private_data;
- struct bcm_vk *vk = container_of(ctx->miscdev, struct bcm_vk, miscdev);
- struct device *dev = &vk->pdev->dev;
- poll_wait(p_file, &ctx->rd_wq, wait);
- cnt = atomic_read(&ctx->pend_cnt);
- if (cnt) {
- ret = (__force __poll_t)(POLLIN | POLLRDNORM);
- if (cnt < 0) {
- dev_err(dev, "Error cnt %d, setting back to 0", cnt);
- atomic_set(&ctx->pend_cnt, 0);
- }
- }
- return ret;
- }
- int bcm_vk_release(struct inode *inode, struct file *p_file)
- {
- int ret;
- struct bcm_vk_ctx *ctx = p_file->private_data;
- struct bcm_vk *vk = container_of(ctx->miscdev, struct bcm_vk, miscdev);
- struct device *dev = &vk->pdev->dev;
- pid_t pid = ctx->pid;
- int dma_cnt;
- unsigned long timeout, start_time;
- /*
- * if there are outstanding DMA transactions, need to delay long enough
- * to ensure that the card side would have stopped touching the host buffer
- * and its SGL list. A race condition could happen if the host app is killed
- * abruptly, eg kill -9, while some DMA transfer orders are still inflight.
- * Nothing could be done except for a delay as host side is running in a
- * completely async fashion.
- */
- start_time = jiffies;
- timeout = start_time + msecs_to_jiffies(BCM_VK_DMA_DRAIN_MAX_MS);
- do {
- if (time_after(jiffies, timeout)) {
- dev_warn(dev, "%d dma still pending for [fd-%d] pid %d\n",
- dma_cnt, ctx->idx, pid);
- break;
- }
- dma_cnt = atomic_read(&ctx->dma_cnt);
- cpu_relax();
- cond_resched();
- } while (dma_cnt);
- dev_dbg(dev, "Draining for [fd-%d] pid %d - delay %d ms\n",
- ctx->idx, pid, jiffies_to_msecs(jiffies - start_time));
- bcm_vk_drain_all_pend(&vk->pdev->dev, &vk->to_v_msg_chan, ctx);
- bcm_vk_drain_all_pend(&vk->pdev->dev, &vk->to_h_msg_chan, ctx);
- ret = bcm_vk_free_ctx(vk, ctx);
- if (ret == 0)
- ret = bcm_vk_handle_last_sess(vk, pid, ctx->q_num);
- else
- ret = 0;
- kref_put(&vk->kref, bcm_vk_release_data);
- return ret;
- }
- int bcm_vk_msg_init(struct bcm_vk *vk)
- {
- struct device *dev = &vk->pdev->dev;
- int ret;
- if (bcm_vk_data_init(vk)) {
- dev_err(dev, "Error initializing internal data structures\n");
- return -EINVAL;
- }
- if (bcm_vk_msg_chan_init(&vk->to_v_msg_chan) ||
- bcm_vk_msg_chan_init(&vk->to_h_msg_chan)) {
- dev_err(dev, "Error initializing communication channel\n");
- return -EIO;
- }
- /* read msgq info if ready */
- ret = bcm_vk_sync_msgq(vk, false);
- if (ret && (ret != -EAGAIN)) {
- dev_err(dev, "Error reading comm msg Q info\n");
- return -EIO;
- }
- return 0;
- }
- void bcm_vk_msg_remove(struct bcm_vk *vk)
- {
- bcm_vk_blk_drv_access(vk);
- /* drain all pending items */
- bcm_vk_drain_all_pend(&vk->pdev->dev, &vk->to_v_msg_chan, NULL);
- bcm_vk_drain_all_pend(&vk->pdev->dev, &vk->to_h_msg_chan, NULL);
- }
|