123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * Driver for HiSilicon PCIe tune and trace device
- *
- * Copyright (c) 2022 HiSilicon Technologies Co., Ltd.
- * Author: Yicong Yang <[email protected]>
- */
- #include <linux/bitfield.h>
- #include <linux/bitops.h>
- #include <linux/cpuhotplug.h>
- #include <linux/delay.h>
- #include <linux/dma-mapping.h>
- #include <linux/interrupt.h>
- #include <linux/io.h>
- #include <linux/iommu.h>
- #include <linux/iopoll.h>
- #include <linux/module.h>
- #include <linux/sysfs.h>
- #include <linux/vmalloc.h>
- #include "hisi_ptt.h"
- /* Dynamic CPU hotplug state used by PTT */
- static enum cpuhp_state hisi_ptt_pmu_online;
- static bool hisi_ptt_wait_tuning_finish(struct hisi_ptt *hisi_ptt)
- {
- u32 val;
- return !readl_poll_timeout(hisi_ptt->iobase + HISI_PTT_TUNING_INT_STAT,
- val, !(val & HISI_PTT_TUNING_INT_STAT_MASK),
- HISI_PTT_WAIT_POLL_INTERVAL_US,
- HISI_PTT_WAIT_TUNE_TIMEOUT_US);
- }
- static ssize_t hisi_ptt_tune_attr_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct hisi_ptt *hisi_ptt = to_hisi_ptt(dev_get_drvdata(dev));
- struct dev_ext_attribute *ext_attr;
- struct hisi_ptt_tune_desc *desc;
- u32 reg;
- u16 val;
- ext_attr = container_of(attr, struct dev_ext_attribute, attr);
- desc = ext_attr->var;
- mutex_lock(&hisi_ptt->tune_lock);
- reg = readl(hisi_ptt->iobase + HISI_PTT_TUNING_CTRL);
- reg &= ~(HISI_PTT_TUNING_CTRL_CODE | HISI_PTT_TUNING_CTRL_SUB);
- reg |= FIELD_PREP(HISI_PTT_TUNING_CTRL_CODE | HISI_PTT_TUNING_CTRL_SUB,
- desc->event_code);
- writel(reg, hisi_ptt->iobase + HISI_PTT_TUNING_CTRL);
- /* Write all 1 to indicates it's the read process */
- writel(~0U, hisi_ptt->iobase + HISI_PTT_TUNING_DATA);
- if (!hisi_ptt_wait_tuning_finish(hisi_ptt)) {
- mutex_unlock(&hisi_ptt->tune_lock);
- return -ETIMEDOUT;
- }
- reg = readl(hisi_ptt->iobase + HISI_PTT_TUNING_DATA);
- reg &= HISI_PTT_TUNING_DATA_VAL_MASK;
- val = FIELD_GET(HISI_PTT_TUNING_DATA_VAL_MASK, reg);
- mutex_unlock(&hisi_ptt->tune_lock);
- return sysfs_emit(buf, "%u\n", val);
- }
- static ssize_t hisi_ptt_tune_attr_store(struct device *dev,
- struct device_attribute *attr,
- const char *buf, size_t count)
- {
- struct hisi_ptt *hisi_ptt = to_hisi_ptt(dev_get_drvdata(dev));
- struct dev_ext_attribute *ext_attr;
- struct hisi_ptt_tune_desc *desc;
- u32 reg;
- u16 val;
- ext_attr = container_of(attr, struct dev_ext_attribute, attr);
- desc = ext_attr->var;
- if (kstrtou16(buf, 10, &val))
- return -EINVAL;
- mutex_lock(&hisi_ptt->tune_lock);
- reg = readl(hisi_ptt->iobase + HISI_PTT_TUNING_CTRL);
- reg &= ~(HISI_PTT_TUNING_CTRL_CODE | HISI_PTT_TUNING_CTRL_SUB);
- reg |= FIELD_PREP(HISI_PTT_TUNING_CTRL_CODE | HISI_PTT_TUNING_CTRL_SUB,
- desc->event_code);
- writel(reg, hisi_ptt->iobase + HISI_PTT_TUNING_CTRL);
- writel(FIELD_PREP(HISI_PTT_TUNING_DATA_VAL_MASK, val),
- hisi_ptt->iobase + HISI_PTT_TUNING_DATA);
- if (!hisi_ptt_wait_tuning_finish(hisi_ptt)) {
- mutex_unlock(&hisi_ptt->tune_lock);
- return -ETIMEDOUT;
- }
- mutex_unlock(&hisi_ptt->tune_lock);
- return count;
- }
- #define HISI_PTT_TUNE_ATTR(_name, _val, _show, _store) \
- static struct hisi_ptt_tune_desc _name##_desc = { \
- .name = #_name, \
- .event_code = (_val), \
- }; \
- static struct dev_ext_attribute hisi_ptt_##_name##_attr = { \
- .attr = __ATTR(_name, 0600, _show, _store), \
- .var = &_name##_desc, \
- }
- #define HISI_PTT_TUNE_ATTR_COMMON(_name, _val) \
- HISI_PTT_TUNE_ATTR(_name, _val, \
- hisi_ptt_tune_attr_show, \
- hisi_ptt_tune_attr_store)
- /*
- * The value of the tuning event are composed of two parts: main event code
- * in BIT[0,15] and subevent code in BIT[16,23]. For example, qox_tx_cpl is
- * a subevent of 'Tx path QoS control' which for tuning the weight of Tx
- * completion TLPs. See hisi_ptt.rst documentation for more information.
- */
- #define HISI_PTT_TUNE_QOS_TX_CPL (0x4 | (3 << 16))
- #define HISI_PTT_TUNE_QOS_TX_NP (0x4 | (4 << 16))
- #define HISI_PTT_TUNE_QOS_TX_P (0x4 | (5 << 16))
- #define HISI_PTT_TUNE_RX_ALLOC_BUF_LEVEL (0x5 | (6 << 16))
- #define HISI_PTT_TUNE_TX_ALLOC_BUF_LEVEL (0x5 | (7 << 16))
- HISI_PTT_TUNE_ATTR_COMMON(qos_tx_cpl, HISI_PTT_TUNE_QOS_TX_CPL);
- HISI_PTT_TUNE_ATTR_COMMON(qos_tx_np, HISI_PTT_TUNE_QOS_TX_NP);
- HISI_PTT_TUNE_ATTR_COMMON(qos_tx_p, HISI_PTT_TUNE_QOS_TX_P);
- HISI_PTT_TUNE_ATTR_COMMON(rx_alloc_buf_level, HISI_PTT_TUNE_RX_ALLOC_BUF_LEVEL);
- HISI_PTT_TUNE_ATTR_COMMON(tx_alloc_buf_level, HISI_PTT_TUNE_TX_ALLOC_BUF_LEVEL);
- static struct attribute *hisi_ptt_tune_attrs[] = {
- &hisi_ptt_qos_tx_cpl_attr.attr.attr,
- &hisi_ptt_qos_tx_np_attr.attr.attr,
- &hisi_ptt_qos_tx_p_attr.attr.attr,
- &hisi_ptt_rx_alloc_buf_level_attr.attr.attr,
- &hisi_ptt_tx_alloc_buf_level_attr.attr.attr,
- NULL,
- };
- static struct attribute_group hisi_ptt_tune_group = {
- .name = "tune",
- .attrs = hisi_ptt_tune_attrs,
- };
- static u16 hisi_ptt_get_filter_val(u16 devid, bool is_port)
- {
- if (is_port)
- return BIT(HISI_PCIE_CORE_PORT_ID(devid & 0xff));
- return devid;
- }
- static bool hisi_ptt_wait_trace_hw_idle(struct hisi_ptt *hisi_ptt)
- {
- u32 val;
- return !readl_poll_timeout_atomic(hisi_ptt->iobase + HISI_PTT_TRACE_STS,
- val, val & HISI_PTT_TRACE_IDLE,
- HISI_PTT_WAIT_POLL_INTERVAL_US,
- HISI_PTT_WAIT_TRACE_TIMEOUT_US);
- }
- static void hisi_ptt_wait_dma_reset_done(struct hisi_ptt *hisi_ptt)
- {
- u32 val;
- readl_poll_timeout_atomic(hisi_ptt->iobase + HISI_PTT_TRACE_WR_STS,
- val, !val, HISI_PTT_RESET_POLL_INTERVAL_US,
- HISI_PTT_RESET_TIMEOUT_US);
- }
- static void hisi_ptt_trace_end(struct hisi_ptt *hisi_ptt)
- {
- writel(0, hisi_ptt->iobase + HISI_PTT_TRACE_CTRL);
- hisi_ptt->trace_ctrl.started = false;
- }
- static int hisi_ptt_trace_start(struct hisi_ptt *hisi_ptt)
- {
- struct hisi_ptt_trace_ctrl *ctrl = &hisi_ptt->trace_ctrl;
- u32 val;
- int i;
- /* Check device idle before start trace */
- if (!hisi_ptt_wait_trace_hw_idle(hisi_ptt)) {
- pci_err(hisi_ptt->pdev, "Failed to start trace, the device is still busy\n");
- return -EBUSY;
- }
- ctrl->started = true;
- /* Reset the DMA before start tracing */
- val = readl(hisi_ptt->iobase + HISI_PTT_TRACE_CTRL);
- val |= HISI_PTT_TRACE_CTRL_RST;
- writel(val, hisi_ptt->iobase + HISI_PTT_TRACE_CTRL);
- hisi_ptt_wait_dma_reset_done(hisi_ptt);
- val = readl(hisi_ptt->iobase + HISI_PTT_TRACE_CTRL);
- val &= ~HISI_PTT_TRACE_CTRL_RST;
- writel(val, hisi_ptt->iobase + HISI_PTT_TRACE_CTRL);
- /* Reset the index of current buffer */
- hisi_ptt->trace_ctrl.buf_index = 0;
- /* Zero the trace buffers */
- for (i = 0; i < HISI_PTT_TRACE_BUF_CNT; i++)
- memset(ctrl->trace_buf[i].addr, 0, HISI_PTT_TRACE_BUF_SIZE);
- /* Clear the interrupt status */
- writel(HISI_PTT_TRACE_INT_STAT_MASK, hisi_ptt->iobase + HISI_PTT_TRACE_INT_STAT);
- writel(0, hisi_ptt->iobase + HISI_PTT_TRACE_INT_MASK);
- /* Set the trace control register */
- val = FIELD_PREP(HISI_PTT_TRACE_CTRL_TYPE_SEL, ctrl->type);
- val |= FIELD_PREP(HISI_PTT_TRACE_CTRL_RXTX_SEL, ctrl->direction);
- val |= FIELD_PREP(HISI_PTT_TRACE_CTRL_DATA_FORMAT, ctrl->format);
- val |= FIELD_PREP(HISI_PTT_TRACE_CTRL_TARGET_SEL, hisi_ptt->trace_ctrl.filter);
- if (!hisi_ptt->trace_ctrl.is_port)
- val |= HISI_PTT_TRACE_CTRL_FILTER_MODE;
- /* Start the Trace */
- val |= HISI_PTT_TRACE_CTRL_EN;
- writel(val, hisi_ptt->iobase + HISI_PTT_TRACE_CTRL);
- return 0;
- }
- static int hisi_ptt_update_aux(struct hisi_ptt *hisi_ptt, int index, bool stop)
- {
- struct hisi_ptt_trace_ctrl *ctrl = &hisi_ptt->trace_ctrl;
- struct perf_output_handle *handle = &ctrl->handle;
- struct perf_event *event = handle->event;
- struct hisi_ptt_pmu_buf *buf;
- size_t size;
- void *addr;
- buf = perf_get_aux(handle);
- if (!buf || !handle->size)
- return -EINVAL;
- addr = ctrl->trace_buf[ctrl->buf_index].addr;
- /*
- * If we're going to stop, read the size of already traced data from
- * HISI_PTT_TRACE_WR_STS. Otherwise we're coming from the interrupt,
- * the data size is always HISI_PTT_TRACE_BUF_SIZE.
- */
- if (stop) {
- u32 reg;
- reg = readl(hisi_ptt->iobase + HISI_PTT_TRACE_WR_STS);
- size = FIELD_GET(HISI_PTT_TRACE_WR_STS_WRITE, reg);
- } else {
- size = HISI_PTT_TRACE_BUF_SIZE;
- }
- memcpy(buf->base + buf->pos, addr, size);
- buf->pos += size;
- /*
- * Just commit the traced data if we're going to stop. Otherwise if the
- * resident AUX buffer cannot contain the data of next trace buffer,
- * apply a new one.
- */
- if (stop) {
- perf_aux_output_end(handle, buf->pos);
- } else if (buf->length - buf->pos < HISI_PTT_TRACE_BUF_SIZE) {
- perf_aux_output_end(handle, buf->pos);
- buf = perf_aux_output_begin(handle, event);
- if (!buf)
- return -EINVAL;
- buf->pos = handle->head % buf->length;
- if (buf->length - buf->pos < HISI_PTT_TRACE_BUF_SIZE) {
- perf_aux_output_end(handle, 0);
- return -EINVAL;
- }
- }
- return 0;
- }
- static irqreturn_t hisi_ptt_isr(int irq, void *context)
- {
- struct hisi_ptt *hisi_ptt = context;
- u32 status, buf_idx;
- status = readl(hisi_ptt->iobase + HISI_PTT_TRACE_INT_STAT);
- if (!(status & HISI_PTT_TRACE_INT_STAT_MASK))
- return IRQ_NONE;
- buf_idx = ffs(status) - 1;
- /* Clear the interrupt status of buffer @buf_idx */
- writel(status, hisi_ptt->iobase + HISI_PTT_TRACE_INT_STAT);
- /*
- * Update the AUX buffer and cache the current buffer index,
- * as we need to know this and save the data when the trace
- * is ended out of the interrupt handler. End the trace
- * if the updating fails.
- */
- if (hisi_ptt_update_aux(hisi_ptt, buf_idx, false))
- hisi_ptt_trace_end(hisi_ptt);
- else
- hisi_ptt->trace_ctrl.buf_index = (buf_idx + 1) % HISI_PTT_TRACE_BUF_CNT;
- return IRQ_HANDLED;
- }
- static void hisi_ptt_irq_free_vectors(void *pdev)
- {
- pci_free_irq_vectors(pdev);
- }
- static int hisi_ptt_register_irq(struct hisi_ptt *hisi_ptt)
- {
- struct pci_dev *pdev = hisi_ptt->pdev;
- int ret;
- ret = pci_alloc_irq_vectors(pdev, 1, 1, PCI_IRQ_MSI);
- if (ret < 0) {
- pci_err(pdev, "failed to allocate irq vector, ret = %d\n", ret);
- return ret;
- }
- ret = devm_add_action_or_reset(&pdev->dev, hisi_ptt_irq_free_vectors, pdev);
- if (ret < 0)
- return ret;
- hisi_ptt->trace_irq = pci_irq_vector(pdev, HISI_PTT_TRACE_DMA_IRQ);
- ret = devm_request_threaded_irq(&pdev->dev, hisi_ptt->trace_irq,
- NULL, hisi_ptt_isr, 0,
- DRV_NAME, hisi_ptt);
- if (ret) {
- pci_err(pdev, "failed to request irq %d, ret = %d\n",
- hisi_ptt->trace_irq, ret);
- return ret;
- }
- return 0;
- }
- static int hisi_ptt_init_filters(struct pci_dev *pdev, void *data)
- {
- struct pci_dev *root_port = pcie_find_root_port(pdev);
- struct hisi_ptt_filter_desc *filter;
- struct hisi_ptt *hisi_ptt = data;
- u32 port_devid;
- if (!root_port)
- return 0;
- port_devid = PCI_DEVID(root_port->bus->number, root_port->devfn);
- if (port_devid < hisi_ptt->lower_bdf ||
- port_devid > hisi_ptt->upper_bdf)
- return 0;
- /*
- * We won't fail the probe if filter allocation failed here. The filters
- * should be partial initialized and users would know which filter fails
- * through the log. Other functions of PTT device are still available.
- */
- filter = kzalloc(sizeof(*filter), GFP_KERNEL);
- if (!filter) {
- pci_err(hisi_ptt->pdev, "failed to add filter %s\n", pci_name(pdev));
- return -ENOMEM;
- }
- filter->devid = PCI_DEVID(pdev->bus->number, pdev->devfn);
- if (pci_pcie_type(pdev) == PCI_EXP_TYPE_ROOT_PORT) {
- filter->is_port = true;
- list_add_tail(&filter->list, &hisi_ptt->port_filters);
- /* Update the available port mask */
- hisi_ptt->port_mask |= hisi_ptt_get_filter_val(filter->devid, true);
- } else {
- list_add_tail(&filter->list, &hisi_ptt->req_filters);
- }
- return 0;
- }
- static void hisi_ptt_release_filters(void *data)
- {
- struct hisi_ptt_filter_desc *filter, *tmp;
- struct hisi_ptt *hisi_ptt = data;
- list_for_each_entry_safe(filter, tmp, &hisi_ptt->req_filters, list) {
- list_del(&filter->list);
- kfree(filter);
- }
- list_for_each_entry_safe(filter, tmp, &hisi_ptt->port_filters, list) {
- list_del(&filter->list);
- kfree(filter);
- }
- }
- static int hisi_ptt_config_trace_buf(struct hisi_ptt *hisi_ptt)
- {
- struct hisi_ptt_trace_ctrl *ctrl = &hisi_ptt->trace_ctrl;
- struct device *dev = &hisi_ptt->pdev->dev;
- int i;
- ctrl->trace_buf = devm_kcalloc(dev, HISI_PTT_TRACE_BUF_CNT,
- sizeof(*ctrl->trace_buf), GFP_KERNEL);
- if (!ctrl->trace_buf)
- return -ENOMEM;
- for (i = 0; i < HISI_PTT_TRACE_BUF_CNT; ++i) {
- ctrl->trace_buf[i].addr = dmam_alloc_coherent(dev, HISI_PTT_TRACE_BUF_SIZE,
- &ctrl->trace_buf[i].dma,
- GFP_KERNEL);
- if (!ctrl->trace_buf[i].addr)
- return -ENOMEM;
- }
- /* Configure the trace DMA buffer */
- for (i = 0; i < HISI_PTT_TRACE_BUF_CNT; i++) {
- writel(lower_32_bits(ctrl->trace_buf[i].dma),
- hisi_ptt->iobase + HISI_PTT_TRACE_ADDR_BASE_LO_0 +
- i * HISI_PTT_TRACE_ADDR_STRIDE);
- writel(upper_32_bits(ctrl->trace_buf[i].dma),
- hisi_ptt->iobase + HISI_PTT_TRACE_ADDR_BASE_HI_0 +
- i * HISI_PTT_TRACE_ADDR_STRIDE);
- }
- writel(HISI_PTT_TRACE_BUF_SIZE, hisi_ptt->iobase + HISI_PTT_TRACE_ADDR_SIZE);
- return 0;
- }
- static int hisi_ptt_init_ctrls(struct hisi_ptt *hisi_ptt)
- {
- struct pci_dev *pdev = hisi_ptt->pdev;
- struct pci_bus *bus;
- int ret;
- u32 reg;
- INIT_LIST_HEAD(&hisi_ptt->port_filters);
- INIT_LIST_HEAD(&hisi_ptt->req_filters);
- ret = hisi_ptt_config_trace_buf(hisi_ptt);
- if (ret)
- return ret;
- /*
- * The device range register provides the information about the root
- * ports which the RCiEP can control and trace. The RCiEP and the root
- * ports which it supports are on the same PCIe core, with same domain
- * number but maybe different bus number. The device range register
- * will tell us which root ports we can support, Bit[31:16] indicates
- * the upper BDF numbers of the root port, while Bit[15:0] indicates
- * the lower.
- */
- reg = readl(hisi_ptt->iobase + HISI_PTT_DEVICE_RANGE);
- hisi_ptt->upper_bdf = FIELD_GET(HISI_PTT_DEVICE_RANGE_UPPER, reg);
- hisi_ptt->lower_bdf = FIELD_GET(HISI_PTT_DEVICE_RANGE_LOWER, reg);
- bus = pci_find_bus(pci_domain_nr(pdev->bus), PCI_BUS_NUM(hisi_ptt->upper_bdf));
- if (bus)
- pci_walk_bus(bus, hisi_ptt_init_filters, hisi_ptt);
- ret = devm_add_action_or_reset(&pdev->dev, hisi_ptt_release_filters, hisi_ptt);
- if (ret)
- return ret;
- hisi_ptt->trace_ctrl.on_cpu = -1;
- return 0;
- }
- static ssize_t cpumask_show(struct device *dev, struct device_attribute *attr,
- char *buf)
- {
- struct hisi_ptt *hisi_ptt = to_hisi_ptt(dev_get_drvdata(dev));
- const cpumask_t *cpumask = cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev));
- return cpumap_print_to_pagebuf(true, buf, cpumask);
- }
- static DEVICE_ATTR_RO(cpumask);
- static struct attribute *hisi_ptt_cpumask_attrs[] = {
- &dev_attr_cpumask.attr,
- NULL
- };
- static const struct attribute_group hisi_ptt_cpumask_attr_group = {
- .attrs = hisi_ptt_cpumask_attrs,
- };
- /*
- * Bit 19 indicates the filter type, 1 for Root Port filter and 0 for Requester
- * filter. Bit[15:0] indicates the filter value, for Root Port filter it's
- * a bit mask of desired ports and for Requester filter it's the Requester ID
- * of the desired PCIe function. Bit[18:16] is reserved for extension.
- *
- * See hisi_ptt.rst documentation for detailed information.
- */
- PMU_FORMAT_ATTR(filter, "config:0-19");
- PMU_FORMAT_ATTR(direction, "config:20-23");
- PMU_FORMAT_ATTR(type, "config:24-31");
- PMU_FORMAT_ATTR(format, "config:32-35");
- static struct attribute *hisi_ptt_pmu_format_attrs[] = {
- &format_attr_filter.attr,
- &format_attr_direction.attr,
- &format_attr_type.attr,
- &format_attr_format.attr,
- NULL
- };
- static struct attribute_group hisi_ptt_pmu_format_group = {
- .name = "format",
- .attrs = hisi_ptt_pmu_format_attrs,
- };
- static const struct attribute_group *hisi_ptt_pmu_groups[] = {
- &hisi_ptt_cpumask_attr_group,
- &hisi_ptt_pmu_format_group,
- &hisi_ptt_tune_group,
- NULL
- };
- static int hisi_ptt_trace_valid_direction(u32 val)
- {
- /*
- * The direction values have different effects according to the data
- * format (specified in the parentheses). TLP set A/B means different
- * set of TLP types. See hisi_ptt.rst documentation for more details.
- */
- static const u32 hisi_ptt_trace_available_direction[] = {
- 0, /* inbound(4DW) or reserved(8DW) */
- 1, /* outbound(4DW) */
- 2, /* {in, out}bound(4DW) or inbound(8DW), TLP set A */
- 3, /* {in, out}bound(4DW) or inbound(8DW), TLP set B */
- };
- int i;
- for (i = 0; i < ARRAY_SIZE(hisi_ptt_trace_available_direction); i++) {
- if (val == hisi_ptt_trace_available_direction[i])
- return 0;
- }
- return -EINVAL;
- }
- static int hisi_ptt_trace_valid_type(u32 val)
- {
- /* Different types can be set simultaneously */
- static const u32 hisi_ptt_trace_available_type[] = {
- 1, /* posted_request */
- 2, /* non-posted_request */
- 4, /* completion */
- };
- int i;
- if (!val)
- return -EINVAL;
- /*
- * Walk the available list and clear the valid bits of
- * the config. If there is any resident bit after the
- * walk then the config is invalid.
- */
- for (i = 0; i < ARRAY_SIZE(hisi_ptt_trace_available_type); i++)
- val &= ~hisi_ptt_trace_available_type[i];
- if (val)
- return -EINVAL;
- return 0;
- }
- static int hisi_ptt_trace_valid_format(u32 val)
- {
- static const u32 hisi_ptt_trace_availble_format[] = {
- 0, /* 4DW */
- 1, /* 8DW */
- };
- int i;
- for (i = 0; i < ARRAY_SIZE(hisi_ptt_trace_availble_format); i++) {
- if (val == hisi_ptt_trace_availble_format[i])
- return 0;
- }
- return -EINVAL;
- }
- static int hisi_ptt_trace_valid_filter(struct hisi_ptt *hisi_ptt, u64 config)
- {
- unsigned long val, port_mask = hisi_ptt->port_mask;
- struct hisi_ptt_filter_desc *filter;
- hisi_ptt->trace_ctrl.is_port = FIELD_GET(HISI_PTT_PMU_FILTER_IS_PORT, config);
- val = FIELD_GET(HISI_PTT_PMU_FILTER_VAL_MASK, config);
- /*
- * Port filters are defined as bit mask. For port filters, check
- * the bits in the @val are within the range of hisi_ptt->port_mask
- * and whether it's empty or not, otherwise user has specified
- * some unsupported root ports.
- *
- * For Requester ID filters, walk the available filter list to see
- * whether we have one matched.
- */
- if (!hisi_ptt->trace_ctrl.is_port) {
- list_for_each_entry(filter, &hisi_ptt->req_filters, list) {
- if (val == hisi_ptt_get_filter_val(filter->devid, filter->is_port))
- return 0;
- }
- } else if (bitmap_subset(&val, &port_mask, BITS_PER_LONG)) {
- return 0;
- }
- return -EINVAL;
- }
- static void hisi_ptt_pmu_init_configs(struct hisi_ptt *hisi_ptt, struct perf_event *event)
- {
- struct hisi_ptt_trace_ctrl *ctrl = &hisi_ptt->trace_ctrl;
- u32 val;
- val = FIELD_GET(HISI_PTT_PMU_FILTER_VAL_MASK, event->attr.config);
- hisi_ptt->trace_ctrl.filter = val;
- val = FIELD_GET(HISI_PTT_PMU_DIRECTION_MASK, event->attr.config);
- ctrl->direction = val;
- val = FIELD_GET(HISI_PTT_PMU_TYPE_MASK, event->attr.config);
- ctrl->type = val;
- val = FIELD_GET(HISI_PTT_PMU_FORMAT_MASK, event->attr.config);
- ctrl->format = val;
- }
- static int hisi_ptt_pmu_event_init(struct perf_event *event)
- {
- struct hisi_ptt *hisi_ptt = to_hisi_ptt(event->pmu);
- int ret;
- u32 val;
- if (event->cpu < 0) {
- dev_dbg(event->pmu->dev, "Per-task mode not supported\n");
- return -EOPNOTSUPP;
- }
- if (event->attr.type != hisi_ptt->hisi_ptt_pmu.type)
- return -ENOENT;
- ret = hisi_ptt_trace_valid_filter(hisi_ptt, event->attr.config);
- if (ret < 0)
- return ret;
- val = FIELD_GET(HISI_PTT_PMU_DIRECTION_MASK, event->attr.config);
- ret = hisi_ptt_trace_valid_direction(val);
- if (ret < 0)
- return ret;
- val = FIELD_GET(HISI_PTT_PMU_TYPE_MASK, event->attr.config);
- ret = hisi_ptt_trace_valid_type(val);
- if (ret < 0)
- return ret;
- val = FIELD_GET(HISI_PTT_PMU_FORMAT_MASK, event->attr.config);
- return hisi_ptt_trace_valid_format(val);
- }
- static void *hisi_ptt_pmu_setup_aux(struct perf_event *event, void **pages,
- int nr_pages, bool overwrite)
- {
- struct hisi_ptt_pmu_buf *buf;
- struct page **pagelist;
- int i;
- if (overwrite) {
- dev_warn(event->pmu->dev, "Overwrite mode is not supported\n");
- return NULL;
- }
- /* If the pages size less than buffers, we cannot start trace */
- if (nr_pages < HISI_PTT_TRACE_TOTAL_BUF_SIZE / PAGE_SIZE)
- return NULL;
- buf = kzalloc(sizeof(*buf), GFP_KERNEL);
- if (!buf)
- return NULL;
- pagelist = kcalloc(nr_pages, sizeof(*pagelist), GFP_KERNEL);
- if (!pagelist)
- goto err;
- for (i = 0; i < nr_pages; i++)
- pagelist[i] = virt_to_page(pages[i]);
- buf->base = vmap(pagelist, nr_pages, VM_MAP, PAGE_KERNEL);
- if (!buf->base) {
- kfree(pagelist);
- goto err;
- }
- buf->nr_pages = nr_pages;
- buf->length = nr_pages * PAGE_SIZE;
- buf->pos = 0;
- kfree(pagelist);
- return buf;
- err:
- kfree(buf);
- return NULL;
- }
- static void hisi_ptt_pmu_free_aux(void *aux)
- {
- struct hisi_ptt_pmu_buf *buf = aux;
- vunmap(buf->base);
- kfree(buf);
- }
- static void hisi_ptt_pmu_start(struct perf_event *event, int flags)
- {
- struct hisi_ptt *hisi_ptt = to_hisi_ptt(event->pmu);
- struct perf_output_handle *handle = &hisi_ptt->trace_ctrl.handle;
- struct hw_perf_event *hwc = &event->hw;
- struct device *dev = event->pmu->dev;
- struct hisi_ptt_pmu_buf *buf;
- int cpu = event->cpu;
- int ret;
- hwc->state = 0;
- /* Serialize the perf process if user specified several CPUs */
- spin_lock(&hisi_ptt->pmu_lock);
- if (hisi_ptt->trace_ctrl.started) {
- dev_dbg(dev, "trace has already started\n");
- goto stop;
- }
- /*
- * Handle the interrupt on the same cpu which starts the trace to avoid
- * context mismatch. Otherwise we'll trigger the WARN from the perf
- * core in event_function_local(). If CPU passed is offline we'll fail
- * here, just log it since we can do nothing here.
- */
- ret = irq_set_affinity(hisi_ptt->trace_irq, cpumask_of(cpu));
- if (ret)
- dev_warn(dev, "failed to set the affinity of trace interrupt\n");
- hisi_ptt->trace_ctrl.on_cpu = cpu;
- buf = perf_aux_output_begin(handle, event);
- if (!buf) {
- dev_dbg(dev, "aux output begin failed\n");
- goto stop;
- }
- buf->pos = handle->head % buf->length;
- hisi_ptt_pmu_init_configs(hisi_ptt, event);
- ret = hisi_ptt_trace_start(hisi_ptt);
- if (ret) {
- dev_dbg(dev, "trace start failed, ret = %d\n", ret);
- perf_aux_output_end(handle, 0);
- goto stop;
- }
- spin_unlock(&hisi_ptt->pmu_lock);
- return;
- stop:
- event->hw.state |= PERF_HES_STOPPED;
- spin_unlock(&hisi_ptt->pmu_lock);
- }
- static void hisi_ptt_pmu_stop(struct perf_event *event, int flags)
- {
- struct hisi_ptt *hisi_ptt = to_hisi_ptt(event->pmu);
- struct hw_perf_event *hwc = &event->hw;
- if (hwc->state & PERF_HES_STOPPED)
- return;
- spin_lock(&hisi_ptt->pmu_lock);
- if (hisi_ptt->trace_ctrl.started) {
- hisi_ptt_trace_end(hisi_ptt);
- if (!hisi_ptt_wait_trace_hw_idle(hisi_ptt))
- dev_warn(event->pmu->dev, "Device is still busy\n");
- hisi_ptt_update_aux(hisi_ptt, hisi_ptt->trace_ctrl.buf_index, true);
- }
- spin_unlock(&hisi_ptt->pmu_lock);
- hwc->state |= PERF_HES_STOPPED;
- perf_event_update_userpage(event);
- hwc->state |= PERF_HES_UPTODATE;
- }
- static int hisi_ptt_pmu_add(struct perf_event *event, int flags)
- {
- struct hisi_ptt *hisi_ptt = to_hisi_ptt(event->pmu);
- struct hw_perf_event *hwc = &event->hw;
- int cpu = event->cpu;
- /* Only allow the cpus on the device's node to add the event */
- if (!cpumask_test_cpu(cpu, cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev))))
- return 0;
- hwc->state = PERF_HES_STOPPED | PERF_HES_UPTODATE;
- if (flags & PERF_EF_START) {
- hisi_ptt_pmu_start(event, PERF_EF_RELOAD);
- if (hwc->state & PERF_HES_STOPPED)
- return -EINVAL;
- }
- return 0;
- }
- static void hisi_ptt_pmu_del(struct perf_event *event, int flags)
- {
- hisi_ptt_pmu_stop(event, PERF_EF_UPDATE);
- }
- static void hisi_ptt_pmu_read(struct perf_event *event)
- {
- }
- static void hisi_ptt_remove_cpuhp_instance(void *hotplug_node)
- {
- cpuhp_state_remove_instance_nocalls(hisi_ptt_pmu_online, hotplug_node);
- }
- static void hisi_ptt_unregister_pmu(void *pmu)
- {
- perf_pmu_unregister(pmu);
- }
- static int hisi_ptt_register_pmu(struct hisi_ptt *hisi_ptt)
- {
- u16 core_id, sicl_id;
- char *pmu_name;
- u32 reg;
- int ret;
- ret = cpuhp_state_add_instance_nocalls(hisi_ptt_pmu_online,
- &hisi_ptt->hotplug_node);
- if (ret)
- return ret;
- ret = devm_add_action_or_reset(&hisi_ptt->pdev->dev,
- hisi_ptt_remove_cpuhp_instance,
- &hisi_ptt->hotplug_node);
- if (ret)
- return ret;
- mutex_init(&hisi_ptt->tune_lock);
- spin_lock_init(&hisi_ptt->pmu_lock);
- hisi_ptt->hisi_ptt_pmu = (struct pmu) {
- .module = THIS_MODULE,
- .capabilities = PERF_PMU_CAP_EXCLUSIVE | PERF_PMU_CAP_ITRACE,
- .task_ctx_nr = perf_sw_context,
- .attr_groups = hisi_ptt_pmu_groups,
- .event_init = hisi_ptt_pmu_event_init,
- .setup_aux = hisi_ptt_pmu_setup_aux,
- .free_aux = hisi_ptt_pmu_free_aux,
- .start = hisi_ptt_pmu_start,
- .stop = hisi_ptt_pmu_stop,
- .add = hisi_ptt_pmu_add,
- .del = hisi_ptt_pmu_del,
- .read = hisi_ptt_pmu_read,
- };
- reg = readl(hisi_ptt->iobase + HISI_PTT_LOCATION);
- core_id = FIELD_GET(HISI_PTT_CORE_ID, reg);
- sicl_id = FIELD_GET(HISI_PTT_SICL_ID, reg);
- pmu_name = devm_kasprintf(&hisi_ptt->pdev->dev, GFP_KERNEL, "hisi_ptt%u_%u",
- sicl_id, core_id);
- if (!pmu_name)
- return -ENOMEM;
- ret = perf_pmu_register(&hisi_ptt->hisi_ptt_pmu, pmu_name, -1);
- if (ret)
- return ret;
- return devm_add_action_or_reset(&hisi_ptt->pdev->dev,
- hisi_ptt_unregister_pmu,
- &hisi_ptt->hisi_ptt_pmu);
- }
- /*
- * The DMA of PTT trace can only use direct mappings due to some
- * hardware restriction. Check whether there is no IOMMU or the
- * policy of the IOMMU domain is passthrough, otherwise the trace
- * cannot work.
- *
- * The PTT device is supposed to behind an ARM SMMUv3, which
- * should have passthrough the device by a quirk.
- */
- static int hisi_ptt_check_iommu_mapping(struct pci_dev *pdev)
- {
- struct iommu_domain *iommu_domain;
- iommu_domain = iommu_get_domain_for_dev(&pdev->dev);
- if (!iommu_domain || iommu_domain->type == IOMMU_DOMAIN_IDENTITY)
- return 0;
- return -EOPNOTSUPP;
- }
- static int hisi_ptt_probe(struct pci_dev *pdev,
- const struct pci_device_id *id)
- {
- struct hisi_ptt *hisi_ptt;
- int ret;
- ret = hisi_ptt_check_iommu_mapping(pdev);
- if (ret) {
- pci_err(pdev, "requires direct DMA mappings\n");
- return ret;
- }
- hisi_ptt = devm_kzalloc(&pdev->dev, sizeof(*hisi_ptt), GFP_KERNEL);
- if (!hisi_ptt)
- return -ENOMEM;
- hisi_ptt->pdev = pdev;
- pci_set_drvdata(pdev, hisi_ptt);
- ret = pcim_enable_device(pdev);
- if (ret) {
- pci_err(pdev, "failed to enable device, ret = %d\n", ret);
- return ret;
- }
- ret = pcim_iomap_regions(pdev, BIT(2), DRV_NAME);
- if (ret) {
- pci_err(pdev, "failed to remap io memory, ret = %d\n", ret);
- return ret;
- }
- hisi_ptt->iobase = pcim_iomap_table(pdev)[2];
- ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64));
- if (ret) {
- pci_err(pdev, "failed to set 64 bit dma mask, ret = %d\n", ret);
- return ret;
- }
- pci_set_master(pdev);
- ret = hisi_ptt_register_irq(hisi_ptt);
- if (ret)
- return ret;
- ret = hisi_ptt_init_ctrls(hisi_ptt);
- if (ret) {
- pci_err(pdev, "failed to init controls, ret = %d\n", ret);
- return ret;
- }
- ret = hisi_ptt_register_pmu(hisi_ptt);
- if (ret) {
- pci_err(pdev, "failed to register PMU device, ret = %d", ret);
- return ret;
- }
- return 0;
- }
- static const struct pci_device_id hisi_ptt_id_tbl[] = {
- { PCI_DEVICE(PCI_VENDOR_ID_HUAWEI, 0xa12e) },
- { }
- };
- MODULE_DEVICE_TABLE(pci, hisi_ptt_id_tbl);
- static struct pci_driver hisi_ptt_driver = {
- .name = DRV_NAME,
- .id_table = hisi_ptt_id_tbl,
- .probe = hisi_ptt_probe,
- };
- static int hisi_ptt_cpu_teardown(unsigned int cpu, struct hlist_node *node)
- {
- struct hisi_ptt *hisi_ptt;
- struct device *dev;
- int target, src;
- hisi_ptt = hlist_entry_safe(node, struct hisi_ptt, hotplug_node);
- src = hisi_ptt->trace_ctrl.on_cpu;
- dev = hisi_ptt->hisi_ptt_pmu.dev;
- if (!hisi_ptt->trace_ctrl.started || src != cpu)
- return 0;
- target = cpumask_any_but(cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev)), cpu);
- if (target >= nr_cpu_ids) {
- dev_err(dev, "no available cpu for perf context migration\n");
- return 0;
- }
- perf_pmu_migrate_context(&hisi_ptt->hisi_ptt_pmu, src, target);
- /*
- * Also make sure the interrupt bind to the migrated CPU as well. Warn
- * the user on failure here.
- */
- if (irq_set_affinity(hisi_ptt->trace_irq, cpumask_of(target)))
- dev_warn(dev, "failed to set the affinity of trace interrupt\n");
- hisi_ptt->trace_ctrl.on_cpu = target;
- return 0;
- }
- static int __init hisi_ptt_init(void)
- {
- int ret;
- ret = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN, DRV_NAME, NULL,
- hisi_ptt_cpu_teardown);
- if (ret < 0)
- return ret;
- hisi_ptt_pmu_online = ret;
- ret = pci_register_driver(&hisi_ptt_driver);
- if (ret)
- cpuhp_remove_multi_state(hisi_ptt_pmu_online);
- return ret;
- }
- module_init(hisi_ptt_init);
- static void __exit hisi_ptt_exit(void)
- {
- pci_unregister_driver(&hisi_ptt_driver);
- cpuhp_remove_multi_state(hisi_ptt_pmu_online);
- }
- module_exit(hisi_ptt_exit);
- MODULE_LICENSE("GPL");
- MODULE_AUTHOR("Yicong Yang <[email protected]>");
- MODULE_DESCRIPTION("Driver for HiSilicon PCIe tune and trace device");
|