123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * core.c - Implementation of core module of MOST Linux driver stack
- *
- * Copyright (C) 2013-2020 Microchip Technology Germany II GmbH & Co. KG
- */
- #include <linux/module.h>
- #include <linux/fs.h>
- #include <linux/slab.h>
- #include <linux/init.h>
- #include <linux/device.h>
- #include <linux/list.h>
- #include <linux/poll.h>
- #include <linux/wait.h>
- #include <linux/kobject.h>
- #include <linux/mutex.h>
- #include <linux/completion.h>
- #include <linux/sysfs.h>
- #include <linux/kthread.h>
- #include <linux/dma-mapping.h>
- #include <linux/idr.h>
- #include <linux/most.h>
- #define MAX_CHANNELS 64
- #define STRING_SIZE 80
- static struct ida mdev_id;
- static int dummy_num_buffers;
- static struct list_head comp_list;
- struct pipe {
- struct most_component *comp;
- int refs;
- int num_buffers;
- };
- struct most_channel {
- struct device dev;
- struct completion cleanup;
- atomic_t mbo_ref;
- atomic_t mbo_nq_level;
- u16 channel_id;
- char name[STRING_SIZE];
- bool is_poisoned;
- struct mutex start_mutex; /* channel activation synchronization */
- struct mutex nq_mutex; /* nq thread synchronization */
- int is_starving;
- struct most_interface *iface;
- struct most_channel_config cfg;
- bool keep_mbo;
- bool enqueue_halt;
- struct list_head fifo;
- spinlock_t fifo_lock; /* fifo access synchronization */
- struct list_head halt_fifo;
- struct list_head list;
- struct pipe pipe0;
- struct pipe pipe1;
- struct list_head trash_fifo;
- struct task_struct *hdm_enqueue_task;
- wait_queue_head_t hdm_fifo_wq;
- };
- #define to_channel(d) container_of(d, struct most_channel, dev)
- struct interface_private {
- int dev_id;
- char name[STRING_SIZE];
- struct most_channel *channel[MAX_CHANNELS];
- struct list_head channel_list;
- };
- static const struct {
- int most_ch_data_type;
- const char *name;
- } ch_data_type[] = {
- { MOST_CH_CONTROL, "control" },
- { MOST_CH_ASYNC, "async" },
- { MOST_CH_SYNC, "sync" },
- { MOST_CH_ISOC, "isoc"},
- { MOST_CH_ISOC, "isoc_avp"},
- };
- /**
- * list_pop_mbo - retrieves the first MBO of the list and removes it
- * @ptr: the list head to grab the MBO from.
- */
- #define list_pop_mbo(ptr) \
- ({ \
- struct mbo *_mbo = list_first_entry(ptr, struct mbo, list); \
- list_del(&_mbo->list); \
- _mbo; \
- })
- /**
- * most_free_mbo_coherent - free an MBO and its coherent buffer
- * @mbo: most buffer
- */
- static void most_free_mbo_coherent(struct mbo *mbo)
- {
- struct most_channel *c = mbo->context;
- u16 const coherent_buf_size = c->cfg.buffer_size + c->cfg.extra_len;
- if (c->iface->dma_free)
- c->iface->dma_free(mbo, coherent_buf_size);
- else
- kfree(mbo->virt_address);
- kfree(mbo);
- if (atomic_sub_and_test(1, &c->mbo_ref))
- complete(&c->cleanup);
- }
- /**
- * flush_channel_fifos - clear the channel fifos
- * @c: pointer to channel object
- */
- static void flush_channel_fifos(struct most_channel *c)
- {
- unsigned long flags, hf_flags;
- struct mbo *mbo, *tmp;
- if (list_empty(&c->fifo) && list_empty(&c->halt_fifo))
- return;
- spin_lock_irqsave(&c->fifo_lock, flags);
- list_for_each_entry_safe(mbo, tmp, &c->fifo, list) {
- list_del(&mbo->list);
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- most_free_mbo_coherent(mbo);
- spin_lock_irqsave(&c->fifo_lock, flags);
- }
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- spin_lock_irqsave(&c->fifo_lock, hf_flags);
- list_for_each_entry_safe(mbo, tmp, &c->halt_fifo, list) {
- list_del(&mbo->list);
- spin_unlock_irqrestore(&c->fifo_lock, hf_flags);
- most_free_mbo_coherent(mbo);
- spin_lock_irqsave(&c->fifo_lock, hf_flags);
- }
- spin_unlock_irqrestore(&c->fifo_lock, hf_flags);
- if (unlikely((!list_empty(&c->fifo) || !list_empty(&c->halt_fifo))))
- dev_warn(&c->dev, "Channel or trash fifo not empty\n");
- }
- /**
- * flush_trash_fifo - clear the trash fifo
- * @c: pointer to channel object
- */
- static int flush_trash_fifo(struct most_channel *c)
- {
- struct mbo *mbo, *tmp;
- unsigned long flags;
- spin_lock_irqsave(&c->fifo_lock, flags);
- list_for_each_entry_safe(mbo, tmp, &c->trash_fifo, list) {
- list_del(&mbo->list);
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- most_free_mbo_coherent(mbo);
- spin_lock_irqsave(&c->fifo_lock, flags);
- }
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- return 0;
- }
- static ssize_t available_directions_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- unsigned int i = c->channel_id;
- strcpy(buf, "");
- if (c->iface->channel_vector[i].direction & MOST_CH_RX)
- strcat(buf, "rx ");
- if (c->iface->channel_vector[i].direction & MOST_CH_TX)
- strcat(buf, "tx ");
- strcat(buf, "\n");
- return strlen(buf);
- }
- static ssize_t available_datatypes_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- unsigned int i = c->channel_id;
- strcpy(buf, "");
- if (c->iface->channel_vector[i].data_type & MOST_CH_CONTROL)
- strcat(buf, "control ");
- if (c->iface->channel_vector[i].data_type & MOST_CH_ASYNC)
- strcat(buf, "async ");
- if (c->iface->channel_vector[i].data_type & MOST_CH_SYNC)
- strcat(buf, "sync ");
- if (c->iface->channel_vector[i].data_type & MOST_CH_ISOC)
- strcat(buf, "isoc ");
- strcat(buf, "\n");
- return strlen(buf);
- }
- static ssize_t number_of_packet_buffers_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- unsigned int i = c->channel_id;
- return snprintf(buf, PAGE_SIZE, "%d\n",
- c->iface->channel_vector[i].num_buffers_packet);
- }
- static ssize_t number_of_stream_buffers_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- unsigned int i = c->channel_id;
- return snprintf(buf, PAGE_SIZE, "%d\n",
- c->iface->channel_vector[i].num_buffers_streaming);
- }
- static ssize_t size_of_packet_buffer_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- unsigned int i = c->channel_id;
- return snprintf(buf, PAGE_SIZE, "%d\n",
- c->iface->channel_vector[i].buffer_size_packet);
- }
- static ssize_t size_of_stream_buffer_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- unsigned int i = c->channel_id;
- return snprintf(buf, PAGE_SIZE, "%d\n",
- c->iface->channel_vector[i].buffer_size_streaming);
- }
- static ssize_t channel_starving_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- return snprintf(buf, PAGE_SIZE, "%d\n", c->is_starving);
- }
- static ssize_t set_number_of_buffers_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- return snprintf(buf, PAGE_SIZE, "%d\n", c->cfg.num_buffers);
- }
- static ssize_t set_buffer_size_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- return snprintf(buf, PAGE_SIZE, "%d\n", c->cfg.buffer_size);
- }
- static ssize_t set_direction_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- if (c->cfg.direction & MOST_CH_TX)
- return snprintf(buf, PAGE_SIZE, "tx\n");
- else if (c->cfg.direction & MOST_CH_RX)
- return snprintf(buf, PAGE_SIZE, "rx\n");
- return snprintf(buf, PAGE_SIZE, "unconfigured\n");
- }
- static ssize_t set_datatype_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- int i;
- struct most_channel *c = to_channel(dev);
- for (i = 0; i < ARRAY_SIZE(ch_data_type); i++) {
- if (c->cfg.data_type & ch_data_type[i].most_ch_data_type)
- return snprintf(buf, PAGE_SIZE, "%s",
- ch_data_type[i].name);
- }
- return snprintf(buf, PAGE_SIZE, "unconfigured\n");
- }
- static ssize_t set_subbuffer_size_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- return snprintf(buf, PAGE_SIZE, "%d\n", c->cfg.subbuffer_size);
- }
- static ssize_t set_packets_per_xact_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_channel *c = to_channel(dev);
- return snprintf(buf, PAGE_SIZE, "%d\n", c->cfg.packets_per_xact);
- }
- static ssize_t set_dbr_size_show(struct device *dev,
- struct device_attribute *attr, char *buf)
- {
- struct most_channel *c = to_channel(dev);
- return snprintf(buf, PAGE_SIZE, "%d\n", c->cfg.dbr_size);
- }
- #define to_dev_attr(a) container_of(a, struct device_attribute, attr)
- static umode_t channel_attr_is_visible(struct kobject *kobj,
- struct attribute *attr, int index)
- {
- struct device_attribute *dev_attr = to_dev_attr(attr);
- struct device *dev = kobj_to_dev(kobj);
- struct most_channel *c = to_channel(dev);
- if (!strcmp(dev_attr->attr.name, "set_dbr_size") &&
- (c->iface->interface != ITYPE_MEDIALB_DIM2))
- return 0;
- if (!strcmp(dev_attr->attr.name, "set_packets_per_xact") &&
- (c->iface->interface != ITYPE_USB))
- return 0;
- return attr->mode;
- }
- #define DEV_ATTR(_name) (&dev_attr_##_name.attr)
- static DEVICE_ATTR_RO(available_directions);
- static DEVICE_ATTR_RO(available_datatypes);
- static DEVICE_ATTR_RO(number_of_packet_buffers);
- static DEVICE_ATTR_RO(number_of_stream_buffers);
- static DEVICE_ATTR_RO(size_of_stream_buffer);
- static DEVICE_ATTR_RO(size_of_packet_buffer);
- static DEVICE_ATTR_RO(channel_starving);
- static DEVICE_ATTR_RO(set_buffer_size);
- static DEVICE_ATTR_RO(set_number_of_buffers);
- static DEVICE_ATTR_RO(set_direction);
- static DEVICE_ATTR_RO(set_datatype);
- static DEVICE_ATTR_RO(set_subbuffer_size);
- static DEVICE_ATTR_RO(set_packets_per_xact);
- static DEVICE_ATTR_RO(set_dbr_size);
- static struct attribute *channel_attrs[] = {
- DEV_ATTR(available_directions),
- DEV_ATTR(available_datatypes),
- DEV_ATTR(number_of_packet_buffers),
- DEV_ATTR(number_of_stream_buffers),
- DEV_ATTR(size_of_stream_buffer),
- DEV_ATTR(size_of_packet_buffer),
- DEV_ATTR(channel_starving),
- DEV_ATTR(set_buffer_size),
- DEV_ATTR(set_number_of_buffers),
- DEV_ATTR(set_direction),
- DEV_ATTR(set_datatype),
- DEV_ATTR(set_subbuffer_size),
- DEV_ATTR(set_packets_per_xact),
- DEV_ATTR(set_dbr_size),
- NULL,
- };
- static const struct attribute_group channel_attr_group = {
- .attrs = channel_attrs,
- .is_visible = channel_attr_is_visible,
- };
- static const struct attribute_group *channel_attr_groups[] = {
- &channel_attr_group,
- NULL,
- };
- static ssize_t description_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_interface *iface = dev_get_drvdata(dev);
- return snprintf(buf, PAGE_SIZE, "%s\n", iface->description);
- }
- static ssize_t interface_show(struct device *dev,
- struct device_attribute *attr,
- char *buf)
- {
- struct most_interface *iface = dev_get_drvdata(dev);
- switch (iface->interface) {
- case ITYPE_LOOPBACK:
- return snprintf(buf, PAGE_SIZE, "loopback\n");
- case ITYPE_I2C:
- return snprintf(buf, PAGE_SIZE, "i2c\n");
- case ITYPE_I2S:
- return snprintf(buf, PAGE_SIZE, "i2s\n");
- case ITYPE_TSI:
- return snprintf(buf, PAGE_SIZE, "tsi\n");
- case ITYPE_HBI:
- return snprintf(buf, PAGE_SIZE, "hbi\n");
- case ITYPE_MEDIALB_DIM:
- return snprintf(buf, PAGE_SIZE, "mlb_dim\n");
- case ITYPE_MEDIALB_DIM2:
- return snprintf(buf, PAGE_SIZE, "mlb_dim2\n");
- case ITYPE_USB:
- return snprintf(buf, PAGE_SIZE, "usb\n");
- case ITYPE_PCIE:
- return snprintf(buf, PAGE_SIZE, "pcie\n");
- }
- return snprintf(buf, PAGE_SIZE, "unknown\n");
- }
- static DEVICE_ATTR_RO(description);
- static DEVICE_ATTR_RO(interface);
- static struct attribute *interface_attrs[] = {
- DEV_ATTR(description),
- DEV_ATTR(interface),
- NULL,
- };
- static const struct attribute_group interface_attr_group = {
- .attrs = interface_attrs,
- };
- static const struct attribute_group *interface_attr_groups[] = {
- &interface_attr_group,
- NULL,
- };
- static struct most_component *match_component(char *name)
- {
- struct most_component *comp;
- list_for_each_entry(comp, &comp_list, list) {
- if (!strcmp(comp->name, name))
- return comp;
- }
- return NULL;
- }
- struct show_links_data {
- int offs;
- char *buf;
- };
- static int print_links(struct device *dev, void *data)
- {
- struct show_links_data *d = data;
- int offs = d->offs;
- char *buf = d->buf;
- struct most_channel *c;
- struct most_interface *iface = dev_get_drvdata(dev);
- list_for_each_entry(c, &iface->p->channel_list, list) {
- if (c->pipe0.comp) {
- offs += scnprintf(buf + offs,
- PAGE_SIZE - offs,
- "%s:%s:%s\n",
- c->pipe0.comp->name,
- dev_name(iface->dev),
- dev_name(&c->dev));
- }
- if (c->pipe1.comp) {
- offs += scnprintf(buf + offs,
- PAGE_SIZE - offs,
- "%s:%s:%s\n",
- c->pipe1.comp->name,
- dev_name(iface->dev),
- dev_name(&c->dev));
- }
- }
- d->offs = offs;
- return 0;
- }
- static int most_match(struct device *dev, struct device_driver *drv)
- {
- if (!strcmp(dev_name(dev), "most"))
- return 0;
- else
- return 1;
- }
- static struct bus_type mostbus = {
- .name = "most",
- .match = most_match,
- };
- static ssize_t links_show(struct device_driver *drv, char *buf)
- {
- struct show_links_data d = { .buf = buf };
- bus_for_each_dev(&mostbus, NULL, &d, print_links);
- return d.offs;
- }
- static ssize_t components_show(struct device_driver *drv, char *buf)
- {
- struct most_component *comp;
- int offs = 0;
- list_for_each_entry(comp, &comp_list, list) {
- offs += scnprintf(buf + offs, PAGE_SIZE - offs, "%s\n",
- comp->name);
- }
- return offs;
- }
- /**
- * get_channel - get pointer to channel
- * @mdev: name of the device interface
- * @mdev_ch: name of channel
- */
- static struct most_channel *get_channel(char *mdev, char *mdev_ch)
- {
- struct device *dev = NULL;
- struct most_interface *iface;
- struct most_channel *c, *tmp;
- dev = bus_find_device_by_name(&mostbus, NULL, mdev);
- if (!dev)
- return NULL;
- put_device(dev);
- iface = dev_get_drvdata(dev);
- list_for_each_entry_safe(c, tmp, &iface->p->channel_list, list) {
- if (!strcmp(dev_name(&c->dev), mdev_ch))
- return c;
- }
- return NULL;
- }
- static
- inline int link_channel_to_component(struct most_channel *c,
- struct most_component *comp,
- char *name,
- char *comp_param)
- {
- int ret;
- struct most_component **comp_ptr;
- if (!c->pipe0.comp)
- comp_ptr = &c->pipe0.comp;
- else if (!c->pipe1.comp)
- comp_ptr = &c->pipe1.comp;
- else
- return -ENOSPC;
- *comp_ptr = comp;
- ret = comp->probe_channel(c->iface, c->channel_id, &c->cfg, name,
- comp_param);
- if (ret) {
- *comp_ptr = NULL;
- return ret;
- }
- return 0;
- }
- int most_set_cfg_buffer_size(char *mdev, char *mdev_ch, u16 val)
- {
- struct most_channel *c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- c->cfg.buffer_size = val;
- return 0;
- }
- int most_set_cfg_subbuffer_size(char *mdev, char *mdev_ch, u16 val)
- {
- struct most_channel *c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- c->cfg.subbuffer_size = val;
- return 0;
- }
- int most_set_cfg_dbr_size(char *mdev, char *mdev_ch, u16 val)
- {
- struct most_channel *c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- c->cfg.dbr_size = val;
- return 0;
- }
- int most_set_cfg_num_buffers(char *mdev, char *mdev_ch, u16 val)
- {
- struct most_channel *c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- c->cfg.num_buffers = val;
- return 0;
- }
- int most_set_cfg_datatype(char *mdev, char *mdev_ch, char *buf)
- {
- int i;
- struct most_channel *c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- for (i = 0; i < ARRAY_SIZE(ch_data_type); i++) {
- if (!strcmp(buf, ch_data_type[i].name)) {
- c->cfg.data_type = ch_data_type[i].most_ch_data_type;
- break;
- }
- }
- if (i == ARRAY_SIZE(ch_data_type))
- dev_warn(&c->dev, "Invalid attribute settings\n");
- return 0;
- }
- int most_set_cfg_direction(char *mdev, char *mdev_ch, char *buf)
- {
- struct most_channel *c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- if (!strcmp(buf, "dir_rx")) {
- c->cfg.direction = MOST_CH_RX;
- } else if (!strcmp(buf, "rx")) {
- c->cfg.direction = MOST_CH_RX;
- } else if (!strcmp(buf, "dir_tx")) {
- c->cfg.direction = MOST_CH_TX;
- } else if (!strcmp(buf, "tx")) {
- c->cfg.direction = MOST_CH_TX;
- } else {
- dev_err(&c->dev, "Invalid direction\n");
- return -ENODATA;
- }
- return 0;
- }
- int most_set_cfg_packets_xact(char *mdev, char *mdev_ch, u16 val)
- {
- struct most_channel *c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- c->cfg.packets_per_xact = val;
- return 0;
- }
- int most_cfg_complete(char *comp_name)
- {
- struct most_component *comp;
- comp = match_component(comp_name);
- if (!comp)
- return -ENODEV;
- return comp->cfg_complete();
- }
- int most_add_link(char *mdev, char *mdev_ch, char *comp_name, char *link_name,
- char *comp_param)
- {
- struct most_channel *c = get_channel(mdev, mdev_ch);
- struct most_component *comp = match_component(comp_name);
- if (!c || !comp)
- return -ENODEV;
- return link_channel_to_component(c, comp, link_name, comp_param);
- }
- int most_remove_link(char *mdev, char *mdev_ch, char *comp_name)
- {
- struct most_channel *c;
- struct most_component *comp;
- comp = match_component(comp_name);
- if (!comp)
- return -ENODEV;
- c = get_channel(mdev, mdev_ch);
- if (!c)
- return -ENODEV;
- if (comp->disconnect_channel(c->iface, c->channel_id))
- return -EIO;
- if (c->pipe0.comp == comp)
- c->pipe0.comp = NULL;
- if (c->pipe1.comp == comp)
- c->pipe1.comp = NULL;
- return 0;
- }
- #define DRV_ATTR(_name) (&driver_attr_##_name.attr)
- static DRIVER_ATTR_RO(links);
- static DRIVER_ATTR_RO(components);
- static struct attribute *mc_attrs[] = {
- DRV_ATTR(links),
- DRV_ATTR(components),
- NULL,
- };
- static const struct attribute_group mc_attr_group = {
- .attrs = mc_attrs,
- };
- static const struct attribute_group *mc_attr_groups[] = {
- &mc_attr_group,
- NULL,
- };
- static struct device_driver mostbus_driver = {
- .name = "most_core",
- .bus = &mostbus,
- .groups = mc_attr_groups,
- };
- static inline void trash_mbo(struct mbo *mbo)
- {
- unsigned long flags;
- struct most_channel *c = mbo->context;
- spin_lock_irqsave(&c->fifo_lock, flags);
- list_add(&mbo->list, &c->trash_fifo);
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- }
- static bool hdm_mbo_ready(struct most_channel *c)
- {
- bool empty;
- if (c->enqueue_halt)
- return false;
- spin_lock_irq(&c->fifo_lock);
- empty = list_empty(&c->halt_fifo);
- spin_unlock_irq(&c->fifo_lock);
- return !empty;
- }
- static void nq_hdm_mbo(struct mbo *mbo)
- {
- unsigned long flags;
- struct most_channel *c = mbo->context;
- spin_lock_irqsave(&c->fifo_lock, flags);
- list_add_tail(&mbo->list, &c->halt_fifo);
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- wake_up_interruptible(&c->hdm_fifo_wq);
- }
- static int hdm_enqueue_thread(void *data)
- {
- struct most_channel *c = data;
- struct mbo *mbo;
- int ret;
- typeof(c->iface->enqueue) enqueue = c->iface->enqueue;
- while (likely(!kthread_should_stop())) {
- wait_event_interruptible(c->hdm_fifo_wq,
- hdm_mbo_ready(c) ||
- kthread_should_stop());
- mutex_lock(&c->nq_mutex);
- spin_lock_irq(&c->fifo_lock);
- if (unlikely(c->enqueue_halt || list_empty(&c->halt_fifo))) {
- spin_unlock_irq(&c->fifo_lock);
- mutex_unlock(&c->nq_mutex);
- continue;
- }
- mbo = list_pop_mbo(&c->halt_fifo);
- spin_unlock_irq(&c->fifo_lock);
- if (c->cfg.direction == MOST_CH_RX)
- mbo->buffer_length = c->cfg.buffer_size;
- ret = enqueue(mbo->ifp, mbo->hdm_channel_id, mbo);
- mutex_unlock(&c->nq_mutex);
- if (unlikely(ret)) {
- dev_err(&c->dev, "Buffer enqueue failed\n");
- nq_hdm_mbo(mbo);
- c->hdm_enqueue_task = NULL;
- return 0;
- }
- }
- return 0;
- }
- static int run_enqueue_thread(struct most_channel *c, int channel_id)
- {
- struct task_struct *task =
- kthread_run(hdm_enqueue_thread, c, "hdm_fifo_%d",
- channel_id);
- if (IS_ERR(task))
- return PTR_ERR(task);
- c->hdm_enqueue_task = task;
- return 0;
- }
- /**
- * arm_mbo - recycle MBO for further usage
- * @mbo: most buffer
- *
- * This puts an MBO back to the list to have it ready for up coming
- * tx transactions.
- *
- * In case the MBO belongs to a channel that recently has been
- * poisoned, the MBO is scheduled to be trashed.
- * Calls the completion handler of an attached component.
- */
- static void arm_mbo(struct mbo *mbo)
- {
- unsigned long flags;
- struct most_channel *c;
- c = mbo->context;
- if (c->is_poisoned) {
- trash_mbo(mbo);
- return;
- }
- spin_lock_irqsave(&c->fifo_lock, flags);
- ++*mbo->num_buffers_ptr;
- list_add_tail(&mbo->list, &c->fifo);
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- if (c->pipe0.refs && c->pipe0.comp->tx_completion)
- c->pipe0.comp->tx_completion(c->iface, c->channel_id);
- if (c->pipe1.refs && c->pipe1.comp->tx_completion)
- c->pipe1.comp->tx_completion(c->iface, c->channel_id);
- }
- /**
- * arm_mbo_chain - helper function that arms an MBO chain for the HDM
- * @c: pointer to interface channel
- * @dir: direction of the channel
- * @compl: pointer to completion function
- *
- * This allocates buffer objects including the containing DMA coherent
- * buffer and puts them in the fifo.
- * Buffers of Rx channels are put in the kthread fifo, hence immediately
- * submitted to the HDM.
- *
- * Returns the number of allocated and enqueued MBOs.
- */
- static int arm_mbo_chain(struct most_channel *c, int dir,
- void (*compl)(struct mbo *))
- {
- unsigned int i;
- struct mbo *mbo;
- unsigned long flags;
- u32 coherent_buf_size = c->cfg.buffer_size + c->cfg.extra_len;
- atomic_set(&c->mbo_nq_level, 0);
- for (i = 0; i < c->cfg.num_buffers; i++) {
- mbo = kzalloc(sizeof(*mbo), GFP_KERNEL);
- if (!mbo)
- goto flush_fifos;
- mbo->context = c;
- mbo->ifp = c->iface;
- mbo->hdm_channel_id = c->channel_id;
- if (c->iface->dma_alloc) {
- mbo->virt_address =
- c->iface->dma_alloc(mbo, coherent_buf_size);
- } else {
- mbo->virt_address =
- kzalloc(coherent_buf_size, GFP_KERNEL);
- }
- if (!mbo->virt_address)
- goto release_mbo;
- mbo->complete = compl;
- mbo->num_buffers_ptr = &dummy_num_buffers;
- if (dir == MOST_CH_RX) {
- nq_hdm_mbo(mbo);
- atomic_inc(&c->mbo_nq_level);
- } else {
- spin_lock_irqsave(&c->fifo_lock, flags);
- list_add_tail(&mbo->list, &c->fifo);
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- }
- }
- return c->cfg.num_buffers;
- release_mbo:
- kfree(mbo);
- flush_fifos:
- flush_channel_fifos(c);
- return 0;
- }
- /**
- * most_submit_mbo - submits an MBO to fifo
- * @mbo: most buffer
- */
- void most_submit_mbo(struct mbo *mbo)
- {
- if (WARN_ONCE(!mbo || !mbo->context,
- "Bad buffer or missing channel reference\n"))
- return;
- nq_hdm_mbo(mbo);
- }
- EXPORT_SYMBOL_GPL(most_submit_mbo);
- /**
- * most_write_completion - write completion handler
- * @mbo: most buffer
- *
- * This recycles the MBO for further usage. In case the channel has been
- * poisoned, the MBO is scheduled to be trashed.
- */
- static void most_write_completion(struct mbo *mbo)
- {
- struct most_channel *c;
- c = mbo->context;
- if (unlikely(c->is_poisoned || (mbo->status == MBO_E_CLOSE)))
- trash_mbo(mbo);
- else
- arm_mbo(mbo);
- }
- int channel_has_mbo(struct most_interface *iface, int id,
- struct most_component *comp)
- {
- struct most_channel *c = iface->p->channel[id];
- unsigned long flags;
- int empty;
- if (unlikely(!c))
- return -EINVAL;
- if (c->pipe0.refs && c->pipe1.refs &&
- ((comp == c->pipe0.comp && c->pipe0.num_buffers <= 0) ||
- (comp == c->pipe1.comp && c->pipe1.num_buffers <= 0)))
- return 0;
- spin_lock_irqsave(&c->fifo_lock, flags);
- empty = list_empty(&c->fifo);
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- return !empty;
- }
- EXPORT_SYMBOL_GPL(channel_has_mbo);
- /**
- * most_get_mbo - get pointer to an MBO of pool
- * @iface: pointer to interface instance
- * @id: channel ID
- * @comp: driver component
- *
- * This attempts to get a free buffer out of the channel fifo.
- * Returns a pointer to MBO on success or NULL otherwise.
- */
- struct mbo *most_get_mbo(struct most_interface *iface, int id,
- struct most_component *comp)
- {
- struct mbo *mbo;
- struct most_channel *c;
- unsigned long flags;
- int *num_buffers_ptr;
- c = iface->p->channel[id];
- if (unlikely(!c))
- return NULL;
- if (c->pipe0.refs && c->pipe1.refs &&
- ((comp == c->pipe0.comp && c->pipe0.num_buffers <= 0) ||
- (comp == c->pipe1.comp && c->pipe1.num_buffers <= 0)))
- return NULL;
- if (comp == c->pipe0.comp)
- num_buffers_ptr = &c->pipe0.num_buffers;
- else if (comp == c->pipe1.comp)
- num_buffers_ptr = &c->pipe1.num_buffers;
- else
- num_buffers_ptr = &dummy_num_buffers;
- spin_lock_irqsave(&c->fifo_lock, flags);
- if (list_empty(&c->fifo)) {
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- return NULL;
- }
- mbo = list_pop_mbo(&c->fifo);
- --*num_buffers_ptr;
- spin_unlock_irqrestore(&c->fifo_lock, flags);
- mbo->num_buffers_ptr = num_buffers_ptr;
- mbo->buffer_length = c->cfg.buffer_size;
- return mbo;
- }
- EXPORT_SYMBOL_GPL(most_get_mbo);
- /**
- * most_put_mbo - return buffer to pool
- * @mbo: most buffer
- */
- void most_put_mbo(struct mbo *mbo)
- {
- struct most_channel *c = mbo->context;
- if (c->cfg.direction == MOST_CH_TX) {
- arm_mbo(mbo);
- return;
- }
- nq_hdm_mbo(mbo);
- atomic_inc(&c->mbo_nq_level);
- }
- EXPORT_SYMBOL_GPL(most_put_mbo);
- /**
- * most_read_completion - read completion handler
- * @mbo: most buffer
- *
- * This function is called by the HDM when data has been received from the
- * hardware and copied to the buffer of the MBO.
- *
- * In case the channel has been poisoned it puts the buffer in the trash queue.
- * Otherwise, it passes the buffer to an component for further processing.
- */
- static void most_read_completion(struct mbo *mbo)
- {
- struct most_channel *c = mbo->context;
- if (unlikely(c->is_poisoned || (mbo->status == MBO_E_CLOSE))) {
- trash_mbo(mbo);
- return;
- }
- if (mbo->status == MBO_E_INVAL) {
- nq_hdm_mbo(mbo);
- atomic_inc(&c->mbo_nq_level);
- return;
- }
- if (atomic_sub_and_test(1, &c->mbo_nq_level))
- c->is_starving = 1;
- if (c->pipe0.refs && c->pipe0.comp->rx_completion &&
- c->pipe0.comp->rx_completion(mbo) == 0)
- return;
- if (c->pipe1.refs && c->pipe1.comp->rx_completion &&
- c->pipe1.comp->rx_completion(mbo) == 0)
- return;
- most_put_mbo(mbo);
- }
- /**
- * most_start_channel - prepares a channel for communication
- * @iface: pointer to interface instance
- * @id: channel ID
- * @comp: driver component
- *
- * This prepares the channel for usage. Cross-checks whether the
- * channel's been properly configured.
- *
- * Returns 0 on success or error code otherwise.
- */
- int most_start_channel(struct most_interface *iface, int id,
- struct most_component *comp)
- {
- int num_buffer;
- int ret;
- struct most_channel *c = iface->p->channel[id];
- if (unlikely(!c))
- return -EINVAL;
- mutex_lock(&c->start_mutex);
- if (c->pipe0.refs + c->pipe1.refs > 0)
- goto out; /* already started by another component */
- if (!try_module_get(iface->mod)) {
- dev_err(&c->dev, "Failed to acquire HDM lock\n");
- mutex_unlock(&c->start_mutex);
- return -ENOLCK;
- }
- c->cfg.extra_len = 0;
- if (c->iface->configure(c->iface, c->channel_id, &c->cfg)) {
- dev_err(&c->dev, "Channel configuration failed. Go check settings...\n");
- ret = -EINVAL;
- goto err_put_module;
- }
- init_waitqueue_head(&c->hdm_fifo_wq);
- if (c->cfg.direction == MOST_CH_RX)
- num_buffer = arm_mbo_chain(c, c->cfg.direction,
- most_read_completion);
- else
- num_buffer = arm_mbo_chain(c, c->cfg.direction,
- most_write_completion);
- if (unlikely(!num_buffer)) {
- ret = -ENOMEM;
- goto err_put_module;
- }
- ret = run_enqueue_thread(c, id);
- if (ret)
- goto err_put_module;
- c->is_starving = 0;
- c->pipe0.num_buffers = c->cfg.num_buffers / 2;
- c->pipe1.num_buffers = c->cfg.num_buffers - c->pipe0.num_buffers;
- atomic_set(&c->mbo_ref, num_buffer);
- out:
- if (comp == c->pipe0.comp)
- c->pipe0.refs++;
- if (comp == c->pipe1.comp)
- c->pipe1.refs++;
- mutex_unlock(&c->start_mutex);
- return 0;
- err_put_module:
- module_put(iface->mod);
- mutex_unlock(&c->start_mutex);
- return ret;
- }
- EXPORT_SYMBOL_GPL(most_start_channel);
- /**
- * most_stop_channel - stops a running channel
- * @iface: pointer to interface instance
- * @id: channel ID
- * @comp: driver component
- */
- int most_stop_channel(struct most_interface *iface, int id,
- struct most_component *comp)
- {
- struct most_channel *c;
- if (unlikely((!iface) || (id >= iface->num_channels) || (id < 0))) {
- pr_err("Bad interface or index out of range\n");
- return -EINVAL;
- }
- c = iface->p->channel[id];
- if (unlikely(!c))
- return -EINVAL;
- mutex_lock(&c->start_mutex);
- if (c->pipe0.refs + c->pipe1.refs >= 2)
- goto out;
- if (c->hdm_enqueue_task)
- kthread_stop(c->hdm_enqueue_task);
- c->hdm_enqueue_task = NULL;
- if (iface->mod)
- module_put(iface->mod);
- c->is_poisoned = true;
- if (c->iface->poison_channel(c->iface, c->channel_id)) {
- dev_err(&c->dev, "Failed to stop channel %d of interface %s\n", c->channel_id,
- c->iface->description);
- mutex_unlock(&c->start_mutex);
- return -EAGAIN;
- }
- flush_trash_fifo(c);
- flush_channel_fifos(c);
- #ifdef CMPL_INTERRUPTIBLE
- if (wait_for_completion_interruptible(&c->cleanup)) {
- dev_err(&c->dev, "Interrupted while cleaning up channel %d\n", c->channel_id);
- mutex_unlock(&c->start_mutex);
- return -EINTR;
- }
- #else
- wait_for_completion(&c->cleanup);
- #endif
- c->is_poisoned = false;
- out:
- if (comp == c->pipe0.comp)
- c->pipe0.refs--;
- if (comp == c->pipe1.comp)
- c->pipe1.refs--;
- mutex_unlock(&c->start_mutex);
- return 0;
- }
- EXPORT_SYMBOL_GPL(most_stop_channel);
- /**
- * most_register_component - registers a driver component with the core
- * @comp: driver component
- */
- int most_register_component(struct most_component *comp)
- {
- if (!comp) {
- pr_err("Bad component\n");
- return -EINVAL;
- }
- list_add_tail(&comp->list, &comp_list);
- return 0;
- }
- EXPORT_SYMBOL_GPL(most_register_component);
- static int disconnect_channels(struct device *dev, void *data)
- {
- struct most_interface *iface;
- struct most_channel *c, *tmp;
- struct most_component *comp = data;
- iface = dev_get_drvdata(dev);
- list_for_each_entry_safe(c, tmp, &iface->p->channel_list, list) {
- if (c->pipe0.comp == comp || c->pipe1.comp == comp)
- comp->disconnect_channel(c->iface, c->channel_id);
- if (c->pipe0.comp == comp)
- c->pipe0.comp = NULL;
- if (c->pipe1.comp == comp)
- c->pipe1.comp = NULL;
- }
- return 0;
- }
- /**
- * most_deregister_component - deregisters a driver component with the core
- * @comp: driver component
- */
- int most_deregister_component(struct most_component *comp)
- {
- if (!comp) {
- pr_err("Bad component\n");
- return -EINVAL;
- }
- bus_for_each_dev(&mostbus, NULL, comp, disconnect_channels);
- list_del(&comp->list);
- return 0;
- }
- EXPORT_SYMBOL_GPL(most_deregister_component);
- static void release_channel(struct device *dev)
- {
- struct most_channel *c = to_channel(dev);
- kfree(c);
- }
- /**
- * most_register_interface - registers an interface with core
- * @iface: device interface
- *
- * Allocates and initializes a new interface instance and all of its channels.
- * Returns a pointer to kobject or an error pointer.
- */
- int most_register_interface(struct most_interface *iface)
- {
- unsigned int i;
- int id;
- struct most_channel *c;
- if (!iface || !iface->enqueue || !iface->configure ||
- !iface->poison_channel || (iface->num_channels > MAX_CHANNELS))
- return -EINVAL;
- id = ida_simple_get(&mdev_id, 0, 0, GFP_KERNEL);
- if (id < 0) {
- dev_err(iface->dev, "Failed to allocate device ID\n");
- return id;
- }
- iface->p = kzalloc(sizeof(*iface->p), GFP_KERNEL);
- if (!iface->p) {
- ida_simple_remove(&mdev_id, id);
- return -ENOMEM;
- }
- INIT_LIST_HEAD(&iface->p->channel_list);
- iface->p->dev_id = id;
- strscpy(iface->p->name, iface->description, sizeof(iface->p->name));
- iface->dev->bus = &mostbus;
- iface->dev->groups = interface_attr_groups;
- dev_set_drvdata(iface->dev, iface);
- if (device_register(iface->dev)) {
- dev_err(iface->dev, "Failed to register interface device\n");
- kfree(iface->p);
- put_device(iface->dev);
- ida_simple_remove(&mdev_id, id);
- return -ENOMEM;
- }
- for (i = 0; i < iface->num_channels; i++) {
- const char *name_suffix = iface->channel_vector[i].name_suffix;
- c = kzalloc(sizeof(*c), GFP_KERNEL);
- if (!c)
- goto err_free_resources;
- if (!name_suffix)
- snprintf(c->name, STRING_SIZE, "ch%d", i);
- else
- snprintf(c->name, STRING_SIZE, "%s", name_suffix);
- c->dev.init_name = c->name;
- c->dev.parent = iface->dev;
- c->dev.groups = channel_attr_groups;
- c->dev.release = release_channel;
- iface->p->channel[i] = c;
- c->is_starving = 0;
- c->iface = iface;
- c->channel_id = i;
- c->keep_mbo = false;
- c->enqueue_halt = false;
- c->is_poisoned = false;
- c->cfg.direction = 0;
- c->cfg.data_type = 0;
- c->cfg.num_buffers = 0;
- c->cfg.buffer_size = 0;
- c->cfg.subbuffer_size = 0;
- c->cfg.packets_per_xact = 0;
- spin_lock_init(&c->fifo_lock);
- INIT_LIST_HEAD(&c->fifo);
- INIT_LIST_HEAD(&c->trash_fifo);
- INIT_LIST_HEAD(&c->halt_fifo);
- init_completion(&c->cleanup);
- atomic_set(&c->mbo_ref, 0);
- mutex_init(&c->start_mutex);
- mutex_init(&c->nq_mutex);
- list_add_tail(&c->list, &iface->p->channel_list);
- if (device_register(&c->dev)) {
- dev_err(&c->dev, "Failed to register channel device\n");
- goto err_free_most_channel;
- }
- }
- most_interface_register_notify(iface->description);
- return 0;
- err_free_most_channel:
- put_device(&c->dev);
- err_free_resources:
- while (i > 0) {
- c = iface->p->channel[--i];
- device_unregister(&c->dev);
- }
- kfree(iface->p);
- device_unregister(iface->dev);
- ida_simple_remove(&mdev_id, id);
- return -ENOMEM;
- }
- EXPORT_SYMBOL_GPL(most_register_interface);
- /**
- * most_deregister_interface - deregisters an interface with core
- * @iface: device interface
- *
- * Before removing an interface instance from the list, all running
- * channels are stopped and poisoned.
- */
- void most_deregister_interface(struct most_interface *iface)
- {
- int i;
- struct most_channel *c;
- for (i = 0; i < iface->num_channels; i++) {
- c = iface->p->channel[i];
- if (c->pipe0.comp)
- c->pipe0.comp->disconnect_channel(c->iface,
- c->channel_id);
- if (c->pipe1.comp)
- c->pipe1.comp->disconnect_channel(c->iface,
- c->channel_id);
- c->pipe0.comp = NULL;
- c->pipe1.comp = NULL;
- list_del(&c->list);
- device_unregister(&c->dev);
- }
- ida_simple_remove(&mdev_id, iface->p->dev_id);
- kfree(iface->p);
- device_unregister(iface->dev);
- }
- EXPORT_SYMBOL_GPL(most_deregister_interface);
- /**
- * most_stop_enqueue - prevents core from enqueueing MBOs
- * @iface: pointer to interface
- * @id: channel id
- *
- * This is called by an HDM that _cannot_ attend to its duties and
- * is imminent to get run over by the core. The core is not going to
- * enqueue any further packets unless the flagging HDM calls
- * most_resume enqueue().
- */
- void most_stop_enqueue(struct most_interface *iface, int id)
- {
- struct most_channel *c = iface->p->channel[id];
- if (!c)
- return;
- mutex_lock(&c->nq_mutex);
- c->enqueue_halt = true;
- mutex_unlock(&c->nq_mutex);
- }
- EXPORT_SYMBOL_GPL(most_stop_enqueue);
- /**
- * most_resume_enqueue - allow core to enqueue MBOs again
- * @iface: pointer to interface
- * @id: channel id
- *
- * This clears the enqueue halt flag and enqueues all MBOs currently
- * sitting in the wait fifo.
- */
- void most_resume_enqueue(struct most_interface *iface, int id)
- {
- struct most_channel *c = iface->p->channel[id];
- if (!c)
- return;
- mutex_lock(&c->nq_mutex);
- c->enqueue_halt = false;
- mutex_unlock(&c->nq_mutex);
- wake_up_interruptible(&c->hdm_fifo_wq);
- }
- EXPORT_SYMBOL_GPL(most_resume_enqueue);
- static int __init most_init(void)
- {
- int err;
- INIT_LIST_HEAD(&comp_list);
- ida_init(&mdev_id);
- err = bus_register(&mostbus);
- if (err) {
- pr_err("Failed to register most bus\n");
- return err;
- }
- err = driver_register(&mostbus_driver);
- if (err) {
- pr_err("Failed to register core driver\n");
- goto err_unregister_bus;
- }
- configfs_init();
- return 0;
- err_unregister_bus:
- bus_unregister(&mostbus);
- return err;
- }
- static void __exit most_exit(void)
- {
- driver_unregister(&mostbus_driver);
- bus_unregister(&mostbus);
- ida_destroy(&mdev_id);
- }
- subsys_initcall(most_init);
- module_exit(most_exit);
- MODULE_LICENSE("GPL");
- MODULE_AUTHOR("Christian Gromm <[email protected]>");
- MODULE_DESCRIPTION("Core module of stacked MOST Linux driver");
|