dfl.c 46 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Driver for FPGA Device Feature List (DFL) Support
  4. *
  5. * Copyright (C) 2017-2018 Intel Corporation, Inc.
  6. *
  7. * Authors:
  8. * Kang Luwei <[email protected]>
  9. * Zhang Yi <[email protected]>
  10. * Wu Hao <[email protected]>
  11. * Xiao Guangrong <[email protected]>
  12. */
  13. #include <linux/dfl.h>
  14. #include <linux/fpga-dfl.h>
  15. #include <linux/module.h>
  16. #include <linux/uaccess.h>
  17. #include "dfl.h"
  18. static DEFINE_MUTEX(dfl_id_mutex);
  19. /*
  20. * when adding a new feature dev support in DFL framework, it's required to
  21. * add a new item in enum dfl_id_type and provide related information in below
  22. * dfl_devs table which is indexed by dfl_id_type, e.g. name string used for
  23. * platform device creation (define name strings in dfl.h, as they could be
  24. * reused by platform device drivers).
  25. *
  26. * if the new feature dev needs chardev support, then it's required to add
  27. * a new item in dfl_chardevs table and configure dfl_devs[i].devt_type as
  28. * index to dfl_chardevs table. If no chardev support just set devt_type
  29. * as one invalid index (DFL_FPGA_DEVT_MAX).
  30. */
  31. enum dfl_fpga_devt_type {
  32. DFL_FPGA_DEVT_FME,
  33. DFL_FPGA_DEVT_PORT,
  34. DFL_FPGA_DEVT_MAX,
  35. };
  36. static struct lock_class_key dfl_pdata_keys[DFL_ID_MAX];
  37. static const char *dfl_pdata_key_strings[DFL_ID_MAX] = {
  38. "dfl-fme-pdata",
  39. "dfl-port-pdata",
  40. };
  41. /**
  42. * dfl_dev_info - dfl feature device information.
  43. * @name: name string of the feature platform device.
  44. * @dfh_id: id value in Device Feature Header (DFH) register by DFL spec.
  45. * @id: idr id of the feature dev.
  46. * @devt_type: index to dfl_chrdevs[].
  47. */
  48. struct dfl_dev_info {
  49. const char *name;
  50. u16 dfh_id;
  51. struct idr id;
  52. enum dfl_fpga_devt_type devt_type;
  53. };
  54. /* it is indexed by dfl_id_type */
  55. static struct dfl_dev_info dfl_devs[] = {
  56. {.name = DFL_FPGA_FEATURE_DEV_FME, .dfh_id = DFH_ID_FIU_FME,
  57. .devt_type = DFL_FPGA_DEVT_FME},
  58. {.name = DFL_FPGA_FEATURE_DEV_PORT, .dfh_id = DFH_ID_FIU_PORT,
  59. .devt_type = DFL_FPGA_DEVT_PORT},
  60. };
  61. /**
  62. * dfl_chardev_info - chardev information of dfl feature device
  63. * @name: nmae string of the char device.
  64. * @devt: devt of the char device.
  65. */
  66. struct dfl_chardev_info {
  67. const char *name;
  68. dev_t devt;
  69. };
  70. /* indexed by enum dfl_fpga_devt_type */
  71. static struct dfl_chardev_info dfl_chrdevs[] = {
  72. {.name = DFL_FPGA_FEATURE_DEV_FME},
  73. {.name = DFL_FPGA_FEATURE_DEV_PORT},
  74. };
  75. static void dfl_ids_init(void)
  76. {
  77. int i;
  78. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  79. idr_init(&dfl_devs[i].id);
  80. }
  81. static void dfl_ids_destroy(void)
  82. {
  83. int i;
  84. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  85. idr_destroy(&dfl_devs[i].id);
  86. }
  87. static int dfl_id_alloc(enum dfl_id_type type, struct device *dev)
  88. {
  89. int id;
  90. WARN_ON(type >= DFL_ID_MAX);
  91. mutex_lock(&dfl_id_mutex);
  92. id = idr_alloc(&dfl_devs[type].id, dev, 0, 0, GFP_KERNEL);
  93. mutex_unlock(&dfl_id_mutex);
  94. return id;
  95. }
  96. static void dfl_id_free(enum dfl_id_type type, int id)
  97. {
  98. WARN_ON(type >= DFL_ID_MAX);
  99. mutex_lock(&dfl_id_mutex);
  100. idr_remove(&dfl_devs[type].id, id);
  101. mutex_unlock(&dfl_id_mutex);
  102. }
  103. static enum dfl_id_type feature_dev_id_type(struct platform_device *pdev)
  104. {
  105. int i;
  106. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  107. if (!strcmp(dfl_devs[i].name, pdev->name))
  108. return i;
  109. return DFL_ID_MAX;
  110. }
  111. static enum dfl_id_type dfh_id_to_type(u16 id)
  112. {
  113. int i;
  114. for (i = 0; i < ARRAY_SIZE(dfl_devs); i++)
  115. if (dfl_devs[i].dfh_id == id)
  116. return i;
  117. return DFL_ID_MAX;
  118. }
  119. /*
  120. * introduce a global port_ops list, it allows port drivers to register ops
  121. * in such list, then other feature devices (e.g. FME), could use the port
  122. * functions even related port platform device is hidden. Below is one example,
  123. * in virtualization case of PCIe-based FPGA DFL device, when SRIOV is
  124. * enabled, port (and it's AFU) is turned into VF and port platform device
  125. * is hidden from system but it's still required to access port to finish FPGA
  126. * reconfiguration function in FME.
  127. */
  128. static DEFINE_MUTEX(dfl_port_ops_mutex);
  129. static LIST_HEAD(dfl_port_ops_list);
  130. /**
  131. * dfl_fpga_port_ops_get - get matched port ops from the global list
  132. * @pdev: platform device to match with associated port ops.
  133. * Return: matched port ops on success, NULL otherwise.
  134. *
  135. * Please note that must dfl_fpga_port_ops_put after use the port_ops.
  136. */
  137. struct dfl_fpga_port_ops *dfl_fpga_port_ops_get(struct platform_device *pdev)
  138. {
  139. struct dfl_fpga_port_ops *ops = NULL;
  140. mutex_lock(&dfl_port_ops_mutex);
  141. if (list_empty(&dfl_port_ops_list))
  142. goto done;
  143. list_for_each_entry(ops, &dfl_port_ops_list, node) {
  144. /* match port_ops using the name of platform device */
  145. if (!strcmp(pdev->name, ops->name)) {
  146. if (!try_module_get(ops->owner))
  147. ops = NULL;
  148. goto done;
  149. }
  150. }
  151. ops = NULL;
  152. done:
  153. mutex_unlock(&dfl_port_ops_mutex);
  154. return ops;
  155. }
  156. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_get);
  157. /**
  158. * dfl_fpga_port_ops_put - put port ops
  159. * @ops: port ops.
  160. */
  161. void dfl_fpga_port_ops_put(struct dfl_fpga_port_ops *ops)
  162. {
  163. if (ops && ops->owner)
  164. module_put(ops->owner);
  165. }
  166. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_put);
  167. /**
  168. * dfl_fpga_port_ops_add - add port_ops to global list
  169. * @ops: port ops to add.
  170. */
  171. void dfl_fpga_port_ops_add(struct dfl_fpga_port_ops *ops)
  172. {
  173. mutex_lock(&dfl_port_ops_mutex);
  174. list_add_tail(&ops->node, &dfl_port_ops_list);
  175. mutex_unlock(&dfl_port_ops_mutex);
  176. }
  177. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_add);
  178. /**
  179. * dfl_fpga_port_ops_del - remove port_ops from global list
  180. * @ops: port ops to del.
  181. */
  182. void dfl_fpga_port_ops_del(struct dfl_fpga_port_ops *ops)
  183. {
  184. mutex_lock(&dfl_port_ops_mutex);
  185. list_del(&ops->node);
  186. mutex_unlock(&dfl_port_ops_mutex);
  187. }
  188. EXPORT_SYMBOL_GPL(dfl_fpga_port_ops_del);
  189. /**
  190. * dfl_fpga_check_port_id - check the port id
  191. * @pdev: port platform device.
  192. * @pport_id: port id to compare.
  193. *
  194. * Return: 1 if port device matches with given port id, otherwise 0.
  195. */
  196. int dfl_fpga_check_port_id(struct platform_device *pdev, void *pport_id)
  197. {
  198. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  199. struct dfl_fpga_port_ops *port_ops;
  200. if (pdata->id != FEATURE_DEV_ID_UNUSED)
  201. return pdata->id == *(int *)pport_id;
  202. port_ops = dfl_fpga_port_ops_get(pdev);
  203. if (!port_ops || !port_ops->get_id)
  204. return 0;
  205. pdata->id = port_ops->get_id(pdev);
  206. dfl_fpga_port_ops_put(port_ops);
  207. return pdata->id == *(int *)pport_id;
  208. }
  209. EXPORT_SYMBOL_GPL(dfl_fpga_check_port_id);
  210. static DEFINE_IDA(dfl_device_ida);
  211. static const struct dfl_device_id *
  212. dfl_match_one_device(const struct dfl_device_id *id, struct dfl_device *ddev)
  213. {
  214. if (id->type == ddev->type && id->feature_id == ddev->feature_id)
  215. return id;
  216. return NULL;
  217. }
  218. static int dfl_bus_match(struct device *dev, struct device_driver *drv)
  219. {
  220. struct dfl_device *ddev = to_dfl_dev(dev);
  221. struct dfl_driver *ddrv = to_dfl_drv(drv);
  222. const struct dfl_device_id *id_entry;
  223. id_entry = ddrv->id_table;
  224. if (id_entry) {
  225. while (id_entry->feature_id) {
  226. if (dfl_match_one_device(id_entry, ddev)) {
  227. ddev->id_entry = id_entry;
  228. return 1;
  229. }
  230. id_entry++;
  231. }
  232. }
  233. return 0;
  234. }
  235. static int dfl_bus_probe(struct device *dev)
  236. {
  237. struct dfl_driver *ddrv = to_dfl_drv(dev->driver);
  238. struct dfl_device *ddev = to_dfl_dev(dev);
  239. return ddrv->probe(ddev);
  240. }
  241. static void dfl_bus_remove(struct device *dev)
  242. {
  243. struct dfl_driver *ddrv = to_dfl_drv(dev->driver);
  244. struct dfl_device *ddev = to_dfl_dev(dev);
  245. if (ddrv->remove)
  246. ddrv->remove(ddev);
  247. }
  248. static int dfl_bus_uevent(struct device *dev, struct kobj_uevent_env *env)
  249. {
  250. struct dfl_device *ddev = to_dfl_dev(dev);
  251. return add_uevent_var(env, "MODALIAS=dfl:t%04Xf%04X",
  252. ddev->type, ddev->feature_id);
  253. }
  254. static ssize_t
  255. type_show(struct device *dev, struct device_attribute *attr, char *buf)
  256. {
  257. struct dfl_device *ddev = to_dfl_dev(dev);
  258. return sprintf(buf, "0x%x\n", ddev->type);
  259. }
  260. static DEVICE_ATTR_RO(type);
  261. static ssize_t
  262. feature_id_show(struct device *dev, struct device_attribute *attr, char *buf)
  263. {
  264. struct dfl_device *ddev = to_dfl_dev(dev);
  265. return sprintf(buf, "0x%x\n", ddev->feature_id);
  266. }
  267. static DEVICE_ATTR_RO(feature_id);
  268. static struct attribute *dfl_dev_attrs[] = {
  269. &dev_attr_type.attr,
  270. &dev_attr_feature_id.attr,
  271. NULL,
  272. };
  273. ATTRIBUTE_GROUPS(dfl_dev);
  274. static struct bus_type dfl_bus_type = {
  275. .name = "dfl",
  276. .match = dfl_bus_match,
  277. .probe = dfl_bus_probe,
  278. .remove = dfl_bus_remove,
  279. .uevent = dfl_bus_uevent,
  280. .dev_groups = dfl_dev_groups,
  281. };
  282. static void release_dfl_dev(struct device *dev)
  283. {
  284. struct dfl_device *ddev = to_dfl_dev(dev);
  285. if (ddev->mmio_res.parent)
  286. release_resource(&ddev->mmio_res);
  287. ida_free(&dfl_device_ida, ddev->id);
  288. kfree(ddev->irqs);
  289. kfree(ddev);
  290. }
  291. static struct dfl_device *
  292. dfl_dev_add(struct dfl_feature_platform_data *pdata,
  293. struct dfl_feature *feature)
  294. {
  295. struct platform_device *pdev = pdata->dev;
  296. struct resource *parent_res;
  297. struct dfl_device *ddev;
  298. int id, i, ret;
  299. ddev = kzalloc(sizeof(*ddev), GFP_KERNEL);
  300. if (!ddev)
  301. return ERR_PTR(-ENOMEM);
  302. id = ida_alloc(&dfl_device_ida, GFP_KERNEL);
  303. if (id < 0) {
  304. dev_err(&pdev->dev, "unable to get id\n");
  305. kfree(ddev);
  306. return ERR_PTR(id);
  307. }
  308. /* freeing resources by put_device() after device_initialize() */
  309. device_initialize(&ddev->dev);
  310. ddev->dev.parent = &pdev->dev;
  311. ddev->dev.bus = &dfl_bus_type;
  312. ddev->dev.release = release_dfl_dev;
  313. ddev->id = id;
  314. ret = dev_set_name(&ddev->dev, "dfl_dev.%d", id);
  315. if (ret)
  316. goto put_dev;
  317. ddev->type = feature_dev_id_type(pdev);
  318. ddev->feature_id = feature->id;
  319. ddev->revision = feature->revision;
  320. ddev->cdev = pdata->dfl_cdev;
  321. /* add mmio resource */
  322. parent_res = &pdev->resource[feature->resource_index];
  323. ddev->mmio_res.flags = IORESOURCE_MEM;
  324. ddev->mmio_res.start = parent_res->start;
  325. ddev->mmio_res.end = parent_res->end;
  326. ddev->mmio_res.name = dev_name(&ddev->dev);
  327. ret = insert_resource(parent_res, &ddev->mmio_res);
  328. if (ret) {
  329. dev_err(&pdev->dev, "%s failed to claim resource: %pR\n",
  330. dev_name(&ddev->dev), &ddev->mmio_res);
  331. goto put_dev;
  332. }
  333. /* then add irq resource */
  334. if (feature->nr_irqs) {
  335. ddev->irqs = kcalloc(feature->nr_irqs,
  336. sizeof(*ddev->irqs), GFP_KERNEL);
  337. if (!ddev->irqs) {
  338. ret = -ENOMEM;
  339. goto put_dev;
  340. }
  341. for (i = 0; i < feature->nr_irqs; i++)
  342. ddev->irqs[i] = feature->irq_ctx[i].irq;
  343. ddev->num_irqs = feature->nr_irqs;
  344. }
  345. ret = device_add(&ddev->dev);
  346. if (ret)
  347. goto put_dev;
  348. dev_dbg(&pdev->dev, "add dfl_dev: %s\n", dev_name(&ddev->dev));
  349. return ddev;
  350. put_dev:
  351. /* calls release_dfl_dev() which does the clean up */
  352. put_device(&ddev->dev);
  353. return ERR_PTR(ret);
  354. }
  355. static void dfl_devs_remove(struct dfl_feature_platform_data *pdata)
  356. {
  357. struct dfl_feature *feature;
  358. dfl_fpga_dev_for_each_feature(pdata, feature) {
  359. if (feature->ddev) {
  360. device_unregister(&feature->ddev->dev);
  361. feature->ddev = NULL;
  362. }
  363. }
  364. }
  365. static int dfl_devs_add(struct dfl_feature_platform_data *pdata)
  366. {
  367. struct dfl_feature *feature;
  368. struct dfl_device *ddev;
  369. int ret;
  370. dfl_fpga_dev_for_each_feature(pdata, feature) {
  371. if (feature->ioaddr)
  372. continue;
  373. if (feature->ddev) {
  374. ret = -EEXIST;
  375. goto err;
  376. }
  377. ddev = dfl_dev_add(pdata, feature);
  378. if (IS_ERR(ddev)) {
  379. ret = PTR_ERR(ddev);
  380. goto err;
  381. }
  382. feature->ddev = ddev;
  383. }
  384. return 0;
  385. err:
  386. dfl_devs_remove(pdata);
  387. return ret;
  388. }
  389. int __dfl_driver_register(struct dfl_driver *dfl_drv, struct module *owner)
  390. {
  391. if (!dfl_drv || !dfl_drv->probe || !dfl_drv->id_table)
  392. return -EINVAL;
  393. dfl_drv->drv.owner = owner;
  394. dfl_drv->drv.bus = &dfl_bus_type;
  395. return driver_register(&dfl_drv->drv);
  396. }
  397. EXPORT_SYMBOL(__dfl_driver_register);
  398. void dfl_driver_unregister(struct dfl_driver *dfl_drv)
  399. {
  400. driver_unregister(&dfl_drv->drv);
  401. }
  402. EXPORT_SYMBOL(dfl_driver_unregister);
  403. #define is_header_feature(feature) ((feature)->id == FEATURE_ID_FIU_HEADER)
  404. /**
  405. * dfl_fpga_dev_feature_uinit - uinit for sub features of dfl feature device
  406. * @pdev: feature device.
  407. */
  408. void dfl_fpga_dev_feature_uinit(struct platform_device *pdev)
  409. {
  410. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  411. struct dfl_feature *feature;
  412. dfl_devs_remove(pdata);
  413. dfl_fpga_dev_for_each_feature(pdata, feature) {
  414. if (feature->ops) {
  415. if (feature->ops->uinit)
  416. feature->ops->uinit(pdev, feature);
  417. feature->ops = NULL;
  418. }
  419. }
  420. }
  421. EXPORT_SYMBOL_GPL(dfl_fpga_dev_feature_uinit);
  422. static int dfl_feature_instance_init(struct platform_device *pdev,
  423. struct dfl_feature_platform_data *pdata,
  424. struct dfl_feature *feature,
  425. struct dfl_feature_driver *drv)
  426. {
  427. void __iomem *base;
  428. int ret = 0;
  429. if (!is_header_feature(feature)) {
  430. base = devm_platform_ioremap_resource(pdev,
  431. feature->resource_index);
  432. if (IS_ERR(base)) {
  433. dev_err(&pdev->dev,
  434. "ioremap failed for feature 0x%x!\n",
  435. feature->id);
  436. return PTR_ERR(base);
  437. }
  438. feature->ioaddr = base;
  439. }
  440. if (drv->ops->init) {
  441. ret = drv->ops->init(pdev, feature);
  442. if (ret)
  443. return ret;
  444. }
  445. feature->ops = drv->ops;
  446. return ret;
  447. }
  448. static bool dfl_feature_drv_match(struct dfl_feature *feature,
  449. struct dfl_feature_driver *driver)
  450. {
  451. const struct dfl_feature_id *ids = driver->id_table;
  452. if (ids) {
  453. while (ids->id) {
  454. if (ids->id == feature->id)
  455. return true;
  456. ids++;
  457. }
  458. }
  459. return false;
  460. }
  461. /**
  462. * dfl_fpga_dev_feature_init - init for sub features of dfl feature device
  463. * @pdev: feature device.
  464. * @feature_drvs: drvs for sub features.
  465. *
  466. * This function will match sub features with given feature drvs list and
  467. * use matched drv to init related sub feature.
  468. *
  469. * Return: 0 on success, negative error code otherwise.
  470. */
  471. int dfl_fpga_dev_feature_init(struct platform_device *pdev,
  472. struct dfl_feature_driver *feature_drvs)
  473. {
  474. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  475. struct dfl_feature_driver *drv = feature_drvs;
  476. struct dfl_feature *feature;
  477. int ret;
  478. while (drv->ops) {
  479. dfl_fpga_dev_for_each_feature(pdata, feature) {
  480. if (dfl_feature_drv_match(feature, drv)) {
  481. ret = dfl_feature_instance_init(pdev, pdata,
  482. feature, drv);
  483. if (ret)
  484. goto exit;
  485. }
  486. }
  487. drv++;
  488. }
  489. ret = dfl_devs_add(pdata);
  490. if (ret)
  491. goto exit;
  492. return 0;
  493. exit:
  494. dfl_fpga_dev_feature_uinit(pdev);
  495. return ret;
  496. }
  497. EXPORT_SYMBOL_GPL(dfl_fpga_dev_feature_init);
  498. static void dfl_chardev_uinit(void)
  499. {
  500. int i;
  501. for (i = 0; i < DFL_FPGA_DEVT_MAX; i++)
  502. if (MAJOR(dfl_chrdevs[i].devt)) {
  503. unregister_chrdev_region(dfl_chrdevs[i].devt,
  504. MINORMASK + 1);
  505. dfl_chrdevs[i].devt = MKDEV(0, 0);
  506. }
  507. }
  508. static int dfl_chardev_init(void)
  509. {
  510. int i, ret;
  511. for (i = 0; i < DFL_FPGA_DEVT_MAX; i++) {
  512. ret = alloc_chrdev_region(&dfl_chrdevs[i].devt, 0,
  513. MINORMASK + 1, dfl_chrdevs[i].name);
  514. if (ret)
  515. goto exit;
  516. }
  517. return 0;
  518. exit:
  519. dfl_chardev_uinit();
  520. return ret;
  521. }
  522. static dev_t dfl_get_devt(enum dfl_fpga_devt_type type, int id)
  523. {
  524. if (type >= DFL_FPGA_DEVT_MAX)
  525. return 0;
  526. return MKDEV(MAJOR(dfl_chrdevs[type].devt), id);
  527. }
  528. /**
  529. * dfl_fpga_dev_ops_register - register cdev ops for feature dev
  530. *
  531. * @pdev: feature dev.
  532. * @fops: file operations for feature dev's cdev.
  533. * @owner: owning module/driver.
  534. *
  535. * Return: 0 on success, negative error code otherwise.
  536. */
  537. int dfl_fpga_dev_ops_register(struct platform_device *pdev,
  538. const struct file_operations *fops,
  539. struct module *owner)
  540. {
  541. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  542. cdev_init(&pdata->cdev, fops);
  543. pdata->cdev.owner = owner;
  544. /*
  545. * set parent to the feature device so that its refcount is
  546. * decreased after the last refcount of cdev is gone, that
  547. * makes sure the feature device is valid during device
  548. * file's life-cycle.
  549. */
  550. pdata->cdev.kobj.parent = &pdev->dev.kobj;
  551. return cdev_add(&pdata->cdev, pdev->dev.devt, 1);
  552. }
  553. EXPORT_SYMBOL_GPL(dfl_fpga_dev_ops_register);
  554. /**
  555. * dfl_fpga_dev_ops_unregister - unregister cdev ops for feature dev
  556. * @pdev: feature dev.
  557. */
  558. void dfl_fpga_dev_ops_unregister(struct platform_device *pdev)
  559. {
  560. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  561. cdev_del(&pdata->cdev);
  562. }
  563. EXPORT_SYMBOL_GPL(dfl_fpga_dev_ops_unregister);
  564. /**
  565. * struct build_feature_devs_info - info collected during feature dev build.
  566. *
  567. * @dev: device to enumerate.
  568. * @cdev: the container device for all feature devices.
  569. * @nr_irqs: number of irqs for all feature devices.
  570. * @irq_table: Linux IRQ numbers for all irqs, indexed by local irq index of
  571. * this device.
  572. * @feature_dev: current feature device.
  573. * @ioaddr: header register region address of current FIU in enumeration.
  574. * @start: register resource start of current FIU.
  575. * @len: max register resource length of current FIU.
  576. * @sub_features: a sub features linked list for feature device in enumeration.
  577. * @feature_num: number of sub features for feature device in enumeration.
  578. */
  579. struct build_feature_devs_info {
  580. struct device *dev;
  581. struct dfl_fpga_cdev *cdev;
  582. unsigned int nr_irqs;
  583. int *irq_table;
  584. struct platform_device *feature_dev;
  585. void __iomem *ioaddr;
  586. resource_size_t start;
  587. resource_size_t len;
  588. struct list_head sub_features;
  589. int feature_num;
  590. };
  591. /**
  592. * struct dfl_feature_info - sub feature info collected during feature dev build
  593. *
  594. * @fid: id of this sub feature.
  595. * @mmio_res: mmio resource of this sub feature.
  596. * @ioaddr: mapped base address of mmio resource.
  597. * @node: node in sub_features linked list.
  598. * @irq_base: start of irq index in this sub feature.
  599. * @nr_irqs: number of irqs of this sub feature.
  600. */
  601. struct dfl_feature_info {
  602. u16 fid;
  603. u8 revision;
  604. struct resource mmio_res;
  605. void __iomem *ioaddr;
  606. struct list_head node;
  607. unsigned int irq_base;
  608. unsigned int nr_irqs;
  609. };
  610. static void dfl_fpga_cdev_add_port_dev(struct dfl_fpga_cdev *cdev,
  611. struct platform_device *port)
  612. {
  613. struct dfl_feature_platform_data *pdata = dev_get_platdata(&port->dev);
  614. mutex_lock(&cdev->lock);
  615. list_add(&pdata->node, &cdev->port_dev_list);
  616. get_device(&pdata->dev->dev);
  617. mutex_unlock(&cdev->lock);
  618. }
  619. /*
  620. * register current feature device, it is called when we need to switch to
  621. * another feature parsing or we have parsed all features on given device
  622. * feature list.
  623. */
  624. static int build_info_commit_dev(struct build_feature_devs_info *binfo)
  625. {
  626. struct platform_device *fdev = binfo->feature_dev;
  627. struct dfl_feature_platform_data *pdata;
  628. struct dfl_feature_info *finfo, *p;
  629. enum dfl_id_type type;
  630. int ret, index = 0, res_idx = 0;
  631. type = feature_dev_id_type(fdev);
  632. if (WARN_ON_ONCE(type >= DFL_ID_MAX))
  633. return -EINVAL;
  634. /*
  635. * we do not need to care for the memory which is associated with
  636. * the platform device. After calling platform_device_unregister(),
  637. * it will be automatically freed by device's release() callback,
  638. * platform_device_release().
  639. */
  640. pdata = kzalloc(struct_size(pdata, features, binfo->feature_num), GFP_KERNEL);
  641. if (!pdata)
  642. return -ENOMEM;
  643. pdata->dev = fdev;
  644. pdata->num = binfo->feature_num;
  645. pdata->dfl_cdev = binfo->cdev;
  646. pdata->id = FEATURE_DEV_ID_UNUSED;
  647. mutex_init(&pdata->lock);
  648. lockdep_set_class_and_name(&pdata->lock, &dfl_pdata_keys[type],
  649. dfl_pdata_key_strings[type]);
  650. /*
  651. * the count should be initialized to 0 to make sure
  652. *__fpga_port_enable() following __fpga_port_disable()
  653. * works properly for port device.
  654. * and it should always be 0 for fme device.
  655. */
  656. WARN_ON(pdata->disable_count);
  657. fdev->dev.platform_data = pdata;
  658. /* each sub feature has one MMIO resource */
  659. fdev->num_resources = binfo->feature_num;
  660. fdev->resource = kcalloc(binfo->feature_num, sizeof(*fdev->resource),
  661. GFP_KERNEL);
  662. if (!fdev->resource)
  663. return -ENOMEM;
  664. /* fill features and resource information for feature dev */
  665. list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) {
  666. struct dfl_feature *feature = &pdata->features[index++];
  667. struct dfl_feature_irq_ctx *ctx;
  668. unsigned int i;
  669. /* save resource information for each feature */
  670. feature->dev = fdev;
  671. feature->id = finfo->fid;
  672. feature->revision = finfo->revision;
  673. /*
  674. * the FIU header feature has some fundamental functions (sriov
  675. * set, port enable/disable) needed for the dfl bus device and
  676. * other sub features. So its mmio resource should be mapped by
  677. * DFL bus device. And we should not assign it to feature
  678. * devices (dfl-fme/afu) again.
  679. */
  680. if (is_header_feature(feature)) {
  681. feature->resource_index = -1;
  682. feature->ioaddr =
  683. devm_ioremap_resource(binfo->dev,
  684. &finfo->mmio_res);
  685. if (IS_ERR(feature->ioaddr))
  686. return PTR_ERR(feature->ioaddr);
  687. } else {
  688. feature->resource_index = res_idx;
  689. fdev->resource[res_idx++] = finfo->mmio_res;
  690. }
  691. if (finfo->nr_irqs) {
  692. ctx = devm_kcalloc(binfo->dev, finfo->nr_irqs,
  693. sizeof(*ctx), GFP_KERNEL);
  694. if (!ctx)
  695. return -ENOMEM;
  696. for (i = 0; i < finfo->nr_irqs; i++)
  697. ctx[i].irq =
  698. binfo->irq_table[finfo->irq_base + i];
  699. feature->irq_ctx = ctx;
  700. feature->nr_irqs = finfo->nr_irqs;
  701. }
  702. list_del(&finfo->node);
  703. kfree(finfo);
  704. }
  705. ret = platform_device_add(binfo->feature_dev);
  706. if (!ret) {
  707. if (type == PORT_ID)
  708. dfl_fpga_cdev_add_port_dev(binfo->cdev,
  709. binfo->feature_dev);
  710. else
  711. binfo->cdev->fme_dev =
  712. get_device(&binfo->feature_dev->dev);
  713. /*
  714. * reset it to avoid build_info_free() freeing their resource.
  715. *
  716. * The resource of successfully registered feature devices
  717. * will be freed by platform_device_unregister(). See the
  718. * comments in build_info_create_dev().
  719. */
  720. binfo->feature_dev = NULL;
  721. }
  722. return ret;
  723. }
  724. static int
  725. build_info_create_dev(struct build_feature_devs_info *binfo,
  726. enum dfl_id_type type)
  727. {
  728. struct platform_device *fdev;
  729. if (type >= DFL_ID_MAX)
  730. return -EINVAL;
  731. /*
  732. * we use -ENODEV as the initialization indicator which indicates
  733. * whether the id need to be reclaimed
  734. */
  735. fdev = platform_device_alloc(dfl_devs[type].name, -ENODEV);
  736. if (!fdev)
  737. return -ENOMEM;
  738. binfo->feature_dev = fdev;
  739. binfo->feature_num = 0;
  740. INIT_LIST_HEAD(&binfo->sub_features);
  741. fdev->id = dfl_id_alloc(type, &fdev->dev);
  742. if (fdev->id < 0)
  743. return fdev->id;
  744. fdev->dev.parent = &binfo->cdev->region->dev;
  745. fdev->dev.devt = dfl_get_devt(dfl_devs[type].devt_type, fdev->id);
  746. return 0;
  747. }
  748. static void build_info_free(struct build_feature_devs_info *binfo)
  749. {
  750. struct dfl_feature_info *finfo, *p;
  751. /*
  752. * it is a valid id, free it. See comments in
  753. * build_info_create_dev()
  754. */
  755. if (binfo->feature_dev && binfo->feature_dev->id >= 0) {
  756. dfl_id_free(feature_dev_id_type(binfo->feature_dev),
  757. binfo->feature_dev->id);
  758. list_for_each_entry_safe(finfo, p, &binfo->sub_features, node) {
  759. list_del(&finfo->node);
  760. kfree(finfo);
  761. }
  762. }
  763. platform_device_put(binfo->feature_dev);
  764. devm_kfree(binfo->dev, binfo);
  765. }
  766. static inline u32 feature_size(u64 value)
  767. {
  768. u32 ofst = FIELD_GET(DFH_NEXT_HDR_OFST, value);
  769. /* workaround for private features with invalid size, use 4K instead */
  770. return ofst ? ofst : 4096;
  771. }
  772. static u16 feature_id(u64 value)
  773. {
  774. u16 id = FIELD_GET(DFH_ID, value);
  775. u8 type = FIELD_GET(DFH_TYPE, value);
  776. if (type == DFH_TYPE_FIU)
  777. return FEATURE_ID_FIU_HEADER;
  778. else if (type == DFH_TYPE_PRIVATE)
  779. return id;
  780. else if (type == DFH_TYPE_AFU)
  781. return FEATURE_ID_AFU;
  782. WARN_ON(1);
  783. return 0;
  784. }
  785. static int parse_feature_irqs(struct build_feature_devs_info *binfo,
  786. resource_size_t ofst, u16 fid,
  787. unsigned int *irq_base, unsigned int *nr_irqs)
  788. {
  789. void __iomem *base = binfo->ioaddr + ofst;
  790. unsigned int i, ibase, inr = 0;
  791. enum dfl_id_type type;
  792. int virq;
  793. u64 v;
  794. type = feature_dev_id_type(binfo->feature_dev);
  795. /*
  796. * Ideally DFL framework should only read info from DFL header, but
  797. * current version DFL only provides mmio resources information for
  798. * each feature in DFL Header, no field for interrupt resources.
  799. * Interrupt resource information is provided by specific mmio
  800. * registers of each private feature which supports interrupt. So in
  801. * order to parse and assign irq resources, DFL framework has to look
  802. * into specific capability registers of these private features.
  803. *
  804. * Once future DFL version supports generic interrupt resource
  805. * information in common DFL headers, the generic interrupt parsing
  806. * code will be added. But in order to be compatible to old version
  807. * DFL, the driver may still fall back to these quirks.
  808. */
  809. if (type == PORT_ID) {
  810. switch (fid) {
  811. case PORT_FEATURE_ID_UINT:
  812. v = readq(base + PORT_UINT_CAP);
  813. ibase = FIELD_GET(PORT_UINT_CAP_FST_VECT, v);
  814. inr = FIELD_GET(PORT_UINT_CAP_INT_NUM, v);
  815. break;
  816. case PORT_FEATURE_ID_ERROR:
  817. v = readq(base + PORT_ERROR_CAP);
  818. ibase = FIELD_GET(PORT_ERROR_CAP_INT_VECT, v);
  819. inr = FIELD_GET(PORT_ERROR_CAP_SUPP_INT, v);
  820. break;
  821. }
  822. } else if (type == FME_ID) {
  823. if (fid == FME_FEATURE_ID_GLOBAL_ERR) {
  824. v = readq(base + FME_ERROR_CAP);
  825. ibase = FIELD_GET(FME_ERROR_CAP_INT_VECT, v);
  826. inr = FIELD_GET(FME_ERROR_CAP_SUPP_INT, v);
  827. }
  828. }
  829. if (!inr) {
  830. *irq_base = 0;
  831. *nr_irqs = 0;
  832. return 0;
  833. }
  834. dev_dbg(binfo->dev, "feature: 0x%x, irq_base: %u, nr_irqs: %u\n",
  835. fid, ibase, inr);
  836. if (ibase + inr > binfo->nr_irqs) {
  837. dev_err(binfo->dev,
  838. "Invalid interrupt number in feature 0x%x\n", fid);
  839. return -EINVAL;
  840. }
  841. for (i = 0; i < inr; i++) {
  842. virq = binfo->irq_table[ibase + i];
  843. if (virq < 0 || virq > NR_IRQS) {
  844. dev_err(binfo->dev,
  845. "Invalid irq table entry for feature 0x%x\n",
  846. fid);
  847. return -EINVAL;
  848. }
  849. }
  850. *irq_base = ibase;
  851. *nr_irqs = inr;
  852. return 0;
  853. }
  854. /*
  855. * when create sub feature instances, for private features, it doesn't need
  856. * to provide resource size and feature id as they could be read from DFH
  857. * register. For afu sub feature, its register region only contains user
  858. * defined registers, so never trust any information from it, just use the
  859. * resource size information provided by its parent FIU.
  860. */
  861. static int
  862. create_feature_instance(struct build_feature_devs_info *binfo,
  863. resource_size_t ofst, resource_size_t size, u16 fid)
  864. {
  865. unsigned int irq_base, nr_irqs;
  866. struct dfl_feature_info *finfo;
  867. u8 revision = 0;
  868. int ret;
  869. u64 v;
  870. if (fid != FEATURE_ID_AFU) {
  871. v = readq(binfo->ioaddr + ofst);
  872. revision = FIELD_GET(DFH_REVISION, v);
  873. /* read feature size and id if inputs are invalid */
  874. size = size ? size : feature_size(v);
  875. fid = fid ? fid : feature_id(v);
  876. }
  877. if (binfo->len - ofst < size)
  878. return -EINVAL;
  879. ret = parse_feature_irqs(binfo, ofst, fid, &irq_base, &nr_irqs);
  880. if (ret)
  881. return ret;
  882. finfo = kzalloc(sizeof(*finfo), GFP_KERNEL);
  883. if (!finfo)
  884. return -ENOMEM;
  885. finfo->fid = fid;
  886. finfo->revision = revision;
  887. finfo->mmio_res.start = binfo->start + ofst;
  888. finfo->mmio_res.end = finfo->mmio_res.start + size - 1;
  889. finfo->mmio_res.flags = IORESOURCE_MEM;
  890. finfo->irq_base = irq_base;
  891. finfo->nr_irqs = nr_irqs;
  892. list_add_tail(&finfo->node, &binfo->sub_features);
  893. binfo->feature_num++;
  894. return 0;
  895. }
  896. static int parse_feature_port_afu(struct build_feature_devs_info *binfo,
  897. resource_size_t ofst)
  898. {
  899. u64 v = readq(binfo->ioaddr + PORT_HDR_CAP);
  900. u32 size = FIELD_GET(PORT_CAP_MMIO_SIZE, v) << 10;
  901. WARN_ON(!size);
  902. return create_feature_instance(binfo, ofst, size, FEATURE_ID_AFU);
  903. }
  904. #define is_feature_dev_detected(binfo) (!!(binfo)->feature_dev)
  905. static int parse_feature_afu(struct build_feature_devs_info *binfo,
  906. resource_size_t ofst)
  907. {
  908. if (!is_feature_dev_detected(binfo)) {
  909. dev_err(binfo->dev, "this AFU does not belong to any FIU.\n");
  910. return -EINVAL;
  911. }
  912. switch (feature_dev_id_type(binfo->feature_dev)) {
  913. case PORT_ID:
  914. return parse_feature_port_afu(binfo, ofst);
  915. default:
  916. dev_info(binfo->dev, "AFU belonging to FIU %s is not supported yet.\n",
  917. binfo->feature_dev->name);
  918. }
  919. return 0;
  920. }
  921. static int build_info_prepare(struct build_feature_devs_info *binfo,
  922. resource_size_t start, resource_size_t len)
  923. {
  924. struct device *dev = binfo->dev;
  925. void __iomem *ioaddr;
  926. if (!devm_request_mem_region(dev, start, len, dev_name(dev))) {
  927. dev_err(dev, "request region fail, start:%pa, len:%pa\n",
  928. &start, &len);
  929. return -EBUSY;
  930. }
  931. ioaddr = devm_ioremap(dev, start, len);
  932. if (!ioaddr) {
  933. dev_err(dev, "ioremap region fail, start:%pa, len:%pa\n",
  934. &start, &len);
  935. return -ENOMEM;
  936. }
  937. binfo->start = start;
  938. binfo->len = len;
  939. binfo->ioaddr = ioaddr;
  940. return 0;
  941. }
  942. static void build_info_complete(struct build_feature_devs_info *binfo)
  943. {
  944. devm_iounmap(binfo->dev, binfo->ioaddr);
  945. devm_release_mem_region(binfo->dev, binfo->start, binfo->len);
  946. }
  947. static int parse_feature_fiu(struct build_feature_devs_info *binfo,
  948. resource_size_t ofst)
  949. {
  950. int ret = 0;
  951. u32 offset;
  952. u16 id;
  953. u64 v;
  954. if (is_feature_dev_detected(binfo)) {
  955. build_info_complete(binfo);
  956. ret = build_info_commit_dev(binfo);
  957. if (ret)
  958. return ret;
  959. ret = build_info_prepare(binfo, binfo->start + ofst,
  960. binfo->len - ofst);
  961. if (ret)
  962. return ret;
  963. }
  964. v = readq(binfo->ioaddr + DFH);
  965. id = FIELD_GET(DFH_ID, v);
  966. /* create platform device for dfl feature dev */
  967. ret = build_info_create_dev(binfo, dfh_id_to_type(id));
  968. if (ret)
  969. return ret;
  970. ret = create_feature_instance(binfo, 0, 0, 0);
  971. if (ret)
  972. return ret;
  973. /*
  974. * find and parse FIU's child AFU via its NEXT_AFU register.
  975. * please note that only Port has valid NEXT_AFU pointer per spec.
  976. */
  977. v = readq(binfo->ioaddr + NEXT_AFU);
  978. offset = FIELD_GET(NEXT_AFU_NEXT_DFH_OFST, v);
  979. if (offset)
  980. return parse_feature_afu(binfo, offset);
  981. dev_dbg(binfo->dev, "No AFUs detected on FIU %d\n", id);
  982. return ret;
  983. }
  984. static int parse_feature_private(struct build_feature_devs_info *binfo,
  985. resource_size_t ofst)
  986. {
  987. if (!is_feature_dev_detected(binfo)) {
  988. dev_err(binfo->dev, "the private feature 0x%x does not belong to any AFU.\n",
  989. feature_id(readq(binfo->ioaddr + ofst)));
  990. return -EINVAL;
  991. }
  992. return create_feature_instance(binfo, ofst, 0, 0);
  993. }
  994. /**
  995. * parse_feature - parse a feature on given device feature list
  996. *
  997. * @binfo: build feature devices information.
  998. * @ofst: offset to current FIU header
  999. */
  1000. static int parse_feature(struct build_feature_devs_info *binfo,
  1001. resource_size_t ofst)
  1002. {
  1003. u64 v;
  1004. u32 type;
  1005. v = readq(binfo->ioaddr + ofst + DFH);
  1006. type = FIELD_GET(DFH_TYPE, v);
  1007. switch (type) {
  1008. case DFH_TYPE_AFU:
  1009. return parse_feature_afu(binfo, ofst);
  1010. case DFH_TYPE_PRIVATE:
  1011. return parse_feature_private(binfo, ofst);
  1012. case DFH_TYPE_FIU:
  1013. return parse_feature_fiu(binfo, ofst);
  1014. default:
  1015. dev_info(binfo->dev,
  1016. "Feature Type %x is not supported.\n", type);
  1017. }
  1018. return 0;
  1019. }
  1020. static int parse_feature_list(struct build_feature_devs_info *binfo,
  1021. resource_size_t start, resource_size_t len)
  1022. {
  1023. resource_size_t end = start + len;
  1024. int ret = 0;
  1025. u32 ofst = 0;
  1026. u64 v;
  1027. ret = build_info_prepare(binfo, start, len);
  1028. if (ret)
  1029. return ret;
  1030. /* walk through the device feature list via DFH's next DFH pointer. */
  1031. for (; start < end; start += ofst) {
  1032. if (end - start < DFH_SIZE) {
  1033. dev_err(binfo->dev, "The region is too small to contain a feature.\n");
  1034. return -EINVAL;
  1035. }
  1036. ret = parse_feature(binfo, start - binfo->start);
  1037. if (ret)
  1038. return ret;
  1039. v = readq(binfo->ioaddr + start - binfo->start + DFH);
  1040. ofst = FIELD_GET(DFH_NEXT_HDR_OFST, v);
  1041. /* stop parsing if EOL(End of List) is set or offset is 0 */
  1042. if ((v & DFH_EOL) || !ofst)
  1043. break;
  1044. }
  1045. /* commit current feature device when reach the end of list */
  1046. build_info_complete(binfo);
  1047. if (is_feature_dev_detected(binfo))
  1048. ret = build_info_commit_dev(binfo);
  1049. return ret;
  1050. }
  1051. struct dfl_fpga_enum_info *dfl_fpga_enum_info_alloc(struct device *dev)
  1052. {
  1053. struct dfl_fpga_enum_info *info;
  1054. get_device(dev);
  1055. info = devm_kzalloc(dev, sizeof(*info), GFP_KERNEL);
  1056. if (!info) {
  1057. put_device(dev);
  1058. return NULL;
  1059. }
  1060. info->dev = dev;
  1061. INIT_LIST_HEAD(&info->dfls);
  1062. return info;
  1063. }
  1064. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_alloc);
  1065. void dfl_fpga_enum_info_free(struct dfl_fpga_enum_info *info)
  1066. {
  1067. struct dfl_fpga_enum_dfl *tmp, *dfl;
  1068. struct device *dev;
  1069. if (!info)
  1070. return;
  1071. dev = info->dev;
  1072. /* remove all device feature lists in the list. */
  1073. list_for_each_entry_safe(dfl, tmp, &info->dfls, node) {
  1074. list_del(&dfl->node);
  1075. devm_kfree(dev, dfl);
  1076. }
  1077. /* remove irq table */
  1078. if (info->irq_table)
  1079. devm_kfree(dev, info->irq_table);
  1080. devm_kfree(dev, info);
  1081. put_device(dev);
  1082. }
  1083. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_free);
  1084. /**
  1085. * dfl_fpga_enum_info_add_dfl - add info of a device feature list to enum info
  1086. *
  1087. * @info: ptr to dfl_fpga_enum_info
  1088. * @start: mmio resource address of the device feature list.
  1089. * @len: mmio resource length of the device feature list.
  1090. *
  1091. * One FPGA device may have one or more Device Feature Lists (DFLs), use this
  1092. * function to add information of each DFL to common data structure for next
  1093. * step enumeration.
  1094. *
  1095. * Return: 0 on success, negative error code otherwise.
  1096. */
  1097. int dfl_fpga_enum_info_add_dfl(struct dfl_fpga_enum_info *info,
  1098. resource_size_t start, resource_size_t len)
  1099. {
  1100. struct dfl_fpga_enum_dfl *dfl;
  1101. dfl = devm_kzalloc(info->dev, sizeof(*dfl), GFP_KERNEL);
  1102. if (!dfl)
  1103. return -ENOMEM;
  1104. dfl->start = start;
  1105. dfl->len = len;
  1106. list_add_tail(&dfl->node, &info->dfls);
  1107. return 0;
  1108. }
  1109. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_add_dfl);
  1110. /**
  1111. * dfl_fpga_enum_info_add_irq - add irq table to enum info
  1112. *
  1113. * @info: ptr to dfl_fpga_enum_info
  1114. * @nr_irqs: number of irqs of the DFL fpga device to be enumerated.
  1115. * @irq_table: Linux IRQ numbers for all irqs, indexed by local irq index of
  1116. * this device.
  1117. *
  1118. * One FPGA device may have several interrupts. This function adds irq
  1119. * information of the DFL fpga device to enum info for next step enumeration.
  1120. * This function should be called before dfl_fpga_feature_devs_enumerate().
  1121. * As we only support one irq domain for all DFLs in the same enum info, adding
  1122. * irq table a second time for the same enum info will return error.
  1123. *
  1124. * If we need to enumerate DFLs which belong to different irq domains, we
  1125. * should fill more enum info and enumerate them one by one.
  1126. *
  1127. * Return: 0 on success, negative error code otherwise.
  1128. */
  1129. int dfl_fpga_enum_info_add_irq(struct dfl_fpga_enum_info *info,
  1130. unsigned int nr_irqs, int *irq_table)
  1131. {
  1132. if (!nr_irqs || !irq_table)
  1133. return -EINVAL;
  1134. if (info->irq_table)
  1135. return -EEXIST;
  1136. info->irq_table = devm_kmemdup(info->dev, irq_table,
  1137. sizeof(int) * nr_irqs, GFP_KERNEL);
  1138. if (!info->irq_table)
  1139. return -ENOMEM;
  1140. info->nr_irqs = nr_irqs;
  1141. return 0;
  1142. }
  1143. EXPORT_SYMBOL_GPL(dfl_fpga_enum_info_add_irq);
  1144. static int remove_feature_dev(struct device *dev, void *data)
  1145. {
  1146. struct platform_device *pdev = to_platform_device(dev);
  1147. enum dfl_id_type type = feature_dev_id_type(pdev);
  1148. int id = pdev->id;
  1149. platform_device_unregister(pdev);
  1150. dfl_id_free(type, id);
  1151. return 0;
  1152. }
  1153. static void remove_feature_devs(struct dfl_fpga_cdev *cdev)
  1154. {
  1155. device_for_each_child(&cdev->region->dev, NULL, remove_feature_dev);
  1156. }
  1157. /**
  1158. * dfl_fpga_feature_devs_enumerate - enumerate feature devices
  1159. * @info: information for enumeration.
  1160. *
  1161. * This function creates a container device (base FPGA region), enumerates
  1162. * feature devices based on the enumeration info and creates platform devices
  1163. * under the container device.
  1164. *
  1165. * Return: dfl_fpga_cdev struct on success, -errno on failure
  1166. */
  1167. struct dfl_fpga_cdev *
  1168. dfl_fpga_feature_devs_enumerate(struct dfl_fpga_enum_info *info)
  1169. {
  1170. struct build_feature_devs_info *binfo;
  1171. struct dfl_fpga_enum_dfl *dfl;
  1172. struct dfl_fpga_cdev *cdev;
  1173. int ret = 0;
  1174. if (!info->dev)
  1175. return ERR_PTR(-ENODEV);
  1176. cdev = devm_kzalloc(info->dev, sizeof(*cdev), GFP_KERNEL);
  1177. if (!cdev)
  1178. return ERR_PTR(-ENOMEM);
  1179. cdev->parent = info->dev;
  1180. mutex_init(&cdev->lock);
  1181. INIT_LIST_HEAD(&cdev->port_dev_list);
  1182. cdev->region = fpga_region_register(info->dev, NULL, NULL);
  1183. if (IS_ERR(cdev->region)) {
  1184. ret = PTR_ERR(cdev->region);
  1185. goto free_cdev_exit;
  1186. }
  1187. /* create and init build info for enumeration */
  1188. binfo = devm_kzalloc(info->dev, sizeof(*binfo), GFP_KERNEL);
  1189. if (!binfo) {
  1190. ret = -ENOMEM;
  1191. goto unregister_region_exit;
  1192. }
  1193. binfo->dev = info->dev;
  1194. binfo->cdev = cdev;
  1195. binfo->nr_irqs = info->nr_irqs;
  1196. if (info->nr_irqs)
  1197. binfo->irq_table = info->irq_table;
  1198. /*
  1199. * start enumeration for all feature devices based on Device Feature
  1200. * Lists.
  1201. */
  1202. list_for_each_entry(dfl, &info->dfls, node) {
  1203. ret = parse_feature_list(binfo, dfl->start, dfl->len);
  1204. if (ret) {
  1205. remove_feature_devs(cdev);
  1206. build_info_free(binfo);
  1207. goto unregister_region_exit;
  1208. }
  1209. }
  1210. build_info_free(binfo);
  1211. return cdev;
  1212. unregister_region_exit:
  1213. fpga_region_unregister(cdev->region);
  1214. free_cdev_exit:
  1215. devm_kfree(info->dev, cdev);
  1216. return ERR_PTR(ret);
  1217. }
  1218. EXPORT_SYMBOL_GPL(dfl_fpga_feature_devs_enumerate);
  1219. /**
  1220. * dfl_fpga_feature_devs_remove - remove all feature devices
  1221. * @cdev: fpga container device.
  1222. *
  1223. * Remove the container device and all feature devices under given container
  1224. * devices.
  1225. */
  1226. void dfl_fpga_feature_devs_remove(struct dfl_fpga_cdev *cdev)
  1227. {
  1228. struct dfl_feature_platform_data *pdata, *ptmp;
  1229. mutex_lock(&cdev->lock);
  1230. if (cdev->fme_dev)
  1231. put_device(cdev->fme_dev);
  1232. list_for_each_entry_safe(pdata, ptmp, &cdev->port_dev_list, node) {
  1233. struct platform_device *port_dev = pdata->dev;
  1234. /* remove released ports */
  1235. if (!device_is_registered(&port_dev->dev)) {
  1236. dfl_id_free(feature_dev_id_type(port_dev),
  1237. port_dev->id);
  1238. platform_device_put(port_dev);
  1239. }
  1240. list_del(&pdata->node);
  1241. put_device(&port_dev->dev);
  1242. }
  1243. mutex_unlock(&cdev->lock);
  1244. remove_feature_devs(cdev);
  1245. fpga_region_unregister(cdev->region);
  1246. devm_kfree(cdev->parent, cdev);
  1247. }
  1248. EXPORT_SYMBOL_GPL(dfl_fpga_feature_devs_remove);
  1249. /**
  1250. * __dfl_fpga_cdev_find_port - find a port under given container device
  1251. *
  1252. * @cdev: container device
  1253. * @data: data passed to match function
  1254. * @match: match function used to find specific port from the port device list
  1255. *
  1256. * Find a port device under container device. This function needs to be
  1257. * invoked with lock held.
  1258. *
  1259. * Return: pointer to port's platform device if successful, NULL otherwise.
  1260. *
  1261. * NOTE: you will need to drop the device reference with put_device() after use.
  1262. */
  1263. struct platform_device *
  1264. __dfl_fpga_cdev_find_port(struct dfl_fpga_cdev *cdev, void *data,
  1265. int (*match)(struct platform_device *, void *))
  1266. {
  1267. struct dfl_feature_platform_data *pdata;
  1268. struct platform_device *port_dev;
  1269. list_for_each_entry(pdata, &cdev->port_dev_list, node) {
  1270. port_dev = pdata->dev;
  1271. if (match(port_dev, data) && get_device(&port_dev->dev))
  1272. return port_dev;
  1273. }
  1274. return NULL;
  1275. }
  1276. EXPORT_SYMBOL_GPL(__dfl_fpga_cdev_find_port);
  1277. static int __init dfl_fpga_init(void)
  1278. {
  1279. int ret;
  1280. ret = bus_register(&dfl_bus_type);
  1281. if (ret)
  1282. return ret;
  1283. dfl_ids_init();
  1284. ret = dfl_chardev_init();
  1285. if (ret) {
  1286. dfl_ids_destroy();
  1287. bus_unregister(&dfl_bus_type);
  1288. }
  1289. return ret;
  1290. }
  1291. /**
  1292. * dfl_fpga_cdev_release_port - release a port platform device
  1293. *
  1294. * @cdev: parent container device.
  1295. * @port_id: id of the port platform device.
  1296. *
  1297. * This function allows user to release a port platform device. This is a
  1298. * mandatory step before turn a port from PF into VF for SRIOV support.
  1299. *
  1300. * Return: 0 on success, negative error code otherwise.
  1301. */
  1302. int dfl_fpga_cdev_release_port(struct dfl_fpga_cdev *cdev, int port_id)
  1303. {
  1304. struct dfl_feature_platform_data *pdata;
  1305. struct platform_device *port_pdev;
  1306. int ret = -ENODEV;
  1307. mutex_lock(&cdev->lock);
  1308. port_pdev = __dfl_fpga_cdev_find_port(cdev, &port_id,
  1309. dfl_fpga_check_port_id);
  1310. if (!port_pdev)
  1311. goto unlock_exit;
  1312. if (!device_is_registered(&port_pdev->dev)) {
  1313. ret = -EBUSY;
  1314. goto put_dev_exit;
  1315. }
  1316. pdata = dev_get_platdata(&port_pdev->dev);
  1317. mutex_lock(&pdata->lock);
  1318. ret = dfl_feature_dev_use_begin(pdata, true);
  1319. mutex_unlock(&pdata->lock);
  1320. if (ret)
  1321. goto put_dev_exit;
  1322. platform_device_del(port_pdev);
  1323. cdev->released_port_num++;
  1324. put_dev_exit:
  1325. put_device(&port_pdev->dev);
  1326. unlock_exit:
  1327. mutex_unlock(&cdev->lock);
  1328. return ret;
  1329. }
  1330. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_release_port);
  1331. /**
  1332. * dfl_fpga_cdev_assign_port - assign a port platform device back
  1333. *
  1334. * @cdev: parent container device.
  1335. * @port_id: id of the port platform device.
  1336. *
  1337. * This function allows user to assign a port platform device back. This is
  1338. * a mandatory step after disable SRIOV support.
  1339. *
  1340. * Return: 0 on success, negative error code otherwise.
  1341. */
  1342. int dfl_fpga_cdev_assign_port(struct dfl_fpga_cdev *cdev, int port_id)
  1343. {
  1344. struct dfl_feature_platform_data *pdata;
  1345. struct platform_device *port_pdev;
  1346. int ret = -ENODEV;
  1347. mutex_lock(&cdev->lock);
  1348. port_pdev = __dfl_fpga_cdev_find_port(cdev, &port_id,
  1349. dfl_fpga_check_port_id);
  1350. if (!port_pdev)
  1351. goto unlock_exit;
  1352. if (device_is_registered(&port_pdev->dev)) {
  1353. ret = -EBUSY;
  1354. goto put_dev_exit;
  1355. }
  1356. ret = platform_device_add(port_pdev);
  1357. if (ret)
  1358. goto put_dev_exit;
  1359. pdata = dev_get_platdata(&port_pdev->dev);
  1360. mutex_lock(&pdata->lock);
  1361. dfl_feature_dev_use_end(pdata);
  1362. mutex_unlock(&pdata->lock);
  1363. cdev->released_port_num--;
  1364. put_dev_exit:
  1365. put_device(&port_pdev->dev);
  1366. unlock_exit:
  1367. mutex_unlock(&cdev->lock);
  1368. return ret;
  1369. }
  1370. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_assign_port);
  1371. static void config_port_access_mode(struct device *fme_dev, int port_id,
  1372. bool is_vf)
  1373. {
  1374. void __iomem *base;
  1375. u64 v;
  1376. base = dfl_get_feature_ioaddr_by_id(fme_dev, FME_FEATURE_ID_HEADER);
  1377. v = readq(base + FME_HDR_PORT_OFST(port_id));
  1378. v &= ~FME_PORT_OFST_ACC_CTRL;
  1379. v |= FIELD_PREP(FME_PORT_OFST_ACC_CTRL,
  1380. is_vf ? FME_PORT_OFST_ACC_VF : FME_PORT_OFST_ACC_PF);
  1381. writeq(v, base + FME_HDR_PORT_OFST(port_id));
  1382. }
  1383. #define config_port_vf_mode(dev, id) config_port_access_mode(dev, id, true)
  1384. #define config_port_pf_mode(dev, id) config_port_access_mode(dev, id, false)
  1385. /**
  1386. * dfl_fpga_cdev_config_ports_pf - configure ports to PF access mode
  1387. *
  1388. * @cdev: parent container device.
  1389. *
  1390. * This function is needed in sriov configuration routine. It could be used to
  1391. * configure the all released ports from VF access mode to PF.
  1392. */
  1393. void dfl_fpga_cdev_config_ports_pf(struct dfl_fpga_cdev *cdev)
  1394. {
  1395. struct dfl_feature_platform_data *pdata;
  1396. mutex_lock(&cdev->lock);
  1397. list_for_each_entry(pdata, &cdev->port_dev_list, node) {
  1398. if (device_is_registered(&pdata->dev->dev))
  1399. continue;
  1400. config_port_pf_mode(cdev->fme_dev, pdata->id);
  1401. }
  1402. mutex_unlock(&cdev->lock);
  1403. }
  1404. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_config_ports_pf);
  1405. /**
  1406. * dfl_fpga_cdev_config_ports_vf - configure ports to VF access mode
  1407. *
  1408. * @cdev: parent container device.
  1409. * @num_vfs: VF device number.
  1410. *
  1411. * This function is needed in sriov configuration routine. It could be used to
  1412. * configure the released ports from PF access mode to VF.
  1413. *
  1414. * Return: 0 on success, negative error code otherwise.
  1415. */
  1416. int dfl_fpga_cdev_config_ports_vf(struct dfl_fpga_cdev *cdev, int num_vfs)
  1417. {
  1418. struct dfl_feature_platform_data *pdata;
  1419. int ret = 0;
  1420. mutex_lock(&cdev->lock);
  1421. /*
  1422. * can't turn multiple ports into 1 VF device, only 1 port for 1 VF
  1423. * device, so if released port number doesn't match VF device number,
  1424. * then reject the request with -EINVAL error code.
  1425. */
  1426. if (cdev->released_port_num != num_vfs) {
  1427. ret = -EINVAL;
  1428. goto done;
  1429. }
  1430. list_for_each_entry(pdata, &cdev->port_dev_list, node) {
  1431. if (device_is_registered(&pdata->dev->dev))
  1432. continue;
  1433. config_port_vf_mode(cdev->fme_dev, pdata->id);
  1434. }
  1435. done:
  1436. mutex_unlock(&cdev->lock);
  1437. return ret;
  1438. }
  1439. EXPORT_SYMBOL_GPL(dfl_fpga_cdev_config_ports_vf);
  1440. static irqreturn_t dfl_irq_handler(int irq, void *arg)
  1441. {
  1442. struct eventfd_ctx *trigger = arg;
  1443. eventfd_signal(trigger, 1);
  1444. return IRQ_HANDLED;
  1445. }
  1446. static int do_set_irq_trigger(struct dfl_feature *feature, unsigned int idx,
  1447. int fd)
  1448. {
  1449. struct platform_device *pdev = feature->dev;
  1450. struct eventfd_ctx *trigger;
  1451. int irq, ret;
  1452. irq = feature->irq_ctx[idx].irq;
  1453. if (feature->irq_ctx[idx].trigger) {
  1454. free_irq(irq, feature->irq_ctx[idx].trigger);
  1455. kfree(feature->irq_ctx[idx].name);
  1456. eventfd_ctx_put(feature->irq_ctx[idx].trigger);
  1457. feature->irq_ctx[idx].trigger = NULL;
  1458. }
  1459. if (fd < 0)
  1460. return 0;
  1461. feature->irq_ctx[idx].name =
  1462. kasprintf(GFP_KERNEL, "fpga-irq[%u](%s-%x)", idx,
  1463. dev_name(&pdev->dev), feature->id);
  1464. if (!feature->irq_ctx[idx].name)
  1465. return -ENOMEM;
  1466. trigger = eventfd_ctx_fdget(fd);
  1467. if (IS_ERR(trigger)) {
  1468. ret = PTR_ERR(trigger);
  1469. goto free_name;
  1470. }
  1471. ret = request_irq(irq, dfl_irq_handler, 0,
  1472. feature->irq_ctx[idx].name, trigger);
  1473. if (!ret) {
  1474. feature->irq_ctx[idx].trigger = trigger;
  1475. return ret;
  1476. }
  1477. eventfd_ctx_put(trigger);
  1478. free_name:
  1479. kfree(feature->irq_ctx[idx].name);
  1480. return ret;
  1481. }
  1482. /**
  1483. * dfl_fpga_set_irq_triggers - set eventfd triggers for dfl feature interrupts
  1484. *
  1485. * @feature: dfl sub feature.
  1486. * @start: start of irq index in this dfl sub feature.
  1487. * @count: number of irqs.
  1488. * @fds: eventfds to bind with irqs. unbind related irq if fds[n] is negative.
  1489. * unbind "count" specified number of irqs if fds ptr is NULL.
  1490. *
  1491. * Bind given eventfds with irqs in this dfl sub feature. Unbind related irq if
  1492. * fds[n] is negative. Unbind "count" specified number of irqs if fds ptr is
  1493. * NULL.
  1494. *
  1495. * Return: 0 on success, negative error code otherwise.
  1496. */
  1497. int dfl_fpga_set_irq_triggers(struct dfl_feature *feature, unsigned int start,
  1498. unsigned int count, int32_t *fds)
  1499. {
  1500. unsigned int i;
  1501. int ret = 0;
  1502. /* overflow */
  1503. if (unlikely(start + count < start))
  1504. return -EINVAL;
  1505. /* exceeds nr_irqs */
  1506. if (start + count > feature->nr_irqs)
  1507. return -EINVAL;
  1508. for (i = 0; i < count; i++) {
  1509. int fd = fds ? fds[i] : -1;
  1510. ret = do_set_irq_trigger(feature, start + i, fd);
  1511. if (ret) {
  1512. while (i--)
  1513. do_set_irq_trigger(feature, start + i, -1);
  1514. break;
  1515. }
  1516. }
  1517. return ret;
  1518. }
  1519. EXPORT_SYMBOL_GPL(dfl_fpga_set_irq_triggers);
  1520. /**
  1521. * dfl_feature_ioctl_get_num_irqs - dfl feature _GET_IRQ_NUM ioctl interface.
  1522. * @pdev: the feature device which has the sub feature
  1523. * @feature: the dfl sub feature
  1524. * @arg: ioctl argument
  1525. *
  1526. * Return: 0 on success, negative error code otherwise.
  1527. */
  1528. long dfl_feature_ioctl_get_num_irqs(struct platform_device *pdev,
  1529. struct dfl_feature *feature,
  1530. unsigned long arg)
  1531. {
  1532. return put_user(feature->nr_irqs, (__u32 __user *)arg);
  1533. }
  1534. EXPORT_SYMBOL_GPL(dfl_feature_ioctl_get_num_irqs);
  1535. /**
  1536. * dfl_feature_ioctl_set_irq - dfl feature _SET_IRQ ioctl interface.
  1537. * @pdev: the feature device which has the sub feature
  1538. * @feature: the dfl sub feature
  1539. * @arg: ioctl argument
  1540. *
  1541. * Return: 0 on success, negative error code otherwise.
  1542. */
  1543. long dfl_feature_ioctl_set_irq(struct platform_device *pdev,
  1544. struct dfl_feature *feature,
  1545. unsigned long arg)
  1546. {
  1547. struct dfl_feature_platform_data *pdata = dev_get_platdata(&pdev->dev);
  1548. struct dfl_fpga_irq_set hdr;
  1549. s32 *fds;
  1550. long ret;
  1551. if (!feature->nr_irqs)
  1552. return -ENOENT;
  1553. if (copy_from_user(&hdr, (void __user *)arg, sizeof(hdr)))
  1554. return -EFAULT;
  1555. if (!hdr.count || (hdr.start + hdr.count > feature->nr_irqs) ||
  1556. (hdr.start + hdr.count < hdr.start))
  1557. return -EINVAL;
  1558. fds = memdup_user((void __user *)(arg + sizeof(hdr)),
  1559. array_size(hdr.count, sizeof(s32)));
  1560. if (IS_ERR(fds))
  1561. return PTR_ERR(fds);
  1562. mutex_lock(&pdata->lock);
  1563. ret = dfl_fpga_set_irq_triggers(feature, hdr.start, hdr.count, fds);
  1564. mutex_unlock(&pdata->lock);
  1565. kfree(fds);
  1566. return ret;
  1567. }
  1568. EXPORT_SYMBOL_GPL(dfl_feature_ioctl_set_irq);
  1569. static void __exit dfl_fpga_exit(void)
  1570. {
  1571. dfl_chardev_uinit();
  1572. dfl_ids_destroy();
  1573. bus_unregister(&dfl_bus_type);
  1574. }
  1575. module_init(dfl_fpga_init);
  1576. module_exit(dfl_fpga_exit);
  1577. MODULE_DESCRIPTION("FPGA Device Feature List (DFL) Support");
  1578. MODULE_AUTHOR("Intel Corporation");
  1579. MODULE_LICENSE("GPL v2");