owl-emac.c 40 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. /*
  3. * Actions Semi Owl SoCs Ethernet MAC driver
  4. *
  5. * Copyright (c) 2012 Actions Semi Inc.
  6. * Copyright (c) 2021 Cristian Ciocaltea <[email protected]>
  7. */
  8. #include <linux/circ_buf.h>
  9. #include <linux/clk.h>
  10. #include <linux/dma-mapping.h>
  11. #include <linux/etherdevice.h>
  12. #include <linux/of_mdio.h>
  13. #include <linux/of_net.h>
  14. #include <linux/platform_device.h>
  15. #include <linux/pm.h>
  16. #include <linux/reset.h>
  17. #include "owl-emac.h"
  18. #define OWL_EMAC_DEFAULT_MSG_ENABLE (NETIF_MSG_DRV | \
  19. NETIF_MSG_PROBE | \
  20. NETIF_MSG_LINK)
  21. static u32 owl_emac_reg_read(struct owl_emac_priv *priv, u32 reg)
  22. {
  23. return readl(priv->base + reg);
  24. }
  25. static void owl_emac_reg_write(struct owl_emac_priv *priv, u32 reg, u32 data)
  26. {
  27. writel(data, priv->base + reg);
  28. }
  29. static u32 owl_emac_reg_update(struct owl_emac_priv *priv,
  30. u32 reg, u32 mask, u32 val)
  31. {
  32. u32 data, old_val;
  33. data = owl_emac_reg_read(priv, reg);
  34. old_val = data & mask;
  35. data &= ~mask;
  36. data |= val & mask;
  37. owl_emac_reg_write(priv, reg, data);
  38. return old_val;
  39. }
  40. static void owl_emac_reg_set(struct owl_emac_priv *priv, u32 reg, u32 bits)
  41. {
  42. owl_emac_reg_update(priv, reg, bits, bits);
  43. }
  44. static void owl_emac_reg_clear(struct owl_emac_priv *priv, u32 reg, u32 bits)
  45. {
  46. owl_emac_reg_update(priv, reg, bits, 0);
  47. }
  48. static struct device *owl_emac_get_dev(struct owl_emac_priv *priv)
  49. {
  50. return priv->netdev->dev.parent;
  51. }
  52. static void owl_emac_irq_enable(struct owl_emac_priv *priv)
  53. {
  54. /* Enable all interrupts except TU.
  55. *
  56. * Note the NIE and AIE bits shall also be set in order to actually
  57. * enable the selected interrupts.
  58. */
  59. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR7,
  60. OWL_EMAC_BIT_MAC_CSR7_NIE |
  61. OWL_EMAC_BIT_MAC_CSR7_AIE |
  62. OWL_EMAC_BIT_MAC_CSR7_ALL_NOT_TUE);
  63. }
  64. static void owl_emac_irq_disable(struct owl_emac_priv *priv)
  65. {
  66. /* Disable all interrupts.
  67. *
  68. * WARNING: Unset only the NIE and AIE bits in CSR7 to workaround an
  69. * unexpected side effect (MAC hardware bug?!) where some bits in the
  70. * status register (CSR5) are cleared automatically before being able
  71. * to read them via owl_emac_irq_clear().
  72. */
  73. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR7,
  74. OWL_EMAC_BIT_MAC_CSR7_ALL_NOT_TUE);
  75. }
  76. static u32 owl_emac_irq_status(struct owl_emac_priv *priv)
  77. {
  78. return owl_emac_reg_read(priv, OWL_EMAC_REG_MAC_CSR5);
  79. }
  80. static u32 owl_emac_irq_clear(struct owl_emac_priv *priv)
  81. {
  82. u32 val = owl_emac_irq_status(priv);
  83. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR5, val);
  84. return val;
  85. }
  86. static dma_addr_t owl_emac_dma_map_rx(struct owl_emac_priv *priv,
  87. struct sk_buff *skb)
  88. {
  89. struct device *dev = owl_emac_get_dev(priv);
  90. /* Buffer pointer for the RX DMA descriptor must be word aligned. */
  91. return dma_map_single(dev, skb_tail_pointer(skb),
  92. skb_tailroom(skb), DMA_FROM_DEVICE);
  93. }
  94. static void owl_emac_dma_unmap_rx(struct owl_emac_priv *priv,
  95. struct sk_buff *skb, dma_addr_t dma_addr)
  96. {
  97. struct device *dev = owl_emac_get_dev(priv);
  98. dma_unmap_single(dev, dma_addr, skb_tailroom(skb), DMA_FROM_DEVICE);
  99. }
  100. static dma_addr_t owl_emac_dma_map_tx(struct owl_emac_priv *priv,
  101. struct sk_buff *skb)
  102. {
  103. struct device *dev = owl_emac_get_dev(priv);
  104. return dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE);
  105. }
  106. static void owl_emac_dma_unmap_tx(struct owl_emac_priv *priv,
  107. struct sk_buff *skb, dma_addr_t dma_addr)
  108. {
  109. struct device *dev = owl_emac_get_dev(priv);
  110. dma_unmap_single(dev, dma_addr, skb_headlen(skb), DMA_TO_DEVICE);
  111. }
  112. static unsigned int owl_emac_ring_num_unused(struct owl_emac_ring *ring)
  113. {
  114. return CIRC_SPACE(ring->head, ring->tail, ring->size);
  115. }
  116. static unsigned int owl_emac_ring_get_next(struct owl_emac_ring *ring,
  117. unsigned int cur)
  118. {
  119. return (cur + 1) & (ring->size - 1);
  120. }
  121. static void owl_emac_ring_push_head(struct owl_emac_ring *ring)
  122. {
  123. ring->head = owl_emac_ring_get_next(ring, ring->head);
  124. }
  125. static void owl_emac_ring_pop_tail(struct owl_emac_ring *ring)
  126. {
  127. ring->tail = owl_emac_ring_get_next(ring, ring->tail);
  128. }
  129. static struct sk_buff *owl_emac_alloc_skb(struct net_device *netdev)
  130. {
  131. struct sk_buff *skb;
  132. int offset;
  133. skb = netdev_alloc_skb(netdev, OWL_EMAC_RX_FRAME_MAX_LEN +
  134. OWL_EMAC_SKB_RESERVE);
  135. if (unlikely(!skb))
  136. return NULL;
  137. /* Ensure 4 bytes DMA alignment. */
  138. offset = ((uintptr_t)skb->data) & (OWL_EMAC_SKB_ALIGN - 1);
  139. if (unlikely(offset))
  140. skb_reserve(skb, OWL_EMAC_SKB_ALIGN - offset);
  141. return skb;
  142. }
  143. static int owl_emac_ring_prepare_rx(struct owl_emac_priv *priv)
  144. {
  145. struct owl_emac_ring *ring = &priv->rx_ring;
  146. struct device *dev = owl_emac_get_dev(priv);
  147. struct net_device *netdev = priv->netdev;
  148. struct owl_emac_ring_desc *desc;
  149. struct sk_buff *skb;
  150. dma_addr_t dma_addr;
  151. int i;
  152. for (i = 0; i < ring->size; i++) {
  153. skb = owl_emac_alloc_skb(netdev);
  154. if (!skb)
  155. return -ENOMEM;
  156. dma_addr = owl_emac_dma_map_rx(priv, skb);
  157. if (dma_mapping_error(dev, dma_addr)) {
  158. dev_kfree_skb(skb);
  159. return -ENOMEM;
  160. }
  161. desc = &ring->descs[i];
  162. desc->status = OWL_EMAC_BIT_RDES0_OWN;
  163. desc->control = skb_tailroom(skb) & OWL_EMAC_MSK_RDES1_RBS1;
  164. desc->buf_addr = dma_addr;
  165. desc->reserved = 0;
  166. ring->skbs[i] = skb;
  167. ring->skbs_dma[i] = dma_addr;
  168. }
  169. desc->control |= OWL_EMAC_BIT_RDES1_RER;
  170. ring->head = 0;
  171. ring->tail = 0;
  172. return 0;
  173. }
  174. static void owl_emac_ring_prepare_tx(struct owl_emac_priv *priv)
  175. {
  176. struct owl_emac_ring *ring = &priv->tx_ring;
  177. struct owl_emac_ring_desc *desc;
  178. int i;
  179. for (i = 0; i < ring->size; i++) {
  180. desc = &ring->descs[i];
  181. desc->status = 0;
  182. desc->control = OWL_EMAC_BIT_TDES1_IC;
  183. desc->buf_addr = 0;
  184. desc->reserved = 0;
  185. }
  186. desc->control |= OWL_EMAC_BIT_TDES1_TER;
  187. memset(ring->skbs_dma, 0, sizeof(dma_addr_t) * ring->size);
  188. ring->head = 0;
  189. ring->tail = 0;
  190. }
  191. static void owl_emac_ring_unprepare_rx(struct owl_emac_priv *priv)
  192. {
  193. struct owl_emac_ring *ring = &priv->rx_ring;
  194. int i;
  195. for (i = 0; i < ring->size; i++) {
  196. ring->descs[i].status = 0;
  197. if (!ring->skbs_dma[i])
  198. continue;
  199. owl_emac_dma_unmap_rx(priv, ring->skbs[i], ring->skbs_dma[i]);
  200. ring->skbs_dma[i] = 0;
  201. dev_kfree_skb(ring->skbs[i]);
  202. ring->skbs[i] = NULL;
  203. }
  204. }
  205. static void owl_emac_ring_unprepare_tx(struct owl_emac_priv *priv)
  206. {
  207. struct owl_emac_ring *ring = &priv->tx_ring;
  208. int i;
  209. for (i = 0; i < ring->size; i++) {
  210. ring->descs[i].status = 0;
  211. if (!ring->skbs_dma[i])
  212. continue;
  213. owl_emac_dma_unmap_tx(priv, ring->skbs[i], ring->skbs_dma[i]);
  214. ring->skbs_dma[i] = 0;
  215. dev_kfree_skb(ring->skbs[i]);
  216. ring->skbs[i] = NULL;
  217. }
  218. }
  219. static int owl_emac_ring_alloc(struct device *dev, struct owl_emac_ring *ring,
  220. unsigned int size)
  221. {
  222. ring->descs = dmam_alloc_coherent(dev,
  223. sizeof(struct owl_emac_ring_desc) * size,
  224. &ring->descs_dma, GFP_KERNEL);
  225. if (!ring->descs)
  226. return -ENOMEM;
  227. ring->skbs = devm_kcalloc(dev, size, sizeof(struct sk_buff *),
  228. GFP_KERNEL);
  229. if (!ring->skbs)
  230. return -ENOMEM;
  231. ring->skbs_dma = devm_kcalloc(dev, size, sizeof(dma_addr_t),
  232. GFP_KERNEL);
  233. if (!ring->skbs_dma)
  234. return -ENOMEM;
  235. ring->size = size;
  236. return 0;
  237. }
  238. static void owl_emac_dma_cmd_resume_rx(struct owl_emac_priv *priv)
  239. {
  240. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR2,
  241. OWL_EMAC_VAL_MAC_CSR2_RPD);
  242. }
  243. static void owl_emac_dma_cmd_resume_tx(struct owl_emac_priv *priv)
  244. {
  245. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR1,
  246. OWL_EMAC_VAL_MAC_CSR1_TPD);
  247. }
  248. static u32 owl_emac_dma_cmd_set_tx(struct owl_emac_priv *priv, u32 status)
  249. {
  250. return owl_emac_reg_update(priv, OWL_EMAC_REG_MAC_CSR6,
  251. OWL_EMAC_BIT_MAC_CSR6_ST, status);
  252. }
  253. static u32 owl_emac_dma_cmd_start_tx(struct owl_emac_priv *priv)
  254. {
  255. return owl_emac_dma_cmd_set_tx(priv, ~0);
  256. }
  257. static u32 owl_emac_dma_cmd_set(struct owl_emac_priv *priv, u32 status)
  258. {
  259. return owl_emac_reg_update(priv, OWL_EMAC_REG_MAC_CSR6,
  260. OWL_EMAC_MSK_MAC_CSR6_STSR, status);
  261. }
  262. static u32 owl_emac_dma_cmd_start(struct owl_emac_priv *priv)
  263. {
  264. return owl_emac_dma_cmd_set(priv, ~0);
  265. }
  266. static u32 owl_emac_dma_cmd_stop(struct owl_emac_priv *priv)
  267. {
  268. return owl_emac_dma_cmd_set(priv, 0);
  269. }
  270. static void owl_emac_set_hw_mac_addr(struct net_device *netdev)
  271. {
  272. struct owl_emac_priv *priv = netdev_priv(netdev);
  273. const u8 *mac_addr = netdev->dev_addr;
  274. u32 addr_high, addr_low;
  275. addr_high = mac_addr[0] << 8 | mac_addr[1];
  276. addr_low = mac_addr[2] << 24 | mac_addr[3] << 16 |
  277. mac_addr[4] << 8 | mac_addr[5];
  278. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR17, addr_high);
  279. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR16, addr_low);
  280. }
  281. static void owl_emac_update_link_state(struct owl_emac_priv *priv)
  282. {
  283. u32 val, status;
  284. if (priv->pause) {
  285. val = OWL_EMAC_BIT_MAC_CSR20_FCE | OWL_EMAC_BIT_MAC_CSR20_TUE;
  286. val |= OWL_EMAC_BIT_MAC_CSR20_TPE | OWL_EMAC_BIT_MAC_CSR20_RPE;
  287. val |= OWL_EMAC_BIT_MAC_CSR20_BPE;
  288. } else {
  289. val = 0;
  290. }
  291. /* Update flow control. */
  292. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR20, val);
  293. val = (priv->speed == SPEED_100) ? OWL_EMAC_VAL_MAC_CSR6_SPEED_100M :
  294. OWL_EMAC_VAL_MAC_CSR6_SPEED_10M;
  295. val <<= OWL_EMAC_OFF_MAC_CSR6_SPEED;
  296. if (priv->duplex == DUPLEX_FULL)
  297. val |= OWL_EMAC_BIT_MAC_CSR6_FD;
  298. spin_lock_bh(&priv->lock);
  299. /* Temporarily stop DMA TX & RX. */
  300. status = owl_emac_dma_cmd_stop(priv);
  301. /* Update operation modes. */
  302. owl_emac_reg_update(priv, OWL_EMAC_REG_MAC_CSR6,
  303. OWL_EMAC_MSK_MAC_CSR6_SPEED |
  304. OWL_EMAC_BIT_MAC_CSR6_FD, val);
  305. /* Restore DMA TX & RX status. */
  306. owl_emac_dma_cmd_set(priv, status);
  307. spin_unlock_bh(&priv->lock);
  308. }
  309. static void owl_emac_adjust_link(struct net_device *netdev)
  310. {
  311. struct owl_emac_priv *priv = netdev_priv(netdev);
  312. struct phy_device *phydev = netdev->phydev;
  313. bool state_changed = false;
  314. if (phydev->link) {
  315. if (!priv->link) {
  316. priv->link = phydev->link;
  317. state_changed = true;
  318. }
  319. if (priv->speed != phydev->speed) {
  320. priv->speed = phydev->speed;
  321. state_changed = true;
  322. }
  323. if (priv->duplex != phydev->duplex) {
  324. priv->duplex = phydev->duplex;
  325. state_changed = true;
  326. }
  327. if (priv->pause != phydev->pause) {
  328. priv->pause = phydev->pause;
  329. state_changed = true;
  330. }
  331. } else {
  332. if (priv->link) {
  333. priv->link = phydev->link;
  334. state_changed = true;
  335. }
  336. }
  337. if (state_changed) {
  338. if (phydev->link)
  339. owl_emac_update_link_state(priv);
  340. if (netif_msg_link(priv))
  341. phy_print_status(phydev);
  342. }
  343. }
  344. static irqreturn_t owl_emac_handle_irq(int irq, void *data)
  345. {
  346. struct net_device *netdev = data;
  347. struct owl_emac_priv *priv = netdev_priv(netdev);
  348. if (netif_running(netdev)) {
  349. owl_emac_irq_disable(priv);
  350. napi_schedule(&priv->napi);
  351. }
  352. return IRQ_HANDLED;
  353. }
  354. static void owl_emac_ether_addr_push(u8 **dst, const u8 *src)
  355. {
  356. u32 *a = (u32 *)(*dst);
  357. const u16 *b = (const u16 *)src;
  358. a[0] = b[0];
  359. a[1] = b[1];
  360. a[2] = b[2];
  361. *dst += 12;
  362. }
  363. static void
  364. owl_emac_setup_frame_prepare(struct owl_emac_priv *priv, struct sk_buff *skb)
  365. {
  366. const u8 bcast_addr[] = { 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF };
  367. const u8 *mac_addr = priv->netdev->dev_addr;
  368. u8 *frame;
  369. int i;
  370. skb_put(skb, OWL_EMAC_SETUP_FRAME_LEN);
  371. frame = skb->data;
  372. memset(frame, 0, skb->len);
  373. owl_emac_ether_addr_push(&frame, mac_addr);
  374. owl_emac_ether_addr_push(&frame, bcast_addr);
  375. /* Fill multicast addresses. */
  376. WARN_ON(priv->mcaddr_list.count >= OWL_EMAC_MAX_MULTICAST_ADDRS);
  377. for (i = 0; i < priv->mcaddr_list.count; i++) {
  378. mac_addr = priv->mcaddr_list.addrs[i];
  379. owl_emac_ether_addr_push(&frame, mac_addr);
  380. }
  381. }
  382. /* The setup frame is a special descriptor which is used to provide physical
  383. * addresses (i.e. mac, broadcast and multicast) to the MAC hardware for
  384. * filtering purposes. To be recognized as a setup frame, the TDES1_SET bit
  385. * must be set in the TX descriptor control field.
  386. */
  387. static int owl_emac_setup_frame_xmit(struct owl_emac_priv *priv)
  388. {
  389. struct owl_emac_ring *ring = &priv->tx_ring;
  390. struct net_device *netdev = priv->netdev;
  391. struct owl_emac_ring_desc *desc;
  392. struct sk_buff *skb;
  393. unsigned int tx_head;
  394. u32 status, control;
  395. dma_addr_t dma_addr;
  396. int ret;
  397. skb = owl_emac_alloc_skb(netdev);
  398. if (!skb)
  399. return -ENOMEM;
  400. owl_emac_setup_frame_prepare(priv, skb);
  401. dma_addr = owl_emac_dma_map_tx(priv, skb);
  402. if (dma_mapping_error(owl_emac_get_dev(priv), dma_addr)) {
  403. ret = -ENOMEM;
  404. goto err_free_skb;
  405. }
  406. spin_lock_bh(&priv->lock);
  407. tx_head = ring->head;
  408. desc = &ring->descs[tx_head];
  409. status = READ_ONCE(desc->status);
  410. control = READ_ONCE(desc->control);
  411. dma_rmb(); /* Ensure data has been read before used. */
  412. if (unlikely(status & OWL_EMAC_BIT_TDES0_OWN) ||
  413. !owl_emac_ring_num_unused(ring)) {
  414. spin_unlock_bh(&priv->lock);
  415. owl_emac_dma_unmap_tx(priv, skb, dma_addr);
  416. ret = -EBUSY;
  417. goto err_free_skb;
  418. }
  419. ring->skbs[tx_head] = skb;
  420. ring->skbs_dma[tx_head] = dma_addr;
  421. control &= OWL_EMAC_BIT_TDES1_IC | OWL_EMAC_BIT_TDES1_TER; /* Maintain bits */
  422. control |= OWL_EMAC_BIT_TDES1_SET;
  423. control |= OWL_EMAC_MSK_TDES1_TBS1 & skb->len;
  424. WRITE_ONCE(desc->control, control);
  425. WRITE_ONCE(desc->buf_addr, dma_addr);
  426. dma_wmb(); /* Flush descriptor before changing ownership. */
  427. WRITE_ONCE(desc->status, OWL_EMAC_BIT_TDES0_OWN);
  428. owl_emac_ring_push_head(ring);
  429. /* Temporarily enable DMA TX. */
  430. status = owl_emac_dma_cmd_start_tx(priv);
  431. /* Trigger setup frame processing. */
  432. owl_emac_dma_cmd_resume_tx(priv);
  433. /* Restore DMA TX status. */
  434. owl_emac_dma_cmd_set_tx(priv, status);
  435. /* Stop regular TX until setup frame is processed. */
  436. netif_stop_queue(netdev);
  437. spin_unlock_bh(&priv->lock);
  438. return 0;
  439. err_free_skb:
  440. dev_kfree_skb(skb);
  441. return ret;
  442. }
  443. static netdev_tx_t owl_emac_ndo_start_xmit(struct sk_buff *skb,
  444. struct net_device *netdev)
  445. {
  446. struct owl_emac_priv *priv = netdev_priv(netdev);
  447. struct device *dev = owl_emac_get_dev(priv);
  448. struct owl_emac_ring *ring = &priv->tx_ring;
  449. struct owl_emac_ring_desc *desc;
  450. unsigned int tx_head;
  451. u32 status, control;
  452. dma_addr_t dma_addr;
  453. dma_addr = owl_emac_dma_map_tx(priv, skb);
  454. if (dma_mapping_error(dev, dma_addr)) {
  455. dev_err_ratelimited(&netdev->dev, "TX DMA mapping failed\n");
  456. dev_kfree_skb(skb);
  457. netdev->stats.tx_dropped++;
  458. return NETDEV_TX_OK;
  459. }
  460. spin_lock_bh(&priv->lock);
  461. tx_head = ring->head;
  462. desc = &ring->descs[tx_head];
  463. status = READ_ONCE(desc->status);
  464. control = READ_ONCE(desc->control);
  465. dma_rmb(); /* Ensure data has been read before used. */
  466. if (!owl_emac_ring_num_unused(ring) ||
  467. unlikely(status & OWL_EMAC_BIT_TDES0_OWN)) {
  468. netif_stop_queue(netdev);
  469. spin_unlock_bh(&priv->lock);
  470. dev_dbg_ratelimited(&netdev->dev, "TX buffer full, status=0x%08x\n",
  471. owl_emac_irq_status(priv));
  472. owl_emac_dma_unmap_tx(priv, skb, dma_addr);
  473. netdev->stats.tx_dropped++;
  474. return NETDEV_TX_BUSY;
  475. }
  476. ring->skbs[tx_head] = skb;
  477. ring->skbs_dma[tx_head] = dma_addr;
  478. control &= OWL_EMAC_BIT_TDES1_IC | OWL_EMAC_BIT_TDES1_TER; /* Maintain bits */
  479. control |= OWL_EMAC_BIT_TDES1_FS | OWL_EMAC_BIT_TDES1_LS;
  480. control |= OWL_EMAC_MSK_TDES1_TBS1 & skb->len;
  481. WRITE_ONCE(desc->control, control);
  482. WRITE_ONCE(desc->buf_addr, dma_addr);
  483. dma_wmb(); /* Flush descriptor before changing ownership. */
  484. WRITE_ONCE(desc->status, OWL_EMAC_BIT_TDES0_OWN);
  485. owl_emac_dma_cmd_resume_tx(priv);
  486. owl_emac_ring_push_head(ring);
  487. /* FIXME: The transmission is currently restricted to a single frame
  488. * at a time as a workaround for a MAC hardware bug that causes random
  489. * freeze of the TX queue processor.
  490. */
  491. netif_stop_queue(netdev);
  492. spin_unlock_bh(&priv->lock);
  493. return NETDEV_TX_OK;
  494. }
  495. static bool owl_emac_tx_complete_tail(struct owl_emac_priv *priv)
  496. {
  497. struct owl_emac_ring *ring = &priv->tx_ring;
  498. struct net_device *netdev = priv->netdev;
  499. struct owl_emac_ring_desc *desc;
  500. struct sk_buff *skb;
  501. unsigned int tx_tail;
  502. u32 status;
  503. tx_tail = ring->tail;
  504. desc = &ring->descs[tx_tail];
  505. status = READ_ONCE(desc->status);
  506. dma_rmb(); /* Ensure data has been read before used. */
  507. if (status & OWL_EMAC_BIT_TDES0_OWN)
  508. return false;
  509. /* Check for errors. */
  510. if (status & OWL_EMAC_BIT_TDES0_ES) {
  511. dev_dbg_ratelimited(&netdev->dev,
  512. "TX complete error status: 0x%08x\n",
  513. status);
  514. netdev->stats.tx_errors++;
  515. if (status & OWL_EMAC_BIT_TDES0_UF)
  516. netdev->stats.tx_fifo_errors++;
  517. if (status & OWL_EMAC_BIT_TDES0_EC)
  518. netdev->stats.tx_aborted_errors++;
  519. if (status & OWL_EMAC_BIT_TDES0_LC)
  520. netdev->stats.tx_window_errors++;
  521. if (status & OWL_EMAC_BIT_TDES0_NC)
  522. netdev->stats.tx_heartbeat_errors++;
  523. if (status & OWL_EMAC_BIT_TDES0_LO)
  524. netdev->stats.tx_carrier_errors++;
  525. } else {
  526. netdev->stats.tx_packets++;
  527. netdev->stats.tx_bytes += ring->skbs[tx_tail]->len;
  528. }
  529. /* Some collisions occurred, but pkt has been transmitted. */
  530. if (status & OWL_EMAC_BIT_TDES0_DE)
  531. netdev->stats.collisions++;
  532. skb = ring->skbs[tx_tail];
  533. owl_emac_dma_unmap_tx(priv, skb, ring->skbs_dma[tx_tail]);
  534. dev_kfree_skb(skb);
  535. ring->skbs[tx_tail] = NULL;
  536. ring->skbs_dma[tx_tail] = 0;
  537. owl_emac_ring_pop_tail(ring);
  538. if (unlikely(netif_queue_stopped(netdev)))
  539. netif_wake_queue(netdev);
  540. return true;
  541. }
  542. static void owl_emac_tx_complete(struct owl_emac_priv *priv)
  543. {
  544. struct owl_emac_ring *ring = &priv->tx_ring;
  545. struct net_device *netdev = priv->netdev;
  546. unsigned int tx_next;
  547. u32 status;
  548. spin_lock(&priv->lock);
  549. while (ring->tail != ring->head) {
  550. if (!owl_emac_tx_complete_tail(priv))
  551. break;
  552. }
  553. /* FIXME: This is a workaround for a MAC hardware bug not clearing
  554. * (sometimes) the OWN bit for a transmitted frame descriptor.
  555. *
  556. * At this point, when TX queue is full, the tail descriptor has the
  557. * OWN bit set, which normally means the frame has not been processed
  558. * or transmitted yet. But if there is at least one descriptor in the
  559. * queue having the OWN bit cleared, we can safely assume the tail
  560. * frame has been also processed by the MAC hardware.
  561. *
  562. * If that's the case, let's force the frame completion by manually
  563. * clearing the OWN bit.
  564. */
  565. if (unlikely(!owl_emac_ring_num_unused(ring))) {
  566. tx_next = ring->tail;
  567. while ((tx_next = owl_emac_ring_get_next(ring, tx_next)) != ring->head) {
  568. status = READ_ONCE(ring->descs[tx_next].status);
  569. dma_rmb(); /* Ensure data has been read before used. */
  570. if (status & OWL_EMAC_BIT_TDES0_OWN)
  571. continue;
  572. netdev_dbg(netdev, "Found uncleared TX desc OWN bit\n");
  573. status = READ_ONCE(ring->descs[ring->tail].status);
  574. dma_rmb(); /* Ensure data has been read before used. */
  575. status &= ~OWL_EMAC_BIT_TDES0_OWN;
  576. WRITE_ONCE(ring->descs[ring->tail].status, status);
  577. owl_emac_tx_complete_tail(priv);
  578. break;
  579. }
  580. }
  581. spin_unlock(&priv->lock);
  582. }
  583. static int owl_emac_rx_process(struct owl_emac_priv *priv, int budget)
  584. {
  585. struct owl_emac_ring *ring = &priv->rx_ring;
  586. struct device *dev = owl_emac_get_dev(priv);
  587. struct net_device *netdev = priv->netdev;
  588. struct owl_emac_ring_desc *desc;
  589. struct sk_buff *curr_skb, *new_skb;
  590. dma_addr_t curr_dma, new_dma;
  591. unsigned int rx_tail, len;
  592. u32 status;
  593. int recv = 0;
  594. while (recv < budget) {
  595. spin_lock(&priv->lock);
  596. rx_tail = ring->tail;
  597. desc = &ring->descs[rx_tail];
  598. status = READ_ONCE(desc->status);
  599. dma_rmb(); /* Ensure data has been read before used. */
  600. if (status & OWL_EMAC_BIT_RDES0_OWN) {
  601. spin_unlock(&priv->lock);
  602. break;
  603. }
  604. curr_skb = ring->skbs[rx_tail];
  605. curr_dma = ring->skbs_dma[rx_tail];
  606. owl_emac_ring_pop_tail(ring);
  607. spin_unlock(&priv->lock);
  608. if (status & (OWL_EMAC_BIT_RDES0_DE | OWL_EMAC_BIT_RDES0_RF |
  609. OWL_EMAC_BIT_RDES0_TL | OWL_EMAC_BIT_RDES0_CS |
  610. OWL_EMAC_BIT_RDES0_DB | OWL_EMAC_BIT_RDES0_CE |
  611. OWL_EMAC_BIT_RDES0_ZERO)) {
  612. dev_dbg_ratelimited(&netdev->dev,
  613. "RX desc error status: 0x%08x\n",
  614. status);
  615. if (status & OWL_EMAC_BIT_RDES0_DE)
  616. netdev->stats.rx_over_errors++;
  617. if (status & (OWL_EMAC_BIT_RDES0_RF | OWL_EMAC_BIT_RDES0_DB))
  618. netdev->stats.rx_frame_errors++;
  619. if (status & OWL_EMAC_BIT_RDES0_TL)
  620. netdev->stats.rx_length_errors++;
  621. if (status & OWL_EMAC_BIT_RDES0_CS)
  622. netdev->stats.collisions++;
  623. if (status & OWL_EMAC_BIT_RDES0_CE)
  624. netdev->stats.rx_crc_errors++;
  625. if (status & OWL_EMAC_BIT_RDES0_ZERO)
  626. netdev->stats.rx_fifo_errors++;
  627. goto drop_skb;
  628. }
  629. len = (status & OWL_EMAC_MSK_RDES0_FL) >> OWL_EMAC_OFF_RDES0_FL;
  630. if (unlikely(len > OWL_EMAC_RX_FRAME_MAX_LEN)) {
  631. netdev->stats.rx_length_errors++;
  632. netdev_err(netdev, "invalid RX frame len: %u\n", len);
  633. goto drop_skb;
  634. }
  635. /* Prepare new skb before receiving the current one. */
  636. new_skb = owl_emac_alloc_skb(netdev);
  637. if (unlikely(!new_skb))
  638. goto drop_skb;
  639. new_dma = owl_emac_dma_map_rx(priv, new_skb);
  640. if (dma_mapping_error(dev, new_dma)) {
  641. dev_kfree_skb(new_skb);
  642. netdev_err(netdev, "RX DMA mapping failed\n");
  643. goto drop_skb;
  644. }
  645. owl_emac_dma_unmap_rx(priv, curr_skb, curr_dma);
  646. skb_put(curr_skb, len - ETH_FCS_LEN);
  647. curr_skb->ip_summed = CHECKSUM_NONE;
  648. curr_skb->protocol = eth_type_trans(curr_skb, netdev);
  649. curr_skb->dev = netdev;
  650. netif_receive_skb(curr_skb);
  651. netdev->stats.rx_packets++;
  652. netdev->stats.rx_bytes += len;
  653. recv++;
  654. goto push_skb;
  655. drop_skb:
  656. netdev->stats.rx_dropped++;
  657. netdev->stats.rx_errors++;
  658. /* Reuse the current skb. */
  659. new_skb = curr_skb;
  660. new_dma = curr_dma;
  661. push_skb:
  662. spin_lock(&priv->lock);
  663. ring->skbs[ring->head] = new_skb;
  664. ring->skbs_dma[ring->head] = new_dma;
  665. WRITE_ONCE(desc->buf_addr, new_dma);
  666. dma_wmb(); /* Flush descriptor before changing ownership. */
  667. WRITE_ONCE(desc->status, OWL_EMAC_BIT_RDES0_OWN);
  668. owl_emac_ring_push_head(ring);
  669. spin_unlock(&priv->lock);
  670. }
  671. return recv;
  672. }
  673. static int owl_emac_poll(struct napi_struct *napi, int budget)
  674. {
  675. int work_done = 0, ru_cnt = 0, recv;
  676. static int tx_err_cnt, rx_err_cnt;
  677. struct owl_emac_priv *priv;
  678. u32 status, proc_status;
  679. priv = container_of(napi, struct owl_emac_priv, napi);
  680. while ((status = owl_emac_irq_clear(priv)) &
  681. (OWL_EMAC_BIT_MAC_CSR5_NIS | OWL_EMAC_BIT_MAC_CSR5_AIS)) {
  682. recv = 0;
  683. /* TX setup frame raises ETI instead of TI. */
  684. if (status & (OWL_EMAC_BIT_MAC_CSR5_TI | OWL_EMAC_BIT_MAC_CSR5_ETI)) {
  685. owl_emac_tx_complete(priv);
  686. tx_err_cnt = 0;
  687. /* Count MAC internal RX errors. */
  688. proc_status = status & OWL_EMAC_MSK_MAC_CSR5_RS;
  689. proc_status >>= OWL_EMAC_OFF_MAC_CSR5_RS;
  690. if (proc_status == OWL_EMAC_VAL_MAC_CSR5_RS_DATA ||
  691. proc_status == OWL_EMAC_VAL_MAC_CSR5_RS_CDES ||
  692. proc_status == OWL_EMAC_VAL_MAC_CSR5_RS_FDES)
  693. rx_err_cnt++;
  694. }
  695. if (status & OWL_EMAC_BIT_MAC_CSR5_RI) {
  696. recv = owl_emac_rx_process(priv, budget - work_done);
  697. rx_err_cnt = 0;
  698. /* Count MAC internal TX errors. */
  699. proc_status = status & OWL_EMAC_MSK_MAC_CSR5_TS;
  700. proc_status >>= OWL_EMAC_OFF_MAC_CSR5_TS;
  701. if (proc_status == OWL_EMAC_VAL_MAC_CSR5_TS_DATA ||
  702. proc_status == OWL_EMAC_VAL_MAC_CSR5_TS_CDES)
  703. tx_err_cnt++;
  704. } else if (status & OWL_EMAC_BIT_MAC_CSR5_RU) {
  705. /* MAC AHB is in suspended state, will return to RX
  706. * descriptor processing when the host changes ownership
  707. * of the descriptor and either an RX poll demand CMD is
  708. * issued or a new frame is recognized by the MAC AHB.
  709. */
  710. if (++ru_cnt == 2)
  711. owl_emac_dma_cmd_resume_rx(priv);
  712. recv = owl_emac_rx_process(priv, budget - work_done);
  713. /* Guard against too many RU interrupts. */
  714. if (ru_cnt > 3)
  715. break;
  716. }
  717. work_done += recv;
  718. if (work_done >= budget)
  719. break;
  720. }
  721. if (work_done < budget) {
  722. napi_complete_done(napi, work_done);
  723. owl_emac_irq_enable(priv);
  724. }
  725. /* Reset MAC when getting too many internal TX or RX errors. */
  726. if (tx_err_cnt > 10 || rx_err_cnt > 10) {
  727. netdev_dbg(priv->netdev, "%s error status: 0x%08x\n",
  728. tx_err_cnt > 10 ? "TX" : "RX", status);
  729. rx_err_cnt = 0;
  730. tx_err_cnt = 0;
  731. schedule_work(&priv->mac_reset_task);
  732. }
  733. return work_done;
  734. }
  735. static void owl_emac_mdio_clock_enable(struct owl_emac_priv *priv)
  736. {
  737. u32 val;
  738. /* Enable MDC clock generation by adjusting CLKDIV according to
  739. * the vendor implementation of the original driver.
  740. */
  741. val = owl_emac_reg_read(priv, OWL_EMAC_REG_MAC_CSR10);
  742. val &= OWL_EMAC_MSK_MAC_CSR10_CLKDIV;
  743. val |= OWL_EMAC_VAL_MAC_CSR10_CLKDIV_128 << OWL_EMAC_OFF_MAC_CSR10_CLKDIV;
  744. val |= OWL_EMAC_BIT_MAC_CSR10_SB;
  745. val |= OWL_EMAC_VAL_MAC_CSR10_OPCODE_CDS << OWL_EMAC_OFF_MAC_CSR10_OPCODE;
  746. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR10, val);
  747. }
  748. static void owl_emac_core_hw_reset(struct owl_emac_priv *priv)
  749. {
  750. /* Trigger hardware reset. */
  751. reset_control_assert(priv->reset);
  752. usleep_range(10, 20);
  753. reset_control_deassert(priv->reset);
  754. usleep_range(100, 200);
  755. }
  756. static int owl_emac_core_sw_reset(struct owl_emac_priv *priv)
  757. {
  758. u32 val;
  759. int ret;
  760. /* Trigger software reset. */
  761. owl_emac_reg_set(priv, OWL_EMAC_REG_MAC_CSR0, OWL_EMAC_BIT_MAC_CSR0_SWR);
  762. ret = readl_poll_timeout(priv->base + OWL_EMAC_REG_MAC_CSR0,
  763. val, !(val & OWL_EMAC_BIT_MAC_CSR0_SWR),
  764. OWL_EMAC_POLL_DELAY_USEC,
  765. OWL_EMAC_RESET_POLL_TIMEOUT_USEC);
  766. if (ret)
  767. return ret;
  768. if (priv->phy_mode == PHY_INTERFACE_MODE_RMII) {
  769. /* Enable RMII and use the 50MHz rmii clk as output to PHY. */
  770. val = 0;
  771. } else {
  772. /* Enable SMII and use the 125MHz rmii clk as output to PHY.
  773. * Additionally set SMII SYNC delay to 4 half cycle.
  774. */
  775. val = 0x04 << OWL_EMAC_OFF_MAC_CTRL_SSDC;
  776. val |= OWL_EMAC_BIT_MAC_CTRL_RSIS;
  777. }
  778. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CTRL, val);
  779. /* MDC is disabled after reset. */
  780. owl_emac_mdio_clock_enable(priv);
  781. /* Set FIFO pause & restart threshold levels. */
  782. val = 0x40 << OWL_EMAC_OFF_MAC_CSR19_FPTL;
  783. val |= 0x10 << OWL_EMAC_OFF_MAC_CSR19_FRTL;
  784. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR19, val);
  785. /* Set flow control pause quanta time to ~100 ms. */
  786. val = 0x4FFF << OWL_EMAC_OFF_MAC_CSR18_PQT;
  787. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR18, val);
  788. /* Setup interrupt mitigation. */
  789. val = 7 << OWL_EMAC_OFF_MAC_CSR11_NRP;
  790. val |= 4 << OWL_EMAC_OFF_MAC_CSR11_RT;
  791. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR11, val);
  792. /* Set RX/TX rings base addresses. */
  793. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR3,
  794. (u32)(priv->rx_ring.descs_dma));
  795. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR4,
  796. (u32)(priv->tx_ring.descs_dma));
  797. /* Setup initial operation mode. */
  798. val = OWL_EMAC_VAL_MAC_CSR6_SPEED_100M << OWL_EMAC_OFF_MAC_CSR6_SPEED;
  799. val |= OWL_EMAC_BIT_MAC_CSR6_FD;
  800. owl_emac_reg_update(priv, OWL_EMAC_REG_MAC_CSR6,
  801. OWL_EMAC_MSK_MAC_CSR6_SPEED |
  802. OWL_EMAC_BIT_MAC_CSR6_FD, val);
  803. owl_emac_reg_clear(priv, OWL_EMAC_REG_MAC_CSR6,
  804. OWL_EMAC_BIT_MAC_CSR6_PR | OWL_EMAC_BIT_MAC_CSR6_PM);
  805. priv->link = 0;
  806. priv->speed = SPEED_UNKNOWN;
  807. priv->duplex = DUPLEX_UNKNOWN;
  808. priv->pause = 0;
  809. priv->mcaddr_list.count = 0;
  810. return 0;
  811. }
  812. static int owl_emac_enable(struct net_device *netdev, bool start_phy)
  813. {
  814. struct owl_emac_priv *priv = netdev_priv(netdev);
  815. int ret;
  816. owl_emac_dma_cmd_stop(priv);
  817. owl_emac_irq_disable(priv);
  818. owl_emac_irq_clear(priv);
  819. owl_emac_ring_prepare_tx(priv);
  820. ret = owl_emac_ring_prepare_rx(priv);
  821. if (ret)
  822. goto err_unprep;
  823. ret = owl_emac_core_sw_reset(priv);
  824. if (ret) {
  825. netdev_err(netdev, "failed to soft reset MAC core: %d\n", ret);
  826. goto err_unprep;
  827. }
  828. owl_emac_set_hw_mac_addr(netdev);
  829. owl_emac_setup_frame_xmit(priv);
  830. netdev_reset_queue(netdev);
  831. napi_enable(&priv->napi);
  832. owl_emac_irq_enable(priv);
  833. owl_emac_dma_cmd_start(priv);
  834. if (start_phy)
  835. phy_start(netdev->phydev);
  836. netif_start_queue(netdev);
  837. return 0;
  838. err_unprep:
  839. owl_emac_ring_unprepare_rx(priv);
  840. owl_emac_ring_unprepare_tx(priv);
  841. return ret;
  842. }
  843. static void owl_emac_disable(struct net_device *netdev, bool stop_phy)
  844. {
  845. struct owl_emac_priv *priv = netdev_priv(netdev);
  846. owl_emac_dma_cmd_stop(priv);
  847. owl_emac_irq_disable(priv);
  848. netif_stop_queue(netdev);
  849. napi_disable(&priv->napi);
  850. if (stop_phy)
  851. phy_stop(netdev->phydev);
  852. owl_emac_ring_unprepare_rx(priv);
  853. owl_emac_ring_unprepare_tx(priv);
  854. }
  855. static int owl_emac_ndo_open(struct net_device *netdev)
  856. {
  857. return owl_emac_enable(netdev, true);
  858. }
  859. static int owl_emac_ndo_stop(struct net_device *netdev)
  860. {
  861. owl_emac_disable(netdev, true);
  862. return 0;
  863. }
  864. static void owl_emac_set_multicast(struct net_device *netdev, int count)
  865. {
  866. struct owl_emac_priv *priv = netdev_priv(netdev);
  867. struct netdev_hw_addr *ha;
  868. int index = 0;
  869. if (count <= 0) {
  870. priv->mcaddr_list.count = 0;
  871. return;
  872. }
  873. netdev_for_each_mc_addr(ha, netdev) {
  874. if (!is_multicast_ether_addr(ha->addr))
  875. continue;
  876. WARN_ON(index >= OWL_EMAC_MAX_MULTICAST_ADDRS);
  877. ether_addr_copy(priv->mcaddr_list.addrs[index++], ha->addr);
  878. }
  879. priv->mcaddr_list.count = index;
  880. owl_emac_setup_frame_xmit(priv);
  881. }
  882. static void owl_emac_ndo_set_rx_mode(struct net_device *netdev)
  883. {
  884. struct owl_emac_priv *priv = netdev_priv(netdev);
  885. u32 status, val = 0;
  886. int mcast_count = 0;
  887. if (netdev->flags & IFF_PROMISC) {
  888. val = OWL_EMAC_BIT_MAC_CSR6_PR;
  889. } else if (netdev->flags & IFF_ALLMULTI) {
  890. val = OWL_EMAC_BIT_MAC_CSR6_PM;
  891. } else if (netdev->flags & IFF_MULTICAST) {
  892. mcast_count = netdev_mc_count(netdev);
  893. if (mcast_count > OWL_EMAC_MAX_MULTICAST_ADDRS) {
  894. val = OWL_EMAC_BIT_MAC_CSR6_PM;
  895. mcast_count = 0;
  896. }
  897. }
  898. spin_lock_bh(&priv->lock);
  899. /* Temporarily stop DMA TX & RX. */
  900. status = owl_emac_dma_cmd_stop(priv);
  901. /* Update operation modes. */
  902. owl_emac_reg_update(priv, OWL_EMAC_REG_MAC_CSR6,
  903. OWL_EMAC_BIT_MAC_CSR6_PR | OWL_EMAC_BIT_MAC_CSR6_PM,
  904. val);
  905. /* Restore DMA TX & RX status. */
  906. owl_emac_dma_cmd_set(priv, status);
  907. spin_unlock_bh(&priv->lock);
  908. /* Set/reset multicast addr list. */
  909. owl_emac_set_multicast(netdev, mcast_count);
  910. }
  911. static int owl_emac_ndo_set_mac_addr(struct net_device *netdev, void *addr)
  912. {
  913. struct sockaddr *skaddr = addr;
  914. if (!is_valid_ether_addr(skaddr->sa_data))
  915. return -EADDRNOTAVAIL;
  916. if (netif_running(netdev))
  917. return -EBUSY;
  918. eth_hw_addr_set(netdev, skaddr->sa_data);
  919. owl_emac_set_hw_mac_addr(netdev);
  920. return owl_emac_setup_frame_xmit(netdev_priv(netdev));
  921. }
  922. static int owl_emac_ndo_eth_ioctl(struct net_device *netdev,
  923. struct ifreq *req, int cmd)
  924. {
  925. if (!netif_running(netdev))
  926. return -EINVAL;
  927. return phy_mii_ioctl(netdev->phydev, req, cmd);
  928. }
  929. static void owl_emac_ndo_tx_timeout(struct net_device *netdev,
  930. unsigned int txqueue)
  931. {
  932. struct owl_emac_priv *priv = netdev_priv(netdev);
  933. schedule_work(&priv->mac_reset_task);
  934. }
  935. static void owl_emac_reset_task(struct work_struct *work)
  936. {
  937. struct owl_emac_priv *priv;
  938. priv = container_of(work, struct owl_emac_priv, mac_reset_task);
  939. netdev_dbg(priv->netdev, "resetting MAC\n");
  940. owl_emac_disable(priv->netdev, false);
  941. owl_emac_enable(priv->netdev, false);
  942. }
  943. static struct net_device_stats *
  944. owl_emac_ndo_get_stats(struct net_device *netdev)
  945. {
  946. /* FIXME: If possible, try to get stats from MAC hardware registers
  947. * instead of tracking them manually in the driver.
  948. */
  949. return &netdev->stats;
  950. }
  951. static const struct net_device_ops owl_emac_netdev_ops = {
  952. .ndo_open = owl_emac_ndo_open,
  953. .ndo_stop = owl_emac_ndo_stop,
  954. .ndo_start_xmit = owl_emac_ndo_start_xmit,
  955. .ndo_set_rx_mode = owl_emac_ndo_set_rx_mode,
  956. .ndo_set_mac_address = owl_emac_ndo_set_mac_addr,
  957. .ndo_validate_addr = eth_validate_addr,
  958. .ndo_eth_ioctl = owl_emac_ndo_eth_ioctl,
  959. .ndo_tx_timeout = owl_emac_ndo_tx_timeout,
  960. .ndo_get_stats = owl_emac_ndo_get_stats,
  961. };
  962. static void owl_emac_ethtool_get_drvinfo(struct net_device *dev,
  963. struct ethtool_drvinfo *info)
  964. {
  965. strscpy(info->driver, OWL_EMAC_DRVNAME, sizeof(info->driver));
  966. }
  967. static u32 owl_emac_ethtool_get_msglevel(struct net_device *netdev)
  968. {
  969. struct owl_emac_priv *priv = netdev_priv(netdev);
  970. return priv->msg_enable;
  971. }
  972. static void owl_emac_ethtool_set_msglevel(struct net_device *ndev, u32 val)
  973. {
  974. struct owl_emac_priv *priv = netdev_priv(ndev);
  975. priv->msg_enable = val;
  976. }
  977. static const struct ethtool_ops owl_emac_ethtool_ops = {
  978. .get_drvinfo = owl_emac_ethtool_get_drvinfo,
  979. .get_link = ethtool_op_get_link,
  980. .get_link_ksettings = phy_ethtool_get_link_ksettings,
  981. .set_link_ksettings = phy_ethtool_set_link_ksettings,
  982. .get_msglevel = owl_emac_ethtool_get_msglevel,
  983. .set_msglevel = owl_emac_ethtool_set_msglevel,
  984. };
  985. static int owl_emac_mdio_wait(struct owl_emac_priv *priv)
  986. {
  987. u32 val;
  988. /* Wait while data transfer is in progress. */
  989. return readl_poll_timeout(priv->base + OWL_EMAC_REG_MAC_CSR10,
  990. val, !(val & OWL_EMAC_BIT_MAC_CSR10_SB),
  991. OWL_EMAC_POLL_DELAY_USEC,
  992. OWL_EMAC_MDIO_POLL_TIMEOUT_USEC);
  993. }
  994. static int owl_emac_mdio_read(struct mii_bus *bus, int addr, int regnum)
  995. {
  996. struct owl_emac_priv *priv = bus->priv;
  997. u32 data, tmp;
  998. int ret;
  999. if (regnum & MII_ADDR_C45)
  1000. return -EOPNOTSUPP;
  1001. data = OWL_EMAC_BIT_MAC_CSR10_SB;
  1002. data |= OWL_EMAC_VAL_MAC_CSR10_OPCODE_RD << OWL_EMAC_OFF_MAC_CSR10_OPCODE;
  1003. tmp = addr << OWL_EMAC_OFF_MAC_CSR10_PHYADD;
  1004. data |= tmp & OWL_EMAC_MSK_MAC_CSR10_PHYADD;
  1005. tmp = regnum << OWL_EMAC_OFF_MAC_CSR10_REGADD;
  1006. data |= tmp & OWL_EMAC_MSK_MAC_CSR10_REGADD;
  1007. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR10, data);
  1008. ret = owl_emac_mdio_wait(priv);
  1009. if (ret)
  1010. return ret;
  1011. data = owl_emac_reg_read(priv, OWL_EMAC_REG_MAC_CSR10);
  1012. data &= OWL_EMAC_MSK_MAC_CSR10_DATA;
  1013. return data;
  1014. }
  1015. static int
  1016. owl_emac_mdio_write(struct mii_bus *bus, int addr, int regnum, u16 val)
  1017. {
  1018. struct owl_emac_priv *priv = bus->priv;
  1019. u32 data, tmp;
  1020. if (regnum & MII_ADDR_C45)
  1021. return -EOPNOTSUPP;
  1022. data = OWL_EMAC_BIT_MAC_CSR10_SB;
  1023. data |= OWL_EMAC_VAL_MAC_CSR10_OPCODE_WR << OWL_EMAC_OFF_MAC_CSR10_OPCODE;
  1024. tmp = addr << OWL_EMAC_OFF_MAC_CSR10_PHYADD;
  1025. data |= tmp & OWL_EMAC_MSK_MAC_CSR10_PHYADD;
  1026. tmp = regnum << OWL_EMAC_OFF_MAC_CSR10_REGADD;
  1027. data |= tmp & OWL_EMAC_MSK_MAC_CSR10_REGADD;
  1028. data |= val & OWL_EMAC_MSK_MAC_CSR10_DATA;
  1029. owl_emac_reg_write(priv, OWL_EMAC_REG_MAC_CSR10, data);
  1030. return owl_emac_mdio_wait(priv);
  1031. }
  1032. static int owl_emac_mdio_init(struct net_device *netdev)
  1033. {
  1034. struct owl_emac_priv *priv = netdev_priv(netdev);
  1035. struct device *dev = owl_emac_get_dev(priv);
  1036. struct device_node *mdio_node;
  1037. int ret;
  1038. mdio_node = of_get_child_by_name(dev->of_node, "mdio");
  1039. if (!mdio_node)
  1040. return -ENODEV;
  1041. if (!of_device_is_available(mdio_node)) {
  1042. ret = -ENODEV;
  1043. goto err_put_node;
  1044. }
  1045. priv->mii = devm_mdiobus_alloc(dev);
  1046. if (!priv->mii) {
  1047. ret = -ENOMEM;
  1048. goto err_put_node;
  1049. }
  1050. snprintf(priv->mii->id, MII_BUS_ID_SIZE, "%s", dev_name(dev));
  1051. priv->mii->name = "owl-emac-mdio";
  1052. priv->mii->parent = dev;
  1053. priv->mii->read = owl_emac_mdio_read;
  1054. priv->mii->write = owl_emac_mdio_write;
  1055. priv->mii->phy_mask = ~0; /* Mask out all PHYs from auto probing. */
  1056. priv->mii->priv = priv;
  1057. ret = devm_of_mdiobus_register(dev, priv->mii, mdio_node);
  1058. err_put_node:
  1059. of_node_put(mdio_node);
  1060. return ret;
  1061. }
  1062. static int owl_emac_phy_init(struct net_device *netdev)
  1063. {
  1064. struct owl_emac_priv *priv = netdev_priv(netdev);
  1065. struct device *dev = owl_emac_get_dev(priv);
  1066. struct phy_device *phy;
  1067. phy = of_phy_get_and_connect(netdev, dev->of_node,
  1068. owl_emac_adjust_link);
  1069. if (!phy)
  1070. return -ENODEV;
  1071. phy_set_sym_pause(phy, true, true, true);
  1072. if (netif_msg_link(priv))
  1073. phy_attached_info(phy);
  1074. return 0;
  1075. }
  1076. static void owl_emac_get_mac_addr(struct net_device *netdev)
  1077. {
  1078. struct device *dev = netdev->dev.parent;
  1079. int ret;
  1080. ret = platform_get_ethdev_address(dev, netdev);
  1081. if (!ret && is_valid_ether_addr(netdev->dev_addr))
  1082. return;
  1083. eth_hw_addr_random(netdev);
  1084. dev_warn(dev, "using random MAC address %pM\n", netdev->dev_addr);
  1085. }
  1086. static __maybe_unused int owl_emac_suspend(struct device *dev)
  1087. {
  1088. struct net_device *netdev = dev_get_drvdata(dev);
  1089. struct owl_emac_priv *priv = netdev_priv(netdev);
  1090. disable_irq(netdev->irq);
  1091. if (netif_running(netdev)) {
  1092. owl_emac_disable(netdev, true);
  1093. netif_device_detach(netdev);
  1094. }
  1095. clk_bulk_disable_unprepare(OWL_EMAC_NCLKS, priv->clks);
  1096. return 0;
  1097. }
  1098. static __maybe_unused int owl_emac_resume(struct device *dev)
  1099. {
  1100. struct net_device *netdev = dev_get_drvdata(dev);
  1101. struct owl_emac_priv *priv = netdev_priv(netdev);
  1102. int ret;
  1103. ret = clk_bulk_prepare_enable(OWL_EMAC_NCLKS, priv->clks);
  1104. if (ret)
  1105. return ret;
  1106. if (netif_running(netdev)) {
  1107. owl_emac_core_hw_reset(priv);
  1108. owl_emac_core_sw_reset(priv);
  1109. ret = owl_emac_enable(netdev, true);
  1110. if (ret) {
  1111. clk_bulk_disable_unprepare(OWL_EMAC_NCLKS, priv->clks);
  1112. return ret;
  1113. }
  1114. netif_device_attach(netdev);
  1115. }
  1116. enable_irq(netdev->irq);
  1117. return 0;
  1118. }
  1119. static void owl_emac_clk_disable_unprepare(void *data)
  1120. {
  1121. struct owl_emac_priv *priv = data;
  1122. clk_bulk_disable_unprepare(OWL_EMAC_NCLKS, priv->clks);
  1123. }
  1124. static int owl_emac_clk_set_rate(struct owl_emac_priv *priv)
  1125. {
  1126. struct device *dev = owl_emac_get_dev(priv);
  1127. unsigned long rate;
  1128. int ret;
  1129. switch (priv->phy_mode) {
  1130. case PHY_INTERFACE_MODE_RMII:
  1131. rate = 50000000;
  1132. break;
  1133. case PHY_INTERFACE_MODE_SMII:
  1134. rate = 125000000;
  1135. break;
  1136. default:
  1137. dev_err(dev, "unsupported phy interface mode %d\n",
  1138. priv->phy_mode);
  1139. return -EOPNOTSUPP;
  1140. }
  1141. ret = clk_set_rate(priv->clks[OWL_EMAC_CLK_RMII].clk, rate);
  1142. if (ret)
  1143. dev_err(dev, "failed to set RMII clock rate: %d\n", ret);
  1144. return ret;
  1145. }
  1146. static int owl_emac_probe(struct platform_device *pdev)
  1147. {
  1148. struct device *dev = &pdev->dev;
  1149. struct net_device *netdev;
  1150. struct owl_emac_priv *priv;
  1151. int ret, i;
  1152. netdev = devm_alloc_etherdev(dev, sizeof(*priv));
  1153. if (!netdev)
  1154. return -ENOMEM;
  1155. platform_set_drvdata(pdev, netdev);
  1156. SET_NETDEV_DEV(netdev, dev);
  1157. priv = netdev_priv(netdev);
  1158. priv->netdev = netdev;
  1159. priv->msg_enable = netif_msg_init(-1, OWL_EMAC_DEFAULT_MSG_ENABLE);
  1160. ret = of_get_phy_mode(dev->of_node, &priv->phy_mode);
  1161. if (ret) {
  1162. dev_err(dev, "failed to get phy mode: %d\n", ret);
  1163. return ret;
  1164. }
  1165. spin_lock_init(&priv->lock);
  1166. ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32));
  1167. if (ret) {
  1168. dev_err(dev, "unsupported DMA mask\n");
  1169. return ret;
  1170. }
  1171. ret = owl_emac_ring_alloc(dev, &priv->rx_ring, OWL_EMAC_RX_RING_SIZE);
  1172. if (ret)
  1173. return ret;
  1174. ret = owl_emac_ring_alloc(dev, &priv->tx_ring, OWL_EMAC_TX_RING_SIZE);
  1175. if (ret)
  1176. return ret;
  1177. priv->base = devm_platform_ioremap_resource(pdev, 0);
  1178. if (IS_ERR(priv->base))
  1179. return PTR_ERR(priv->base);
  1180. netdev->irq = platform_get_irq(pdev, 0);
  1181. if (netdev->irq < 0)
  1182. return netdev->irq;
  1183. ret = devm_request_irq(dev, netdev->irq, owl_emac_handle_irq,
  1184. IRQF_SHARED, netdev->name, netdev);
  1185. if (ret) {
  1186. dev_err(dev, "failed to request irq: %d\n", netdev->irq);
  1187. return ret;
  1188. }
  1189. for (i = 0; i < OWL_EMAC_NCLKS; i++)
  1190. priv->clks[i].id = owl_emac_clk_names[i];
  1191. ret = devm_clk_bulk_get(dev, OWL_EMAC_NCLKS, priv->clks);
  1192. if (ret)
  1193. return ret;
  1194. ret = clk_bulk_prepare_enable(OWL_EMAC_NCLKS, priv->clks);
  1195. if (ret)
  1196. return ret;
  1197. ret = devm_add_action_or_reset(dev, owl_emac_clk_disable_unprepare, priv);
  1198. if (ret)
  1199. return ret;
  1200. ret = owl_emac_clk_set_rate(priv);
  1201. if (ret)
  1202. return ret;
  1203. priv->reset = devm_reset_control_get_exclusive(dev, NULL);
  1204. if (IS_ERR(priv->reset))
  1205. return dev_err_probe(dev, PTR_ERR(priv->reset),
  1206. "failed to get reset control");
  1207. owl_emac_get_mac_addr(netdev);
  1208. owl_emac_core_hw_reset(priv);
  1209. owl_emac_mdio_clock_enable(priv);
  1210. ret = owl_emac_mdio_init(netdev);
  1211. if (ret) {
  1212. dev_err(dev, "failed to initialize MDIO bus\n");
  1213. return ret;
  1214. }
  1215. ret = owl_emac_phy_init(netdev);
  1216. if (ret) {
  1217. dev_err(dev, "failed to initialize PHY\n");
  1218. return ret;
  1219. }
  1220. INIT_WORK(&priv->mac_reset_task, owl_emac_reset_task);
  1221. netdev->min_mtu = OWL_EMAC_MTU_MIN;
  1222. netdev->max_mtu = OWL_EMAC_MTU_MAX;
  1223. netdev->watchdog_timeo = OWL_EMAC_TX_TIMEOUT;
  1224. netdev->netdev_ops = &owl_emac_netdev_ops;
  1225. netdev->ethtool_ops = &owl_emac_ethtool_ops;
  1226. netif_napi_add(netdev, &priv->napi, owl_emac_poll);
  1227. ret = devm_register_netdev(dev, netdev);
  1228. if (ret) {
  1229. netif_napi_del(&priv->napi);
  1230. phy_disconnect(netdev->phydev);
  1231. return ret;
  1232. }
  1233. return 0;
  1234. }
  1235. static int owl_emac_remove(struct platform_device *pdev)
  1236. {
  1237. struct owl_emac_priv *priv = platform_get_drvdata(pdev);
  1238. netif_napi_del(&priv->napi);
  1239. phy_disconnect(priv->netdev->phydev);
  1240. cancel_work_sync(&priv->mac_reset_task);
  1241. return 0;
  1242. }
  1243. static const struct of_device_id owl_emac_of_match[] = {
  1244. { .compatible = "actions,owl-emac", },
  1245. { }
  1246. };
  1247. MODULE_DEVICE_TABLE(of, owl_emac_of_match);
  1248. static SIMPLE_DEV_PM_OPS(owl_emac_pm_ops,
  1249. owl_emac_suspend, owl_emac_resume);
  1250. static struct platform_driver owl_emac_driver = {
  1251. .driver = {
  1252. .name = OWL_EMAC_DRVNAME,
  1253. .of_match_table = owl_emac_of_match,
  1254. .pm = &owl_emac_pm_ops,
  1255. },
  1256. .probe = owl_emac_probe,
  1257. .remove = owl_emac_remove,
  1258. };
  1259. module_platform_driver(owl_emac_driver);
  1260. MODULE_DESCRIPTION("Actions Semi Owl SoCs Ethernet MAC Driver");
  1261. MODULE_AUTHOR("Actions Semi Inc.");
  1262. MODULE_AUTHOR("Cristian Ciocaltea <[email protected]>");
  1263. MODULE_LICENSE("GPL");