sas_scsi_host.c 32 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279
  1. // SPDX-License-Identifier: GPL-2.0-only
  2. /*
  3. * Serial Attached SCSI (SAS) class SCSI Host glue.
  4. *
  5. * Copyright (C) 2005 Adaptec, Inc. All rights reserved.
  6. * Copyright (C) 2005 Luben Tuikov <[email protected]>
  7. */
  8. #include <linux/kthread.h>
  9. #include <linux/firmware.h>
  10. #include <linux/export.h>
  11. #include <linux/ctype.h>
  12. #include <linux/kernel.h>
  13. #include "sas_internal.h"
  14. #include <scsi/scsi_host.h>
  15. #include <scsi/scsi_device.h>
  16. #include <scsi/scsi_tcq.h>
  17. #include <scsi/scsi.h>
  18. #include <scsi/scsi_eh.h>
  19. #include <scsi/scsi_transport.h>
  20. #include <scsi/scsi_transport_sas.h>
  21. #include <scsi/sas_ata.h>
  22. #include "scsi_sas_internal.h"
  23. #include "scsi_transport_api.h"
  24. #include "scsi_priv.h"
  25. #include <linux/err.h>
  26. #include <linux/blkdev.h>
  27. #include <linux/freezer.h>
  28. #include <linux/gfp.h>
  29. #include <linux/scatterlist.h>
  30. #include <linux/libata.h>
  31. /* record final status and free the task */
  32. static void sas_end_task(struct scsi_cmnd *sc, struct sas_task *task)
  33. {
  34. struct task_status_struct *ts = &task->task_status;
  35. enum scsi_host_status hs = DID_OK;
  36. enum exec_status stat = SAS_SAM_STAT_GOOD;
  37. if (ts->resp == SAS_TASK_UNDELIVERED) {
  38. /* transport error */
  39. hs = DID_NO_CONNECT;
  40. } else { /* ts->resp == SAS_TASK_COMPLETE */
  41. /* task delivered, what happened afterwards? */
  42. switch (ts->stat) {
  43. case SAS_DEV_NO_RESPONSE:
  44. case SAS_INTERRUPTED:
  45. case SAS_PHY_DOWN:
  46. case SAS_NAK_R_ERR:
  47. case SAS_OPEN_TO:
  48. hs = DID_NO_CONNECT;
  49. break;
  50. case SAS_DATA_UNDERRUN:
  51. scsi_set_resid(sc, ts->residual);
  52. if (scsi_bufflen(sc) - scsi_get_resid(sc) < sc->underflow)
  53. hs = DID_ERROR;
  54. break;
  55. case SAS_DATA_OVERRUN:
  56. hs = DID_ERROR;
  57. break;
  58. case SAS_QUEUE_FULL:
  59. hs = DID_SOFT_ERROR; /* retry */
  60. break;
  61. case SAS_DEVICE_UNKNOWN:
  62. hs = DID_BAD_TARGET;
  63. break;
  64. case SAS_OPEN_REJECT:
  65. if (ts->open_rej_reason == SAS_OREJ_RSVD_RETRY)
  66. hs = DID_SOFT_ERROR; /* retry */
  67. else
  68. hs = DID_ERROR;
  69. break;
  70. case SAS_PROTO_RESPONSE:
  71. pr_notice("LLDD:%s sent SAS_PROTO_RESP for an SSP task; please report this\n",
  72. task->dev->port->ha->sas_ha_name);
  73. break;
  74. case SAS_ABORTED_TASK:
  75. hs = DID_ABORT;
  76. break;
  77. case SAS_SAM_STAT_CHECK_CONDITION:
  78. memcpy(sc->sense_buffer, ts->buf,
  79. min(SCSI_SENSE_BUFFERSIZE, ts->buf_valid_size));
  80. stat = SAS_SAM_STAT_CHECK_CONDITION;
  81. break;
  82. default:
  83. stat = ts->stat;
  84. break;
  85. }
  86. }
  87. sc->result = (hs << 16) | stat;
  88. ASSIGN_SAS_TASK(sc, NULL);
  89. sas_free_task(task);
  90. }
  91. static void sas_scsi_task_done(struct sas_task *task)
  92. {
  93. struct scsi_cmnd *sc = task->uldd_task;
  94. struct domain_device *dev = task->dev;
  95. struct sas_ha_struct *ha = dev->port->ha;
  96. unsigned long flags;
  97. spin_lock_irqsave(&dev->done_lock, flags);
  98. if (test_bit(SAS_HA_FROZEN, &ha->state))
  99. task = NULL;
  100. else
  101. ASSIGN_SAS_TASK(sc, NULL);
  102. spin_unlock_irqrestore(&dev->done_lock, flags);
  103. if (unlikely(!task)) {
  104. /* task will be completed by the error handler */
  105. pr_debug("task done but aborted\n");
  106. return;
  107. }
  108. if (unlikely(!sc)) {
  109. pr_debug("task_done called with non existing SCSI cmnd!\n");
  110. sas_free_task(task);
  111. return;
  112. }
  113. sas_end_task(sc, task);
  114. scsi_done(sc);
  115. }
  116. static struct sas_task *sas_create_task(struct scsi_cmnd *cmd,
  117. struct domain_device *dev,
  118. gfp_t gfp_flags)
  119. {
  120. struct sas_task *task = sas_alloc_task(gfp_flags);
  121. struct scsi_lun lun;
  122. if (!task)
  123. return NULL;
  124. task->uldd_task = cmd;
  125. ASSIGN_SAS_TASK(cmd, task);
  126. task->dev = dev;
  127. task->task_proto = task->dev->tproto; /* BUG_ON(!SSP) */
  128. task->ssp_task.retry_count = 1;
  129. int_to_scsilun(cmd->device->lun, &lun);
  130. memcpy(task->ssp_task.LUN, &lun.scsi_lun, 8);
  131. task->ssp_task.task_attr = TASK_ATTR_SIMPLE;
  132. task->ssp_task.cmd = cmd;
  133. task->scatter = scsi_sglist(cmd);
  134. task->num_scatter = scsi_sg_count(cmd);
  135. task->total_xfer_len = scsi_bufflen(cmd);
  136. task->data_dir = cmd->sc_data_direction;
  137. task->task_done = sas_scsi_task_done;
  138. return task;
  139. }
  140. int sas_queuecommand(struct Scsi_Host *host, struct scsi_cmnd *cmd)
  141. {
  142. struct sas_internal *i = to_sas_internal(host->transportt);
  143. struct domain_device *dev = cmd_to_domain_dev(cmd);
  144. struct sas_task *task;
  145. int res = 0;
  146. /* If the device fell off, no sense in issuing commands */
  147. if (test_bit(SAS_DEV_GONE, &dev->state)) {
  148. cmd->result = DID_BAD_TARGET << 16;
  149. goto out_done;
  150. }
  151. if (dev_is_sata(dev)) {
  152. spin_lock_irq(dev->sata_dev.ap->lock);
  153. res = ata_sas_queuecmd(cmd, dev->sata_dev.ap);
  154. spin_unlock_irq(dev->sata_dev.ap->lock);
  155. return res;
  156. }
  157. task = sas_create_task(cmd, dev, GFP_ATOMIC);
  158. if (!task)
  159. return SCSI_MLQUEUE_HOST_BUSY;
  160. res = i->dft->lldd_execute_task(task, GFP_ATOMIC);
  161. if (res)
  162. goto out_free_task;
  163. return 0;
  164. out_free_task:
  165. pr_debug("lldd_execute_task returned: %d\n", res);
  166. ASSIGN_SAS_TASK(cmd, NULL);
  167. sas_free_task(task);
  168. if (res == -SAS_QUEUE_FULL)
  169. cmd->result = DID_SOFT_ERROR << 16; /* retry */
  170. else
  171. cmd->result = DID_ERROR << 16;
  172. out_done:
  173. scsi_done(cmd);
  174. return 0;
  175. }
  176. EXPORT_SYMBOL_GPL(sas_queuecommand);
  177. static void sas_eh_finish_cmd(struct scsi_cmnd *cmd)
  178. {
  179. struct sas_ha_struct *sas_ha = SHOST_TO_SAS_HA(cmd->device->host);
  180. struct domain_device *dev = cmd_to_domain_dev(cmd);
  181. struct sas_task *task = TO_SAS_TASK(cmd);
  182. /* At this point, we only get called following an actual abort
  183. * of the task, so we should be guaranteed not to be racing with
  184. * any completions from the LLD. Task is freed after this.
  185. */
  186. sas_end_task(cmd, task);
  187. if (dev_is_sata(dev)) {
  188. /* defer commands to libata so that libata EH can
  189. * handle ata qcs correctly
  190. */
  191. list_move_tail(&cmd->eh_entry, &sas_ha->eh_ata_q);
  192. return;
  193. }
  194. /* now finish the command and move it on to the error
  195. * handler done list, this also takes it off the
  196. * error handler pending list.
  197. */
  198. scsi_eh_finish_cmd(cmd, &sas_ha->eh_done_q);
  199. }
  200. static void sas_scsi_clear_queue_lu(struct list_head *error_q, struct scsi_cmnd *my_cmd)
  201. {
  202. struct scsi_cmnd *cmd, *n;
  203. list_for_each_entry_safe(cmd, n, error_q, eh_entry) {
  204. if (cmd->device->sdev_target == my_cmd->device->sdev_target &&
  205. cmd->device->lun == my_cmd->device->lun)
  206. sas_eh_finish_cmd(cmd);
  207. }
  208. }
  209. static void sas_scsi_clear_queue_I_T(struct list_head *error_q,
  210. struct domain_device *dev)
  211. {
  212. struct scsi_cmnd *cmd, *n;
  213. list_for_each_entry_safe(cmd, n, error_q, eh_entry) {
  214. struct domain_device *x = cmd_to_domain_dev(cmd);
  215. if (x == dev)
  216. sas_eh_finish_cmd(cmd);
  217. }
  218. }
  219. static void sas_scsi_clear_queue_port(struct list_head *error_q,
  220. struct asd_sas_port *port)
  221. {
  222. struct scsi_cmnd *cmd, *n;
  223. list_for_each_entry_safe(cmd, n, error_q, eh_entry) {
  224. struct domain_device *dev = cmd_to_domain_dev(cmd);
  225. struct asd_sas_port *x = dev->port;
  226. if (x == port)
  227. sas_eh_finish_cmd(cmd);
  228. }
  229. }
  230. enum task_disposition {
  231. TASK_IS_DONE,
  232. TASK_IS_ABORTED,
  233. TASK_IS_AT_LU,
  234. TASK_IS_NOT_AT_LU,
  235. TASK_ABORT_FAILED,
  236. };
  237. static enum task_disposition sas_scsi_find_task(struct sas_task *task)
  238. {
  239. unsigned long flags;
  240. int i, res;
  241. struct sas_internal *si =
  242. to_sas_internal(task->dev->port->ha->core.shost->transportt);
  243. for (i = 0; i < 5; i++) {
  244. pr_notice("%s: aborting task 0x%p\n", __func__, task);
  245. res = si->dft->lldd_abort_task(task);
  246. spin_lock_irqsave(&task->task_state_lock, flags);
  247. if (task->task_state_flags & SAS_TASK_STATE_DONE) {
  248. spin_unlock_irqrestore(&task->task_state_lock, flags);
  249. pr_debug("%s: task 0x%p is done\n", __func__, task);
  250. return TASK_IS_DONE;
  251. }
  252. spin_unlock_irqrestore(&task->task_state_lock, flags);
  253. if (res == TMF_RESP_FUNC_COMPLETE) {
  254. pr_notice("%s: task 0x%p is aborted\n",
  255. __func__, task);
  256. return TASK_IS_ABORTED;
  257. } else if (si->dft->lldd_query_task) {
  258. pr_notice("%s: querying task 0x%p\n", __func__, task);
  259. res = si->dft->lldd_query_task(task);
  260. switch (res) {
  261. case TMF_RESP_FUNC_SUCC:
  262. pr_notice("%s: task 0x%p at LU\n", __func__,
  263. task);
  264. return TASK_IS_AT_LU;
  265. case TMF_RESP_FUNC_COMPLETE:
  266. pr_notice("%s: task 0x%p not at LU\n",
  267. __func__, task);
  268. return TASK_IS_NOT_AT_LU;
  269. case TMF_RESP_FUNC_FAILED:
  270. pr_notice("%s: task 0x%p failed to abort\n",
  271. __func__, task);
  272. return TASK_ABORT_FAILED;
  273. default:
  274. pr_notice("%s: task 0x%p result code %d not handled\n",
  275. __func__, task, res);
  276. }
  277. }
  278. }
  279. return TASK_ABORT_FAILED;
  280. }
  281. static int sas_recover_lu(struct domain_device *dev, struct scsi_cmnd *cmd)
  282. {
  283. int res = TMF_RESP_FUNC_FAILED;
  284. struct scsi_lun lun;
  285. struct sas_internal *i =
  286. to_sas_internal(dev->port->ha->core.shost->transportt);
  287. int_to_scsilun(cmd->device->lun, &lun);
  288. pr_notice("eh: device %016llx LUN 0x%llx has the task\n",
  289. SAS_ADDR(dev->sas_addr),
  290. cmd->device->lun);
  291. if (i->dft->lldd_abort_task_set)
  292. res = i->dft->lldd_abort_task_set(dev, lun.scsi_lun);
  293. if (res == TMF_RESP_FUNC_FAILED) {
  294. if (i->dft->lldd_clear_task_set)
  295. res = i->dft->lldd_clear_task_set(dev, lun.scsi_lun);
  296. }
  297. if (res == TMF_RESP_FUNC_FAILED) {
  298. if (i->dft->lldd_lu_reset)
  299. res = i->dft->lldd_lu_reset(dev, lun.scsi_lun);
  300. }
  301. return res;
  302. }
  303. static int sas_recover_I_T(struct domain_device *dev)
  304. {
  305. int res = TMF_RESP_FUNC_FAILED;
  306. struct sas_internal *i =
  307. to_sas_internal(dev->port->ha->core.shost->transportt);
  308. pr_notice("I_T nexus reset for dev %016llx\n",
  309. SAS_ADDR(dev->sas_addr));
  310. if (i->dft->lldd_I_T_nexus_reset)
  311. res = i->dft->lldd_I_T_nexus_reset(dev);
  312. return res;
  313. }
  314. /* take a reference on the last known good phy for this device */
  315. struct sas_phy *sas_get_local_phy(struct domain_device *dev)
  316. {
  317. struct sas_ha_struct *ha = dev->port->ha;
  318. struct sas_phy *phy;
  319. unsigned long flags;
  320. /* a published domain device always has a valid phy, it may be
  321. * stale, but it is never NULL
  322. */
  323. BUG_ON(!dev->phy);
  324. spin_lock_irqsave(&ha->phy_port_lock, flags);
  325. phy = dev->phy;
  326. get_device(&phy->dev);
  327. spin_unlock_irqrestore(&ha->phy_port_lock, flags);
  328. return phy;
  329. }
  330. EXPORT_SYMBOL_GPL(sas_get_local_phy);
  331. static void sas_wait_eh(struct domain_device *dev)
  332. {
  333. struct sas_ha_struct *ha = dev->port->ha;
  334. DEFINE_WAIT(wait);
  335. if (dev_is_sata(dev)) {
  336. ata_port_wait_eh(dev->sata_dev.ap);
  337. return;
  338. }
  339. retry:
  340. spin_lock_irq(&ha->lock);
  341. while (test_bit(SAS_DEV_EH_PENDING, &dev->state)) {
  342. prepare_to_wait(&ha->eh_wait_q, &wait, TASK_UNINTERRUPTIBLE);
  343. spin_unlock_irq(&ha->lock);
  344. schedule();
  345. spin_lock_irq(&ha->lock);
  346. }
  347. finish_wait(&ha->eh_wait_q, &wait);
  348. spin_unlock_irq(&ha->lock);
  349. /* make sure SCSI EH is complete */
  350. if (scsi_host_in_recovery(ha->core.shost)) {
  351. msleep(10);
  352. goto retry;
  353. }
  354. }
  355. static int sas_queue_reset(struct domain_device *dev, int reset_type,
  356. u64 lun, int wait)
  357. {
  358. struct sas_ha_struct *ha = dev->port->ha;
  359. int scheduled = 0, tries = 100;
  360. /* ata: promote lun reset to bus reset */
  361. if (dev_is_sata(dev)) {
  362. sas_ata_schedule_reset(dev);
  363. if (wait)
  364. sas_ata_wait_eh(dev);
  365. return SUCCESS;
  366. }
  367. while (!scheduled && tries--) {
  368. spin_lock_irq(&ha->lock);
  369. if (!test_bit(SAS_DEV_EH_PENDING, &dev->state) &&
  370. !test_bit(reset_type, &dev->state)) {
  371. scheduled = 1;
  372. ha->eh_active++;
  373. list_add_tail(&dev->ssp_dev.eh_list_node, &ha->eh_dev_q);
  374. set_bit(SAS_DEV_EH_PENDING, &dev->state);
  375. set_bit(reset_type, &dev->state);
  376. int_to_scsilun(lun, &dev->ssp_dev.reset_lun);
  377. scsi_schedule_eh(ha->core.shost);
  378. }
  379. spin_unlock_irq(&ha->lock);
  380. if (wait)
  381. sas_wait_eh(dev);
  382. if (scheduled)
  383. return SUCCESS;
  384. }
  385. pr_warn("%s reset of %s failed\n",
  386. reset_type == SAS_DEV_LU_RESET ? "LUN" : "Bus",
  387. dev_name(&dev->rphy->dev));
  388. return FAILED;
  389. }
  390. int sas_eh_abort_handler(struct scsi_cmnd *cmd)
  391. {
  392. int res = TMF_RESP_FUNC_FAILED;
  393. struct sas_task *task = TO_SAS_TASK(cmd);
  394. struct Scsi_Host *host = cmd->device->host;
  395. struct domain_device *dev = cmd_to_domain_dev(cmd);
  396. struct sas_internal *i = to_sas_internal(host->transportt);
  397. unsigned long flags;
  398. if (!i->dft->lldd_abort_task)
  399. return FAILED;
  400. spin_lock_irqsave(host->host_lock, flags);
  401. /* We cannot do async aborts for SATA devices */
  402. if (dev_is_sata(dev) && !host->host_eh_scheduled) {
  403. spin_unlock_irqrestore(host->host_lock, flags);
  404. return FAILED;
  405. }
  406. spin_unlock_irqrestore(host->host_lock, flags);
  407. if (task)
  408. res = i->dft->lldd_abort_task(task);
  409. else
  410. pr_notice("no task to abort\n");
  411. if (res == TMF_RESP_FUNC_SUCC || res == TMF_RESP_FUNC_COMPLETE)
  412. return SUCCESS;
  413. return FAILED;
  414. }
  415. EXPORT_SYMBOL_GPL(sas_eh_abort_handler);
  416. /* Attempt to send a LUN reset message to a device */
  417. int sas_eh_device_reset_handler(struct scsi_cmnd *cmd)
  418. {
  419. int res;
  420. struct scsi_lun lun;
  421. struct Scsi_Host *host = cmd->device->host;
  422. struct domain_device *dev = cmd_to_domain_dev(cmd);
  423. struct sas_internal *i = to_sas_internal(host->transportt);
  424. if (current != host->ehandler)
  425. return sas_queue_reset(dev, SAS_DEV_LU_RESET, cmd->device->lun, 0);
  426. int_to_scsilun(cmd->device->lun, &lun);
  427. if (!i->dft->lldd_lu_reset)
  428. return FAILED;
  429. res = i->dft->lldd_lu_reset(dev, lun.scsi_lun);
  430. if (res == TMF_RESP_FUNC_SUCC || res == TMF_RESP_FUNC_COMPLETE)
  431. return SUCCESS;
  432. return FAILED;
  433. }
  434. EXPORT_SYMBOL_GPL(sas_eh_device_reset_handler);
  435. int sas_eh_target_reset_handler(struct scsi_cmnd *cmd)
  436. {
  437. int res;
  438. struct Scsi_Host *host = cmd->device->host;
  439. struct domain_device *dev = cmd_to_domain_dev(cmd);
  440. struct sas_internal *i = to_sas_internal(host->transportt);
  441. if (current != host->ehandler)
  442. return sas_queue_reset(dev, SAS_DEV_RESET, 0, 0);
  443. if (!i->dft->lldd_I_T_nexus_reset)
  444. return FAILED;
  445. res = i->dft->lldd_I_T_nexus_reset(dev);
  446. if (res == TMF_RESP_FUNC_SUCC || res == TMF_RESP_FUNC_COMPLETE ||
  447. res == -ENODEV)
  448. return SUCCESS;
  449. return FAILED;
  450. }
  451. EXPORT_SYMBOL_GPL(sas_eh_target_reset_handler);
  452. /* Try to reset a device */
  453. static int try_to_reset_cmd_device(struct scsi_cmnd *cmd)
  454. {
  455. int res;
  456. struct Scsi_Host *shost = cmd->device->host;
  457. if (!shost->hostt->eh_device_reset_handler)
  458. goto try_target_reset;
  459. res = shost->hostt->eh_device_reset_handler(cmd);
  460. if (res == SUCCESS)
  461. return res;
  462. try_target_reset:
  463. if (shost->hostt->eh_target_reset_handler)
  464. return shost->hostt->eh_target_reset_handler(cmd);
  465. return FAILED;
  466. }
  467. static void sas_eh_handle_sas_errors(struct Scsi_Host *shost, struct list_head *work_q)
  468. {
  469. struct scsi_cmnd *cmd, *n;
  470. enum task_disposition res = TASK_IS_DONE;
  471. int tmf_resp, need_reset;
  472. struct sas_internal *i = to_sas_internal(shost->transportt);
  473. unsigned long flags;
  474. struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
  475. LIST_HEAD(done);
  476. /* clean out any commands that won the completion vs eh race */
  477. list_for_each_entry_safe(cmd, n, work_q, eh_entry) {
  478. struct domain_device *dev = cmd_to_domain_dev(cmd);
  479. struct sas_task *task;
  480. spin_lock_irqsave(&dev->done_lock, flags);
  481. /* by this point the lldd has either observed
  482. * SAS_HA_FROZEN and is leaving the task alone, or has
  483. * won the race with eh and decided to complete it
  484. */
  485. task = TO_SAS_TASK(cmd);
  486. spin_unlock_irqrestore(&dev->done_lock, flags);
  487. if (!task)
  488. list_move_tail(&cmd->eh_entry, &done);
  489. }
  490. Again:
  491. list_for_each_entry_safe(cmd, n, work_q, eh_entry) {
  492. struct sas_task *task = TO_SAS_TASK(cmd);
  493. list_del_init(&cmd->eh_entry);
  494. spin_lock_irqsave(&task->task_state_lock, flags);
  495. need_reset = task->task_state_flags & SAS_TASK_NEED_DEV_RESET;
  496. spin_unlock_irqrestore(&task->task_state_lock, flags);
  497. if (need_reset) {
  498. pr_notice("%s: task 0x%p requests reset\n",
  499. __func__, task);
  500. goto reset;
  501. }
  502. pr_debug("trying to find task 0x%p\n", task);
  503. res = sas_scsi_find_task(task);
  504. switch (res) {
  505. case TASK_IS_DONE:
  506. pr_notice("%s: task 0x%p is done\n", __func__,
  507. task);
  508. sas_eh_finish_cmd(cmd);
  509. continue;
  510. case TASK_IS_ABORTED:
  511. pr_notice("%s: task 0x%p is aborted\n",
  512. __func__, task);
  513. sas_eh_finish_cmd(cmd);
  514. continue;
  515. case TASK_IS_AT_LU:
  516. pr_info("task 0x%p is at LU: lu recover\n", task);
  517. reset:
  518. tmf_resp = sas_recover_lu(task->dev, cmd);
  519. if (tmf_resp == TMF_RESP_FUNC_COMPLETE) {
  520. pr_notice("dev %016llx LU 0x%llx is recovered\n",
  521. SAS_ADDR(task->dev),
  522. cmd->device->lun);
  523. sas_eh_finish_cmd(cmd);
  524. sas_scsi_clear_queue_lu(work_q, cmd);
  525. goto Again;
  526. }
  527. fallthrough;
  528. case TASK_IS_NOT_AT_LU:
  529. case TASK_ABORT_FAILED:
  530. pr_notice("task 0x%p is not at LU: I_T recover\n",
  531. task);
  532. tmf_resp = sas_recover_I_T(task->dev);
  533. if (tmf_resp == TMF_RESP_FUNC_COMPLETE ||
  534. tmf_resp == -ENODEV) {
  535. struct domain_device *dev = task->dev;
  536. pr_notice("I_T %016llx recovered\n",
  537. SAS_ADDR(task->dev->sas_addr));
  538. sas_eh_finish_cmd(cmd);
  539. sas_scsi_clear_queue_I_T(work_q, dev);
  540. goto Again;
  541. }
  542. /* Hammer time :-) */
  543. try_to_reset_cmd_device(cmd);
  544. if (i->dft->lldd_clear_nexus_port) {
  545. struct asd_sas_port *port = task->dev->port;
  546. pr_debug("clearing nexus for port:%d\n",
  547. port->id);
  548. res = i->dft->lldd_clear_nexus_port(port);
  549. if (res == TMF_RESP_FUNC_COMPLETE) {
  550. pr_notice("clear nexus port:%d succeeded\n",
  551. port->id);
  552. sas_eh_finish_cmd(cmd);
  553. sas_scsi_clear_queue_port(work_q,
  554. port);
  555. goto Again;
  556. }
  557. }
  558. if (i->dft->lldd_clear_nexus_ha) {
  559. pr_debug("clear nexus ha\n");
  560. res = i->dft->lldd_clear_nexus_ha(ha);
  561. if (res == TMF_RESP_FUNC_COMPLETE) {
  562. pr_notice("clear nexus ha succeeded\n");
  563. sas_eh_finish_cmd(cmd);
  564. goto clear_q;
  565. }
  566. }
  567. /* If we are here -- this means that no amount
  568. * of effort could recover from errors. Quite
  569. * possibly the HA just disappeared.
  570. */
  571. pr_err("error from device %016llx, LUN 0x%llx couldn't be recovered in any way\n",
  572. SAS_ADDR(task->dev->sas_addr),
  573. cmd->device->lun);
  574. sas_eh_finish_cmd(cmd);
  575. goto clear_q;
  576. }
  577. }
  578. out:
  579. list_splice_tail(&done, work_q);
  580. list_splice_tail_init(&ha->eh_ata_q, work_q);
  581. return;
  582. clear_q:
  583. pr_debug("--- Exit %s -- clear_q\n", __func__);
  584. list_for_each_entry_safe(cmd, n, work_q, eh_entry)
  585. sas_eh_finish_cmd(cmd);
  586. goto out;
  587. }
  588. static void sas_eh_handle_resets(struct Scsi_Host *shost)
  589. {
  590. struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
  591. struct sas_internal *i = to_sas_internal(shost->transportt);
  592. /* handle directed resets to sas devices */
  593. spin_lock_irq(&ha->lock);
  594. while (!list_empty(&ha->eh_dev_q)) {
  595. struct domain_device *dev;
  596. struct ssp_device *ssp;
  597. ssp = list_entry(ha->eh_dev_q.next, typeof(*ssp), eh_list_node);
  598. list_del_init(&ssp->eh_list_node);
  599. dev = container_of(ssp, typeof(*dev), ssp_dev);
  600. kref_get(&dev->kref);
  601. WARN_ONCE(dev_is_sata(dev), "ssp reset to ata device?\n");
  602. spin_unlock_irq(&ha->lock);
  603. if (test_and_clear_bit(SAS_DEV_LU_RESET, &dev->state))
  604. i->dft->lldd_lu_reset(dev, ssp->reset_lun.scsi_lun);
  605. if (test_and_clear_bit(SAS_DEV_RESET, &dev->state))
  606. i->dft->lldd_I_T_nexus_reset(dev);
  607. sas_put_device(dev);
  608. spin_lock_irq(&ha->lock);
  609. clear_bit(SAS_DEV_EH_PENDING, &dev->state);
  610. ha->eh_active--;
  611. }
  612. spin_unlock_irq(&ha->lock);
  613. }
  614. void sas_scsi_recover_host(struct Scsi_Host *shost)
  615. {
  616. struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
  617. LIST_HEAD(eh_work_q);
  618. int tries = 0;
  619. bool retry;
  620. retry:
  621. tries++;
  622. retry = true;
  623. spin_lock_irq(shost->host_lock);
  624. list_splice_init(&shost->eh_cmd_q, &eh_work_q);
  625. spin_unlock_irq(shost->host_lock);
  626. pr_notice("Enter %s busy: %d failed: %d\n",
  627. __func__, scsi_host_busy(shost), shost->host_failed);
  628. /*
  629. * Deal with commands that still have SAS tasks (i.e. they didn't
  630. * complete via the normal sas_task completion mechanism),
  631. * SAS_HA_FROZEN gives eh dominion over all sas_task completion.
  632. */
  633. set_bit(SAS_HA_FROZEN, &ha->state);
  634. sas_eh_handle_sas_errors(shost, &eh_work_q);
  635. clear_bit(SAS_HA_FROZEN, &ha->state);
  636. if (list_empty(&eh_work_q))
  637. goto out;
  638. /*
  639. * Now deal with SCSI commands that completed ok but have a an error
  640. * code (and hopefully sense data) attached. This is roughly what
  641. * scsi_unjam_host does, but we skip scsi_eh_abort_cmds because any
  642. * command we see here has no sas_task and is thus unknown to the HA.
  643. */
  644. sas_ata_eh(shost, &eh_work_q);
  645. if (!scsi_eh_get_sense(&eh_work_q, &ha->eh_done_q))
  646. scsi_eh_ready_devs(shost, &eh_work_q, &ha->eh_done_q);
  647. out:
  648. sas_eh_handle_resets(shost);
  649. /* now link into libata eh --- if we have any ata devices */
  650. sas_ata_strategy_handler(shost);
  651. scsi_eh_flush_done_q(&ha->eh_done_q);
  652. /* check if any new eh work was scheduled during the last run */
  653. spin_lock_irq(&ha->lock);
  654. if (ha->eh_active == 0) {
  655. shost->host_eh_scheduled = 0;
  656. retry = false;
  657. }
  658. spin_unlock_irq(&ha->lock);
  659. if (retry)
  660. goto retry;
  661. pr_notice("--- Exit %s: busy: %d failed: %d tries: %d\n",
  662. __func__, scsi_host_busy(shost),
  663. shost->host_failed, tries);
  664. }
  665. int sas_ioctl(struct scsi_device *sdev, unsigned int cmd, void __user *arg)
  666. {
  667. struct domain_device *dev = sdev_to_domain_dev(sdev);
  668. if (dev_is_sata(dev))
  669. return ata_sas_scsi_ioctl(dev->sata_dev.ap, sdev, cmd, arg);
  670. return -EINVAL;
  671. }
  672. EXPORT_SYMBOL_GPL(sas_ioctl);
  673. struct domain_device *sas_find_dev_by_rphy(struct sas_rphy *rphy)
  674. {
  675. struct Scsi_Host *shost = dev_to_shost(rphy->dev.parent);
  676. struct sas_ha_struct *ha = SHOST_TO_SAS_HA(shost);
  677. struct domain_device *found_dev = NULL;
  678. int i;
  679. unsigned long flags;
  680. spin_lock_irqsave(&ha->phy_port_lock, flags);
  681. for (i = 0; i < ha->num_phys; i++) {
  682. struct asd_sas_port *port = ha->sas_port[i];
  683. struct domain_device *dev;
  684. spin_lock(&port->dev_list_lock);
  685. list_for_each_entry(dev, &port->dev_list, dev_list_node) {
  686. if (rphy == dev->rphy) {
  687. found_dev = dev;
  688. spin_unlock(&port->dev_list_lock);
  689. goto found;
  690. }
  691. }
  692. spin_unlock(&port->dev_list_lock);
  693. }
  694. found:
  695. spin_unlock_irqrestore(&ha->phy_port_lock, flags);
  696. return found_dev;
  697. }
  698. int sas_target_alloc(struct scsi_target *starget)
  699. {
  700. struct sas_rphy *rphy = dev_to_rphy(starget->dev.parent);
  701. struct domain_device *found_dev = sas_find_dev_by_rphy(rphy);
  702. if (!found_dev)
  703. return -ENODEV;
  704. kref_get(&found_dev->kref);
  705. starget->hostdata = found_dev;
  706. return 0;
  707. }
  708. EXPORT_SYMBOL_GPL(sas_target_alloc);
  709. #define SAS_DEF_QD 256
  710. int sas_slave_configure(struct scsi_device *scsi_dev)
  711. {
  712. struct domain_device *dev = sdev_to_domain_dev(scsi_dev);
  713. BUG_ON(dev->rphy->identify.device_type != SAS_END_DEVICE);
  714. if (dev_is_sata(dev)) {
  715. ata_sas_slave_configure(scsi_dev, dev->sata_dev.ap);
  716. return 0;
  717. }
  718. sas_read_port_mode_page(scsi_dev);
  719. if (scsi_dev->tagged_supported) {
  720. scsi_change_queue_depth(scsi_dev, SAS_DEF_QD);
  721. } else {
  722. pr_notice("device %016llx, LUN 0x%llx doesn't support TCQ\n",
  723. SAS_ADDR(dev->sas_addr), scsi_dev->lun);
  724. scsi_change_queue_depth(scsi_dev, 1);
  725. }
  726. scsi_dev->allow_restart = 1;
  727. return 0;
  728. }
  729. EXPORT_SYMBOL_GPL(sas_slave_configure);
  730. int sas_change_queue_depth(struct scsi_device *sdev, int depth)
  731. {
  732. struct domain_device *dev = sdev_to_domain_dev(sdev);
  733. if (dev_is_sata(dev))
  734. return ata_change_queue_depth(dev->sata_dev.ap,
  735. sas_to_ata_dev(dev), sdev, depth);
  736. if (!sdev->tagged_supported)
  737. depth = 1;
  738. return scsi_change_queue_depth(sdev, depth);
  739. }
  740. EXPORT_SYMBOL_GPL(sas_change_queue_depth);
  741. int sas_bios_param(struct scsi_device *scsi_dev,
  742. struct block_device *bdev,
  743. sector_t capacity, int *hsc)
  744. {
  745. hsc[0] = 255;
  746. hsc[1] = 63;
  747. sector_div(capacity, 255*63);
  748. hsc[2] = capacity;
  749. return 0;
  750. }
  751. EXPORT_SYMBOL_GPL(sas_bios_param);
  752. void sas_task_internal_done(struct sas_task *task)
  753. {
  754. del_timer(&task->slow_task->timer);
  755. complete(&task->slow_task->completion);
  756. }
  757. void sas_task_internal_timedout(struct timer_list *t)
  758. {
  759. struct sas_task_slow *slow = from_timer(slow, t, timer);
  760. struct sas_task *task = slow->task;
  761. bool is_completed = true;
  762. unsigned long flags;
  763. spin_lock_irqsave(&task->task_state_lock, flags);
  764. if (!(task->task_state_flags & SAS_TASK_STATE_DONE)) {
  765. task->task_state_flags |= SAS_TASK_STATE_ABORTED;
  766. is_completed = false;
  767. }
  768. spin_unlock_irqrestore(&task->task_state_lock, flags);
  769. if (!is_completed)
  770. complete(&task->slow_task->completion);
  771. }
  772. #define TASK_TIMEOUT (20 * HZ)
  773. #define TASK_RETRY 3
  774. static int sas_execute_internal_abort(struct domain_device *device,
  775. enum sas_internal_abort type, u16 tag,
  776. unsigned int qid, void *data)
  777. {
  778. struct sas_ha_struct *ha = device->port->ha;
  779. struct sas_internal *i = to_sas_internal(ha->core.shost->transportt);
  780. struct sas_task *task = NULL;
  781. int res, retry;
  782. for (retry = 0; retry < TASK_RETRY; retry++) {
  783. task = sas_alloc_slow_task(GFP_KERNEL);
  784. if (!task)
  785. return -ENOMEM;
  786. task->dev = device;
  787. task->task_proto = SAS_PROTOCOL_INTERNAL_ABORT;
  788. task->task_done = sas_task_internal_done;
  789. task->slow_task->timer.function = sas_task_internal_timedout;
  790. task->slow_task->timer.expires = jiffies + TASK_TIMEOUT;
  791. add_timer(&task->slow_task->timer);
  792. task->abort_task.tag = tag;
  793. task->abort_task.type = type;
  794. task->abort_task.qid = qid;
  795. res = i->dft->lldd_execute_task(task, GFP_KERNEL);
  796. if (res) {
  797. del_timer_sync(&task->slow_task->timer);
  798. pr_err("Executing internal abort failed %016llx (%d)\n",
  799. SAS_ADDR(device->sas_addr), res);
  800. break;
  801. }
  802. wait_for_completion(&task->slow_task->completion);
  803. res = TMF_RESP_FUNC_FAILED;
  804. /* Even if the internal abort timed out, return direct. */
  805. if (task->task_state_flags & SAS_TASK_STATE_ABORTED) {
  806. bool quit = true;
  807. if (i->dft->lldd_abort_timeout)
  808. quit = i->dft->lldd_abort_timeout(task, data);
  809. else
  810. pr_err("Internal abort: timeout %016llx\n",
  811. SAS_ADDR(device->sas_addr));
  812. res = -EIO;
  813. if (quit)
  814. break;
  815. }
  816. if (task->task_status.resp == SAS_TASK_COMPLETE &&
  817. task->task_status.stat == SAS_SAM_STAT_GOOD) {
  818. res = TMF_RESP_FUNC_COMPLETE;
  819. break;
  820. }
  821. if (task->task_status.resp == SAS_TASK_COMPLETE &&
  822. task->task_status.stat == TMF_RESP_FUNC_SUCC) {
  823. res = TMF_RESP_FUNC_SUCC;
  824. break;
  825. }
  826. pr_err("Internal abort: task to dev %016llx response: 0x%x status 0x%x\n",
  827. SAS_ADDR(device->sas_addr), task->task_status.resp,
  828. task->task_status.stat);
  829. sas_free_task(task);
  830. task = NULL;
  831. }
  832. BUG_ON(retry == TASK_RETRY && task != NULL);
  833. sas_free_task(task);
  834. return res;
  835. }
  836. int sas_execute_internal_abort_single(struct domain_device *device, u16 tag,
  837. unsigned int qid, void *data)
  838. {
  839. return sas_execute_internal_abort(device, SAS_INTERNAL_ABORT_SINGLE,
  840. tag, qid, data);
  841. }
  842. EXPORT_SYMBOL_GPL(sas_execute_internal_abort_single);
  843. int sas_execute_internal_abort_dev(struct domain_device *device,
  844. unsigned int qid, void *data)
  845. {
  846. return sas_execute_internal_abort(device, SAS_INTERNAL_ABORT_DEV,
  847. SCSI_NO_TAG, qid, data);
  848. }
  849. EXPORT_SYMBOL_GPL(sas_execute_internal_abort_dev);
  850. int sas_execute_tmf(struct domain_device *device, void *parameter,
  851. int para_len, int force_phy_id,
  852. struct sas_tmf_task *tmf)
  853. {
  854. struct sas_task *task;
  855. struct sas_internal *i =
  856. to_sas_internal(device->port->ha->core.shost->transportt);
  857. int res, retry;
  858. for (retry = 0; retry < TASK_RETRY; retry++) {
  859. task = sas_alloc_slow_task(GFP_KERNEL);
  860. if (!task)
  861. return -ENOMEM;
  862. task->dev = device;
  863. task->task_proto = device->tproto;
  864. if (dev_is_sata(device)) {
  865. task->ata_task.device_control_reg_update = 1;
  866. if (force_phy_id >= 0) {
  867. task->ata_task.force_phy = true;
  868. task->ata_task.force_phy_id = force_phy_id;
  869. }
  870. memcpy(&task->ata_task.fis, parameter, para_len);
  871. } else {
  872. memcpy(&task->ssp_task, parameter, para_len);
  873. }
  874. task->task_done = sas_task_internal_done;
  875. task->tmf = tmf;
  876. task->slow_task->timer.function = sas_task_internal_timedout;
  877. task->slow_task->timer.expires = jiffies + TASK_TIMEOUT;
  878. add_timer(&task->slow_task->timer);
  879. res = i->dft->lldd_execute_task(task, GFP_KERNEL);
  880. if (res) {
  881. del_timer_sync(&task->slow_task->timer);
  882. pr_err("executing TMF task failed %016llx (%d)\n",
  883. SAS_ADDR(device->sas_addr), res);
  884. break;
  885. }
  886. wait_for_completion(&task->slow_task->completion);
  887. if (i->dft->lldd_tmf_exec_complete)
  888. i->dft->lldd_tmf_exec_complete(device);
  889. res = TMF_RESP_FUNC_FAILED;
  890. if ((task->task_state_flags & SAS_TASK_STATE_ABORTED)) {
  891. if (!(task->task_state_flags & SAS_TASK_STATE_DONE)) {
  892. pr_err("TMF task timeout for %016llx and not done\n",
  893. SAS_ADDR(device->sas_addr));
  894. if (i->dft->lldd_tmf_aborted)
  895. i->dft->lldd_tmf_aborted(task);
  896. break;
  897. }
  898. pr_warn("TMF task timeout for %016llx and done\n",
  899. SAS_ADDR(device->sas_addr));
  900. }
  901. if (task->task_status.resp == SAS_TASK_COMPLETE &&
  902. task->task_status.stat == TMF_RESP_FUNC_COMPLETE) {
  903. res = TMF_RESP_FUNC_COMPLETE;
  904. break;
  905. }
  906. if (task->task_status.resp == SAS_TASK_COMPLETE &&
  907. task->task_status.stat == TMF_RESP_FUNC_SUCC) {
  908. res = TMF_RESP_FUNC_SUCC;
  909. break;
  910. }
  911. if (task->task_status.resp == SAS_TASK_COMPLETE &&
  912. task->task_status.stat == SAS_DATA_UNDERRUN) {
  913. /* no error, but return the number of bytes of
  914. * underrun
  915. */
  916. pr_warn("TMF task to dev %016llx resp: 0x%x sts 0x%x underrun\n",
  917. SAS_ADDR(device->sas_addr),
  918. task->task_status.resp,
  919. task->task_status.stat);
  920. res = task->task_status.residual;
  921. break;
  922. }
  923. if (task->task_status.resp == SAS_TASK_COMPLETE &&
  924. task->task_status.stat == SAS_DATA_OVERRUN) {
  925. pr_warn("TMF task blocked task error %016llx\n",
  926. SAS_ADDR(device->sas_addr));
  927. res = -EMSGSIZE;
  928. break;
  929. }
  930. if (task->task_status.resp == SAS_TASK_COMPLETE &&
  931. task->task_status.stat == SAS_OPEN_REJECT) {
  932. pr_warn("TMF task open reject failed %016llx\n",
  933. SAS_ADDR(device->sas_addr));
  934. res = -EIO;
  935. } else {
  936. pr_warn("TMF task to dev %016llx resp: 0x%x status 0x%x\n",
  937. SAS_ADDR(device->sas_addr),
  938. task->task_status.resp,
  939. task->task_status.stat);
  940. }
  941. sas_free_task(task);
  942. task = NULL;
  943. }
  944. if (retry == TASK_RETRY)
  945. pr_warn("executing TMF for %016llx failed after %d attempts!\n",
  946. SAS_ADDR(device->sas_addr), TASK_RETRY);
  947. sas_free_task(task);
  948. return res;
  949. }
  950. static int sas_execute_ssp_tmf(struct domain_device *device, u8 *lun,
  951. struct sas_tmf_task *tmf)
  952. {
  953. struct sas_ssp_task ssp_task;
  954. if (!(device->tproto & SAS_PROTOCOL_SSP))
  955. return TMF_RESP_FUNC_ESUPP;
  956. memcpy(ssp_task.LUN, lun, 8);
  957. return sas_execute_tmf(device, &ssp_task, sizeof(ssp_task), -1, tmf);
  958. }
  959. int sas_abort_task_set(struct domain_device *dev, u8 *lun)
  960. {
  961. struct sas_tmf_task tmf_task = {
  962. .tmf = TMF_ABORT_TASK_SET,
  963. };
  964. return sas_execute_ssp_tmf(dev, lun, &tmf_task);
  965. }
  966. EXPORT_SYMBOL_GPL(sas_abort_task_set);
  967. int sas_clear_task_set(struct domain_device *dev, u8 *lun)
  968. {
  969. struct sas_tmf_task tmf_task = {
  970. .tmf = TMF_CLEAR_TASK_SET,
  971. };
  972. return sas_execute_ssp_tmf(dev, lun, &tmf_task);
  973. }
  974. EXPORT_SYMBOL_GPL(sas_clear_task_set);
  975. int sas_lu_reset(struct domain_device *dev, u8 *lun)
  976. {
  977. struct sas_tmf_task tmf_task = {
  978. .tmf = TMF_LU_RESET,
  979. };
  980. return sas_execute_ssp_tmf(dev, lun, &tmf_task);
  981. }
  982. EXPORT_SYMBOL_GPL(sas_lu_reset);
  983. int sas_query_task(struct sas_task *task, u16 tag)
  984. {
  985. struct sas_tmf_task tmf_task = {
  986. .tmf = TMF_QUERY_TASK,
  987. .tag_of_task_to_be_managed = tag,
  988. };
  989. struct scsi_cmnd *cmnd = task->uldd_task;
  990. struct domain_device *dev = task->dev;
  991. struct scsi_lun lun;
  992. int_to_scsilun(cmnd->device->lun, &lun);
  993. return sas_execute_ssp_tmf(dev, lun.scsi_lun, &tmf_task);
  994. }
  995. EXPORT_SYMBOL_GPL(sas_query_task);
  996. int sas_abort_task(struct sas_task *task, u16 tag)
  997. {
  998. struct sas_tmf_task tmf_task = {
  999. .tmf = TMF_ABORT_TASK,
  1000. .tag_of_task_to_be_managed = tag,
  1001. };
  1002. struct scsi_cmnd *cmnd = task->uldd_task;
  1003. struct domain_device *dev = task->dev;
  1004. struct scsi_lun lun;
  1005. int_to_scsilun(cmnd->device->lun, &lun);
  1006. return sas_execute_ssp_tmf(dev, lun.scsi_lun, &tmf_task);
  1007. }
  1008. EXPORT_SYMBOL_GPL(sas_abort_task);
  1009. /*
  1010. * Tell an upper layer that it needs to initiate an abort for a given task.
  1011. * This should only ever be called by an LLDD.
  1012. */
  1013. void sas_task_abort(struct sas_task *task)
  1014. {
  1015. struct scsi_cmnd *sc = task->uldd_task;
  1016. /* Escape for libsas internal commands */
  1017. if (!sc) {
  1018. struct sas_task_slow *slow = task->slow_task;
  1019. if (!slow)
  1020. return;
  1021. if (!del_timer(&slow->timer))
  1022. return;
  1023. slow->timer.function(&slow->timer);
  1024. return;
  1025. }
  1026. if (dev_is_sata(task->dev))
  1027. sas_ata_task_abort(task);
  1028. else
  1029. blk_abort_request(scsi_cmd_to_rq(sc));
  1030. }
  1031. EXPORT_SYMBOL_GPL(sas_task_abort);
  1032. int sas_slave_alloc(struct scsi_device *sdev)
  1033. {
  1034. if (dev_is_sata(sdev_to_domain_dev(sdev)) && sdev->lun)
  1035. return -ENXIO;
  1036. return 0;
  1037. }
  1038. EXPORT_SYMBOL_GPL(sas_slave_alloc);
  1039. void sas_target_destroy(struct scsi_target *starget)
  1040. {
  1041. struct domain_device *found_dev = starget->hostdata;
  1042. if (!found_dev)
  1043. return;
  1044. starget->hostdata = NULL;
  1045. sas_put_device(found_dev);
  1046. }
  1047. EXPORT_SYMBOL_GPL(sas_target_destroy);
  1048. #define SAS_STRING_ADDR_SIZE 16
  1049. int sas_request_addr(struct Scsi_Host *shost, u8 *addr)
  1050. {
  1051. int res;
  1052. const struct firmware *fw;
  1053. res = request_firmware(&fw, "sas_addr", &shost->shost_gendev);
  1054. if (res)
  1055. return res;
  1056. if (fw->size < SAS_STRING_ADDR_SIZE) {
  1057. res = -ENODEV;
  1058. goto out;
  1059. }
  1060. res = hex2bin(addr, fw->data, strnlen(fw->data, SAS_ADDR_SIZE * 2) / 2);
  1061. if (res)
  1062. goto out;
  1063. out:
  1064. release_firmware(fw);
  1065. return res;
  1066. }
  1067. EXPORT_SYMBOL_GPL(sas_request_addr);