rnbd-srv.c 22 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. /*
  3. * RDMA Network Block Driver
  4. *
  5. * Copyright (c) 2014 - 2018 ProfitBricks GmbH. All rights reserved.
  6. * Copyright (c) 2018 - 2019 1&1 IONOS Cloud GmbH. All rights reserved.
  7. * Copyright (c) 2019 - 2020 1&1 IONOS SE. All rights reserved.
  8. */
  9. #undef pr_fmt
  10. #define pr_fmt(fmt) KBUILD_MODNAME " L" __stringify(__LINE__) ": " fmt
  11. #include <linux/module.h>
  12. #include <linux/blkdev.h>
  13. #include "rnbd-srv.h"
  14. #include "rnbd-srv-trace.h"
  15. MODULE_DESCRIPTION("RDMA Network Block Device Server");
  16. MODULE_LICENSE("GPL");
  17. static u16 port_nr = RTRS_PORT;
  18. module_param_named(port_nr, port_nr, ushort, 0444);
  19. MODULE_PARM_DESC(port_nr,
  20. "The port number the server is listening on (default: "
  21. __stringify(RTRS_PORT)")");
  22. #define DEFAULT_DEV_SEARCH_PATH "/"
  23. static char dev_search_path[PATH_MAX] = DEFAULT_DEV_SEARCH_PATH;
  24. static int dev_search_path_set(const char *val, const struct kernel_param *kp)
  25. {
  26. const char *p = strrchr(val, '\n') ? : val + strlen(val);
  27. if (strlen(val) >= sizeof(dev_search_path))
  28. return -EINVAL;
  29. snprintf(dev_search_path, sizeof(dev_search_path), "%.*s",
  30. (int)(p - val), val);
  31. pr_info("dev_search_path changed to '%s'\n", dev_search_path);
  32. return 0;
  33. }
  34. static struct kparam_string dev_search_path_kparam_str = {
  35. .maxlen = sizeof(dev_search_path),
  36. .string = dev_search_path
  37. };
  38. static const struct kernel_param_ops dev_search_path_ops = {
  39. .set = dev_search_path_set,
  40. .get = param_get_string,
  41. };
  42. module_param_cb(dev_search_path, &dev_search_path_ops,
  43. &dev_search_path_kparam_str, 0444);
  44. MODULE_PARM_DESC(dev_search_path,
  45. "Sets the dev_search_path. When a device is mapped this path is prepended to the device path from the map device operation. If %SESSNAME% is specified in a path, then device will be searched in a session namespace. (default: "
  46. DEFAULT_DEV_SEARCH_PATH ")");
  47. static DEFINE_MUTEX(sess_lock);
  48. static DEFINE_SPINLOCK(dev_lock);
  49. static LIST_HEAD(sess_list);
  50. static LIST_HEAD(dev_list);
  51. struct rnbd_io_private {
  52. struct rtrs_srv_op *id;
  53. struct rnbd_srv_sess_dev *sess_dev;
  54. };
  55. static void rnbd_sess_dev_release(struct kref *kref)
  56. {
  57. struct rnbd_srv_sess_dev *sess_dev;
  58. sess_dev = container_of(kref, struct rnbd_srv_sess_dev, kref);
  59. complete(sess_dev->destroy_comp);
  60. }
  61. static inline void rnbd_put_sess_dev(struct rnbd_srv_sess_dev *sess_dev)
  62. {
  63. kref_put(&sess_dev->kref, rnbd_sess_dev_release);
  64. }
  65. static struct rnbd_srv_sess_dev *
  66. rnbd_get_sess_dev(int dev_id, struct rnbd_srv_session *srv_sess)
  67. {
  68. struct rnbd_srv_sess_dev *sess_dev;
  69. int ret = 0;
  70. rcu_read_lock();
  71. sess_dev = xa_load(&srv_sess->index_idr, dev_id);
  72. if (sess_dev)
  73. ret = kref_get_unless_zero(&sess_dev->kref);
  74. rcu_read_unlock();
  75. if (!sess_dev || !ret)
  76. return ERR_PTR(-ENXIO);
  77. return sess_dev;
  78. }
  79. static void rnbd_dev_bi_end_io(struct bio *bio)
  80. {
  81. struct rnbd_io_private *rnbd_priv = bio->bi_private;
  82. struct rnbd_srv_sess_dev *sess_dev = rnbd_priv->sess_dev;
  83. rnbd_put_sess_dev(sess_dev);
  84. rtrs_srv_resp_rdma(rnbd_priv->id, blk_status_to_errno(bio->bi_status));
  85. kfree(rnbd_priv);
  86. bio_put(bio);
  87. }
  88. static int process_rdma(struct rnbd_srv_session *srv_sess,
  89. struct rtrs_srv_op *id, void *data, u32 datalen,
  90. const void *usr, size_t usrlen)
  91. {
  92. const struct rnbd_msg_io *msg = usr;
  93. struct rnbd_io_private *priv;
  94. struct rnbd_srv_sess_dev *sess_dev;
  95. u32 dev_id;
  96. int err;
  97. struct bio *bio;
  98. short prio;
  99. trace_process_rdma(srv_sess, msg, id, datalen, usrlen);
  100. priv = kmalloc(sizeof(*priv), GFP_KERNEL);
  101. if (!priv)
  102. return -ENOMEM;
  103. dev_id = le32_to_cpu(msg->device_id);
  104. sess_dev = rnbd_get_sess_dev(dev_id, srv_sess);
  105. if (IS_ERR(sess_dev)) {
  106. pr_err_ratelimited("Got I/O request on session %s for unknown device id %d\n",
  107. srv_sess->sessname, dev_id);
  108. err = -ENOTCONN;
  109. goto err;
  110. }
  111. priv->sess_dev = sess_dev;
  112. priv->id = id;
  113. bio = bio_alloc(sess_dev->bdev, 1,
  114. rnbd_to_bio_flags(le32_to_cpu(msg->rw)), GFP_KERNEL);
  115. if (bio_add_page(bio, virt_to_page(data), datalen,
  116. offset_in_page(data)) != datalen) {
  117. rnbd_srv_err(sess_dev, "Failed to map data to bio\n");
  118. err = -EINVAL;
  119. goto bio_put;
  120. }
  121. bio->bi_end_io = rnbd_dev_bi_end_io;
  122. bio->bi_private = priv;
  123. bio->bi_iter.bi_sector = le64_to_cpu(msg->sector);
  124. bio->bi_iter.bi_size = le32_to_cpu(msg->bi_size);
  125. prio = srv_sess->ver < RNBD_PROTO_VER_MAJOR ||
  126. usrlen < sizeof(*msg) ? 0 : le16_to_cpu(msg->prio);
  127. bio_set_prio(bio, prio);
  128. submit_bio(bio);
  129. return 0;
  130. bio_put:
  131. bio_put(bio);
  132. rnbd_put_sess_dev(sess_dev);
  133. err:
  134. kfree(priv);
  135. return err;
  136. }
  137. static void destroy_device(struct kref *kref)
  138. {
  139. struct rnbd_srv_dev *dev = container_of(kref, struct rnbd_srv_dev, kref);
  140. WARN_ONCE(!list_empty(&dev->sess_dev_list),
  141. "Device %s is being destroyed but still in use!\n",
  142. dev->id);
  143. spin_lock(&dev_lock);
  144. list_del(&dev->list);
  145. spin_unlock(&dev_lock);
  146. mutex_destroy(&dev->lock);
  147. if (dev->dev_kobj.state_in_sysfs)
  148. /*
  149. * Destroy kobj only if it was really created.
  150. */
  151. rnbd_srv_destroy_dev_sysfs(dev);
  152. else
  153. kfree(dev);
  154. }
  155. static void rnbd_put_srv_dev(struct rnbd_srv_dev *dev)
  156. {
  157. kref_put(&dev->kref, destroy_device);
  158. }
  159. void rnbd_destroy_sess_dev(struct rnbd_srv_sess_dev *sess_dev, bool keep_id)
  160. {
  161. DECLARE_COMPLETION_ONSTACK(dc);
  162. if (keep_id)
  163. /* free the resources for the id but don't */
  164. /* allow to re-use the id itself because it */
  165. /* is still used by the client */
  166. xa_cmpxchg(&sess_dev->sess->index_idr, sess_dev->device_id,
  167. sess_dev, NULL, 0);
  168. else
  169. xa_erase(&sess_dev->sess->index_idr, sess_dev->device_id);
  170. synchronize_rcu();
  171. sess_dev->destroy_comp = &dc;
  172. rnbd_put_sess_dev(sess_dev);
  173. wait_for_completion(&dc); /* wait for inflights to drop to zero */
  174. blkdev_put(sess_dev->bdev, sess_dev->open_flags);
  175. mutex_lock(&sess_dev->dev->lock);
  176. list_del(&sess_dev->dev_list);
  177. if (sess_dev->open_flags & FMODE_WRITE)
  178. sess_dev->dev->open_write_cnt--;
  179. mutex_unlock(&sess_dev->dev->lock);
  180. rnbd_put_srv_dev(sess_dev->dev);
  181. rnbd_srv_info(sess_dev, "Device closed\n");
  182. kfree(sess_dev);
  183. }
  184. static void destroy_sess(struct rnbd_srv_session *srv_sess)
  185. {
  186. struct rnbd_srv_sess_dev *sess_dev;
  187. unsigned long index;
  188. if (xa_empty(&srv_sess->index_idr))
  189. goto out;
  190. trace_destroy_sess(srv_sess);
  191. mutex_lock(&srv_sess->lock);
  192. xa_for_each(&srv_sess->index_idr, index, sess_dev)
  193. rnbd_srv_destroy_dev_session_sysfs(sess_dev);
  194. mutex_unlock(&srv_sess->lock);
  195. out:
  196. xa_destroy(&srv_sess->index_idr);
  197. pr_info("RTRS Session %s disconnected\n", srv_sess->sessname);
  198. mutex_lock(&sess_lock);
  199. list_del(&srv_sess->list);
  200. mutex_unlock(&sess_lock);
  201. mutex_destroy(&srv_sess->lock);
  202. kfree(srv_sess);
  203. }
  204. static int create_sess(struct rtrs_srv_sess *rtrs)
  205. {
  206. struct rnbd_srv_session *srv_sess;
  207. char pathname[NAME_MAX];
  208. int err;
  209. err = rtrs_srv_get_path_name(rtrs, pathname, sizeof(pathname));
  210. if (err) {
  211. pr_err("rtrs_srv_get_path_name(%s): %d\n", pathname, err);
  212. return err;
  213. }
  214. srv_sess = kzalloc(sizeof(*srv_sess), GFP_KERNEL);
  215. if (!srv_sess)
  216. return -ENOMEM;
  217. srv_sess->queue_depth = rtrs_srv_get_queue_depth(rtrs);
  218. xa_init_flags(&srv_sess->index_idr, XA_FLAGS_ALLOC);
  219. mutex_init(&srv_sess->lock);
  220. mutex_lock(&sess_lock);
  221. list_add(&srv_sess->list, &sess_list);
  222. mutex_unlock(&sess_lock);
  223. srv_sess->rtrs = rtrs;
  224. strscpy(srv_sess->sessname, pathname, sizeof(srv_sess->sessname));
  225. rtrs_srv_set_sess_priv(rtrs, srv_sess);
  226. trace_create_sess(srv_sess);
  227. return 0;
  228. }
  229. static int rnbd_srv_link_ev(struct rtrs_srv_sess *rtrs,
  230. enum rtrs_srv_link_ev ev, void *priv)
  231. {
  232. struct rnbd_srv_session *srv_sess = priv;
  233. switch (ev) {
  234. case RTRS_SRV_LINK_EV_CONNECTED:
  235. return create_sess(rtrs);
  236. case RTRS_SRV_LINK_EV_DISCONNECTED:
  237. if (WARN_ON_ONCE(!srv_sess))
  238. return -EINVAL;
  239. destroy_sess(srv_sess);
  240. return 0;
  241. default:
  242. pr_warn("Received unknown RTRS session event %d from session %s\n",
  243. ev, srv_sess->sessname);
  244. return -EINVAL;
  245. }
  246. }
  247. void rnbd_srv_sess_dev_force_close(struct rnbd_srv_sess_dev *sess_dev,
  248. struct kobj_attribute *attr)
  249. {
  250. struct rnbd_srv_session *sess = sess_dev->sess;
  251. /* It is already started to close by client's close message. */
  252. if (!mutex_trylock(&sess->lock))
  253. return;
  254. sess_dev->keep_id = true;
  255. /* first remove sysfs itself to avoid deadlock */
  256. sysfs_remove_file_self(&sess_dev->kobj, &attr->attr);
  257. rnbd_srv_destroy_dev_session_sysfs(sess_dev);
  258. mutex_unlock(&sess->lock);
  259. }
  260. static void process_msg_close(struct rnbd_srv_session *srv_sess,
  261. void *data, size_t datalen, const void *usr,
  262. size_t usrlen)
  263. {
  264. const struct rnbd_msg_close *close_msg = usr;
  265. struct rnbd_srv_sess_dev *sess_dev;
  266. trace_process_msg_close(srv_sess, close_msg);
  267. sess_dev = rnbd_get_sess_dev(le32_to_cpu(close_msg->device_id),
  268. srv_sess);
  269. if (IS_ERR(sess_dev))
  270. return;
  271. rnbd_put_sess_dev(sess_dev);
  272. mutex_lock(&srv_sess->lock);
  273. rnbd_srv_destroy_dev_session_sysfs(sess_dev);
  274. mutex_unlock(&srv_sess->lock);
  275. }
  276. static int process_msg_open(struct rnbd_srv_session *srv_sess,
  277. const void *msg, size_t len,
  278. void *data, size_t datalen);
  279. static int process_msg_sess_info(struct rnbd_srv_session *srv_sess,
  280. const void *msg, size_t len,
  281. void *data, size_t datalen);
  282. static int rnbd_srv_rdma_ev(void *priv, struct rtrs_srv_op *id,
  283. void *data, size_t datalen,
  284. const void *usr, size_t usrlen)
  285. {
  286. struct rnbd_srv_session *srv_sess = priv;
  287. const struct rnbd_msg_hdr *hdr = usr;
  288. int ret = 0;
  289. u16 type;
  290. if (WARN_ON_ONCE(!srv_sess))
  291. return -ENODEV;
  292. type = le16_to_cpu(hdr->type);
  293. switch (type) {
  294. case RNBD_MSG_IO:
  295. return process_rdma(srv_sess, id, data, datalen, usr, usrlen);
  296. case RNBD_MSG_CLOSE:
  297. process_msg_close(srv_sess, data, datalen, usr, usrlen);
  298. break;
  299. case RNBD_MSG_OPEN:
  300. ret = process_msg_open(srv_sess, usr, usrlen, data, datalen);
  301. break;
  302. case RNBD_MSG_SESS_INFO:
  303. ret = process_msg_sess_info(srv_sess, usr, usrlen, data,
  304. datalen);
  305. break;
  306. default:
  307. pr_warn("Received unexpected message type %d from session %s\n",
  308. type, srv_sess->sessname);
  309. return -EINVAL;
  310. }
  311. /*
  312. * Since ret is passed to rtrs to handle the failure case, we
  313. * just return 0 at the end otherwise callers in rtrs would call
  314. * send_io_resp_imm again to print redundant err message.
  315. */
  316. rtrs_srv_resp_rdma(id, ret);
  317. return 0;
  318. }
  319. static struct rnbd_srv_sess_dev
  320. *rnbd_sess_dev_alloc(struct rnbd_srv_session *srv_sess)
  321. {
  322. struct rnbd_srv_sess_dev *sess_dev;
  323. int error;
  324. sess_dev = kzalloc(sizeof(*sess_dev), GFP_KERNEL);
  325. if (!sess_dev)
  326. return ERR_PTR(-ENOMEM);
  327. error = xa_alloc(&srv_sess->index_idr, &sess_dev->device_id, sess_dev,
  328. xa_limit_32b, GFP_NOWAIT);
  329. if (error < 0) {
  330. pr_warn("Allocating idr failed, err: %d\n", error);
  331. kfree(sess_dev);
  332. return ERR_PTR(error);
  333. }
  334. return sess_dev;
  335. }
  336. static struct rnbd_srv_dev *rnbd_srv_init_srv_dev(struct block_device *bdev)
  337. {
  338. struct rnbd_srv_dev *dev;
  339. dev = kzalloc(sizeof(*dev), GFP_KERNEL);
  340. if (!dev)
  341. return ERR_PTR(-ENOMEM);
  342. snprintf(dev->id, sizeof(dev->id), "%pg", bdev);
  343. kref_init(&dev->kref);
  344. INIT_LIST_HEAD(&dev->sess_dev_list);
  345. mutex_init(&dev->lock);
  346. return dev;
  347. }
  348. static struct rnbd_srv_dev *
  349. rnbd_srv_find_or_add_srv_dev(struct rnbd_srv_dev *new_dev)
  350. {
  351. struct rnbd_srv_dev *dev;
  352. spin_lock(&dev_lock);
  353. list_for_each_entry(dev, &dev_list, list) {
  354. if (!strncmp(dev->id, new_dev->id, sizeof(dev->id))) {
  355. if (!kref_get_unless_zero(&dev->kref))
  356. /*
  357. * We lost the race, device is almost dead.
  358. * Continue traversing to find a valid one.
  359. */
  360. continue;
  361. spin_unlock(&dev_lock);
  362. return dev;
  363. }
  364. }
  365. list_add(&new_dev->list, &dev_list);
  366. spin_unlock(&dev_lock);
  367. return new_dev;
  368. }
  369. static int rnbd_srv_check_update_open_perm(struct rnbd_srv_dev *srv_dev,
  370. struct rnbd_srv_session *srv_sess,
  371. enum rnbd_access_mode access_mode)
  372. {
  373. int ret = -EPERM;
  374. mutex_lock(&srv_dev->lock);
  375. switch (access_mode) {
  376. case RNBD_ACCESS_RO:
  377. ret = 0;
  378. break;
  379. case RNBD_ACCESS_RW:
  380. if (srv_dev->open_write_cnt == 0) {
  381. srv_dev->open_write_cnt++;
  382. ret = 0;
  383. } else {
  384. pr_err("Mapping device '%s' for session %s with RW permissions failed. Device already opened as 'RW' by %d client(s), access mode %s.\n",
  385. srv_dev->id, srv_sess->sessname,
  386. srv_dev->open_write_cnt,
  387. rnbd_access_mode_str(access_mode));
  388. }
  389. break;
  390. case RNBD_ACCESS_MIGRATION:
  391. if (srv_dev->open_write_cnt < 2) {
  392. srv_dev->open_write_cnt++;
  393. ret = 0;
  394. } else {
  395. pr_err("Mapping device '%s' for session %s with migration permissions failed. Device already opened as 'RW' by %d client(s), access mode %s.\n",
  396. srv_dev->id, srv_sess->sessname,
  397. srv_dev->open_write_cnt,
  398. rnbd_access_mode_str(access_mode));
  399. }
  400. break;
  401. default:
  402. pr_err("Received mapping request for device '%s' on session %s with invalid access mode: %d\n",
  403. srv_dev->id, srv_sess->sessname, access_mode);
  404. ret = -EINVAL;
  405. }
  406. mutex_unlock(&srv_dev->lock);
  407. return ret;
  408. }
  409. static struct rnbd_srv_dev *
  410. rnbd_srv_get_or_create_srv_dev(struct block_device *bdev,
  411. struct rnbd_srv_session *srv_sess,
  412. enum rnbd_access_mode access_mode)
  413. {
  414. int ret;
  415. struct rnbd_srv_dev *new_dev, *dev;
  416. new_dev = rnbd_srv_init_srv_dev(bdev);
  417. if (IS_ERR(new_dev))
  418. return new_dev;
  419. dev = rnbd_srv_find_or_add_srv_dev(new_dev);
  420. if (dev != new_dev)
  421. kfree(new_dev);
  422. ret = rnbd_srv_check_update_open_perm(dev, srv_sess, access_mode);
  423. if (ret) {
  424. rnbd_put_srv_dev(dev);
  425. return ERR_PTR(ret);
  426. }
  427. return dev;
  428. }
  429. static void rnbd_srv_fill_msg_open_rsp(struct rnbd_msg_open_rsp *rsp,
  430. struct rnbd_srv_sess_dev *sess_dev)
  431. {
  432. struct block_device *bdev = sess_dev->bdev;
  433. rsp->hdr.type = cpu_to_le16(RNBD_MSG_OPEN_RSP);
  434. rsp->device_id = cpu_to_le32(sess_dev->device_id);
  435. rsp->nsectors = cpu_to_le64(bdev_nr_sectors(bdev));
  436. rsp->logical_block_size = cpu_to_le16(bdev_logical_block_size(bdev));
  437. rsp->physical_block_size = cpu_to_le16(bdev_physical_block_size(bdev));
  438. rsp->max_segments = cpu_to_le16(bdev_max_segments(bdev));
  439. rsp->max_hw_sectors =
  440. cpu_to_le32(queue_max_hw_sectors(bdev_get_queue(bdev)));
  441. rsp->max_write_same_sectors = 0;
  442. rsp->max_discard_sectors = cpu_to_le32(bdev_max_discard_sectors(bdev));
  443. rsp->discard_granularity = cpu_to_le32(bdev_discard_granularity(bdev));
  444. rsp->discard_alignment = cpu_to_le32(bdev_discard_alignment(bdev));
  445. rsp->secure_discard = cpu_to_le16(bdev_max_secure_erase_sectors(bdev));
  446. rsp->cache_policy = 0;
  447. if (bdev_write_cache(bdev))
  448. rsp->cache_policy |= RNBD_WRITEBACK;
  449. if (bdev_fua(bdev))
  450. rsp->cache_policy |= RNBD_FUA;
  451. }
  452. static struct rnbd_srv_sess_dev *
  453. rnbd_srv_create_set_sess_dev(struct rnbd_srv_session *srv_sess,
  454. const struct rnbd_msg_open *open_msg,
  455. struct block_device *bdev, fmode_t open_flags,
  456. struct rnbd_srv_dev *srv_dev)
  457. {
  458. struct rnbd_srv_sess_dev *sdev = rnbd_sess_dev_alloc(srv_sess);
  459. if (IS_ERR(sdev))
  460. return sdev;
  461. kref_init(&sdev->kref);
  462. strscpy(sdev->pathname, open_msg->dev_name, sizeof(sdev->pathname));
  463. sdev->bdev = bdev;
  464. sdev->sess = srv_sess;
  465. sdev->dev = srv_dev;
  466. sdev->open_flags = open_flags;
  467. sdev->access_mode = open_msg->access_mode;
  468. return sdev;
  469. }
  470. static char *rnbd_srv_get_full_path(struct rnbd_srv_session *srv_sess,
  471. const char *dev_name)
  472. {
  473. char *full_path;
  474. char *a, *b;
  475. full_path = kmalloc(PATH_MAX, GFP_KERNEL);
  476. if (!full_path)
  477. return ERR_PTR(-ENOMEM);
  478. /*
  479. * Replace %SESSNAME% with a real session name in order to
  480. * create device namespace.
  481. */
  482. a = strnstr(dev_search_path, "%SESSNAME%", sizeof(dev_search_path));
  483. if (a) {
  484. int len = a - dev_search_path;
  485. len = snprintf(full_path, PATH_MAX, "%.*s/%s/%s", len,
  486. dev_search_path, srv_sess->sessname, dev_name);
  487. if (len >= PATH_MAX) {
  488. pr_err("Too long path: %s, %s, %s\n",
  489. dev_search_path, srv_sess->sessname, dev_name);
  490. kfree(full_path);
  491. return ERR_PTR(-EINVAL);
  492. }
  493. } else {
  494. snprintf(full_path, PATH_MAX, "%s/%s",
  495. dev_search_path, dev_name);
  496. }
  497. /* eliminitate duplicated slashes */
  498. a = strchr(full_path, '/');
  499. b = a;
  500. while (*b != '\0') {
  501. if (*b == '/' && *a == '/') {
  502. b++;
  503. } else {
  504. a++;
  505. *a = *b;
  506. b++;
  507. }
  508. }
  509. a++;
  510. *a = '\0';
  511. return full_path;
  512. }
  513. static int process_msg_sess_info(struct rnbd_srv_session *srv_sess,
  514. const void *msg, size_t len,
  515. void *data, size_t datalen)
  516. {
  517. const struct rnbd_msg_sess_info *sess_info_msg = msg;
  518. struct rnbd_msg_sess_info_rsp *rsp = data;
  519. srv_sess->ver = min_t(u8, sess_info_msg->ver, RNBD_PROTO_VER_MAJOR);
  520. trace_process_msg_sess_info(srv_sess, sess_info_msg);
  521. rsp->hdr.type = cpu_to_le16(RNBD_MSG_SESS_INFO_RSP);
  522. rsp->ver = srv_sess->ver;
  523. return 0;
  524. }
  525. /**
  526. * find_srv_sess_dev() - a dev is already opened by this name
  527. * @srv_sess: the session to search.
  528. * @dev_name: string containing the name of the device.
  529. *
  530. * Return struct rnbd_srv_sess_dev if srv_sess already opened the dev_name
  531. * NULL if the session didn't open the device yet.
  532. */
  533. static struct rnbd_srv_sess_dev *
  534. find_srv_sess_dev(struct rnbd_srv_session *srv_sess, const char *dev_name)
  535. {
  536. struct rnbd_srv_sess_dev *sess_dev;
  537. unsigned long index;
  538. if (xa_empty(&srv_sess->index_idr))
  539. return NULL;
  540. xa_for_each(&srv_sess->index_idr, index, sess_dev)
  541. if (!strcmp(sess_dev->pathname, dev_name))
  542. return sess_dev;
  543. return NULL;
  544. }
  545. static int process_msg_open(struct rnbd_srv_session *srv_sess,
  546. const void *msg, size_t len,
  547. void *data, size_t datalen)
  548. {
  549. int ret;
  550. struct rnbd_srv_dev *srv_dev;
  551. struct rnbd_srv_sess_dev *srv_sess_dev;
  552. const struct rnbd_msg_open *open_msg = msg;
  553. struct block_device *bdev;
  554. fmode_t open_flags;
  555. char *full_path;
  556. struct rnbd_msg_open_rsp *rsp = data;
  557. trace_process_msg_open(srv_sess, open_msg);
  558. open_flags = FMODE_READ;
  559. if (open_msg->access_mode != RNBD_ACCESS_RO)
  560. open_flags |= FMODE_WRITE;
  561. mutex_lock(&srv_sess->lock);
  562. srv_sess_dev = find_srv_sess_dev(srv_sess, open_msg->dev_name);
  563. if (srv_sess_dev)
  564. goto fill_response;
  565. if ((strlen(dev_search_path) + strlen(open_msg->dev_name))
  566. >= PATH_MAX) {
  567. pr_err("Opening device for session %s failed, device path too long. '%s/%s' is longer than PATH_MAX (%d)\n",
  568. srv_sess->sessname, dev_search_path, open_msg->dev_name,
  569. PATH_MAX);
  570. ret = -EINVAL;
  571. goto reject;
  572. }
  573. if (strstr(open_msg->dev_name, "..")) {
  574. pr_err("Opening device for session %s failed, device path %s contains relative path ..\n",
  575. srv_sess->sessname, open_msg->dev_name);
  576. ret = -EINVAL;
  577. goto reject;
  578. }
  579. full_path = rnbd_srv_get_full_path(srv_sess, open_msg->dev_name);
  580. if (IS_ERR(full_path)) {
  581. ret = PTR_ERR(full_path);
  582. pr_err("Opening device '%s' for client %s failed, failed to get device full path, err: %d\n",
  583. open_msg->dev_name, srv_sess->sessname, ret);
  584. goto reject;
  585. }
  586. bdev = blkdev_get_by_path(full_path, open_flags, THIS_MODULE);
  587. if (IS_ERR(bdev)) {
  588. ret = PTR_ERR(bdev);
  589. pr_err("Opening device '%s' on session %s failed, failed to open the block device, err: %d\n",
  590. full_path, srv_sess->sessname, ret);
  591. goto free_path;
  592. }
  593. srv_dev = rnbd_srv_get_or_create_srv_dev(bdev, srv_sess,
  594. open_msg->access_mode);
  595. if (IS_ERR(srv_dev)) {
  596. pr_err("Opening device '%s' on session %s failed, creating srv_dev failed, err: %ld\n",
  597. full_path, srv_sess->sessname, PTR_ERR(srv_dev));
  598. ret = PTR_ERR(srv_dev);
  599. goto blkdev_put;
  600. }
  601. srv_sess_dev = rnbd_srv_create_set_sess_dev(srv_sess, open_msg,
  602. bdev, open_flags,
  603. srv_dev);
  604. if (IS_ERR(srv_sess_dev)) {
  605. pr_err("Opening device '%s' on session %s failed, creating sess_dev failed, err: %ld\n",
  606. full_path, srv_sess->sessname, PTR_ERR(srv_sess_dev));
  607. ret = PTR_ERR(srv_sess_dev);
  608. goto srv_dev_put;
  609. }
  610. /* Create the srv_dev sysfs files if they haven't been created yet. The
  611. * reason to delay the creation is not to create the sysfs files before
  612. * we are sure the device can be opened.
  613. */
  614. mutex_lock(&srv_dev->lock);
  615. if (!srv_dev->dev_kobj.state_in_sysfs) {
  616. ret = rnbd_srv_create_dev_sysfs(srv_dev, bdev);
  617. if (ret) {
  618. mutex_unlock(&srv_dev->lock);
  619. rnbd_srv_err(srv_sess_dev,
  620. "Opening device failed, failed to create device sysfs files, err: %d\n",
  621. ret);
  622. goto free_srv_sess_dev;
  623. }
  624. }
  625. ret = rnbd_srv_create_dev_session_sysfs(srv_sess_dev);
  626. if (ret) {
  627. mutex_unlock(&srv_dev->lock);
  628. rnbd_srv_err(srv_sess_dev,
  629. "Opening device failed, failed to create dev client sysfs files, err: %d\n",
  630. ret);
  631. goto free_srv_sess_dev;
  632. }
  633. list_add(&srv_sess_dev->dev_list, &srv_dev->sess_dev_list);
  634. mutex_unlock(&srv_dev->lock);
  635. rnbd_srv_info(srv_sess_dev, "Opened device '%s'\n", srv_dev->id);
  636. kfree(full_path);
  637. fill_response:
  638. rnbd_srv_fill_msg_open_rsp(rsp, srv_sess_dev);
  639. mutex_unlock(&srv_sess->lock);
  640. return 0;
  641. free_srv_sess_dev:
  642. xa_erase(&srv_sess->index_idr, srv_sess_dev->device_id);
  643. synchronize_rcu();
  644. kfree(srv_sess_dev);
  645. srv_dev_put:
  646. if (open_msg->access_mode != RNBD_ACCESS_RO) {
  647. mutex_lock(&srv_dev->lock);
  648. srv_dev->open_write_cnt--;
  649. mutex_unlock(&srv_dev->lock);
  650. }
  651. rnbd_put_srv_dev(srv_dev);
  652. blkdev_put:
  653. blkdev_put(bdev, open_flags);
  654. free_path:
  655. kfree(full_path);
  656. reject:
  657. mutex_unlock(&srv_sess->lock);
  658. return ret;
  659. }
  660. static struct rtrs_srv_ctx *rtrs_ctx;
  661. static struct rtrs_srv_ops rtrs_ops;
  662. static int __init rnbd_srv_init_module(void)
  663. {
  664. int err;
  665. BUILD_BUG_ON(sizeof(struct rnbd_msg_hdr) != 4);
  666. BUILD_BUG_ON(sizeof(struct rnbd_msg_sess_info) != 36);
  667. BUILD_BUG_ON(sizeof(struct rnbd_msg_sess_info_rsp) != 36);
  668. BUILD_BUG_ON(sizeof(struct rnbd_msg_open) != 264);
  669. BUILD_BUG_ON(sizeof(struct rnbd_msg_close) != 8);
  670. BUILD_BUG_ON(sizeof(struct rnbd_msg_open_rsp) != 56);
  671. rtrs_ops = (struct rtrs_srv_ops) {
  672. .rdma_ev = rnbd_srv_rdma_ev,
  673. .link_ev = rnbd_srv_link_ev,
  674. };
  675. rtrs_ctx = rtrs_srv_open(&rtrs_ops, port_nr);
  676. if (IS_ERR(rtrs_ctx)) {
  677. err = PTR_ERR(rtrs_ctx);
  678. pr_err("rtrs_srv_open(), err: %d\n", err);
  679. return err;
  680. }
  681. err = rnbd_srv_create_sysfs_files();
  682. if (err) {
  683. pr_err("rnbd_srv_create_sysfs_files(), err: %d\n", err);
  684. rtrs_srv_close(rtrs_ctx);
  685. return err;
  686. }
  687. return 0;
  688. }
  689. static void __exit rnbd_srv_cleanup_module(void)
  690. {
  691. rtrs_srv_close(rtrs_ctx);
  692. WARN_ON(!list_empty(&sess_list));
  693. rnbd_srv_destroy_sysfs_files();
  694. }
  695. module_init(rnbd_srv_init_module);
  696. module_exit(rnbd_srv_cleanup_module);