svcrdma: Add gfp flags to svc_rdma_post_recv()
svc_rdma_post_recv() allocates pages for receive buffers on-demand. It uses GFP_KERNEL so the allocator tries hard, and may sleep. But I'm about to add a call to svc_rdma_post_recv() from a function that may not sleep. Since all svc_rdma_post_recv() call sites can tolerate its failure, allow it to fail if the page allocator returns nothing. Longer term, receive buffers, being a finite resource per-connection, should be pre-allocated and re-used. Signed-off-by: Chuck Lever <chuck.lever@oracle.com> Acked-by: Bruce Fields <bfields@fieldses.org> Signed-off-by: Doug Ledford <dledford@redhat.com>
This commit is contained in:

committed by
Doug Ledford

parent
71810ef327
commit
39b09a1a12
@@ -221,7 +221,7 @@ extern struct rpcrdma_read_chunk *
|
|||||||
extern int svc_rdma_send(struct svcxprt_rdma *, struct ib_send_wr *);
|
extern int svc_rdma_send(struct svcxprt_rdma *, struct ib_send_wr *);
|
||||||
extern void svc_rdma_send_error(struct svcxprt_rdma *, struct rpcrdma_msg *,
|
extern void svc_rdma_send_error(struct svcxprt_rdma *, struct rpcrdma_msg *,
|
||||||
enum rpcrdma_errcode);
|
enum rpcrdma_errcode);
|
||||||
extern int svc_rdma_post_recv(struct svcxprt_rdma *);
|
extern int svc_rdma_post_recv(struct svcxprt_rdma *, gfp_t);
|
||||||
extern int svc_rdma_create_listen(struct svc_serv *, int, struct sockaddr *);
|
extern int svc_rdma_create_listen(struct svc_serv *, int, struct sockaddr *);
|
||||||
extern struct svc_rdma_op_ctxt *svc_rdma_get_context(struct svcxprt_rdma *);
|
extern struct svc_rdma_op_ctxt *svc_rdma_get_context(struct svcxprt_rdma *);
|
||||||
extern void svc_rdma_put_context(struct svc_rdma_op_ctxt *, int);
|
extern void svc_rdma_put_context(struct svc_rdma_op_ctxt *, int);
|
||||||
|
@@ -465,7 +465,7 @@ static int send_reply(struct svcxprt_rdma *rdma,
|
|||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
/* Post a recv buffer to handle another request. */
|
/* Post a recv buffer to handle another request. */
|
||||||
ret = svc_rdma_post_recv(rdma);
|
ret = svc_rdma_post_recv(rdma, GFP_KERNEL);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
printk(KERN_INFO
|
printk(KERN_INFO
|
||||||
"svcrdma: could not post a receive buffer, err=%d."
|
"svcrdma: could not post a receive buffer, err=%d."
|
||||||
|
@@ -668,7 +668,7 @@ static struct svcxprt_rdma *rdma_create_xprt(struct svc_serv *serv,
|
|||||||
return cma_xprt;
|
return cma_xprt;
|
||||||
}
|
}
|
||||||
|
|
||||||
int svc_rdma_post_recv(struct svcxprt_rdma *xprt)
|
int svc_rdma_post_recv(struct svcxprt_rdma *xprt, gfp_t flags)
|
||||||
{
|
{
|
||||||
struct ib_recv_wr recv_wr, *bad_recv_wr;
|
struct ib_recv_wr recv_wr, *bad_recv_wr;
|
||||||
struct svc_rdma_op_ctxt *ctxt;
|
struct svc_rdma_op_ctxt *ctxt;
|
||||||
@@ -686,7 +686,9 @@ int svc_rdma_post_recv(struct svcxprt_rdma *xprt)
|
|||||||
pr_err("svcrdma: Too many sges (%d)\n", sge_no);
|
pr_err("svcrdma: Too many sges (%d)\n", sge_no);
|
||||||
goto err_put_ctxt;
|
goto err_put_ctxt;
|
||||||
}
|
}
|
||||||
page = alloc_page(GFP_KERNEL | __GFP_NOFAIL);
|
page = alloc_page(flags);
|
||||||
|
if (!page)
|
||||||
|
goto err_put_ctxt;
|
||||||
ctxt->pages[sge_no] = page;
|
ctxt->pages[sge_no] = page;
|
||||||
pa = ib_dma_map_page(xprt->sc_cm_id->device,
|
pa = ib_dma_map_page(xprt->sc_cm_id->device,
|
||||||
page, 0, PAGE_SIZE,
|
page, 0, PAGE_SIZE,
|
||||||
@@ -1182,7 +1184,7 @@ static struct svc_xprt *svc_rdma_accept(struct svc_xprt *xprt)
|
|||||||
|
|
||||||
/* Post receive buffers */
|
/* Post receive buffers */
|
||||||
for (i = 0; i < newxprt->sc_max_requests; i++) {
|
for (i = 0; i < newxprt->sc_max_requests; i++) {
|
||||||
ret = svc_rdma_post_recv(newxprt);
|
ret = svc_rdma_post_recv(newxprt, GFP_KERNEL);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
dprintk("svcrdma: failure posting receive buffers\n");
|
dprintk("svcrdma: failure posting receive buffers\n");
|
||||||
goto errout;
|
goto errout;
|
||||||
|
Reference in New Issue
Block a user