Lines Matching refs:rqstp

30 static int svc_deferred_recv(struct svc_rqst *rqstp);
325 void svc_xprt_copy_addrs(struct svc_rqst *rqstp, struct svc_xprt *xprt) in svc_xprt_copy_addrs() argument
327 memcpy(&rqstp->rq_addr, &xprt->xpt_remote, xprt->xpt_remotelen); in svc_xprt_copy_addrs()
328 rqstp->rq_addrlen = xprt->xpt_remotelen; in svc_xprt_copy_addrs()
334 memcpy(&rqstp->rq_daddr, &xprt->xpt_local, xprt->xpt_locallen); in svc_xprt_copy_addrs()
335 rqstp->rq_daddrlen = xprt->xpt_locallen; in svc_xprt_copy_addrs()
346 char *svc_print_addr(struct svc_rqst *rqstp, char *buf, size_t len) in svc_print_addr() argument
348 return __svc_print_addr(svc_addr(rqstp), buf, len); in svc_print_addr()
360 static bool svc_xprt_reserve_slot(struct svc_rqst *rqstp, struct svc_xprt *xprt) in svc_xprt_reserve_slot() argument
362 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_reserve_slot()
366 set_bit(RQ_DATA, &rqstp->rq_flags); in svc_xprt_reserve_slot()
371 static void svc_xprt_release_slot(struct svc_rqst *rqstp) in svc_xprt_release_slot() argument
373 struct svc_xprt *xprt = rqstp->rq_xprt; in svc_xprt_release_slot()
374 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_release_slot()
411 struct svc_rqst *rqstp = NULL; in svc_xprt_do_enqueue() local
437 list_for_each_entry_rcu(rqstp, &pool->sp_all_threads, rq_all) { in svc_xprt_do_enqueue()
438 if (test_and_set_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_xprt_do_enqueue()
441 rqstp->rq_qtime = ktime_get(); in svc_xprt_do_enqueue()
442 wake_up_process(rqstp->rq_task); in svc_xprt_do_enqueue()
446 rqstp = NULL; in svc_xprt_do_enqueue()
450 trace_svc_xprt_do_enqueue(xprt, rqstp); in svc_xprt_do_enqueue()
499 void svc_reserve(struct svc_rqst *rqstp, int space) in svc_reserve() argument
501 struct svc_xprt *xprt = rqstp->rq_xprt; in svc_reserve()
503 space += rqstp->rq_res.head[0].iov_len; in svc_reserve()
505 if (xprt && space < rqstp->rq_reserved) { in svc_reserve()
506 atomic_sub((rqstp->rq_reserved - space), &xprt->xpt_reserved); in svc_reserve()
507 rqstp->rq_reserved = space; in svc_reserve()
514 static void svc_xprt_release(struct svc_rqst *rqstp) in svc_xprt_release() argument
516 struct svc_xprt *xprt = rqstp->rq_xprt; in svc_xprt_release()
518 xprt->xpt_ops->xpo_release_rqst(rqstp); in svc_xprt_release()
520 kfree(rqstp->rq_deferred); in svc_xprt_release()
521 rqstp->rq_deferred = NULL; in svc_xprt_release()
523 svc_free_res_pages(rqstp); in svc_xprt_release()
524 rqstp->rq_res.page_len = 0; in svc_xprt_release()
525 rqstp->rq_res.page_base = 0; in svc_xprt_release()
532 if ((rqstp->rq_res.len) > rqstp->rq_reserved) in svc_xprt_release()
534 rqstp->rq_reserved, in svc_xprt_release()
535 rqstp->rq_res.len); in svc_xprt_release()
537 rqstp->rq_res.head[0].iov_len = 0; in svc_xprt_release()
538 svc_reserve(rqstp, 0); in svc_xprt_release()
539 svc_xprt_release_slot(rqstp); in svc_xprt_release()
540 rqstp->rq_xprt = NULL; in svc_xprt_release()
553 struct svc_rqst *rqstp; in svc_wake_up() local
559 list_for_each_entry_rcu(rqstp, &pool->sp_all_threads, rq_all) { in svc_wake_up()
561 if (test_bit(RQ_BUSY, &rqstp->rq_flags)) in svc_wake_up()
564 wake_up_process(rqstp->rq_task); in svc_wake_up()
565 trace_svc_wake_up(rqstp->rq_task->pid); in svc_wake_up()
642 static int svc_alloc_arg(struct svc_rqst *rqstp) in svc_alloc_arg() argument
644 struct svc_serv *serv = rqstp->rq_server; in svc_alloc_arg()
658 while (rqstp->rq_pages[i] == NULL) { in svc_alloc_arg()
668 rqstp->rq_pages[i] = p; in svc_alloc_arg()
670 rqstp->rq_page_end = &rqstp->rq_pages[i]; in svc_alloc_arg()
671 rqstp->rq_pages[i++] = NULL; /* this might be seen in nfs_read_actor */ in svc_alloc_arg()
674 arg = &rqstp->rq_arg; in svc_alloc_arg()
675 arg->head[0].iov_base = page_address(rqstp->rq_pages[0]); in svc_alloc_arg()
677 arg->pages = rqstp->rq_pages + 1; in svc_alloc_arg()
687 rqst_should_sleep(struct svc_rqst *rqstp) in rqst_should_sleep() argument
689 struct svc_pool *pool = rqstp->rq_pool; in rqst_should_sleep()
710 static struct svc_xprt *svc_get_next_xprt(struct svc_rqst *rqstp, long timeout) in svc_get_next_xprt() argument
712 struct svc_pool *pool = rqstp->rq_pool; in svc_get_next_xprt()
716 WARN_ON_ONCE(rqstp->rq_xprt); in svc_get_next_xprt()
718 rqstp->rq_xprt = svc_xprt_dequeue(pool); in svc_get_next_xprt()
719 if (rqstp->rq_xprt) in svc_get_next_xprt()
729 clear_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt()
732 if (likely(rqst_should_sleep(rqstp))) in svc_get_next_xprt()
739 set_bit(RQ_BUSY, &rqstp->rq_flags); in svc_get_next_xprt()
741 rqstp->rq_xprt = svc_xprt_dequeue(pool); in svc_get_next_xprt()
742 if (rqstp->rq_xprt) in svc_get_next_xprt()
756 rqstp->rq_chandle.thread_wait = 5*HZ; in svc_get_next_xprt()
758 rqstp->rq_chandle.thread_wait = 1*HZ; in svc_get_next_xprt()
759 trace_svc_xprt_dequeue(rqstp); in svc_get_next_xprt()
760 return rqstp->rq_xprt; in svc_get_next_xprt()
779 static int svc_handle_xprt(struct svc_rqst *rqstp, struct svc_xprt *xprt) in svc_handle_xprt() argument
781 struct svc_serv *serv = rqstp->rq_server; in svc_handle_xprt()
806 } else if (svc_xprt_reserve_slot(rqstp, xprt)) { in svc_handle_xprt()
809 rqstp, rqstp->rq_pool->sp_id, xprt, in svc_handle_xprt()
811 rqstp->rq_deferred = svc_deferred_dequeue(xprt); in svc_handle_xprt()
812 if (rqstp->rq_deferred) in svc_handle_xprt()
813 len = svc_deferred_recv(rqstp); in svc_handle_xprt()
815 len = xprt->xpt_ops->xpo_recvfrom(rqstp); in svc_handle_xprt()
817 trace_svc_xdr_recvfrom(rqstp, &rqstp->rq_arg); in svc_handle_xprt()
818 rqstp->rq_stime = ktime_get(); in svc_handle_xprt()
819 rqstp->rq_reserved = serv->sv_max_mesg; in svc_handle_xprt()
820 atomic_add(rqstp->rq_reserved, &xprt->xpt_reserved); in svc_handle_xprt()
834 int svc_recv(struct svc_rqst *rqstp, long timeout) in svc_recv() argument
837 struct svc_serv *serv = rqstp->rq_server; in svc_recv()
840 err = svc_alloc_arg(rqstp); in svc_recv()
850 xprt = svc_get_next_xprt(rqstp, timeout); in svc_recv()
856 len = svc_handle_xprt(rqstp, xprt); in svc_recv()
865 xprt->xpt_ops->xpo_secure_port(rqstp); in svc_recv()
866 rqstp->rq_chandle.defer = svc_defer; in svc_recv()
867 rqstp->rq_xid = svc_getu32(&rqstp->rq_arg.head[0]); in svc_recv()
871 trace_svc_recv(rqstp, len); in svc_recv()
874 rqstp->rq_res.len = 0; in svc_recv()
875 svc_xprt_release(rqstp); in svc_recv()
884 void svc_drop(struct svc_rqst *rqstp) in svc_drop() argument
886 trace_svc_drop(rqstp); in svc_drop()
887 svc_xprt_release(rqstp); in svc_drop()
894 int svc_send(struct svc_rqst *rqstp) in svc_send() argument
900 xprt = rqstp->rq_xprt; in svc_send()
905 xb = &rqstp->rq_res; in svc_send()
909 trace_svc_xdr_sendto(rqstp, xb); in svc_send()
910 trace_svc_stats_latency(rqstp); in svc_send()
912 len = xprt->xpt_ops->xpo_sendto(rqstp); in svc_send()
914 trace_svc_send(rqstp, len); in svc_send()
915 svc_xprt_release(rqstp); in svc_send()
1172 struct svc_rqst *rqstp = container_of(req, struct svc_rqst, rq_chandle); in svc_defer() local
1175 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags)) in svc_defer()
1177 if (rqstp->rq_deferred) { in svc_defer()
1178 dr = rqstp->rq_deferred; in svc_defer()
1179 rqstp->rq_deferred = NULL; in svc_defer()
1184 size = sizeof(struct svc_deferred_req) + rqstp->rq_arg.len; in svc_defer()
1189 dr->handle.owner = rqstp->rq_server; in svc_defer()
1190 dr->prot = rqstp->rq_prot; in svc_defer()
1191 memcpy(&dr->addr, &rqstp->rq_addr, rqstp->rq_addrlen); in svc_defer()
1192 dr->addrlen = rqstp->rq_addrlen; in svc_defer()
1193 dr->daddr = rqstp->rq_daddr; in svc_defer()
1194 dr->argslen = rqstp->rq_arg.len >> 2; in svc_defer()
1195 dr->xprt_hlen = rqstp->rq_xprt_hlen; in svc_defer()
1198 skip = rqstp->rq_arg.len - rqstp->rq_arg.head[0].iov_len; in svc_defer()
1199 memcpy(dr->args, rqstp->rq_arg.head[0].iov_base - skip, in svc_defer()
1202 trace_svc_defer(rqstp); in svc_defer()
1203 svc_xprt_get(rqstp->rq_xprt); in svc_defer()
1204 dr->xprt = rqstp->rq_xprt; in svc_defer()
1205 set_bit(RQ_DROPME, &rqstp->rq_flags); in svc_defer()
1214 static noinline int svc_deferred_recv(struct svc_rqst *rqstp) in svc_deferred_recv() argument
1216 struct svc_deferred_req *dr = rqstp->rq_deferred; in svc_deferred_recv()
1221 rqstp->rq_arg.head[0].iov_base = dr->args + (dr->xprt_hlen>>2); in svc_deferred_recv()
1223 rqstp->rq_arg.head[0].iov_len = (dr->argslen<<2) - dr->xprt_hlen; in svc_deferred_recv()
1224 rqstp->rq_arg.page_len = 0; in svc_deferred_recv()
1226 rqstp->rq_arg.len = dr->argslen<<2; in svc_deferred_recv()
1227 rqstp->rq_prot = dr->prot; in svc_deferred_recv()
1228 memcpy(&rqstp->rq_addr, &dr->addr, dr->addrlen); in svc_deferred_recv()
1229 rqstp->rq_addrlen = dr->addrlen; in svc_deferred_recv()
1231 rqstp->rq_xprt_hlen = dr->xprt_hlen; in svc_deferred_recv()
1232 rqstp->rq_daddr = dr->daddr; in svc_deferred_recv()
1233 rqstp->rq_respages = rqstp->rq_pages; in svc_deferred_recv()