Home
last modified time | relevance | path

Searched refs:num_sge (Results 1 – 60 of 60) sorted by relevance

/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_srq.c61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in ipath_post_srq_receive()
81 wqe->num_sge = wr->num_sge; in ipath_post_srq_receive()
82 for (i = 0; i < wr->num_sge; i++) in ipath_post_srq_receive()
291 p->num_sge = wqe->num_sge; in ipath_modify_srq()
292 for (i = 0; i < wqe->num_sge; i++) in ipath_modify_srq()
Dipath_ruc.c130 for (i = j = 0; i < wqe->num_sge; i++) { in ipath_init_sge()
140 ss->num_sge = j; in ipath_init_sge()
330 sqp->s_sge.num_sge = wqe->wr.num_sge; in ipath_ruc_loopback()
372 qp->r_sge.num_sge = wqe->wr.num_sge; in ipath_ruc_loopback()
413 if (--sqp->s_sge.num_sge) in ipath_ruc_loopback()
Dipath_verbs.c188 if (--ss->num_sge) in ipath_copy_sge()
227 if (--ss->num_sge) in ipath_skip_sge()
253 u8 num_sge = ss->num_sge; in ipath_count_sge() local
274 if (--num_sge) in ipath_count_sge()
313 if (--ss->num_sge) in ipath_copy_from_sge()
360 if (wr->num_sge > qp->s_max_sge) in ipath_post_one_send()
382 (wr->num_sge == 0 || in ipath_post_one_send()
400 if (wr->num_sge) { in ipath_post_one_send()
403 for (i = 0, j = 0; i < wr->num_sge; i++) { in ipath_post_one_send()
416 wqe->wr.num_sge = j; in ipath_post_one_send()
[all …]
Dipath_keys.c227 ss->num_sge = 1; in ipath_rkey_ok()
264 ss->num_sge = 1; in ipath_rkey_ok()
Dipath_ud.c198 if (--swqe->wr.num_sge) in ipath_ud_loopback()
325 qp->s_sge.num_sge = wqe->wr.num_sge; in ipath_make_ud_req()
Dipath_uc.c100 qp->s_sge.num_sge = wqe->wr.num_sge; in ipath_make_uc_req()
Dipath_rc.c50 ss->num_sge = wqe->wr.num_sge; in restart_sge()
463 qp->s_sge.num_sge = wqe->wr.num_sge; in ipath_make_rc_req()
1453 e->rdma_sge.num_sge = 0; in ipath_rc_rcv_error()
1840 e->rdma_sge.num_sge = 0; in ipath_rc_rcv()
Dipath_verbs.h295 u8 num_sge; member
331 u8 num_sge; member
Dipath_sdma.c763 if (--ss->num_sge) in ipath_sdma_verbs_send()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_srq.c61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in qib_post_srq_receive()
81 wqe->num_sge = wr->num_sge; in qib_post_srq_receive()
82 for (i = 0; i < wr->num_sge; i++) in qib_post_srq_receive()
287 p->num_sge = wqe->num_sge; in qib_modify_srq()
288 for (i = 0; i < wqe->num_sge; i++) in qib_modify_srq()
Dqib_ruc.c94 for (i = j = 0; i < wqe->num_sge; i++) { in qib_init_sge()
104 ss->num_sge = j; in qib_init_sge()
115 ss->num_sge = 0; in qib_init_sge()
429 sqp->s_sge.num_sge = wqe->wr.num_sge; in qib_ruc_loopback()
466 qp->r_sge.num_sge = 1; in qib_ruc_loopback()
480 sqp->s_sge.num_sge = 1; in qib_ruc_loopback()
483 qp->r_sge.num_sge = wqe->wr.num_sge; in qib_ruc_loopback()
505 qp->r_sge.num_sge = 0; in qib_ruc_loopback()
529 if (--sqp->s_sge.num_sge) in qib_ruc_loopback()
779 for (i = 0; i < wqe->wr.num_sge; i++) { in qib_send_complete()
Dqib_uc.c101 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_uc_req()
282 qp->r_sge.num_sge = 0; in qib_uc_rcv()
445 qp->r_sge.num_sge = 1; in qib_uc_rcv()
447 qp->r_sge.num_sge = 0; in qib_uc_rcv()
527 qp->r_sge.num_sge = 0; in qib_uc_rcv()
Dqib_ud.c174 ssge.num_sge = swqe->wr.num_sge; in qib_ud_loopback()
189 if (--ssge.num_sge) in qib_ud_loopback()
322 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_ud_req()
Dqib_verbs.c188 if (--ss->num_sge) in qib_copy_sge()
229 if (--ss->num_sge) in qib_skip_sge()
255 u8 num_sge = ss->num_sge; in qib_count_sge() local
276 if (--num_sge) in qib_count_sge()
314 if (--ss->num_sge) in qib_copy_from_sge()
357 if (wr->num_sge > qp->s_max_sge) in qib_post_one_send()
382 (wr->num_sge == 0 || in qib_post_one_send()
403 if (wr->num_sge) { in qib_post_one_send()
406 for (i = 0; i < wr->num_sge; i++) { in qib_post_one_send()
419 wqe->wr.num_sge = j; in qib_post_one_send()
[all …]
Dqib_rc.c51 ss->num_sge = wqe->wr.num_sge; in restart_sge()
138 qp->s_ack_rdma_sge.num_sge = 1; in qib_make_rc_ack()
489 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_rc_req()
1012 for (i = 0; i < wqe->wr.num_sge; i++) { in qib_rc_send_complete()
1068 for (i = 0; i < wqe->wr.num_sge; i++) { in do_rc_completion()
1580 WARN_ON(qp->s_rdma_read_sge.num_sge); in qib_rc_rcv_resp()
2074 qp->r_sge.num_sge = 1; in qib_rc_rcv()
2076 qp->r_sge.num_sge = 0; in qib_rc_rcv()
2210 qp->r_sge.num_sge = 0; in qib_rc_rcv()
Dqib_verbs.h356 u8 num_sge; member
394 u8 num_sge; member
1068 while (ss->num_sge) { in qib_put_ss()
1070 if (--ss->num_sge) in qib_put_ss()
Dqib_qp.c415 qp->r_sge.num_sge = 0; in qib_reset_qp()
432 for (i = 0; i < wqe->wr.num_sge; i++) { in clear_mr_refs()
Dqib_sdma.c621 if (--ss->num_sge) in qib_sdma_verbs_send()
/linux-4.1.27/net/rds/
Diw_send.c143 send->s_wr.num_sge = 1; in rds_iw_send_init_ring()
269 send->s_wr.num_sge = 1; in rds_iw_send_cq_comp_handler()
459 send->s_wr.num_sge = 2; in rds_iw_xmit_populate_wr()
474 send->s_wr.num_sge = 1; in rds_iw_xmit_populate_wr()
687 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); in rds_iw_xmit()
805 int num_sge; in rds_iw_xmit_rdma() local
858 num_sge = op->op_count; in rds_iw_xmit_rdma()
886 if (num_sge > rds_iwdev->max_sge) { in rds_iw_xmit_rdma()
887 send->s_wr.num_sge = rds_iwdev->max_sge; in rds_iw_xmit_rdma()
888 num_sge -= rds_iwdev->max_sge; in rds_iw_xmit_rdma()
[all …]
Dib_send.c669 send->s_wr.num_sge = 1; in rds_ib_xmit()
684 send->s_wr.num_sge = 2; in rds_ib_xmit()
709 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); in rds_ib_xmit()
825 send->s_wr.num_sge = 1; in rds_ib_xmit_atomic()
892 int num_sge; in rds_ib_xmit_rdma() local
929 num_sge = op->op_count; in rds_ib_xmit_rdma()
942 if (num_sge > max_sge) { in rds_ib_xmit_rdma()
943 send->s_wr.num_sge = max_sge; in rds_ib_xmit_rdma()
944 num_sge -= max_sge; in rds_ib_xmit_rdma()
946 send->s_wr.num_sge = num_sge; in rds_ib_xmit_rdma()
[all …]
Diw_recv.c96 recv->r_wr.num_sge = RDS_IW_RECV_SGE; in rds_iw_recv_init_ring()
357 wr->num_sge = 1; in rds_iw_recv_init_ack()
Dib_recv.c60 recv->r_wr.num_sge = RDS_IB_RECV_SGE; in rds_ib_recv_init_ring()
526 wr->num_sge = 1; in rds_ib_recv_init_ack()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_reqs.c76 if (unlikely((recv_wr->num_sge < 0) || in ehca_write_rwqe()
77 (recv_wr->num_sge > ipz_rqueue->act_nr_of_sg))) { in ehca_write_rwqe()
80 recv_wr->num_sge, ipz_rqueue->act_nr_of_sg); in ehca_write_rwqe()
88 wqe_p->nr_of_data_seg = recv_wr->num_sge; in ehca_write_rwqe()
90 for (cnt_ds = 0; cnt_ds < recv_wr->num_sge; cnt_ds++) { in ehca_write_rwqe()
122 send_wr->num_sge, send_wr->send_flags, in trace_send_wr_ud()
137 for (j = 0; j < send_wr->num_sge; j++) { in trace_send_wr_ud()
166 if (unlikely((send_wr->num_sge < 0) || in ehca_write_swqe()
167 (send_wr->num_sge > qp->ipz_squeue.act_nr_of_sg))) { in ehca_write_swqe()
170 send_wr->num_sge, qp->ipz_squeue.act_nr_of_sg); in ehca_write_swqe()
[all …]
/linux-4.1.27/drivers/infiniband/hw/amso1100/
Dc2_qp.c836 sizeof(struct c2_data_addr) * ib_wr->num_sge; in c2_post_send()
837 if (ib_wr->num_sge > qp->send_sgl_depth) { in c2_post_send()
846 ib_wr->num_sge, in c2_post_send()
854 (sizeof(struct c2_data_addr) * ib_wr->num_sge); in c2_post_send()
855 if (ib_wr->num_sge > qp->rdma_write_sgl_depth) { in c2_post_send()
869 ib_wr->num_sge, in c2_post_send()
879 if (ib_wr->num_sge > 1) { in c2_post_send()
965 if (ib_wr->num_sge > qp->recv_sgl_depth) { in c2_post_receive()
978 BUG_ON(ib_wr->num_sge >= 256); in c2_post_receive()
981 ib_wr->num_sge, &tot_len, &actual_sge_count); in c2_post_receive()
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
Diwch_qp.c66 if (wr->num_sge > T3_MAX_SGE) in build_rdma_send()
72 for (i = 0; i < wr->num_sge; i++) { in build_rdma_send()
81 wqe->send.num_sgle = cpu_to_be32(wr->num_sge); in build_rdma_send()
82 *flit_cnt = 4 + ((wr->num_sge) << 1); in build_rdma_send()
92 if (wr->num_sge > T3_MAX_SGE) in build_rdma_write()
109 for (i = 0; i < wr->num_sge; i++) { in build_rdma_write()
121 wqe->write.num_sgle = cpu_to_be32(wr->num_sge); in build_rdma_write()
122 *flit_cnt = 5 + ((wr->num_sge) << 1); in build_rdma_write()
131 if (wr->num_sge > 1) in build_rdma_read()
254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, in build_rdma_recv()
[all …]
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dqp.c373 for (i = 0; i < wr->num_sge; i++) { in build_immd()
405 int num_sge, u32 *plenp) in build_isgl() argument
412 for (i = 0; i < num_sge; i++) { in build_isgl()
427 isglp->nsge = cpu_to_be16(num_sge); in build_isgl()
441 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_send()
470 if (wr->num_sge) { in build_rdma_send()
482 wr->sg_list, wr->num_sge, &plen); in build_rdma_send()
486 wr->num_sge * sizeof(struct fw_ri_sge); in build_rdma_send()
508 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_write()
513 if (wr->num_sge) { in build_rdma_write()
[all …]
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_srq.c514 if (unlikely(wr->num_sge > srq->max_gs)) { in mthca_tavor_post_srq_recv()
521 for (i = 0; i < wr->num_sge; ++i) { in mthca_tavor_post_srq_recv()
607 if (unlikely(wr->num_sge > srq->max_gs)) { in mthca_arbel_post_srq_recv()
613 for (i = 0; i < wr->num_sge; ++i) { in mthca_arbel_post_srq_recv()
Dmthca_qp.c1729 if (wr->num_sge > qp->sq.max_gs) { in mthca_tavor_post_send()
1736 for (i = 0; i < wr->num_sge; ++i) { in mthca_tavor_post_send()
1855 if (unlikely(wr->num_sge > qp->rq.max_gs)) { in mthca_tavor_post_receive()
1861 for (i = 0; i < wr->num_sge; ++i) { in mthca_tavor_post_receive()
2070 if (wr->num_sge > qp->sq.max_gs) { in mthca_arbel_post_send()
2077 for (i = 0; i < wr->num_sge; ++i) { in mthca_arbel_post_send()
2191 if (unlikely(wr->num_sge > qp->rq.max_gs)) { in mthca_arbel_post_receive()
2197 for (i = 0; i < wr->num_sge; ++i) { in mthca_arbel_post_receive()
/linux-4.1.27/drivers/infiniband/ulp/iser/
Diser_initiator.c156 iser_task->desc.num_sge = 2; in iser_prepare_write_cmd()
174 tx_desc->num_sge = 1; in iser_create_send_desc()
487 tx_desc->num_sge = 2; in iser_send_data_out()
549 mdesc->num_sge = 2; in iser_send_control()
Diser_verbs.c1016 rx_wr.num_sge = 1; in iser_post_recvl()
1040 rx_wr->num_sge = 1; in iser_post_recvm()
1077 send_wr.num_sge = tx_desc->num_sge; in iser_post_send()
Discsi_iser.h281 int num_sge; member
Diser_memory.c679 sig_wr.num_sge = 1; in iser_reg_sig_mr()
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c1943 struct ocrdma_sge *sge, int num_sge, in ocrdma_build_sges() argument
1948 for (i = 0; i < num_sge; i++) { in ocrdma_build_sges()
1955 if (num_sge == 0) in ocrdma_build_sges()
1959 static inline uint32_t ocrdma_sglist_len(struct ib_sge *sg_list, int num_sge) in ocrdma_sglist_len() argument
1963 for (i = 0; i < num_sge; i++) in ocrdma_sglist_len()
1978 hdr->total_len = ocrdma_sglist_len(wr->sg_list, wr->num_sge); in ocrdma_build_inline_sges()
1986 for (i = 0; i < wr->num_sge; i++) { in ocrdma_build_inline_sges()
1998 ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list); in ocrdma_build_inline_sges()
1999 if (wr->num_sge) in ocrdma_build_inline_sges()
2000 wqe_size += (wr->num_sge * sizeof(struct ocrdma_sge)); in ocrdma_build_inline_sges()
[all …]
/linux-4.1.27/drivers/infiniband/ulp/isert/
Dib_isert.c1020 rx_wr->num_sge = 1; in isert_post_recv()
1054 send_wr.num_sge = tx_desc->num_sge; in isert_post_send()
1079 tx_desc->num_sge = 1; in isert_create_send_desc()
1125 send_wr->num_sge = isert_cmd->tx_desc.num_sge; in isert_init_send_wr()
1147 rx_wr.num_sge = 1; in isert_rdma_post_recvl()
1190 tx_desc->num_sge = 2; in isert_put_login_tx()
2206 isert_cmd->tx_desc.num_sge = 2; in isert_put_response()
2334 isert_cmd->tx_desc.num_sge = 2; in isert_put_reject()
2375 isert_cmd->tx_desc.num_sge = 2; in isert_put_text_rsp()
2427 send_wr->num_sge = ++i; in isert_build_rdma_wr()
[all …]
Dib_isert.h74 int num_sge; member
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dqp.c1989 if (!wr->num_sge) in set_reg_umr_segment()
2106 for (i = 0; i < wr->num_sge; i++) { in set_data_inl_seg()
2380 if (unlikely(wr->num_sge != 1) || in set_sig_umr_wr()
2605 int num_sge; in mlx5_ib_post_send() local
2623 num_sge = wr->num_sge; in mlx5_ib_post_send()
2624 if (unlikely(num_sge > qp->sq.max_gs)) { in mlx5_ib_post_send()
2675 num_sge = 0; in mlx5_ib_post_send()
2688 num_sge = 0; in mlx5_ib_post_send()
2755 num_sge = 0; in mlx5_ib_post_send()
2812 if (wr->send_flags & IB_SEND_INLINE && num_sge) { in mlx5_ib_post_send()
[all …]
Dsrq.c440 if (unlikely(wr->num_sge > srq->msrq.max_gs)) { in mlx5_ib_post_srq_recv()
458 for (i = 0; i < wr->num_sge; i++) { in mlx5_ib_post_srq_recv()
Dmr.c704 wr->num_sge = 1; in prep_umr_reg_wqe()
706 wr->num_sge = 0; in prep_umr_reg_wqe()
934 wr.num_sge = 1; in mlx5_ib_update_mtt()
/linux-4.1.27/drivers/scsi/lpfc/
Dlpfc_scsi.c2264 int i = 0, num_sge = 0, status; in lpfc_bg_setup_sgl() local
2330 num_sge++; in lpfc_bg_setup_sgl()
2350 num_sge++; in lpfc_bg_setup_sgl()
2354 return num_sge; in lpfc_bg_setup_sgl()
2418 int num_sge = 0; in lpfc_bg_setup_sgl_prot() local
2451 if (num_sge >= (phba->cfg_total_seg_cnt - 2)) in lpfc_bg_setup_sgl_prot()
2452 return num_sge + 3; in lpfc_bg_setup_sgl_prot()
2499 num_sge++; in lpfc_bg_setup_sgl_prot()
2530 num_sge++; in lpfc_bg_setup_sgl_prot()
2537 if (num_sge >= phba->cfg_total_seg_cnt) in lpfc_bg_setup_sgl_prot()
[all …]
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dsrq.c330 if (unlikely(wr->num_sge > srq->msrq.max_gs)) { in mlx4_ib_post_srq_recv()
348 for (i = 0; i < wr->num_sge; ++i) { in mlx4_ib_post_srq_recv()
Dqp.c2054 for (i = 0; i < wr->num_sge; ++i) in build_sriov_qp0_header()
2177 for (i = 0; i < wr->num_sge; ++i) in build_mlx_header()
2607 wr->num_sge > qp->sq.max_gs - (halign >> 4))) in build_lso_seg()
2680 if (unlikely(wr->num_sge > qp->sq.max_gs)) { in mlx4_ib_post_send()
2863 dseg += wr->num_sge - 1; in mlx4_ib_post_send()
2864 size += wr->num_sge * (sizeof (struct mlx4_wqe_data_seg) / 16); in mlx4_ib_post_send()
2875 for (i = wr->num_sge - 1; i >= 0; --i, --dseg) in mlx4_ib_post_send()
2985 if (unlikely(wr->num_sge > qp->rq.max_gs)) { in mlx4_ib_post_recv()
3008 for (i = 0; i < wr->num_sge; ++i) in mlx4_ib_post_recv()
Dmad.c598 wr.num_sge = 1; in mlx4_ib_send_to_slave()
1134 recv_wr.num_sge = 1; in mlx4_ib_post_pv_qp_buf()
1251 wr.num_sge = 1; in mlx4_ib_send_to_wire()
/linux-4.1.27/drivers/infiniband/core/
Duverbs_cmd.c2221 if (user_wr->num_sge + sg_ind > cmd.sge_count) { in ib_uverbs_post_send()
2227 user_wr->num_sge * sizeof (struct ib_sge), in ib_uverbs_post_send()
2242 next->num_sge = user_wr->num_sge; in ib_uverbs_post_send()
2300 if (next->num_sge) { in ib_uverbs_post_send()
2307 next->num_sge * sizeof (struct ib_sge))) { in ib_uverbs_post_send()
2311 sg_ind += next->num_sge; in ib_uverbs_post_send()
2378 if (user_wr->num_sge + sg_ind > sge_count) { in ib_uverbs_unmarshall_recv()
2384 user_wr->num_sge * sizeof (struct ib_sge), in ib_uverbs_unmarshall_recv()
2399 next->num_sge = user_wr->num_sge; in ib_uverbs_unmarshall_recv()
2401 if (next->num_sge) { in ib_uverbs_unmarshall_recv()
[all …]
Dmad.c962 mad_send_wr->send_wr.num_sge = 2; in ib_create_send_mad()
2658 recv_wr.num_sge = 1; in ib_mad_post_receive_mads()
/linux-4.1.27/net/sunrpc/xprtrdma/
Dsvc_rdma_sendto.c214 write_wr.num_sge = sge_no; in send_write()
474 send_wr.num_sge = sge_no; in send_reply()
Dsvc_rdma_recvfrom.c200 read_wr.num_sge = pages_needed; in rdma_read_chunk_lcl()
321 read_wr.num_sge = 1; in rdma_read_chunk_frmr()
Dverbs.c1595 send_wr.num_sge = req->rl_niovs; in rpcrdma_ep_post()
1597 if (send_wr.num_sge == 4) /* no need to sync any pad (constant) */ in rpcrdma_ep_post()
1637 recv_wr.num_sge = 1; in rpcrdma_ep_post_recv()
Dsvc_rdma_transport.c539 recv_wr.num_sge = ctxt->count; in svc_rdma_post_recv()
1354 err_wr.num_sge = 1; in svc_rdma_send_error()
/linux-4.1.27/include/uapi/rdma/
Dib_user_verbs.h639 __u32 num_sge; member
683 __u32 num_sge; member
/linux-4.1.27/net/9p/
Dtrans_rdma.c424 wr.num_sge = 1; in post_recv()
517 wr.num_sge = 1; in rdma_request()
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib_verbs.c233 priv->rx_wr.num_sge = 1; in ipoib_transport_dev_init()
Dipoib_cm.c343 wr->num_sge = priv->cm.num_frags; in ipoib_cm_init_rx_wr()
704 priv->tx_wr.num_sge = 1; in post_send()
Dipoib_ib.c534 priv->tx_wr.num_sge = nr_frags + off; in post_send()
/linux-4.1.27/drivers/infiniband/hw/nes/
Dnes_verbs.c3237 for (sge_index = 0; sge_index < ib_wr->num_sge; sge_index++) { in fill_wqe_sg_send()
3322 if (ib_wr->num_sge > nesdev->nesadapter->max_sge) { in nes_post_send()
3345 if (ib_wr->num_sge > nesdev->nesadapter->max_sge) { in nes_post_send()
3347 ib_wr->num_sge, nesdev->nesadapter->max_sge); in nes_post_send()
3378 if (ib_wr->num_sge > 1) { in nes_post_send()
3380 ib_wr->num_sge); in nes_post_send()
3445 " ib_wr=%u, max=1\n", ib_wr->num_sge); in nes_post_send()
3572 if (ib_wr->num_sge > nesdev->nesadapter->max_sge) { in nes_post_recv()
3582 nes_debug(NES_DBG_IW_RX, "ibwr sge count = %u.\n", ib_wr->num_sge); in nes_post_recv()
3592 for (sge_index=0; sge_index < ib_wr->num_sge; sge_index++) { in nes_post_recv()
/linux-4.1.27/drivers/scsi/be2iscsi/
Dbe_main.h293 unsigned int num_sge; member
/linux-4.1.27/include/rdma/
Dib_verbs.h1032 int num_sge; member
1091 int num_sge; member
/linux-4.1.27/drivers/infiniband/ulp/srpt/
Dib_srpt.c790 wr.num_sge = 1; in srpt_post_recv()
826 wr.num_sge = 1; in srpt_post_send()
2817 wr.num_sge = riu->sge_cnt; in srpt_perform_rdmas()
2833 wr.num_sge = 0; in srpt_perform_rdmas()
/linux-4.1.27/drivers/message/fusion/
Dmptbase.c4310 int scale, num_sge, numSGE; in initChainBuffers() local
4348 num_sge = scale + (ioc->req_sz - 60) / ioc->SGE_size; in initChainBuffers()
4350 num_sge = 1 + scale + (ioc->req_sz - 64) / ioc->SGE_size; in initChainBuffers()
4360 ioc->name, num_sge, numSGE)); in initChainBuffers()
4371 while (numSGE - num_sge > 0) { in initChainBuffers()
4373 num_sge += (scale - 1); in initChainBuffers()
4378 ioc->name, numSGE, num_sge, num_chain)); in initChainBuffers()
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c169 rx->rx_wrq.num_sge = 1; in kiblnd_post_rx()
1078 wrq->num_sge = 1; in kiblnd_init_tx_msg()
1142 wrq->num_sge = 1; in kiblnd_init_rdma()
/linux-4.1.27/drivers/infiniband/ulp/srp/
Dib_srp.c1045 .num_sge = 0, in srp_inv_rkey()
1697 wr.num_sge = 1; in srp_post_send()
1717 wr.num_sge = 1; in srp_post_recv()