Lines Matching refs:wr

696 static void prep_umr_reg_wqe(struct ib_pd *pd, struct ib_send_wr *wr,  in prep_umr_reg_wqe()  argument
702 struct mlx5_umr_wr *umrwr = umr_wr(wr); in prep_umr_reg_wqe()
708 wr->next = NULL; in prep_umr_reg_wqe()
709 wr->send_flags = 0; in prep_umr_reg_wqe()
710 wr->sg_list = sg; in prep_umr_reg_wqe()
712 wr->num_sge = 1; in prep_umr_reg_wqe()
714 wr->num_sge = 0; in prep_umr_reg_wqe()
716 wr->opcode = MLX5_IB_WR_UMR; in prep_umr_reg_wqe()
728 struct ib_send_wr *wr, u32 key) in prep_umr_unreg_wqe() argument
730 struct mlx5_umr_wr *umrwr = umr_wr(wr); in prep_umr_unreg_wqe()
732 wr->send_flags = MLX5_IB_SEND_UMR_UNREG | MLX5_IB_SEND_UMR_FAIL_IF_FREE; in prep_umr_unreg_wqe()
733 wr->opcode = MLX5_IB_WR_UMR; in prep_umr_unreg_wqe()
815 umrwr.wr.wr_id = (u64)(unsigned long)&umr_context; in reg_umr()
816 prep_umr_reg_wqe(pd, &umrwr.wr, &sg, dma, npages, mr->mmr.key, in reg_umr()
821 err = ib_post_send(umrc->qp, &umrwr.wr, &bad); in reg_umr()
868 struct mlx5_umr_wr wr; in mlx5_ib_update_mtt() local
932 memset(&wr, 0, sizeof(wr)); in mlx5_ib_update_mtt()
933 wr.wr.wr_id = (u64)(unsigned long)&umr_context; in mlx5_ib_update_mtt()
940 wr.wr.send_flags = MLX5_IB_SEND_UMR_FAIL_IF_FREE | in mlx5_ib_update_mtt()
942 wr.wr.sg_list = &sg; in mlx5_ib_update_mtt()
943 wr.wr.num_sge = 1; in mlx5_ib_update_mtt()
944 wr.wr.opcode = MLX5_IB_WR_UMR; in mlx5_ib_update_mtt()
945 wr.npages = sg.length / sizeof(u64); in mlx5_ib_update_mtt()
946 wr.page_shift = PAGE_SHIFT; in mlx5_ib_update_mtt()
947 wr.mkey = mr->mmr.key; in mlx5_ib_update_mtt()
948 wr.target.offset = start_page_index; in mlx5_ib_update_mtt()
952 err = ib_post_send(umrc->qp, &wr.wr, &bad); in mlx5_ib_update_mtt()
1142 memset(&umrwr.wr, 0, sizeof(umrwr)); in unreg_umr()
1143 umrwr.wr.wr_id = (u64)(unsigned long)&umr_context; in unreg_umr()
1144 prep_umr_unreg_wqe(dev, &umrwr.wr, mr->mmr.key); in unreg_umr()
1148 err = ib_post_send(umrc->qp, &umrwr.wr, &bad); in unreg_umr()