Lines Matching refs:rq

190 	return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift));  in get_recv_wqe()
389 qp->rq.wqe_cnt = qp->rq.max_gs = 0; in set_rq_size()
395 qp->rq.wqe_cnt = roundup_pow_of_two(max(1U, cap->max_recv_wr)); in set_rq_size()
396 qp->rq.max_gs = roundup_pow_of_two(max(1U, cap->max_recv_sge)); in set_rq_size()
397 qp->rq.wqe_shift = ilog2(qp->rq.max_gs * sizeof (struct mlx4_wqe_data_seg)); in set_rq_size()
402 cap->max_recv_wr = qp->rq.max_post = qp->rq.wqe_cnt; in set_rq_size()
403 cap->max_recv_sge = qp->rq.max_gs; in set_rq_size()
405 cap->max_recv_wr = qp->rq.max_post = in set_rq_size()
406 min(dev->dev->caps.max_wqes - MLX4_IB_SQ_MAX_SPARE, qp->rq.wqe_cnt); in set_rq_size()
407 cap->max_recv_sge = min(qp->rq.max_gs, in set_rq_size()
509 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size()
511 if (qp->rq.wqe_shift > qp->sq.wqe_shift) { in set_kernel_sq_size()
512 qp->rq.offset = 0; in set_kernel_sq_size()
513 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size()
515 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in set_kernel_sq_size()
544 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_user_sq_size()
555 kmalloc(sizeof (struct mlx4_ib_buf) * qp->rq.wqe_cnt, in alloc_proxy_bufs()
559 for (i = 0; i < qp->rq.wqe_cnt; i++) { in alloc_proxy_bufs()
593 for (i = 0; i < qp->rq.wqe_cnt; i++) { in free_proxy_bufs()
698 spin_lock_init(&qp->rq.lock); in create_qp_common()
790 qp->rq.wrid = kmalloc(qp->rq.wqe_cnt * sizeof (u64), gfp); in create_qp_common()
791 if (!qp->sq.wrid || !qp->rq.wrid) { in create_qp_common()
878 kfree(qp->rq.wrid); in create_qp_common()
1048 if (qp->rq.wqe_cnt) in destroy_qp_common()
1054 kfree(qp->rq.wrid); in destroy_qp_common()
1059 if (qp->rq.wqe_cnt) in destroy_qp_common()
1510 if (qp->rq.wqe_cnt) in __mlx4_ib_modify_qp()
1511 context->rq_size_stride = ilog2(qp->rq.wqe_cnt) << 3; in __mlx4_ib_modify_qp()
1512 context->rq_size_stride |= qp->rq.wqe_shift - 4; in __mlx4_ib_modify_qp()
1673 if (qp->rq.wqe_cnt && cur_state == IB_QPS_RESET && new_state == IB_QPS_INIT) in __mlx4_ib_modify_qp()
1811 qp->rq.head = 0; in __mlx4_ib_modify_qp()
1812 qp->rq.tail = 0; in __mlx4_ib_modify_qp()
1816 if (qp->rq.wqe_cnt) in __mlx4_ib_modify_qp()
2966 max_gs = qp->rq.max_gs; in mlx4_ib_post_recv()
2967 spin_lock_irqsave(&qp->rq.lock, flags); in mlx4_ib_post_recv()
2976 ind = qp->rq.head & (qp->rq.wqe_cnt - 1); in mlx4_ib_post_recv()
2979 if (mlx4_wq_overflow(&qp->rq, nreq, qp->ibqp.recv_cq)) { in mlx4_ib_post_recv()
2985 if (unlikely(wr->num_sge > qp->rq.max_gs)) { in mlx4_ib_post_recv()
3017 qp->rq.wrid[ind] = wr->wr_id; in mlx4_ib_post_recv()
3019 ind = (ind + 1) & (qp->rq.wqe_cnt - 1); in mlx4_ib_post_recv()
3024 qp->rq.head += nreq; in mlx4_ib_post_recv()
3032 *qp->db.db = cpu_to_be32(qp->rq.head & 0xffff); in mlx4_ib_post_recv()
3035 spin_unlock_irqrestore(&qp->rq.lock, flags); in mlx4_ib_post_recv()
3180 qp_attr->cap.max_recv_wr = qp->rq.wqe_cnt; in mlx4_ib_query_qp()
3181 qp_attr->cap.max_recv_sge = qp->rq.max_gs; in mlx4_ib_query_qp()