Home
last modified time | relevance | path

Searched refs:wr (Results 1 – 200 of 237) sorted by relevance

12

/linux-4.1.27/drivers/media/dvb-frontends/
Ddib3000mb.c152 wr(DIB3000MB_REG_LOCK1_MASK, DIB3000MB_LOCK1_SEARCH_4); in dib3000mb_set_frontend()
158 wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_2K); in dib3000mb_set_frontend()
162 wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_8K); in dib3000mb_set_frontend()
175 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_32); in dib3000mb_set_frontend()
179 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_16); in dib3000mb_set_frontend()
183 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_8); in dib3000mb_set_frontend()
187 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_4); in dib3000mb_set_frontend()
200 wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_OFF); in dib3000mb_set_frontend()
207 wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_ON); in dib3000mb_set_frontend()
217 wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_QPSK); in dib3000mb_set_frontend()
[all …]
Ddib3000mb_priv.h24 #define wr(reg,val) if (dib3000_write_reg(state,reg,val)) \ macro
31 wr(a[i],v[i]); \
34 #define set_or(reg,val) wr(reg,rd(reg) | val)
36 #define set_and(reg,val) wr(reg,rd(reg) & val)
Dlgdt330x.c110 u8 wr [] = { reg }; in i2c_read_demod_bytes() local
113 .flags = 0, .buf = wr, .len = 1 }, in i2c_read_demod_bytes()
/linux-4.1.27/drivers/infiniband/hw/amso1100/
Dc2_cm.c46 struct c2wr_qp_connect_req *wr; /* variable size needs a malloc. */ in c2_llp_connect() local
81 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in c2_llp_connect()
82 if (!wr) { in c2_llp_connect()
93 c2_wr_set_id(wr, CCWR_QP_CONNECT); in c2_llp_connect()
94 wr->hdr.context = 0; in c2_llp_connect()
95 wr->rnic_handle = c2dev->adapter_handle; in c2_llp_connect()
96 wr->qp_handle = qp->adapter_handle; in c2_llp_connect()
98 wr->remote_addr = raddr->sin_addr.s_addr; in c2_llp_connect()
99 wr->remote_port = raddr->sin_port; in c2_llp_connect()
106 wr->private_data_length = in c2_llp_connect()
[all …]
Dc2_mm.c56 struct c2wr_nsmr_pbl_req *wr; /* PBL WR ptr */ in send_pbl_messages() local
74 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); in send_pbl_messages()
75 if (!wr) { in send_pbl_messages()
78 c2_wr_set_id(wr, CCWR_NSMR_PBL); in send_pbl_messages()
85 wr->hdr.context = 0; in send_pbl_messages()
86 wr->rnic_handle = c2dev->adapter_handle; in send_pbl_messages()
87 wr->stag_index = stag_index; /* already swapped */ in send_pbl_messages()
88 wr->flags = 0; in send_pbl_messages()
92 wr->addrs_length = cpu_to_be32(count); in send_pbl_messages()
105 wr->flags = cpu_to_be32(MEM_PBL_COMPLETE); in send_pbl_messages()
[all …]
Dc2_qp.c137 struct c2wr_qp_modify_req wr; in c2_qp_modify() local
154 c2_wr_set_id(&wr, CCWR_QP_MODIFY); in c2_qp_modify()
155 wr.hdr.context = (unsigned long) vq_req; in c2_qp_modify()
156 wr.rnic_handle = c2dev->adapter_handle; in c2_qp_modify()
157 wr.qp_handle = qp->adapter_handle; in c2_qp_modify()
158 wr.ord = cpu_to_be32(C2_QP_NO_ATTR_CHANGE); in c2_qp_modify()
159 wr.ird = cpu_to_be32(C2_QP_NO_ATTR_CHANGE); in c2_qp_modify()
160 wr.sq_depth = cpu_to_be32(C2_QP_NO_ATTR_CHANGE); in c2_qp_modify()
161 wr.rq_depth = cpu_to_be32(C2_QP_NO_ATTR_CHANGE); in c2_qp_modify()
170 wr.next_qp_state = cpu_to_be32(to_c2_state(attr->qp_state)); in c2_qp_modify()
[all …]
Dc2_rnic.c83 struct c2wr_init_req wr; in c2_adapter_init() local
86 memset(&wr, 0, sizeof(wr)); in c2_adapter_init()
87 c2_wr_set_id(&wr, CCWR_INIT); in c2_adapter_init()
88 wr.hdr.context = 0; in c2_adapter_init()
89 wr.hint_count = cpu_to_be64(c2dev->hint_count_dma); in c2_adapter_init()
90 wr.q0_host_shared = cpu_to_be64(c2dev->req_vq.shared_dma); in c2_adapter_init()
91 wr.q1_host_shared = cpu_to_be64(c2dev->rep_vq.shared_dma); in c2_adapter_init()
92 wr.q1_host_msg_pool = cpu_to_be64(c2dev->rep_vq.host_dma); in c2_adapter_init()
93 wr.q2_host_shared = cpu_to_be64(c2dev->aeq.shared_dma); in c2_adapter_init()
94 wr.q2_host_msg_pool = cpu_to_be64(c2dev->aeq.host_dma); in c2_adapter_init()
[all …]
Dc2_ae.c150 union c2wr *wr; in c2_ae_event() local
164 wr = c2_mq_consume(mq); in c2_ae_event()
165 if (!wr) in c2_ae_event()
171 event_id = c2_wr_get_id(wr); in c2_ae_event()
172 resource_indicator = be32_to_cpu(wr->ae.ae_generic.resource_type); in c2_ae_event()
174 (void *) (unsigned long) wr->ae.ae_generic.user_context; in c2_ae_event()
176 status = cm_event.status = c2_convert_cm_status(c2_wr_get_result(wr)); in c2_ae_event()
200 (unsigned long long) wr->ae.ae_generic.user_context, in c2_ae_event()
201 be32_to_cpu(wr->ae.ae_generic.resource_type), in c2_ae_event()
202 be32_to_cpu(wr->ae.ae_generic.resource), in c2_ae_event()
[all …]
Dc2_cq.c292 struct c2wr_cq_create_req wr; in c2_init_cq() local
320 memset(&wr, 0, sizeof(wr)); in c2_init_cq()
321 c2_wr_set_id(&wr, CCWR_CQ_CREATE); in c2_init_cq()
322 wr.hdr.context = (unsigned long) vq_req; in c2_init_cq()
323 wr.rnic_handle = c2dev->adapter_handle; in c2_init_cq()
324 wr.msg_size = cpu_to_be32(cq->mq.msg_size); in c2_init_cq()
325 wr.depth = cpu_to_be32(cq->mq.q_size); in c2_init_cq()
326 wr.shared_ht = cpu_to_be64(cq->mq.shared_dma); in c2_init_cq()
327 wr.msg_pool = cpu_to_be64(cq->mq.host_dma); in c2_init_cq()
328 wr.user_context = (u64) (unsigned long) (cq); in c2_init_cq()
[all …]
Dc2_wr.h1479 static __inline__ u8 c2_wr_get_id(void *wr) in c2_wr_get_id() argument
1481 return ((struct c2wr_hdr *) wr)->id; in c2_wr_get_id()
1483 static __inline__ void c2_wr_set_id(void *wr, u8 id) in c2_wr_set_id() argument
1485 ((struct c2wr_hdr *) wr)->id = id; in c2_wr_set_id()
1487 static __inline__ u8 c2_wr_get_result(void *wr) in c2_wr_get_result() argument
1489 return ((struct c2wr_hdr *) wr)->result; in c2_wr_get_result()
1491 static __inline__ void c2_wr_set_result(void *wr, u8 result) in c2_wr_set_result() argument
1493 ((struct c2wr_hdr *) wr)->result = result; in c2_wr_set_result()
1495 static __inline__ u8 c2_wr_get_flags(void *wr) in c2_wr_get_flags() argument
1497 return ((struct c2wr_hdr *) wr)->flags; in c2_wr_get_flags()
[all …]
Dc2_vq.c178 int vq_send_wr(struct c2_dev *c2dev, union c2wr *wr) in vq_send_wr() argument
226 memcpy(msg, wr, c2dev->req_vq.msg_size); in vq_send_wr()
Dc2_vq.h57 extern int vq_send_wr(struct c2_dev *c2dev, union c2wr * wr);
/linux-4.1.27/lib/
Ddecompress_unlzma.c292 static inline size_t INIT get_pos(struct writer *wr) in get_pos() argument
295 wr->global_pos + wr->buffer_pos; in get_pos()
298 static inline uint8_t INIT peek_old_byte(struct writer *wr, in peek_old_byte() argument
301 if (!wr->flush) { in peek_old_byte()
303 while (offs > wr->header->dict_size) in peek_old_byte()
304 offs -= wr->header->dict_size; in peek_old_byte()
305 pos = wr->buffer_pos - offs; in peek_old_byte()
306 return wr->buffer[pos]; in peek_old_byte()
308 uint32_t pos = wr->buffer_pos - offs; in peek_old_byte()
309 while (pos >= wr->header->dict_size) in peek_old_byte()
[all …]
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
Diwch_qp.c42 static int build_rdma_send(union t3_wr *wqe, struct ib_send_wr *wr, in build_rdma_send() argument
48 switch (wr->opcode) { in build_rdma_send()
50 if (wr->send_flags & IB_SEND_SOLICITED) in build_rdma_send()
57 if (wr->send_flags & IB_SEND_SOLICITED) in build_rdma_send()
61 wqe->send.rem_stag = cpu_to_be32(wr->ex.invalidate_rkey); in build_rdma_send()
66 if (wr->num_sge > T3_MAX_SGE) in build_rdma_send()
72 for (i = 0; i < wr->num_sge; i++) { in build_rdma_send()
73 if ((plen + wr->sg_list[i].length) < plen) in build_rdma_send()
76 plen += wr->sg_list[i].length; in build_rdma_send()
77 wqe->send.sgl[i].stag = cpu_to_be32(wr->sg_list[i].lkey); in build_rdma_send()
[all …]
Diwch_cm.c179 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in release_tid()
194 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in iwch_quiesce_tid()
195 req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); in iwch_quiesce_tid()
215 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in iwch_resume_tid()
216 req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); in iwch_resume_tid()
401 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_CLOSE_CON)); in send_halfclose()
402 req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); in send_halfclose()
422 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_REQ)); in send_abort()
423 req->wr.wr_lo = htonl(V_WR_TID(ep->hwtid)); in send_abort()
461 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in send_connect()
[all …]
Diwch_provider.h327 int iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
329 int iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
Dcxio_hal.h169 int cxio_peek_cq(struct t3_wq *wr, struct t3_cq *cq, int opcode);
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dqp.c1851 struct ib_send_wr *wr) in set_datagram_seg() argument
1853 memcpy(&dseg->av, &to_mah(wr->wr.ud.ah)->av, sizeof(struct mlx5_av)); in set_datagram_seg()
1854 dseg->av.dqp_dct = cpu_to_be32(wr->wr.ud.remote_qpn | MLX5_EXTENDED_UD_AV); in set_datagram_seg()
1855 dseg->av.key.qkey.qkey = cpu_to_be32(wr->wr.ud.remote_qkey); in set_datagram_seg()
1912 struct ib_send_wr *wr, int li) in set_frwr_umr_segment() argument
1923 umr->klm_octowords = get_klm_octo(wr->wr.fast_reg.page_list_len); in set_frwr_umr_segment()
1965 struct ib_send_wr *wr) in set_reg_umr_segment() argument
1967 struct mlx5_umr_wr *umrwr = (struct mlx5_umr_wr *)&wr->wr.fast_reg; in set_reg_umr_segment()
1971 if (wr->send_flags & MLX5_IB_SEND_UMR_FAIL_IF_FREE) in set_reg_umr_segment()
1976 if (!(wr->send_flags & MLX5_IB_SEND_UMR_UNREG)) { in set_reg_umr_segment()
[all …]
Dmr.c687 static void prep_umr_reg_wqe(struct ib_pd *pd, struct ib_send_wr *wr, in prep_umr_reg_wqe() argument
694 struct mlx5_umr_wr *umrwr = (struct mlx5_umr_wr *)&wr->wr.fast_reg; in prep_umr_reg_wqe()
700 wr->next = NULL; in prep_umr_reg_wqe()
701 wr->send_flags = 0; in prep_umr_reg_wqe()
702 wr->sg_list = sg; in prep_umr_reg_wqe()
704 wr->num_sge = 1; in prep_umr_reg_wqe()
706 wr->num_sge = 0; in prep_umr_reg_wqe()
708 wr->opcode = MLX5_IB_WR_UMR; in prep_umr_reg_wqe()
720 struct ib_send_wr *wr, u32 key) in prep_umr_unreg_wqe() argument
722 struct mlx5_umr_wr *umrwr = (struct mlx5_umr_wr *)&wr->wr.fast_reg; in prep_umr_unreg_wqe()
[all …]
Dsrq.c426 int mlx5_ib_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr, in mlx5_ib_post_srq_recv() argument
439 for (nreq = 0; wr; nreq++, wr = wr->next) { in mlx5_ib_post_srq_recv()
440 if (unlikely(wr->num_sge > srq->msrq.max_gs)) { in mlx5_ib_post_srq_recv()
442 *bad_wr = wr; in mlx5_ib_post_srq_recv()
448 *bad_wr = wr; in mlx5_ib_post_srq_recv()
452 srq->wrid[srq->head] = wr->wr_id; in mlx5_ib_post_srq_recv()
458 for (i = 0; i < wr->num_sge; i++) { in mlx5_ib_post_srq_recv()
459 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx5_ib_post_srq_recv()
460 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx5_ib_post_srq_recv()
461 scat[i].addr = cpu_to_be64(wr->sg_list[i].addr); in mlx5_ib_post_srq_recv()
Dmlx5_ib.h373 struct ib_send_wr wr[2]; member
542 int mlx5_ib_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr,
552 int mlx5_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
554 int mlx5_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
/linux-4.1.27/include/linux/
Dhdlcdrv.h27 unsigned rd, wr; member
34 unsigned int wr; member
48 buf->buffer[buf->wr] = buf->shreg; in hdlcdrv_add_bitbuffer()
49 buf->wr = (buf->wr+1) % sizeof(buf->buffer); in hdlcdrv_add_bitbuffer()
57 buf->buffer[buf->wr] = bits & 0xff; in hdlcdrv_add_bitbuffer_word()
58 buf->wr = (buf->wr+1) % sizeof(buf->buffer); in hdlcdrv_add_bitbuffer_word()
59 buf->buffer[buf->wr] = (bits >> 8) & 0xff; in hdlcdrv_add_bitbuffer_word()
60 buf->wr = (buf->wr+1) % sizeof(buf->buffer); in hdlcdrv_add_bitbuffer_word()
163 ret = !((HDLCDRV_HDLCBUFFER - 1 + hb->rd - hb->wr) % HDLCDRV_HDLCBUFFER); in hdlcdrv_hbuf_full()
176 ret = (hb->rd == hb->wr); in hdlcdrv_hbuf_empty()
[all …]
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_qp.c1479 int ind, struct ib_send_wr *wr, in build_mlx_header() argument
1488 mthca_ah_grh_present(to_mah(wr->wr.ud.ah)), 0, in build_mlx_header()
1491 err = mthca_read_ah(dev, to_mah(wr->wr.ud.ah), &sqp->ud_header); in build_mlx_header()
1502 switch (wr->opcode) { in build_mlx_header()
1510 sqp->ud_header.immediate_data = wr->ex.imm_data; in build_mlx_header()
1519 sqp->ud_header.bth.solicited_event = !!(wr->send_flags & IB_SEND_SOLICITED); in build_mlx_header()
1525 wr->wr.ud.pkey_index, &pkey); in build_mlx_header()
1527 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->wr.ud.remote_qpn); in build_mlx_header()
1529 sqp->ud_header.deth.qkey = cpu_to_be32(wr->wr.ud.remote_qkey & 0x80000000 ? in build_mlx_header()
1530 sqp->qkey : wr->wr.ud.remote_qkey); in build_mlx_header()
[all …]
Dmthca_srq.c475 int mthca_tavor_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr, in mthca_tavor_post_srq_recv() argument
494 for (nreq = 0; wr; wr = wr->next) { in mthca_tavor_post_srq_recv()
502 *bad_wr = wr; in mthca_tavor_post_srq_recv()
514 if (unlikely(wr->num_sge > srq->max_gs)) { in mthca_tavor_post_srq_recv()
516 *bad_wr = wr; in mthca_tavor_post_srq_recv()
521 for (i = 0; i < wr->num_sge; ++i) { in mthca_tavor_post_srq_recv()
522 mthca_set_data_seg(wqe, wr->sg_list + i); in mthca_tavor_post_srq_recv()
532 srq->wrid[ind] = wr->wr_id; in mthca_tavor_post_srq_recv()
575 int mthca_arbel_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr, in mthca_arbel_post_srq_recv() argument
590 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mthca_arbel_post_srq_recv()
[all …]
Dmthca_dev.h522 int mthca_tavor_post_srq_recv(struct ib_srq *srq, struct ib_recv_wr *wr,
524 int mthca_arbel_post_srq_recv(struct ib_srq *srq, struct ib_recv_wr *wr,
533 int mthca_tavor_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
535 int mthca_tavor_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
537 int mthca_arbel_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
539 int mthca_arbel_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
/linux-4.1.27/arch/mips/mm/
Dtlbex.c1869 struct work_registers wr = build_get_work_registers(p); in build_r4000_tlbchange_handler_head() local
1872 build_get_pmde64(p, l, r, wr.r1, wr.r2); /* get pmd in ptr */ in build_r4000_tlbchange_handler_head()
1874 build_get_pgde32(p, wr.r1, wr.r2); /* get pgd in ptr */ in build_r4000_tlbchange_handler_head()
1883 build_is_huge_pte(p, r, wr.r1, wr.r2, label_tlb_huge_update); in build_r4000_tlbchange_handler_head()
1886 UASM_i_MFC0(p, wr.r1, C0_BADVADDR); in build_r4000_tlbchange_handler_head()
1887 UASM_i_LW(p, wr.r2, 0, wr.r2); in build_r4000_tlbchange_handler_head()
1888 UASM_i_SRL(p, wr.r1, wr.r1, PAGE_SHIFT + PTE_ORDER - PTE_T_LOG2); in build_r4000_tlbchange_handler_head()
1889 uasm_i_andi(p, wr.r1, wr.r1, (PTRS_PER_PTE - 1) << PTE_T_LOG2); in build_r4000_tlbchange_handler_head()
1890 UASM_i_ADDU(p, wr.r2, wr.r2, wr.r1); in build_r4000_tlbchange_handler_head()
1895 iPTE_LW(p, wr.r1, wr.r2); /* get even pte */ in build_r4000_tlbchange_handler_head()
[all …]
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dqp.c2034 struct ib_send_wr *wr, in build_sriov_qp0_header() argument
2041 struct mlx4_ib_ah *ah = to_mah(wr->wr.ud.ah); in build_sriov_qp0_header()
2049 if (wr->opcode != IB_WR_SEND) in build_sriov_qp0_header()
2054 for (i = 0; i < wr->num_sge; ++i) in build_sriov_qp0_header()
2055 send_size += wr->sg_list[i].length; in build_sriov_qp0_header()
2080 sqp->ud_header.bth.solicited_event = !!(wr->send_flags & IB_SEND_SOLICITED); in build_sriov_qp0_header()
2084 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->wr.ud.remote_qpn); in build_sriov_qp0_header()
2156 static int build_mlx_header(struct mlx4_ib_sqp *sqp, struct ib_send_wr *wr, in build_mlx_header() argument
2163 struct mlx4_ib_ah *ah = to_mah(wr->wr.ud.ah); in build_mlx_header()
2177 for (i = 0; i < wr->num_sge; ++i) in build_mlx_header()
[all …]
Dsrq.c309 int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr, in mlx4_ib_post_srq_recv() argument
324 *bad_wr = wr; in mlx4_ib_post_srq_recv()
329 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mlx4_ib_post_srq_recv()
330 if (unlikely(wr->num_sge > srq->msrq.max_gs)) { in mlx4_ib_post_srq_recv()
332 *bad_wr = wr; in mlx4_ib_post_srq_recv()
338 *bad_wr = wr; in mlx4_ib_post_srq_recv()
342 srq->wrid[srq->head] = wr->wr_id; in mlx4_ib_post_srq_recv()
348 for (i = 0; i < wr->num_sge; ++i) { in mlx4_ib_post_srq_recv()
349 scat[i].byte_count = cpu_to_be32(wr->sg_list[i].length); in mlx4_ib_post_srq_recv()
350 scat[i].lkey = cpu_to_be32(wr->sg_list[i].lkey); in mlx4_ib_post_srq_recv()
[all …]
Dmr.c324 struct ib_send_wr wr; in mlx4_ib_bind_mw() local
328 memset(&wr, 0, sizeof(wr)); in mlx4_ib_bind_mw()
329 wr.opcode = IB_WR_BIND_MW; in mlx4_ib_bind_mw()
330 wr.wr_id = mw_bind->wr_id; in mlx4_ib_bind_mw()
331 wr.send_flags = mw_bind->send_flags; in mlx4_ib_bind_mw()
332 wr.wr.bind_mw.mw = mw; in mlx4_ib_bind_mw()
333 wr.wr.bind_mw.bind_info = mw_bind->bind_info; in mlx4_ib_bind_mw()
334 wr.wr.bind_mw.rkey = ib_inc_rkey(mw->rkey); in mlx4_ib_bind_mw()
336 ret = mlx4_ib_post_send(qp, &wr, &bad_wr); in mlx4_ib_bind_mw()
338 mw->rkey = wr.wr.bind_mw.rkey; in mlx4_ib_bind_mw()
Dmad.c466 struct ib_send_wr wr, *bad_wr; in mlx4_ib_send_to_slave() local
591 wr.wr.ud.ah = ah; in mlx4_ib_send_to_slave()
592 wr.wr.ud.port_num = port; in mlx4_ib_send_to_slave()
593 wr.wr.ud.remote_qkey = IB_QP_SET_QKEY; in mlx4_ib_send_to_slave()
594 wr.wr.ud.remote_qpn = dqpn; in mlx4_ib_send_to_slave()
595 wr.next = NULL; in mlx4_ib_send_to_slave()
596 wr.wr_id = ((u64) tun_tx_ix) | MLX4_TUN_SET_WRID_QPN(dest_qpt); in mlx4_ib_send_to_slave()
597 wr.sg_list = &list; in mlx4_ib_send_to_slave()
598 wr.num_sge = 1; in mlx4_ib_send_to_slave()
599 wr.opcode = IB_WR_SEND; in mlx4_ib_send_to_slave()
[all …]
Dmlx4_ib.h692 int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr,
703 int mlx4_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
705 int mlx4_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_keys.c341 int qib_fast_reg_mr(struct qib_qp *qp, struct ib_send_wr *wr) in qib_fast_reg_mr() argument
346 u32 rkey = wr->wr.fast_reg.rkey; in qib_fast_reg_mr()
363 if (wr->wr.fast_reg.page_list_len > mr->max_segs) in qib_fast_reg_mr()
366 ps = 1UL << wr->wr.fast_reg.page_shift; in qib_fast_reg_mr()
367 if (wr->wr.fast_reg.length > ps * wr->wr.fast_reg.page_list_len) in qib_fast_reg_mr()
370 mr->user_base = wr->wr.fast_reg.iova_start; in qib_fast_reg_mr()
371 mr->iova = wr->wr.fast_reg.iova_start; in qib_fast_reg_mr()
373 mr->length = wr->wr.fast_reg.length; in qib_fast_reg_mr()
374 mr->access_flags = wr->wr.fast_reg.access_flags; in qib_fast_reg_mr()
375 page_list = wr->wr.fast_reg.page_list->page_list; in qib_fast_reg_mr()
[all …]
Dqib_ud.c62 qp = qib_lookup_qpn(ibp, swqe->wr.wr.ud.remote_qpn); in qib_ud_loopback()
79 ah_attr = &to_iah(swqe->wr.wr.ud.ah)->attr; in qib_ud_loopback()
109 qkey = (int)swqe->wr.wr.ud.remote_qkey < 0 ? in qib_ud_loopback()
110 sqp->qkey : swqe->wr.wr.ud.remote_qkey; in qib_ud_loopback()
133 if (swqe->wr.opcode == IB_WR_SEND_WITH_IMM) { in qib_ud_loopback()
135 wc.ex.imm_data = swqe->wr.ex.imm_data; in qib_ud_loopback()
174 ssge.num_sge = swqe->wr.num_sge; in qib_ud_loopback()
213 swqe->wr.wr.ud.pkey_index : 0; in qib_ud_loopback()
220 swqe->wr.send_flags & IB_SEND_SOLICITED); in qib_ud_loopback()
280 ah_attr = &to_iah(wqe->wr.wr.ud.ah)->attr; in qib_make_ud_req()
[all …]
Dqib_ruc.c429 sqp->s_sge.num_sge = wqe->wr.num_sge; in qib_ruc_loopback()
431 switch (wqe->wr.opcode) { in qib_ruc_loopback()
434 wc.ex.imm_data = wqe->wr.ex.imm_data; in qib_ruc_loopback()
448 wc.ex.imm_data = wqe->wr.ex.imm_data; in qib_ruc_loopback()
461 wqe->wr.wr.rdma.remote_addr, in qib_ruc_loopback()
462 wqe->wr.wr.rdma.rkey, in qib_ruc_loopback()
474 wqe->wr.wr.rdma.remote_addr, in qib_ruc_loopback()
475 wqe->wr.wr.rdma.rkey, in qib_ruc_loopback()
483 qp->r_sge.num_sge = wqe->wr.num_sge; in qib_ruc_loopback()
492 wqe->wr.wr.atomic.remote_addr, in qib_ruc_loopback()
[all …]
Dqib_rc.c51 ss->num_sge = wqe->wr.num_sge; in restart_sge()
318 if ((wqe->wr.send_flags & IB_SEND_FENCE) && in qib_make_rc_req()
334 switch (wqe->wr.opcode) { in qib_make_rc_req()
350 if (wqe->wr.opcode == IB_WR_SEND) in qib_make_rc_req()
355 ohdr->u.imm_data = wqe->wr.ex.imm_data; in qib_make_rc_req()
358 if (wqe->wr.send_flags & IB_SEND_SOLICITED) in qib_make_rc_req()
377 cpu_to_be64(wqe->wr.wr.rdma.remote_addr); in qib_make_rc_req()
379 cpu_to_be32(wqe->wr.wr.rdma.rkey); in qib_make_rc_req()
389 if (wqe->wr.opcode == IB_WR_RDMA_WRITE) in qib_make_rc_req()
395 ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; in qib_make_rc_req()
[all …]
Dqib_uc.c101 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_uc_req()
105 switch (wqe->wr.opcode) { in qib_make_uc_req()
113 if (wqe->wr.opcode == IB_WR_SEND) in qib_make_uc_req()
119 ohdr->u.imm_data = wqe->wr.ex.imm_data; in qib_make_uc_req()
122 if (wqe->wr.send_flags & IB_SEND_SOLICITED) in qib_make_uc_req()
132 cpu_to_be64(wqe->wr.wr.rdma.remote_addr); in qib_make_uc_req()
134 cpu_to_be32(wqe->wr.wr.rdma.rkey); in qib_make_uc_req()
142 if (wqe->wr.opcode == IB_WR_RDMA_WRITE) in qib_make_uc_req()
148 ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; in qib_make_uc_req()
150 if (wqe->wr.send_flags & IB_SEND_SOLICITED) in qib_make_uc_req()
[all …]
Dqib_srq.c48 int qib_post_srq_receive(struct ib_srq *ibsrq, struct ib_recv_wr *wr, in qib_post_srq_receive() argument
56 for (; wr; wr = wr->next) { in qib_post_srq_receive()
61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in qib_post_srq_receive()
62 *bad_wr = wr; in qib_post_srq_receive()
74 *bad_wr = wr; in qib_post_srq_receive()
80 wqe->wr_id = wr->wr_id; in qib_post_srq_receive()
81 wqe->num_sge = wr->num_sge; in qib_post_srq_receive()
82 for (i = 0; i < wr->num_sge; i++) in qib_post_srq_receive()
83 wqe->sg_list[i] = wr->sg_list[i]; in qib_post_srq_receive()
Dqib_verbs.c337 static int qib_post_one_send(struct qib_qp *qp, struct ib_send_wr *wr, in qib_post_one_send() argument
357 if (wr->num_sge > qp->s_max_sge) in qib_post_one_send()
365 if (wr->opcode == IB_WR_FAST_REG_MR) { in qib_post_one_send()
366 if (qib_fast_reg_mr(qp, wr)) in qib_post_one_send()
369 if ((unsigned) wr->opcode >= IB_WR_RDMA_READ) in qib_post_one_send()
373 if (wr->opcode != IB_WR_SEND && in qib_post_one_send()
374 wr->opcode != IB_WR_SEND_WITH_IMM) in qib_post_one_send()
377 if (qp->ibqp.pd != wr->wr.ud.ah->pd) in qib_post_one_send()
379 } else if ((unsigned) wr->opcode > IB_WR_ATOMIC_FETCH_AND_ADD) in qib_post_one_send()
381 else if (wr->opcode >= IB_WR_ATOMIC_CMP_AND_SWP && in qib_post_one_send()
[all …]
Dqib_verbs.h341 struct ib_send_wr wr; /* don't use wr.sg_list */ member
974 int qib_post_ud_send(struct qib_qp *qp, struct ib_send_wr *wr);
989 int qib_post_srq_receive(struct ib_srq *ibsrq, struct ib_recv_wr *wr,
1041 int qib_fast_reg_mr(struct qib_qp *qp, struct ib_send_wr *wr);
Dqib_qp.c432 for (i = 0; i < wqe->wr.num_sge; i++) { in clear_mr_refs()
440 atomic_dec(&to_iah(wqe->wr.wr.ud.ah)->refcount); in clear_mr_refs()
1377 wqe->wr.opcode, in qib_qp_iter_print()
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dqp.c365 struct ib_send_wr *wr, int max, u32 *plenp) in build_immd() argument
373 for (i = 0; i < wr->num_sge; i++) { in build_immd()
374 if ((plen + wr->sg_list[i].length) > max) in build_immd()
376 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; in build_immd()
377 plen += wr->sg_list[i].length; in build_immd()
378 rem = wr->sg_list[i].length; in build_immd()
435 struct ib_send_wr *wr, u8 *len16) in build_rdma_send() argument
441 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_send()
443 switch (wr->opcode) { in build_rdma_send()
445 if (wr->send_flags & IB_SEND_SOLICITED) in build_rdma_send()
[all …]
Dmem.c84 req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR) | in _c4iw_write_mem_dma_aligned()
86 req->wr.wr_lo = wait ? (__force __be64)(unsigned long) &wr_wait : 0L; in _c4iw_write_mem_dma_aligned()
87 req->wr.wr_mid = cpu_to_be32(FW_WR_LEN16_V(DIV_ROUND_UP(wr_len, 16))); in _c4iw_write_mem_dma_aligned()
91 req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), 16)); in _c4iw_write_mem_dma_aligned()
145 req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR) | in _c4iw_write_mem_inline()
147 req->wr.wr_lo = (__force __be64)&wr_wait; in _c4iw_write_mem_inline()
149 req->wr.wr_hi = cpu_to_be32(FW_WR_OP_V(FW_ULPTX_WR)); in _c4iw_write_mem_inline()
150 req->wr.wr_mid = cpu_to_be32( in _c4iw_write_mem_inline()
156 req->len16 = cpu_to_be32(DIV_ROUND_UP(wr_len-sizeof(req->wr), in _c4iw_write_mem_inline()
/linux-4.1.27/arch/sparc/kernel/
Dtrampoline_32.S46 wr %g1, 0x0, %psr ! traps off though
51 wr %g1, 0x0, %wim
55 wr %g3, 0x0, %tbr
70 wr %g1, PSR_ET, %psr ! traps on
100 wr %g1, 0x0, %psr ! traps off though
105 wr %g1, 0x0, %wim
110 wr %g1, 0x0, %tbr
131 wr %g1, PSR_ET, %psr ! traps on
159 wr %g1, 0x0, %psr ! traps off though
164 wr %g1, 0x0, %wim
[all …]
Dentry.S157 wr %l0, 0x0, %psr
179 wr %l4, 0x0, %psr
181 wr %l4, PSR_ET, %psr
198 wr %l0, PSR_ET, %psr
228 wr %g2, 0x0, %psr
230 wr %g2, PSR_ET, %psr
237 wr %g2, PSR_ET, %psr ! keep ET up
247 wr %g2, 0x0, %psr
249 wr %g2, PSR_ET, %psr
253 wr %l0, PSR_ET, %psr
[all …]
Drtrap_32.S57 wr %t_psr, 0x0, %psr
87 wr %t_psr, 0x0, %psr
95 wr %t_psr, PSR_ET, %psr
129 wr %glob_tmp, 0x0, %wim
153 wr %t_psr, 0x0, %psr
164 wr %t_wim, 0x0, %wim ! or else...
166 wr %t_psr, PSR_ET, %psr
191 wr %g1, 0x0, %wim
209 wr %t_psr, 0x0, %psr
216 wr %t_wim, 0x0, %wim
[all …]
Dwuf.S92 wr %twin_tmp1, 0x0, %wim /* Make window 'I' invalid */
121 wr %t_psr, 0x0, %psr
179 wr %t_wim, 0x0, %wim
187 wr %t_psr, PSR_ET, %psr ! enable traps
213 wr %t_psr, 0x0, %psr
299 wr %t_psr, 0x0, %psr
Dwof.S110 wr %glob_tmp, 0x0, %wim ! set new %wim, this is safe now
121 wr %t_psr, 0x0, %psr ! restore condition codes in %psr
147 wr %glob_tmp, 0x0, %wim ! Now it is safe to set new %wim
190 wr %t_psr, 0x0, %psr
252 wr %t_psr, PSR_ET, %psr
285 wr %t_psr, 0x0, %psr
Duna_asm_64.S12 wr %o3, 0, %asi
45 wr %o4, 0x0, %asi
70 wr %o4, 0, %asi
127 wr %o5, 0x0, %asi
Drtrap_64.S192 wr %o3, %g0, %y
231 wr %g0, ASI_AIUP, %asi
336 wr %g1, FPRS_FEF, %fprs
350 wr %g1, 0, %gsr
360 5: wr %g0, FPRS_FEF, %fprs
369 wr %g0, FPRS_DU, %fprs
Dhead_32.S243 wr %g2, 0x0, %psr
292 wr %g3, 0x0, %psr ! tick tock, tick tock
510 wr %g2, 0x0, %psr
560 wr %g0, 0x0, %wim ! so we do not get a trap
573 wr %g1, 0x0, %wim ! make window 1 invalid
649 wr %g3, 0x0, %tbr
654 wr %g3, 0x0, %psr
657 wr %g3, PSR_ET, %psr
Dhvtramp.S94 wr %g0, 0, %fprs
95 wr %g0, ASI_P, %asi
Dtrampoline_64.S56 wr %g1, %asr18
86 wr %g5, %asr25
102 wr %g2, 0, %tick_cmpr
251 wr %g0, 0, %fprs
253 wr %g0, ASI_P, %asi
Dhead_64.S488 wr %g1, %asr18
532 wr %g0, 0, %fprs
676 wr %g0, ASI_P, %asi
821 wr %o2, 0, %tick_cmpr
832 wr %o2, %asr25
950 wr %g0, ASI_AIUS, %asi
956 wr %g0, ASI_AIUS, %asi
Divec.S34 wr %g0, 1 << PIL_DEVICE_IRQ, %set_softint
Dfpu_traps.S24 wr %g0, FPRS_FEF, %fprs
186 wr %g7, 0, %gsr
191 wr %g0, FPRS_FEF, %fprs ! clean DU/DL bits
383 wr %g0, 0, %fprs
Detrap_32.S120 wr %g2, 0x0, %wim
214 wr %g2, 0x0, %wim
Detrap_64.S49 wr %g0, 0, %fprs
79 wr %g3, 0x0, %asi
Dtsb.S558 wr %g0, ASI_BLK_INIT_QUAD_LDD_P, %asi
581 wr %g2, 0x0, %asi
/linux-4.1.27/drivers/infiniband/ulp/isert/
Dib_isert.c54 struct isert_rdma_wr *wr);
59 struct isert_rdma_wr *wr);
1697 struct isert_rdma_wr *wr = &isert_cmd->rdma_wr; in isert_unmap_cmd() local
1701 if (wr->data.sg) { in isert_unmap_cmd()
1703 isert_unmap_data_buf(isert_conn, &wr->data); in isert_unmap_cmd()
1706 if (wr->send_wr) { in isert_unmap_cmd()
1708 kfree(wr->send_wr); in isert_unmap_cmd()
1709 wr->send_wr = NULL; in isert_unmap_cmd()
1712 if (wr->ib_sge) { in isert_unmap_cmd()
1714 kfree(wr->ib_sge); in isert_unmap_cmd()
[all …]
Dib_isert.h220 struct isert_rdma_wr *wr);
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_ruc.c330 sqp->s_sge.num_sge = wqe->wr.num_sge; in ipath_ruc_loopback()
332 switch (wqe->wr.opcode) { in ipath_ruc_loopback()
335 wc.ex.imm_data = wqe->wr.ex.imm_data; in ipath_ruc_loopback()
346 wc.ex.imm_data = wqe->wr.ex.imm_data; in ipath_ruc_loopback()
356 wqe->wr.wr.rdma.remote_addr, in ipath_ruc_loopback()
357 wqe->wr.wr.rdma.rkey, in ipath_ruc_loopback()
366 wqe->wr.wr.rdma.remote_addr, in ipath_ruc_loopback()
367 wqe->wr.wr.rdma.rkey, in ipath_ruc_loopback()
372 qp->r_sge.num_sge = wqe->wr.num_sge; in ipath_ruc_loopback()
380 wqe->wr.wr.atomic.remote_addr, in ipath_ruc_loopback()
[all …]
Dipath_ud.c68 qp = ipath_lookup_qpn(&dev->qp_table, swqe->wr.wr.ud.remote_qpn); in ipath_ud_loopback()
80 ((int) swqe->wr.wr.ud.remote_qkey < 0 ? in ipath_ud_loopback()
81 sqp->qkey : swqe->wr.wr.ud.remote_qkey) != qp->qkey)) { in ipath_ud_loopback()
96 if (swqe->wr.opcode == IB_WR_SEND_WITH_IMM) { in ipath_ud_loopback()
98 wc.ex.imm_data = swqe->wr.ex.imm_data; in ipath_ud_loopback()
178 ah_attr = &to_iah(swqe->wr.wr.ud.ah)->attr; in ipath_ud_loopback()
198 if (--swqe->wr.num_sge) in ipath_ud_loopback()
228 swqe->wr.send_flags & IB_SEND_SOLICITED); in ipath_ud_loopback()
283 ah_attr = &to_iah(wqe->wr.wr.ud.ah)->attr; in ipath_make_ud_req()
325 qp->s_sge.num_sge = wqe->wr.num_sge; in ipath_make_ud_req()
[all …]
Dipath_rc.c50 ss->num_sge = wqe->wr.num_sge; in restart_sge()
294 if ((wqe->wr.send_flags & IB_SEND_FENCE) && in ipath_make_rc_req()
310 switch (wqe->wr.opcode) { in ipath_make_rc_req()
326 if (wqe->wr.opcode == IB_WR_SEND) in ipath_make_rc_req()
331 ohdr->u.imm_data = wqe->wr.ex.imm_data; in ipath_make_rc_req()
334 if (wqe->wr.send_flags & IB_SEND_SOLICITED) in ipath_make_rc_req()
353 cpu_to_be64(wqe->wr.wr.rdma.remote_addr); in ipath_make_rc_req()
355 cpu_to_be32(wqe->wr.wr.rdma.rkey); in ipath_make_rc_req()
365 if (wqe->wr.opcode == IB_WR_RDMA_WRITE) in ipath_make_rc_req()
371 ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; in ipath_make_rc_req()
[all …]
Dipath_uc.c100 qp->s_sge.num_sge = wqe->wr.num_sge; in ipath_make_uc_req()
102 switch (wqe->wr.opcode) { in ipath_make_uc_req()
110 if (wqe->wr.opcode == IB_WR_SEND) in ipath_make_uc_req()
116 ohdr->u.imm_data = wqe->wr.ex.imm_data; in ipath_make_uc_req()
119 if (wqe->wr.send_flags & IB_SEND_SOLICITED) in ipath_make_uc_req()
129 cpu_to_be64(wqe->wr.wr.rdma.remote_addr); in ipath_make_uc_req()
131 cpu_to_be32(wqe->wr.wr.rdma.rkey); in ipath_make_uc_req()
139 if (wqe->wr.opcode == IB_WR_RDMA_WRITE) in ipath_make_uc_req()
145 ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; in ipath_make_uc_req()
147 if (wqe->wr.send_flags & IB_SEND_SOLICITED) in ipath_make_uc_req()
[all …]
Dipath_srq.c48 int ipath_post_srq_receive(struct ib_srq *ibsrq, struct ib_recv_wr *wr, in ipath_post_srq_receive() argument
56 for (; wr; wr = wr->next) { in ipath_post_srq_receive()
61 if ((unsigned) wr->num_sge > srq->rq.max_sge) { in ipath_post_srq_receive()
62 *bad_wr = wr; in ipath_post_srq_receive()
74 *bad_wr = wr; in ipath_post_srq_receive()
80 wqe->wr_id = wr->wr_id; in ipath_post_srq_receive()
81 wqe->num_sge = wr->num_sge; in ipath_post_srq_receive()
82 for (i = 0; i < wr->num_sge; i++) in ipath_post_srq_receive()
83 wqe->sg_list[i] = wr->sg_list[i]; in ipath_post_srq_receive()
Dipath_verbs.c336 static int ipath_post_one_send(struct ipath_qp *qp, struct ib_send_wr *wr) in ipath_post_one_send() argument
360 if (wr->num_sge > qp->s_max_sge) in ipath_post_one_send()
369 if ((unsigned) wr->opcode >= IB_WR_RDMA_READ) in ipath_post_one_send()
373 if (wr->opcode != IB_WR_SEND && in ipath_post_one_send()
374 wr->opcode != IB_WR_SEND_WITH_IMM) in ipath_post_one_send()
377 if (qp->ibqp.pd != wr->wr.ud.ah->pd) in ipath_post_one_send()
379 } else if ((unsigned) wr->opcode > IB_WR_ATOMIC_FETCH_AND_ADD) in ipath_post_one_send()
381 else if (wr->opcode >= IB_WR_ATOMIC_CMP_AND_SWP && in ipath_post_one_send()
382 (wr->num_sge == 0 || in ipath_post_one_send()
383 wr->sg_list[0].length < sizeof(u64) || in ipath_post_one_send()
[all …]
Dipath_verbs.h280 struct ib_send_wr wr; /* don't use wr.sg_list */ member
775 int ipath_post_ud_send(struct ipath_qp *qp, struct ib_send_wr *wr);
791 int ipath_post_srq_receive(struct ib_srq *ibsrq, struct ib_recv_wr *wr,
/linux-4.1.27/arch/arm/boot/dts/
Domap3430-sdp.dts68 gpmc,cs-wr-off-ns = <186>;
71 gpmc,adv-wr-off-ns = <48>;
77 gpmc,wr-cycle-ns = <186>;
82 gpmc,wr-data-mux-bus-ns = <90>;
83 gpmc,wr-access-ns = <186>;
114 gpmc,cs-wr-off-ns = <36>;
117 gpmc,adv-wr-off-ns = <36>;
123 gpmc,wr-cycle-ns = <72>;
125 gpmc,wr-access-ns = <30>;
159 gpmc,cs-wr-off-ns = <72>;
[all …]
Domap-gpmc-smsc911x.dtsi31 gpmc,cs-wr-off-ns = <150>;
34 gpmc,adv-wr-off-ns = <40>;
40 gpmc,wr-cycle-ns = <155>;
47 gpmc,wr-data-mux-bus-ns = <0>;
48 gpmc,wr-access-ns = <0>;
Domap-gpmc-smsc9221.dtsi34 gpmc,cs-wr-off-ns = <36>;
37 gpmc,adv-wr-off-ns = <12>;
43 gpmc,wr-cycle-ns = <54>;
48 gpmc,wr-data-mux-bus-ns = <18>;
49 gpmc,wr-access-ns = <42>;
Domap2430-sdp.dts53 gpmc,cs-wr-off-ns = <187>;
56 gpmc,adv-wr-off-ns = <48>;
62 gpmc,wr-cycle-ns = <187>;
69 gpmc,wr-data-mux-bus-ns = <0>;
70 gpmc,wr-access-ns = <0>;
Domap-zoom-common.dtsi33 gpmc,cs-wr-off-ns = <155>;
36 gpmc,adv-wr-off-ns = <40>;
42 gpmc,wr-cycle-ns = <155>;
49 gpmc,wr-data-mux-bus-ns = <45>;
50 gpmc,wr-access-ns = <145>;
Dorion5x-rd88f5182-nas.dts58 devbus,wr-high-ps = <90000>;
59 devbus,wr-low-ps = <90000>;
60 devbus,ale-wr-ps = <90000>;
80 devbus,wr-high-ps = <90000>;
81 devbus,wr-low-ps = <90000>;
82 devbus,ale-wr-ps = <90000>;
Domap3-sb-t35.dtsi115 gpmc,cs-wr-off-ns = <150>;
118 gpmc,adv-wr-off-ns = <40>;
124 gpmc,wr-cycle-ns = <155>;
131 gpmc,wr-data-mux-bus-ns = <0>;
132 gpmc,wr-access-ns = <0>;
Domap3-devkit8000.dts119 gpmc,cs-wr-off-ns = <44>;
122 gpmc,adv-wr-off-ns = <44>;
127 gpmc,wr-cycle-ns = <82>;
128 gpmc,wr-access-ns = <40>;
129 gpmc,wr-data-mux-bus-ns = <0>;
Ddm8168-evm.dts98 gpmc,cs-wr-off-ns = <44>;
101 gpmc,adv-wr-off-ns = <44>;
108 gpmc,wr-cycle-ns = <82>;
115 gpmc,wr-access-ns = <40>;
116 gpmc,wr-data-mux-bus-ns = <0>;
Domap4-duovero-parlor.dts149 gpmc,cs-wr-off-ns = <50>;
152 gpmc,adv-wr-off-ns = <10>;
158 gpmc,wr-cycle-ns = <50>;
163 gpmc,wr-data-mux-bus-ns = <35>;
164 gpmc,wr-access-ns = <50>;
Domap3-n950-n9.dtsi140 gpmc,cs-wr-off-ns = <87>;
143 gpmc,adv-wr-off-ns = <10>;
149 gpmc,wr-cycle-ns = <112>;
156 gpmc,wr-data-mux-bus-ns = <30>;
157 gpmc,wr-access-ns = <81>;
Domap2420-h4.dts36 gpmc,cs-wr-off-ns = <160>;
39 gpmc,adv-wr-off-ns = <50>;
45 gpmc,wr-cycle-ns = <170>;
Domap3-lilly-a83x.dtsi377 gpmc,cs-wr-off-ns = <100>;
380 gpmc,adv-wr-off-ns = <100>;
386 gpmc,wr-cycle-ns = <100>;
392 gpmc,wr-data-mux-bus-ns = <75>;
393 gpmc,wr-access-ns = <155>;
430 gpmc,cs-wr-off-ns = <60>;
433 gpmc,adv-wr-off-ns = <10>;
439 gpmc,wr-cycle-ns = <100>;
444 gpmc,wr-data-mux-bus-ns = <15>;
445 gpmc,wr-access-ns = <75>;
Domap3-igep.dtsi113 gpmc,cs-wr-off-ns = <44>;
116 gpmc,adv-wr-off-ns = <44>;
121 gpmc,wr-cycle-ns = <82>;
122 gpmc,wr-access-ns = <40>;
123 gpmc,wr-data-mux-bus-ns = <0>;
Domap2420-n8x0-common.dtsi60 gpmc,cs-wr-off-ns = <109>;
63 gpmc,adv-wr-off-ns = <18>;
69 gpmc,wr-cycle-ns = <136>;
Dam335x-chilisom.dtsi219 gpmc,cs-wr-off-ns = <44>;
222 gpmc,adv-wr-off-ns = <44>;
229 gpmc,wr-cycle-ns = <82>;
236 gpmc,wr-access-ns = <40>;
237 gpmc,wr-data-mux-bus-ns = <0>;
Domap3-evm-37xx.dts170 gpmc,cs-wr-off-ns = <44>;
173 gpmc,adv-wr-off-ns = <44>;
178 gpmc,wr-cycle-ns = <82>;
179 gpmc,wr-access-ns = <40>;
180 gpmc,wr-data-mux-bus-ns = <0>;
Domap3-lilly-dbb056.dts141 gpmc,cs-wr-off-ns = <65>;
144 gpmc,adv-wr-off-ns = <10>;
150 gpmc,wr-cycle-ns = <100>;
155 gpmc,wr-data-mux-bus-ns = <15>;
156 gpmc,wr-access-ns = <75>;
Dam335x-igep0033.dtsi139 gpmc,cs-wr-off-ns = <44>;
142 gpmc,adv-wr-off-ns = <44>;
149 gpmc,wr-cycle-ns = <82>;
156 gpmc,wr-access-ns = <40>;
157 gpmc,wr-data-mux-bus-ns = <0>;
Domap3-ldp.dts113 gpmc,cs-wr-off-ns = <44>;
116 gpmc,adv-wr-off-ns = <44>;
121 gpmc,wr-cycle-ns = <82>;
122 gpmc,wr-access-ns = <40>;
123 gpmc,wr-data-mux-bus-ns = <0>;
Domap3-cm-t3x.dtsi274 gpmc,cs-wr-off-ns = <120>;
278 gpmc,adv-wr-off-ns = <120>;
291 gpmc,wr-cycle-ns = <120>;
292 gpmc,wr-access-ns = <186>;
293 gpmc,wr-data-mux-bus-ns = <90>;
Dorion5x-lacie-ethernet-disk-mini-v2.dts78 devbus,wr-high-ps = <90000>;
79 devbus,wr-low-ps = <90000>;
80 devbus,ale-wr-ps = <90000>;
Dam335x-nano.dts265 gpmc,cs-wr-off-ns = <160>;
268 gpmc,adv-wr-off-ns = <30>;
274 gpmc,wr-cycle-ns = <160>;
279 gpmc,wr-data-mux-bus-ns = <70>;
280 gpmc,wr-access-ns = <80>;
Domap3-tao3530.dtsi288 gpmc,cs-wr-off-ns = <36>;
291 gpmc,adv-wr-off-ns = <36>;
297 gpmc,wr-cycle-ns = <72>;
299 gpmc,wr-access-ns = <30>;
Domap3-gta04.dtsi474 gpmc,cs-wr-off-ns = <44>;
477 gpmc,adv-wr-off-ns = <44>;
482 gpmc,wr-cycle-ns = <82>;
483 gpmc,wr-access-ns = <40>;
484 gpmc,wr-data-mux-bus-ns = <0>;
Domap3-n900.dts728 gpmc,cs-wr-off-ns = <87>;
731 gpmc,adv-wr-off-ns = <10>;
737 gpmc,wr-cycle-ns = <112>;
744 gpmc,wr-data-mux-bus-ns = <30>;
745 gpmc,wr-access-ns = <81>;
794 gpmc,cs-wr-off-ns = <24>;
797 gpmc,adv-wr-off-ns = <0>;
805 gpmc,wr-cycle-ns = <180>;
810 gpmc,wr-access-ns = <0>;
811 gpmc,wr-data-mux-bus-ns = <12>;
Domap3-beagle.dts397 gpmc,cs-wr-off-ns = <36>;
400 gpmc,adv-wr-off-ns = <36>;
406 gpmc,wr-cycle-ns = <72>;
408 gpmc,wr-access-ns = <30>;
Darmada-xp-openblocks-ax3-4.dts86 devbus,wr-high-ps = <60000>;
87 devbus,wr-low-ps = <60000>;
88 devbus,ale-wr-ps = <60000>;
Darmada-xp-db.dts96 devbus,wr-high-ps = <60000>;
97 devbus,wr-low-ps = <60000>;
98 devbus,ale-wr-ps = <60000>;
Ddra72-evm.dts306 gpmc,cs-wr-off-ns = <80>;
309 gpmc,adv-wr-off-ns = <60>;
315 gpmc,wr-access-ns = <80>;
317 gpmc,wr-cycle-ns = <80>;
322 gpmc,wr-data-mux-bus-ns = <0>;
Domap3-pandora-common.dtsi510 gpmc,cs-wr-off-ns = <44>;
513 gpmc,adv-wr-off-ns = <44>;
518 gpmc,wr-cycle-ns = <82>;
519 gpmc,wr-access-ns = <40>;
520 gpmc,wr-data-mux-bus-ns = <0>;
Darmada-xp-gp.dts115 devbus,wr-high-ps = <60000>;
116 devbus,wr-low-ps = <60000>;
117 devbus,ale-wr-ps = <60000>;
Ddra7-evm.dts582 gpmc,cs-wr-off-ns = <80>;
585 gpmc,adv-wr-off-ns = <60>;
591 gpmc,wr-access-ns = <80>;
593 gpmc,wr-cycle-ns = <80>;
598 gpmc,wr-data-mux-bus-ns = <0>;
Dam335x-evm.dts457 gpmc,cs-wr-off-ns = <44>;
460 gpmc,adv-wr-off-ns = <44>;
467 gpmc,wr-cycle-ns = <82>;
474 gpmc,wr-access-ns = <40>;
475 gpmc,wr-data-mux-bus-ns = <0>;
Dam437x-gp-evm.dts527 gpmc,cs-wr-off-ns = <40>;
530 gpmc,adv-wr-off-ns = <25>;
537 gpmc,wr-cycle-ns = <40>;
543 gpmc,wr-access-ns = <40>;
544 gpmc,wr-data-mux-bus-ns = <0>;
Dam43x-epos-evm.dts486 gpmc,cs-wr-off-ns = <40>;
489 gpmc,adv-wr-off-ns = <25>; /* min( tALH + tALS + 1) */
496 gpmc,wr-cycle-ns = <40>;
502 gpmc,wr-access-ns = <40>;
503 gpmc,wr-data-mux-bus-ns = <0>;
Dexynos3250-rinato.dts209 wr-setup = <0>;
210 wr-act = <1>;
211 wr-hold = <0>;
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.c1926 struct ib_send_wr *wr) in ocrdma_build_ud_hdr() argument
1930 struct ocrdma_ah *ah = get_ocrdma_ah(wr->wr.ud.ah); in ocrdma_build_ud_hdr()
1932 ud_hdr->rsvd_dest_qpn = wr->wr.ud.remote_qpn; in ocrdma_build_ud_hdr()
1936 ud_hdr->qkey = wr->wr.ud.remote_qkey; in ocrdma_build_ud_hdr()
1972 struct ib_send_wr *wr, u32 wqe_size) in ocrdma_build_inline_sges() argument
1977 if (wr->send_flags & IB_SEND_INLINE && qp->qp_type != IB_QPT_UD) { in ocrdma_build_inline_sges()
1978 hdr->total_len = ocrdma_sglist_len(wr->sg_list, wr->num_sge); in ocrdma_build_inline_sges()
1986 for (i = 0; i < wr->num_sge; i++) { in ocrdma_build_inline_sges()
1988 (void *)(unsigned long)wr->sg_list[i].addr, in ocrdma_build_inline_sges()
1989 wr->sg_list[i].length); in ocrdma_build_inline_sges()
[all …]
/linux-4.1.27/drivers/scsi/csiostor/
Dcsio_scsi.c205 struct fw_scsi_cmd_wr *wr = (struct fw_scsi_cmd_wr *)addr; in csio_scsi_init_cmd_wr() local
209 wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_CMD_WR) | in csio_scsi_init_cmd_wr()
211 wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) | in csio_scsi_init_cmd_wr()
215 wr->cookie = (uintptr_t) req; in csio_scsi_init_cmd_wr()
216 wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx)); in csio_scsi_init_cmd_wr()
217 wr->tmo_val = (uint8_t) req->tmo; in csio_scsi_init_cmd_wr()
218 wr->r3 = 0; in csio_scsi_init_cmd_wr()
219 memset(&wr->r5, 0, 8); in csio_scsi_init_cmd_wr()
225 wr->rsp_dmalen = cpu_to_be32(dma_buf->len); in csio_scsi_init_cmd_wr()
226 wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr); in csio_scsi_init_cmd_wr()
[all …]
Dcsio_lnode.c1415 csio_ln_mgmt_wr_handler(struct csio_hw *hw, void *wr, uint32_t len) in csio_ln_mgmt_wr_handler() argument
1422 wr_cmd = (struct fw_fcoe_els_ct_wr *) wr; in csio_ln_mgmt_wr_handler()
1472 struct fw_wr_hdr *wr; in csio_fcoe_fwevt_handler() local
1508 wr = (struct fw_wr_hdr *) (cmd + 4); in csio_fcoe_fwevt_handler()
1509 if (FW_WR_OP_G(be32_to_cpu(wr->hi)) in csio_fcoe_fwevt_handler()
1577 FW_WR_OP_G(be32_to_cpu((wr->hi)))); in csio_fcoe_fwevt_handler()
1581 wr = (struct fw_wr_hdr *) (cmd); in csio_fcoe_fwevt_handler()
1582 if (FW_WR_OP_G(be32_to_cpu(wr->hi)) == FW_FCOE_ELS_CT_WR) { in csio_fcoe_fwevt_handler()
1583 csio_ln_mgmt_wr_handler(hw, wr, in csio_fcoe_fwevt_handler()
1587 FW_WR_OP_G(be32_to_cpu((wr->hi)))); in csio_fcoe_fwevt_handler()
[all …]
Dcsio_wr.c762 void *wr; in csio_wr_cleanup_iq_ftr() local
771 wr = (void *)((uintptr_t)q->vstart + in csio_wr_cleanup_iq_ftr()
774 ftr = (struct csio_iqwr_footer *)((uintptr_t)wr + in csio_wr_cleanup_iq_ftr()
1037 void *wr, uint32_t len_to_qid, in csio_wr_process_fl() argument
1087 iq_handler(hw, wr, q->wr_sz - sizeof(struct csio_iqwr_footer), in csio_wr_process_fl()
1130 void *wr = (void *)((uintptr_t)q->vstart + (q->cidx * q->wr_sz)); in csio_wr_process_iq() local
1139 ftr = (struct csio_iqwr_footer *)((uintptr_t)wr + in csio_wr_process_iq()
1148 CSIO_DB_ASSERT(((uintptr_t)wr + q->wr_sz) <= in csio_wr_process_iq()
1156 iq_handler(hw, wr, q->wr_sz - sizeof(*ftr), NULL, priv); in csio_wr_process_iq()
1159 csio_wr_process_fl(hw, q, wr, in csio_wr_process_iq()
[all …]
Dcsio_mb.c195 const u32 *params, u32 *val, bool wr, in csio_mb_params() argument
207 (wr ? FW_CMD_WRITE_F : FW_CMD_READ_F) | in csio_mb_params()
213 if (wr) { in csio_mb_params()
301 bool wr, bool init, bool tgt, bool cofld, in csio_mb_caps_config() argument
307 CSIO_INIT_MBP(mbp, cmdp, tmo, hw, cbfn, wr ? 0 : 1); in csio_mb_caps_config()
311 (wr ? FW_CMD_WRITE_F : FW_CMD_READ_F)); in csio_mb_caps_config()
315 if (!wr) in csio_mb_caps_config()
347 uint8_t portid, bool wr, uint32_t fc, uint16_t caps, in csio_mb_port() argument
357 (wr ? FW_CMD_EXEC_F : FW_CMD_READ_F) | in csio_mb_port()
359 if (!wr) { in csio_mb_port()
Dcsio_isr.c131 csio_fwevt_intx_handler(struct csio_hw *hw, void *wr, uint32_t len, in csio_fwevt_intx_handler() argument
146 csio_process_scsi_cmpl(struct csio_hw *hw, void *wr, uint32_t len, in csio_process_scsi_cmpl() argument
155 ioreq = csio_scsi_cmpl_handler(hw, wr, len, flb, NULL, &scsiwr); in csio_process_scsi_cmpl()
285 csio_scsi_intx_handler(struct csio_hw *hw, void *wr, uint32_t len, in csio_scsi_intx_handler() argument
/linux-4.1.27/arch/sparc/include/asm/
Dvisasm.h22 297: wr %g0, FPRS_FEF, %fprs; \
25 wr %g0, 0, %fprs;
42 297: wr %o5, FPRS_FEF, %fprs;
45 wr %o5, 0, %fprs;
Dbackoff.h62 wr tmp, 0, %asr27; \
Dwinmacro.h51 wr %scratch, 0x0, %y;
/linux-4.1.27/drivers/infiniband/ulp/iser/
Diser_memory.c659 struct ib_send_wr *bad_wr, *wr = NULL; in iser_reg_sig_mr() local
672 wr = &inv_wr; in iser_reg_sig_mr()
680 sig_wr.wr.sig_handover.sig_attrs = &sig_attrs; in iser_reg_sig_mr()
681 sig_wr.wr.sig_handover.sig_mr = pi_ctx->sig_mr; in iser_reg_sig_mr()
683 sig_wr.wr.sig_handover.prot = &prot_reg->sge; in iser_reg_sig_mr()
684 sig_wr.wr.sig_handover.access_flags = IB_ACCESS_LOCAL_WRITE | in iser_reg_sig_mr()
688 if (!wr) in iser_reg_sig_mr()
689 wr = &sig_wr; in iser_reg_sig_mr()
691 wr->next = &sig_wr; in iser_reg_sig_mr()
693 ret = ib_post_send(ib_conn->qp, wr, &bad_wr); in iser_reg_sig_mr()
[all …]
/linux-4.1.27/Documentation/devicetree/bindings/mtd/
Dgpmc-nor.txt16 - gpmc,cs-wr-off-ns: Chip-select de-assertion time for writes
23 - gpmc,wr-cycle-ns: Total write cycle time
62 gpmc,cs-wr-off-ns = <186>;
65 gpmc,adv-wr-off-ns = <48>;
71 gpmc,wr-cycle-ns = <186>;
76 gpmc,wr-data-mux-bus-ns = <90>;
77 gpmc,wr-access-ns = <186>;
Dlpc32xx-mlc.txt38 nxp,wr-high = <40000000>;
39 nxp,wr-low = <83333333>;
Dgpmc-nand.txt76 gpmc,cs-wr-off-ns = <44>;
79 gpmc,adv-wr-off-ns = <44>;
84 gpmc,wr-cycle-ns = <82>;
85 gpmc,wr-access-ns = <40>;
86 gpmc,wr-data-mux-bus-ns = <0>;
/linux-4.1.27/net/9p/
Dtrans_rdma.c407 struct ib_recv_wr wr, *bad_wr; in post_recv() local
420 wr.next = NULL; in post_recv()
422 wr.wr_id = (unsigned long) c; in post_recv()
423 wr.sg_list = &sge; in post_recv()
424 wr.num_sge = 1; in post_recv()
425 return ib_post_recv(rdma->qp, &wr, &bad_wr); in post_recv()
435 struct ib_send_wr wr, *bad_wr; in rdma_request() local
511 wr.next = NULL; in rdma_request()
513 wr.wr_id = (unsigned long) c; in rdma_request()
514 wr.opcode = IB_WR_SEND; in rdma_request()
[all …]
Dtrans_fd.c158 struct file *wr; member
246 if (!ts->wr->f_op->poll) in p9_fd_poll()
253 if (ts->rd != ts->wr) { in p9_fd_poll()
254 n = ts->wr->f_op->poll(ts->wr, pt); in p9_fd_poll()
425 if (!(ts->wr->f_flags & O_NONBLOCK)) in p9_fd_write()
431 ret = vfs_write(ts->wr, (__force void __user *)v, len, &ts->wr->f_pos); in p9_fd_write()
794 ts->wr = fget(wfd); in p9_fd_open()
795 if (!ts->rd || !ts->wr) { in p9_fd_open()
798 if (ts->wr) in p9_fd_open()
799 fput(ts->wr); in p9_fd_open()
[all …]
/linux-4.1.27/arch/mips/kvm/
Dmsa.S93 .macro kvm_restore_msa_upper wr, off, base
98 insert_d \wr, 1
101 insert_w \wr, 2
103 insert_w \wr, 3
106 insert_w \wr, 2
108 insert_w \wr, 3
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/
Dcxgb4_uld.h55 (w)->wr.wr_hi = htonl(FW_WR_OP_V(FW_TP_WR) | \
56 FW_WR_IMMDLEN_V(sizeof(*w) - sizeof(w->wr))); \
57 (w)->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(sizeof(*w), 16)) | \
59 (w)->wr.wr_lo = cpu_to_be64(0); \
68 (w)->wr.wr_hi = htonl(FW_WR_OP_V(FW_ULPTX_WR) | \
70 (w)->wr.wr_mid = htonl(FW_WR_LEN16_V(DIV_ROUND_UP(wrlen, 16)) | \
72 (w)->wr.wr_lo = cpu_to_be64(0); \
Dsge.c972 u64 *wr = (u64 *)&q->desc[index]; in ring_tx_db() local
976 wr); in ring_tx_db()
1141 struct fw_eth_tx_pkt_wr *wr; in t4_eth_xmit() local
1200 wr = (void *)&q->q.desc[q->q.pidx]; in t4_eth_xmit()
1201 wr->equiq_to_len16 = htonl(wr_mid); in t4_eth_xmit()
1202 wr->r3 = cpu_to_be64(0); in t4_eth_xmit()
1203 end = (u64 *)wr + flits; in t4_eth_xmit()
1208 struct cpl_tx_pkt_lso *lso = (void *)wr; in t4_eth_xmit()
1214 wr->op_immdlen = htonl(FW_WR_OP_V(FW_ETH_TX_PKT_WR) | in t4_eth_xmit()
1237 wr->op_immdlen = htonl(FW_WR_OP_V(FW_ETH_TX_PKT_WR) | in t4_eth_xmit()
[all …]
/linux-4.1.27/drivers/media/pci/solo6x10/
Dsolo6x10-p2m.c37 int solo_p2m_dma(struct solo_dev *solo_dev, int wr, in solo_p2m_dma() argument
50 wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE); in solo_p2m_dma()
54 ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size, in solo_p2m_dma()
58 wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE); in solo_p2m_dma()
136 void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr, in solo_p2m_fill_desc() argument
145 (wr ? SOLO_P2M_WRITE : 0) | SOLO_P2M_TRANS_ON; in solo_p2m_fill_desc()
157 int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr, in solo_p2m_dma_t() argument
163 solo_p2m_fill_desc(&desc[1], wr, dma_addr, ext_addr, size, repeat, in solo_p2m_dma_t()
Dsolo6x10.h368 int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr,
371 int solo_p2m_dma(struct solo_dev *solo_dev, int wr,
374 void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr,
Dsolo6x10-i2c.c69 static void solo_i2c_flush(struct solo_dev *solo_dev, int wr) in solo_i2c_flush() argument
78 if (wr) { in solo_i2c_flush()
/linux-4.1.27/arch/sparc/lib/
Dxor.S32 0: wr %g0, FPRS_FEF, %fprs
34 wr %g0, ASI_BLK_P, %asi
89 wr %g1, %g0, %asi
91 wr %g0, 0, %fprs
101 0: wr %g0, FPRS_FEF, %fprs
103 wr %g0, ASI_BLK_P, %asi
155 wr %g1, %g0, %asi
157 wr %g0, 0, %fprs
167 0: wr %g0, FPRS_FEF, %fprs
169 wr %g0, ASI_BLK_P, %asi
[all …]
DNGpage.S21 wr %g0, ASI_BLK_INIT_QUAD_LDD_P, %asi
56 wr %g3, 0x0, %asi
67 wr %g0, ASI_BLK_INIT_QUAD_LDD_P, %asi
105 wr %g3, 0x0, %asi
DNGbzero.S44 wr %o4, 0x0, %asi
71 wr %g7, 0x0, %asi
88 wr %o4, 0x0, %asi
110 wr %o5, 0x0, %asi
DGENbzero.S43 wr %o4, 0x0, %asi
86 wr %o4, 0x0, %asi
108 wr %o5, 0x0, %asi
Dcopy_page.S174 wr %g0, ASI_BLK_P, %asi
175 wr %g0, ASI_BLK_COMMIT_P, %asi
225 wr %g3, 0x0, %asi
DNG4memcpy.S21 wr %g0, FPRS_FEF, %fprs; \
27 #define VISExitHalf and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
30 #define VISExitHalf and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
97 wr %g0, 0x80, %asi
DU3memcpy.S14 #define VISEntryHalf rd %fprs, %o5; wr %g0, FPRS_FEF, %fprs; \
16 #define VISExitHalf and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
18 #define VISEntryHalf rd %fprs, %o5; wr %g0, FPRS_FEF, %fprs
19 #define VISExitHalf and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
Dudivdi3.S64 wr %g0, 0, %y
202 wr %g0,%o1,%y ! SPARC has 0-3 delay insn after a wr
Ddivdi3.S86 wr %g0, 0, %y
214 wr %g0,%o1,%y ! SPARC has 0-3 delay insn after a wr
Dmuldi3.S25 wr %g0, %i1, %y
DNG2memcpy.S16 #define VISEntryHalf rd %fprs, %o5; wr %g0, FPRS_FEF, %fprs; \
18 #define VISExitHalf and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
20 #define VISEntryHalf rd %fprs, %o5; wr %g0, FPRS_FEF, %fprs
21 #define VISExitHalf and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
DNGmemcpy.S12 wr TMP, 0x0, %asi;
16 wr %g0, ASI_PNF, %asi
120 wr %g0, STORE_ASI, %asi
DU1memcpy.S16 #define VISEntry rd %fprs, %o5; wr %g0, FPRS_FEF, %fprs; \
18 #define VISExit and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
20 #define VISEntry rd %fprs, %o5; wr %g0, FPRS_FEF, %fprs
21 #define VISExit and %o5, FPRS_FEF, %o5; wr %o5, 0x0, %fprs
/linux-4.1.27/drivers/staging/fbtft/
Dfbtft-io.c147 gpio_set_value(par->gpio.wr, 0); in fbtft_write_gpio8_wr()
152 gpio_set_value(par->gpio.wr, 0); /* used as delay */ in fbtft_write_gpio8_wr()
170 gpio_set_value(par->gpio.wr, 1); in fbtft_write_gpio8_wr()
197 gpio_set_value(par->gpio.wr, 0); in fbtft_write_gpio16_wr()
202 gpio_set_value(par->gpio.wr, 0); /* used as delay */ in fbtft_write_gpio16_wr()
220 gpio_set_value(par->gpio.wr, 1); in fbtft_write_gpio16_wr()
Dfbtft_device.c1137 gpio_set_value(par->gpio.wr, 0); in write_gpio16_wr_slow()
1142 gpio_set_value(par->gpio.wr, 0); /* used as delay */ in write_gpio16_wr_slow()
1160 gpio_set_value(par->gpio.wr, 1); in write_gpio16_wr_slow()
Dfbtft.h238 int wr; member
/linux-4.1.27/Documentation/devicetree/bindings/net/
Dgpmc-eth.txt31 - gpmc,cs-wr-off-ns: Chip-select de-assertion time for writes
38 - gpmc,wr-cycle-ns: Total write cycle time
70 gpmc,cs-wr-off-ns = <186>;
73 gpmc,adv-wr-off-ns = <48>;
79 gpmc,wr-cycle-ns = <186>;
84 gpmc,wr-data-mux-bus-ns = <90>;
85 gpmc,wr-access-ns = <186>;
/linux-4.1.27/drivers/infiniband/core/
Duverbs_cmd.c2186 struct ib_send_wr *wr = NULL, *last, *next, *bad_wr; in ib_uverbs_post_send() local
2235 wr = next; in ib_uverbs_post_send()
2253 next->wr.ud.ah = idr_read_ah(user_wr->wr.ud.ah, in ib_uverbs_post_send()
2255 if (!next->wr.ud.ah) { in ib_uverbs_post_send()
2259 next->wr.ud.remote_qpn = user_wr->wr.ud.remote_qpn; in ib_uverbs_post_send()
2260 next->wr.ud.remote_qkey = user_wr->wr.ud.remote_qkey; in ib_uverbs_post_send()
2271 next->wr.rdma.remote_addr = in ib_uverbs_post_send()
2272 user_wr->wr.rdma.remote_addr; in ib_uverbs_post_send()
2273 next->wr.rdma.rkey = in ib_uverbs_post_send()
2274 user_wr->wr.rdma.rkey; in ib_uverbs_post_send()
[all …]
Dmad.c742 port_num = send_wr->wr.ud.port_num; in handle_outgoing_dr_smp()
784 send_wr->wr.ud.pkey_index, in handle_outgoing_dr_smp()
785 send_wr->wr.ud.port_num, &mad_wc); in handle_outgoing_dr_smp()
965 mad_send_wr->send_wr.wr.ud.remote_qpn = remote_qpn; in ib_create_send_mad()
966 mad_send_wr->send_wr.wr.ud.remote_qkey = IB_QP_SET_QKEY; in ib_create_send_mad()
967 mad_send_wr->send_wr.wr.ud.pkey_index = pkey_index; in ib_create_send_mad()
1165 mad_send_wr->send_wr.wr.ud.ah = send_buf->ah; in ib_post_send_mad()
1750 static inline int rcv_has_same_class(struct ib_mad_send_wr_private *wr, in rcv_has_same_class() argument
1753 return ((struct ib_mad *)(wr->send_buf.mad))->mad_hdr.mgmt_class == in rcv_has_same_class()
1758 struct ib_mad_send_wr_private *wr, in rcv_has_same_gid() argument
[all …]
Dmad_rmpp.c619 static inline void adjust_last_ack(struct ib_mad_send_wr_private *wr, in adjust_last_ack() argument
624 wr->last_ack = seg_num; in adjust_last_ack()
625 list = &wr->last_ack_seg->list; in adjust_last_ack()
626 list_for_each_entry(wr->last_ack_seg, list, list) in adjust_last_ack()
627 if (wr->last_ack_seg->num == seg_num) in adjust_last_ack()
Dagent.c124 mad_send_wr->send_wr.wr.ud.port_num = port_num; in agent_send_response()
/linux-4.1.27/drivers/media/radio/
Dradio-tea5764.c175 struct tea5764_write_regs wr; in tea5764_i2c_write() local
180 .len = sizeof(wr), in tea5764_i2c_write()
181 .buf = (void *)&wr in tea5764_i2c_write()
184 wr.intreg = r->intreg & 0xff; in tea5764_i2c_write()
185 wr.frqset = __cpu_to_be16(r->frqset); in tea5764_i2c_write()
186 wr.tnctrl = __cpu_to_be16(r->tnctrl); in tea5764_i2c_write()
187 wr.testreg = __cpu_to_be16(r->testreg); in tea5764_i2c_write()
188 wr.rdsctrl = __cpu_to_be16(r->rdsctrl); in tea5764_i2c_write()
189 wr.rdsbbl = __cpu_to_be16(r->rdsbbl); in tea5764_i2c_write()
/linux-4.1.27/net/sunrpc/xprtrdma/
Dfrwr_ops.c224 fastreg_wr.wr.fast_reg.iova_start = seg1->mr_dma + pageoff; in frwr_op_map()
225 fastreg_wr.wr.fast_reg.page_list = frmr->fr_pgl; in frwr_op_map()
226 fastreg_wr.wr.fast_reg.page_shift = PAGE_SHIFT; in frwr_op_map()
227 fastreg_wr.wr.fast_reg.page_list_len = page_no; in frwr_op_map()
228 fastreg_wr.wr.fast_reg.length = len; in frwr_op_map()
229 fastreg_wr.wr.fast_reg.access_flags = writing ? in frwr_op_map()
234 fastreg_wr.wr.fast_reg.rkey = mr->rkey; in frwr_op_map()
Dsvc_rdma_recvfrom.c197 read_wr.wr.rdma.rkey = rs_handle; in rdma_read_chunk_lcl()
198 read_wr.wr.rdma.remote_addr = rs_offset; in rdma_read_chunk_lcl()
306 fastreg_wr.wr.fast_reg.iova_start = (unsigned long)frmr->kva; in rdma_read_chunk_frmr()
307 fastreg_wr.wr.fast_reg.page_list = frmr->page_list; in rdma_read_chunk_frmr()
308 fastreg_wr.wr.fast_reg.page_list_len = frmr->page_list_len; in rdma_read_chunk_frmr()
309 fastreg_wr.wr.fast_reg.page_shift = PAGE_SHIFT; in rdma_read_chunk_frmr()
310 fastreg_wr.wr.fast_reg.length = frmr->map_len; in rdma_read_chunk_frmr()
311 fastreg_wr.wr.fast_reg.access_flags = frmr->access_flags; in rdma_read_chunk_frmr()
312 fastreg_wr.wr.fast_reg.rkey = frmr->mr->lkey; in rdma_read_chunk_frmr()
318 read_wr.wr.rdma.rkey = rs_handle; in rdma_read_chunk_frmr()
[all …]
Dsvc_rdma_transport.c1249 fastreg_wr.wr.fast_reg.iova_start = (unsigned long)frmr->kva; in svc_rdma_fastreg()
1250 fastreg_wr.wr.fast_reg.page_list = frmr->page_list; in svc_rdma_fastreg()
1251 fastreg_wr.wr.fast_reg.page_list_len = frmr->page_list_len; in svc_rdma_fastreg()
1252 fastreg_wr.wr.fast_reg.page_shift = PAGE_SHIFT; in svc_rdma_fastreg()
1253 fastreg_wr.wr.fast_reg.length = frmr->map_len; in svc_rdma_fastreg()
1254 fastreg_wr.wr.fast_reg.access_flags = frmr->access_flags; in svc_rdma_fastreg()
1255 fastreg_wr.wr.fast_reg.rkey = frmr->mr->lkey; in svc_rdma_fastreg()
1259 int svc_rdma_send(struct svcxprt_rdma *xprt, struct ib_send_wr *wr) in svc_rdma_send() argument
1270 for (n_wr = wr->next; n_wr; n_wr = n_wr->next) in svc_rdma_send()
1297 ret = ib_post_send(xprt->sc_qp, wr, &bad_wr); in svc_rdma_send()
Dsvc_rdma_sendto.c217 write_wr.wr.rdma.rkey = rmr; in send_write()
218 write_wr.wr.rdma.remote_addr = to; in send_write()
/linux-4.1.27/drivers/gpu/drm/exynos/
Dexynos_drm_dmabuf.c75 struct scatterlist *rd, *wr; in exynos_gem_map_dma_buf() local
101 wr = sgt->sgl; in exynos_gem_map_dma_buf()
103 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in exynos_gem_map_dma_buf()
105 wr = sg_next(wr); in exynos_gem_map_dma_buf()
/linux-4.1.27/drivers/net/hamradio/
Dscc.c275 static inline void wr(struct scc_channel *scc, unsigned char reg, in wr() function
721 wr(scc,R12,tc & 255); /* brg rate LOW */ in set_brg()
722 wr(scc,R13,tc >> 8); /* brg rate HIGH */ in set_brg()
742 wr(scc, R14, BRSRC); /* BRG source = PCLK */ in init_brg()
799 wr(scc,R4,X1CLK|SDLC); /* *1 clock, SDLC mode */ in init_channel()
800 wr(scc,R1,0); /* no W/REQ operation */ in init_channel()
801 wr(scc,R3,Rx8|RxCRC_ENAB); /* RX 8 bits/char, CRC, disabled */ in init_channel()
802 wr(scc,R5,Tx8|DTR|TxCRC_ENAB); /* TX 8 bits/char, disabled, DTR */ in init_channel()
803 wr(scc,R6,0); /* SDLC address zero (not used) */ in init_channel()
804 wr(scc,R7,FLAG); /* SDLC flag value */ in init_channel()
[all …]
Dhdlcdrv.c455 s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0; in hdlcdrv_open()
459 s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0; in hdlcdrv_open()
586 if (s->bitbuf_channel.rd == s->bitbuf_channel.wr) in hdlcdrv_ioctl()
599 if (s->bitbuf_hdlc.rd == s->bitbuf_hdlc.wr) in hdlcdrv_ioctl()
651 s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0; in hdlcdrv_setup()
656 s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0; in hdlcdrv_setup()
666 s->bitbuf_channel.rd = s->bitbuf_channel.wr = 0; in hdlcdrv_setup()
669 s->bitbuf_hdlc.rd = s->bitbuf_hdlc.wr = 0; in hdlcdrv_setup()
/linux-4.1.27/drivers/gpu/drm/udl/
Dudl_dmabuf.c79 struct scatterlist *rd, *wr; in udl_map_dma_buf() local
118 wr = sgt->sgl; in udl_map_dma_buf()
120 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in udl_map_dma_buf()
122 wr = sg_next(wr); in udl_map_dma_buf()
/linux-4.1.27/Documentation/devicetree/bindings/video/
Dsamsung-fimd.txt52 - wr-setup: clock cycles for the active period of CS signal is enabled until
55 - wr-active: clock cycles for the active period of CS is enabled.
57 - wr-hold: clock cycles for the active period of CS is disabled until write
69 | wr-setup+1 | | wr-hold+1 |
72 | wr-active+1|
/linux-4.1.27/drivers/net/ethernet/apm/xgene/
Dxgene_enet_xgmac.c49 static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr, in xgene_enet_wr_indirect() argument
57 iowrite32(wr_data, wr); in xgene_enet_wr_indirect()
75 void __iomem *addr, *wr, *cmd, *cmd_done; in xgene_enet_wr_mac() local
78 wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET; in xgene_enet_wr_mac()
82 if (!xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data)) in xgene_enet_wr_mac()
Dxgene_enet_hw.c252 static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr, in xgene_enet_wr_indirect() argument
260 iowrite32(wr_data, wr); in xgene_enet_wr_indirect()
278 void __iomem *addr, *wr, *cmd, *cmd_done; in xgene_enet_wr_mcx_mac() local
281 wr = pdata->mcx_mac_addr + MAC_WRITE_REG_OFFSET; in xgene_enet_wr_mcx_mac()
285 if (!xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data)) in xgene_enet_wr_mcx_mac()
/linux-4.1.27/Documentation/devicetree/bindings/memory-controllers/
Dmvebu-devbus.txt85 - devbus,ale-wr-ps: Defines the time delay from the ALE[0] negation cycle
89 - devbus,wr-low-ps: Defines the time during which DEV_WEn is active.
95 - devbus,wr-high-ps: Defines the time during which DEV_WEn is kept
98 <wr-high-ps> - <tick> ps.
151 devbus,wr-high-ps = <60000>;
152 devbus,wr-low-ps = <60000>;
153 devbus,ale-wr-ps = <60000>;
/linux-4.1.27/net/rds/
Dib_send.c813 send->s_wr.wr.atomic.compare_add = op->op_m_cswp.compare; in rds_ib_xmit_atomic()
814 send->s_wr.wr.atomic.swap = op->op_m_cswp.swap; in rds_ib_xmit_atomic()
815 send->s_wr.wr.atomic.compare_add_mask = op->op_m_cswp.compare_mask; in rds_ib_xmit_atomic()
816 send->s_wr.wr.atomic.swap_mask = op->op_m_cswp.swap_mask; in rds_ib_xmit_atomic()
819 send->s_wr.wr.atomic.compare_add = op->op_m_fadd.add; in rds_ib_xmit_atomic()
820 send->s_wr.wr.atomic.swap = 0; in rds_ib_xmit_atomic()
821 send->s_wr.wr.atomic.compare_add_mask = op->op_m_fadd.nocarry_mask; in rds_ib_xmit_atomic()
822 send->s_wr.wr.atomic.swap_mask = 0; in rds_ib_xmit_atomic()
827 send->s_wr.wr.atomic.remote_addr = op->op_remote_addr; in rds_ib_xmit_atomic()
828 send->s_wr.wr.atomic.rkey = op->op_rkey; in rds_ib_xmit_atomic()
[all …]
Diw_send.c777 send->s_wr.wr.fast_reg.length = len; in rds_iw_build_send_fastreg()
778 send->s_wr.wr.fast_reg.rkey = send->s_mr->rkey; in rds_iw_build_send_fastreg()
779 send->s_wr.wr.fast_reg.page_list = send->s_page_list; in rds_iw_build_send_fastreg()
780 send->s_wr.wr.fast_reg.page_list_len = nent; in rds_iw_build_send_fastreg()
781 send->s_wr.wr.fast_reg.page_shift = PAGE_SHIFT; in rds_iw_build_send_fastreg()
782 send->s_wr.wr.fast_reg.access_flags = IB_ACCESS_REMOTE_WRITE; in rds_iw_build_send_fastreg()
783 send->s_wr.wr.fast_reg.iova_start = sg_addr; in rds_iw_build_send_fastreg()
882 send->s_wr.wr.rdma.remote_addr = remote_addr; in rds_iw_xmit_rdma()
883 send->s_wr.wr.rdma.rkey = op->op_rkey; in rds_iw_xmit_rdma()
Diw_rdma.c713 f_wr.wr.fast_reg.length = mapping->m_sg.bytes;
714 f_wr.wr.fast_reg.rkey = mapping->m_rkey;
715 f_wr.wr.fast_reg.page_list = ibmr->page_list;
716 f_wr.wr.fast_reg.page_list_len = mapping->m_sg.dma_len;
717 f_wr.wr.fast_reg.page_shift = PAGE_SHIFT;
718 f_wr.wr.fast_reg.access_flags = IB_ACCESS_LOCAL_WRITE |
721 f_wr.wr.fast_reg.iova_start = 0;
Diw_recv.c349 struct ib_send_wr *wr = &ic->i_ack_wr; in rds_iw_recv_init_ack() local
356 wr->sg_list = sge; in rds_iw_recv_init_ack()
357 wr->num_sge = 1; in rds_iw_recv_init_ack()
358 wr->opcode = IB_WR_SEND; in rds_iw_recv_init_ack()
359 wr->wr_id = RDS_IW_ACK_WR_ID; in rds_iw_recv_init_ack()
360 wr->send_flags = IB_SEND_SIGNALED | IB_SEND_SOLICITED; in rds_iw_recv_init_ack()
Dib_recv.c518 struct ib_send_wr *wr = &ic->i_ack_wr; in rds_ib_recv_init_ack() local
525 wr->sg_list = sge; in rds_ib_recv_init_ack()
526 wr->num_sge = 1; in rds_ib_recv_init_ack()
527 wr->opcode = IB_WR_SEND; in rds_ib_recv_init_ack()
528 wr->wr_id = RDS_IB_ACK_WR_ID; in rds_ib_recv_init_ack()
529 wr->send_flags = IB_SEND_SIGNALED | IB_SEND_SOLICITED; in rds_ib_recv_init_ack()
/linux-4.1.27/arch/ia64/hp/sim/
Dsimscsi.c71 static int rd, wr; variable
322 queue[wr].sc = sc; in simscsi_queuecommand_lck()
323 wr = (wr + 1) % SIMSCSI_REQ_QUEUE_LEN; in simscsi_queuecommand_lck()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_reqs.c118 struct ib_mad_hdr *mad_hdr = send_wr->wr.ud.mad_hdr; in trace_send_wr_ud()
163 u32 remote_qkey = send_wr->wr.ud.remote_qkey; in ehca_write_swqe()
226 if (send_wr->wr.ud.remote_qkey & 0x80000000) in ehca_write_swqe()
229 wqe_p->destination_qp_number = send_wr->wr.ud.remote_qpn << 8; in ehca_write_swqe()
231 if (unlikely(!send_wr->wr.ud.ah)) { in ehca_write_swqe()
235 if (unlikely(send_wr->wr.ud.remote_qpn == 0)) { in ehca_write_swqe()
239 my_av = container_of(send_wr->wr.ud.ah, struct ehca_av, ib_ah); in ehca_write_swqe()
258 wqe_p->pkeyi = send_wr->wr.ud.pkey_index; in ehca_write_swqe()
272 send_wr->wr.rdma.remote_addr; in ehca_write_swqe()
273 wqe_p->u.nud.rkey = send_wr->wr.rdma.rkey; in ehca_write_swqe()
/linux-4.1.27/drivers/infiniband/ulp/srp/
Dib_srp.c469 static struct ib_recv_wr wr = { .wr_id = SRP_LAST_WR_ID }; in srp_destroy_qp() local
482 ret = ib_post_recv(ch->qp, &wr, &bad_wr); in srp_destroy_qp()
1041 struct ib_send_wr wr = { in srp_inv_rkey() local
1050 return ib_post_send(ch->qp, &wr, &bad_wr); in srp_inv_rkey()
1298 struct ib_send_wr wr; in srp_map_finish_fr() local
1312 memset(&wr, 0, sizeof(wr)); in srp_map_finish_fr()
1313 wr.opcode = IB_WR_FAST_REG_MR; in srp_map_finish_fr()
1314 wr.wr_id = FAST_REG_WR_ID_MASK; in srp_map_finish_fr()
1315 wr.wr.fast_reg.iova_start = state->base_dma_addr; in srp_map_finish_fr()
1316 wr.wr.fast_reg.page_list = desc->frpl; in srp_map_finish_fr()
[all …]
/linux-4.1.27/drivers/infiniband/ulp/srpt/
Dib_srpt.c779 struct ib_recv_wr wr, *bad_wr; in srpt_post_recv() local
782 wr.wr_id = encode_wr_id(SRPT_RECV, ioctx->ioctx.index); in srpt_post_recv()
788 wr.next = NULL; in srpt_post_recv()
789 wr.sg_list = &list; in srpt_post_recv()
790 wr.num_sge = 1; in srpt_post_recv()
792 return ib_post_srq_recv(sdev->srq, &wr, &bad_wr); in srpt_post_recv()
804 struct ib_send_wr wr, *bad_wr; in srpt_post_send() local
823 wr.next = NULL; in srpt_post_send()
824 wr.wr_id = encode_wr_id(SRPT_SEND, ioctx->ioctx.index); in srpt_post_send()
825 wr.sg_list = &list; in srpt_post_send()
[all …]
/linux-4.1.27/drivers/infiniband/hw/usnic/
Dusnic_ib_verbs.h63 int usnic_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
65 int usnic_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
/linux-4.1.27/arch/sparc/mm/
Dswift.S61 wr %g3, 0x0, %psr
99 wr %g1, 0x0, %psr
127 wr %g3, 0x0, %psr
165 wr %g1, 0x0, %psr
Dultra.S780 wr %g0, (1 << PIL_SMP_CALL_FUNC), %set_softint
785 wr %g0, (1 << PIL_SMP_CALL_FUNC_SNGL), %set_softint
790 wr %g0, (1 << PIL_SMP_RECEIVE_SIGNAL), %set_softint
795 wr %g0, (1 << PIL_SMP_CAPTURE), %set_softint
800 wr %g0, (1 << PIL_SMP_CTX_NEW_VERSION), %set_softint
806 wr %g0, (1 << PIL_KGDB_CAPTURE), %set_softint
/linux-4.1.27/drivers/gpu/drm/nouveau/nvif/
Dobject.c102 struct nvif_ioctl_wr_v0 wr; in nvif_object_wr() member
105 .wr.size = size, in nvif_object_wr()
106 .wr.addr = addr, in nvif_object_wr()
107 .wr.data = data, in nvif_object_wr()
/linux-4.1.27/Documentation/devicetree/bindings/bus/
Dti-gpmc.txt43 - gpmc,cs-wr-off-ns: Write deassertion time
48 - gpmc,adv-wr-off-ns: Write deassertion time
63 - gpmc,wr-cycle-ns: Total write cycle time
83 - gpmc,wr-access-ns: In synchronous write mode, for single or
88 - gpmc,wr-data-mux-bus-ns: In address-data multiplex mode, specifies
/linux-4.1.27/drivers/mtd/nand/
Dmpc5121_nfc.c407 u8 *buffer, uint size, int wr) in mpc5121_nfc_copy_spare() argument
444 if (wr) in mpc5121_nfc_copy_spare()
459 int wr) in mpc5121_nfc_buf_copy() argument
473 mpc5121_nfc_copy_spare(mtd, c, buf, len, wr); in mpc5121_nfc_buf_copy()
484 if (wr) in mpc5121_nfc_buf_copy()
493 mpc5121_nfc_buf_copy(mtd, buf, len, wr); in mpc5121_nfc_buf_copy()
/linux-4.1.27/drivers/i2c/busses/
Di2c-ocores.c301 u32 curr, wr; in oc_setreg_grlib() local
308 wr = (curr & 0xff00) | value; in oc_setreg_grlib()
310 wr = (((u32)value) << 8) | (curr & 0xff); in oc_setreg_grlib()
312 wr = value; in oc_setreg_grlib()
314 iowrite32be(wr, i2c->base + (rreg << i2c->reg_shift)); in oc_setreg_grlib()
/linux-4.1.27/fs/autofs4/
Dwaitq.c63 ssize_t wr = 0; in autofs4_write() local
73 (wr = __vfs_write(file,data,bytes,&file->f_pos)) > 0) { in autofs4_write()
74 data += wr; in autofs4_write()
75 bytes -= wr; in autofs4_write()
83 if (wr == -EPIPE && !sigpipe) { in autofs4_write()
/linux-4.1.27/arch/sparc/power/
Dhibernate_asm.S64 wr %g0, ASI_PHYS_USE_EC, %asi
119 wr %g1, %g0, %asi
/linux-4.1.27/Documentation/devicetree/bindings/mips/cavium/
Dbootbus.txt42 - cavium,t-wr-hld: A cell specifying the WR_HLD timing (in nS).
93 cavium,t-wr-hld = <45>;
113 cavium,t-wr-hld = <70>;
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/
Dsge.c1168 struct fw_eth_tx_pkt_vm_wr *wr; in t4vf_eth_xmit() local
1172 const size_t fw_hdr_copy_len = (sizeof(wr->ethmacdst) + in t4vf_eth_xmit()
1173 sizeof(wr->ethmacsrc) + in t4vf_eth_xmit()
1174 sizeof(wr->ethtype) + in t4vf_eth_xmit()
1175 sizeof(wr->vlantci)); in t4vf_eth_xmit()
1257 wr = (void *)&txq->q.desc[txq->q.pidx]; in t4vf_eth_xmit()
1258 wr->equiq_to_len16 = cpu_to_be32(wr_mid); in t4vf_eth_xmit()
1259 wr->r3[0] = cpu_to_be32(0); in t4vf_eth_xmit()
1260 wr->r3[1] = cpu_to_be32(0); in t4vf_eth_xmit()
1261 skb_copy_from_linear_data(skb, (void *)wr->ethmacdst, fw_hdr_copy_len); in t4vf_eth_xmit()
[all …]
/linux-4.1.27/drivers/scsi/cxgbi/cxgb3i/
Dcxgb3i.c164 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in send_act_open_req()
209 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_CLOSE_CON)); in send_close_req()
210 req->wr.wr_lo = htonl(V_WR_TID(tid)); in send_close_req()
253 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_REQ)); in send_abort_req()
254 req->wr.wr_lo = htonl(V_WR_TID(csk->tid)); in send_abort_req()
284 rpl->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_OFLD_HOST_ABORT_CON_RPL)); in send_abort_rpl()
285 rpl->wr.wr_lo = htonl(V_WR_TID(csk->tid)); in send_abort_rpl()
312 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in send_rx_credits()
1074 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_BYPASS)); in ulp_mem_io_set_hdr()
1153 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in ddp_setup_conn_pgidx()
[all …]
/linux-4.1.27/drivers/media/v4l2-core/
Dvideobuf2-dma-sg.c445 struct scatterlist *rd, *wr; in vb2_dma_sg_dmabuf_ops_attach() local
465 wr = sgt->sgl; in vb2_dma_sg_dmabuf_ops_attach()
467 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in vb2_dma_sg_dmabuf_ops_attach()
469 wr = sg_next(wr); in vb2_dma_sg_dmabuf_ops_attach()
Dvideobuf2-dma-contig.c242 struct scatterlist *rd, *wr; in vb2_dc_dmabuf_ops_attach() local
262 wr = sgt->sgl; in vb2_dc_dmabuf_ops_attach()
264 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in vb2_dc_dmabuf_ops_attach()
266 wr = sg_next(wr); in vb2_dc_dmabuf_ops_attach()
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib_cm.c116 struct ib_recv_wr *wr, in ipoib_cm_post_receive_nonsrq() argument
123 wr->wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_nonsrq()
128 ret = ib_post_recv(rx->qp, wr, &bad_wr); in ipoib_cm_post_receive_nonsrq()
328 struct ib_recv_wr *wr, in ipoib_cm_init_rx_wr() argument
341 wr->next = NULL; in ipoib_cm_init_rx_wr()
342 wr->sg_list = sge; in ipoib_cm_init_rx_wr()
343 wr->num_sge = priv->cm.num_frags; in ipoib_cm_init_rx_wr()
351 struct ib_recv_wr wr; in ipoib_cm_nonsrq_init_rx() member
370 ipoib_cm_init_rx_wr(dev, &t->wr, t->sge); in ipoib_cm_nonsrq_init_rx()
392 ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i); in ipoib_cm_nonsrq_init_rx()
Dipoib_ib.c536 priv->tx_wr.wr.ud.remote_qpn = qpn; in post_send()
537 priv->tx_wr.wr.ud.ah = address; in post_send()
540 priv->tx_wr.wr.ud.mss = skb_shinfo(skb)->gso_size; in post_send()
541 priv->tx_wr.wr.ud.header = head; in post_send()
542 priv->tx_wr.wr.ud.hlen = hlen; in post_send()
/linux-4.1.27/arch/cris/include/arch-v32/arch/hwregs/
Dmarb_defs.h359 unsigned int wr : 1; member
407 unsigned int wr : 1; member
/linux-4.1.27/arch/cris/include/arch-v32/mach-a3/mach/hwregs/
Dmarb_bar_defs.h381 unsigned int wr : 1; member
424 unsigned int wr : 1; member
Dmarb_foo_defs.h507 unsigned int wr : 1; member
556 unsigned int wr : 1; member
Dddr2_defs.h104 unsigned int wr : 3; member
/linux-4.1.27/arch/cris/include/arch-v32/mach-fs/mach/hwregs/
Dmarb_defs.h359 unsigned int wr : 1; member
407 unsigned int wr : 1; member
/linux-4.1.27/drivers/pcmcia/
Dm32r_pcc.c84 void pcc_iorw(int sock, unsigned long port, void *buf, size_t size, size_t nmemb, int wr, int flag) in pcc_iorw() argument
160 if (wr) { in pcc_iorw()
180 if (wr) { in pcc_iorw()
/linux-4.1.27/Documentation/ia64/
Derr_inject.txt464 int wr(char *fn, unsigned long data)
572 wr(fn, err_type_info.err_type_info);
574 wr(fn, 0x0);
580 wr(fn, mode);
649 wr(fn, err_type_info.err_type_info);
651 wr(fn, err_struct_info.err_struct_info);
657 wr(fn,mode);
692 if (wr(fn,virt_addr)<0)
/linux-4.1.27/drivers/infiniband/hw/nes/
Dnes_verbs.c3356 ib_wr->wr.rdma.rkey); in nes_post_send()
3358 ib_wr->wr.rdma.remote_addr); in nes_post_send()
3393 ib_wr->wr.rdma.remote_addr); in nes_post_send()
3395 ib_wr->wr.rdma.rkey); in nes_post_send()
3412 int flags = ib_wr->wr.fast_reg.access_flags; in nes_post_send()
3414 container_of(ib_wr->wr.fast_reg.page_list, in nes_post_send()
3420 if (ib_wr->wr.fast_reg.page_list_len > in nes_post_send()
3429 ib_wr->wr.fast_reg.iova_start); in nes_post_send()
3432 ib_wr->wr.fast_reg.length); in nes_post_send()
3437 ib_wr->wr.fast_reg.rkey); in nes_post_send()
[all …]
/linux-4.1.27/arch/mips/boot/dts/cavium-octeon/
Docteon_68xx.dts471 cavium,t-wr-hld = <35>;
488 cavium,t-wr-hld = <320>;
505 cavium,t-wr-hld = <300>;
522 cavium,t-wr-hld = <30>;
Docteon_3xxx.dts435 cavium,t-wr-hld = <45>;
452 cavium,t-wr-hld = <320>;
469 cavium,t-wr-hld = <30>;
486 cavium,t-wr-hld = <70>;
/linux-4.1.27/drivers/staging/lustre/lustre/llite/
Dvvp_io.c344 start = io->u.ci_wr.wr.crw_pos; in vvp_io_write_lock()
345 end = start + io->u.ci_wr.wr.crw_count - 1; in vvp_io_write_lock()
573 loff_t pos = io->u.ci_wr.wr.crw_pos; in vvp_io_write_start()
574 size_t cnt = io->u.ci_wr.wr.crw_count; in vvp_io_write_start()
584 pos = io->u.ci_wr.wr.crw_pos = i_size_read(inode); in vvp_io_write_start()
/linux-4.1.27/arch/microblaze/boot/dts/
Dsystem.dts69 xlnx,allow-dcache-wr = <0x1>;
70 xlnx,allow-icache-wr = <0x1>;
110 xlnx,number-of-wr-addr-brk = <0x0>;
/linux-4.1.27/arch/powerpc/boot/dts/
Dvirtex440-ml507.dts76 xlnx,dcu-wr-flush-plb-prio = <0>;
77 xlnx,dcu-wr-store-plb-prio = <0>;
78 xlnx,dcu-wr-urgent-plb-prio = <0>;
Dvirtex440-ml510.dts71 xlnx,dcu-wr-flush-plb-prio = <0x0>;
72 xlnx,dcu-wr-store-plb-prio = <0x0>;
73 xlnx,dcu-wr-urgent-plb-prio = <0x0>;
/linux-4.1.27/sound/pci/ctxfi/
Dcthardware.h95 int (*src_set_wr)(void *blk, unsigned int wr);
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/
Dcxgb3_offload.c556 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in mk_tid_release()
878 rpl->wr.wr_hi = in do_abort_req_rss()
880 rpl->wr.wr_lo = htonl(V_WR_TID(tid)); in do_abort_req_rss()
1100 req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); in set_l2t_ix()

12