qid               329 arch/mips/cavium-octeon/executive/cvmx-helper-rgmii.c 		pko_mem_queue_qos.s.qid = queue;
qid                83 arch/mips/cavium-octeon/executive/cvmx-pko.c 		config.s.qid		= base_queue + queue;
qid               186 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define PIC_IRT_MSG_Q_INDEX(qid)	((qid) + PIC_IRT_MSG_Q0_INDEX)
qid              1023 arch/mips/include/asm/octeon/cvmx-pko-defs.h 		uint64_t qid:8;
qid              1025 arch/mips/include/asm/octeon/cvmx-pko-defs.h 		uint64_t qid:8;
qid              1047 arch/mips/include/asm/octeon/cvmx-pko-defs.h 		uint64_t qid:8;
qid              1049 arch/mips/include/asm/octeon/cvmx-pko-defs.h 		uint64_t qid:8;
qid              1199 arch/mips/include/asm/octeon/cvmx-pko-defs.h 		uint64_t qid:7;
qid              1201 arch/mips/include/asm/octeon/cvmx-pko-defs.h 		uint64_t qid:7;
qid                78 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_tapq(ap_qid_t qid, unsigned long *info)
qid                80 arch/s390/include/asm/ap.h 	register unsigned long reg0 asm ("0") = qid;
qid               101 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_test_queue(ap_qid_t qid,
qid               106 arch/s390/include/asm/ap.h 		qid |= 1UL << 23; /* set T bit*/
qid               107 arch/s390/include/asm/ap.h 	return ap_tapq(qid, info);
qid               116 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_rapq(ap_qid_t qid)
qid               118 arch/s390/include/asm/ap.h 	register unsigned long reg0 asm ("0") = qid | (1UL << 24);
qid               135 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_zapq(ap_qid_t qid)
qid               137 arch/s390/include/asm/ap.h 	register unsigned long reg0 asm ("0") = qid | (2UL << 24);
qid               219 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_aqic(ap_qid_t qid,
qid               223 arch/s390/include/asm/ap.h 	register unsigned long reg0 asm ("0") = qid | (3UL << 24);
qid               268 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_qact(ap_qid_t qid, int ifbit,
qid               271 arch/s390/include/asm/ap.h 	register unsigned long reg0 asm ("0") = qid | (5UL << 24)
qid               302 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_nqap(ap_qid_t qid,
qid               306 arch/s390/include/asm/ap.h 	register unsigned long reg0 asm ("0") = qid | 0x40000000UL;
qid               340 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_dqap(ap_qid_t qid,
qid               344 arch/s390/include/asm/ap.h 	register unsigned long reg0 asm("0") = qid | 0x80000000UL;
qid               223 arch/s390/include/uapi/asm/zcrypt.h 	unsigned int qid:16;
qid               341 arch/s390/include/uapi/asm/zcrypt.h 	unsigned int qid:14;
qid               249 drivers/block/virtio_blk.c 	int qid = vq->index;
qid               254 drivers/block/virtio_blk.c 	spin_lock_irqsave(&vblk->vqs[qid].lock, flags);
qid               257 drivers/block/virtio_blk.c 		while ((vbr = virtqueue_get_buf(vblk->vqs[qid].vq, &len)) != NULL) {
qid               270 drivers/block/virtio_blk.c 	spin_unlock_irqrestore(&vblk->vqs[qid].lock, flags);
qid               295 drivers/block/virtio_blk.c 	int qid = hctx->queue_num;
qid               351 drivers/block/virtio_blk.c 	spin_lock_irqsave(&vblk->vqs[qid].lock, flags);
qid               353 drivers/block/virtio_blk.c 		err = virtblk_add_req_scsi(vblk->vqs[qid].vq, vbr, vbr->sg, num);
qid               355 drivers/block/virtio_blk.c 		err = virtblk_add_req(vblk->vqs[qid].vq, vbr, vbr->sg, num);
qid               357 drivers/block/virtio_blk.c 		virtqueue_kick(vblk->vqs[qid].vq);
qid               363 drivers/block/virtio_blk.c 		spin_unlock_irqrestore(&vblk->vqs[qid].lock, flags);
qid               374 drivers/block/virtio_blk.c 	if (bd->last && virtqueue_kick_prepare(vblk->vqs[qid].vq))
qid               376 drivers/block/virtio_blk.c 	spin_unlock_irqrestore(&vblk->vqs[qid].lock, flags);
qid               379 drivers/block/virtio_blk.c 		virtqueue_notify(vblk->vqs[qid].vq);
qid               886 drivers/block/xen-blkfront.c 	int qid = hctx->queue_num;
qid               890 drivers/block/xen-blkfront.c 	BUG_ON(info->nr_rings <= qid);
qid               891 drivers/block/xen-blkfront.c 	rinfo = &info->rinfo[qid];
qid               207 drivers/clk/zynqmp/clkc.c 	qdata.qid = PM_QID_CLOCK_GET_NUM_CLOCKS;
qid               231 drivers/clk/zynqmp/clkc.c 	qdata.qid = PM_QID_CLOCK_GET_NAME;
qid               264 drivers/clk/zynqmp/clkc.c 	qdata.qid = PM_QID_CLOCK_GET_TOPOLOGY;
qid               296 drivers/clk/zynqmp/clkc.c 	qdata.qid = PM_QID_CLOCK_GET_FIXEDFACTOR_PARAMS;
qid               338 drivers/clk/zynqmp/clkc.c 	qdata.qid = PM_QID_CLOCK_GET_PARENTS;
qid               364 drivers/clk/zynqmp/clkc.c 	qdata.qid = PM_QID_CLOCK_GET_ATTRIBUTES;
qid               527 drivers/crypto/cavium/zip/zip_regs.h 		u64 qid                         : 3;
qid               531 drivers/crypto/cavium/zip/zip_regs.h 		u64 qid                         : 3;
qid               388 drivers/crypto/chelsio/chcr_algo.c static inline void dsgl_walk_end(struct dsgl_walk *walk, unsigned short qid,
qid               405 drivers/crypto/chelsio/chcr_algo.c 	phys_cpl->rss_hdr_int.qid = htons(qid);
qid               728 drivers/crypto/chelsio/chcr_algo.c 	int qid = u_ctx->lldi.rxq_ids[ctx->rx_qidx];
qid               738 drivers/crypto/chelsio/chcr_algo.c 		FILL_WR_RX_Q_ID(ctx->tx_chan_id, qid,
qid               742 drivers/crypto/chelsio/chcr_algo.c 						       qid);
qid               833 drivers/crypto/chelsio/chcr_algo.c 	chcr_add_cipher_dst_ent(wrparam->req, phys_cpl, wrparam, wrparam->qid);
qid              1165 drivers/crypto/chelsio/chcr_algo.c 	wrparam.qid = u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx];
qid              1189 drivers/crypto/chelsio/chcr_algo.c 				  unsigned short qid,
qid              1281 drivers/crypto/chelsio/chcr_algo.c 	wrparam.qid = qid;
qid              2323 drivers/crypto/chelsio/chcr_algo.c 					 unsigned short qid,
qid              2444 drivers/crypto/chelsio/chcr_algo.c 	chcr_add_aead_dst_ent(req, phys_cpl, qid);
qid              2562 drivers/crypto/chelsio/chcr_algo.c 			   unsigned short qid)
qid              2576 drivers/crypto/chelsio/chcr_algo.c 	dsgl_walk_end(&dsgl_walk, qid, ctx->pci_chan_id);
qid              2605 drivers/crypto/chelsio/chcr_algo.c 			     unsigned short qid)
qid              2618 drivers/crypto/chelsio/chcr_algo.c 	dsgl_walk_end(&dsgl_walk, qid, ctx->pci_chan_id);
qid              2884 drivers/crypto/chelsio/chcr_algo.c 					  unsigned short qid,
qid              2961 drivers/crypto/chelsio/chcr_algo.c 	chcr_add_aead_dst_ent(req, phys_cpl, qid);
qid              2981 drivers/crypto/chelsio/chcr_algo.c 				     unsigned short qid,
qid              3078 drivers/crypto/chelsio/chcr_algo.c 	chcr_add_aead_dst_ent(req, phys_cpl, qid);
qid               230 drivers/crypto/chelsio/chcr_algo.h #define FILL_WR_RX_Q_ID(cid, qid, lcb, fid) \
qid               233 drivers/crypto/chelsio/chcr_algo.h 			FW_CRYPTO_LOOKASIDE_WR_RX_Q_ID_V((qid)) | \
qid               238 drivers/crypto/chelsio/chcr_algo.h #define FILL_ULPTX_CMD_DEST(cid, qid) \
qid               244 drivers/crypto/chelsio/chcr_algo.h 	      ULP_TXPKT_FID_V(qid))
qid               293 drivers/crypto/chelsio/chcr_algo.h 	unsigned short qid;
qid               322 drivers/crypto/chelsio/chcr_crypto.h 			   unsigned short qid);
qid               332 drivers/crypto/chelsio/chcr_crypto.h 			     unsigned short qid);
qid               545 drivers/crypto/chelsio/chcr_ipsec.c 	int qid = q->q.cntxt_id;
qid               578 drivers/crypto/chelsio/chcr_ipsec.c 	wr->req.ulptx.cmd_dest = FILL_ULPTX_CMD_DEST(pi->port_id, qid);
qid                31 drivers/crypto/virtio/virtio_crypto_core.c 	unsigned int qid = vq->index;
qid                33 drivers/crypto/virtio/virtio_crypto_core.c 	spin_lock_irqsave(&vcrypto->data_vq[qid].lock, flags);
qid                38 drivers/crypto/virtio/virtio_crypto_core.c 				&vcrypto->data_vq[qid].lock, flags);
qid                42 drivers/crypto/virtio/virtio_crypto_core.c 				&vcrypto->data_vq[qid].lock, flags);
qid                45 drivers/crypto/virtio/virtio_crypto_core.c 	spin_unlock_irqrestore(&vcrypto->data_vq[qid].lock, flags);
qid               100 drivers/firmware/xilinx/zynqmp-debug.c 		qdata.qid = pm_api_arg[0];
qid               109 drivers/firmware/xilinx/zynqmp-debug.c 		switch (qdata.qid) {
qid               288 drivers/firmware/xilinx/zynqmp.c 	ret = zynqmp_pm_invoke_fn(PM_QUERY_DATA, qdata.qid, qdata.arg1,
qid               296 drivers/firmware/xilinx/zynqmp.c 	return qdata.qid == PM_QID_CLOCK_GET_NAME ? 0 : ret;
qid               181 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c 	unsigned int qid;
qid               188 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c 				&properties, &qid);
qid               195 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c 	pr_debug("DIQ Created with queue id: %d\n", qid);
qid               197 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c 	kq = pqm_get_kernel_queue(dbgdev->pqm, qid);
qid               201 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c 		pqm_destroy_queue(dbgdev->pqm, qid);
qid               906 drivers/gpu/drm/amd/amdkfd/kfd_priv.h 			    unsigned int *qid);
qid               907 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_destroy_queue(struct process_queue_manager *pqm, unsigned int qid);
qid               908 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_update_queue(struct process_queue_manager *pqm, unsigned int qid,
qid               910 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_set_cu_mask(struct process_queue_manager *pqm, unsigned int qid,
qid               912 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid,
qid               915 drivers/gpu/drm/amd/amdkfd/kfd_priv.h 						unsigned int qid);
qid               917 drivers/gpu/drm/amd/amdkfd/kfd_priv.h 		       unsigned int qid,
qid                32 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 			struct process_queue_manager *pqm, unsigned int qid)
qid                37 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		if ((pqn->q && pqn->q->properties.queue_id == qid) ||
qid                38 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		    (pqn->kq && pqn->kq->queue->properties.queue_id == qid))
qid                46 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 					unsigned int *qid)
qid                62 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	*qid = found;
qid                78 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid,
qid                87 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	pqn = get_queue_by_qid(pqm, qid);
qid               168 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 				struct file *f, unsigned int qid)
qid               177 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	q_properties->queue_id = qid;
qid               195 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 			    unsigned int *qid)
qid               226 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	retval = find_available_queue_slot(pqm, qid);
qid               253 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		retval = create_cp_queue(pqm, dev, &q, properties, f, *qid);
qid               274 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		retval = create_cp_queue(pqm, dev, &q, properties, f, *qid);
qid               289 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		kq->queue->properties.queue_id = *qid;
qid               330 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	clear_bit(*qid, pqm->queue_slot_bitmap);
qid               337 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_destroy_queue(struct process_queue_manager *pqm, unsigned int qid)
qid               349 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	pqn = get_queue_by_qid(pqm, qid);
qid               400 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	clear_bit(qid, pqm->queue_slot_bitmap);
qid               410 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_update_queue(struct process_queue_manager *pqm, unsigned int qid,
qid               416 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	pqn = get_queue_by_qid(pqm, qid);
qid               418 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		pr_debug("No queue %d exists for update operation\n", qid);
qid               435 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_set_cu_mask(struct process_queue_manager *pqm, unsigned int qid,
qid               441 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	pqn = get_queue_by_qid(pqm, qid);
qid               443 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		pr_debug("No queue %d exists for update operation\n", qid);
qid               465 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 					unsigned int qid)
qid               469 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	pqn = get_queue_by_qid(pqm, qid);
qid               477 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 		       unsigned int qid,
qid               484 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 	pqn = get_queue_by_qid(pqm, qid);
qid               487 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c 			 qid);
qid              1879 drivers/infiniband/hw/cxgb4/cm.c 			       __func__, ep->com.qp->wq.sq.qid, ep,
qid              3032 drivers/infiniband/hw/cxgb4/cm.c 				ep->com.qp->wq.sq.qid);
qid               195 drivers/infiniband/hw/cxgb4/cq.c 				 CQE_QPID_V(wq->sq.qid));
qid               229 drivers/infiniband/hw/cxgb4/cq.c 				 CQE_QPID_V(wq->sq.qid));
qid               425 drivers/infiniband/hw/cxgb4/cq.c 		WARN_ONCE(1, "Unexpected DRAIN CQE qp id %u!\n", wq->sq.qid);
qid               454 drivers/infiniband/hw/cxgb4/cq.c 		    (CQE_QPID(cqe) == wq->sq.qid) && cqe_completes_wr(cqe, wq))
qid               107 drivers/infiniband/hw/cxgb4/device.c 		le.qid = wq->sq.qid;
qid               113 drivers/infiniband/hw/cxgb4/device.c 		le.qid = wq->rq.qid;
qid               152 drivers/infiniband/hw/cxgb4/device.c 				   lep->qid, lep->opcode,
qid               250 drivers/infiniband/hw/cxgb4/device.c 	if (id != qp->wq.sq.qid)
qid               271 drivers/infiniband/hw/cxgb4/device.c 				      qp->wq.sq.qid, qp->srq ? "srq" : "rq",
qid               272 drivers/infiniband/hw/cxgb4/device.c 				      qp->srq ? qp->srq->idx : qp->wq.rq.qid,
qid               292 drivers/infiniband/hw/cxgb4/device.c 				      qp->wq.sq.qid, qp->wq.rq.qid,
qid               306 drivers/infiniband/hw/cxgb4/device.c 			      qp->wq.sq.qid, qp->wq.rq.qid,
qid               479 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.qid.total, dev->rdev.stats.qid.cur,
qid               480 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.qid.max, dev->rdev.stats.qid.fail);
qid               526 drivers/infiniband/hw/cxgb4/device.c 	dev->rdev.stats.qid.max = 0;
qid               527 drivers/infiniband/hw/cxgb4/device.c 	dev->rdev.stats.qid.fail = 0;
qid               757 drivers/infiniband/hw/cxgb4/device.c 		if (!(entry->qid & rdev->qpmask)) {
qid               759 drivers/infiniband/hw/cxgb4/device.c 					  entry->qid);
qid               761 drivers/infiniband/hw/cxgb4/device.c 			rdev->stats.qid.cur -= rdev->qpmask + 1;
qid               848 drivers/infiniband/hw/cxgb4/device.c 	rdev->stats.qid.total = rdev->lldi.vr->qp.size;
qid              1185 drivers/infiniband/hw/cxgb4/device.c 		u32 qid = be32_to_cpu(rc->pldbuflen_qid);
qid              1186 drivers/infiniband/hw/cxgb4/device.c 		c4iw_ev_handler(dev, qid);
qid              1378 drivers/infiniband/hw/cxgb4/device.c 					  qp->wq.sq.qid,
qid              1383 drivers/infiniband/hw/cxgb4/device.c 			       pci_name(ctx->lldi.pdev), qp->wq.sq.qid);
qid              1391 drivers/infiniband/hw/cxgb4/device.c 					  qp->wq.rq.qid,
qid              1397 drivers/infiniband/hw/cxgb4/device.c 			       pci_name(ctx->lldi.pdev), qp->wq.rq.qid);
qid               211 drivers/infiniband/hw/cxgb4/ev.c 		       CQE_STATUS(err_cqe), qhp->wq.sq.qid);
qid               223 drivers/infiniband/hw/cxgb4/ev.c int c4iw_ev_handler(struct c4iw_dev *dev, u32 qid)
qid               229 drivers/infiniband/hw/cxgb4/ev.c 	chp = xa_load(&dev->cqs, qid);
qid               240 drivers/infiniband/hw/cxgb4/ev.c 		pr_debug("unknown cqid 0x%x\n", qid);
qid               105 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	u32 qid;
qid               129 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	struct c4iw_stat qid;
qid               165 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	u16 qid;
qid              1025 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_ev_handler(struct c4iw_dev *rnicp, u32 qid);
qid              1029 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_put_cqid(struct c4iw_rdev *rdev, u32 qid,
qid              1032 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_put_qpid(struct c4iw_rdev *rdev, u32 qid,
qid              1038 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void __iomem *c4iw_bar2_addrs(struct c4iw_rdev *rdev, unsigned int qid,
qid               159 drivers/infiniband/hw/cxgb4/qp.c 	c4iw_put_qpid(rdev, wq->sq.qid, uctx);
qid               167 drivers/infiniband/hw/cxgb4/qp.c 		c4iw_put_qpid(rdev, wq->rq.qid, uctx);
qid               177 drivers/infiniband/hw/cxgb4/qp.c void __iomem *c4iw_bar2_addrs(struct c4iw_rdev *rdev, unsigned int qid,
qid               184 drivers/infiniband/hw/cxgb4/qp.c 	ret = cxgb4_bar2_sge_qregs(rdev->lldi.ports[0], qid, qtype,
qid               213 drivers/infiniband/hw/cxgb4/qp.c 	wq->sq.qid = c4iw_get_qpid(rdev, uctx);
qid               214 drivers/infiniband/hw/cxgb4/qp.c 	if (!wq->sq.qid)
qid               218 drivers/infiniband/hw/cxgb4/qp.c 		wq->rq.qid = c4iw_get_qpid(rdev, uctx);
qid               219 drivers/infiniband/hw/cxgb4/qp.c 		if (!wq->rq.qid) {
qid               282 drivers/infiniband/hw/cxgb4/qp.c 	wq->sq.bar2_va = c4iw_bar2_addrs(rdev, wq->sq.qid,
qid               287 drivers/infiniband/hw/cxgb4/qp.c 		wq->rq.bar2_va = c4iw_bar2_addrs(rdev, wq->rq.qid,
qid               297 drivers/infiniband/hw/cxgb4/qp.c 			pci_name(rdev->lldi.pdev), wq->sq.qid, wq->rq.qid);
qid               347 drivers/infiniband/hw/cxgb4/qp.c 	res->u.sqrq.eqid = cpu_to_be32(wq->sq.qid);
qid               376 drivers/infiniband/hw/cxgb4/qp.c 		res->u.sqrq.eqid = cpu_to_be32(wq->rq.qid);
qid               381 drivers/infiniband/hw/cxgb4/qp.c 	ret = c4iw_ref_send_wait(rdev, skb, wr_waitp, 0, wq->sq.qid, __func__);
qid               386 drivers/infiniband/hw/cxgb4/qp.c 		 wq->sq.qid, wq->rq.qid, wq->db,
qid               407 drivers/infiniband/hw/cxgb4/qp.c 		c4iw_put_qpid(rdev, wq->rq.qid, uctx);
qid               409 drivers/infiniband/hw/cxgb4/qp.c 	c4iw_put_qpid(rdev, wq->sq.qid, uctx);
qid              1000 drivers/infiniband/hw/cxgb4/qp.c 				 CQE_QPID_V(qhp->wq.sq.qid));
qid              1051 drivers/infiniband/hw/cxgb4/qp.c 				 CQE_QPID_V(qhp->wq.sq.qid));
qid              1566 drivers/infiniband/hw/cxgb4/qp.c 	pr_debug("qhp %p qid 0x%x tid %u\n", qhp, qhp->wq.sq.qid,
qid              1698 drivers/infiniband/hw/cxgb4/qp.c 	pr_debug("qhp %p qid 0x%x tid %u\n", qhp, qhp->wq.sq.qid, ep->hwtid);
qid              1718 drivers/infiniband/hw/cxgb4/qp.c 				 qhp->ep->hwtid, qhp->wq.sq.qid, __func__);
qid              1755 drivers/infiniband/hw/cxgb4/qp.c 		 qhp->wq.sq.qid, qhp->ep->hwtid, qhp->ep->ird, qhp->ep->ord);
qid              1800 drivers/infiniband/hw/cxgb4/qp.c 	wqe->u.init.qpid = cpu_to_be32(qhp->wq.sq.qid);
qid              1801 drivers/infiniband/hw/cxgb4/qp.c 	wqe->u.init.sq_eqid = cpu_to_be32(qhp->wq.sq.qid);
qid              1806 drivers/infiniband/hw/cxgb4/qp.c 		wqe->u.init.rq_eqid = cpu_to_be32(qhp->wq.rq.qid);
qid              1821 drivers/infiniband/hw/cxgb4/qp.c 				 qhp->ep->hwtid, qhp->wq.sq.qid, __func__);
qid              1845 drivers/infiniband/hw/cxgb4/qp.c 		 qhp, qhp->wq.sq.qid, qhp->wq.rq.qid, qhp->ep, qhp->attr.state,
qid              2031 drivers/infiniband/hw/cxgb4/qp.c 		 qhp->wq.sq.qid);
qid              2089 drivers/infiniband/hw/cxgb4/qp.c 	__xa_erase(&rhp->qps, qhp->wq.sq.qid);
qid              2099 drivers/infiniband/hw/cxgb4/qp.c 	pr_debug("ib_qp %p qpid 0x%0x\n", ib_qp, qhp->wq.sq.qid);
qid              2220 drivers/infiniband/hw/cxgb4/qp.c 	ret = xa_insert_irq(&rhp->qps, qhp->wq.sq.qid, qhp, GFP_KERNEL);
qid              2263 drivers/infiniband/hw/cxgb4/qp.c 		uresp.sqid = qhp->wq.sq.qid;
qid              2267 drivers/infiniband/hw/cxgb4/qp.c 			uresp.rqid = qhp->wq.rq.qid;
qid              2334 drivers/infiniband/hw/cxgb4/qp.c 	qhp->ibqp.qp_num = qhp->wq.sq.qid;
qid              2339 drivers/infiniband/hw/cxgb4/qp.c 		 qhp->wq.sq.qid, qhp->wq.sq.size, qhp->wq.sq.memsize,
qid              2340 drivers/infiniband/hw/cxgb4/qp.c 		 attrs->cap.max_send_wr, qhp->wq.rq.qid, qhp->wq.rq.size,
qid              2356 drivers/infiniband/hw/cxgb4/qp.c 	xa_erase_irq(&rhp->qps, qhp->wq.sq.qid);
qid              2504 drivers/infiniband/hw/cxgb4/qp.c 	res->u.srq.eqid = cpu_to_be32(wq->qid);
qid              2514 drivers/infiniband/hw/cxgb4/qp.c 	c4iw_put_qpid(rdev, wq->qid, uctx);
qid              2530 drivers/infiniband/hw/cxgb4/qp.c 	wq->qid = c4iw_get_qpid(rdev, uctx);
qid              2531 drivers/infiniband/hw/cxgb4/qp.c 	if (!wq->qid)
qid              2560 drivers/infiniband/hw/cxgb4/qp.c 	wq->bar2_va = c4iw_bar2_addrs(rdev, wq->qid, CXGB4_BAR2_QTYPE_EGRESS,
qid              2570 drivers/infiniband/hw/cxgb4/qp.c 			pci_name(rdev->lldi.pdev), wq->qid);
qid              2599 drivers/infiniband/hw/cxgb4/qp.c 	res->u.srq.eqid = cpu_to_be32(wq->qid);
qid              2623 drivers/infiniband/hw/cxgb4/qp.c 	ret = c4iw_ref_send_wait(rdev, skb, wr_waitp, 0, wq->qid, __func__);
qid              2629 drivers/infiniband/hw/cxgb4/qp.c 			__func__, srq->idx, wq->qid, srq->pdid, wq->queue,
qid              2647 drivers/infiniband/hw/cxgb4/qp.c 	c4iw_put_qpid(rdev, wq->qid, uctx);
qid              2754 drivers/infiniband/hw/cxgb4/qp.c 		uresp.srqid = srq->wq.qid;
qid              2778 drivers/infiniband/hw/cxgb4/qp.c 		 __func__, srq->wq.qid, srq->idx, srq->wq.size,
qid              2809 drivers/infiniband/hw/cxgb4/qp.c 	pr_debug("%s id %d\n", __func__, srq->wq.qid);
qid               111 drivers/infiniband/hw/cxgb4/resource.c 	u32 qid;
qid               119 drivers/infiniband/hw/cxgb4/resource.c 		qid = entry->qid;
qid               122 drivers/infiniband/hw/cxgb4/resource.c 		qid = c4iw_get_resource(&rdev->resource.qid_table);
qid               123 drivers/infiniband/hw/cxgb4/resource.c 		if (!qid)
qid               126 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.cur += rdev->qpmask + 1;
qid               128 drivers/infiniband/hw/cxgb4/resource.c 		for (i = qid+1; i & rdev->qpmask; i++) {
qid               132 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
qid               143 drivers/infiniband/hw/cxgb4/resource.c 		entry->qid = qid;
qid               145 drivers/infiniband/hw/cxgb4/resource.c 		for (i = qid+1; i & rdev->qpmask; i++) {
qid               149 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
qid               155 drivers/infiniband/hw/cxgb4/resource.c 	pr_debug("qid 0x%x\n", qid);
qid               157 drivers/infiniband/hw/cxgb4/resource.c 	if (rdev->stats.qid.cur > rdev->stats.qid.max)
qid               158 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.max = rdev->stats.qid.cur;
qid               160 drivers/infiniband/hw/cxgb4/resource.c 	return qid;
qid               163 drivers/infiniband/hw/cxgb4/resource.c void c4iw_put_cqid(struct c4iw_rdev *rdev, u32 qid,
qid               171 drivers/infiniband/hw/cxgb4/resource.c 	pr_debug("qid 0x%x\n", qid);
qid               172 drivers/infiniband/hw/cxgb4/resource.c 	entry->qid = qid;
qid               181 drivers/infiniband/hw/cxgb4/resource.c 	u32 qid;
qid               189 drivers/infiniband/hw/cxgb4/resource.c 		qid = entry->qid;
qid               192 drivers/infiniband/hw/cxgb4/resource.c 		qid = c4iw_get_resource(&rdev->resource.qid_table);
qid               193 drivers/infiniband/hw/cxgb4/resource.c 		if (!qid) {
qid               195 drivers/infiniband/hw/cxgb4/resource.c 			rdev->stats.qid.fail++;
qid               200 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.cur += rdev->qpmask + 1;
qid               202 drivers/infiniband/hw/cxgb4/resource.c 		for (i = qid+1; i & rdev->qpmask; i++) {
qid               206 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
qid               217 drivers/infiniband/hw/cxgb4/resource.c 		entry->qid = qid;
qid               219 drivers/infiniband/hw/cxgb4/resource.c 		for (i = qid; i & rdev->qpmask; i++) {
qid               223 drivers/infiniband/hw/cxgb4/resource.c 			entry->qid = i;
qid               229 drivers/infiniband/hw/cxgb4/resource.c 	pr_debug("qid 0x%x\n", qid);
qid               231 drivers/infiniband/hw/cxgb4/resource.c 	if (rdev->stats.qid.cur > rdev->stats.qid.max)
qid               232 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.max = rdev->stats.qid.cur;
qid               234 drivers/infiniband/hw/cxgb4/resource.c 	return qid;
qid               237 drivers/infiniband/hw/cxgb4/resource.c void c4iw_put_qpid(struct c4iw_rdev *rdev, u32 qid,
qid               245 drivers/infiniband/hw/cxgb4/resource.c 	pr_debug("qid 0x%x\n", qid);
qid               246 drivers/infiniband/hw/cxgb4/resource.c 	entry->qid = qid;
qid                42 drivers/infiniband/hw/cxgb4/restrack.c 	if (rdma_nl_put_driver_u32(msg, "sqid", wq->sq.qid))
qid                70 drivers/infiniband/hw/cxgb4/restrack.c 	if (rdma_nl_put_driver_u32(msg, "rqid", wq->rq.qid))
qid                51 drivers/infiniband/hw/cxgb4/t4.h 	__be16 qid;
qid               343 drivers/infiniband/hw/cxgb4/t4.h 	u32 qid;
qid               370 drivers/infiniband/hw/cxgb4/t4.h 	u32 qid;
qid               407 drivers/infiniband/hw/cxgb4/t4.h 	u32 qid;
qid               632 drivers/infiniband/hw/cxgb4/t4.h 	writel(QID_V(wq->sq.qid) | PIDX_V(inc), wq->db);
qid               657 drivers/infiniband/hw/cxgb4/t4.h 	writel(QID_V(wq->rq.qid) | PIDX_V(inc), wq->db);
qid              1316 drivers/net/bonding/bond_options.c 	u16 qid;
qid              1327 drivers/net/bonding/bond_options.c 	if (sscanf(++delim, "%hd\n", &qid) != 1)
qid              1332 drivers/net/bonding/bond_options.c 	    qid > bond->dev->real_num_tx_queues)
qid              1348 drivers/net/bonding/bond_options.c 		else if (qid && qid == slave->queue_id) {
qid              1357 drivers/net/bonding/bond_options.c 	update_slave->queue_id = qid;
qid              1275 drivers/net/ethernet/amazon/ena/ena_com.c 	u16 qid;
qid              1279 drivers/net/ethernet/amazon/ena/ena_com.c 		qid = rss->host_rss_ind_tbl[i];
qid              1280 drivers/net/ethernet/amazon/ena/ena_com.c 		if (qid >= ENA_TOTAL_NUM_QUEUES)
qid              1283 drivers/net/ethernet/amazon/ena/ena_com.c 		io_sq = &ena_dev->io_sq_queues[qid];
qid              1417 drivers/net/ethernet/amazon/ena/ena_com.c int ena_com_get_io_handlers(struct ena_com_dev *ena_dev, u16 qid,
qid              1421 drivers/net/ethernet/amazon/ena/ena_com.c 	if (qid >= ENA_TOTAL_NUM_QUEUES) {
qid              1422 drivers/net/ethernet/amazon/ena/ena_com.c 		pr_err("Invalid queue number %d but the max is %d\n", qid,
qid              1427 drivers/net/ethernet/amazon/ena/ena_com.c 	*io_sq = &ena_dev->io_sq_queues[qid];
qid              1428 drivers/net/ethernet/amazon/ena/ena_com.c 	*io_cq = &ena_dev->io_cq_queues[qid];
qid              1825 drivers/net/ethernet/amazon/ena/ena_com.c 	if (ctx->qid >= ENA_TOTAL_NUM_QUEUES) {
qid              1827 drivers/net/ethernet/amazon/ena/ena_com.c 		       ctx->qid, ENA_TOTAL_NUM_QUEUES);
qid              1831 drivers/net/ethernet/amazon/ena/ena_com.c 	io_sq = &ena_dev->io_sq_queues[ctx->qid];
qid              1832 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq = &ena_dev->io_cq_queues[ctx->qid];
qid              1840 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq->qid = ctx->qid;
qid              1846 drivers/net/ethernet/amazon/ena/ena_com.c 	io_sq->qid = ctx->qid;
qid              1879 drivers/net/ethernet/amazon/ena/ena_com.c void ena_com_destroy_io_queue(struct ena_com_dev *ena_dev, u16 qid)
qid              1884 drivers/net/ethernet/amazon/ena/ena_com.c 	if (qid >= ENA_TOTAL_NUM_QUEUES) {
qid              1885 drivers/net/ethernet/amazon/ena/ena_com.c 		pr_err("Qid (%d) is bigger than max num of queues (%d)\n", qid,
qid              1890 drivers/net/ethernet/amazon/ena/ena_com.c 	io_sq = &ena_dev->io_sq_queues[qid];
qid              1891 drivers/net/ethernet/amazon/ena/ena_com.c 	io_cq = &ena_dev->io_cq_queues[qid];
qid               156 drivers/net/ethernet/amazon/ena/ena_com.h 	u16 qid;
qid               197 drivers/net/ethernet/amazon/ena/ena_com.h 	u16 qid;
qid               370 drivers/net/ethernet/amazon/ena/ena_com.h 	u16 qid;
qid               459 drivers/net/ethernet/amazon/ena/ena_com.h void ena_com_destroy_io_queue(struct ena_com_dev *ena_dev, u16 qid);
qid               469 drivers/net/ethernet/amazon/ena/ena_com.h int ena_com_get_io_handlers(struct ena_com_dev *ena_dev, u16 qid,
qid                93 drivers/net/ethernet/amazon/ena/ena_eth_com.c 			 io_sq->qid, io_sq->entries_in_tx_burst_left);
qid               279 drivers/net/ethernet/amazon/ena/ena_eth_com.c 			 io_cq->qid, *first_cdesc_idx, count);
qid               532 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	pr_debug("fetch rx packet: queue %d completed desc: %d\n", io_cq->qid,
qid               552 drivers/net/ethernet/amazon/ena/ena_eth_com.c 	pr_debug("[%s][QID#%d] Updating SQ head to: %d\n", __func__, io_sq->qid,
qid               168 drivers/net/ethernet/amazon/ena/ena_eth_com.h 	pr_debug("queue: %d num_descs: %d num_entries_needed: %d\n", io_sq->qid,
qid               180 drivers/net/ethernet/amazon/ena/ena_eth_com.h 		 io_sq->qid, tail);
qid               186 drivers/net/ethernet/amazon/ena/ena_eth_com.h 			 io_sq->qid, max_entries_in_tx_burst);
qid               205 drivers/net/ethernet/amazon/ena/ena_eth_com.h 				 io_cq->qid, head);
qid               151 drivers/net/ethernet/amazon/ena/ena_netdev.c 				     struct ena_ring *ring, u16 qid)
qid               153 drivers/net/ethernet/amazon/ena/ena_netdev.c 	ring->qid = qid;
qid               157 drivers/net/ethernet/amazon/ena/ena_netdev.c 	ring->napi = &adapter->ena_napi[qid].napi;
qid               208 drivers/net/ethernet/amazon/ena/ena_netdev.c static int ena_setup_tx_resources(struct ena_adapter *adapter, int qid)
qid               210 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct ena_ring *tx_ring = &adapter->tx_ring[qid];
qid               211 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct ena_irq *ena_irq = &adapter->irq_tbl[ENA_IO_IRQ_IDX(qid)];
qid               274 drivers/net/ethernet/amazon/ena/ena_netdev.c static void ena_free_tx_resources(struct ena_adapter *adapter, int qid)
qid               276 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct ena_ring *tx_ring = &adapter->tx_ring[qid];
qid               354 drivers/net/ethernet/amazon/ena/ena_netdev.c 				  u32 qid)
qid               356 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct ena_ring *rx_ring = &adapter->rx_ring[qid];
qid               357 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct ena_irq *ena_irq = &adapter->irq_tbl[ENA_IO_IRQ_IDX(qid)];
qid               411 drivers/net/ethernet/amazon/ena/ena_netdev.c 				  u32 qid)
qid               413 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct ena_ring *rx_ring = &adapter->rx_ring[qid];
qid               547 drivers/net/ethernet/amazon/ena/ena_netdev.c 				   rx_ring->qid);
qid               556 drivers/net/ethernet/amazon/ena/ena_netdev.c 				   rx_ring->qid);
qid               569 drivers/net/ethernet/amazon/ena/ena_netdev.c 			    rx_ring->qid, i, num);
qid               582 drivers/net/ethernet/amazon/ena/ena_netdev.c 			     u32 qid)
qid               584 drivers/net/ethernet/amazon/ena/ena_netdev.c 	struct ena_ring *rx_ring = &adapter->rx_ring[qid];
qid               670 drivers/net/ethernet/amazon/ena/ena_netdev.c 				      tx_ring->qid, i);
qid               675 drivers/net/ethernet/amazon/ena/ena_netdev.c 				   tx_ring->qid, i);
qid               683 drivers/net/ethernet/amazon/ena/ena_netdev.c 						  tx_ring->qid));
qid               765 drivers/net/ethernet/amazon/ena/ena_netdev.c 	txq = netdev_get_tx_queue(tx_ring->netdev, tx_ring->qid);
qid               792 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  "tx_poll: q %d skb %p completed\n", tx_ring->qid,
qid               813 drivers/net/ethernet/amazon/ena/ena_netdev.c 		  tx_ring->qid, tx_pkts);
qid              1064 drivers/net/ethernet/amazon/ena/ena_netdev.c 		  "%s qid %d\n", __func__, rx_ring->qid);
qid              1082 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  rx_ring->qid, ena_rx_ctx.descs, ena_rx_ctx.l3_proto,
qid              1105 drivers/net/ethernet/amazon/ena/ena_netdev.c 		skb_record_rx_queue(skb, rx_ring->qid);
qid              1554 drivers/net/ethernet/amazon/ena/ena_netdev.c 		napi->qid = i;
qid              1628 drivers/net/ethernet/amazon/ena/ena_netdev.c static int ena_create_io_tx_queue(struct ena_adapter *adapter, int qid)
qid              1639 drivers/net/ethernet/amazon/ena/ena_netdev.c 	tx_ring = &adapter->tx_ring[qid];
qid              1640 drivers/net/ethernet/amazon/ena/ena_netdev.c 	msix_vector = ENA_IO_IRQ_IDX(qid);
qid              1641 drivers/net/ethernet/amazon/ena/ena_netdev.c 	ena_qid = ENA_IO_TXQ_IDX(qid);
qid              1646 drivers/net/ethernet/amazon/ena/ena_netdev.c 	ctx.qid = ena_qid;
qid              1656 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  qid, rc);
qid              1666 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  qid, rc);
qid              1695 drivers/net/ethernet/amazon/ena/ena_netdev.c static int ena_create_io_rx_queue(struct ena_adapter *adapter, int qid)
qid              1706 drivers/net/ethernet/amazon/ena/ena_netdev.c 	rx_ring = &adapter->rx_ring[qid];
qid              1707 drivers/net/ethernet/amazon/ena/ena_netdev.c 	msix_vector = ENA_IO_IRQ_IDX(qid);
qid              1708 drivers/net/ethernet/amazon/ena/ena_netdev.c 	ena_qid = ENA_IO_RXQ_IDX(qid);
qid              1712 drivers/net/ethernet/amazon/ena/ena_netdev.c 	ctx.qid = ena_qid;
qid              1723 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  qid, rc);
qid              1733 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  qid, rc);
qid              2255 drivers/net/ethernet/amazon/ena/ena_netdev.c 	int qid, rc, nb_hw_desc;
qid              2259 drivers/net/ethernet/amazon/ena/ena_netdev.c 	qid = skb_get_queue_mapping(skb);
qid              2260 drivers/net/ethernet/amazon/ena/ena_netdev.c 	tx_ring = &adapter->tx_ring[qid];
qid              2261 drivers/net/ethernet/amazon/ena/ena_netdev.c 	txq = netdev_get_tx_queue(dev, qid);
qid              2293 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  qid);
qid              2341 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  __func__, qid);
qid              2391 drivers/net/ethernet/amazon/ena/ena_netdev.c 	u16 qid;
qid              2397 drivers/net/ethernet/amazon/ena/ena_netdev.c 		qid = skb_get_rx_queue(skb);
qid              2399 drivers/net/ethernet/amazon/ena/ena_netdev.c 		qid = netdev_pick_tx(dev, skb, NULL);
qid              2401 drivers/net/ethernet/amazon/ena/ena_netdev.c 	return qid;
qid              2865 drivers/net/ethernet/amazon/ena/ena_netdev.c 			  rx_ring->qid);
qid              2898 drivers/net/ethernet/amazon/ena/ena_netdev.c 				  tx_ring->qid);
qid              2910 drivers/net/ethernet/amazon/ena/ena_netdev.c 					     tx_ring->qid, i);
qid               158 drivers/net/ethernet/amazon/ena/ena_netdev.h 	u32 qid;
qid               265 drivers/net/ethernet/amazon/ena/ena_netdev.h 	u16 qid;
qid               235 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 				 struct bnx2x_virtf *vf, int qid,
qid               241 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid);
qid               245 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	q_params->q_obj = &bnx2x_vfq(vf, qid, sp_obj);
qid               268 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	bnx2x_vf_igu_ack_sb(bp, vf, vf_igu_sb(vf, bnx2x_vfq(vf, qid, sb_idx)),
qid               275 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 				  int qid)
qid               287 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	q_params.q_obj = &bnx2x_vfq(vf, qid, sp_obj);
qid               307 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	if (bnx2x_vfq(vf, qid, cxt)) {
qid               308 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0;
qid               309 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0;
qid               356 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 				   int qid, bool drv_only, int type)
qid               369 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_mac_obj);
qid               372 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, mac_obj);
qid               374 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_obj);
qid               400 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 				    struct bnx2x_virtf *vf, int qid,
qid               415 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_mac_obj);
qid               420 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_obj);
qid               424 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, mac_obj);
qid               457 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 				  int qid, bool drv_only)
qid               468 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		rc = bnx2x_vf_mac_vlan_config(bp, vf, qid,
qid               482 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			bnx2x_vf_mac_vlan_config(bp, vf, qid,
qid               494 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int bnx2x_vf_queue_setup(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid,
qid               499 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid);
qid               501 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	rc = bnx2x_vf_queue_create(bp, vf, qid, qctor);
qid               510 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	BNX2X_ERR("QSETUP[%d:%d] error: rc %d\n", vf->abs_vfid, qid, rc);
qid               515 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			       int qid)
qid               519 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid);
qid               522 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	if ((qid == LEADING_IDX) &&
qid               524 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, true,
qid               528 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, true,
qid               532 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, true,
qid               539 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	if (bnx2x_vfq(vf, qid, sp_obj).state != BNX2X_Q_STATE_RESET) {
qid               543 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		qstate.q_obj = &bnx2x_vfq(vf, qid, sp_obj);
qid               554 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	BNX2X_ERR("vf[%d:%d] error: rc %d\n", vf->abs_vfid, qid, rc);
qid               608 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c static void bnx2x_vf_prep_rx_mode(struct bnx2x *bp, u8 qid,
qid               613 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	struct bnx2x_vf_queue *vfq = vfq_get(vf, qid);
qid               634 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		    int qid, unsigned long accept_flags)
qid               640 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	bnx2x_vf_prep_rx_mode(bp, qid, &ramrod, vf, accept_flags);
qid               642 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	vfq_get(vf, qid)->accept_flags = ramrod.rx_accept_flags;
qid               646 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int bnx2x_vf_queue_teardown(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid)
qid               650 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid);
qid               653 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	if (qid == LEADING_IDX) {
qid               654 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		rc = bnx2x_vf_rxmode(bp, vf, qid, 0);
qid               660 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid,
qid               665 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid,
qid               670 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 			rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid,
qid               682 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	rc = bnx2x_vf_queue_destroy(bp, vf, qid);
qid               688 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		  vf->abs_vfid, qid, rc);
qid              1968 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c static void bnx2x_vf_qtbl_set_q(struct bnx2x *bp, u8 abs_vfid, u8 qid,
qid              1971 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	u32 reg = PXP_REG_HST_ZONE_PERMISSION_TABLE + qid * 4;
qid              2262 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	int qid, rc = 0;
qid              2273 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	for (qid = 0; qid < vf_rxq_count(vf); qid++) {
qid              2274 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		qstate.q_obj = &bnx2x_vfq(vf, qid, sp_obj);
qid              2275 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		qstate.params.update_tpa.sge_map = sge_addr[qid];
qid              2277 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		   vf->abs_vfid, qid, U64_HI(sge_addr[qid]),
qid              2278 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		   U64_LO(sge_addr[qid]));
qid              2282 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 				  U64_HI(sge_addr[qid]), U64_LO(sge_addr[qid]),
qid              2283 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 				  vf->abs_vfid, qid);
qid               453 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h 				  int qid, bool drv_only);
qid               455 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h int bnx2x_vf_queue_setup(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid,
qid               458 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h int bnx2x_vf_queue_teardown(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid);
qid               464 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h 		    int qid, unsigned long accept_flags);
qid              1947 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	int qid = mbx->msg->req.q_op.vf_qid;
qid              1951 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	   vf->abs_vfid, qid);
qid              1953 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 	rc = bnx2x_vf_queue_teardown(bp, vf, qid);
qid                27 drivers/net/ethernet/brocade/bna/bfi.h 			u8	qid;
qid                36 drivers/net/ethernet/brocade/bna/bfi.h #define bfi_mhdr_2_qid(_mh)	((_mh)->mtag.h2i.qid)
qid              1257 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	int qid = skb_get_queue_mapping(skb);
qid              1258 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	struct netdev_queue *txq = netdev_get_tx_queue(netdev, qid);
qid              1274 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		qid += nic->xdp_tx_queues;
qid              1278 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	if (qid >= MAX_SND_QUEUES_PER_QS) {
qid              1279 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		tmp = qid / MAX_SND_QUEUES_PER_QS;
qid              1288 drivers/net/ethernet/cavium/thunder/nicvf_main.c 		qid = qid % MAX_SND_QUEUES_PER_QS;
qid              1291 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	sq = &snic->qs->sq[qid];
qid              1293 drivers/net/ethernet/cavium/thunder/nicvf_main.c 	    !nicvf_sq_append_skb(snic, sq, skb, qid)) {
qid              1305 drivers/net/ethernet/cavium/thunder/nicvf_main.c 				   "Transmit ring full, stopping SQ%d\n", qid);
qid              1660 drivers/net/ethernet/chelsio/cxgb/sge.c 		     unsigned int qid, struct net_device *dev)
qid              1663 drivers/net/ethernet/chelsio/cxgb/sge.c 	struct cmdQ *q = &sge->cmdQ[qid];
qid              1697 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (sge->tx_sched && !qid && skb->dev) {
qid              1732 drivers/net/ethernet/chelsio/cxgb/sge.c 	if (qid)
qid              1423 drivers/net/ethernet/chelsio/cxgb3/sge.c 				   unsigned int qid)
qid              1430 drivers/net/ethernet/chelsio/cxgb3/sge.c 		struct sge_qset *qs = txq_to_qset(q, qid);
qid              1432 drivers/net/ethernet/chelsio/cxgb3/sge.c 		set_bit(qid, &qs->txq_stopped);
qid              1436 drivers/net/ethernet/chelsio/cxgb3/sge.c 		    test_and_clear_bit(qid, &qs->txq_stopped))
qid              1167 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h 	__u8 qid:4;
qid              1171 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h 	__u8 qid:4;
qid               336 drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h 	u32 qid;
qid               566 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 			      struct cudbg_error *cudbg_err, int qid)
qid               580 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	no_of_read_words = t4_read_cim_ibq(padap, qid,
qid               637 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c u32 cudbg_cim_obq_size(struct adapter *padap, int qid)
qid               642 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 		     QUENUMSELECT_V(qid));
qid               650 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 			      struct cudbg_error *cudbg_err, int qid)
qid               658 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	qsize =  cudbg_cim_obq_size(padap, qid);
qid               664 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c 	no_of_read_words = t4_read_cim_obq(padap, qid,
qid               169 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h u32 cudbg_cim_obq_size(struct adapter *padap, int qid);
qid               242 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qid = txq->cntxt_id;
qid               254 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qid = rxq->cntxt_id;
qid               266 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h 	entry->qid = flq->cntxt_id;
qid              1668 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 		      unsigned int qid,
qid              1708 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_read_cim_ibq(struct adapter *adap, unsigned int qid, u32 *data,
qid              1710 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_read_cim_obq(struct adapter *adap, unsigned int qid, u32 *data,
qid              1751 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid);
qid               395 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	unsigned int qid = (uintptr_t)inode->i_private & 7;
qid               396 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	struct adapter *adap = inode->i_private - qid;
qid               402 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	ret = t4_read_cim_ibq(adap, qid, (u32 *)p->data, CIM_IBQ_SIZE * 4);
qid               422 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	unsigned int qid = (uintptr_t)inode->i_private & 7;
qid               423 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	struct adapter *adap = inode->i_private - qid;
qid               429 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 	ret = t4_read_cim_obq(adap, qid, (u32 *)p->data, 6 * CIM_OBQ_SIZE * 4);
qid               575 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		unsigned int qid = EGR_QID_G(ntohl(p->opcode_qid));
qid               578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start];
qid               896 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid)
qid               898 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	qid -= p->ingr_start;
qid               899 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	return netdev2pinfo(p->ingr_map[qid]->netdev)->tx_chan;
qid              1814 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static int read_eq_indices(struct adapter *adap, u16 qid, u16 *pidx, u16 *cidx)
qid              1816 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	u32 addr = t4_read_reg(adap, SGE_DBQ_CTXT_BADDR_A) + 24 * qid + 8;
qid              1832 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c int cxgb4_sync_txq_pidx(struct net_device *dev, u16 qid, u16 pidx,
qid              1839 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 	ret = read_eq_indices(adap, qid, &hw_pidx, &hw_cidx);
qid              1858 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			     QID_V(qid) | val);
qid              1955 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 			 unsigned int qid,
qid              1962 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				 qid,
qid              2193 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		u16 qid = (dropped_db >> 15) & 0x1ffff;
qid              2199 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 		ret = t4_bar2_sge_qregs(adap, qid, T4_BAR2_QTYPE_EGRESS,
qid              2203 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c 				"qid=%d, pidx_inc=%d\n", qid, pidx_inc);
qid               184 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h void cxgb4_remove_tid(struct tid_info *t, unsigned int qid, unsigned int tid,
qid               414 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h int cxgb4_sync_txq_pidx(struct net_device *dev, u16 qid, u16 pidx, u16 size);
qid               421 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h 			 unsigned int qid,
qid               109 drivers/net/ethernet/chelsio/cxgb4/sched.c 						 const unsigned int qid,
qid               127 drivers/net/ethernet/chelsio/cxgb4/sched.c 			if (qe->cntxt_id == qid) {
qid               149 drivers/net/ethernet/chelsio/cxgb4/sched.c 	unsigned int qid;
qid               157 drivers/net/ethernet/chelsio/cxgb4/sched.c 	qid = txq->q.cntxt_id;
qid               160 drivers/net/ethernet/chelsio/cxgb4/sched.c 	e = t4_sched_queue_lookup(pi, qid, &index);
qid               191 drivers/net/ethernet/chelsio/cxgb4/sched.c 	unsigned int qid;
qid               202 drivers/net/ethernet/chelsio/cxgb4/sched.c 	qid = txq->q.cntxt_id;
qid               210 drivers/net/ethernet/chelsio/cxgb4/sched.c 	qe->cntxt_id = qid;
qid              3312 drivers/net/ethernet/chelsio/cxgb4/sge.c 			unsigned int qid = ntohl(rc->pldbuflen_qid);
qid              3314 drivers/net/ethernet/chelsio/cxgb4/sge.c 			qid -= adap->sge.ingr_start;
qid              3315 drivers/net/ethernet/chelsio/cxgb4/sge.c 			napi_schedule(&adap->sge.ingr_map[qid]->napi);
qid              3493 drivers/net/ethernet/chelsio/cxgb4/sge.c 				  unsigned int qid,
qid              3500 drivers/net/ethernet/chelsio/cxgb4/sge.c 	ret = t4_bar2_sge_qregs(adapter, qid, qtype, 0,
qid              6502 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid)
qid              6508 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 				    FW_FILTER_WR_NOREPLY_V(qid < 0));
qid              6510 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	if (qid >= 0)
qid              6512 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			cpu_to_be16(FW_FILTER_WR_RX_RPL_IQ_V(qid));
qid              9247 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		      unsigned int qid,
qid              9278 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	bar2_page_offset = ((u64)(qid >> qpp_shift) << page_shift);
qid              9279 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	bar2_qid = qid & qpp_mask;
qid              9756 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int t4_read_cim_ibq(struct adapter *adap, unsigned int qid, u32 *data, size_t n)
qid              9762 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	if (qid > 5 || (n & 3))
qid              9765 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	addr = qid * nwords;
qid              9798 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int t4_read_cim_obq(struct adapter *adap, unsigned int qid, u32 *data, size_t n)
qid              9805 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	if ((qid > (cim_num_obq - 1)) || (n & 3))
qid              9809 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		     QUENUMSELECT_V(qid));
qid               150 drivers/net/ethernet/chelsio/cxgb4/t4_hw.h 	__be32 qid;
qid               235 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h 	__be16 qid;
qid               535 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		unsigned int qid = EGR_QID_G(be32_to_cpu(p->opcode_qid));
qid               548 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		eq_idx = EQ_IDX(s, qid);
qid               551 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				"Egress Update QID %d out of range\n", qid);
qid               557 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				"Egress Update QID %d TXQ=NULL\n", qid);
qid               561 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		if (unlikely(tq->abs_id != qid)) {
qid               564 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				qid, tq->abs_id);
qid              1947 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		unsigned int qid, iq_idx;
qid              1979 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		qid = RSPD_QID_G(be32_to_cpu(rc->pldbuflen_qid));
qid              1980 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		iq_idx = IQ_IDX(s, qid);
qid              1983 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				"Ingress QID %d out of range\n", qid);
qid              1989 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				"Ingress QID %d RSPQ=NULL\n", qid);
qid              1992 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (unlikely(rspq->abs_id != qid)) {
qid              1995 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				qid, rspq->abs_id);
qid              2179 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				  unsigned int qid,
qid              2186 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	ret = t4vf_bar2_sge_qregs(adapter, qid, qtype,
qid               374 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h 			unsigned int qid,
qid               768 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 			unsigned int qid,
qid               799 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	bar2_page_offset = ((u64)(qid >> qpp_shift) << page_shift);
qid               800 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	bar2_qid = qid & qpp_mask;
qid               971 drivers/net/ethernet/emulex/benet/be.h void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm,
qid               208 drivers/net/ethernet/emulex/benet/be_main.c static void be_rxq_notify(struct be_adapter *adapter, u16 qid, u16 posted)
qid               215 drivers/net/ethernet/emulex/benet/be_main.c 	val |= qid & DB_RQ_RING_ID_MASK;
qid               237 drivers/net/ethernet/emulex/benet/be_main.c static void be_eq_notify(struct be_adapter *adapter, u16 qid,
qid               243 drivers/net/ethernet/emulex/benet/be_main.c 	val |= qid & DB_EQ_RING_ID_MASK;
qid               244 drivers/net/ethernet/emulex/benet/be_main.c 	val |= ((qid & DB_EQ_RING_ID_EXT_MASK) << DB_EQ_RING_ID_EXT_MASK_SHIFT);
qid               259 drivers/net/ethernet/emulex/benet/be_main.c void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm, u16 num_popped)
qid               263 drivers/net/ethernet/emulex/benet/be_main.c 	val |= qid & DB_CQ_RING_ID_MASK;
qid               264 drivers/net/ethernet/emulex/benet/be_main.c 	val |= ((qid & DB_CQ_RING_ID_EXT_MASK) <<
qid              2554 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	struct dpni_queue_id qid = {0};
qid              2573 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 					     &queue, &qid);
qid              2577 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 			fq->tx_fqid[j] = qid.fqid;
qid              2685 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	struct dpni_queue_id qid;
qid              2689 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 			     DPNI_QUEUE_RX, 0, fq->flowid, &queue, &qid);
qid              2695 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	fq->fqid = qid.fqid;
qid              2733 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	struct dpni_queue_id qid;
qid              2739 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 				     &queue, &qid);
qid              2744 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		fq->tx_fqid[i] = qid.fqid;
qid              2748 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	fq->tx_qdbin = qid.qdbin;
qid              2752 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 			     &queue, &qid);
qid              2758 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 	fq->fqid = qid.fqid;
qid              1428 drivers/net/ethernet/freescale/dpaa2/dpni.c 		   struct dpni_queue_id *qid)
qid              1461 drivers/net/ethernet/freescale/dpaa2/dpni.c 	qid->fqid = le32_to_cpu(rsp_params->fqid);
qid              1462 drivers/net/ethernet/freescale/dpaa2/dpni.c 	qid->qdbin = le16_to_cpu(rsp_params->qdbin);
qid               834 drivers/net/ethernet/freescale/dpaa2/dpni.h 		   struct dpni_queue_id	*qid);
qid               461 drivers/net/ethernet/freescale/fec.h 	int qid;
qid               424 drivers/net/ethernet/freescale/fec_main.c 				estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
qid               557 drivers/net/ethernet/freescale/fec_main.c 			estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
qid               637 drivers/net/ethernet/freescale/fec_main.c 			estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
qid               700 drivers/net/ethernet/freescale/fec_main.c 			estatus |= FEC_TX_BD_FTYPE(txq->bd.qid);
qid              3277 drivers/net/ethernet/freescale/fec_main.c 		rxq->bd.qid = i;
qid              3293 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.qid = i;
qid               192 drivers/net/ethernet/hisilicon/hns/hnae.h 			__u8 qid;
qid               168 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.c static void hns_ppe_set_qid(struct ppe_common_cb *ppe_common, u32 qid)
qid               175 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.c 			       PPE_CFG_QID_MODE_DEF_QID_S, qid);
qid               605 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c 	u16 i = 0, j = 0, qid = 0;
qid               608 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c 	for (qid = 0; qid < nic_dev->num_qps; qid++) {
qid               612 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c 		hinic_txq_get_stats(&nic_dev->txqs[qid], &txq_stats);
qid               621 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c 	for (qid = 0; qid < nic_dev->num_qps; qid++) {
qid               625 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c 		hinic_rxq_get_stats(&nic_dev->rxqs[qid], &rxq_stats);
qid              3114 drivers/net/ethernet/intel/i40e/i40e_main.c 	int qid = ring->queue_index;
qid              3117 drivers/net/ethernet/intel/i40e/i40e_main.c 		qid -= ring->vsi->alloc_queue_pairs;
qid              3119 drivers/net/ethernet/intel/i40e/i40e_main.c 	if (!xdp_on || !test_bit(qid, ring->vsi->af_xdp_zc_qps))
qid              3122 drivers/net/ethernet/intel/i40e/i40e_main.c 	return xdp_get_umem_from_qid(ring->vsi->netdev, qid);
qid               190 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 					    u16 qid)
qid               195 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	return (vsi && (qid < vsi->alloc_queue_pairs));
qid               824 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 				u16 qid = i40e_vc_get_pf_queue_id(vf,
qid               827 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 				reg = qid;
qid               828 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 				qid = i40e_vc_get_pf_queue_id(vf, vsi_id,
qid               830 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 				reg |= qid << 16;
qid               852 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	u16 vsi_id, qid;
qid               868 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 			qid = i40e_vc_get_pf_queue_id(vf, vsi_id, j);
qid               870 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 			reg = (qid & I40E_VPLAN_QTABLE_QINDEX_MASK);
qid                78 drivers/net/ethernet/intel/i40e/i40e_xsk.c 				u16 qid)
qid                88 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	if (qid >= vsi->num_queue_pairs)
qid                91 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	if (qid >= netdev->real_num_rx_queues ||
qid                92 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	    qid >= netdev->real_num_tx_queues)
qid               105 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	set_bit(qid, vsi->af_xdp_zc_qps);
qid               110 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		err = i40e_queue_pair_disable(vsi, qid);
qid               114 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		err = i40e_queue_pair_enable(vsi, qid);
qid               119 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		err = i40e_xsk_wakeup(vsi->netdev, qid, XDP_WAKEUP_RX);
qid               134 drivers/net/ethernet/intel/i40e/i40e_xsk.c static int i40e_xsk_umem_disable(struct i40e_vsi *vsi, u16 qid)
qid               141 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	umem = xdp_get_umem_from_qid(netdev, qid);
qid               148 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		err = i40e_queue_pair_disable(vsi, qid);
qid               153 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	clear_bit(qid, vsi->af_xdp_zc_qps);
qid               157 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		err = i40e_queue_pair_enable(vsi, qid);
qid               176 drivers/net/ethernet/intel/i40e/i40e_xsk.c 			u16 qid)
qid               178 drivers/net/ethernet/intel/i40e/i40e_xsk.c 	return umem ? i40e_xsk_umem_enable(vsi, umem, qid) :
qid               179 drivers/net/ethernet/intel/i40e/i40e_xsk.c 		i40e_xsk_umem_disable(vsi, qid);
qid                14 drivers/net/ethernet/intel/i40e/i40e_xsk.h 			u16 qid);
qid              1747 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c static bool ice_vc_isvalid_q_id(struct ice_vf *vf, u16 vsi_id, u8 qid)
qid              1751 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 	return (vsi && (qid < vsi->alloc_txq));
qid                34 drivers/net/ethernet/intel/ixgbe/ixgbe_txrx_common.h 			 u16 qid);
qid                15 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	int qid = ring->ring_idx;
qid                17 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	if (!xdp_on || !test_bit(qid, adapter->af_xdp_zc_qps))
qid                20 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	return xdp_get_umem_from_qid(adapter->netdev, qid);
qid                67 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 				 u16 qid)
qid                74 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	if (qid >= adapter->num_rx_queues)
qid                77 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	if (qid >= netdev->real_num_rx_queues ||
qid                78 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	    qid >= netdev->real_num_tx_queues)
qid                95 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		ixgbe_txrx_ring_disable(adapter, qid);
qid                97 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	set_bit(qid, adapter->af_xdp_zc_qps);
qid               100 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		ixgbe_txrx_ring_enable(adapter, qid);
qid               103 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		err = ixgbe_xsk_wakeup(adapter->netdev, qid, XDP_WAKEUP_RX);
qid               111 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c static int ixgbe_xsk_umem_disable(struct ixgbe_adapter *adapter, u16 qid)
qid               116 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	umem = xdp_get_umem_from_qid(adapter->netdev, qid);
qid               124 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		ixgbe_txrx_ring_disable(adapter, qid);
qid               126 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	clear_bit(qid, adapter->af_xdp_zc_qps);
qid               130 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		ixgbe_txrx_ring_enable(adapter, qid);
qid               136 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 			 u16 qid)
qid               138 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	return umem ? ixgbe_xsk_umem_enable(adapter, umem, qid) :
qid               139 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 		ixgbe_xsk_umem_disable(adapter, qid);
qid               698 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c int ixgbe_xsk_wakeup(struct net_device *dev, u32 qid, u32 flags)
qid               709 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	if (qid >= adapter->num_xdp_queues)
qid               712 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	ring = adapter->xdp_ring[qid];
qid                44 drivers/net/ethernet/mellanox/mlx5/core/en/params.h 						u16 qid,
qid                49 drivers/net/ethernet/mellanox/mlx5/core/en/params.h 	int ch = qid - nch * group;
qid                59 drivers/net/ethernet/mellanox/mlx5/core/en/params.h 					      u16 qid,
qid                65 drivers/net/ethernet/mellanox/mlx5/core/en/params.h 	*ix = qid % nch;
qid                66 drivers/net/ethernet/mellanox/mlx5/core/en/params.h 	*group = qid / nch;
qid                70 drivers/net/ethernet/mellanox/mlx5/core/en/params.h 				      struct mlx5e_params *params, u64 qid)
qid                72 drivers/net/ethernet/mellanox/mlx5/core/en/params.h 	return qid < params->num_channels * profile->rq_groups;
qid                10 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c int mlx5e_xsk_wakeup(struct net_device *dev, u32 qid, u32 flags)
qid                20 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c 	if (unlikely(!mlx5e_qid_get_ch_if_in_group(params, qid, MLX5E_RQ_GROUP_XSK, &ix)))
qid                12 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.h int mlx5e_xsk_wakeup(struct net_device *dev, u32 qid, u32 flags);
qid               231 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.c int mlx5e_xsk_setup_umem(struct net_device *dev, struct xdp_umem *umem, u16 qid)
qid               237 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.c 	if (unlikely(!mlx5e_qid_get_ch_if_in_group(params, qid, MLX5E_RQ_GROUP_XSK, &ix)))
qid                25 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.h int mlx5e_xsk_setup_umem(struct net_device *dev, struct xdp_umem *umem, u16 qid);
qid                54 drivers/net/ethernet/netronome/nfp/abm/ctrl.c 	unsigned int qid;
qid                58 drivers/net/ethernet/netronome/nfp/abm/ctrl.c 	qid = band * NFP_NET_MAX_RX_RINGS + alink->queue_base + queue;
qid                60 drivers/net/ethernet/netronome/nfp/abm/ctrl.c 	sym_offset = qid * stride + offset;
qid               134 drivers/net/ethernet/netronome/nfp/abm/ctrl.c 	unsigned int qid;
qid               136 drivers/net/ethernet/netronome/nfp/abm/ctrl.c 	qid = band * NFP_NET_MAX_RX_RINGS + alink->queue_base + queue;
qid               138 drivers/net/ethernet/netronome/nfp/abm/ctrl.c 	return __nfp_abm_ctrl_set_q_act(alink->abm, qid, act);
qid              1535 drivers/net/ethernet/pensando/ionic/ionic_if.h 	__le32 qid;
qid              2068 drivers/net/ethernet/qlogic/qed/qed_l2.c 		if (p_params->qid != QED_RFS_NTUPLE_QID_RSS) {
qid              2069 drivers/net/ethernet/qlogic/qed/qed_l2.c 			rc = qed_fw_l2_queue(p_hwfn, p_params->qid,
qid              2849 drivers/net/ethernet/qlogic/qed/qed_l2.c 		params->qid = QED_RFS_NTUPLE_QID_RSS;
qid              1489 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	u8 qid;
qid              1543 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	qid = p_hwfn->hw_info.resc_start[QED_LL2_QUEUE] + connection_handle;
qid              1544 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_conn->queue_id = qid;
qid              1545 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	p_ll2_conn->tx_stats_id = qid;
qid              1548 drivers/net/ethernet/qlogic/qed/qed_ll2.c 					    TSTORM_LL2_RX_PRODS_OFFSET(qid);
qid              2102 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	u8 qid = p_ll2_conn->queue_id;
qid              2107 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		      CORE_LL2_TSTORM_PER_QUEUE_STAT_OFFSET(qid);
qid              2121 drivers/net/ethernet/qlogic/qed/qed_ll2.c 	u8 qid = p_ll2_conn->queue_id;
qid              2126 drivers/net/ethernet/qlogic/qed/qed_ll2.c 		      CORE_LL2_USTORM_PER_QUEUE_STAT_OFFSET(qid);
qid               223 drivers/net/ethernet/qlogic/qed/qed_sriov.c 					u16 qid,
qid               235 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		p_qcid = &p_vf->vf_queues[qid].cids[i];
qid               865 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	int qid;
qid               867 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	for (qid = 0; qid < vf->num_rxqs; qid++) {
qid               868 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		qed_fw_l2_queue(p_hwfn, vf->vf_queues[qid].fw_rx_qid,
qid               896 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	int qid = 0;
qid               907 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	for (qid = 0; qid < num_rx_queues; qid++) {
qid               909 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		vf->igu_sbs[qid] = p_block->igu_sb_id;
qid               911 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		SET_FIELD(val, IGU_MAPPING_LINE_VECTOR_NUMBER, qid);
qid              1004 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	u16 qid, num_irqs;
qid              1027 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		qid = p_params->req_rx_queue[i];
qid              1028 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		if (qid < min_vf_qzone || qid > max_vf_qzone) {
qid              1031 drivers/net/ethernet/qlogic/qed/qed_sriov.c 				  qid,
qid              1037 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		qid = p_params->req_tx_queue[i];
qid              1038 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		if (qid > max_vf_qzone) {
qid              1041 drivers/net/ethernet/qlogic/qed/qed_sriov.c 				  qid, p_params->rel_vf_id, max_vf_qzone);
qid              1046 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		if (qid < min_vf_qzone)
qid              1050 drivers/net/ethernet/qlogic/qed/qed_sriov.c 				   p_params->rel_vf_id, qid, i);
qid              2094 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	if (p_qid_tlv->qid >= MAX_QUEUES_PER_QZONE) {
qid              2097 drivers/net/ethernet/qlogic/qed/qed_sriov.c 			   p_vf->relative_vf_id, p_qid_tlv->qid);
qid              2101 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	return p_qid_tlv->qid;
qid              2709 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		u16 qid = req->rx_qid + i;
qid              2711 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		handlers[i] = vf->vf_queues[qid].cids[qid_usage_idx].p_cid;
qid              3434 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	u16 coal = 0, qid, i;
qid              3441 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	qid = req->qid;
qid              3445 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		if (!qed_iov_validate_rxq(p_hwfn, p_vf, qid,
qid              3449 drivers/net/ethernet/qlogic/qed/qed_sriov.c 				   p_vf->abs_vf_id, qid);
qid              3453 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		p_cid = qed_iov_get_vf_rx_queue_cid(&p_vf->vf_queues[qid]);
qid              3458 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		if (!qed_iov_validate_txq(p_hwfn, p_vf, qid,
qid              3462 drivers/net/ethernet/qlogic/qed/qed_sriov.c 				   p_vf->abs_vf_id, qid);
qid              3466 drivers/net/ethernet/qlogic/qed/qed_sriov.c 			p_queue = &p_vf->vf_queues[qid];
qid              3503 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	u16 qid;
qid              3509 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	qid = req->qid;
qid              3511 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	if (!qed_iov_validate_rxq(p_hwfn, vf, qid,
qid              3515 drivers/net/ethernet/qlogic/qed/qed_sriov.c 			   vf->abs_vf_id, qid);
qid              3519 drivers/net/ethernet/qlogic/qed/qed_sriov.c 	if (!qed_iov_validate_txq(p_hwfn, vf, qid,
qid              3523 drivers/net/ethernet/qlogic/qed/qed_sriov.c 			   vf->abs_vf_id, qid);
qid              3530 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		   vf->abs_vf_id, rx_coal, tx_coal, qid);
qid              3533 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		p_cid = qed_iov_get_vf_rx_queue_cid(&vf->vf_queues[qid]);
qid              3540 drivers/net/ethernet/qlogic/qed/qed_sriov.c 				   vf->abs_vf_id, vf->vf_queues[qid].fw_rx_qid);
qid              3547 drivers/net/ethernet/qlogic/qed/qed_sriov.c 		struct qed_vf_queue *p_queue = &vf->vf_queues[qid];
qid               169 drivers/net/ethernet/qlogic/qed/qed_vf.c 	p_qid_tlv->qid = p_cid->qid_usage_idx;
qid               854 drivers/net/ethernet/qlogic/qed/qed_vf.c 	u16 qid = p_cid->rel.queue_id;
qid               860 drivers/net/ethernet/qlogic/qed/qed_vf.c 	req->tx_qid = qid;
qid               890 drivers/net/ethernet/qlogic/qed/qed_vf.c 		u8 cid = p_iov->acquire_resp.resc.cid[qid];
qid               899 drivers/net/ethernet/qlogic/qed/qed_vf.c 		   qid, p_cid->qid_usage_idx, *pp_doorbell, resp->offset);
qid              1365 drivers/net/ethernet/qlogic/qed/qed_vf.c 	req->qid = p_cid->rel.queue_id;
qid              1429 drivers/net/ethernet/qlogic/qed/qed_vf.c 	req->qid = p_cid->rel.queue_id;
qid              1434 drivers/net/ethernet/qlogic/qed/qed_vf.c 		   rx_coal, tx_coal, req->qid);
qid               260 drivers/net/ethernet/qlogic/qed/qed_vf.h 	u8 qid;
qid               509 drivers/net/ethernet/qlogic/qed/qed_vf.h 	u16 qid;
qid               515 drivers/net/ethernet/qlogic/qed/qed_vf.h 	u16 qid;
qid               127 drivers/net/ethernet/qlogic/qede/qede_filter.c 	params.qid = rxq_id;
qid              1094 drivers/net/ethernet/sfc/falcon/farch.c 	int qid;
qid              1096 drivers/net/ethernet/sfc/falcon/farch.c 	qid = EF4_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_SUBDATA);
qid              1097 drivers/net/ethernet/sfc/falcon/farch.c 	if (qid < EF4_TXQ_TYPES * efx->n_tx_channels) {
qid              1098 drivers/net/ethernet/sfc/falcon/farch.c 		tx_queue = ef4_get_tx_queue(efx, qid / EF4_TXQ_TYPES,
qid              1099 drivers/net/ethernet/sfc/falcon/farch.c 					    qid % EF4_TXQ_TYPES);
qid              1116 drivers/net/ethernet/sfc/falcon/farch.c 	int qid;
qid              1119 drivers/net/ethernet/sfc/falcon/farch.c 	qid = EF4_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_RX_DESCQ_ID);
qid              1121 drivers/net/ethernet/sfc/falcon/farch.c 	if (qid >= efx->n_channels)
qid              1123 drivers/net/ethernet/sfc/falcon/farch.c 	channel = ef4_get_channel(efx, qid);
qid              1130 drivers/net/ethernet/sfc/falcon/farch.c 			   "RXQ %d flush retry\n", qid);
qid              1082 drivers/net/ethernet/sfc/farch.c 	int qid;
qid              1084 drivers/net/ethernet/sfc/farch.c 	qid = EFX_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_SUBDATA);
qid              1085 drivers/net/ethernet/sfc/farch.c 	if (qid < EFX_TXQ_TYPES * (efx->n_tx_channels + efx->n_extra_tx_channels)) {
qid              1086 drivers/net/ethernet/sfc/farch.c 		tx_queue = efx_get_tx_queue(efx, qid / EFX_TXQ_TYPES,
qid              1087 drivers/net/ethernet/sfc/farch.c 					    qid % EFX_TXQ_TYPES);
qid              1104 drivers/net/ethernet/sfc/farch.c 	int qid;
qid              1107 drivers/net/ethernet/sfc/farch.c 	qid = EFX_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_RX_DESCQ_ID);
qid              1109 drivers/net/ethernet/sfc/farch.c 	if (qid >= efx->n_channels)
qid              1111 drivers/net/ethernet/sfc/farch.c 	channel = efx_get_channel(efx, qid);
qid              1118 drivers/net/ethernet/sfc/farch.c 			   "RXQ %d flush retry\n", qid);
qid              1405 drivers/net/ethernet/sfc/siena_sriov.c 	unsigned qid, seq, type, data;
qid              1407 drivers/net/ethernet/sfc/siena_sriov.c 	qid = EFX_QWORD_FIELD(*event, FSF_CZ_USER_QID);
qid              1417 drivers/net/ethernet/sfc/siena_sriov.c 		   qid, seq, type, data);
qid              1419 drivers/net/ethernet/sfc/siena_sriov.c 	if (map_vi_index(efx, qid, &vf, NULL))
qid              1493 drivers/net/ethernet/sfc/siena_sriov.c 	unsigned queue, qid;
qid              1496 drivers/net/ethernet/sfc/siena_sriov.c 	if (map_vi_index(efx, queue, &vf, &qid))
qid              1499 drivers/net/ethernet/sfc/siena_sriov.c 	if (!test_bit(qid, vf->txq_mask))
qid              1502 drivers/net/ethernet/sfc/siena_sriov.c 	__clear_bit(qid, vf->txq_mask);
qid              1512 drivers/net/ethernet/sfc/siena_sriov.c 	unsigned ev_failed, queue, qid;
qid              1517 drivers/net/ethernet/sfc/siena_sriov.c 	if (map_vi_index(efx, queue, &vf, &qid))
qid              1519 drivers/net/ethernet/sfc/siena_sriov.c 	if (!test_bit(qid, vf->rxq_mask))
qid              1523 drivers/net/ethernet/sfc/siena_sriov.c 		set_bit(qid, vf->rxq_retry_mask);
qid              1526 drivers/net/ethernet/sfc/siena_sriov.c 		__clear_bit(qid, vf->rxq_mask);
qid               139 drivers/net/hyperv/netvsc_trace.h 		__field(  u16,	qid	      )
qid               146 drivers/net/hyperv/netvsc_trace.h 		__entry->qid = chan->offermsg.offer.sub_channel_index;
qid               152 drivers/net/hyperv/netvsc_trace.h 		  __get_str(name), __entry->qid,
qid               164 drivers/net/hyperv/netvsc_trace.h 		__field(  u16,	qid	    )
qid               169 drivers/net/hyperv/netvsc_trace.h 		__entry->qid = chan->offermsg.offer.sub_channel_index;
qid               173 drivers/net/hyperv/netvsc_trace.h 		  __get_str(name), __entry->qid,
qid              1137 drivers/net/vmxnet3/vmxnet3_drv.c 				       VMXNET3_REG_TXPROD + tq->qid * 8,
qid              1316 drivers/net/vmxnet3/vmxnet3_drv.c 		BUG_ON(rcd->rqID != rq->qid && rcd->rqID != rq->qid2 &&
qid              1338 drivers/net/vmxnet3/vmxnet3_drv.c 			       (rcd->rqID != rq->qid &&
qid              1566 drivers/net/vmxnet3/vmxnet3_drv.c 					       rxprod_reg[ring_idx] + rq->qid * 8,
qid              2158 drivers/net/vmxnet3/vmxnet3_drv.c 			rq->qid = i;
qid              2835 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->qid = i;
qid               244 drivers/net/vmxnet3/vmxnet3_int.h 	int				qid;
qid               290 drivers/net/vmxnet3/vmxnet3_int.h 	u32 qid;            /* rqID in RCD for buffer from 1st ring */
qid               444 drivers/net/wireless/ath/ar5523/ar5523.c 	__be32 qid = cpu_to_be32(0);
qid               448 drivers/net/wireless/ath/ar5523/ar5523.c 				 &qid, sizeof(qid), 0);
qid               476 drivers/net/wireless/ath/ar5523/ar5523.c 	qinfo.qid	     = cpu_to_be32(0);
qid               284 drivers/net/wireless/ath/ar5523/ar5523_hw.h 	__be32	qid;
qid               380 drivers/net/wireless/ath/ath9k/ar9003_mac.c 	ts->qid = MS(ads->ds_info, AR_TxQcuNum);
qid               113 drivers/net/wireless/ath/ath9k/mac.h 	u8 qid;
qid              2667 drivers/net/wireless/ath/ath9k/xmit.c 		if (ts.qid == sc->beacon.beaconq) {
qid              2680 drivers/net/wireless/ath/ath9k/xmit.c 		txq = &sc->tx.txq[ts.qid];
qid               870 drivers/net/wireless/ath/wil6210/wmi.c 		    data->info.qid, data->info.mid, data->info.cid);
qid              1438 drivers/net/wireless/ath/wil6210/wmi.c 		    d_len, data->info.qid, data->info.mid, data->info.cid);
qid              2442 drivers/net/wireless/ath/wil6210/wmi.h 	u8 qid;
qid              1186 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	int ret, qid;
qid              1196 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	qid = le16_to_cpu(rsp->queue_number);
qid              1199 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (qid >= ARRAY_SIZE(trans_pcie->txq)) {
qid              1200 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		WARN_ONCE(1, "queue index %d unsupported", qid);
qid              1205 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	if (test_and_set_bit(qid, trans_pcie->queue_used)) {
qid              1206 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 		WARN_ONCE(1, "queue %d already used", qid);
qid              1211 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	txq->id = qid;
qid              1212 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	trans_pcie->txq[qid] = txq;
qid              1219 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	IWL_DEBUG_TX_QUEUES(trans, "Activate queue %d\n", qid);
qid              1222 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c 	return qid;
qid               139 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_cleanup(struct mt76_dev *dev, enum mt76_txq_id qid, bool flush)
qid               141 drivers/net/wireless/mediatek/mt76/dma.c 	struct mt76_sw_queue *sq = &dev->q_tx[qid];
qid               160 drivers/net/wireless/mediatek/mt76/dma.c 			n_swq_queued[entry.qid]++;
qid               166 drivers/net/wireless/mediatek/mt76/dma.c 			dev->drv->tx_complete_skb(dev, qid, &entry);
qid               192 drivers/net/wireless/mediatek/mt76/dma.c 	       qid < IEEE80211_NUM_ACS && q->queued < q->ndesc - 8;
qid               202 drivers/net/wireless/mediatek/mt76/dma.c 		ieee80211_wake_queue(dev->hw, qid);
qid               257 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_queue_skb_raw(struct mt76_dev *dev, enum mt76_txq_id qid,
qid               260 drivers/net/wireless/mediatek/mt76/dma.c 	struct mt76_queue *q = dev->q_tx[qid].q;
qid               281 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
qid               285 drivers/net/wireless/mediatek/mt76/dma.c 	struct mt76_queue *q = dev->q_tx[qid].q;
qid               333 drivers/net/wireless/mediatek/mt76/dma.c 	ret = dev->drv->tx_prepare_skb(dev, txwi, qid, wcid, sta, &tx_info);
qid               355 drivers/net/wireless/mediatek/mt76/dma.c 	dev->drv->tx_complete_skb(dev, qid, &e);
qid               425 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_rx_reset(struct mt76_dev *dev, enum mt76_rxq_id qid)
qid               427 drivers/net/wireless/mediatek/mt76/dma.c 	struct mt76_queue *q = &dev->q_rx[qid];
qid               530 drivers/net/wireless/mediatek/mt76/dma.c 	int qid, done = 0, cur;
qid               533 drivers/net/wireless/mediatek/mt76/dma.c 	qid = napi - dev->napi;
qid               538 drivers/net/wireless/mediatek/mt76/dma.c 		cur = mt76_dma_rx_process(dev, &dev->q_rx[qid], budget - done);
qid               539 drivers/net/wireless/mediatek/mt76/dma.c 		mt76_rx_poll_complete(dev, qid, napi);
qid               547 drivers/net/wireless/mediatek/mt76/dma.c 		dev->drv->rx_poll_complete(dev, qid);
qid                95 drivers/net/wireless/mediatek/mt76/mt76.h 	enum mt76_txq_id qid;
qid               159 drivers/net/wireless/mediatek/mt76/mt76.h 	int (*tx_queue_skb)(struct mt76_dev *dev, enum mt76_txq_id qid,
qid               163 drivers/net/wireless/mediatek/mt76/mt76.h 	int (*tx_queue_skb_raw)(struct mt76_dev *dev, enum mt76_txq_id qid,
qid               169 drivers/net/wireless/mediatek/mt76/mt76.h 	void (*rx_reset)(struct mt76_dev *dev, enum mt76_rxq_id qid);
qid               171 drivers/net/wireless/mediatek/mt76/mt76.h 	void (*tx_cleanup)(struct mt76_dev *dev, enum mt76_txq_id qid,
qid               295 drivers/net/wireless/mediatek/mt76/mt76.h 			      enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid               299 drivers/net/wireless/mediatek/mt76/mt76.h 	void (*tx_complete_skb)(struct mt76_dev *dev, enum mt76_txq_id qid,
qid               711 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_txq_schedule(struct mt76_dev *dev, enum mt76_txq_id qid);
qid               792 drivers/net/wireless/mediatek/mt76/mt76.h static inline u8 q2ep(u8 qid)
qid               795 drivers/net/wireless/mediatek/mt76/mt76.h 	return qid + 1;
qid               779 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		      struct sk_buff *skb, enum mt76_txq_id qid,
qid               788 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	struct mt76_queue *q = dev->mt76.q_tx[qid].q;
qid               803 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		if (vif_idx && qid >= MT_TXQ_BEACON)
qid               877 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	if (qid >= MT_TXQ_BEACON)
qid               908 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 			  enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid               940 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	mt7603_mac_write_txwi(dev, txwi_ptr, tx_info->skb, qid, wcid,
qid              1153 drivers/net/wireless/mediatek/mt76/mt7603/mac.c void mt7603_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid              1164 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	if (qid < 4)
qid               228 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h 			  enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid               232 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h void mt7603_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid               222 drivers/net/wireless/mediatek/mt76/mt7615/mac.c void mt7615_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid               763 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 			  enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid                56 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	enum mt76_txq_id qid;
qid               108 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 		qid = MT_TXQ_MCU;
qid               110 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 		qid = MT_TXQ_FWDL;
qid               112 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c 	return mt76_tx_queue_skb_raw(dev, qid, skb, 0);
qid               254 drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h 			  enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid               258 drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h void mt7615_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid               177 drivers/net/wireless/mediatek/mt76/mt76x02.h 			   enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid               826 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c void mt76x02_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid               190 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h void mt76x02_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid               139 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c 			   enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid               149 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c 	if (qid == MT_TXQ_PSD && wcid && wcid->idx < 128)
qid                18 drivers/net/wireless/mediatek/mt76/mt76x02_usb.h 			    enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid                21 drivers/net/wireless/mediatek/mt76/mt76x02_usb.h void mt76x02u_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid                18 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c void mt76x02u_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid,
qid                64 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c 			    enum mt76_txq_id qid, struct mt76_wcid *wcid,
qid                69 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c 	int pid, len = tx_info->skb->len, ep = q2ep(mdev->q_tx[qid].q->hw_idx);
qid               489 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	u8 cw_min = 5, cw_max = 10, qid;
qid               492 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	qid = dev->mt76.q_tx[queue].q->hw_idx;
qid               503 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	mt76_wr(dev, MT_EDCA_CFG_AC(qid), val);
qid               505 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val = mt76_rr(dev, MT_WMM_TXOP(qid));
qid               506 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val &= ~(MT_WMM_TXOP_MASK << MT_WMM_TXOP_SHIFT(qid));
qid               507 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val |= params->txop << MT_WMM_TXOP_SHIFT(qid);
qid               508 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	mt76_wr(dev, MT_WMM_TXOP(qid), val);
qid               511 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val &= ~(MT_WMM_AIFSN_MASK << MT_WMM_AIFSN_SHIFT(qid));
qid               512 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val |= params->aifs << MT_WMM_AIFSN_SHIFT(qid);
qid               516 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val &= ~(MT_WMM_CWMIN_MASK << MT_WMM_CWMIN_SHIFT(qid));
qid               517 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val |= cw_min << MT_WMM_CWMIN_SHIFT(qid);
qid               521 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val &= ~(MT_WMM_CWMAX_MASK << MT_WMM_CWMAX_SHIFT(qid));
qid               522 drivers/net/wireless/mediatek/mt76/mt76x02_util.c 	val |= cw_max << MT_WMM_CWMAX_SHIFT(qid);
qid               254 drivers/net/wireless/mediatek/mt76/tx.c 	int qid = skb_get_queue_mapping(skb);
qid               256 drivers/net/wireless/mediatek/mt76/tx.c 	if (WARN_ON(qid >= MT_TXQ_PSD)) {
qid               257 drivers/net/wireless/mediatek/mt76/tx.c 		qid = MT_TXQ_BE;
qid               258 drivers/net/wireless/mediatek/mt76/tx.c 		skb_set_queue_mapping(skb, qid);
qid               278 drivers/net/wireless/mediatek/mt76/tx.c 	q = dev->q_tx[qid].q;
qid               281 drivers/net/wireless/mediatek/mt76/tx.c 	dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, sta);
qid               384 drivers/net/wireless/mediatek/mt76/tx.c 	enum mt76_txq_id qid = mt76_txq_get_qid(txq);
qid               419 drivers/net/wireless/mediatek/mt76/tx.c 	idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, txq->sta);
qid               453 drivers/net/wireless/mediatek/mt76/tx.c 		idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid,
qid               462 drivers/net/wireless/mediatek/mt76/tx.c 		hwq->entry[idx].qid = sq - dev->q_tx;
qid               473 drivers/net/wireless/mediatek/mt76/tx.c mt76_txq_schedule_list(struct mt76_dev *dev, enum mt76_txq_id qid)
qid               475 drivers/net/wireless/mediatek/mt76/tx.c 	struct mt76_sw_queue *sq = &dev->q_tx[qid];
qid               494 drivers/net/wireless/mediatek/mt76/tx.c 		txq = ieee80211_next_txq(dev->hw, qid);
qid               526 drivers/net/wireless/mediatek/mt76/tx.c void mt76_txq_schedule(struct mt76_dev *dev, enum mt76_txq_id qid)
qid               528 drivers/net/wireless/mediatek/mt76/tx.c 	struct mt76_sw_queue *sq = &dev->q_tx[qid];
qid               531 drivers/net/wireless/mediatek/mt76/tx.c 	if (qid >= 4)
qid               540 drivers/net/wireless/mediatek/mt76/tx.c 		ieee80211_txq_schedule_start(dev->hw, qid);
qid               541 drivers/net/wireless/mediatek/mt76/tx.c 		len = mt76_txq_schedule_list(dev, qid);
qid               542 drivers/net/wireless/mediatek/mt76/tx.c 		ieee80211_txq_schedule_end(dev->hw, qid);
qid               768 drivers/net/wireless/mediatek/mt76/usb.c mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid,
qid               772 drivers/net/wireless/mediatek/mt76/usb.c 	struct mt76_queue *q = dev->q_tx[qid].q;
qid               783 drivers/net/wireless/mediatek/mt76/usb.c 	err = dev->drv->tx_prepare_skb(dev, NULL, qid, wcid, sta, &tx_info);
qid               341 drivers/net/wireless/mediatek/mt7601u/dma.c static u8 q2ep(u8 qid)
qid               344 drivers/net/wireless/mediatek/mt7601u/dma.c 	return qid + 1;
qid                29 drivers/net/wireless/mediatek/mt7601u/tx.c 	int qid = skb_get_queue_mapping(skb);
qid                31 drivers/net/wireless/mediatek/mt7601u/tx.c 	if (WARN_ON(qid >= MT_TXQ_PSD)) {
qid                32 drivers/net/wireless/mediatek/mt7601u/tx.c 		qid = MT_TXQ_BE;
qid                33 drivers/net/wireless/mediatek/mt7601u/tx.c 		skb_set_queue_mapping(skb, qid);
qid                36 drivers/net/wireless/mediatek/mt7601u/tx.c 	return q2hwq(qid);
qid               632 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	switch (queue->qid) {
qid               655 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	switch (queue->qid) {
qid               681 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	switch (queue->qid) {
qid               719 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	if (entry->queue->qid == QID_RX) {
qid               737 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	if (entry->queue->qid == QID_RX) {
qid              1758 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	switch (queue->qid) {
qid               721 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	switch (queue->qid) {
qid               744 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	switch (queue->qid) {
qid               770 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	switch (queue->qid) {
qid               808 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	if (entry->queue->qid == QID_RX) {
qid               826 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	if (entry->queue->qid == QID_RX) {
qid              2057 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	switch (queue->qid) {
qid               722 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	switch (queue->qid) {
qid               745 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	switch (queue->qid) {
qid              1848 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	switch (queue->qid) {
qid               816 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	rt2x00_set_field32(&word, TXWI_W1_PACKETID_QUEUE, entry->queue->qid);
qid               989 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			   entry->queue->qid, entry->entry_idx);
qid              1103 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 	u8 qid;
qid              1111 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		qid = rt2x00_get_field32(reg, TX_STA_FIFO_PID_QUEUE);
qid              1112 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		queue = rt2x00queue_get_tx_queue(rt2x00dev, qid);
qid              1116 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 				   qid);
qid              1125 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 				    entry->entry_idx, qid);
qid              1153 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			   entry->entry_idx, entry->queue->qid);
qid              1241 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		switch (queue->qid) {
qid                31 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	int idx, qid;
qid                33 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	switch (queue->qid) {
qid                38 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		qid = queue->qid;
qid                39 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		idx = rt2x00mmio_register_read(rt2x00dev, TX_DTX_IDX(qid));
qid               425 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	switch (queue->qid) {
qid               456 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	switch (queue->qid) {
qid               463 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 		rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(queue->qid),
qid               485 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	switch (queue->qid) {
qid               528 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	switch (queue->qid) {
qid               567 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	switch (queue->qid) {
qid               612 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	if (entry->queue->qid == QID_RX) {
qid               631 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	if (entry->queue->qid == QID_RX) {
qid                51 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	switch (queue->qid) {
qid                74 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	switch (queue->qid) {
qid               395 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	if (entry->queue->qid == QID_BEACON)
qid               719 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	switch (queue->qid) {
qid               171 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c 	dump_hdr->queue_index = entry->queue->qid;
qid               324 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c 				queue->qid, (unsigned int)queue->flags,
qid                95 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 	enum data_queue_qid qid = skb_get_queue_mapping(skb);
qid               112 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 		qid = QID_ATIM;
qid               114 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 	queue = rt2x00queue_get_tx_queue(rt2x00dev, qid);
qid               118 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 			   "Please file bug report to %s\n", qid, DRV_PROJECT);
qid               502 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 			   entry->queue->qid, DRV_PROJECT);
qid               667 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 			   queue->qid);
qid               679 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 			   queue->qid, DRV_PROJECT);
qid               894 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	switch (queue->qid) {
qid               903 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		ieee80211_stop_queue(queue->rt2x00dev->hw, queue->qid);
qid               927 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	switch (queue->qid) {
qid               936 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		ieee80211_wake_queue(queue->rt2x00dev->hw, queue->qid);
qid               990 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		(queue->qid == QID_AC_VO) ||
qid               991 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		(queue->qid == QID_AC_VI) ||
qid               992 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		(queue->qid == QID_AC_BE) ||
qid               993 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		(queue->qid == QID_AC_BK);
qid              1020 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 			    queue->qid);
qid              1214 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 			     struct data_queue *queue, enum data_queue_qid qid)
qid              1221 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	queue->qid = qid;
qid              1235 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	enum data_queue_qid qid;
qid              1271 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	qid = QID_AC_VO;
qid              1273 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		rt2x00queue_init(rt2x00dev, queue, qid++);
qid               455 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h 	enum data_queue_qid qid;
qid               426 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	switch (queue->qid) {
qid               466 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if ((entry->queue->qid == QID_BEACON) &&
qid               485 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	switch (queue->qid) {
qid               526 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 		    queue->qid);
qid               571 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (entry->queue->qid == QID_RX)
qid               584 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (queue->qid == QID_RX) {
qid               665 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (queue->qid != QID_BEACON ||
qid               700 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (queue->qid != QID_BEACON ||
qid              1046 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	switch (queue->qid) {
qid              1069 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	switch (queue->qid) {
qid              1100 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	switch (queue->qid) {
qid              1292 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	if (entry->queue->qid == QID_RX) {
qid              1310 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	if (entry->queue->qid == QID_RX) {
qid              1787 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid);
qid              1812 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00_set_field32(&word, TXD_W5_PID_TYPE, entry->queue->qid);
qid              1819 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	if (entry->queue->qid != QID_BEACON) {
qid              1865 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	skbdesc->desc_len = (entry->queue->qid == QID_BEACON) ? TXINFO_SIZE :
qid              2930 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	switch (queue->qid) {
qid              1018 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	switch (queue->qid) {
qid              1041 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	switch (queue->qid) {
qid              1481 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid);
qid              2350 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	switch (queue->qid) {
qid               304 drivers/net/wireless/ti/wl1251/acx.h 	u8 qid;
qid               217 drivers/net/wireless/ti/wl1251/init.c static int wl1251_hw_init_txq_fill(u8 qid,
qid               221 drivers/net/wireless/ti/wl1251/init.c 	config->qid = qid;
qid               223 drivers/net/wireless/ti/wl1251/init.c 	switch (qid) {
qid               249 drivers/net/wireless/ti/wl1251/init.c 		wl1251_error("Invalid TX queue id: %d", qid);
qid               480 drivers/nvme/host/core.c 		struct nvme_command *cmd, blk_mq_req_flags_t flags, int qid)
qid               485 drivers/nvme/host/core.c 	if (qid == NVME_QID_ANY) {
qid               489 drivers/nvme/host/core.c 				qid ? qid - 1 : 0);
qid               820 drivers/nvme/host/core.c 		unsigned timeout, int qid, int at_head,
qid               826 drivers/nvme/host/core.c 	req = nvme_alloc_request(q, cmd, flags, qid);
qid               309 drivers/nvme/host/fabrics.c 			case (offsetof(struct nvmf_connect_command, qid)):
qid               312 drivers/nvme/host/fabrics.c 					inv_sqe, cmd->connect.qid);
qid               377 drivers/nvme/host/fabrics.c 	cmd.connect.qid = 0;
qid               437 drivers/nvme/host/fabrics.c int nvmf_connect_io_queue(struct nvme_ctrl *ctrl, u16 qid, bool poll)
qid               447 drivers/nvme/host/fabrics.c 	cmd.connect.qid = cpu_to_le16(qid);
qid               463 drivers/nvme/host/fabrics.c 			data, sizeof(*data), 0, qid, 1,
qid               169 drivers/nvme/host/fabrics.h int nvmf_connect_io_queue(struct nvme_ctrl *ctrl, u16 qid, bool poll);
qid              1327 drivers/nvme/host/fc.c 	conn_rqst->connect_cmd.qid  = cpu_to_be16(queue->qnum);
qid               484 drivers/nvme/host/nvme.h 		struct nvme_command *cmd, blk_mq_req_flags_t flags, int qid);
qid               492 drivers/nvme/host/nvme.h 		unsigned timeout, int qid, int at_head,
qid               144 drivers/nvme/host/pci.c static inline unsigned int sq_idx(unsigned int qid, u32 stride)
qid               146 drivers/nvme/host/pci.c 	return qid * 2 * stride;
qid               149 drivers/nvme/host/pci.c static inline unsigned int cq_idx(unsigned int qid, u32 stride)
qid               151 drivers/nvme/host/pci.c 	return (qid * 2 + 1) * stride;
qid               179 drivers/nvme/host/pci.c 	u16 qid;
qid               271 drivers/nvme/host/pci.c 			    struct nvme_queue *nvmeq, int qid)
qid               273 drivers/nvme/host/pci.c 	if (!dev->dbbuf_dbs || !qid)
qid               276 drivers/nvme/host/pci.c 	nvmeq->dbbuf_sq_db = &dev->dbbuf_dbs[sq_idx(qid, dev->db_stride)];
qid               277 drivers/nvme/host/pci.c 	nvmeq->dbbuf_cq_db = &dev->dbbuf_dbs[cq_idx(qid, dev->db_stride)];
qid               278 drivers/nvme/host/pci.c 	nvmeq->dbbuf_sq_ei = &dev->dbbuf_eis[sq_idx(qid, dev->db_stride)];
qid               279 drivers/nvme/host/pci.c 	nvmeq->dbbuf_cq_ei = &dev->dbbuf_eis[cq_idx(qid, dev->db_stride)];
qid               517 drivers/nvme/host/pci.c 	if (!iod->nvmeq->qid)
qid               810 drivers/nvme/host/pci.c 			if (iod->nvmeq->qid &&
qid               942 drivers/nvme/host/pci.c 	if (!nvmeq->qid)
qid               944 drivers/nvme/host/pci.c 	return nvmeq->dev->tagset.tags[nvmeq->qid - 1];
qid               965 drivers/nvme/host/pci.c 	if (unlikely(nvmeq->qid == 0 &&
qid              1111 drivers/nvme/host/pci.c 	c.delete_queue.qid = cpu_to_le16(id);
qid              1116 drivers/nvme/host/pci.c static int adapter_alloc_cq(struct nvme_dev *dev, u16 qid,
qid              1132 drivers/nvme/host/pci.c 	c.create_cq.cqid = cpu_to_le16(qid);
qid              1140 drivers/nvme/host/pci.c static int adapter_alloc_sq(struct nvme_dev *dev, u16 qid,
qid              1162 drivers/nvme/host/pci.c 	c.create_sq.sqid = cpu_to_le16(qid);
qid              1165 drivers/nvme/host/pci.c 	c.create_sq.cqid = cpu_to_le16(qid);
qid              1267 drivers/nvme/host/pci.c 			 req->tag, nvmeq->qid);
qid              1284 drivers/nvme/host/pci.c 			 req->tag, nvmeq->qid);
qid              1299 drivers/nvme/host/pci.c 	if (!nvmeq->qid || iod->aborted) {
qid              1302 drivers/nvme/host/pci.c 			 req->tag, nvmeq->qid);
qid              1319 drivers/nvme/host/pci.c 	cmd.abort.sqid = cpu_to_le16(nvmeq->qid);
qid              1323 drivers/nvme/host/pci.c 		 req->tag, nvmeq->qid);
qid              1383 drivers/nvme/host/pci.c 	if (!nvmeq->qid && nvmeq->dev->ctrl.admin_q)
qid              1452 drivers/nvme/host/pci.c 				int qid)
qid              1456 drivers/nvme/host/pci.c 	if (qid && dev->cmb_use_sqes && (dev->cmbsz & NVME_CMBSZ_SQS)) {
qid              1477 drivers/nvme/host/pci.c static int nvme_alloc_queue(struct nvme_dev *dev, int qid, int depth)
qid              1479 drivers/nvme/host/pci.c 	struct nvme_queue *nvmeq = &dev->queues[qid];
qid              1481 drivers/nvme/host/pci.c 	if (dev->ctrl.queue_count > qid)
qid              1484 drivers/nvme/host/pci.c 	nvmeq->sqes = qid ? dev->io_sqes : NVME_ADM_SQES;
qid              1491 drivers/nvme/host/pci.c 	if (nvme_alloc_sq_cmds(dev, nvmeq, qid))
qid              1499 drivers/nvme/host/pci.c 	nvmeq->q_db = &dev->dbs[qid * 2 * dev->db_stride];
qid              1500 drivers/nvme/host/pci.c 	nvmeq->qid = qid;
qid              1519 drivers/nvme/host/pci.c 				nvme_irq, nvmeq, "nvme%dq%d", nr, nvmeq->qid);
qid              1522 drivers/nvme/host/pci.c 				NULL, nvmeq, "nvme%dq%d", nr, nvmeq->qid);
qid              1526 drivers/nvme/host/pci.c static void nvme_init_queue(struct nvme_queue *nvmeq, u16 qid)
qid              1534 drivers/nvme/host/pci.c 	nvmeq->q_db = &dev->dbs[qid * 2 * dev->db_stride];
qid              1536 drivers/nvme/host/pci.c 	nvme_dbbuf_init(dev, nvmeq, qid);
qid              1541 drivers/nvme/host/pci.c static int nvme_create_queue(struct nvme_queue *nvmeq, int qid, bool polled)
qid              1554 drivers/nvme/host/pci.c 		vector = dev->num_vecs == 1 ? 0 : qid;
qid              1558 drivers/nvme/host/pci.c 	result = adapter_alloc_cq(dev, qid, nvmeq, vector);
qid              1562 drivers/nvme/host/pci.c 	result = adapter_alloc_sq(dev, qid, nvmeq);
qid              1569 drivers/nvme/host/pci.c 	nvme_init_queue(nvmeq, qid);
qid              1582 drivers/nvme/host/pci.c 	adapter_delete_sq(dev, qid);
qid              1584 drivers/nvme/host/pci.c 	adapter_delete_cq(dev, qid);
qid              2224 drivers/nvme/host/pci.c 	cmd.delete_queue.qid = cpu_to_le16(nvmeq->qid);
qid              1603 drivers/nvme/host/rdma.c 	priv.qid = cpu_to_le16(nvme_rdma_queue_idx(queue));
qid              1608 drivers/nvme/host/rdma.c 	if (priv.qid == 0) {
qid              1136 drivers/nvme/host/tcp.c static void nvme_tcp_free_queue(struct nvme_ctrl *nctrl, int qid)
qid              1139 drivers/nvme/host/tcp.c 	struct nvme_tcp_queue *queue = &ctrl->queues[qid];
qid              1250 drivers/nvme/host/tcp.c 		int qid, size_t queue_size)
qid              1253 drivers/nvme/host/tcp.c 	struct nvme_tcp_queue *queue = &ctrl->queues[qid];
qid              1263 drivers/nvme/host/tcp.c 	if (qid > 0)
qid              1323 drivers/nvme/host/tcp.c 	if (!qid)
qid              1326 drivers/nvme/host/tcp.c 		n = (qid - 1) % num_online_cpus();
qid              1341 drivers/nvme/host/tcp.c 				qid, ret);
qid              1352 drivers/nvme/host/tcp.c 				"failed to allocate queue %d crypto\n", qid);
qid              1431 drivers/nvme/host/tcp.c static void nvme_tcp_stop_queue(struct nvme_ctrl *nctrl, int qid)
qid              1434 drivers/nvme/host/tcp.c 	struct nvme_tcp_queue *queue = &ctrl->queues[qid];
qid               199 drivers/nvme/host/trace.c 	u16 qid = get_unaligned_le16(spc + 2);
qid               205 drivers/nvme/host/trace.c 			 recfmt, qid, sqsize, cattr, kato);
qid                26 drivers/nvme/host/trace.h #define parse_nvme_cmd(qid, opcode, fctype, cdw10)			\
qid                29 drivers/nvme/host/trace.h 	((qid) ?							\
qid                53 drivers/nvme/host/trace.h 		__field(int, qid)
qid                64 drivers/nvme/host/trace.h 		__entry->qid = nvme_req_qid(req);
qid                77 drivers/nvme/host/trace.h 		      __entry->qid, __entry->cid, __entry->nsid,
qid                79 drivers/nvme/host/trace.h 		      show_opcode_name(__entry->qid, __entry->opcode,
qid                81 drivers/nvme/host/trace.h 		      parse_nvme_cmd(__entry->qid, __entry->opcode,
qid                91 drivers/nvme/host/trace.h 		__field(int, qid)
qid               100 drivers/nvme/host/trace.h 		__entry->qid = nvme_req_qid(req);
qid               110 drivers/nvme/host/trace.h 		      __entry->qid, __entry->cid, __entry->result,
qid               150 drivers/nvme/host/trace.h 		__field(int, qid)
qid               157 drivers/nvme/host/trace.h 		__entry->qid = nvme_req_qid(req);
qid               163 drivers/nvme/host/trace.h 		__entry->qid, __entry->sq_head, __entry->sq_tail
qid               891 drivers/nvme/target/admin-cmd.c 	       req->sq->qid);
qid               692 drivers/nvme/target/core.c 	new_error_slot->sqid = cpu_to_le16(req->sq->qid);
qid               708 drivers/nvme/target/core.c 	req->cqe->sq_id = cpu_to_le16(req->sq->qid);
qid               729 drivers/nvme/target/core.c 		u16 qid, u16 size)
qid               731 drivers/nvme/target/core.c 	cq->qid = qid;
qid               734 drivers/nvme/target/core.c 	ctrl->cqs[qid] = cq;
qid               738 drivers/nvme/target/core.c 		u16 qid, u16 size)
qid               741 drivers/nvme/target/core.c 	sq->qid = qid;
qid               744 drivers/nvme/target/core.c 	ctrl->sqs[qid] = sq;
qid               898 drivers/nvme/target/core.c 	else if (likely(req->sq->qid != 0))
qid               954 drivers/nvme/target/core.c 		if (req->sq->qid && p2p_dev) {
qid              1133 drivers/nvme/target/core.c 		       cmd->common.opcode, req->sq->qid);
qid              1139 drivers/nvme/target/core.c 		       cmd->common.opcode, req->sq->qid);
qid               105 drivers/nvme/target/fabrics-cmd.c 	u16 qid = le16_to_cpu(c->qid);
qid               124 drivers/nvme/target/fabrics-cmd.c 	nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1);
qid               125 drivers/nvme/target/fabrics-cmd.c 	nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1);
qid               136 drivers/nvme/target/fabrics-cmd.c 				qid, ctrl->cntlid, ret);
qid               216 drivers/nvme/target/fabrics-cmd.c 	u16 qid = le16_to_cpu(c->qid);
qid               245 drivers/nvme/target/fabrics-cmd.c 	if (unlikely(qid > ctrl->subsys->max_qid)) {
qid               246 drivers/nvme/target/fabrics-cmd.c 		pr_warn("invalid queue id (%d)\n", qid);
qid               248 drivers/nvme/target/fabrics-cmd.c 		req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid);
qid               259 drivers/nvme/target/fabrics-cmd.c 	pr_debug("adding queue %d to ctrl %d.\n", qid, ctrl->cntlid);
qid               290 drivers/nvme/target/fabrics-cmd.c 	if (cmd->connect.qid == 0)
qid               119 drivers/nvme/target/fc.c 	u16				qid;
qid               178 drivers/nvme/target/fc.c nvmet_fc_makeconnid(struct nvmet_fc_tgt_assoc *assoc, u16 qid)
qid               180 drivers/nvme/target/fc.c 	return (assoc->association_id | qid);
qid               498 drivers/nvme/target/fc.c 	fcpreq->hwqid = queue->qid ?
qid               499 drivers/nvme/target/fc.c 			((queue->qid - 1) % tgtport->ops->max_hw_queues) : 0;
qid               582 drivers/nvme/target/fc.c 			u16 qid, u16 sqsize)
qid               588 drivers/nvme/target/fc.c 	if (qid > NVMET_NR_QUEUES)
qid               600 drivers/nvme/target/fc.c 				assoc->a_id, qid);
qid               604 drivers/nvme/target/fc.c 	queue->qid = qid;
qid               623 drivers/nvme/target/fc.c 	WARN_ON(assoc->queues[qid]);
qid               625 drivers/nvme/target/fc.c 	assoc->queues[qid] = queue;
qid               649 drivers/nvme/target/fc.c 	queue->assoc->queues[queue->qid] = NULL;
qid               758 drivers/nvme/target/fc.c 	u16 qid = nvmet_fc_getqueueid(connection_id);
qid               761 drivers/nvme/target/fc.c 	if (qid > NVMET_NR_QUEUES)
qid               767 drivers/nvme/target/fc.c 			queue = assoc->queues[qid];
qid              1451 drivers/nvme/target/fc.c 					be16_to_cpu(rqst->connect_cmd.qid),
qid              1491 drivers/nvme/target/fc.c 				be16_to_cpu(rqst->connect_cmd.qid)));
qid              2106 drivers/nvme/target/fc.c 		cqe->sq_id = cpu_to_le16(fod->queue->qid);
qid               339 drivers/nvme/target/io-cmd-bdev.c 		       req->sq->qid);
qid               391 drivers/nvme/target/io-cmd-file.c 				cmd->common.opcode, req->sq->qid);
qid                92 drivers/nvme/target/nvmet.h 	u16			qid;
qid                99 drivers/nvme/target/nvmet.h 	u16			qid;
qid               385 drivers/nvme/target/nvmet.h void nvmet_cq_setup(struct nvmet_ctrl *ctrl, struct nvmet_cq *cq, u16 qid,
qid               387 drivers/nvme/target/nvmet.h void nvmet_sq_setup(struct nvmet_ctrl *ctrl, struct nvmet_sq *sq, u16 qid,
qid              1105 drivers/nvme/target/rdma.c 	queue->host_qid = le16_to_cpu(req->qid);
qid              1689 drivers/nvme/target/tcp.c 	if (sq->qid == 0) {
qid               140 drivers/nvme/target/trace.c 	u16 qid = get_unaligned_le16(spc + 2);
qid               146 drivers/nvme/target/trace.c 			 recfmt, qid, sqsize, cattr, kato);
qid                28 drivers/nvme/target/trace.h #define parse_nvme_cmd(qid, opcode, fctype, cdw10)			\
qid                31 drivers/nvme/target/trace.h 	(qid ?								\
qid                55 drivers/nvme/target/trace.h 	if ((init && req->sq->qid) || (!init && req->cq->qid)) {
qid                72 drivers/nvme/target/trace.h 		__field(int, qid)
qid                85 drivers/nvme/target/trace.h 		__entry->qid = req->sq->qid;
qid                99 drivers/nvme/target/trace.h 		__entry->qid, __entry->cid, __entry->nsid,
qid               101 drivers/nvme/target/trace.h 		show_opcode_name(__entry->qid, __entry->opcode,
qid               103 drivers/nvme/target/trace.h 		parse_nvme_cmd(__entry->qid, __entry->opcode,
qid               113 drivers/nvme/target/trace.h 		__field(int, qid)
qid               120 drivers/nvme/target/trace.h 		__entry->qid = req->cq->qid;
qid               129 drivers/nvme/target/trace.h 		__entry->qid, __entry->cid, __entry->result, __entry->status)
qid               295 drivers/s390/crypto/ap_bus.c static int ap_query_queue(ap_qid_t qid, int *queue_depth, int *device_type,
qid               302 drivers/s390/crypto/ap_bus.c 	if (!ap_test_config_card_id(AP_QID_CARD(qid)))
qid               305 drivers/s390/crypto/ap_bus.c 	status = ap_test_queue(qid, ap_apft_available(), &info);
qid               638 drivers/s390/crypto/ap_bus.c 	    AP_QID_CARD(to_ap_queue(dev)->qid) == (int)(long) data)
qid               711 drivers/s390/crypto/ap_bus.c 		card = AP_QID_CARD(to_ap_queue(dev)->qid);
qid               712 drivers/s390/crypto/ap_bus.c 		queue = AP_QID_QUEUE(to_ap_queue(dev)->qid);
qid               787 drivers/s390/crypto/ap_bus.c 		card = AP_QID_CARD(to_ap_queue(dev)->qid);
qid               788 drivers/s390/crypto/ap_bus.c 		queue = AP_QID_QUEUE(to_ap_queue(dev)->qid);
qid              1316 drivers/s390/crypto/ap_bus.c static int ap_get_compatible_type(ap_qid_t qid, int rawtype, unsigned int func)
qid              1337 drivers/s390/crypto/ap_bus.c 		status = ap_qact(qid, 0, &apinfo);
qid              1345 drivers/s390/crypto/ap_bus.c 		       AP_QID_CARD(qid), AP_QID_QUEUE(qid), rawtype);
qid              1348 drivers/s390/crypto/ap_bus.c 		       AP_QID_CARD(qid), AP_QID_QUEUE(qid), rawtype, comp_type);
qid              1367 drivers/s390/crypto/ap_bus.c 	return is_queue_dev(dev) && to_ap_queue(dev)->qid == (int)(long) data;
qid              1377 drivers/s390/crypto/ap_bus.c 		&& AP_QID_QUEUE(to_ap_queue(dev)->qid) == (int)(long) data;
qid              1386 drivers/s390/crypto/ap_bus.c 	ap_qid_t qid;
qid              1418 drivers/s390/crypto/ap_bus.c 			qid = AP_MKQID(id, dom);
qid              1419 drivers/s390/crypto/ap_bus.c 			if (ap_query_queue(qid, &depth, &type, &func) == 0)
qid              1455 drivers/s390/crypto/ap_bus.c 		qid = AP_MKQID(id, dom);
qid              1457 drivers/s390/crypto/ap_bus.c 				      (void *)(long) qid,
qid              1471 drivers/s390/crypto/ap_bus.c 		rc = ap_query_queue(qid, &depth, &type, &func);
qid              1493 drivers/s390/crypto/ap_bus.c 		comp_type = ap_get_compatible_type(qid, type, func);
qid              1515 drivers/s390/crypto/ap_bus.c 		aq = ap_queue_create(qid, comp_type);
qid               176 drivers/s390/crypto/ap_bus.h 	ap_qid_t qid;			/* AP queue id. */
qid                37 drivers/s390/crypto/ap_queue.c 	status = ap_aqic(aq->qid, qirqctrl, ind);
qid                47 drivers/s390/crypto/ap_queue.c 		       AP_QID_CARD(aq->qid),
qid                48 drivers/s390/crypto/ap_queue.c 		       AP_QID_QUEUE(aq->qid));
qid                71 drivers/s390/crypto/ap_queue.c __ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length,
qid                75 drivers/s390/crypto/ap_queue.c 		qid |= 0x400000UL;
qid                76 drivers/s390/crypto/ap_queue.c 	return ap_nqap(qid, psmid, msg, length);
qid                79 drivers/s390/crypto/ap_queue.c int ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length)
qid                83 drivers/s390/crypto/ap_queue.c 	status = __ap_send(qid, psmid, msg, length, 0);
qid                98 drivers/s390/crypto/ap_queue.c int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length)
qid               104 drivers/s390/crypto/ap_queue.c 	status = ap_dqap(qid, psmid, msg, length);
qid               139 drivers/s390/crypto/ap_queue.c 	status = ap_dqap(aq->qid, &aq->reply->psmid,
qid               243 drivers/s390/crypto/ap_queue.c 	status = __ap_send(aq->qid, ap_msg->psmid,
qid               298 drivers/s390/crypto/ap_queue.c 	status = ap_rapq(aq->qid);
qid               332 drivers/s390/crypto/ap_queue.c 		status = ap_tapq(aq->qid, NULL);
qid               370 drivers/s390/crypto/ap_queue.c 		status = ap_tapq(aq->qid, NULL);
qid               569 drivers/s390/crypto/ap_queue.c 	       AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
qid               630 drivers/s390/crypto/ap_queue.c struct ap_queue *ap_queue_create(ap_qid_t qid, int device_type)
qid               640 drivers/s390/crypto/ap_queue.c 	aq->qid = qid;
qid               769 drivers/s390/crypto/ap_queue.c 	ap_zapq(aq->qid);
qid                60 drivers/s390/crypto/vfio_ap_drv.c 	q->apqn = to_ap_queue(&apdev->device)->qid;
qid               440 drivers/s390/crypto/vfio_ap_ops.c 	ap_qid_t qid;
qid               444 drivers/s390/crypto/vfio_ap_ops.c 		qid = AP_MKQID(*qres->apid, *qres->apqi);
qid               445 drivers/s390/crypto/vfio_ap_ops.c 		if (qid == ap_queue->qid)
qid               448 drivers/s390/crypto/vfio_ap_ops.c 		id = AP_QID_CARD(ap_queue->qid);
qid               452 drivers/s390/crypto/vfio_ap_ops.c 		id = AP_QID_QUEUE(ap_queue->qid);
qid               638 drivers/s390/crypto/zcrypt_api.c 	int qid = 0, rc = -ENODEV;
qid               684 drivers/s390/crypto/zcrypt_api.c 						AP_QID_QUEUE(zq->queue->qid)))
qid               702 drivers/s390/crypto/zcrypt_api.c 	qid = pref_zq->queue->qid;
qid               711 drivers/s390/crypto/zcrypt_api.c 			      AP_QID_CARD(qid), AP_QID_QUEUE(qid));
qid               722 drivers/s390/crypto/zcrypt_api.c 	int qid = 0, rc = -ENODEV;
qid               768 drivers/s390/crypto/zcrypt_api.c 						AP_QID_QUEUE(zq->queue->qid)))
qid               786 drivers/s390/crypto/zcrypt_api.c 	qid = pref_zq->queue->qid;
qid               795 drivers/s390/crypto/zcrypt_api.c 			      AP_QID_CARD(qid), AP_QID_QUEUE(qid));
qid               808 drivers/s390/crypto/zcrypt_api.c 	int qid = 0, rc = -ENODEV;
qid               853 drivers/s390/crypto/zcrypt_api.c 			     tdom != AP_QID_QUEUE(zq->queue->qid)))
qid               857 drivers/s390/crypto/zcrypt_api.c 						AP_QID_QUEUE(zq->queue->qid)))
qid               876 drivers/s390/crypto/zcrypt_api.c 	qid = pref_zq->queue->qid;
qid               878 drivers/s390/crypto/zcrypt_api.c 		*domain = AP_QID_QUEUE(qid);
qid               889 drivers/s390/crypto/zcrypt_api.c 			      AP_QID_CARD(qid), AP_QID_QUEUE(qid));
qid               933 drivers/s390/crypto/zcrypt_api.c 	int qid = 0, rc = -ENODEV;
qid               990 drivers/s390/crypto/zcrypt_api.c 			     !is_desired_ep11_queue(zq->queue->qid,
qid               995 drivers/s390/crypto/zcrypt_api.c 						AP_QID_QUEUE(zq->queue->qid)))
qid              1013 drivers/s390/crypto/zcrypt_api.c 	qid = pref_zq->queue->qid;
qid              1025 drivers/s390/crypto/zcrypt_api.c 			      AP_QID_CARD(qid), AP_QID_QUEUE(qid));
qid              1037 drivers/s390/crypto/zcrypt_api.c 	int qid = 0, rc = -ENODEV;
qid              1078 drivers/s390/crypto/zcrypt_api.c 	qid = pref_zq->queue->qid;
qid              1088 drivers/s390/crypto/zcrypt_api.c 			      AP_QID_CARD(qid), AP_QID_QUEUE(qid));
qid              1105 drivers/s390/crypto/zcrypt_api.c 			card = AP_QID_CARD(zq->queue->qid);
qid              1108 drivers/s390/crypto/zcrypt_api.c 			queue = AP_QID_QUEUE(zq->queue->qid);
qid              1112 drivers/s390/crypto/zcrypt_api.c 			stat->qid = zq->queue->qid;
qid              1132 drivers/s390/crypto/zcrypt_api.c 			card = AP_QID_CARD(zq->queue->qid);
qid              1133 drivers/s390/crypto/zcrypt_api.c 			queue = AP_QID_QUEUE(zq->queue->qid);
qid              1137 drivers/s390/crypto/zcrypt_api.c 			stat->qid = zq->queue->qid;
qid              1156 drivers/s390/crypto/zcrypt_api.c 			if (card == AP_QID_CARD(zq->queue->qid) &&
qid              1157 drivers/s390/crypto/zcrypt_api.c 			    queue == AP_QID_QUEUE(zq->queue->qid)) {
qid              1160 drivers/s390/crypto/zcrypt_api.c 				devstat->qid = zq->queue->qid;
qid              1183 drivers/s390/crypto/zcrypt_api.c 			card = AP_QID_CARD(zq->queue->qid);
qid              1184 drivers/s390/crypto/zcrypt_api.c 			if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index
qid              1204 drivers/s390/crypto/zcrypt_api.c 			card = AP_QID_CARD(zq->queue->qid);
qid              1205 drivers/s390/crypto/zcrypt_api.c 			if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index
qid              1231 drivers/s390/crypto/zcrypt_api.c 			card = AP_QID_CARD(zq->queue->qid);
qid              1232 drivers/s390/crypto/zcrypt_api.c 			if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index
qid              1256 drivers/s390/crypto/zcrypt_api.c 			if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index)
qid              1279 drivers/s390/crypto/zcrypt_api.c 			if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index)
qid              1580 drivers/s390/crypto/zcrypt_ccamisc.c 		card = AP_QID_CARD(device_status[i].qid);
qid              1581 drivers/s390/crypto/zcrypt_ccamisc.c 		dom = AP_QID_QUEUE(device_status[i].qid);
qid              1612 drivers/s390/crypto/zcrypt_ccamisc.c 			card = AP_QID_CARD(device_status[i].qid);
qid              1613 drivers/s390/crypto/zcrypt_ccamisc.c 			dom = AP_QID_QUEUE(device_status[i].qid);
qid              1630 drivers/s390/crypto/zcrypt_ccamisc.c 			card = AP_QID_CARD(device_status[oi].qid);
qid              1631 drivers/s390/crypto/zcrypt_ccamisc.c 			dom = AP_QID_QUEUE(device_status[oi].qid);
qid              1699 drivers/s390/crypto/zcrypt_ccamisc.c 			card = AP_QID_CARD(device_status[i].qid);
qid              1700 drivers/s390/crypto/zcrypt_ccamisc.c 			dom = AP_QID_QUEUE(device_status[i].qid);
qid                97 drivers/s390/crypto/zcrypt_cex2c.c 	msg->cprbx.domain = AP_QID_QUEUE(aq->qid);
qid                99 drivers/s390/crypto/zcrypt_cex2c.c 	rc = ap_send(aq->qid, 0x0102030405060708ULL, ap_msg.message,
qid               107 drivers/s390/crypto/zcrypt_cex2c.c 		rc = ap_recv(aq->qid, &psmid, ap_msg.message, 4096);
qid               223 drivers/s390/crypto/zcrypt_cex2c.c 	ap_rapq(aq->qid);
qid               117 drivers/s390/crypto/zcrypt_cex4.c 	cca_get_info(AP_QID_CARD(zq->queue->qid),
qid               118 drivers/s390/crypto/zcrypt_cex4.c 		     AP_QID_QUEUE(zq->queue->qid),
qid                84 drivers/s390/crypto/zcrypt_error.h 	int card = AP_QID_CARD(zq->queue->qid);
qid                85 drivers/s390/crypto/zcrypt_error.h 	int queue = AP_QID_QUEUE(zq->queue->qid);
qid               360 drivers/s390/crypto/zcrypt_msgtype50.c 		       AP_QID_CARD(zq->queue->qid),
qid               361 drivers/s390/crypto/zcrypt_msgtype50.c 		       AP_QID_QUEUE(zq->queue->qid));
qid               364 drivers/s390/crypto/zcrypt_msgtype50.c 			   AP_QID_CARD(zq->queue->qid),
qid               365 drivers/s390/crypto/zcrypt_msgtype50.c 			   AP_QID_QUEUE(zq->queue->qid),
qid               397 drivers/s390/crypto/zcrypt_msgtype50.c 		       AP_QID_CARD(zq->queue->qid),
qid               398 drivers/s390/crypto/zcrypt_msgtype50.c 		       AP_QID_QUEUE(zq->queue->qid));
qid               401 drivers/s390/crypto/zcrypt_msgtype50.c 			   AP_QID_CARD(zq->queue->qid),
qid               402 drivers/s390/crypto/zcrypt_msgtype50.c 			   AP_QID_QUEUE(zq->queue->qid),
qid               297 drivers/s390/crypto/zcrypt_msgtype6.c 	msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid);
qid               367 drivers/s390/crypto/zcrypt_msgtype6.c 	msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid);
qid               657 drivers/s390/crypto/zcrypt_msgtype6.c 				   AP_QID_CARD(zq->queue->qid),
qid               658 drivers/s390/crypto/zcrypt_msgtype6.c 				   AP_QID_QUEUE(zq->queue->qid),
qid               664 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_CARD(zq->queue->qid),
qid               665 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_QUEUE(zq->queue->qid));
qid               668 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_CARD(zq->queue->qid),
qid               669 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_QUEUE(zq->queue->qid),
qid               808 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_CARD(zq->queue->qid),
qid               809 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_QUEUE(zq->queue->qid));
qid               812 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_CARD(zq->queue->qid),
qid               813 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_QUEUE(zq->queue->qid),
qid               842 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_CARD(zq->queue->qid),
qid               843 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_QUEUE(zq->queue->qid));
qid               846 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_CARD(zq->queue->qid),
qid               847 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_QUEUE(zq->queue->qid),
qid               871 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_CARD(zq->queue->qid),
qid               872 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_QUEUE(zq->queue->qid));
qid               875 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_CARD(zq->queue->qid),
qid               876 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_QUEUE(zq->queue->qid),
qid               901 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_CARD(zq->queue->qid),
qid               902 drivers/s390/crypto/zcrypt_msgtype6.c 		       AP_QID_QUEUE(zq->queue->qid));
qid               905 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_CARD(zq->queue->qid),
qid               906 drivers/s390/crypto/zcrypt_msgtype6.c 			   AP_QID_QUEUE(zq->queue->qid),
qid              1217 drivers/s390/crypto/zcrypt_msgtype6.c 					AP_QID_QUEUE(zq->queue->qid);
qid              1235 drivers/s390/crypto/zcrypt_msgtype6.c 					AP_QID_QUEUE(zq->queue->qid);
qid              1297 drivers/s390/crypto/zcrypt_msgtype6.c 	msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid);
qid                64 drivers/s390/crypto/zcrypt_queue.c 		   AP_QID_CARD(zq->queue->qid),
qid                65 drivers/s390/crypto/zcrypt_queue.c 		   AP_QID_QUEUE(zq->queue->qid),
qid               168 drivers/s390/crypto/zcrypt_queue.c 		   AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid));
qid               211 drivers/s390/crypto/zcrypt_queue.c 		   AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid));
qid              2712 drivers/scsi/aacraid/aacraid.h int aac_queue_get(struct aac_dev * dev, u32 * index, u32 qid, struct hw_fib * hw_fib, int wait, struct fib * fibptr, unsigned long *nonotify);
qid               357 drivers/scsi/aacraid/commsup.c static int aac_get_entry (struct aac_dev * dev, u32 qid, struct aac_entry **entry, u32 * index, unsigned long *nonotify)
qid               369 drivers/scsi/aacraid/commsup.c 	q = &dev->queues->queue[qid];
qid               375 drivers/scsi/aacraid/commsup.c 			if (qid == AdapNormCmdQueue)
qid               384 drivers/scsi/aacraid/commsup.c 	if (qid == AdapNormCmdQueue) {
qid               395 drivers/scsi/aacraid/commsup.c 				qid, atomic_read(&q->numpending));
qid               419 drivers/scsi/aacraid/commsup.c int aac_queue_get(struct aac_dev * dev, u32 * index, u32 qid, struct hw_fib * hw_fib, int wait, struct fib * fibptr, unsigned long *nonotify)
qid               424 drivers/scsi/aacraid/commsup.c 	if (qid == AdapNormCmdQueue) {
qid               426 drivers/scsi/aacraid/commsup.c 		while (!aac_get_entry(dev, qid, &entry, index, nonotify)) {
qid               435 drivers/scsi/aacraid/commsup.c 		while (!aac_get_entry(dev, qid, &entry, index, nonotify)) {
qid               830 drivers/scsi/aacraid/commsup.c void aac_consumer_free(struct aac_dev * dev, struct aac_queue *q, u32 qid)
qid               844 drivers/scsi/aacraid/commsup.c 		switch (qid) {
qid               684 drivers/scsi/be2iscsi/be_main.h 	u8 qid[10];
qid                51 drivers/scsi/bfa/bfa.h 		(__mh).mtag.h2i.qid     = (__bfa)->iocfc.hw_qid[__reqq];\
qid               692 drivers/scsi/bfa/bfa_core.c bfa_reqq_resume(struct bfa_s *bfa, int qid)
qid               697 drivers/scsi/bfa/bfa_core.c 	waitq = bfa_reqq(bfa, qid);
qid               702 drivers/scsi/bfa/bfa_core.c 		if (bfa_reqq_full(bfa, qid))
qid               712 drivers/scsi/bfa/bfa_core.c bfa_isr_rspq(struct bfa_s *bfa, int qid)
qid               719 drivers/scsi/bfa/bfa_core.c 	ci = bfa_rspq_ci(bfa, qid);
qid               720 drivers/scsi/bfa/bfa_core.c 	pi = bfa_rspq_pi(bfa, qid);
qid               725 drivers/scsi/bfa/bfa_core.c 		m = bfa_rspq_elem(bfa, qid, ci);
qid               735 drivers/scsi/bfa/bfa_core.c 	bfa_isr_rspq_ack(bfa, qid, ci);
qid               740 drivers/scsi/bfa/bfa_core.c 	waitq = bfa_reqq(bfa, qid);
qid               742 drivers/scsi/bfa/bfa_core.c 		bfa_reqq_resume(bfa, qid);
qid               748 drivers/scsi/bfa/bfa_core.c bfa_isr_reqq(struct bfa_s *bfa, int qid)
qid               752 drivers/scsi/bfa/bfa_core.c 	bfa_isr_reqq_ack(bfa, qid);
qid               757 drivers/scsi/bfa/bfa_core.c 	waitq = bfa_reqq(bfa, qid);
qid               759 drivers/scsi/bfa/bfa_core.c 		bfa_reqq_resume(bfa, qid);
qid                52 drivers/scsi/bfa/bfi.h 			u8	qid;
qid              1137 drivers/scsi/csiostor/csio_wr.c 	uint32_t wr_type, fw_qid, qid;
qid              1170 drivers/scsi/csiostor/csio_wr.c 			qid = fw_qid - wrm->fw_iq_start;
qid              1171 drivers/scsi/csiostor/csio_wr.c 			q_completed = hw->wrm.intr_map[qid];
qid              1173 drivers/scsi/csiostor/csio_wr.c 			if (unlikely(qid ==
qid               274 drivers/scsi/csiostor/csio_wr.h 	__be32 qid;
qid               562 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_wq_by_id(struct lpfc_hba *phba, int qid)
qid               567 drivers/scsi/lpfc/lpfc_debugfs.h 		if (phba->sli4_hba.hdwq[wq_idx].io_wq->queue_id == qid)
qid               570 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("IO WQ[Idx:%d|Qid:%d]\n", wq_idx, qid);
qid               575 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.els_wq->queue_id == qid) {
qid               576 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("ELS WQ[Qid:%d]\n", qid);
qid               581 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.nvmels_wq->queue_id == qid) {
qid               582 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("NVME LS WQ[Qid:%d]\n", qid);
qid               596 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_mq_by_id(struct lpfc_hba *phba, int qid)
qid               598 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.mbx_wq->queue_id == qid) {
qid               599 drivers/scsi/lpfc/lpfc_debugfs.h 		printk(KERN_ERR "MBX WQ[Qid:%d]\n", qid);
qid               613 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_rq_by_id(struct lpfc_hba *phba, int qid)
qid               615 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.hdr_rq->queue_id == qid) {
qid               616 drivers/scsi/lpfc/lpfc_debugfs.h 		printk(KERN_ERR "HDR RQ[Qid:%d]\n", qid);
qid               620 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.dat_rq->queue_id == qid) {
qid               621 drivers/scsi/lpfc/lpfc_debugfs.h 		printk(KERN_ERR "DAT RQ[Qid:%d]\n", qid);
qid               635 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_cq_by_id(struct lpfc_hba *phba, int qid)
qid               640 drivers/scsi/lpfc/lpfc_debugfs.h 		if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid)
qid               644 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("IO CQ[Idx:%d|Qid:%d]\n", cq_idx, qid);
qid               649 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.els_cq->queue_id == qid) {
qid               650 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("ELS CQ[Qid:%d]\n", qid);
qid               655 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.nvmels_cq->queue_id == qid) {
qid               656 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("NVME LS CQ[Qid:%d]\n", qid);
qid               661 drivers/scsi/lpfc/lpfc_debugfs.h 	if (phba->sli4_hba.mbx_cq->queue_id == qid) {
qid               662 drivers/scsi/lpfc/lpfc_debugfs.h 		pr_err("MBX CQ[Qid:%d]\n", qid);
qid               676 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_eq_by_id(struct lpfc_hba *phba, int qid)
qid               681 drivers/scsi/lpfc/lpfc_debugfs.h 		if (phba->sli4_hba.hdwq[eq_idx].hba_eq->queue_id == qid)
qid               685 drivers/scsi/lpfc/lpfc_debugfs.h 		printk(KERN_ERR "FCP EQ[Idx:%d|Qid:%d]\n", eq_idx, qid);
qid              4013 drivers/scsi/qla2xxx/qla_nx.c 	uint32_t r_stride, r_value, r_cnt, qid = 0;
qid              4025 drivers/scsi/qla2xxx/qla_nx.c 		qla82xx_md_rw_32(ha, s_addr, qid, 1);
qid              4032 drivers/scsi/qla2xxx/qla_nx.c 		qid += q_hdr->q_strd.queue_id_stride;
qid              2627 drivers/scsi/qla2xxx/qla_nx2.c 	uint32_t r_stride, r_value, r_cnt, qid = 0;
qid              2640 drivers/scsi/qla2xxx/qla_nx2.c 		qla8044_wr_reg_indirect(vha, s_addr, qid);
qid              2647 drivers/scsi/qla2xxx/qla_nx2.c 		qid += q_hdr->q_strd.queue_id_stride;
qid              3853 drivers/scsi/qla2xxx/qla_target.c 	int qid = GET_QID(handle);
qid              3859 drivers/scsi/qla2xxx/qla_target.c 	if (qid == rsp->req->id) {
qid              3861 drivers/scsi/qla2xxx/qla_target.c 	} else if (vha->hw->req_q_map[qid]) {
qid              3865 drivers/scsi/qla2xxx/qla_target.c 		req = vha->hw->req_q_map[qid];
qid              2415 drivers/scsi/qla4xxx/ql4_nx.c 	uint32_t r_stride, r_value, r_cnt, qid = 0;
qid              2428 drivers/scsi/qla4xxx/ql4_nx.c 		ha->isp_ops->wr_reg_indirect(ha, s_addr, qid);
qid              2435 drivers/scsi/qla4xxx/ql4_nx.c 		qid += q_hdr->q_strd.queue_id_stride;
qid               101 drivers/staging/media/ipu3/ipu3-css.c 	enum imgu_abi_queue_id qid;
qid              1943 drivers/staging/media/ipu3/ipu3-css.c 	if (b->queue >= IPU3_CSS_QUEUES || !imgu_css_queues[b->queue].qid)
qid              1946 drivers/staging/media/ipu3/ipu3-css.c 	b->queue_pos = imgu_css_queue_pos(css, imgu_css_queues[b->queue].qid,
qid              1976 drivers/staging/media/ipu3/ipu3-css.c 	r = imgu_css_queue_data(css, imgu_css_queues[b->queue].qid,
qid              1982 drivers/staging/media/ipu3/ipu3-css.c 					      imgu_css_queues[b->queue].qid);
qid              2015 drivers/staging/media/ipu3/ipu3-css.c 	int evtype, pipe, pipeid, queue, qid, r;
qid              2038 drivers/staging/media/ipu3/ipu3-css.c 		qid = imgu_css_queues[queue].qid;
qid              2045 drivers/staging/media/ipu3/ipu3-css.c 		if (qid >= IMGU_ABI_QUEUE_NUM) {
qid              2046 drivers/staging/media/ipu3/ipu3-css.c 			dev_err(css->dev, "Invalid qid: %i\n", qid);
qid              2054 drivers/staging/media/ipu3/ipu3-css.c 		r = imgu_css_dequeue_data(css, qid, &daddr);
qid              2062 drivers/staging/media/ipu3/ipu3-css.c 					IMGU_ABI_EVENT_BUFFER_DEQUEUED(qid));
qid                86 fs/9p/cache.c  	if (buflen != sizeof(v9inode->qid.version))
qid                89 fs/9p/cache.c  	if (memcmp(buffer, &v9inode->qid.version,
qid                90 fs/9p/cache.c  		   sizeof(v9inode->qid.version)))
qid               117 fs/9p/cache.c  						  &v9inode->qid.path,
qid               118 fs/9p/cache.c  						  sizeof(v9inode->qid.path),
qid               119 fs/9p/cache.c  						  &v9inode->qid.version,
qid               120 fs/9p/cache.c  						  sizeof(v9inode->qid.version),
qid               138 fs/9p/cache.c  	fscache_relinquish_cookie(v9inode->fscache, &v9inode->qid.version,
qid               190 fs/9p/cache.c  						  &v9inode->qid.path,
qid               191 fs/9p/cache.c  						  sizeof(v9inode->qid.path),
qid               192 fs/9p/cache.c  						  &v9inode->qid.version,
qid               193 fs/9p/cache.c  						  sizeof(v9inode->qid.version),
qid               627 fs/9p/v9fs.c   	memset(&v9inode->qid, 0, sizeof(v9inode->qid));
qid               115 fs/9p/v9fs.h   	struct p9_qid qid;
qid                51 fs/9p/v9fs_vfs.h ino_t v9fs_qid2ino(struct p9_qid *qid);
qid               131 fs/9p/vfs_dir.c 					 v9fs_qid2ino(&st.qid), dt_type(&st));
qid               188 fs/9p/vfs_dir.c 				      v9fs_qid2ino(&curdirent.qid),
qid               202 fs/9p/vfs_inode.c 	wstat->qid.type = ~0;
qid               203 fs/9p/vfs_inode.c 	wstat->qid.version = ~0;
qid               204 fs/9p/vfs_inode.c 	*((long long *)&wstat->qid.path) = ~0;
qid               459 fs/9p/vfs_inode.c 	if (memcmp(&v9inode->qid.version,
qid               460 fs/9p/vfs_inode.c 		   &st->qid.version, sizeof(v9inode->qid.version)))
qid               463 fs/9p/vfs_inode.c 	if (v9inode->qid.type != st->qid.type)
qid               466 fs/9p/vfs_inode.c 	if (v9inode->qid.path != st->qid.path)
qid               481 fs/9p/vfs_inode.c 	memcpy(&v9inode->qid, &st->qid, sizeof(st->qid));
qid               486 fs/9p/vfs_inode.c 				   struct p9_qid *qid,
qid               503 fs/9p/vfs_inode.c 	i_ino = v9fs_qid2ino(qid);
qid               541 fs/9p/vfs_inode.c 	inode = v9fs_qid_iget(sb, &st->qid, st, new);
qid              1214 fs/9p/vfs_inode.c ino_t v9fs_qid2ino(struct p9_qid *qid)
qid              1216 fs/9p/vfs_inode.c 	u64 path = qid->path + 2;
qid                69 fs/9p/vfs_inode_dotl.c 	if (memcmp(&v9inode->qid.version,
qid                70 fs/9p/vfs_inode_dotl.c 		   &st->qid.version, sizeof(v9inode->qid.version)))
qid                73 fs/9p/vfs_inode_dotl.c 	if (v9inode->qid.type != st->qid.type)
qid                76 fs/9p/vfs_inode_dotl.c 	if (v9inode->qid.path != st->qid.path)
qid                92 fs/9p/vfs_inode_dotl.c 	memcpy(&v9inode->qid, &st->qid, sizeof(st->qid));
qid                98 fs/9p/vfs_inode_dotl.c 					struct p9_qid *qid,
qid               114 fs/9p/vfs_inode_dotl.c 	i_ino = v9fs_qid2ino(qid);
qid               156 fs/9p/vfs_inode_dotl.c 	inode = v9fs_qid_iget_dotl(sb, &st->qid, fid, st, new);
qid               235 fs/9p/vfs_inode_dotl.c 	struct p9_qid qid;
qid               289 fs/9p/vfs_inode_dotl.c 				    mode, gid, &qid);
qid               378 fs/9p/vfs_inode_dotl.c 	struct p9_qid qid;
qid               407 fs/9p/vfs_inode_dotl.c 	err = p9_client_mkdir_dotl(dfid, name, mode, gid, &qid);
qid               674 fs/9p/vfs_inode_dotl.c 	struct p9_qid qid;
qid               694 fs/9p/vfs_inode_dotl.c 	err = p9_client_symlink(dfid, name, symname, gid, &qid);
qid               812 fs/9p/vfs_inode_dotl.c 	struct p9_qid qid;
qid               839 fs/9p/vfs_inode_dotl.c 	err = p9_client_mknod_dotl(dfid, name, mode, rdev, gid, &qid);
qid               163 fs/9p/vfs_super.c 		d_inode(root)->i_ino = v9fs_qid2ino(&st->qid);
qid               174 fs/9p/vfs_super.c 		d_inode(root)->i_ino = v9fs_qid2ino(&st->qid);
qid              1388 fs/ext4/super.c static int ext4_get_next_id(struct super_block *sb, struct kqid *qid);
qid              5575 fs/ext4/super.c 	struct kqid qid;
qid              5580 fs/ext4/super.c 	qid = make_kqid_projid(projid);
qid              5581 fs/ext4/super.c 	dquot = dqget(sb, qid);
qid              6071 fs/ext4/super.c static int ext4_get_next_id(struct super_block *sb, struct kqid *qid)
qid              6075 fs/ext4/super.c 	if (!sb_has_quota_loaded(sb, qid->type))
qid              6077 fs/ext4/super.c 	ops = sb_dqopt(sb)->ops[qid->type];
qid              6080 fs/ext4/super.c 	return dquot_get_next_id(sb, qid);
qid              1205 fs/f2fs/super.c 	struct kqid qid;
qid              1210 fs/f2fs/super.c 	qid = make_kqid_projid(projid);
qid              1211 fs/f2fs/super.c 	dquot = dqget(sb, qid);
qid                86 fs/gfs2/quota.c 				 const struct kqid qid)
qid                91 fs/gfs2/quota.c 	h = jhash(&qid, sizeof(struct kqid), h);
qid               196 fs/gfs2/quota.c 	struct kqid qid = qd->qd_id;
qid               197 fs/gfs2/quota.c 	return (2 * (u64)from_kqid(&init_user_ns, qid)) +
qid               198 fs/gfs2/quota.c 		((qid.type == USRQUOTA) ? 0 : 1);
qid               211 fs/gfs2/quota.c static struct gfs2_quota_data *qd_alloc(unsigned hash, struct gfs2_sbd *sdp, struct kqid qid)
qid               223 fs/gfs2/quota.c 	qd->qd_id = qid;
qid               242 fs/gfs2/quota.c 						     struct kqid qid)
qid               248 fs/gfs2/quota.c 		if (!qid_eq(qd->qd_id, qid))
qid               262 fs/gfs2/quota.c static int qd_get(struct gfs2_sbd *sdp, struct kqid qid,
qid               266 fs/gfs2/quota.c 	unsigned int hash = gfs2_qd_hash(sdp, qid);
qid               269 fs/gfs2/quota.c 	*qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
qid               275 fs/gfs2/quota.c 	new_qd = qd_alloc(hash, sdp, qid);
qid               281 fs/gfs2/quota.c 	*qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
qid               494 fs/gfs2/quota.c static int qdsb_get(struct gfs2_sbd *sdp, struct kqid qid,
qid               499 fs/gfs2/quota.c 	error = qd_get(sdp, qid, qdp);
qid              1317 fs/gfs2/quota.c int gfs2_quota_refresh(struct gfs2_sbd *sdp, struct kqid qid)
qid              1323 fs/gfs2/quota.c 	error = qd_get(sdp, qid, &qd);
qid              1610 fs/gfs2/quota.c static int gfs2_get_dqblk(struct super_block *sb, struct kqid qid,
qid              1624 fs/gfs2/quota.c 	if ((qid.type != USRQUOTA) &&
qid              1625 fs/gfs2/quota.c 	    (qid.type != GRPQUOTA))
qid              1628 fs/gfs2/quota.c 	error = qd_get(sdp, qid, &qd);
qid              1649 fs/gfs2/quota.c static int gfs2_set_dqblk(struct super_block *sb, struct kqid qid,
qid              1665 fs/gfs2/quota.c 	if ((qid.type != USRQUOTA) &&
qid              1666 fs/gfs2/quota.c 	    (qid.type != GRPQUOTA))
qid              1672 fs/gfs2/quota.c 	error = qd_get(sdp, qid, &qd);
qid                32 fs/gfs2/quota.h extern int gfs2_quota_refresh(struct gfs2_sbd *sdp, struct kqid qid);
qid               185 fs/gfs2/sys.c  	struct kqid qid;
qid               196 fs/gfs2/sys.c  	qid = make_kqid(current_user_ns(), USRQUOTA, id);
qid               197 fs/gfs2/sys.c  	if (!qid_valid(qid))
qid               200 fs/gfs2/sys.c  	error = gfs2_quota_refresh(sdp, qid);
qid               207 fs/gfs2/sys.c  	struct kqid qid;
qid               218 fs/gfs2/sys.c  	qid = make_kqid(current_user_ns(), GRPQUOTA, id);
qid               219 fs/gfs2/sys.c  	if (!qid_valid(qid))
qid               222 fs/gfs2/sys.c  	error = gfs2_quota_refresh(sdp, qid);
qid                27 fs/ntfs/quota.c 	const le32 qid = QUOTA_DEFAULTS_ID;
qid                43 fs/ntfs/quota.c 	err = ntfs_index_lookup(&qid, sizeof(qid), ictx);
qid               877 fs/ocfs2/quota_global.c static int ocfs2_get_next_id(struct super_block *sb, struct kqid *qid)
qid               879 fs/ocfs2/quota_global.c 	int type = qid->type;
qid               883 fs/ocfs2/quota_global.c 	trace_ocfs2_get_next_id(from_kqid(&init_user_ns, *qid), type);
qid               894 fs/ocfs2/quota_global.c 	status = qtree_get_next_id(&info->dqi_gi, qid);
qid               262 fs/quota/dquot.c hashfn(const struct super_block *sb, struct kqid qid)
qid               264 fs/quota/dquot.c 	unsigned int id = from_kqid(&init_user_ns, qid);
qid               265 fs/quota/dquot.c 	int type = qid.type;
qid               288 fs/quota/dquot.c 				struct kqid qid)
qid               295 fs/quota/dquot.c 		if (dquot->dq_sb == sb && qid_eq(dquot->dq_id, qid))
qid               848 fs/quota/dquot.c struct dquot *dqget(struct super_block *sb, struct kqid qid)
qid               850 fs/quota/dquot.c 	unsigned int hashent = hashfn(sb, qid);
qid               853 fs/quota/dquot.c 	if (!qid_has_mapping(sb->s_user_ns, qid))
qid               856 fs/quota/dquot.c         if (!sb_has_quota_active(sb, qid.type))
qid               861 fs/quota/dquot.c 	if (!sb_has_quota_active(sb, qid.type)) {
qid               869 fs/quota/dquot.c 	dquot = find_dquot(hashent, sb, qid);
qid               873 fs/quota/dquot.c 			empty = get_empty_dquot(sb, qid.type);
qid               880 fs/quota/dquot.c 		dquot->dq_id = qid;
qid              1446 fs/quota/dquot.c 		struct kqid qid;
qid              1468 fs/quota/dquot.c 			qid = make_kqid_uid(inode->i_uid);
qid              1471 fs/quota/dquot.c 			qid = make_kqid_gid(inode->i_gid);
qid              1477 fs/quota/dquot.c 			qid = make_kqid_projid(projid);
qid              1480 fs/quota/dquot.c 		dquot = dqget(sb, qid);
qid              2126 fs/quota/dquot.c int dquot_get_next_id(struct super_block *sb, struct kqid *qid)
qid              2130 fs/quota/dquot.c 	if (!sb_has_quota_active(sb, qid->type))
qid              2132 fs/quota/dquot.c 	if (!dqopt->ops[qid->type]->get_next_id)
qid              2134 fs/quota/dquot.c 	return dqopt->ops[qid->type]->get_next_id(sb, qid);
qid              2630 fs/quota/dquot.c int dquot_get_dqblk(struct super_block *sb, struct kqid qid,
qid              2635 fs/quota/dquot.c 	dquot = dqget(sb, qid);
qid              2645 fs/quota/dquot.c int dquot_get_next_dqblk(struct super_block *sb, struct kqid *qid,
qid              2653 fs/quota/dquot.c 	err = sb->dq_op->get_next_id(sb, qid);
qid              2656 fs/quota/dquot.c 	dquot = dqget(sb, *qid);
qid              2763 fs/quota/dquot.c int dquot_set_dqblk(struct super_block *sb, struct kqid qid,
qid              2769 fs/quota/dquot.c 	dquot = dqget(sb, qid);
qid               120 fs/quota/kqid.c bool qid_valid(struct kqid qid)
qid               122 fs/quota/kqid.c 	switch (qid.type) {
qid               124 fs/quota/kqid.c 		return uid_valid(qid.uid);
qid               126 fs/quota/kqid.c 		return gid_valid(qid.gid);
qid               128 fs/quota/kqid.c 		return projid_valid(qid.projid);
qid                37 fs/quota/netlink.c void quota_send_warning(struct kqid qid, dev_t dev,
qid                63 fs/quota/netlink.c 	ret = nla_put_u32(skb, QUOTA_NL_A_QTYPE, qid.type);
qid                67 fs/quota/netlink.c 				from_kqid_munged(&init_user_ns, qid),
qid               202 fs/quota/quota.c 	struct kqid qid;
qid               209 fs/quota/quota.c 	qid = make_kqid(current_user_ns(), type, id);
qid               210 fs/quota/quota.c 	if (!qid_has_mapping(sb->s_user_ns, qid))
qid               212 fs/quota/quota.c 	ret = sb->s_qcop->get_dqblk(sb, qid, &fdq);
qid               228 fs/quota/quota.c 	struct kqid qid;
qid               235 fs/quota/quota.c 	qid = make_kqid(current_user_ns(), type, id);
qid               236 fs/quota/quota.c 	if (!qid_has_mapping(sb->s_user_ns, qid))
qid               238 fs/quota/quota.c 	ret = sb->s_qcop->get_nextdqblk(sb, &qid, &fdq);
qid               243 fs/quota/quota.c 	idq.dqb_id = from_kqid(current_user_ns(), qid);
qid               280 fs/quota/quota.c 	struct kqid qid;
qid               286 fs/quota/quota.c 	qid = make_kqid(current_user_ns(), type, id);
qid               287 fs/quota/quota.c 	if (!qid_has_mapping(sb->s_user_ns, qid))
qid               290 fs/quota/quota.c 	return sb->s_qcop->set_dqblk(sb, qid, &fdq);
qid               565 fs/quota/quota.c 	struct kqid qid;
qid               571 fs/quota/quota.c 	qid = make_kqid(current_user_ns(), type, id);
qid               572 fs/quota/quota.c 	if (!qid_has_mapping(sb->s_user_ns, qid))
qid               575 fs/quota/quota.c 	if (from_kqid(sb->s_user_ns, qid) == 0 &&
qid               590 fs/quota/quota.c 	return sb->s_qcop->set_dqblk(sb, qid, &qdq);
qid               627 fs/quota/quota.c 	struct kqid qid;
qid               632 fs/quota/quota.c 	qid = make_kqid(current_user_ns(), type, id);
qid               633 fs/quota/quota.c 	if (!qid_has_mapping(sb->s_user_ns, qid))
qid               635 fs/quota/quota.c 	ret = sb->s_qcop->get_dqblk(sb, qid, &qdq);
qid               653 fs/quota/quota.c 	struct kqid qid;
qid               659 fs/quota/quota.c 	qid = make_kqid(current_user_ns(), type, id);
qid               660 fs/quota/quota.c 	if (!qid_has_mapping(sb->s_user_ns, qid))
qid               662 fs/quota/quota.c 	ret = sb->s_qcop->get_nextdqblk(sb, &qid, &qdq);
qid               665 fs/quota/quota.c 	id_out = from_kqid(current_user_ns(), qid);
qid                36 fs/quota/quota_tree.c static int get_index(struct qtree_mem_dqinfo *info, struct kqid qid, int depth)
qid                38 fs/quota/quota_tree.c 	qid_t id = from_kqid(&init_user_ns, qid);
qid               723 fs/quota/quota_tree.c int qtree_get_next_id(struct qtree_mem_dqinfo *info, struct kqid *qid)
qid               725 fs/quota/quota_tree.c 	qid_t id = from_kqid(&init_user_ns, *qid);
qid               731 fs/quota/quota_tree.c 	*qid = make_kqid(&init_user_ns, qid->type, id);
qid               363 fs/quota/quota_v2.c static int v2_get_next_id(struct super_block *sb, struct kqid *qid)
qid               369 fs/quota/quota_v2.c 	ret = qtree_get_next_id(sb_dqinfo(sb, qid->type)->dqi_priv, qid);
qid               220 fs/xfs/xfs_quotaops.c 	struct kqid		qid,
qid               231 fs/xfs/xfs_quotaops.c 	id = from_kqid(&init_user_ns, qid);
qid               232 fs/xfs/xfs_quotaops.c 	return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq);
qid               239 fs/xfs/xfs_quotaops.c 	struct kqid		*qid,
qid               251 fs/xfs/xfs_quotaops.c 	id = from_kqid(&init_user_ns, *qid);
qid               252 fs/xfs/xfs_quotaops.c 	ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type),
qid               258 fs/xfs/xfs_quotaops.c 	*qid = make_kqid(current_user_ns(), qid->type, id);
qid               265 fs/xfs/xfs_quotaops.c 	struct kqid		qid,
qid               277 fs/xfs/xfs_quotaops.c 	return xfs_qm_scall_setqlim(mp, from_kqid(&init_user_ns, qid),
qid               278 fs/xfs/xfs_quotaops.c 				     xfs_quota_type(qid.type), qdq);
qid                57 include/linux/dqblk_qtree.h int qtree_get_next_id(struct qtree_mem_dqinfo *info, struct kqid *qid);
qid               261 include/linux/firmware/xlnx-zynqmp.h 	u32 qid;
qid               207 include/linux/nvme-fc.h 	__be16	qid;
qid                58 include/linux/nvme-rdma.h 	__le16		qid;
qid               973 include/linux/nvme.h 	__le16			qid;
qid              1072 include/linux/nvme.h #define show_opcode_name(qid, opcode, fctype)			\
qid              1075 include/linux/nvme.h 	((qid) ?						\
qid              1148 include/linux/nvme.h 	__le16		qid;
qid                89 include/linux/qed/qed_eth_if.h 	u16 qid;
qid                79 include/linux/quota.h extern qid_t from_kqid(struct user_namespace *to, struct kqid qid);
qid                80 include/linux/quota.h extern qid_t from_kqid_munged(struct user_namespace *to, struct kqid qid);
qid                81 include/linux/quota.h extern bool qid_valid(struct kqid qid);
qid                98 include/linux/quota.h 				    enum quota_type type, qid_t qid)
qid               105 include/linux/quota.h 		kqid.uid = make_kuid(from, qid);
qid               108 include/linux/quota.h 		kqid.gid = make_kgid(from, qid);
qid               111 include/linux/quota.h 		kqid.projid = make_kprojid(from, qid);
qid               187 include/linux/quota.h static inline bool qid_has_mapping(struct user_namespace *ns, struct kqid qid)
qid               189 include/linux/quota.h 	return from_kqid(ns, qid) != (qid_t) -1;
qid               318 include/linux/quota.h 	int (*get_next_id)(struct super_block *sb, struct kqid *qid);	/* Get next ID with existing structure in the quota file */
qid               337 include/linux/quota.h 	int (*get_next_id) (struct super_block *sb, struct kqid *qid);
qid               508 include/linux/quota.h extern void quota_send_warning(struct kqid qid, dev_t dev,
qid               511 include/linux/quota.h static inline void quota_send_warning(struct kqid qid, dev_t dev,
qid                48 include/linux/quotaops.h struct dquot *dqget(struct super_block *sb, struct kqid qid);
qid                95 include/linux/quotaops.h int dquot_get_next_id(struct super_block *sb, struct kqid *qid);
qid               388 include/net/9p/9p.h 	struct p9_qid qid;
qid               405 include/net/9p/9p.h 	struct p9_qid qid;
qid               150 include/net/9p/client.h 	struct p9_qid qid;
qid               168 include/net/9p/client.h 	struct p9_qid qid;
qid               195 include/net/9p/client.h 		kgid_t gid, struct p9_qid *qid);
qid               197 include/net/9p/client.h 		kgid_t gid, struct p9_qid *qid);
qid               898 net/9p/client.c 	memset(&fid->qid, 0, sizeof(struct p9_qid));
qid              1122 net/9p/client.c 	struct p9_qid qid;
qid              1141 net/9p/client.c 	err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", &qid);
qid              1149 net/9p/client.c 		 qid.type, (unsigned long long)qid.path, qid.version);
qid              1151 net/9p/client.c 	memmove(&fid->qid, &qid, sizeof(struct p9_qid));
qid              1220 net/9p/client.c 		memmove(&fid->qid, &wqids[nwqids - 1], sizeof(struct p9_qid));
qid              1222 net/9p/client.c 		fid->qid = oldfid->qid;
qid              1245 net/9p/client.c 	struct p9_qid qid;
qid              1265 net/9p/client.c 	err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", &qid, &iounit);
qid              1272 net/9p/client.c 		p9_is_proto_dotl(clnt) ? "RLOPEN" : "ROPEN",  qid.type,
qid              1273 net/9p/client.c 		(unsigned long long)qid.path, qid.version, iounit);
qid              1286 net/9p/client.c 		kgid_t gid, struct p9_qid *qid)
qid              1309 net/9p/client.c 	err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", qid, &iounit);
qid              1316 net/9p/client.c 			qid->type,
qid              1317 net/9p/client.c 			(unsigned long long)qid->path,
qid              1318 net/9p/client.c 			qid->version, iounit);
qid              1336 net/9p/client.c 	struct p9_qid qid;
qid              1354 net/9p/client.c 	err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", &qid, &iounit);
qid              1361 net/9p/client.c 				qid.type,
qid              1362 net/9p/client.c 				(unsigned long long)qid.path,
qid              1363 net/9p/client.c 				qid.version, iounit);
qid              1376 net/9p/client.c 		const char *symtgt, kgid_t gid, struct p9_qid *qid)
qid              1393 net/9p/client.c 	err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", qid);
qid              1400 net/9p/client.c 			qid->type, (unsigned long long)qid->path, qid->version);
qid              1720 net/9p/client.c 		ret->size, ret->type, ret->dev, ret->qid.type,
qid              1721 net/9p/client.c 		(unsigned long long)ret->qid.path, ret->qid.version, ret->mode,
qid              1779 net/9p/client.c 		ret->st_result_mask, ret->qid.type, ret->qid.path,
qid              1780 net/9p/client.c 		ret->qid.version, ret->st_mode, ret->st_nlink,
qid              1842 net/9p/client.c 		wst->size, wst->type, wst->dev, wst->qid.type,
qid              1843 net/9p/client.c 		(unsigned long long)wst->qid.path, wst->qid.version, wst->mode,
qid              2135 net/9p/client.c 			dev_t rdev, kgid_t gid, struct p9_qid *qid)
qid              2150 net/9p/client.c 	err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", qid);
qid              2155 net/9p/client.c 	p9_debug(P9_DEBUG_9P, "<<< RMKNOD qid %x.%llx.%x\n", qid->type,
qid              2156 net/9p/client.c 				(unsigned long long)qid->path, qid->version);
qid              2166 net/9p/client.c 				kgid_t gid, struct p9_qid *qid)
qid              2181 net/9p/client.c 	err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", qid);
qid              2186 net/9p/client.c 	p9_debug(P9_DEBUG_9P, "<<< RMKDIR qid %x.%llx.%x\n", qid->type,
qid              2187 net/9p/client.c 				(unsigned long long)qid->path, qid->version);
qid               181 net/9p/protocol.c 				struct p9_qid *qid =
qid               185 net/9p/protocol.c 						      &qid->type, &qid->version,
qid               186 net/9p/protocol.c 						      &qid->path);
qid               201 net/9p/protocol.c 						&stbuf->dev, &stbuf->qid,
qid               315 net/9p/protocol.c 					&stbuf->qid,
qid               411 net/9p/protocol.c 				const struct p9_qid *qid =
qid               415 net/9p/protocol.c 						 qid->type, qid->version,
qid               416 net/9p/protocol.c 						 qid->path);
qid               425 net/9p/protocol.c 						 stbuf->dev, &stbuf->qid,
qid               606 net/9p/protocol.c 	ret = p9pdu_readf(&fake_pdu, clnt->proto_version, "Qqbs", &dirent->qid,
qid              1988 net/sched/sch_api.c 	u32 qid;
qid              2021 net/sched/sch_api.c 	qid = TC_H_MAJ(clid);
qid              2026 net/sched/sch_api.c 		if (qid && qid1) {
qid              2028 net/sched/sch_api.c 			if (qid != qid1)
qid              2031 net/sched/sch_api.c 			qid = qid1;
qid              2032 net/sched/sch_api.c 		} else if (qid == 0)
qid              2033 net/sched/sch_api.c 			qid = dev->qdisc->handle;
qid              2041 net/sched/sch_api.c 			portid = TC_H_MAKE(qid, portid);
qid              2043 net/sched/sch_api.c 		if (qid == 0)
qid              2044 net/sched/sch_api.c 			qid = dev->qdisc->handle;
qid              2048 net/sched/sch_api.c 	q = qdisc_lookup(dev, qid);
qid              2060 net/sched/sch_api.c 			clid = qid;
qid              2062 net/sched/sch_api.c 		clid = TC_H_MAKE(qid, clid);
qid               616 net/xdp/xsk.c  	u32 flags, qid;
qid               647 net/xdp/xsk.c  	qid = sxdp->sxdp_queue_id;
qid               678 net/xdp/xsk.c  		if (umem_xs->dev != dev || umem_xs->queue_id != qid) {
qid               697 net/xdp/xsk.c  		err = xdp_umem_assign_dev(xs->umem, dev, qid, flags);
qid               706 net/xdp/xsk.c  	xs->queue_id = qid;