/linux-4.1.27/drivers/s390/char/ |
D | tape_std.c | 37 struct tape_request * request; in tape_std_assign_timeout() local 41 request = (struct tape_request *) data; in tape_std_assign_timeout() 42 device = request->device; in tape_std_assign_timeout() 47 rc = tape_cancel_io(device, request); in tape_std_assign_timeout() 58 struct tape_request *request; in tape_std_assign() local 60 request = tape_alloc_request(2, 11); in tape_std_assign() 61 if (IS_ERR(request)) in tape_std_assign() 62 return PTR_ERR(request); in tape_std_assign() 64 request->op = TO_ASSIGN; in tape_std_assign() 65 tape_ccw_cc(request->cpaddr, ASSIGN, 11, request->cpdata); in tape_std_assign() [all …]
|
D | tape_34xx.c | 55 static void __tape_34xx_medium_sense(struct tape_request *request) in __tape_34xx_medium_sense() argument 57 struct tape_device *device = request->device; in __tape_34xx_medium_sense() 60 if (request->rc == 0) { in __tape_34xx_medium_sense() 61 sense = request->cpdata; in __tape_34xx_medium_sense() 80 request->rc); in __tape_34xx_medium_sense() 81 tape_free_request(request); in __tape_34xx_medium_sense() 86 struct tape_request *request; in tape_34xx_medium_sense() local 89 request = tape_alloc_request(1, 32); in tape_34xx_medium_sense() 90 if (IS_ERR(request)) { in tape_34xx_medium_sense() 92 return PTR_ERR(request); in tape_34xx_medium_sense() [all …]
|
D | tape_3590.c | 203 struct tape_request *request; in tape_3592_kekl_query() local 212 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query() 213 if (IS_ERR(request)) { in tape_3592_kekl_query() 214 rc = PTR_ERR(request); in tape_3592_kekl_query() 217 order = request->cpdata; in tape_3592_kekl_query() 221 request->op = TO_KEKL_QUERY; in tape_3592_kekl_query() 222 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query() 223 tape_ccw_end(request->cpaddr + 1, READ_SS_DATA, sizeof(*int_kekls), in tape_3592_kekl_query() 225 rc = tape_do_io(device, request); in tape_3592_kekl_query() 232 tape_free_request(request); in tape_3592_kekl_query() [all …]
|
D | tape_core.c | 291 __tape_cancel_io(struct tape_device *device, struct tape_request *request) in __tape_cancel_io() argument 297 if (request->callback == NULL) in __tape_cancel_io() 302 rc = ccw_device_clear(device->cdev, (long) request); in __tape_cancel_io() 306 request->status = TAPE_REQUEST_DONE; in __tape_cancel_io() 309 request->status = TAPE_REQUEST_CANCEL; in __tape_cancel_io() 641 struct tape_request * request; in __tape_discard_requests() local 645 request = list_entry(l, struct tape_request, list); in __tape_discard_requests() 646 if (request->status == TAPE_REQUEST_IN_IO) in __tape_discard_requests() 647 request->status = TAPE_REQUEST_DONE; in __tape_discard_requests() 648 list_del(&request->list); in __tape_discard_requests() [all …]
|
D | sclp_vt220.c | 97 static int __sclp_vt220_emit(struct sclp_vt220_request *request); 118 sclp_vt220_process_queue(struct sclp_vt220_request *request) in sclp_vt220_process_queue() argument 125 page = request->sclp_req.sccb; in sclp_vt220_process_queue() 128 list_del(&request->list); in sclp_vt220_process_queue() 131 request = NULL; in sclp_vt220_process_queue() 133 request = list_entry(sclp_vt220_outqueue.next, in sclp_vt220_process_queue() 135 if (!request || sclp_vt220_suspended) { in sclp_vt220_process_queue() 141 } while (__sclp_vt220_emit(request)); in sclp_vt220_process_queue() 142 if (request == NULL && sclp_vt220_flush_later) in sclp_vt220_process_queue() 154 sclp_vt220_callback(struct sclp_req *request, void *data) in sclp_vt220_callback() argument [all …]
|
D | sclp_sdias.c | 53 static void sdias_callback(struct sclp_req *request, void *data) in sdias_callback() argument 107 struct sclp_req request; in sclp_sdias_blk_count() local 113 memset(&request, 0, sizeof(request)); in sclp_sdias_blk_count() 123 request.sccb = &sccb; in sclp_sdias_blk_count() 124 request.command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_sdias_blk_count() 125 request.status = SCLP_REQ_FILLED; in sclp_sdias_blk_count() 126 request.callback = sdias_callback; in sclp_sdias_blk_count() 128 rc = sdias_sclp_send(&request); in sclp_sdias_blk_count() 166 struct sclp_req request; in sclp_sdias_copy() local 172 memset(&request, 0, sizeof(request)); in sclp_sdias_copy() [all …]
|
D | sclp_async.c | 25 static struct sclp_req *request; variable 132 request->command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_async_send_wait() 133 request->sccb = sccb; in sclp_async_send_wait() 134 request->status = SCLP_REQ_FILLED; in sclp_async_send_wait() 145 rc = sclp_add_request(request); in sclp_async_send_wait() 149 while (request->status != SCLP_REQ_DONE && in sclp_async_send_wait() 150 request->status != SCLP_REQ_FAILED) { in sclp_async_send_wait() 154 if (request->status != SCLP_REQ_DONE) in sclp_async_send_wait() 157 request->sccb)->header.response_code; in sclp_async_send_wait() 179 request = kzalloc(sizeof(struct sclp_req), GFP_KERNEL); in sclp_async_init() [all …]
|
D | tape_char.c | 131 struct tape_request *request; in tapechar_read() local 165 request = device->discipline->read_block(device, block_size); in tapechar_read() 166 if (IS_ERR(request)) in tapechar_read() 167 return PTR_ERR(request); in tapechar_read() 169 rc = tape_do_io(device, request); in tapechar_read() 171 rc = block_size - request->rescnt; in tapechar_read() 178 tape_free_request(request); in tapechar_read() 189 struct tape_request *request; in tapechar_write() local 218 request = device->discipline->write_block(device, block_size); in tapechar_write() 219 if (IS_ERR(request)) in tapechar_write() [all …]
|
D | sclp_rw.c | 377 sclp_writedata_callback(struct sclp_req *request, void *data) in sclp_writedata_callback() argument 386 if (request->status == SCLP_REQ_FAILED) { in sclp_writedata_callback() 407 buffer->request.status = SCLP_REQ_FILLED; in sclp_writedata_callback() 408 rc = sclp_add_request(request); in sclp_writedata_callback() 423 buffer->request.status = SCLP_REQ_FILLED; in sclp_writedata_callback() 424 rc = sclp_add_request(request); in sclp_writedata_callback() 462 buffer->request.command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_emit_buffer() 463 buffer->request.status = SCLP_REQ_FILLED; in sclp_emit_buffer() 464 buffer->request.callback = sclp_writedata_callback; in sclp_emit_buffer() 465 buffer->request.callback_data = buffer; in sclp_emit_buffer() [all …]
|
D | sclp_cmd.c | 46 struct sclp_req *request; in sclp_sync_request_timeout() local 49 request = kzalloc(sizeof(*request), GFP_KERNEL); in sclp_sync_request_timeout() 50 if (!request) in sclp_sync_request_timeout() 53 request->queue_timeout = timeout; in sclp_sync_request_timeout() 54 request->command = cmd; in sclp_sync_request_timeout() 55 request->sccb = sccb; in sclp_sync_request_timeout() 56 request->status = SCLP_REQ_FILLED; in sclp_sync_request_timeout() 57 request->callback = sclp_sync_callback; in sclp_sync_request_timeout() 58 request->callback_data = &completion; in sclp_sync_request_timeout() 62 rc = sclp_add_request(request); in sclp_sync_request_timeout() [all …]
|
D | tape.h | 242 tape_do_io_free(struct tape_device *device, struct tape_request *request) in tape_do_io_free() argument 246 rc = tape_do_io(device, request); in tape_do_io_free() 247 tape_free_request(request); in tape_do_io_free() 252 tape_do_io_async_free(struct tape_device *device, struct tape_request *request) in tape_do_io_async_free() argument 254 request->callback = (void *) tape_free_request; in tape_do_io_async_free() 255 request->callback_data = NULL; in tape_do_io_async_free() 256 tape_do_io_async(device, request); in tape_do_io_async_free()
|
D | tape_proc.c | 41 struct tape_request *request; in tape_proc_show() local 71 request = list_entry(device->req_queue.next, in tape_proc_show() 73 str = tape_op_verbose[request->op]; in tape_proc_show()
|
/linux-4.1.27/include/linux/ |
D | elevator.h | 12 typedef int (elevator_merge_fn) (struct request_queue *, struct request **, 15 typedef void (elevator_merge_req_fn) (struct request_queue *, struct request *, struct request *); 17 typedef void (elevator_merged_fn) (struct request_queue *, struct request *, int); 19 typedef int (elevator_allow_merge_fn) (struct request_queue *, struct request *, struct bio *); 22 struct request *, struct bio *); 26 typedef void (elevator_add_req_fn) (struct request_queue *, struct request *); 27 typedef struct request *(elevator_request_list_fn) (struct request_queue *, struct request *); 28 typedef void (elevator_completed_req_fn) (struct request_queue *, struct request *); 33 typedef int (elevator_set_req_fn) (struct request_queue *, struct request *, 35 typedef void (elevator_put_req_fn) (struct request *); [all …]
|
D | blk-mq.h | 82 struct request *rq; 89 typedef enum blk_eh_timer_return (timeout_fn)(struct request *, bool); 92 typedef int (init_request_fn)(void *, struct request *, unsigned int, 94 typedef void (exit_request_fn)(void *, struct request *, unsigned int, 97 typedef void (busy_iter_fn)(struct blk_mq_hw_ctx *, struct request *, void *, 178 void blk_mq_insert_request(struct request *, bool, bool, bool); 179 void blk_mq_free_request(struct request *rq); 180 void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *, struct request *rq); 182 struct request *blk_mq_alloc_request(struct request_queue *q, int rw, 184 struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag); [all …]
|
D | blkdev.h | 35 struct request; 50 struct request; 51 typedef void (rq_end_io_fn)(struct request *, int); 100 struct request { struct 211 struct request *next_rq; argument 214 static inline unsigned short req_get_ioprio(struct request *req) in req_get_ioprio() 238 typedef int (prep_rq_fn) (struct request_queue *, struct request *); 239 typedef void (unprep_rq_fn) (struct request_queue *, struct request *); 250 typedef void (softirq_done_fn)(struct request *); 251 typedef int (dma_drain_needed_fn)(struct request *); [all …]
|
D | bsg-lib.h | 28 struct request; 41 struct request *req; 44 void *request; member
|
D | ide.h | 309 struct request *rq; /* copy of request */ 337 struct request *rq; 367 ide_startstop_t (*do_request)(struct ide_drive_s *, struct request *, 488 struct request *rq; /* current request */ 569 struct request sense_rq; 697 void (*rw_disk)(ide_drive_t *, struct request *); 754 struct request *rq; 1002 void ide_complete_power_step(ide_drive_t *, struct request *); 1003 ide_startstop_t ide_start_power_step(ide_drive_t *, struct request *); 1004 void ide_complete_pm_rq(ide_drive_t *, struct request *); [all …]
|
D | blktrace_api.h | 60 extern void blk_add_driver_data(struct request_queue *q, struct request *rq, 106 static inline int blk_cmd_buf_len(struct request *rq) in blk_cmd_buf_len() 111 extern void blk_dump_cmd(char *buf, struct request *rq);
|
D | device-mapper.h | 49 typedef int (*dm_map_request_fn) (struct dm_target *ti, struct request *clone, 52 struct request *rq, 54 struct request **clone); 55 typedef void (*dm_release_clone_request_fn) (struct request *clone); 68 struct request *clone, int error, 410 union map_info *dm_get_rq_mapinfo(struct request *rq);
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
D | radeon_ioc32.c | 137 drm_radeon_stipple_t __user *request; in compat_radeon_cp_stipple() local 143 request = compat_alloc_user_space(sizeof(*request)); in compat_radeon_cp_stipple() 144 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_radeon_cp_stipple() 146 &request->mask)) in compat_radeon_cp_stipple() 149 return drm_ioctl(file, DRM_IOCTL_RADEON_STIPPLE, (unsigned long)request); in compat_radeon_cp_stipple() 171 drm_radeon_texture_t __user *request; in compat_radeon_cp_texture() local 183 request = compat_alloc_user_space(sizeof(*request) + sizeof(*image)); in compat_radeon_cp_texture() 184 if (!access_ok(VERIFY_WRITE, request, in compat_radeon_cp_texture() 185 sizeof(*request) + sizeof(*image))) in compat_radeon_cp_texture() 187 image = (drm_radeon_tex_image_t __user *) (request + 1); in compat_radeon_cp_texture() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/ |
D | niobuf.c | 488 int ptl_send_rpc(struct ptlrpc_request *request, int noreply) in ptl_send_rpc() argument 496 struct obd_device *obd = request->rq_import->imp_obd; in ptl_send_rpc() 501 LASSERT(request->rq_type == PTL_RPC_MSG_REQUEST); in ptl_send_rpc() 502 LASSERT(request->rq_wait_ctx == 0); in ptl_send_rpc() 506 LASSERT(!request->rq_receiving_reply); in ptl_send_rpc() 507 LASSERT(!((lustre_msg_get_flags(request->rq_reqmsg) & MSG_REPLAY) && in ptl_send_rpc() 508 (request->rq_import->imp_state == LUSTRE_IMP_FULL))); in ptl_send_rpc() 514 spin_lock(&request->rq_lock); in ptl_send_rpc() 515 request->rq_err = 1; in ptl_send_rpc() 516 spin_unlock(&request->rq_lock); in ptl_send_rpc() [all …]
|
D | client.c | 530 struct ptlrpc_request *request; in ptlrpc_prep_req_from_pool() local 547 request = list_entry(pool->prp_req_list.next, struct ptlrpc_request, in ptlrpc_prep_req_from_pool() 549 list_del_init(&request->rq_list); in ptlrpc_prep_req_from_pool() 552 LASSERT(request->rq_reqbuf); in ptlrpc_prep_req_from_pool() 553 LASSERT(request->rq_pool); in ptlrpc_prep_req_from_pool() 555 reqbuf = request->rq_reqbuf; in ptlrpc_prep_req_from_pool() 556 memset(request, 0, sizeof(*request)); in ptlrpc_prep_req_from_pool() 557 request->rq_reqbuf = reqbuf; in ptlrpc_prep_req_from_pool() 558 request->rq_reqbuf_len = pool->prp_rq_size; in ptlrpc_prep_req_from_pool() 559 request->rq_pool = pool; in ptlrpc_prep_req_from_pool() [all …]
|
D | import.c | 108 struct ptlrpc_request *request, 638 struct ptlrpc_request *request; in ptlrpc_connect_import() local 697 request = ptlrpc_request_alloc(imp, &RQF_MDS_CONNECT); in ptlrpc_connect_import() 698 if (request == NULL) { in ptlrpc_connect_import() 703 rc = ptlrpc_request_bufs_pack(request, LUSTRE_OBD_VERSION, in ptlrpc_connect_import() 706 ptlrpc_request_free(request); in ptlrpc_connect_import() 712 lustre_msg_set_service_time(request->rq_reqmsg, in ptlrpc_connect_import() 713 at_timeout2est(request->rq_timeout)); in ptlrpc_connect_import() 720 request->rq_timeout = INITIAL_CONNECT_TIMEOUT; in ptlrpc_connect_import() 721 lustre_msg_set_timeout(request->rq_reqmsg, request->rq_timeout); in ptlrpc_connect_import() [all …]
|
D | service.c | 1872 struct ptlrpc_request *request; in ptlrpc_server_handle_request() local 1879 request = ptlrpc_server_request_get(svcpt, false); in ptlrpc_server_handle_request() 1880 if (request == NULL) in ptlrpc_server_handle_request() 1889 if (request->rq_export && request->rq_ops) in ptlrpc_server_handle_request() 1893 ptlrpc_rqphase_move(request, RQ_PHASE_INTERPRET); in ptlrpc_server_handle_request() 1899 timediff = cfs_timeval_sub(&work_start, &request->rq_arrival_time, in ptlrpc_server_handle_request() 1912 rc = lu_context_init(&request->rq_session, LCT_SESSION | LCT_NOREF); in ptlrpc_server_handle_request() 1917 request->rq_session.lc_thread = thread; in ptlrpc_server_handle_request() 1918 request->rq_session.lc_cookie = 0x5; in ptlrpc_server_handle_request() 1919 lu_context_enter(&request->rq_session); in ptlrpc_server_handle_request() [all …]
|
/linux-4.1.27/drivers/usb/musb/ |
D | musb_gadget.c | 56 static inline void map_dma_buffer(struct musb_request *request, in map_dma_buffer() argument 62 request->map_state = UN_MAPPED; in map_dma_buffer() 73 musb_ep->packet_sz, request->request.buf, in map_dma_buffer() 74 request->request.length); in map_dma_buffer() 78 if (request->request.dma == DMA_ADDR_INVALID) { in map_dma_buffer() 84 request->request.buf, in map_dma_buffer() 85 request->request.length, in map_dma_buffer() 86 request->tx in map_dma_buffer() 93 request->request.dma = dma_addr; in map_dma_buffer() 94 request->map_state = MUSB_MAPPED; in map_dma_buffer() [all …]
|
D | musb_gadget_ep0.c | 261 struct musb_request *request; in service_zero_data_request() local 304 request = next_request(musb_ep); in service_zero_data_request() 305 if (!musb_ep->busy && request) { in service_zero_data_request() 307 musb_ep_restart(musb, request); in service_zero_data_request() 488 struct musb_request *request; in ep0_rxstate() local 492 request = next_ep0_request(musb); in ep0_rxstate() 493 req = &request->request; in ep0_rxstate() 546 struct usb_request *request; in ep0_txstate() local 557 request = &req->request; in ep0_txstate() 560 fifo_src = (u8 *) request->buf + request->actual; in ep0_txstate() [all …]
|
/linux-4.1.27/drivers/gpu/drm/ |
D | drm_ioc32.c | 441 struct drm_buf_info __user *request; in compat_drm_infobufs() local 458 nbytes = sizeof(*request) + count * sizeof(struct drm_buf_desc); in compat_drm_infobufs() 459 request = compat_alloc_user_space(nbytes); in compat_drm_infobufs() 460 if (!access_ok(VERIFY_WRITE, request, nbytes)) in compat_drm_infobufs() 462 list = (struct drm_buf_desc *) (request + 1); in compat_drm_infobufs() 464 if (__put_user(count, &request->count) in compat_drm_infobufs() 465 || __put_user(list, &request->list)) in compat_drm_infobufs() 468 err = drm_ioctl(file, DRM_IOCTL_INFO_BUFS, (unsigned long)request); in compat_drm_infobufs() 472 if (__get_user(actual, &request->count)) in compat_drm_infobufs() 505 struct drm_buf_map __user *request; in compat_drm_mapbufs() local [all …]
|
D | drm_agpsupport.c | 199 int drm_agp_alloc(struct drm_device *dev, struct drm_agp_buffer *request) in drm_agp_alloc() argument 211 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_agp_alloc() 212 type = (u32) request->type; in drm_agp_alloc() 224 request->handle = entry->handle; in drm_agp_alloc() 225 request->physical = memory->physical; in drm_agp_alloc() 235 struct drm_agp_buffer *request = data; in drm_agp_alloc_ioctl() local 237 return drm_agp_alloc(dev, request); in drm_agp_alloc_ioctl() 273 int drm_agp_unbind(struct drm_device *dev, struct drm_agp_binding *request) in drm_agp_unbind() argument 280 if (!(entry = drm_agp_lookup_entry(dev, request->handle))) in drm_agp_unbind() 295 struct drm_agp_binding *request = data; in drm_agp_unbind_ioctl() local [all …]
|
D | drm_bufs.c | 515 struct drm_map *request = data; in drm_legacy_rmmap_ioctl() local 523 r_list->user_token == (unsigned long)request->handle && in drm_legacy_rmmap_ioctl() 598 struct drm_buf_desc *request) in drm_legacy_addbufs_agp() argument 619 count = request->count; in drm_legacy_addbufs_agp() 620 order = order_base_2(request->size); in drm_legacy_addbufs_agp() 623 alignment = (request->flags & _DRM_PAGE_ALIGN) in drm_legacy_addbufs_agp() 629 agp_offset = dev->agp->base + request->agp_start; in drm_legacy_addbufs_agp() 750 request->count = entry->buf_count; in drm_legacy_addbufs_agp() 751 request->size = size; in drm_legacy_addbufs_agp() 762 struct drm_buf_desc *request) in drm_legacy_addbufs_pci() argument [all …]
|
D | drm_context.c | 160 struct drm_ctx_priv_map *request = data; in drm_legacy_getsareactx() local 166 map = idr_find(&dev->ctx_idr, request->ctx_id); in drm_legacy_getsareactx() 172 request->handle = NULL; in drm_legacy_getsareactx() 175 request->handle = in drm_legacy_getsareactx() 183 if (request->handle == NULL) in drm_legacy_getsareactx() 204 struct drm_ctx_priv_map *request = data; in drm_legacy_setsareactx() local 211 && r_list->user_token == (unsigned long) request->handle) in drm_legacy_setsareactx() 223 if (IS_ERR(idr_replace(&dev->ctx_idr, map, request->ctx_id))) in drm_legacy_setsareactx()
|
D | drm_scatter.c | 85 struct drm_scatter_gather *request = data; in drm_legacy_sg_alloc() local 104 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_legacy_sg_alloc() 105 DRM_DEBUG("size=%ld pages=%ld\n", request->size, pages); in drm_legacy_sg_alloc() 147 request->handle = entry->handle; in drm_legacy_sg_alloc() 201 struct drm_scatter_gather *request = data; in drm_legacy_sg_free() local 213 if (!entry || entry->handle != request->handle) in drm_legacy_sg_free()
|
D | drm_irq.c | 1478 e->event.user_data = vblwait->request.signal; in drm_queue_vblank_event() 1505 if ((vblwait->request.type & _DRM_VBLANK_NEXTONMISS) && in drm_queue_vblank_event() 1506 (seq - vblwait->request.sequence) <= (1 << 23)) { in drm_queue_vblank_event() 1507 vblwait->request.sequence = seq + 1; in drm_queue_vblank_event() 1508 vblwait->reply.sequence = vblwait->request.sequence; in drm_queue_vblank_event() 1512 vblwait->request.sequence, seq, pipe); in drm_queue_vblank_event() 1515 vblwait->request.sequence); in drm_queue_vblank_event() 1517 e->event.sequence = vblwait->request.sequence; in drm_queue_vblank_event() 1518 if ((seq - vblwait->request.sequence) <= (1 << 23)) { in drm_queue_vblank_event() 1525 vblwait->reply.sequence = vblwait->request.sequence; in drm_queue_vblank_event() [all …]
|
/linux-4.1.27/block/ |
D | blk.h | 24 struct request *flush_rq; 58 void init_request_from_bio(struct request *req, struct bio *bio); 59 void blk_rq_bio_prep(struct request_queue *q, struct request *rq, 61 int blk_rq_append_bio(struct request_queue *q, struct request *rq, 65 void blk_dequeue_request(struct request *rq); 67 bool __blk_end_bidi_request(struct request *rq, int error, 72 void blk_add_timer(struct request *req); 73 void blk_delete_timer(struct request *); 76 bool bio_attempt_front_merge(struct request_queue *q, struct request *req, 78 bool bio_attempt_back_merge(struct request_queue *q, struct request *req, [all …]
|
D | noop-iosched.c | 15 static void noop_merged_requests(struct request_queue *q, struct request *rq, in noop_merged_requests() 16 struct request *next) in noop_merged_requests() 26 struct request *rq; in noop_dispatch() 27 rq = list_entry(nd->queue.next, struct request, queuelist); in noop_dispatch() 35 static void noop_add_request(struct request_queue *q, struct request *rq) in noop_add_request() 42 static struct request * 43 noop_former_request(struct request_queue *q, struct request *rq) in noop_former_request() 49 return list_entry(rq->queuelist.prev, struct request, queuelist); in noop_former_request() 52 static struct request * 53 noop_latter_request(struct request_queue *q, struct request *rq) in noop_latter_request() [all …]
|
D | elevator.c | 56 static int elv_iosched_allow_merge(struct request *rq, struct bio *bio) in elv_iosched_allow_merge() 70 bool elv_rq_merge_ok(struct request *rq, struct bio *bio) in elv_rq_merge_ok() 245 static inline void __elv_rqhash_del(struct request *rq) in __elv_rqhash_del() 251 static void elv_rqhash_del(struct request_queue *q, struct request *rq) in elv_rqhash_del() 257 static void elv_rqhash_add(struct request_queue *q, struct request *rq) in elv_rqhash_add() 266 static void elv_rqhash_reposition(struct request_queue *q, struct request *rq) in elv_rqhash_reposition() 272 static struct request *elv_rqhash_find(struct request_queue *q, sector_t offset) in elv_rqhash_find() 276 struct request *rq; in elv_rqhash_find() 297 void elv_rb_add(struct rb_root *root, struct request *rq) in elv_rb_add() 301 struct request *__rq; in elv_rb_add() [all …]
|
D | deadline-iosched.c | 40 struct request *next_rq[2]; 54 static void deadline_move_request(struct deadline_data *, struct request *); 57 deadline_rb_root(struct deadline_data *dd, struct request *rq) in deadline_rb_root() 65 static inline struct request * 66 deadline_latter_request(struct request *rq) in deadline_latter_request() 77 deadline_add_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_add_rq_rb() 85 deadline_del_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_del_rq_rb() 99 deadline_add_request(struct request_queue *q, struct request *rq) in deadline_add_request() 116 static void deadline_remove_request(struct request_queue *q, struct request *rq) in deadline_remove_request() 125 deadline_merge(struct request_queue *q, struct request **req, struct bio *bio) in deadline_merge() [all …]
|
D | blk-merge.c | 89 void blk_recalc_rq_segments(struct request *rq) in blk_recalc_rq_segments() 247 int blk_rq_map_sg(struct request_queue *q, struct request *rq, in blk_rq_map_sg() 287 struct request *req, in ll_new_hw_segment() 312 int ll_back_merge_fn(struct request_queue *q, struct request *req, in ll_back_merge_fn() 330 int ll_front_merge_fn(struct request_queue *q, struct request *req, in ll_front_merge_fn() 352 static bool req_no_special_merge(struct request *req) in req_no_special_merge() 359 static int req_gap_to_prev(struct request *req, struct request *next) in req_gap_to_prev() 367 static int ll_merge_requests_fn(struct request_queue *q, struct request *req, in ll_merge_requests_fn() 368 struct request *next) in ll_merge_requests_fn() 421 void blk_rq_set_mixed_merge(struct request *rq) in blk_rq_set_mixed_merge() [all …]
|
D | blk-flush.c | 97 static unsigned int blk_flush_policy(unsigned int fflags, struct request *rq) in blk_flush_policy() 113 static unsigned int blk_flush_cur_seq(struct request *rq) in blk_flush_cur_seq() 118 static void blk_flush_restore_request(struct request *rq) in blk_flush_restore_request() 132 static bool blk_flush_queue_rq(struct request *rq, bool add_front) in blk_flush_queue_rq() 165 static bool blk_flush_complete_seq(struct request *rq, in blk_flush_complete_seq() 219 static void flush_end_io(struct request *flush_rq, int error) in flush_end_io() 224 struct request *rq, *n; in flush_end_io() 287 struct request *first_rq = in blk_kick_flush() 288 list_first_entry(pending, struct request, flush.list); in blk_kick_flush() 289 struct request *flush_rq = fq->flush_rq; in blk_kick_flush() [all …]
|
D | blk-core.c | 97 void blk_rq_init(struct request_queue *q, struct request *rq) in blk_rq_init() 117 static void req_bio_endio(struct request *rq, struct bio *bio, in req_bio_endio() 135 void blk_dump_rq_flags(struct request *rq, char *msg) in blk_dump_rq_flags() 799 static inline void blk_free_request(struct request_list *rl, struct request *rq) in blk_free_request() 974 static struct request *__get_request(struct request_list *rl, int rw_flags, in __get_request() 978 struct request *rq; in __get_request() 1151 static struct request *get_request(struct request_queue *q, int rw_flags, in get_request() 1157 struct request *rq; in get_request() 1192 static struct request *blk_old_get_request(struct request_queue *q, int rw, in blk_old_get_request() 1195 struct request *rq; in blk_old_get_request() [all …]
|
D | blk-softirq.c | 31 struct request *rq; in blk_done_softirq() 33 rq = list_entry(local_list.next, struct request, ipi_list); in blk_done_softirq() 42 struct request *rq = data; in trigger_softirq() 59 static int raise_blk_irq(int cpu, struct request *rq) in raise_blk_irq() 75 static int raise_blk_irq(int cpu, struct request *rq) in raise_blk_irq() 105 void __blk_complete_request(struct request *req) in __blk_complete_request() 166 void blk_complete_request(struct request *req) in blk_complete_request()
|
D | bsg-lib.c | 55 struct request *req = job->req; in bsg_job_done() 56 struct request *rsp = req->next_rq; in bsg_job_done() 82 static void bsg_softirq_done(struct request *rq) in bsg_softirq_done() 90 static int bsg_map_buffer(struct bsg_buffer *buf, struct request *req) in bsg_map_buffer() 110 static int bsg_create_job(struct device *dev, struct request *req) in bsg_create_job() 112 struct request *rsp = req->next_rq; in bsg_create_job() 127 job->request = req->cmd; in bsg_create_job() 166 struct request *req; in bsg_request_fn()
|
D | blk-mq.c | 183 struct request *rq, unsigned int rw_flags) in blk_mq_rq_ctx_init() 230 static struct request * 233 struct request *rq; in __blk_mq_alloc_request() 253 struct request *blk_mq_alloc_request(struct request_queue *q, int rw, gfp_t gfp, in blk_mq_alloc_request() 258 struct request *rq; in blk_mq_alloc_request() 293 struct blk_mq_ctx *ctx, struct request *rq) in __blk_mq_free_request() 307 void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *hctx, struct request *rq) in blk_mq_free_hctx_request() 317 void blk_mq_free_request(struct request *rq) in blk_mq_free_request() 327 inline void __blk_mq_end_request(struct request *rq, int error) in __blk_mq_end_request() 341 void blk_mq_end_request(struct request *rq, int error) in blk_mq_end_request() [all …]
|
D | blk-timeout.c | 78 void blk_delete_timer(struct request *req) in blk_delete_timer() 83 static void blk_rq_timed_out(struct request *req) in blk_rq_timed_out() 113 static void blk_rq_check_expired(struct request *rq, unsigned long *next_timeout, in blk_rq_check_expired() 134 struct request *rq, *tmp; in blk_rq_timed_out_timer() 157 void blk_abort_request(struct request *req) in blk_abort_request() 188 void blk_add_timer(struct request *req) in blk_add_timer()
|
D | blk-mq.h | 28 void __blk_mq_complete_request(struct request *rq); 32 void blk_mq_clone_flush_request(struct request *flush_rq, 33 struct request *orig_rq); 63 extern void blk_mq_rq_timed_out(struct request *req, bool reserved);
|
D | blk-tag.c | 23 struct request *blk_queue_find_tag(struct request_queue *q, int tag) in blk_queue_find_tag() 91 struct request **tag_index; in init_tag_map() 101 tag_index = kzalloc(depth * sizeof(struct request *), GFP_ATOMIC); in init_tag_map() 206 struct request **tag_index; in blk_queue_resize_tags() 241 memcpy(bqt->tag_index, tag_index, max_depth * sizeof(struct request *)); in blk_queue_resize_tags() 265 void blk_queue_end_tag(struct request_queue *q, struct request *rq) in blk_queue_end_tag() 313 int blk_queue_start_tag(struct request_queue *q, struct request *rq) in blk_queue_start_tag()
|
D | blk-exec.c | 23 static void blk_end_sync_rq(struct request *rq, int error) in blk_end_sync_rq() 52 struct request *rq, int at_head, in blk_execute_rq_nowait() 110 struct request *rq, int at_head) in blk_execute_rq()
|
D | blk-map.c | 12 int blk_rq_append_bio(struct request_queue *q, struct request *rq, in blk_rq_append_bio() 63 int blk_rq_map_user_iov(struct request_queue *q, struct request *rq, in blk_rq_map_user_iov() 121 int blk_rq_map_user(struct request_queue *q, struct request *rq, in blk_rq_map_user() 181 int blk_rq_map_kern(struct request_queue *q, struct request *rq, void *kbuf, in blk_rq_map_kern()
|
/linux-4.1.27/drivers/gpu/drm/i915/ |
D | i915_ioc32.c | 116 drm_i915_irq_emit_t __user *request; in compat_i915_irq_emit() local 121 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_irq_emit() 122 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_irq_emit() 124 &request->irq_seq)) in compat_i915_irq_emit() 128 (unsigned long)request); in compat_i915_irq_emit() 139 drm_i915_getparam_t __user *request; in compat_i915_getparam() local 144 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_getparam() 145 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_getparam() 146 || __put_user(req32.param, &request->param) in compat_i915_getparam() 148 &request->value)) in compat_i915_getparam() [all …]
|
D | intel_lrc.c | 520 struct drm_i915_gem_request *request) in execlists_context_queue() argument 530 if (!request) { in execlists_context_queue() 535 request = kzalloc(sizeof(*request), GFP_KERNEL); in execlists_context_queue() 536 if (request == NULL) in execlists_context_queue() 538 request->ring = ring; in execlists_context_queue() 539 request->ctx = to; in execlists_context_queue() 540 kref_init(&request->ref); in execlists_context_queue() 541 request->uniq = dev_priv->request_uniq++; in execlists_context_queue() 542 i915_gem_context_reference(request->ctx); in execlists_context_queue() 544 i915_gem_request_reference(request); in execlists_context_queue() [all …]
|
D | i915_gem.c | 2333 struct drm_i915_gem_request *request; in __i915_add_request() local 2338 request = ring->outstanding_lazy_request; in __i915_add_request() 2339 if (WARN_ON(request == NULL)) in __i915_add_request() 2343 ringbuf = request->ctx->engine[ring->id].ringbuf; in __i915_add_request() 2356 ret = logical_ring_flush_all_caches(ringbuf, request->ctx); in __i915_add_request() 2370 request->postfix = intel_ring_get_tail(ringbuf); in __i915_add_request() 2373 ret = ring->emit_request(ringbuf, request); in __i915_add_request() 2381 request->tail = intel_ring_get_tail(ringbuf); in __i915_add_request() 2384 request->head = request_start; in __i915_add_request() 2392 request->batch_obj = obj; in __i915_add_request() [all …]
|
/linux-4.1.27/drivers/net/hyperv/ |
D | rndis_filter.c | 80 struct rndis_request *request; in get_rndis_request() local 85 request = kzalloc(sizeof(struct rndis_request), GFP_KERNEL); in get_rndis_request() 86 if (!request) in get_rndis_request() 89 init_completion(&request->wait_event); in get_rndis_request() 91 rndis_msg = &request->request_msg; in get_rndis_request() 95 request->pkt.q_idx = 0; in get_rndis_request() 107 list_add_tail(&request->list_ent, &dev->req_list); in get_rndis_request() 110 return request; in get_rndis_request() 247 struct rndis_request *request) in rndis_set_link_state() argument 252 query_complete = &request->response_msg.msg.query_complete; in rndis_set_link_state() [all …]
|
/linux-4.1.27/sound/pci/mixart/ |
D | mixart_hwdep.c | 145 struct mixart_msg request; in mixart_enum_connectors() local 162 request.message_id = MSG_SYSTEM_ENUM_PLAY_CONNECTOR; in mixart_enum_connectors() 163 request.uid = (struct mixart_uid){0,0}; /* board num = 0 */ in mixart_enum_connectors() 164 request.data = NULL; in mixart_enum_connectors() 165 request.size = 0; in mixart_enum_connectors() 167 err = snd_mixart_send_msg(mgr, &request, sizeof(*connector), connector); in mixart_enum_connectors() 192 request.message_id = MSG_CONNECTOR_GET_AUDIO_INFO; in mixart_enum_connectors() 193 request.uid = connector->uid[k]; in mixart_enum_connectors() 194 request.data = audio_info_req; in mixart_enum_connectors() 195 request.size = sizeof(*audio_info_req); in mixart_enum_connectors() [all …]
|
D | mixart.c | 77 struct mixart_msg request; in mixart_set_pipe_state() local 99 request.message_id = MSG_SYSTEM_WAIT_SYNCHRO_CMD; in mixart_set_pipe_state() 100 request.uid = (struct mixart_uid){0,0}; in mixart_set_pipe_state() 101 request.data = &system_msg_uid; in mixart_set_pipe_state() 102 request.size = sizeof(system_msg_uid); in mixart_set_pipe_state() 104 err = snd_mixart_send_msg_wait_notif(mgr, &request, system_msg_uid); in mixart_set_pipe_state() 118 request.message_id = MSG_STREAM_START_STREAM_GRP_PACKET; in mixart_set_pipe_state() 120 request.message_id = MSG_STREAM_STOP_STREAM_GRP_PACKET; in mixart_set_pipe_state() 122 request.uid = pipe->group_uid; /*(struct mixart_uid){0,0};*/ in mixart_set_pipe_state() 123 request.data = &group_state; in mixart_set_pipe_state() [all …]
|
D | mixart_mixer.c | 306 struct mixart_msg request; in mixart_update_analog_audio_level() local 324 if(is_capture) request.uid = chip->uid_in_analog_physio; in mixart_update_analog_audio_level() 325 else request.uid = chip->uid_out_analog_physio; in mixart_update_analog_audio_level() 326 request.message_id = MSG_PHYSICALIO_SET_LEVEL; in mixart_update_analog_audio_level() 327 request.data = &io_level; in mixart_update_analog_audio_level() 328 request.size = sizeof(io_level); in mixart_update_analog_audio_level() 330 err = snd_mixart_send_msg(chip->mgr, &request, sizeof(resp), &resp); in mixart_update_analog_audio_level() 727 struct mixart_msg request; in mixart_update_playback_stream_level() local 760 request.message_id = MSG_STREAM_SET_OUT_STREAM_LEVEL; in mixart_update_playback_stream_level() 761 request.uid = (struct mixart_uid){0,0}; in mixart_update_playback_stream_level() [all …]
|
/linux-4.1.27/drivers/staging/rtl8712/ |
D | usb_ops.c | 39 u8 request; in usb_read8() local 47 request = 0x05; in usb_read8() 52 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read8() 59 u8 request; in usb_read16() local 67 request = 0x05; in usb_read16() 72 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read16() 79 u8 request; in usb_read32() local 87 request = 0x05; in usb_read32() 92 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read32() 99 u8 request; in usb_write8() local [all …]
|
/linux-4.1.27/security/apparmor/ |
D | file.c | 70 if (sa->aad->fs.request & AA_AUDIT_FILE_MASK) { in file_audit_cb() 72 audit_file_mask(ab, sa->aad->fs.request); in file_audit_cb() 78 if (sa->aad->fs.request & AA_AUDIT_FILE_MASK) { in file_audit_cb() 107 gfp_t gfp, int op, u32 request, const char *name, in aa_audit_file() argument 116 aad.fs.request = request; in aa_audit_file() 130 sa.aad->fs.request &= mask; in aa_audit_file() 132 if (likely(!sa.aad->fs.request)) in aa_audit_file() 137 sa.aad->fs.request = sa.aad->fs.request & ~perms->allow; in aa_audit_file() 139 if (sa.aad->fs.request & perms->kill) in aa_audit_file() 143 if ((sa.aad->fs.request & perms->quiet) && in aa_audit_file() [all …]
|
D | domain.c | 97 const char *name, u32 request, in change_profile_perms() argument 113 if (COMBINED_PERM_MASK(perms) & request) in change_profile_perms() 756 u32 request; in aa_change_profile() local 762 request = AA_MAY_ONEXEC; in aa_change_profile() 765 request = AA_MAY_CHANGE_PROFILE; in aa_change_profile() 806 perms = change_profile_perms(profile, ns, hname, request, in aa_change_profile() 808 if (!(perms.allow & request)) { in aa_change_profile() 846 error = aa_audit_file(profile, &perms, GFP_KERNEL, op, request, in aa_change_profile()
|
/linux-4.1.27/arch/um/drivers/ |
D | mconsole_user.c | 66 if (!strncmp(req->request.data, cmd->command, in mconsole_parse() 84 req->len = recvfrom(fd, &req->request, sizeof(req->request), 0, in mconsole_get_request() 91 if (req->request.magic != MCONSOLE_MAGIC) { in mconsole_get_request() 93 len = MIN(sizeof(req->request.data) - 1, in mconsole_get_request() 94 strlen((char *) &req->request)); in mconsole_get_request() 95 memmove(req->request.data, &req->request, len); in mconsole_get_request() 96 req->request.data[len] = '\0'; in mconsole_get_request() 98 req->request.magic = MCONSOLE_MAGIC; in mconsole_get_request() 99 req->request.version = 0; in mconsole_get_request() 100 req->request.len = len; in mconsole_get_request() [all …]
|
D | mconsole_kern.c | 67 req->request.cmd->handler(&req->request); in mc_work_proc() 90 new->request = req; in mconsole_interrupt() 91 new->request.regs = get_irq_regs()->regs; in mconsole_interrupt() 115 char *ptr = req->request.data; in mconsole_log() 119 len = req->len - (ptr - req->request.data); in mconsole_log() 131 char *ptr = req->request.data; in mconsole_proc() 473 char *ptr = req->request.data, *name, *error_string = ""; in mconsole_config() 499 char *ptr = req->request.data, *err_msg = ""; in mconsole_remove() 625 char *ptr = req->request.data; in mconsole_sysrq() 661 char *ptr = req->request.data; in mconsole_stack()
|
/linux-4.1.27/Documentation/block/ |
D | request.txt | 2 struct request documentation 9 2.0 Struct request members classification 11 2.1 struct request members explanation 17 Short explanation of request members 45 request type, etc. 57 unsigned long nr_sectors DBI Total number of sectors in request 62 segments in a request 65 segments in a request 68 of request 80 on request completion [all …]
|
D | biodoc.txt | 59 2.3 Changes in the request structure 63 3.2.1 Traversing segments and completion units in a request 117 a per-queue level (e.g maximum request size, maximum number of segments in 134 Sets two variables that limit the size of the request. 136 - The request queue's max_sectors, which is a soft size in 140 - The request queue's max_hw_sectors, which is a hard limit 141 and reflects the maximum size request a driver can handle 148 Maximum physical segments you can handle in a request. 128 152 Maximum dma segments the hardware can handle in a request. 128 176 setting the queue bounce limit for the request queue for the device [all …]
|
D | cfq-iosched.txt | 7 CFQ maintains the per process queue for the processes which request I/O 17 This specifies how long CFQ should idle for next request on certain cfq queues 52 backward distance of request is just 1/back_seek_penalty from a "front" 53 request, then the seeking cost of two requests is considered equivalent. 55 So scheduler will not bias toward one or the other request (otherwise scheduler 56 will bias toward front request). Default value of back_seek_penalty is 2. 101 time for each process to issue I/O request before the cfq queue is switched. 117 This parameter is used to limit the dispatching of asynchronous request to 118 device request queue in queue's slice time. The maximum number of request that 136 This specifies the number of request dispatched to the device queue. In a [all …]
|
D | queue-sysfs.txt | 62 for a filesystem request. Must be smaller than or equal to the maximum 92 To avoid priority inversion through request starvation, a request 93 queue maintains a separate request pool per each cgroup when 95 per-block-cgroup request pool. IOW, if there are N block cgroups, 96 each request queue may have up to N request pools, each independently 119 If this option is '1', the block layer will migrate request completions to the 120 cpu "group" that originally submitted the request. For some workloads this
|
D | deadline-iosched.txt | 21 service time for a request. As we focus mainly on read latencies, this is 22 tunable. When a read request first enters the io scheduler, it is assigned 41 This parameter tunes the balance between per-request latency and aggregate 61 Sometimes it happens that a request enters the io scheduler that is contiguous 62 with a request that is already on the queue. Either it fits in the back of that 63 request, or it fits at the front. That is called either a back merge candidate
|
/linux-4.1.27/drivers/firewire/ |
D | core-transaction.c | 633 struct fw_request *request; in free_response_callback() local 635 request = container_of(packet, struct fw_request, response); in free_response_callback() 636 kfree(request); in free_response_callback() 751 struct fw_request *request; in allocate_request() local 784 request = kmalloc(sizeof(*request) + length, GFP_ATOMIC); in allocate_request() 785 if (request == NULL) in allocate_request() 788 request->response.speed = p->speed; in allocate_request() 789 request->response.timestamp = in allocate_request() 791 request->response.generation = p->generation; in allocate_request() 792 request->response.ack = 0; in allocate_request() [all …]
|
D | core-cdev.c | 125 struct fw_request *request; member 189 struct fw_cdev_event_request request; member 589 struct fw_cdev_send_request *request, in init_request() argument 595 if (request->tcode != TCODE_STREAM_DATA && in init_request() 596 (request->length > 4096 || request->length > 512 << speed)) in init_request() 599 if (request->tcode == TCODE_WRITE_QUADLET_REQUEST && in init_request() 600 request->length < 4) in init_request() 603 e = kmalloc(sizeof(*e) + request->length, GFP_KERNEL); in init_request() 608 e->response.length = request->length; in init_request() 609 e->response.closure = request->closure; in init_request() [all …]
|
D | sbp2.c | 296 } request; member 324 } request; member 418 static void sbp2_status_write(struct fw_card *card, struct fw_request *request, in sbp2_status_write() argument 430 fw_send_response(card, request, RCODE_TYPE_ERROR); in sbp2_status_write() 443 fw_send_response(card, request, RCODE_COMPLETE); in sbp2_status_write() 466 fw_send_response(card, request, RCODE_COMPLETE); in sbp2_status_write() 583 orb->request.response.high = 0; in sbp2_send_management_orb() 584 orb->request.response.low = cpu_to_be32(orb->response_bus); in sbp2_send_management_orb() 586 orb->request.misc = cpu_to_be32( in sbp2_send_management_orb() 590 orb->request.length = cpu_to_be32( in sbp2_send_management_orb() [all …]
|
/linux-4.1.27/drivers/s390/cio/ |
D | chsc_sch.c | 60 struct chsc_request *request = private->request; in chsc_subchannel_irq() local 68 if (!request) { in chsc_subchannel_irq() 73 private->request = NULL; in chsc_subchannel_irq() 74 memcpy(&request->irb, irb, sizeof(*irb)); in chsc_subchannel_irq() 76 complete(&request->completion); in chsc_subchannel_irq() 114 if (private->request) { in chsc_subchannel_remove() 115 complete(&private->request->completion); in chsc_subchannel_remove() 243 struct chsc_request *request) in chsc_async() argument 255 if (private->request) { in chsc_async() 273 private->request = request; in chsc_async() [all …]
|
D | chsc.c | 70 struct chsc_header request; member 104 ssd_area->request.length = 0x0010; in chsc_get_ssd_info() 105 ssd_area->request.code = 0x0004; in chsc_get_ssd_info() 159 ssqd->request.length = 0x0010; in chsc_ssqd() 160 ssqd->request.code = 0x0024; in chsc_ssqd() 185 scssc->request.length = 0x0fe0; in chsc_sadc() 186 scssc->request.code = 0x0021; in chsc_sadc() 331 struct chsc_header request; member 548 sei->request.length = 0x0010; in chsc_process_event_information() 549 sei->request.code = 0x000e; in chsc_process_event_information() [all …]
|
D | chsc.h | 68 struct chsc_header request; member 82 struct chsc_header request; member 102 struct chsc_header request; member 161 struct chsc_header request; member 197 struct chsc_header request; member
|
/linux-4.1.27/Documentation/mmc/ |
D | mmc-async-req.txt | 7 preparations for the next request are done in parallel with the current 10 time between when an MMC request ends and another MMC request begins. 14 MMC request. 22 a request and how fast the memory is. The faster the MMC/SD is the 23 more significant the prepare request time becomes. Roughly the expected 38 It starts a new MMC command request for a host. The function isn't 39 truly non-blocking. If there is an ongoing async request it waits 40 for completion of that request and starts the new one and returns. It 41 doesn't wait for the new request to complete. If there is no ongoing 42 request it starts the new request and returns immediately. [all …]
|
/linux-4.1.27/kernel/ |
D | ptrace.c | 296 static int ptrace_attach(struct task_struct *task, long request, in ptrace_attach() argument 300 bool seize = (request == PTRACE_SEIZE); in ptrace_attach() 702 #define is_singlestep(request) ((request) == PTRACE_SINGLESTEP) argument 704 #define is_singlestep(request) 0 argument 708 #define is_singleblock(request) ((request) == PTRACE_SINGLEBLOCK) argument 710 #define is_singleblock(request) 0 argument 714 #define is_sysemu_singlestep(request) ((request) == PTRACE_SYSEMU_SINGLESTEP) argument 716 #define is_sysemu_singlestep(request) 0 argument 719 static int ptrace_resume(struct task_struct *child, long request, in ptrace_resume() argument 727 if (request == PTRACE_SYSCALL) in ptrace_resume() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/llite/ |
D | symlink.c | 46 struct ptlrpc_request **request, char **symname) in ll_readlink_internal() argument 54 *request = NULL; in ll_readlink_internal() 75 rc = md_getattr(sbi->ll_md_exp, op_data, request); in ll_readlink_internal() 83 body = req_capsule_server_get(&(*request)->rq_pill, &RMF_MDT_BODY); in ll_readlink_internal() 99 *symname = req_capsule_server_get(&(*request)->rq_pill, &RMF_MDT_MD); in ll_readlink_internal() 124 struct ptlrpc_request *request = NULL; in ll_follow_link() local 138 rc = ll_readlink_internal(inode, &request, &symname); in ll_follow_link() 142 ptlrpc_req_finished(request); in ll_follow_link() 143 request = NULL; in ll_follow_link() 151 return request; in ll_follow_link()
|
D | namei.c | 408 static int ll_lookup_it_finish(struct ptlrpc_request *request, in ll_lookup_it_finish() argument 421 rc = ll_prep_inode(&inode, request, (*de)->d_sb, it); in ll_lookup_it_finish() 678 struct ptlrpc_request *request = NULL; in ll_create_node() local 685 request = it->d.lustre.it_data; in ll_create_node() 687 rc = ll_prep_inode(&inode, request, dir->i_sb, it); in ll_create_node() 702 ptlrpc_req_finished(request); in ll_create_node() 742 static void ll_update_times(struct ptlrpc_request *request, in ll_update_times() argument 745 struct mdt_body *body = req_capsule_server_get(&request->rq_pill, in ll_update_times() 764 struct ptlrpc_request *request = NULL; in ll_new_node() local 786 cfs_curproc_cap_pack(), rdev, &request); in ll_new_node() [all …]
|
D | dir.c | 148 struct ptlrpc_request *request; in ll_dir_filler() local 184 rc = md_readpage(exp, op_data, page_pool, &request); in ll_dir_filler() 190 body = req_capsule_server_get(&request->rq_pill, &RMF_MDT_BODY); in ll_dir_filler() 197 nrdpgs = (request->rq_bulk->bd_nob_transferred+PAGE_CACHE_SIZE-1) in ll_dir_filler() 202 ptlrpc_req_finished(request); in ll_dir_filler() 365 struct ptlrpc_request *request; in ll_get_dir_page() local 378 request = (struct ptlrpc_request *)it.d.lustre.it_data; in ll_get_dir_page() 379 if (request) in ll_get_dir_page() 380 ptlrpc_req_finished(request); in ll_get_dir_page() 661 struct ptlrpc_request *request = NULL; in ll_dir_setdirstripe() local [all …]
|
D | xattr.c | 479 struct ptlrpc_request *request = NULL; in ll_getxattr() local 496 &lmmsize, &request); in ll_getxattr() 536 if (request) in ll_getxattr() 537 ptlrpc_req_finished(request); in ll_getxattr() 551 struct ptlrpc_request *request = NULL; in ll_listxattr() local 592 rc2 = ll_dir_getstripe(inode, &lmm, &lmmsize, &request); in ll_listxattr() 604 ptlrpc_req_finished(request); in ll_listxattr() 617 ptlrpc_req_finished(request); in ll_listxattr()
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
D | mthca_profile.c | 65 struct mthca_profile *request, in mthca_make_profile() argument 100 profile[MTHCA_RES_UARC].size = request->uarc_size; in mthca_make_profile() 102 profile[MTHCA_RES_QP].num = request->num_qp; in mthca_make_profile() 103 profile[MTHCA_RES_SRQ].num = request->num_srq; in mthca_make_profile() 104 profile[MTHCA_RES_EQP].num = request->num_qp; in mthca_make_profile() 105 profile[MTHCA_RES_RDB].num = request->num_qp * request->rdb_per_qp; in mthca_make_profile() 106 profile[MTHCA_RES_CQ].num = request->num_cq; in mthca_make_profile() 108 profile[MTHCA_RES_MCG].num = request->num_mcg; in mthca_make_profile() 109 profile[MTHCA_RES_MPT].num = request->num_mpt; in mthca_make_profile() 110 profile[MTHCA_RES_MTT].num = request->num_mtt; in mthca_make_profile() [all …]
|
/linux-4.1.27/drivers/isdn/hardware/eicon/ |
D | dadapter.c | 112 if (d->request) { in diva_didd_add_descriptor() 113 MAdapter.request = d->request; in diva_didd_add_descriptor() 114 dprintf = (DIVA_DI_PRINTF)d->request; in diva_didd_add_descriptor() 116 DBG_TRC(("DIMAINT registered, dprintf=%08x", d->request)) in diva_didd_add_descriptor() 120 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_add_descriptor() 132 DBG_TRC(("Add adapter[%d], request=%08x", (i + 1), d->request)) in diva_didd_add_descriptor() 145 static int diva_didd_remove_descriptor(IDI_CALL request) { in diva_didd_remove_descriptor() argument 148 if (request == MAdapter.request) { in diva_didd_remove_descriptor() 152 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_remove_descriptor() 156 if (HandleTable[i].request == request) { in diva_didd_remove_descriptor() [all …]
|
D | divasfunc.c | 72 d.request = Requests[card - 1]; in diva_xdi_didd_register_adapter() 84 DAdapter.request((ENTITY *)&req); in diva_xdi_didd_register_adapter() 106 DAdapter.request((ENTITY *)&req); in diva_xdi_didd_remove_adapter() 146 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in didd_callback() 174 DAdapter.request((ENTITY *)&req); in connect_didd() 182 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in connect_didd() 206 DAdapter.request((ENTITY *)&req); in disconnect_didd()
|
D | idifunc.c | 70 if (card->d.request == d->request) { in find_card_in_list() 98 card->d.request((ENTITY *)&sync_req); in um_new_card() 168 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in didd_callback() 202 DAdapter.request((ENTITY *)&req); in connect_didd() 210 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in connect_didd() 237 DAdapter.request((ENTITY *)&req); in disconnect_didd()
|
D | debug.c | 25 static void diva_free_dma_descriptor(IDI_CALL request, int nr); 26 static int diva_get_dma_descriptor(IDI_CALL request, dword *dma_magic); 42 IDI_CALL request; member 833 …(*(clients[id].request))((ENTITY *)(*(clients[id].pIdiLib->DivaSTraceGetHandle))(clients[id].pIdiL… in diva_set_driver_dbg_mask() 841 static int diva_get_idi_adapter_info(IDI_CALL request, dword *serial, dword *logical) { in diva_get_idi_adapter_info() argument 846 (*request)((ENTITY *)&sync_req); in diva_get_idi_adapter_info() 852 (*request)((ENTITY *)&sync_req); in diva_get_idi_adapter_info() 872 diva_get_idi_adapter_info(d->request, &serial, &logical); in diva_mnt_add_xdi_adapter() 891 if (clients[id].hDbg && (clients[id].request == d->request)) { in diva_mnt_add_xdi_adapter() 921 clients[id].request = d->request; in diva_mnt_add_xdi_adapter() [all …]
|
D | mntfunc.c | 59 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in didd_callback() 93 DAdapter.request((ENTITY *)&req); in connect_didd() 103 DAdapter.request((ENTITY *)&req); in connect_didd() 124 DAdapter.request((ENTITY *)&req); in disconnect_didd() 129 (IDI_CALL) MaintDescriptor.request; in disconnect_didd() 130 DAdapter.request((ENTITY *)&req); in disconnect_didd()
|
D | capifunc.c | 376 if (adapter[i].request) in clean_adapter() 383 while ((max_adapter != 0) && !adapter[max_adapter - 1].request) in clean_adapter() 407 if (card->d.request == d->request) { in divacapi_remove_card() 462 d.request = card->d.request; in divacapi_remove_cards() 507 card->d.request((ENTITY *)&sync_req); in diva_add_card() 536 card->d.request((ENTITY *)&sync_req); in diva_add_card() 557 card->d.request((ENTITY *)&sync_req); in diva_add_card() 562 a->request = DIRequest; /* card->d.request; */ in diva_add_card() 620 if (adapter[i].request) in diva_add_card() 679 if (adapter[i].request) in diva_add_card() [all …]
|
/linux-4.1.27/arch/mips/jz4740/ |
D | gpio.c | 138 int jz_gpio_bulk_request(const struct jz_gpio_bulk_request *request, size_t num) in jz_gpio_bulk_request() argument 143 for (i = 0; i < num; ++i, ++request) { in jz_gpio_bulk_request() 144 ret = gpio_request(request->gpio, request->name); in jz_gpio_bulk_request() 147 jz_gpio_set_function(request->gpio, request->function); in jz_gpio_bulk_request() 153 for (--request; i > 0; --i, --request) { in jz_gpio_bulk_request() 154 gpio_free(request->gpio); in jz_gpio_bulk_request() 155 jz_gpio_set_function(request->gpio, JZ_GPIO_FUNC_NONE); in jz_gpio_bulk_request() 162 void jz_gpio_bulk_free(const struct jz_gpio_bulk_request *request, size_t num) in jz_gpio_bulk_free() argument 166 for (i = 0; i < num; ++i, ++request) { in jz_gpio_bulk_free() 167 gpio_free(request->gpio); in jz_gpio_bulk_free() [all …]
|
/linux-4.1.27/drivers/s390/block/ |
D | scm_blk_cluster.c | 60 static bool clusters_intersect(struct request *A, struct request *B) in clusters_intersect() 77 struct request *req = scmrq->request[scmrq->aob->request.msb_count]; in scm_reserve_cluster() 95 for (pos = 0; pos < iter->aob->request.msb_count; pos++) { in scm_reserve_cluster() 96 if (clusters_intersect(req, iter->request[pos]) && in scm_reserve_cluster() 98 rq_data_dir(iter->request[pos]) == WRITE)) { in scm_reserve_cluster() 134 struct request *req = scmrq->request[0]; in scm_prepare_cluster_request() 164 scmrq->aob->request.msb_count = 1; in scm_prepare_cluster_request() 199 int pos = scmrq->aob->request.msb_count; in scm_need_cluster_request() 201 if (rq_data_dir(scmrq->request[pos]) == READ) in scm_need_cluster_request() 204 return blk_rq_bytes(scmrq->request[pos]) < CLUSTER_SIZE; in scm_need_cluster_request()
|
D | scm_blk.c | 46 kfree(scmrq->request); in __scm_free_rq() 80 scmrq->request = kcalloc(nr_requests_per_io, sizeof(scmrq->request[0]), in __scm_alloc_rq() 82 if (!scmrq->request) in __scm_alloc_rq() 134 for (i = 0; i < nr_requests_per_io && scmrq->request[i]; i++) { in scm_request_done() 148 static bool scm_permit_request(struct scm_blk_dev *bdev, struct request *req) in scm_permit_request() 185 int pos = scmrq->aob->request.msb_count; in scm_request_prepare() 187 struct request *req = scmrq->request[pos]; in scm_request_prepare() 197 scmrq->aob->request.msb_count++; in scm_request_prepare() 215 struct request *req) in scm_request_set() 217 scmrq->request[scmrq->aob->request.msb_count] = req; in scm_request_set() [all …]
|
/linux-4.1.27/arch/s390/pci/ |
D | pci_clp.c | 88 rrb->request.hdr.len = sizeof(rrb->request); in clp_query_pci_fngrp() 89 rrb->request.hdr.cmd = CLP_QUERY_PCI_FNGRP; in clp_query_pci_fngrp() 91 rrb->request.pfgid = pfgid; in clp_query_pci_fngrp() 141 rrb->request.hdr.len = sizeof(rrb->request); in clp_query_pci_fn() 142 rrb->request.hdr.cmd = CLP_QUERY_PCI_FN; in clp_query_pci_fn() 144 rrb->request.fh = fh; in clp_query_pci_fn() 210 rrb->request.hdr.len = sizeof(rrb->request); in clp_set_pci_fn() 211 rrb->request.hdr.cmd = CLP_SET_PCI_FN; in clp_set_pci_fn() 213 rrb->request.fh = *fh; in clp_set_pci_fn() 214 rrb->request.oc = command; in clp_set_pci_fn() [all …]
|
/linux-4.1.27/Documentation/filesystems/nfs/ |
D | idmapper.txt | 7 performing an upcall to userspace to request the information. There are two 8 ways NFS could obtain this information: placing a call to /sbin/request-key 11 NFS will attempt to call /sbin/request-key first. If this succeeds, the 12 result will be cached using the generic request-key cache. This call should 13 only fail if /etc/request-key.conf is not configured for the id_resolver key 14 type, see the "Configuring" section below if you wish to use the request-key 17 If the call to /sbin/request-key fails (if /etc/request-key.conf is not 26 The file /etc/request-key.conf will need to be modified so /sbin/request-key can 46 would edit your request-key.conf so it look similar to this: 54 request-key will find the first matching line and corresponding program. In [all …]
|
/linux-4.1.27/drivers/scsi/ |
D | storvsc_drv.c | 821 struct storvsc_cmd_request *request; in handle_multichannel_storage() local 830 request = &stor_device->init_request; in handle_multichannel_storage() 831 vstor_packet = &request->vstor_packet; in handle_multichannel_storage() 851 memset(request, 0, sizeof(struct storvsc_cmd_request)); in handle_multichannel_storage() 852 init_completion(&request->wait_event); in handle_multichannel_storage() 860 (unsigned long)request, in handle_multichannel_storage() 867 t = wait_for_completion_timeout(&request->wait_event, 10*HZ); in handle_multichannel_storage() 886 struct storvsc_cmd_request *request; in storvsc_channel_init() local 896 request = &stor_device->init_request; in storvsc_channel_init() 897 vstor_packet = &request->vstor_packet; in storvsc_channel_init() [all …]
|
D | qlogicfas408.c | 96 static int ql_pdma(struct qlogicfas408_priv *priv, int phase, char *request, int reqlen) in ql_pdma() argument 106 insl(qbase + 4, request, 32); in ql_pdma() 108 request += 128; in ql_pdma() 113 insl(qbase + 4, request, 21); in ql_pdma() 115 request += 84; in ql_pdma() 118 insl(qbase + 4, request, 11); in ql_pdma() 120 request += 44; in ql_pdma() 132 *request++ = inb(qbase + 4); in ql_pdma() 143 outsl(qbase + 4, request, 32); in ql_pdma() 145 request += 128; in ql_pdma() [all …]
|
D | scsi_lib.c | 122 struct request_queue *q = cmd->request->q; in scsi_mq_requeue_cmd() 124 blk_mq_requeue_request(cmd->request); in scsi_mq_requeue_cmd() 171 blk_requeue_request(q, cmd->request); in __scsi_queue_insert() 220 struct request *req; in scsi_execute() 531 struct request *req = cmd->request; in scsi_requeue_command() 616 if (cmd->request->cmd_type == REQ_TYPE_FS) { in scsi_uninit_cmd() 628 if (cmd->request->next_rq && cmd->request->next_rq->special) in scsi_mq_free_sgtables() 629 scsi_free_sgtable(cmd->request->next_rq->special, true); in scsi_mq_free_sgtables() 680 struct scsi_data_buffer *bidi_sdb = cmd->request->next_rq->special; in scsi_release_bidi_buffers() 684 cmd->request->next_rq->special = NULL; in scsi_release_bidi_buffers() [all …]
|
D | sd_dif.c | 117 sdkp = scsi_disk(scmd->request->rq_disk); in sd_dif_prepare() 124 __rq_for_each_bio(bio, scmd->request) { in sd_dif_prepare() 168 sdkp = scsi_disk(scmd->request->rq_disk); in sd_dif_complete() 176 __rq_for_each_bio(bio, scmd->request) { in sd_dif_complete()
|
/linux-4.1.27/drivers/media/usb/pwc/ |
D | pwc-ctrl.c | 108 u8 request, u16 value, int recv_count) in recv_control_msg() argument 113 request, in recv_control_msg() 119 rc, request, value); in recv_control_msg() 144 u8 request, u16 value, void *buf, int buflen) in send_control_msg() argument 147 request, in send_control_msg() 412 int pwc_get_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_u8_ctrl() argument 416 ret = recv_control_msg(pdev, request, value, 1); in pwc_get_u8_ctrl() 424 int pwc_set_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, u8 data) in pwc_set_u8_ctrl() argument 429 ret = send_control_msg(pdev, request, value, pdev->ctrl_buf, 1); in pwc_set_u8_ctrl() 436 int pwc_get_s8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_s8_ctrl() argument [all …]
|
D | pwc.h | 372 u8 request, u16 value, void *buf, int buflen); 375 int pwc_get_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data); 376 int pwc_set_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, u8 data); 377 int pwc_get_s8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data); 379 int pwc_get_u16_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *dat); 380 int pwc_set_u16_ctrl(struct pwc_device *pdev, u8 request, u16 value, u16 data);
|
/linux-4.1.27/drivers/input/misc/ |
D | uinput.c | 63 struct uinput_request *request) in uinput_request_alloc_id() argument 72 request->id = id; in uinput_request_alloc_id() 73 udev->requests[id] = request; in uinput_request_alloc_id() 94 struct uinput_request *request) in uinput_request_reserve_slot() argument 98 uinput_request_alloc_id(udev, request)); in uinput_request_reserve_slot() 102 struct uinput_request *request) in uinput_request_done() argument 105 udev->requests[request->id] = NULL; in uinput_request_done() 108 complete(&request->done); in uinput_request_done() 112 struct uinput_request *request) in uinput_request_send() argument 125 init_completion(&request->done); in uinput_request_send() [all …]
|
/linux-4.1.27/drivers/clk/qcom/ |
D | clk-rcg2.c | 339 s64 request; in clk_edp_pixel_set_rate() local 349 request = rate; in clk_edp_pixel_set_rate() 350 request *= frac->den; in clk_edp_pixel_set_rate() 351 request = div_s64(request, frac->num); in clk_edp_pixel_set_rate() 352 if ((src_rate < (request - delta)) || in clk_edp_pixel_set_rate() 353 (src_rate > (request + delta))) in clk_edp_pixel_set_rate() 387 s64 request; in clk_edp_pixel_determine_rate() local 401 request = rate; in clk_edp_pixel_determine_rate() 402 request *= frac->den; in clk_edp_pixel_determine_rate() 403 request = div_s64(request, frac->num); in clk_edp_pixel_determine_rate() [all …]
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
D | profile.c | 70 struct mlx4_profile *request, in mlx4_make_profile() argument 106 request->num_mtt = in mlx4_make_profile() 107 roundup_pow_of_two(max_t(unsigned, request->num_mtt, in mlx4_make_profile() 123 profile[MLX4_RES_QP].num = request->num_qp; in mlx4_make_profile() 124 profile[MLX4_RES_RDMARC].num = request->num_qp * request->rdmarc_per_qp; in mlx4_make_profile() 125 profile[MLX4_RES_ALTC].num = request->num_qp; in mlx4_make_profile() 126 profile[MLX4_RES_AUXC].num = request->num_qp; in mlx4_make_profile() 127 profile[MLX4_RES_SRQ].num = request->num_srq; in mlx4_make_profile() 128 profile[MLX4_RES_CQ].num = request->num_cq; in mlx4_make_profile() 131 profile[MLX4_RES_DMPT].num = request->num_mpt; in mlx4_make_profile() [all …]
|
/linux-4.1.27/Documentation/scsi/ |
D | hptiop.txt | 85 A request packet can be allocated in either IOP or host memory. 87 To send a request to the controller: 89 - Get a free request packet by reading the inbound queue port or 90 allocate a free request in host DMA coherent memory. 104 - The IOP process the request. When the request is completed, it 108 For requests allocated in IOP memory, the request offset is posted to 113 flag is set in the request, the low 32-bit context value will be 116 - The host read the outbound queue and complete the request. 118 For requests allocated in IOP memory, the host driver free the request 131 To send a request to the controller: [all …]
|
D | osd.txt | 75 its state throughout execution. Each request goes through these stages: 77 a. osd_start_request() allocates the request. 79 b. Any of the osd_req_* methods is used to encode a request of the specified 84 can be called multiple times on the same request. However, only one 88 and signs the request using the provided capability key and integrity- 91 e. osd_execute_request() may be called to execute the request via the block 92 layer and wait for its completion. The request can be executed 95 f. After execution, osd_req_decode_sense() can be called to decode the request's 101 h. osd_end_request() must be called to deallocate the request and any resource 102 associated with it. Note that osd_end_request cleans up the request at any [all …]
|
/linux-4.1.27/net/wireless/ |
D | reg.c | 280 static void reg_free_request(struct regulatory_request *request) in reg_free_request() argument 282 if (request != get_last_request()) in reg_free_request() 283 kfree(request); in reg_free_request() 294 static void reg_update_last_request(struct regulatory_request *request) in reg_update_last_request() argument 299 if (lr == request) in reg_update_last_request() 303 rcu_assign_pointer(last_request, request); in reg_update_last_request() 466 struct reg_regdb_search_request *request; in reg_regdb_search() local 474 request = list_first_entry(®_regdb_search_list, in reg_regdb_search() 477 list_del(&request->list); in reg_regdb_search() 482 if (alpha2_equal(request->alpha2, curdom->alpha2)) { in reg_regdb_search() [all …]
|
D | nl80211.h | 21 struct regulatory_request *request); 24 nl80211_send_reg_change_event(struct regulatory_request *request) in nl80211_send_reg_change_event() argument 26 nl80211_common_reg_change_event(NL80211_CMD_REG_CHANGE, request); in nl80211_send_reg_change_event() 30 nl80211_send_wiphy_reg_change_event(struct regulatory_request *request) in nl80211_send_wiphy_reg_change_event() argument 32 nl80211_common_reg_change_event(NL80211_CMD_WIPHY_REG_CHANGE, request); in nl80211_send_wiphy_reg_change_event()
|
D | sme.c | 63 struct cfg80211_scan_request *request; in cfg80211_conn_scan() local 77 request = kzalloc(sizeof(*request) + sizeof(request->ssids[0]) + in cfg80211_conn_scan() 78 sizeof(request->channels[0]) * n_channels, in cfg80211_conn_scan() 80 if (!request) in cfg80211_conn_scan() 89 kfree(request); in cfg80211_conn_scan() 92 request->channels[0] = wdev->conn->params.channel; in cfg80211_conn_scan() 93 request->rates[band] = (1 << sband->n_bitrates) - 1; in cfg80211_conn_scan() 108 request->channels[i++] = channel; in cfg80211_conn_scan() 110 request->rates[band] = (1 << bands->n_bitrates) - 1; in cfg80211_conn_scan() 114 request->n_channels = n_channels; in cfg80211_conn_scan() [all …]
|
D | scan.c | 168 struct cfg80211_scan_request *request; in ___cfg80211_scan_done() local 183 request = rdev->scan_req; in ___cfg80211_scan_done() 184 if (!request) in ___cfg80211_scan_done() 187 wdev = request->wdev; in ___cfg80211_scan_done() 197 if (!request->aborted && in ___cfg80211_scan_done() 198 request->flags & NL80211_SCAN_FLAG_FLUSH) { in ___cfg80211_scan_done() 201 __cfg80211_bss_expire(rdev, request->scan_start); in ___cfg80211_scan_done() 205 msg = nl80211_build_scan_msg(rdev, wdev, request->aborted); in ___cfg80211_scan_done() 208 if (wdev->netdev && !request->aborted) { in ___cfg80211_scan_done() 219 kfree(request); in ___cfg80211_scan_done() [all …]
|
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/ |
D | framework.c | 369 sfw_get_stats(srpc_stat_reqst_t *request, srpc_stat_reply_t *reply) in sfw_get_stats() argument 378 if (request->str_sid.ses_nid == LNET_NID_ANY) { in sfw_get_stats() 383 if (sn == NULL || !sfw_sid_equal(request->str_sid, sn->sn_id)) { in sfw_get_stats() 412 sfw_make_session(srpc_mksn_reqst_t *request, srpc_mksn_reply_t *reply) in sfw_make_session() argument 415 srpc_msg_t *msg = container_of(request, srpc_msg_t, in sfw_make_session() 419 if (request->mksn_sid.ses_nid == LNET_NID_ANY) { in sfw_make_session() 430 if (sfw_sid_equal(request->mksn_sid, sn->sn_id)) { in sfw_make_session() 435 if (!request->mksn_force) { in sfw_make_session() 463 sfw_init_session(sn, request->mksn_sid, in sfw_make_session() 464 msg->msg_ses_feats, &request->mksn_name[0]); in sfw_make_session() [all …]
|
/linux-4.1.27/arch/microblaze/kernel/ |
D | ptrace.c | 77 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 83 switch (request) { in arch_ptrace() 89 if (addr >= PT_SIZE && request == PTRACE_PEEKUSR) { in arch_ptrace() 106 if (request == PTRACE_PEEKUSR) in arch_ptrace() 126 if (rval == 0 && request == PTRACE_PEEKUSR) in arch_ptrace() 130 rval = ptrace_request(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/security/smack/ |
D | smack_access.c | 126 int request, struct smk_audit_info *a) in smk_access() argument 163 if ((request & MAY_ANYREAD) == request || in smk_access() 164 (request & MAY_LOCK) == request) { in smk_access() 182 if (may <= 0 || (request & may) != request) { in smk_access() 210 request, rc, a); in smk_access() 332 if (sad->request[0] == '\0') in smack_log_callback() 335 audit_log_format(ab, " requested=%s", sad->request); in smack_log_callback() 349 void smack_log(char *subject_label, char *object_label, int request, in smack_log() argument 372 smack_str_from_perm(request_buffer, request); in smack_log() 390 sad->request = request_buffer; in smack_log() [all …]
|
/linux-4.1.27/include/drm/ |
D | drm_agpsupport.h | 56 int drm_agp_alloc(struct drm_device *dev, struct drm_agp_buffer *request); 59 int drm_agp_free(struct drm_device *dev, struct drm_agp_buffer *request); 62 int drm_agp_unbind(struct drm_device *dev, struct drm_agp_binding *request); 65 int drm_agp_bind(struct drm_device *dev, struct drm_agp_binding *request); 150 struct drm_agp_buffer *request) in drm_agp_alloc() argument 162 struct drm_agp_buffer *request) in drm_agp_free() argument 174 struct drm_agp_binding *request) in drm_agp_unbind() argument 186 struct drm_agp_binding *request) in drm_agp_bind() argument
|
/linux-4.1.27/drivers/block/ |
D | nbd.c | 50 struct request *active_req; 99 static void nbd_end_request(struct nbd_device *nbd, struct request *req) in nbd_end_request() 228 static int nbd_send_req(struct nbd_device *nbd, struct request *req) in nbd_send_req() 231 struct nbd_request request; in nbd_send_req() local 234 memset(&request, 0, sizeof(request)); in nbd_send_req() 235 request.magic = htonl(NBD_REQUEST_MAGIC); in nbd_send_req() 236 request.type = htonl(nbd_cmd(req)); in nbd_send_req() 239 request.from = cpu_to_be64((u64)blk_rq_pos(req) << 9); in nbd_send_req() 240 request.len = htonl(size); in nbd_send_req() 242 memcpy(request.handle, &req, sizeof(req)); in nbd_send_req() [all …]
|
/linux-4.1.27/tools/usb/usbip/src/ |
D | usbip_attach.c | 123 struct op_import_request request; in query_import_device() local 127 memset(&request, 0, sizeof(request)); in query_import_device() 137 strncpy(request.busid, busid, SYSFS_BUS_ID_SIZE-1); in query_import_device() 139 PACK_OP_IMPORT_REQUEST(0, &request); in query_import_device() 141 rc = usbip_net_send(sockfd, (void *) &request, sizeof(request)); in query_import_device()
|
D | usbip_network.h | 78 #define PACK_OP_IMPORT_REQUEST(pack, request) do {\ argument 100 #define PACK_OP_EXPORT_REQUEST(pack, request) do {\ argument 101 usbip_net_pack_usb_device(pack, &(request)->udev);\ 121 #define PACK_OP_UNEXPORT_REQUEST(pack, request) do {\ argument 122 usbip_net_pack_usb_device(pack, &(request)->udev);\ 163 #define PACK_OP_DEVLIST_REQUEST(pack, request) do {\ argument
|
/linux-4.1.27/drivers/hid/ |
D | hid-hyperv.c | 136 struct synthhid_protocol_request request; member 386 struct mousevsc_prt_msg *request; in mousevsc_connect_to_vsp() local 389 request = &input_dev->protocol_req; in mousevsc_connect_to_vsp() 390 memset(request, 0, sizeof(struct mousevsc_prt_msg)); in mousevsc_connect_to_vsp() 392 request->type = PIPE_MESSAGE_DATA; in mousevsc_connect_to_vsp() 393 request->size = sizeof(struct synthhid_protocol_request); in mousevsc_connect_to_vsp() 394 request->request.header.type = SYNTH_HID_PROTOCOL_REQUEST; in mousevsc_connect_to_vsp() 395 request->request.header.size = sizeof(unsigned int); in mousevsc_connect_to_vsp() 396 request->request.version_requested.version = SYNTHHID_INPUT_VERSION; in mousevsc_connect_to_vsp() 398 ret = vmbus_sendpacket(device->channel, request, in mousevsc_connect_to_vsp() [all …]
|
/linux-4.1.27/Documentation/ABI/testing/ |
D | sysfs-driver-ppi | 21 What: /sys/devices/pnp0/<bus-num>/ppi/request 25 This attribute shows the request for an operation to be 27 the OS to the pre-OS environment. The request should be an 28 integer value range from 1 to 160, and 0 means no request. 36 request it acted upon. The format is "<request> <response num> 54 This attribute shows whether it is allowed to request an 57 The format is "<request> <status num>: <status description>". 65 This attribute shows whether it is allowed to request an
|
D | configfs-usb-gadget-uac1 | 11 req_buf_size - ISO OUT endpoint request buffer size 12 req_count - ISO OUT endpoint request count
|
/linux-4.1.27/fs/ |
D | locks.c | 865 static int flock_lock_inode(struct inode *inode, struct file_lock *request) in flock_lock_inode() argument 874 ctx = locks_get_lock_context(inode, request->fl_type); in flock_lock_inode() 876 if (request->fl_type != F_UNLCK) in flock_lock_inode() 878 return (request->fl_flags & FL_EXISTS) ? -ENOENT : 0; in flock_lock_inode() 881 if (!(request->fl_flags & FL_ACCESS) && (request->fl_type != F_UNLCK)) { in flock_lock_inode() 888 if (request->fl_flags & FL_ACCESS) in flock_lock_inode() 892 if (request->fl_file != fl->fl_file) in flock_lock_inode() 894 if (request->fl_type == fl->fl_type) in flock_lock_inode() 901 if (request->fl_type == F_UNLCK) { in flock_lock_inode() 902 if ((request->fl_flags & FL_EXISTS) && !found) in flock_lock_inode() [all …]
|
/linux-4.1.27/tools/firewire/ |
D | decode-fcp.c | 147 (struct avc_frame *) t->request->packet.write_block.data; in decode_avc() 176 (struct avc_frame *) t->request->packet.write_block.data; in decode_fcp() 178 ((unsigned long long) t->request->packet.common.offset_high << 32) | in decode_fcp() 179 t->request->packet.common.offset_low; in decode_fcp() 181 if (t->request->packet.common.tcode != TCODE_WRITE_BLOCK_REQUEST) in decode_fcp()
|
/linux-4.1.27/drivers/media/rc/img-ir/ |
D | img-ir-rc5.c | 16 struct img_ir_scancode_req *request) in img_ir_rc5_scancode() argument 36 request->protocol = RC_TYPE_RC5; in img_ir_rc5_scancode() 37 request->scancode = addr << 8 | cmd; in img_ir_rc5_scancode() 38 request->toggle = tgl; in img_ir_rc5_scancode()
|
D | img-ir-rc6.c | 16 struct img_ir_scancode_req *request) in img_ir_rc6_scancode() argument 57 request->protocol = RC_TYPE_RC6_0; in img_ir_rc6_scancode() 58 request->scancode = addr << 8 | cmd; in img_ir_rc6_scancode() 59 request->toggle = trl2; in img_ir_rc6_scancode()
|
D | img-ir-sony.c | 16 struct img_ir_scancode_req *request) in img_ir_sony_scancode() argument 28 request->protocol = RC_TYPE_SONY12; in img_ir_sony_scancode() 37 request->protocol = RC_TYPE_SONY15; in img_ir_sony_scancode() 47 request->protocol = RC_TYPE_SONY20; in img_ir_sony_scancode() 52 request->scancode = dev << 16 | subdev << 8 | func; in img_ir_sony_scancode()
|
D | img-ir-nec.c | 17 struct img_ir_scancode_req *request) in img_ir_nec_scancode() argument 33 request->scancode = bitrev8(addr) << 24 | in img_ir_nec_scancode() 40 request->scancode = addr << 16 | in img_ir_nec_scancode() 46 request->scancode = addr << 8 | in img_ir_nec_scancode() 49 request->protocol = RC_TYPE_NEC; in img_ir_nec_scancode()
|
D | img-ir-jvc.c | 16 struct img_ir_scancode_req *request) in img_ir_jvc_scancode() argument 26 request->protocol = RC_TYPE_JVC; in img_ir_jvc_scancode() 27 request->scancode = cust << 8 | data; in img_ir_jvc_scancode()
|
D | img-ir-sharp.c | 16 struct img_ir_scancode_req *request) in img_ir_sharp_scancode() argument 35 request->protocol = RC_TYPE_SHARP; in img_ir_sharp_scancode() 36 request->scancode = addr << 8 | cmd; in img_ir_sharp_scancode()
|
D | img-ir-sanyo.c | 27 struct img_ir_scancode_req *request) in img_ir_sanyo_scancode() argument 47 request->protocol = RC_TYPE_SANYO; in img_ir_sanyo_scancode() 48 request->scancode = addr << 8 | data; in img_ir_sanyo_scancode()
|
/linux-4.1.27/include/scsi/ |
D | scsi_cmnd.h | 110 struct request *request; /* The command we are member 155 return *(struct scsi_driver **)cmd->request->rq_disk->private_data; in scsi_cmd_to_driver() 201 return blk_bidi_rq(cmd->request) && in scsi_bidi_cmnd() 202 (cmd->request->next_rq->special != NULL); in scsi_bidi_cmnd() 208 cmd->request->next_rq->special : &cmd->sdb; in scsi_in() 294 return blk_rq_pos(scmd->request); in scsi_get_lba() 304 return blk_rq_pos(scmd->request) >> in scsi_prot_ref_tag()
|
D | scsi_tcq.h | 17 struct request *req = NULL; in scsi_mq_find_tag() 35 struct request *req; in scsi_find_tag() 89 struct request *req; in scsi_host_find_tag()
|
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/ |
D | hif-ops.h | 25 u32 len, u32 request) in hif_read_write_sync() argument 29 (request & HIF_WRITE) ? "write" : "read", in hif_read_write_sync() 30 addr, buf, len, request); in hif_read_write_sync() 32 return ar->hif_ops->read_write_sync(ar, addr, buf, len, request); in hif_read_write_sync() 36 u32 length, u32 request, in hif_write_async() argument 41 address, buffer, length, request); in hif_write_async() 44 request, packet); in hif_write_async()
|
D | sdio.c | 148 static int ath6kl_sdio_io(struct sdio_func *func, u32 request, u32 addr, in ath6kl_sdio_io() argument 155 if (request & HIF_WRITE) { in ath6kl_sdio_io() 165 if (request & HIF_FIXED_ADDRESS) in ath6kl_sdio_io() 170 if (request & HIF_FIXED_ADDRESS) in ath6kl_sdio_io() 179 request & HIF_WRITE ? "wr" : "rd", addr, in ath6kl_sdio_io() 180 request & HIF_FIXED_ADDRESS ? " (fixed)" : "", buf, len); in ath6kl_sdio_io() 183 trace_ath6kl_sdio(addr, request, buf, len); in ath6kl_sdio_io() 408 u32 len, u32 request) in ath6kl_sdio_read_write_sync() argument 415 if (request & HIF_BLOCK_BASIS) in ath6kl_sdio_read_write_sync() 424 if (request & HIF_WRITE) in ath6kl_sdio_read_write_sync() [all …]
|
/linux-4.1.27/drivers/scsi/isci/ |
D | task.h | 169 struct isci_request *request, 173 struct isci_request *request); 176 struct isci_request *request); 180 struct isci_request *request); 183 struct isci_request *request);
|
D | request.c | 2502 struct isci_request *request, in isci_request_set_open_reject_status() argument 2509 set_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_set_open_reject_status() 2526 struct isci_request *request, in isci_request_handle_controller_specific_errors() argument 2533 cstatus = request->scu_status; in isci_request_handle_controller_specific_errors() 2535 dev_dbg(&request->isci_host->pdev->dev, in isci_request_handle_controller_specific_errors() 2538 __func__, request, cstatus); in isci_request_handle_controller_specific_errors() 2569 set_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_handle_controller_specific_errors() 2579 clear_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_handle_controller_specific_errors() 2607 set_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_handle_controller_specific_errors() 2617 request, task, response_ptr, status_ptr, in isci_request_handle_controller_specific_errors() [all …]
|
/linux-4.1.27/drivers/net/wireless/ath/ |
D | regd.c | 466 struct regulatory_request *request) in __ath_reg_dyn_country() argument 470 if (request->initiator == NL80211_REGDOM_SET_BY_COUNTRY_IE && in __ath_reg_dyn_country() 474 country_code = ath_regd_find_country_by_name(request->alpha2); in __ath_reg_dyn_country() 483 ath_reg_apply_world_flags(wiphy, request->initiator, reg); in __ath_reg_dyn_country() 490 struct regulatory_request *request) in ath_reg_dyn_country() argument 492 if (__ath_reg_dyn_country(wiphy, reg, request)) in ath_reg_dyn_country() 498 reg_initiator_name(request->initiator)); in ath_reg_dyn_country() 502 struct regulatory_request *request, in ath_reg_notifier_apply() argument 515 if (!request) in ath_reg_notifier_apply() 518 reg->region = request->dfs_region; in ath_reg_notifier_apply() [all …]
|
/linux-4.1.27/sound/usb/6fire/ |
D | comm.c | 59 static void usb6fire_comm_init_buffer(u8 *buffer, u8 id, u8 request, in usb6fire_comm_init_buffer() argument 63 buffer[2] = request; in usb6fire_comm_init_buffer() 65 switch (request) { in usb6fire_comm_init_buffer() 110 static int usb6fire_comm_write8(struct comm_runtime *rt, u8 request, in usb6fire_comm_write8() argument 121 usb6fire_comm_init_buffer(buffer, 0x00, request, reg, value, 0x00); in usb6fire_comm_write8() 128 static int usb6fire_comm_write16(struct comm_runtime *rt, u8 request, in usb6fire_comm_write16() argument 139 usb6fire_comm_init_buffer(buffer, 0x00, request, reg, vl, vh); in usb6fire_comm_write16()
|
D | comm.h | 34 int (*write8)(struct comm_runtime *rt, u8 request, u8 reg, u8 value); 35 int (*write16)(struct comm_runtime *rt, u8 request, u8 reg,
|
/linux-4.1.27/drivers/gpu/drm/via/ |
D | via_irq.c | 358 if (irqwait->request.irq >= dev_priv->num_irqs) { in via_wait_irq() 360 irqwait->request.irq); in via_wait_irq() 364 cur_irq += irqwait->request.irq; in via_wait_irq() 366 switch (irqwait->request.type & ~VIA_IRQ_FLAGS_MASK) { in via_wait_irq() 368 irqwait->request.sequence += in via_wait_irq() 370 irqwait->request.type &= ~_DRM_VBLANK_RELATIVE; in via_wait_irq() 377 if (irqwait->request.type & VIA_IRQ_SIGNAL) { in via_wait_irq() 382 force_sequence = (irqwait->request.type & VIA_IRQ_FORCE_SEQUENCE); in via_wait_irq() 384 ret = via_driver_irq_wait(dev, irqwait->request.irq, force_sequence, in via_wait_irq() 385 &irqwait->request.sequence); in via_wait_irq()
|
/linux-4.1.27/drivers/staging/lustre/lustre/mdc/ |
D | mdc_reint.c | 47 static int mdc_reint(struct ptlrpc_request *request, in mdc_reint() argument 53 request->rq_send_state = level; in mdc_reint() 56 rc = ptlrpc_queue_wait(request); in mdc_reint() 60 else if (!req_capsule_server_get(&request->rq_pill, &RMF_MDT_BODY)) in mdc_reint() 105 struct ptlrpc_request **request, struct md_open_data **mod) in mdc_setattr() argument 197 *request = req; in mdc_setattr() 210 struct ptlrpc_request **request) in mdc_create() argument 314 *request = req; in mdc_create() 319 struct ptlrpc_request **request) in mdc_unlink() argument 323 struct ptlrpc_request *req = *request; in mdc_unlink() [all …]
|
D | mdc_internal.h | 135 struct ptlrpc_request **request); 137 struct ptlrpc_request **request); 140 struct ptlrpc_request **request); 143 struct ptlrpc_request **request, struct md_open_data **mod); 145 struct ptlrpc_request **request);
|
/linux-4.1.27/fs/dlm/ |
D | requestqueue.c | 24 struct dlm_message request; member 47 memcpy(&e->request, ms, ms->m_header.h_length); in dlm_add_requestqueue() 82 ms = &e->request; in dlm_process_requestqueue() 90 dlm_receive_message_saved(ls, &e->request, e->recover_seq); in dlm_process_requestqueue() 162 ms = &e->request; in dlm_purge_requestqueue()
|
/linux-4.1.27/arch/parisc/kernel/ |
D | ptrace.c | 114 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 120 switch (request) { in arch_ptrace() 169 ret = ptrace_request(child, request, addr, data); in arch_ptrace() 204 long compat_arch_ptrace(struct task_struct *child, compat_long_t request, in compat_arch_ptrace() argument 210 switch (request) { in compat_arch_ptrace() 238 ret = arch_ptrace(child, request, addr, data); in compat_arch_ptrace() 262 ret = compat_ptrace_request(child, request, addr, data); in compat_arch_ptrace()
|
/linux-4.1.27/drivers/usb/dwc3/ |
D | ep0.c | 113 req->request.actual = 0; in __dwc3_gadget_ep0_queue() 114 req->request.status = -EINPROGRESS; in __dwc3_gadget_ep0_queue() 212 int dwc3_gadget_ep0_queue(struct usb_ep *ep, struct usb_request *request, in dwc3_gadget_ep0_queue() argument 215 struct dwc3_request *req = to_dwc3_request(request); in dwc3_gadget_ep0_queue() 227 request, dep->name); in dwc3_gadget_ep0_queue() 240 request, dep->name, request->length, in dwc3_gadget_ep0_queue() 383 dwc->ep0_usb_req.request.length = sizeof(*response_pkt); in dwc3_ep0_handle_status() 384 dwc->ep0_usb_req.request.buf = dwc->setup_buf; in dwc3_ep0_handle_status() 385 dwc->ep0_usb_req.request.complete = dwc3_ep0_status_cmpl; in dwc3_ep0_handle_status() 667 dwc->ep0_usb_req.request.length = dep->endpoint.maxpacket; in dwc3_ep0_set_sel() [all …]
|
D | gadget.c | 253 } while(++i < req->request.num_mapped_sgs); in dwc3_gadget_giveback() 259 if (req->request.status == -EINPROGRESS) in dwc3_gadget_giveback() 260 req->request.status = status; in dwc3_gadget_giveback() 265 usb_gadget_unmap_request(&dwc->gadget, &req->request, in dwc3_gadget_giveback() 269 req, dep->name, req->request.actual, in dwc3_gadget_giveback() 270 req->request.length, status); in dwc3_gadget_giveback() 274 usb_gadget_giveback_request(&dep->endpoint, &req->request); in dwc3_gadget_giveback() 759 return &req->request; in dwc3_gadget_ep_alloc_request() 763 struct usb_request *request) in dwc3_gadget_ep_free_request() argument 765 struct dwc3_request *req = to_dwc3_request(request); in dwc3_gadget_ep_free_request() [all …]
|
/linux-4.1.27/drivers/net/ethernet/cisco/enic/ |
D | enic_pp.c | 228 if (pp->request != PORT_REQUEST_ASSOCIATE) { in enic_pp_preassociate_rr() 246 if (pp->request != PORT_REQUEST_ASSOCIATE) { in enic_pp_preassociate_rr() 269 if (prev_pp->request != PORT_REQUEST_PREASSOCIATE_RR || in enic_pp_associate() 270 (prev_pp->request == PORT_REQUEST_PREASSOCIATE_RR && in enic_pp_associate() 314 if (pp->request >= enic_pp_handlers_count in enic_process_set_pp_request() 315 || !enic_pp_handlers[pp->request]) in enic_process_set_pp_request() 318 return enic_pp_handlers[pp->request](enic, vf, prev_pp, restore_pp); in enic_process_set_pp_request() 322 int request, u16 *response) in enic_process_get_pp_request() argument 326 switch (request) { in enic_process_get_pp_request()
|
/linux-4.1.27/arch/powerpc/platforms/pseries/ |
D | msi.c | 231 int request; member 261 req = counts->request; in count_spare_msis() 283 static int msi_quota_for_device(struct pci_dev *dev, int request) in msi_quota_for_device() argument 290 request); in msi_quota_for_device() 315 if (request <= counts.quota) in msi_quota_for_device() 320 counts.request = request; in msi_quota_for_device() 332 request = min(counts.quota, request); in msi_quota_for_device() 334 pr_debug("rtas_msi: request clamped to quota %d\n", request); in msi_quota_for_device() 338 return request; in msi_quota_for_device()
|
/linux-4.1.27/Documentation/filesystems/ |
D | fuse.txt | 151 If a process issuing a FUSE filesystem request is interrupted, the 154 1) If the request is not yet sent to userspace AND the signal is 155 fatal (SIGKILL or unhandled fatal signal), then the request is 158 2) If the request is not yet sent to userspace AND the signal is not 159 fatal, then an 'interrupted' flag is set for the request. When 160 the request has been successfully transferred to userspace and 161 this flag is set, an INTERRUPT request is queued. 163 3) If the request is already sent to userspace, then an INTERRUPT 164 request is queued. 170 or may honor them by sending a reply to the _original_ request, with [all …]
|
/linux-4.1.27/drivers/media/usb/dvb-usb/ |
D | dtv5100.c | 37 u8 request; in dtv5100_i2c_msg() local 45 request = (addr == DTV5100_DEMOD_ADDR ? DTV5100_DEMOD_READ : in dtv5100_i2c_msg() 52 request = (addr == DTV5100_DEMOD_ADDR ? DTV5100_DEMOD_WRITE : in dtv5100_i2c_msg() 64 return usb_control_msg(d->udev, usb_rcvctrlpipe(d->udev, 0), request, in dtv5100_i2c_msg() 150 for (i = 0; dtv5100_init[i].request; i++) { in dtv5100_probe() 152 dtv5100_init[i].request, in dtv5100_probe()
|
D | opera1.c | 52 static int opera1_xilinx_rw(struct usb_device *dev, u8 request, u16 value, in opera1_xilinx_rw() argument 68 ret = usb_control_msg(dev, pipe, request, in opera1_xilinx_rw() 72 if (request == OPERA_TUNER_REQ) { in opera1_xilinx_rw() 95 u8 request; in opera1_usb_i2c_msgxfer() local 107 request=0xb6; in opera1_usb_i2c_msgxfer() 111 request=0xb6; in opera1_usb_i2c_msgxfer() 115 request=0xb1; in opera1_usb_i2c_msgxfer() 119 request=0xb8; in opera1_usb_i2c_msgxfer() 123 request=0xb1; in opera1_usb_i2c_msgxfer() 126 ret = opera1_xilinx_rw(dev->udev, request, in opera1_usb_i2c_msgxfer()
|
/linux-4.1.27/drivers/staging/rtl8188eu/os_dep/ |
D | usb_ops_linux.c | 230 static int usbctrl_vendorreq(struct adapter *adapt, u8 request, u16 value, u16 index, void *pdata, … in usbctrl_vendorreq() argument 275 …status = usb_control_msg(udev, pipe, request, reqtype, value, index, pIo_buf, len, RTW_USB_CONTROL… in usbctrl_vendorreq() 315 u8 request; in usb_read8() local 323 request = 0x05; in usb_read8() 330 usbctrl_vendorreq(adapter, request, wvalue, index, &data, len, requesttype); in usb_read8() 339 u8 request; in usb_read16() local 346 request = 0x05; in usb_read16() 351 usbctrl_vendorreq(adapter, request, wvalue, index, &data, len, requesttype); in usb_read16() 358 u8 request; in usb_read32() local 366 request = 0x05; in usb_read32() [all …]
|
/linux-4.1.27/fs/afs/ |
D | cmservice.c | 138 ASSERT(call->server && call->count && call->request); in afs_cm_destructor() 139 afs_break_callbacks(call->server, call->count, call->request); in afs_cm_destructor() 164 afs_break_callbacks(call->server, call->count, call->request); in SRXAFSCB_CallBack() 220 call->request = kcalloc(call->count, in afs_deliver_cb_callback() 223 if (!call->request) in afs_deliver_cb_callback() 226 cb = call->request; in afs_deliver_cb_callback() 259 ret = afs_extract_data(call, skb, last, call->request, in afs_deliver_cb_callback() 268 cb = call->request; in afs_deliver_cb_callback() 432 struct afs_uuid *r = call->request; in SRXAFSCB_ProbeUuid() 487 call->request = kmalloc(sizeof(struct afs_uuid), GFP_KERNEL); in afs_deliver_cb_probe_uuid() [all …]
|
/linux-4.1.27/sound/soc/intel/baytrail/ |
D | sst-baytrail-ipc.c | 154 struct sst_byt_alloc_params request; member 379 stream->request.pcm_params.pcm_wd_sz = bits; in sst_byt_stream_set_bits() 386 stream->request.pcm_params.num_chan = channels; in sst_byt_stream_set_channels() 393 stream->request.pcm_params.sfreq = rate; in sst_byt_stream_set_rate() 401 stream->request.str_type.codec_type = codec_type; in sst_byt_stream_type() 402 stream->request.str_type.str_type = stream_type; in sst_byt_stream_type() 403 stream->request.str_type.operation = operation; in sst_byt_stream_type() 404 stream->request.str_type.time_slots = 0xc; in sst_byt_stream_type() 412 stream->request.frame_info.num_entries = 1; in sst_byt_stream_buffer() 413 stream->request.frame_info.ring_buf_info[0].addr = buffer_addr; in sst_byt_stream_buffer() [all …]
|
/linux-4.1.27/drivers/media/usb/as102/ |
D | as102_fw.c | 125 fw_pkt.u.request[0] = 0x00; in as102_firmware_upload() 126 fw_pkt.u.request[1] = 0x03; in as102_firmware_upload() 137 fw_pkt.u.request[0] = 0x00; in as102_firmware_upload() 138 fw_pkt.u.request[1] = 0x01; in as102_firmware_upload() 140 data_len += sizeof(fw_pkt.u.request); in as102_firmware_upload()
|
/linux-4.1.27/Documentation/video4linux/bttv/ |
D | README.quirks | 25 demultiplexed by the audio request signal. Thus the arbiter defaults to 27 bus access. This is desirable since the video will request the bus more 29 the audio will have first access to the bus even when issuing a request 30 after the video request but before the PCI external arbiter has granted 44 (2) Do not reassert REQ to request another bus transaction until after 57 only the granted request to the REQ pin. The arbiter decision lock 61 responsibility to remove its request at the same time. It is the 62 arbiters responsibility to allow this request to flow through to REQ and 63 not allow the other request to hold REQ asserted. The decision lock may
|
/linux-4.1.27/drivers/media/usb/dvb-usb-v2/ |
D | ec168.c | 32 u8 request, requesttype; in ec168_ctrl_msg() local 41 request = req->cmd; in ec168_ctrl_msg() 45 request = req->cmd; in ec168_ctrl_msg() 49 request = CONFIG; in ec168_ctrl_msg() 53 request = CONFIG; in ec168_ctrl_msg() 57 request = DEMOD_RW; in ec168_ctrl_msg() 61 request = DEMOD_RW; in ec168_ctrl_msg() 87 ret = usb_control_msg(d->udev, pipe, request, requesttype, req->value, in ec168_ctrl_msg() 90 dvb_usb_dbg_usb_control_msg(d->udev, request, requesttype, req->value, in ec168_ctrl_msg()
|
D | au6610.c | 82 u8 request; in au6610_i2c_msg() local 86 request = AU6610_REQ_I2C_WRITE; in au6610_i2c_msg() 88 request = AU6610_REQ_I2C_READ; in au6610_i2c_msg() 91 return au6610_usb_msg(d, request, addr, wbuf, wlen, rbuf, rlen); in au6610_i2c_msg()
|
/linux-4.1.27/Documentation/devicetree/bindings/dma/ |
D | fsl-edma.txt | 5 specific DMA request source can only be multiplexed by any channel of certain 23 Specific request source can only be multiplexed by specific channels 25 The 2nd cell specifies the request source(slot) ID. 26 See the SoC's reference manual for all the supported request sources. 61 specifies the channel group(DMAMUX) in which this request can be multiplexed, 62 and the 2nd specifies the request source.
|
D | fsl-imx-dma.txt | 4 DMA request bindings as described in dma/dma.txt . 35 - dmas: List of one or more DMA request specifiers. One DMA request specifier 37 specifying the request line.
|
D | arm-pl330.txt | 34 [property name] = <[phandle of the dma controller] [dma request id]>; 37 where 'dma request id' is the dma request number which is connected 40 names correspond 1:1 with the dma request ids in the dmas property.
|
D | k3dma.txt | 12 have specific request line 30 Use specific request line passing from dmax 31 For example, i2c0 read channel request line is 18, while write channel use 19
|
D | moxa,moxart-dma.txt | 11 - #dma-cells : Should be 1, a single cell holding a line request number 34 Use specific request line passing from dma 35 For example, MMC request line is 5
|
D | sirfsoc-dma.txt | 26 Fill the specific dma request line in dmas. In the below example, spi0 read 27 channel request line is 9 of the 2nd dma controller, while write channel uses 28 4 of the 2nd dma controller; spi1 read channel request line is 12 of the 1st
|
D | dma.txt | 1 * Generic DMA Controller and DMA request bindings 4 DMA request or channel information that goes from a hardware device to a DMA 17 - dma-requests: Number of DMA request signals supported by the 46 typically contains a DMA request line number or a
|
/linux-4.1.27/arch/c6x/kernel/ |
D | ptrace.c | 142 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 147 switch (request) { in arch_ptrace() 153 if (ret == 0 && request == PTRACE_POKETEXT) in arch_ptrace() 157 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/drivers/md/ |
D | dm-target.c | 134 static int io_err_map_rq(struct dm_target *ti, struct request *clone, in io_err_map_rq() 140 static int io_err_clone_and_map_rq(struct dm_target *ti, struct request *rq, in io_err_clone_and_map_rq() 142 struct request **clone) in io_err_clone_and_map_rq() 147 static void io_err_release_clone_rq(struct request *clone) in io_err_release_clone_rq()
|
D | dm.c | 85 struct request *orig, *clone; 105 union map_info *dm_get_rq_mapinfo(struct request *rq) in dm_get_rq_mapinfo() 323 _rq_cache = kmem_cache_create("dm_clone_request", sizeof(struct request), in local_init() 324 __alignof__(struct request), 0, NULL); in local_init() 623 static struct request *alloc_clone_request(struct mapped_device *md, in alloc_clone_request() 629 static void free_clone_request(struct mapped_device *md, struct request *rq) in free_clone_request() 1044 static struct dm_rq_target_io *tio_from_request(struct request *rq) in tio_from_request() 1077 static void free_rq_clone(struct request *clone) in free_rq_clone() 1105 static void dm_end_request(struct request *clone, int error) in dm_end_request() 1110 struct request *rq = tio->orig; in dm_end_request() [all …]
|
/linux-4.1.27/drivers/net/wireless/rt2x00/ |
D | rt2x00usb.h | 103 const u8 request, const u8 requesttype, 126 const u8 request, const u8 requesttype, 144 const u8 request, const u8 requesttype, 161 const u8 request, in rt2x00usb_vendor_request_sw() argument 166 return rt2x00usb_vendor_request(rt2x00dev, request, in rt2x00usb_vendor_request_sw()
|
/linux-4.1.27/drivers/media/rc/ |
D | igorplugusb.c | 46 struct usb_ctrlrequest request; member 133 ir->request.bRequest = cmd; in igorplugusb_cmd() 179 ir->request.bRequest = GET_INFRACODE; in igorplugusb_probe() 180 ir->request.bRequestType = USB_TYPE_VENDOR | USB_DIR_IN; in igorplugusb_probe() 181 ir->request.wLength = cpu_to_le16(sizeof(ir->buf_in)); in igorplugusb_probe() 188 usb_rcvctrlpipe(udev, 0), (uint8_t *)&ir->request, in igorplugusb_probe()
|
/linux-4.1.27/arch/tile/kernel/ |
D | ptrace.c | 150 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 159 switch (request) { in arch_ptrace() 221 ret = ptrace_request(child, request, addr, data); in arch_ptrace() 233 ret = compat_ptrace_request(child, request, in arch_ptrace() 238 ret = ptrace_request(child, request, addr, data); in arch_ptrace() 247 long compat_arch_ptrace(struct task_struct *child, compat_long_t request, in compat_arch_ptrace() argument
|
/linux-4.1.27/arch/um/kernel/ |
D | ptrace.c | 43 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 50 switch (request) { in arch_ptrace() 105 ret = ptrace_request(child, request, addr, data); in arch_ptrace() 107 ret = subarch_ptrace(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/drivers/input/serio/ |
D | hyperv-keyboard.c | 288 struct synth_kbd_protocol_request *request; in hv_kbd_connect_to_vsp() local 293 request = &kbd_dev->protocol_req; in hv_kbd_connect_to_vsp() 294 memset(request, 0, sizeof(struct synth_kbd_protocol_request)); in hv_kbd_connect_to_vsp() 295 request->header.type = __cpu_to_le32(SYNTH_KBD_PROTOCOL_REQUEST); in hv_kbd_connect_to_vsp() 296 request->version_requested.version = __cpu_to_le32(SYNTH_KBD_VERSION); in hv_kbd_connect_to_vsp() 298 error = vmbus_sendpacket(hv_dev->channel, request, in hv_kbd_connect_to_vsp() 300 (unsigned long)request, in hv_kbd_connect_to_vsp()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/ |
D | regd.c | 307 struct regulatory_request *request, in _rtl_reg_notifier_apply() argument 313 switch (request->initiator) { in _rtl_reg_notifier_apply() 319 _rtl_reg_apply_world_flags(wiphy, request->initiator, reg); in _rtl_reg_notifier_apply() 360 request)) in _rtl_regd_init_wiphy() 389 struct regulatory_request *request)) in rtl_regd_init() argument 431 void rtl_reg_notifier(struct wiphy *wiphy, struct regulatory_request *request) in rtl_reg_notifier() argument 438 _rtl_reg_notifier_apply(wiphy, request, &rtlpriv->regd); in rtl_reg_notifier()
|
D | regd.h | 59 struct regulatory_request *request)); 60 void rtl_reg_notifier(struct wiphy *wiphy, struct regulatory_request *request);
|
/linux-4.1.27/drivers/mmc/card/ |
D | queue.h | 6 struct request; 35 struct request *req; 54 int (*issue_fn)(struct mmc_queue *, struct request *);
|
/linux-4.1.27/drivers/ide/ |
D | ide-pm.c | 10 struct request *rq; in generic_ide_suspend() 46 struct request *rq; in generic_ide_resume() 81 void ide_complete_power_step(ide_drive_t *drive, struct request *rq) in ide_complete_power_step() 111 ide_startstop_t ide_start_power_step(ide_drive_t *drive, struct request *rq) in ide_start_power_step() 182 void ide_complete_pm_rq(ide_drive_t *drive, struct request *rq) in ide_complete_pm_rq() 209 void ide_check_pm_state(ide_drive_t *drive, struct request *rq) in ide_check_pm_state()
|
D | ide-io.c | 57 int ide_end_rq(ide_drive_t *drive, struct request *rq, int error, in ide_end_rq() 78 struct request *rq = cmd->rq; in ide_complete_cmd() 118 struct request *rq = hwif->rq; in ide_complete_rq() 136 void ide_kill_rq(ide_drive_t *drive, struct request *rq) in ide_kill_rq() 227 struct request *rq = cmd->rq; in ide_map_sg() 254 struct request *rq) in execute_drive_cmd() 280 static ide_startstop_t ide_special_rq(ide_drive_t *drive, struct request *rq) in ide_special_rq() 306 static ide_startstop_t start_request (ide_drive_t *drive, struct request *rq) in start_request() 433 static void __ide_requeue_and_plug(struct request_queue *q, struct request *rq) in __ide_requeue_and_plug() 443 void ide_requeue_and_plug(ide_drive_t *drive, struct request *rq) in ide_requeue_and_plug() [all …]
|
D | ide-atapi.c | 92 struct request *rq; in ide_queue_pc_tail() 171 void ide_prep_sense(ide_drive_t *drive, struct request *rq) in ide_prep_sense() 174 struct request *sense_rq = &drive->sense_rq; in ide_prep_sense() 249 struct request *failed_rq = drive->hwif->rq; in ide_retry_pc() 250 struct request *sense_rq = &drive->sense_rq; in ide_retry_pc() 278 struct request *rq = drive->hwif->rq; in ide_cd_expiry() 308 int ide_cd_get_xferlen(struct request *rq) in ide_cd_get_xferlen() 343 int ide_check_ireason(ide_drive_t *drive, struct request *rq, int len, in ide_check_ireason() 394 struct request *rq = hwif->rq; in ide_pc_intr() 586 struct request *rq = hwif->rq; in ide_transfer_pc() [all …]
|
/linux-4.1.27/drivers/staging/wlan-ng/ |
D | cfg80211.c | 337 struct cfg80211_scan_request *request) in prism2_scan() argument 352 if (!request) in prism2_scan() 355 dev = request->wdev->netdev; in prism2_scan() 358 if (priv->scan_request && priv->scan_request != request) in prism2_scan() 366 priv->scan_request = request; in prism2_scan() 375 if (request->n_ssids > 0) { in prism2_scan() 377 msg1.ssid.data.len = request->ssids->ssid_len; in prism2_scan() 379 request->ssids->ssid, request->ssids->ssid_len); in prism2_scan() 386 (i < request->n_channels) && i < ARRAY_SIZE(prism2_channels); in prism2_scan() 390 request->channels[i]->center_freq); in prism2_scan() [all …]
|
/linux-4.1.27/drivers/media/usb/au0828/ |
D | au0828-core.c | 50 static int send_control_msg(struct au0828_dev *dev, u16 request, u32 value, 52 static int recv_control_msg(struct au0828_dev *dev, u16 request, u32 value, 75 static int send_control_msg(struct au0828_dev *dev, u16 request, u32 value, in send_control_msg() argument 85 request, in send_control_msg() 102 static int recv_control_msg(struct au0828_dev *dev, u16 request, u32 value, in recv_control_msg() argument 110 request, in recv_control_msg()
|
/linux-4.1.27/arch/s390/include/asm/ |
D | pci_clp.h | 167 struct clp_req_list_pci request; member 172 struct clp_req_set_pci request; member 177 struct clp_req_query_pci request; member 182 struct clp_req_query_pci_grp request; member
|
/linux-4.1.27/Documentation/power/ |
D | pm_qos_interface.txt | 30 changes to the request list or elements of the list. Typically the 31 aggregated target value is simply the max or min of the request values held 54 the request. 60 Returns if the request is still active, i.e. it has not been removed from a 72 Only processes can register a pm_qos request. To provide for automatic 80 request on the parameter. 87 To remove the user mode request for a target value simply close the device 96 Values are updated in response to changes of the request list. 99 simply the minimum of the request values held in the parameter list elements. 125 the request. [all …]
|
/linux-4.1.27/include/trace/events/ |
D | block.h | 66 TP_PROTO(struct request_queue *q, struct request *rq), 110 TP_PROTO(struct request_queue *q, struct request *rq), 126 TP_PROTO(struct request_queue *q, struct request *rq), 145 TP_PROTO(struct request_queue *q, struct request *rq, 178 TP_PROTO(struct request_queue *q, struct request *rq), 225 TP_PROTO(struct request_queue *q, struct request *rq), 240 TP_PROTO(struct request_queue *q, struct request *rq), 324 TP_PROTO(struct request_queue *q, struct request *rq, struct bio *bio), 361 TP_PROTO(struct request_queue *q, struct request *rq, struct bio *bio), 377 TP_PROTO(struct request_queue *q, struct request *rq, struct bio *bio), [all …]
|
/linux-4.1.27/Documentation/security/ |
D | keys-request-key.txt | 5 The key request service is part of the key retention service (refer to 54 own upcall mechanism rather than using /sbin/request-key. 67 forking and execution of /sbin/request-key. 74 A request proceeds in the following manner: 93 (4) request_key() then forks and executes /sbin/request-key with a new session 96 (5) /sbin/request-key assumes the authority associated with key U. 98 (6) /sbin/request-key execs an appropriate program to perform the actual 121 3) and another copy of /sbin/request-key spawned (as per step 4); but the 126 /sbin/request-key at the appropriate places because (a) execve will discard two 140 This is provided to prevent excessive repeated spawning of /sbin/request-key [all …]
|
/linux-4.1.27/sound/usb/ |
D | helper.c | 83 int snd_usb_ctl_msg(struct usb_device *dev, unsigned int pipe, __u8 request, in snd_usb_ctl_msg() argument 102 err = usb_control_msg(dev, pipe, request, requesttype, in snd_usb_ctl_msg() 110 snd_usb_ctl_msg_quirk(dev, pipe, request, requesttype, in snd_usb_ctl_msg()
|
/linux-4.1.27/arch/unicore32/kernel/ |
D | ptrace.c | 87 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 93 switch (request) { in arch_ptrace() 108 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/drivers/net/wireless/orinoco/ |
D | cfg.c | 143 struct cfg80211_scan_request *request) in orinoco_scan() argument 148 if (!request) in orinoco_scan() 151 if (priv->scan_request && priv->scan_request != request) in orinoco_scan() 154 priv->scan_request = request; in orinoco_scan() 156 err = orinoco_hw_trigger_scan(priv, request->ssids); in orinoco_scan()
|
/linux-4.1.27/drivers/gpu/drm/msm/edp/ |
D | edp_aux.c | 45 bool native = msg->request & (DP_AUX_NATIVE_WRITE & DP_AUX_NATIVE_READ); in edp_msg_fifo_tx() 46 bool read = msg->request & (DP_AUX_I2C_READ & DP_AUX_NATIVE_READ); in edp_msg_fifo_tx() 122 bool native = msg->request & (DP_AUX_NATIVE_WRITE & DP_AUX_NATIVE_READ); in edp_aux_transfer() 123 bool read = msg->request & (DP_AUX_I2C_READ & DP_AUX_NATIVE_READ); in edp_aux_transfer() 136 __func__, msg->size, msg->request); in edp_aux_transfer()
|
/linux-4.1.27/arch/arc/kernel/ |
D | ptrace.c | 141 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 146 pr_debug("REQ=%ld: ADDR =0x%lx, DATA=0x%lx)\n", request, addr, data); in arch_ptrace() 148 switch (request) { in arch_ptrace() 154 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/Documentation/networking/ |
D | dns_resolver.txt | 21 upcalled to userspace through /sbin/request-key. 24 request-key. It is under development and does not yet provide the full feature 49 To set up this facility, the /etc/request-key.conf file must be altered so that 50 /sbin/request-key can appropriately direct the upcalls. For example, to handle 130 directed by means of configuration lines in /etc/request-key.conf that tell 131 /sbin/request-key what program to run to instantiate the key. 146 See <file:Documentation/security/keys-request-key.txt> for further 147 information about request-key function.
|
/linux-4.1.27/fs/ncpfs/ |
D | ioctl.c | 261 struct ncp_ioctl_request request; in __ncp_ioctl() local 275 request.function = request32.function; in __ncp_ioctl() 276 request.size = request32.size; in __ncp_ioctl() 277 request.data = compat_ptr(request32.data); in __ncp_ioctl() 280 if (copy_from_user(&request, argp, sizeof(request))) in __ncp_ioctl() 283 if ((request.function > 255) in __ncp_ioctl() 284 || (request.size > in __ncp_ioctl() 291 if (copy_from_user(bouncebuffer, request.data, request.size)) { in __ncp_ioctl() 301 server->current_size = request.size; in __ncp_ioctl() 302 memcpy(server->packet, bouncebuffer, request.size); in __ncp_ioctl() [all …]
|
/linux-4.1.27/arch/cris/arch-v10/kernel/ |
D | ptrace.c | 78 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 85 switch (request) { in arch_ptrace() 174 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/drivers/soc/tegra/ |
D | pmc.c | 452 static int tegra_io_rail_prepare(int id, unsigned long *request, in tegra_io_rail_prepare() argument 469 *request = IO_DPD_REQ; in tegra_io_rail_prepare() 472 *request = IO_DPD2_REQ; in tegra_io_rail_prepare() 517 unsigned long request, status, value; in tegra_io_rail_power_on() local 521 err = tegra_io_rail_prepare(id, &request, &status, &bit); in tegra_io_rail_power_on() 527 value = tegra_pmc_readl(request); in tegra_io_rail_power_on() 531 tegra_pmc_writel(value, request); in tegra_io_rail_power_on() 545 unsigned long request, status, value; in tegra_io_rail_power_off() local 549 err = tegra_io_rail_prepare(id, &request, &status, &bit); in tegra_io_rail_power_off() 555 value = tegra_pmc_readl(request); in tegra_io_rail_power_off() [all …]
|
/linux-4.1.27/arch/um/kernel/skas/ |
D | process.c | 40 init_task.thread.request.u.thread.proc = start_kernel_proc; in start_uml() 41 init_task.thread.request.u.thread.arg = NULL; in start_uml()
|
/linux-4.1.27/Documentation/s390/ |
D | cds.txt | 64 initiate an I/O request. 70 terminate the current I/O request processed on the device. 76 interrupt handler according to the rules (flags) defined during I/O request 143 armed) interrupt request lines (IRQs) to drive its devices with. Where 150 and initiates an I/O request on behalf of the device driver. The 156 For long running I/O request to be canceled, the CDS layer provides the 180 The ccw_device_start() routines is the I/O request front-end processor. All 187 with the associated I/O request when calling ccw_device_start(). 219 particular I/O request. 221 request. A value of 0 will make cio use the opm. [all …]
|
/linux-4.1.27/drivers/staging/lustre/lustre/include/ |
D | obd_class.h | 1483 struct ptlrpc_request **request) in md_getattr() argument 1489 rc = MDP(exp->exp_obd, getattr)(exp, op_data, request); in md_getattr() 1518 struct ptlrpc_request **request) in md_close() argument 1524 rc = MDP(exp->exp_obd, close)(exp, op_data, mod, request); in md_close() 1531 struct ptlrpc_request **request) in md_create() argument 1538 uid, gid, cap_effective, rdev, request); in md_create() 1574 struct ptlrpc_request **request) in md_getattr_name() argument 1580 rc = MDP(exp->exp_obd, getattr_name)(exp, op_data, request); in md_getattr_name() 1602 struct ptlrpc_request **request) in md_link() argument 1608 rc = MDP(exp->exp_obd, link)(exp, op_data, request); in md_link() [all …]
|
/linux-4.1.27/arch/openrisc/kernel/ |
D | ptrace.c | 159 long arch_ptrace(struct task_struct *child, long request, unsigned long addr, in arch_ptrace() argument 164 switch (request) { in arch_ptrace() 166 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/drivers/scsi/libsas/ |
D | sas_internal.h | 103 extern int sas_smp_host_handler(struct Scsi_Host *shost, struct request *req, 104 struct request *rsp); 107 struct request *req, in sas_smp_host_handler() 108 struct request *rsp) in sas_smp_host_handler()
|
/linux-4.1.27/sound/soc/intel/haswell/ |
D | sst-haswell-ipc.c | 220 struct sst_hsw_ipc_stream_alloc_req request; member 1050 stream->request.format.bitdepth = bits; in sst_hsw_stream_set_bits() 1062 stream->request.format.ch_num = channels; in sst_hsw_stream_set_channels() 1074 stream->request.format.frequency = rate; in sst_hsw_stream_set_rate() 1087 stream->request.format.map = map; in sst_hsw_stream_set_map_config() 1088 stream->request.format.config = config; in sst_hsw_stream_set_map_config() 1100 stream->request.format.style = style; in sst_hsw_stream_set_style() 1112 stream->request.format.valid_bit = bits; in sst_hsw_stream_set_valid() 1127 stream->request.path_id = path_id; in sst_hsw_stream_format() 1128 stream->request.stream_type = stream_type; in sst_hsw_stream_format() [all …]
|
/linux-4.1.27/drivers/net/wireless/ |
D | zd1201.c | 517 unsigned char *request; in zd1201_setconfig() local 528 request = kmalloc(16, gfp_mask); in zd1201_setconfig() 529 if (!request) in zd1201_setconfig() 533 kfree(request); in zd1201_setconfig() 536 memset(request, 0, 16); in zd1201_setconfig() 538 request[0] = ZD1201_USB_RESREQ; in zd1201_setconfig() 539 request[1] = seq; in zd1201_setconfig() 540 request[2] = 0; in zd1201_setconfig() 541 request[3] = 0; in zd1201_setconfig() 542 if (request[1] == 0) { in zd1201_setconfig() [all …]
|
/linux-4.1.27/drivers/base/power/ |
D | runtime.c | 109 dev->power.request = RPM_REQ_NONE; in pm_runtime_cancel_pending() 250 && dev->power.request == RPM_REQ_RESUME)) in rpm_check_suspend_allowed() 317 dev->power.request > RPM_REQ_IDLE) in rpm_idle() 327 dev->power.request = RPM_REQ_NONE; in rpm_idle() 334 dev->power.request = RPM_REQ_IDLE; in rpm_idle() 443 dev->power.request = RPM_REQ_NONE; in rpm_suspend() 504 dev->power.request = (rpmflags & RPM_AUTO) ? in rpm_suspend() 620 dev->power.request = RPM_REQ_NONE; in rpm_resume() 688 dev->power.request = RPM_REQ_RESUME; in rpm_resume() 783 req = dev->power.request; in pm_runtime_work() [all …]
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | ptrace32.c | 47 long compat_arch_ptrace(struct task_struct *child, compat_long_t request, in compat_arch_ptrace() argument 54 switch (request) { in compat_arch_ptrace() 308 ret = arch_ptrace(child, request, addr, data); in compat_arch_ptrace() 312 ret = compat_ptrace_request(child, request, addr, data); in compat_arch_ptrace()
|
/linux-4.1.27/drivers/usb/serial/ |
D | cp210x.c | 319 static int cp210x_get_config(struct usb_serial_port *port, u8 request, in cp210x_get_config() argument 336 request, REQTYPE_INTERFACE_TO_HOST, 0x0000, in cp210x_get_config() 348 __func__, request, size, result); in cp210x_get_config() 364 static int cp210x_set_config(struct usb_serial_port *port, u8 request, in cp210x_set_config() argument 386 request, REQTYPE_HOST_TO_INTERFACE, 0x0000, in cp210x_set_config() 392 request, REQTYPE_HOST_TO_INTERFACE, data[0], in cp210x_set_config() 401 __func__, request, size, result); in cp210x_set_config() 417 u8 request, unsigned int data) in cp210x_set_config_single() argument 419 return cp210x_set_config(port, request, &data, 2); in cp210x_set_config_single()
|
/linux-4.1.27/arch/blackfin/kernel/ |
D | ptrace.c | 247 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument 254 switch (request) { in arch_ptrace() 352 request = PTRACE_GETFDPIC; in arch_ptrace() 356 request = PTRACE_GETFDPIC; in arch_ptrace() 387 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
|
/linux-4.1.27/drivers/gpu/drm/tegra/ |
D | dpaux.c | 122 switch (msg->request & ~DP_AUX_I2C_MOT) { in tegra_dpaux_transfer() 136 switch (msg->request & ~DP_AUX_I2C_MOT) { in tegra_dpaux_transfer() 138 if (msg->request & DP_AUX_I2C_MOT) in tegra_dpaux_transfer() 146 if (msg->request & DP_AUX_I2C_MOT) in tegra_dpaux_transfer() 154 if (msg->request & DP_AUX_I2C_MOT) in tegra_dpaux_transfer() 176 if ((msg->request & DP_AUX_I2C_READ) == 0) { in tegra_dpaux_transfer() 225 if (msg->request & DP_AUX_I2C_READ) { in tegra_dpaux_transfer()
|
/linux-4.1.27/Documentation/device-mapper/ |
D | log-writes.txt | 7 There is a log_write_entry written for every WRITE request and the target is 17 next REQ_FLUSH request. This is to make it easier for userspace to replay the 22 Once we see a REQ_FLUSH request we splice this list onto the request and once 23 the FLUSH request completes we log all of the WRITEs and then the FLUSH. Only 43 request. Consider the following example: 138 And that will replay the log until it sees a FUA request, run the fsck command
|
/linux-4.1.27/Documentation/devicetree/bindings/sound/ |
D | sirf-audio-port.txt | 6 - dmas: List of DMA controller phandle and DMA request line ordered pairs. 7 - dma-names: Identifier string for each DMA request line in the dmas property.
|