Home
last modified time | relevance | path

Searched refs:request (Results 1 – 200 of 1169) sorted by relevance

123456

/linux-4.1.27/drivers/s390/char/
Dtape_std.c37 struct tape_request * request; in tape_std_assign_timeout() local
41 request = (struct tape_request *) data; in tape_std_assign_timeout()
42 device = request->device; in tape_std_assign_timeout()
47 rc = tape_cancel_io(device, request); in tape_std_assign_timeout()
58 struct tape_request *request; in tape_std_assign() local
60 request = tape_alloc_request(2, 11); in tape_std_assign()
61 if (IS_ERR(request)) in tape_std_assign()
62 return PTR_ERR(request); in tape_std_assign()
64 request->op = TO_ASSIGN; in tape_std_assign()
65 tape_ccw_cc(request->cpaddr, ASSIGN, 11, request->cpdata); in tape_std_assign()
[all …]
Dtape_34xx.c55 static void __tape_34xx_medium_sense(struct tape_request *request) in __tape_34xx_medium_sense() argument
57 struct tape_device *device = request->device; in __tape_34xx_medium_sense()
60 if (request->rc == 0) { in __tape_34xx_medium_sense()
61 sense = request->cpdata; in __tape_34xx_medium_sense()
80 request->rc); in __tape_34xx_medium_sense()
81 tape_free_request(request); in __tape_34xx_medium_sense()
86 struct tape_request *request; in tape_34xx_medium_sense() local
89 request = tape_alloc_request(1, 32); in tape_34xx_medium_sense()
90 if (IS_ERR(request)) { in tape_34xx_medium_sense()
92 return PTR_ERR(request); in tape_34xx_medium_sense()
[all …]
Dtape_3590.c203 struct tape_request *request; in tape_3592_kekl_query() local
212 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query()
213 if (IS_ERR(request)) { in tape_3592_kekl_query()
214 rc = PTR_ERR(request); in tape_3592_kekl_query()
217 order = request->cpdata; in tape_3592_kekl_query()
221 request->op = TO_KEKL_QUERY; in tape_3592_kekl_query()
222 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query()
223 tape_ccw_end(request->cpaddr + 1, READ_SS_DATA, sizeof(*int_kekls), in tape_3592_kekl_query()
225 rc = tape_do_io(device, request); in tape_3592_kekl_query()
232 tape_free_request(request); in tape_3592_kekl_query()
[all …]
Dtape_core.c291 __tape_cancel_io(struct tape_device *device, struct tape_request *request) in __tape_cancel_io() argument
297 if (request->callback == NULL) in __tape_cancel_io()
302 rc = ccw_device_clear(device->cdev, (long) request); in __tape_cancel_io()
306 request->status = TAPE_REQUEST_DONE; in __tape_cancel_io()
309 request->status = TAPE_REQUEST_CANCEL; in __tape_cancel_io()
641 struct tape_request * request; in __tape_discard_requests() local
645 request = list_entry(l, struct tape_request, list); in __tape_discard_requests()
646 if (request->status == TAPE_REQUEST_IN_IO) in __tape_discard_requests()
647 request->status = TAPE_REQUEST_DONE; in __tape_discard_requests()
648 list_del(&request->list); in __tape_discard_requests()
[all …]
Dsclp_vt220.c97 static int __sclp_vt220_emit(struct sclp_vt220_request *request);
118 sclp_vt220_process_queue(struct sclp_vt220_request *request) in sclp_vt220_process_queue() argument
125 page = request->sclp_req.sccb; in sclp_vt220_process_queue()
128 list_del(&request->list); in sclp_vt220_process_queue()
131 request = NULL; in sclp_vt220_process_queue()
133 request = list_entry(sclp_vt220_outqueue.next, in sclp_vt220_process_queue()
135 if (!request || sclp_vt220_suspended) { in sclp_vt220_process_queue()
141 } while (__sclp_vt220_emit(request)); in sclp_vt220_process_queue()
142 if (request == NULL && sclp_vt220_flush_later) in sclp_vt220_process_queue()
154 sclp_vt220_callback(struct sclp_req *request, void *data) in sclp_vt220_callback() argument
[all …]
Dsclp_sdias.c53 static void sdias_callback(struct sclp_req *request, void *data) in sdias_callback() argument
107 struct sclp_req request; in sclp_sdias_blk_count() local
113 memset(&request, 0, sizeof(request)); in sclp_sdias_blk_count()
123 request.sccb = &sccb; in sclp_sdias_blk_count()
124 request.command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_sdias_blk_count()
125 request.status = SCLP_REQ_FILLED; in sclp_sdias_blk_count()
126 request.callback = sdias_callback; in sclp_sdias_blk_count()
128 rc = sdias_sclp_send(&request); in sclp_sdias_blk_count()
166 struct sclp_req request; in sclp_sdias_copy() local
172 memset(&request, 0, sizeof(request)); in sclp_sdias_copy()
[all …]
Dsclp_async.c25 static struct sclp_req *request; variable
132 request->command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_async_send_wait()
133 request->sccb = sccb; in sclp_async_send_wait()
134 request->status = SCLP_REQ_FILLED; in sclp_async_send_wait()
145 rc = sclp_add_request(request); in sclp_async_send_wait()
149 while (request->status != SCLP_REQ_DONE && in sclp_async_send_wait()
150 request->status != SCLP_REQ_FAILED) { in sclp_async_send_wait()
154 if (request->status != SCLP_REQ_DONE) in sclp_async_send_wait()
157 request->sccb)->header.response_code; in sclp_async_send_wait()
179 request = kzalloc(sizeof(struct sclp_req), GFP_KERNEL); in sclp_async_init()
[all …]
Dtape_char.c131 struct tape_request *request; in tapechar_read() local
165 request = device->discipline->read_block(device, block_size); in tapechar_read()
166 if (IS_ERR(request)) in tapechar_read()
167 return PTR_ERR(request); in tapechar_read()
169 rc = tape_do_io(device, request); in tapechar_read()
171 rc = block_size - request->rescnt; in tapechar_read()
178 tape_free_request(request); in tapechar_read()
189 struct tape_request *request; in tapechar_write() local
218 request = device->discipline->write_block(device, block_size); in tapechar_write()
219 if (IS_ERR(request)) in tapechar_write()
[all …]
Dsclp_rw.c377 sclp_writedata_callback(struct sclp_req *request, void *data) in sclp_writedata_callback() argument
386 if (request->status == SCLP_REQ_FAILED) { in sclp_writedata_callback()
407 buffer->request.status = SCLP_REQ_FILLED; in sclp_writedata_callback()
408 rc = sclp_add_request(request); in sclp_writedata_callback()
423 buffer->request.status = SCLP_REQ_FILLED; in sclp_writedata_callback()
424 rc = sclp_add_request(request); in sclp_writedata_callback()
462 buffer->request.command = SCLP_CMDW_WRITE_EVENT_DATA; in sclp_emit_buffer()
463 buffer->request.status = SCLP_REQ_FILLED; in sclp_emit_buffer()
464 buffer->request.callback = sclp_writedata_callback; in sclp_emit_buffer()
465 buffer->request.callback_data = buffer; in sclp_emit_buffer()
[all …]
Dsclp_cmd.c46 struct sclp_req *request; in sclp_sync_request_timeout() local
49 request = kzalloc(sizeof(*request), GFP_KERNEL); in sclp_sync_request_timeout()
50 if (!request) in sclp_sync_request_timeout()
53 request->queue_timeout = timeout; in sclp_sync_request_timeout()
54 request->command = cmd; in sclp_sync_request_timeout()
55 request->sccb = sccb; in sclp_sync_request_timeout()
56 request->status = SCLP_REQ_FILLED; in sclp_sync_request_timeout()
57 request->callback = sclp_sync_callback; in sclp_sync_request_timeout()
58 request->callback_data = &completion; in sclp_sync_request_timeout()
62 rc = sclp_add_request(request); in sclp_sync_request_timeout()
[all …]
Dtape.h242 tape_do_io_free(struct tape_device *device, struct tape_request *request) in tape_do_io_free() argument
246 rc = tape_do_io(device, request); in tape_do_io_free()
247 tape_free_request(request); in tape_do_io_free()
252 tape_do_io_async_free(struct tape_device *device, struct tape_request *request) in tape_do_io_async_free() argument
254 request->callback = (void *) tape_free_request; in tape_do_io_async_free()
255 request->callback_data = NULL; in tape_do_io_async_free()
256 tape_do_io_async(device, request); in tape_do_io_async_free()
Dtape_proc.c41 struct tape_request *request; in tape_proc_show() local
71 request = list_entry(device->req_queue.next, in tape_proc_show()
73 str = tape_op_verbose[request->op]; in tape_proc_show()
/linux-4.1.27/include/linux/
Delevator.h12 typedef int (elevator_merge_fn) (struct request_queue *, struct request **,
15 typedef void (elevator_merge_req_fn) (struct request_queue *, struct request *, struct request *);
17 typedef void (elevator_merged_fn) (struct request_queue *, struct request *, int);
19 typedef int (elevator_allow_merge_fn) (struct request_queue *, struct request *, struct bio *);
22 struct request *, struct bio *);
26 typedef void (elevator_add_req_fn) (struct request_queue *, struct request *);
27 typedef struct request *(elevator_request_list_fn) (struct request_queue *, struct request *);
28 typedef void (elevator_completed_req_fn) (struct request_queue *, struct request *);
33 typedef int (elevator_set_req_fn) (struct request_queue *, struct request *,
35 typedef void (elevator_put_req_fn) (struct request *);
[all …]
Dblk-mq.h82 struct request *rq;
89 typedef enum blk_eh_timer_return (timeout_fn)(struct request *, bool);
92 typedef int (init_request_fn)(void *, struct request *, unsigned int,
94 typedef void (exit_request_fn)(void *, struct request *, unsigned int,
97 typedef void (busy_iter_fn)(struct blk_mq_hw_ctx *, struct request *, void *,
178 void blk_mq_insert_request(struct request *, bool, bool, bool);
179 void blk_mq_free_request(struct request *rq);
180 void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *, struct request *rq);
182 struct request *blk_mq_alloc_request(struct request_queue *q, int rw,
184 struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag);
[all …]
Dblkdev.h35 struct request;
50 struct request;
51 typedef void (rq_end_io_fn)(struct request *, int);
100 struct request { struct
211 struct request *next_rq; argument
214 static inline unsigned short req_get_ioprio(struct request *req) in req_get_ioprio()
238 typedef int (prep_rq_fn) (struct request_queue *, struct request *);
239 typedef void (unprep_rq_fn) (struct request_queue *, struct request *);
250 typedef void (softirq_done_fn)(struct request *);
251 typedef int (dma_drain_needed_fn)(struct request *);
[all …]
Dbsg-lib.h28 struct request;
41 struct request *req;
44 void *request; member
Dide.h309 struct request *rq; /* copy of request */
337 struct request *rq;
367 ide_startstop_t (*do_request)(struct ide_drive_s *, struct request *,
488 struct request *rq; /* current request */
569 struct request sense_rq;
697 void (*rw_disk)(ide_drive_t *, struct request *);
754 struct request *rq;
1002 void ide_complete_power_step(ide_drive_t *, struct request *);
1003 ide_startstop_t ide_start_power_step(ide_drive_t *, struct request *);
1004 void ide_complete_pm_rq(ide_drive_t *, struct request *);
[all …]
Dblktrace_api.h60 extern void blk_add_driver_data(struct request_queue *q, struct request *rq,
106 static inline int blk_cmd_buf_len(struct request *rq) in blk_cmd_buf_len()
111 extern void blk_dump_cmd(char *buf, struct request *rq);
Ddevice-mapper.h49 typedef int (*dm_map_request_fn) (struct dm_target *ti, struct request *clone,
52 struct request *rq,
54 struct request **clone);
55 typedef void (*dm_release_clone_request_fn) (struct request *clone);
68 struct request *clone, int error,
410 union map_info *dm_get_rq_mapinfo(struct request *rq);
/linux-4.1.27/drivers/gpu/drm/radeon/
Dradeon_ioc32.c137 drm_radeon_stipple_t __user *request; in compat_radeon_cp_stipple() local
143 request = compat_alloc_user_space(sizeof(*request)); in compat_radeon_cp_stipple()
144 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_radeon_cp_stipple()
146 &request->mask)) in compat_radeon_cp_stipple()
149 return drm_ioctl(file, DRM_IOCTL_RADEON_STIPPLE, (unsigned long)request); in compat_radeon_cp_stipple()
171 drm_radeon_texture_t __user *request; in compat_radeon_cp_texture() local
183 request = compat_alloc_user_space(sizeof(*request) + sizeof(*image)); in compat_radeon_cp_texture()
184 if (!access_ok(VERIFY_WRITE, request, in compat_radeon_cp_texture()
185 sizeof(*request) + sizeof(*image))) in compat_radeon_cp_texture()
187 image = (drm_radeon_tex_image_t __user *) (request + 1); in compat_radeon_cp_texture()
[all …]
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/
Dniobuf.c488 int ptl_send_rpc(struct ptlrpc_request *request, int noreply) in ptl_send_rpc() argument
496 struct obd_device *obd = request->rq_import->imp_obd; in ptl_send_rpc()
501 LASSERT(request->rq_type == PTL_RPC_MSG_REQUEST); in ptl_send_rpc()
502 LASSERT(request->rq_wait_ctx == 0); in ptl_send_rpc()
506 LASSERT(!request->rq_receiving_reply); in ptl_send_rpc()
507 LASSERT(!((lustre_msg_get_flags(request->rq_reqmsg) & MSG_REPLAY) && in ptl_send_rpc()
508 (request->rq_import->imp_state == LUSTRE_IMP_FULL))); in ptl_send_rpc()
514 spin_lock(&request->rq_lock); in ptl_send_rpc()
515 request->rq_err = 1; in ptl_send_rpc()
516 spin_unlock(&request->rq_lock); in ptl_send_rpc()
[all …]
Dclient.c530 struct ptlrpc_request *request; in ptlrpc_prep_req_from_pool() local
547 request = list_entry(pool->prp_req_list.next, struct ptlrpc_request, in ptlrpc_prep_req_from_pool()
549 list_del_init(&request->rq_list); in ptlrpc_prep_req_from_pool()
552 LASSERT(request->rq_reqbuf); in ptlrpc_prep_req_from_pool()
553 LASSERT(request->rq_pool); in ptlrpc_prep_req_from_pool()
555 reqbuf = request->rq_reqbuf; in ptlrpc_prep_req_from_pool()
556 memset(request, 0, sizeof(*request)); in ptlrpc_prep_req_from_pool()
557 request->rq_reqbuf = reqbuf; in ptlrpc_prep_req_from_pool()
558 request->rq_reqbuf_len = pool->prp_rq_size; in ptlrpc_prep_req_from_pool()
559 request->rq_pool = pool; in ptlrpc_prep_req_from_pool()
[all …]
Dimport.c108 struct ptlrpc_request *request,
638 struct ptlrpc_request *request; in ptlrpc_connect_import() local
697 request = ptlrpc_request_alloc(imp, &RQF_MDS_CONNECT); in ptlrpc_connect_import()
698 if (request == NULL) { in ptlrpc_connect_import()
703 rc = ptlrpc_request_bufs_pack(request, LUSTRE_OBD_VERSION, in ptlrpc_connect_import()
706 ptlrpc_request_free(request); in ptlrpc_connect_import()
712 lustre_msg_set_service_time(request->rq_reqmsg, in ptlrpc_connect_import()
713 at_timeout2est(request->rq_timeout)); in ptlrpc_connect_import()
720 request->rq_timeout = INITIAL_CONNECT_TIMEOUT; in ptlrpc_connect_import()
721 lustre_msg_set_timeout(request->rq_reqmsg, request->rq_timeout); in ptlrpc_connect_import()
[all …]
Dservice.c1872 struct ptlrpc_request *request; in ptlrpc_server_handle_request() local
1879 request = ptlrpc_server_request_get(svcpt, false); in ptlrpc_server_handle_request()
1880 if (request == NULL) in ptlrpc_server_handle_request()
1889 if (request->rq_export && request->rq_ops) in ptlrpc_server_handle_request()
1893 ptlrpc_rqphase_move(request, RQ_PHASE_INTERPRET); in ptlrpc_server_handle_request()
1899 timediff = cfs_timeval_sub(&work_start, &request->rq_arrival_time, in ptlrpc_server_handle_request()
1912 rc = lu_context_init(&request->rq_session, LCT_SESSION | LCT_NOREF); in ptlrpc_server_handle_request()
1917 request->rq_session.lc_thread = thread; in ptlrpc_server_handle_request()
1918 request->rq_session.lc_cookie = 0x5; in ptlrpc_server_handle_request()
1919 lu_context_enter(&request->rq_session); in ptlrpc_server_handle_request()
[all …]
/linux-4.1.27/drivers/usb/musb/
Dmusb_gadget.c56 static inline void map_dma_buffer(struct musb_request *request, in map_dma_buffer() argument
62 request->map_state = UN_MAPPED; in map_dma_buffer()
73 musb_ep->packet_sz, request->request.buf, in map_dma_buffer()
74 request->request.length); in map_dma_buffer()
78 if (request->request.dma == DMA_ADDR_INVALID) { in map_dma_buffer()
84 request->request.buf, in map_dma_buffer()
85 request->request.length, in map_dma_buffer()
86 request->tx in map_dma_buffer()
93 request->request.dma = dma_addr; in map_dma_buffer()
94 request->map_state = MUSB_MAPPED; in map_dma_buffer()
[all …]
Dmusb_gadget_ep0.c261 struct musb_request *request; in service_zero_data_request() local
304 request = next_request(musb_ep); in service_zero_data_request()
305 if (!musb_ep->busy && request) { in service_zero_data_request()
307 musb_ep_restart(musb, request); in service_zero_data_request()
488 struct musb_request *request; in ep0_rxstate() local
492 request = next_ep0_request(musb); in ep0_rxstate()
493 req = &request->request; in ep0_rxstate()
546 struct usb_request *request; in ep0_txstate() local
557 request = &req->request; in ep0_txstate()
560 fifo_src = (u8 *) request->buf + request->actual; in ep0_txstate()
[all …]
/linux-4.1.27/drivers/gpu/drm/
Ddrm_ioc32.c441 struct drm_buf_info __user *request; in compat_drm_infobufs() local
458 nbytes = sizeof(*request) + count * sizeof(struct drm_buf_desc); in compat_drm_infobufs()
459 request = compat_alloc_user_space(nbytes); in compat_drm_infobufs()
460 if (!access_ok(VERIFY_WRITE, request, nbytes)) in compat_drm_infobufs()
462 list = (struct drm_buf_desc *) (request + 1); in compat_drm_infobufs()
464 if (__put_user(count, &request->count) in compat_drm_infobufs()
465 || __put_user(list, &request->list)) in compat_drm_infobufs()
468 err = drm_ioctl(file, DRM_IOCTL_INFO_BUFS, (unsigned long)request); in compat_drm_infobufs()
472 if (__get_user(actual, &request->count)) in compat_drm_infobufs()
505 struct drm_buf_map __user *request; in compat_drm_mapbufs() local
[all …]
Ddrm_agpsupport.c199 int drm_agp_alloc(struct drm_device *dev, struct drm_agp_buffer *request) in drm_agp_alloc() argument
211 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_agp_alloc()
212 type = (u32) request->type; in drm_agp_alloc()
224 request->handle = entry->handle; in drm_agp_alloc()
225 request->physical = memory->physical; in drm_agp_alloc()
235 struct drm_agp_buffer *request = data; in drm_agp_alloc_ioctl() local
237 return drm_agp_alloc(dev, request); in drm_agp_alloc_ioctl()
273 int drm_agp_unbind(struct drm_device *dev, struct drm_agp_binding *request) in drm_agp_unbind() argument
280 if (!(entry = drm_agp_lookup_entry(dev, request->handle))) in drm_agp_unbind()
295 struct drm_agp_binding *request = data; in drm_agp_unbind_ioctl() local
[all …]
Ddrm_bufs.c515 struct drm_map *request = data; in drm_legacy_rmmap_ioctl() local
523 r_list->user_token == (unsigned long)request->handle && in drm_legacy_rmmap_ioctl()
598 struct drm_buf_desc *request) in drm_legacy_addbufs_agp() argument
619 count = request->count; in drm_legacy_addbufs_agp()
620 order = order_base_2(request->size); in drm_legacy_addbufs_agp()
623 alignment = (request->flags & _DRM_PAGE_ALIGN) in drm_legacy_addbufs_agp()
629 agp_offset = dev->agp->base + request->agp_start; in drm_legacy_addbufs_agp()
750 request->count = entry->buf_count; in drm_legacy_addbufs_agp()
751 request->size = size; in drm_legacy_addbufs_agp()
762 struct drm_buf_desc *request) in drm_legacy_addbufs_pci() argument
[all …]
Ddrm_context.c160 struct drm_ctx_priv_map *request = data; in drm_legacy_getsareactx() local
166 map = idr_find(&dev->ctx_idr, request->ctx_id); in drm_legacy_getsareactx()
172 request->handle = NULL; in drm_legacy_getsareactx()
175 request->handle = in drm_legacy_getsareactx()
183 if (request->handle == NULL) in drm_legacy_getsareactx()
204 struct drm_ctx_priv_map *request = data; in drm_legacy_setsareactx() local
211 && r_list->user_token == (unsigned long) request->handle) in drm_legacy_setsareactx()
223 if (IS_ERR(idr_replace(&dev->ctx_idr, map, request->ctx_id))) in drm_legacy_setsareactx()
Ddrm_scatter.c85 struct drm_scatter_gather *request = data; in drm_legacy_sg_alloc() local
104 pages = (request->size + PAGE_SIZE - 1) / PAGE_SIZE; in drm_legacy_sg_alloc()
105 DRM_DEBUG("size=%ld pages=%ld\n", request->size, pages); in drm_legacy_sg_alloc()
147 request->handle = entry->handle; in drm_legacy_sg_alloc()
201 struct drm_scatter_gather *request = data; in drm_legacy_sg_free() local
213 if (!entry || entry->handle != request->handle) in drm_legacy_sg_free()
Ddrm_irq.c1478 e->event.user_data = vblwait->request.signal; in drm_queue_vblank_event()
1505 if ((vblwait->request.type & _DRM_VBLANK_NEXTONMISS) && in drm_queue_vblank_event()
1506 (seq - vblwait->request.sequence) <= (1 << 23)) { in drm_queue_vblank_event()
1507 vblwait->request.sequence = seq + 1; in drm_queue_vblank_event()
1508 vblwait->reply.sequence = vblwait->request.sequence; in drm_queue_vblank_event()
1512 vblwait->request.sequence, seq, pipe); in drm_queue_vblank_event()
1515 vblwait->request.sequence); in drm_queue_vblank_event()
1517 e->event.sequence = vblwait->request.sequence; in drm_queue_vblank_event()
1518 if ((seq - vblwait->request.sequence) <= (1 << 23)) { in drm_queue_vblank_event()
1525 vblwait->reply.sequence = vblwait->request.sequence; in drm_queue_vblank_event()
[all …]
/linux-4.1.27/block/
Dblk.h24 struct request *flush_rq;
58 void init_request_from_bio(struct request *req, struct bio *bio);
59 void blk_rq_bio_prep(struct request_queue *q, struct request *rq,
61 int blk_rq_append_bio(struct request_queue *q, struct request *rq,
65 void blk_dequeue_request(struct request *rq);
67 bool __blk_end_bidi_request(struct request *rq, int error,
72 void blk_add_timer(struct request *req);
73 void blk_delete_timer(struct request *);
76 bool bio_attempt_front_merge(struct request_queue *q, struct request *req,
78 bool bio_attempt_back_merge(struct request_queue *q, struct request *req,
[all …]
Dnoop-iosched.c15 static void noop_merged_requests(struct request_queue *q, struct request *rq, in noop_merged_requests()
16 struct request *next) in noop_merged_requests()
26 struct request *rq; in noop_dispatch()
27 rq = list_entry(nd->queue.next, struct request, queuelist); in noop_dispatch()
35 static void noop_add_request(struct request_queue *q, struct request *rq) in noop_add_request()
42 static struct request *
43 noop_former_request(struct request_queue *q, struct request *rq) in noop_former_request()
49 return list_entry(rq->queuelist.prev, struct request, queuelist); in noop_former_request()
52 static struct request *
53 noop_latter_request(struct request_queue *q, struct request *rq) in noop_latter_request()
[all …]
Delevator.c56 static int elv_iosched_allow_merge(struct request *rq, struct bio *bio) in elv_iosched_allow_merge()
70 bool elv_rq_merge_ok(struct request *rq, struct bio *bio) in elv_rq_merge_ok()
245 static inline void __elv_rqhash_del(struct request *rq) in __elv_rqhash_del()
251 static void elv_rqhash_del(struct request_queue *q, struct request *rq) in elv_rqhash_del()
257 static void elv_rqhash_add(struct request_queue *q, struct request *rq) in elv_rqhash_add()
266 static void elv_rqhash_reposition(struct request_queue *q, struct request *rq) in elv_rqhash_reposition()
272 static struct request *elv_rqhash_find(struct request_queue *q, sector_t offset) in elv_rqhash_find()
276 struct request *rq; in elv_rqhash_find()
297 void elv_rb_add(struct rb_root *root, struct request *rq) in elv_rb_add()
301 struct request *__rq; in elv_rb_add()
[all …]
Ddeadline-iosched.c40 struct request *next_rq[2];
54 static void deadline_move_request(struct deadline_data *, struct request *);
57 deadline_rb_root(struct deadline_data *dd, struct request *rq) in deadline_rb_root()
65 static inline struct request *
66 deadline_latter_request(struct request *rq) in deadline_latter_request()
77 deadline_add_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_add_rq_rb()
85 deadline_del_rq_rb(struct deadline_data *dd, struct request *rq) in deadline_del_rq_rb()
99 deadline_add_request(struct request_queue *q, struct request *rq) in deadline_add_request()
116 static void deadline_remove_request(struct request_queue *q, struct request *rq) in deadline_remove_request()
125 deadline_merge(struct request_queue *q, struct request **req, struct bio *bio) in deadline_merge()
[all …]
Dblk-merge.c89 void blk_recalc_rq_segments(struct request *rq) in blk_recalc_rq_segments()
247 int blk_rq_map_sg(struct request_queue *q, struct request *rq, in blk_rq_map_sg()
287 struct request *req, in ll_new_hw_segment()
312 int ll_back_merge_fn(struct request_queue *q, struct request *req, in ll_back_merge_fn()
330 int ll_front_merge_fn(struct request_queue *q, struct request *req, in ll_front_merge_fn()
352 static bool req_no_special_merge(struct request *req) in req_no_special_merge()
359 static int req_gap_to_prev(struct request *req, struct request *next) in req_gap_to_prev()
367 static int ll_merge_requests_fn(struct request_queue *q, struct request *req, in ll_merge_requests_fn()
368 struct request *next) in ll_merge_requests_fn()
421 void blk_rq_set_mixed_merge(struct request *rq) in blk_rq_set_mixed_merge()
[all …]
Dblk-flush.c97 static unsigned int blk_flush_policy(unsigned int fflags, struct request *rq) in blk_flush_policy()
113 static unsigned int blk_flush_cur_seq(struct request *rq) in blk_flush_cur_seq()
118 static void blk_flush_restore_request(struct request *rq) in blk_flush_restore_request()
132 static bool blk_flush_queue_rq(struct request *rq, bool add_front) in blk_flush_queue_rq()
165 static bool blk_flush_complete_seq(struct request *rq, in blk_flush_complete_seq()
219 static void flush_end_io(struct request *flush_rq, int error) in flush_end_io()
224 struct request *rq, *n; in flush_end_io()
287 struct request *first_rq = in blk_kick_flush()
288 list_first_entry(pending, struct request, flush.list); in blk_kick_flush()
289 struct request *flush_rq = fq->flush_rq; in blk_kick_flush()
[all …]
Dblk-core.c97 void blk_rq_init(struct request_queue *q, struct request *rq) in blk_rq_init()
117 static void req_bio_endio(struct request *rq, struct bio *bio, in req_bio_endio()
135 void blk_dump_rq_flags(struct request *rq, char *msg) in blk_dump_rq_flags()
799 static inline void blk_free_request(struct request_list *rl, struct request *rq) in blk_free_request()
974 static struct request *__get_request(struct request_list *rl, int rw_flags, in __get_request()
978 struct request *rq; in __get_request()
1151 static struct request *get_request(struct request_queue *q, int rw_flags, in get_request()
1157 struct request *rq; in get_request()
1192 static struct request *blk_old_get_request(struct request_queue *q, int rw, in blk_old_get_request()
1195 struct request *rq; in blk_old_get_request()
[all …]
Dblk-softirq.c31 struct request *rq; in blk_done_softirq()
33 rq = list_entry(local_list.next, struct request, ipi_list); in blk_done_softirq()
42 struct request *rq = data; in trigger_softirq()
59 static int raise_blk_irq(int cpu, struct request *rq) in raise_blk_irq()
75 static int raise_blk_irq(int cpu, struct request *rq) in raise_blk_irq()
105 void __blk_complete_request(struct request *req) in __blk_complete_request()
166 void blk_complete_request(struct request *req) in blk_complete_request()
Dbsg-lib.c55 struct request *req = job->req; in bsg_job_done()
56 struct request *rsp = req->next_rq; in bsg_job_done()
82 static void bsg_softirq_done(struct request *rq) in bsg_softirq_done()
90 static int bsg_map_buffer(struct bsg_buffer *buf, struct request *req) in bsg_map_buffer()
110 static int bsg_create_job(struct device *dev, struct request *req) in bsg_create_job()
112 struct request *rsp = req->next_rq; in bsg_create_job()
127 job->request = req->cmd; in bsg_create_job()
166 struct request *req; in bsg_request_fn()
Dblk-mq.c183 struct request *rq, unsigned int rw_flags) in blk_mq_rq_ctx_init()
230 static struct request *
233 struct request *rq; in __blk_mq_alloc_request()
253 struct request *blk_mq_alloc_request(struct request_queue *q, int rw, gfp_t gfp, in blk_mq_alloc_request()
258 struct request *rq; in blk_mq_alloc_request()
293 struct blk_mq_ctx *ctx, struct request *rq) in __blk_mq_free_request()
307 void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *hctx, struct request *rq) in blk_mq_free_hctx_request()
317 void blk_mq_free_request(struct request *rq) in blk_mq_free_request()
327 inline void __blk_mq_end_request(struct request *rq, int error) in __blk_mq_end_request()
341 void blk_mq_end_request(struct request *rq, int error) in blk_mq_end_request()
[all …]
Dblk-timeout.c78 void blk_delete_timer(struct request *req) in blk_delete_timer()
83 static void blk_rq_timed_out(struct request *req) in blk_rq_timed_out()
113 static void blk_rq_check_expired(struct request *rq, unsigned long *next_timeout, in blk_rq_check_expired()
134 struct request *rq, *tmp; in blk_rq_timed_out_timer()
157 void blk_abort_request(struct request *req) in blk_abort_request()
188 void blk_add_timer(struct request *req) in blk_add_timer()
Dblk-mq.h28 void __blk_mq_complete_request(struct request *rq);
32 void blk_mq_clone_flush_request(struct request *flush_rq,
33 struct request *orig_rq);
63 extern void blk_mq_rq_timed_out(struct request *req, bool reserved);
Dblk-tag.c23 struct request *blk_queue_find_tag(struct request_queue *q, int tag) in blk_queue_find_tag()
91 struct request **tag_index; in init_tag_map()
101 tag_index = kzalloc(depth * sizeof(struct request *), GFP_ATOMIC); in init_tag_map()
206 struct request **tag_index; in blk_queue_resize_tags()
241 memcpy(bqt->tag_index, tag_index, max_depth * sizeof(struct request *)); in blk_queue_resize_tags()
265 void blk_queue_end_tag(struct request_queue *q, struct request *rq) in blk_queue_end_tag()
313 int blk_queue_start_tag(struct request_queue *q, struct request *rq) in blk_queue_start_tag()
Dblk-exec.c23 static void blk_end_sync_rq(struct request *rq, int error) in blk_end_sync_rq()
52 struct request *rq, int at_head, in blk_execute_rq_nowait()
110 struct request *rq, int at_head) in blk_execute_rq()
Dblk-map.c12 int blk_rq_append_bio(struct request_queue *q, struct request *rq, in blk_rq_append_bio()
63 int blk_rq_map_user_iov(struct request_queue *q, struct request *rq, in blk_rq_map_user_iov()
121 int blk_rq_map_user(struct request_queue *q, struct request *rq, in blk_rq_map_user()
181 int blk_rq_map_kern(struct request_queue *q, struct request *rq, void *kbuf, in blk_rq_map_kern()
/linux-4.1.27/drivers/gpu/drm/i915/
Di915_ioc32.c116 drm_i915_irq_emit_t __user *request; in compat_i915_irq_emit() local
121 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_irq_emit()
122 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_irq_emit()
124 &request->irq_seq)) in compat_i915_irq_emit()
128 (unsigned long)request); in compat_i915_irq_emit()
139 drm_i915_getparam_t __user *request; in compat_i915_getparam() local
144 request = compat_alloc_user_space(sizeof(*request)); in compat_i915_getparam()
145 if (!access_ok(VERIFY_WRITE, request, sizeof(*request)) in compat_i915_getparam()
146 || __put_user(req32.param, &request->param) in compat_i915_getparam()
148 &request->value)) in compat_i915_getparam()
[all …]
Dintel_lrc.c520 struct drm_i915_gem_request *request) in execlists_context_queue() argument
530 if (!request) { in execlists_context_queue()
535 request = kzalloc(sizeof(*request), GFP_KERNEL); in execlists_context_queue()
536 if (request == NULL) in execlists_context_queue()
538 request->ring = ring; in execlists_context_queue()
539 request->ctx = to; in execlists_context_queue()
540 kref_init(&request->ref); in execlists_context_queue()
541 request->uniq = dev_priv->request_uniq++; in execlists_context_queue()
542 i915_gem_context_reference(request->ctx); in execlists_context_queue()
544 i915_gem_request_reference(request); in execlists_context_queue()
[all …]
Di915_gem.c2333 struct drm_i915_gem_request *request; in __i915_add_request() local
2338 request = ring->outstanding_lazy_request; in __i915_add_request()
2339 if (WARN_ON(request == NULL)) in __i915_add_request()
2343 ringbuf = request->ctx->engine[ring->id].ringbuf; in __i915_add_request()
2356 ret = logical_ring_flush_all_caches(ringbuf, request->ctx); in __i915_add_request()
2370 request->postfix = intel_ring_get_tail(ringbuf); in __i915_add_request()
2373 ret = ring->emit_request(ringbuf, request); in __i915_add_request()
2381 request->tail = intel_ring_get_tail(ringbuf); in __i915_add_request()
2384 request->head = request_start; in __i915_add_request()
2392 request->batch_obj = obj; in __i915_add_request()
[all …]
/linux-4.1.27/drivers/net/hyperv/
Drndis_filter.c80 struct rndis_request *request; in get_rndis_request() local
85 request = kzalloc(sizeof(struct rndis_request), GFP_KERNEL); in get_rndis_request()
86 if (!request) in get_rndis_request()
89 init_completion(&request->wait_event); in get_rndis_request()
91 rndis_msg = &request->request_msg; in get_rndis_request()
95 request->pkt.q_idx = 0; in get_rndis_request()
107 list_add_tail(&request->list_ent, &dev->req_list); in get_rndis_request()
110 return request; in get_rndis_request()
247 struct rndis_request *request) in rndis_set_link_state() argument
252 query_complete = &request->response_msg.msg.query_complete; in rndis_set_link_state()
[all …]
/linux-4.1.27/sound/pci/mixart/
Dmixart_hwdep.c145 struct mixart_msg request; in mixart_enum_connectors() local
162 request.message_id = MSG_SYSTEM_ENUM_PLAY_CONNECTOR; in mixart_enum_connectors()
163 request.uid = (struct mixart_uid){0,0}; /* board num = 0 */ in mixart_enum_connectors()
164 request.data = NULL; in mixart_enum_connectors()
165 request.size = 0; in mixart_enum_connectors()
167 err = snd_mixart_send_msg(mgr, &request, sizeof(*connector), connector); in mixart_enum_connectors()
192 request.message_id = MSG_CONNECTOR_GET_AUDIO_INFO; in mixart_enum_connectors()
193 request.uid = connector->uid[k]; in mixart_enum_connectors()
194 request.data = audio_info_req; in mixart_enum_connectors()
195 request.size = sizeof(*audio_info_req); in mixart_enum_connectors()
[all …]
Dmixart.c77 struct mixart_msg request; in mixart_set_pipe_state() local
99 request.message_id = MSG_SYSTEM_WAIT_SYNCHRO_CMD; in mixart_set_pipe_state()
100 request.uid = (struct mixart_uid){0,0}; in mixart_set_pipe_state()
101 request.data = &system_msg_uid; in mixart_set_pipe_state()
102 request.size = sizeof(system_msg_uid); in mixart_set_pipe_state()
104 err = snd_mixart_send_msg_wait_notif(mgr, &request, system_msg_uid); in mixart_set_pipe_state()
118 request.message_id = MSG_STREAM_START_STREAM_GRP_PACKET; in mixart_set_pipe_state()
120 request.message_id = MSG_STREAM_STOP_STREAM_GRP_PACKET; in mixart_set_pipe_state()
122 request.uid = pipe->group_uid; /*(struct mixart_uid){0,0};*/ in mixart_set_pipe_state()
123 request.data = &group_state; in mixart_set_pipe_state()
[all …]
Dmixart_mixer.c306 struct mixart_msg request; in mixart_update_analog_audio_level() local
324 if(is_capture) request.uid = chip->uid_in_analog_physio; in mixart_update_analog_audio_level()
325 else request.uid = chip->uid_out_analog_physio; in mixart_update_analog_audio_level()
326 request.message_id = MSG_PHYSICALIO_SET_LEVEL; in mixart_update_analog_audio_level()
327 request.data = &io_level; in mixart_update_analog_audio_level()
328 request.size = sizeof(io_level); in mixart_update_analog_audio_level()
330 err = snd_mixart_send_msg(chip->mgr, &request, sizeof(resp), &resp); in mixart_update_analog_audio_level()
727 struct mixart_msg request; in mixart_update_playback_stream_level() local
760 request.message_id = MSG_STREAM_SET_OUT_STREAM_LEVEL; in mixart_update_playback_stream_level()
761 request.uid = (struct mixart_uid){0,0}; in mixart_update_playback_stream_level()
[all …]
/linux-4.1.27/drivers/staging/rtl8712/
Dusb_ops.c39 u8 request; in usb_read8() local
47 request = 0x05; in usb_read8()
52 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read8()
59 u8 request; in usb_read16() local
67 request = 0x05; in usb_read16()
72 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read16()
79 u8 request; in usb_read32() local
87 request = 0x05; in usb_read32()
92 r8712_usbctrl_vendorreq(pintfpriv, request, wvalue, index, &data, len, in usb_read32()
99 u8 request; in usb_write8() local
[all …]
/linux-4.1.27/security/apparmor/
Dfile.c70 if (sa->aad->fs.request & AA_AUDIT_FILE_MASK) { in file_audit_cb()
72 audit_file_mask(ab, sa->aad->fs.request); in file_audit_cb()
78 if (sa->aad->fs.request & AA_AUDIT_FILE_MASK) { in file_audit_cb()
107 gfp_t gfp, int op, u32 request, const char *name, in aa_audit_file() argument
116 aad.fs.request = request; in aa_audit_file()
130 sa.aad->fs.request &= mask; in aa_audit_file()
132 if (likely(!sa.aad->fs.request)) in aa_audit_file()
137 sa.aad->fs.request = sa.aad->fs.request & ~perms->allow; in aa_audit_file()
139 if (sa.aad->fs.request & perms->kill) in aa_audit_file()
143 if ((sa.aad->fs.request & perms->quiet) && in aa_audit_file()
[all …]
Ddomain.c97 const char *name, u32 request, in change_profile_perms() argument
113 if (COMBINED_PERM_MASK(perms) & request) in change_profile_perms()
756 u32 request; in aa_change_profile() local
762 request = AA_MAY_ONEXEC; in aa_change_profile()
765 request = AA_MAY_CHANGE_PROFILE; in aa_change_profile()
806 perms = change_profile_perms(profile, ns, hname, request, in aa_change_profile()
808 if (!(perms.allow & request)) { in aa_change_profile()
846 error = aa_audit_file(profile, &perms, GFP_KERNEL, op, request, in aa_change_profile()
/linux-4.1.27/arch/um/drivers/
Dmconsole_user.c66 if (!strncmp(req->request.data, cmd->command, in mconsole_parse()
84 req->len = recvfrom(fd, &req->request, sizeof(req->request), 0, in mconsole_get_request()
91 if (req->request.magic != MCONSOLE_MAGIC) { in mconsole_get_request()
93 len = MIN(sizeof(req->request.data) - 1, in mconsole_get_request()
94 strlen((char *) &req->request)); in mconsole_get_request()
95 memmove(req->request.data, &req->request, len); in mconsole_get_request()
96 req->request.data[len] = '\0'; in mconsole_get_request()
98 req->request.magic = MCONSOLE_MAGIC; in mconsole_get_request()
99 req->request.version = 0; in mconsole_get_request()
100 req->request.len = len; in mconsole_get_request()
[all …]
Dmconsole_kern.c67 req->request.cmd->handler(&req->request); in mc_work_proc()
90 new->request = req; in mconsole_interrupt()
91 new->request.regs = get_irq_regs()->regs; in mconsole_interrupt()
115 char *ptr = req->request.data; in mconsole_log()
119 len = req->len - (ptr - req->request.data); in mconsole_log()
131 char *ptr = req->request.data; in mconsole_proc()
473 char *ptr = req->request.data, *name, *error_string = ""; in mconsole_config()
499 char *ptr = req->request.data, *err_msg = ""; in mconsole_remove()
625 char *ptr = req->request.data; in mconsole_sysrq()
661 char *ptr = req->request.data; in mconsole_stack()
/linux-4.1.27/Documentation/block/
Drequest.txt2 struct request documentation
9 2.0 Struct request members classification
11 2.1 struct request members explanation
17 Short explanation of request members
45 request type, etc.
57 unsigned long nr_sectors DBI Total number of sectors in request
62 segments in a request
65 segments in a request
68 of request
80 on request completion
[all …]
Dbiodoc.txt59 2.3 Changes in the request structure
63 3.2.1 Traversing segments and completion units in a request
117 a per-queue level (e.g maximum request size, maximum number of segments in
134 Sets two variables that limit the size of the request.
136 - The request queue's max_sectors, which is a soft size in
140 - The request queue's max_hw_sectors, which is a hard limit
141 and reflects the maximum size request a driver can handle
148 Maximum physical segments you can handle in a request. 128
152 Maximum dma segments the hardware can handle in a request. 128
176 setting the queue bounce limit for the request queue for the device
[all …]
Dcfq-iosched.txt7 CFQ maintains the per process queue for the processes which request I/O
17 This specifies how long CFQ should idle for next request on certain cfq queues
52 backward distance of request is just 1/back_seek_penalty from a "front"
53 request, then the seeking cost of two requests is considered equivalent.
55 So scheduler will not bias toward one or the other request (otherwise scheduler
56 will bias toward front request). Default value of back_seek_penalty is 2.
101 time for each process to issue I/O request before the cfq queue is switched.
117 This parameter is used to limit the dispatching of asynchronous request to
118 device request queue in queue's slice time. The maximum number of request that
136 This specifies the number of request dispatched to the device queue. In a
[all …]
Dqueue-sysfs.txt62 for a filesystem request. Must be smaller than or equal to the maximum
92 To avoid priority inversion through request starvation, a request
93 queue maintains a separate request pool per each cgroup when
95 per-block-cgroup request pool. IOW, if there are N block cgroups,
96 each request queue may have up to N request pools, each independently
119 If this option is '1', the block layer will migrate request completions to the
120 cpu "group" that originally submitted the request. For some workloads this
Ddeadline-iosched.txt21 service time for a request. As we focus mainly on read latencies, this is
22 tunable. When a read request first enters the io scheduler, it is assigned
41 This parameter tunes the balance between per-request latency and aggregate
61 Sometimes it happens that a request enters the io scheduler that is contiguous
62 with a request that is already on the queue. Either it fits in the back of that
63 request, or it fits at the front. That is called either a back merge candidate
/linux-4.1.27/drivers/firewire/
Dcore-transaction.c633 struct fw_request *request; in free_response_callback() local
635 request = container_of(packet, struct fw_request, response); in free_response_callback()
636 kfree(request); in free_response_callback()
751 struct fw_request *request; in allocate_request() local
784 request = kmalloc(sizeof(*request) + length, GFP_ATOMIC); in allocate_request()
785 if (request == NULL) in allocate_request()
788 request->response.speed = p->speed; in allocate_request()
789 request->response.timestamp = in allocate_request()
791 request->response.generation = p->generation; in allocate_request()
792 request->response.ack = 0; in allocate_request()
[all …]
Dcore-cdev.c125 struct fw_request *request; member
189 struct fw_cdev_event_request request; member
589 struct fw_cdev_send_request *request, in init_request() argument
595 if (request->tcode != TCODE_STREAM_DATA && in init_request()
596 (request->length > 4096 || request->length > 512 << speed)) in init_request()
599 if (request->tcode == TCODE_WRITE_QUADLET_REQUEST && in init_request()
600 request->length < 4) in init_request()
603 e = kmalloc(sizeof(*e) + request->length, GFP_KERNEL); in init_request()
608 e->response.length = request->length; in init_request()
609 e->response.closure = request->closure; in init_request()
[all …]
Dsbp2.c296 } request; member
324 } request; member
418 static void sbp2_status_write(struct fw_card *card, struct fw_request *request, in sbp2_status_write() argument
430 fw_send_response(card, request, RCODE_TYPE_ERROR); in sbp2_status_write()
443 fw_send_response(card, request, RCODE_COMPLETE); in sbp2_status_write()
466 fw_send_response(card, request, RCODE_COMPLETE); in sbp2_status_write()
583 orb->request.response.high = 0; in sbp2_send_management_orb()
584 orb->request.response.low = cpu_to_be32(orb->response_bus); in sbp2_send_management_orb()
586 orb->request.misc = cpu_to_be32( in sbp2_send_management_orb()
590 orb->request.length = cpu_to_be32( in sbp2_send_management_orb()
[all …]
/linux-4.1.27/drivers/s390/cio/
Dchsc_sch.c60 struct chsc_request *request = private->request; in chsc_subchannel_irq() local
68 if (!request) { in chsc_subchannel_irq()
73 private->request = NULL; in chsc_subchannel_irq()
74 memcpy(&request->irb, irb, sizeof(*irb)); in chsc_subchannel_irq()
76 complete(&request->completion); in chsc_subchannel_irq()
114 if (private->request) { in chsc_subchannel_remove()
115 complete(&private->request->completion); in chsc_subchannel_remove()
243 struct chsc_request *request) in chsc_async() argument
255 if (private->request) { in chsc_async()
273 private->request = request; in chsc_async()
[all …]
Dchsc.c70 struct chsc_header request; member
104 ssd_area->request.length = 0x0010; in chsc_get_ssd_info()
105 ssd_area->request.code = 0x0004; in chsc_get_ssd_info()
159 ssqd->request.length = 0x0010; in chsc_ssqd()
160 ssqd->request.code = 0x0024; in chsc_ssqd()
185 scssc->request.length = 0x0fe0; in chsc_sadc()
186 scssc->request.code = 0x0021; in chsc_sadc()
331 struct chsc_header request; member
548 sei->request.length = 0x0010; in chsc_process_event_information()
549 sei->request.code = 0x000e; in chsc_process_event_information()
[all …]
Dchsc.h68 struct chsc_header request; member
82 struct chsc_header request; member
102 struct chsc_header request; member
161 struct chsc_header request; member
197 struct chsc_header request; member
/linux-4.1.27/Documentation/mmc/
Dmmc-async-req.txt7 preparations for the next request are done in parallel with the current
10 time between when an MMC request ends and another MMC request begins.
14 MMC request.
22 a request and how fast the memory is. The faster the MMC/SD is the
23 more significant the prepare request time becomes. Roughly the expected
38 It starts a new MMC command request for a host. The function isn't
39 truly non-blocking. If there is an ongoing async request it waits
40 for completion of that request and starts the new one and returns. It
41 doesn't wait for the new request to complete. If there is no ongoing
42 request it starts the new request and returns immediately.
[all …]
/linux-4.1.27/kernel/
Dptrace.c296 static int ptrace_attach(struct task_struct *task, long request, in ptrace_attach() argument
300 bool seize = (request == PTRACE_SEIZE); in ptrace_attach()
702 #define is_singlestep(request) ((request) == PTRACE_SINGLESTEP) argument
704 #define is_singlestep(request) 0 argument
708 #define is_singleblock(request) ((request) == PTRACE_SINGLEBLOCK) argument
710 #define is_singleblock(request) 0 argument
714 #define is_sysemu_singlestep(request) ((request) == PTRACE_SYSEMU_SINGLESTEP) argument
716 #define is_sysemu_singlestep(request) 0 argument
719 static int ptrace_resume(struct task_struct *child, long request, in ptrace_resume() argument
727 if (request == PTRACE_SYSCALL) in ptrace_resume()
[all …]
/linux-4.1.27/drivers/staging/lustre/lustre/llite/
Dsymlink.c46 struct ptlrpc_request **request, char **symname) in ll_readlink_internal() argument
54 *request = NULL; in ll_readlink_internal()
75 rc = md_getattr(sbi->ll_md_exp, op_data, request); in ll_readlink_internal()
83 body = req_capsule_server_get(&(*request)->rq_pill, &RMF_MDT_BODY); in ll_readlink_internal()
99 *symname = req_capsule_server_get(&(*request)->rq_pill, &RMF_MDT_MD); in ll_readlink_internal()
124 struct ptlrpc_request *request = NULL; in ll_follow_link() local
138 rc = ll_readlink_internal(inode, &request, &symname); in ll_follow_link()
142 ptlrpc_req_finished(request); in ll_follow_link()
143 request = NULL; in ll_follow_link()
151 return request; in ll_follow_link()
Dnamei.c408 static int ll_lookup_it_finish(struct ptlrpc_request *request, in ll_lookup_it_finish() argument
421 rc = ll_prep_inode(&inode, request, (*de)->d_sb, it); in ll_lookup_it_finish()
678 struct ptlrpc_request *request = NULL; in ll_create_node() local
685 request = it->d.lustre.it_data; in ll_create_node()
687 rc = ll_prep_inode(&inode, request, dir->i_sb, it); in ll_create_node()
702 ptlrpc_req_finished(request); in ll_create_node()
742 static void ll_update_times(struct ptlrpc_request *request, in ll_update_times() argument
745 struct mdt_body *body = req_capsule_server_get(&request->rq_pill, in ll_update_times()
764 struct ptlrpc_request *request = NULL; in ll_new_node() local
786 cfs_curproc_cap_pack(), rdev, &request); in ll_new_node()
[all …]
Ddir.c148 struct ptlrpc_request *request; in ll_dir_filler() local
184 rc = md_readpage(exp, op_data, page_pool, &request); in ll_dir_filler()
190 body = req_capsule_server_get(&request->rq_pill, &RMF_MDT_BODY); in ll_dir_filler()
197 nrdpgs = (request->rq_bulk->bd_nob_transferred+PAGE_CACHE_SIZE-1) in ll_dir_filler()
202 ptlrpc_req_finished(request); in ll_dir_filler()
365 struct ptlrpc_request *request; in ll_get_dir_page() local
378 request = (struct ptlrpc_request *)it.d.lustre.it_data; in ll_get_dir_page()
379 if (request) in ll_get_dir_page()
380 ptlrpc_req_finished(request); in ll_get_dir_page()
661 struct ptlrpc_request *request = NULL; in ll_dir_setdirstripe() local
[all …]
Dxattr.c479 struct ptlrpc_request *request = NULL; in ll_getxattr() local
496 &lmmsize, &request); in ll_getxattr()
536 if (request) in ll_getxattr()
537 ptlrpc_req_finished(request); in ll_getxattr()
551 struct ptlrpc_request *request = NULL; in ll_listxattr() local
592 rc2 = ll_dir_getstripe(inode, &lmm, &lmmsize, &request); in ll_listxattr()
604 ptlrpc_req_finished(request); in ll_listxattr()
617 ptlrpc_req_finished(request); in ll_listxattr()
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_profile.c65 struct mthca_profile *request, in mthca_make_profile() argument
100 profile[MTHCA_RES_UARC].size = request->uarc_size; in mthca_make_profile()
102 profile[MTHCA_RES_QP].num = request->num_qp; in mthca_make_profile()
103 profile[MTHCA_RES_SRQ].num = request->num_srq; in mthca_make_profile()
104 profile[MTHCA_RES_EQP].num = request->num_qp; in mthca_make_profile()
105 profile[MTHCA_RES_RDB].num = request->num_qp * request->rdb_per_qp; in mthca_make_profile()
106 profile[MTHCA_RES_CQ].num = request->num_cq; in mthca_make_profile()
108 profile[MTHCA_RES_MCG].num = request->num_mcg; in mthca_make_profile()
109 profile[MTHCA_RES_MPT].num = request->num_mpt; in mthca_make_profile()
110 profile[MTHCA_RES_MTT].num = request->num_mtt; in mthca_make_profile()
[all …]
/linux-4.1.27/drivers/isdn/hardware/eicon/
Ddadapter.c112 if (d->request) { in diva_didd_add_descriptor()
113 MAdapter.request = d->request; in diva_didd_add_descriptor()
114 dprintf = (DIVA_DI_PRINTF)d->request; in diva_didd_add_descriptor()
116 DBG_TRC(("DIMAINT registered, dprintf=%08x", d->request)) in diva_didd_add_descriptor()
120 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_add_descriptor()
132 DBG_TRC(("Add adapter[%d], request=%08x", (i + 1), d->request)) in diva_didd_add_descriptor()
145 static int diva_didd_remove_descriptor(IDI_CALL request) { in diva_didd_remove_descriptor() argument
148 if (request == MAdapter.request) { in diva_didd_remove_descriptor()
152 MAdapter.request = (IDI_CALL)no_printf; in diva_didd_remove_descriptor()
156 if (HandleTable[i].request == request) { in diva_didd_remove_descriptor()
[all …]
Ddivasfunc.c72 d.request = Requests[card - 1]; in diva_xdi_didd_register_adapter()
84 DAdapter.request((ENTITY *)&req); in diva_xdi_didd_register_adapter()
106 DAdapter.request((ENTITY *)&req); in diva_xdi_didd_remove_adapter()
146 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in didd_callback()
174 DAdapter.request((ENTITY *)&req); in connect_didd()
182 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in connect_didd()
206 DAdapter.request((ENTITY *)&req); in disconnect_didd()
Didifunc.c70 if (card->d.request == d->request) { in find_card_in_list()
98 card->d.request((ENTITY *)&sync_req); in um_new_card()
168 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in didd_callback()
202 DAdapter.request((ENTITY *)&req); in connect_didd()
210 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in connect_didd()
237 DAdapter.request((ENTITY *)&req); in disconnect_didd()
Ddebug.c25 static void diva_free_dma_descriptor(IDI_CALL request, int nr);
26 static int diva_get_dma_descriptor(IDI_CALL request, dword *dma_magic);
42 IDI_CALL request; member
833 …(*(clients[id].request))((ENTITY *)(*(clients[id].pIdiLib->DivaSTraceGetHandle))(clients[id].pIdiL… in diva_set_driver_dbg_mask()
841 static int diva_get_idi_adapter_info(IDI_CALL request, dword *serial, dword *logical) { in diva_get_idi_adapter_info() argument
846 (*request)((ENTITY *)&sync_req); in diva_get_idi_adapter_info()
852 (*request)((ENTITY *)&sync_req); in diva_get_idi_adapter_info()
872 diva_get_idi_adapter_info(d->request, &serial, &logical); in diva_mnt_add_xdi_adapter()
891 if (clients[id].hDbg && (clients[id].request == d->request)) { in diva_mnt_add_xdi_adapter()
921 clients[id].request = d->request; in diva_mnt_add_xdi_adapter()
[all …]
Dmntfunc.c59 dprintf = (DIVA_DI_PRINTF) MAdapter.request; in didd_callback()
93 DAdapter.request((ENTITY *)&req); in connect_didd()
103 DAdapter.request((ENTITY *)&req); in connect_didd()
124 DAdapter.request((ENTITY *)&req); in disconnect_didd()
129 (IDI_CALL) MaintDescriptor.request; in disconnect_didd()
130 DAdapter.request((ENTITY *)&req); in disconnect_didd()
Dcapifunc.c376 if (adapter[i].request) in clean_adapter()
383 while ((max_adapter != 0) && !adapter[max_adapter - 1].request) in clean_adapter()
407 if (card->d.request == d->request) { in divacapi_remove_card()
462 d.request = card->d.request; in divacapi_remove_cards()
507 card->d.request((ENTITY *)&sync_req); in diva_add_card()
536 card->d.request((ENTITY *)&sync_req); in diva_add_card()
557 card->d.request((ENTITY *)&sync_req); in diva_add_card()
562 a->request = DIRequest; /* card->d.request; */ in diva_add_card()
620 if (adapter[i].request) in diva_add_card()
679 if (adapter[i].request) in diva_add_card()
[all …]
/linux-4.1.27/arch/mips/jz4740/
Dgpio.c138 int jz_gpio_bulk_request(const struct jz_gpio_bulk_request *request, size_t num) in jz_gpio_bulk_request() argument
143 for (i = 0; i < num; ++i, ++request) { in jz_gpio_bulk_request()
144 ret = gpio_request(request->gpio, request->name); in jz_gpio_bulk_request()
147 jz_gpio_set_function(request->gpio, request->function); in jz_gpio_bulk_request()
153 for (--request; i > 0; --i, --request) { in jz_gpio_bulk_request()
154 gpio_free(request->gpio); in jz_gpio_bulk_request()
155 jz_gpio_set_function(request->gpio, JZ_GPIO_FUNC_NONE); in jz_gpio_bulk_request()
162 void jz_gpio_bulk_free(const struct jz_gpio_bulk_request *request, size_t num) in jz_gpio_bulk_free() argument
166 for (i = 0; i < num; ++i, ++request) { in jz_gpio_bulk_free()
167 gpio_free(request->gpio); in jz_gpio_bulk_free()
[all …]
/linux-4.1.27/drivers/s390/block/
Dscm_blk_cluster.c60 static bool clusters_intersect(struct request *A, struct request *B) in clusters_intersect()
77 struct request *req = scmrq->request[scmrq->aob->request.msb_count]; in scm_reserve_cluster()
95 for (pos = 0; pos < iter->aob->request.msb_count; pos++) { in scm_reserve_cluster()
96 if (clusters_intersect(req, iter->request[pos]) && in scm_reserve_cluster()
98 rq_data_dir(iter->request[pos]) == WRITE)) { in scm_reserve_cluster()
134 struct request *req = scmrq->request[0]; in scm_prepare_cluster_request()
164 scmrq->aob->request.msb_count = 1; in scm_prepare_cluster_request()
199 int pos = scmrq->aob->request.msb_count; in scm_need_cluster_request()
201 if (rq_data_dir(scmrq->request[pos]) == READ) in scm_need_cluster_request()
204 return blk_rq_bytes(scmrq->request[pos]) < CLUSTER_SIZE; in scm_need_cluster_request()
Dscm_blk.c46 kfree(scmrq->request); in __scm_free_rq()
80 scmrq->request = kcalloc(nr_requests_per_io, sizeof(scmrq->request[0]), in __scm_alloc_rq()
82 if (!scmrq->request) in __scm_alloc_rq()
134 for (i = 0; i < nr_requests_per_io && scmrq->request[i]; i++) { in scm_request_done()
148 static bool scm_permit_request(struct scm_blk_dev *bdev, struct request *req) in scm_permit_request()
185 int pos = scmrq->aob->request.msb_count; in scm_request_prepare()
187 struct request *req = scmrq->request[pos]; in scm_request_prepare()
197 scmrq->aob->request.msb_count++; in scm_request_prepare()
215 struct request *req) in scm_request_set()
217 scmrq->request[scmrq->aob->request.msb_count] = req; in scm_request_set()
[all …]
/linux-4.1.27/arch/s390/pci/
Dpci_clp.c88 rrb->request.hdr.len = sizeof(rrb->request); in clp_query_pci_fngrp()
89 rrb->request.hdr.cmd = CLP_QUERY_PCI_FNGRP; in clp_query_pci_fngrp()
91 rrb->request.pfgid = pfgid; in clp_query_pci_fngrp()
141 rrb->request.hdr.len = sizeof(rrb->request); in clp_query_pci_fn()
142 rrb->request.hdr.cmd = CLP_QUERY_PCI_FN; in clp_query_pci_fn()
144 rrb->request.fh = fh; in clp_query_pci_fn()
210 rrb->request.hdr.len = sizeof(rrb->request); in clp_set_pci_fn()
211 rrb->request.hdr.cmd = CLP_SET_PCI_FN; in clp_set_pci_fn()
213 rrb->request.fh = *fh; in clp_set_pci_fn()
214 rrb->request.oc = command; in clp_set_pci_fn()
[all …]
/linux-4.1.27/Documentation/filesystems/nfs/
Didmapper.txt7 performing an upcall to userspace to request the information. There are two
8 ways NFS could obtain this information: placing a call to /sbin/request-key
11 NFS will attempt to call /sbin/request-key first. If this succeeds, the
12 result will be cached using the generic request-key cache. This call should
13 only fail if /etc/request-key.conf is not configured for the id_resolver key
14 type, see the "Configuring" section below if you wish to use the request-key
17 If the call to /sbin/request-key fails (if /etc/request-key.conf is not
26 The file /etc/request-key.conf will need to be modified so /sbin/request-key can
46 would edit your request-key.conf so it look similar to this:
54 request-key will find the first matching line and corresponding program. In
[all …]
/linux-4.1.27/drivers/scsi/
Dstorvsc_drv.c821 struct storvsc_cmd_request *request; in handle_multichannel_storage() local
830 request = &stor_device->init_request; in handle_multichannel_storage()
831 vstor_packet = &request->vstor_packet; in handle_multichannel_storage()
851 memset(request, 0, sizeof(struct storvsc_cmd_request)); in handle_multichannel_storage()
852 init_completion(&request->wait_event); in handle_multichannel_storage()
860 (unsigned long)request, in handle_multichannel_storage()
867 t = wait_for_completion_timeout(&request->wait_event, 10*HZ); in handle_multichannel_storage()
886 struct storvsc_cmd_request *request; in storvsc_channel_init() local
896 request = &stor_device->init_request; in storvsc_channel_init()
897 vstor_packet = &request->vstor_packet; in storvsc_channel_init()
[all …]
Dqlogicfas408.c96 static int ql_pdma(struct qlogicfas408_priv *priv, int phase, char *request, int reqlen) in ql_pdma() argument
106 insl(qbase + 4, request, 32); in ql_pdma()
108 request += 128; in ql_pdma()
113 insl(qbase + 4, request, 21); in ql_pdma()
115 request += 84; in ql_pdma()
118 insl(qbase + 4, request, 11); in ql_pdma()
120 request += 44; in ql_pdma()
132 *request++ = inb(qbase + 4); in ql_pdma()
143 outsl(qbase + 4, request, 32); in ql_pdma()
145 request += 128; in ql_pdma()
[all …]
Dscsi_lib.c122 struct request_queue *q = cmd->request->q; in scsi_mq_requeue_cmd()
124 blk_mq_requeue_request(cmd->request); in scsi_mq_requeue_cmd()
171 blk_requeue_request(q, cmd->request); in __scsi_queue_insert()
220 struct request *req; in scsi_execute()
531 struct request *req = cmd->request; in scsi_requeue_command()
616 if (cmd->request->cmd_type == REQ_TYPE_FS) { in scsi_uninit_cmd()
628 if (cmd->request->next_rq && cmd->request->next_rq->special) in scsi_mq_free_sgtables()
629 scsi_free_sgtable(cmd->request->next_rq->special, true); in scsi_mq_free_sgtables()
680 struct scsi_data_buffer *bidi_sdb = cmd->request->next_rq->special; in scsi_release_bidi_buffers()
684 cmd->request->next_rq->special = NULL; in scsi_release_bidi_buffers()
[all …]
Dsd_dif.c117 sdkp = scsi_disk(scmd->request->rq_disk); in sd_dif_prepare()
124 __rq_for_each_bio(bio, scmd->request) { in sd_dif_prepare()
168 sdkp = scsi_disk(scmd->request->rq_disk); in sd_dif_complete()
176 __rq_for_each_bio(bio, scmd->request) { in sd_dif_complete()
/linux-4.1.27/drivers/media/usb/pwc/
Dpwc-ctrl.c108 u8 request, u16 value, int recv_count) in recv_control_msg() argument
113 request, in recv_control_msg()
119 rc, request, value); in recv_control_msg()
144 u8 request, u16 value, void *buf, int buflen) in send_control_msg() argument
147 request, in send_control_msg()
412 int pwc_get_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_u8_ctrl() argument
416 ret = recv_control_msg(pdev, request, value, 1); in pwc_get_u8_ctrl()
424 int pwc_set_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, u8 data) in pwc_set_u8_ctrl() argument
429 ret = send_control_msg(pdev, request, value, pdev->ctrl_buf, 1); in pwc_set_u8_ctrl()
436 int pwc_get_s8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data) in pwc_get_s8_ctrl() argument
[all …]
Dpwc.h372 u8 request, u16 value, void *buf, int buflen);
375 int pwc_get_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data);
376 int pwc_set_u8_ctrl(struct pwc_device *pdev, u8 request, u16 value, u8 data);
377 int pwc_get_s8_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *data);
379 int pwc_get_u16_ctrl(struct pwc_device *pdev, u8 request, u16 value, int *dat);
380 int pwc_set_u16_ctrl(struct pwc_device *pdev, u8 request, u16 value, u16 data);
/linux-4.1.27/drivers/input/misc/
Duinput.c63 struct uinput_request *request) in uinput_request_alloc_id() argument
72 request->id = id; in uinput_request_alloc_id()
73 udev->requests[id] = request; in uinput_request_alloc_id()
94 struct uinput_request *request) in uinput_request_reserve_slot() argument
98 uinput_request_alloc_id(udev, request)); in uinput_request_reserve_slot()
102 struct uinput_request *request) in uinput_request_done() argument
105 udev->requests[request->id] = NULL; in uinput_request_done()
108 complete(&request->done); in uinput_request_done()
112 struct uinput_request *request) in uinput_request_send() argument
125 init_completion(&request->done); in uinput_request_send()
[all …]
/linux-4.1.27/drivers/clk/qcom/
Dclk-rcg2.c339 s64 request; in clk_edp_pixel_set_rate() local
349 request = rate; in clk_edp_pixel_set_rate()
350 request *= frac->den; in clk_edp_pixel_set_rate()
351 request = div_s64(request, frac->num); in clk_edp_pixel_set_rate()
352 if ((src_rate < (request - delta)) || in clk_edp_pixel_set_rate()
353 (src_rate > (request + delta))) in clk_edp_pixel_set_rate()
387 s64 request; in clk_edp_pixel_determine_rate() local
401 request = rate; in clk_edp_pixel_determine_rate()
402 request *= frac->den; in clk_edp_pixel_determine_rate()
403 request = div_s64(request, frac->num); in clk_edp_pixel_determine_rate()
[all …]
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Dprofile.c70 struct mlx4_profile *request, in mlx4_make_profile() argument
106 request->num_mtt = in mlx4_make_profile()
107 roundup_pow_of_two(max_t(unsigned, request->num_mtt, in mlx4_make_profile()
123 profile[MLX4_RES_QP].num = request->num_qp; in mlx4_make_profile()
124 profile[MLX4_RES_RDMARC].num = request->num_qp * request->rdmarc_per_qp; in mlx4_make_profile()
125 profile[MLX4_RES_ALTC].num = request->num_qp; in mlx4_make_profile()
126 profile[MLX4_RES_AUXC].num = request->num_qp; in mlx4_make_profile()
127 profile[MLX4_RES_SRQ].num = request->num_srq; in mlx4_make_profile()
128 profile[MLX4_RES_CQ].num = request->num_cq; in mlx4_make_profile()
131 profile[MLX4_RES_DMPT].num = request->num_mpt; in mlx4_make_profile()
[all …]
/linux-4.1.27/Documentation/scsi/
Dhptiop.txt85 A request packet can be allocated in either IOP or host memory.
87 To send a request to the controller:
89 - Get a free request packet by reading the inbound queue port or
90 allocate a free request in host DMA coherent memory.
104 - The IOP process the request. When the request is completed, it
108 For requests allocated in IOP memory, the request offset is posted to
113 flag is set in the request, the low 32-bit context value will be
116 - The host read the outbound queue and complete the request.
118 For requests allocated in IOP memory, the host driver free the request
131 To send a request to the controller:
[all …]
Dosd.txt75 its state throughout execution. Each request goes through these stages:
77 a. osd_start_request() allocates the request.
79 b. Any of the osd_req_* methods is used to encode a request of the specified
84 can be called multiple times on the same request. However, only one
88 and signs the request using the provided capability key and integrity-
91 e. osd_execute_request() may be called to execute the request via the block
92 layer and wait for its completion. The request can be executed
95 f. After execution, osd_req_decode_sense() can be called to decode the request's
101 h. osd_end_request() must be called to deallocate the request and any resource
102 associated with it. Note that osd_end_request cleans up the request at any
[all …]
/linux-4.1.27/net/wireless/
Dreg.c280 static void reg_free_request(struct regulatory_request *request) in reg_free_request() argument
282 if (request != get_last_request()) in reg_free_request()
283 kfree(request); in reg_free_request()
294 static void reg_update_last_request(struct regulatory_request *request) in reg_update_last_request() argument
299 if (lr == request) in reg_update_last_request()
303 rcu_assign_pointer(last_request, request); in reg_update_last_request()
466 struct reg_regdb_search_request *request; in reg_regdb_search() local
474 request = list_first_entry(&reg_regdb_search_list, in reg_regdb_search()
477 list_del(&request->list); in reg_regdb_search()
482 if (alpha2_equal(request->alpha2, curdom->alpha2)) { in reg_regdb_search()
[all …]
Dnl80211.h21 struct regulatory_request *request);
24 nl80211_send_reg_change_event(struct regulatory_request *request) in nl80211_send_reg_change_event() argument
26 nl80211_common_reg_change_event(NL80211_CMD_REG_CHANGE, request); in nl80211_send_reg_change_event()
30 nl80211_send_wiphy_reg_change_event(struct regulatory_request *request) in nl80211_send_wiphy_reg_change_event() argument
32 nl80211_common_reg_change_event(NL80211_CMD_WIPHY_REG_CHANGE, request); in nl80211_send_wiphy_reg_change_event()
Dsme.c63 struct cfg80211_scan_request *request; in cfg80211_conn_scan() local
77 request = kzalloc(sizeof(*request) + sizeof(request->ssids[0]) + in cfg80211_conn_scan()
78 sizeof(request->channels[0]) * n_channels, in cfg80211_conn_scan()
80 if (!request) in cfg80211_conn_scan()
89 kfree(request); in cfg80211_conn_scan()
92 request->channels[0] = wdev->conn->params.channel; in cfg80211_conn_scan()
93 request->rates[band] = (1 << sband->n_bitrates) - 1; in cfg80211_conn_scan()
108 request->channels[i++] = channel; in cfg80211_conn_scan()
110 request->rates[band] = (1 << bands->n_bitrates) - 1; in cfg80211_conn_scan()
114 request->n_channels = n_channels; in cfg80211_conn_scan()
[all …]
Dscan.c168 struct cfg80211_scan_request *request; in ___cfg80211_scan_done() local
183 request = rdev->scan_req; in ___cfg80211_scan_done()
184 if (!request) in ___cfg80211_scan_done()
187 wdev = request->wdev; in ___cfg80211_scan_done()
197 if (!request->aborted && in ___cfg80211_scan_done()
198 request->flags & NL80211_SCAN_FLAG_FLUSH) { in ___cfg80211_scan_done()
201 __cfg80211_bss_expire(rdev, request->scan_start); in ___cfg80211_scan_done()
205 msg = nl80211_build_scan_msg(rdev, wdev, request->aborted); in ___cfg80211_scan_done()
208 if (wdev->netdev && !request->aborted) { in ___cfg80211_scan_done()
219 kfree(request); in ___cfg80211_scan_done()
[all …]
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/
Dframework.c369 sfw_get_stats(srpc_stat_reqst_t *request, srpc_stat_reply_t *reply) in sfw_get_stats() argument
378 if (request->str_sid.ses_nid == LNET_NID_ANY) { in sfw_get_stats()
383 if (sn == NULL || !sfw_sid_equal(request->str_sid, sn->sn_id)) { in sfw_get_stats()
412 sfw_make_session(srpc_mksn_reqst_t *request, srpc_mksn_reply_t *reply) in sfw_make_session() argument
415 srpc_msg_t *msg = container_of(request, srpc_msg_t, in sfw_make_session()
419 if (request->mksn_sid.ses_nid == LNET_NID_ANY) { in sfw_make_session()
430 if (sfw_sid_equal(request->mksn_sid, sn->sn_id)) { in sfw_make_session()
435 if (!request->mksn_force) { in sfw_make_session()
463 sfw_init_session(sn, request->mksn_sid, in sfw_make_session()
464 msg->msg_ses_feats, &request->mksn_name[0]); in sfw_make_session()
[all …]
/linux-4.1.27/arch/microblaze/kernel/
Dptrace.c77 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
83 switch (request) { in arch_ptrace()
89 if (addr >= PT_SIZE && request == PTRACE_PEEKUSR) { in arch_ptrace()
106 if (request == PTRACE_PEEKUSR) in arch_ptrace()
126 if (rval == 0 && request == PTRACE_PEEKUSR) in arch_ptrace()
130 rval = ptrace_request(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/security/smack/
Dsmack_access.c126 int request, struct smk_audit_info *a) in smk_access() argument
163 if ((request & MAY_ANYREAD) == request || in smk_access()
164 (request & MAY_LOCK) == request) { in smk_access()
182 if (may <= 0 || (request & may) != request) { in smk_access()
210 request, rc, a); in smk_access()
332 if (sad->request[0] == '\0') in smack_log_callback()
335 audit_log_format(ab, " requested=%s", sad->request); in smack_log_callback()
349 void smack_log(char *subject_label, char *object_label, int request, in smack_log() argument
372 smack_str_from_perm(request_buffer, request); in smack_log()
390 sad->request = request_buffer; in smack_log()
[all …]
/linux-4.1.27/include/drm/
Ddrm_agpsupport.h56 int drm_agp_alloc(struct drm_device *dev, struct drm_agp_buffer *request);
59 int drm_agp_free(struct drm_device *dev, struct drm_agp_buffer *request);
62 int drm_agp_unbind(struct drm_device *dev, struct drm_agp_binding *request);
65 int drm_agp_bind(struct drm_device *dev, struct drm_agp_binding *request);
150 struct drm_agp_buffer *request) in drm_agp_alloc() argument
162 struct drm_agp_buffer *request) in drm_agp_free() argument
174 struct drm_agp_binding *request) in drm_agp_unbind() argument
186 struct drm_agp_binding *request) in drm_agp_bind() argument
/linux-4.1.27/drivers/block/
Dnbd.c50 struct request *active_req;
99 static void nbd_end_request(struct nbd_device *nbd, struct request *req) in nbd_end_request()
228 static int nbd_send_req(struct nbd_device *nbd, struct request *req) in nbd_send_req()
231 struct nbd_request request; in nbd_send_req() local
234 memset(&request, 0, sizeof(request)); in nbd_send_req()
235 request.magic = htonl(NBD_REQUEST_MAGIC); in nbd_send_req()
236 request.type = htonl(nbd_cmd(req)); in nbd_send_req()
239 request.from = cpu_to_be64((u64)blk_rq_pos(req) << 9); in nbd_send_req()
240 request.len = htonl(size); in nbd_send_req()
242 memcpy(request.handle, &req, sizeof(req)); in nbd_send_req()
[all …]
/linux-4.1.27/tools/usb/usbip/src/
Dusbip_attach.c123 struct op_import_request request; in query_import_device() local
127 memset(&request, 0, sizeof(request)); in query_import_device()
137 strncpy(request.busid, busid, SYSFS_BUS_ID_SIZE-1); in query_import_device()
139 PACK_OP_IMPORT_REQUEST(0, &request); in query_import_device()
141 rc = usbip_net_send(sockfd, (void *) &request, sizeof(request)); in query_import_device()
Dusbip_network.h78 #define PACK_OP_IMPORT_REQUEST(pack, request) do {\ argument
100 #define PACK_OP_EXPORT_REQUEST(pack, request) do {\ argument
101 usbip_net_pack_usb_device(pack, &(request)->udev);\
121 #define PACK_OP_UNEXPORT_REQUEST(pack, request) do {\ argument
122 usbip_net_pack_usb_device(pack, &(request)->udev);\
163 #define PACK_OP_DEVLIST_REQUEST(pack, request) do {\ argument
/linux-4.1.27/drivers/hid/
Dhid-hyperv.c136 struct synthhid_protocol_request request; member
386 struct mousevsc_prt_msg *request; in mousevsc_connect_to_vsp() local
389 request = &input_dev->protocol_req; in mousevsc_connect_to_vsp()
390 memset(request, 0, sizeof(struct mousevsc_prt_msg)); in mousevsc_connect_to_vsp()
392 request->type = PIPE_MESSAGE_DATA; in mousevsc_connect_to_vsp()
393 request->size = sizeof(struct synthhid_protocol_request); in mousevsc_connect_to_vsp()
394 request->request.header.type = SYNTH_HID_PROTOCOL_REQUEST; in mousevsc_connect_to_vsp()
395 request->request.header.size = sizeof(unsigned int); in mousevsc_connect_to_vsp()
396 request->request.version_requested.version = SYNTHHID_INPUT_VERSION; in mousevsc_connect_to_vsp()
398 ret = vmbus_sendpacket(device->channel, request, in mousevsc_connect_to_vsp()
[all …]
/linux-4.1.27/Documentation/ABI/testing/
Dsysfs-driver-ppi21 What: /sys/devices/pnp0/<bus-num>/ppi/request
25 This attribute shows the request for an operation to be
27 the OS to the pre-OS environment. The request should be an
28 integer value range from 1 to 160, and 0 means no request.
36 request it acted upon. The format is "<request> <response num>
54 This attribute shows whether it is allowed to request an
57 The format is "<request> <status num>: <status description>".
65 This attribute shows whether it is allowed to request an
Dconfigfs-usb-gadget-uac111 req_buf_size - ISO OUT endpoint request buffer size
12 req_count - ISO OUT endpoint request count
/linux-4.1.27/fs/
Dlocks.c865 static int flock_lock_inode(struct inode *inode, struct file_lock *request) in flock_lock_inode() argument
874 ctx = locks_get_lock_context(inode, request->fl_type); in flock_lock_inode()
876 if (request->fl_type != F_UNLCK) in flock_lock_inode()
878 return (request->fl_flags & FL_EXISTS) ? -ENOENT : 0; in flock_lock_inode()
881 if (!(request->fl_flags & FL_ACCESS) && (request->fl_type != F_UNLCK)) { in flock_lock_inode()
888 if (request->fl_flags & FL_ACCESS) in flock_lock_inode()
892 if (request->fl_file != fl->fl_file) in flock_lock_inode()
894 if (request->fl_type == fl->fl_type) in flock_lock_inode()
901 if (request->fl_type == F_UNLCK) { in flock_lock_inode()
902 if ((request->fl_flags & FL_EXISTS) && !found) in flock_lock_inode()
[all …]
/linux-4.1.27/tools/firewire/
Ddecode-fcp.c147 (struct avc_frame *) t->request->packet.write_block.data; in decode_avc()
176 (struct avc_frame *) t->request->packet.write_block.data; in decode_fcp()
178 ((unsigned long long) t->request->packet.common.offset_high << 32) | in decode_fcp()
179 t->request->packet.common.offset_low; in decode_fcp()
181 if (t->request->packet.common.tcode != TCODE_WRITE_BLOCK_REQUEST) in decode_fcp()
/linux-4.1.27/drivers/media/rc/img-ir/
Dimg-ir-rc5.c16 struct img_ir_scancode_req *request) in img_ir_rc5_scancode() argument
36 request->protocol = RC_TYPE_RC5; in img_ir_rc5_scancode()
37 request->scancode = addr << 8 | cmd; in img_ir_rc5_scancode()
38 request->toggle = tgl; in img_ir_rc5_scancode()
Dimg-ir-rc6.c16 struct img_ir_scancode_req *request) in img_ir_rc6_scancode() argument
57 request->protocol = RC_TYPE_RC6_0; in img_ir_rc6_scancode()
58 request->scancode = addr << 8 | cmd; in img_ir_rc6_scancode()
59 request->toggle = trl2; in img_ir_rc6_scancode()
Dimg-ir-sony.c16 struct img_ir_scancode_req *request) in img_ir_sony_scancode() argument
28 request->protocol = RC_TYPE_SONY12; in img_ir_sony_scancode()
37 request->protocol = RC_TYPE_SONY15; in img_ir_sony_scancode()
47 request->protocol = RC_TYPE_SONY20; in img_ir_sony_scancode()
52 request->scancode = dev << 16 | subdev << 8 | func; in img_ir_sony_scancode()
Dimg-ir-nec.c17 struct img_ir_scancode_req *request) in img_ir_nec_scancode() argument
33 request->scancode = bitrev8(addr) << 24 | in img_ir_nec_scancode()
40 request->scancode = addr << 16 | in img_ir_nec_scancode()
46 request->scancode = addr << 8 | in img_ir_nec_scancode()
49 request->protocol = RC_TYPE_NEC; in img_ir_nec_scancode()
Dimg-ir-jvc.c16 struct img_ir_scancode_req *request) in img_ir_jvc_scancode() argument
26 request->protocol = RC_TYPE_JVC; in img_ir_jvc_scancode()
27 request->scancode = cust << 8 | data; in img_ir_jvc_scancode()
Dimg-ir-sharp.c16 struct img_ir_scancode_req *request) in img_ir_sharp_scancode() argument
35 request->protocol = RC_TYPE_SHARP; in img_ir_sharp_scancode()
36 request->scancode = addr << 8 | cmd; in img_ir_sharp_scancode()
Dimg-ir-sanyo.c27 struct img_ir_scancode_req *request) in img_ir_sanyo_scancode() argument
47 request->protocol = RC_TYPE_SANYO; in img_ir_sanyo_scancode()
48 request->scancode = addr << 8 | data; in img_ir_sanyo_scancode()
/linux-4.1.27/include/scsi/
Dscsi_cmnd.h110 struct request *request; /* The command we are member
155 return *(struct scsi_driver **)cmd->request->rq_disk->private_data; in scsi_cmd_to_driver()
201 return blk_bidi_rq(cmd->request) && in scsi_bidi_cmnd()
202 (cmd->request->next_rq->special != NULL); in scsi_bidi_cmnd()
208 cmd->request->next_rq->special : &cmd->sdb; in scsi_in()
294 return blk_rq_pos(scmd->request); in scsi_get_lba()
304 return blk_rq_pos(scmd->request) >> in scsi_prot_ref_tag()
Dscsi_tcq.h17 struct request *req = NULL; in scsi_mq_find_tag()
35 struct request *req; in scsi_find_tag()
89 struct request *req; in scsi_host_find_tag()
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/
Dhif-ops.h25 u32 len, u32 request) in hif_read_write_sync() argument
29 (request & HIF_WRITE) ? "write" : "read", in hif_read_write_sync()
30 addr, buf, len, request); in hif_read_write_sync()
32 return ar->hif_ops->read_write_sync(ar, addr, buf, len, request); in hif_read_write_sync()
36 u32 length, u32 request, in hif_write_async() argument
41 address, buffer, length, request); in hif_write_async()
44 request, packet); in hif_write_async()
Dsdio.c148 static int ath6kl_sdio_io(struct sdio_func *func, u32 request, u32 addr, in ath6kl_sdio_io() argument
155 if (request & HIF_WRITE) { in ath6kl_sdio_io()
165 if (request & HIF_FIXED_ADDRESS) in ath6kl_sdio_io()
170 if (request & HIF_FIXED_ADDRESS) in ath6kl_sdio_io()
179 request & HIF_WRITE ? "wr" : "rd", addr, in ath6kl_sdio_io()
180 request & HIF_FIXED_ADDRESS ? " (fixed)" : "", buf, len); in ath6kl_sdio_io()
183 trace_ath6kl_sdio(addr, request, buf, len); in ath6kl_sdio_io()
408 u32 len, u32 request) in ath6kl_sdio_read_write_sync() argument
415 if (request & HIF_BLOCK_BASIS) in ath6kl_sdio_read_write_sync()
424 if (request & HIF_WRITE) in ath6kl_sdio_read_write_sync()
[all …]
/linux-4.1.27/drivers/scsi/isci/
Dtask.h169 struct isci_request *request,
173 struct isci_request *request);
176 struct isci_request *request);
180 struct isci_request *request);
183 struct isci_request *request);
Drequest.c2502 struct isci_request *request, in isci_request_set_open_reject_status() argument
2509 set_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_set_open_reject_status()
2526 struct isci_request *request, in isci_request_handle_controller_specific_errors() argument
2533 cstatus = request->scu_status; in isci_request_handle_controller_specific_errors()
2535 dev_dbg(&request->isci_host->pdev->dev, in isci_request_handle_controller_specific_errors()
2538 __func__, request, cstatus); in isci_request_handle_controller_specific_errors()
2569 set_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_handle_controller_specific_errors()
2579 clear_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_handle_controller_specific_errors()
2607 set_bit(IREQ_COMPLETE_IN_TARGET, &request->flags); in isci_request_handle_controller_specific_errors()
2617 request, task, response_ptr, status_ptr, in isci_request_handle_controller_specific_errors()
[all …]
/linux-4.1.27/drivers/net/wireless/ath/
Dregd.c466 struct regulatory_request *request) in __ath_reg_dyn_country() argument
470 if (request->initiator == NL80211_REGDOM_SET_BY_COUNTRY_IE && in __ath_reg_dyn_country()
474 country_code = ath_regd_find_country_by_name(request->alpha2); in __ath_reg_dyn_country()
483 ath_reg_apply_world_flags(wiphy, request->initiator, reg); in __ath_reg_dyn_country()
490 struct regulatory_request *request) in ath_reg_dyn_country() argument
492 if (__ath_reg_dyn_country(wiphy, reg, request)) in ath_reg_dyn_country()
498 reg_initiator_name(request->initiator)); in ath_reg_dyn_country()
502 struct regulatory_request *request, in ath_reg_notifier_apply() argument
515 if (!request) in ath_reg_notifier_apply()
518 reg->region = request->dfs_region; in ath_reg_notifier_apply()
[all …]
/linux-4.1.27/sound/usb/6fire/
Dcomm.c59 static void usb6fire_comm_init_buffer(u8 *buffer, u8 id, u8 request, in usb6fire_comm_init_buffer() argument
63 buffer[2] = request; in usb6fire_comm_init_buffer()
65 switch (request) { in usb6fire_comm_init_buffer()
110 static int usb6fire_comm_write8(struct comm_runtime *rt, u8 request, in usb6fire_comm_write8() argument
121 usb6fire_comm_init_buffer(buffer, 0x00, request, reg, value, 0x00); in usb6fire_comm_write8()
128 static int usb6fire_comm_write16(struct comm_runtime *rt, u8 request, in usb6fire_comm_write16() argument
139 usb6fire_comm_init_buffer(buffer, 0x00, request, reg, vl, vh); in usb6fire_comm_write16()
Dcomm.h34 int (*write8)(struct comm_runtime *rt, u8 request, u8 reg, u8 value);
35 int (*write16)(struct comm_runtime *rt, u8 request, u8 reg,
/linux-4.1.27/drivers/gpu/drm/via/
Dvia_irq.c358 if (irqwait->request.irq >= dev_priv->num_irqs) { in via_wait_irq()
360 irqwait->request.irq); in via_wait_irq()
364 cur_irq += irqwait->request.irq; in via_wait_irq()
366 switch (irqwait->request.type & ~VIA_IRQ_FLAGS_MASK) { in via_wait_irq()
368 irqwait->request.sequence += in via_wait_irq()
370 irqwait->request.type &= ~_DRM_VBLANK_RELATIVE; in via_wait_irq()
377 if (irqwait->request.type & VIA_IRQ_SIGNAL) { in via_wait_irq()
382 force_sequence = (irqwait->request.type & VIA_IRQ_FORCE_SEQUENCE); in via_wait_irq()
384 ret = via_driver_irq_wait(dev, irqwait->request.irq, force_sequence, in via_wait_irq()
385 &irqwait->request.sequence); in via_wait_irq()
/linux-4.1.27/drivers/staging/lustre/lustre/mdc/
Dmdc_reint.c47 static int mdc_reint(struct ptlrpc_request *request, in mdc_reint() argument
53 request->rq_send_state = level; in mdc_reint()
56 rc = ptlrpc_queue_wait(request); in mdc_reint()
60 else if (!req_capsule_server_get(&request->rq_pill, &RMF_MDT_BODY)) in mdc_reint()
105 struct ptlrpc_request **request, struct md_open_data **mod) in mdc_setattr() argument
197 *request = req; in mdc_setattr()
210 struct ptlrpc_request **request) in mdc_create() argument
314 *request = req; in mdc_create()
319 struct ptlrpc_request **request) in mdc_unlink() argument
323 struct ptlrpc_request *req = *request; in mdc_unlink()
[all …]
Dmdc_internal.h135 struct ptlrpc_request **request);
137 struct ptlrpc_request **request);
140 struct ptlrpc_request **request);
143 struct ptlrpc_request **request, struct md_open_data **mod);
145 struct ptlrpc_request **request);
/linux-4.1.27/fs/dlm/
Drequestqueue.c24 struct dlm_message request; member
47 memcpy(&e->request, ms, ms->m_header.h_length); in dlm_add_requestqueue()
82 ms = &e->request; in dlm_process_requestqueue()
90 dlm_receive_message_saved(ls, &e->request, e->recover_seq); in dlm_process_requestqueue()
162 ms = &e->request; in dlm_purge_requestqueue()
/linux-4.1.27/arch/parisc/kernel/
Dptrace.c114 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
120 switch (request) { in arch_ptrace()
169 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
204 long compat_arch_ptrace(struct task_struct *child, compat_long_t request, in compat_arch_ptrace() argument
210 switch (request) { in compat_arch_ptrace()
238 ret = arch_ptrace(child, request, addr, data); in compat_arch_ptrace()
262 ret = compat_ptrace_request(child, request, addr, data); in compat_arch_ptrace()
/linux-4.1.27/drivers/usb/dwc3/
Dep0.c113 req->request.actual = 0; in __dwc3_gadget_ep0_queue()
114 req->request.status = -EINPROGRESS; in __dwc3_gadget_ep0_queue()
212 int dwc3_gadget_ep0_queue(struct usb_ep *ep, struct usb_request *request, in dwc3_gadget_ep0_queue() argument
215 struct dwc3_request *req = to_dwc3_request(request); in dwc3_gadget_ep0_queue()
227 request, dep->name); in dwc3_gadget_ep0_queue()
240 request, dep->name, request->length, in dwc3_gadget_ep0_queue()
383 dwc->ep0_usb_req.request.length = sizeof(*response_pkt); in dwc3_ep0_handle_status()
384 dwc->ep0_usb_req.request.buf = dwc->setup_buf; in dwc3_ep0_handle_status()
385 dwc->ep0_usb_req.request.complete = dwc3_ep0_status_cmpl; in dwc3_ep0_handle_status()
667 dwc->ep0_usb_req.request.length = dep->endpoint.maxpacket; in dwc3_ep0_set_sel()
[all …]
Dgadget.c253 } while(++i < req->request.num_mapped_sgs); in dwc3_gadget_giveback()
259 if (req->request.status == -EINPROGRESS) in dwc3_gadget_giveback()
260 req->request.status = status; in dwc3_gadget_giveback()
265 usb_gadget_unmap_request(&dwc->gadget, &req->request, in dwc3_gadget_giveback()
269 req, dep->name, req->request.actual, in dwc3_gadget_giveback()
270 req->request.length, status); in dwc3_gadget_giveback()
274 usb_gadget_giveback_request(&dep->endpoint, &req->request); in dwc3_gadget_giveback()
759 return &req->request; in dwc3_gadget_ep_alloc_request()
763 struct usb_request *request) in dwc3_gadget_ep_free_request() argument
765 struct dwc3_request *req = to_dwc3_request(request); in dwc3_gadget_ep_free_request()
[all …]
/linux-4.1.27/drivers/net/ethernet/cisco/enic/
Denic_pp.c228 if (pp->request != PORT_REQUEST_ASSOCIATE) { in enic_pp_preassociate_rr()
246 if (pp->request != PORT_REQUEST_ASSOCIATE) { in enic_pp_preassociate_rr()
269 if (prev_pp->request != PORT_REQUEST_PREASSOCIATE_RR || in enic_pp_associate()
270 (prev_pp->request == PORT_REQUEST_PREASSOCIATE_RR && in enic_pp_associate()
314 if (pp->request >= enic_pp_handlers_count in enic_process_set_pp_request()
315 || !enic_pp_handlers[pp->request]) in enic_process_set_pp_request()
318 return enic_pp_handlers[pp->request](enic, vf, prev_pp, restore_pp); in enic_process_set_pp_request()
322 int request, u16 *response) in enic_process_get_pp_request() argument
326 switch (request) { in enic_process_get_pp_request()
/linux-4.1.27/arch/powerpc/platforms/pseries/
Dmsi.c231 int request; member
261 req = counts->request; in count_spare_msis()
283 static int msi_quota_for_device(struct pci_dev *dev, int request) in msi_quota_for_device() argument
290 request); in msi_quota_for_device()
315 if (request <= counts.quota) in msi_quota_for_device()
320 counts.request = request; in msi_quota_for_device()
332 request = min(counts.quota, request); in msi_quota_for_device()
334 pr_debug("rtas_msi: request clamped to quota %d\n", request); in msi_quota_for_device()
338 return request; in msi_quota_for_device()
/linux-4.1.27/Documentation/filesystems/
Dfuse.txt151 If a process issuing a FUSE filesystem request is interrupted, the
154 1) If the request is not yet sent to userspace AND the signal is
155 fatal (SIGKILL or unhandled fatal signal), then the request is
158 2) If the request is not yet sent to userspace AND the signal is not
159 fatal, then an 'interrupted' flag is set for the request. When
160 the request has been successfully transferred to userspace and
161 this flag is set, an INTERRUPT request is queued.
163 3) If the request is already sent to userspace, then an INTERRUPT
164 request is queued.
170 or may honor them by sending a reply to the _original_ request, with
[all …]
/linux-4.1.27/drivers/media/usb/dvb-usb/
Ddtv5100.c37 u8 request; in dtv5100_i2c_msg() local
45 request = (addr == DTV5100_DEMOD_ADDR ? DTV5100_DEMOD_READ : in dtv5100_i2c_msg()
52 request = (addr == DTV5100_DEMOD_ADDR ? DTV5100_DEMOD_WRITE : in dtv5100_i2c_msg()
64 return usb_control_msg(d->udev, usb_rcvctrlpipe(d->udev, 0), request, in dtv5100_i2c_msg()
150 for (i = 0; dtv5100_init[i].request; i++) { in dtv5100_probe()
152 dtv5100_init[i].request, in dtv5100_probe()
Dopera1.c52 static int opera1_xilinx_rw(struct usb_device *dev, u8 request, u16 value, in opera1_xilinx_rw() argument
68 ret = usb_control_msg(dev, pipe, request, in opera1_xilinx_rw()
72 if (request == OPERA_TUNER_REQ) { in opera1_xilinx_rw()
95 u8 request; in opera1_usb_i2c_msgxfer() local
107 request=0xb6; in opera1_usb_i2c_msgxfer()
111 request=0xb6; in opera1_usb_i2c_msgxfer()
115 request=0xb1; in opera1_usb_i2c_msgxfer()
119 request=0xb8; in opera1_usb_i2c_msgxfer()
123 request=0xb1; in opera1_usb_i2c_msgxfer()
126 ret = opera1_xilinx_rw(dev->udev, request, in opera1_usb_i2c_msgxfer()
/linux-4.1.27/drivers/staging/rtl8188eu/os_dep/
Dusb_ops_linux.c230 static int usbctrl_vendorreq(struct adapter *adapt, u8 request, u16 value, u16 index, void *pdata, … in usbctrl_vendorreq() argument
275 …status = usb_control_msg(udev, pipe, request, reqtype, value, index, pIo_buf, len, RTW_USB_CONTROL… in usbctrl_vendorreq()
315 u8 request; in usb_read8() local
323 request = 0x05; in usb_read8()
330 usbctrl_vendorreq(adapter, request, wvalue, index, &data, len, requesttype); in usb_read8()
339 u8 request; in usb_read16() local
346 request = 0x05; in usb_read16()
351 usbctrl_vendorreq(adapter, request, wvalue, index, &data, len, requesttype); in usb_read16()
358 u8 request; in usb_read32() local
366 request = 0x05; in usb_read32()
[all …]
/linux-4.1.27/fs/afs/
Dcmservice.c138 ASSERT(call->server && call->count && call->request); in afs_cm_destructor()
139 afs_break_callbacks(call->server, call->count, call->request); in afs_cm_destructor()
164 afs_break_callbacks(call->server, call->count, call->request); in SRXAFSCB_CallBack()
220 call->request = kcalloc(call->count, in afs_deliver_cb_callback()
223 if (!call->request) in afs_deliver_cb_callback()
226 cb = call->request; in afs_deliver_cb_callback()
259 ret = afs_extract_data(call, skb, last, call->request, in afs_deliver_cb_callback()
268 cb = call->request; in afs_deliver_cb_callback()
432 struct afs_uuid *r = call->request; in SRXAFSCB_ProbeUuid()
487 call->request = kmalloc(sizeof(struct afs_uuid), GFP_KERNEL); in afs_deliver_cb_probe_uuid()
[all …]
/linux-4.1.27/sound/soc/intel/baytrail/
Dsst-baytrail-ipc.c154 struct sst_byt_alloc_params request; member
379 stream->request.pcm_params.pcm_wd_sz = bits; in sst_byt_stream_set_bits()
386 stream->request.pcm_params.num_chan = channels; in sst_byt_stream_set_channels()
393 stream->request.pcm_params.sfreq = rate; in sst_byt_stream_set_rate()
401 stream->request.str_type.codec_type = codec_type; in sst_byt_stream_type()
402 stream->request.str_type.str_type = stream_type; in sst_byt_stream_type()
403 stream->request.str_type.operation = operation; in sst_byt_stream_type()
404 stream->request.str_type.time_slots = 0xc; in sst_byt_stream_type()
412 stream->request.frame_info.num_entries = 1; in sst_byt_stream_buffer()
413 stream->request.frame_info.ring_buf_info[0].addr = buffer_addr; in sst_byt_stream_buffer()
[all …]
/linux-4.1.27/drivers/media/usb/as102/
Das102_fw.c125 fw_pkt.u.request[0] = 0x00; in as102_firmware_upload()
126 fw_pkt.u.request[1] = 0x03; in as102_firmware_upload()
137 fw_pkt.u.request[0] = 0x00; in as102_firmware_upload()
138 fw_pkt.u.request[1] = 0x01; in as102_firmware_upload()
140 data_len += sizeof(fw_pkt.u.request); in as102_firmware_upload()
/linux-4.1.27/Documentation/video4linux/bttv/
DREADME.quirks25 demultiplexed by the audio request signal. Thus the arbiter defaults to
27 bus access. This is desirable since the video will request the bus more
29 the audio will have first access to the bus even when issuing a request
30 after the video request but before the PCI external arbiter has granted
44 (2) Do not reassert REQ to request another bus transaction until after
57 only the granted request to the REQ pin. The arbiter decision lock
61 responsibility to remove its request at the same time. It is the
62 arbiters responsibility to allow this request to flow through to REQ and
63 not allow the other request to hold REQ asserted. The decision lock may
/linux-4.1.27/drivers/media/usb/dvb-usb-v2/
Dec168.c32 u8 request, requesttype; in ec168_ctrl_msg() local
41 request = req->cmd; in ec168_ctrl_msg()
45 request = req->cmd; in ec168_ctrl_msg()
49 request = CONFIG; in ec168_ctrl_msg()
53 request = CONFIG; in ec168_ctrl_msg()
57 request = DEMOD_RW; in ec168_ctrl_msg()
61 request = DEMOD_RW; in ec168_ctrl_msg()
87 ret = usb_control_msg(d->udev, pipe, request, requesttype, req->value, in ec168_ctrl_msg()
90 dvb_usb_dbg_usb_control_msg(d->udev, request, requesttype, req->value, in ec168_ctrl_msg()
Dau6610.c82 u8 request; in au6610_i2c_msg() local
86 request = AU6610_REQ_I2C_WRITE; in au6610_i2c_msg()
88 request = AU6610_REQ_I2C_READ; in au6610_i2c_msg()
91 return au6610_usb_msg(d, request, addr, wbuf, wlen, rbuf, rlen); in au6610_i2c_msg()
/linux-4.1.27/Documentation/devicetree/bindings/dma/
Dfsl-edma.txt5 specific DMA request source can only be multiplexed by any channel of certain
23 Specific request source can only be multiplexed by specific channels
25 The 2nd cell specifies the request source(slot) ID.
26 See the SoC's reference manual for all the supported request sources.
61 specifies the channel group(DMAMUX) in which this request can be multiplexed,
62 and the 2nd specifies the request source.
Dfsl-imx-dma.txt4 DMA request bindings as described in dma/dma.txt .
35 - dmas: List of one or more DMA request specifiers. One DMA request specifier
37 specifying the request line.
Darm-pl330.txt34 [property name] = <[phandle of the dma controller] [dma request id]>;
37 where 'dma request id' is the dma request number which is connected
40 names correspond 1:1 with the dma request ids in the dmas property.
Dk3dma.txt12 have specific request line
30 Use specific request line passing from dmax
31 For example, i2c0 read channel request line is 18, while write channel use 19
Dmoxa,moxart-dma.txt11 - #dma-cells : Should be 1, a single cell holding a line request number
34 Use specific request line passing from dma
35 For example, MMC request line is 5
Dsirfsoc-dma.txt26 Fill the specific dma request line in dmas. In the below example, spi0 read
27 channel request line is 9 of the 2nd dma controller, while write channel uses
28 4 of the 2nd dma controller; spi1 read channel request line is 12 of the 1st
Ddma.txt1 * Generic DMA Controller and DMA request bindings
4 DMA request or channel information that goes from a hardware device to a DMA
17 - dma-requests: Number of DMA request signals supported by the
46 typically contains a DMA request line number or a
/linux-4.1.27/arch/c6x/kernel/
Dptrace.c142 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
147 switch (request) { in arch_ptrace()
153 if (ret == 0 && request == PTRACE_POKETEXT) in arch_ptrace()
157 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/drivers/md/
Ddm-target.c134 static int io_err_map_rq(struct dm_target *ti, struct request *clone, in io_err_map_rq()
140 static int io_err_clone_and_map_rq(struct dm_target *ti, struct request *rq, in io_err_clone_and_map_rq()
142 struct request **clone) in io_err_clone_and_map_rq()
147 static void io_err_release_clone_rq(struct request *clone) in io_err_release_clone_rq()
Ddm.c85 struct request *orig, *clone;
105 union map_info *dm_get_rq_mapinfo(struct request *rq) in dm_get_rq_mapinfo()
323 _rq_cache = kmem_cache_create("dm_clone_request", sizeof(struct request), in local_init()
324 __alignof__(struct request), 0, NULL); in local_init()
623 static struct request *alloc_clone_request(struct mapped_device *md, in alloc_clone_request()
629 static void free_clone_request(struct mapped_device *md, struct request *rq) in free_clone_request()
1044 static struct dm_rq_target_io *tio_from_request(struct request *rq) in tio_from_request()
1077 static void free_rq_clone(struct request *clone) in free_rq_clone()
1105 static void dm_end_request(struct request *clone, int error) in dm_end_request()
1110 struct request *rq = tio->orig; in dm_end_request()
[all …]
/linux-4.1.27/drivers/net/wireless/rt2x00/
Drt2x00usb.h103 const u8 request, const u8 requesttype,
126 const u8 request, const u8 requesttype,
144 const u8 request, const u8 requesttype,
161 const u8 request, in rt2x00usb_vendor_request_sw() argument
166 return rt2x00usb_vendor_request(rt2x00dev, request, in rt2x00usb_vendor_request_sw()
/linux-4.1.27/drivers/media/rc/
Digorplugusb.c46 struct usb_ctrlrequest request; member
133 ir->request.bRequest = cmd; in igorplugusb_cmd()
179 ir->request.bRequest = GET_INFRACODE; in igorplugusb_probe()
180 ir->request.bRequestType = USB_TYPE_VENDOR | USB_DIR_IN; in igorplugusb_probe()
181 ir->request.wLength = cpu_to_le16(sizeof(ir->buf_in)); in igorplugusb_probe()
188 usb_rcvctrlpipe(udev, 0), (uint8_t *)&ir->request, in igorplugusb_probe()
/linux-4.1.27/arch/tile/kernel/
Dptrace.c150 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
159 switch (request) { in arch_ptrace()
221 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
233 ret = compat_ptrace_request(child, request, in arch_ptrace()
238 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
247 long compat_arch_ptrace(struct task_struct *child, compat_long_t request, in compat_arch_ptrace() argument
/linux-4.1.27/arch/um/kernel/
Dptrace.c43 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
50 switch (request) { in arch_ptrace()
105 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
107 ret = subarch_ptrace(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/drivers/input/serio/
Dhyperv-keyboard.c288 struct synth_kbd_protocol_request *request; in hv_kbd_connect_to_vsp() local
293 request = &kbd_dev->protocol_req; in hv_kbd_connect_to_vsp()
294 memset(request, 0, sizeof(struct synth_kbd_protocol_request)); in hv_kbd_connect_to_vsp()
295 request->header.type = __cpu_to_le32(SYNTH_KBD_PROTOCOL_REQUEST); in hv_kbd_connect_to_vsp()
296 request->version_requested.version = __cpu_to_le32(SYNTH_KBD_VERSION); in hv_kbd_connect_to_vsp()
298 error = vmbus_sendpacket(hv_dev->channel, request, in hv_kbd_connect_to_vsp()
300 (unsigned long)request, in hv_kbd_connect_to_vsp()
/linux-4.1.27/drivers/net/wireless/rtlwifi/
Dregd.c307 struct regulatory_request *request, in _rtl_reg_notifier_apply() argument
313 switch (request->initiator) { in _rtl_reg_notifier_apply()
319 _rtl_reg_apply_world_flags(wiphy, request->initiator, reg); in _rtl_reg_notifier_apply()
360 request)) in _rtl_regd_init_wiphy()
389 struct regulatory_request *request)) in rtl_regd_init() argument
431 void rtl_reg_notifier(struct wiphy *wiphy, struct regulatory_request *request) in rtl_reg_notifier() argument
438 _rtl_reg_notifier_apply(wiphy, request, &rtlpriv->regd); in rtl_reg_notifier()
Dregd.h59 struct regulatory_request *request));
60 void rtl_reg_notifier(struct wiphy *wiphy, struct regulatory_request *request);
/linux-4.1.27/drivers/mmc/card/
Dqueue.h6 struct request;
35 struct request *req;
54 int (*issue_fn)(struct mmc_queue *, struct request *);
/linux-4.1.27/drivers/ide/
Dide-pm.c10 struct request *rq; in generic_ide_suspend()
46 struct request *rq; in generic_ide_resume()
81 void ide_complete_power_step(ide_drive_t *drive, struct request *rq) in ide_complete_power_step()
111 ide_startstop_t ide_start_power_step(ide_drive_t *drive, struct request *rq) in ide_start_power_step()
182 void ide_complete_pm_rq(ide_drive_t *drive, struct request *rq) in ide_complete_pm_rq()
209 void ide_check_pm_state(ide_drive_t *drive, struct request *rq) in ide_check_pm_state()
Dide-io.c57 int ide_end_rq(ide_drive_t *drive, struct request *rq, int error, in ide_end_rq()
78 struct request *rq = cmd->rq; in ide_complete_cmd()
118 struct request *rq = hwif->rq; in ide_complete_rq()
136 void ide_kill_rq(ide_drive_t *drive, struct request *rq) in ide_kill_rq()
227 struct request *rq = cmd->rq; in ide_map_sg()
254 struct request *rq) in execute_drive_cmd()
280 static ide_startstop_t ide_special_rq(ide_drive_t *drive, struct request *rq) in ide_special_rq()
306 static ide_startstop_t start_request (ide_drive_t *drive, struct request *rq) in start_request()
433 static void __ide_requeue_and_plug(struct request_queue *q, struct request *rq) in __ide_requeue_and_plug()
443 void ide_requeue_and_plug(ide_drive_t *drive, struct request *rq) in ide_requeue_and_plug()
[all …]
Dide-atapi.c92 struct request *rq; in ide_queue_pc_tail()
171 void ide_prep_sense(ide_drive_t *drive, struct request *rq) in ide_prep_sense()
174 struct request *sense_rq = &drive->sense_rq; in ide_prep_sense()
249 struct request *failed_rq = drive->hwif->rq; in ide_retry_pc()
250 struct request *sense_rq = &drive->sense_rq; in ide_retry_pc()
278 struct request *rq = drive->hwif->rq; in ide_cd_expiry()
308 int ide_cd_get_xferlen(struct request *rq) in ide_cd_get_xferlen()
343 int ide_check_ireason(ide_drive_t *drive, struct request *rq, int len, in ide_check_ireason()
394 struct request *rq = hwif->rq; in ide_pc_intr()
586 struct request *rq = hwif->rq; in ide_transfer_pc()
[all …]
/linux-4.1.27/drivers/staging/wlan-ng/
Dcfg80211.c337 struct cfg80211_scan_request *request) in prism2_scan() argument
352 if (!request) in prism2_scan()
355 dev = request->wdev->netdev; in prism2_scan()
358 if (priv->scan_request && priv->scan_request != request) in prism2_scan()
366 priv->scan_request = request; in prism2_scan()
375 if (request->n_ssids > 0) { in prism2_scan()
377 msg1.ssid.data.len = request->ssids->ssid_len; in prism2_scan()
379 request->ssids->ssid, request->ssids->ssid_len); in prism2_scan()
386 (i < request->n_channels) && i < ARRAY_SIZE(prism2_channels); in prism2_scan()
390 request->channels[i]->center_freq); in prism2_scan()
[all …]
/linux-4.1.27/drivers/media/usb/au0828/
Dau0828-core.c50 static int send_control_msg(struct au0828_dev *dev, u16 request, u32 value,
52 static int recv_control_msg(struct au0828_dev *dev, u16 request, u32 value,
75 static int send_control_msg(struct au0828_dev *dev, u16 request, u32 value, in send_control_msg() argument
85 request, in send_control_msg()
102 static int recv_control_msg(struct au0828_dev *dev, u16 request, u32 value, in recv_control_msg() argument
110 request, in recv_control_msg()
/linux-4.1.27/arch/s390/include/asm/
Dpci_clp.h167 struct clp_req_list_pci request; member
172 struct clp_req_set_pci request; member
177 struct clp_req_query_pci request; member
182 struct clp_req_query_pci_grp request; member
/linux-4.1.27/Documentation/power/
Dpm_qos_interface.txt30 changes to the request list or elements of the list. Typically the
31 aggregated target value is simply the max or min of the request values held
54 the request.
60 Returns if the request is still active, i.e. it has not been removed from a
72 Only processes can register a pm_qos request. To provide for automatic
80 request on the parameter.
87 To remove the user mode request for a target value simply close the device
96 Values are updated in response to changes of the request list.
99 simply the minimum of the request values held in the parameter list elements.
125 the request.
[all …]
/linux-4.1.27/include/trace/events/
Dblock.h66 TP_PROTO(struct request_queue *q, struct request *rq),
110 TP_PROTO(struct request_queue *q, struct request *rq),
126 TP_PROTO(struct request_queue *q, struct request *rq),
145 TP_PROTO(struct request_queue *q, struct request *rq,
178 TP_PROTO(struct request_queue *q, struct request *rq),
225 TP_PROTO(struct request_queue *q, struct request *rq),
240 TP_PROTO(struct request_queue *q, struct request *rq),
324 TP_PROTO(struct request_queue *q, struct request *rq, struct bio *bio),
361 TP_PROTO(struct request_queue *q, struct request *rq, struct bio *bio),
377 TP_PROTO(struct request_queue *q, struct request *rq, struct bio *bio),
[all …]
/linux-4.1.27/Documentation/security/
Dkeys-request-key.txt5 The key request service is part of the key retention service (refer to
54 own upcall mechanism rather than using /sbin/request-key.
67 forking and execution of /sbin/request-key.
74 A request proceeds in the following manner:
93 (4) request_key() then forks and executes /sbin/request-key with a new session
96 (5) /sbin/request-key assumes the authority associated with key U.
98 (6) /sbin/request-key execs an appropriate program to perform the actual
121 3) and another copy of /sbin/request-key spawned (as per step 4); but the
126 /sbin/request-key at the appropriate places because (a) execve will discard two
140 This is provided to prevent excessive repeated spawning of /sbin/request-key
[all …]
/linux-4.1.27/sound/usb/
Dhelper.c83 int snd_usb_ctl_msg(struct usb_device *dev, unsigned int pipe, __u8 request, in snd_usb_ctl_msg() argument
102 err = usb_control_msg(dev, pipe, request, requesttype, in snd_usb_ctl_msg()
110 snd_usb_ctl_msg_quirk(dev, pipe, request, requesttype, in snd_usb_ctl_msg()
/linux-4.1.27/arch/unicore32/kernel/
Dptrace.c87 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
93 switch (request) { in arch_ptrace()
108 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/drivers/net/wireless/orinoco/
Dcfg.c143 struct cfg80211_scan_request *request) in orinoco_scan() argument
148 if (!request) in orinoco_scan()
151 if (priv->scan_request && priv->scan_request != request) in orinoco_scan()
154 priv->scan_request = request; in orinoco_scan()
156 err = orinoco_hw_trigger_scan(priv, request->ssids); in orinoco_scan()
/linux-4.1.27/drivers/gpu/drm/msm/edp/
Dedp_aux.c45 bool native = msg->request & (DP_AUX_NATIVE_WRITE & DP_AUX_NATIVE_READ); in edp_msg_fifo_tx()
46 bool read = msg->request & (DP_AUX_I2C_READ & DP_AUX_NATIVE_READ); in edp_msg_fifo_tx()
122 bool native = msg->request & (DP_AUX_NATIVE_WRITE & DP_AUX_NATIVE_READ); in edp_aux_transfer()
123 bool read = msg->request & (DP_AUX_I2C_READ & DP_AUX_NATIVE_READ); in edp_aux_transfer()
136 __func__, msg->size, msg->request); in edp_aux_transfer()
/linux-4.1.27/arch/arc/kernel/
Dptrace.c141 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
146 pr_debug("REQ=%ld: ADDR =0x%lx, DATA=0x%lx)\n", request, addr, data); in arch_ptrace()
148 switch (request) { in arch_ptrace()
154 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/Documentation/networking/
Ddns_resolver.txt21 upcalled to userspace through /sbin/request-key.
24 request-key. It is under development and does not yet provide the full feature
49 To set up this facility, the /etc/request-key.conf file must be altered so that
50 /sbin/request-key can appropriately direct the upcalls. For example, to handle
130 directed by means of configuration lines in /etc/request-key.conf that tell
131 /sbin/request-key what program to run to instantiate the key.
146 See <file:Documentation/security/keys-request-key.txt> for further
147 information about request-key function.
/linux-4.1.27/fs/ncpfs/
Dioctl.c261 struct ncp_ioctl_request request; in __ncp_ioctl() local
275 request.function = request32.function; in __ncp_ioctl()
276 request.size = request32.size; in __ncp_ioctl()
277 request.data = compat_ptr(request32.data); in __ncp_ioctl()
280 if (copy_from_user(&request, argp, sizeof(request))) in __ncp_ioctl()
283 if ((request.function > 255) in __ncp_ioctl()
284 || (request.size > in __ncp_ioctl()
291 if (copy_from_user(bouncebuffer, request.data, request.size)) { in __ncp_ioctl()
301 server->current_size = request.size; in __ncp_ioctl()
302 memcpy(server->packet, bouncebuffer, request.size); in __ncp_ioctl()
[all …]
/linux-4.1.27/arch/cris/arch-v10/kernel/
Dptrace.c78 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
85 switch (request) { in arch_ptrace()
174 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/drivers/soc/tegra/
Dpmc.c452 static int tegra_io_rail_prepare(int id, unsigned long *request, in tegra_io_rail_prepare() argument
469 *request = IO_DPD_REQ; in tegra_io_rail_prepare()
472 *request = IO_DPD2_REQ; in tegra_io_rail_prepare()
517 unsigned long request, status, value; in tegra_io_rail_power_on() local
521 err = tegra_io_rail_prepare(id, &request, &status, &bit); in tegra_io_rail_power_on()
527 value = tegra_pmc_readl(request); in tegra_io_rail_power_on()
531 tegra_pmc_writel(value, request); in tegra_io_rail_power_on()
545 unsigned long request, status, value; in tegra_io_rail_power_off() local
549 err = tegra_io_rail_prepare(id, &request, &status, &bit); in tegra_io_rail_power_off()
555 value = tegra_pmc_readl(request); in tegra_io_rail_power_off()
[all …]
/linux-4.1.27/arch/um/kernel/skas/
Dprocess.c40 init_task.thread.request.u.thread.proc = start_kernel_proc; in start_uml()
41 init_task.thread.request.u.thread.arg = NULL; in start_uml()
/linux-4.1.27/Documentation/s390/
Dcds.txt64 initiate an I/O request.
70 terminate the current I/O request processed on the device.
76 interrupt handler according to the rules (flags) defined during I/O request
143 armed) interrupt request lines (IRQs) to drive its devices with. Where
150 and initiates an I/O request on behalf of the device driver. The
156 For long running I/O request to be canceled, the CDS layer provides the
180 The ccw_device_start() routines is the I/O request front-end processor. All
187 with the associated I/O request when calling ccw_device_start().
219 particular I/O request.
221 request. A value of 0 will make cio use the opm.
[all …]
/linux-4.1.27/drivers/staging/lustre/lustre/include/
Dobd_class.h1483 struct ptlrpc_request **request) in md_getattr() argument
1489 rc = MDP(exp->exp_obd, getattr)(exp, op_data, request); in md_getattr()
1518 struct ptlrpc_request **request) in md_close() argument
1524 rc = MDP(exp->exp_obd, close)(exp, op_data, mod, request); in md_close()
1531 struct ptlrpc_request **request) in md_create() argument
1538 uid, gid, cap_effective, rdev, request); in md_create()
1574 struct ptlrpc_request **request) in md_getattr_name() argument
1580 rc = MDP(exp->exp_obd, getattr_name)(exp, op_data, request); in md_getattr_name()
1602 struct ptlrpc_request **request) in md_link() argument
1608 rc = MDP(exp->exp_obd, link)(exp, op_data, request); in md_link()
[all …]
/linux-4.1.27/arch/openrisc/kernel/
Dptrace.c159 long arch_ptrace(struct task_struct *child, long request, unsigned long addr, in arch_ptrace() argument
164 switch (request) { in arch_ptrace()
166 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/drivers/scsi/libsas/
Dsas_internal.h103 extern int sas_smp_host_handler(struct Scsi_Host *shost, struct request *req,
104 struct request *rsp);
107 struct request *req, in sas_smp_host_handler()
108 struct request *rsp) in sas_smp_host_handler()
/linux-4.1.27/sound/soc/intel/haswell/
Dsst-haswell-ipc.c220 struct sst_hsw_ipc_stream_alloc_req request; member
1050 stream->request.format.bitdepth = bits; in sst_hsw_stream_set_bits()
1062 stream->request.format.ch_num = channels; in sst_hsw_stream_set_channels()
1074 stream->request.format.frequency = rate; in sst_hsw_stream_set_rate()
1087 stream->request.format.map = map; in sst_hsw_stream_set_map_config()
1088 stream->request.format.config = config; in sst_hsw_stream_set_map_config()
1100 stream->request.format.style = style; in sst_hsw_stream_set_style()
1112 stream->request.format.valid_bit = bits; in sst_hsw_stream_set_valid()
1127 stream->request.path_id = path_id; in sst_hsw_stream_format()
1128 stream->request.stream_type = stream_type; in sst_hsw_stream_format()
[all …]
/linux-4.1.27/drivers/net/wireless/
Dzd1201.c517 unsigned char *request; in zd1201_setconfig() local
528 request = kmalloc(16, gfp_mask); in zd1201_setconfig()
529 if (!request) in zd1201_setconfig()
533 kfree(request); in zd1201_setconfig()
536 memset(request, 0, 16); in zd1201_setconfig()
538 request[0] = ZD1201_USB_RESREQ; in zd1201_setconfig()
539 request[1] = seq; in zd1201_setconfig()
540 request[2] = 0; in zd1201_setconfig()
541 request[3] = 0; in zd1201_setconfig()
542 if (request[1] == 0) { in zd1201_setconfig()
[all …]
/linux-4.1.27/drivers/base/power/
Druntime.c109 dev->power.request = RPM_REQ_NONE; in pm_runtime_cancel_pending()
250 && dev->power.request == RPM_REQ_RESUME)) in rpm_check_suspend_allowed()
317 dev->power.request > RPM_REQ_IDLE) in rpm_idle()
327 dev->power.request = RPM_REQ_NONE; in rpm_idle()
334 dev->power.request = RPM_REQ_IDLE; in rpm_idle()
443 dev->power.request = RPM_REQ_NONE; in rpm_suspend()
504 dev->power.request = (rpmflags & RPM_AUTO) ? in rpm_suspend()
620 dev->power.request = RPM_REQ_NONE; in rpm_resume()
688 dev->power.request = RPM_REQ_RESUME; in rpm_resume()
783 req = dev->power.request; in pm_runtime_work()
[all …]
/linux-4.1.27/arch/powerpc/kernel/
Dptrace32.c47 long compat_arch_ptrace(struct task_struct *child, compat_long_t request, in compat_arch_ptrace() argument
54 switch (request) { in compat_arch_ptrace()
308 ret = arch_ptrace(child, request, addr, data); in compat_arch_ptrace()
312 ret = compat_ptrace_request(child, request, addr, data); in compat_arch_ptrace()
/linux-4.1.27/drivers/usb/serial/
Dcp210x.c319 static int cp210x_get_config(struct usb_serial_port *port, u8 request, in cp210x_get_config() argument
336 request, REQTYPE_INTERFACE_TO_HOST, 0x0000, in cp210x_get_config()
348 __func__, request, size, result); in cp210x_get_config()
364 static int cp210x_set_config(struct usb_serial_port *port, u8 request, in cp210x_set_config() argument
386 request, REQTYPE_HOST_TO_INTERFACE, 0x0000, in cp210x_set_config()
392 request, REQTYPE_HOST_TO_INTERFACE, data[0], in cp210x_set_config()
401 __func__, request, size, result); in cp210x_set_config()
417 u8 request, unsigned int data) in cp210x_set_config_single() argument
419 return cp210x_set_config(port, request, &data, 2); in cp210x_set_config_single()
/linux-4.1.27/arch/blackfin/kernel/
Dptrace.c247 long arch_ptrace(struct task_struct *child, long request, in arch_ptrace() argument
254 switch (request) { in arch_ptrace()
352 request = PTRACE_GETFDPIC; in arch_ptrace()
356 request = PTRACE_GETFDPIC; in arch_ptrace()
387 ret = ptrace_request(child, request, addr, data); in arch_ptrace()
/linux-4.1.27/drivers/gpu/drm/tegra/
Ddpaux.c122 switch (msg->request & ~DP_AUX_I2C_MOT) { in tegra_dpaux_transfer()
136 switch (msg->request & ~DP_AUX_I2C_MOT) { in tegra_dpaux_transfer()
138 if (msg->request & DP_AUX_I2C_MOT) in tegra_dpaux_transfer()
146 if (msg->request & DP_AUX_I2C_MOT) in tegra_dpaux_transfer()
154 if (msg->request & DP_AUX_I2C_MOT) in tegra_dpaux_transfer()
176 if ((msg->request & DP_AUX_I2C_READ) == 0) { in tegra_dpaux_transfer()
225 if (msg->request & DP_AUX_I2C_READ) { in tegra_dpaux_transfer()
/linux-4.1.27/Documentation/device-mapper/
Dlog-writes.txt7 There is a log_write_entry written for every WRITE request and the target is
17 next REQ_FLUSH request. This is to make it easier for userspace to replay the
22 Once we see a REQ_FLUSH request we splice this list onto the request and once
23 the FLUSH request completes we log all of the WRITEs and then the FLUSH. Only
43 request. Consider the following example:
138 And that will replay the log until it sees a FUA request, run the fsck command
/linux-4.1.27/Documentation/devicetree/bindings/sound/
Dsirf-audio-port.txt6 - dmas: List of DMA controller phandle and DMA request line ordered pairs.
7 - dma-names: Identifier string for each DMA request line in the dmas property.

123456