Home
last modified time | relevance | path

Searched refs:sg_page (Results 1 – 140 of 140) sorted by relevance

/linux-4.1.27/include/crypto/
Dscatterwalk.h54 return !(((sg_page(walk_in->sg) - sg_page(walk_out->sg)) << PAGE_SHIFT) + in scatterwalk_samebuf()
86 return sg_page(walk->sg) + (walk->offset >> PAGE_SHIFT); in scatterwalk_page()
/linux-4.1.27/include/linux/
Dscatterlist.h95 static inline struct page *sg_page(struct scatterlist *sg) in sg_page() function
205 return page_to_phys(sg_page(sg)) + sg->offset; in sg_phys()
220 return page_address(sg_page(sg)) + sg->offset; in sg_virt()
288 return nth_page(sg_page(piter->sg), piter->sg_pgoffset); in sg_page_iter_page()
/linux-4.1.27/net/rds/
Dmessage.c65 rdsdebug("putting data page %p\n", (void *)sg_page(&rm->data.op_sg[i])); in rds_message_purge()
67 __free_page(sg_page(&rm->data.op_sg[i])); in rds_message_purge()
283 if (!sg_page(sg)) { in rds_message_copy_from_user()
296 nbytes = copy_page_from_iter(sg_page(sg), sg->offset + sg_off, in rds_message_copy_from_user()
333 ret = copy_page_to_iter(sg_page(sg), sg->offset + vec_off, in rds_message_inc_copy_to_user()
Dpage.c133 get_page(sg_page(scat)); in rds_page_remainder_alloc()
176 ret ? NULL : sg_page(scat), ret ? 0 : scat->offset, in rds_page_remainder_alloc()
Dtcp_send.c121 sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit()
125 rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit()
Dib_recv.c193 rdsdebug("frag %p page %p\n", frag, sg_page(&frag->f_sg)); in rds_ib_frag_free()
383 recv->r_ibinc, sg_page(&recv->r_frag->f_sg), in rds_ib_recv_refill()
501 ret = copy_page_to_iter(sg_page(&frag->f_sg), in rds_ib_inc_copy_to_user()
755 addr = kmap_atomic(sg_page(&frag->f_sg)); in rds_ib_cong_recv()
Drdma.c269 put_page(sg_page(&sg[i])); in __rds_rdma_map()
448 struct page *page = sg_page(&ro->op_sg[i]); in rds_rdma_free_op()
467 struct page *page = sg_page(ao->op_sg); in rds_atomic_free_op()
Dib_rdma.c484 struct page *page = sg_page(&ibmr->sg[i]); in __rds_ib_teardown_mr()
/linux-4.1.27/crypto/
Dalgif_aead.c86 if (!sg_page(sg + i)) in aead_put_sgl()
89 put_page(sg_page(sg + i)); in aead_put_sgl()
222 err = memcpy_from_msg(page_address(sg_page(sg)) + in aead_sendmsg()
261 if (!sg_page(sg)) in aead_sendmsg()
264 err = memcpy_from_msg(page_address(sg_page(sg)), in aead_sendmsg()
267 __free_page(sg_page(sg)); in aead_sendmsg()
466 sg_set_page(assoc + i, sg_page(sg), in aead_recvmsg()
482 sg_set_page(assoc + i, sg_page(sg), in aead_recvmsg()
Dalgif_skcipher.c98 put_page(sg_page(sg)); in skcipher_free_async_sgls()
177 if (!sg_page(sg + i)) in skcipher_pull_sgl()
189 put_page(sg_page(sg + i)); in skcipher_pull_sgl()
365 err = memcpy_from_msg(page_address(sg_page(sg)) + in skcipher_sendmsg()
403 if (!sg_page(sg + i)) in skcipher_sendmsg()
406 err = memcpy_from_msg(page_address(sg_page(sg + i)), in skcipher_sendmsg()
409 __free_page(sg_page(sg + i)); in skcipher_sendmsg()
576 sg_set_page(&tmp[x], sg_page(&sreq->tsg[x]), in skcipher_recvmsg_async()
586 sg_set_page(sreq->tsg + txbufs++, sg_page(sg), sg->length, in skcipher_recvmsg_async()
Dauthencesn.c404 dstp = sg_page(dst); in crypto_authenc_esn_genicv()
427 sg_set_page(hsg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_esn_genicv()
428 sg_set_page(hsg + 1, sg_page(assoc2), assoc2->length, assoc2->offset); in crypto_authenc_esn_genicv()
431 sg_set_page(tsg, sg_page(assoc1), assoc1->length, assoc1->offset); in crypto_authenc_esn_genicv()
572 srcp = sg_page(src); in crypto_authenc_esn_iverify()
595 sg_set_page(hsg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_esn_iverify()
596 sg_set_page(hsg + 1, sg_page(assoc2), assoc2->length, assoc2->offset); in crypto_authenc_esn_iverify()
599 sg_set_page(tsg, sg_page(assoc1), assoc1->length, assoc1->offset); in crypto_authenc_esn_iverify()
Deseqiv.c94 srcp = sg_page(osrc); in eseqiv_givencrypt()
95 dstp = sg_page(odst); in eseqiv_givencrypt()
Dscatterwalk.c56 page = sg_page(walk->sg) + ((walk->offset - 1) >> PAGE_SHIFT); in scatterwalk_pagedone()
Dauthenc.c340 dstp = sg_page(dst); in crypto_authenc_genicv()
354 sg_set_page(asg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_genicv()
495 srcp = sg_page(src); in crypto_authenc_iverify()
509 sg_set_page(asg, sg_page(assoc), assoc->length, assoc->offset); in crypto_authenc_iverify()
Dshash.c285 data = kmap_atomic(sg_page(sg)); in shash_ahash_digest()
425 data = kmap_atomic(sg_page(sg)); in shash_compat_digest()
Dahash.c73 walk->pg = sg_page(walk->sg) + (walk->offset >> PAGE_SHIFT); in hash_walk_new_entry()
Dgcm.c1150 srcp = sg_page(src); in crypto_rfc4543_crypt()
1160 sg_set_page(assoc, sg_page(req->assoc), req->assoc->length, in crypto_rfc4543_crypt()
/linux-4.1.27/arch/mips/mm/
Ddma-default.c272 __dma_sync(sg_page(sg), sg->offset, sg->length, in mips_dma_map_sg()
277 sg->dma_address = plat_map_dma_mem_page(dev, sg_page(sg)) + in mips_dma_map_sg()
303 __dma_sync(sg_page(sg), sg->offset, sg->length, in mips_dma_unmap_sg()
333 __dma_sync(sg_page(sg), sg->offset, sg->length, in mips_dma_sync_sg_for_cpu()
345 __dma_sync(sg_page(sg), sg->offset, sg->length, in mips_dma_sync_sg_for_device()
/linux-4.1.27/drivers/target/
Dtarget_core_sbc.c364 addr = kmap_atomic(sg_page(sg)); in xdreadwrite_callback()
479 addr = (unsigned char *)kmap_atomic(sg_page(sg)); in compare_and_write_callback()
1196 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate()
1197 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_generate()
1204 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_generate()
1292 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_copy_prot()
1296 addr = kmap_atomic(sg_page(sg)) + sg->offset + offset; in sbc_dif_copy_prot()
1331 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_verify_write()
1332 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_verify_write()
1339 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_verify_write()
[all …]
Dtarget_core_iblock.c497 while (bio_add_page(bio, sg_page(sg), sg->length, sg->offset) in iblock_execute_write_same()
660 rc = bio_integrity_add_page(bio, sg_page(sg), sg->length, in iblock_alloc_bip()
668 sg_page(sg), sg->length, sg->offset); in iblock_alloc_bip()
756 while (bio_add_page(bio, sg_page(sg), sg->length, sg->offset) in iblock_execute_rw()
Dtarget_core_transport.c1475 buf = kmap(sg_page(sgl)) + sgl->offset; in target_submit_cmd_map_sgls()
1479 kunmap(sg_page(sgl)); in target_submit_cmd_map_sgls()
2162 __free_page(sg_page(sg)); in transport_free_sgl()
2246 return kmap(sg_page(sg)) + sg->offset; in transport_kmap_data_sg()
2255 pages[i] = sg_page(sg); in transport_kmap_data_sg()
2272 kunmap(sg_page(cmd->t_data_sg)); in transport_kunmap_data_sg()
2315 __free_page(sg_page(&sg[i])); in target_alloc_sgl()
Dtarget_core_file.c339 bvec[i].bv_page = sg_page(sg); in fd_do_rw()
465 bvec[i].bv_page = sg_page(&cmd->t_data_sg[0]); in fd_execute_write_same()
Dtarget_core_user.c366 void *from = kmap_atomic(sg_page(sg)) + sg->offset; in tcmu_queue_cmd_ring()
496 to = kmap_atomic(sg_page(sg)) + sg->offset; in tcmu_handle_completion()
Dtarget_core_rd.c94 pg = sg_page(&sg[j]); in rd_release_sgl_table()
Dtarget_core_pscsi.c891 page = sg_page(sg); in pscsi_map_sg()
/linux-4.1.27/drivers/target/tcm_fc/
Dtfc_io.c103 page = sg_page(sg); in ft_queue_data_in()
121 page = sg_page(sg); in ft_queue_data_in()
300 page = sg_page(sg); in ft_recv_write_data()
308 page = sg_page(sg); in ft_recv_write_data()
Dtfc_cmd.c71 sg_page(sg), sg->length, sg->offset); in _ft_dump_cmd()
/linux-4.1.27/tools/virtio/linux/
Dscatterlist.h66 static inline struct page *sg_page(struct scatterlist *sg) in sg_page() function
174 return page_to_phys(sg_page(sg)) + sg->offset; in sg_phys()
/linux-4.1.27/drivers/staging/android/ion/
Dion_carveout_heap.c64 struct page *page = sg_page(table->sgl); in ion_carveout_heap_phys()
113 struct page *page = sg_page(table->sgl); in ion_carveout_heap_free()
Dion_chunk_heap.c84 gen_pool_free(chunk_heap->pool, page_to_phys(sg_page(sg)), in ion_chunk_heap_allocate()
112 gen_pool_free(chunk_heap->pool, page_to_phys(sg_page(sg)), in ion_chunk_heap_free()
Dion_system_heap.c194 free_buffer_page(sys_heap, buffer, sg_page(sg)); in ion_system_heap_free()
364 struct page *page = sg_page(table->sgl); in ion_system_contig_heap_free()
379 struct page *page = sg_page(table->sgl); in ion_system_contig_heap_phys()
Dion_heap.c50 struct page *page = sg_page(sg); in ion_heap_map_kernel()
82 struct page *page = sg_page(sg); in ion_heap_map_user()
Dion.c233 struct page *page = sg_page(sg); in ion_buffer_create()
/linux-4.1.27/drivers/crypto/qce/
Ddma.c115 if (!sg_page(sg)) in qce_sgtable_add()
124 sg_set_page(sg, sg_page(new_sgl), new_sgl->length, in qce_sgtable_add()
/linux-4.1.27/arch/metag/include/asm/
Ddma-mapping.h58 BUG_ON(!sg_page(sg)); in dma_map_sg()
97 BUG_ON(!sg_page(sg)); in dma_unmap_sg()
/linux-4.1.27/drivers/scsi/
Dstorvsc_drv.c565 page_buf = sg_page((&sgl[i])); in destroy_bounce_buffer()
654 kmap_atomic(sg_page(cur_dest_sgl)) + in copy_from_bounce_buffer()
661 sg_page(cur_src_sgl)); in copy_from_bounce_buffer()
702 sg_page(cur_src_sgl)); in copy_from_bounce_buffer()
742 kmap_atomic(sg_page(cur_src_sgl)) + in copy_to_bounce_buffer()
749 kmap_atomic(sg_page(cur_dest_sgl)); in copy_to_bounce_buffer()
776 sg_page(cur_dest_sgl)); in copy_to_bounce_buffer()
1667 page_to_pfn(sg_page((cur_sgl))); in storvsc_queuecommand()
Dlibiscsi_tcp.c133 if (page_count(sg_page(sg)) >= 1 && !recv) in iscsi_tcp_segment_map()
138 segment->sg_mapped = kmap_atomic(sg_page(sg)); in iscsi_tcp_segment_map()
142 segment->sg_mapped = kmap(sg_page(sg)); in iscsi_tcp_segment_map()
154 kunmap(sg_page(segment->sg)); in iscsi_tcp_segment_unmap()
211 sg_set_page(&sg, sg_page(segment->sg), copied, in iscsi_tcp_segment_done()
Dwd7000.c1121 any2scsi(sgb[i].ptr, isa_page_to_bus(sg_page(sg)) + sg->offset); in wd7000_queuecommand_lck()
1128 any2scsi(scb->dataptr, isa_page_to_bus(sg_page(sg)) + sg->offset); in wd7000_queuecommand_lck()
Dosst.c384 pages[i] = sg_page(sg); in osst_execute()
624 memset(page_address(sg_page(&STp->buffer->sg[i])), in osst_verify_frame()
4520 …STp->buffer->aux = (os_aux_t *) (page_address(sg_page(&STp->buffer->sg[i])) + OS_DATA_SIZE - b_siz… in __os_scsi_tape_open()
5364 if (sg_page(&STbuffer->sg[0]) == NULL) { in enlarge_buffer()
5414 __free_pages(sg_page(&STbuffer->sg[i]), order); in normalize_buffer()
5442 res = copy_from_user(page_address(sg_page(&st_bp->sg[i])) + offset, ubp, cnt); in append_to_buffer()
5475 res = copy_to_user(ubp, page_address(sg_page(&st_bp->sg[i])) + offset, cnt); in from_buffer()
5508 memset(page_address(sg_page(&st_bp->sg[i])) + offset, 0, cnt); in osst_zero_buffer_tail()
5528 memcpy(page_address(sg_page(&st_bp->sg[i])), ptr, cnt); in osst_copy_to_buffer()
5549 memcpy(ptr, page_address(sg_page(&st_bp->sg[i])), cnt); in osst_copy_from_buffer()
Discsi_tcp.c288 r = tcp_sw_conn->sendpage(sk, sg_page(sg), offset, in iscsi_sw_tcp_xmit_segment()
Dultrastor.c691 mscp->sglist[i].address = isa_page_to_bus(sg_page(sg)) + sg->offset; in build_sg_list()
Daha1542.c457 any2scsi(cptr[i].dataptr, isa_page_to_bus(sg_page(sg)) in aha1542_queuecommand()
Dpmcraid.c3251 __free_pages(sg_page(&(sglist->scatterlist[i])), in pmcraid_free_sglist()
3305 __free_pages(sg_page(&scatterlist[j]), order); in pmcraid_alloc_sglist()
3349 struct page *page = sg_page(&scatterlist[i]); in pmcraid_copy_sglist()
3370 struct page *page = sg_page(&scatterlist[i]); in pmcraid_copy_sglist()
Dxen-scsifront.c441 page = sg_page(sg); in map_data_for_request()
Dmegaraid.c666 buf = kmap_atomic(sg_page(sg)) + sg->offset; in mega_build_cmd()
1553 if( sg_page(sgl) ) { in mega_cmd_done()
Dgdth.c2298 if (!sg_page(sl)) { in gdth_copy_internal_data()
2304 address = kmap_atomic(sg_page(sl)) + sl->offset; in gdth_copy_internal_data()
2306 flush_dcache_page(sg_page(sl)); in gdth_copy_internal_data()
Datari_NCR5380.c430 page_address(sg_page(&cmd->SCp.buffer[1])), endaddr); in merge_contiguous_buffers()
Dscsi_lib.c3116 page = nth_page(sg_page(sg), (*offset >> PAGE_SHIFT)); in scsi_kmap_atomic_sg()
Dipr.c3783 __free_pages(sg_page(&scatterlist[j]), order); in ipr_alloc_ucode_buffer()
3809 __free_pages(sg_page(&sglist->scatterlist[i]), sglist->order); in ipr_free_ucode_buffer()
3839 struct page *page = sg_page(&scatterlist[i]); in ipr_copy_ucode_buffer()
3854 struct page *page = sg_page(&scatterlist[i]); in ipr_copy_ucode_buffer()
Dips.c1513 buffer = kmap_atomic(sg_page(sg)) + sg->offset; in ips_is_passthru()
/linux-4.1.27/arch/sparc/mm/
Diommu.c246 sg->dma_address = iommu_get_one(dev, sg_page(sg), n) + sg->offset; in iommu_get_scsi_sgl_gflush()
267 if ((page = (unsigned long) page_address(sg_page(sg))) != 0) { in iommu_get_scsi_sgl_pflush()
277 sg->dma_address = iommu_get_one(dev, sg_page(sg), n) + sg->offset; in iommu_get_scsi_sgl_pflush()
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_memfree.c73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages()
83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent()
309 page = sg_page(&chunk->mem[i]); in mthca_table_find()
493 put_page(sg_page(&db_tab->page[i].mem)); in mthca_map_user_db()
559 put_page(sg_page(&db_tab->page[i].mem)); in mthca_cleanup_user_db_tab()
/linux-4.1.27/arch/powerpc/kernel/
Ddma.c138 __dma_sync_page(sg_page(sg), sg->offset, sg->length, direction); in dma_direct_map_sg()
203 __dma_sync_page(sg_page(sg), sg->offset, sg->length, direction); in dma_direct_sync_sg()
/linux-4.1.27/arch/frv/mb93090-mb00/
Dpci-dma.c65 vaddr = kmap_atomic_primary(sg_page(&sg[i])); in dma_map_sg()
/linux-4.1.27/arch/x86/kernel/
Dpci-nommu.c66 BUG_ON(!sg_page(s)); in nommu_map_sg()
Dpci-calgary_64.c377 BUG_ON(!sg_page(s)); in calgary_map_sg()
/linux-4.1.27/arch/sh/kernel/
Ddma-nommu.c36 BUG_ON(!sg_page(s)); in nommu_map_sg()
/linux-4.1.27/drivers/infiniband/ulp/iser/
Diser_memory.c49 __free_page(sg_page(sg)); in iser_free_bounce_sg()
93 __free_page(sg_page(&sg[i - 1])); in iser_alloc_bounce_sg()
111 oaddr = kmap_atomic(sg_page(osg)) + osg->offset; in iser_copy_bounce()
116 baddr = kmap_atomic(sg_page(bsg)) + bsg->offset; in iser_copy_bounce()
348 sg_page(sg), sg->offset, in iser_data_buf_dump()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Dicm.c63 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages()
74 lowmem_page_address(sg_page(&chunk->mem[i])), in mlx4_free_icm_coherent()
341 page = sg_page(&chunk->mem[i]); in mlx4_table_find()
/linux-4.1.27/drivers/crypto/ccp/
Dccp-crypto-main.c306 if (!sg_page(sg)) in ccp_crypto_sg_table_add()
311 sg_set_page(sg, sg_page(sg_add), sg_add->length, in ccp_crypto_sg_table_add()
/linux-4.1.27/net/sunrpc/auth_gss/
Dgss_krb5_crypto.c420 in_page = sg_page(sg); in encryptor()
424 sg_set_page(&desc->outfrags[desc->fragno], sg_page(sg), sg->length, in encryptor()
448 sg_set_page(&desc->outfrags[0], sg_page(sg), fraglen, in encryptor()
505 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor()
526 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, in decryptor()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_dma.c105 addr = (u64) page_address(sg_page(sg)); in qib_map_sg()
Dqib_mr.c272 vaddr = page_address(sg_page(sg)); in qib_reg_user_mr()
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_dma.c112 addr = (u64) page_address(sg_page(sg)); in ipath_map_sg()
Dipath_mr.c227 vaddr = page_address(sg_page(sg)); in ipath_reg_user_mr()
/linux-4.1.27/drivers/gpu/drm/i915/
Di915_gem_render_state.c88 page = sg_page(so->obj->pages->sgl); in render_state_setup()
Di915_gem_dmabuf.c67 sg_set_page(dst, sg_page(src), src->length, 0); in i915_gem_map_dma_buf()
Dintel_ringbuffer.c659 kunmap(sg_page(ring->scratch.obj->pages->sgl)); in intel_fini_pipe_control()
690 ring->scratch.cpu_page = kmap(sg_page(ring->scratch.obj->pages->sgl)); in intel_init_pipe_control()
1825 kunmap(sg_page(obj->pages->sgl)); in cleanup_status_page()
1873 ring->status_page.page_addr = kmap(sg_page(obj->pages->sgl)); in init_status_page()
Dintel_lrc.c1451 kunmap(sg_page(ring->status_page.obj->pages->sgl)); in intel_logical_ring_cleanup()
1906 kmap(sg_page(default_ctx_obj->pages->sgl)); in lrc_setup_hardware_status_page()
/linux-4.1.27/arch/alpha/kernel/
Dpci-noop.c150 BUG_ON(!sg_page(sg)); in alpha_noop_map_sg()
/linux-4.1.27/drivers/crypto/
Dhifn_795x.c1382 spage = sg_page(src); in hifn_setup_dma()
1396 BUG_ON(!sg_page(t)); in hifn_setup_dma()
1397 dpage = sg_page(t); in hifn_setup_dma()
1401 BUG_ON(!sg_page(dst)); in hifn_setup_dma()
1402 dpage = sg_page(dst); in hifn_setup_dma()
1452 __free_page(sg_page(s)); in hifn_cipher_walk_exit()
1734 daddr = kmap_atomic(sg_page(dst)); in ablkcipher_get()
1787 __func__, sg_page(t), t->length, in hifn_process_ready()
1788 sg_page(dst), dst->length, nbytes); in hifn_process_ready()
1796 saddr = kmap_atomic(sg_page(t)); in hifn_process_ready()
Dixp4xx_crypto.c794 ptr = page_address(sg_page(sg)) + sg->offset; in chainup_buffers()
Domap-sham.c642 vaddr = kmap_atomic(sg_page(ctx->sg)); in omap_sham_append_sg()
/linux-4.1.27/arch/mn10300/include/asm/
Ddma-mapping.h61 BUG_ON(!sg_page(sg)); in dma_map_sg()
/linux-4.1.27/lib/
Ddma-debug.c1371 entry->pfn = page_to_pfn(sg_page(s)); in debug_dma_map_sg()
1379 if (!PageHighMem(sg_page(s))) { in debug_dma_map_sg()
1422 .pfn = page_to_pfn(sg_page(s)), in debug_dma_unmap_sg()
1583 .pfn = page_to_pfn(sg_page(s)), in debug_dma_sync_sg_for_cpu()
1616 .pfn = page_to_pfn(sg_page(s)), in debug_dma_sync_sg_for_device()
/linux-4.1.27/arch/openrisc/kernel/
Ddma.c187 s->dma_address = or1k_map_page(dev, sg_page(s), s->offset, in or1k_map_sg()
/linux-4.1.27/arch/microblaze/kernel/
Ddma.c64 __dma_sync(page_to_phys(sg_page(sg)) + sg->offset, in dma_direct_map_sg()
/linux-4.1.27/drivers/crypto/amcc/
Dcrypto4xx_core.c560 addr = dma_map_page(dev->core_dev->device, sg_page(sg), in crypto4xx_copy_pkt_to_dst()
643 addr = dma_map_page(dev->core_dev->device, sg_page(dst), in crypto4xx_ablkcipher_done()
899 addr = dma_map_page(dev->core_dev->device, sg_page(sg), in crypto4xx_build_pd()
913 pd->src = (u32)dma_map_page(dev->core_dev->device, sg_page(src), in crypto4xx_build_pd()
940 sg_page(dst), dst->offset, in crypto4xx_build_pd()
/linux-4.1.27/arch/xtensa/include/asm/
Ddma-mapping.h63 BUG_ON(!sg_page(sg)); in dma_map_sg()
/linux-4.1.27/arch/avr32/include/asm/
Ddma-mapping.h220 sg[i].dma_address = page_to_bus(sg_page(&sg[i])) + sg[i].offset; in dma_map_sg()
/linux-4.1.27/drivers/mmc/host/
Dtmio_mmc.h122 return kmap_atomic(sg_page(sg)) + sg->offset; in tmio_mmc_kmap_atomic()
Dmmc_spi.c926 dma_addr = dma_map_page(dma_dev, sg_page(sg), 0, in mmc_spi_data_do()
935 kmap_addr = kmap(sg_page(sg)); in mmc_spi_data_do()
968 flush_kernel_dcache_page(sg_page(sg)); in mmc_spi_data_do()
969 kunmap(sg_page(sg)); in mmc_spi_data_do()
Dtifm_sd.c195 pg = nth_page(sg_page(&sg[host->sg_pos]), off >> PAGE_SHIFT); in tifm_sd_transfer_data()
244 pg = nth_page(sg_page(&sg[host->sg_pos]), off >> PAGE_SHIFT); in tifm_sd_bounce_block()
251 tifm_sd_copy_page(sg_page(&host->bounce_buf), in tifm_sd_bounce_block()
255 tifm_sd_copy_page(pg, p_off, sg_page(&host->bounce_buf), in tifm_sd_bounce_block()
Dsdricoh_cs.c324 page = sg_page(data->sg); in sdricoh_request()
Dusdhi6rol0.c364 host->pg.page = sg_page(sg); in usdhi6_sg_map()
502 host->pg.page = nth_page(sg_page(host->sg), host->page_idx); in usdhi6_sg_advance()
Datmel-mci.c1873 flush_dcache_page(sg_page(sg)); in atmci_read_data_pio()
1887 flush_dcache_page(sg_page(sg)); in atmci_read_data_pio()
Dsdhci.c449 return kmap_atomic(sg_page(sg)) + sg->offset; in sdhci_kmap_atomic()
/linux-4.1.27/arch/arc/include/asm/
Ddma-mapping.h129 s->dma_address = dma_map_page(dev, sg_page(s), s->offset, in dma_map_sg()
/linux-4.1.27/arch/arm/mm/
Ddma-mapping.c912 s->dma_address = ops->map_page(dev, sg_page(s), s->offset, in arm_dma_map_sg()
1512 phys_addr_t phys = page_to_phys(sg_page(s)); in __map_sg_chunk()
1517 __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); in __map_sg_chunk()
1631 __dma_page_dev_to_cpu(sg_page(s), s->offset, in __iommu_unmap_sg()
1682 __dma_page_dev_to_cpu(sg_page(s), s->offset, s->length, dir); in arm_iommu_sync_sg_for_cpu()
1700 __dma_page_cpu_to_dev(sg_page(s), s->offset, s->length, dir); in arm_iommu_sync_sg_for_device()
/linux-4.1.27/drivers/gpu/drm/armada/
Darmada_gem.c482 page_cache_release(sg_page(sg)); in armada_gem_prime_map_dma_buf()
503 page_cache_release(sg_page(sg)); in armada_gem_prime_unmap_dma_buf()
/linux-4.1.27/drivers/gpu/drm/exynos/
Dexynos_drm_dmabuf.c103 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in exynos_gem_map_dma_buf()
/linux-4.1.27/drivers/gpu/drm/udl/
Dudl_dmabuf.c120 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in udl_map_dma_buf()
/linux-4.1.27/drivers/scsi/libfc/
Dfc_libfc.c144 page_addr = kmap_atomic(sg_page(sg) + (off >> PAGE_SHIFT)); in fc_copy_buffer_to_sglist()
Dfc_fcp.c647 page = sg_page(sg) + (off >> PAGE_SHIFT); in fc_fcp_send_data()
/linux-4.1.27/drivers/infiniband/core/
Dumem.c60 page = sg_page(sg); in __ib_umem_release()
/linux-4.1.27/drivers/media/v4l2-core/
Dvideobuf2-dma-contig.c60 struct page *page = sg_page(s); in vb2_dc_sgt_foreach_page()
264 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in vb2_dc_dmabuf_ops_attach()
Dvideobuf2-dma-sg.c467 sg_set_page(wr, sg_page(rd), rd->length, rd->offset); in vb2_dma_sg_dmabuf_ops_attach()
/linux-4.1.27/drivers/md/
Ddm-crypt.c585 src = kmap_atomic(sg_page(&dmreq->sg_in)); in crypt_iv_lmk_gen()
603 dst = kmap_atomic(sg_page(&dmreq->sg_out)); in crypt_iv_lmk_post()
727 src = kmap_atomic(sg_page(&dmreq->sg_in)); in crypt_iv_tcw_gen()
751 dst = kmap_atomic(sg_page(&dmreq->sg_out)); in crypt_iv_tcw_post()
/linux-4.1.27/arch/s390/pci/
Dpci_dma.c387 struct page *page = sg_page(s); in s390_dma_map_sg()
/linux-4.1.27/drivers/gpu/drm/
Ddrm_prime.c722 page = sg_page(sg); in drm_prime_sg_to_page_addr_arrays()
/linux-4.1.27/drivers/infiniband/hw/usnic/
Dusnic_uiom.c89 page = sg_page(sg); in usnic_uiom_put_pages()
/linux-4.1.27/drivers/block/
Dxen-blkfront.c525 gnt_list_entry = get_grant(&gref_head, page_to_pfn(sg_page(sg)), info); in blkif_queue_request()
537 bvec_data = kmap_atomic(sg_page(sg)); in blkif_queue_request()
1071 bvec_data = kmap_atomic(sg_page(sg)); in blkif_completion()
Dskd_main.c1671 page = sg_page(&skspcl->req.sg[sksg_ix++]); in skd_sg_io_copy_buffer()
2663 if (!sg_page(sg)) in skd_sg_1st_page_ptr()
2949 struct page *page = sg_page(&skspcl->req.sg[i]); in skd_release_special()
Dcpqarray.c947 sg_page(&tmp_sg[i]), in do_ida_request()
Dnvme-core.c1743 put_page(sg_page(&iod->sg[i])); in nvme_unmap_user_pages()
Dcciss.c3361 temp64.val = (__u64) pci_map_page(h->pdev, sg_page(&tmp_sg[i]), in do_cciss_request()
/linux-4.1.27/drivers/scsi/isci/
Drequest.c1427 struct page *page = sg_page(sg); in sci_stp_request_pio_data_in_copy_data_buffer()
1780 kaddr = kmap_atomic(sg_page(sg)); in sci_io_request_frame_handler()
2927 kaddr = kmap_atomic(sg_page(sg)); in isci_request_io_request_complete()
3194 kaddr = kmap_atomic(sg_page(sg)); in sci_io_request_construct_smp()
/linux-4.1.27/drivers/memstick/core/
Dms_block.c56 sg_set_page(sg_to, sg_page(sg_from), in msb_sg_copy()
74 sg_set_page(sg_to, sg_page(sg_from), in msb_sg_copy()
85 sg_set_page(sg_to, sg_page(sg_from), len, sg_from->offset); in msb_sg_copy()
Dmspro_block.c639 nth_page(sg_page(&(msb->req_sg[msb->current_seg])), in h_mspro_block_transfer_data()
/linux-4.1.27/drivers/vhost/
Dscsi.c479 put_page(sg_page(&tv_cmd->tvc_sgl[i])); in vhost_scsi_release_cmd()
483 put_page(sg_page(&tv_cmd->tvc_prot_sgl[i])); in vhost_scsi_release_cmd()
874 struct page *page = sg_page(&sg[i]); in vhost_scsi_iov_to_sgl()
/linux-4.1.27/drivers/scsi/cxgbi/
Dlibcxgbi.c1382 struct page *sgpage = sg_page(sg); in ddp_make_gl()
1416 struct page *page = sg_page(sg); in ddp_make_gl()
2044 struct page *page = sg_page(sg); in sgl_read_to_frags()
2060 page = sg_page(sg); in sgl_read_to_frags()
/linux-4.1.27/drivers/ide/
Dide-taskfile.c242 page = sg_page(cursg); in ide_pio_bytes()
/linux-4.1.27/drivers/staging/rts5208/
Drtsx_transport.c79 struct page *page = sg_page(sg) + in rtsx_stor_access_xfer_buf()
/linux-4.1.27/drivers/memstick/host/
Dtifm_ms.c206 pg = nth_page(sg_page(&host->req->sg), in tifm_ms_transfer_data()
Djmb38x_ms.c321 pg = nth_page(sg_page(&host->req->sg), in jmb38x_ms_transfer_data()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_mrmw.c1864 pgaddr = page_to_pfn(sg_page(*sg)) in ehca_set_pagebuf_user1()
1903 u64 pgaddr = page_to_pfn(sg_page(*sg)) << PAGE_SHIFT; in ehca_check_kpages_per_ate()
1933 pgaddr = (page_to_pfn(sg_page(*sg)) in ehca_set_pagebuf_user2()
/linux-4.1.27/drivers/usb/mon/
Dmon_text.c165 if (PageHighMem(sg_page(sg))) in mon_text_get_data()
Dmon_bin.c426 if (length == 0 || PageHighMem(sg_page(sg))) in mon_bin_get_data()
/linux-4.1.27/drivers/iommu/
Diommu.c1150 phys_addr_t phys = page_to_phys(sg_page(s)) + s->offset; in default_iommu_map_sg()
Dintel-iommu.c2024 pteval = page_to_phys(sg_page(sg)) | prot; in __domain_mapping()
3326 BUG_ON(!sg_page(sg)); in intel_nontranslate_map_sg()
3327 sg->dma_address = page_to_phys(sg_page(sg)) + sg->offset; in intel_nontranslate_map_sg()
/linux-4.1.27/arch/sparc/kernel/
Dldc.c2054 unsigned long base = page_to_pfn(sg_page(sg)) << PAGE_SHIFT; in sg_count_one()
2116 fill_cookies(&state, page_to_pfn(sg_page(&sg[i])) << PAGE_SHIFT, in ldc_map_sg()
/linux-4.1.27/drivers/ata/
Dlibata-sff.c709 page = sg_page(qc->cursg); in ata_pio_sector()
848 page = sg_page(sg); in __atapi_pio_bytes()
Dlibata-scsi.c3153 buf = page_address(sg_page(scsi_sglist(scmd))); in ata_scsi_write_same_xlat()
3279 p = page_address(sg_page(scsi_sglist(scmd))); in ata_scsi_mode_select_xlat()
/linux-4.1.27/arch/um/drivers/
Dubd_kern.c1318 sg->offset, sg->length, sg_page(sg)); in do_ubd_request()
/linux-4.1.27/drivers/scsi/megaraid/
Dmegaraid_mbox.c1567 if (sg_page(sgl)) { in DEF_SCSI_QCMD()
2309 if (sg_page(sgl)) { in megaraid_mbox_dpc()
/linux-4.1.27/drivers/staging/unisys/virthba/
Dvirthba.c1054 thispage_orig = kmap_atomic(sg_page(sg + i)); in do_scsi_nolinuxstat()
/linux-4.1.27/drivers/target/iscsi/
Discsi_target_util.c1230 sg_page(sg), sg->offset + offset, sub_len, 0); in iscsit_fe_sendpage_sg()
Discsi_target.c757 iov[i].iov_base = kmap(sg_page(sg)) + sg->offset + page_off; in iscsit_map_iovec()
779 kunmap(sg_page(&sg[i])); in iscsit_unmap_iovec()
/linux-4.1.27/drivers/usb/core/
Dmessage.c433 if (!PageHighMem(sg_page(sg))) in usb_sg_init()
Dhcd.c1497 sg_page(sg), in usb_hcd_map_urb_for_dma()
Ddevio.c297 if (sg_page(&as->urb->sg[i])) in free_async()
/linux-4.1.27/drivers/scsi/qla2xxx/
Dqla_target.c2797 spt = page_address(sg_page(sg)) + sg->offset; in qlt_handle_dif_error()
4351 i, sg, sg_page(sg), sg->length, sg->offset); in qlt_set_data_offset()
4387 sg_set_page(sgp, sg_page(sg), in qlt_set_data_offset()
4391 sg_set_page(sgp, sg_page(sg), sg->length, 0); in qlt_set_data_offset()
Dqla_isr.c1804 spt = page_address(sg_page(sg)) + sg->offset; in qla2x00_handle_dif_error()
/linux-4.1.27/drivers/char/
Dvirtio_console.c370 struct page *page = sg_page(&buf->sg[i]); in free_buf()
/linux-4.1.27/drivers/scsi/arcmsr/
Darcmsr_hba.c2239 buffer = kmap_atomic(sg_page(sg)) + sg->offset; in arcmsr_iop_message_xfer()
2515 buffer = kmap_atomic(sg_page(sg)) + sg->offset; in arcmsr_handle_virtual_command()
/linux-4.1.27/drivers/scsi/csiostor/
Dcsio_scsi.c1528 sg_addr = kmap_atomic(sg_page(sg) + (sg_off >> PAGE_SHIFT)); in csio_scsi_copy_to_sgl()
/linux-4.1.27/drivers/scsi/mvsas/
Dmv_sas.c1796 to = kmap_atomic(sg_page(sg_resp)); in mvs_slot_complete()
/linux-4.1.27/drivers/usb/misc/
Dusbtest.c484 if (!sg_page(&sg[i])) in free_sglist()
/linux-4.1.27/drivers/usb/wusbcore/
Dwa-xfer.c1071 sg_set_page(current_seg_sg, sg_page(current_xfer_sg), in wa_xfer_create_subset_sg()
/linux-4.1.27/drivers/infiniband/hw/nes/
Dnes_verbs.c2619 nespbl->page = sg_page(sg); in nes_reg_user_mr()