cpa 56 arch/s390/include/asm/scsw.h __u32 cpa; cpa 258 arch/x86/mm/pageattr.c static unsigned long __cpa_addr(struct cpa_data *cpa, unsigned long idx) cpa 260 arch/x86/mm/pageattr.c if (cpa->flags & CPA_PAGES_ARRAY) { cpa 261 arch/x86/mm/pageattr.c struct page *page = cpa->pages[idx]; cpa 269 arch/x86/mm/pageattr.c if (cpa->flags & CPA_ARRAY) cpa 270 arch/x86/mm/pageattr.c return cpa->vaddr[idx]; cpa 272 arch/x86/mm/pageattr.c return *cpa->vaddr + idx * PAGE_SIZE; cpa 337 arch/x86/mm/pageattr.c struct cpa_data *cpa = data; cpa 340 arch/x86/mm/pageattr.c for (i = 0; i < cpa->numpages; i++) cpa 341 arch/x86/mm/pageattr.c __flush_tlb_one_kernel(fix_addr(__cpa_addr(cpa, i))); cpa 346 arch/x86/mm/pageattr.c struct cpa_data *cpa = data; cpa 356 arch/x86/mm/pageattr.c if (cpa->force_flush_all || cpa->numpages > tlb_single_page_flush_ceiling) cpa 359 arch/x86/mm/pageattr.c on_each_cpu(__cpa_flush_tlb, cpa, 1); cpa 365 arch/x86/mm/pageattr.c for (i = 0; i < cpa->numpages; i++) { cpa 366 arch/x86/mm/pageattr.c unsigned long addr = __cpa_addr(cpa, i); cpa 622 arch/x86/mm/pageattr.c static pte_t *_lookup_address_cpa(struct cpa_data *cpa, unsigned long address, cpa 625 arch/x86/mm/pageattr.c if (cpa->pgd) cpa 626 arch/x86/mm/pageattr.c return lookup_address_in_pgd(cpa->pgd + pgd_index(address), cpa 747 arch/x86/mm/pageattr.c struct cpa_data *cpa) cpa 758 arch/x86/mm/pageattr.c tmp = _lookup_address_cpa(cpa, address, &level); cpa 786 arch/x86/mm/pageattr.c if (numpages < cpa->numpages) cpa 787 arch/x86/mm/pageattr.c cpa->numpages = numpages; cpa 798 arch/x86/mm/pageattr.c pgprot_val(req_prot) &= ~pgprot_val(cpa->mask_clr); cpa 799 arch/x86/mm/pageattr.c pgprot_val(req_prot) |= pgprot_val(cpa->mask_set); cpa 816 arch/x86/mm/pageattr.c cpa->pfn = pfn; cpa 838 arch/x86/mm/pageattr.c cpa->force_static_prot = 1; cpa 859 arch/x86/mm/pageattr.c if (address != lpaddr || cpa->numpages != numpages) cpa 884 arch/x86/mm/pageattr.c cpa->flags |= CPA_FLUSHTLB; cpa 890 arch/x86/mm/pageattr.c struct cpa_data *cpa) cpa 894 arch/x86/mm/pageattr.c if (cpa->force_split) cpa 898 arch/x86/mm/pageattr.c do_split = __should_split_large_page(kpte, address, cpa); cpa 904 arch/x86/mm/pageattr.c static void split_set_pte(struct cpa_data *cpa, pte_t *pte, unsigned long pfn, cpa 915 arch/x86/mm/pageattr.c if (!cpa->force_static_prot) cpa 941 arch/x86/mm/pageattr.c __split_large_page(struct cpa_data *cpa, pte_t *kpte, unsigned long address, cpa 955 arch/x86/mm/pageattr.c tmp = _lookup_address_cpa(cpa, address, &level); cpa 1003 arch/x86/mm/pageattr.c split_set_pte(cpa, pbase + i, pfn, ref_prot, lpaddr, lpinc); cpa 1045 arch/x86/mm/pageattr.c static int split_large_page(struct cpa_data *cpa, pte_t *kpte, cpa 1058 arch/x86/mm/pageattr.c if (__split_large_page(cpa, kpte, address, base)) cpa 1221 arch/x86/mm/pageattr.c static void populate_pte(struct cpa_data *cpa, cpa 1232 arch/x86/mm/pageattr.c set_pte(pte, pfn_pte(cpa->pfn, pgprot)); cpa 1235 arch/x86/mm/pageattr.c cpa->pfn++; cpa 1240 arch/x86/mm/pageattr.c static long populate_pmd(struct cpa_data *cpa, cpa 1267 arch/x86/mm/pageattr.c populate_pte(cpa, start, pre_end, cur_pages, pmd, pgprot); cpa 1291 arch/x86/mm/pageattr.c set_pmd(pmd, pmd_mkhuge(pfn_pmd(cpa->pfn, cpa 1295 arch/x86/mm/pageattr.c cpa->pfn += PMD_SIZE >> PAGE_SHIFT; cpa 1308 arch/x86/mm/pageattr.c populate_pte(cpa, start, end, num_pages - cur_pages, cpa 1314 arch/x86/mm/pageattr.c static int populate_pud(struct cpa_data *cpa, unsigned long start, p4d_t *p4d, cpa 1322 arch/x86/mm/pageattr.c end = start + (cpa->numpages << PAGE_SHIFT); cpa 1334 arch/x86/mm/pageattr.c cur_pages = min_t(int, (int)cpa->numpages, cur_pages); cpa 1345 arch/x86/mm/pageattr.c cur_pages = populate_pmd(cpa, start, pre_end, cur_pages, cpa 1354 arch/x86/mm/pageattr.c if (cpa->numpages == cur_pages) cpa 1364 arch/x86/mm/pageattr.c set_pud(pud, pud_mkhuge(pfn_pud(cpa->pfn, cpa 1368 arch/x86/mm/pageattr.c cpa->pfn += PUD_SIZE >> PAGE_SHIFT; cpa 1382 arch/x86/mm/pageattr.c tmp = populate_pmd(cpa, start, end, cpa->numpages - cur_pages, cpa 1396 arch/x86/mm/pageattr.c static int populate_pgd(struct cpa_data *cpa, unsigned long addr) cpa 1404 arch/x86/mm/pageattr.c pgd_entry = cpa->pgd + pgd_index(addr); cpa 1426 arch/x86/mm/pageattr.c pgprot_val(pgprot) &= ~pgprot_val(cpa->mask_clr); cpa 1427 arch/x86/mm/pageattr.c pgprot_val(pgprot) |= pgprot_val(cpa->mask_set); cpa 1429 arch/x86/mm/pageattr.c ret = populate_pud(cpa, addr, p4d, pgprot); cpa 1437 arch/x86/mm/pageattr.c addr + (cpa->numpages << PAGE_SHIFT)); cpa 1441 arch/x86/mm/pageattr.c cpa->numpages = ret; cpa 1445 arch/x86/mm/pageattr.c static int __cpa_process_fault(struct cpa_data *cpa, unsigned long vaddr, cpa 1448 arch/x86/mm/pageattr.c if (cpa->pgd) { cpa 1454 arch/x86/mm/pageattr.c return populate_pgd(cpa, vaddr); cpa 1461 arch/x86/mm/pageattr.c cpa->numpages = 1; cpa 1474 arch/x86/mm/pageattr.c cpa->numpages = 1; cpa 1475 arch/x86/mm/pageattr.c cpa->pfn = __pa(vaddr) >> PAGE_SHIFT; cpa 1478 arch/x86/mm/pageattr.c } else if (__cpa_pfn_in_highmap(cpa->pfn)) { cpa 1484 arch/x86/mm/pageattr.c *cpa->vaddr); cpa 1490 arch/x86/mm/pageattr.c static int __change_page_attr(struct cpa_data *cpa, int primary) cpa 1497 arch/x86/mm/pageattr.c address = __cpa_addr(cpa, cpa->curpage); cpa 1499 arch/x86/mm/pageattr.c kpte = _lookup_address_cpa(cpa, address, &level); cpa 1501 arch/x86/mm/pageattr.c return __cpa_process_fault(cpa, address, primary); cpa 1505 arch/x86/mm/pageattr.c return __cpa_process_fault(cpa, address, primary); cpa 1512 arch/x86/mm/pageattr.c pgprot_val(new_prot) &= ~pgprot_val(cpa->mask_clr); cpa 1513 arch/x86/mm/pageattr.c pgprot_val(new_prot) |= pgprot_val(cpa->mask_set); cpa 1528 arch/x86/mm/pageattr.c cpa->pfn = pfn; cpa 1534 arch/x86/mm/pageattr.c cpa->flags |= CPA_FLUSHTLB; cpa 1536 arch/x86/mm/pageattr.c cpa->numpages = 1; cpa 1544 arch/x86/mm/pageattr.c do_split = should_split_large_page(kpte, address, cpa); cpa 1556 arch/x86/mm/pageattr.c err = split_large_page(cpa, kpte, address); cpa 1563 arch/x86/mm/pageattr.c static int __change_page_attr_set_clr(struct cpa_data *cpa, int checkalias); cpa 1565 arch/x86/mm/pageattr.c static int cpa_process_alias(struct cpa_data *cpa) cpa 1568 arch/x86/mm/pageattr.c unsigned long laddr = (unsigned long)__va(cpa->pfn << PAGE_SHIFT); cpa 1572 arch/x86/mm/pageattr.c if (!pfn_range_is_mapped(cpa->pfn, cpa->pfn + 1)) cpa 1579 arch/x86/mm/pageattr.c vaddr = __cpa_addr(cpa, cpa->curpage); cpa 1583 arch/x86/mm/pageattr.c alias_cpa = *cpa; cpa 1588 arch/x86/mm/pageattr.c cpa->force_flush_all = 1; cpa 1602 arch/x86/mm/pageattr.c __cpa_pfn_in_highmap(cpa->pfn)) { cpa 1603 arch/x86/mm/pageattr.c unsigned long temp_cpa_vaddr = (cpa->pfn << PAGE_SHIFT) + cpa 1605 arch/x86/mm/pageattr.c alias_cpa = *cpa; cpa 1610 arch/x86/mm/pageattr.c cpa->force_flush_all = 1; cpa 1622 arch/x86/mm/pageattr.c static int __change_page_attr_set_clr(struct cpa_data *cpa, int checkalias) cpa 1624 arch/x86/mm/pageattr.c unsigned long numpages = cpa->numpages; cpa 1633 arch/x86/mm/pageattr.c cpa->numpages = rempages; cpa 1635 arch/x86/mm/pageattr.c if (cpa->flags & (CPA_ARRAY | CPA_PAGES_ARRAY)) cpa 1636 arch/x86/mm/pageattr.c cpa->numpages = 1; cpa 1640 arch/x86/mm/pageattr.c ret = __change_page_attr(cpa, checkalias); cpa 1647 arch/x86/mm/pageattr.c ret = cpa_process_alias(cpa); cpa 1657 arch/x86/mm/pageattr.c BUG_ON(cpa->numpages > rempages || !cpa->numpages); cpa 1658 arch/x86/mm/pageattr.c rempages -= cpa->numpages; cpa 1659 arch/x86/mm/pageattr.c cpa->curpage += cpa->numpages; cpa 1664 arch/x86/mm/pageattr.c cpa->numpages = numpages; cpa 1673 arch/x86/mm/pageattr.c struct cpa_data cpa; cpa 1676 arch/x86/mm/pageattr.c memset(&cpa, 0, sizeof(cpa)); cpa 1715 arch/x86/mm/pageattr.c cpa.vaddr = addr; cpa 1716 arch/x86/mm/pageattr.c cpa.pages = pages; cpa 1717 arch/x86/mm/pageattr.c cpa.numpages = numpages; cpa 1718 arch/x86/mm/pageattr.c cpa.mask_set = mask_set; cpa 1719 arch/x86/mm/pageattr.c cpa.mask_clr = mask_clr; cpa 1720 arch/x86/mm/pageattr.c cpa.flags = 0; cpa 1721 arch/x86/mm/pageattr.c cpa.curpage = 0; cpa 1722 arch/x86/mm/pageattr.c cpa.force_split = force_split; cpa 1725 arch/x86/mm/pageattr.c cpa.flags |= in_flag; cpa 1733 arch/x86/mm/pageattr.c ret = __change_page_attr_set_clr(&cpa, checkalias); cpa 1738 arch/x86/mm/pageattr.c if (!(cpa.flags & CPA_FLUSHTLB)) cpa 1755 arch/x86/mm/pageattr.c cpa_flush(&cpa, cache); cpa 1945 arch/x86/mm/pageattr.c struct cpa_data cpa; cpa 1956 arch/x86/mm/pageattr.c memset(&cpa, 0, sizeof(cpa)); cpa 1957 arch/x86/mm/pageattr.c cpa.vaddr = &addr; cpa 1958 arch/x86/mm/pageattr.c cpa.numpages = numpages; cpa 1959 arch/x86/mm/pageattr.c cpa.mask_set = enc ? __pgprot(_PAGE_ENC) : __pgprot(0); cpa 1960 arch/x86/mm/pageattr.c cpa.mask_clr = enc ? __pgprot(0) : __pgprot(_PAGE_ENC); cpa 1961 arch/x86/mm/pageattr.c cpa.pgd = init_mm.pgd; cpa 1970 arch/x86/mm/pageattr.c cpa_flush(&cpa, 1); cpa 1972 arch/x86/mm/pageattr.c ret = __change_page_attr_set_clr(&cpa, 1); cpa 1981 arch/x86/mm/pageattr.c cpa_flush(&cpa, 0); cpa 2120 arch/x86/mm/pageattr.c struct cpa_data cpa = { .vaddr = &tempaddr, cpa 2133 arch/x86/mm/pageattr.c return __change_page_attr_set_clr(&cpa, 0); cpa 2139 arch/x86/mm/pageattr.c struct cpa_data cpa = { .vaddr = &tempaddr, cpa 2152 arch/x86/mm/pageattr.c return __change_page_attr_set_clr(&cpa, 0); cpa 2216 arch/x86/mm/pageattr.c struct cpa_data cpa = { cpa 2232 arch/x86/mm/pageattr.c cpa.mask_clr = pgprot_encrypted(cpa.mask_clr); cpa 2234 arch/x86/mm/pageattr.c cpa.mask_set = __pgprot(_PAGE_PRESENT | page_flags); cpa 2236 arch/x86/mm/pageattr.c retval = __change_page_attr_set_clr(&cpa, 0); cpa 2259 arch/x86/mm/pageattr.c struct cpa_data cpa = { cpa 2271 arch/x86/mm/pageattr.c retval = __change_page_attr_set_clr(&cpa, 0); cpa 1607 drivers/s390/block/dasd_3990_erp.c __u32 cpa = 0; cpa 1644 drivers/s390/block/dasd_3990_erp.c cpa = default_erp->refers->irb.scsw.cmd.cpa; cpa 1646 drivers/s390/block/dasd_3990_erp.c if (cpa == 0) { cpa 1725 drivers/s390/block/dasd_3990_erp.c ccw->cda = cpa; cpa 1766 drivers/s390/block/dasd_3990_erp.c __u32 cpa = 0; cpa 1802 drivers/s390/block/dasd_3990_erp.c cpa = previous_erp->irb.scsw.cmd.cpa; cpa 1804 drivers/s390/block/dasd_3990_erp.c if (cpa == 0) { cpa 1852 drivers/s390/block/dasd_3990_erp.c ccw->cda = cpa; cpa 5476 drivers/s390/block/dasd_eckd.c (void *) (addr_t) irb->scsw.cmd.cpa); cpa 5527 drivers/s390/block/dasd_eckd.c irb->scsw.cmd.cpa; /* failing CCW */ cpa 690 drivers/s390/block/dasd_fba.c (void *) (addr_t) irb->scsw.cmd.cpa); cpa 732 drivers/s390/block/dasd_fba.c if (act < ((struct ccw1 *)(addr_t) irb->scsw.cmd.cpa) - 2) { cpa 733 drivers/s390/block/dasd_fba.c act = ((struct ccw1 *)(addr_t) irb->scsw.cmd.cpa) - 2; cpa 736 drivers/s390/block/dasd_fba.c end = min((struct ccw1 *)(addr_t) irb->scsw.cmd.cpa + 2, last); cpa 212 drivers/s390/char/vmur.c static void free_chan_prog(struct ccw1 *cpa) cpa 214 drivers/s390/char/vmur.c struct ccw1 *ptr = cpa; cpa 220 drivers/s390/char/vmur.c kfree(cpa); cpa 234 drivers/s390/char/vmur.c struct ccw1 *cpa; cpa 245 drivers/s390/char/vmur.c cpa = kcalloc(rec_count + 1, sizeof(struct ccw1), cpa 247 drivers/s390/char/vmur.c if (!cpa) cpa 251 drivers/s390/char/vmur.c cpa[i].cmd_code = WRITE_CCW_CMD; cpa 252 drivers/s390/char/vmur.c cpa[i].flags = CCW_FLAG_CC | CCW_FLAG_SLI; cpa 253 drivers/s390/char/vmur.c cpa[i].count = reclen; cpa 256 drivers/s390/char/vmur.c free_chan_prog(cpa); cpa 259 drivers/s390/char/vmur.c cpa[i].cda = (u32)(addr_t) kbuf; cpa 261 drivers/s390/char/vmur.c free_chan_prog(cpa); cpa 267 drivers/s390/char/vmur.c cpa[i].cmd_code = CCW_CMD_NOOP; cpa 268 drivers/s390/char/vmur.c return cpa; cpa 271 drivers/s390/char/vmur.c static int do_ur_io(struct urdev *urd, struct ccw1 *cpa) cpa 277 drivers/s390/char/vmur.c TRACE("do_ur_io: cpa=%p\n", cpa); cpa 286 drivers/s390/char/vmur.c rc = ccw_device_start(cdev, cpa, 1, 0, 0); cpa 419 drivers/s390/char/vmur.c struct ccw1 *cpa; cpa 422 drivers/s390/char/vmur.c cpa = alloc_chan_prog(udata, count / reclen, reclen); cpa 423 drivers/s390/char/vmur.c if (IS_ERR(cpa)) cpa 424 drivers/s390/char/vmur.c return PTR_ERR(cpa); cpa 426 drivers/s390/char/vmur.c rc = do_ur_io(urd, cpa); cpa 438 drivers/s390/char/vmur.c free_chan_prog(cpa); cpa 124 drivers/s390/cio/cio.c struct ccw1 * cpa, /* logical channel prog addr */ cpa 151 drivers/s390/cio/cio.c orb->cmd.cpa = (__u32) __pa(cpa); cpa 176 drivers/s390/cio/cio.c cio_start (struct subchannel *sch, struct ccw1 *cpa, __u8 lpm) cpa 178 drivers/s390/cio/cio.c return cio_start_key(sch, cpa, lpm, PAGE_DEFAULT_KEY); cpa 70 drivers/s390/cio/device_fsm.c if ((void *)(addr_t)orb->cmd.cpa == cpa 72 drivers/s390/cio/device_fsm.c (void *)(addr_t)orb->cmd.cpa == cpa 80 drivers/s390/cio/device_fsm.c (void *)(addr_t)orb->cmd.cpa, cpa 191 drivers/s390/cio/device_ops.c int ccw_device_start_timeout_key(struct ccw_device *cdev, struct ccw1 *cpa, cpa 229 drivers/s390/cio/device_ops.c ret = cio_start_key (sch, cpa, lpm, key); cpa 271 drivers/s390/cio/device_ops.c int ccw_device_start_key(struct ccw_device *cdev, struct ccw1 *cpa, cpa 275 drivers/s390/cio/device_ops.c return ccw_device_start_timeout_key(cdev, cpa, intparm, lpm, key, cpa 305 drivers/s390/cio/device_ops.c int ccw_device_start(struct ccw_device *cdev, struct ccw1 *cpa, cpa 308 drivers/s390/cio/device_ops.c return ccw_device_start_key(cdev, cpa, intparm, lpm, cpa 343 drivers/s390/cio/device_ops.c int ccw_device_start_timeout(struct ccw_device *cdev, struct ccw1 *cpa, cpa 347 drivers/s390/cio/device_ops.c return ccw_device_start_timeout_key(cdev, cpa, intparm, lpm, cpa 271 drivers/s390/cio/device_status.c cdev_irb->scsw.cmd.cpa = irb->scsw.cmd.cpa; cpa 37 drivers/s390/cio/orb.h u32 cpa; /* channel program address */ cpa 298 drivers/s390/cio/vfio_ccw_cp.c static inline int is_cpa_within_range(u32 cpa, u32 head, int len) cpa 302 drivers/s390/cio/vfio_ccw_cp.c return (head <= cpa && cpa <= tail); cpa 651 drivers/s390/cio/vfio_ccw_cp.c ret = ccwchain_handle_ccw(orb->cmd.cpa, cp); cpa 771 drivers/s390/cio/vfio_ccw_cp.c struct ccw1 *cpa; cpa 787 drivers/s390/cio/vfio_ccw_cp.c cpa = chain->ch_ccw; cpa 788 drivers/s390/cio/vfio_ccw_cp.c orb->cmd.cpa = (__u32) __pa(cpa); cpa 810 drivers/s390/cio/vfio_ccw_cp.c u32 cpa = scsw->cmd.cpa; cpa 828 drivers/s390/cio/vfio_ccw_cp.c if (is_cpa_within_range(cpa, ccw_head, chain->ch_len + 1)) { cpa 835 drivers/s390/cio/vfio_ccw_cp.c cpa = chain->ch_iova + (cpa - ccw_head); cpa 840 drivers/s390/cio/vfio_ccw_cp.c scsw->cmd.cpa = cpa; cpa 1405 drivers/s390/net/lcs.c (irb->scsw.cmd.cpa != 0)) { cpa 1406 drivers/s390/net/lcs.c index = (struct ccw1 *) __va((addr_t) irb->scsw.cmd.cpa) cpa 3398 fs/ext4/mballoc.c struct ext4_prealloc_space *cpa) cpa 3402 fs/ext4/mballoc.c if (cpa == NULL) { cpa 3406 fs/ext4/mballoc.c cur_distance = abs(goal_block - cpa->pa_pstart); cpa 3410 fs/ext4/mballoc.c return cpa; cpa 3413 fs/ext4/mballoc.c atomic_dec(&cpa->pa_count); cpa 3428 fs/ext4/mballoc.c struct ext4_prealloc_space *pa, *cpa = NULL; cpa 3492 fs/ext4/mballoc.c cpa = ext4_mb_check_group_pa(goal_block, cpa 3493 fs/ext4/mballoc.c pa, cpa); cpa 3499 fs/ext4/mballoc.c if (cpa) { cpa 3500 fs/ext4/mballoc.c ext4_mb_use_group_pa(ac, cpa); cpa 138 sound/soc/sof/intel/hda-dsp.c unsigned int cpa; cpa 148 sound/soc/sof/intel/hda-dsp.c cpa = HDA_DSP_ADSPCS_CPA_MASK(core_mask); cpa 151 sound/soc/sof/intel/hda-dsp.c (adspcs & cpa) == cpa,