Lines Matching refs:port_mmio

650 static int mv_stop_edma_engine(void __iomem *port_mmio);
957 void __iomem *port_mmio = mv_ap_base(ap); in mv_save_cached_regs() local
960 pp->cached.fiscfg = readl(port_mmio + FISCFG); in mv_save_cached_regs()
961 pp->cached.ltmode = readl(port_mmio + LTMODE); in mv_save_cached_regs()
962 pp->cached.haltcond = readl(port_mmio + EDMA_HALTCOND); in mv_save_cached_regs()
963 pp->cached.unknown_rsvd = readl(port_mmio + EDMA_UNKNOWN_RSVD); in mv_save_cached_regs()
1001 static void mv_set_edma_ptrs(void __iomem *port_mmio, in mv_set_edma_ptrs() argument
1014 writel((pp->crqb_dma >> 16) >> 16, port_mmio + EDMA_REQ_Q_BASE_HI); in mv_set_edma_ptrs()
1016 port_mmio + EDMA_REQ_Q_IN_PTR); in mv_set_edma_ptrs()
1017 writelfl(index, port_mmio + EDMA_REQ_Q_OUT_PTR); in mv_set_edma_ptrs()
1026 writel((pp->crpb_dma >> 16) >> 16, port_mmio + EDMA_RSP_Q_BASE_HI); in mv_set_edma_ptrs()
1027 writelfl(index, port_mmio + EDMA_RSP_Q_IN_PTR); in mv_set_edma_ptrs()
1029 port_mmio + EDMA_RSP_Q_OUT_PTR); in mv_set_edma_ptrs()
1077 void __iomem *port_mmio, in mv_clear_and_enable_port_irqs() argument
1087 writelfl(0, port_mmio + EDMA_ERR_IRQ_CAUSE); in mv_clear_and_enable_port_irqs()
1095 writelfl(0, port_mmio + FIS_IRQ_CAUSE); in mv_clear_and_enable_port_irqs()
1173 static void mv_start_edma(struct ata_port *ap, void __iomem *port_mmio, in mv_start_edma() argument
1188 mv_set_edma_ptrs(port_mmio, hpriv, pp); in mv_start_edma()
1189 mv_clear_and_enable_port_irqs(ap, port_mmio, DONE_IRQ|ERR_IRQ); in mv_start_edma()
1191 writelfl(EDMA_EN, port_mmio + EDMA_CMD); in mv_start_edma()
1198 void __iomem *port_mmio = mv_ap_base(ap); in mv_wait_for_edma_empty_idle() local
1211 u32 edma_stat = readl(port_mmio + EDMA_STATUS); in mv_wait_for_edma_empty_idle()
1226 static int mv_stop_edma_engine(void __iomem *port_mmio) in mv_stop_edma_engine() argument
1231 writelfl(EDMA_DS, port_mmio + EDMA_CMD); in mv_stop_edma_engine()
1235 u32 reg = readl(port_mmio + EDMA_CMD); in mv_stop_edma_engine()
1245 void __iomem *port_mmio = mv_ap_base(ap); in mv_stop_edma() local
1253 if (mv_stop_edma_engine(port_mmio)) { in mv_stop_edma()
1494 void __iomem *port_mmio; in mv_config_fbs() local
1514 port_mmio = mv_ap_base(ap); in mv_config_fbs()
1515 mv_write_cached_reg(port_mmio + FISCFG, old_fiscfg, fiscfg); in mv_config_fbs()
1516 mv_write_cached_reg(port_mmio + LTMODE, old_ltmode, ltmode); in mv_config_fbs()
1517 mv_write_cached_reg(port_mmio + EDMA_HALTCOND, old_haltcond, haltcond); in mv_config_fbs()
1619 void __iomem *port_mmio = mv_ap_base(ap); in mv_edma_cfg() local
1675 writelfl(cfg, port_mmio + EDMA_CFG); in mv_edma_cfg()
1899 void __iomem *port_mmio = mv_ap_base(ap); in mv_bmdma_setup() local
1905 writel(0, port_mmio + BMDMA_CMD); in mv_bmdma_setup()
1909 port_mmio + BMDMA_PRD_HIGH); in mv_bmdma_setup()
1911 port_mmio + BMDMA_PRD_LOW); in mv_bmdma_setup()
1927 void __iomem *port_mmio = mv_ap_base(ap); in mv_bmdma_start() local
1932 writelfl(cmd, port_mmio + BMDMA_CMD); in mv_bmdma_start()
1946 void __iomem *port_mmio = mv_ap_base(ap); in mv_bmdma_stop_ap() local
1950 cmd = readl(port_mmio + BMDMA_CMD); in mv_bmdma_stop_ap()
1953 writelfl(cmd, port_mmio + BMDMA_CMD); in mv_bmdma_stop_ap()
1976 void __iomem *port_mmio = mv_ap_base(ap); in mv_bmdma_status() local
1983 reg = readl(port_mmio + BMDMA_STATUS); in mv_bmdma_status()
2242 void __iomem *port_mmio = mv_ap_base(ap); in mv_send_fis() local
2247 old_ifctl = readl(port_mmio + SATA_IFCTL); in mv_send_fis()
2249 writelfl(ifctl, port_mmio + SATA_IFCTL); in mv_send_fis()
2253 writel(fis[i], port_mmio + VENDOR_UNIQUE_FIS); in mv_send_fis()
2256 writelfl(ifctl | 0x200, port_mmio + SATA_IFCTL); in mv_send_fis()
2257 writelfl(fis[final_word], port_mmio + VENDOR_UNIQUE_FIS); in mv_send_fis()
2264 ifstat = readl(port_mmio + SATA_IFSTAT); in mv_send_fis()
2268 writelfl(old_ifctl, port_mmio + SATA_IFCTL); in mv_send_fis()
2349 void __iomem *port_mmio = mv_ap_base(ap); in mv_qc_issue() local
2365 mv_start_edma(ap, port_mmio, pp, qc->tf.protocol); in mv_qc_issue()
2371 port_mmio + EDMA_REQ_Q_IN_PTR); in mv_qc_issue()
2476 void __iomem *port_mmio = mv_ap_base(ap); in mv_get_err_pmp_map() local
2478 return readl(port_mmio + SATA_TESTCTL) >> 16; in mv_get_err_pmp_map()
2508 void __iomem *port_mmio = mv_ap_base(ap); in mv_req_q_empty() local
2511 in_ptr = (readl(port_mmio + EDMA_REQ_Q_IN_PTR) in mv_req_q_empty()
2513 out_ptr = (readl(port_mmio + EDMA_REQ_Q_OUT_PTR) in mv_req_q_empty()
2653 void __iomem *port_mmio = mv_ap_base(ap); in mv_err_intr() local
2671 edma_err_cause = readl(port_mmio + EDMA_ERR_IRQ_CAUSE); in mv_err_intr()
2673 fis_cause = readl(port_mmio + FIS_IRQ_CAUSE); in mv_err_intr()
2674 writelfl(~fis_cause, port_mmio + FIS_IRQ_CAUSE); in mv_err_intr()
2676 writelfl(~edma_err_cause, port_mmio + EDMA_ERR_IRQ_CAUSE); in mv_err_intr()
2816 void __iomem *port_mmio = mv_ap_base(ap); in mv_process_crpb_entries() local
2824 in_index = (readl(port_mmio + EDMA_RSP_Q_IN_PTR) in mv_process_crpb_entries()
2852 port_mmio + EDMA_RSP_Q_OUT_PTR); in mv_process_crpb_entries()
3172 #define ZERO(reg) writel(0, port_mmio + (reg))
3176 void __iomem *port_mmio = mv_port_base(mmio, port); in mv5_reset_hc_port() local
3181 writel(0x11f, port_mmio + EDMA_CFG); in mv5_reset_hc_port()
3192 writel(0xbc, port_mmio + EDMA_IORDY_TMOUT); in mv5_reset_hc_port()
3334 void __iomem *port_mmio; in mv6_read_preamp() local
3344 port_mmio = mv_port_base(mmio, idx); in mv6_read_preamp()
3345 tmp = readl(port_mmio + PHY_MODE2); in mv6_read_preamp()
3359 void __iomem *port_mmio = mv_port_base(mmio, port); in mv6_phy_errata() local
3369 m2 = readl(port_mmio + PHY_MODE2); in mv6_phy_errata()
3372 writel(m2, port_mmio + PHY_MODE2); in mv6_phy_errata()
3376 m2 = readl(port_mmio + PHY_MODE2); in mv6_phy_errata()
3378 writel(m2, port_mmio + PHY_MODE2); in mv6_phy_errata()
3387 m3 = readl(port_mmio + PHY_MODE3); in mv6_phy_errata()
3395 u32 m4 = readl(port_mmio + PHY_MODE4); in mv6_phy_errata()
3405 writel(m4, port_mmio + PHY_MODE4); in mv6_phy_errata()
3413 writel(m3, port_mmio + PHY_MODE3); in mv6_phy_errata()
3416 m2 = readl(port_mmio + PHY_MODE2); in mv6_phy_errata()
3429 writel(m2, port_mmio + PHY_MODE2); in mv6_phy_errata()
3443 void __iomem *port_mmio; in mv_soc_read_preamp() local
3446 port_mmio = mv_port_base(mmio, idx); in mv_soc_read_preamp()
3447 tmp = readl(port_mmio + PHY_MODE2); in mv_soc_read_preamp()
3454 #define ZERO(reg) writel(0, port_mmio + (reg))
3458 void __iomem *port_mmio = mv_port_base(mmio, port); in mv_soc_reset_hc_port() local
3463 writel(0x101f, port_mmio + EDMA_CFG); in mv_soc_reset_hc_port()
3474 writel(0x800, port_mmio + EDMA_IORDY_TMOUT); in mv_soc_reset_hc_port()
3520 void __iomem *port_mmio = mv_port_base(mmio, port); in mv_soc_65n_phy_errata() local
3523 reg = readl(port_mmio + PHY_MODE3); in mv_soc_65n_phy_errata()
3528 writel(reg, port_mmio + PHY_MODE3); in mv_soc_65n_phy_errata()
3530 reg = readl(port_mmio + PHY_MODE4); in mv_soc_65n_phy_errata()
3533 writel(reg, port_mmio + PHY_MODE4); in mv_soc_65n_phy_errata()
3535 reg = readl(port_mmio + PHY_MODE9_GEN2); in mv_soc_65n_phy_errata()
3539 writel(reg, port_mmio + PHY_MODE9_GEN2); in mv_soc_65n_phy_errata()
3541 reg = readl(port_mmio + PHY_MODE9_GEN1); in mv_soc_65n_phy_errata()
3545 writel(reg, port_mmio + PHY_MODE9_GEN1); in mv_soc_65n_phy_errata()
3564 static void mv_setup_ifcfg(void __iomem *port_mmio, int want_gen2i) in mv_setup_ifcfg() argument
3566 u32 ifcfg = readl(port_mmio + SATA_IFCFG); in mv_setup_ifcfg()
3571 writelfl(ifcfg, port_mmio + SATA_IFCFG); in mv_setup_ifcfg()
3577 void __iomem *port_mmio = mv_port_base(mmio, port_no); in mv_reset_channel() local
3584 mv_stop_edma_engine(port_mmio); in mv_reset_channel()
3585 writelfl(EDMA_RESET, port_mmio + EDMA_CMD); in mv_reset_channel()
3589 mv_setup_ifcfg(port_mmio, 1); in mv_reset_channel()
3596 writelfl(EDMA_RESET, port_mmio + EDMA_CMD); in mv_reset_channel()
3598 writelfl(0, port_mmio + EDMA_CMD); in mv_reset_channel()
3609 void __iomem *port_mmio = mv_ap_base(ap); in mv_pmp_select() local
3610 u32 reg = readl(port_mmio + SATA_IFCTL); in mv_pmp_select()
3615 writelfl(reg, port_mmio + SATA_IFCTL); in mv_pmp_select()
3686 void __iomem *port_mmio = mv_ap_base(ap); in mv_eh_thaw() local
3690 writel(0, port_mmio + EDMA_ERR_IRQ_CAUSE); in mv_eh_thaw()
3711 static void mv_port_init(struct ata_ioports *port, void __iomem *port_mmio) in mv_port_init() argument
3713 void __iomem *serr, *shd_base = port_mmio + SHD_BLK; in mv_port_init()
3731 serr = port_mmio + mv_scr_offset(SCR_ERROR); in mv_port_init()
3733 writelfl(0, port_mmio + EDMA_ERR_IRQ_CAUSE); in mv_port_init()
3736 writelfl(~EDMA_ERR_IRQ_TRANSIENT, port_mmio + EDMA_ERR_IRQ_MASK); in mv_port_init()
3739 readl(port_mmio + EDMA_CFG), in mv_port_init()
3740 readl(port_mmio + EDMA_ERR_IRQ_CAUSE), in mv_port_init()
3741 readl(port_mmio + EDMA_ERR_IRQ_MASK)); in mv_port_init()
3978 void __iomem *port_mmio = mv_port_base(mmio, port); in mv_init_host() local
3980 mv_port_init(&ap->ioaddr, port_mmio); in mv_init_host()
4444 void __iomem *port_mmio = mv_port_base(hpriv->base, port); in mv_pci_init_one() local
4445 unsigned int offset = port_mmio - hpriv->base; in mv_pci_init_one()