idx 133 arch/alpha/kernel/perf_event.c int idx; idx 158 arch/alpha/kernel/perf_event.c idx0 = ev67_mapping[evtype[0]-1].idx; idx 199 arch/alpha/kernel/perf_event.c event[0]->hw.idx = idx0; idx 202 arch/alpha/kernel/perf_event.c event[1]->hw.idx = idx0 ^ 1; idx 233 arch/alpha/kernel/perf_event.c static inline void alpha_write_pmc(int idx, unsigned long val) idx 235 arch/alpha/kernel/perf_event.c val &= alpha_pmu->pmc_count_mask[idx]; idx 236 arch/alpha/kernel/perf_event.c val <<= alpha_pmu->pmc_count_shift[idx]; idx 237 arch/alpha/kernel/perf_event.c val |= (1<<idx); idx 241 arch/alpha/kernel/perf_event.c static inline unsigned long alpha_read_pmc(int idx) idx 246 arch/alpha/kernel/perf_event.c val >>= alpha_pmu->pmc_count_shift[idx]; idx 247 arch/alpha/kernel/perf_event.c val &= alpha_pmu->pmc_count_mask[idx]; idx 253 arch/alpha/kernel/perf_event.c struct hw_perf_event *hwc, int idx) idx 277 arch/alpha/kernel/perf_event.c if (unlikely(left < alpha_pmu->pmc_left[idx])) idx 278 arch/alpha/kernel/perf_event.c left = alpha_pmu->pmc_left[idx]; idx 280 arch/alpha/kernel/perf_event.c if (left > (long)alpha_pmu->pmc_max_period[idx]) idx 281 arch/alpha/kernel/perf_event.c left = alpha_pmu->pmc_max_period[idx]; idx 285 arch/alpha/kernel/perf_event.c alpha_write_pmc(idx, (unsigned long)(-left)); idx 308 arch/alpha/kernel/perf_event.c struct hw_perf_event *hwc, int idx, long ovf) idx 315 arch/alpha/kernel/perf_event.c new_raw_count = alpha_read_pmc(idx); idx 321 arch/alpha/kernel/perf_event.c delta = (new_raw_count - (prev_raw_count & alpha_pmu->pmc_count_mask[idx])) + ovf; idx 327 arch/alpha/kernel/perf_event.c delta += alpha_pmu->pmc_max_period[idx] + 1; idx 403 arch/alpha/kernel/perf_event.c cpuc->current_idx[j] != pe->hw.idx) { idx 414 arch/alpha/kernel/perf_event.c int idx = hwc->idx; idx 417 arch/alpha/kernel/perf_event.c alpha_perf_event_set_period(pe, hwc, idx); idx 418 arch/alpha/kernel/perf_event.c cpuc->current_idx[j] = idx; idx 497 arch/alpha/kernel/perf_event.c int idx = cpuc->current_idx[j]; idx 510 arch/alpha/kernel/perf_event.c alpha_perf_event_update(event, hwc, idx, 0); idx 513 arch/alpha/kernel/perf_event.c cpuc->idx_mask &= ~(1UL<<idx); idx 528 arch/alpha/kernel/perf_event.c alpha_perf_event_update(event, hwc, hwc->idx, 0); idx 538 arch/alpha/kernel/perf_event.c cpuc->idx_mask &= ~(1UL<<hwc->idx); idx 543 arch/alpha/kernel/perf_event.c alpha_perf_event_update(event, hwc, hwc->idx, 0); idx 548 arch/alpha/kernel/perf_event.c wrperfmon(PERFMON_CMD_DISABLE, (1UL<<hwc->idx)); idx 562 arch/alpha/kernel/perf_event.c alpha_perf_event_set_period(event, hwc, hwc->idx); idx 567 arch/alpha/kernel/perf_event.c cpuc->idx_mask |= 1UL<<hwc->idx; idx 569 arch/alpha/kernel/perf_event.c wrperfmon(PERFMON_CMD_ENABLE, (1UL<<hwc->idx)); idx 664 arch/alpha/kernel/perf_event.c hwc->idx = PMC_NO_INDEX; idx 811 arch/alpha/kernel/perf_event.c int idx, j; idx 832 arch/alpha/kernel/perf_event.c idx = la_ptr; idx 835 arch/alpha/kernel/perf_event.c if (cpuc->current_idx[j] == idx) idx 850 arch/alpha/kernel/perf_event.c pr_warning("PMI: No event at index %d!\n", idx); idx 856 arch/alpha/kernel/perf_event.c alpha_perf_event_update(event, hwc, idx, alpha_pmu->pmc_max_period[idx]+1); idx 859 arch/alpha/kernel/perf_event.c if (alpha_perf_event_set_period(event, hwc, idx)) { idx 100 arch/arc/kernel/perf_event.c static u64 arc_pmu_read_counter(int idx) idx 109 arch/arc/kernel/perf_event.c write_aux_reg(ARC_REG_PCT_INDEX, idx); idx 119 arch/arc/kernel/perf_event.c struct hw_perf_event *hwc, int idx) idx 122 arch/arc/kernel/perf_event.c u64 new_raw_count = arc_pmu_read_counter(idx); idx 136 arch/arc/kernel/perf_event.c arc_perf_event_update(event, &event->hw, event->hw.idx); idx 248 arch/arc/kernel/perf_event.c int idx = hwc->idx; idx 273 arch/arc/kernel/perf_event.c write_aux_reg(ARC_REG_PCT_INDEX, idx); idx 292 arch/arc/kernel/perf_event.c int idx = hwc->idx; idx 294 arch/arc/kernel/perf_event.c if (WARN_ON_ONCE(idx == -1)) idx 307 arch/arc/kernel/perf_event.c read_aux_reg(ARC_REG_PCT_INT_CTRL) | BIT(idx)); idx 310 arch/arc/kernel/perf_event.c write_aux_reg(ARC_REG_PCT_INDEX, idx); /* counter # */ idx 317 arch/arc/kernel/perf_event.c int idx = hwc->idx; idx 325 arch/arc/kernel/perf_event.c write_aux_reg(ARC_REG_PCT_INT_ACT, BIT(idx)); idx 327 arch/arc/kernel/perf_event.c read_aux_reg(ARC_REG_PCT_INT_CTRL) & ~BIT(idx)); idx 332 arch/arc/kernel/perf_event.c write_aux_reg(ARC_REG_PCT_INDEX, idx); idx 342 arch/arc/kernel/perf_event.c arc_perf_event_update(event, &event->hw, idx); idx 352 arch/arc/kernel/perf_event.c __clear_bit(event->hw.idx, pmu_cpu->used_mask); idx 354 arch/arc/kernel/perf_event.c pmu_cpu->act_counter[event->hw.idx] = 0; idx 364 arch/arc/kernel/perf_event.c int idx = hwc->idx; idx 366 arch/arc/kernel/perf_event.c idx = ffz(pmu_cpu->used_mask[0]); idx 367 arch/arc/kernel/perf_event.c if (idx == arc_pmu->n_counters) idx 370 arch/arc/kernel/perf_event.c __set_bit(idx, pmu_cpu->used_mask); idx 371 arch/arc/kernel/perf_event.c hwc->idx = idx; idx 373 arch/arc/kernel/perf_event.c write_aux_reg(ARC_REG_PCT_INDEX, idx); idx 375 arch/arc/kernel/perf_event.c pmu_cpu->act_counter[idx] = event; idx 406 arch/arc/kernel/perf_event.c int idx; idx 420 arch/arc/kernel/perf_event.c idx = __ffs(active_ints); idx 423 arch/arc/kernel/perf_event.c write_aux_reg(ARC_REG_PCT_INT_ACT, BIT(idx)); idx 431 arch/arc/kernel/perf_event.c read_aux_reg(ARC_REG_PCT_INT_CTRL) | BIT(idx)); idx 433 arch/arc/kernel/perf_event.c event = pmu_cpu->act_counter[idx]; idx 436 arch/arc/kernel/perf_event.c WARN_ON_ONCE(hwc->idx != idx); idx 438 arch/arc/kernel/perf_event.c arc_perf_event_update(event, &event->hw, event->hw.idx); idx 445 arch/arc/kernel/perf_event.c active_ints &= ~BIT(idx); idx 64 arch/arc/mm/highmem.c int idx, cpu_idx; idx 73 arch/arc/mm/highmem.c idx = cpu_idx + KM_TYPE_NR * smp_processor_id(); idx 74 arch/arc/mm/highmem.c vaddr = FIXMAP_ADDR(idx); idx 76 arch/arc/mm/highmem.c set_pte_at(&init_mm, vaddr, fixmap_page_table + idx, idx 96 arch/arc/mm/highmem.c int idx = cpu_idx + KM_TYPE_NR * smp_processor_id(); idx 98 arch/arc/mm/highmem.c WARN_ON(kvaddr != FIXMAP_ADDR(idx)); idx 100 arch/arc/mm/highmem.c pte_clear(&init_mm, kvaddr, fixmap_page_table + idx); idx 125 arch/arc/mm/tlb.c unsigned int idx; idx 130 arch/arc/mm/tlb.c idx = read_aux_reg(ARC_REG_TLBINDEX); idx 132 arch/arc/mm/tlb.c return idx; idx 137 arch/arc/mm/tlb.c unsigned int idx; idx 140 arch/arc/mm/tlb.c idx = tlb_entry_lkup(vaddr_n_asid); idx 143 arch/arc/mm/tlb.c if (likely(!(idx & TLB_LKUP_ERR))) { idx 147 arch/arc/mm/tlb.c WARN(idx == TLB_DUP_ERR, "Probe returned Dup PD for %x\n", idx 175 arch/arc/mm/tlb.c unsigned int idx; idx 178 arch/arc/mm/tlb.c idx = read_aux_reg(ARC_REG_TLBINDEX); idx 181 arch/arc/mm/tlb.c if (unlikely(idx & TLB_LKUP_ERR)) idx 192 arch/arc/mm/tlb.c unsigned int idx; idx 198 arch/arc/mm/tlb.c idx = tlb_entry_lkup(pd0); idx 206 arch/arc/mm/tlb.c if (likely(idx & TLB_LKUP_ERR)) idx 104 arch/arm/boot/compressed/atags_to_fdt.c int idx; idx 106 arch/arm/boot/compressed/atags_to_fdt.c for (idx = 7; idx >= 0; idx--) { idx 82 arch/arm/include/asm/barrier.h static inline unsigned long array_index_mask_nospec(unsigned long idx, idx 92 arch/arm/include/asm/barrier.h : "r" (idx), "Ir" (sz) idx 56 arch/arm/include/asm/fixmap.h void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot); idx 42 arch/arm/kernel/module-plts.c int idx = 0; idx 51 arch/arm/kernel/module-plts.c idx = (pltsec->plt_count - 1) % PLT_ENT_COUNT; idx 53 arch/arm/kernel/module-plts.c if (plt->lit[idx] == val) idx 54 arch/arm/kernel/module-plts.c return (u32)&plt->ldr[idx]; idx 56 arch/arm/kernel/module-plts.c idx = (idx + 1) % PLT_ENT_COUNT; idx 57 arch/arm/kernel/module-plts.c if (!idx) idx 64 arch/arm/kernel/module-plts.c if (!idx) idx 71 arch/arm/kernel/module-plts.c plt->lit[idx] = val; idx 73 arch/arm/kernel/module-plts.c return (u32)&plt->ldr[idx]; idx 239 arch/arm/kernel/perf_event_v6.c int counter = hwc->idx; idx 257 arch/arm/kernel/perf_event_v6.c int counter = hwc->idx; idx 275 arch/arm/kernel/perf_event_v6.c int idx = hwc->idx; idx 277 arch/arm/kernel/perf_event_v6.c if (ARMV6_CYCLE_COUNTER == idx) { idx 280 arch/arm/kernel/perf_event_v6.c } else if (ARMV6_COUNTER0 == idx) { idx 284 arch/arm/kernel/perf_event_v6.c } else if (ARMV6_COUNTER1 == idx) { idx 289 arch/arm/kernel/perf_event_v6.c WARN_ONCE(1, "invalid counter number (%d)\n", idx); idx 312 arch/arm/kernel/perf_event_v6.c int idx; idx 326 arch/arm/kernel/perf_event_v6.c for (idx = 0; idx < cpu_pmu->num_events; ++idx) { idx 327 arch/arm/kernel/perf_event_v6.c struct perf_event *event = cpuc->events[idx]; idx 338 arch/arm/kernel/perf_event_v6.c if (!armv6_pmcr_counter_has_overflowed(pmcr, idx)) idx 417 arch/arm/kernel/perf_event_v6.c clear_bit(event->hw.idx, cpuc->used_mask); idx 426 arch/arm/kernel/perf_event_v6.c int idx = hwc->idx; idx 428 arch/arm/kernel/perf_event_v6.c if (ARMV6_CYCLE_COUNTER == idx) { idx 431 arch/arm/kernel/perf_event_v6.c } else if (ARMV6_COUNTER0 == idx) { idx 434 arch/arm/kernel/perf_event_v6.c } else if (ARMV6_COUNTER1 == idx) { idx 438 arch/arm/kernel/perf_event_v6.c WARN_ONCE(1, "invalid counter number (%d)\n", idx); idx 461 arch/arm/kernel/perf_event_v6.c int idx = hwc->idx; idx 463 arch/arm/kernel/perf_event_v6.c if (ARMV6_CYCLE_COUNTER == idx) { idx 465 arch/arm/kernel/perf_event_v6.c } else if (ARMV6_COUNTER0 == idx) { idx 467 arch/arm/kernel/perf_event_v6.c } else if (ARMV6_COUNTER1 == idx) { idx 470 arch/arm/kernel/perf_event_v6.c WARN_ONCE(1, "invalid counter number (%d)\n", idx); idx 728 arch/arm/kernel/perf_event_v7.c static inline int armv7_pmnc_counter_valid(struct arm_pmu *cpu_pmu, int idx) idx 730 arch/arm/kernel/perf_event_v7.c return idx >= ARMV7_IDX_CYCLE_COUNTER && idx 731 arch/arm/kernel/perf_event_v7.c idx <= ARMV7_IDX_COUNTER_LAST(cpu_pmu); idx 734 arch/arm/kernel/perf_event_v7.c static inline int armv7_pmnc_counter_has_overflowed(u32 pmnc, int idx) idx 736 arch/arm/kernel/perf_event_v7.c return pmnc & BIT(ARMV7_IDX_TO_COUNTER(idx)); idx 739 arch/arm/kernel/perf_event_v7.c static inline void armv7_pmnc_select_counter(int idx) idx 741 arch/arm/kernel/perf_event_v7.c u32 counter = ARMV7_IDX_TO_COUNTER(idx); idx 750 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 753 arch/arm/kernel/perf_event_v7.c if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) { idx 755 arch/arm/kernel/perf_event_v7.c smp_processor_id(), idx); idx 756 arch/arm/kernel/perf_event_v7.c } else if (idx == ARMV7_IDX_CYCLE_COUNTER) { idx 759 arch/arm/kernel/perf_event_v7.c armv7_pmnc_select_counter(idx); idx 770 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 772 arch/arm/kernel/perf_event_v7.c if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) { idx 774 arch/arm/kernel/perf_event_v7.c smp_processor_id(), idx); idx 775 arch/arm/kernel/perf_event_v7.c } else if (idx == ARMV7_IDX_CYCLE_COUNTER) { idx 778 arch/arm/kernel/perf_event_v7.c armv7_pmnc_select_counter(idx); idx 783 arch/arm/kernel/perf_event_v7.c static inline void armv7_pmnc_write_evtsel(int idx, u32 val) idx 785 arch/arm/kernel/perf_event_v7.c armv7_pmnc_select_counter(idx); idx 790 arch/arm/kernel/perf_event_v7.c static inline void armv7_pmnc_enable_counter(int idx) idx 792 arch/arm/kernel/perf_event_v7.c u32 counter = ARMV7_IDX_TO_COUNTER(idx); idx 796 arch/arm/kernel/perf_event_v7.c static inline void armv7_pmnc_disable_counter(int idx) idx 798 arch/arm/kernel/perf_event_v7.c u32 counter = ARMV7_IDX_TO_COUNTER(idx); idx 802 arch/arm/kernel/perf_event_v7.c static inline void armv7_pmnc_enable_intens(int idx) idx 804 arch/arm/kernel/perf_event_v7.c u32 counter = ARMV7_IDX_TO_COUNTER(idx); idx 808 arch/arm/kernel/perf_event_v7.c static inline void armv7_pmnc_disable_intens(int idx) idx 810 arch/arm/kernel/perf_event_v7.c u32 counter = ARMV7_IDX_TO_COUNTER(idx); idx 877 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 879 arch/arm/kernel/perf_event_v7.c if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) { idx 881 arch/arm/kernel/perf_event_v7.c smp_processor_id(), idx); idx 894 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_counter(idx); idx 901 arch/arm/kernel/perf_event_v7.c if (cpu_pmu->set_event_filter || idx != ARMV7_IDX_CYCLE_COUNTER) idx 902 arch/arm/kernel/perf_event_v7.c armv7_pmnc_write_evtsel(idx, hwc->config_base); idx 907 arch/arm/kernel/perf_event_v7.c armv7_pmnc_enable_intens(idx); idx 912 arch/arm/kernel/perf_event_v7.c armv7_pmnc_enable_counter(idx); idx 923 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 925 arch/arm/kernel/perf_event_v7.c if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) { idx 927 arch/arm/kernel/perf_event_v7.c smp_processor_id(), idx); idx 939 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_counter(idx); idx 944 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_intens(idx); idx 955 arch/arm/kernel/perf_event_v7.c int idx; idx 973 arch/arm/kernel/perf_event_v7.c for (idx = 0; idx < cpu_pmu->num_events; ++idx) { idx 974 arch/arm/kernel/perf_event_v7.c struct perf_event *event = cpuc->events[idx]; idx 985 arch/arm/kernel/perf_event_v7.c if (!armv7_pmnc_counter_has_overflowed(pmnc, idx)) idx 1035 arch/arm/kernel/perf_event_v7.c int idx; idx 1052 arch/arm/kernel/perf_event_v7.c for (idx = ARMV7_IDX_COUNTER0; idx < cpu_pmu->num_events; ++idx) { idx 1053 arch/arm/kernel/perf_event_v7.c if (!test_and_set_bit(idx, cpuc->used_mask)) idx 1054 arch/arm/kernel/perf_event_v7.c return idx; idx 1064 arch/arm/kernel/perf_event_v7.c clear_bit(event->hw.idx, cpuc->used_mask); idx 1096 arch/arm/kernel/perf_event_v7.c u32 idx, nb_cnt = cpu_pmu->num_events, val; idx 1105 arch/arm/kernel/perf_event_v7.c for (idx = ARMV7_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) { idx 1106 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_counter(idx); idx 1107 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_intens(idx); idx 1415 arch/arm/kernel/perf_event_v7.c static void krait_evt_setup(int idx, u32 config_base) idx 1437 arch/arm/kernel/perf_event_v7.c armv7_pmnc_write_evtsel(idx, val); idx 1497 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 1505 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_counter(idx); idx 1514 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_intens(idx); idx 1523 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 1534 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_counter(idx); idx 1542 arch/arm/kernel/perf_event_v7.c krait_evt_setup(idx, hwc->config_base); idx 1544 arch/arm/kernel/perf_event_v7.c armv7_pmnc_write_evtsel(idx, hwc->config_base); idx 1547 arch/arm/kernel/perf_event_v7.c armv7_pmnc_enable_intens(idx); idx 1550 arch/arm/kernel/perf_event_v7.c armv7_pmnc_enable_counter(idx); idx 1559 arch/arm/kernel/perf_event_v7.c u32 idx, nb_cnt = cpu_pmu->num_events; idx 1573 arch/arm/kernel/perf_event_v7.c for (idx = ARMV7_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) { idx 1574 arch/arm/kernel/perf_event_v7.c armv7_pmnc_select_counter(idx); idx 1609 arch/arm/kernel/perf_event_v7.c int idx; idx 1630 arch/arm/kernel/perf_event_v7.c idx = armv7pmu_get_event_idx(cpuc, event); idx 1631 arch/arm/kernel/perf_event_v7.c if (idx < 0 && bit >= 0) idx 1634 arch/arm/kernel/perf_event_v7.c return idx; idx 1762 arch/arm/kernel/perf_event_v7.c static void scorpion_evt_setup(int idx, u32 config_base) idx 1784 arch/arm/kernel/perf_event_v7.c armv7_pmnc_write_evtsel(idx, val); idx 1830 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 1838 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_counter(idx); idx 1847 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_intens(idx); idx 1856 arch/arm/kernel/perf_event_v7.c int idx = hwc->idx; idx 1867 arch/arm/kernel/perf_event_v7.c armv7_pmnc_disable_counter(idx); idx 1875 arch/arm/kernel/perf_event_v7.c scorpion_evt_setup(idx, hwc->config_base); idx 1876 arch/arm/kernel/perf_event_v7.c else if (idx != ARMV7_IDX_CYCLE_COUNTER) idx 1877 arch/arm/kernel/perf_event_v7.c armv7_pmnc_write_evtsel(idx, hwc->config_base); idx 1880 arch/arm/kernel/perf_event_v7.c armv7_pmnc_enable_intens(idx); idx 1883 arch/arm/kernel/perf_event_v7.c armv7_pmnc_enable_counter(idx); idx 1892 arch/arm/kernel/perf_event_v7.c u32 idx, nb_cnt = cpu_pmu->num_events; idx 1907 arch/arm/kernel/perf_event_v7.c for (idx = ARMV7_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) { idx 1908 arch/arm/kernel/perf_event_v7.c armv7_pmnc_select_counter(idx); idx 1942 arch/arm/kernel/perf_event_v7.c int idx; idx 1960 arch/arm/kernel/perf_event_v7.c idx = armv7pmu_get_event_idx(cpuc, event); idx 1961 arch/arm/kernel/perf_event_v7.c if (idx < 0 && bit >= 0) idx 1964 arch/arm/kernel/perf_event_v7.c return idx; idx 151 arch/arm/kernel/perf_event_xscale.c int idx; idx 173 arch/arm/kernel/perf_event_xscale.c for (idx = 0; idx < cpu_pmu->num_events; ++idx) { idx 174 arch/arm/kernel/perf_event_xscale.c struct perf_event *event = cpuc->events[idx]; idx 180 arch/arm/kernel/perf_event_xscale.c if (!xscale1_pmnc_counter_has_overflowed(pmnc, idx)) idx 210 arch/arm/kernel/perf_event_xscale.c int idx = hwc->idx; idx 212 arch/arm/kernel/perf_event_xscale.c switch (idx) { idx 228 arch/arm/kernel/perf_event_xscale.c WARN_ONCE(1, "invalid counter number (%d)\n", idx); idx 246 arch/arm/kernel/perf_event_xscale.c int idx = hwc->idx; idx 248 arch/arm/kernel/perf_event_xscale.c switch (idx) { idx 262 arch/arm/kernel/perf_event_xscale.c WARN_ONCE(1, "invalid counter number (%d)\n", idx); idx 298 arch/arm/kernel/perf_event_xscale.c clear_bit(event->hw.idx, cpuc->used_mask); idx 328 arch/arm/kernel/perf_event_xscale.c int counter = hwc->idx; idx 349 arch/arm/kernel/perf_event_xscale.c int counter = hwc->idx; idx 503 arch/arm/kernel/perf_event_xscale.c int idx; idx 519 arch/arm/kernel/perf_event_xscale.c for (idx = 0; idx < cpu_pmu->num_events; ++idx) { idx 520 arch/arm/kernel/perf_event_xscale.c struct perf_event *event = cpuc->events[idx]; idx 526 arch/arm/kernel/perf_event_xscale.c if (!xscale2_pmnc_counter_has_overflowed(of_flags, idx)) idx 556 arch/arm/kernel/perf_event_xscale.c int idx = hwc->idx; idx 561 arch/arm/kernel/perf_event_xscale.c switch (idx) { idx 586 arch/arm/kernel/perf_event_xscale.c WARN_ONCE(1, "invalid counter number (%d)\n", idx); idx 602 arch/arm/kernel/perf_event_xscale.c int idx = hwc->idx; idx 607 arch/arm/kernel/perf_event_xscale.c switch (idx) { idx 637 arch/arm/kernel/perf_event_xscale.c WARN_ONCE(1, "invalid counter number (%d)\n", idx); idx 652 arch/arm/kernel/perf_event_xscale.c int idx = xscale1pmu_get_event_idx(cpuc, event); idx 653 arch/arm/kernel/perf_event_xscale.c if (idx >= 0) idx 657 arch/arm/kernel/perf_event_xscale.c idx = XSCALE_COUNTER3; idx 659 arch/arm/kernel/perf_event_xscale.c idx = XSCALE_COUNTER2; idx 661 arch/arm/kernel/perf_event_xscale.c return idx; idx 691 arch/arm/kernel/perf_event_xscale.c int counter = hwc->idx; idx 718 arch/arm/kernel/perf_event_xscale.c int counter = hwc->idx; idx 11 arch/arm/kernel/perf_regs.c u64 perf_reg_value(struct pt_regs *regs, int idx) idx 13 arch/arm/kernel/perf_regs.c if (WARN_ON_ONCE((u32)idx >= PERF_REG_ARM_MAX)) idx 16 arch/arm/kernel/perf_regs.c return regs->uregs[idx]; idx 367 arch/arm/kernel/ptrace.c static long ptrace_hbp_idx_to_num(int idx) idx 370 arch/arm/kernel/ptrace.c long num = (idx << 1) + 1; idx 462 arch/arm/kernel/ptrace.c int idx, ret = 0; idx 469 arch/arm/kernel/ptrace.c idx = ptrace_hbp_num_to_idx(num); idx 470 arch/arm/kernel/ptrace.c if (idx < 0 || idx >= ARM_MAX_HBP_SLOTS) { idx 475 arch/arm/kernel/ptrace.c bp = tsk->thread.debug.hbp[idx]; idx 507 arch/arm/kernel/ptrace.c int idx, gen_len, gen_type, implied_type, ret = 0; idx 520 arch/arm/kernel/ptrace.c idx = ptrace_hbp_num_to_idx(num); idx 521 arch/arm/kernel/ptrace.c if (idx < 0 || idx >= ARM_MAX_HBP_SLOTS) { idx 531 arch/arm/kernel/ptrace.c bp = tsk->thread.debug.hbp[idx]; idx 538 arch/arm/kernel/ptrace.c tsk->thread.debug.hbp[idx] = bp; idx 368 arch/arm/kernel/signal.c unsigned int idx, thumb = 0; idx 427 arch/arm/kernel/signal.c idx = 6 + thumb * 3; idx 429 arch/arm/kernel/signal.c idx += 5; idx 430 arch/arm/kernel/signal.c if (__put_user(sigreturn_codes[idx], rc ) || idx 431 arch/arm/kernel/signal.c __put_user(sigreturn_codes[idx+1], rc+1) || idx 432 arch/arm/kernel/signal.c __put_user(sigreturn_codes[idx+2], rc+2) || idx 438 arch/arm/kernel/signal.c idx = thumb << 1; idx 440 arch/arm/kernel/signal.c idx += 3; idx 446 arch/arm/kernel/signal.c if (__put_user(sigreturn_codes[idx], rc) || idx 447 arch/arm/kernel/signal.c __put_user(sigreturn_codes[idx+1], rc+1)) idx 461 arch/arm/kernel/signal.c (idx << 2) + thumb; idx 172 arch/arm/kernel/unwind.c const struct unwind_idx *idx = NULL; idx 184 arch/arm/kernel/unwind.c idx = search_index(addr, __start_unwind_idx, idx 195 arch/arm/kernel/unwind.c idx = search_index(addr, table->start, idx 206 arch/arm/kernel/unwind.c pr_debug("%s: idx = %p\n", __func__, idx); idx 207 arch/arm/kernel/unwind.c return idx; idx 378 arch/arm/kernel/unwind.c const struct unwind_idx *idx; idx 391 arch/arm/kernel/unwind.c idx = unwind_find_idx(frame->pc); idx 392 arch/arm/kernel/unwind.c if (!idx) { idx 402 arch/arm/kernel/unwind.c if (idx->insn == 1) idx 405 arch/arm/kernel/unwind.c else if ((idx->insn & 0x80000000) == 0) idx 407 arch/arm/kernel/unwind.c ctrl.insn = (unsigned long *)prel31_to_addr(&idx->insn); idx 408 arch/arm/kernel/unwind.c else if ((idx->insn & 0xff000000) == 0x80000000) idx 410 arch/arm/kernel/unwind.c ctrl.insn = &idx->insn; idx 413 arch/arm/kernel/unwind.c idx->insn, idx); idx 12 arch/arm/kvm/hyp/cp15-sr.c static u64 *cp15_64(struct kvm_cpu_context *ctxt, int idx) idx 14 arch/arm/kvm/hyp/cp15-sr.c return (u64 *)(ctxt->cp15 + idx); idx 523 arch/arm/mach-at91/pm.c int idx = 0; idx 528 arch/arm/mach-at91/pm.c soc_pm.data.ramc[idx] = of_iomap(np, 0); idx 529 arch/arm/mach-at91/pm.c if (!soc_pm.data.ramc[idx]) idx 530 arch/arm/mach-at91/pm.c panic(pr_fmt("unable to map ramc[%d] cpu registers\n"), idx); idx 537 arch/arm/mach-at91/pm.c idx++; idx 540 arch/arm/mach-at91/pm.c if (!idx) idx 85 arch/arm/mach-imx/avic.c int idx = d->hwirq >> 5; idx 87 arch/arm/mach-imx/avic.c avic_saved_mask_reg[idx] = imx_readl(avic_base + ct->regs.mask); idx 107 arch/arm/mach-imx/avic.c int idx = d->hwirq >> 5; idx 109 arch/arm/mach-imx/avic.c imx_writel(avic_saved_mask_reg[idx], avic_base + ct->regs.mask); idx 124 arch/arm/mach-imx/avic.c static __init void avic_init_gc(int idx, unsigned int irq_start) idx 141 arch/arm/mach-imx/avic.c ct->regs.mask = !idx ? AVIC_INTENABLEL : AVIC_INTENABLEH; idx 91 arch/arm/mach-imx/gpc.c unsigned int idx = d->hwirq / 32; idx 95 arch/arm/mach-imx/gpc.c gpc_wake_irqs[idx] = on ? gpc_wake_irqs[idx] | mask : idx 96 arch/arm/mach-imx/gpc.c gpc_wake_irqs[idx] & ~mask; idx 76 arch/arm/mach-imx/tzic.c int idx = d->hwirq >> 5; idx 78 arch/arm/mach-imx/tzic.c imx_writel(gc->wake_active, tzic_base + TZIC_WAKEUP0(idx)); idx 83 arch/arm/mach-imx/tzic.c int idx = d->hwirq >> 5; idx 85 arch/arm/mach-imx/tzic.c imx_writel(imx_readl(tzic_base + TZIC_ENSET0(idx)), idx 86 arch/arm/mach-imx/tzic.c tzic_base + TZIC_WAKEUP0(idx)); idx 100 arch/arm/mach-imx/tzic.c static __init void tzic_init_gc(int idx, unsigned int irq_start) idx 116 arch/arm/mach-imx/tzic.c ct->regs.disable = TZIC_ENCLEAR0(idx); idx 117 arch/arm/mach-imx/tzic.c ct->regs.enable = TZIC_ENSET0(idx); idx 50 arch/arm/mach-omap2/cm2xxx_3xxx.h static inline u32 omap2_cm_read_mod_reg(s16 module, u16 idx) idx 52 arch/arm/mach-omap2/cm2xxx_3xxx.h return readl_relaxed(cm_base.va + module + idx); idx 55 arch/arm/mach-omap2/cm2xxx_3xxx.h static inline void omap2_cm_write_mod_reg(u32 val, s16 module, u16 idx) idx 57 arch/arm/mach-omap2/cm2xxx_3xxx.h writel_relaxed(val, cm_base.va + module + idx); idx 62 arch/arm/mach-omap2/cm2xxx_3xxx.h s16 idx) idx 66 arch/arm/mach-omap2/cm2xxx_3xxx.h v = omap2_cm_read_mod_reg(module, idx); idx 69 arch/arm/mach-omap2/cm2xxx_3xxx.h omap2_cm_write_mod_reg(v, module, idx); idx 75 arch/arm/mach-omap2/cm2xxx_3xxx.h static inline u32 omap2_cm_read_mod_bits_shift(s16 domain, s16 idx, u32 mask) idx 79 arch/arm/mach-omap2/cm2xxx_3xxx.h v = omap2_cm_read_mod_reg(domain, idx); idx 86 arch/arm/mach-omap2/cm2xxx_3xxx.h static inline u32 omap2_cm_set_mod_reg_bits(u32 bits, s16 module, s16 idx) idx 88 arch/arm/mach-omap2/cm2xxx_3xxx.h return omap2_cm_rmw_mod_reg_bits(bits, bits, module, idx); idx 91 arch/arm/mach-omap2/cm2xxx_3xxx.h static inline u32 omap2_cm_clear_mod_reg_bits(u32 bits, s16 module, s16 idx) idx 93 arch/arm/mach-omap2/cm2xxx_3xxx.h return omap2_cm_rmw_mod_reg_bits(bits, 0x0, module, idx); idx 51 arch/arm/mach-omap2/cm33xx.c static inline u32 am33xx_cm_read_reg(u16 inst, u16 idx) idx 53 arch/arm/mach-omap2/cm33xx.c return readl_relaxed(cm_base.va + inst + idx); idx 57 arch/arm/mach-omap2/cm33xx.c static inline void am33xx_cm_write_reg(u32 val, u16 inst, u16 idx) idx 59 arch/arm/mach-omap2/cm33xx.c writel_relaxed(val, cm_base.va + inst + idx); idx 63 arch/arm/mach-omap2/cm33xx.c static inline u32 am33xx_cm_rmw_reg_bits(u32 mask, u32 bits, s16 inst, s16 idx) idx 67 arch/arm/mach-omap2/cm33xx.c v = am33xx_cm_read_reg(inst, idx); idx 70 arch/arm/mach-omap2/cm33xx.c am33xx_cm_write_reg(v, inst, idx); idx 75 arch/arm/mach-omap2/cm33xx.c static inline u32 am33xx_cm_read_reg_bits(u16 inst, s16 idx, u32 mask) idx 79 arch/arm/mach-omap2/cm33xx.c v = am33xx_cm_read_reg(inst, idx); idx 74 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx); idx 113 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx) idx 118 arch/arm/mach-omap2/cminst44xx.c return readl_relaxed(_cm_bases[part].va + inst + idx); idx 122 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_write_inst_reg(u32 val, u8 part, u16 inst, u16 idx) idx 127 arch/arm/mach-omap2/cminst44xx.c writel_relaxed(val, _cm_bases[part].va + inst + idx); idx 132 arch/arm/mach-omap2/cminst44xx.c s16 idx) idx 136 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, idx); idx 139 arch/arm/mach-omap2/cminst44xx.c omap4_cminst_write_inst_reg(v, part, inst, idx); idx 144 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_set_inst_reg_bits(u32 bits, u8 part, u16 inst, s16 idx) idx 146 arch/arm/mach-omap2/cminst44xx.c return omap4_cminst_rmw_inst_reg_bits(bits, bits, part, inst, idx); idx 150 arch/arm/mach-omap2/cminst44xx.c s16 idx) idx 152 arch/arm/mach-omap2/cminst44xx.c return omap4_cminst_rmw_inst_reg_bits(bits, 0x0, part, inst, idx); idx 155 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg_bits(u8 part, u16 inst, s16 idx, u32 mask) idx 159 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, idx); idx 170 arch/arm/mach-omap2/cpuidle34xx.c int idx; idx 193 arch/arm/mach-omap2/cpuidle34xx.c for (idx = index - 1; idx >= 0; idx--) { idx 194 arch/arm/mach-omap2/cpuidle34xx.c cx = &omap3_idle_data[idx]; idx 197 arch/arm/mach-omap2/cpuidle34xx.c next_index = idx; idx 33 arch/arm/mach-omap2/omap-secure.c u32 omap_secure_dispatcher(u32 idx, u32 flag, u32 nargs, u32 arg1, u32 arg2, idx 51 arch/arm/mach-omap2/omap-secure.c ret = omap_smc2(idx, flag, __pa(param)); idx 106 arch/arm/mach-omap2/omap-secure.c u32 rx51_secure_dispatcher(u32 idx, u32 process, u32 flag, u32 nargs, idx 126 arch/arm/mach-omap2/omap-secure.c ret = omap_smc3(idx, process, flag, __pa(param)); idx 61 arch/arm/mach-omap2/omap-secure.h extern u32 omap_secure_dispatcher(u32 idx, u32 flag, u32 nargs, idx 70 arch/arm/mach-omap2/omap-secure.h extern u32 rx51_secure_dispatcher(u32 idx, u32 process, u32 flag, u32 nargs, idx 75 arch/arm/mach-omap2/omap-wakeupgen.c static inline u32 wakeupgen_readl(u8 idx, u32 cpu) idx 78 arch/arm/mach-omap2/omap-wakeupgen.c (cpu * CPU_ENA_OFFSET) + (idx * 4)); idx 81 arch/arm/mach-omap2/omap-wakeupgen.c static inline void wakeupgen_writel(u32 val, u8 idx, u32 cpu) idx 84 arch/arm/mach-omap2/omap-wakeupgen.c (cpu * CPU_ENA_OFFSET) + (idx * 4)); idx 87 arch/arm/mach-omap2/omap-wakeupgen.c static inline void sar_writel(u32 val, u32 offset, u8 idx) idx 89 arch/arm/mach-omap2/omap-wakeupgen.c writel_relaxed(val, sar_base + offset + (idx * 4)); idx 27 arch/arm/mach-omap2/prcm_mpu_44xx_54xx.h extern u32 omap4_prcm_mpu_read_inst_reg(s16 inst, u16 idx); idx 28 arch/arm/mach-omap2/prcm_mpu_44xx_54xx.h extern void omap4_prcm_mpu_write_inst_reg(u32 val, s16 inst, u16 idx); idx 30 arch/arm/mach-omap2/prcm_mpu_44xx_54xx.h s16 idx); idx 138 arch/arm/mach-omap2/prm.h bool (*was_any_context_lost_old)(u8 part, s16 inst, u16 idx); idx 139 arch/arm/mach-omap2/prm.h void (*clear_context_loss_flags_old)(u8 part, s16 inst, u16 idx); idx 160 arch/arm/mach-omap2/prm.h extern bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx); idx 161 arch/arm/mach-omap2/prm.h extern void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx); idx 53 arch/arm/mach-omap2/prm2xxx_3xxx.h static inline u32 omap2_prm_read_mod_reg(s16 module, u16 idx) idx 55 arch/arm/mach-omap2/prm2xxx_3xxx.h return readl_relaxed(prm_base.va + module + idx); idx 58 arch/arm/mach-omap2/prm2xxx_3xxx.h static inline void omap2_prm_write_mod_reg(u32 val, s16 module, u16 idx) idx 60 arch/arm/mach-omap2/prm2xxx_3xxx.h writel_relaxed(val, prm_base.va + module + idx); idx 65 arch/arm/mach-omap2/prm2xxx_3xxx.h s16 idx) idx 69 arch/arm/mach-omap2/prm2xxx_3xxx.h v = omap2_prm_read_mod_reg(module, idx); idx 72 arch/arm/mach-omap2/prm2xxx_3xxx.h omap2_prm_write_mod_reg(v, module, idx); idx 78 arch/arm/mach-omap2/prm2xxx_3xxx.h static inline u32 omap2_prm_read_mod_bits_shift(s16 domain, s16 idx, u32 mask) idx 82 arch/arm/mach-omap2/prm2xxx_3xxx.h v = omap2_prm_read_mod_reg(domain, idx); idx 89 arch/arm/mach-omap2/prm2xxx_3xxx.h static inline u32 omap2_prm_set_mod_reg_bits(u32 bits, s16 module, s16 idx) idx 91 arch/arm/mach-omap2/prm2xxx_3xxx.h return omap2_prm_rmw_mod_reg_bits(bits, bits, module, idx); idx 94 arch/arm/mach-omap2/prm2xxx_3xxx.h static inline u32 omap2_prm_clear_mod_reg_bits(u32 bits, s16 module, s16 idx) idx 96 arch/arm/mach-omap2/prm2xxx_3xxx.h return omap2_prm_rmw_mod_reg_bits(bits, 0x0, module, idx); idx 31 arch/arm/mach-omap2/prm33xx.c static u32 am33xx_prm_read_reg(s16 inst, u16 idx) idx 33 arch/arm/mach-omap2/prm33xx.c return readl_relaxed(prm_base.va + inst + idx); idx 37 arch/arm/mach-omap2/prm33xx.c static void am33xx_prm_write_reg(u32 val, s16 inst, u16 idx) idx 39 arch/arm/mach-omap2/prm33xx.c writel_relaxed(val, prm_base.va + inst + idx); idx 43 arch/arm/mach-omap2/prm33xx.c static u32 am33xx_prm_rmw_reg_bits(u32 mask, u32 bits, s16 inst, s16 idx) idx 47 arch/arm/mach-omap2/prm33xx.c v = am33xx_prm_read_reg(inst, idx); idx 50 arch/arm/mach-omap2/prm33xx.c am33xx_prm_write_reg(v, inst, idx); idx 404 arch/arm/mach-omap2/prm44xx.c static bool omap44xx_prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx) idx 406 arch/arm/mach-omap2/prm44xx.c return (omap4_prminst_read_inst_reg(part, inst, idx)) ? 1 : 0; idx 420 arch/arm/mach-omap2/prm44xx.c u16 idx) idx 422 arch/arm/mach-omap2/prm44xx.c omap4_prminst_write_inst_reg(0xffffffff, part, inst, idx); idx 394 arch/arm/mach-omap2/prm_common.c bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx) idx 399 arch/arm/mach-omap2/prm_common.c ret = prm_ll_data->was_any_context_lost_old(part, inst, idx); idx 418 arch/arm/mach-omap2/prm_common.c void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx) idx 421 arch/arm/mach-omap2/prm_common.c prm_ll_data->clear_context_loss_flags_old(part, inst, idx); idx 58 arch/arm/mach-omap2/prminst44xx.c u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx) idx 63 arch/arm/mach-omap2/prminst44xx.c return readl_relaxed(_prm_bases[part].va + inst + idx); idx 67 arch/arm/mach-omap2/prminst44xx.c void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx) idx 72 arch/arm/mach-omap2/prminst44xx.c writel_relaxed(val, _prm_bases[part].va + inst + idx); idx 77 arch/arm/mach-omap2/prminst44xx.c u16 idx) idx 81 arch/arm/mach-omap2/prminst44xx.c v = omap4_prminst_read_inst_reg(part, inst, idx); idx 84 arch/arm/mach-omap2/prminst44xx.c omap4_prminst_write_inst_reg(v, part, inst, idx); idx 20 arch/arm/mach-omap2/prminst44xx.h extern u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx); idx 21 arch/arm/mach-omap2/prminst44xx.h extern void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx); idx 23 arch/arm/mach-omap2/prminst44xx.h s16 inst, u16 idx); idx 41 arch/arm/mach-shmobile/setup-rcar-gen2.c int idx = 0; idx 48 arch/arm/mach-shmobile/setup-rcar-gen2.c idx = of_property_match_string(cpg, "clock-names", match->data); idx 49 arch/arm/mach-shmobile/setup-rcar-gen2.c extal = of_parse_phandle(cpg, "clocks", idx); idx 279 arch/arm/mach-vexpress/spc.c int idx, max_opp = info->num_opps[cluster]; idx 284 arch/arm/mach-vexpress/spc.c for (idx = 0; idx < max_opp; idx++, opps++) { idx 302 arch/arm/mach-vexpress/spc.c int idx, max_opp = info->num_opps[cluster]; idx 305 arch/arm/mach-vexpress/spc.c for (idx = 0; idx < max_opp; idx++, opps++) idx 308 arch/arm/mach-vexpress/spc.c return (idx == max_opp) ? -EINVAL : idx; idx 403 arch/arm/mach-vexpress/spc.c uint32_t data = 0, off, ret, idx; idx 413 arch/arm/mach-vexpress/spc.c for (idx = 0; idx < MAX_OPPS; idx++, off += 4, opps++) { idx 422 arch/arm/mach-vexpress/spc.c info->num_opps[cluster] = idx; idx 430 arch/arm/mach-vexpress/spc.c int idx, ret = 0, max_opp; idx 439 arch/arm/mach-vexpress/spc.c for (idx = 0; idx < max_opp; idx++, opps++) { idx 64 arch/arm/mm/cache-l2x0-pmu.c static void l2x0_pmu_counter_config_write(int idx, u32 val) idx 66 arch/arm/mm/cache-l2x0-pmu.c writel_relaxed(val, l2x0_base + L2X0_EVENT_CNT0_CFG - 4 * idx); idx 69 arch/arm/mm/cache-l2x0-pmu.c static u32 l2x0_pmu_counter_read(int idx) idx 71 arch/arm/mm/cache-l2x0-pmu.c return readl_relaxed(l2x0_base + L2X0_EVENT_CNT0_VAL - 4 * idx); idx 74 arch/arm/mm/cache-l2x0-pmu.c static void l2x0_pmu_counter_write(int idx, u32 val) idx 76 arch/arm/mm/cache-l2x0-pmu.c writel_relaxed(val, l2x0_base + L2X0_EVENT_CNT0_VAL - 4 * idx); idx 124 arch/arm/mm/cache-l2x0-pmu.c new_count = l2x0_pmu_counter_read(hw->idx); idx 148 arch/arm/mm/cache-l2x0-pmu.c l2x0_pmu_counter_write(hw->idx, 0); idx 177 arch/arm/mm/cache-l2x0-pmu.c static void __l2x0_pmu_event_enable(int idx, u32 event) idx 183 arch/arm/mm/cache-l2x0-pmu.c l2x0_pmu_counter_config_write(idx, val); idx 200 arch/arm/mm/cache-l2x0-pmu.c __l2x0_pmu_event_enable(hw->idx, hw->config_base); idx 203 arch/arm/mm/cache-l2x0-pmu.c static void __l2x0_pmu_event_disable(int idx) idx 209 arch/arm/mm/cache-l2x0-pmu.c l2x0_pmu_counter_config_write(idx, val); idx 219 arch/arm/mm/cache-l2x0-pmu.c __l2x0_pmu_event_disable(hw->idx); idx 232 arch/arm/mm/cache-l2x0-pmu.c int idx = l2x0_pmu_find_idx(); idx 234 arch/arm/mm/cache-l2x0-pmu.c if (idx == -1) idx 246 arch/arm/mm/cache-l2x0-pmu.c events[idx] = event; idx 247 arch/arm/mm/cache-l2x0-pmu.c hw->idx = idx; idx 265 arch/arm/mm/cache-l2x0-pmu.c events[hw->idx] = NULL; idx 266 arch/arm/mm/cache-l2x0-pmu.c hw->idx = -1; idx 18 arch/arm/mm/highmem.c static inline void set_fixmap_pte(int idx, pte_t pte) idx 20 arch/arm/mm/highmem.c unsigned long vaddr = __fix_to_virt(idx); idx 54 arch/arm/mm/highmem.c unsigned int idx; idx 79 arch/arm/mm/highmem.c idx = FIX_KMAP_BEGIN + type + KM_TYPE_NR * smp_processor_id(); idx 80 arch/arm/mm/highmem.c vaddr = __fix_to_virt(idx); idx 93 arch/arm/mm/highmem.c set_fixmap_pte(idx, mk_pte(page, kmap_prot)); idx 102 arch/arm/mm/highmem.c int idx, type; idx 106 arch/arm/mm/highmem.c idx = FIX_KMAP_BEGIN + type + KM_TYPE_NR * smp_processor_id(); idx 111 arch/arm/mm/highmem.c BUG_ON(vaddr != __fix_to_virt(idx)); idx 112 arch/arm/mm/highmem.c set_fixmap_pte(idx, __pte(0)); idx 114 arch/arm/mm/highmem.c (void) idx; /* to kill a warning */ idx 129 arch/arm/mm/highmem.c int idx, type; idx 138 arch/arm/mm/highmem.c idx = FIX_KMAP_BEGIN + type + KM_TYPE_NR * smp_processor_id(); idx 139 arch/arm/mm/highmem.c vaddr = __fix_to_virt(idx); idx 143 arch/arm/mm/highmem.c set_fixmap_pte(idx, pfn_pte(pfn, kmap_prot)); idx 406 arch/arm/mm/mmu.c void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot) idx 408 arch/arm/mm/mmu.c unsigned long vaddr = __fix_to_virt(idx); idx 414 arch/arm/mm/mmu.c BUG_ON(idx >= __end_of_fixed_addresses); idx 190 arch/arm/net/bpf_jit_32.c unsigned int idx; idx 225 arch/arm/net/bpf_jit_32.c ctx->target[ctx->idx] = inst; idx 227 arch/arm/net/bpf_jit_32.c ctx->idx++; idx 383 arch/arm/net/bpf_jit_32.c imm = offset - (8 + ctx->idx * 4); idx 460 arch/arm/net/bpf_jit_32.c from = ctx->idx; idx 1156 arch/arm/net/bpf_jit_32.c const int idx0 = ctx->idx; idx 1157 arch/arm/net/bpf_jit_32.c #define cur_offset (ctx->idx - idx0) idx 1838 arch/arm/net/bpf_jit_32.c ctx->offsets[i] = ctx->idx; idx 1843 arch/arm/net/bpf_jit_32.c ctx->offsets[i] = ctx->idx; idx 1856 arch/arm/net/bpf_jit_32.c for (i = 0; i < ctx->idx; i++) { idx 1931 arch/arm/net/bpf_jit_32.c tmp_idx = ctx.idx; idx 1933 arch/arm/net/bpf_jit_32.c ctx.prologue_bytes = (ctx.idx - tmp_idx) * 4; idx 1935 arch/arm/net/bpf_jit_32.c ctx.epilogue_offset = ctx.idx; idx 1938 arch/arm/net/bpf_jit_32.c tmp_idx = ctx.idx; idx 1940 arch/arm/net/bpf_jit_32.c ctx.epilogue_bytes = (ctx.idx - tmp_idx) * 4; idx 1942 arch/arm/net/bpf_jit_32.c ctx.idx += ctx.imm_count; idx 1962 arch/arm/net/bpf_jit_32.c image_size = sizeof(u32) * ctx.idx; idx 1977 arch/arm/net/bpf_jit_32.c ctx.idx = 0; idx 1999 arch/arm/net/bpf_jit_32.c flush_icache_range((u32)header, (u32)(ctx.target + ctx.idx)); idx 341 arch/arm/xen/enlighten.c xatp.idx = 0; idx 44 arch/arm64/include/asm/barrier.h static inline unsigned long array_index_mask_nospec(unsigned long idx, idx 53 arch/arm64/include/asm/barrier.h : "r" (idx), "Ir" (sz) idx 105 arch/arm64/include/asm/fixmap.h #define __late_clear_fixmap(idx) __set_fixmap((idx), 0, FIXMAP_PAGE_CLEAR) idx 107 arch/arm64/include/asm/fixmap.h extern void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot); idx 83 arch/arm64/kernel/cacheinfo.c unsigned int level, idx; idx 88 arch/arm64/kernel/cacheinfo.c for (idx = 0, level = 1; level <= this_cpu_ci->num_levels && idx 89 arch/arm64/kernel/cacheinfo.c idx < this_cpu_ci->num_leaves; idx++, level++) { idx 355 arch/arm64/kernel/perf_event.c int idx = event->hw.idx; idx 357 arch/arm64/kernel/perf_event.c return !WARN_ON(idx < 0) && idx 359 arch/arm64/kernel/perf_event.c (idx != ARMV8_IDX_CYCLE_COUNTER); idx 389 arch/arm64/kernel/perf_event.c static inline int armv8pmu_counter_valid(struct arm_pmu *cpu_pmu, int idx) idx 391 arch/arm64/kernel/perf_event.c return idx >= ARMV8_IDX_CYCLE_COUNTER && idx 392 arch/arm64/kernel/perf_event.c idx <= ARMV8_IDX_COUNTER_LAST(cpu_pmu); idx 395 arch/arm64/kernel/perf_event.c static inline int armv8pmu_counter_has_overflowed(u32 pmnc, int idx) idx 397 arch/arm64/kernel/perf_event.c return pmnc & BIT(ARMV8_IDX_TO_COUNTER(idx)); idx 400 arch/arm64/kernel/perf_event.c static inline void armv8pmu_select_counter(int idx) idx 402 arch/arm64/kernel/perf_event.c u32 counter = ARMV8_IDX_TO_COUNTER(idx); idx 407 arch/arm64/kernel/perf_event.c static inline u32 armv8pmu_read_evcntr(int idx) idx 409 arch/arm64/kernel/perf_event.c armv8pmu_select_counter(idx); idx 415 arch/arm64/kernel/perf_event.c int idx = event->hw.idx; idx 418 arch/arm64/kernel/perf_event.c val = armv8pmu_read_evcntr(idx); idx 420 arch/arm64/kernel/perf_event.c val = (val << 32) | armv8pmu_read_evcntr(idx - 1); idx 428 arch/arm64/kernel/perf_event.c int idx = hwc->idx; idx 431 arch/arm64/kernel/perf_event.c if (!armv8pmu_counter_valid(cpu_pmu, idx)) idx 433 arch/arm64/kernel/perf_event.c smp_processor_id(), idx); idx 434 arch/arm64/kernel/perf_event.c else if (idx == ARMV8_IDX_CYCLE_COUNTER) idx 442 arch/arm64/kernel/perf_event.c static inline void armv8pmu_write_evcntr(int idx, u32 value) idx 444 arch/arm64/kernel/perf_event.c armv8pmu_select_counter(idx); idx 451 arch/arm64/kernel/perf_event.c int idx = event->hw.idx; idx 454 arch/arm64/kernel/perf_event.c armv8pmu_write_evcntr(idx, upper_32_bits(value)); idx 455 arch/arm64/kernel/perf_event.c armv8pmu_write_evcntr(idx - 1, lower_32_bits(value)); idx 457 arch/arm64/kernel/perf_event.c armv8pmu_write_evcntr(idx, value); idx 465 arch/arm64/kernel/perf_event.c int idx = hwc->idx; idx 467 arch/arm64/kernel/perf_event.c if (!armv8pmu_counter_valid(cpu_pmu, idx)) idx 469 arch/arm64/kernel/perf_event.c smp_processor_id(), idx); idx 470 arch/arm64/kernel/perf_event.c else if (idx == ARMV8_IDX_CYCLE_COUNTER) { idx 484 arch/arm64/kernel/perf_event.c static inline void armv8pmu_write_evtype(int idx, u32 val) idx 486 arch/arm64/kernel/perf_event.c armv8pmu_select_counter(idx); idx 494 arch/arm64/kernel/perf_event.c int idx = hwc->idx; idx 505 arch/arm64/kernel/perf_event.c armv8pmu_write_evtype(idx - 1, hwc->config_base); idx 506 arch/arm64/kernel/perf_event.c armv8pmu_write_evtype(idx, chain_evt); idx 508 arch/arm64/kernel/perf_event.c armv8pmu_write_evtype(idx, hwc->config_base); idx 512 arch/arm64/kernel/perf_event.c static inline int armv8pmu_enable_counter(int idx) idx 514 arch/arm64/kernel/perf_event.c u32 counter = ARMV8_IDX_TO_COUNTER(idx); idx 516 arch/arm64/kernel/perf_event.c return idx; idx 522 arch/arm64/kernel/perf_event.c int idx = event->hw.idx; idx 523 arch/arm64/kernel/perf_event.c u32 counter_bits = BIT(ARMV8_IDX_TO_COUNTER(idx)); idx 526 arch/arm64/kernel/perf_event.c counter_bits |= BIT(ARMV8_IDX_TO_COUNTER(idx - 1)); idx 532 arch/arm64/kernel/perf_event.c armv8pmu_enable_counter(idx); idx 534 arch/arm64/kernel/perf_event.c armv8pmu_enable_counter(idx - 1); idx 538 arch/arm64/kernel/perf_event.c static inline int armv8pmu_disable_counter(int idx) idx 540 arch/arm64/kernel/perf_event.c u32 counter = ARMV8_IDX_TO_COUNTER(idx); idx 542 arch/arm64/kernel/perf_event.c return idx; idx 549 arch/arm64/kernel/perf_event.c int idx = hwc->idx; idx 550 arch/arm64/kernel/perf_event.c u32 counter_bits = BIT(ARMV8_IDX_TO_COUNTER(idx)); idx 553 arch/arm64/kernel/perf_event.c counter_bits |= BIT(ARMV8_IDX_TO_COUNTER(idx - 1)); idx 560 arch/arm64/kernel/perf_event.c armv8pmu_disable_counter(idx - 1); idx 561 arch/arm64/kernel/perf_event.c armv8pmu_disable_counter(idx); idx 565 arch/arm64/kernel/perf_event.c static inline int armv8pmu_enable_intens(int idx) idx 567 arch/arm64/kernel/perf_event.c u32 counter = ARMV8_IDX_TO_COUNTER(idx); idx 569 arch/arm64/kernel/perf_event.c return idx; idx 574 arch/arm64/kernel/perf_event.c return armv8pmu_enable_intens(event->hw.idx); idx 577 arch/arm64/kernel/perf_event.c static inline int armv8pmu_disable_intens(int idx) idx 579 arch/arm64/kernel/perf_event.c u32 counter = ARMV8_IDX_TO_COUNTER(idx); idx 586 arch/arm64/kernel/perf_event.c return idx; idx 591 arch/arm64/kernel/perf_event.c return armv8pmu_disable_intens(event->hw.idx); idx 695 arch/arm64/kernel/perf_event.c int idx; idx 718 arch/arm64/kernel/perf_event.c for (idx = 0; idx < cpu_pmu->num_events; ++idx) { idx 719 arch/arm64/kernel/perf_event.c struct perf_event *event = cpuc->events[idx]; idx 730 arch/arm64/kernel/perf_event.c if (!armv8pmu_counter_has_overflowed(pmovsr, idx)) idx 759 arch/arm64/kernel/perf_event.c int idx; idx 761 arch/arm64/kernel/perf_event.c for (idx = ARMV8_IDX_COUNTER0; idx < cpu_pmu->num_events; idx ++) { idx 762 arch/arm64/kernel/perf_event.c if (!test_and_set_bit(idx, cpuc->used_mask)) idx 763 arch/arm64/kernel/perf_event.c return idx; idx 771 arch/arm64/kernel/perf_event.c int idx; idx 777 arch/arm64/kernel/perf_event.c for (idx = ARMV8_IDX_COUNTER0 + 1; idx < cpu_pmu->num_events; idx += 2) { idx 778 arch/arm64/kernel/perf_event.c if (!test_and_set_bit(idx, cpuc->used_mask)) { idx 780 arch/arm64/kernel/perf_event.c if (!test_and_set_bit(idx - 1, cpuc->used_mask)) idx 781 arch/arm64/kernel/perf_event.c return idx; idx 783 arch/arm64/kernel/perf_event.c clear_bit(idx, cpuc->used_mask); idx 814 arch/arm64/kernel/perf_event.c int idx = event->hw.idx; idx 816 arch/arm64/kernel/perf_event.c clear_bit(idx, cpuc->used_mask); idx 818 arch/arm64/kernel/perf_event.c clear_bit(idx - 1, cpuc->used_mask); idx 877 arch/arm64/kernel/perf_event.c u32 idx, nb_cnt = cpu_pmu->num_events; idx 880 arch/arm64/kernel/perf_event.c for (idx = ARMV8_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) { idx 881 arch/arm64/kernel/perf_event.c armv8pmu_disable_counter(idx); idx 882 arch/arm64/kernel/perf_event.c armv8pmu_disable_intens(idx); idx 12 arch/arm64/kernel/perf_regs.c u64 perf_reg_value(struct pt_regs *regs, int idx) idx 14 arch/arm64/kernel/perf_regs.c if (WARN_ON_ONCE((u32)idx >= PERF_REG_ARM64_MAX)) idx 23 arch/arm64/kernel/perf_regs.c if ((u32)idx == PERF_REG_ARM64_SP) idx 25 arch/arm64/kernel/perf_regs.c if ((u32)idx == PERF_REG_ARM64_LR) idx 29 arch/arm64/kernel/perf_regs.c if ((u32)idx == PERF_REG_ARM64_SP) idx 32 arch/arm64/kernel/perf_regs.c if ((u32)idx == PERF_REG_ARM64_PC) idx 35 arch/arm64/kernel/perf_regs.c return regs->regs[idx]; idx 236 arch/arm64/kernel/ptrace.c unsigned long idx) idx 242 arch/arm64/kernel/ptrace.c if (idx >= ARM_MAX_BRP) idx 244 arch/arm64/kernel/ptrace.c idx = array_index_nospec(idx, ARM_MAX_BRP); idx 245 arch/arm64/kernel/ptrace.c bp = tsk->thread.debug.hbp_break[idx]; idx 248 arch/arm64/kernel/ptrace.c if (idx >= ARM_MAX_WRP) idx 250 arch/arm64/kernel/ptrace.c idx = array_index_nospec(idx, ARM_MAX_WRP); idx 251 arch/arm64/kernel/ptrace.c bp = tsk->thread.debug.hbp_watch[idx]; idx 261 arch/arm64/kernel/ptrace.c unsigned long idx, idx 268 arch/arm64/kernel/ptrace.c if (idx >= ARM_MAX_BRP) idx 270 arch/arm64/kernel/ptrace.c idx = array_index_nospec(idx, ARM_MAX_BRP); idx 271 arch/arm64/kernel/ptrace.c tsk->thread.debug.hbp_break[idx] = bp; idx 275 arch/arm64/kernel/ptrace.c if (idx >= ARM_MAX_WRP) idx 277 arch/arm64/kernel/ptrace.c idx = array_index_nospec(idx, ARM_MAX_WRP); idx 278 arch/arm64/kernel/ptrace.c tsk->thread.debug.hbp_watch[idx] = bp; idx 289 arch/arm64/kernel/ptrace.c unsigned long idx) idx 321 arch/arm64/kernel/ptrace.c err = ptrace_hbp_set_event(note_type, tsk, idx, bp); idx 388 arch/arm64/kernel/ptrace.c unsigned long idx, idx 391 arch/arm64/kernel/ptrace.c struct perf_event *bp = ptrace_hbp_get_event(note_type, tsk, idx); idx 402 arch/arm64/kernel/ptrace.c unsigned long idx, idx 405 arch/arm64/kernel/ptrace.c struct perf_event *bp = ptrace_hbp_get_event(note_type, tsk, idx); idx 416 arch/arm64/kernel/ptrace.c unsigned long idx) idx 418 arch/arm64/kernel/ptrace.c struct perf_event *bp = ptrace_hbp_get_event(note_type, tsk, idx); idx 421 arch/arm64/kernel/ptrace.c bp = ptrace_hbp_create(note_type, tsk, idx); idx 428 arch/arm64/kernel/ptrace.c unsigned long idx, idx 436 arch/arm64/kernel/ptrace.c bp = ptrace_hbp_get_initialised_bp(note_type, tsk, idx); idx 453 arch/arm64/kernel/ptrace.c unsigned long idx, idx 460 arch/arm64/kernel/ptrace.c bp = ptrace_hbp_get_initialised_bp(note_type, tsk, idx); idx 482 arch/arm64/kernel/ptrace.c int ret, idx = 0, offset, limit; idx 507 arch/arm64/kernel/ptrace.c ret = ptrace_hbp_get_addr(note_type, target, idx, &addr); idx 516 arch/arm64/kernel/ptrace.c ret = ptrace_hbp_get_ctrl(note_type, target, idx, &ctrl); idx 531 arch/arm64/kernel/ptrace.c idx++; idx 543 arch/arm64/kernel/ptrace.c int ret, idx = 0, offset, limit; idx 562 arch/arm64/kernel/ptrace.c ret = ptrace_hbp_set_addr(note_type, target, idx, addr); idx 573 arch/arm64/kernel/ptrace.c ret = ptrace_hbp_set_ctrl(note_type, target, idx, ctrl); idx 584 arch/arm64/kernel/ptrace.c idx++; idx 1259 arch/arm64/kernel/ptrace.c unsigned int idx = start + i; idx 1262 arch/arm64/kernel/ptrace.c switch (idx) { idx 1274 arch/arm64/kernel/ptrace.c reg = task_pt_regs(target)->regs[idx]; idx 1315 arch/arm64/kernel/ptrace.c unsigned int idx = start + i; idx 1331 arch/arm64/kernel/ptrace.c switch (idx) { idx 1343 arch/arm64/kernel/ptrace.c newregs.regs[idx] = reg; idx 1622 arch/arm64/kernel/ptrace.c int err, idx = compat_ptrace_hbp_num_to_idx(num); idx 1625 arch/arm64/kernel/ptrace.c err = ptrace_hbp_get_addr(note_type, tsk, idx, &addr); idx 1628 arch/arm64/kernel/ptrace.c err = ptrace_hbp_get_ctrl(note_type, tsk, idx, &ctrl); idx 1643 arch/arm64/kernel/ptrace.c int err, idx = compat_ptrace_hbp_num_to_idx(num); idx 1647 arch/arm64/kernel/ptrace.c err = ptrace_hbp_set_addr(note_type, tsk, idx, addr); idx 1650 arch/arm64/kernel/ptrace.c err = ptrace_hbp_set_ctrl(note_type, tsk, idx, ctrl); idx 369 arch/arm64/kernel/signal32.c unsigned int idx = thumb << 1; idx 372 arch/arm64/kernel/signal32.c idx += 3; idx 375 arch/arm64/kernel/signal32.c (idx << 2) + thumb; idx 56 arch/arm64/kvm/pmu.c #define PMEVTYPER_READ_CASE(idx) \ idx 57 arch/arm64/kvm/pmu.c case idx: \ idx 58 arch/arm64/kvm/pmu.c return read_sysreg(pmevtyper##idx##_el0) idx 60 arch/arm64/kvm/pmu.c #define PMEVTYPER_WRITE_CASE(idx) \ idx 61 arch/arm64/kvm/pmu.c case idx: \ idx 62 arch/arm64/kvm/pmu.c write_sysreg(val, pmevtyper##idx##_el0); \ idx 102 arch/arm64/kvm/pmu.c static u64 kvm_vcpu_pmu_read_evtype_direct(int idx) idx 104 arch/arm64/kvm/pmu.c switch (idx) { idx 119 arch/arm64/kvm/pmu.c static void kvm_vcpu_pmu_write_evtype_direct(int idx, u32 val) idx 121 arch/arm64/kvm/pmu.c switch (idx) { idx 744 arch/arm64/kvm/sys_regs.c static bool pmu_counter_idx_valid(struct kvm_vcpu *vcpu, u64 idx) idx 750 arch/arm64/kvm/sys_regs.c if (idx >= val && idx != ARMV8_PMU_CYCLE_IDX) { idx 762 arch/arm64/kvm/sys_regs.c u64 idx; idx 773 arch/arm64/kvm/sys_regs.c idx = __vcpu_sys_reg(vcpu, PMSELR_EL0) idx 780 arch/arm64/kvm/sys_regs.c idx = ARMV8_PMU_CYCLE_IDX; idx 789 arch/arm64/kvm/sys_regs.c idx = ARMV8_PMU_CYCLE_IDX; idx 795 arch/arm64/kvm/sys_regs.c idx = ((r->CRm & 3) << 3) | (r->Op2 & 7); idx 800 arch/arm64/kvm/sys_regs.c if (!pmu_counter_idx_valid(vcpu, idx)) idx 807 arch/arm64/kvm/sys_regs.c kvm_pmu_set_counter_value(vcpu, idx, p->regval); idx 809 arch/arm64/kvm/sys_regs.c p->regval = kvm_pmu_get_counter_value(vcpu, idx); idx 818 arch/arm64/kvm/sys_regs.c u64 idx, reg; idx 828 arch/arm64/kvm/sys_regs.c idx = __vcpu_sys_reg(vcpu, PMSELR_EL0) & ARMV8_PMU_COUNTER_MASK; idx 829 arch/arm64/kvm/sys_regs.c reg = PMEVTYPER0_EL0 + idx; idx 831 arch/arm64/kvm/sys_regs.c idx = ((r->CRm & 3) << 3) | (r->Op2 & 7); idx 832 arch/arm64/kvm/sys_regs.c if (idx == ARMV8_PMU_CYCLE_IDX) idx 836 arch/arm64/kvm/sys_regs.c reg = PMEVTYPER0_EL0 + idx; idx 841 arch/arm64/kvm/sys_regs.c if (!pmu_counter_idx_valid(vcpu, idx)) idx 845 arch/arm64/kvm/sys_regs.c kvm_pmu_set_counter_event_type(vcpu, p->regval, idx); idx 184 arch/arm64/kvm/vgic-sys-reg-v3.c struct sys_reg_params *p, u8 apr, u8 idx) idx 190 arch/arm64/kvm/vgic-sys-reg-v3.c ap_reg = &vgicv3->vgic_ap1r[idx]; idx 192 arch/arm64/kvm/vgic-sys-reg-v3.c ap_reg = &vgicv3->vgic_ap0r[idx]; idx 203 arch/arm64/kvm/vgic-sys-reg-v3.c u8 idx = r->Op2 & 3; idx 205 arch/arm64/kvm/vgic-sys-reg-v3.c if (idx > vgic_v3_max_apr_idx(vcpu)) idx 208 arch/arm64/kvm/vgic-sys-reg-v3.c vgic_v3_access_apr_reg(vcpu, p, apr, idx); idx 35 arch/arm64/mm/context.c #define idx2asid(idx) (((idx) << 1) & ~ASID_MASK) idx 39 arch/arm64/mm/context.c #define idx2asid(idx) asid2idx(idx) idx 864 arch/arm64/mm/mmu.c void __set_fixmap(enum fixed_addresses idx, idx 867 arch/arm64/mm/mmu.c unsigned long addr = __fix_to_virt(idx); idx 870 arch/arm64/mm/mmu.c BUG_ON(idx <= FIX_HOLE || idx >= __end_of_fixed_addresses); idx 56 arch/arm64/net/bpf_jit_comp.c int idx; idx 66 arch/arm64/net/bpf_jit_comp.c ctx->image[ctx->idx] = cpu_to_le32(insn); idx 68 arch/arm64/net/bpf_jit_comp.c ctx->idx++; idx 165 arch/arm64/net/bpf_jit_comp.c int from = ctx->idx; idx 185 arch/arm64/net/bpf_jit_comp.c const int idx0 = ctx->idx; idx 227 arch/arm64/net/bpf_jit_comp.c cur_offset = ctx->idx - idx0; idx 252 arch/arm64/net/bpf_jit_comp.c const int idx0 = ctx->idx; idx 253 arch/arm64/net/bpf_jit_comp.c #define cur_offset (ctx->idx - idx0) idx 791 arch/arm64/net/bpf_jit_comp.c ctx->offset[i] = ctx->idx; idx 795 arch/arm64/net/bpf_jit_comp.c ctx->offset[i] = ctx->idx; idx 807 arch/arm64/net/bpf_jit_comp.c for (i = 0; i < ctx->idx; i++) { idx 868 arch/arm64/net/bpf_jit_comp.c image_size = sizeof(u32) * ctx.idx; idx 893 arch/arm64/net/bpf_jit_comp.c ctx.epilogue_offset = ctx.idx; idx 897 arch/arm64/net/bpf_jit_comp.c image_size = sizeof(u32) * ctx.idx; idx 909 arch/arm64/net/bpf_jit_comp.c ctx.idx = 0; idx 932 arch/arm64/net/bpf_jit_comp.c bpf_flush_icache(header, ctx.image + ctx.idx); idx 935 arch/arm64/net/bpf_jit_comp.c if (extra_pass && ctx.idx != jit_data->ctx.idx) { idx 937 arch/arm64/net/bpf_jit_comp.c ctx.idx, jit_data->ctx.idx); idx 99 arch/c6x/platforms/megamod-pic.c int n, idx; idx 104 arch/c6x/platforms/megamod-pic.c idx = cascade->index; idx 106 arch/c6x/platforms/megamod-pic.c while ((events = soc_readl(&pic->regs->mevtflag[idx])) != 0) { idx 109 arch/c6x/platforms/megamod-pic.c irq = irq_linear_revmap(pic->irqhost, idx * 32 + n); idx 111 arch/c6x/platforms/megamod-pic.c soc_writel(1 << n, &pic->regs->evtclr[idx]); idx 908 arch/csky/kernel/perf_event.c if (hw_raw_write_mapping[hwc->idx] != NULL) idx 909 arch/csky/kernel/perf_event.c hw_raw_write_mapping[hwc->idx]((u64)(-left) & idx 912 arch/csky/kernel/perf_event.c cpwcr(HPOFSR, ~BIT(hwc->idx) & cprcr(HPOFSR)); idx 928 arch/csky/kernel/perf_event.c hw_raw_read_mapping[hwc->idx](), csky_pmu.count_width - 1); idx 980 arch/csky/kernel/perf_event.c hwc->idx = ret; idx 986 arch/csky/kernel/perf_event.c hwc->idx = ret; idx 991 arch/csky/kernel/perf_event.c hwc->idx = event->attr.config; idx 1025 arch/csky/kernel/perf_event.c int idx = hwc->idx; idx 1027 arch/csky/kernel/perf_event.c if (WARN_ON_ONCE(idx == -1)) idx 1039 arch/csky/kernel/perf_event.c cpwcr(HPINTENR, BIT(idx) | cprcr(HPINTENR)); idx 1040 arch/csky/kernel/perf_event.c cpwcr(HPCNTENR, BIT(idx) | cprcr(HPCNTENR)); idx 1049 arch/csky/kernel/perf_event.c int idx = hwc->idx; idx 1053 arch/csky/kernel/perf_event.c cpwcr(HPINTENR, ~BIT(idx) & cprcr(HPINTENR)); idx 1054 arch/csky/kernel/perf_event.c cpwcr(HPCNTENR, ~BIT(idx) & cprcr(HPCNTENR)); idx 1080 arch/csky/kernel/perf_event.c hw_events->events[hwc->idx] = NULL; idx 1091 arch/csky/kernel/perf_event.c hw_events->events[hwc->idx] = event; idx 1108 arch/csky/kernel/perf_event.c int idx; idx 1123 arch/csky/kernel/perf_event.c for (idx = 0; idx < CSKY_PMU_MAX_EVENTS; ++idx) { idx 1124 arch/csky/kernel/perf_event.c struct perf_event *event = cpuc->events[idx]; idx 1134 arch/csky/kernel/perf_event.c if (!(cprcr(HPOFSR) & BIT(idx))) idx 11 arch/csky/kernel/perf_regs.c u64 perf_reg_value(struct pt_regs *regs, int idx) idx 13 arch/csky/kernel/perf_regs.c if (WARN_ON_ONCE((u32)idx >= PERF_REG_CSKY_MAX)) idx 16 arch/csky/kernel/perf_regs.c return (u64)*((u32 *)regs + idx); idx 22 arch/csky/mm/asid.c #define idx2asid(info, idx) (((idx) << (info)->ctxt_shift) & ~ASID_MASK(info)) idx 42 arch/csky/mm/highmem.c int idx, type; idx 50 arch/csky/mm/highmem.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 51 arch/csky/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 53 arch/csky/mm/highmem.c BUG_ON(!pte_none(*(kmap_pte - idx))); idx 55 arch/csky/mm/highmem.c set_pte(kmap_pte-idx, mk_pte(page, PAGE_KERNEL)); idx 65 arch/csky/mm/highmem.c int idx; idx 71 arch/csky/mm/highmem.c idx = KM_TYPE_NR*smp_processor_id() + kmap_atomic_idx(); idx 73 arch/csky/mm/highmem.c BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); idx 75 arch/csky/mm/highmem.c pte_clear(&init_mm, vaddr, kmap_pte - idx); idx 78 arch/csky/mm/highmem.c (void) idx; /* to kill a warning */ idx 94 arch/csky/mm/highmem.c int idx, type; idx 99 arch/csky/mm/highmem.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 100 arch/csky/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 101 arch/csky/mm/highmem.c set_pte(kmap_pte-idx, pfn_pte(pfn, PAGE_KERNEL)); idx 109 arch/csky/mm/highmem.c unsigned long idx, vaddr = (unsigned long)ptr; idx 115 arch/csky/mm/highmem.c idx = virt_to_fix(vaddr); idx 116 arch/csky/mm/highmem.c pte = kmap_pte - (idx - FIX_KMAP_BEGIN); idx 69 arch/csky/mm/tlb.c int idx; idx 74 arch/csky/mm/tlb.c idx = read_mmu_index(); idx 75 arch/csky/mm/tlb.c if (idx >= 0) idx 103 arch/csky/mm/tlb.c int idx; idx 108 arch/csky/mm/tlb.c idx = read_mmu_index(); idx 109 arch/csky/mm/tlb.c if (idx >= 0) idx 129 arch/csky/mm/tlb.c int oldpid, idx; idx 136 arch/csky/mm/tlb.c idx = read_mmu_index(); idx 137 arch/csky/mm/tlb.c if (idx >= 0) idx 155 arch/csky/mm/tlb.c int oldpid, idx; idx 162 arch/csky/mm/tlb.c idx = read_mmu_index(); idx 163 arch/csky/mm/tlb.c if (idx >= 0) idx 1305 arch/ia64/hp/common/sba_iommu.c int idx; idx 1404 arch/ia64/hp/common/sba_iommu.c idx = sba_alloc_range(ioc, dev, dma_len); idx 1405 arch/ia64/hp/common/sba_iommu.c if (idx < 0) { idx 1409 arch/ia64/hp/common/sba_iommu.c dma_sg->dma_address = (dma_addr_t)(PIDE_FLAG | (idx << iovp_shift) idx 766 arch/ia64/kernel/perfmon.c int idx, next; idx 773 arch/ia64/kernel/perfmon.c idx = ctx->ctx_msgq_tail; idx 776 arch/ia64/kernel/perfmon.c DPRINT(("ctx=%p head=%d tail=%d msg=%d\n", ctx, ctx->ctx_msgq_head, ctx->ctx_msgq_tail, idx)); idx 778 arch/ia64/kernel/perfmon.c return ctx->ctx_msgq+idx; idx 328 arch/ia64/pci/pci.c int idx; idx 333 arch/ia64/pci/pci.c for (idx = 0; idx < PCI_BRIDGE_RESOURCES; idx++) { idx 334 arch/ia64/pci/pci.c struct resource *r = &dev->resource[idx]; idx 339 arch/ia64/pci/pci.c pci_claim_resource(dev, idx); idx 346 arch/ia64/pci/pci.c int idx; idx 351 arch/ia64/pci/pci.c for (idx = PCI_BRIDGE_RESOURCES; idx < PCI_NUM_RESOURCES; idx++) { idx 352 arch/ia64/pci/pci.c struct resource *r = &dev->resource[idx]; idx 357 arch/ia64/pci/pci.c pci_claim_bridge_resource(dev, idx); idx 82 arch/m68k/kernel/asm-offsets.c DEFINE(FONT_DESC_IDX, offsetof(struct font_desc, idx)); idx 49 arch/m68k/kernel/pcibios.c int idx; idx 54 arch/m68k/kernel/pcibios.c for (idx = 0; idx < 6; idx++) { idx 56 arch/m68k/kernel/pcibios.c if (!(mask & (1 << idx))) idx 59 arch/m68k/kernel/pcibios.c r = dev->resource + idx; idx 58 arch/microblaze/include/asm/fixmap.h extern void __set_fixmap(enum fixed_addresses idx, idx 39 arch/microblaze/mm/highmem.c int idx, type; idx 48 arch/microblaze/mm/highmem.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 49 arch/microblaze/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 51 arch/microblaze/mm/highmem.c BUG_ON(!pte_none(*(kmap_pte-idx))); idx 53 arch/microblaze/mm/highmem.c set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); idx 64 arch/microblaze/mm/highmem.c unsigned int idx; idx 74 arch/microblaze/mm/highmem.c idx = type + KM_TYPE_NR * smp_processor_id(); idx 76 arch/microblaze/mm/highmem.c BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); idx 82 arch/microblaze/mm/highmem.c pte_clear(&init_mm, vaddr, kmap_pte-idx); idx 91 arch/microblaze/mm/init.c int idx; idx 94 arch/microblaze/mm/init.c for (idx = 0; idx < __end_of_fixed_addresses; idx++) idx 95 arch/microblaze/mm/init.c clear_fixmap(idx); idx 247 arch/microblaze/mm/pgtable.c void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t flags) idx 249 arch/microblaze/mm/pgtable.c unsigned long address = __fix_to_virt(idx); idx 251 arch/microblaze/mm/pgtable.c if (idx >= __end_of_fixed_addresses) idx 745 arch/microblaze/pci/pci-common.c static inline void alloc_resource(struct pci_dev *dev, int idx) idx 747 arch/microblaze/pci/pci-common.c struct resource *pr, *r = &dev->resource[idx]; idx 750 arch/microblaze/pci/pci-common.c pci_name(dev), idx, idx 758 arch/microblaze/pci/pci-common.c pr_warn("PCI: Cannot allocate resource region %d ", idx); idx 776 arch/microblaze/pci/pci-common.c int idx, disabled; idx 782 arch/microblaze/pci/pci-common.c for (idx = 0; idx <= PCI_ROM_RESOURCE; idx++) { idx 783 arch/microblaze/pci/pci-common.c r = &dev->resource[idx]; idx 791 arch/microblaze/pci/pci-common.c if (idx == PCI_ROM_RESOURCE) idx 798 arch/microblaze/pci/pci-common.c alloc_resource(dev, idx); idx 122 arch/mips/alchemy/devboards/pm.c int idx; idx 137 arch/mips/alchemy/devboards/pm.c idx = (attr->attr.name)[4] - '0'; idx 139 arch/mips/alchemy/devboards/pm.c !!(db1x_pm_wakemsk & SYS_WAKEMSK_GPIO(idx))); idx 732 arch/mips/cavium-octeon/executive/cvmx-l2c.c uint64_t idx = addr >> CVMX_L2C_IDX_ADDR_SHIFT; idx 749 arch/mips/cavium-octeon/executive/cvmx-l2c.c uint32_t a_14_12 = (idx / (CVMX_L2C_MEMBANK_SELECT_SIZE/(1<<CVMX_L2C_IDX_ADDR_SHIFT))) & 0x7; idx 751 arch/mips/cavium-octeon/executive/cvmx-l2c.c idx ^= idx / cvmx_l2c_get_num_sets(); idx 752 arch/mips/cavium-octeon/executive/cvmx-l2c.c idx ^= a_14_12; idx 754 arch/mips/cavium-octeon/executive/cvmx-l2c.c idx ^= ((addr & CVMX_L2C_ALIAS_MASK) >> CVMX_L2C_TAG_ADDR_ALIAS_SHIFT); idx 757 arch/mips/cavium-octeon/executive/cvmx-l2c.c idx &= CVMX_L2C_IDX_MASK; idx 758 arch/mips/cavium-octeon/executive/cvmx-l2c.c return idx; idx 678 arch/mips/cavium-octeon/octeon-platform.c static void __init octeon_fdt_pip_iface(int pip, int idx) idx 685 arch/mips/cavium-octeon/octeon-platform.c snprintf(name_buffer, sizeof(name_buffer), "interface@%d", idx); idx 690 arch/mips/cavium-octeon/octeon-platform.c if (cvmx_helper_interface_enumerate(idx) == 0) idx 691 arch/mips/cavium-octeon/octeon-platform.c count = cvmx_helper_ports_on_interface(idx); idx 694 arch/mips/cavium-octeon/octeon-platform.c octeon_fdt_pip_port(iface, idx, p, count - 1); idx 106 arch/mips/fw/cfe/cfe_api.c int cfe_enumenv(int idx, char *name, int namelen, char *val, int vallen) idx 115 arch/mips/fw/cfe/cfe_api.c xiocb.plist.xiocb_envbuf.enum_idx = idx; idx 127 arch/mips/fw/cfe/cfe_api.c cfe_enummem(int idx, int flags, u64 *start, u64 *length, u64 *type) idx 136 arch/mips/fw/cfe/cfe_api.c xiocb.plist.xiocb_meminfo.mi_idx = idx; idx 20 arch/mips/generic/board-ocelot.c int idx; idx 27 arch/mips/generic/board-ocelot.c idx = read_c0_index(); idx 28 arch/mips/generic/board-ocelot.c if (idx < 0) idx 87 arch/mips/include/asm/fw/cfe/cfe_api.h int cfe_enumenv(int idx, char *name, int namelen, char *val, int vallen); idx 88 arch/mips/include/asm/fw/cfe/cfe_api.h int cfe_enummem(int idx, int flags, uint64_t * start, uint64_t * length, idx 40 arch/mips/include/asm/maar.h static inline void write_maar_pair(unsigned idx, phys_addr_t lower, idx 52 arch/mips/include/asm/maar.h write_c0_maari(idx << 1); idx 58 arch/mips/include/asm/maar.h write_c0_maari((idx << 1) | 0x1); idx 19 arch/mips/include/asm/msa.h extern void read_msa_wr_b(unsigned idx, union fpureg *to); idx 20 arch/mips/include/asm/msa.h extern void read_msa_wr_h(unsigned idx, union fpureg *to); idx 21 arch/mips/include/asm/msa.h extern void read_msa_wr_w(unsigned idx, union fpureg *to); idx 22 arch/mips/include/asm/msa.h extern void read_msa_wr_d(unsigned idx, union fpureg *to); idx 33 arch/mips/include/asm/msa.h static inline void read_msa_wr(unsigned idx, union fpureg *to, idx 38 arch/mips/include/asm/msa.h read_msa_wr_b(idx, to); idx 42 arch/mips/include/asm/msa.h read_msa_wr_h(idx, to); idx 46 arch/mips/include/asm/msa.h read_msa_wr_w(idx, to); idx 50 arch/mips/include/asm/msa.h read_msa_wr_d(idx, to); idx 58 arch/mips/include/asm/msa.h extern void write_msa_wr_b(unsigned idx, union fpureg *from); idx 59 arch/mips/include/asm/msa.h extern void write_msa_wr_h(unsigned idx, union fpureg *from); idx 60 arch/mips/include/asm/msa.h extern void write_msa_wr_w(unsigned idx, union fpureg *from); idx 61 arch/mips/include/asm/msa.h extern void write_msa_wr_d(unsigned idx, union fpureg *from); idx 72 arch/mips/include/asm/msa.h static inline void write_msa_wr(unsigned idx, union fpureg *from, idx 77 arch/mips/include/asm/msa.h write_msa_wr_b(idx, from); idx 81 arch/mips/include/asm/msa.h write_msa_wr_h(idx, from); idx 85 arch/mips/include/asm/msa.h write_msa_wr_w(idx, from); idx 89 arch/mips/include/asm/msa.h write_msa_wr_d(idx, from); idx 100 arch/mips/include/asm/processor.h # define FPR_IDX(width, idx) (idx) idx 102 arch/mips/include/asm/processor.h # define FPR_IDX(width, idx) ((idx) ^ ((64 / (width)) - 1)) idx 106 arch/mips/include/asm/processor.h static inline u##width get_fpr##width(union fpureg *fpr, unsigned idx) \ idx 108 arch/mips/include/asm/processor.h return fpr->val##width[FPR_IDX(width, idx)]; \ idx 111 arch/mips/include/asm/processor.h static inline void set_fpr##width(union fpureg *fpr, unsigned idx, \ idx 114 arch/mips/include/asm/processor.h fpr->val##width[FPR_IDX(width, idx)] = val; \ idx 539 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_HR_BASE(idx) (A_BCM1480_HR_BASE_0 + ((idx)*BCM1480_HR_REGISTER_SPACING)) idx 540 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_HR_REGISTER(idx, reg) (A_BCM1480_HR_BASE(idx) + (reg)) idx 550 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HR_RULE_OP(idx) (BCM1480_HR_OP_OFFSET + ((idx)*BCM1480_HR_RULE_SPACING)) idx 551 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HR_RULE_TYPE(idx) (BCM1480_HR_TYPE_OFFSET + ((idx)*BCM1480_HR_RULE_SPACING)) idx 556 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HR_HA_LEAF0(idx) (BCM1480_HR_LEAF_OFFSET + ((idx)*BCM1480_HR_LEAF_SPACING)) idx 563 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HR_PATH(idx) (BCM1480_HR_PATH_OFFSET + ((idx)*BCM1480_HR_PATH_SPACING)) idx 570 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HR_RT_WORD(idx) (BCM1480_HR_ROUTE_OFFSET + ((idx)*BCM1480_HR_ROUTE_SPACING)) idx 588 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_PMI_LCL_BASE(idx) (A_BCM1480_PMI_LCL_0 + ((idx)*BCM1480_PM_LCL_REGISTER_SPACING)) idx 589 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_PMI_LCL_REGISTER(idx, reg) (A_BCM1480_PMI_LCL_BASE(idx) + (reg)) idx 590 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_PMO_LCL_BASE(idx) (A_BCM1480_PMO_LCL_0 + ((idx)*BCM1480_PM_LCL_REGISTER_SPACING)) idx 591 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_PMO_LCL_REGISTER(idx, reg) (A_BCM1480_PMO_LCL_BASE(idx) + (reg)) idx 710 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_HSP_BASE(idx) (A_BCM1480_HSP_BASE_0 + ((idx)*BCM1480_HSP_REGISTER_SPACING)) idx 711 arch/mips/include/asm/sibyte/bcm1480_regs.h #define A_BCM1480_HSP_REGISTER(idx, reg) (A_BCM1480_HSP_BASE(idx) + (reg)) idx 742 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HSP_RX_PKT_RAMALLOC(idx) (R_BCM1480_HSP_RX_PKT_RAMALLOC_0 + 8*(idx)) idx 760 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HSP_RX_SPI_WATERMARK(idx) (R_BCM1480_HSP_RX_SPI_WATERMARK_0 + 8*(idx)) idx 782 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HSP_TX_PKT_RAMALLOC(idx) (R_BCM1480_HSP_TX_PKT_RAMALLOC_0 + 8*(idx)) idx 794 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HSP_TX_PKT_RXPHITCNT(idx) (R_BCM1480_HSP_TX_PKT_RXPHITCNT_0 + 8*(idx)) idx 802 arch/mips/include/asm/sibyte/bcm1480_regs.h #define R_BCM1480_HSP_TX_PKT_TXPHITCNT(idx) (R_BCM1480_HSP_TX_PKT_TXPHITCNT_0 + 8*(idx)) idx 576 arch/mips/include/asm/sibyte/sb1250_regs.h #define A_SMB_BASE(idx) (A_SMB_0+(idx)*SMB_REGISTER_SPACING) idx 577 arch/mips/include/asm/sibyte/sb1250_regs.h #define A_SMB_REGISTER(idx, reg) (A_SMB_BASE(idx)+(reg)) idx 810 arch/mips/include/asm/sibyte/sb1250_regs.h #define A_DM_BASE(idx) (A_DM_0 + ((idx) * DM_REGISTER_SPACING)) idx 811 arch/mips/include/asm/sibyte/sb1250_regs.h #define A_DM_REGISTER(idx, reg) (A_DM_BASE(idx) + (reg)) idx 824 arch/mips/include/asm/sibyte/sb1250_regs.h #define A_DM_PARTIAL(idx) (A_DM_PARTIAL_0 + ((idx) * DM_PARTIAL_REGISTER_SPACING)) idx 832 arch/mips/include/asm/sibyte/sb1250_regs.h #define A_DM_CRC_BASE(idx) (A_DM_CRC_0 + ((idx) * DM_CRC_REGISTER_SPACING)) idx 833 arch/mips/include/asm/sibyte/sb1250_regs.h #define A_DM_CRC_REGISTER(idx, reg) (A_DM_CRC_BASE(idx) + (reg)) idx 8 arch/mips/include/asm/tlb.h #define _UNIQUE_ENTRYHI(base, idx) \ idx 9 arch/mips/include/asm/tlb.h (((base) + ((idx) << (PAGE_SHIFT + 1))) | \ idx 11 arch/mips/include/asm/tlb.h #define UNIQUE_ENTRYHI(idx) _UNIQUE_ENTRYHI(CKSEG0, idx) idx 12 arch/mips/include/asm/tlb.h #define UNIQUE_GUEST_ENTRYHI(idx) _UNIQUE_ENTRYHI(CKSEG1, idx) idx 427 arch/mips/include/asm/txx9/tx4939.h #define TX4939_CRYPTO_DESC_HASH_IDX(idx) ((idx) << 4) idx 429 arch/mips/include/asm/txx9/tx4939.h #define TX4939_CRYPTO_DESC_ENCRYPT_IDX(idx) ((idx) << 0) idx 82 arch/mips/kernel/perf_event_mipsxx.c u64 (*read_counter)(unsigned int idx); idx 83 arch/mips/kernel/perf_event_mipsxx.c void (*write_counter)(unsigned int idx, u64 val); idx 157 arch/mips/kernel/perf_event_mipsxx.c static unsigned int mipsxx_pmu_swizzle_perf_idx(unsigned int idx) idx 160 arch/mips/kernel/perf_event_mipsxx.c idx = (idx + 2) & 3; idx 161 arch/mips/kernel/perf_event_mipsxx.c return idx; idx 164 arch/mips/kernel/perf_event_mipsxx.c static u64 mipsxx_pmu_read_counter(unsigned int idx) idx 166 arch/mips/kernel/perf_event_mipsxx.c idx = mipsxx_pmu_swizzle_perf_idx(idx); idx 168 arch/mips/kernel/perf_event_mipsxx.c switch (idx) { idx 182 arch/mips/kernel/perf_event_mipsxx.c WARN_ONCE(1, "Invalid performance counter number (%d)\n", idx); idx 187 arch/mips/kernel/perf_event_mipsxx.c static u64 mipsxx_pmu_read_counter_64(unsigned int idx) idx 189 arch/mips/kernel/perf_event_mipsxx.c idx = mipsxx_pmu_swizzle_perf_idx(idx); idx 191 arch/mips/kernel/perf_event_mipsxx.c switch (idx) { idx 201 arch/mips/kernel/perf_event_mipsxx.c WARN_ONCE(1, "Invalid performance counter number (%d)\n", idx); idx 206 arch/mips/kernel/perf_event_mipsxx.c static void mipsxx_pmu_write_counter(unsigned int idx, u64 val) idx 208 arch/mips/kernel/perf_event_mipsxx.c idx = mipsxx_pmu_swizzle_perf_idx(idx); idx 210 arch/mips/kernel/perf_event_mipsxx.c switch (idx) { idx 226 arch/mips/kernel/perf_event_mipsxx.c static void mipsxx_pmu_write_counter_64(unsigned int idx, u64 val) idx 228 arch/mips/kernel/perf_event_mipsxx.c idx = mipsxx_pmu_swizzle_perf_idx(idx); idx 230 arch/mips/kernel/perf_event_mipsxx.c switch (idx) { idx 246 arch/mips/kernel/perf_event_mipsxx.c static unsigned int mipsxx_pmu_read_control(unsigned int idx) idx 248 arch/mips/kernel/perf_event_mipsxx.c idx = mipsxx_pmu_swizzle_perf_idx(idx); idx 250 arch/mips/kernel/perf_event_mipsxx.c switch (idx) { idx 260 arch/mips/kernel/perf_event_mipsxx.c WARN_ONCE(1, "Invalid performance counter number (%d)\n", idx); idx 265 arch/mips/kernel/perf_event_mipsxx.c static void mipsxx_pmu_write_control(unsigned int idx, unsigned int val) idx 267 arch/mips/kernel/perf_event_mipsxx.c idx = mipsxx_pmu_swizzle_perf_idx(idx); idx 269 arch/mips/kernel/perf_event_mipsxx.c switch (idx) { idx 315 arch/mips/kernel/perf_event_mipsxx.c static void mipsxx_pmu_enable_event(struct hw_perf_event *evt, int idx) idx 321 arch/mips/kernel/perf_event_mipsxx.c WARN_ON(idx < 0 || idx >= mipspmu.num_counters); idx 323 arch/mips/kernel/perf_event_mipsxx.c cpuc->saved_ctrl[idx] = M_PERFCTL_EVENT(evt->event_base & 0xff) | idx 330 arch/mips/kernel/perf_event_mipsxx.c cpuc->saved_ctrl[idx] |= idx 335 arch/mips/kernel/perf_event_mipsxx.c cpuc->saved_ctrl[idx] |= M_TC_EN_ALL; idx 348 arch/mips/kernel/perf_event_mipsxx.c cpuc->saved_ctrl[idx] |= ctrl; idx 356 arch/mips/kernel/perf_event_mipsxx.c static void mipsxx_pmu_disable_event(int idx) idx 361 arch/mips/kernel/perf_event_mipsxx.c WARN_ON(idx < 0 || idx >= mipspmu.num_counters); idx 364 arch/mips/kernel/perf_event_mipsxx.c cpuc->saved_ctrl[idx] = mipsxx_pmu_read_control(idx) & idx 366 arch/mips/kernel/perf_event_mipsxx.c mipsxx_pmu_write_control(idx, cpuc->saved_ctrl[idx]); idx 372 arch/mips/kernel/perf_event_mipsxx.c int idx) idx 399 arch/mips/kernel/perf_event_mipsxx.c mipspmu.write_counter(idx, mipspmu.overflow - left); idx 408 arch/mips/kernel/perf_event_mipsxx.c int idx) idx 415 arch/mips/kernel/perf_event_mipsxx.c new_raw_count = mipspmu.read_counter(idx); idx 437 arch/mips/kernel/perf_event_mipsxx.c mipspmu_event_set_period(event, hwc, hwc->idx); idx 440 arch/mips/kernel/perf_event_mipsxx.c mipsxx_pmu_enable_event(hwc, hwc->idx); idx 449 arch/mips/kernel/perf_event_mipsxx.c mipsxx_pmu_disable_event(hwc->idx); idx 451 arch/mips/kernel/perf_event_mipsxx.c mipspmu_event_update(event, hwc, hwc->idx); idx 460 arch/mips/kernel/perf_event_mipsxx.c int idx; idx 466 arch/mips/kernel/perf_event_mipsxx.c idx = mipsxx_pmu_alloc_counter(cpuc, hwc); idx 467 arch/mips/kernel/perf_event_mipsxx.c if (idx < 0) { idx 468 arch/mips/kernel/perf_event_mipsxx.c err = idx; idx 476 arch/mips/kernel/perf_event_mipsxx.c event->hw.idx = idx; idx 477 arch/mips/kernel/perf_event_mipsxx.c mipsxx_pmu_disable_event(idx); idx 478 arch/mips/kernel/perf_event_mipsxx.c cpuc->events[idx] = event; idx 496 arch/mips/kernel/perf_event_mipsxx.c int idx = hwc->idx; idx 498 arch/mips/kernel/perf_event_mipsxx.c WARN_ON(idx < 0 || idx >= mipspmu.num_counters); idx 501 arch/mips/kernel/perf_event_mipsxx.c cpuc->events[idx] = NULL; idx 502 arch/mips/kernel/perf_event_mipsxx.c clear_bit(idx, cpuc->used_mask); idx 512 arch/mips/kernel/perf_event_mipsxx.c if (hwc->idx < 0) idx 515 arch/mips/kernel/perf_event_mipsxx.c mipspmu_event_update(event, hwc, hwc->idx); idx 674 arch/mips/kernel/perf_event_mipsxx.c static const struct mips_perf_event *mipspmu_map_general_event(int idx) idx 677 arch/mips/kernel/perf_event_mipsxx.c if ((*mipspmu.general_event_map)[idx].cntr_mask == 0) idx 679 arch/mips/kernel/perf_event_mipsxx.c return &(*mipspmu.general_event_map)[idx]; idx 734 arch/mips/kernel/perf_event_mipsxx.c int idx, struct perf_sample_data *data, idx 737 arch/mips/kernel/perf_event_mipsxx.c struct perf_event *event = cpuc->events[idx]; idx 740 arch/mips/kernel/perf_event_mipsxx.c mipspmu_event_update(event, hwc, idx); idx 742 arch/mips/kernel/perf_event_mipsxx.c if (!mipspmu_event_set_period(event, hwc, idx)) idx 746 arch/mips/kernel/perf_event_mipsxx.c mipsxx_pmu_disable_event(idx); idx 1333 arch/mips/kernel/perf_event_mipsxx.c hwc->idx = -1; idx 652 arch/mips/kvm/mips.c unsigned int idx; idx 675 arch/mips/kvm/mips.c idx = reg->id - KVM_REG_MIPS_FPR_32(0); idx 678 arch/mips/kvm/mips.c v = get_fpr32(&fpu->fpr[idx], 0); idx 680 arch/mips/kvm/mips.c v = get_fpr32(&fpu->fpr[idx & ~1], idx & 1); idx 685 arch/mips/kvm/mips.c idx = reg->id - KVM_REG_MIPS_FPR_64(0); idx 687 arch/mips/kvm/mips.c if (idx & 1 && !(kvm_read_c0_guest_status(cop0) & ST0_FR)) idx 689 arch/mips/kvm/mips.c v = get_fpr64(&fpu->fpr[idx], 0); idx 709 arch/mips/kvm/mips.c idx = reg->id - KVM_REG_MIPS_VEC_128(0); idx 712 arch/mips/kvm/mips.c vs[0] = get_fpr64(&fpu->fpr[idx], 0); idx 713 arch/mips/kvm/mips.c vs[1] = get_fpr64(&fpu->fpr[idx], 1); idx 716 arch/mips/kvm/mips.c vs[0] = get_fpr64(&fpu->fpr[idx], 1); idx 717 arch/mips/kvm/mips.c vs[1] = get_fpr64(&fpu->fpr[idx], 0); idx 763 arch/mips/kvm/mips.c unsigned int idx; idx 809 arch/mips/kvm/mips.c idx = reg->id - KVM_REG_MIPS_FPR_32(0); idx 812 arch/mips/kvm/mips.c set_fpr32(&fpu->fpr[idx], 0, v); idx 814 arch/mips/kvm/mips.c set_fpr32(&fpu->fpr[idx & ~1], idx & 1, v); idx 819 arch/mips/kvm/mips.c idx = reg->id - KVM_REG_MIPS_FPR_64(0); idx 821 arch/mips/kvm/mips.c if (idx & 1 && !(kvm_read_c0_guest_status(cop0) & ST0_FR)) idx 823 arch/mips/kvm/mips.c set_fpr64(&fpu->fpr[idx], 0, v); idx 840 arch/mips/kvm/mips.c idx = reg->id - KVM_REG_MIPS_VEC_128(0); idx 843 arch/mips/kvm/mips.c set_fpr64(&fpu->fpr[idx], 0, vs[0]); idx 844 arch/mips/kvm/mips.c set_fpr64(&fpu->fpr[idx], 1, vs[1]); idx 847 arch/mips/kvm/mips.c set_fpr64(&fpu->fpr[idx], 1, vs[0]); idx 848 arch/mips/kvm/mips.c set_fpr64(&fpu->fpr[idx], 0, vs[1]); idx 1009 arch/mips/kvm/mmu.c int idx; idx 1019 arch/mips/kvm/mmu.c idx = (badvaddr >> PAGE_SHIFT) & 1; idx 1020 arch/mips/kvm/mmu.c if (kvm_mips_map_page(vcpu, gpa, write_fault, &pte_gpa[idx], idx 1021 arch/mips/kvm/mmu.c &pte_gpa[!idx]) < 0) idx 1048 arch/mips/kvm/mmu.c unsigned int idx = TLB_LO_IDX(*tlb, gva); idx 1062 arch/mips/kvm/mmu.c if (kvm_mips_map_page(vcpu, mips3_tlbpfn_to_paddr(tlb_lo[idx]), idx 1063 arch/mips/kvm/mmu.c write_fault, &pte_gpa[idx], NULL) < 0) idx 1067 arch/mips/kvm/mmu.c pte_gpa[!idx] = pfn_pte(0, __pgprot(0)); idx 1068 arch/mips/kvm/mmu.c if (tlb_lo[!idx] & ENTRYLO_V) { idx 1071 arch/mips/kvm/mmu.c mips3_tlbpfn_to_paddr(tlb_lo[!idx])); idx 1073 arch/mips/kvm/mmu.c pte_gpa[!idx] = *ptep_buddy; idx 141 arch/mips/kvm/tlb.c int idx; idx 148 arch/mips/kvm/tlb.c idx = read_c0_index(); idx 150 arch/mips/kvm/tlb.c if (idx >= current_cpu_data.tlbsize) idx 153 arch/mips/kvm/tlb.c if (idx >= 0) { idx 154 arch/mips/kvm/tlb.c write_c0_entryhi(UNIQUE_ENTRYHI(idx)); idx 163 arch/mips/kvm/tlb.c return idx; idx 253 arch/mips/kvm/tlb.c int idx; idx 264 arch/mips/kvm/tlb.c idx = _kvm_mips_host_tlb_inv((va & VPN2_MASK) | idx 281 arch/mips/kvm/tlb.c if (idx > 0) idx 284 arch/mips/kvm/tlb.c kvm_mips_get_root_asid(vcpu), idx); idx 1827 arch/mips/kvm/vz.c unsigned int idx; idx 1985 arch/mips/kvm/vz.c idx = reg->id - KVM_REG_MIPS_CP0_MAAR(0); idx 1986 arch/mips/kvm/vz.c if (idx >= ARRAY_SIZE(vcpu->arch.maar)) idx 1988 arch/mips/kvm/vz.c *v = vcpu->arch.maar[idx]; idx 2004 arch/mips/kvm/vz.c idx = reg->id - KVM_REG_MIPS_CP0_KSCRATCH1 + 2; idx 2005 arch/mips/kvm/vz.c if (!cpu_guest_has_kscr(idx)) idx 2007 arch/mips/kvm/vz.c switch (idx) { idx 2048 arch/mips/kvm/vz.c unsigned int idx; idx 2254 arch/mips/kvm/vz.c idx = reg->id - KVM_REG_MIPS_CP0_MAAR(0); idx 2255 arch/mips/kvm/vz.c if (idx >= ARRAY_SIZE(vcpu->arch.maar)) idx 2257 arch/mips/kvm/vz.c vcpu->arch.maar[idx] = mips_process_maar(dmtc_op, v); idx 2273 arch/mips/kvm/vz.c idx = reg->id - KVM_REG_MIPS_CP0_KSCRATCH1 + 2; idx 2274 arch/mips/kvm/vz.c if (!cpu_guest_has_kscr(idx)) idx 2276 arch/mips/kvm/vz.c switch (idx) { idx 131 arch/mips/loongson64/common/mem.c int idx; idx 138 arch/mips/loongson64/common/mem.c for (idx = 0; idx < PCI_NUM_RESOURCES; idx++) { idx 139 arch/mips/loongson64/common/mem.c r = &dev->resource[idx]; idx 78 arch/mips/math-emu/dsemul.c int idx; idx 91 arch/mips/math-emu/dsemul.c idx = BD_EMUFRAME_NONE; idx 97 arch/mips/math-emu/dsemul.c idx = bitmap_find_free_region(mm_ctx->bd_emupage_allocmap, idx 99 arch/mips/math-emu/dsemul.c if (idx < 0) { idx 119 arch/mips/math-emu/dsemul.c pr_debug("allocate emuframe %d to %d\n", idx, current->pid); idx 122 arch/mips/math-emu/dsemul.c return idx; idx 125 arch/mips/math-emu/dsemul.c static void free_emuframe(int idx, struct mm_struct *mm) idx 131 arch/mips/math-emu/dsemul.c pr_debug("free emuframe %d from %d\n", idx, current->pid); idx 132 arch/mips/math-emu/dsemul.c bitmap_clear(mm_ctx->bd_emupage_allocmap, idx, 1); idx 50 arch/mips/mm/highmem.c int idx, type; idx 58 arch/mips/mm/highmem.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 59 arch/mips/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 61 arch/mips/mm/highmem.c BUG_ON(!pte_none(*(kmap_pte - idx))); idx 63 arch/mips/mm/highmem.c set_pte(kmap_pte-idx, mk_pte(page, PAGE_KERNEL)); idx 84 arch/mips/mm/highmem.c int idx = type + KM_TYPE_NR * smp_processor_id(); idx 86 arch/mips/mm/highmem.c BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); idx 92 arch/mips/mm/highmem.c pte_clear(&init_mm, vaddr, kmap_pte-idx); idx 109 arch/mips/mm/highmem.c int idx, type; idx 115 arch/mips/mm/highmem.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 116 arch/mips/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 117 arch/mips/mm/highmem.c set_pte(kmap_pte-idx, pfn_pte(pfn, PAGE_KERNEL)); idx 86 arch/mips/mm/init.c enum fixed_addresses idx; idx 97 arch/mips/mm/init.c idx = (addr >> PAGE_SHIFT) & (FIX_N_COLOURS - 1); idx 98 arch/mips/mm/init.c idx += in_interrupt() ? FIX_N_COLOURS : 0; idx 99 arch/mips/mm/init.c vaddr = __fix_to_virt(FIX_CMAP_END - idx); idx 94 arch/mips/mm/tlb-r3k.c int idx; idx 99 arch/mips/mm/tlb-r3k.c idx = read_c0_index(); idx 102 arch/mips/mm/tlb-r3k.c if (idx < 0) /* BARRIER */ idx 131 arch/mips/mm/tlb-r3k.c int idx; idx 136 arch/mips/mm/tlb-r3k.c idx = read_c0_index(); idx 139 arch/mips/mm/tlb-r3k.c if (idx < 0) /* BARRIER */ idx 157 arch/mips/mm/tlb-r3k.c int oldpid, newpid, idx; idx 169 arch/mips/mm/tlb-r3k.c idx = read_c0_index(); idx 172 arch/mips/mm/tlb-r3k.c if (idx < 0) /* BARRIER */ idx 186 arch/mips/mm/tlb-r3k.c int idx, pid; idx 208 arch/mips/mm/tlb-r3k.c idx = read_c0_index(); idx 211 arch/mips/mm/tlb-r3k.c if (idx < 0) { /* BARRIER */ idx 134 arch/mips/mm/tlb-r4k.c int idx; idx 144 arch/mips/mm/tlb-r4k.c idx = read_c0_index(); idx 147 arch/mips/mm/tlb-r4k.c if (idx < 0) idx 150 arch/mips/mm/tlb-r4k.c write_c0_entryhi(UNIQUE_ENTRYHI(idx)); idx 185 arch/mips/mm/tlb-r4k.c int idx; idx 192 arch/mips/mm/tlb-r4k.c idx = read_c0_index(); idx 195 arch/mips/mm/tlb-r4k.c if (idx < 0) idx 198 arch/mips/mm/tlb-r4k.c write_c0_entryhi(UNIQUE_ENTRYHI(idx)); idx 219 arch/mips/mm/tlb-r4k.c int idx; idx 235 arch/mips/mm/tlb-r4k.c idx = read_c0_index(); idx 238 arch/mips/mm/tlb-r4k.c if (idx < 0) idx 241 arch/mips/mm/tlb-r4k.c write_c0_entryhi(UNIQUE_ENTRYHI(idx)); idx 263 arch/mips/mm/tlb-r4k.c int oldpid, idx; idx 273 arch/mips/mm/tlb-r4k.c idx = read_c0_index(); idx 276 arch/mips/mm/tlb-r4k.c if (idx >= 0) { idx 278 arch/mips/mm/tlb-r4k.c write_c0_entryhi(UNIQUE_ENTRYHI(idx)); idx 301 arch/mips/mm/tlb-r4k.c int idx, pid; idx 325 arch/mips/mm/tlb-r4k.c idx = read_c0_index(); idx 337 arch/mips/mm/tlb-r4k.c if (idx < 0) idx 367 arch/mips/mm/tlb-r4k.c if (idx < 0) idx 103 arch/mips/net/ebpf_jit.c u32 idx; idx 129 arch/mips/net/ebpf_jit.c u32 *p = &(ctx)->target[ctx->idx]; \ idx 135 arch/mips/net/ebpf_jit.c (ctx)->idx++; \ idx 177 arch/mips/net/ebpf_jit.c (ctx->idx * 4) - 4; idx 423 arch/mips/net/ebpf_jit.c int idx) idx 464 arch/mips/net/ebpf_jit.c get_reg_val_type(ctx, idx, insn->dst_reg) == REG_32BIT) idx 470 arch/mips/net/ebpf_jit.c get_reg_val_type(ctx, idx, insn->dst_reg) != REG_32BIT) idx 1533 arch/mips/net/ebpf_jit.c ctx->offsets[i] = (ctx->offsets[i] & OFFSETS_B_CONV) | (ctx->idx * 4); idx 1542 arch/mips/net/ebpf_jit.c ctx->offsets[i] = ctx->idx * 4; idx 1553 arch/mips/net/ebpf_jit.c ctx->offsets[i] = ctx->idx * 4; idx 1566 arch/mips/net/ebpf_jit.c int idx; idx 1569 arch/mips/net/ebpf_jit.c for (idx = start_idx; idx < prog->len; idx++) { idx 1570 arch/mips/net/ebpf_jit.c rvt[idx] = (rvt[idx] & RVT_VISITED_MASK) | exit_rvt; idx 1571 arch/mips/net/ebpf_jit.c insn = prog->insnsi + idx; idx 1608 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1627 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1642 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1643 arch/mips/net/ebpf_jit.c idx++; idx 1660 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1675 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1680 arch/mips/net/ebpf_jit.c rvt[idx] = RVT_DONE | exit_rvt; idx 1682 arch/mips/net/ebpf_jit.c return idx; idx 1684 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1685 arch/mips/net/ebpf_jit.c idx += insn->off; idx 1699 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_BRANCH_TAKEN; idx 1700 arch/mips/net/ebpf_jit.c idx += insn->off; idx 1703 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_FALL_THROUGH; idx 1712 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1716 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1721 arch/mips/net/ebpf_jit.c rvt[idx] |= RVT_DONE; idx 1725 arch/mips/net/ebpf_jit.c return idx; idx 1875 arch/mips/net/ebpf_jit.c ctx.idx = 0; idx 1886 arch/mips/net/ebpf_jit.c image_size = 4 * ctx.idx; idx 1896 arch/mips/net/ebpf_jit.c ctx.idx = 0; idx 1906 arch/mips/net/ebpf_jit.c (unsigned long)&ctx.target[ctx.idx]); idx 28 arch/mips/oprofile/op_model_loongson2.c #define LOONGSON2_PERFCTRL_EVENT(idx, event) \ idx 29 arch/mips/oprofile/op_model_loongson2.c (((event) & 0x0f) << ((idx) ? 9 : 5)) idx 28 arch/mips/oprofile/op_model_loongson3.c #define LOONGSON3_PERFCTRL_EVENT(idx, event) \ idx 29 arch/mips/oprofile/op_model_loongson3.c (((event) & (idx ? 0x0f : 0x3f)) << 5) idx 250 arch/mips/pci/pci-legacy.c int idx; idx 255 arch/mips/pci/pci-legacy.c for (idx=0; idx < PCI_NUM_RESOURCES; idx++) { idx 257 arch/mips/pci/pci-legacy.c if (!(mask & (1<<idx))) idx 260 arch/mips/pci/pci-legacy.c r = &dev->resource[idx]; idx 263 arch/mips/pci/pci-legacy.c if ((idx == PCI_ROM_RESOURCE) && idx 89 arch/mips/sibyte/common/cfe.c unsigned int idx; idx 105 arch/mips/sibyte/common/cfe.c for (idx = 0; cfe_enummem(idx, mem_flags, &addr, &size, &type) != CFE_ERR_NOMORE; idx 106 arch/mips/sibyte/common/cfe.c idx++) { idx 171 arch/mips/sibyte/common/cfe.c int idx; idx 176 arch/mips/sibyte/common/cfe.c for (idx = 0; idx < sizeof(rdarg)-1; idx++) { idx 177 arch/mips/sibyte/common/cfe.c if (!str[idx] || (str[idx] == ' ')) break; idx 178 arch/mips/sibyte/common/cfe.c rdarg[idx] = str[idx]; idx 181 arch/mips/sibyte/common/cfe.c rdarg[idx] = 0; idx 26 arch/nds32/include/asm/fixmap.h void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot); idx 190 arch/nds32/include/asm/pmu.h int idx; idx 193 arch/nds32/include/asm/pmu.h idx = 0; idx 195 arch/nds32/include/asm/pmu.h idx = 1; idx 197 arch/nds32/include/asm/pmu.h idx = 2; idx 203 arch/nds32/include/asm/pmu.h return idx; idx 39 arch/nds32/kernel/cacheinfo.c unsigned int level, idx; idx 43 arch/nds32/kernel/cacheinfo.c for (idx = 0, level = 1; level <= this_cpu_ci->num_levels && idx 44 arch/nds32/kernel/cacheinfo.c idx < this_cpu_ci->num_leaves; idx++, level++) { idx 84 arch/nds32/kernel/perf_event_cpu.c int idx = config >> 8; idx 86 arch/nds32/kernel/perf_event_cpu.c switch (idx) { idx 158 arch/nds32/kernel/perf_event_cpu.c static inline int nds32_pfm_counter_has_overflowed(u32 pfm, int idx) idx 162 arch/nds32/kernel/perf_event_cpu.c switch (idx) { idx 232 arch/nds32/kernel/perf_event_cpu.c int idx; idx 250 arch/nds32/kernel/perf_event_cpu.c for (idx = 0; idx < cpu_pmu->num_events; ++idx) { idx 251 arch/nds32/kernel/perf_event_cpu.c struct perf_event *event = cpuc->events[idx]; idx 262 arch/nds32/kernel/perf_event_cpu.c if (!nds32_pfm_counter_has_overflowed(pfm, idx)) idx 287 arch/nds32/kernel/perf_event_cpu.c static inline int nds32_pfm_counter_valid(struct nds32_pmu *cpu_pmu, int idx) idx 289 arch/nds32/kernel/perf_event_cpu.c return ((idx >= 0) && (idx < cpu_pmu->num_events)); idx 292 arch/nds32/kernel/perf_event_cpu.c static inline int nds32_pfm_disable_counter(int idx) idx 297 arch/nds32/kernel/perf_event_cpu.c mask = PFM_CTL_EN[idx]; idx 301 arch/nds32/kernel/perf_event_cpu.c return idx; idx 311 arch/nds32/kernel/perf_event_cpu.c int idx = event->idx; idx 315 arch/nds32/kernel/perf_event_cpu.c if (idx == -1) idx 318 arch/nds32/kernel/perf_event_cpu.c no_kernel_tracing = PFM_CTL_KS[idx]; idx 319 arch/nds32/kernel/perf_event_cpu.c no_user_tracing = PFM_CTL_KU[idx]; idx 337 arch/nds32/kernel/perf_event_cpu.c static inline void nds32_pfm_write_evtsel(int idx, u32 evnum) idx 346 arch/nds32/kernel/perf_event_cpu.c offset = PFM_CTL_OFFSEL[idx]; idx 348 arch/nds32/kernel/perf_event_cpu.c no_kernel_mask = PFM_CTL_KS[idx]; idx 349 arch/nds32/kernel/perf_event_cpu.c no_user_mask = PFM_CTL_KU[idx]; idx 359 arch/nds32/kernel/perf_event_cpu.c ev_mask = PFM_CTL_SEL[idx]; idx 370 arch/nds32/kernel/perf_event_cpu.c static inline int nds32_pfm_enable_counter(int idx) idx 375 arch/nds32/kernel/perf_event_cpu.c mask = PFM_CTL_EN[idx]; idx 379 arch/nds32/kernel/perf_event_cpu.c return idx; idx 382 arch/nds32/kernel/perf_event_cpu.c static inline int nds32_pfm_enable_intens(int idx) idx 387 arch/nds32/kernel/perf_event_cpu.c mask = PFM_CTL_IE[idx]; idx 391 arch/nds32/kernel/perf_event_cpu.c return idx; idx 394 arch/nds32/kernel/perf_event_cpu.c static inline int nds32_pfm_disable_intens(int idx) idx 399 arch/nds32/kernel/perf_event_cpu.c mask = PFM_CTL_IE[idx]; idx 403 arch/nds32/kernel/perf_event_cpu.c return idx; idx 419 arch/nds32/kernel/perf_event_cpu.c int idx = hwc->idx; idx 421 arch/nds32/kernel/perf_event_cpu.c if (!nds32_pfm_counter_valid(cpu_pmu, idx)) { idx 435 arch/nds32/kernel/perf_event_cpu.c nds32_pfm_disable_counter(idx); idx 449 arch/nds32/kernel/perf_event_cpu.c nds32_pfm_write_evtsel(idx, evnum); idx 454 arch/nds32/kernel/perf_event_cpu.c nds32_pfm_enable_intens(idx); idx 459 arch/nds32/kernel/perf_event_cpu.c nds32_pfm_enable_counter(idx); idx 470 arch/nds32/kernel/perf_event_cpu.c int idx = hwc->idx; idx 472 arch/nds32/kernel/perf_event_cpu.c if (!nds32_pfm_counter_valid(cpu_pmu, idx)) { idx 473 arch/nds32/kernel/perf_event_cpu.c pr_err("CPU disabling wrong pfm counter IRQ enable %d\n", idx); idx 485 arch/nds32/kernel/perf_event_cpu.c nds32_pfm_disable_counter(idx); idx 490 arch/nds32/kernel/perf_event_cpu.c nds32_pfm_disable_intens(idx); idx 499 arch/nds32/kernel/perf_event_cpu.c int idx = hwc->idx; idx 502 arch/nds32/kernel/perf_event_cpu.c if (!nds32_pfm_counter_valid(cpu_pmu, idx)) { idx 503 arch/nds32/kernel/perf_event_cpu.c pr_err("CPU reading wrong counter %d\n", idx); idx 505 arch/nds32/kernel/perf_event_cpu.c switch (idx) { idx 518 arch/nds32/kernel/perf_event_cpu.c __func__, idx); idx 528 arch/nds32/kernel/perf_event_cpu.c int idx = hwc->idx; idx 530 arch/nds32/kernel/perf_event_cpu.c if (!nds32_pfm_counter_valid(cpu_pmu, idx)) { idx 531 arch/nds32/kernel/perf_event_cpu.c pr_err("CPU writing wrong counter %d\n", idx); idx 533 arch/nds32/kernel/perf_event_cpu.c switch (idx) { idx 546 arch/nds32/kernel/perf_event_cpu.c __func__, idx); idx 554 arch/nds32/kernel/perf_event_cpu.c int idx; idx 569 arch/nds32/kernel/perf_event_cpu.c idx = get_converted_event_idx(evtype); idx 574 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(idx, cpuc->used_mask)) idx 575 arch/nds32/kernel/perf_event_cpu.c return idx; idx 581 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(idx, cpuc->used_mask)) idx 582 arch/nds32/kernel/perf_event_cpu.c return idx; idx 589 arch/nds32/kernel/perf_event_cpu.c if (!test_and_set_bit(idx, cpuc->used_mask)) idx 590 arch/nds32/kernel/perf_event_cpu.c return idx; idx 816 arch/nds32/kernel/perf_event_cpu.c hwc->idx = -1; idx 911 arch/nds32/kernel/perf_event_cpu.c int idx; idx 917 arch/nds32/kernel/perf_event_cpu.c idx = nds32_pmu->get_event_idx(hw_events, event); idx 918 arch/nds32/kernel/perf_event_cpu.c if (idx < 0) { idx 919 arch/nds32/kernel/perf_event_cpu.c err = idx; idx 927 arch/nds32/kernel/perf_event_cpu.c event->hw.idx = idx; idx 929 arch/nds32/kernel/perf_event_cpu.c hw_events->events[idx] = event; idx 989 arch/nds32/kernel/perf_event_cpu.c int idx = hwc->idx; idx 992 arch/nds32/kernel/perf_event_cpu.c hw_events->events[idx] = NULL; idx 993 arch/nds32/kernel/perf_event_cpu.c clear_bit(idx, hw_events->used_mask); idx 155 arch/nds32/mm/alignment.c static inline unsigned long *idx_to_addr(struct pt_regs *regs, int idx) idx 158 arch/nds32/mm/alignment.c if (idx >= 0 && idx <= 25) /* R0-R25 */ idx 159 arch/nds32/mm/alignment.c return ®s->uregs[0] + idx; idx 160 arch/nds32/mm/alignment.c else if (idx >= 28 && idx <= 30) /* FP, GP, LP */ idx 161 arch/nds32/mm/alignment.c return ®s->fp + (idx - 28); idx 162 arch/nds32/mm/alignment.c else if (idx == 31) /* SP */ idx 37 arch/nds32/mm/highmem.c unsigned int idx; idx 49 arch/nds32/mm/highmem.c idx = type + KM_TYPE_NR * smp_processor_id(); idx 50 arch/nds32/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 255 arch/nds32/mm/init.c void __set_fixmap(enum fixed_addresses idx, idx 258 arch/nds32/mm/init.c unsigned long addr = __fix_to_virt(idx); idx 261 arch/nds32/mm/init.c BUG_ON(idx <= FIX_HOLE || idx >= __end_of_fixed_addresses); idx 61 arch/openrisc/include/asm/fixmap.h static __always_inline unsigned long fix_to_virt(const unsigned int idx) idx 72 arch/openrisc/include/asm/fixmap.h if (idx >= __end_of_fixed_addresses) idx 75 arch/openrisc/include/asm/fixmap.h return __fix_to_virt(idx); idx 44 arch/parisc/include/asm/fixmap.h void set_fixmap(enum fixed_addresses idx, phys_addr_t phys); idx 45 arch/parisc/include/asm/fixmap.h void clear_fixmap(enum fixed_addresses idx); idx 225 arch/parisc/kernel/pci-dma.c #define PCXL_SEARCH_LOOP(idx, mask, size) \ idx 230 arch/parisc/kernel/pci-dma.c idx = (int)((u_long)res_ptr - (u_long)pcxl_res_map); \ idx 231 arch/parisc/kernel/pci-dma.c pcxl_res_hint = idx + (size >> 3); \ idx 236 arch/parisc/kernel/pci-dma.c #define PCXL_FIND_FREE_MAPPING(idx, mask, size) { \ idx 239 arch/parisc/kernel/pci-dma.c PCXL_SEARCH_LOOP(idx, mask, size); \ idx 241 arch/parisc/kernel/pci-dma.c PCXL_SEARCH_LOOP(idx, mask, size); \ idx 292 arch/parisc/kernel/pci-dma.c #define PCXL_FREE_MAPPINGS(idx, m, size) \ idx 293 arch/parisc/kernel/pci-dma.c u##size *res_ptr = (u##size *)&(pcxl_res_map[(idx) + (((size >> 3) - 1) & (~((size >> 3) - 1)))]); \ idx 13 arch/parisc/mm/fixmap.c void notrace set_fixmap(enum fixed_addresses idx, phys_addr_t phys) idx 15 arch/parisc/mm/fixmap.c unsigned long vaddr = __fix_to_virt(idx); idx 31 arch/parisc/mm/fixmap.c void notrace clear_fixmap(enum fixed_addresses idx) idx 33 arch/parisc/mm/fixmap.c unsigned long vaddr = __fix_to_virt(idx); idx 97 arch/powerpc/include/asm/book3s/64/hugetlb.h unsigned long idx = (addr & ((1UL << pdshift) - 1)) >> hugepd_shift(hpd); idx 99 arch/powerpc/include/asm/book3s/64/hugetlb.h return hugepd_page(hpd) + idx; idx 77 arch/powerpc/include/asm/fixmap.h static inline void __set_fixmap(enum fixed_addresses idx, idx 80 arch/powerpc/include/asm/fixmap.h if (__builtin_constant_p(idx)) idx 81 arch/powerpc/include/asm/fixmap.h BUILD_BUG_ON(idx >= __end_of_fixed_addresses); idx 82 arch/powerpc/include/asm/fixmap.h else if (WARN_ON(idx >= __end_of_fixed_addresses)) idx 85 arch/powerpc/include/asm/fixmap.h map_kernel_page(__fix_to_virt(idx), phys, flags); idx 22 arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h unsigned long idx = (addr & ((1UL << pdshift) - 1)) >> PAGE_SHIFT; idx 24 arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h return hugepd_page(hpd) + idx; idx 1435 arch/powerpc/include/asm/reg.h static inline void mtsrin(u32 val, u32 idx) idx 1437 arch/powerpc/include/asm/reg.h asm volatile("mtsrin %0, %1" : : "r" (val), "r" (idx)); idx 75 arch/powerpc/include/asm/xive.h u32 idx; idx 1163 arch/powerpc/kernel/fadump.c int i, j, idx; idx 1171 arch/powerpc/kernel/fadump.c idx = i; idx 1173 arch/powerpc/kernel/fadump.c if (mem_ranges[idx].base > mem_ranges[j].base) idx 1174 arch/powerpc/kernel/fadump.c idx = j; idx 1176 arch/powerpc/kernel/fadump.c if (idx != i) { idx 1177 arch/powerpc/kernel/fadump.c tmp_range = mem_ranges[idx]; idx 1178 arch/powerpc/kernel/fadump.c mem_ranges[idx] = mem_ranges[i]; idx 1184 arch/powerpc/kernel/fadump.c idx = 0; idx 1189 arch/powerpc/kernel/fadump.c mem_ranges[idx].size += mem_ranges[i].size; idx 1191 arch/powerpc/kernel/fadump.c idx++; idx 1192 arch/powerpc/kernel/fadump.c if (i == idx) idx 1195 arch/powerpc/kernel/fadump.c mem_ranges[idx] = mem_ranges[i]; idx 1198 arch/powerpc/kernel/fadump.c mrange_info->mem_range_cnt = idx + 1; idx 449 arch/powerpc/kernel/irq.c unsigned int idx = (srr1 & SRR1_WAKEMASK_P8) >> 18; idx 450 arch/powerpc/kernel/irq.c u8 reason = srr1_to_lazyirq[idx]; idx 1230 arch/powerpc/kernel/pci-common.c static inline void alloc_resource(struct pci_dev *dev, int idx) idx 1232 arch/powerpc/kernel/pci-common.c struct resource *pr, *r = &dev->resource[idx]; idx 1235 arch/powerpc/kernel/pci-common.c pci_name(dev), idx, r); idx 1241 arch/powerpc/kernel/pci-common.c " of device %s, will remap\n", idx, pci_name(dev)); idx 1254 arch/powerpc/kernel/pci-common.c int idx, disabled; idx 1260 arch/powerpc/kernel/pci-common.c for (idx = 0; idx <= PCI_ROM_RESOURCE; idx++) { idx 1261 arch/powerpc/kernel/pci-common.c r = &dev->resource[idx]; idx 1269 arch/powerpc/kernel/pci-common.c if (idx == PCI_ROM_RESOURCE ) idx 1276 arch/powerpc/kernel/pci-common.c alloc_resource(dev, idx); idx 764 arch/powerpc/kernel/smp.c int idx = group_start + j; idx 766 arch/powerpc/kernel/smp.c if (tg->thread_list[idx] == hw_cpu_id) idx 439 arch/powerpc/kvm/book3s_64_mmu_hv.c int idx, ret; idx 441 arch/powerpc/kvm/book3s_64_mmu_hv.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 444 arch/powerpc/kvm/book3s_64_mmu_hv.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 1241 arch/powerpc/kvm/book3s_64_mmu_hv.c unsigned long idx) idx 1256 arch/powerpc/kvm/book3s_64_mmu_hv.c hptep = (__be64 *)(old->virt + (idx << 4)); idx 1281 arch/powerpc/kvm/book3s_64_mmu_hv.c rev = &old->rev[idx]; idx 1300 arch/powerpc/kvm/book3s_64_mmu_hv.c kvmppc_unmap_hpte(kvm, idx, memslot, rmapp, gfn); idx 1325 arch/powerpc/kvm/book3s_64_mmu_hv.c pteg = idx / HPTES_PER_GROUP; idx 1356 arch/powerpc/kvm/book3s_64_mmu_hv.c new_idx = new_pteg * HPTES_PER_GROUP + (idx % HPTES_PER_GROUP); idx 389 arch/powerpc/kvm/book3s_64_vio.c unsigned long idx, unsigned long tce) idx 395 arch/powerpc/kvm/book3s_64_vio.c idx -= stt->offset; idx 396 arch/powerpc/kvm/book3s_64_vio.c sttpage = idx / TCES_PER_PAGE; idx 410 arch/powerpc/kvm/book3s_64_vio.c tbl[idx % TCES_PER_PAGE] = tce; idx 544 arch/powerpc/kvm/book3s_64_vio.c long ret, idx; idx 560 arch/powerpc/kvm/book3s_64_vio.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 594 arch/powerpc/kvm/book3s_64_vio.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 605 arch/powerpc/kvm/book3s_64_vio.c long i, ret = H_SUCCESS, idx; idx 630 arch/powerpc/kvm/book3s_64_vio.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 691 arch/powerpc/kvm/book3s_64_vio.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 172 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long idx, unsigned long tce) idx 177 arch/powerpc/kvm/book3s_64_vio_hv.c idx -= stt->offset; idx 178 arch/powerpc/kvm/book3s_64_vio_hv.c page = stt->pages[idx / TCES_PER_PAGE]; idx 186 arch/powerpc/kvm/book3s_64_vio_hv.c tbl[idx % TCES_PER_PAGE] = tce; idx 198 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long i, idx, sttpage, sttpages; idx 210 arch/powerpc/kvm/book3s_64_vio_hv.c idx = (ioba >> stt->page_shift) - stt->offset; idx 211 arch/powerpc/kvm/book3s_64_vio_hv.c sttpage = idx / TCES_PER_PAGE; idx 212 arch/powerpc/kvm/book3s_64_vio_hv.c sttpages = _ALIGN_UP(idx % TCES_PER_PAGE + npages, TCES_PER_PAGE) / idx 658 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long idx; idx 670 arch/powerpc/kvm/book3s_64_vio_hv.c idx = (ioba >> stt->page_shift) - stt->offset; idx 671 arch/powerpc/kvm/book3s_64_vio_hv.c page = stt->pages[idx / TCES_PER_PAGE]; idx 678 arch/powerpc/kvm/book3s_64_vio_hv.c vcpu->arch.regs.gpr[4] = tbl[idx % TCES_PER_PAGE]; idx 915 arch/powerpc/kvm/book3s_hv.c int idx, rc; idx 959 arch/powerpc/kvm/book3s_hv.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 961 arch/powerpc/kvm/book3s_hv.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 1164 arch/powerpc/kvm/book3s_pr.c int idx = srcu_read_lock(&vcpu->kvm->srcu); idx 1166 arch/powerpc/kvm/book3s_pr.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 1214 arch/powerpc/kvm/book3s_pr.c int idx = srcu_read_lock(&vcpu->kvm->srcu); idx 1216 arch/powerpc/kvm/book3s_pr.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 356 arch/powerpc/kvm/book3s_pr_papr.c int rc, idx; idx 399 arch/powerpc/kvm/book3s_pr_papr.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 401 arch/powerpc/kvm/book3s_pr_papr.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 1191 arch/powerpc/kvm/book3s_xics.c u16 idx; idx 1195 arch/powerpc/kvm/book3s_xics.c ics = kvmppc_xics_find_ics(xics, irq, &idx); idx 1199 arch/powerpc/kvm/book3s_xics.c irqp = &ics->irq_state[idx]; idx 1240 arch/powerpc/kvm/book3s_xics.c u16 idx; idx 1249 arch/powerpc/kvm/book3s_xics.c ics = kvmppc_xics_find_ics(xics, irq, &idx); idx 1255 arch/powerpc/kvm/book3s_xics.c irqp = &ics->irq_state[idx]; idx 1442 arch/powerpc/kvm/book3s_xics.c u16 idx; idx 1444 arch/powerpc/kvm/book3s_xics.c ics = kvmppc_xics_find_ics(xics, irq, &idx); idx 1448 arch/powerpc/kvm/book3s_xics.c ics->irq_state[idx].host_irq = host_irq; idx 1449 arch/powerpc/kvm/book3s_xics.c ics->irq_state[idx].intr_cpu = -1; idx 1458 arch/powerpc/kvm/book3s_xics.c u16 idx; idx 1460 arch/powerpc/kvm/book3s_xics.c ics = kvmppc_xics_find_ics(xics, irq, &idx); idx 1464 arch/powerpc/kvm/book3s_xics.c ics->irq_state[idx].host_irq = 0; idx 621 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 641 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 644 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 720 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 725 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 728 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 742 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 747 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 750 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 781 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 786 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 789 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 806 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 808 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 811 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 905 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 914 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, guest_irq, &idx); idx 917 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 1000 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 1009 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, guest_irq, &idx); idx 1012 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 1353 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 1355 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 1359 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 1425 arch/powerpc/kvm/book3s_xive.c u32 idx = q->idx; idx 1430 arch/powerpc/kvm/book3s_xive.c irq = __xive_read_eq(q->qpage, q->msk, &idx, &toggle); idx 1500 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 1502 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 1506 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 1642 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 1654 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 1663 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 1806 arch/powerpc/kvm/book3s_xive.c u16 idx; idx 1811 arch/powerpc/kvm/book3s_xive.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 1816 arch/powerpc/kvm/book3s_xive.c state = &sb->irq_state[idx]; idx 2051 arch/powerpc/kvm/book3s_xive.c u32 i0, i1, idx; idx 2059 arch/powerpc/kvm/book3s_xive.c idx = q->idx; idx 2060 arch/powerpc/kvm/book3s_xive.c i0 = be32_to_cpup(q->qpage + idx); idx 2061 arch/powerpc/kvm/book3s_xive.c idx = (idx + 1) & q->msk; idx 2062 arch/powerpc/kvm/book3s_xive.c i1 = be32_to_cpup(q->qpage + idx); idx 256 arch/powerpc/kvm/book3s_xive.h static inline u32 __xive_read_eq(__be32 *qpage, u32 msk, u32 *idx, u32 *toggle) idx 262 arch/powerpc/kvm/book3s_xive.h cur = be32_to_cpup(qpage + *idx); idx 265 arch/powerpc/kvm/book3s_xive.h *idx = (*idx + 1) & msk; idx 266 arch/powerpc/kvm/book3s_xive.h if (*idx == 0) idx 338 arch/powerpc/kvm/book3s_xive_native.c u16 idx; idx 346 arch/powerpc/kvm/book3s_xive_native.c sb = kvmppc_xive_find_source(xive, irq, &idx); idx 355 arch/powerpc/kvm/book3s_xive_native.c state = &sb->irq_state[idx]; idx 121 arch/powerpc/kvm/book3s_xive_template.c u32 idx, toggle; idx 141 arch/powerpc/kvm/book3s_xive_template.c idx = q->idx; idx 157 arch/powerpc/kvm/book3s_xive_template.c hirq = __xive_read_eq(qpage, q->msk, &idx, &toggle); idx 177 arch/powerpc/kvm/book3s_xive_template.c q->idx = idx; idx 224 arch/powerpc/kvm/book3s_xive_template.c q->idx = idx; idx 383 arch/powerpc/kvm/book3s_xive_template.c u32 idx, toggle, entry, irq, hw_num; idx 388 arch/powerpc/kvm/book3s_xive_template.c idx = q->idx; idx 396 arch/powerpc/kvm/book3s_xive_template.c entry = be32_to_cpup(qpage + idx); idx 419 arch/powerpc/kvm/book3s_xive_template.c qpage[idx] = cpu_to_be32((entry & 0x80000000) | XICS_DUMMY); idx 432 arch/powerpc/kvm/book3s_xive_template.c idx = (idx + 1) & q->msk; idx 433 arch/powerpc/kvm/book3s_xive_template.c if (idx == 0) idx 989 arch/powerpc/kvm/booke.c int idx; idx 1265 arch/powerpc/kvm/booke.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 1289 arch/powerpc/kvm/booke.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 1313 arch/powerpc/kvm/booke.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 1331 arch/powerpc/kvm/booke.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 396 arch/powerpc/kvm/e500_mmu.c int idx; idx 431 arch/powerpc/kvm/e500_mmu.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 447 arch/powerpc/kvm/e500_mmu.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 544 arch/powerpc/kvm/e500_mmu_host.c unsigned int idx = vcpu_e500->h2g_tlb1_rmap[sesel] - 1; idx 545 arch/powerpc/kvm/e500_mmu_host.c vcpu_e500->g2h_tlb1_map[idx] &= ~(1ULL << sesel); idx 126 arch/powerpc/kvm/mpic.c u32 val, int idx); idx 128 arch/powerpc/kvm/mpic.c u32 *ptr, int idx); idx 730 arch/powerpc/kvm/mpic.c int idx; idx 731 arch/powerpc/kvm/mpic.c idx = (addr - 0x10A0) >> 4; idx 732 arch/powerpc/kvm/mpic.c write_IRQreg_ivpr(opp, opp->irq_ipi0 + idx, val); idx 789 arch/powerpc/kvm/mpic.c int idx; idx 790 arch/powerpc/kvm/mpic.c idx = (addr - 0x10A0) >> 4; idx 791 arch/powerpc/kvm/mpic.c retval = read_IRQreg_ivpr(opp, opp->irq_ipi0 + idx); idx 810 arch/powerpc/kvm/mpic.c int idx; idx 824 arch/powerpc/kvm/mpic.c idx = (addr >> 6) & 0x3; idx 831 arch/powerpc/kvm/mpic.c if ((opp->timers[idx].tccr & TCCR_TOG) != 0 && idx 833 arch/powerpc/kvm/mpic.c (opp->timers[idx].tbcr & TBCR_CI) != 0) idx 834 arch/powerpc/kvm/mpic.c opp->timers[idx].tccr &= ~TCCR_TOG; idx 836 arch/powerpc/kvm/mpic.c opp->timers[idx].tbcr = val; idx 839 arch/powerpc/kvm/mpic.c write_IRQreg_ivpr(opp, opp->irq_tim0 + idx, val); idx 842 arch/powerpc/kvm/mpic.c write_IRQreg_idr(opp, opp->irq_tim0 + idx, val); idx 853 arch/powerpc/kvm/mpic.c int idx; idx 859 arch/powerpc/kvm/mpic.c idx = (addr >> 6) & 0x3; idx 868 arch/powerpc/kvm/mpic.c retval = opp->timers[idx].tccr; idx 871 arch/powerpc/kvm/mpic.c retval = opp->timers[idx].tbcr; idx 874 arch/powerpc/kvm/mpic.c retval = read_IRQreg_ivpr(opp, opp->irq_tim0 + idx); idx 877 arch/powerpc/kvm/mpic.c retval = read_IRQreg_idr(opp, opp->irq_tim0 + idx); idx 890 arch/powerpc/kvm/mpic.c int idx; idx 895 arch/powerpc/kvm/mpic.c idx = addr >> 5; idx 899 arch/powerpc/kvm/mpic.c write_IRQreg_ivpr(opp, idx, val); idx 902 arch/powerpc/kvm/mpic.c write_IRQreg_idr(opp, idx, val); idx 905 arch/powerpc/kvm/mpic.c write_IRQreg_ilr(opp, idx, val); idx 916 arch/powerpc/kvm/mpic.c int idx; idx 922 arch/powerpc/kvm/mpic.c idx = addr >> 5; idx 926 arch/powerpc/kvm/mpic.c retval = read_IRQreg_ivpr(opp, idx); idx 929 arch/powerpc/kvm/mpic.c retval = read_IRQreg_idr(opp, idx); idx 932 arch/powerpc/kvm/mpic.c retval = read_IRQreg_ilr(opp, idx); idx 944 arch/powerpc/kvm/mpic.c int idx = opp->irq_msi; idx 954 arch/powerpc/kvm/mpic.c idx += srs; idx 957 arch/powerpc/kvm/mpic.c openpic_set_irq(opp, idx, 1); idx 1025 arch/powerpc/kvm/mpic.c u32 val, int idx) idx 1032 arch/powerpc/kvm/mpic.c pr_debug("%s: cpu %d addr %#llx <= 0x%08x\n", __func__, idx, idx 1035 arch/powerpc/kvm/mpic.c if (idx < 0) idx 1041 arch/powerpc/kvm/mpic.c dst = &opp->dst[idx]; idx 1048 arch/powerpc/kvm/mpic.c idx = (addr - 0x40) >> 4; idx 1050 arch/powerpc/kvm/mpic.c opp->src[opp->irq_ipi0 + idx].destmask |= val; idx 1051 arch/powerpc/kvm/mpic.c openpic_set_irq(opp, opp->irq_ipi0 + idx, 1); idx 1052 arch/powerpc/kvm/mpic.c openpic_set_irq(opp, opp->irq_ipi0 + idx, 0); idx 1058 arch/powerpc/kvm/mpic.c __func__, idx, dst->ctpr, dst->raised.priority, idx 1063 arch/powerpc/kvm/mpic.c __func__, idx); idx 1067 arch/powerpc/kvm/mpic.c __func__, idx, dst->raised.next); idx 1102 arch/powerpc/kvm/mpic.c idx, n_IRQ); idx 1192 arch/powerpc/kvm/mpic.c u32 *ptr, int idx) idx 1198 arch/powerpc/kvm/mpic.c pr_debug("%s: cpu %d addr %#llx\n", __func__, idx, addr); idx 1201 arch/powerpc/kvm/mpic.c if (idx < 0) idx 1207 arch/powerpc/kvm/mpic.c dst = &opp->dst[idx]; idx 1214 arch/powerpc/kvm/mpic.c retval = idx; idx 1217 arch/powerpc/kvm/mpic.c retval = openpic_iack(opp, dst, idx); idx 1226 arch/powerpc/kvm/powerpc.c int idx, ret; idx 1251 arch/powerpc/kvm/powerpc.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 1256 arch/powerpc/kvm/powerpc.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 1314 arch/powerpc/kvm/powerpc.c int idx, ret; idx 1355 arch/powerpc/kvm/powerpc.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 1360 arch/powerpc/kvm/powerpc.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 1089 arch/powerpc/lib/sstep.c unsigned char perm, idx; idx 1094 arch/powerpc/lib/sstep.c idx = (v1 >> (i * 8)) & 0xff; idx 1095 arch/powerpc/lib/sstep.c if (idx < 64) idx 1096 arch/powerpc/lib/sstep.c if (v2 & PPC_BIT(idx)) idx 232 arch/powerpc/math-emu/math.c int idx = 0; idx 351 arch/powerpc/math-emu/math.c idx = (insn >> 16) & 0x1f; idx 354 arch/powerpc/math-emu/math.c op1 = (void *)((idx ? regs->gpr[idx] : 0) + sdisp); idx 358 arch/powerpc/math-emu/math.c idx = (insn >> 16) & 0x1f; idx 359 arch/powerpc/math-emu/math.c if (!idx) idx 364 arch/powerpc/math-emu/math.c op1 = (void *)(regs->gpr[idx] + sdisp); idx 382 arch/powerpc/math-emu/math.c idx = (insn >> 16) & 0x1f; idx 384 arch/powerpc/math-emu/math.c op1 = (void *)((idx ? regs->gpr[idx] : 0) idx 389 arch/powerpc/math-emu/math.c idx = (insn >> 16) & 0x1f; idx 390 arch/powerpc/math-emu/math.c if (!idx) idx 393 arch/powerpc/math-emu/math.c op1 = (void *)(regs->gpr[idx] idx 449 arch/powerpc/math-emu/math.c regs->gpr[idx] = (unsigned long)op1; idx 150 arch/powerpc/mm/book3s32/mmu.c int idx; idx 152 arch/powerpc/mm/book3s32/mmu.c while ((idx = find_free_bat()) != -1 && base != top) { idx 157 arch/powerpc/mm/book3s32/mmu.c setbat(idx, PAGE_OFFSET + base, base, size, PAGE_KERNEL_X); idx 417 arch/powerpc/mm/book3s64/hash_pgtable.c unsigned long idx; idx 432 arch/powerpc/mm/book3s64/hash_pgtable.c for (idx = start; idx < end; idx += step) idx 434 arch/powerpc/mm/book3s64/hash_pgtable.c mmu_hash_ops.hpte_updateboltedpp(newpp, idx, mmu_linear_psize, idx 388 arch/powerpc/mm/book3s64/hash_utils.c int idx = -1; idx 392 arch/powerpc/mm/book3s64/hash_utils.c idx = MMU_PAGE_4K; idx 395 arch/powerpc/mm/book3s64/hash_utils.c idx = MMU_PAGE_64K; idx 398 arch/powerpc/mm/book3s64/hash_utils.c idx = MMU_PAGE_1M; idx 401 arch/powerpc/mm/book3s64/hash_utils.c idx = MMU_PAGE_16M; idx 404 arch/powerpc/mm/book3s64/hash_utils.c idx = MMU_PAGE_16G; idx 407 arch/powerpc/mm/book3s64/hash_utils.c return idx; idx 434 arch/powerpc/mm/book3s64/hash_utils.c int idx, base_idx; idx 469 arch/powerpc/mm/book3s64/hash_utils.c idx = get_idx_from_shift(shift); idx 470 arch/powerpc/mm/book3s64/hash_utils.c if (idx < 0) idx 477 arch/powerpc/mm/book3s64/hash_utils.c def->penc[idx] = penc; idx 481 arch/powerpc/mm/book3s64/hash_utils.c def->avpnm, def->tlbiel, def->penc[idx]); idx 173 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long idx; idx 185 arch/powerpc/mm/book3s64/radix_pgtable.c for (idx = start; idx < end; idx += PAGE_SIZE) { idx 186 arch/powerpc/mm/book3s64/radix_pgtable.c pgdp = pgd_offset_k(idx); idx 187 arch/powerpc/mm/book3s64/radix_pgtable.c pudp = pud_alloc(&init_mm, pgdp, idx); idx 194 arch/powerpc/mm/book3s64/radix_pgtable.c pmdp = pmd_alloc(&init_mm, pudp, idx); idx 201 arch/powerpc/mm/book3s64/radix_pgtable.c ptep = pte_alloc_kernel(pmdp, idx); idx 205 arch/powerpc/mm/book3s64/radix_pgtable.c radix__pte_update(&init_mm, idx, ptep, clear, 0, 0); idx 404 arch/powerpc/mm/book3s64/radix_pgtable.c int idx = -1; idx 408 arch/powerpc/mm/book3s64/radix_pgtable.c idx = MMU_PAGE_4K; idx 411 arch/powerpc/mm/book3s64/radix_pgtable.c idx = MMU_PAGE_64K; idx 414 arch/powerpc/mm/book3s64/radix_pgtable.c idx = MMU_PAGE_2M; idx 417 arch/powerpc/mm/book3s64/radix_pgtable.c idx = MMU_PAGE_1G; idx 420 arch/powerpc/mm/book3s64/radix_pgtable.c return idx; idx 428 arch/powerpc/mm/book3s64/radix_pgtable.c int shift, idx; idx 457 arch/powerpc/mm/book3s64/radix_pgtable.c idx = get_idx_from_shift(shift); idx 458 arch/powerpc/mm/book3s64/radix_pgtable.c if (idx < 0) idx 461 arch/powerpc/mm/book3s64/radix_pgtable.c def = &mmu_psize_defs[idx]; idx 272 arch/powerpc/mm/book3s64/slb.c unsigned char idx; idx 274 arch/powerpc/mm/book3s64/slb.c idx = (ti->slb_preload_tail + i) % SLB_PRELOAD_NR; idx 275 arch/powerpc/mm/book3s64/slb.c if (esid == ti->slb_preload_esid[idx]) idx 283 arch/powerpc/mm/book3s64/slb.c unsigned char idx; idx 297 arch/powerpc/mm/book3s64/slb.c idx = (ti->slb_preload_tail + ti->slb_preload_nr) % SLB_PRELOAD_NR; idx 298 arch/powerpc/mm/book3s64/slb.c ti->slb_preload_esid[idx] = esid; idx 492 arch/powerpc/mm/book3s64/slb.c unsigned char idx; idx 495 arch/powerpc/mm/book3s64/slb.c idx = (ti->slb_preload_tail + i) % SLB_PRELOAD_NR; idx 496 arch/powerpc/mm/book3s64/slb.c ea = (unsigned long)ti->slb_preload_esid[idx] << SID_SHIFT; idx 36 arch/powerpc/mm/highmem.c int idx, type; idx 44 arch/powerpc/mm/highmem.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 45 arch/powerpc/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 46 arch/powerpc/mm/highmem.c WARN_ON(IS_ENABLED(CONFIG_DEBUG_HIGHMEM) && !pte_none(*(kmap_pte - idx))); idx 47 arch/powerpc/mm/highmem.c __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); idx 66 arch/powerpc/mm/highmem.c unsigned int idx; idx 68 arch/powerpc/mm/highmem.c idx = type + KM_TYPE_NR * smp_processor_id(); idx 69 arch/powerpc/mm/highmem.c WARN_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); idx 75 arch/powerpc/mm/highmem.c pte_clear(&init_mm, vaddr, kmap_pte-idx); idx 65 arch/powerpc/mm/nohash/fsl_booke.c unsigned long tlbcam_sz(int idx) idx 67 arch/powerpc/mm/nohash/fsl_booke.c return tlbcam_addrs[idx].limit - tlbcam_addrs[idx].start + 1; idx 24 arch/powerpc/mm/ptdump/bats.c static void bat_show_601(struct seq_file *m, int idx, u32 lower, u32 upper) idx 33 arch/powerpc/mm/ptdump/bats.c seq_printf(m, "%d: ", idx); idx 71 arch/powerpc/mm/ptdump/bats.c static void bat_show_603(struct seq_file *m, int idx, u32 lower, u32 upper, bool is_d) idx 79 arch/powerpc/mm/ptdump/bats.c seq_printf(m, "%d: ", idx); idx 282 arch/powerpc/mm/ptdump/hashpagetable.c int penc = -2, idx = 0, shift; idx 298 arch/powerpc/mm/ptdump/hashpagetable.c while (idx < MMU_PAGE_COUNT) { idx 299 arch/powerpc/mm/ptdump/hashpagetable.c penc = entry.penc[idx]; idx 300 arch/powerpc/mm/ptdump/hashpagetable.c if ((penc != -1) && (mmu_psize_defs[idx].shift)) { idx 301 arch/powerpc/mm/ptdump/hashpagetable.c shift = mmu_psize_defs[idx].shift - HPTE_R_RPN_SHIFT; idx 304 arch/powerpc/mm/ptdump/hashpagetable.c *aps = mmu_psize_to_shift(idx); idx 310 arch/powerpc/mm/ptdump/hashpagetable.c idx++; idx 31 arch/powerpc/net/bpf_jit.h #define PLANT_INSTR(d, idx, instr) \ idx 32 arch/powerpc/net/bpf_jit.h do { if (d) { (d)[idx] = instr; } idx++; } while (0) idx 33 arch/powerpc/net/bpf_jit.h #define EMIT(instr) PLANT_INSTR(image, ctx->idx, instr) idx 184 arch/powerpc/net/bpf_jit.h (((dest) - (ctx->idx * 4)) & 0x03fffffc)) idx 188 arch/powerpc/net/bpf_jit.h (((dest) - (ctx->idx * 4)) & \ idx 241 arch/powerpc/net/bpf_jit.h if (is_nearbranch((dest) - (ctx->idx * 4))) { \ idx 246 arch/powerpc/net/bpf_jit.h PPC_BCC_SHORT(cond ^ COND_CMP_TRUE, (ctx->idx+2)*4); \ idx 133 arch/powerpc/net/bpf_jit32.h unsigned int idx; idx 102 arch/powerpc/net/bpf_jit64.h unsigned int idx; idx 131 arch/powerpc/net/bpf_jit_comp.c addrs[i] = ctx->idx * 4; idx 176 arch/powerpc/net/bpf_jit_comp.c PPC_BCC_SHORT(COND_NE, (ctx->idx*4)+12); idx 354 arch/powerpc/net/bpf_jit_comp.c PPC_BCC_SHORT(COND_NE, ctx->idx * 4 + 12); idx 546 arch/powerpc/net/bpf_jit_comp.c addrs[i] = ctx->idx * 4; idx 619 arch/powerpc/net/bpf_jit_comp.c cgctx.idx = 0; idx 635 arch/powerpc/net/bpf_jit_comp.c proglen = cgctx.idx * 4; idx 646 arch/powerpc/net/bpf_jit_comp.c cgctx.idx = 0; idx 653 arch/powerpc/net/bpf_jit_comp.c proglen - (cgctx.idx * 4), cgctx.seen); idx 193 arch/powerpc/net/bpf_jit_comp64.c unsigned int i, ctx_idx = ctx->idx; idx 208 arch/powerpc/net/bpf_jit_comp64.c for (i = ctx->idx - ctx_idx; i < 5; i++) idx 319 arch/powerpc/net/bpf_jit_comp64.c addrs[i] = ctx->idx * 4; idx 505 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 514 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 523 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 531 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 569 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 634 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 690 arch/powerpc/net/bpf_jit_comp64.c tmp_idx = ctx->idx * 4; idx 703 arch/powerpc/net/bpf_jit_comp64.c tmp_idx = ctx->idx * 4; idx 717 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 723 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 729 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 744 arch/powerpc/net/bpf_jit_comp64.c addrs[++i] = ctx->idx * 4; idx 1008 arch/powerpc/net/bpf_jit_comp64.c addrs[i] = ctx->idx * 4; idx 1047 arch/powerpc/net/bpf_jit_comp64.c tmp_idx = ctx->idx; idx 1048 arch/powerpc/net/bpf_jit_comp64.c ctx->idx = addrs[i] / 4; idx 1055 arch/powerpc/net/bpf_jit_comp64.c ctx->idx = tmp_idx; idx 1150 arch/powerpc/net/bpf_jit_comp64.c cgctx.idx = 0; idx 1165 arch/powerpc/net/bpf_jit_comp64.c proglen = cgctx.idx * 4; idx 1197 arch/powerpc/net/bpf_jit_comp64.c cgctx.idx = 0; idx 1204 arch/powerpc/net/bpf_jit_comp64.c proglen - (cgctx.idx * 4), cgctx.seen); idx 729 arch/powerpc/perf/core-book3s.c static unsigned long read_pmc(int idx) idx 733 arch/powerpc/perf/core-book3s.c switch (idx) { idx 761 arch/powerpc/perf/core-book3s.c printk(KERN_ERR "oops trying to read PMC%d\n", idx); idx 770 arch/powerpc/perf/core-book3s.c static void write_pmc(int idx, unsigned long val) idx 772 arch/powerpc/perf/core-book3s.c switch (idx) { idx 800 arch/powerpc/perf/core-book3s.c printk(KERN_ERR "oops trying to write PMC%d\n", idx); idx 1060 arch/powerpc/perf/core-book3s.c if (!event->hw.idx) idx 1064 arch/powerpc/perf/core-book3s.c val = read_pmc(event->hw.idx); idx 1077 arch/powerpc/perf/core-book3s.c val = read_pmc(event->hw.idx); idx 1122 arch/powerpc/perf/core-book3s.c if (!event->hw.idx) idx 1124 arch/powerpc/perf/core-book3s.c val = (event->hw.idx == 5) ? pmc5 : pmc6; idx 1126 arch/powerpc/perf/core-book3s.c event->hw.idx = 0; idx 1142 arch/powerpc/perf/core-book3s.c event->hw.idx = cpuhw->limited_hwidx[i]; idx 1143 arch/powerpc/perf/core-book3s.c val = (event->hw.idx == 5) ? pmc5 : pmc6; idx 1284 arch/powerpc/perf/core-book3s.c int idx; idx 1367 arch/powerpc/perf/core-book3s.c if (event->hw.idx && event->hw.idx != hwc_index[i] + 1) { idx 1369 arch/powerpc/perf/core-book3s.c write_pmc(event->hw.idx, 0); idx 1370 arch/powerpc/perf/core-book3s.c event->hw.idx = 0; idx 1380 arch/powerpc/perf/core-book3s.c if (event->hw.idx) idx 1382 arch/powerpc/perf/core-book3s.c idx = hwc_index[i] + 1; idx 1383 arch/powerpc/perf/core-book3s.c if (is_limited_pmc(idx)) { idx 1385 arch/powerpc/perf/core-book3s.c cpuhw->limited_hwidx[n_lim] = idx; idx 1402 arch/powerpc/perf/core-book3s.c event->hw.idx = idx; idx 1405 arch/powerpc/perf/core-book3s.c write_pmc(idx, val); idx 1558 arch/powerpc/perf/core-book3s.c ppmu->disable_pmc(event->hw.idx - 1, cpuhw->mmcr); idx 1559 arch/powerpc/perf/core-book3s.c if (event->hw.idx) { idx 1560 arch/powerpc/perf/core-book3s.c write_pmc(event->hw.idx, 0); idx 1561 arch/powerpc/perf/core-book3s.c event->hw.idx = 0; idx 1600 arch/powerpc/perf/core-book3s.c if (!event->hw.idx || !event->hw.sample_period) idx 1619 arch/powerpc/perf/core-book3s.c write_pmc(event->hw.idx, val); idx 1630 arch/powerpc/perf/core-book3s.c if (!event->hw.idx || !event->hw.sample_period) idx 1641 arch/powerpc/perf/core-book3s.c write_pmc(event->hw.idx, 0); idx 1889 arch/powerpc/perf/core-book3s.c event->hw.idx = 0; idx 2004 arch/powerpc/perf/core-book3s.c return event->hw.idx; idx 2046 arch/powerpc/perf/core-book3s.c write_pmc(event->hw.idx, 0); idx 2075 arch/powerpc/perf/core-book3s.c write_pmc(event->hw.idx, val); idx 2213 arch/powerpc/perf/core-book3s.c if (event->hw.idx == (i + 1)) { idx 2227 arch/powerpc/perf/core-book3s.c if (!event->hw.idx || is_limited_pmc(event->hw.idx)) idx 2229 arch/powerpc/perf/core-book3s.c if (pmc_overflow_power7(val[event->hw.idx - 1])) { idx 2233 arch/powerpc/perf/core-book3s.c val[event->hw.idx - 1], idx 52 arch/powerpc/perf/core-fsl-emb.c static unsigned long read_pmc(int idx) idx 56 arch/powerpc/perf/core-fsl-emb.c switch (idx) { idx 76 arch/powerpc/perf/core-fsl-emb.c printk(KERN_ERR "oops trying to read PMC%d\n", idx); idx 85 arch/powerpc/perf/core-fsl-emb.c static void write_pmc(int idx, unsigned long val) idx 87 arch/powerpc/perf/core-fsl-emb.c switch (idx) { idx 107 arch/powerpc/perf/core-fsl-emb.c printk(KERN_ERR "oops trying to write PMC%d\n", idx); idx 116 arch/powerpc/perf/core-fsl-emb.c static void write_pmlca(int idx, unsigned long val) idx 118 arch/powerpc/perf/core-fsl-emb.c switch (idx) { idx 138 arch/powerpc/perf/core-fsl-emb.c printk(KERN_ERR "oops trying to write PMLCA%d\n", idx); idx 147 arch/powerpc/perf/core-fsl-emb.c static void write_pmlcb(int idx, unsigned long val) idx 149 arch/powerpc/perf/core-fsl-emb.c switch (idx) { idx 169 arch/powerpc/perf/core-fsl-emb.c printk(KERN_ERR "oops trying to write PMLCB%d\n", idx); idx 190 arch/powerpc/perf/core-fsl-emb.c val = read_pmc(event->hw.idx); idx 317 arch/powerpc/perf/core-fsl-emb.c event->hw.idx = i; idx 353 arch/powerpc/perf/core-fsl-emb.c int i = event->hw.idx; idx 363 arch/powerpc/perf/core-fsl-emb.c WARN_ON(event != cpuhw->event[event->hw.idx]); idx 370 arch/powerpc/perf/core-fsl-emb.c event->hw.idx = -1; idx 393 arch/powerpc/perf/core-fsl-emb.c if (event->hw.idx < 0 || !event->hw.sample_period) idx 410 arch/powerpc/perf/core-fsl-emb.c write_pmc(event->hw.idx, val); idx 421 arch/powerpc/perf/core-fsl-emb.c if (event->hw.idx < 0 || !event->hw.sample_period) idx 432 arch/powerpc/perf/core-fsl-emb.c write_pmc(event->hw.idx, 0); idx 547 arch/powerpc/perf/core-fsl-emb.c event->hw.idx = -1; idx 610 arch/powerpc/perf/core-fsl-emb.c write_pmc(event->hw.idx, 0); idx 637 arch/powerpc/perf/core-fsl-emb.c write_pmc(event->hw.idx, val); idx 1120 arch/powerpc/perf/hv-24x7.c u16 idx; idx 1134 arch/powerpc/perf/hv-24x7.c idx = event_get_chip(event); idx 1137 arch/powerpc/perf/hv-24x7.c idx = event_get_core(event); idx 1140 arch/powerpc/perf/hv-24x7.c idx = event_get_vcpu(event); idx 1153 arch/powerpc/perf/hv-24x7.c req->starting_ix = cpu_to_be16(idx); idx 1160 arch/powerpc/perf/hv-24x7.c req->starting_thread_group_ix = idx % 2; idx 782 arch/powerpc/perf/imc-pmu.c event->hw.idx = -1; idx 908 arch/powerpc/perf/imc-pmu.c event->hw.idx = -1; idx 1317 arch/powerpc/perf/imc-pmu.c event->hw.idx = -1; idx 155 arch/powerpc/perf/isa207-common.c static inline u64 isa207_find_source(u64 idx, u32 sub_idx) idx 159 arch/powerpc/perf/isa207-common.c switch(idx) { idx 206 arch/powerpc/perf/isa207-common.c u64 idx; idx 220 arch/powerpc/perf/isa207-common.c idx = (sier & ISA207_SIER_LDST_MASK) >> ISA207_SIER_LDST_SHIFT; idx 223 arch/powerpc/perf/isa207-common.c dsrc->val = isa207_find_source(idx, sub_idx); idx 72 arch/powerpc/perf/perf_regs.c u64 perf_reg_value(struct pt_regs *regs, int idx) idx 74 arch/powerpc/perf/perf_regs.c if (WARN_ON_ONCE(idx >= PERF_REG_POWERPC_MAX)) idx 77 arch/powerpc/perf/perf_regs.c if (idx == PERF_REG_POWERPC_SIER && idx 83 arch/powerpc/perf/perf_regs.c if (idx == PERF_REG_POWERPC_MMCRA && idx 88 arch/powerpc/perf/perf_regs.c return regs_get_register(regs, pt_regs_offset[idx]); idx 595 arch/powerpc/platforms/512x/clock-commonclk.c static void mpc512x_clk_setup_mclk(struct mclk_setup_data *entry, size_t idx) idx 604 arch/powerpc/platforms/512x/clock-commonclk.c clks_idx_pub = MPC512x_CLK_PSC0_MCLK + idx; idx 606 arch/powerpc/platforms/512x/clock-commonclk.c + (idx) * MCLK_MAX_IDX; idx 607 arch/powerpc/platforms/512x/clock-commonclk.c mccr_reg = &clkregs->psc_ccr[idx]; idx 610 arch/powerpc/platforms/512x/clock-commonclk.c clks_idx_pub = MPC512x_CLK_MSCAN0_MCLK + idx; idx 612 arch/powerpc/platforms/512x/clock-commonclk.c + (NR_PSCS + idx) * MCLK_MAX_IDX; idx 613 arch/powerpc/platforms/512x/clock-commonclk.c mccr_reg = &clkregs->mscan_ccr[idx]; idx 622 arch/powerpc/platforms/512x/clock-commonclk.c clks_idx_pub = MPC512x_CLK_OUT0_CLK + idx; idx 624 arch/powerpc/platforms/512x/clock-commonclk.c + (NR_PSCS + NR_MSCANS + NR_SPDIFS + idx) idx 626 arch/powerpc/platforms/512x/clock-commonclk.c mccr_reg = &clkregs->out_ccr[idx]; idx 894 arch/powerpc/platforms/512x/clock-commonclk.c size_t idx; /* used as mclk_idx, just to trim line length */ idx 895 arch/powerpc/platforms/512x/clock-commonclk.c for (idx = 0; idx < ARRAY_SIZE(mclk_outclk_data); idx++) idx 896 arch/powerpc/platforms/512x/clock-commonclk.c mpc512x_clk_setup_mclk(&mclk_outclk_data[idx], idx); idx 1030 arch/powerpc/platforms/512x/clock-commonclk.c int idx; idx 1037 arch/powerpc/platforms/512x/clock-commonclk.c idx = (res.start >> 8) & 0xf; idx 1038 arch/powerpc/platforms/512x/clock-commonclk.c NODE_CHK("ipg", clks[MPC512x_CLK_PSC0 + idx], 0, PSC); idx 1039 arch/powerpc/platforms/512x/clock-commonclk.c NODE_CHK("mclk", clks[MPC512x_CLK_PSC0_MCLK + idx], 0, PSC); idx 1054 arch/powerpc/platforms/512x/clock-commonclk.c idx = 0; idx 1055 arch/powerpc/platforms/512x/clock-commonclk.c idx += (res.start & 0x2000) ? 2 : 0; idx 1056 arch/powerpc/platforms/512x/clock-commonclk.c idx += (res.start & 0x0080) ? 1 : 0; idx 1058 arch/powerpc/platforms/512x/clock-commonclk.c NODE_CHK("mclk", clks[MPC512x_CLK_MSCAN0_MCLK + idx], 0, CAN); idx 1129 arch/powerpc/platforms/512x/clock-commonclk.c idx = MPC512x_CLK_FEC2; idx 1131 arch/powerpc/platforms/512x/clock-commonclk.c idx = MPC512x_CLK_FEC; idx 1132 arch/powerpc/platforms/512x/clock-commonclk.c NODE_CHK("per", clks[idx], 0, FEC); idx 1137 arch/powerpc/platforms/512x/clock-commonclk.c idx = (res.start & 0x4000) ? 1 : 0; idx 1138 arch/powerpc/platforms/512x/clock-commonclk.c NODE_CHK("ipg", clks[MPC512x_CLK_USB1 + idx], 0, USB); idx 97 arch/powerpc/platforms/cell/axon_msi.c int idx; idx 107 arch/powerpc/platforms/cell/axon_msi.c idx = msic->read_offset / sizeof(__le32); idx 108 arch/powerpc/platforms/cell/axon_msi.c msi = le32_to_cpu(msic->fifo_virt[idx]); idx 116 arch/powerpc/platforms/cell/axon_msi.c msic->fifo_virt[idx] = cpu_to_le32(0xffffffff); idx 618 arch/powerpc/platforms/cell/spufs/switch.c u64 idx, ch_indices[] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; idx 630 arch/powerpc/platforms/cell/spufs/switch.c idx = ch_indices[i]; idx 631 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnlcntptr_RW, idx); idx 633 arch/powerpc/platforms/cell/spufs/switch.c csa->spu_chnldata_RW[idx] = in_be64(&priv2->spu_chnldata_RW); idx 634 arch/powerpc/platforms/cell/spufs/switch.c csa->spu_chnlcnt_RW[idx] = in_be64(&priv2->spu_chnlcnt_RW); idx 677 arch/powerpc/platforms/cell/spufs/switch.c u64 idx; idx 684 arch/powerpc/platforms/cell/spufs/switch.c idx = ch_indices[i]; idx 685 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnlcntptr_RW, idx); idx 1069 arch/powerpc/platforms/cell/spufs/switch.c u64 idx; idx 1081 arch/powerpc/platforms/cell/spufs/switch.c idx = ch_indices[i]; idx 1082 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnlcntptr_RW, idx); idx 1095 arch/powerpc/platforms/cell/spufs/switch.c u64 idx; idx 1102 arch/powerpc/platforms/cell/spufs/switch.c idx = ch_indices[i]; idx 1103 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnlcntptr_RW, idx); idx 1560 arch/powerpc/platforms/cell/spufs/switch.c u64 idx, ch_indices[] = { 0UL, 3UL, 4UL, 24UL, 25UL, 27UL }; idx 1567 arch/powerpc/platforms/cell/spufs/switch.c idx = ch_indices[i]; idx 1568 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnlcntptr_RW, idx); idx 1570 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnldata_RW, csa->spu_chnldata_RW[idx]); idx 1571 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnlcnt_RW, csa->spu_chnlcnt_RW[idx]); idx 1581 arch/powerpc/platforms/cell/spufs/switch.c u64 idx; idx 1591 arch/powerpc/platforms/cell/spufs/switch.c idx = ch_indices[i]; idx 1592 arch/powerpc/platforms/cell/spufs/switch.c out_be64(&priv2->spu_chnlcntptr_RW, idx); idx 39 arch/powerpc/platforms/embedded6xx/wii.c #define HW_GPIO_BASE(idx) (idx * 0x20) idx 40 arch/powerpc/platforms/embedded6xx/wii.c #define HW_GPIO_OUT(idx) (HW_GPIO_BASE(idx) + 0) idx 41 arch/powerpc/platforms/embedded6xx/wii.c #define HW_GPIO_DIR(idx) (HW_GPIO_BASE(idx) + 4) idx 94 arch/powerpc/platforms/powernv/ocxl.c u8 idx; idx 99 arch/powerpc/platforms/powernv/ocxl.c &idx); idx 100 arch/powerpc/platforms/powernv/ocxl.c if (idx == afu_idx) idx 321 arch/powerpc/platforms/powernv/ocxl.c int shift, idx; idx 324 arch/powerpc/platforms/powernv/ocxl.c idx = (PNV_OCXL_TL_MAX_TEMPLATE - templ) / 2; idx 326 arch/powerpc/platforms/powernv/ocxl.c buf[idx] |= rate << shift; idx 140 arch/powerpc/platforms/powernv/opal-core.c int idx = 0; idx 145 arch/powerpc/platforms/powernv/opal-core.c bufp[idx++] = cpu_to_be64(AT_ENTRY); idx 146 arch/powerpc/platforms/powernv/opal-core.c bufp[idx++] = cpu_to_be64(opal_boot_entry); idx 149 arch/powerpc/platforms/powernv/opal-core.c bufp[idx++] = cpu_to_be64(AT_NULL); idx 452 arch/powerpc/platforms/powernv/opal-core.c u32 idx, cpu_data_version; idx 509 arch/powerpc/platforms/powernv/opal-core.c idx = be32_to_cpu(opalc_metadata->region_cnt); idx 510 arch/powerpc/platforms/powernv/opal-core.c if (idx > MAX_PT_LOAD_CNT) { idx 512 arch/powerpc/platforms/powernv/opal-core.c MAX_PT_LOAD_CNT, idx); idx 513 arch/powerpc/platforms/powernv/opal-core.c idx = MAX_PT_LOAD_CNT; idx 515 arch/powerpc/platforms/powernv/opal-core.c for (i = 0; i < idx; i++) { idx 55 arch/powerpc/platforms/powernv/pci-ioda-tce.c static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc) idx 63 arch/powerpc/platforms/powernv/pci-ioda-tce.c int n = (idx & mask) >> (level * shift); idx 88 arch/powerpc/platforms/powernv/pci-ioda-tce.c idx &= ~mask; idx 93 arch/powerpc/platforms/powernv/pci-ioda-tce.c return tmp + idx; idx 110 arch/powerpc/platforms/powernv/pci-ioda-tce.c unsigned long idx = index - tbl->it_offset + i; idx 112 arch/powerpc/platforms/powernv/pci-ioda-tce.c *(pnv_tce(tbl, false, idx, true)) = cpu_to_be64(newtce); idx 125 arch/powerpc/platforms/powernv/pci-ioda-tce.c unsigned long idx = index - tbl->it_offset; idx 131 arch/powerpc/platforms/powernv/pci-ioda-tce.c ptce = pnv_tce(tbl, false, idx, false); idx 139 arch/powerpc/platforms/powernv/pci-ioda-tce.c ptce = pnv_tce(tbl, false, idx, alloc); idx 168 arch/powerpc/platforms/powernv/pci-ioda-tce.c unsigned long idx = index - tbl->it_offset + i; idx 169 arch/powerpc/platforms/powernv/pci-ioda-tce.c __be64 *ptce = pnv_tce(tbl, false, idx, false); idx 3424 arch/powerpc/platforms/powernv/pci-ioda.c unsigned int idx; idx 3428 arch/powerpc/platforms/powernv/pci-ioda.c for (idx = 0; idx < phb->ioda.dma32_count; idx++) { idx 3429 arch/powerpc/platforms/powernv/pci-ioda.c if (phb->ioda.dma32_segmap[idx] != pe->pe_number) idx 3433 arch/powerpc/platforms/powernv/pci-ioda.c idx, 0, 0ul, 0ul, 0ul); idx 3436 arch/powerpc/platforms/powernv/pci-ioda.c rc, idx); idx 3440 arch/powerpc/platforms/powernv/pci-ioda.c phb->ioda.dma32_segmap[idx] = IODA_INVALID_PE; idx 3501 arch/powerpc/platforms/powernv/pci-ioda.c int idx; idx 3504 arch/powerpc/platforms/powernv/pci-ioda.c for (idx = 0; idx < phb->ioda.total_pe_num; idx++) { idx 3505 arch/powerpc/platforms/powernv/pci-ioda.c if (map[idx] != pe->pe_number) idx 3511 arch/powerpc/platforms/powernv/pci-ioda.c idx / PNV_IODA1_M64_SEGS, idx 3512 arch/powerpc/platforms/powernv/pci-ioda.c idx % PNV_IODA1_M64_SEGS); idx 3515 arch/powerpc/platforms/powernv/pci-ioda.c phb->ioda.reserved_pe_idx, win, 0, idx); idx 3519 arch/powerpc/platforms/powernv/pci-ioda.c rc, win, idx); idx 3521 arch/powerpc/platforms/powernv/pci-ioda.c map[idx] = IODA_INVALID_PE; idx 397 arch/powerpc/platforms/ps3/os-area.c struct db_index *idx; idx 425 arch/powerpc/platforms/ps3/os-area.c i->idx = (void *)db + db->index_64; idx 426 arch/powerpc/platforms/ps3/os-area.c i->last_idx = i->idx + db->count_64; idx 430 arch/powerpc/platforms/ps3/os-area.c i->idx++; idx 434 arch/powerpc/platforms/ps3/os-area.c if (i->idx >= i->last_idx) { idx 440 arch/powerpc/platforms/ps3/os-area.c && i->match_id.owner != (int)i->idx->owner) idx 443 arch/powerpc/platforms/ps3/os-area.c && i->match_id.key != (int)i->idx->key) idx 456 arch/powerpc/platforms/ps3/os-area.c i.idx->owner, i.idx->key, idx 459 arch/powerpc/platforms/ps3/os-area.c i.idx->owner = 0; idx 460 arch/powerpc/platforms/ps3/os-area.c i.idx->key = 0; idx 487 arch/powerpc/platforms/ps3/os-area.c i.idx->owner, i.idx->key, idx 490 arch/powerpc/platforms/ps3/os-area.c i.idx->owner = id->owner; idx 491 arch/powerpc/platforms/ps3/os-area.c i.idx->key = id->key; idx 495 arch/powerpc/platforms/ps3/os-area.c i.idx->owner, i.idx->key, idx 502 arch/powerpc/platforms/pseries/cmm.c unsigned long idx; idx 509 arch/powerpc/platforms/pseries/cmm.c for (idx = 0; idx < pa->index; idx++) idx 510 arch/powerpc/platforms/pseries/cmm.c if (pa->page[idx] >= start && pa->page[idx] < end) idx 556 arch/powerpc/platforms/pseries/cmm.c unsigned long idx; idx 566 arch/powerpc/platforms/pseries/cmm.c for (idx = (pa_curr->index - 1); (idx + 1) > 0; idx--) { idx 567 arch/powerpc/platforms/pseries/cmm.c if ((pa_curr->page[idx] < start_page) || idx 568 arch/powerpc/platforms/pseries/cmm.c (pa_curr->page[idx] >= end_page)) idx 571 arch/powerpc/platforms/pseries/cmm.c plpar_page_set_active(__pa(pa_curr->page[idx])); idx 572 arch/powerpc/platforms/pseries/cmm.c free_page(pa_curr->page[idx]); idx 576 arch/powerpc/platforms/pseries/cmm.c pa_curr->page[idx] = pa_last->page[--pa_last->index]; idx 1023 arch/powerpc/platforms/pseries/lpar.c static unsigned long call_block_remove(unsigned long idx, unsigned long *param, idx 1029 arch/powerpc/platforms/pseries/lpar.c if (idx < 2) { idx 1035 arch/powerpc/platforms/pseries/lpar.c if (idx > PLPAR_HCALL9_BUFSIZE) { idx 1036 arch/powerpc/platforms/pseries/lpar.c pr_err("Too many PTEs (%lu) for H_BLOCK_REMOVE", idx); idx 1037 arch/powerpc/platforms/pseries/lpar.c idx = PLPAR_HCALL9_BUFSIZE; idx 1038 arch/powerpc/platforms/pseries/lpar.c } else if (idx < PLPAR_HCALL9_BUFSIZE) idx 1039 arch/powerpc/platforms/pseries/lpar.c param[idx] = HBR_END; idx 1051 arch/powerpc/platforms/pseries/lpar.c for (i = 0; i < idx-1; i++) { idx 1068 arch/powerpc/platforms/pseries/lpar.c idx = new_idx + 1; idx 1420 arch/powerpc/platforms/pseries/lpar.c int call_status, len, idx, bpsize; idx 1452 arch/powerpc/platforms/pseries/lpar.c idx = 2; idx 1453 arch/powerpc/platforms/pseries/lpar.c while (idx < len) { idx 1454 arch/powerpc/platforms/pseries/lpar.c u8 block_shift = local_buffer[idx++]; idx 1463 arch/powerpc/platforms/pseries/lpar.c for (npsize = local_buffer[idx++]; idx 1464 arch/powerpc/platforms/pseries/lpar.c npsize > 0 && idx < len; npsize--) idx 1465 arch/powerpc/platforms/pseries/lpar.c check_lp_set_hblkrm((unsigned int) local_buffer[idx++], idx 1470 arch/powerpc/platforms/pseries/lpar.c for (idx = 0; idx < MMU_PAGE_COUNT; idx++) idx 1471 arch/powerpc/platforms/pseries/lpar.c if (hblkrm_size[bpsize][idx]) idx 1473 arch/powerpc/platforms/pseries/lpar.c bpsize, idx, hblkrm_size[bpsize][idx]); idx 318 arch/powerpc/platforms/pseries/lparcfg.c int idx, w_idx; idx 333 arch/powerpc/platforms/pseries/lparcfg.c idx = 0; idx 334 arch/powerpc/platforms/pseries/lparcfg.c while ((*local_buffer) && (idx < splpar_strlen)) { idx 335 arch/powerpc/platforms/pseries/lparcfg.c workbuffer[w_idx++] = local_buffer[idx++]; idx 336 arch/powerpc/platforms/pseries/lparcfg.c if ((local_buffer[idx] == ',') idx 337 arch/powerpc/platforms/pseries/lparcfg.c || (local_buffer[idx] == '\0')) { idx 344 arch/powerpc/platforms/pseries/lparcfg.c idx++; /* skip the comma */ idx 346 arch/powerpc/platforms/pseries/lparcfg.c } else if (local_buffer[idx] == '=') { idx 268 arch/powerpc/sysdev/mpic.c unsigned int idx = src_no & mpic->isu_mask; idx 272 arch/powerpc/sysdev/mpic.c reg + (idx * MPIC_INFO(IRQ_STRIDE))); idx 285 arch/powerpc/sysdev/mpic.c unsigned int idx = src_no & mpic->isu_mask; idx 288 arch/powerpc/sysdev/mpic.c reg + (idx * MPIC_INFO(IRQ_STRIDE)), value); idx 87 arch/powerpc/sysdev/xive/common.c cur = be32_to_cpup(q->qpage + q->idx); idx 96 arch/powerpc/sysdev/xive/common.c q->idx = (q->idx + 1) & q->msk; idx 99 arch/powerpc/sysdev/xive/common.c if (q->idx == 0) idx 226 arch/powerpc/sysdev/xive/common.c u32 i0, i1, idx; idx 230 arch/powerpc/sysdev/xive/common.c idx = q->idx; idx 231 arch/powerpc/sysdev/xive/common.c i0 = be32_to_cpup(q->qpage + idx); idx 232 arch/powerpc/sysdev/xive/common.c idx = (idx + 1) & q->msk; idx 233 arch/powerpc/sysdev/xive/common.c i1 = be32_to_cpup(q->qpage + idx); idx 235 arch/powerpc/sysdev/xive/common.c q->idx, q->toggle, i0, i1); idx 148 arch/powerpc/sysdev/xive/native.c q->idx = 0; idx 479 arch/powerpc/sysdev/xive/spapr.c q->idx = 0; idx 38 arch/riscv/include/asm/fixmap.h #define __late_clear_fixmap(idx) __set_fixmap((idx), 0, FIXMAP_PAGE_CLEAR) idx 40 arch/riscv/include/asm/fixmap.h extern void __set_fixmap(enum fixed_addresses idx, idx 63 arch/riscv/kernel/module-sections.c static bool duplicate_rela(const Elf_Rela *rela, int idx) idx 66 arch/riscv/kernel/module-sections.c for (i = 0; i < idx; i++) { idx 67 arch/riscv/kernel/module-sections.c if (is_rela_equal(&rela[i], &rela[idx])) idx 182 arch/riscv/kernel/perf_event.c static inline u64 read_counter(int idx) idx 186 arch/riscv/kernel/perf_event.c switch (idx) { idx 194 arch/riscv/kernel/perf_event.c WARN_ON_ONCE(idx < 0 || idx > RISCV_MAX_COUNTERS); idx 201 arch/riscv/kernel/perf_event.c static inline void write_counter(int idx, u64 value) idx 219 arch/riscv/kernel/perf_event.c int idx = hwc->idx; idx 224 arch/riscv/kernel/perf_event.c new_raw_count = read_counter(idx); idx 291 arch/riscv/kernel/perf_event.c local64_set(&hwc->prev_count, read_counter(hwc->idx)); idx 313 arch/riscv/kernel/perf_event.c hwc->idx = hwc->config; idx 314 arch/riscv/kernel/perf_event.c cpuc->events[hwc->idx] = event; idx 333 arch/riscv/kernel/perf_event.c cpuc->events[hwc->idx] = NULL; idx 428 arch/riscv/kernel/perf_event.c hwc->idx = -1; idx 11 arch/riscv/kernel/perf_regs.c u64 perf_reg_value(struct pt_regs *regs, int idx) idx 13 arch/riscv/kernel/perf_regs.c if (WARN_ON_ONCE((u32)idx >= PERF_REG_RISCV_MAX)) idx 16 arch/riscv/kernel/perf_regs.c return ((unsigned long *)regs)[idx]; idx 161 arch/riscv/mm/init.c void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot) idx 163 arch/riscv/mm/init.c unsigned long addr = __fix_to_virt(idx); idx 166 arch/riscv/mm/init.c BUG_ON(idx <= FIX_HOLE || idx >= __end_of_fixed_addresses); idx 32 arch/s390/include/asm/pci_io.h #define ZPCI_ADDR(idx) (ZPCI_IOMAP_ADDR_BASE | ((u64) idx << ZPCI_IOMAP_SHIFT)) idx 72 arch/s390/kernel/cache.c int idx; idx 77 arch/s390/kernel/cache.c for (idx = 0; idx < this_cpu_ci->num_leaves; idx++) { idx 78 arch/s390/kernel/cache.c cache = this_cpu_ci->info_list + idx; idx 79 arch/s390/kernel/cache.c seq_printf(m, "cache%-11d: ", idx); idx 156 arch/s390/kernel/cache.c unsigned int level, idx, pvt; idx 163 arch/s390/kernel/cache.c for (idx = 0, level = 0; level < this_cpu_ci->num_levels && idx 164 arch/s390/kernel/cache.c idx < this_cpu_ci->num_leaves; idx++, level++) { idx 550 arch/s390/kernel/crash_dump.c u64 idx; idx 552 arch/s390/kernel/crash_dump.c for_each_mem_range(idx, &memblock.physmem, &oldmem_type, NUMA_NO_NODE, idx 564 arch/s390/kernel/crash_dump.c u64 idx; idx 566 arch/s390/kernel/crash_dump.c for_each_mem_range(idx, &memblock.physmem, &oldmem_type, NUMA_NO_NODE, idx 1027 arch/s390/kernel/debug.c int numargs, idx; idx 1044 arch/s390/kernel/debug.c for (idx = 0; idx < min(numargs, (int)(id->buf_size / sizeof(long)) - 1); idx++) idx 1045 arch/s390/kernel/debug.c curr_event->args[idx] = va_arg(ap, long); idx 1062 arch/s390/kernel/debug.c int numargs, idx; idx 1080 arch/s390/kernel/debug.c for (idx = 0; idx < min(numargs, (int)(id->buf_size / sizeof(long)) - 1); idx++) idx 1081 arch/s390/kernel/debug.c curr_event->args[idx] = va_arg(ap, long); idx 304 arch/s390/kernel/irq.c int idx; idx 306 arch/s390/kernel/irq.c for (idx = 0; idx < ARRAY_SIZE(ext_int_hash); idx++) idx 307 arch/s390/kernel/irq.c INIT_HLIST_HEAD(&ext_int_hash[idx]); idx 1373 arch/s390/kernel/perf_cpum_sf.c unsigned long i, range_scan, idx; idx 1382 arch/s390/kernel/perf_cpum_sf.c for (i = 0, idx = aux->head; i < range_scan; i++, idx++) { idx 1383 arch/s390/kernel/perf_cpum_sf.c te = aux_sdb_trailer(aux, idx); idx 1411 arch/s390/kernel/perf_cpum_sf.c unsigned long i, range_scan, idx; idx 1429 arch/s390/kernel/perf_cpum_sf.c idx = aux->empty_mark + 1; idx 1430 arch/s390/kernel/perf_cpum_sf.c for (i = 0; i < range_scan; i++, idx++) { idx 1431 arch/s390/kernel/perf_cpum_sf.c te = aux_sdb_trailer(aux, idx); idx 1523 arch/s390/kernel/perf_cpum_sf.c unsigned long i, range_scan, idx; idx 1548 arch/s390/kernel/perf_cpum_sf.c idx = aux->empty_mark + 1; idx 1549 arch/s390/kernel/perf_cpum_sf.c for (i = 0; i < range_scan; i++, idx++) { idx 1550 arch/s390/kernel/perf_cpum_sf.c te = aux_sdb_trailer(aux, idx); idx 1555 arch/s390/kernel/perf_cpum_sf.c if (idx == aux->alert_mark) idx 11 arch/s390/kernel/perf_regs.c u64 perf_reg_value(struct pt_regs *regs, int idx) idx 15 arch/s390/kernel/perf_regs.c if (idx >= PERF_REG_S390_R0 && idx <= PERF_REG_S390_R15) idx 16 arch/s390/kernel/perf_regs.c return regs->gprs[idx]; idx 18 arch/s390/kernel/perf_regs.c if (idx >= PERF_REG_S390_FP0 && idx <= PERF_REG_S390_FP15) { idx 22 arch/s390/kernel/perf_regs.c idx -= PERF_REG_S390_FP0; idx 23 arch/s390/kernel/perf_regs.c fp = MACHINE_HAS_VX ? *(freg_t *)(current->thread.fpu.vxrs + idx) idx 24 arch/s390/kernel/perf_regs.c : current->thread.fpu.fprs[idx]; idx 28 arch/s390/kernel/perf_regs.c if (idx == PERF_REG_S390_MASK) idx 30 arch/s390/kernel/perf_regs.c if (idx == PERF_REG_S390_PC) idx 33 arch/s390/kernel/perf_regs.c WARN_ON_ONCE((u32)idx >= PERF_REG_S390_MAX); idx 834 arch/s390/kvm/gaccess.c unsigned long _len, nr_pages, gpa, idx; idx 857 arch/s390/kvm/gaccess.c for (idx = 0; idx < nr_pages && !rc; idx++) { idx 858 arch/s390/kvm/gaccess.c gpa = *(pages + idx) + (ga & ~PAGE_MASK); idx 2723 arch/s390/kvm/interrupt.c int summary_set, idx; idx 2733 arch/s390/kvm/interrupt.c idx = srcu_read_lock(&kvm->srcu); idx 2738 arch/s390/kvm/interrupt.c srcu_read_unlock(&kvm->srcu, idx); idx 2747 arch/s390/kvm/interrupt.c srcu_read_unlock(&kvm->srcu, idx); idx 814 arch/s390/kvm/kvm-s390.c unsigned int idx; idx 845 arch/s390/kvm/kvm-s390.c idx = srcu_read_lock(&kvm->srcu); idx 847 arch/s390/kvm/kvm-s390.c srcu_read_unlock(&kvm->srcu, idx); idx 4350 arch/s390/kvm/kvm-s390.c int idx; idx 4357 arch/s390/kvm/kvm-s390.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 4359 arch/s390/kvm/kvm-s390.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 269 arch/s390/pci/pci.c int idx; idx 271 arch/s390/pci/pci.c idx = zdev->bars[bar].map_idx; idx 274 arch/s390/pci/pci.c WARN_ON(!++zpci_iomap_start[idx].count); idx 275 arch/s390/pci/pci.c zpci_iomap_start[idx].fh = zdev->fh; idx 276 arch/s390/pci/pci.c zpci_iomap_start[idx].bar = bar; idx 279 arch/s390/pci/pci.c return (void __iomem *) ZPCI_ADDR(idx) + offset; idx 345 arch/s390/pci/pci.c unsigned int idx = ZPCI_IDX(addr); idx 349 arch/s390/pci/pci.c WARN_ON(!zpci_iomap_start[idx].count); idx 350 arch/s390/pci/pci.c if (!--zpci_iomap_start[idx].count) { idx 351 arch/s390/pci/pci.c zpci_iomap_start[idx].fh = 0; idx 352 arch/s390/pci/pci.c zpci_iomap_start[idx].bar = 0; idx 167 arch/sh/boards/of-generic.c void __init __weak arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 114 arch/sh/drivers/pci/pcie-sh7786.c #define DEFINE_CONTROLLER(start, idx) \ idx 117 arch/sh/drivers/pci/pcie-sh7786.c .resources = sh7786_pci##idx##_resources, \ idx 118 arch/sh/drivers/pci/pcie-sh7786.c .nr_resources = ARRAY_SIZE(sh7786_pci##idx##_resources), \ idx 75 arch/sh/include/asm/fixmap.h extern void __set_fixmap(enum fixed_addresses idx, idx 77 arch/sh/include/asm/fixmap.h extern void __clear_fixmap(enum fixed_addresses idx, pgprot_t flags); idx 34 arch/sh/kernel/cpu/sh2/clock-sh7619.c int idx = (__raw_readw(FREQCR) & 0x0007); idx 35 arch/sh/kernel/cpu/sh2/clock-sh7619.c return clk->parent->rate / pfc_divisors[idx]; idx 62 arch/sh/kernel/cpu/sh2/clock-sh7619.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 72 arch/sh/kernel/cpu/sh2/clock-sh7619.c if (idx < ARRAY_SIZE(sh7619_clk_ops)) idx 73 arch/sh/kernel/cpu/sh2/clock-sh7619.c *ops = sh7619_clk_ops[idx]; idx 36 arch/sh/kernel/cpu/sh2a/clock-sh7201.c int idx = (__raw_readw(FREQCR) & 0x0007); idx 37 arch/sh/kernel/cpu/sh2a/clock-sh7201.c return clk->parent->rate / pfc_divisors[idx]; idx 46 arch/sh/kernel/cpu/sh2a/clock-sh7201.c int idx = (__raw_readw(FREQCR) & 0x0007); idx 47 arch/sh/kernel/cpu/sh2a/clock-sh7201.c return clk->parent->rate / pfc_divisors[idx]; idx 56 arch/sh/kernel/cpu/sh2a/clock-sh7201.c int idx = ((__raw_readw(FREQCR) >> 4) & 0x0007); idx 57 arch/sh/kernel/cpu/sh2a/clock-sh7201.c return clk->parent->rate / ifc_divisors[idx]; idx 71 arch/sh/kernel/cpu/sh2a/clock-sh7201.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 80 arch/sh/kernel/cpu/sh2a/clock-sh7201.c if (idx < ARRAY_SIZE(sh7201_clk_ops)) idx 81 arch/sh/kernel/cpu/sh2a/clock-sh7201.c *ops = sh7201_clk_ops[idx]; idx 38 arch/sh/kernel/cpu/sh2a/clock-sh7203.c int idx = (__raw_readw(FREQCR) & 0x0007); idx 39 arch/sh/kernel/cpu/sh2a/clock-sh7203.c return clk->parent->rate / pfc_divisors[idx]; idx 48 arch/sh/kernel/cpu/sh2a/clock-sh7203.c int idx = (__raw_readw(FREQCR) & 0x0007); idx 49 arch/sh/kernel/cpu/sh2a/clock-sh7203.c return clk->parent->rate / pfc_divisors[idx-2]; idx 67 arch/sh/kernel/cpu/sh2a/clock-sh7203.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 76 arch/sh/kernel/cpu/sh2a/clock-sh7203.c if (idx < ARRAY_SIZE(sh7203_clk_ops)) idx 77 arch/sh/kernel/cpu/sh2a/clock-sh7203.c *ops = sh7203_clk_ops[idx]; idx 35 arch/sh/kernel/cpu/sh2a/clock-sh7206.c int idx = (__raw_readw(FREQCR) & 0x0007); idx 36 arch/sh/kernel/cpu/sh2a/clock-sh7206.c return clk->parent->rate / pfc_divisors[idx]; idx 54 arch/sh/kernel/cpu/sh2a/clock-sh7206.c int idx = (__raw_readw(FREQCR) & 0x0007); idx 55 arch/sh/kernel/cpu/sh2a/clock-sh7206.c return clk->parent->rate / ifc_divisors[idx]; idx 69 arch/sh/kernel/cpu/sh2a/clock-sh7206.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 78 arch/sh/kernel/cpu/sh2a/clock-sh7206.c if (idx < ARRAY_SIZE(sh7206_clk_ops)) idx 79 arch/sh/kernel/cpu/sh2a/clock-sh7206.c *ops = sh7206_clk_ops[idx]; idx 29 arch/sh/kernel/cpu/sh3/clock-sh3.c int idx = ((frqcr & 0x2000) >> 11) | (frqcr & 0x0003); idx 31 arch/sh/kernel/cpu/sh3/clock-sh3.c clk->rate *= pfc_divisors[idx]; idx 41 arch/sh/kernel/cpu/sh3/clock-sh3.c int idx = ((frqcr & 0x2000) >> 11) | (frqcr & 0x0003); idx 43 arch/sh/kernel/cpu/sh3/clock-sh3.c return clk->parent->rate / pfc_divisors[idx]; idx 53 arch/sh/kernel/cpu/sh3/clock-sh3.c int idx = ((frqcr & 0x8000) >> 13) | ((frqcr & 0x0030) >> 4); idx 55 arch/sh/kernel/cpu/sh3/clock-sh3.c return clk->parent->rate / stc_multipliers[idx]; idx 65 arch/sh/kernel/cpu/sh3/clock-sh3.c int idx = ((frqcr & 0x4000) >> 12) | ((frqcr & 0x000c) >> 2); idx 67 arch/sh/kernel/cpu/sh3/clock-sh3.c return clk->parent->rate / ifc_divisors[idx]; idx 81 arch/sh/kernel/cpu/sh3/clock-sh3.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 83 arch/sh/kernel/cpu/sh3/clock-sh3.c if (idx < ARRAY_SIZE(sh3_clk_ops)) idx 84 arch/sh/kernel/cpu/sh3/clock-sh3.c *ops = sh3_clk_ops[idx]; idx 41 arch/sh/kernel/cpu/sh3/clock-sh7705.c int idx = __raw_readw(FRQCR) & 0x0003; idx 42 arch/sh/kernel/cpu/sh3/clock-sh7705.c return clk->parent->rate / pfc_divisors[idx]; idx 51 arch/sh/kernel/cpu/sh3/clock-sh7705.c int idx = (__raw_readw(FRQCR) & 0x0300) >> 8; idx 52 arch/sh/kernel/cpu/sh3/clock-sh7705.c return clk->parent->rate / stc_multipliers[idx]; idx 61 arch/sh/kernel/cpu/sh3/clock-sh7705.c int idx = (__raw_readw(FRQCR) & 0x0030) >> 4; idx 62 arch/sh/kernel/cpu/sh3/clock-sh7705.c return clk->parent->rate / ifc_divisors[idx]; idx 76 arch/sh/kernel/cpu/sh3/clock-sh7705.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 78 arch/sh/kernel/cpu/sh3/clock-sh7705.c if (idx < ARRAY_SIZE(sh7705_clk_ops)) idx 79 arch/sh/kernel/cpu/sh3/clock-sh7705.c *ops = sh7705_clk_ops[idx]; idx 25 arch/sh/kernel/cpu/sh3/clock-sh7706.c int idx = ((frqcr & 0x2000) >> 11) | (frqcr & 0x0003); idx 27 arch/sh/kernel/cpu/sh3/clock-sh7706.c clk->rate *= pfc_divisors[idx]; idx 37 arch/sh/kernel/cpu/sh3/clock-sh7706.c int idx = ((frqcr & 0x2000) >> 11) | (frqcr & 0x0003); idx 39 arch/sh/kernel/cpu/sh3/clock-sh7706.c return clk->parent->rate / pfc_divisors[idx]; idx 49 arch/sh/kernel/cpu/sh3/clock-sh7706.c int idx = ((frqcr & 0x8000) >> 13) | ((frqcr & 0x0030) >> 4); idx 51 arch/sh/kernel/cpu/sh3/clock-sh7706.c return clk->parent->rate / stc_multipliers[idx]; idx 61 arch/sh/kernel/cpu/sh3/clock-sh7706.c int idx = ((frqcr & 0x4000) >> 12) | ((frqcr & 0x000c) >> 2); idx 63 arch/sh/kernel/cpu/sh3/clock-sh7706.c return clk->parent->rate / ifc_divisors[idx]; idx 77 arch/sh/kernel/cpu/sh3/clock-sh7706.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 79 arch/sh/kernel/cpu/sh3/clock-sh7706.c if (idx < ARRAY_SIZE(sh7706_clk_ops)) idx 80 arch/sh/kernel/cpu/sh3/clock-sh7706.c *ops = sh7706_clk_ops[idx]; idx 25 arch/sh/kernel/cpu/sh3/clock-sh7709.c int idx = ((frqcr & 0x2000) >> 11) | (frqcr & 0x0003); idx 27 arch/sh/kernel/cpu/sh3/clock-sh7709.c clk->rate *= pfc_divisors[idx]; idx 37 arch/sh/kernel/cpu/sh3/clock-sh7709.c int idx = ((frqcr & 0x2000) >> 11) | (frqcr & 0x0003); idx 39 arch/sh/kernel/cpu/sh3/clock-sh7709.c return clk->parent->rate / pfc_divisors[idx]; idx 49 arch/sh/kernel/cpu/sh3/clock-sh7709.c int idx = (frqcr & 0x0080) ? idx 52 arch/sh/kernel/cpu/sh3/clock-sh7709.c return clk->parent->rate * stc_multipliers[idx]; idx 62 arch/sh/kernel/cpu/sh3/clock-sh7709.c int idx = ((frqcr & 0x4000) >> 12) | ((frqcr & 0x000c) >> 2); idx 64 arch/sh/kernel/cpu/sh3/clock-sh7709.c return clk->parent->rate / ifc_divisors[idx]; idx 78 arch/sh/kernel/cpu/sh3/clock-sh7709.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 80 arch/sh/kernel/cpu/sh3/clock-sh7709.c if (idx < ARRAY_SIZE(sh7709_clk_ops)) idx 81 arch/sh/kernel/cpu/sh3/clock-sh7709.c *ops = sh7709_clk_ops[idx]; idx 35 arch/sh/kernel/cpu/sh3/clock-sh7710.c int idx = (__raw_readw(FRQCR) & 0x0007); idx 36 arch/sh/kernel/cpu/sh3/clock-sh7710.c return clk->parent->rate / md_table[idx]; idx 45 arch/sh/kernel/cpu/sh3/clock-sh7710.c int idx = (__raw_readw(FRQCR) & 0x0700) >> 8; idx 46 arch/sh/kernel/cpu/sh3/clock-sh7710.c return clk->parent->rate / md_table[idx]; idx 55 arch/sh/kernel/cpu/sh3/clock-sh7710.c int idx = (__raw_readw(FRQCR) & 0x0070) >> 4; idx 56 arch/sh/kernel/cpu/sh3/clock-sh7710.c return clk->parent->rate / md_table[idx]; idx 70 arch/sh/kernel/cpu/sh3/clock-sh7710.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 72 arch/sh/kernel/cpu/sh3/clock-sh7710.c if (idx < ARRAY_SIZE(sh7710_clk_ops)) idx 73 arch/sh/kernel/cpu/sh3/clock-sh7710.c *ops = sh7710_clk_ops[idx]; idx 24 arch/sh/kernel/cpu/sh3/clock-sh7712.c int idx = (frqcr & 0x0300) >> 8; idx 26 arch/sh/kernel/cpu/sh3/clock-sh7712.c clk->rate *= multipliers[idx]; idx 36 arch/sh/kernel/cpu/sh3/clock-sh7712.c int idx = frqcr & 0x0007; idx 38 arch/sh/kernel/cpu/sh3/clock-sh7712.c return clk->parent->rate / divisors[idx]; idx 48 arch/sh/kernel/cpu/sh3/clock-sh7712.c int idx = (frqcr & 0x0030) >> 4; idx 50 arch/sh/kernel/cpu/sh3/clock-sh7712.c return clk->parent->rate / divisors[idx]; idx 63 arch/sh/kernel/cpu/sh3/clock-sh7712.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 65 arch/sh/kernel/cpu/sh3/clock-sh7712.c if (idx < ARRAY_SIZE(sh7712_clk_ops)) idx 66 arch/sh/kernel/cpu/sh3/clock-sh7712.c *ops = sh7712_clk_ops[idx]; idx 24 arch/sh/kernel/cpu/sh4/clock-sh4-202.c int idx = __raw_readl(CPG2_FRQCR3) & 0x0007; idx 25 arch/sh/kernel/cpu/sh4/clock-sh4-202.c return clk->parent->rate / frqcr3_divisors[idx]; idx 52 arch/sh/kernel/cpu/sh4/clock-sh4-202.c int idx = (__raw_readl(CPG2_FRQCR3) >> 3) & 0x0007; idx 53 arch/sh/kernel/cpu/sh4/clock-sh4-202.c return clk->parent->rate / frqcr3_divisors[idx]; idx 90 arch/sh/kernel/cpu/sh4/clock-sh4-202.c int idx = (__raw_readl(CPG2_FRQCR3) >> 6) & 0x0007; idx 91 arch/sh/kernel/cpu/sh4/clock-sh4-202.c return clk->parent->rate / frqcr3_divisors[idx]; idx 37 arch/sh/kernel/cpu/sh4/clock-sh4.c int idx = (__raw_readw(FRQCR) & 0x0007); idx 38 arch/sh/kernel/cpu/sh4/clock-sh4.c return clk->parent->rate / pfc_divisors[idx]; idx 47 arch/sh/kernel/cpu/sh4/clock-sh4.c int idx = (__raw_readw(FRQCR) >> 3) & 0x0007; idx 48 arch/sh/kernel/cpu/sh4/clock-sh4.c return clk->parent->rate / bfc_divisors[idx]; idx 57 arch/sh/kernel/cpu/sh4/clock-sh4.c int idx = (__raw_readw(FRQCR) >> 6) & 0x0007; idx 58 arch/sh/kernel/cpu/sh4/clock-sh4.c return clk->parent->rate / ifc_divisors[idx]; idx 72 arch/sh/kernel/cpu/sh4/clock-sh4.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 74 arch/sh/kernel/cpu/sh4/clock-sh4.c if (idx < ARRAY_SIZE(sh4_clk_ops)) idx 75 arch/sh/kernel/cpu/sh4/clock-sh4.c *ops = sh4_clk_ops[idx]; idx 202 arch/sh/kernel/cpu/sh4/perf_event.c static u64 sh7750_pmu_read(int idx) idx 204 arch/sh/kernel/cpu/sh4/perf_event.c return (u64)((u64)(__raw_readl(PMCTRH(idx)) & 0xffff) << 32) | idx 205 arch/sh/kernel/cpu/sh4/perf_event.c __raw_readl(PMCTRL(idx)); idx 208 arch/sh/kernel/cpu/sh4/perf_event.c static void sh7750_pmu_disable(struct hw_perf_event *hwc, int idx) idx 212 arch/sh/kernel/cpu/sh4/perf_event.c tmp = __raw_readw(PMCR(idx)); idx 214 arch/sh/kernel/cpu/sh4/perf_event.c __raw_writew(tmp, PMCR(idx)); idx 217 arch/sh/kernel/cpu/sh4/perf_event.c static void sh7750_pmu_enable(struct hw_perf_event *hwc, int idx) idx 219 arch/sh/kernel/cpu/sh4/perf_event.c __raw_writew(__raw_readw(PMCR(idx)) | PMCR_PMCLR, PMCR(idx)); idx 220 arch/sh/kernel/cpu/sh4/perf_event.c __raw_writew(hwc->config | PMCR_PMEN | PMCR_PMST, PMCR(idx)); idx 33 arch/sh/kernel/cpu/sh4a/clock-sh7763.c int idx = ((__raw_readl(FRQCR) >> 4) & 0x07); idx 34 arch/sh/kernel/cpu/sh4a/clock-sh7763.c return clk->parent->rate / p0fc_divisors[idx]; idx 43 arch/sh/kernel/cpu/sh4a/clock-sh7763.c int idx = ((__raw_readl(FRQCR) >> 16) & 0x07); idx 44 arch/sh/kernel/cpu/sh4a/clock-sh7763.c return clk->parent->rate / bfc_divisors[idx]; idx 62 arch/sh/kernel/cpu/sh4a/clock-sh7763.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 64 arch/sh/kernel/cpu/sh4a/clock-sh7763.c if (idx < ARRAY_SIZE(sh7763_clk_ops)) idx 65 arch/sh/kernel/cpu/sh4a/clock-sh7763.c *ops = sh7763_clk_ops[idx]; idx 70 arch/sh/kernel/cpu/sh4a/clock-sh7763.c int idx = ((__raw_readl(FRQCR) >> 20) & 0x07); idx 71 arch/sh/kernel/cpu/sh4a/clock-sh7763.c return clk->parent->rate / cfc_divisors[idx]; idx 30 arch/sh/kernel/cpu/sh4a/clock-sh7770.c int idx = ((__raw_readl(FRQCR) >> 28) & 0x000f); idx 31 arch/sh/kernel/cpu/sh4a/clock-sh7770.c return clk->parent->rate / pfc_divisors[idx]; idx 40 arch/sh/kernel/cpu/sh4a/clock-sh7770.c int idx = (__raw_readl(FRQCR) & 0x000f); idx 41 arch/sh/kernel/cpu/sh4a/clock-sh7770.c return clk->parent->rate / bfc_divisors[idx]; idx 50 arch/sh/kernel/cpu/sh4a/clock-sh7770.c int idx = ((__raw_readl(FRQCR) >> 24) & 0x000f); idx 51 arch/sh/kernel/cpu/sh4a/clock-sh7770.c return clk->parent->rate / ifc_divisors[idx]; idx 65 arch/sh/kernel/cpu/sh4a/clock-sh7770.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 67 arch/sh/kernel/cpu/sh4a/clock-sh7770.c if (idx < ARRAY_SIZE(sh7770_clk_ops)) idx 68 arch/sh/kernel/cpu/sh4a/clock-sh7770.c *ops = sh7770_clk_ops[idx]; idx 33 arch/sh/kernel/cpu/sh4a/clock-sh7780.c int idx = (__raw_readl(FRQCR) & 0x0003); idx 34 arch/sh/kernel/cpu/sh4a/clock-sh7780.c return clk->parent->rate / pfc_divisors[idx]; idx 43 arch/sh/kernel/cpu/sh4a/clock-sh7780.c int idx = ((__raw_readl(FRQCR) >> 16) & 0x0007); idx 44 arch/sh/kernel/cpu/sh4a/clock-sh7780.c return clk->parent->rate / bfc_divisors[idx]; idx 53 arch/sh/kernel/cpu/sh4a/clock-sh7780.c int idx = ((__raw_readl(FRQCR) >> 24) & 0x0001); idx 54 arch/sh/kernel/cpu/sh4a/clock-sh7780.c return clk->parent->rate / ifc_divisors[idx]; idx 68 arch/sh/kernel/cpu/sh4a/clock-sh7780.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 70 arch/sh/kernel/cpu/sh4a/clock-sh7780.c if (idx < ARRAY_SIZE(sh7780_clk_ops)) idx 71 arch/sh/kernel/cpu/sh4a/clock-sh7780.c *ops = sh7780_clk_ops[idx]; idx 76 arch/sh/kernel/cpu/sh4a/clock-sh7780.c int idx = ((__raw_readl(FRQCR) >> 20) & 0x0007); idx 77 arch/sh/kernel/cpu/sh4a/clock-sh7780.c return clk->parent->rate / cfc_divisors[idx]; idx 14 arch/sh/kernel/cpu/sh4a/perf_event.c #define PPC_CCBR(idx) (0xff200800 + (sizeof(u32) * idx)) idx 15 arch/sh/kernel/cpu/sh4a/perf_event.c #define PPC_PMCTR(idx) (0xfc100000 + (sizeof(u32) * idx)) idx 227 arch/sh/kernel/cpu/sh4a/perf_event.c static u64 sh4a_pmu_read(int idx) idx 229 arch/sh/kernel/cpu/sh4a/perf_event.c return __raw_readl(PPC_PMCTR(idx)); idx 232 arch/sh/kernel/cpu/sh4a/perf_event.c static void sh4a_pmu_disable(struct hw_perf_event *hwc, int idx) idx 236 arch/sh/kernel/cpu/sh4a/perf_event.c tmp = __raw_readl(PPC_CCBR(idx)); idx 238 arch/sh/kernel/cpu/sh4a/perf_event.c __raw_writel(tmp, PPC_CCBR(idx)); idx 241 arch/sh/kernel/cpu/sh4a/perf_event.c static void sh4a_pmu_enable(struct hw_perf_event *hwc, int idx) idx 247 arch/sh/kernel/cpu/sh4a/perf_event.c tmp |= idx ? PMCAT_CLR1 : PMCAT_CLR0; idx 250 arch/sh/kernel/cpu/sh4a/perf_event.c tmp = __raw_readl(PPC_CCBR(idx)); idx 252 arch/sh/kernel/cpu/sh4a/perf_event.c __raw_writel(tmp, PPC_CCBR(idx)); idx 254 arch/sh/kernel/cpu/sh4a/perf_event.c __raw_writel(__raw_readl(PPC_CCBR(idx)) | CCBR_DUC, PPC_CCBR(idx)); idx 15 arch/sh/kernel/cpu/sh4a/ubc.c #define UBC_CBR(idx) (0xff200000 + (0x20 * idx)) idx 16 arch/sh/kernel/cpu/sh4a/ubc.c #define UBC_CRR(idx) (0xff200004 + (0x20 * idx)) idx 17 arch/sh/kernel/cpu/sh4a/ubc.c #define UBC_CAR(idx) (0xff200008 + (0x20 * idx)) idx 18 arch/sh/kernel/cpu/sh4a/ubc.c #define UBC_CAMR(idx) (0xff20000c + (0x20 * idx)) idx 32 arch/sh/kernel/cpu/sh4a/ubc.c static void sh4a_ubc_enable(struct arch_hw_breakpoint *info, int idx) idx 34 arch/sh/kernel/cpu/sh4a/ubc.c __raw_writel(UBC_CBR_CE | info->len | info->type, UBC_CBR(idx)); idx 35 arch/sh/kernel/cpu/sh4a/ubc.c __raw_writel(info->address, UBC_CAR(idx)); idx 38 arch/sh/kernel/cpu/sh4a/ubc.c static void sh4a_ubc_disable(struct arch_hw_breakpoint *info, int idx) idx 40 arch/sh/kernel/cpu/sh4a/ubc.c __raw_writel(0, UBC_CBR(idx)); idx 41 arch/sh/kernel/cpu/sh4a/ubc.c __raw_writel(0, UBC_CAR(idx)); idx 24 arch/sh/kernel/cpu/sh5/clock-sh5.c int idx = (__raw_readl(cprc_base + 0x00) >> 6) & 0x0007; idx 25 arch/sh/kernel/cpu/sh5/clock-sh5.c clk->rate *= ifc_table[idx]; idx 34 arch/sh/kernel/cpu/sh5/clock-sh5.c int idx = (__raw_readw(cprc_base) >> 12) & 0x0007; idx 35 arch/sh/kernel/cpu/sh5/clock-sh5.c return clk->parent->rate / ifc_table[idx]; idx 44 arch/sh/kernel/cpu/sh5/clock-sh5.c int idx = (__raw_readw(cprc_base) >> 3) & 0x0007; idx 45 arch/sh/kernel/cpu/sh5/clock-sh5.c return clk->parent->rate / ifc_table[idx]; idx 54 arch/sh/kernel/cpu/sh5/clock-sh5.c int idx = (__raw_readw(cprc_base) & 0x0007); idx 55 arch/sh/kernel/cpu/sh5/clock-sh5.c return clk->parent->rate / ifc_table[idx]; idx 69 arch/sh/kernel/cpu/sh5/clock-sh5.c void __init arch_init_clk_ops(struct sh_clk_ops **ops, int idx) idx 74 arch/sh/kernel/cpu/sh5/clock-sh5.c if (idx < ARRAY_SIZE(sh5_clk_ops)) idx 75 arch/sh/kernel/cpu/sh5/clock-sh5.c *ops = sh5_clk_ops[idx]; idx 177 arch/sh/kernel/perf_event.c struct hw_perf_event *hwc, int idx) idx 197 arch/sh/kernel/perf_event.c new_raw_count = sh_pmu->read(idx); idx 221 arch/sh/kernel/perf_event.c int idx = hwc->idx; idx 224 arch/sh/kernel/perf_event.c sh_pmu->disable(hwc, idx); idx 225 arch/sh/kernel/perf_event.c cpuc->events[idx] = NULL; idx 230 arch/sh/kernel/perf_event.c sh_perf_event_update(event, &event->hw, idx); idx 239 arch/sh/kernel/perf_event.c int idx = hwc->idx; idx 241 arch/sh/kernel/perf_event.c if (WARN_ON_ONCE(idx == -1)) idx 247 arch/sh/kernel/perf_event.c cpuc->events[idx] = event; idx 249 arch/sh/kernel/perf_event.c sh_pmu->enable(hwc, idx); idx 257 arch/sh/kernel/perf_event.c __clear_bit(event->hw.idx, cpuc->used_mask); idx 266 arch/sh/kernel/perf_event.c int idx = hwc->idx; idx 271 arch/sh/kernel/perf_event.c if (__test_and_set_bit(idx, cpuc->used_mask)) { idx 272 arch/sh/kernel/perf_event.c idx = find_first_zero_bit(cpuc->used_mask, sh_pmu->num_events); idx 273 arch/sh/kernel/perf_event.c if (idx == sh_pmu->num_events) idx 276 arch/sh/kernel/perf_event.c __set_bit(idx, cpuc->used_mask); idx 277 arch/sh/kernel/perf_event.c hwc->idx = idx; idx 280 arch/sh/kernel/perf_event.c sh_pmu->disable(hwc, idx); idx 295 arch/sh/kernel/perf_event.c sh_perf_event_update(event, &event->hw, event->hw.idx); idx 102 arch/sh/mm/init.c void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) idx 104 arch/sh/mm/init.c unsigned long address = __fix_to_virt(idx); idx 106 arch/sh/mm/init.c if (idx >= __end_of_fixed_addresses) { idx 114 arch/sh/mm/init.c void __clear_fixmap(enum fixed_addresses idx, pgprot_t prot) idx 116 arch/sh/mm/init.c unsigned long address = __fix_to_virt(idx); idx 118 arch/sh/mm/init.c if (idx >= __end_of_fixed_addresses) { idx 50 arch/sh/mm/ioremap_fixed.c enum fixed_addresses idx0, idx; idx 87 arch/sh/mm/ioremap_fixed.c idx = idx0; idx 90 arch/sh/mm/ioremap_fixed.c __set_fixmap(idx, phys_addr, prot); idx 92 arch/sh/mm/ioremap_fixed.c idx++; idx 102 arch/sh/mm/ioremap_fixed.c enum fixed_addresses idx; idx 124 arch/sh/mm/ioremap_fixed.c idx = FIX_IOREMAP_BEGIN + slot + nrpages - 1; idx 126 arch/sh/mm/ioremap_fixed.c __clear_fixmap(idx, __pgprot(_PAGE_WIRED)); idx 127 arch/sh/mm/ioremap_fixed.c --idx; idx 33 arch/sh/mm/kmap.c enum fixed_addresses idx; idx 41 arch/sh/mm/kmap.c idx = FIX_CMAP_END - idx 45 arch/sh/mm/kmap.c vaddr = __fix_to_virt(idx); idx 47 arch/sh/mm/kmap.c BUG_ON(!pte_none(*(kmap_coherent_pte - idx))); idx 48 arch/sh/mm/kmap.c set_pte(kmap_coherent_pte - idx, mk_pte(page, PAGE_KERNEL)); idx 57 arch/sh/mm/kmap.c enum fixed_addresses idx = __virt_to_fix(vaddr); idx 62 arch/sh/mm/kmap.c pte_clear(&init_mm, vaddr, kmap_coherent_pte - idx); idx 97 arch/sh/mm/nommu.c void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) idx 922 arch/sparc/kernel/mdesc.c int idx; idx 925 arch/sparc/kernel/mdesc.c idx = 1; idx 943 arch/sparc/kernel/mdesc.c mark_core_ids(hp, mp, idx); idx 944 arch/sparc/kernel/mdesc.c idx++; idx 951 arch/sparc/kernel/mdesc.c int idx = 1; idx 964 arch/sparc/kernel/mdesc.c mark_max_cache_ids(hp, mp, idx); idx 965 arch/sparc/kernel/mdesc.c idx++; idx 973 arch/sparc/kernel/mdesc.c int idx = 1; idx 989 arch/sparc/kernel/mdesc.c cpu_data(*id).sock_id = idx; idx 991 arch/sparc/kernel/mdesc.c idx++; idx 1033 arch/sparc/kernel/mdesc.c int idx; idx 1036 arch/sparc/kernel/mdesc.c idx = 0; idx 1046 arch/sparc/kernel/mdesc.c mark_proc_ids(hp, mp, idx); idx 1047 arch/sparc/kernel/mdesc.c idx++; idx 364 arch/sparc/kernel/pci.c u32 idx, first, last; idx 368 arch/sparc/kernel/pci.c for (idx = 0; idx < 8; idx++) { idx 369 arch/sparc/kernel/pci.c if ((map & (1 << idx)) != 0) { idx 370 arch/sparc/kernel/pci.c if (first > idx) idx 371 arch/sparc/kernel/pci.c first = idx; idx 372 arch/sparc/kernel/pci.c if (last < idx) idx 373 arch/sparc/kernel/pci.c last = idx; idx 180 arch/sparc/kernel/perf_event.c static u32 sparc_default_read_pmc(int idx) idx 185 arch/sparc/kernel/perf_event.c if (idx == PIC_UPPER_INDEX) idx 191 arch/sparc/kernel/perf_event.c static void sparc_default_write_pmc(int idx, u64 val) idx 196 arch/sparc/kernel/perf_event.c if (idx == PIC_UPPER_INDEX) idx 731 arch/sparc/kernel/perf_event.c static u32 sparc_vt_read_pmc(int idx) idx 733 arch/sparc/kernel/perf_event.c u64 val = pcr_ops->read_pic(idx); idx 738 arch/sparc/kernel/perf_event.c static void sparc_vt_write_pmc(int idx, u64 val) idx 742 arch/sparc/kernel/perf_event.c pcr = pcr_ops->read_pcr(idx); idx 746 arch/sparc/kernel/perf_event.c pcr_ops->write_pic(idx, val & 0xffffffff); idx 748 arch/sparc/kernel/perf_event.c pcr_ops->write_pcr(idx, pcr); idx 806 arch/sparc/kernel/perf_event.c static u64 event_encoding(u64 event_id, int idx) idx 808 arch/sparc/kernel/perf_event.c if (idx == PIC_UPPER_INDEX) idx 815 arch/sparc/kernel/perf_event.c static u64 mask_for_index(int idx) idx 817 arch/sparc/kernel/perf_event.c return event_encoding(sparc_pmu->event_mask, idx); idx 820 arch/sparc/kernel/perf_event.c static u64 nop_for_index(int idx) idx 822 arch/sparc/kernel/perf_event.c return event_encoding(idx == PIC_UPPER_INDEX ? idx 824 arch/sparc/kernel/perf_event.c sparc_pmu->lower_nop, idx); idx 827 arch/sparc/kernel/perf_event.c static inline void sparc_pmu_enable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, int idx) idx 829 arch/sparc/kernel/perf_event.c u64 enc, val, mask = mask_for_index(idx); idx 833 arch/sparc/kernel/perf_event.c pcr_index = idx; idx 835 arch/sparc/kernel/perf_event.c enc = perf_event_get_enc(cpuc->events[idx]); idx 839 arch/sparc/kernel/perf_event.c val |= event_encoding(enc, idx); idx 845 arch/sparc/kernel/perf_event.c static inline void sparc_pmu_disable_event(struct cpu_hw_events *cpuc, struct hw_perf_event *hwc, int idx) idx 847 arch/sparc/kernel/perf_event.c u64 mask = mask_for_index(idx); idx 848 arch/sparc/kernel/perf_event.c u64 nop = nop_for_index(idx); idx 853 arch/sparc/kernel/perf_event.c pcr_index = idx; idx 864 arch/sparc/kernel/perf_event.c struct hw_perf_event *hwc, int idx) idx 872 arch/sparc/kernel/perf_event.c new_raw_count = sparc_pmu->read_pmc(idx); idx 888 arch/sparc/kernel/perf_event.c struct hw_perf_event *hwc, int idx) idx 916 arch/sparc/kernel/perf_event.c sparc_pmu->write_pmc(idx, (u64)(-left) & 0xffffffff); idx 931 arch/sparc/kernel/perf_event.c cpuc->current_idx[i] != cp->hw.idx) { idx 958 arch/sparc/kernel/perf_event.c int idx = hwc->idx; idx 964 arch/sparc/kernel/perf_event.c sparc_perf_event_set_period(cp, hwc, idx); idx 965 arch/sparc/kernel/perf_event.c cpuc->current_idx[i] = idx; idx 968 arch/sparc/kernel/perf_event.c cpuc->pcr[0] &= ~mask_for_index(idx); idx 970 arch/sparc/kernel/perf_event.c cpuc->pcr[0] |= nop_for_index(idx); idx 972 arch/sparc/kernel/perf_event.c cpuc->pcr[0] |= event_encoding(enc, idx); idx 993 arch/sparc/kernel/perf_event.c int idx = hwc->idx; idx 998 arch/sparc/kernel/perf_event.c cpuc->current_idx[i] = idx; idx 1008 arch/sparc/kernel/perf_event.c int idx = cp->hw.idx; idx 1010 arch/sparc/kernel/perf_event.c cpuc->pcr[idx] |= cp->hw.config_base; idx 1084 arch/sparc/kernel/perf_event.c int idx = active_event_index(cpuc, event); idx 1088 arch/sparc/kernel/perf_event.c sparc_perf_event_set_period(event, &event->hw, idx); idx 1093 arch/sparc/kernel/perf_event.c sparc_pmu_enable_event(cpuc, &event->hw, idx); idx 1101 arch/sparc/kernel/perf_event.c int idx = active_event_index(cpuc, event); idx 1104 arch/sparc/kernel/perf_event.c sparc_pmu_disable_event(cpuc, &event->hw, idx); idx 1109 arch/sparc/kernel/perf_event.c sparc_perf_event_update(event, &event->hw, idx); idx 1152 arch/sparc/kernel/perf_event.c int idx = active_event_index(cpuc, event); idx 1155 arch/sparc/kernel/perf_event.c sparc_perf_event_update(event, hwc, idx); idx 1261 arch/sparc/kernel/perf_event.c evts[i]->hw.idx = i; idx 1308 arch/sparc/kernel/perf_event.c evts[0]->hw.idx = idx0; idx 1310 arch/sparc/kernel/perf_event.c evts[1]->hw.idx = idx0 ^ 1; idx 1490 arch/sparc/kernel/perf_event.c hwc->idx = PIC_NO_INDEX; idx 1654 arch/sparc/kernel/perf_event.c int idx = cpuc->current_idx[i]; idx 1660 arch/sparc/kernel/perf_event.c pcr_ops->write_pcr(idx, cpuc->pcr[idx]); idx 1663 arch/sparc/kernel/perf_event.c val = sparc_perf_event_update(event, hwc, idx); idx 1668 arch/sparc/kernel/perf_event.c if (!sparc_perf_event_set_period(event, hwc, idx)) idx 110 arch/sparc/kernel/sun4d_irq.c unsigned int idx, mask; idx 124 arch/sparc/kernel/sun4d_irq.c for (idx = 0; mask != 0; idx++, slot <<= 1) { idx 132 arch/sparc/kernel/sun4d_irq.c pil = sun4d_encode_irq(sbino, sbusl, idx); idx 55 arch/sparc/mm/highmem.c long idx, type; idx 63 arch/sparc/mm/highmem.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 64 arch/sparc/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 74 arch/sparc/mm/highmem.c BUG_ON(!pte_none(*(kmap_pte-idx))); idx 76 arch/sparc/mm/highmem.c set_pte(kmap_pte-idx, mk_pte(page, kmap_prot)); idx 103 arch/sparc/mm/highmem.c unsigned long idx; idx 105 arch/sparc/mm/highmem.c idx = type + KM_TYPE_NR * smp_processor_id(); idx 106 arch/sparc/mm/highmem.c BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN+idx)); idx 119 arch/sparc/mm/highmem.c pte_clear(&init_mm, vaddr, kmap_pte-idx); idx 35 arch/sparc/mm/tsb.c unsigned long idx; idx 37 arch/sparc/mm/tsb.c for (idx = 0; idx < KERNEL_TSB_NENTRIES; idx++) { idx 38 arch/sparc/mm/tsb.c struct tsb *ent = &swapper_tsb[idx]; idx 39 arch/sparc/mm/tsb.c unsigned long match = idx << 13; idx 192 arch/sparc/net/bpf_jit_comp_64.c int idx; idx 239 arch/sparc/net/bpf_jit_comp_64.c ctx->image[ctx->idx] = insn; idx 241 arch/sparc/net/bpf_jit_comp_64.c ctx->idx++; idx 247 arch/sparc/net/bpf_jit_comp_64.c void *here = &ctx->image[ctx->idx]; idx 251 arch/sparc/net/bpf_jit_comp_64.c ctx->image[ctx->idx] = CALL | ((off >> 2) & 0x3fffffff); idx 253 arch/sparc/net/bpf_jit_comp_64.c ctx->idx++; idx 667 arch/sparc/net/bpf_jit_comp_64.c if (!is_simm10(branch_dst - ctx->idx) || idx 740 arch/sparc/net/bpf_jit_comp_64.c emit_branch(br_opcode, ctx->idx, branch_dst, ctx); idx 784 arch/sparc/net/bpf_jit_comp_64.c emit_cbcondi(cbcond_opcode, ctx->idx, branch_dst, idx 787 arch/sparc/net/bpf_jit_comp_64.c emit_cbcond(cbcond_opcode, ctx->idx, branch_dst, idx 841 arch/sparc/net/bpf_jit_comp_64.c ctx->epilogue_offset = ctx->idx; idx 863 arch/sparc/net/bpf_jit_comp_64.c emit_branch(BGEU, ctx->idx, ctx->idx + OFFSET1, ctx); idx 870 arch/sparc/net/bpf_jit_comp_64.c emit_branch(BGU, ctx->idx, ctx->idx + OFFSET2, ctx); idx 884 arch/sparc/net/bpf_jit_comp_64.c emit_branch(BE, ctx->idx, ctx->idx + OFFSET3, ctx); idx 1171 arch/sparc/net/bpf_jit_comp_64.c emit_branch(BA, ctx->idx, ctx->offset[i + off], ctx); idx 1238 arch/sparc/net/bpf_jit_comp_64.c emit_branch(BA, ctx->idx, ctx->epilogue_offset, ctx); idx 1437 arch/sparc/net/bpf_jit_comp_64.c ctx->offset[i] = ctx->idx; idx 1440 arch/sparc/net/bpf_jit_comp_64.c ctx->offset[i] = ctx->idx; idx 1506 arch/sparc/net/bpf_jit_comp_64.c image_size = sizeof(u32) * ctx.idx; idx 1529 arch/sparc/net/bpf_jit_comp_64.c ctx.idx = 0; idx 1540 arch/sparc/net/bpf_jit_comp_64.c ctx.idx * 4, idx 1548 arch/sparc/net/bpf_jit_comp_64.c if (ctx.idx * 4 == prev_image_size) idx 1550 arch/sparc/net/bpf_jit_comp_64.c prev_image_size = ctx.idx * 4; idx 1555 arch/sparc/net/bpf_jit_comp_64.c image_size = sizeof(u32) * ctx.idx; idx 1565 arch/sparc/net/bpf_jit_comp_64.c ctx.idx = 0; idx 1577 arch/sparc/net/bpf_jit_comp_64.c if (ctx.idx * 4 != prev_image_size) { idx 1579 arch/sparc/net/bpf_jit_comp_64.c prev_image_size, ctx.idx * 4); idx 40 arch/um/include/asm/fixmap.h extern void __set_fixmap (enum fixed_addresses idx, idx 347 arch/unicore32/kernel/pci.c int idx; idx 352 arch/unicore32/kernel/pci.c for (idx = 0; idx < 6; idx++) { idx 354 arch/unicore32/kernel/pci.c if (!(mask & (1 << idx))) idx 357 arch/unicore32/kernel/pci.c r = dev->resource + idx; idx 218 arch/unicore32/kernel/signal.c unsigned int idx = 0; idx 221 arch/unicore32/kernel/signal.c idx += 1; idx 223 arch/unicore32/kernel/signal.c if (__put_user(sigreturn_codes[idx], rc) || idx 224 arch/unicore32/kernel/signal.c __put_user(sigreturn_codes[idx+1], rc+1)) idx 227 arch/unicore32/kernel/signal.c retcode = KERN_SIGRETURN_CODE + (idx << 2); idx 78 arch/x86/boot/early_serial_console.c int idx = 0; idx 84 arch/x86/boot/early_serial_console.c idx = 1; idx 86 arch/x86/boot/early_serial_console.c port = bases[idx]; idx 438 arch/x86/events/amd/core.c int idx, new = -1; idx 456 arch/x86/events/amd/core.c for_each_set_bit(idx, c->idxmsk, x86_pmu.num_counters) { idx 457 arch/x86/events/amd/core.c if (new == -1 || hwc->idx == idx) idx 459 arch/x86/events/amd/core.c old = cmpxchg(nb->owners + idx, NULL, event); idx 460 arch/x86/events/amd/core.c else if (nb->owners[idx] == event) idx 472 arch/x86/events/amd/core.c new = idx; idx 581 arch/x86/events/amd/core.c static void amd_pmu_wait_on_overflow(int idx) idx 592 arch/x86/events/amd/core.c rdmsrl(x86_pmu_event_addr(idx), counter); idx 604 arch/x86/events/amd/core.c int idx; idx 622 arch/x86/events/amd/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 623 arch/x86/events/amd/core.c if (!test_bit(idx, cpuc->active_mask)) idx 626 arch/x86/events/amd/core.c amd_pmu_wait_on_overflow(idx); idx 644 arch/x86/events/amd/core.c amd_pmu_wait_on_overflow(event->hw.idx); idx 698 arch/x86/events/amd/core.c amd_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 807 arch/x86/events/amd/core.c amd_get_event_constraints_f15h(struct cpu_hw_events *cpuc, int idx, idx 882 arch/x86/events/amd/core.c amd_get_event_constraints_f17h(struct cpu_hw_events *cpuc, int idx, idx 412 arch/x86/events/amd/iommu.c static __init int init_one_iommu(unsigned int idx) idx 424 arch/x86/events/amd/iommu.c perf_iommu->iommu = get_amd_iommu(idx); idx 425 arch/x86/events/amd/iommu.c perf_iommu->max_banks = amd_iommu_pc_get_max_banks(idx); idx 426 arch/x86/events/amd/iommu.c perf_iommu->max_counters = amd_iommu_pc_get_max_counters(idx); idx 435 arch/x86/events/amd/iommu.c snprintf(perf_iommu->name, IOMMU_NAME_SIZE, "amd_iommu_%u", idx); idx 440 arch/x86/events/amd/iommu.c idx, perf_iommu->max_banks, perf_iommu->max_counters); idx 443 arch/x86/events/amd/iommu.c pr_warn("Error initializing IOMMU %d.\n", idx); idx 31 arch/x86/events/amd/iommu.h extern u8 amd_iommu_pc_get_max_banks(unsigned int idx); idx 33 arch/x86/events/amd/iommu.h extern u8 amd_iommu_pc_get_max_counters(unsigned int idx); idx 41 arch/x86/events/amd/iommu.h extern struct amd_iommu *get_amd_iommu(int idx); idx 133 arch/x86/events/amd/uncore.c if (hwc->idx != -1 && uncore->events[hwc->idx] == event) idx 138 arch/x86/events/amd/uncore.c hwc->idx = i; idx 144 arch/x86/events/amd/uncore.c hwc->idx = -1; idx 147 arch/x86/events/amd/uncore.c hwc->idx = i; idx 153 arch/x86/events/amd/uncore.c if (hwc->idx == -1) idx 156 arch/x86/events/amd/uncore.c hwc->config_base = uncore->msr_base + (2 * hwc->idx); idx 157 arch/x86/events/amd/uncore.c hwc->event_base = uncore->msr_base + 1 + (2 * hwc->idx); idx 158 arch/x86/events/amd/uncore.c hwc->event_base_rdpmc = uncore->rdpmc_base + hwc->idx; idx 180 arch/x86/events/amd/uncore.c hwc->idx = -1; idx 200 arch/x86/events/amd/uncore.c hwc->idx = -1; idx 73 arch/x86/events/core.c int idx = hwc->idx; idx 76 arch/x86/events/core.c if (idx == INTEL_PMC_IDX_FIXED_BTS) idx 133 arch/x86/events/core.c reg->idx = er->idx; idx 603 arch/x86/events/core.c event->hw.idx = -1; idx 608 arch/x86/events/core.c event->hw.extra_reg.idx = EXTRA_REG_NONE; idx 609 arch/x86/events/core.c event->hw.branch_reg.idx = EXTRA_REG_NONE; idx 617 arch/x86/events/core.c int idx; idx 619 arch/x86/events/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 622 arch/x86/events/core.c if (!test_bit(idx, cpuc->active_mask)) idx 624 arch/x86/events/core.c rdmsrl(x86_pmu_config_addr(idx), val); idx 628 arch/x86/events/core.c wrmsrl(x86_pmu_config_addr(idx), val); idx 665 arch/x86/events/core.c int idx; idx 667 arch/x86/events/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 668 arch/x86/events/core.c struct hw_perf_event *hwc = &cpuc->events[idx]->hw; idx 670 arch/x86/events/core.c if (!test_bit(idx, cpuc->active_mask)) idx 723 arch/x86/events/core.c int idx; idx 731 arch/x86/events/core.c for (idx = 0; idx < num; idx++) { idx 732 arch/x86/events/core.c if (constraints[idx]->weight == wmin) idx 736 arch/x86/events/core.c sched->state.event = idx; /* start with min weight */ idx 771 arch/x86/events/core.c int idx; idx 782 arch/x86/events/core.c idx = INTEL_PMC_IDX_FIXED; idx 783 arch/x86/events/core.c for_each_set_bit_from(idx, c->idxmsk, X86_PMC_IDX_MAX) { idx 784 arch/x86/events/core.c if (!__test_and_set_bit(idx, sched->state.used)) idx 790 arch/x86/events/core.c idx = sched->state.counter; idx 791 arch/x86/events/core.c for_each_set_bit_from(idx, c->idxmsk, INTEL_PMC_IDX_FIXED) { idx 792 arch/x86/events/core.c if (!__test_and_set_bit(idx, sched->state.used)) { idx 803 arch/x86/events/core.c sched->state.counter = idx; idx 926 arch/x86/events/core.c if (hwc->idx == -1) idx 930 arch/x86/events/core.c if (!test_bit(hwc->idx, c->idxmsk)) idx 934 arch/x86/events/core.c if (test_bit(hwc->idx, used_mask)) idx 937 arch/x86/events/core.c __set_bit(hwc->idx, used_mask); idx 939 arch/x86/events/core.c assign[i] = hwc->idx; idx 1063 arch/x86/events/core.c hwc->idx = cpuc->assign[i]; idx 1067 arch/x86/events/core.c if (hwc->idx == INTEL_PMC_IDX_FIXED_BTS) { idx 1070 arch/x86/events/core.c } else if (hwc->idx >= INTEL_PMC_IDX_FIXED) { idx 1072 arch/x86/events/core.c hwc->event_base = MSR_ARCH_PERFMON_FIXED_CTR0 + (hwc->idx - INTEL_PMC_IDX_FIXED); idx 1073 arch/x86/events/core.c hwc->event_base_rdpmc = (hwc->idx - INTEL_PMC_IDX_FIXED) | 1<<30; idx 1075 arch/x86/events/core.c hwc->config_base = x86_pmu_config_addr(hwc->idx); idx 1076 arch/x86/events/core.c hwc->event_base = x86_pmu_event_addr(hwc->idx); idx 1077 arch/x86/events/core.c hwc->event_base_rdpmc = x86_pmu_rdpmc_index(hwc->idx); idx 1106 arch/x86/events/core.c return hwc->idx == cpuc->assign[i] && idx 1144 arch/x86/events/core.c if (hwc->idx == -1 || idx 1196 arch/x86/events/core.c int ret = 0, idx = hwc->idx; idx 1198 arch/x86/events/core.c if (idx == INTEL_PMC_IDX_FIXED_BTS) idx 1229 arch/x86/events/core.c per_cpu(pmc_prev_left[idx], smp_processor_id()) = left; idx 1330 arch/x86/events/core.c int idx = event->hw.idx; idx 1335 arch/x86/events/core.c if (WARN_ON_ONCE(idx == -1)) idx 1345 arch/x86/events/core.c cpuc->events[idx] = event; idx 1346 arch/x86/events/core.c __set_bit(idx, cpuc->active_mask); idx 1347 arch/x86/events/core.c __set_bit(idx, cpuc->running); idx 1358 arch/x86/events/core.c int cpu, idx; idx 1390 arch/x86/events/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 1391 arch/x86/events/core.c rdmsrl(x86_pmu_config_addr(idx), pmc_ctrl); idx 1392 arch/x86/events/core.c rdmsrl(x86_pmu_event_addr(idx), pmc_count); idx 1394 arch/x86/events/core.c prev_left = per_cpu(pmc_prev_left[idx], cpu); idx 1397 arch/x86/events/core.c cpu, idx, pmc_ctrl); idx 1399 arch/x86/events/core.c cpu, idx, pmc_count); idx 1401 arch/x86/events/core.c cpu, idx, prev_left); idx 1403 arch/x86/events/core.c for (idx = 0; idx < x86_pmu.num_counters_fixed; idx++) { idx 1404 arch/x86/events/core.c rdmsrl(MSR_ARCH_PERFMON_FIXED_CTR0 + idx, pmc_count); idx 1407 arch/x86/events/core.c cpu, idx, pmc_count); idx 1417 arch/x86/events/core.c if (test_bit(hwc->idx, cpuc->active_mask)) { idx 1419 arch/x86/events/core.c __clear_bit(hwc->idx, cpuc->active_mask); idx 1420 arch/x86/events/core.c cpuc->events[hwc->idx] = NULL; idx 1496 arch/x86/events/core.c int idx, handled = 0; idx 1511 arch/x86/events/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 1512 arch/x86/events/core.c if (!test_bit(idx, cpuc->active_mask)) idx 1515 arch/x86/events/core.c event = cpuc->events[idx]; idx 1719 arch/x86/events/core.c is_visible(struct kobject *kobj, struct attribute *attr, int idx) idx 1723 arch/x86/events/core.c if (idx >= x86_pmu.max_events) idx 1728 arch/x86/events/core.c return pmu_attr->event_str || x86_pmu.event_map(idx) ? attr->mode : 0; idx 2158 arch/x86/events/core.c int idx = event->hw.idx; idx 2163 arch/x86/events/core.c if (x86_pmu.num_counters_fixed && idx >= INTEL_PMC_IDX_FIXED) { idx 2164 arch/x86/events/core.c idx -= INTEL_PMC_IDX_FIXED; idx 2165 arch/x86/events/core.c idx |= 1 << 30; idx 2168 arch/x86/events/core.c return idx + 1; idx 2400 arch/x86/events/core.c unsigned int idx = segment >> 3; idx 2408 arch/x86/events/core.c if (!ldt || idx >= ldt->nr_entries) idx 2411 arch/x86/events/core.c desc = &ldt->entries[idx]; idx 2416 arch/x86/events/core.c if (idx >= GDT_ENTRIES) idx 2419 arch/x86/events/core.c desc = raw_cpu_ptr(gdt_page.gdt) + idx; idx 138 arch/x86/events/intel/bts.c static unsigned long bts_buffer_offset(struct bts_buffer *buf, unsigned int idx) idx 140 arch/x86/events/intel/bts.c return buf->buf[idx].offset + buf->buf[idx].displacement; idx 2083 arch/x86/events/intel/core.c static void intel_tfa_commit_scheduling(struct cpu_hw_events *cpuc, int idx, int cntr) idx 2134 arch/x86/events/intel/core.c int idx = hwc->idx - INTEL_PMC_IDX_FIXED; idx 2137 arch/x86/events/intel/core.c mask = 0xfULL << (idx * 4); idx 2154 arch/x86/events/intel/core.c if (unlikely(hwc->idx == INTEL_PMC_IDX_FIXED_BTS)) { idx 2160 arch/x86/events/intel/core.c cpuc->intel_ctrl_guest_mask &= ~(1ull << hwc->idx); idx 2161 arch/x86/events/intel/core.c cpuc->intel_ctrl_host_mask &= ~(1ull << hwc->idx); idx 2162 arch/x86/events/intel/core.c cpuc->intel_cp_status &= ~(1ull << hwc->idx); idx 2196 arch/x86/events/intel/core.c int idx = hwc->idx - INTEL_PMC_IDX_FIXED; idx 2217 arch/x86/events/intel/core.c bits <<= (idx * 4); idx 2218 arch/x86/events/intel/core.c mask = 0xfULL << (idx * 4); idx 2221 arch/x86/events/intel/core.c bits |= ICL_FIXED_0_ADAPTIVE << (idx * 4); idx 2222 arch/x86/events/intel/core.c mask |= ICL_FIXED_0_ADAPTIVE << (idx * 4); idx 2236 arch/x86/events/intel/core.c if (unlikely(hwc->idx == INTEL_PMC_IDX_FIXED_BTS)) { idx 2245 arch/x86/events/intel/core.c cpuc->intel_ctrl_guest_mask |= (1ull << hwc->idx); idx 2247 arch/x86/events/intel/core.c cpuc->intel_ctrl_host_mask |= (1ull << hwc->idx); idx 2250 arch/x86/events/intel/core.c cpuc->intel_cp_status |= (1ull << hwc->idx); idx 2296 arch/x86/events/intel/core.c int idx; idx 2305 arch/x86/events/intel/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 2306 arch/x86/events/intel/core.c wrmsrl_safe(x86_pmu_config_addr(idx), 0ull); idx 2307 arch/x86/events/intel/core.c wrmsrl_safe(x86_pmu_event_addr(idx), 0ull); idx 2309 arch/x86/events/intel/core.c for (idx = 0; idx < x86_pmu.num_counters_fixed; idx++) idx 2310 arch/x86/events/intel/core.c wrmsrl_safe(MSR_ARCH_PERFMON_FIXED_CTR0 + idx, 0ull); idx 2598 arch/x86/events/intel/core.c static int intel_alt_er(int idx, u64 config) idx 2600 arch/x86/events/intel/core.c int alt_idx = idx; idx 2603 arch/x86/events/intel/core.c return idx; idx 2605 arch/x86/events/intel/core.c if (idx == EXTRA_REG_RSP_0) idx 2608 arch/x86/events/intel/core.c if (idx == EXTRA_REG_RSP_1) idx 2612 arch/x86/events/intel/core.c return idx; idx 2617 arch/x86/events/intel/core.c static void intel_fixup_er(struct perf_event *event, int idx) idx 2619 arch/x86/events/intel/core.c event->hw.extra_reg.idx = idx; idx 2621 arch/x86/events/intel/core.c if (idx == EXTRA_REG_RSP_0) { idx 2625 arch/x86/events/intel/core.c } else if (idx == EXTRA_REG_RSP_1) { idx 2647 arch/x86/events/intel/core.c int idx = reg->idx; idx 2658 arch/x86/events/intel/core.c era = &cpuc->shared_regs->regs[idx]; idx 2678 arch/x86/events/intel/core.c if (idx != reg->idx) idx 2679 arch/x86/events/intel/core.c intel_fixup_er(event, idx); idx 2703 arch/x86/events/intel/core.c idx = intel_alt_er(idx, reg->config); idx 2704 arch/x86/events/intel/core.c if (idx != reg->idx) { idx 2731 arch/x86/events/intel/core.c era = &cpuc->shared_regs->regs[reg->idx]; idx 2748 arch/x86/events/intel/core.c if (xreg->idx != EXTRA_REG_NONE) { idx 2754 arch/x86/events/intel/core.c if (breg->idx != EXTRA_REG_NONE) { idx 2765 arch/x86/events/intel/core.c x86_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 2783 arch/x86/events/intel/core.c __intel_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 2800 arch/x86/events/intel/core.c return x86_get_event_constraints(cpuc, idx, event); idx 2833 arch/x86/events/intel/core.c static void intel_commit_scheduling(struct cpu_hw_events *cpuc, int idx, int cntr) idx 2836 arch/x86/events/intel/core.c struct event_constraint *c = cpuc->event_constraint[idx]; idx 2887 arch/x86/events/intel/core.c dyn_constraint(struct cpu_hw_events *cpuc, struct event_constraint *c, int idx) idx 2897 arch/x86/events/intel/core.c cx = &cpuc->constraint_list[idx]; idx 2917 arch/x86/events/intel/core.c int idx, struct event_constraint *c) idx 2945 arch/x86/events/intel/core.c c = dyn_constraint(cpuc, c, idx); idx 3016 arch/x86/events/intel/core.c intel_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 3021 arch/x86/events/intel/core.c c1 = cpuc->event_constraint[idx]; idx 3028 arch/x86/events/intel/core.c c2 = __intel_get_event_constraints(cpuc, idx, event); idx 3037 arch/x86/events/intel/core.c return intel_get_excl_constraints(cpuc, event, idx, c2); idx 3069 arch/x86/events/intel/core.c if (hwc->idx >= 0) { idx 3080 arch/x86/events/intel/core.c xl->state[hwc->idx] = INTEL_EXCL_UNUSED; idx 3094 arch/x86/events/intel/core.c if (reg->idx != EXTRA_REG_NONE) idx 3098 arch/x86/events/intel/core.c if (reg->idx != EXTRA_REG_NONE) idx 3371 arch/x86/events/intel/core.c int idx; idx 3373 arch/x86/events/intel/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 3374 arch/x86/events/intel/core.c struct perf_event *event = cpuc->events[idx]; idx 3376 arch/x86/events/intel/core.c arr[idx].msr = x86_pmu_config_addr(idx); idx 3377 arch/x86/events/intel/core.c arr[idx].host = arr[idx].guest = 0; idx 3379 arch/x86/events/intel/core.c if (!test_bit(idx, cpuc->active_mask)) idx 3382 arch/x86/events/intel/core.c arr[idx].host = arr[idx].guest = idx 3386 arch/x86/events/intel/core.c arr[idx].host &= ~ARCH_PERFMON_EVENTSEL_ENABLE; idx 3388 arch/x86/events/intel/core.c arr[idx].guest &= ~ARCH_PERFMON_EVENTSEL_ENABLE; idx 3404 arch/x86/events/intel/core.c int idx; idx 3406 arch/x86/events/intel/core.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 3407 arch/x86/events/intel/core.c struct hw_perf_event *hwc = &cpuc->events[idx]->hw; idx 3409 arch/x86/events/intel/core.c if (!test_bit(idx, cpuc->active_mask) || idx 3410 arch/x86/events/intel/core.c cpuc->events[idx]->attr.exclude_host) idx 3467 arch/x86/events/intel/core.c hsw_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 3472 arch/x86/events/intel/core.c c = intel_get_event_constraints(cpuc, idx, event); idx 3485 arch/x86/events/intel/core.c icl_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 3496 arch/x86/events/intel/core.c return hsw_get_event_constraints(cpuc, idx, event); idx 3500 arch/x86/events/intel/core.c glp_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 3509 arch/x86/events/intel/core.c c = intel_get_event_constraints(cpuc, idx, event); idx 3515 arch/x86/events/intel/core.c tnt_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 3532 arch/x86/events/intel/core.c c = intel_get_event_constraints(cpuc, idx, event); idx 3540 arch/x86/events/intel/core.c tfa_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 3543 arch/x86/events/intel/core.c struct event_constraint *c = hsw_get_event_constraints(cpuc, idx, event); idx 3549 arch/x86/events/intel/core.c c = dyn_constraint(cpuc, c, idx); idx 5154 arch/x86/events/intel/core.c if ((er->idx == EXTRA_REG_LBR) && !er->extra_msr_access) idx 166 arch/x86/events/intel/cstate.c static bool test_msr(int idx, void *data) idx 168 arch/x86/events/intel/cstate.c return test_bit(idx, (unsigned long *) data); idx 346 arch/x86/events/intel/cstate.c event->hw.idx = -1; idx 1099 arch/x86/events/intel/ds.c wrmsrl(MSR_RELOAD_PMC0 + hwc->idx, ds->pebs_event_reset[hwc->idx]); idx 1110 arch/x86/events/intel/ds.c cpuc->pebs_enabled |= 1ULL << hwc->idx; idx 1113 arch/x86/events/intel/ds.c cpuc->pebs_enabled |= 1ULL << (hwc->idx + 32); idx 1130 arch/x86/events/intel/ds.c unsigned int idx = hwc->idx; idx 1132 arch/x86/events/intel/ds.c if (idx >= INTEL_PMC_IDX_FIXED) idx 1133 arch/x86/events/intel/ds.c idx = MAX_PEBS_EVENTS + (idx - INTEL_PMC_IDX_FIXED); idx 1134 arch/x86/events/intel/ds.c ds->pebs_event_reset[idx] = idx 1137 arch/x86/events/intel/ds.c ds->pebs_event_reset[hwc->idx] = 0; idx 1167 arch/x86/events/intel/ds.c cpuc->pebs_enabled &= ~(1ULL << hwc->idx); idx 1171 arch/x86/events/intel/ds.c cpuc->pebs_enabled &= ~(1ULL << (hwc->idx + 32)); idx 185 arch/x86/events/intel/knc.c (void)wrmsrl_safe(hwc->config_base + hwc->idx, val); idx 196 arch/x86/events/intel/knc.c (void)wrmsrl_safe(hwc->config_base + hwc->idx, val); idx 311 arch/x86/events/intel/lbr.c static inline void wrlbr_from(unsigned int idx, u64 val) idx 314 arch/x86/events/intel/lbr.c wrmsrl(x86_pmu.lbr_from + idx, val); idx 317 arch/x86/events/intel/lbr.c static inline void wrlbr_to(unsigned int idx, u64 val) idx 319 arch/x86/events/intel/lbr.c wrmsrl(x86_pmu.lbr_to + idx, val); idx 322 arch/x86/events/intel/lbr.c static inline u64 rdlbr_from(unsigned int idx) idx 326 arch/x86/events/intel/lbr.c rdmsrl(x86_pmu.lbr_from + idx, val); idx 331 arch/x86/events/intel/lbr.c static inline u64 rdlbr_to(unsigned int idx) idx 335 arch/x86/events/intel/lbr.c rdmsrl(x86_pmu.lbr_to + idx, val); idx 776 arch/x86/events/intel/lbr.c reg->idx = EXTRA_REG_LBR; idx 919 arch/x86/events/intel/p4.c int idx; idx 921 arch/x86/events/intel/p4.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 922 arch/x86/events/intel/p4.c struct perf_event *event = cpuc->events[idx]; idx 923 arch/x86/events/intel/p4.c if (!test_bit(idx, cpuc->active_mask)) idx 935 arch/x86/events/intel/p4.c unsigned int idx; idx 939 arch/x86/events/intel/p4.c idx = p4_config_unpack_metric(config); idx 940 arch/x86/events/intel/p4.c if (idx == P4_PEBS_METRIC__none) idx 943 arch/x86/events/intel/p4.c bind = &p4_pebs_bind_map[idx]; idx 954 arch/x86/events/intel/p4.c unsigned int idx = p4_config_unpack_event(hwc->config); idx 958 arch/x86/events/intel/p4.c bind = &p4_event_bind_map[idx]; idx 966 arch/x86/events/intel/p4.c WARN_ON_ONCE(hwc->idx == 1); idx 988 arch/x86/events/intel/p4.c int idx; idx 990 arch/x86/events/intel/p4.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 991 arch/x86/events/intel/p4.c struct perf_event *event = cpuc->events[idx]; idx 992 arch/x86/events/intel/p4.c if (!test_bit(idx, cpuc->active_mask)) idx 1004 arch/x86/events/intel/p4.c int idx, handled = 0; idx 1009 arch/x86/events/intel/p4.c for (idx = 0; idx < x86_pmu.num_counters; idx++) { idx 1012 arch/x86/events/intel/p4.c if (!test_bit(idx, cpuc->active_mask)) { idx 1014 arch/x86/events/intel/p4.c if (__test_and_clear_bit(idx, cpuc->running)) idx 1019 arch/x86/events/intel/p4.c event = cpuc->events[idx]; idx 1022 arch/x86/events/intel/p4.c WARN_ON_ONCE(hwc->idx != idx); idx 1180 arch/x86/events/intel/p4.c unsigned int idx = P4_ESCR_MSR_IDX(addr); idx 1182 arch/x86/events/intel/p4.c if (unlikely(idx >= P4_ESCR_MSR_TABLE_SIZE || idx 1183 arch/x86/events/intel/p4.c !p4_escr_table[idx] || idx 1184 arch/x86/events/intel/p4.c p4_escr_table[idx] != addr)) { idx 1189 arch/x86/events/intel/p4.c return idx; idx 1241 arch/x86/events/intel/p4.c if (hwc->idx != -1 && !p4_should_swap_ts(hwc->config, cpu)) { idx 1242 arch/x86/events/intel/p4.c cntr_idx = hwc->idx; idx 1244 arch/x86/events/intel/p4.c assign[i] = hwc->idx; idx 1277 arch/x86/events/intel/p4.c hwc->idx = -1; idx 925 arch/x86/events/intel/pt.c unsigned int idx, cur_pg = 0, z_pg = 0, start_idx = 0; idx 969 arch/x86/events/intel/pt.c idx = (pg - cur_pg) / TOPA_ENTRY_PAGES(topa, 0); idx 970 arch/x86/events/intel/pt.c return &tp->table[idx]; idx 976 arch/x86/events/intel/pt.c for (idx = start_idx, cur_pg += z_pg; idx < topa->last; idx++) { idx 977 arch/x86/events/intel/pt.c if (cur_pg + TOPA_ENTRY_PAGES(topa, idx) > pg) idx 978 arch/x86/events/intel/pt.c return &tp->table[idx]; idx 980 arch/x86/events/intel/pt.c cur_pg += TOPA_ENTRY_PAGES(topa, idx); idx 1031 arch/x86/events/intel/pt.c unsigned long idx, npages, wakeup; idx 1060 arch/x86/events/intel/pt.c idx = (head >> PAGE_SHIFT) + npages; idx 1061 arch/x86/events/intel/pt.c idx &= buf->nr_pages - 1; idx 1063 arch/x86/events/intel/pt.c if (idx != buf->stop_pos) { idx 1064 arch/x86/events/intel/pt.c buf->stop_pos = idx; idx 1065 arch/x86/events/intel/pt.c buf->stop_te = pt_topa_entry_for_page(buf, idx); idx 1072 arch/x86/events/intel/pt.c idx = (head >> PAGE_SHIFT) + npages - 1; idx 1073 arch/x86/events/intel/pt.c if (idx > wakeup) idx 1074 arch/x86/events/intel/pt.c idx = wakeup; idx 1076 arch/x86/events/intel/pt.c idx &= buf->nr_pages - 1; idx 1077 arch/x86/events/intel/pt.c if (idx != buf->intr_pos) { idx 1078 arch/x86/events/intel/pt.c buf->intr_pos = idx; idx 1079 arch/x86/events/intel/pt.c buf->intr_te = pt_topa_entry_for_page(buf, idx); idx 371 arch/x86/events/intel/rapl.c event->hw.idx = bit; idx 511 arch/x86/events/intel/rapl.c static bool test_msr(int idx, void *data) idx 513 arch/x86/events/intel/rapl.c return test_bit(idx, (unsigned long *) data); idx 155 arch/x86/events/intel/uncore.c if (reg1->idx == EXTRA_REG_NONE || idx 159 arch/x86/events/intel/uncore.c er = &box->shared_regs[reg1->idx]; idx 195 arch/x86/events/intel/uncore.c er = &box->shared_regs[reg1->idx]; idx 200 arch/x86/events/intel/uncore.c u64 uncore_shared_reg_config(struct intel_uncore_box *box, int idx) idx 206 arch/x86/events/intel/uncore.c er = &box->shared_regs[idx]; idx 216 arch/x86/events/intel/uncore.c struct perf_event *event, int idx) idx 220 arch/x86/events/intel/uncore.c hwc->idx = idx; idx 221 arch/x86/events/intel/uncore.c hwc->last_tag = ++box->tags[idx]; idx 223 arch/x86/events/intel/uncore.c if (uncore_pmc_fixed(hwc->idx)) { idx 229 arch/x86/events/intel/uncore.c hwc->config_base = uncore_event_ctl(box, hwc->idx); idx 230 arch/x86/events/intel/uncore.c hwc->event_base = uncore_perf_ctr(box, hwc->idx); idx 238 arch/x86/events/intel/uncore.c if (uncore_pmc_freerunning(event->hw.idx)) idx 240 arch/x86/events/intel/uncore.c else if (uncore_pmc_fixed(event->hw.idx)) idx 444 arch/x86/events/intel/uncore.c if (hwc->idx == -1) idx 448 arch/x86/events/intel/uncore.c if (!test_bit(hwc->idx, c->idxmsk)) idx 452 arch/x86/events/intel/uncore.c if (test_bit(hwc->idx, used_mask)) idx 455 arch/x86/events/intel/uncore.c __set_bit(hwc->idx, used_mask); idx 457 arch/x86/events/intel/uncore.c assign[i] = hwc->idx; idx 474 arch/x86/events/intel/uncore.c int idx = event->hw.idx; idx 476 arch/x86/events/intel/uncore.c if (WARN_ON_ONCE(idx == -1 || idx >= UNCORE_PMC_IDX_MAX)) idx 485 arch/x86/events/intel/uncore.c if (uncore_pmc_freerunning(event->hw.idx)) { idx 498 arch/x86/events/intel/uncore.c box->events[idx] = event; idx 500 arch/x86/events/intel/uncore.c __set_bit(idx, box->active_mask); idx 515 arch/x86/events/intel/uncore.c if (uncore_pmc_freerunning(hwc->idx)) { idx 523 arch/x86/events/intel/uncore.c if (__test_and_clear_bit(hwc->idx, box->active_mask)) { idx 526 arch/x86/events/intel/uncore.c box->events[hwc->idx] = NULL; idx 559 arch/x86/events/intel/uncore.c if (uncore_pmc_freerunning(hwc->idx)) { idx 582 arch/x86/events/intel/uncore.c if (hwc->idx == assign[i] && idx 600 arch/x86/events/intel/uncore.c if (hwc->idx != assign[i] || idx 628 arch/x86/events/intel/uncore.c if (uncore_pmc_freerunning(event->hw.idx)) idx 643 arch/x86/events/intel/uncore.c event->hw.idx = -1; idx 665 arch/x86/events/intel/uncore.c if (uncore_pmc_freerunning(event->hw.idx)) idx 729 arch/x86/events/intel/uncore.c event->hw.idx = -1; idx 731 arch/x86/events/intel/uncore.c event->hw.extra_reg.idx = EXTRA_REG_NONE; idx 732 arch/x86/events/intel/uncore.c event->hw.branch_reg.idx = EXTRA_REG_NONE; idx 751 arch/x86/events/intel/uncore.c event->hw.idx = UNCORE_PMC_IDX_FREERUNNING; idx 1000 arch/x86/events/intel/uncore.c int idx = UNCORE_PCI_DEV_IDX(id->driver_data); idx 1002 arch/x86/events/intel/uncore.c uncore_extra_pci_dev[die].dev[idx] = pdev; idx 24 arch/x86/events/intel/uncore.h #define UNCORE_PCI_DEV_FULL_DATA(dev, func, type, idx) \ idx 25 arch/x86/events/intel/uncore.h ((dev << 24) | (func << 16) | (type << 8) | idx) idx 26 arch/x86/events/intel/uncore.h #define UNCORE_PCI_DEV_DATA(type, idx) ((type << 8) | idx) idx 188 arch/x86/events/intel/uncore.h static inline bool uncore_pmc_fixed(int idx) idx 190 arch/x86/events/intel/uncore.h return idx == UNCORE_PMC_IDX_FIXED; idx 193 arch/x86/events/intel/uncore.h static inline bool uncore_pmc_freerunning(int idx) idx 195 arch/x86/events/intel/uncore.h return idx == UNCORE_PMC_IDX_FREERUNNING; idx 221 arch/x86/events/intel/uncore.h unsigned uncore_pci_event_ctl(struct intel_uncore_box *box, int idx) idx 224 arch/x86/events/intel/uncore.h return idx * 8 + box->pmu->type->event_ctl; idx 226 arch/x86/events/intel/uncore.h return idx * 4 + box->pmu->type->event_ctl; idx 230 arch/x86/events/intel/uncore.h unsigned uncore_pci_perf_ctr(struct intel_uncore_box *box, int idx) idx 232 arch/x86/events/intel/uncore.h return idx * 8 + box->pmu->type->perf_ctr; idx 308 arch/x86/events/intel/uncore.h unsigned int idx = uncore_freerunning_idx(event->hw.config); idx 312 arch/x86/events/intel/uncore.h pmu->type->freerunning[type].counter_offset * idx + idx 317 arch/x86/events/intel/uncore.h unsigned uncore_msr_event_ctl(struct intel_uncore_box *box, int idx) idx 321 arch/x86/events/intel/uncore.h (box->pmu->type->pair_ctr_ctl ? 2 * idx : idx); idx 324 arch/x86/events/intel/uncore.h (box->pmu->type->pair_ctr_ctl ? 2 * idx : idx) + idx 330 arch/x86/events/intel/uncore.h unsigned uncore_msr_perf_ctr(struct intel_uncore_box *box, int idx) idx 334 arch/x86/events/intel/uncore.h (box->pmu->type->pair_ctr_ctl ? 2 * idx : idx); idx 337 arch/x86/events/intel/uncore.h (box->pmu->type->pair_ctr_ctl ? 2 * idx : idx) + idx 361 arch/x86/events/intel/uncore.h unsigned uncore_event_ctl(struct intel_uncore_box *box, int idx) idx 364 arch/x86/events/intel/uncore.h return uncore_pci_event_ctl(box, idx); idx 366 arch/x86/events/intel/uncore.h return uncore_msr_event_ctl(box, idx); idx 370 arch/x86/events/intel/uncore.h unsigned uncore_perf_ctr(struct intel_uncore_box *box, int idx) idx 373 arch/x86/events/intel/uncore.h return uncore_pci_perf_ctr(box, idx); idx 375 arch/x86/events/intel/uncore.h return uncore_msr_perf_ctr(box, idx); idx 415 arch/x86/events/intel/uncore.h unsigned int idx = uncore_freerunning_idx(event->hw.config); idx 418 arch/x86/events/intel/uncore.h (idx < uncore_num_freerunning(box, event)); idx 509 arch/x86/events/intel/uncore.h u64 uncore_shared_reg_config(struct intel_uncore_box *box, int idx); idx 249 arch/x86/events/intel/uncore_nhmex.c if (hwc->idx == UNCORE_PMC_IDX_FIXED) idx 371 arch/x86/events/intel/uncore_nhmex.c reg1->idx = 0; idx 383 arch/x86/events/intel/uncore_nhmex.c if (reg1->idx != EXTRA_REG_NONE) { idx 456 arch/x86/events/intel/uncore_nhmex.c reg1->idx = 0; idx 468 arch/x86/events/intel/uncore_nhmex.c if (reg1->idx != EXTRA_REG_NONE) { idx 553 arch/x86/events/intel/uncore_nhmex.c static bool nhmex_mbox_get_shared_reg(struct intel_uncore_box *box, int idx, u64 config) idx 560 arch/x86/events/intel/uncore_nhmex.c if (idx < EXTRA_REG_NHMEX_M_ZDP_CTL_FVC) { idx 561 arch/x86/events/intel/uncore_nhmex.c er = &box->shared_regs[idx]; idx 577 arch/x86/events/intel/uncore_nhmex.c idx -= EXTRA_REG_NHMEX_M_ZDP_CTL_FVC; idx 578 arch/x86/events/intel/uncore_nhmex.c if (WARN_ON_ONCE(idx >= 4)) idx 590 arch/x86/events/intel/uncore_nhmex.c if (__BITS_VALUE(atomic_read(&er->ref), idx, 8)) { idx 592 arch/x86/events/intel/uncore_nhmex.c mask |= NHMEX_M_PMON_ZDP_CTL_FVC_EVENT_MASK(idx); idx 594 arch/x86/events/intel/uncore_nhmex.c mask |= WSMEX_M_PMON_ZDP_CTL_FVC_EVENT_MASK(idx); idx 598 arch/x86/events/intel/uncore_nhmex.c atomic_add(1 << (idx * 8), &er->ref); idx 601 arch/x86/events/intel/uncore_nhmex.c NHMEX_M_PMON_ZDP_CTL_FVC_EVENT_MASK(idx); idx 604 arch/x86/events/intel/uncore_nhmex.c WSMEX_M_PMON_ZDP_CTL_FVC_EVENT_MASK(idx); idx 614 arch/x86/events/intel/uncore_nhmex.c static void nhmex_mbox_put_shared_reg(struct intel_uncore_box *box, int idx) idx 618 arch/x86/events/intel/uncore_nhmex.c if (idx < EXTRA_REG_NHMEX_M_ZDP_CTL_FVC) { idx 619 arch/x86/events/intel/uncore_nhmex.c er = &box->shared_regs[idx]; idx 624 arch/x86/events/intel/uncore_nhmex.c idx -= EXTRA_REG_NHMEX_M_ZDP_CTL_FVC; idx 626 arch/x86/events/intel/uncore_nhmex.c atomic_sub(1 << (idx * 8), &er->ref); idx 633 arch/x86/events/intel/uncore_nhmex.c u64 idx, orig_idx = __BITS_VALUE(reg1->idx, 0, 8); idx 637 arch/x86/events/intel/uncore_nhmex.c idx = orig_idx - EXTRA_REG_NHMEX_M_ZDP_CTL_FVC; idx 639 arch/x86/events/intel/uncore_nhmex.c config &= NHMEX_M_PMON_ZDP_CTL_FVC_EVENT_MASK(idx); idx 641 arch/x86/events/intel/uncore_nhmex.c config &= WSMEX_M_PMON_ZDP_CTL_FVC_EVENT_MASK(idx); idx 643 arch/x86/events/intel/uncore_nhmex.c idx = new_idx - orig_idx; idx 644 arch/x86/events/intel/uncore_nhmex.c config <<= 3 * idx; idx 646 arch/x86/events/intel/uncore_nhmex.c idx = orig_idx - new_idx; idx 647 arch/x86/events/intel/uncore_nhmex.c config >>= 3 * idx; idx 659 arch/x86/events/intel/uncore_nhmex.c hwc->config += idx << NHMEX_M_PMON_CTL_INC_SEL_SHIFT; idx 661 arch/x86/events/intel/uncore_nhmex.c hwc->config -= idx << NHMEX_M_PMON_CTL_INC_SEL_SHIFT; idx 663 arch/x86/events/intel/uncore_nhmex.c reg1->idx = ~0xff | new_idx; idx 673 arch/x86/events/intel/uncore_nhmex.c int i, idx[2], alloc = 0; idx 676 arch/x86/events/intel/uncore_nhmex.c idx[0] = __BITS_VALUE(reg1->idx, 0, 8); idx 677 arch/x86/events/intel/uncore_nhmex.c idx[1] = __BITS_VALUE(reg1->idx, 1, 8); idx 681 arch/x86/events/intel/uncore_nhmex.c idx[i] = 0xff; idx 683 arch/x86/events/intel/uncore_nhmex.c if (idx[i] == 0xff) idx 686 arch/x86/events/intel/uncore_nhmex.c if (!nhmex_mbox_get_shared_reg(box, idx[i], idx 693 arch/x86/events/intel/uncore_nhmex.c if (reg2->idx != EXTRA_REG_NONE && idx 695 arch/x86/events/intel/uncore_nhmex.c !nhmex_mbox_get_shared_reg(box, reg2->idx, reg2->config)) idx 705 arch/x86/events/intel/uncore_nhmex.c if (idx[0] != 0xff && idx[0] != __BITS_VALUE(reg1->idx, 0, 8)) idx 706 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_alter_er(event, idx[0], true); idx 708 arch/x86/events/intel/uncore_nhmex.c if (reg2->idx != EXTRA_REG_NONE) idx 713 arch/x86/events/intel/uncore_nhmex.c if (idx[0] != 0xff && !(alloc & 0x1) && idx 714 arch/x86/events/intel/uncore_nhmex.c idx[0] >= EXTRA_REG_NHMEX_M_ZDP_CTL_FVC) { idx 721 arch/x86/events/intel/uncore_nhmex.c BUG_ON(__BITS_VALUE(reg1->idx, 1, 8) != 0xff); idx 722 arch/x86/events/intel/uncore_nhmex.c idx[0] -= EXTRA_REG_NHMEX_M_ZDP_CTL_FVC; idx 723 arch/x86/events/intel/uncore_nhmex.c idx[0] = (idx[0] + 1) % 4; idx 724 arch/x86/events/intel/uncore_nhmex.c idx[0] += EXTRA_REG_NHMEX_M_ZDP_CTL_FVC; idx 725 arch/x86/events/intel/uncore_nhmex.c if (idx[0] != __BITS_VALUE(reg1->idx, 0, 8)) { idx 726 arch/x86/events/intel/uncore_nhmex.c config1 = nhmex_mbox_alter_er(event, idx[0], false); idx 732 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_put_shared_reg(box, idx[0]); idx 734 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_put_shared_reg(box, idx[1]); idx 747 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_put_shared_reg(box, __BITS_VALUE(reg1->idx, 0, 8)); idx 749 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_put_shared_reg(box, __BITS_VALUE(reg1->idx, 1, 8)); idx 753 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_put_shared_reg(box, reg2->idx); idx 760 arch/x86/events/intel/uncore_nhmex.c if (er->idx < EXTRA_REG_NHMEX_M_ZDP_CTL_FVC) idx 761 arch/x86/events/intel/uncore_nhmex.c return er->idx; idx 762 arch/x86/events/intel/uncore_nhmex.c return er->idx + (er->event >> NHMEX_M_PMON_CTL_INC_SEL_SHIFT) - 0xd; idx 785 arch/x86/events/intel/uncore_nhmex.c if (WARN_ON_ONCE(msr >= 0xffff || er->idx >= 0xff)) idx 789 arch/x86/events/intel/uncore_nhmex.c if (er->idx == EXTRA_REG_NHMEX_M_PLD) idx 794 arch/x86/events/intel/uncore_nhmex.c reg1->idx &= ~(0xff << (reg_idx * 8)); idx 796 arch/x86/events/intel/uncore_nhmex.c reg1->idx |= nhmex_mbox_extra_reg_idx(er) << (reg_idx * 8); idx 806 arch/x86/events/intel/uncore_nhmex.c reg2->idx = EXTRA_REG_NHMEX_M_FILTER; idx 819 arch/x86/events/intel/uncore_nhmex.c static u64 nhmex_mbox_shared_reg_config(struct intel_uncore_box *box, int idx) idx 825 arch/x86/events/intel/uncore_nhmex.c if (idx < EXTRA_REG_NHMEX_M_ZDP_CTL_FVC) idx 826 arch/x86/events/intel/uncore_nhmex.c return box->shared_regs[idx].config; idx 840 arch/x86/events/intel/uncore_nhmex.c int idx; idx 842 arch/x86/events/intel/uncore_nhmex.c idx = __BITS_VALUE(reg1->idx, 0, 8); idx 843 arch/x86/events/intel/uncore_nhmex.c if (idx != 0xff) idx 845 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_shared_reg_config(box, idx)); idx 846 arch/x86/events/intel/uncore_nhmex.c idx = __BITS_VALUE(reg1->idx, 1, 8); idx 847 arch/x86/events/intel/uncore_nhmex.c if (idx != 0xff) idx 849 arch/x86/events/intel/uncore_nhmex.c nhmex_mbox_shared_reg_config(box, idx)); idx 851 arch/x86/events/intel/uncore_nhmex.c if (reg2->idx != EXTRA_REG_NONE) { idx 950 arch/x86/events/intel/uncore_nhmex.c if (reg1->idx % 2) { idx 951 arch/x86/events/intel/uncore_nhmex.c reg1->idx--; idx 954 arch/x86/events/intel/uncore_nhmex.c reg1->idx++; idx 959 arch/x86/events/intel/uncore_nhmex.c switch (reg1->idx % 6) { idx 985 arch/x86/events/intel/uncore_nhmex.c int idx, er_idx; idx 992 arch/x86/events/intel/uncore_nhmex.c idx = reg1->idx % 6; idx 995 arch/x86/events/intel/uncore_nhmex.c er_idx = idx; idx 999 arch/x86/events/intel/uncore_nhmex.c er_idx += (reg1->idx / 6) * 5; idx 1003 arch/x86/events/intel/uncore_nhmex.c if (idx < 2) { idx 1009 arch/x86/events/intel/uncore_nhmex.c } else if (idx == 2 || idx == 3) { idx 1014 arch/x86/events/intel/uncore_nhmex.c u64 mask = 0xff << ((idx - 2) * 8); idx 1015 arch/x86/events/intel/uncore_nhmex.c if (!__BITS_VALUE(atomic_read(&er->ref), idx - 2, 8) || idx 1017 arch/x86/events/intel/uncore_nhmex.c atomic_add(1 << ((idx - 2) * 8), &er->ref); idx 1043 arch/x86/events/intel/uncore_nhmex.c idx ^= 1; idx 1044 arch/x86/events/intel/uncore_nhmex.c if (idx != reg1->idx % 6) { idx 1045 arch/x86/events/intel/uncore_nhmex.c if (idx == 2) idx 1047 arch/x86/events/intel/uncore_nhmex.c else if (idx == 3) idx 1053 arch/x86/events/intel/uncore_nhmex.c if (idx != reg1->idx % 6) idx 1066 arch/x86/events/intel/uncore_nhmex.c int idx, er_idx; idx 1071 arch/x86/events/intel/uncore_nhmex.c idx = reg1->idx % 6; idx 1072 arch/x86/events/intel/uncore_nhmex.c er_idx = idx; idx 1075 arch/x86/events/intel/uncore_nhmex.c er_idx += (reg1->idx / 6) * 5; idx 1078 arch/x86/events/intel/uncore_nhmex.c if (idx == 2 || idx == 3) idx 1079 arch/x86/events/intel/uncore_nhmex.c atomic_sub(1 << ((idx - 2) * 8), &er->ref); idx 1091 arch/x86/events/intel/uncore_nhmex.c int idx; idx 1093 arch/x86/events/intel/uncore_nhmex.c idx = (event->hw.config & NHMEX_R_PMON_CTL_EV_SEL_MASK) >> idx 1095 arch/x86/events/intel/uncore_nhmex.c if (idx >= 0x18) idx 1098 arch/x86/events/intel/uncore_nhmex.c reg1->idx = idx; idx 1101 arch/x86/events/intel/uncore_nhmex.c switch (idx % 6) { idx 1116 arch/x86/events/intel/uncore_nhmex.c int idx, port; idx 1118 arch/x86/events/intel/uncore_nhmex.c idx = reg1->idx; idx 1119 arch/x86/events/intel/uncore_nhmex.c port = idx / 6 + box->pmu->pmu_idx * 4; idx 1121 arch/x86/events/intel/uncore_nhmex.c switch (idx % 6) { idx 1131 arch/x86/events/intel/uncore_nhmex.c uncore_shared_reg_config(box, 2 + (idx / 6) * 5)); idx 125 arch/x86/events/intel/uncore_snb.c if (hwc->idx < UNCORE_PMC_IDX_FIXED) idx 454 arch/x86/events/intel/uncore_snb.c int idx, base; idx 492 arch/x86/events/intel/uncore_snb.c event->hw.idx = -1; idx 494 arch/x86/events/intel/uncore_snb.c event->hw.extra_reg.idx = EXTRA_REG_NONE; idx 495 arch/x86/events/intel/uncore_snb.c event->hw.branch_reg.idx = EXTRA_REG_NONE; idx 502 arch/x86/events/intel/uncore_snb.c idx = UNCORE_PMC_IDX_FREERUNNING; idx 506 arch/x86/events/intel/uncore_snb.c idx = UNCORE_PMC_IDX_FREERUNNING; idx 514 arch/x86/events/intel/uncore_snb.c event->hw.idx = idx; idx 937 arch/x86/events/intel/uncore_snb.c if (hwc->idx < UNCORE_PMC_IDX_FIXED) idx 105 arch/x86/events/intel/uncore_snbep.c .idx = (i) \ idx 554 arch/x86/events/intel/uncore_snbep.c if (reg1->idx != EXTRA_REG_NONE) idx 867 arch/x86/events/intel/uncore_snbep.c if (reg1->idx == EXTRA_REG_NONE) idx 872 arch/x86/events/intel/uncore_snbep.c if (!(reg1->idx & (0x1 << i))) idx 930 arch/x86/events/intel/uncore_snbep.c int idx = 0; idx 935 arch/x86/events/intel/uncore_snbep.c idx |= er->idx; idx 938 arch/x86/events/intel/uncore_snbep.c if (idx) { idx 941 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & snbep_cbox_filter_mask(idx); idx 942 arch/x86/events/intel/uncore_snbep.c reg1->idx = idx; idx 976 arch/x86/events/intel/uncore_snbep.c if (new_idx > reg1->idx) idx 977 arch/x86/events/intel/uncore_snbep.c config <<= 8 * (new_idx - reg1->idx); idx 979 arch/x86/events/intel/uncore_snbep.c config >>= 8 * (reg1->idx - new_idx); idx 982 arch/x86/events/intel/uncore_snbep.c hwc->config += new_idx - reg1->idx; idx 984 arch/x86/events/intel/uncore_snbep.c reg1->idx = new_idx; idx 995 arch/x86/events/intel/uncore_snbep.c int idx = reg1->idx; idx 999 arch/x86/events/intel/uncore_snbep.c if (reg1->idx == EXTRA_REG_NONE || idx 1003 arch/x86/events/intel/uncore_snbep.c mask = 0xffULL << (idx * 8); idx 1005 arch/x86/events/intel/uncore_snbep.c if (!__BITS_VALUE(atomic_read(&er->ref), idx, 8) || idx 1007 arch/x86/events/intel/uncore_snbep.c atomic_add(1 << (idx * 8), &er->ref); idx 1015 arch/x86/events/intel/uncore_snbep.c idx = (idx + 1) % 4; idx 1016 arch/x86/events/intel/uncore_snbep.c if (idx != reg1->idx) { idx 1017 arch/x86/events/intel/uncore_snbep.c config1 = snbep_pcu_alter_er(event, idx, false); idx 1024 arch/x86/events/intel/uncore_snbep.c if (idx != reg1->idx) idx 1025 arch/x86/events/intel/uncore_snbep.c snbep_pcu_alter_er(event, idx, true); idx 1039 arch/x86/events/intel/uncore_snbep.c atomic_sub(1 << (reg1->idx * 8), &er->ref); idx 1051 arch/x86/events/intel/uncore_snbep.c reg1->idx = ev_sel - 0xb; idx 1052 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & (0xff << (reg1->idx * 8)); idx 1106 arch/x86/events/intel/uncore_snbep.c reg1->idx = 0; idx 1122 arch/x86/events/intel/uncore_snbep.c if (reg1->idx != EXTRA_REG_NONE) { idx 1123 arch/x86/events/intel/uncore_snbep.c int idx = box->pmu->pmu_idx + SNBEP_PCI_QPI_PORT0_FILTER; idx 1125 arch/x86/events/intel/uncore_snbep.c struct pci_dev *filter_pdev = uncore_extra_pci_dev[die].dev[idx]; idx 1614 arch/x86/events/intel/uncore_snbep.c int idx = 0; idx 1619 arch/x86/events/intel/uncore_snbep.c idx |= er->idx; idx 1622 arch/x86/events/intel/uncore_snbep.c if (idx) { idx 1625 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & ivbep_cbox_filter_mask(idx); idx 1626 arch/x86/events/intel/uncore_snbep.c reg1->idx = idx; idx 1636 arch/x86/events/intel/uncore_snbep.c if (reg1->idx != EXTRA_REG_NONE) { idx 1737 arch/x86/events/intel/uncore_snbep.c pci_write_config_dword(pdev, ivbep_uncore_irp_ctls[hwc->idx], idx 1746 arch/x86/events/intel/uncore_snbep.c pci_write_config_dword(pdev, ivbep_uncore_irp_ctls[hwc->idx], hwc->config); idx 1755 arch/x86/events/intel/uncore_snbep.c pci_read_config_dword(pdev, ivbep_uncore_irp_ctrs[hwc->idx], (u32 *)&count); idx 1756 arch/x86/events/intel/uncore_snbep.c pci_read_config_dword(pdev, ivbep_uncore_irp_ctrs[hwc->idx] + 4, (u32 *)&count + 1); idx 2042 arch/x86/events/intel/uncore_snbep.c int idx = 0; idx 2047 arch/x86/events/intel/uncore_snbep.c idx |= er->idx; idx 2050 arch/x86/events/intel/uncore_snbep.c if (idx) { idx 2053 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & knl_cha_filter_mask(idx); idx 2058 arch/x86/events/intel/uncore_snbep.c reg1->idx = idx; idx 2465 arch/x86/events/intel/uncore_snbep.c reg1->idx = 0; idx 2597 arch/x86/events/intel/uncore_snbep.c int idx = 0; idx 2602 arch/x86/events/intel/uncore_snbep.c idx |= er->idx; idx 2605 arch/x86/events/intel/uncore_snbep.c if (idx) { idx 2608 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & hswep_cbox_filter_mask(idx); idx 2609 arch/x86/events/intel/uncore_snbep.c reg1->idx = idx; idx 2620 arch/x86/events/intel/uncore_snbep.c if (reg1->idx != EXTRA_REG_NONE) { idx 2718 arch/x86/events/intel/uncore_snbep.c reg1->idx = ev_sel - 0xb; idx 2719 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & (0xff << reg1->idx); idx 2812 arch/x86/events/intel/uncore_snbep.c pci_read_config_dword(pdev, hswep_uncore_irp_ctrs[hwc->idx], (u32 *)&count); idx 2813 arch/x86/events/intel/uncore_snbep.c pci_read_config_dword(pdev, hswep_uncore_irp_ctrs[hwc->idx] + 4, (u32 *)&count + 1); idx 3490 arch/x86/events/intel/uncore_snbep.c int idx = 0; idx 3495 arch/x86/events/intel/uncore_snbep.c idx |= er->idx; idx 3498 arch/x86/events/intel/uncore_snbep.c if (idx) { idx 3501 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & skx_cha_filter_mask(idx); idx 3502 arch/x86/events/intel/uncore_snbep.c reg1->idx = idx; idx 4074 arch/x86/events/intel/uncore_snbep.c reg1->idx = 0; idx 4085 arch/x86/events/intel/uncore_snbep.c if (reg1->idx != EXTRA_REG_NONE) idx 4183 arch/x86/events/intel/uncore_snbep.c reg1->idx = ev_sel - 0xb; idx 4184 arch/x86/events/intel/uncore_snbep.c reg1->config = event->attr.config1 & (0xff << reg1->idx); idx 20 arch/x86/events/msr.c static bool test_aperfmperf(int idx, void *data) idx 25 arch/x86/events/msr.c static bool test_ptsc(int idx, void *data) idx 30 arch/x86/events/msr.c static bool test_irperf(int idx, void *data) idx 35 arch/x86/events/msr.c static bool test_therm_status(int idx, void *data) idx 40 arch/x86/events/msr.c static bool test_intel(int idx, void *data) idx 84 arch/x86/events/msr.c if (idx == PERF_MSR_SMI) idx 101 arch/x86/events/msr.c if (idx == PERF_MSR_SMI || idx == PERF_MSR_PPERF) idx 208 arch/x86/events/msr.c event->hw.idx = -1; idx 481 arch/x86/events/perf_event.h int idx; /* per_xxx->regs[] reg index */ idx 490 arch/x86/events/perf_event.h .idx = EXTRA_REG_##i, \ idx 494 arch/x86/events/perf_event.h #define INTEL_EVENT_EXTRA_REG(event, msr, vm, idx) \ idx 495 arch/x86/events/perf_event.h EVENT_EXTRA_REG(event, msr, ARCH_PERFMON_EVENTSEL_EVENT, vm, idx) idx 497 arch/x86/events/perf_event.h #define INTEL_UEVENT_EXTRA_REG(event, msr, vm, idx) \ idx 499 arch/x86/events/perf_event.h ARCH_PERFMON_EVENTSEL_UMASK, vm, idx) idx 601 arch/x86/events/perf_event.h int idx, idx 609 arch/x86/events/perf_event.h void (*commit_scheduling)(struct cpu_hw_events *cpuc, int idx, int cntr); idx 949 arch/x86/events/perf_event.h x86_get_event_constraints(struct cpu_hw_events *cpuc, int idx, idx 9 arch/x86/events/probe.h bool (*test)(int idx, void *data); idx 158 arch/x86/include/asm/fixmap.h void __native_set_fixmap(enum fixed_addresses idx, pte_t pte); idx 159 arch/x86/include/asm/fixmap.h void native_set_fixmap(unsigned /* enum fixed_addresses */ idx, idx 163 arch/x86/include/asm/fixmap.h static inline void __set_fixmap(enum fixed_addresses idx, idx 166 arch/x86/include/asm/fixmap.h native_set_fixmap(idx, phys, flags); idx 192 arch/x86/include/asm/fixmap.h #define __late_set_fixmap(idx, phys, flags) __set_fixmap(idx, phys, flags) idx 193 arch/x86/include/asm/fixmap.h #define __late_clear_fixmap(idx) __set_fixmap(idx, 0, __pgprot(0)) idx 195 arch/x86/include/asm/fixmap.h void __early_set_fixmap(enum fixed_addresses idx, idx 452 arch/x86/include/asm/kvm_host.h u8 idx; idx 630 arch/x86/include/asm/paravirt.h static inline void __set_fixmap(unsigned /* enum fixed_addresses */ idx, idx 633 arch/x86/include/asm/paravirt.h pv_ops.mmu.set_fixmap(idx, phys, flags); idx 130 arch/x86/include/asm/paravirt_types.h void (*load_gs_index)(unsigned int idx); idx 301 arch/x86/include/asm/paravirt_types.h void (*set_fixmap)(unsigned /* enum fixed_addresses */ idx, idx 359 arch/x86/include/asm/ptrace.h extern int do_get_thread_area(struct task_struct *p, int idx, idx 361 arch/x86/include/asm/ptrace.h extern int do_set_thread_area(struct task_struct *p, int idx, idx 1112 arch/x86/kernel/acpi/boot.c int idx; idx 1128 arch/x86/kernel/acpi/boot.c for (idx = 0; idx < mp_irq_entries; idx++) { idx 1129 arch/x86/kernel/acpi/boot.c struct mpc_intsrc *irq = mp_irqs + idx; idx 1140 arch/x86/kernel/acpi/boot.c if (idx != mp_irq_entries) { idx 69 arch/x86/kernel/apic/io_apic.c #define for_each_ioapic(idx) \ idx 70 arch/x86/kernel/apic/io_apic.c for ((idx) = 0; (idx) < nr_ioapics; (idx)++) idx 71 arch/x86/kernel/apic/io_apic.c #define for_each_ioapic_reverse(idx) \ idx 72 arch/x86/kernel/apic/io_apic.c for ((idx) = nr_ioapics - 1; (idx) >= 0; (idx)--) idx 73 arch/x86/kernel/apic/io_apic.c #define for_each_pin(idx, pin) \ idx 74 arch/x86/kernel/apic/io_apic.c for ((pin) = 0; (pin) < ioapics[(idx)].nr_registers; (pin)++) idx 75 arch/x86/kernel/apic/io_apic.c #define for_each_ioapic_pin(idx, pin) \ idx 76 arch/x86/kernel/apic/io_apic.c for_each_ioapic((idx)) \ idx 77 arch/x86/kernel/apic/io_apic.c for_each_pin((idx), (pin)) idx 234 arch/x86/kernel/apic/io_apic.c static void alloc_ioapic_saved_registers(int idx) idx 238 arch/x86/kernel/apic/io_apic.c if (ioapics[idx].saved_registers) idx 241 arch/x86/kernel/apic/io_apic.c size = sizeof(struct IO_APIC_route_entry) * ioapics[idx].nr_registers; idx 242 arch/x86/kernel/apic/io_apic.c ioapics[idx].saved_registers = kzalloc(size, GFP_KERNEL); idx 243 arch/x86/kernel/apic/io_apic.c if (!ioapics[idx].saved_registers) idx 244 arch/x86/kernel/apic/io_apic.c pr_err("IOAPIC %d: suspend/resume impossible!\n", idx); idx 247 arch/x86/kernel/apic/io_apic.c static void free_ioapic_saved_registers(int idx) idx 249 arch/x86/kernel/apic/io_apic.c kfree(ioapics[idx].saved_registers); idx 250 arch/x86/kernel/apic/io_apic.c ioapics[idx].saved_registers = NULL; idx 274 arch/x86/kernel/apic/io_apic.c static __attribute_const__ struct io_apic __iomem *io_apic_base(int idx) idx 276 arch/x86/kernel/apic/io_apic.c return (void __iomem *) __fix_to_virt(FIX_IO_APIC_BASE_0 + idx) idx 277 arch/x86/kernel/apic/io_apic.c + (mpc_ioapic_addr(idx) & ~PAGE_MASK); idx 781 arch/x86/kernel/apic/io_apic.c #define default_ISA_trigger(idx) (IOAPIC_EDGE) idx 782 arch/x86/kernel/apic/io_apic.c #define default_ISA_polarity(idx) (IOAPIC_POL_HIGH) idx 789 arch/x86/kernel/apic/io_apic.c #define default_EISA_trigger(idx) (EISA_ELCR(mp_irqs[idx].srcbusirq)) idx 790 arch/x86/kernel/apic/io_apic.c #define default_EISA_polarity(idx) default_ISA_polarity(idx) idx 795 arch/x86/kernel/apic/io_apic.c #define default_PCI_trigger(idx) (IOAPIC_LEVEL) idx 796 arch/x86/kernel/apic/io_apic.c #define default_PCI_polarity(idx) (IOAPIC_POL_LOW) idx 798 arch/x86/kernel/apic/io_apic.c static int irq_polarity(int idx) idx 800 arch/x86/kernel/apic/io_apic.c int bus = mp_irqs[idx].srcbus; idx 805 arch/x86/kernel/apic/io_apic.c switch (mp_irqs[idx].irqflag & MP_IRQPOL_MASK) { idx 809 arch/x86/kernel/apic/io_apic.c return default_ISA_polarity(idx); idx 811 arch/x86/kernel/apic/io_apic.c return default_PCI_polarity(idx); idx 824 arch/x86/kernel/apic/io_apic.c static int eisa_irq_trigger(int idx, int bus, int trigger) idx 831 arch/x86/kernel/apic/io_apic.c return default_EISA_trigger(idx); idx 837 arch/x86/kernel/apic/io_apic.c static inline int eisa_irq_trigger(int idx, int bus, int trigger) idx 843 arch/x86/kernel/apic/io_apic.c static int irq_trigger(int idx) idx 845 arch/x86/kernel/apic/io_apic.c int bus = mp_irqs[idx].srcbus; idx 851 arch/x86/kernel/apic/io_apic.c switch (mp_irqs[idx].irqflag & MP_IRQTRIG_MASK) { idx 855 arch/x86/kernel/apic/io_apic.c trigger = default_ISA_trigger(idx); idx 857 arch/x86/kernel/apic/io_apic.c trigger = default_PCI_trigger(idx); idx 859 arch/x86/kernel/apic/io_apic.c return eisa_irq_trigger(idx, bus, trigger); idx 1034 arch/x86/kernel/apic/io_apic.c static int mp_map_pin_to_irq(u32 gsi, int idx, int ioapic, int pin, idx 1046 arch/x86/kernel/apic/io_apic.c if (idx >= 0 && test_bit(mp_irqs[idx].srcbus, mp_bus_not_pci)) { idx 1047 arch/x86/kernel/apic/io_apic.c irq = mp_irqs[idx].srcbusirq; idx 1077 arch/x86/kernel/apic/io_apic.c static int pin_2_irq(int idx, int ioapic, int pin, unsigned int flags) idx 1084 arch/x86/kernel/apic/io_apic.c if (mp_irqs[idx].dstirq != pin) idx 1107 arch/x86/kernel/apic/io_apic.c return mp_map_pin_to_irq(gsi, idx, ioapic, pin, flags, NULL); idx 1112 arch/x86/kernel/apic/io_apic.c int ioapic, pin, idx; idx 1119 arch/x86/kernel/apic/io_apic.c idx = find_irq_entry(ioapic, pin, mp_INT); idx 1120 arch/x86/kernel/apic/io_apic.c if ((flags & IOAPIC_MAP_CHECK) && idx < 0) idx 1123 arch/x86/kernel/apic/io_apic.c return mp_map_pin_to_irq(gsi, idx, ioapic, pin, flags, info); idx 1212 arch/x86/kernel/apic/io_apic.c int idx; idx 1217 arch/x86/kernel/apic/io_apic.c idx = find_irq_entry(ioapic, pin, mp_INT); idx 1218 arch/x86/kernel/apic/io_apic.c if (idx < 0) idx 1223 arch/x86/kernel/apic/io_apic.c pin_2_irq(idx, ioapic, pin, idx 2194 arch/x86/kernel/apic/io_apic.c int idx; idx 2195 arch/x86/kernel/apic/io_apic.c idx = find_irq_entry(apic1, pin1, mp_INT); idx 2196 arch/x86/kernel/apic/io_apic.c if (idx != -1 && irq_trigger(idx)) idx 2349 arch/x86/kernel/apic/io_apic.c static void ioapic_destroy_irqdomain(int idx) idx 2351 arch/x86/kernel/apic/io_apic.c if (ioapics[idx].irqdomain) { idx 2352 arch/x86/kernel/apic/io_apic.c irq_domain_remove(ioapics[idx].irqdomain); idx 2353 arch/x86/kernel/apic/io_apic.c ioapics[idx].irqdomain = NULL; idx 2529 arch/x86/kernel/apic/io_apic.c static u8 io_apic_unique_id(int idx, u8 id) idx 2533 arch/x86/kernel/apic/io_apic.c return io_apic_get_unique_id(idx, id); idx 2538 arch/x86/kernel/apic/io_apic.c static u8 io_apic_unique_id(int idx, u8 id) idx 2559 arch/x86/kernel/apic/io_apic.c reg_00.raw = io_apic_read(idx, 0); idx 2565 arch/x86/kernel/apic/io_apic.c idx, new_id, id); idx 2575 arch/x86/kernel/apic/io_apic.c io_apic_write(idx, 0, reg_00.raw); idx 2576 arch/x86/kernel/apic/io_apic.c reg_00.raw = io_apic_read(idx, 0); idx 2599 arch/x86/kernel/apic/io_apic.c int ioapic, pin, idx; idx 2612 arch/x86/kernel/apic/io_apic.c idx = find_irq_entry(ioapic, pin, mp_INT); idx 2613 arch/x86/kernel/apic/io_apic.c if (idx < 0) idx 2616 arch/x86/kernel/apic/io_apic.c *trigger = irq_trigger(idx); idx 2617 arch/x86/kernel/apic/io_apic.c *polarity = irq_polarity(idx); idx 2664 arch/x86/kernel/apic/io_apic.c unsigned long ioapic_phys, idx = FIX_IO_APIC_BASE_0; idx 2694 arch/x86/kernel/apic/io_apic.c set_fixmap_nocache(idx, ioapic_phys); idx 2696 arch/x86/kernel/apic/io_apic.c __fix_to_virt(idx) + (ioapic_phys & ~PAGE_MASK), idx 2698 arch/x86/kernel/apic/io_apic.c idx++; idx 2756 arch/x86/kernel/apic/io_apic.c static int bad_ioapic_register(int idx) idx 2762 arch/x86/kernel/apic/io_apic.c reg_00.raw = io_apic_read(idx, 0); idx 2763 arch/x86/kernel/apic/io_apic.c reg_01.raw = io_apic_read(idx, 1); idx 2764 arch/x86/kernel/apic/io_apic.c reg_02.raw = io_apic_read(idx, 2); idx 2768 arch/x86/kernel/apic/io_apic.c mpc_ioapic_addr(idx)); idx 2777 arch/x86/kernel/apic/io_apic.c int idx; idx 2779 arch/x86/kernel/apic/io_apic.c for (idx = 0; idx < MAX_IO_APICS; idx++) idx 2780 arch/x86/kernel/apic/io_apic.c if (ioapics[idx].nr_registers == 0) idx 2781 arch/x86/kernel/apic/io_apic.c return idx; idx 2798 arch/x86/kernel/apic/io_apic.c int idx, ioapic, entries; idx 2812 arch/x86/kernel/apic/io_apic.c idx = find_free_ioapic_entry(); idx 2813 arch/x86/kernel/apic/io_apic.c if (idx >= MAX_IO_APICS) { idx 2815 arch/x86/kernel/apic/io_apic.c MAX_IO_APICS, idx); idx 2819 arch/x86/kernel/apic/io_apic.c ioapics[idx].mp_config.type = MP_IOAPIC; idx 2820 arch/x86/kernel/apic/io_apic.c ioapics[idx].mp_config.flags = MPC_APIC_USABLE; idx 2821 arch/x86/kernel/apic/io_apic.c ioapics[idx].mp_config.apicaddr = address; idx 2823 arch/x86/kernel/apic/io_apic.c set_fixmap_nocache(FIX_IO_APIC_BASE_0 + idx, address); idx 2824 arch/x86/kernel/apic/io_apic.c if (bad_ioapic_register(idx)) { idx 2825 arch/x86/kernel/apic/io_apic.c clear_fixmap(FIX_IO_APIC_BASE_0 + idx); idx 2829 arch/x86/kernel/apic/io_apic.c ioapics[idx].mp_config.apicid = io_apic_unique_id(idx, id); idx 2830 arch/x86/kernel/apic/io_apic.c ioapics[idx].mp_config.apicver = io_apic_get_version(idx); idx 2836 arch/x86/kernel/apic/io_apic.c entries = io_apic_get_redir_entries(idx); idx 2847 arch/x86/kernel/apic/io_apic.c clear_fixmap(FIX_IO_APIC_BASE_0 + idx); idx 2851 arch/x86/kernel/apic/io_apic.c gsi_cfg = mp_ioapic_gsi_routing(idx); idx 2855 arch/x86/kernel/apic/io_apic.c ioapics[idx].irqdomain = NULL; idx 2856 arch/x86/kernel/apic/io_apic.c ioapics[idx].irqdomain_cfg = *cfg; idx 2864 arch/x86/kernel/apic/io_apic.c if (mp_irqdomain_create(idx)) { idx 2865 arch/x86/kernel/apic/io_apic.c clear_fixmap(FIX_IO_APIC_BASE_0 + idx); idx 2868 arch/x86/kernel/apic/io_apic.c alloc_ioapic_saved_registers(idx); idx 2873 arch/x86/kernel/apic/io_apic.c if (nr_ioapics <= idx) idx 2874 arch/x86/kernel/apic/io_apic.c nr_ioapics = idx + 1; idx 2877 arch/x86/kernel/apic/io_apic.c ioapics[idx].nr_registers = entries; idx 2880 arch/x86/kernel/apic/io_apic.c idx, mpc_ioapic_id(idx), idx 2881 arch/x86/kernel/apic/io_apic.c mpc_ioapic_ver(idx), mpc_ioapic_addr(idx), idx 374 arch/x86/kernel/cpu/cacheinfo.c unsigned slot, unsigned long idx) idx 378 arch/x86/kernel/cpu/cacheinfo.c idx |= BIT(30); idx 384 arch/x86/kernel/cpu/cacheinfo.c u32 reg = idx | (i << 20); idx 1019 arch/x86/kernel/cpu/cacheinfo.c unsigned int idx, ret; idx 1024 arch/x86/kernel/cpu/cacheinfo.c for (idx = 0; idx < this_cpu_ci->num_leaves; idx++) { idx 1025 arch/x86/kernel/cpu/cacheinfo.c ret = cpuid4_cache_lookup_regs(idx, &id4_regs); idx 1030 arch/x86/kernel/cpu/cacheinfo.c __cache_cpumap_setup(cpu, idx, &id4_regs); idx 122 arch/x86/kernel/cpu/mtrr/generic.c int idx; idx 129 arch/x86/kernel/cpu/mtrr/generic.c idx = 0; idx 130 arch/x86/kernel/cpu/mtrr/generic.c idx += (start >> 16); idx 131 arch/x86/kernel/cpu/mtrr/generic.c return mtrr_state.fixed_ranges[idx]; idx 134 arch/x86/kernel/cpu/mtrr/generic.c idx = 1 * 8; idx 135 arch/x86/kernel/cpu/mtrr/generic.c idx += ((start - 0x80000) >> 14); idx 136 arch/x86/kernel/cpu/mtrr/generic.c return mtrr_state.fixed_ranges[idx]; idx 140 arch/x86/kernel/cpu/mtrr/generic.c idx = 3 * 8; idx 141 arch/x86/kernel/cpu/mtrr/generic.c idx += ((start - 0xC0000) >> 12); idx 142 arch/x86/kernel/cpu/mtrr/generic.c return mtrr_state.fixed_ranges[idx]; idx 303 arch/x86/kernel/cpu/resctrl/core.c static void rdt_get_cache_alloc_cfg(int idx, struct rdt_resource *r) idx 309 arch/x86/kernel/cpu/resctrl/core.c cpuid_count(0x00000010, idx, &eax.full, &ebx, &ecx, &edx.full); idx 754 arch/x86/kernel/cpu/resctrl/core.c #define RDT_OPT(idx, n, f) \ idx 755 arch/x86/kernel/cpu/resctrl/core.c [idx] = { \ idx 215 arch/x86/kernel/head64.c int idx = i + (physaddr >> PMD_SHIFT); idx 217 arch/x86/kernel/head64.c pmd[idx % PTRS_PER_PMD] = pmd_entry + i * PMD_SIZE; idx 487 arch/x86/kernel/pci-calgary_64.c size_t idx = busno_to_phbid(num); idx 489 arch/x86/kernel/pci-calgary_64.c return split_queue_offsets[idx]; idx 494 arch/x86/kernel/pci-calgary_64.c size_t idx = busno_to_phbid(num); idx 496 arch/x86/kernel/pci-calgary_64.c return tar_offsets[idx]; idx 501 arch/x86/kernel/pci-calgary_64.c size_t idx = busno_to_phbid(num); idx 503 arch/x86/kernel/pci-calgary_64.c return phb_offsets[idx]; idx 60 arch/x86/kernel/perf_regs.c u64 perf_reg_value(struct pt_regs *regs, int idx) idx 64 arch/x86/kernel/perf_regs.c if (idx >= PERF_REG_X86_XMM0 && idx < PERF_REG_X86_XMM_MAX) { idx 68 arch/x86/kernel/perf_regs.c return perf_regs->xmm_regs[idx - PERF_REG_X86_XMM0]; idx 71 arch/x86/kernel/perf_regs.c if (WARN_ON_ONCE(idx >= ARRAY_SIZE(pt_regs_offset))) idx 74 arch/x86/kernel/perf_regs.c return regs_get_register(regs, pt_regs_offset[idx]); idx 292 arch/x86/kernel/process_64.c unsigned short idx = selector >> 3; idx 296 arch/x86/kernel/process_64.c if (unlikely(idx >= GDT_ENTRIES)) idx 303 arch/x86/kernel/process_64.c if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) idx 306 arch/x86/kernel/process_64.c idx -= GDT_ENTRY_TLS_MIN; idx 307 arch/x86/kernel/process_64.c base = get_desc_base(&task->thread.tls_array[idx]); idx 319 arch/x86/kernel/process_64.c if (unlikely(idx >= ldt->nr_entries)) idx 322 arch/x86/kernel/process_64.c base = get_desc_base(ldt->entries + idx); idx 24 arch/x86/kernel/tls.c int idx; idx 26 arch/x86/kernel/tls.c for (idx = 0; idx < GDT_ENTRY_TLS_ENTRIES; idx++) idx 27 arch/x86/kernel/tls.c if (desc_empty(&t->tls_array[idx])) idx 28 arch/x86/kernel/tls.c return idx + GDT_ENTRY_TLS_MIN; idx 84 arch/x86/kernel/tls.c static void set_tls_desc(struct task_struct *p, int idx, idx 88 arch/x86/kernel/tls.c struct desc_struct *desc = &t->tls_array[idx - GDT_ENTRY_TLS_MIN]; idx 114 arch/x86/kernel/tls.c int do_set_thread_area(struct task_struct *p, int idx, idx 127 arch/x86/kernel/tls.c if (idx == -1) idx 128 arch/x86/kernel/tls.c idx = info.entry_number; idx 134 arch/x86/kernel/tls.c if (idx == -1 && can_allocate) { idx 135 arch/x86/kernel/tls.c idx = get_free_idx(); idx 136 arch/x86/kernel/tls.c if (idx < 0) idx 137 arch/x86/kernel/tls.c return idx; idx 138 arch/x86/kernel/tls.c if (put_user(idx, &u_info->entry_number)) idx 142 arch/x86/kernel/tls.c if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) idx 145 arch/x86/kernel/tls.c set_tls_desc(p, idx, &info, 1); idx 152 arch/x86/kernel/tls.c modified_sel = (idx << 3) | 3; idx 201 arch/x86/kernel/tls.c static void fill_user_desc(struct user_desc *info, int idx, idx 206 arch/x86/kernel/tls.c info->entry_number = idx; idx 220 arch/x86/kernel/tls.c int do_get_thread_area(struct task_struct *p, int idx, idx 226 arch/x86/kernel/tls.c if (idx == -1 && get_user(idx, &u_info->entry_number)) idx 229 arch/x86/kernel/tls.c if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) idx 232 arch/x86/kernel/tls.c index = idx - GDT_ENTRY_TLS_MIN; idx 236 arch/x86/kernel/tls.c fill_user_desc(&info, idx, &p->thread.tls_array[index]); idx 64 arch/x86/kernel/tsc.c int seq, idx; idx 70 arch/x86/kernel/tsc.c idx = seq & 1; idx 72 arch/x86/kernel/tsc.c data->cyc2ns_offset = this_cpu_read(cyc2ns.data[idx].cyc2ns_offset); idx 73 arch/x86/kernel/tsc.c data->cyc2ns_mul = this_cpu_read(cyc2ns.data[idx].cyc2ns_mul); idx 74 arch/x86/kernel/tsc.c data->cyc2ns_shift = this_cpu_read(cyc2ns.data[idx].cyc2ns_shift); idx 150 arch/x86/kernel/unwind_orc.c unsigned int idx, start, stop; idx 152 arch/x86/kernel/unwind_orc.c idx = (ip - LOOKUP_START_IP) / LOOKUP_BLOCK_SIZE; idx 154 arch/x86/kernel/unwind_orc.c if (unlikely((idx >= lookup_num_blocks-1))) { idx 156 arch/x86/kernel/unwind_orc.c idx, lookup_num_blocks, (void *)ip); idx 160 arch/x86/kernel/unwind_orc.c start = orc_lookup[idx]; idx 161 arch/x86/kernel/unwind_orc.c stop = orc_lookup[idx + 1] + 1; idx 166 arch/x86/kernel/unwind_orc.c idx, lookup_num_blocks, start, stop, (void *)ip); idx 645 arch/x86/kvm/cpuid.c int idx, i; idx 655 arch/x86/kvm/cpuid.c for (idx = 1, i = 1; idx < 64; ++idx) { idx 656 arch/x86/kvm/cpuid.c u64 mask = ((u64)1 << idx); idx 660 arch/x86/kvm/cpuid.c do_host_cpuid(&entry[i], function, idx); idx 661 arch/x86/kvm/cpuid.c if (idx == 1) { idx 168 arch/x86/kvm/hyperv.c int gsi, idx; idx 173 arch/x86/kvm/hyperv.c for (idx = 0; idx < ARRAY_SIZE(hv_vcpu->stimer); idx++) { idx 174 arch/x86/kvm/hyperv.c stimer = &hv_vcpu->stimer[idx]; idx 181 arch/x86/kvm/hyperv.c idx = srcu_read_lock(&kvm->irq_srcu); idx 185 arch/x86/kvm/hyperv.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 226 arch/x86/kvm/ioapic.c u32 idx; idx 229 arch/x86/kvm/ioapic.c for_each_set_bit(idx, &irr, IOAPIC_NUM_PINS) idx 230 arch/x86/kvm/ioapic.c ioapic_set_irq(ioapic, idx, 1, true); idx 92 arch/x86/kvm/irq_comm.c int idx = kvm_vector_to_index(irq->vector, dest_vcpus, idx 95 arch/x86/kvm/irq_comm.c lowest = kvm_get_vcpu(kvm, idx); idx 255 arch/x86/kvm/irq_comm.c int idx, gsi; idx 257 arch/x86/kvm/irq_comm.c idx = srcu_read_lock(&kvm->irq_srcu); idx 263 arch/x86/kvm/irq_comm.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 404 arch/x86/kvm/irq_comm.c int idx; idx 406 arch/x86/kvm/irq_comm.c idx = srcu_read_lock(&kvm->irq_srcu); idx 424 arch/x86/kvm/irq_comm.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 832 arch/x86/kvm/lapic.c int i, idx = -1; idx 837 arch/x86/kvm/lapic.c idx = find_next_bit(bitmap, bitmap_size, idx + 1); idx 838 arch/x86/kvm/lapic.c BUG_ON(idx == bitmap_size); idx 841 arch/x86/kvm/lapic.c return idx; idx 1185 arch/x86/kvm/mmu.c unsigned long idx; idx 1187 arch/x86/kvm/mmu.c idx = gfn_to_index(gfn, slot->base_gfn, level); idx 1188 arch/x86/kvm/mmu.c return &slot->arch.lpage_info[level - 2][idx]; idx 1472 arch/x86/kvm/mmu.c unsigned long idx; idx 1474 arch/x86/kvm/mmu.c idx = gfn_to_index(gfn, slot->base_gfn, level); idx 1475 arch/x86/kvm/mmu.c return &slot->arch.rmap[level - PT_PAGE_TABLE_LEVEL][idx]; idx 2246 arch/x86/kvm/mmu.c unsigned int idx; idx 2252 arch/x86/kvm/mmu.c int idx) idx 2262 arch/x86/kvm/mmu.c pvec->page[pvec->nr].idx = idx; idx 2267 arch/x86/kvm/mmu.c static inline void clear_unsync_child_bit(struct kvm_mmu_page *sp, int idx) idx 2271 arch/x86/kvm/mmu.c __clear_bit(idx, sp->unsync_child_bitmap); idx 2433 arch/x86/kvm/mmu.c unsigned int idx[PT64_ROOT_MAX_LEVEL]; idx 2449 arch/x86/kvm/mmu.c unsigned idx = pvec->page[n].idx; idx 2452 arch/x86/kvm/mmu.c parents->idx[level-1] = idx; idx 2471 arch/x86/kvm/mmu.c WARN_ON(pvec->page[0].idx != INVALID_INDEX); idx 2492 arch/x86/kvm/mmu.c unsigned int idx = parents->idx[level]; idx 2497 arch/x86/kvm/mmu.c WARN_ON(idx == INVALID_INDEX); idx 2498 arch/x86/kvm/mmu.c clear_unsync_child_bit(sp, idx); idx 6177 arch/x86/kvm/mmu.c int idx; idx 6198 arch/x86/kvm/mmu.c idx = srcu_read_lock(&kvm->srcu); idx 6213 arch/x86/kvm/mmu.c srcu_read_unlock(&kvm->srcu, idx); idx 228 arch/x86/kvm/page_track.c int idx; idx 235 arch/x86/kvm/page_track.c idx = srcu_read_lock(&head->track_srcu); idx 239 arch/x86/kvm/page_track.c srcu_read_unlock(&head->track_srcu, idx); idx 253 arch/x86/kvm/page_track.c int idx; idx 260 arch/x86/kvm/page_track.c idx = srcu_read_lock(&head->track_srcu); idx 264 arch/x86/kvm/page_track.c srcu_read_unlock(&head->track_srcu, idx); idx 65 arch/x86/kvm/pmu.c if (!test_and_set_bit(pmc->idx, idx 67 arch/x86/kvm/pmu.c __set_bit(pmc->idx, (unsigned long *)&pmu->global_status); idx 79 arch/x86/kvm/pmu.c if (!test_and_set_bit(pmc->idx, idx 81 arch/x86/kvm/pmu.c __set_bit(pmc->idx, (unsigned long *)&pmu->global_status); idx 135 arch/x86/kvm/pmu.c PTR_ERR(event), pmc->idx); idx 140 arch/x86/kvm/pmu.c clear_bit(pmc->idx, (unsigned long*)&pmc_to_pmu(pmc)->reprogram_pmi); idx 205 arch/x86/kvm/pmu.c void reprogram_fixed_counter(struct kvm_pmc *pmc, u8 ctrl, int idx) idx 220 arch/x86/kvm/pmu.c test_bit(idx, (ulong *)&filter->fixed_counter_bitmap)) idx 223 arch/x86/kvm/pmu.c !test_bit(idx, (ulong *)&filter->fixed_counter_bitmap)) idx 228 arch/x86/kvm/pmu.c kvm_x86_ops->pmu_ops->find_fixed_event(idx), idx 245 arch/x86/kvm/pmu.c int idx = pmc_idx - INTEL_PMC_IDX_FIXED; idx 246 arch/x86/kvm/pmu.c u8 ctrl = fixed_ctrl_field(pmu->fixed_ctr_ctrl, idx); idx 248 arch/x86/kvm/pmu.c reprogram_fixed_counter(pmc, ctrl, idx); idx 274 arch/x86/kvm/pmu.c int kvm_pmu_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx) idx 276 arch/x86/kvm/pmu.c return kvm_x86_ops->pmu_ops->is_valid_msr_idx(vcpu, idx); idx 290 arch/x86/kvm/pmu.c static int kvm_pmu_rdpmc_vmware(struct kvm_vcpu *vcpu, unsigned idx, u64 *data) idx 294 arch/x86/kvm/pmu.c switch (idx) { idx 313 arch/x86/kvm/pmu.c int kvm_pmu_rdpmc(struct kvm_vcpu *vcpu, unsigned idx, u64 *data) idx 315 arch/x86/kvm/pmu.c bool fast_mode = idx & (1u << 31); idx 323 arch/x86/kvm/pmu.c if (is_vmware_backdoor_pmc(idx)) idx 324 arch/x86/kvm/pmu.c return kvm_pmu_rdpmc_vmware(vcpu, idx, data); idx 326 arch/x86/kvm/pmu.c pmc = kvm_x86_ops->pmu_ops->msr_idx_to_pmc(vcpu, idx, &mask); idx 12 arch/x86/kvm/pmu.h #define fixed_ctrl_field(ctrl_reg, idx) (((ctrl_reg) >> ((idx)*4)) & 0xf) idx 27 arch/x86/kvm/pmu.h unsigned (*find_fixed_event)(int idx); idx 30 arch/x86/kvm/pmu.h struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, unsigned idx, idx 32 arch/x86/kvm/pmu.h int (*is_valid_msr_idx)(struct kvm_vcpu *vcpu, unsigned idx); idx 123 arch/x86/kvm/pmu.h int kvm_pmu_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx); idx 147 arch/x86/kvm/pmu_amd.c static unsigned amd_find_fixed_event(int idx) idx 177 arch/x86/kvm/pmu_amd.c static int amd_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx) idx 181 arch/x86/kvm/pmu_amd.c idx &= ~(3u << 30); idx 183 arch/x86/kvm/pmu_amd.c return (idx >= pmu->nr_arch_gp_counters); idx 187 arch/x86/kvm/pmu_amd.c static struct kvm_pmc *amd_msr_idx_to_pmc(struct kvm_vcpu *vcpu, unsigned idx, u64 *mask) idx 192 arch/x86/kvm/pmu_amd.c idx &= ~(3u << 30); idx 193 arch/x86/kvm/pmu_amd.c if (idx >= pmu->nr_arch_gp_counters) idx 197 arch/x86/kvm/pmu_amd.c return &counters[idx]; idx 287 arch/x86/kvm/pmu_amd.c pmu->gp_counters[i].idx = i; idx 5325 arch/x86/kvm/svm.c int idx, ret = -EINVAL; idx 5334 arch/x86/kvm/svm.c idx = srcu_read_lock(&kvm->irq_srcu); idx 5416 arch/x86/kvm/svm.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 6495 arch/x86/kvm/svm.c static unsigned long get_num_contig_pages(unsigned long idx, idx 6499 arch/x86/kvm/svm.c unsigned long i = idx + 1, pages = 1; idx 6502 arch/x86/kvm/svm.c paddr = __sme_page_pa(inpages[idx]); idx 85 arch/x86/kvm/vmx/pmu_intel.c static unsigned intel_find_fixed_event(int idx) idx 90 arch/x86/kvm/vmx/pmu_intel.c if (idx >= size) idx 93 arch/x86/kvm/vmx/pmu_intel.c event = fixed_pmc_events[array_index_nospec(idx, size)]; idx 102 arch/x86/kvm/vmx/pmu_intel.c return test_bit(pmc->idx, (unsigned long *)&pmu->global_ctrl); idx 111 arch/x86/kvm/vmx/pmu_intel.c u32 idx = pmc_idx - INTEL_PMC_IDX_FIXED; idx 113 arch/x86/kvm/vmx/pmu_intel.c return get_fixed_pmc(pmu, idx + MSR_CORE_PERF_FIXED_CTR0); idx 118 arch/x86/kvm/vmx/pmu_intel.c static int intel_is_valid_msr_idx(struct kvm_vcpu *vcpu, unsigned idx) idx 121 arch/x86/kvm/vmx/pmu_intel.c bool fixed = idx & (1u << 30); idx 123 arch/x86/kvm/vmx/pmu_intel.c idx &= ~(3u << 30); idx 125 arch/x86/kvm/vmx/pmu_intel.c return (!fixed && idx >= pmu->nr_arch_gp_counters) || idx 126 arch/x86/kvm/vmx/pmu_intel.c (fixed && idx >= pmu->nr_arch_fixed_counters); idx 130 arch/x86/kvm/vmx/pmu_intel.c unsigned idx, u64 *mask) idx 133 arch/x86/kvm/vmx/pmu_intel.c bool fixed = idx & (1u << 30); idx 137 arch/x86/kvm/vmx/pmu_intel.c idx &= ~(3u << 30); idx 145 arch/x86/kvm/vmx/pmu_intel.c if (idx >= num_counters) idx 148 arch/x86/kvm/vmx/pmu_intel.c return &counters[array_index_nospec(idx, num_counters)]; idx 338 arch/x86/kvm/vmx/pmu_intel.c pmu->gp_counters[i].idx = i; idx 344 arch/x86/kvm/vmx/pmu_intel.c pmu->fixed_counters[i].idx = i + INTEL_PMC_IDX_FIXED; idx 3430 arch/x86/kvm/vmx/vmx.c int idx, r; idx 3432 arch/x86/kvm/vmx/vmx.c idx = srcu_read_lock(&kvm->srcu); idx 3453 arch/x86/kvm/vmx/vmx.c srcu_read_unlock(&kvm->srcu, idx); idx 3460 arch/x86/kvm/vmx/vmx.c int i, idx, r = 0; idx 3479 arch/x86/kvm/vmx/vmx.c idx = srcu_read_lock(&kvm->srcu); idx 3495 arch/x86/kvm/vmx/vmx.c srcu_read_unlock(&kvm->srcu, idx); idx 7466 arch/x86/kvm/vmx/vmx.c int idx, ret = 0; idx 7473 arch/x86/kvm/vmx/vmx.c idx = srcu_read_lock(&kvm->irq_srcu); idx 7539 arch/x86/kvm/vmx/vmx.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 3568 arch/x86/kvm/x86.c int idx; idx 3586 arch/x86/kvm/x86.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 3588 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 4278 arch/x86/kvm/x86.c int idx = srcu_read_lock(&vcpu->kvm->srcu); idx 4280 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 4284 arch/x86/kvm/x86.c int idx = srcu_read_lock(&vcpu->kvm->srcu); idx 4286 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 4306 arch/x86/kvm/x86.c int idx; idx 4314 arch/x86/kvm/x86.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 4316 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 4496 arch/x86/kvm/x86.c int idx; idx 4520 arch/x86/kvm/x86.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 4522 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 8869 arch/x86/kvm/x86.c int pending_vec, max_bits, idx; idx 8909 arch/x86/kvm/x86.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 8914 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 9028 arch/x86/kvm/x86.c int idx; idx 9032 arch/x86/kvm/x86.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 9034 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 9523 arch/x86/kvm/x86.c int idx; idx 9529 arch/x86/kvm/x86.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 9531 arch/x86/kvm/x86.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 71 arch/x86/lib/insn-eval.c int idx = INAT_SEG_REG_DEFAULT; idx 83 arch/x86/lib/insn-eval.c idx = INAT_SEG_REG_CS; idx 87 arch/x86/lib/insn-eval.c idx = INAT_SEG_REG_SS; idx 91 arch/x86/lib/insn-eval.c idx = INAT_SEG_REG_DS; idx 95 arch/x86/lib/insn-eval.c idx = INAT_SEG_REG_ES; idx 99 arch/x86/lib/insn-eval.c idx = INAT_SEG_REG_FS; idx 103 arch/x86/lib/insn-eval.c idx = INAT_SEG_REG_GS; idx 114 arch/x86/lib/insn-eval.c return idx; idx 260 arch/x86/lib/insn-eval.c int idx; idx 281 arch/x86/lib/insn-eval.c idx = get_seg_reg_override_idx(insn); idx 282 arch/x86/lib/insn-eval.c if (idx < 0) idx 283 arch/x86/lib/insn-eval.c return idx; idx 285 arch/x86/lib/insn-eval.c if (idx == INAT_SEG_REG_DEFAULT) idx 293 arch/x86/lib/insn-eval.c if (idx != INAT_SEG_REG_FS && idx 294 arch/x86/lib/insn-eval.c idx != INAT_SEG_REG_GS) idx 295 arch/x86/lib/insn-eval.c idx = INAT_SEG_REG_IGNORE; idx 298 arch/x86/lib/insn-eval.c return idx; idx 504 arch/x86/mm/dump_pagetables.c static inline bool is_hypervisor_range(int idx) idx 511 arch/x86/mm/dump_pagetables.c return (idx >= pgd_index(GUARD_HOLE_BASE_ADDR)) && idx 512 arch/x86/mm/dump_pagetables.c (idx < pgd_index(GUARD_HOLE_END_ADDR)); idx 37 arch/x86/mm/highmem_32.c int idx, type; idx 46 arch/x86/mm/highmem_32.c idx = type + KM_TYPE_NR*smp_processor_id(); idx 47 arch/x86/mm/highmem_32.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 48 arch/x86/mm/highmem_32.c BUG_ON(!pte_none(*(kmap_pte-idx))); idx 49 arch/x86/mm/highmem_32.c set_pte(kmap_pte-idx, mk_pte(page, prot)); idx 78 arch/x86/mm/highmem_32.c int idx, type; idx 81 arch/x86/mm/highmem_32.c idx = type + KM_TYPE_NR * smp_processor_id(); idx 84 arch/x86/mm/highmem_32.c WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); idx 92 arch/x86/mm/highmem_32.c kpte_clear_flush(kmap_pte-idx, vaddr); idx 50 arch/x86/mm/iomap_32.c int idx, type; idx 56 arch/x86/mm/iomap_32.c idx = type + KM_TYPE_NR * smp_processor_id(); idx 57 arch/x86/mm/iomap_32.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 58 arch/x86/mm/iomap_32.c set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); idx 95 arch/x86/mm/iomap_32.c int idx, type; idx 98 arch/x86/mm/iomap_32.c idx = type + KM_TYPE_NR * smp_processor_id(); idx 101 arch/x86/mm/iomap_32.c WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); idx 109 arch/x86/mm/iomap_32.c kpte_clear_flush(kmap_pte-idx, vaddr); idx 862 arch/x86/mm/ioremap.c void __init __early_set_fixmap(enum fixed_addresses idx, idx 865 arch/x86/mm/ioremap.c unsigned long addr = __fix_to_virt(idx); idx 868 arch/x86/mm/ioremap.c if (idx >= __end_of_fixed_addresses) { idx 164 arch/x86/mm/numa.c void __init numa_remove_memblk_from(int idx, struct numa_meminfo *mi) idx 167 arch/x86/mm/numa.c memmove(&mi->blk[idx], &mi->blk[idx + 1], idx 168 arch/x86/mm/numa.c (mi->nr_blks - idx) * sizeof(mi->blk[0])); idx 19 arch/x86/mm/numa_internal.h void __init numa_remove_memblk_from(int idx, struct numa_meminfo *mi); idx 258 arch/x86/mm/pageattr.c static unsigned long __cpa_addr(struct cpa_data *cpa, unsigned long idx) idx 261 arch/x86/mm/pageattr.c struct page *page = cpa->pages[idx]; idx 270 arch/x86/mm/pageattr.c return cpa->vaddr[idx]; idx 272 arch/x86/mm/pageattr.c return *cpa->vaddr + idx * PAGE_SIZE; idx 625 arch/x86/mm/pgtable.c void __native_set_fixmap(enum fixed_addresses idx, pte_t pte) idx 627 arch/x86/mm/pgtable.c unsigned long address = __fix_to_virt(idx); idx 638 arch/x86/mm/pgtable.c if (idx >= __end_of_fixed_addresses) { idx 646 arch/x86/mm/pgtable.c void native_set_fixmap(unsigned /* enum fixed_addresses */ idx, idx 652 arch/x86/mm/pgtable.c __native_set_fixmap(idx, pfn_pte(phys >> PAGE_SHIFT, flags)); idx 72 arch/x86/pci/i386.c pcibios_save_fw_addr(struct pci_dev *dev, int idx, resource_size_t fw_addr) idx 89 arch/x86/pci/i386.c map->fw_addr[idx] = fw_addr; idx 95 arch/x86/pci/i386.c map->fw_addr[idx] = fw_addr; idx 99 arch/x86/pci/i386.c resource_size_t pcibios_retrieve_fw_addr(struct pci_dev *dev, int idx) idx 111 arch/x86/pci/i386.c fw_addr = map->fw_addr[idx]; idx 211 arch/x86/pci/i386.c int idx; idx 214 arch/x86/pci/i386.c for (idx = PCI_BRIDGE_RESOURCES; idx < PCI_NUM_RESOURCES; idx++) { idx 215 arch/x86/pci/i386.c r = &dev->resource[idx]; idx 220 arch/x86/pci/i386.c if (!r->start || pci_claim_bridge_resource(dev, idx) < 0) { idx 251 arch/x86/pci/i386.c int idx, disabled, i; idx 264 arch/x86/pci/i386.c for (idx = idx_range[i].start; idx <= idx_range[i].end; idx++) { idx 265 arch/x86/pci/i386.c r = &dev->resource[idx]; idx 277 arch/x86/pci/i386.c idx, r, disabled, pass); idx 278 arch/x86/pci/i386.c if (pci_claim_resource(dev, idx) < 0) { idx 281 arch/x86/pci/i386.c idx, r); idx 285 arch/x86/pci/i386.c idx, r->start); idx 51 arch/x86/um/asm/ptrace.h extern int ptrace_get_thread_area(struct task_struct *child, int idx, idx 54 arch/x86/um/asm/ptrace.h extern int ptrace_set_thread_area(struct task_struct *child, int idx, idx 70 arch/x86/um/asm/ptrace.h static inline int ptrace_get_thread_area(struct task_struct *child, int idx, idx 76 arch/x86/um/asm/ptrace.h static inline int ptrace_set_thread_area(struct task_struct *child, int idx, idx 66 arch/x86/um/tls_32.c int idx; idx 71 arch/x86/um/tls_32.c for (idx = 0; idx < GDT_ENTRY_TLS_ENTRIES; idx++) idx 72 arch/x86/um/tls_32.c if (!t->arch.tls_array[idx].present) idx 73 arch/x86/um/tls_32.c return idx + GDT_ENTRY_TLS_MIN; idx 95 arch/x86/um/tls_32.c int idx; idx 97 arch/x86/um/tls_32.c for (idx = GDT_ENTRY_TLS_MIN; idx < GDT_ENTRY_TLS_MAX; idx++) { idx 99 arch/x86/um/tls_32.c &to->thread.arch.tls_array[idx - GDT_ENTRY_TLS_MIN]; idx 108 arch/x86/um/tls_32.c curr->tls.entry_number = idx; idx 204 arch/x86/um/tls_32.c int idx, int flushed) idx 208 arch/x86/um/tls_32.c if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) idx 211 arch/x86/um/tls_32.c t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].tls = *info; idx 212 arch/x86/um/tls_32.c t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].present = 1; idx 213 arch/x86/um/tls_32.c t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].flushed = flushed; idx 221 arch/x86/um/tls_32.c int idx, ret = -EFAULT; idx 230 arch/x86/um/tls_32.c idx = info.entry_number; idx 232 arch/x86/um/tls_32.c ret = set_tls_entry(new, &info, idx, 0); idx 239 arch/x86/um/tls_32.c int idx) idx 246 arch/x86/um/tls_32.c if (idx < GDT_ENTRY_TLS_MIN || idx > GDT_ENTRY_TLS_MAX) idx 249 arch/x86/um/tls_32.c if (!t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].present) idx 252 arch/x86/um/tls_32.c *info = t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].tls; idx 260 arch/x86/um/tls_32.c !t->arch.tls_array[idx - GDT_ENTRY_TLS_MIN].flushed)) { idx 273 arch/x86/um/tls_32.c info->entry_number = idx; idx 280 arch/x86/um/tls_32.c int idx, ret; idx 288 arch/x86/um/tls_32.c idx = info.entry_number; idx 290 arch/x86/um/tls_32.c if (idx == -1) { idx 291 arch/x86/um/tls_32.c idx = get_free_idx(current); idx 292 arch/x86/um/tls_32.c if (idx < 0) idx 293 arch/x86/um/tls_32.c return idx; idx 294 arch/x86/um/tls_32.c info.entry_number = idx; idx 296 arch/x86/um/tls_32.c if (put_user(idx, &user_desc->entry_number)) idx 303 arch/x86/um/tls_32.c return set_tls_entry(current, &info, idx, 1); idx 311 arch/x86/um/tls_32.c int ptrace_set_thread_area(struct task_struct *child, int idx, idx 322 arch/x86/um/tls_32.c return set_tls_entry(child, &info, idx, 0); idx 328 arch/x86/um/tls_32.c int idx, ret; idx 333 arch/x86/um/tls_32.c if (get_user(idx, &user_desc->entry_number)) idx 336 arch/x86/um/tls_32.c ret = get_tls_entry(current, &info, idx); idx 350 arch/x86/um/tls_32.c int ptrace_get_thread_area(struct task_struct *child, int idx, idx 359 arch/x86/um/tls_32.c ret = get_tls_entry(child, &info, idx); idx 35 arch/x86/xen/enlighten_hvm.c xatp.idx = 0; idx 574 arch/x86/xen/enlighten_pv.c static void xen_load_gs_index(unsigned int idx) idx 576 arch/x86/xen/enlighten_pv.c if (HYPERVISOR_set_segment_base(SEGBASE_GS_USER_SEL, idx)) idx 2295 arch/x86/xen/mmu_pv.c static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot) idx 2301 arch/x86/xen/mmu_pv.c switch (idx) { idx 2343 arch/x86/xen/mmu_pv.c __native_set_fixmap(idx, pte); idx 2348 arch/x86/xen/mmu_pv.c if (idx == VSYSCALL_PAGE) { idx 2349 arch/x86/xen/mmu_pv.c unsigned long vaddr = __fix_to_virt(idx); idx 60 arch/xtensa/include/asm/fixmap.h static __always_inline unsigned long fix_to_virt(const unsigned int idx) idx 67 arch/xtensa/include/asm/fixmap.h BUILD_BUG_ON(idx >= __end_of_fixed_addresses); idx 68 arch/xtensa/include/asm/fixmap.h return __fix_to_virt(idx); idx 126 arch/xtensa/kernel/perf_event.c static inline uint32_t xtensa_pmu_read_counter(int idx) idx 128 arch/xtensa/kernel/perf_event.c return get_er(XTENSA_PMU_PM(idx)); idx 131 arch/xtensa/kernel/perf_event.c static inline void xtensa_pmu_write_counter(int idx, uint32_t v) idx 133 arch/xtensa/kernel/perf_event.c set_er(v, XTENSA_PMU_PM(idx)); idx 137 arch/xtensa/kernel/perf_event.c struct hw_perf_event *hwc, int idx) idx 144 arch/xtensa/kernel/perf_event.c new_raw_count = xtensa_pmu_read_counter(event->hw.idx); idx 155 arch/xtensa/kernel/perf_event.c struct hw_perf_event *hwc, int idx) idx 182 arch/xtensa/kernel/perf_event.c xtensa_pmu_write_counter(idx, -left); idx 243 arch/xtensa/kernel/perf_event.c int idx = hwc->idx; idx 245 arch/xtensa/kernel/perf_event.c if (WARN_ON_ONCE(idx == -1)) idx 250 arch/xtensa/kernel/perf_event.c xtensa_perf_event_set_period(event, hwc, idx); idx 255 arch/xtensa/kernel/perf_event.c set_er(hwc->config, XTENSA_PMU_PMCTRL(idx)); idx 261 arch/xtensa/kernel/perf_event.c int idx = hwc->idx; idx 264 arch/xtensa/kernel/perf_event.c set_er(0, XTENSA_PMU_PMCTRL(idx)); idx 265 arch/xtensa/kernel/perf_event.c set_er(get_er(XTENSA_PMU_PMSTAT(idx)), idx 266 arch/xtensa/kernel/perf_event.c XTENSA_PMU_PMSTAT(idx)); idx 272 arch/xtensa/kernel/perf_event.c xtensa_perf_event_update(event, &event->hw, idx); idx 285 arch/xtensa/kernel/perf_event.c int idx = hwc->idx; idx 287 arch/xtensa/kernel/perf_event.c if (__test_and_set_bit(idx, ev->used_mask)) { idx 288 arch/xtensa/kernel/perf_event.c idx = find_first_zero_bit(ev->used_mask, idx 290 arch/xtensa/kernel/perf_event.c if (idx == XCHAL_NUM_PERF_COUNTERS) idx 293 arch/xtensa/kernel/perf_event.c __set_bit(idx, ev->used_mask); idx 294 arch/xtensa/kernel/perf_event.c hwc->idx = idx; idx 296 arch/xtensa/kernel/perf_event.c ev->event[idx] = event; idx 312 arch/xtensa/kernel/perf_event.c __clear_bit(event->hw.idx, ev->used_mask); idx 318 arch/xtensa/kernel/perf_event.c xtensa_perf_event_update(event, &event->hw, event->hw.idx); idx 421 arch/xtensa/kernel/ptrace.c unsigned idx = addr >> 1; idx 423 arch/xtensa/kernel/ptrace.c if ((!dbreak && idx >= XCHAL_NUM_IBREAK) || idx 424 arch/xtensa/kernel/ptrace.c (dbreak && idx >= XCHAL_NUM_DBREAK)) idx 428 arch/xtensa/kernel/ptrace.c bp = child->thread.ptrace_wp[idx]; idx 430 arch/xtensa/kernel/ptrace.c bp = child->thread.ptrace_bp[idx]; idx 456 arch/xtensa/kernel/ptrace.c unsigned idx = addr >> 1; idx 459 arch/xtensa/kernel/ptrace.c if ((!dbreak && idx >= XCHAL_NUM_IBREAK) || idx 460 arch/xtensa/kernel/ptrace.c (dbreak && idx >= XCHAL_NUM_DBREAK)) idx 467 arch/xtensa/kernel/ptrace.c bp = child->thread.ptrace_wp[idx]; idx 473 arch/xtensa/kernel/ptrace.c bp = child->thread.ptrace_bp[idx]; idx 483 arch/xtensa/kernel/ptrace.c child->thread.ptrace_wp[idx] = bp; idx 485 arch/xtensa/kernel/ptrace.c child->thread.ptrace_bp[idx] = bp; idx 42 arch/xtensa/mm/highmem.c enum fixed_addresses idx; idx 50 arch/xtensa/mm/highmem.c idx = kmap_idx(kmap_atomic_idx_push(), idx 52 arch/xtensa/mm/highmem.c vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); idx 54 arch/xtensa/mm/highmem.c BUG_ON(!pte_none(*(kmap_pte + idx))); idx 56 arch/xtensa/mm/highmem.c set_pte(kmap_pte + idx, mk_pte(page, PAGE_KERNEL_EXEC)); idx 66 arch/xtensa/mm/highmem.c int idx = kmap_idx(kmap_atomic_idx(), idx 75 arch/xtensa/mm/highmem.c pte_clear(&init_mm, kvaddr, kmap_pte + idx); idx 705 block/bfq-wf2q.c unsigned int idx = bfq_class_idx(entity); idx 707 block/bfq-wf2q.c return sched_data->service_tree + idx; idx 59 block/bio-integrity.c unsigned long idx = 0; idx 61 block/bio-integrity.c bip->bip_vec = bvec_alloc(gfp_mask, nr_vecs, &idx, idx 65 block/bio-integrity.c bip->bip_max_vcnt = bvec_nr_vecs(idx); idx 66 block/bio-integrity.c bip->bip_slab = idx; idx 147 block/bio.c unsigned int bvec_nr_vecs(unsigned short idx) idx 149 block/bio.c return bvec_slabs[--idx].nr_vecs; idx 152 block/bio.c void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned int idx) idx 154 block/bio.c if (!idx) idx 156 block/bio.c idx--; idx 158 block/bio.c BIO_BUG_ON(idx >= BVEC_POOL_NR); idx 160 block/bio.c if (idx == BVEC_POOL_MAX) { idx 163 block/bio.c struct biovec_slab *bvs = bvec_slabs + idx; idx 169 block/bio.c struct bio_vec *bvec_alloc(gfp_t gfp_mask, int nr, unsigned long *idx, idx 179 block/bio.c *idx = 0; idx 182 block/bio.c *idx = 1; idx 185 block/bio.c *idx = 2; idx 188 block/bio.c *idx = 3; idx 191 block/bio.c *idx = 4; idx 194 block/bio.c *idx = 5; idx 204 block/bio.c if (*idx == BVEC_POOL_MAX) { idx 208 block/bio.c struct biovec_slab *bvs = bvec_slabs + *idx; idx 224 block/bio.c *idx = BVEC_POOL_MAX; idx 229 block/bio.c (*idx)++; idx 498 block/bio.c unsigned long idx = 0; idx 500 block/bio.c bvl = bvec_alloc(gfp_mask, nr_iovecs, &idx, &bs->bvec_pool); idx 504 block/bio.c bvl = bvec_alloc(gfp_mask, nr_iovecs, &idx, &bs->bvec_pool); idx 510 block/bio.c bio->bi_flags |= idx << BVEC_POOL_OFFSET; idx 202 block/blk-core.c int idx = (__force int)status; idx 204 block/blk-core.c if (WARN_ON_ONCE(idx >= ARRAY_SIZE(blk_errors))) idx 206 block/blk-core.c return blk_errors[idx].errno; idx 213 block/blk-core.c int idx = (__force int)status; idx 215 block/blk-core.c if (WARN_ON_ONCE(idx >= ARRAY_SIZE(blk_errors))) idx 221 block/blk-core.c caller, blk_errors[idx].name, idx 721 block/blk-iocost.c int idx = ioc->autop_idx; idx 722 block/blk-iocost.c const struct ioc_params *p = &autop[idx]; idx 735 block/blk-iocost.c if (idx < AUTOP_SSD_DFL) idx 740 block/blk-iocost.c return idx; idx 751 block/blk-iocost.c return idx + 1; idx 760 block/blk-iocost.c return idx - 1; idx 765 block/blk-iocost.c return idx; idx 819 block/blk-iocost.c int idx; idx 823 block/blk-iocost.c idx = ioc_autop_idx(ioc); idx 824 block/blk-iocost.c p = &autop[idx]; idx 826 block/blk-iocost.c if (idx == ioc->autop_idx && !force) idx 829 block/blk-iocost.c if (idx != ioc->autop_idx) idx 832 block/blk-iocost.c ioc->autop_idx = idx; idx 1589 block/blk-iocost.c int idx = min_t(int, abs(ioc->busy_level), idx 1591 block/blk-iocost.c u32 adj_pct = vrate_adj_pct[idx]; idx 121 block/blk-mq-sched.c unsigned short idx = ctx->index_hw[hctx->type]; idx 123 block/blk-mq-sched.c if (++idx == hctx->nr_ctx) idx 124 block/blk-mq-sched.c idx = 0; idx 126 block/blk-mq-sched.c return hctx->ctxs[idx]; idx 171 block/genhd.c piter->idx = ptbl->len - 1; idx 173 block/genhd.c piter->idx = 0; idx 175 block/genhd.c piter->idx = 1; idx 219 block/genhd.c for (; piter->idx != end; piter->idx += inc) { idx 222 block/genhd.c part = rcu_dereference(ptbl->part[piter->idx]); idx 228 block/genhd.c piter->idx == 0)) idx 233 block/genhd.c piter->idx += inc; idx 491 block/genhd.c int idx; idx 503 block/genhd.c idx = idr_alloc(&ext_devt_idr, part, 0, NR_EXT_DEVT, GFP_NOWAIT); idx 507 block/genhd.c if (idx < 0) idx 508 block/genhd.c return idx == -ENOSPC ? -EBUSY : idx; idx 510 block/genhd.c *devt = MKDEV(BLOCK_EXT_MAJOR, blk_mangle_minor(idx)); idx 1242 crypto/ecc.c unsigned int idx; idx 1254 crypto/ecc.c idx = (!!vli_test_bit(u1, i)) | ((!!vli_test_bit(u2, i)) << 1); idx 1255 crypto/ecc.c point = points[idx]; idx 1264 crypto/ecc.c idx = (!!vli_test_bit(u1, i)) | ((!!vli_test_bit(u2, i)) << 1); idx 1265 crypto/ecc.c point = points[idx]; idx 67 drivers/acpi/acpi_extlog.c #define ELOG_ENTRY_DATA(idx) \ idx 68 drivers/acpi/acpi_extlog.c (*(l1_entry_base + (idx))) idx 75 drivers/acpi/acpi_extlog.c int idx; idx 80 drivers/acpi/acpi_extlog.c idx = ELOG_IDX(cpu, bank); idx 81 drivers/acpi/acpi_extlog.c data = ELOG_ENTRY_DATA(idx); idx 599 drivers/acpi/arm64/iort.c unsigned int idx, int *its_id) idx 614 drivers/acpi/arm64/iort.c if (idx >= its->its_count) { idx 616 drivers/acpi/arm64/iort.c idx, its->its_count); idx 620 drivers/acpi/arm64/iort.c *its_id = its->identifiers[idx]; idx 1390 drivers/acpi/arm64/iort.c int idx; idx 1392 drivers/acpi/arm64/iort.c idx = acpi_match_platform_list(pmcg_plat_info); idx 1393 drivers/acpi/arm64/iort.c if (idx >= 0) idx 1394 drivers/acpi/arm64/iort.c model = pmcg_plat_info[idx].data; idx 463 drivers/acpi/hmat/hmat.c unsigned int ipds, tpds, i, idx = 0, tdx = 0; idx 475 drivers/acpi/hmat/hmat.c idx = i; idx 492 drivers/acpi/hmat/hmat.c return hmat_normalize(entries[idx * tpds + tdx], idx 679 drivers/acpi/property.c int ret, idx = 0; idx 765 drivers/acpi/property.c if (idx == index) { idx 776 drivers/acpi/property.c if (idx == index) idx 783 drivers/acpi/property.c idx++; idx 845 drivers/acpi/utils.c int idx = 0; idx 850 drivers/acpi/utils.c for (; plat->oem_id[0]; plat++, idx++) { idx 864 drivers/acpi/utils.c return idx; idx 7031 drivers/ata/libata-core.c int idx = 0, size = 1; idx 7065 drivers/ata/libata-core.c ata_force_tbl[idx++] = te; idx 7071 drivers/ata/libata-core.c ata_force_tbl_size = idx; idx 326 drivers/ata/libata-eh.c int idx; idx 331 drivers/ata/libata-eh.c idx = ehc->cmd_timeout_idx[dev->devno][ent]; idx 332 drivers/ata/libata-eh.c return ata_eh_cmd_timeout_table[ent].timeouts[idx]; idx 351 drivers/ata/libata-eh.c int idx; idx 356 drivers/ata/libata-eh.c idx = ehc->cmd_timeout_idx[dev->devno][ent]; idx 357 drivers/ata/libata-eh.c if (ata_eh_cmd_timeout_table[ent].timeouts[idx + 1] != ULONG_MAX) idx 390 drivers/ata/libata-eh.c int idx, rc = 0; idx 393 drivers/ata/libata-eh.c idx = ering->cursor; idx 395 drivers/ata/libata-eh.c ent = &ering->ring[idx]; idx 401 drivers/ata/libata-eh.c idx = (idx - 1 + ATA_ERING_SIZE) % ATA_ERING_SIZE; idx 402 drivers/ata/libata-eh.c } while (idx != ering->cursor); idx 580 drivers/ata/sata_mv.c void (*read_preamp)(struct mv_host_priv *hpriv, int idx, idx 607 drivers/ata/sata_mv.c static void mv5_read_preamp(struct mv_host_priv *hpriv, int idx, idx 617 drivers/ata/sata_mv.c static void mv6_read_preamp(struct mv_host_priv *hpriv, int idx, idx 624 drivers/ata/sata_mv.c static void mv_soc_read_preamp(struct mv_host_priv *hpriv, int idx, idx 3102 drivers/ata/sata_mv.c static void mv5_read_preamp(struct mv_host_priv *hpriv, int idx, idx 3105 drivers/ata/sata_mv.c void __iomem *phy_mmio = mv5_phy_base(mmio, idx); idx 3110 drivers/ata/sata_mv.c hpriv->signal[idx].pre = tmp & 0x1800; /* bits 12:11 */ idx 3111 drivers/ata/sata_mv.c hpriv->signal[idx].amps = tmp & 0xe0; /* bits 7:5 */ idx 3314 drivers/ata/sata_mv.c static void mv6_read_preamp(struct mv_host_priv *hpriv, int idx, idx 3322 drivers/ata/sata_mv.c hpriv->signal[idx].amps = 0x7 << 8; idx 3323 drivers/ata/sata_mv.c hpriv->signal[idx].pre = 0x1 << 5; idx 3327 drivers/ata/sata_mv.c port_mmio = mv_port_base(mmio, idx); idx 3330 drivers/ata/sata_mv.c hpriv->signal[idx].amps = tmp & 0x700; /* bits 10:8 */ idx 3331 drivers/ata/sata_mv.c hpriv->signal[idx].pre = tmp & 0xe0; /* bits 7:5 */ idx 3423 drivers/ata/sata_mv.c static void mv_soc_read_preamp(struct mv_host_priv *hpriv, int idx, idx 3429 drivers/ata/sata_mv.c port_mmio = mv_port_base(mmio, idx); idx 3432 drivers/ata/sata_mv.c hpriv->signal[idx].amps = tmp & 0x700; /* bits 10:8 */ idx 3433 drivers/ata/sata_mv.c hpriv->signal[idx].pre = tmp & 0xe0; /* bits 7:5 */ idx 770 drivers/ata/sata_nv.c unsigned int idx = 0; idx 774 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_ERR << 8) | tf->hob_feature | WNB); idx 775 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_NSECT << 8) | tf->hob_nsect); idx 776 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_LBAL << 8) | tf->hob_lbal); idx 777 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_LBAM << 8) | tf->hob_lbam); idx 778 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_LBAH << 8) | tf->hob_lbah); idx 779 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_ERR << 8) | tf->feature); idx 781 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_ERR << 8) | tf->feature | WNB); idx 783 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_NSECT << 8) | tf->nsect); idx 784 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_LBAL << 8) | tf->lbal); idx 785 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_LBAM << 8) | tf->lbam); idx 786 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_LBAH << 8) | tf->lbah); idx 790 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_DEVICE << 8) | tf->device); idx 792 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16((ATA_REG_CMD << 8) | tf->command | CMDEND); idx 794 drivers/ata/sata_nv.c while (idx < 12) idx 795 drivers/ata/sata_nv.c cpb[idx++] = cpu_to_le16(IGN); idx 797 drivers/ata/sata_nv.c return idx; idx 1293 drivers/ata/sata_nv.c int idx, idx 1299 drivers/ata/sata_nv.c if (idx == qc->n_elem - 1) idx 1301 drivers/ata/sata_nv.c else if (idx != 4) idx 1972 drivers/ata/sata_nv.c unsigned int si, idx; idx 1976 drivers/ata/sata_nv.c idx = 0; idx 1990 drivers/ata/sata_nv.c prd[idx].addr = cpu_to_le32(addr); idx 1991 drivers/ata/sata_nv.c prd[idx].flags_len = cpu_to_le32(len & 0xffff); idx 1993 drivers/ata/sata_nv.c idx++; idx 1999 drivers/ata/sata_nv.c prd[idx - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); idx 579 drivers/ata/sata_promise.c unsigned int si, idx; idx 585 drivers/ata/sata_promise.c idx = 0; idx 603 drivers/ata/sata_promise.c prd[idx].addr = cpu_to_le32(addr); idx 604 drivers/ata/sata_promise.c prd[idx].flags_len = cpu_to_le32(len & 0xffff); idx 605 drivers/ata/sata_promise.c VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", idx, addr, len); idx 607 drivers/ata/sata_promise.c idx++; idx 613 drivers/ata/sata_promise.c len = le32_to_cpu(prd[idx - 1].flags_len); idx 620 drivers/ata/sata_promise.c addr = le32_to_cpu(prd[idx - 1].addr); idx 621 drivers/ata/sata_promise.c prd[idx - 1].flags_len = cpu_to_le32(len - SG_COUNT_ASIC_BUG); idx 622 drivers/ata/sata_promise.c VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", idx - 1, addr, SG_COUNT_ASIC_BUG); idx 626 drivers/ata/sata_promise.c prd[idx].addr = cpu_to_le32(addr); idx 627 drivers/ata/sata_promise.c prd[idx].flags_len = cpu_to_le32(len); idx 628 drivers/ata/sata_promise.c VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", idx, addr, len); idx 630 drivers/ata/sata_promise.c idx++; idx 633 drivers/ata/sata_promise.c prd[idx - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); idx 445 drivers/ata/sata_sx4.c unsigned int i, si, idx, total_len = 0, sgt_len; idx 458 drivers/ata/sata_sx4.c idx = 0; idx 460 drivers/ata/sata_sx4.c buf[idx++] = cpu_to_le32(sg_dma_address(sg)); idx 461 drivers/ata/sata_sx4.c buf[idx++] = cpu_to_le32(sg_dma_len(sg)); idx 464 drivers/ata/sata_sx4.c buf[idx - 1] |= cpu_to_le32(ATA_PRD_EOT); idx 465 drivers/ata/sata_sx4.c sgt_len = idx * 4; idx 571 drivers/ata/sata_sx4.c unsigned int idx = pp->hdma_prod & PDC_HDMA_Q_MASK; idx 579 drivers/ata/sata_sx4.c pp->hdma[idx].qc = qc; idx 580 drivers/ata/sata_sx4.c pp->hdma[idx].seq = seq; idx 581 drivers/ata/sata_sx4.c pp->hdma[idx].pkt_ofs = pkt_ofs; idx 589 drivers/ata/sata_sx4.c unsigned int idx = pp->hdma_cons & PDC_HDMA_Q_MASK; idx 597 drivers/ata/sata_sx4.c __pdc20621_push_hdma(pp->hdma[idx].qc, pp->hdma[idx].seq, idx 598 drivers/ata/sata_sx4.c pp->hdma[idx].pkt_ofs); idx 985 drivers/ata/sata_sx4.c u16 idx; idx 996 drivers/ata/sata_sx4.c idx = (u16) (offset / window_size); idx 1000 drivers/ata/sata_sx4.c writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); idx 1003 drivers/ata/sata_sx4.c offset -= (idx * window_size); idx 1004 drivers/ata/sata_sx4.c idx++; idx 1014 drivers/ata/sata_sx4.c writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); idx 1019 drivers/ata/sata_sx4.c idx++; idx 1025 drivers/ata/sata_sx4.c writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); idx 1037 drivers/ata/sata_sx4.c u16 idx; idx 1048 drivers/ata/sata_sx4.c idx = (u16) (offset / window_size); idx 1050 drivers/ata/sata_sx4.c writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); idx 1052 drivers/ata/sata_sx4.c offset -= (idx * window_size); idx 1053 drivers/ata/sata_sx4.c idx++; idx 1063 drivers/ata/sata_sx4.c writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); idx 1070 drivers/ata/sata_sx4.c idx++; idx 1074 drivers/ata/sata_sx4.c writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR); idx 830 drivers/atm/he.c he_dev->rbpl_base[i].idx = i << RBP_IDX_OFFSET; idx 1916 drivers/atm/he.c new_tail->idx = i << RBP_IDX_OFFSET; idx 211 drivers/atm/he.h volatile u32 idx; /* virt */ idx 1442 drivers/auxdisplay/panel.c const char *idx; idx 1444 drivers/auxdisplay/panel.c idx = strchr(sigtab, *name); idx 1445 drivers/auxdisplay/panel.c if (!idx) idx 1448 drivers/auxdisplay/panel.c in = idx - sigtab; idx 167 drivers/base/base.h extern void device_links_read_unlock(int idx); idx 358 drivers/base/cacheinfo.c #define per_cache_index_dev(cpu, idx) ((per_cpu_index_dev(cpu))[idx]) idx 68 drivers/base/core.c void device_links_read_unlock(int idx) idx 70 drivers/base/core.c srcu_read_unlock(&device_links_srcu, idx); idx 216 drivers/base/core.c int idx; idx 218 drivers/base/core.c idx = device_links_read_lock(); idx 222 drivers/base/core.c device_links_read_unlock(idx); idx 1980 drivers/base/power/domain.c unsigned int idx = genpdspec->args[0]; idx 1985 drivers/base/power/domain.c if (idx >= genpd_data->num_domains) { idx 1986 drivers/base/power/domain.c pr_err("%s: invalid domain index %u\n", __func__, idx); idx 1990 drivers/base/power/domain.c if (!genpd_data->domains[idx]) idx 1993 drivers/base/power/domain.c return genpd_data->domains[idx]; idx 258 drivers/base/power/main.c int idx; idx 260 drivers/base/power/main.c idx = device_links_read_lock(); idx 273 drivers/base/power/main.c device_links_read_unlock(idx); idx 313 drivers/base/power/main.c int idx; idx 315 drivers/base/power/main.c idx = device_links_read_lock(); idx 330 drivers/base/power/main.c device_links_read_unlock(idx); idx 1231 drivers/base/power/main.c int idx; idx 1236 drivers/base/power/main.c idx = device_links_read_lock(); idx 1241 drivers/base/power/main.c device_links_read_unlock(idx); idx 1688 drivers/base/power/main.c int idx; idx 1696 drivers/base/power/main.c idx = device_links_read_lock(); idx 1704 drivers/base/power/main.c device_links_read_unlock(idx); idx 331 drivers/base/power/runtime.c int retval, idx; idx 347 drivers/base/power/runtime.c idx = device_links_read_lock(); idx 353 drivers/base/power/runtime.c device_links_read_unlock(idx); idx 372 drivers/base/power/runtime.c idx = device_links_read_lock(); idx 377 drivers/base/power/runtime.c device_links_read_unlock(idx); idx 1165 drivers/base/power/runtime.c int idx = device_links_read_lock(); idx 1171 drivers/base/power/runtime.c device_links_read_unlock(idx); idx 1221 drivers/base/power/runtime.c int idx = device_links_read_lock(); idx 1225 drivers/base/power/runtime.c device_links_read_unlock(idx); idx 1641 drivers/base/power/runtime.c int idx; idx 1643 drivers/base/power/runtime.c idx = device_links_read_lock(); idx 1654 drivers/base/power/runtime.c device_links_read_unlock(idx); idx 1664 drivers/base/power/runtime.c int idx; idx 1666 drivers/base/power/runtime.c idx = device_links_read_lock(); idx 1676 drivers/base/power/runtime.c device_links_read_unlock(idx); idx 1686 drivers/base/power/runtime.c int idx; idx 1688 drivers/base/power/runtime.c idx = device_links_read_lock(); idx 1698 drivers/base/power/runtime.c device_links_read_unlock(idx); idx 250 drivers/base/regmap/internal.h unsigned int idx) idx 252 drivers/base/regmap/internal.h return base + (map->cache_word_size * idx); idx 256 drivers/base/regmap/internal.h unsigned int idx); idx 257 drivers/base/regmap/internal.h bool regcache_set_val(struct regmap *map, void *base, unsigned int idx, idx 49 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode, unsigned int idx) idx 51 drivers/base/regmap/regcache-rbtree.c return regcache_get_val(map, rbnode->block, idx); idx 56 drivers/base/regmap/regcache-rbtree.c unsigned int idx, unsigned int val) idx 58 drivers/base/regmap/regcache-rbtree.c set_bit(idx, rbnode->cache_present); idx 59 drivers/base/regmap/regcache-rbtree.c regcache_set_val(map, rbnode->block, idx, val); idx 548 drivers/base/regmap/regcache.c bool regcache_set_val(struct regmap *map, void *base, unsigned int idx, idx 551 drivers/base/regmap/regcache.c if (regcache_get_val(map, base, idx) == val) idx 556 drivers/base/regmap/regcache.c map->format.format_val(base + (map->cache_word_size * idx), idx 565 drivers/base/regmap/regcache.c cache[idx] = val; idx 571 drivers/base/regmap/regcache.c cache[idx] = val; idx 577 drivers/base/regmap/regcache.c cache[idx] = val; idx 584 drivers/base/regmap/regcache.c cache[idx] = val; idx 595 drivers/base/regmap/regcache.c unsigned int idx) idx 603 drivers/base/regmap/regcache.c idx)); idx 609 drivers/base/regmap/regcache.c return cache[idx]; idx 614 drivers/base/regmap/regcache.c return cache[idx]; idx 619 drivers/base/regmap/regcache.c return cache[idx]; idx 625 drivers/base/regmap/regcache.c return cache[idx]; idx 660 drivers/base/regmap/regcache.c static bool regcache_reg_present(unsigned long *cache_present, unsigned int idx) idx 665 drivers/base/regmap/regcache.c return test_bit(idx, cache_present); idx 58 drivers/block/brd.c pgoff_t idx; idx 73 drivers/block/brd.c idx = sector >> PAGE_SECTORS_SHIFT; /* sector to page index */ idx 74 drivers/block/brd.c page = radix_tree_lookup(&brd->brd_pages, idx); idx 77 drivers/block/brd.c BUG_ON(page && page->index != idx); idx 89 drivers/block/brd.c pgoff_t idx; idx 112 drivers/block/brd.c idx = sector >> PAGE_SECTORS_SHIFT; idx 113 drivers/block/brd.c page->index = idx; idx 114 drivers/block/brd.c if (radix_tree_insert(&brd->brd_pages, idx, page)) { idx 116 drivers/block/brd.c page = radix_tree_lookup(&brd->brd_pages, idx); idx 118 drivers/block/brd.c BUG_ON(page->index != idx); idx 365 drivers/block/drbd/drbd_actlog.c unsigned idx = device->al_tr_cycle + i; idx 366 drivers/block/drbd/drbd_actlog.c extent_nr = lc_element_by_index(device->act_log, idx)->lc_number; idx 197 drivers/block/drbd/drbd_bitmap.c static void bm_store_page_idx(struct page *page, unsigned long idx) idx 199 drivers/block/drbd/drbd_bitmap.c BUG_ON(0 != (idx & ~BM_PAGE_IDX_MASK)); idx 200 drivers/block/drbd/drbd_bitmap.c set_page_private(page, idx); idx 312 drivers/block/drbd/drbd_bitmap.c static unsigned long *__bm_map_pidx(struct drbd_bitmap *b, unsigned int idx) idx 314 drivers/block/drbd/drbd_bitmap.c struct page *page = b->bm_pages[idx]; idx 318 drivers/block/drbd/drbd_bitmap.c static unsigned long *bm_map_pidx(struct drbd_bitmap *b, unsigned int idx) idx 320 drivers/block/drbd/drbd_bitmap.c return __bm_map_pidx(b, idx); idx 561 drivers/block/drbd/drbd_bitmap.c int idx, last_word; idx 564 drivers/block/drbd/drbd_bitmap.c for (idx = 0; idx < b->bm_number_of_pages - 1; idx++) { idx 565 drivers/block/drbd/drbd_bitmap.c p_addr = __bm_map_pidx(b, idx); idx 572 drivers/block/drbd/drbd_bitmap.c p_addr = __bm_map_pidx(b, idx); idx 587 drivers/block/drbd/drbd_bitmap.c unsigned int idx; idx 599 drivers/block/drbd/drbd_bitmap.c idx = bm_word_to_page_idx(b, offset); idx 600 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, idx); idx 608 drivers/block/drbd/drbd_bitmap.c bm_set_page_need_writeout(b->bm_pages[idx]); idx 816 drivers/block/drbd/drbd_bitmap.c unsigned int idx; idx 833 drivers/block/drbd/drbd_bitmap.c idx = bm_word_to_page_idx(b, offset); idx 834 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, idx); idx 844 drivers/block/drbd/drbd_bitmap.c bm_set_page_need_writeout(b->bm_pages[idx]); idx 944 drivers/block/drbd/drbd_bitmap.c unsigned int idx = bm_page_to_idx(bio_first_page_all(bio)); idx 947 drivers/block/drbd/drbd_bitmap.c !bm_test_page_unchanged(b->bm_pages[idx])) idx 948 drivers/block/drbd/drbd_bitmap.c drbd_warn(device, "bitmap page idx %u changed during IO!\n", idx); idx 954 drivers/block/drbd/drbd_bitmap.c bm_set_page_io_err(b->bm_pages[idx]); idx 959 drivers/block/drbd/drbd_bitmap.c bio->bi_status, idx); idx 961 drivers/block/drbd/drbd_bitmap.c bm_clear_page_io_err(b->bm_pages[idx]); idx 962 drivers/block/drbd/drbd_bitmap.c dynamic_drbd_dbg(device, "bitmap page idx %u completed\n", idx); idx 965 drivers/block/drbd/drbd_bitmap.c bm_page_unlock_io(device, idx); idx 1599 drivers/block/drbd/drbd_bitmap.c unsigned int idx = bm_bit_to_page_idx(b, bitnr); idx 1600 drivers/block/drbd/drbd_bitmap.c if (page_nr != idx) { idx 1601 drivers/block/drbd/drbd_bitmap.c page_nr = idx; idx 1604 drivers/block/drbd/drbd_bitmap.c p_addr = bm_map_pidx(b, idx); idx 722 drivers/block/drbd/drbd_debugfs.c enum drbd_uuid_index idx; idx 729 drivers/block/drbd/drbd_debugfs.c for (idx = UI_CURRENT; idx <= UI_HISTORY_END; idx++) { idx 730 drivers/block/drbd/drbd_debugfs.c seq_printf(m, "0x%016llX\n", md->uuid[idx]); idx 1130 drivers/block/drbd/drbd_int.h extern void drbd_uuid_set(struct drbd_device *device, int idx, u64 val) __must_hold(local); idx 1131 drivers/block/drbd/drbd_int.h extern void _drbd_uuid_set(struct drbd_device *device, int idx, u64 val) __must_hold(local); idx 1135 drivers/block/drbd/drbd_int.h extern void __drbd_uuid_set(struct drbd_device *device, int idx, u64 val) __must_hold(local); idx 3442 drivers/block/drbd/drbd_main.c void __drbd_uuid_set(struct drbd_device *device, int idx, u64 val) __must_hold(local) idx 3444 drivers/block/drbd/drbd_main.c if (idx == UI_CURRENT) { idx 3453 drivers/block/drbd/drbd_main.c device->ldev->md.uuid[idx] = val; idx 3457 drivers/block/drbd/drbd_main.c void _drbd_uuid_set(struct drbd_device *device, int idx, u64 val) __must_hold(local) idx 3461 drivers/block/drbd/drbd_main.c __drbd_uuid_set(device, idx, val); idx 3465 drivers/block/drbd/drbd_main.c void drbd_uuid_set(struct drbd_device *device, int idx, u64 val) __must_hold(local) idx 3469 drivers/block/drbd/drbd_main.c if (device->ldev->md.uuid[idx]) { idx 3471 drivers/block/drbd/drbd_main.c device->ldev->md.uuid[UI_HISTORY_START] = device->ldev->md.uuid[idx]; idx 3473 drivers/block/drbd/drbd_main.c __drbd_uuid_set(device, idx, val); idx 693 drivers/block/null_blk_main.c u64 idx; idx 698 drivers/block/null_blk_main.c idx = sector >> PAGE_SECTORS_SHIFT; idx 701 drivers/block/null_blk_main.c t_page = radix_tree_lookup(root, idx); idx 706 drivers/block/null_blk_main.c ret = radix_tree_delete_item(root, idx, t_page); idx 715 drivers/block/null_blk_main.c static struct nullb_page *null_radix_tree_insert(struct nullb *nullb, u64 idx, idx 722 drivers/block/null_blk_main.c if (radix_tree_insert(root, idx, t_page)) { idx 724 drivers/block/null_blk_main.c t_page = radix_tree_lookup(root, idx); idx 725 drivers/block/null_blk_main.c WARN_ON(!t_page || t_page->page->index != idx); idx 765 drivers/block/null_blk_main.c u64 idx; idx 769 drivers/block/null_blk_main.c idx = sector >> PAGE_SECTORS_SHIFT; idx 773 drivers/block/null_blk_main.c t_page = radix_tree_lookup(root, idx); idx 774 drivers/block/null_blk_main.c WARN_ON(t_page && t_page->page->index != idx); idx 799 drivers/block/null_blk_main.c u64 idx; idx 816 drivers/block/null_blk_main.c idx = sector >> PAGE_SECTORS_SHIFT; idx 817 drivers/block/null_blk_main.c t_page->page->index = idx; idx 818 drivers/block/null_blk_main.c t_page = null_radix_tree_insert(nullb, idx, t_page, !ignore_cache); idx 833 drivers/block/null_blk_main.c u64 idx; idx 837 drivers/block/null_blk_main.c idx = c_page->page->index; idx 839 drivers/block/null_blk_main.c t_page = null_insert_page(nullb, idx << PAGE_SECTORS_SHIFT, true); idx 846 drivers/block/null_blk_main.c idx, t_page); idx 871 drivers/block/null_blk_main.c ret = radix_tree_delete_item(&nullb->dev->cache, idx, c_page); idx 357 drivers/block/pktcdvd.c int idx; idx 359 drivers/block/pktcdvd.c for (idx = 0; idx < MAX_WRITERS; idx++) { idx 360 drivers/block/pktcdvd.c struct pktcdvd_device *pd = pkt_devs[idx]; idx 2700 drivers/block/pktcdvd.c int idx; idx 2707 drivers/block/pktcdvd.c for (idx = 0; idx < MAX_WRITERS; idx++) idx 2708 drivers/block/pktcdvd.c if (!pkt_devs[idx]) idx 2710 drivers/block/pktcdvd.c if (idx == MAX_WRITERS) { idx 2733 drivers/block/pktcdvd.c sprintf(pd->name, DRIVER_NAME"%d", idx); idx 2746 drivers/block/pktcdvd.c disk->first_minor = idx; idx 2756 drivers/block/pktcdvd.c pd->pkt_dev = MKDEV(pktdev_major, idx); idx 2769 drivers/block/pktcdvd.c pkt_devs[idx] = pd; idx 2793 drivers/block/pktcdvd.c int idx; idx 2798 drivers/block/pktcdvd.c for (idx = 0; idx < MAX_WRITERS; idx++) { idx 2799 drivers/block/pktcdvd.c pd = pkt_devs[idx]; idx 2803 drivers/block/pktcdvd.c if (idx == MAX_WRITERS) { idx 2816 drivers/block/pktcdvd.c pkt_devs[idx] = NULL; idx 352 drivers/block/rsxx/dma.c if (ctrl->cmd.idx != ioread32(ctrl->regmap + SW_CMD_IDX)) { idx 359 drivers/block/rsxx/dma.c iowrite32(ctrl->cmd.idx, ctrl->regmap + SW_CMD_IDX); idx 451 drivers/block/rsxx/dma.c hw_cmd_buf[ctrl->cmd.idx].command = dma->cmd; idx 452 drivers/block/rsxx/dma.c hw_cmd_buf[ctrl->cmd.idx].tag = tag; idx 453 drivers/block/rsxx/dma.c hw_cmd_buf[ctrl->cmd.idx]._rsvd = 0; idx 454 drivers/block/rsxx/dma.c hw_cmd_buf[ctrl->cmd.idx].sub_page = idx 458 drivers/block/rsxx/dma.c hw_cmd_buf[ctrl->cmd.idx].device_addr = idx 461 drivers/block/rsxx/dma.c hw_cmd_buf[ctrl->cmd.idx].host_addr = idx 466 drivers/block/rsxx/dma.c ctrl->id, dma->laddr, tag, ctrl->cmd.idx); idx 468 drivers/block/rsxx/dma.c ctrl->cmd.idx = (ctrl->cmd.idx + 1) & RSXX_CS_IDX_MASK; idx 490 drivers/block/rsxx/dma.c iowrite32(ctrl->cmd.idx, ctrl->regmap + SW_CMD_IDX); idx 510 drivers/block/rsxx/dma.c count = le16_to_cpu(hw_st_buf[ctrl->status.idx].count); idx 522 drivers/block/rsxx/dma.c status = hw_st_buf[ctrl->status.idx].status; idx 523 drivers/block/rsxx/dma.c tag = hw_st_buf[ctrl->status.idx].tag; idx 534 drivers/block/rsxx/dma.c tag, ctrl->status.idx, ctrl->id); idx 542 drivers/block/rsxx/dma.c ctrl->status.idx); idx 556 drivers/block/rsxx/dma.c ctrl->status.idx = (ctrl->status.idx + 1) & idx 560 drivers/block/rsxx/dma.c count = le16_to_cpu(hw_st_buf[ctrl->status.idx].count); idx 782 drivers/block/rsxx/dma.c ctrl->status.idx = ioread32(ctrl->regmap + HW_STATUS_CNT); idx 783 drivers/block/rsxx/dma.c if (ctrl->status.idx > RSXX_MAX_OUTSTANDING_CMDS) { idx 785 drivers/block/rsxx/dma.c ctrl->status.idx); idx 788 drivers/block/rsxx/dma.c iowrite32(ctrl->status.idx, ctrl->regmap + HW_STATUS_CNT); idx 789 drivers/block/rsxx/dma.c iowrite32(ctrl->status.idx, ctrl->regmap + SW_STATUS_CNT); idx 791 drivers/block/rsxx/dma.c ctrl->cmd.idx = ioread32(ctrl->regmap + HW_CMD_IDX); idx 792 drivers/block/rsxx/dma.c if (ctrl->cmd.idx > RSXX_MAX_OUTSTANDING_CMDS) { idx 794 drivers/block/rsxx/dma.c ctrl->status.idx); idx 797 drivers/block/rsxx/dma.c iowrite32(ctrl->cmd.idx, ctrl->regmap + HW_CMD_IDX); idx 798 drivers/block/rsxx/dma.c iowrite32(ctrl->cmd.idx, ctrl->regmap + SW_CMD_IDX); idx 65 drivers/block/rsxx/rsxx_priv.h u32 idx; idx 1102 drivers/block/sunvdc.c u32 idx; idx 1104 drivers/block/sunvdc.c for (idx = dr->cons; idx != dr->prod; idx = vio_dring_next(dr, idx)) { idx 1105 drivers/block/sunvdc.c struct vio_disk_desc *desc = vio_dring_entry(dr, idx); idx 1106 drivers/block/sunvdc.c struct vdc_req_entry *rqe = &port->rq_arr[idx]; idx 1111 drivers/block/sunvdc.c dr->cons = vio_dring_next(dr, idx); idx 73 drivers/block/sx8.c #define NEXT_RESP(idx) ((idx + 1) % RMSG_Q_LEN) idx 587 drivers/block/sx8.c unsigned int idx, void *mem) idx 596 drivers/block/sx8.c st->handle = cpu_to_le32(TAG_ENCODE(idx)); idx 603 drivers/block/sx8.c unsigned int idx, void *mem) idx 610 drivers/block/sx8.c ab->handle = cpu_to_le32(TAG_ENCODE(idx)); idx 628 drivers/block/sx8.c unsigned int idx, void *mem) idx 631 drivers/block/sx8.c u32 msg_data = (u32) (carm_ref_msg_dma(host, idx) + idx 637 drivers/block/sx8.c ioc->handle = cpu_to_le32(TAG_ENCODE(idx)); idx 648 drivers/block/sx8.c unsigned int idx, void *mem) idx 651 drivers/block/sx8.c u32 msg_data = (u32) (carm_ref_msg_dma(host, idx) + sizeof(*ioc)); idx 656 drivers/block/sx8.c ioc->handle = cpu_to_le32(TAG_ENCODE(idx)); idx 665 drivers/block/sx8.c unsigned int idx = host->wait_q_prod % CARM_MAX_WAIT_Q; idx 670 drivers/block/sx8.c host->wait_q[idx] = q; idx 677 drivers/block/sx8.c unsigned int idx; idx 682 drivers/block/sx8.c idx = host->wait_q_cons % CARM_MAX_WAIT_Q; idx 685 drivers/block/sx8.c return host->wait_q[idx]; idx 1016 drivers/block/sx8.c unsigned int idx = host->resp_idx % RMSG_Q_LEN; idx 1019 drivers/block/sx8.c u32 status = le32_to_cpu(resp[idx].status); idx 1022 drivers/block/sx8.c VPRINTK("ending response on index %u\n", idx); idx 1023 drivers/block/sx8.c writel(idx << 3, mmio + CARM_RESP_IDX); idx 1029 drivers/block/sx8.c VPRINTK("handling msg response on index %u\n", idx); idx 1030 drivers/block/sx8.c carm_handle_resp(host, resp[idx].ret_handle, status); idx 1031 drivers/block/sx8.c resp[idx].status = cpu_to_le32(0xffffffff); idx 1036 drivers/block/sx8.c u8 *evt_type_ptr = (u8 *) &resp[idx]; idx 1040 drivers/block/sx8.c resp[idx].status = cpu_to_le32(0xffffffff); idx 1043 drivers/block/sx8.c idx = NEXT_RESP(idx); idx 97 drivers/bluetooth/bpa10x.c bool idx = usb_pipebulk(urb->pipe); idx 99 drivers/bluetooth/bpa10x.c data->rx_skb[idx] = h4_recv_buf(hdev, data->rx_skb[idx], idx 104 drivers/bluetooth/bpa10x.c if (IS_ERR(data->rx_skb[idx])) { idx 107 drivers/bluetooth/bpa10x.c data->rx_skb[idx] = NULL; idx 1394 drivers/bluetooth/btmrvl_sdio.c u8 dump_num = 0, idx, i, read_reg, doneflag = 0; idx 1408 drivers/bluetooth/btmrvl_sdio.c for (idx = 0; idx < ARRAY_SIZE(mem_type_mapping_tbl); idx++) { idx 1409 drivers/bluetooth/btmrvl_sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; idx 1437 drivers/bluetooth/btmrvl_sdio.c for (idx = 0; idx < dump_num; idx++) { idx 1438 drivers/bluetooth/btmrvl_sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; idx 1537 drivers/bluetooth/btmrvl_sdio.c for (idx = 0; idx < dump_num; idx++) { idx 1538 drivers/bluetooth/btmrvl_sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; idx 1556 drivers/bluetooth/btmrvl_sdio.c vfree(mem_type_mapping_tbl[idx].mem_ptr); idx 1557 drivers/bluetooth/btmrvl_sdio.c mem_type_mapping_tbl[idx].mem_ptr = NULL; idx 130 drivers/bluetooth/btqca.c int idx, length; idx 184 drivers/bluetooth/btqca.c idx = 0; idx 186 drivers/bluetooth/btqca.c while (idx < length) { idx 187 drivers/bluetooth/btqca.c tlv_nvm = (struct tlv_type_nvm *)(data + idx); idx 215 drivers/bluetooth/btqca.c idx += (sizeof(u16) + sizeof(u16) + 8 + tag_len); idx 127 drivers/bus/moxtet.c mdev->idx == new_dev->idx) idx 169 drivers/bus/moxtet.c if (dev->idx >= TURRIS_MOX_MAX_MODULES || dev->id > 0xf) idx 173 drivers/bus/moxtet.c dev->idx); idx 225 drivers/bus/moxtet.c dev->idx = val; idx 227 drivers/bus/moxtet.c if (dev->idx >= TURRIS_MOX_MAX_MODULES) { idx 229 drivers/bus/moxtet.c nc, dev->idx); idx 234 drivers/bus/moxtet.c dev->id = moxtet->modules[dev->idx]; idx 238 drivers/bus/moxtet.c dev->idx); idx 296 drivers/bus/moxtet.c dev->idx = i; idx 312 drivers/bus/moxtet.c static int moxtet_set_irq(struct moxtet *moxtet, int idx, int id, int nsame) idx 325 drivers/bus/moxtet.c pos->idx = idx; idx 411 drivers/bus/moxtet.c if (mdev->idx >= moxtet->count) idx 418 drivers/bus/moxtet.c return buf[mdev->idx + 1] >> 4; idx 428 drivers/bus/moxtet.c if (mdev->idx >= moxtet->count) idx 433 drivers/bus/moxtet.c moxtet->tx[moxtet->count - mdev->idx] = val; idx 449 drivers/bus/moxtet.c if (mdev->idx >= moxtet->count) idx 452 drivers/bus/moxtet.c return moxtet->tx[moxtet->count - mdev->idx]; idx 661 drivers/bus/moxtet.c id = moxtet->modules[pos->idx]; idx 663 drivers/bus/moxtet.c seq_printf(p, " moxtet-%s.%i#%i", mox_module_name(id), pos->idx, idx 687 drivers/bus/moxtet.c if (!(buf[pos[i].idx + 1] & BIT(4 + pos[i].bit))) idx 403 drivers/char/agp/i460-agp.c int i, start_offset, end_offset, idx, pg, num_entries; idx 429 drivers/char/agp/i460-agp.c for (idx = ((lp == start) ? start_offset : 0); idx 430 drivers/char/agp/i460-agp.c idx < ((lp == end) ? (end_offset + 1) : I460_KPAGES_PER_IOPAGE); idx 431 drivers/char/agp/i460-agp.c idx++) idx 433 drivers/char/agp/i460-agp.c if (test_bit(idx, lp->alloced_map)) idx 449 drivers/char/agp/i460-agp.c for (idx = ((lp == start) ? start_offset : 0); idx 450 drivers/char/agp/i460-agp.c idx < ((lp == end) ? (end_offset + 1) : I460_KPAGES_PER_IOPAGE); idx 451 drivers/char/agp/i460-agp.c idx++, i++) idx 454 drivers/char/agp/i460-agp.c __set_bit(idx, lp->alloced_map); idx 464 drivers/char/agp/i460-agp.c int i, pg, start_offset, end_offset, idx, num_entries; idx 478 drivers/char/agp/i460-agp.c for (idx = ((lp == start) ? start_offset : 0); idx 479 drivers/char/agp/i460-agp.c idx < ((lp == end) ? (end_offset + 1) : I460_KPAGES_PER_IOPAGE); idx 480 drivers/char/agp/i460-agp.c idx++, i++) idx 483 drivers/char/agp/i460-agp.c __clear_bit(idx, lp->alloced_map); idx 2797 drivers/char/ipmi/ipmi_msghandler.c struct attribute *attr, int idx) idx 61 drivers/char/ipmi/ipmi_si_mem_io.c int idx; idx 63 drivers/char/ipmi/ipmi_si_mem_io.c for (idx = 0; idx < num; idx++) idx 64 drivers/char/ipmi/ipmi_si_mem_io.c release_mem_region(addr + idx * io->regspacing, idx 79 drivers/char/ipmi/ipmi_si_mem_io.c int mapsize, idx; idx 119 drivers/char/ipmi/ipmi_si_mem_io.c for (idx = 0; idx < io->io_size; idx++) { idx 120 drivers/char/ipmi/ipmi_si_mem_io.c if (request_mem_region(addr + idx * io->regspacing, idx 123 drivers/char/ipmi/ipmi_si_mem_io.c mem_region_cleanup(io, idx); idx 54 drivers/char/ipmi/ipmi_si_port_io.c int idx; idx 57 drivers/char/ipmi/ipmi_si_port_io.c for (idx = 0; idx < io->io_size; idx++) idx 58 drivers/char/ipmi/ipmi_si_port_io.c release_region(addr + idx * io->regspacing, idx 66 drivers/char/ipmi/ipmi_si_port_io.c int idx; idx 100 drivers/char/ipmi/ipmi_si_port_io.c for (idx = 0; idx < io->io_size; idx++) { idx 101 drivers/char/ipmi/ipmi_si_port_io.c if (request_region(addr + idx * io->regspacing, idx 104 drivers/char/ipmi/ipmi_si_port_io.c while (idx--) idx 105 drivers/char/ipmi/ipmi_si_port_io.c release_region(addr + idx * io->regspacing, idx 1296 drivers/char/random.c unsigned int idx; idx 1300 drivers/char/random.c idx = READ_ONCE(f->reg_idx); idx 1301 drivers/char/random.c if (idx >= sizeof(struct pt_regs) / sizeof(__u32)) idx 1302 drivers/char/random.c idx = 0; idx 1303 drivers/char/random.c ptr += idx++; idx 1304 drivers/char/random.c WRITE_ONCE(f->reg_idx, idx); idx 50 drivers/clk/at91/pmc.c unsigned int idx = clkspec->args[1]; idx 55 drivers/clk/at91/pmc.c if (idx < pmc_data->ncore) idx 56 drivers/clk/at91/pmc.c return pmc_data->chws[idx]; idx 59 drivers/clk/at91/pmc.c if (idx < pmc_data->nsystem) idx 60 drivers/clk/at91/pmc.c return pmc_data->shws[idx]; idx 63 drivers/clk/at91/pmc.c if (idx < pmc_data->nperiph) idx 64 drivers/clk/at91/pmc.c return pmc_data->phws[idx]; idx 67 drivers/clk/at91/pmc.c if (idx < pmc_data->ngck) idx 68 drivers/clk/at91/pmc.c return pmc_data->ghws[idx]; idx 74 drivers/clk/at91/pmc.c pr_err("%s: invalid type (%u) or index (%u)\n", __func__, type, idx); idx 787 drivers/clk/bcm/clk-kona-setup.c unsigned int idx = clkspec->args[0]; idx 789 drivers/clk/bcm/clk-kona-setup.c if (idx >= ccu->clk_num) { idx 790 drivers/clk/bcm/clk-kona-setup.c pr_err("%s: invalid index %u\n", __func__, idx); idx 794 drivers/clk/bcm/clk-kona-setup.c return &ccu->kona_clks[idx].hw; idx 64 drivers/clk/clk-asm9260.c unsigned int idx; idx 71 drivers/clk/clk-asm9260.c unsigned int idx; idx 309 drivers/clk/clk-asm9260.c hws[dc->idx] = clk_hw_register_divider(NULL, dc->name, idx 319 drivers/clk/clk-asm9260.c hws[gd->idx] = clk_hw_register_gate(NULL, gd->name, idx 197 drivers/clk/clk-ast2600.c static u32 get_bit(u8 idx) idx 199 drivers/clk/clk-ast2600.c return BIT(idx % 32); idx 517 drivers/clk/clk-axm5516.c unsigned int idx = clkspec->args[0]; idx 519 drivers/clk/clk-axm5516.c if (idx >= ARRAY_SIZE(axmclk_clocks)) { idx 520 drivers/clk/clk-axm5516.c pr_err("%s: invalid index %u\n", __func__, idx); idx 524 drivers/clk/clk-axm5516.c return &axmclk_clocks[idx]->hw; idx 69 drivers/clk/clk-cdce706.c unsigned idx; idx 169 drivers/clk/clk-cdce706.c __func__, hwd->idx, hwd->mux, hwd->mul, hwd->div); idx 204 drivers/clk/clk-cdce706.c __func__, hwd->idx, mul, div); idx 220 drivers/clk/clk-cdce706.c __func__, hwd->idx, mul, div); idx 223 drivers/clk/clk-cdce706.c CDCE706_PLL_HI(hwd->idx), idx 232 drivers/clk/clk-cdce706.c CDCE706_PLL_M_LOW(hwd->idx), idx 238 drivers/clk/clk-cdce706.c CDCE706_PLL_N_LOW(hwd->idx), idx 245 drivers/clk/clk-cdce706.c CDCE706_PLL_FVCO_MASK(hwd->idx), idx 247 drivers/clk/clk-cdce706.c CDCE706_PLL_FVCO_MASK(hwd->idx) : 0); idx 265 drivers/clk/clk-cdce706.c CDCE706_DIVIDER_PLL(hwd->idx), idx 266 drivers/clk/clk-cdce706.c CDCE706_DIVIDER_PLL_MASK(hwd->idx), idx 267 drivers/clk/clk-cdce706.c index << CDCE706_DIVIDER_PLL_SHIFT(hwd->idx)); idx 284 drivers/clk/clk-cdce706.c __func__, hwd->idx, hwd->div); idx 353 drivers/clk/clk-cdce706.c __func__, hwd->idx, div); idx 365 drivers/clk/clk-cdce706.c __func__, hwd->idx, hwd->div); idx 368 drivers/clk/clk-cdce706.c CDCE706_DIVIDER(hwd->idx), idx 385 drivers/clk/clk-cdce706.c return cdce706_reg_update(hwd->dev_data, CDCE706_CLKOUT(hwd->idx), idx 394 drivers/clk/clk-cdce706.c cdce706_reg_update(hwd->dev_data, CDCE706_CLKOUT(hwd->idx), idx 406 drivers/clk/clk-cdce706.c CDCE706_CLKOUT(hwd->idx), idx 457 drivers/clk/clk-cdce706.c hw->idx = i; idx 620 drivers/clk/clk-cdce706.c unsigned int idx = clkspec->args[0]; idx 622 drivers/clk/clk-cdce706.c if (idx >= ARRAY_SIZE(cdce->clkout)) { idx 623 drivers/clk/clk-cdce706.c pr_err("%s: invalid index %u\n", __func__, idx); idx 627 drivers/clk/clk-cdce706.c return &cdce->clkout[idx].hw; idx 596 drivers/clk/clk-cdce925.c unsigned int idx = clkspec->args[0]; idx 598 drivers/clk/clk-cdce925.c if (idx >= ARRAY_SIZE(data->clk)) { idx 599 drivers/clk/clk-cdce925.c pr_err("%s: invalid index %u\n", __func__, idx); idx 603 drivers/clk/clk-cdce925.c return &data->clk[idx].hw; idx 220 drivers/clk/clk-lochnagar.c unsigned int idx = clkspec->args[0]; idx 222 drivers/clk/clk-lochnagar.c if (idx >= ARRAY_SIZE(priv->lclks)) { idx 223 drivers/clk/clk-lochnagar.c dev_err(priv->dev, "Invalid index %u\n", idx); idx 227 drivers/clk/clk-lochnagar.c return &priv->lclks[idx].hw; idx 151 drivers/clk/clk-max77686.c unsigned int idx = clkspec->args[0]; idx 153 drivers/clk/clk-max77686.c if (idx >= drv_data->num_clks) { idx 154 drivers/clk/clk-max77686.c pr_err("%s: invalid index %u\n", __func__, idx); idx 158 drivers/clk/clk-max77686.c return &drv_data->max_clk_data[idx].hw; idx 252 drivers/clk/clk-max9485.c unsigned int idx = clkspec->args[0]; idx 254 drivers/clk/clk-max9485.c return &drvdata->hw[idx].hw; idx 786 drivers/clk/clk-qoriq.c static int mux_set_parent(struct clk_hw *hw, u8 idx) idx 791 drivers/clk/clk-qoriq.c if (idx >= hwc->num_parents) idx 794 drivers/clk/clk-qoriq.c clksel = hwc->parent_to_clksel[idx]; idx 832 drivers/clk/clk-qoriq.c int idx) idx 836 drivers/clk/clk-qoriq.c if (!(hwc->info->clksel[idx].flags & CLKSEL_VALID)) idx 839 drivers/clk/clk-qoriq.c pll = hwc->info->clksel[idx].pll; idx 840 drivers/clk/clk-qoriq.c div = hwc->info->clksel[idx].div; idx 851 drivers/clk/clk-qoriq.c const char *fmt, int idx) idx 860 drivers/clk/clk-qoriq.c snprintf(name, sizeof(name), fmt, idx); idx 906 drivers/clk/clk-qoriq.c static struct clk * __init create_one_cmux(struct clockgen *cg, int idx) idx 919 drivers/clk/clk-qoriq.c hwc->reg = cg->regs + 0x70000 + 0x20 * idx; idx 921 drivers/clk/clk-qoriq.c hwc->reg = cg->regs + 0x20 * idx; idx 923 drivers/clk/clk-qoriq.c hwc->info = cg->info.cmux_groups[cg->info.cmux_to_group[idx]]; idx 951 drivers/clk/clk-qoriq.c pct80_rate, "cg-cmux%d", idx); idx 954 drivers/clk/clk-qoriq.c static struct clk * __init create_one_hwaccel(struct clockgen *cg, int idx) idx 962 drivers/clk/clk-qoriq.c hwc->reg = cg->regs + 0x20 * idx + 0x10; idx 963 drivers/clk/clk-qoriq.c hwc->info = cg->info.hwaccel[idx]; idx 966 drivers/clk/clk-qoriq.c "cg-hwaccel%d", idx); idx 1012 drivers/clk/clk-qoriq.c int idx, rc; idx 1019 drivers/clk/clk-qoriq.c idx = (res.start & 0xf0) >> 5; idx 1020 drivers/clk/clk-qoriq.c clk = clockgen.cmux[idx]; idx 1068 drivers/clk/clk-qoriq.c static struct clk __init *input_clock_by_index(const char *name, int idx) idx 1141 drivers/clk/clk-qoriq.c static void __init create_one_pll(struct clockgen *cg, int idx) idx 1145 drivers/clk/clk-qoriq.c struct clockgen_pll *pll = &cg->pll[idx]; idx 1149 drivers/clk/clk-qoriq.c if (!(cg->info.pll_mask & (1 << idx))) idx 1152 drivers/clk/clk-qoriq.c if (cg->coreclk && idx != PLATFORM_PLL) { idx 1160 drivers/clk/clk-qoriq.c switch (idx) { idx 1177 drivers/clk/clk-qoriq.c WARN_ONCE(1, "index %d\n", idx); idx 1181 drivers/clk/clk-qoriq.c if (idx == PLATFORM_PLL) idx 1184 drivers/clk/clk-qoriq.c reg = cg->regs + 0x800 + 0x20 * (idx - 1); idx 1197 drivers/clk/clk-qoriq.c ((cg->info.flags & CG_PLL_8BIT) && idx != PLATFORM_PLL)) idx 1210 drivers/clk/clk-qoriq.c if (idx != PLATFORM_PLL && i >= 4) idx 1214 drivers/clk/clk-qoriq.c "cg-pll%d-div%d", idx, i + 1); idx 1241 drivers/clk/clk-qoriq.c static void __init legacy_pll_init(struct device_node *np, int idx) idx 1250 drivers/clk/clk-qoriq.c pll = &clockgen.pll[idx]; idx 1300 drivers/clk/clk-qoriq.c int idx; idx 1312 drivers/clk/clk-qoriq.c idx = (res.start & 0xf0) >> 5; idx 1313 drivers/clk/clk-qoriq.c legacy_pll_init(np, CGA_PLL1 + idx); idx 1322 drivers/clk/clk-qoriq.c u32 type, idx; idx 1330 drivers/clk/clk-qoriq.c idx = clkspec->args[1]; idx 1334 drivers/clk/clk-qoriq.c if (idx != 0) idx 1339 drivers/clk/clk-qoriq.c if (idx >= ARRAY_SIZE(cg->cmux)) idx 1341 drivers/clk/clk-qoriq.c clk = cg->cmux[idx]; idx 1344 drivers/clk/clk-qoriq.c if (idx >= ARRAY_SIZE(cg->hwaccel)) idx 1346 drivers/clk/clk-qoriq.c clk = cg->hwaccel[idx]; idx 1349 drivers/clk/clk-qoriq.c if (idx >= ARRAY_SIZE(cg->fman)) idx 1351 drivers/clk/clk-qoriq.c clk = cg->fman[idx]; idx 1355 drivers/clk/clk-qoriq.c if (idx >= ARRAY_SIZE(pll->div)) idx 1357 drivers/clk/clk-qoriq.c clk = pll->div[idx].clk; idx 1360 drivers/clk/clk-qoriq.c if (idx != 0) idx 1375 drivers/clk/clk-qoriq.c pr_err("%s: Bad phandle args %u %u\n", __func__, type, idx); idx 81 drivers/clk/clk-rk808.c unsigned int idx = clkspec->args[0]; idx 83 drivers/clk/clk-rk808.c if (idx >= 2) { idx 84 drivers/clk/clk-rk808.c pr_err("%s: invalid index %u\n", __func__, idx); idx 88 drivers/clk/clk-rk808.c return idx ? &rk808_clkout->clkout2_hw : &rk808_clkout->clkout1_hw; idx 123 drivers/clk/clk-scmi.c int idx, count, err; idx 147 drivers/clk/clk-scmi.c for (idx = 0; idx < count; idx++) { idx 154 drivers/clk/clk-scmi.c sclk->info = handle->clk_ops->info_get(handle, idx); idx 156 drivers/clk/clk-scmi.c dev_dbg(dev, "invalid clock info for idx %d\n", idx); idx 160 drivers/clk/clk-scmi.c sclk->id = idx; idx 165 drivers/clk/clk-scmi.c dev_err(dev, "failed to register clock %d\n", idx); idx 167 drivers/clk/clk-scmi.c hws[idx] = NULL; idx 170 drivers/clk/clk-scmi.c hws[idx] = &sclk->hw; idx 65 drivers/clk/clk-scpi.c int idx; idx 69 drivers/clk/clk-scpi.c for (idx = 0; idx < clk->info->count; idx++, opp++) { idx 86 drivers/clk/clk-scpi.c int idx = clk->scpi_ops->dvfs_get_idx(clk->id); idx 89 drivers/clk/clk-scpi.c if (idx < 0) idx 92 drivers/clk/clk-scpi.c opp = clk->info->opps + idx; idx 106 drivers/clk/clk-scpi.c int idx, max_opp = clk->info->count; idx 109 drivers/clk/clk-scpi.c for (idx = 0; idx < max_opp; idx++, opp++) idx 111 drivers/clk/clk-scpi.c return idx; idx 180 drivers/clk/clk-scpi.c unsigned int idx = clkspec->args[0], count; idx 184 drivers/clk/clk-scpi.c if (idx == sclk->id) idx 194 drivers/clk/clk-scpi.c int idx, count, err; idx 213 drivers/clk/clk-scpi.c for (idx = 0; idx < count; idx++) { idx 223 drivers/clk/clk-scpi.c idx, &name)) { idx 229 drivers/clk/clk-scpi.c idx, &val)) { idx 243 drivers/clk/clk-scpi.c clk_data->clk[idx] = sclk; idx 789 drivers/clk/clk-si5341.c unsigned int idx = clkspec->args[1]; idx 794 drivers/clk/clk-si5341.c if (idx >= data->num_outputs) { idx 796 drivers/clk/clk-si5341.c "invalid output index %u\n", idx); idx 799 drivers/clk/clk-si5341.c return &data->clk[idx].hw; idx 801 drivers/clk/clk-si5341.c if (idx >= data->num_synth) { idx 803 drivers/clk/clk-si5341.c "invalid synthesizer index %u\n", idx); idx 806 drivers/clk/clk-si5341.c return &data->synth[idx].hw; idx 808 drivers/clk/clk-si5341.c if (idx > 0) { idx 810 drivers/clk/clk-si5341.c "invalid PLL index %u\n", idx); idx 1341 drivers/clk/clk-si5351.c unsigned int idx = clkspec->args[0]; idx 1343 drivers/clk/clk-si5351.c if (idx >= drvdata->num_clkout) { idx 1344 drivers/clk/clk-si5351.c pr_err("%s: invalid index %u\n", __func__, idx); idx 1348 drivers/clk/clk-si5351.c return &drvdata->clkout[idx].hw; idx 528 drivers/clk/clk-stm32f4.c int idx; idx 1155 drivers/clk/clk-stm32f4.c int idx; idx 1780 drivers/clk/clk-stm32f4.c if (post_div->idx != NO_IDX) idx 1781 drivers/clk/clk-stm32f4.c clks[post_div->idx] = hw; idx 1814 drivers/clk/clk-stm32f4.c int idx; idx 1819 drivers/clk/clk-stm32f4.c idx = stm32f4_rcc_lookup_clk_idx(0, secondary); idx 1821 drivers/clk/clk-stm32f4.c if (idx < 0) idx 1824 drivers/clk/clk-stm32f4.c clks[idx] = clk_hw_register_gate( idx 1828 drivers/clk/clk-stm32f4.c if (IS_ERR(clks[idx])) { idx 1886 drivers/clk/clk-stm32f4.c if (aux_clk->idx != NO_IDX) idx 1887 drivers/clk/clk-stm32f4.c clks[aux_clk->idx] = hw; idx 1317 drivers/clk/clk-stm32h7.c int idx = n * 3 + odf; idx 1322 drivers/clk/clk-stm32h7.c hws[ODF_BANK + idx] = clk_hw_register_composite(NULL, idx 22 drivers/clk/clk-tango4.c static void __init make_pll(int idx, const char *parent, void __iomem *base) idx 27 drivers/clk/clk-tango4.c sprintf(name, "pll%d", idx); idx 28 drivers/clk/clk-tango4.c val = readl(base + idx * 8); idx 36 drivers/clk/clk-tango4.c static void __init make_cd(int idx, void __iomem *base) idx 41 drivers/clk/clk-tango4.c sprintf(name, "cd%d", idx); idx 42 drivers/clk/clk-tango4.c val = readl(base + idx * 8); idx 73 drivers/clk/clk-versaclock5.c #define VC5_OUT_DIV_CONTROL(idx) (0x21 + ((idx) * 0x10)) idx 80 drivers/clk/clk-versaclock5.c #define VC5_OUT_DIV_FRAC(idx, n) (0x22 + ((idx) * 0x10) + (n)) idx 83 drivers/clk/clk-versaclock5.c #define VC5_OUT_DIV_STEP_SPREAD(idx, n) (0x26 + ((idx) * 0x10) + (n)) idx 84 drivers/clk/clk-versaclock5.c #define VC5_OUT_DIV_SPREAD_MOD(idx, n) (0x29 + ((idx) * 0x10) + (n)) idx 85 drivers/clk/clk-versaclock5.c #define VC5_OUT_DIV_SKEW_INT(idx, n) (0x2b + ((idx) * 0x10) + (n)) idx 86 drivers/clk/clk-versaclock5.c #define VC5_OUT_DIV_INT(idx, n) (0x2d + ((idx) * 0x10) + (n)) idx 87 drivers/clk/clk-versaclock5.c #define VC5_OUT_DIV_SKEW_FRAC(idx) (0x2f + ((idx) * 0x10)) idx 91 drivers/clk/clk-versaclock5.c #define VC5_CLK_OUTPUT_CFG(idx, n) (0x60 + ((idx) * 0x2) + (n)) idx 668 drivers/clk/clk-versaclock5.c unsigned int idx = clkspec->args[0]; idx 670 drivers/clk/clk-versaclock5.c if (idx >= vc5->chip_info->clk_out_cnt) idx 673 drivers/clk/clk-versaclock5.c return &vc5->clk_out[idx].hw; idx 699 drivers/clk/clk-versaclock5.c unsigned int n, idx = 0; idx 814 drivers/clk/clk-versaclock5.c idx = vc5_map_index_to_output(vc5->chip_info->model, n); idx 816 drivers/clk/clk-versaclock5.c init.name = vc5_fod_names[idx]; idx 821 drivers/clk/clk-versaclock5.c vc5->clk_fod[n].num = idx; idx 839 drivers/clk/clk-versaclock5.c vc5->clk_out[0].num = idx; idx 851 drivers/clk/clk-versaclock5.c idx = vc5_map_index_to_output(vc5->chip_info->model, n - 1); idx 852 drivers/clk/clk-versaclock5.c parent_names[0] = vc5_fod_names[idx]; idx 859 drivers/clk/clk-versaclock5.c init.name = vc5_clk_out_names[idx + 1]; idx 864 drivers/clk/clk-versaclock5.c vc5->clk_out[n].num = idx; idx 4274 drivers/clk/clk.c unsigned int idx = clkspec->args[0]; idx 4276 drivers/clk/clk.c if (idx >= clk_data->clk_num) { idx 4277 drivers/clk/clk.c pr_err("%s: invalid clock index %u\n", __func__, idx); idx 4281 drivers/clk/clk.c return clk_data->clks[idx]; idx 4289 drivers/clk/clk.c unsigned int idx = clkspec->args[0]; idx 4291 drivers/clk/clk.c if (idx >= hw_data->num) { idx 4292 drivers/clk/clk.c pr_err("%s: invalid index %u\n", __func__, idx); idx 4296 drivers/clk/clk.c return hw_data->hws[idx]; idx 4827 drivers/clk/clk.c uint32_t idx; idx 4832 drivers/clk/clk.c of_property_for_each_u32(np, "clock-critical", prop, cur, idx) idx 4833 drivers/clk/clk.c if (index == idx) idx 114 drivers/clk/hisilicon/clk-hi3660-stub.c unsigned int idx = clkspec->args[0]; idx 116 drivers/clk/hisilicon/clk-hi3660-stub.c if (idx >= HI3660_CLK_STUB_NUM) { idx 117 drivers/clk/hisilicon/clk-hi3660-stub.c pr_err("%s: invalid index %u\n", __func__, idx); idx 121 drivers/clk/hisilicon/clk-hi3660-stub.c return &hi3660_stub_clks[idx].hw; idx 26 drivers/clk/imx/clk-pfd.c u8 idx; idx 39 drivers/clk/imx/clk-pfd.c writel_relaxed(1 << ((pfd->idx + 1) * 8 - 1), pfd->reg + CLR); idx 48 drivers/clk/imx/clk-pfd.c writel_relaxed(1 << ((pfd->idx + 1) * 8 - 1), pfd->reg + SET); idx 56 drivers/clk/imx/clk-pfd.c u8 frac = (readl_relaxed(pfd->reg) >> (pfd->idx * 8)) & 0x3f; idx 99 drivers/clk/imx/clk-pfd.c writel_relaxed(0x3f << (pfd->idx * 8), pfd->reg + CLR); idx 100 drivers/clk/imx/clk-pfd.c writel_relaxed(frac << (pfd->idx * 8), pfd->reg + SET); idx 109 drivers/clk/imx/clk-pfd.c if (readl_relaxed(pfd->reg) & (1 << ((pfd->idx + 1) * 8 - 1))) idx 125 drivers/clk/imx/clk-pfd.c void __iomem *reg, u8 idx) idx 137 drivers/clk/imx/clk-pfd.c pfd->idx = idx; idx 170 drivers/clk/imx/clk-pfdv2.c void __iomem *reg, u8 idx) idx 177 drivers/clk/imx/clk-pfdv2.c WARN_ON(idx > 3); idx 184 drivers/clk/imx/clk-pfdv2.c pfd->gate_bit = (idx + 1) * 8 - 1; idx 186 drivers/clk/imx/clk-pfdv2.c pfd->frac_off = idx * 8; idx 64 drivers/clk/imx/clk.h #define imx_clk_pfd(name, parent_name, reg, idx) \ idx 65 drivers/clk/imx/clk.h to_clk(imx_clk_hw_pfd(name, parent_name, reg, idx)) idx 171 drivers/clk/imx/clk.h void __iomem *reg, u8 idx); idx 174 drivers/clk/imx/clk.h void __iomem *reg, u8 idx); idx 82 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 151 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 292 drivers/clk/ingenic/cgu.c u8 i, hw_idx, idx = 0; idx 294 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 307 drivers/clk/ingenic/cgu.c idx++; idx 311 drivers/clk/ingenic/cgu.c return idx; idx 314 drivers/clk/ingenic/cgu.c static int ingenic_clk_set_parent(struct clk_hw *hw, u8 idx) idx 323 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 337 drivers/clk/ingenic/cgu.c if (curr_idx == idx) idx 343 drivers/clk/ingenic/cgu.c BUG_ON(curr_idx != idx); idx 360 drivers/clk/ingenic/cgu.c return idx ? -EINVAL : 0; idx 372 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 446 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 469 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 528 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 550 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 568 drivers/clk/ingenic/cgu.c clk_info = &cgu->clock_info[ingenic_clk->idx]; idx 596 drivers/clk/ingenic/cgu.c static int ingenic_register_clock(struct ingenic_cgu *cgu, unsigned idx) idx 598 drivers/clk/ingenic/cgu.c const struct ingenic_cgu_clk_info *clk_info = &cgu->clock_info[idx]; idx 621 drivers/clk/ingenic/cgu.c cgu->clocks.clks[idx] = clk; idx 639 drivers/clk/ingenic/cgu.c ingenic_clk->idx = idx; idx 733 drivers/clk/ingenic/cgu.c cgu->clocks.clks[idx] = clk; idx 199 drivers/clk/ingenic/cgu.h unsigned idx; idx 99 drivers/clk/ingenic/jz4780-cgu.c static int jz4780_otg_phy_set_parent(struct clk_hw *hw, u8 idx) idx 104 drivers/clk/ingenic/jz4780-cgu.c if (idx > 0) idx 44 drivers/clk/ingenic/tcu.c unsigned int idx; idx 134 drivers/clk/ingenic/tcu.c WARN_ONCE(ret < 0, "Unable to read TCSR %d", tcu_clk->idx); idx 139 drivers/clk/ingenic/tcu.c static int ingenic_tcu_set_parent(struct clk_hw *hw, u8 idx) idx 149 drivers/clk/ingenic/tcu.c TCU_TCSR_PARENT_CLOCK_MASK, BIT(idx)); idx 150 drivers/clk/ingenic/tcu.c WARN_ONCE(ret < 0, "Unable to update TCSR %d", tcu_clk->idx); idx 167 drivers/clk/ingenic/tcu.c WARN_ONCE(ret < 0, "Unable to read TCSR %d", tcu_clk->idx); idx 213 drivers/clk/ingenic/tcu.c WARN_ONCE(ret < 0, "Unable to update TCSR %d", tcu_clk->idx); idx 270 drivers/clk/ingenic/tcu.c unsigned int idx, enum tcu_clk_parent parent, idx 282 drivers/clk/ingenic/tcu.c tcu_clk->idx = idx; idx 297 drivers/clk/ingenic/tcu.c clocks->hws[idx] = &tcu_clk->hw; idx 200 drivers/clk/mvebu/cp110-system-controller.c unsigned int idx = clkspec->args[1]; idx 203 drivers/clk/mvebu/cp110-system-controller.c if (idx >= CP110_MAX_CORE_CLOCKS) idx 205 drivers/clk/mvebu/cp110-system-controller.c return clk_data->hws[idx]; idx 207 drivers/clk/mvebu/cp110-system-controller.c if (idx >= CP110_MAX_GATABLE_CLOCKS) idx 209 drivers/clk/mvebu/cp110-system-controller.c return clk_data->hws[CP110_MAX_CORE_CLOCKS + idx]; idx 26 drivers/clk/mxs/clk-ref.c u8 idx; idx 35 drivers/clk/mxs/clk-ref.c writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + CLR); idx 44 drivers/clk/mxs/clk-ref.c writel_relaxed(1 << ((ref->idx + 1) * 8 - 1), ref->reg + SET); idx 52 drivers/clk/mxs/clk-ref.c u8 frac = (readl_relaxed(ref->reg) >> (ref->idx * 8)) & 0x3f; idx 90 drivers/clk/mxs/clk-ref.c u8 frac, shift = ref->idx * 8; idx 122 drivers/clk/mxs/clk-ref.c void __iomem *reg, u8 idx) idx 139 drivers/clk/mxs/clk-ref.c ref->idx = idx; idx 25 drivers/clk/mxs/clk.h void __iomem *reg, u8 idx); idx 558 drivers/clk/qcom/clk-rpm.c unsigned int idx = clkspec->args[0]; idx 560 drivers/clk/qcom/clk-rpm.c if (idx >= rcc->num_clks) { idx 561 drivers/clk/qcom/clk-rpm.c pr_err("%s: invalid index %u\n", __func__, idx); idx 565 drivers/clk/qcom/clk-rpm.c return rcc->clks[idx] ? &rcc->clks[idx]->hw : ERR_PTR(-ENOENT); idx 398 drivers/clk/qcom/clk-rpmh.c unsigned int idx = clkspec->args[0]; idx 400 drivers/clk/qcom/clk-rpmh.c if (idx >= rpmh->num_clks) { idx 401 drivers/clk/qcom/clk-rpmh.c pr_err("%s: invalid index %u\n", __func__, idx); idx 405 drivers/clk/qcom/clk-rpmh.c return rpmh->clks[idx]; idx 732 drivers/clk/qcom/clk-smd-rpm.c unsigned int idx = clkspec->args[0]; idx 734 drivers/clk/qcom/clk-smd-rpm.c if (idx >= rcc->num_clks) { idx 735 drivers/clk/qcom/clk-smd-rpm.c pr_err("%s: invalid index %u\n", __func__, idx); idx 739 drivers/clk/qcom/clk-smd-rpm.c return rcc->clks[idx] ? &rcc->clks[idx]->hw : ERR_PTR(-ENOENT); idx 187 drivers/clk/qcom/clk-spmi-pmic-div.c int idx = clkspec->args[0] - 1; /* Start at 1 instead of 0 */ idx 189 drivers/clk/qcom/clk-spmi-pmic-div.c if (idx < 0 || idx >= cc->nclks) { idx 195 drivers/clk/qcom/clk-spmi-pmic-div.c return &cc->clks[idx].hw; idx 217 drivers/clk/qcom/common.c unsigned int idx = clkspec->args[0]; idx 219 drivers/clk/qcom/common.c if (idx >= cc->num_rclks) { idx 220 drivers/clk/qcom/common.c pr_err("%s: invalid index %u\n", __func__, idx); idx 224 drivers/clk/qcom/common.c return cc->rclks[idx] ? &cc->rclks[idx]->hw : ERR_PTR(-ENOENT); idx 272 drivers/clk/qcom/krait-cc.c unsigned int idx = clkspec->args[0]; idx 275 drivers/clk/qcom/krait-cc.c if (idx >= 5) { idx 276 drivers/clk/qcom/krait-cc.c pr_err("%s: invalid clock index %d\n", __func__, idx); idx 280 drivers/clk/qcom/krait-cc.c return clks[idx] ? : ERR_PTR(-ENODEV); idx 257 drivers/clk/renesas/renesas-cpg-mssr.c unsigned int idx; idx 276 drivers/clk/renesas/renesas-cpg-mssr.c idx = MOD_CLK_PACK_10(clkidx); idx 279 drivers/clk/renesas/renesas-cpg-mssr.c idx = MOD_CLK_PACK(clkidx); idx 282 drivers/clk/renesas/renesas-cpg-mssr.c if (range_check < 0 || idx >= priv->num_mod_clks) { idx 287 drivers/clk/renesas/renesas-cpg-mssr.c clk = priv->clks[priv->num_core_clks + idx]; idx 641 drivers/clk/renesas/renesas-cpg-mssr.c unsigned int idx = MOD_CLK_PACK(unpacked); idx 643 drivers/clk/renesas/renesas-cpg-mssr.c if (unpacked % 100 > 31 || idx >= rcdev->nr_resets) { idx 648 drivers/clk/renesas/renesas-cpg-mssr.c return idx; idx 418 drivers/clk/rockchip/clk.c int idx; idx 420 drivers/clk/rockchip/clk.c for (idx = 0; idx < nr_pll; idx++, list++) { idx 443 drivers/clk/rockchip/clk.c unsigned int idx; idx 446 drivers/clk/rockchip/clk.c for (idx = 0; idx < nr_clk; idx++, list++) { idx 102 drivers/clk/samsung/clk.c unsigned int idx, ret; idx 104 drivers/clk/samsung/clk.c for (idx = 0; idx < nr_clk; idx++, list++) { idx 107 drivers/clk/samsung/clk.c idx); idx 132 drivers/clk/samsung/clk.c unsigned int idx, ret; idx 134 drivers/clk/samsung/clk.c for (idx = 0; idx < nr_clk; idx++, list++) { idx 161 drivers/clk/samsung/clk.c unsigned int idx; idx 163 drivers/clk/samsung/clk.c for (idx = 0; idx < nr_clk; idx++, list++) { idx 182 drivers/clk/samsung/clk.c unsigned int idx; idx 184 drivers/clk/samsung/clk.c for (idx = 0; idx < nr_clk; idx++, list++) { idx 205 drivers/clk/samsung/clk.c unsigned int idx; idx 207 drivers/clk/samsung/clk.c for (idx = 0; idx < nr_clk; idx++, list++) { idx 235 drivers/clk/samsung/clk.c unsigned int idx; idx 237 drivers/clk/samsung/clk.c for (idx = 0; idx < nr_clk; idx++, list++) { idx 200 drivers/clk/st/clk-flexgen.c void __iomem *reg, spinlock_t *lock, u32 idx, idx 218 drivers/clk/st/clk-flexgen.c xbar_reg = reg + 0x18 + (idx & ~0x3); idx 219 drivers/clk/st/clk-flexgen.c xbar_shift = (idx % 4) * 0x8; idx 220 drivers/clk/st/clk-flexgen.c fdiv_reg = reg + 0x164 + idx * 4; idx 237 drivers/clk/st/clk-flexgen.c fgxbar->pdiv.reg = reg + 0x58 + idx * 4; idx 41 drivers/clk/sunxi/clk-sun8i-bus-gates.c int idx = of_property_match_string(node, "clock-names", idx 43 drivers/clk/sunxi/clk-sun8i-bus-gates.c if (idx < 0) idx 46 drivers/clk/sunxi/clk-sun8i-bus-gates.c parents[i] = of_clk_get_parent_name(node, idx); idx 106 drivers/clk/ti/composite.c static inline struct clk_hw *_get_hw(struct clk_hw_omap_comp *clk, int idx) idx 111 drivers/clk/ti/composite.c if (!clk->comp_clks[idx]) idx 114 drivers/clk/ti/composite.c return clk->comp_clks[idx]->hw; idx 65 drivers/clk/uniphier/clk-uniphier-core.c clk_num = max(clk_num, p->idx + 1); idx 82 drivers/clk/uniphier/clk-uniphier-core.c dev_dbg(dev, "register %s (index=%d)\n", p->name, p->idx); idx 87 drivers/clk/uniphier/clk-uniphier-core.c if (p->idx >= 0) idx 88 drivers/clk/uniphier/clk-uniphier-core.c hw_data->hws[p->idx] = hw; idx 25 drivers/clk/uniphier/clk-uniphier-mio.c .idx = -1, \ idx 63 drivers/clk/uniphier/clk-uniphier-mio.c #define UNIPHIER_MIO_CLK_USB2(idx, ch) \ idx 64 drivers/clk/uniphier/clk-uniphier-mio.c UNIPHIER_CLK_GATE("usb2" #ch, (idx), "usb2", 0x20 + 0x200 * (ch), 28) idx 66 drivers/clk/uniphier/clk-uniphier-mio.c #define UNIPHIER_MIO_CLK_USB2_PHY(idx, ch) \ idx 67 drivers/clk/uniphier/clk-uniphier-mio.c UNIPHIER_CLK_GATE("usb2" #ch "-phy", (idx), "usb2", 0x20 + 0x200 * (ch), 29) idx 9 drivers/clk/uniphier/clk-uniphier-peri.c #define UNIPHIER_PERI_CLK_UART(idx, ch) \ idx 10 drivers/clk/uniphier/clk-uniphier-peri.c UNIPHIER_CLK_GATE("uart" #ch, (idx), "uart", 0x24, 19 + (ch)) idx 15 drivers/clk/uniphier/clk-uniphier-peri.c #define UNIPHIER_PERI_CLK_I2C(idx, ch) \ idx 16 drivers/clk/uniphier/clk-uniphier-peri.c UNIPHIER_CLK_GATE("i2c" #ch, (idx), "i2c-common", 0x24, 5 + (ch)) idx 18 drivers/clk/uniphier/clk-uniphier-peri.c #define UNIPHIER_PERI_CLK_FI2C(idx, ch) \ idx 19 drivers/clk/uniphier/clk-uniphier-peri.c UNIPHIER_CLK_GATE("i2c" #ch, (idx), "i2c", 0x24, 24 + (ch)) idx 21 drivers/clk/uniphier/clk-uniphier-peri.c #define UNIPHIER_PERI_CLK_SCSSI(idx, ch) \ idx 22 drivers/clk/uniphier/clk-uniphier-peri.c UNIPHIER_CLK_GATE("scssi" #ch, (idx), "spi", 0x20, 17 + (ch)) idx 24 drivers/clk/uniphier/clk-uniphier-peri.c #define UNIPHIER_PERI_CLK_MCSSI(idx) \ idx 25 drivers/clk/uniphier/clk-uniphier-peri.c UNIPHIER_CLK_GATE("mcssi", (idx), "spi", 0x24, 14) idx 23 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD4_SYS_CLK_NAND(idx) \ idx 25 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x2104, 2) idx 27 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_PRO5_SYS_CLK_NAND(idx) \ idx 29 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x2104, 2) idx 31 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_NAND(idx) \ idx 33 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("nand", (idx), "nand-50m", 0x210c, 0) idx 35 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_SYS_CLK_NAND_4X(idx) \ idx 36 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_FACTOR("nand-4x", (idx), "nand", 4, 1) idx 38 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_EMMC(idx) \ idx 39 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("emmc", (idx), NULL, 0x210c, 2) idx 41 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD4_SYS_CLK_STDMAC(idx) \ idx 42 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("stdmac", (idx), NULL, 0x2104, 10) idx 44 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_STDMAC(idx) \ idx 45 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("stdmac", (idx), NULL, 0x210c, 8) idx 47 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_HSC(idx) \ idx 48 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("hsc", (idx), NULL, 0x210c, 9) idx 50 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_PRO4_SYS_CLK_GIO(idx) \ idx 51 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("gio", (idx), NULL, 0x2104, 6) idx 53 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_PRO4_SYS_CLK_USB3(idx, ch) \ idx 54 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("usb3" #ch, (idx), NULL, 0x2104, 16 + (ch)) idx 56 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_PRO4_SYS_CLK_AIO(idx) \ idx 58 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("aio", (idx), "aio-io200m", 0x2104, 13) idx 60 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_PRO5_SYS_CLK_AIO(idx) \ idx 62 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("aio", (idx), "aio-io200m", 0x2104, 13) idx 64 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_AIO(idx) \ idx 66 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("aio", (idx), "aio-io200m", 0x2108, 0) idx 68 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_EVEA(idx) \ idx 70 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("evea", (idx), "evea-io100m", 0x2108, 1) idx 72 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_EXIV(idx) \ idx 74 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("exiv", (idx), "exiv-io200m", 0x2110, 2) idx 76 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_PRO4_SYS_CLK_ETHER(idx) \ idx 77 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("ether", (idx), NULL, 0x2104, 12) idx 79 drivers/clk/uniphier/clk-uniphier-sys.c #define UNIPHIER_LD11_SYS_CLK_ETHER(idx) \ idx 80 drivers/clk/uniphier/clk-uniphier-sys.c UNIPHIER_CLK_GATE("ether", (idx), NULL, 0x210c, 6) idx 59 drivers/clk/uniphier/clk-uniphier.h int idx; idx 74 drivers/clk/uniphier/clk-uniphier.h .idx = (_idx), \ idx 87 drivers/clk/uniphier/clk-uniphier.h .idx = (_idx), \ idx 99 drivers/clk/uniphier/clk-uniphier.h .idx = (_idx), \ idx 69 drivers/clk/zte/clk.c int idx; idx 71 drivers/clk/zte/clk.c idx = hw_to_idx(zx_pll); idx 72 drivers/clk/zte/clk.c if (unlikely(idx == -EINVAL)) idx 75 drivers/clk/zte/clk.c return zx_pll->lookup_table[idx].rate; idx 82 drivers/clk/zte/clk.c int idx; idx 84 drivers/clk/zte/clk.c idx = rate_to_idx(zx_pll, rate); idx 86 drivers/clk/zte/clk.c return zx_pll->lookup_table[idx].rate; idx 95 drivers/clk/zte/clk.c int idx; idx 97 drivers/clk/zte/clk.c idx = rate_to_idx(zx_pll, rate); idx 98 drivers/clk/zte/clk.c config = &zx_pll->lookup_table[idx]; idx 313 drivers/clk/zynq/clkc.c int idx = of_property_match_string(np, "clock-names", idx 315 drivers/clk/zynq/clkc.c if (idx >= 0) idx 317 drivers/clk/zynq/clkc.c of_clk_get_parent_name(np, idx); idx 382 drivers/clk/zynq/clkc.c int idx = of_property_match_string(np, "clock-names", idx 384 drivers/clk/zynq/clkc.c if (idx >= 0) idx 386 drivers/clk/zynq/clkc.c idx); idx 407 drivers/clk/zynq/clkc.c int idx = of_property_match_string(np, "clock-names", idx 409 drivers/clk/zynq/clkc.c if (idx >= 0) idx 411 drivers/clk/zynq/clkc.c idx); idx 434 drivers/clk/zynq/clkc.c int idx; idx 437 drivers/clk/zynq/clkc.c idx = of_property_match_string(np, "clock-names", clk_name); idx 438 drivers/clk/zynq/clkc.c if (idx >= 0) idx 440 drivers/clk/zynq/clkc.c idx); idx 479 drivers/clk/zynq/clkc.c int idx = of_property_match_string(np, "clock-names", idx 481 drivers/clk/zynq/clkc.c if (idx >= 0) idx 483 drivers/clk/zynq/clkc.c idx); idx 56 drivers/clocksource/timer-atlas7.c static inline void sirfsoc_timer_count_disable(int idx) idx 58 drivers/clocksource/timer-atlas7.c writel_relaxed(readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_32COUNTER_0_CTRL + 4 * idx) & ~0x7, idx 59 drivers/clocksource/timer-atlas7.c sirfsoc_timer_base + SIRFSOC_TIMER_32COUNTER_0_CTRL + 4 * idx); idx 63 drivers/clocksource/timer-atlas7.c static inline void sirfsoc_timer_count_enable(int idx) idx 65 drivers/clocksource/timer-atlas7.c writel_relaxed(readl_relaxed(sirfsoc_timer_base + SIRFSOC_TIMER_32COUNTER_0_CTRL + 4 * idx) | 0x3, idx 66 drivers/clocksource/timer-atlas7.c sirfsoc_timer_base + SIRFSOC_TIMER_32COUNTER_0_CTRL + 4 * idx); idx 306 drivers/clocksource/timer-tegra.c unsigned int idx = tegra_irq_idx_for_cpu(cpu, tegra20); idx 307 drivers/clocksource/timer-tegra.c unsigned int irq = irq_of_parse_and_map(np, idx); idx 56 drivers/connector/cn_queue.c return ((i1->idx == i2->idx) && (i1->val == i2->val)); idx 88 drivers/connector/cn_queue.c cbq->group = cbq->id.id.idx; idx 242 drivers/connector/connector.c cbq->id.id.idx, idx 536 drivers/cpufreq/cpufreq.c int idx; idx 538 drivers/cpufreq/cpufreq.c idx = cpufreq_frequency_table_target(policy, target_freq, idx 540 drivers/cpufreq/cpufreq.c policy->cached_resolved_idx = idx; idx 541 drivers/cpufreq/cpufreq.c return policy->freq_table[idx].frequency; idx 35 drivers/cpufreq/davinci-cpufreq.c static int davinci_target(struct cpufreq_policy *policy, unsigned int idx) idx 43 drivers/cpufreq/davinci-cpufreq.c new_freq = pdata->freq_table[idx].frequency; idx 47 drivers/cpufreq/davinci-cpufreq.c ret = pdata->set_voltage(idx); idx 64 drivers/cpufreq/davinci-cpufreq.c pdata->set_voltage(idx); idx 209 drivers/cpufreq/freq_table.c int idx; idx 216 drivers/cpufreq/freq_table.c cpufreq_for_each_valid_entry_idx(pos, table, idx) idx 218 drivers/cpufreq/freq_table.c return idx; idx 2683 drivers/cpufreq/intel_pstate.c int idx; idx 2694 drivers/cpufreq/intel_pstate.c idx = acpi_match_platform_list(plat_info); idx 2695 drivers/cpufreq/intel_pstate.c if (idx < 0) idx 2698 drivers/cpufreq/intel_pstate.c switch (plat_info[idx].data) { idx 128 drivers/cpufreq/pasemi-cpufreq.c int cur_astate, idx; idx 186 drivers/cpufreq/pasemi-cpufreq.c cpufreq_for_each_entry_idx(pos, pas_freqs, idx) { idx 188 drivers/cpufreq/pasemi-cpufreq.c pr_debug("%d: %d\n", idx, pos->frequency); idx 188 drivers/cpufreq/pxa2xx-cpufreq.c static int pxa_set_target(struct cpufreq_policy *policy, unsigned int idx) idx 199 drivers/cpufreq/pxa2xx-cpufreq.c new_freq_cpu = pxa_freq_settings[idx].khz; idx 206 drivers/cpufreq/pxa2xx-cpufreq.c ret = pxa_cpufreq_change_voltage(&pxa_freq_settings[idx]); idx 223 drivers/cpufreq/pxa2xx-cpufreq.c ret = pxa_cpufreq_change_voltage(&pxa_freq_settings[idx]); idx 118 drivers/cpufreq/s3c2416-cpufreq.c static int s3c2416_cpufreq_enter_dvs(struct s3c2416_data *s3c_freq, int idx) idx 141 drivers/cpufreq/s3c2416-cpufreq.c dvfs = &s3c2416_dvfs_table[idx]; idx 160 drivers/cpufreq/s3c2416-cpufreq.c static int s3c2416_cpufreq_leave_dvs(struct s3c2416_data *s3c_freq, int idx) idx 174 drivers/cpufreq/s3c2416-cpufreq.c dvfs = &s3c2416_dvfs_table[idx]; idx 221 drivers/cpufreq/s3c2416-cpufreq.c int idx, ret, to_dvs = 0; idx 225 drivers/cpufreq/s3c2416-cpufreq.c idx = s3c_freq->freq_table[index].driver_data; idx 227 drivers/cpufreq/s3c2416-cpufreq.c if (idx == SOURCE_HCLK) idx 247 drivers/cpufreq/s3c2416-cpufreq.c ret = s3c2416_cpufreq_enter_dvs(s3c_freq, idx); idx 250 drivers/cpufreq/s3c2416-cpufreq.c ret = s3c2416_cpufreq_leave_dvs(s3c_freq, idx); idx 34 drivers/cpuidle/cpuidle-arm.c struct cpuidle_driver *drv, int idx) idx 41 drivers/cpuidle/cpuidle-arm.c return CPU_PM_CPU_IDLE_ENTER(arm_cpuidle_suspend, idx); idx 27 drivers/cpuidle/cpuidle-big_little.c struct cpuidle_driver *drv, int idx); idx 124 drivers/cpuidle/cpuidle-big_little.c struct cpuidle_driver *drv, int idx) idx 135 drivers/cpuidle/cpuidle-big_little.c return idx; idx 28 drivers/cpuidle/cpuidle-psci.c struct cpuidle_driver *drv, int idx) idx 33 drivers/cpuidle/cpuidle-psci.c idx, state[idx - 1]); idx 97 drivers/cpuidle/dt_idle_states.c static bool idle_state_valid(struct device_node *state_node, unsigned int idx, idx 115 drivers/cpuidle/dt_idle_states.c idx); idx 281 drivers/cpuidle/governors/menu.c int idx; idx 349 drivers/cpuidle/governors/menu.c idx = -1; idx 357 drivers/cpuidle/governors/menu.c if (idx == -1) idx 358 drivers/cpuidle/governors/menu.c idx = i; /* first enabled state */ idx 365 drivers/cpuidle/governors/menu.c if ((drv->states[idx].flags & CPUIDLE_FLAG_POLLING) && idx 369 drivers/cpuidle/governors/menu.c idx = i; idx 382 drivers/cpuidle/governors/menu.c predicted_us = drv->states[idx].target_residency; idx 392 drivers/cpuidle/governors/menu.c if (drv->states[idx].target_residency < TICK_USEC && idx 394 drivers/cpuidle/governors/menu.c idx = i; idx 396 drivers/cpuidle/governors/menu.c return idx; idx 401 drivers/cpuidle/governors/menu.c idx = i; idx 404 drivers/cpuidle/governors/menu.c if (idx == -1) idx 405 drivers/cpuidle/governors/menu.c idx = 0; /* No states enabled. Must use 0. */ idx 411 drivers/cpuidle/governors/menu.c if (((drv->states[idx].flags & CPUIDLE_FLAG_POLLING) || idx 417 drivers/cpuidle/governors/menu.c if (idx > 0 && drv->states[idx].target_residency > delta_next_us) { idx 424 drivers/cpuidle/governors/menu.c for (i = idx - 1; i >= 0; i--) { idx 429 drivers/cpuidle/governors/menu.c idx = i; idx 436 drivers/cpuidle/governors/menu.c return idx; idx 237 drivers/cpuidle/governors/teo.c int max_early_idx, prev_max_early_idx, constraint_idx, idx, i; idx 256 drivers/cpuidle/governors/teo.c idx = -1; idx 282 drivers/cpuidle/governors/teo.c idx < 0) idx 292 drivers/cpuidle/governors/teo.c if (max_early_idx == idx) { idx 306 drivers/cpuidle/governors/teo.c drv->states[idx].target_residency < TICK_USEC)) { idx 309 drivers/cpuidle/governors/teo.c max_early_idx = idx; idx 315 drivers/cpuidle/governors/teo.c if (idx < 0) { idx 316 drivers/cpuidle/governors/teo.c idx = i; /* first enabled state */ idx 327 drivers/cpuidle/governors/teo.c idx = i; idx 354 drivers/cpuidle/governors/teo.c if (idx == max_early_idx) idx 358 drivers/cpuidle/governors/teo.c idx = max_early_idx; idx 359 drivers/cpuidle/governors/teo.c duration_us = drv->states[idx].target_residency; idx 367 drivers/cpuidle/governors/teo.c if (constraint_idx < idx) idx 368 drivers/cpuidle/governors/teo.c idx = constraint_idx; idx 370 drivers/cpuidle/governors/teo.c if (idx < 0) { idx 371 drivers/cpuidle/governors/teo.c idx = 0; /* No states enabled. Must use 0. */ idx 372 drivers/cpuidle/governors/teo.c } else if (idx > 0) { idx 403 drivers/cpuidle/governors/teo.c if (drv->states[idx].target_residency > avg_us) idx 404 drivers/cpuidle/governors/teo.c idx = teo_find_shallower_state(drv, dev, idx 405 drivers/cpuidle/governors/teo.c idx, avg_us); idx 414 drivers/cpuidle/governors/teo.c if (((drv->states[idx].flags & CPUIDLE_FLAG_POLLING) || idx 426 drivers/cpuidle/governors/teo.c if (idx > 0 && drv->states[idx].target_residency > delta_tick_us) idx 427 drivers/cpuidle/governors/teo.c idx = teo_find_shallower_state(drv, dev, idx, delta_tick_us); idx 430 drivers/cpuidle/governors/teo.c return idx; idx 252 drivers/crypto/amcc/crypto4xx_core.c static u32 crypto4xx_put_pd_to_pdr(struct crypto4xx_device *dev, u32 idx) idx 254 drivers/crypto/amcc/crypto4xx_core.c struct pd_uinfo *pd_uinfo = &dev->pdr_uinfo[idx]; idx 341 drivers/crypto/amcc/crypto4xx_core.c dma_addr_t *gd_dma, u32 idx) idx 343 drivers/crypto/amcc/crypto4xx_core.c *gd_dma = dev->gdr_pa + sizeof(struct ce_gd) * idx; idx 345 drivers/crypto/amcc/crypto4xx_core.c return &dev->gdr[idx]; idx 438 drivers/crypto/amcc/crypto4xx_core.c dma_addr_t *sd_dma, u32 idx) idx 440 drivers/crypto/amcc/crypto4xx_core.c *sd_dma = dev->sdr_pa + sizeof(struct ce_sd) * idx; idx 442 drivers/crypto/amcc/crypto4xx_core.c return &dev->sdr[idx]; idx 634 drivers/crypto/amcc/crypto4xx_core.c static void crypto4xx_pd_done(struct crypto4xx_device *dev, u32 idx) idx 636 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd = &dev->pdr[idx]; idx 637 drivers/crypto/amcc/crypto4xx_core.c struct pd_uinfo *pd_uinfo = &dev->pdr_uinfo[idx]; idx 275 drivers/crypto/bcm/util.c unsigned int idx = skip; idx 282 drivers/crypto/bcm/util.c sg_copy_part_to_buf(sg, dbuf, count, idx); idx 286 drivers/crypto/bcm/util.c idx += 16; idx 223 drivers/crypto/caam/error.c u8 idx = (status & JRSTA_DECOERR_INDEX_MASK) >> idx 258 drivers/crypto/caam/error.c error, idx_str, idx, cha_str, cha_err_code, idx 277 drivers/crypto/caam/error.c u8 idx = (status & JRSTA_DECOERR_INDEX_MASK) >> idx 299 drivers/crypto/caam/error.c status, error, idx_str, idx, err_str, err_err_code); idx 20 drivers/crypto/caam/key_gen.h u32 idx; idx 22 drivers/crypto/caam/key_gen.h idx = (hash & OP_ALG_ALGSEL_SUBMASK) >> OP_ALG_ALGSEL_SHIFT; idx 24 drivers/crypto/caam/key_gen.h return (u32)(mdpadlen[idx] * 2); idx 53 drivers/crypto/cavium/cpt/cptvf.h u32 idx; /* Command queue host write idx */ idx 196 drivers/crypto/cavium/cpt/cptvf_main.c queue->idx = 0; idx 243 drivers/crypto/cavium/cpt/cptvf_reqmanager.c ent = &queue->qhead->head[queue->idx * qinfo->cmd_size]; idx 246 drivers/crypto/cavium/cpt/cptvf_reqmanager.c if (++queue->idx >= queue->qhead->size / 64) { idx 259 drivers/crypto/cavium/cpt/cptvf_reqmanager.c queue->idx = 0; idx 23 drivers/crypto/cavium/nitrox/nitrox_debugfs.c seq_printf(s, "NITROX [%d]\n", ndev->idx); idx 40 drivers/crypto/cavium/nitrox/nitrox_debugfs.c seq_printf(s, "NITROX [%d] Request Statistics\n", ndev->idx); idx 232 drivers/crypto/cavium/nitrox/nitrox_dev.h u8 idx; idx 259 drivers/crypto/cavium/nitrox/nitrox_main.c ndev->idx = num_devices++; idx 69 drivers/crypto/cavium/nitrox/nitrox_mbx.c msg.id.chipid = ndev->idx; idx 288 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c int idx; idx 293 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c idx = cmdq->write_idx; idx 295 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c ent = cmdq->base + (idx * cmdq->instr_size); idx 307 drivers/crypto/cavium/nitrox/nitrox_reqmgr.c cmdq->write_idx = incr_index(idx, 1, ndev->qlen); idx 75 drivers/crypto/cavium/zip/zip_main.c int idx; idx 77 drivers/crypto/cavium/zip/zip_main.c for (idx = 0; idx < MAX_ZIP_DEVICES; idx++) { idx 78 drivers/crypto/cavium/zip/zip_main.c if (!zip_dev[idx]) idx 83 drivers/crypto/cavium/zip/zip_main.c if (idx < MAX_ZIP_DEVICES) idx 89 drivers/crypto/cavium/zip/zip_main.c zip_dev[idx] = zip; idx 90 drivers/crypto/cavium/zip/zip_main.c zip->index = idx; idx 306 drivers/crypto/ccree/cc_aead.c unsigned int idx = 0; idx 312 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 313 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 314 drivers/crypto/ccree/cc_aead.c set_din_sram(&desc[idx], idx 318 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 319 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 320 drivers/crypto/ccree/cc_aead.c idx++; idx 323 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 324 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 325 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0, ctx->hash_len); idx 326 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 327 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 328 drivers/crypto/ccree/cc_aead.c idx++; idx 331 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 332 drivers/crypto/ccree/cc_aead.c set_xor_val(&desc[idx], hmac_pad_const[i]); idx 333 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 334 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 335 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 336 drivers/crypto/ccree/cc_aead.c idx++; idx 339 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 340 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 343 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 344 drivers/crypto/ccree/cc_aead.c set_xor_active(&desc[idx]); idx 345 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_HASH); idx 346 drivers/crypto/ccree/cc_aead.c idx++; idx 349 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 350 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 351 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], idx 354 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 355 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 356 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); idx 357 drivers/crypto/ccree/cc_aead.c idx++; idx 362 drivers/crypto/ccree/cc_aead.c return idx; idx 424 drivers/crypto/ccree/cc_aead.c unsigned int idx = 0; idx 460 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 461 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hashmode); idx 462 drivers/crypto/ccree/cc_aead.c set_din_sram(&desc[idx], larval_addr, digestsize); idx 463 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 464 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 465 drivers/crypto/ccree/cc_aead.c idx++; idx 468 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 469 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hashmode); idx 470 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0, ctx->hash_len); idx 471 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 472 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 473 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 474 drivers/crypto/ccree/cc_aead.c idx++; idx 476 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 477 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 479 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_HASH); idx 480 drivers/crypto/ccree/cc_aead.c idx++; idx 483 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 484 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hashmode); idx 485 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], padded_authkey_dma_addr, idx 487 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 488 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 489 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); idx 490 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], idx 492 drivers/crypto/ccree/cc_aead.c idx++; idx 494 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 495 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0, (blocksize - digestsize)); idx 496 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], BYPASS); idx 497 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], (padded_authkey_dma_addr + idx 500 drivers/crypto/ccree/cc_aead.c idx++; idx 502 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 503 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, key_dma_addr, idx 505 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], BYPASS); idx 506 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], padded_authkey_dma_addr, idx 508 drivers/crypto/ccree/cc_aead.c idx++; idx 511 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 512 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0, idx 514 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], BYPASS); idx 515 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], idx 519 drivers/crypto/ccree/cc_aead.c idx++; idx 523 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 524 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0, (blocksize - keylen)); idx 525 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], BYPASS); idx 526 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], padded_authkey_dma_addr, idx 528 drivers/crypto/ccree/cc_aead.c idx++; idx 531 drivers/crypto/ccree/cc_aead.c rc = cc_send_sync_request(ctx->drvdata, &cc_req, desc, idx); idx 738 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 744 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 745 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, sg_dma_address(areq->src), idx 747 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], flow_mode); idx 750 drivers/crypto/ccree/cc_aead.c set_din_not_last_indication(&desc[idx]); idx 754 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 755 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_MLLI, areq_ctx->assoc.sram_addr, idx 757 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], flow_mode); idx 760 drivers/crypto/ccree/cc_aead.c set_din_not_last_indication(&desc[idx]); idx 767 drivers/crypto/ccree/cc_aead.c *seq_size = (++idx); idx 777 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 793 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 794 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 797 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], flow_mode); idx 820 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 821 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_MLLI, mlli_addr, mlli_nents, idx 823 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], flow_mode); idx 831 drivers/crypto/ccree/cc_aead.c *seq_size = (++idx); idx 839 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 852 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 853 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 857 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], idx 861 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], flow_mode); idx 865 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 866 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_MLLI, areq_ctx->src.sram_addr, idx 868 drivers/crypto/ccree/cc_aead.c set_dout_mlli(&desc[idx], areq_ctx->dst.sram_addr, idx 870 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], flow_mode); idx 877 drivers/crypto/ccree/cc_aead.c *seq_size = (++idx); idx 887 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 894 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 895 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 896 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 897 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], req_ctx->icv_dma_addr, ctx->authsize, idx 899 drivers/crypto/ccree/cc_aead.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 901 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 902 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 904 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], idx 906 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 910 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 911 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 912 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 913 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], req_ctx->mac_buf_dma_addr, idx 915 drivers/crypto/ccree/cc_aead.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 916 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], idx 918 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); idx 920 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 921 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 923 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 927 drivers/crypto/ccree/cc_aead.c *seq_size = (++idx); idx 938 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 942 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 943 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], direct); idx 944 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], ctx->flow_mode); idx 945 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->gen_ctx.iv_dma_addr, idx 948 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 950 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 951 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], ctx->cipher_mode); idx 952 drivers/crypto/ccree/cc_aead.c idx++; idx 955 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 956 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], direct); idx 957 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 958 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], ctx->flow_mode); idx 960 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, ctx->enckey_dma_addr, idx 963 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 965 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, ctx->enckey_dma_addr, idx 967 drivers/crypto/ccree/cc_aead.c set_key_size_des(&desc[idx], ctx->enc_keylen); idx 969 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], ctx->cipher_mode); idx 970 drivers/crypto/ccree/cc_aead.c idx++; idx 972 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 980 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 985 drivers/crypto/ccree/cc_aead.c cc_set_cipher_desc(req, desc, &idx); idx 986 drivers/crypto/ccree/cc_aead.c cc_proc_cipher_desc(req, data_flow_mode, desc, &idx); idx 989 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 990 drivers/crypto/ccree/cc_aead.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 991 drivers/crypto/ccree/cc_aead.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 992 drivers/crypto/ccree/cc_aead.c idx++; idx 995 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1007 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1010 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1011 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 1012 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1015 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1016 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 1017 drivers/crypto/ccree/cc_aead.c idx++; idx 1020 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1021 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 1022 drivers/crypto/ccree/cc_aead.c set_din_sram(&desc[idx], cc_digest_len_addr(ctx->drvdata, hash_mode), idx 1024 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1025 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1026 drivers/crypto/ccree/cc_aead.c idx++; idx 1028 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1036 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1039 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1040 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0, CC_AES_BLOCK_SIZE); idx 1041 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 1042 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 1043 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1044 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 1045 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1046 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1047 drivers/crypto/ccree/cc_aead.c idx++; idx 1050 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1051 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1054 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1055 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 1056 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1057 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 1058 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1059 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1060 drivers/crypto/ccree/cc_aead.c idx++; idx 1063 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1064 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1067 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 1068 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 1069 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1070 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 1071 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1072 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1073 drivers/crypto/ccree/cc_aead.c idx++; idx 1076 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1077 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1080 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE2); idx 1081 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 1082 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1083 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 1084 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1085 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1086 drivers/crypto/ccree/cc_aead.c idx++; idx 1088 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1096 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1100 drivers/crypto/ccree/cc_aead.c cc_set_assoc_desc(req, DIN_HASH, desc, &idx); idx 1103 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1117 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1119 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1120 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 1121 drivers/crypto/ccree/cc_aead.c set_dout_sram(&desc[idx], aead_handle->sram_workspace_addr, idx 1123 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 1124 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE1); idx 1125 drivers/crypto/ccree/cc_aead.c set_cipher_do(&desc[idx], DO_PAD); idx 1126 drivers/crypto/ccree/cc_aead.c idx++; idx 1129 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1130 drivers/crypto/ccree/cc_aead.c set_dout_sram(&desc[idx], aead_handle->sram_workspace_addr, idx 1132 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 1133 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 1134 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], HASH_DIGEST_RESULT_LITTLE_ENDIAN); idx 1135 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 1136 drivers/crypto/ccree/cc_aead.c idx++; idx 1139 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1140 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 1141 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1144 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1145 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 1146 drivers/crypto/ccree/cc_aead.c idx++; idx 1149 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1150 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], hash_mode); idx 1151 drivers/crypto/ccree/cc_aead.c set_din_sram(&desc[idx], cc_digest_len_addr(ctx->drvdata, hash_mode), idx 1153 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 1154 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1155 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1156 drivers/crypto/ccree/cc_aead.c idx++; idx 1159 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1160 drivers/crypto/ccree/cc_aead.c set_din_sram(&desc[idx], aead_handle->sram_workspace_addr, idx 1162 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_HASH); idx 1163 drivers/crypto/ccree/cc_aead.c idx++; idx 1165 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1425 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1438 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1439 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_CTR); idx 1440 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, ctx->enckey_dma_addr, idx 1443 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1444 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1445 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1446 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1447 drivers/crypto/ccree/cc_aead.c idx++; idx 1450 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1451 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_CTR); idx 1452 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1453 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1455 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1456 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 1457 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1458 drivers/crypto/ccree/cc_aead.c idx++; idx 1461 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1462 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_CBC_MAC); idx 1463 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, ctx->enckey_dma_addr, idx 1466 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1467 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1468 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1469 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1470 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1471 drivers/crypto/ccree/cc_aead.c idx++; idx 1474 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1475 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_CBC_MAC); idx 1476 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1477 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->mac_buf_dma_addr, idx 1479 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1480 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 1481 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1482 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1483 drivers/crypto/ccree/cc_aead.c idx++; idx 1487 drivers/crypto/ccree/cc_aead.c cc_set_assoc_desc(req, DIN_HASH, desc, &idx); idx 1489 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1490 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1493 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_HASH); idx 1494 drivers/crypto/ccree/cc_aead.c idx++; idx 1499 drivers/crypto/ccree/cc_aead.c cc_proc_cipher_desc(req, cipher_flow_mode, desc, &idx); idx 1502 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1503 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_CBC_MAC); idx 1504 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], req_ctx->mac_buf_dma_addr, ctx->authsize, idx 1506 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 1507 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], HASH_DIGEST_RESULT_LITTLE_ENDIAN); idx 1508 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 1509 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1510 drivers/crypto/ccree/cc_aead.c idx++; idx 1513 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1514 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_CTR); idx 1515 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_ENCRYPT); idx 1516 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->ccm_iv0_dma_addr, idx 1518 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1519 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 1520 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1521 drivers/crypto/ccree/cc_aead.c idx++; idx 1523 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1524 drivers/crypto/ccree/cc_aead.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 1525 drivers/crypto/ccree/cc_aead.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 1526 drivers/crypto/ccree/cc_aead.c idx++; idx 1529 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1530 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->mac_buf_dma_addr, idx 1532 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], mac_result, ctx->authsize, NS_BIT, 1); idx 1533 drivers/crypto/ccree/cc_aead.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 1534 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 1535 drivers/crypto/ccree/cc_aead.c idx++; idx 1537 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1630 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1633 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1634 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_ECB); idx 1635 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_ENCRYPT); idx 1636 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, ctx->enckey_dma_addr, idx 1638 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1639 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1640 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1641 drivers/crypto/ccree/cc_aead.c idx++; idx 1644 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1645 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0x0, AES_BLOCK_SIZE); idx 1646 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], req_ctx->hkey_dma_addr, AES_BLOCK_SIZE, idx 1648 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 1649 drivers/crypto/ccree/cc_aead.c idx++; idx 1652 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1653 drivers/crypto/ccree/cc_aead.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 1654 drivers/crypto/ccree/cc_aead.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 1655 drivers/crypto/ccree/cc_aead.c idx++; idx 1658 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1659 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->hkey_dma_addr, idx 1661 drivers/crypto/ccree/cc_aead.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 1662 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1663 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1664 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_HASH_HW_GHASH); idx 1665 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 1666 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1667 drivers/crypto/ccree/cc_aead.c idx++; idx 1674 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1675 drivers/crypto/ccree/cc_aead.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 1676 drivers/crypto/ccree/cc_aead.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 1677 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1678 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1679 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_HASH_HW_GHASH); idx 1680 drivers/crypto/ccree/cc_aead.c set_cipher_do(&desc[idx], 1); //1=AES_SK RKEK idx 1681 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_ENCRYPT); idx 1682 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 1683 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1684 drivers/crypto/ccree/cc_aead.c idx++; idx 1689 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1690 drivers/crypto/ccree/cc_aead.c set_din_const(&desc[idx], 0x0, AES_BLOCK_SIZE); idx 1691 drivers/crypto/ccree/cc_aead.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 1692 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 1693 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1694 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_HASH_HW_GHASH); idx 1695 drivers/crypto/ccree/cc_aead.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 1696 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 1697 drivers/crypto/ccree/cc_aead.c idx++; idx 1699 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1708 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1711 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1712 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_GCTR); idx 1713 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_ENCRYPT); idx 1714 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, ctx->enckey_dma_addr, idx 1716 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1717 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1718 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1719 drivers/crypto/ccree/cc_aead.c idx++; idx 1723 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1724 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_GCTR); idx 1725 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1726 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, idx 1729 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_ENCRYPT); idx 1730 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 1731 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1732 drivers/crypto/ccree/cc_aead.c idx++; idx 1735 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 1746 drivers/crypto/ccree/cc_aead.c unsigned int idx = *seq_size; idx 1755 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1756 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->gcm_block_len_dma_addr, idx 1758 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_HASH); idx 1759 drivers/crypto/ccree/cc_aead.c idx++; idx 1762 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1763 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_HASH_HW_GHASH); idx 1764 drivers/crypto/ccree/cc_aead.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 1765 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], req_ctx->mac_buf_dma_addr, AES_BLOCK_SIZE, idx 1767 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 1768 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 1769 drivers/crypto/ccree/cc_aead.c set_aes_not_hash_mode(&desc[idx]); idx 1771 drivers/crypto/ccree/cc_aead.c idx++; idx 1774 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1775 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_GCTR); idx 1776 drivers/crypto/ccree/cc_aead.c set_key_size_aes(&desc[idx], ctx->enc_keylen); idx 1777 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->gcm_iv_inc1_dma_addr, idx 1779 drivers/crypto/ccree/cc_aead.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_ENCRYPT); idx 1780 drivers/crypto/ccree/cc_aead.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 1781 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1782 drivers/crypto/ccree/cc_aead.c idx++; idx 1785 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1786 drivers/crypto/ccree/cc_aead.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 1787 drivers/crypto/ccree/cc_aead.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 1788 drivers/crypto/ccree/cc_aead.c idx++; idx 1791 drivers/crypto/ccree/cc_aead.c hw_desc_init(&desc[idx]); idx 1792 drivers/crypto/ccree/cc_aead.c set_cipher_mode(&desc[idx], DRV_CIPHER_GCTR); idx 1793 drivers/crypto/ccree/cc_aead.c set_din_type(&desc[idx], DMA_DLLI, req_ctx->mac_buf_dma_addr, idx 1795 drivers/crypto/ccree/cc_aead.c set_dout_dlli(&desc[idx], mac_result, ctx->authsize, NS_BIT, 1); idx 1796 drivers/crypto/ccree/cc_aead.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 1797 drivers/crypto/ccree/cc_aead.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 1798 drivers/crypto/ccree/cc_aead.c idx++; idx 1800 drivers/crypto/ccree/cc_aead.c *seq_size = idx; idx 335 drivers/crypto/ccree/cc_hash.c int idx) idx 343 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 344 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 346 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_result_dma_addr, digestsize, idx 348 drivers/crypto/ccree/cc_hash.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 349 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 350 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 351 drivers/crypto/ccree/cc_hash.c set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); idx 352 drivers/crypto/ccree/cc_hash.c cc_set_endianity(ctx->hash_mode, &desc[idx]); idx 353 drivers/crypto/ccree/cc_hash.c idx++; idx 355 drivers/crypto/ccree/cc_hash.c return idx; idx 359 drivers/crypto/ccree/cc_hash.c int idx) idx 367 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 368 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 369 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_buff_dma_addr, digestsize, idx 371 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 372 drivers/crypto/ccree/cc_hash.c cc_set_endianity(ctx->hash_mode, &desc[idx]); idx 373 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 374 drivers/crypto/ccree/cc_hash.c idx++; idx 377 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 378 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 379 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->opad_digest_dma_addr, idx 381 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 382 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 383 drivers/crypto/ccree/cc_hash.c idx++; idx 386 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 387 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 388 drivers/crypto/ccree/cc_hash.c set_din_sram(&desc[idx], idx 391 drivers/crypto/ccree/cc_hash.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 392 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 393 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 394 drivers/crypto/ccree/cc_hash.c idx++; idx 397 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 398 drivers/crypto/ccree/cc_hash.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 399 drivers/crypto/ccree/cc_hash.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 400 drivers/crypto/ccree/cc_hash.c idx++; idx 403 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 404 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->digest_buff_dma_addr, idx 406 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_HASH); idx 407 drivers/crypto/ccree/cc_hash.c idx++; idx 409 drivers/crypto/ccree/cc_hash.c return idx; idx 427 drivers/crypto/ccree/cc_hash.c int idx = 0; idx 462 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 463 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 465 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->digest_buff_dma_addr, idx 468 drivers/crypto/ccree/cc_hash.c set_din_sram(&desc[idx], larval_digest_addr, idx 471 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 472 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 473 drivers/crypto/ccree/cc_hash.c idx++; idx 476 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 477 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 480 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 484 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0, ctx->hash_len); idx 486 drivers/crypto/ccree/cc_hash.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 488 drivers/crypto/ccree/cc_hash.c set_cipher_do(&desc[idx], DO_PAD); idx 490 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 491 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 492 drivers/crypto/ccree/cc_hash.c idx++; idx 494 drivers/crypto/ccree/cc_hash.c cc_set_desc(state, ctx, DIN_HASH, desc, false, &idx); idx 498 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 499 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 500 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_buff_dma_addr, idx 502 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 503 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE1); idx 504 drivers/crypto/ccree/cc_hash.c set_cipher_do(&desc[idx], DO_PAD); idx 505 drivers/crypto/ccree/cc_hash.c idx++; idx 507 drivers/crypto/ccree/cc_hash.c idx = cc_fin_hmac(desc, req, idx); idx 510 drivers/crypto/ccree/cc_hash.c idx = cc_fin_result(desc, req, idx); idx 512 drivers/crypto/ccree/cc_hash.c rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base); idx 523 drivers/crypto/ccree/cc_hash.c struct ahash_req_ctx *state, unsigned int idx) idx 526 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 527 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 528 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->digest_buff_dma_addr, idx 530 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 531 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 532 drivers/crypto/ccree/cc_hash.c idx++; idx 535 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 536 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 537 drivers/crypto/ccree/cc_hash.c set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); idx 538 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->digest_bytes_len_dma_addr, idx 540 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 541 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 542 drivers/crypto/ccree/cc_hash.c idx++; idx 544 drivers/crypto/ccree/cc_hash.c cc_set_desc(state, ctx, DIN_HASH, desc, false, &idx); idx 546 drivers/crypto/ccree/cc_hash.c return idx; idx 560 drivers/crypto/ccree/cc_hash.c u32 idx = 0; idx 595 drivers/crypto/ccree/cc_hash.c idx = cc_restore_hash(desc, ctx, state, idx); idx 598 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 599 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 600 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_buff_dma_addr, idx 602 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 603 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 604 drivers/crypto/ccree/cc_hash.c idx++; idx 607 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 608 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 609 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_bytes_len_dma_addr, idx 611 drivers/crypto/ccree/cc_hash.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 612 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 613 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE1); idx 614 drivers/crypto/ccree/cc_hash.c idx++; idx 616 drivers/crypto/ccree/cc_hash.c rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base); idx 638 drivers/crypto/ccree/cc_hash.c unsigned int idx = 0; idx 667 drivers/crypto/ccree/cc_hash.c idx = cc_restore_hash(desc, ctx, state, idx); idx 670 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 671 drivers/crypto/ccree/cc_hash.c set_cipher_do(&desc[idx], DO_PAD); idx 672 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], ctx->hw_mode, ctx->hash_mode); idx 673 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_bytes_len_dma_addr, idx 675 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE1); idx 676 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 677 drivers/crypto/ccree/cc_hash.c idx++; idx 680 drivers/crypto/ccree/cc_hash.c idx = cc_fin_hmac(desc, req, idx); idx 682 drivers/crypto/ccree/cc_hash.c idx = cc_fin_result(desc, req, idx); idx 684 drivers/crypto/ccree/cc_hash.c rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base); idx 727 drivers/crypto/ccree/cc_hash.c int i, idx = 0, rc = 0; idx 768 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 769 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 770 drivers/crypto/ccree/cc_hash.c set_din_sram(&desc[idx], larval_addr, idx 772 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 773 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 774 drivers/crypto/ccree/cc_hash.c idx++; idx 777 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 778 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 779 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0, ctx->hash_len); idx 780 drivers/crypto/ccree/cc_hash.c set_cipher_config1(&desc[idx], HASH_PADDING_ENABLED); idx 781 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 782 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 783 drivers/crypto/ccree/cc_hash.c idx++; idx 785 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 786 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 789 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_HASH); idx 790 drivers/crypto/ccree/cc_hash.c idx++; idx 793 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 794 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 795 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], ctx->opad_tmp_keys_dma_addr, idx 797 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 798 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 799 drivers/crypto/ccree/cc_hash.c set_cipher_config1(&desc[idx], HASH_PADDING_DISABLED); idx 800 drivers/crypto/ccree/cc_hash.c cc_set_endianity(ctx->hash_mode, &desc[idx]); idx 801 drivers/crypto/ccree/cc_hash.c idx++; idx 803 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 804 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0, (blocksize - digestsize)); idx 805 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], BYPASS); idx 806 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], idx 810 drivers/crypto/ccree/cc_hash.c idx++; idx 812 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 813 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 816 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], BYPASS); idx 817 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], ctx->opad_tmp_keys_dma_addr, idx 819 drivers/crypto/ccree/cc_hash.c idx++; idx 822 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 823 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0, idx 825 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], BYPASS); idx 826 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], idx 830 drivers/crypto/ccree/cc_hash.c idx++; idx 834 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 835 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0, blocksize); idx 836 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], BYPASS); idx 837 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], (ctx->opad_tmp_keys_dma_addr), idx 839 drivers/crypto/ccree/cc_hash.c idx++; idx 842 drivers/crypto/ccree/cc_hash.c rc = cc_send_sync_request(ctx->drvdata, &cc_req, desc, idx); idx 849 drivers/crypto/ccree/cc_hash.c for (idx = 0, i = 0; i < 2; i++) { idx 851 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 852 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 853 drivers/crypto/ccree/cc_hash.c set_din_sram(&desc[idx], larval_addr, ctx->inter_digestsize); idx 854 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 855 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 856 drivers/crypto/ccree/cc_hash.c idx++; idx 859 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 860 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 861 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0, ctx->hash_len); idx 862 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 863 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 864 drivers/crypto/ccree/cc_hash.c idx++; idx 867 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 868 drivers/crypto/ccree/cc_hash.c set_xor_val(&desc[idx], hmac_pad_const[i]); idx 869 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 870 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_HASH); idx 871 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 872 drivers/crypto/ccree/cc_hash.c idx++; idx 875 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 876 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, ctx->opad_tmp_keys_dma_addr, idx 878 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 879 drivers/crypto/ccree/cc_hash.c set_xor_active(&desc[idx]); idx 880 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_HASH); idx 881 drivers/crypto/ccree/cc_hash.c idx++; idx 886 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 887 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 889 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], ctx->opad_tmp_keys_dma_addr, idx 892 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], ctx->digest_buff_dma_addr, idx 894 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_HASH_to_DOUT); idx 895 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 896 drivers/crypto/ccree/cc_hash.c idx++; idx 899 drivers/crypto/ccree/cc_hash.c rc = cc_send_sync_request(ctx->drvdata, &cc_req, desc, idx); idx 924 drivers/crypto/ccree/cc_hash.c unsigned int idx = 0; idx 957 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 958 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, ctx->key_params.key_dma_addr, idx 960 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], DRV_CIPHER_ECB); idx 961 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_ENCRYPT); idx 962 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], keylen); idx 963 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 964 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 965 drivers/crypto/ccree/cc_hash.c idx++; idx 967 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 968 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0x01010101, CC_AES_128_BIT_KEY_SIZE); idx 969 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 970 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], idx 973 drivers/crypto/ccree/cc_hash.c idx++; idx 975 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 976 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0x02020202, CC_AES_128_BIT_KEY_SIZE); idx 977 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 978 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], idx 981 drivers/crypto/ccree/cc_hash.c idx++; idx 983 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 984 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0x03030303, CC_AES_128_BIT_KEY_SIZE); idx 985 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 986 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], idx 989 drivers/crypto/ccree/cc_hash.c idx++; idx 991 drivers/crypto/ccree/cc_hash.c rc = cc_send_sync_request(ctx->drvdata, &cc_req, desc, idx); idx 1159 drivers/crypto/ccree/cc_hash.c u32 idx = 0; idx 1188 drivers/crypto/ccree/cc_hash.c cc_setup_xcbc(req, desc, &idx); idx 1190 drivers/crypto/ccree/cc_hash.c cc_setup_cmac(req, desc, &idx); idx 1192 drivers/crypto/ccree/cc_hash.c cc_set_desc(state, ctx, DIN_AES_DOUT, desc, true, &idx); idx 1195 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1196 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 1197 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_buff_dma_addr, idx 1199 drivers/crypto/ccree/cc_hash.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 1200 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_AES_to_DOUT); idx 1201 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 1202 drivers/crypto/ccree/cc_hash.c idx++; idx 1208 drivers/crypto/ccree/cc_hash.c rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base); idx 1225 drivers/crypto/ccree/cc_hash.c int idx = 0; idx 1268 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1269 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], DRV_CIPHER_ECB); idx 1270 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DRV_CRYPTO_DIRECTION_DECRYPT); idx 1271 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 1274 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], key_len); idx 1275 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1276 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 1277 drivers/crypto/ccree/cc_hash.c idx++; idx 1282 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1283 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->digest_buff_dma_addr, idx 1285 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_buff_dma_addr, idx 1287 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 1288 drivers/crypto/ccree/cc_hash.c idx++; idx 1291 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1292 drivers/crypto/ccree/cc_hash.c set_din_no_dma(&desc[idx], 0, 0xfffff0); idx 1293 drivers/crypto/ccree/cc_hash.c set_dout_no_dma(&desc[idx], 0, 0, 1); idx 1294 drivers/crypto/ccree/cc_hash.c idx++; idx 1298 drivers/crypto/ccree/cc_hash.c cc_setup_xcbc(req, desc, &idx); idx 1300 drivers/crypto/ccree/cc_hash.c cc_setup_cmac(req, desc, &idx); idx 1303 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1304 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 1305 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], key_len); idx 1306 drivers/crypto/ccree/cc_hash.c set_cmac_size0_mode(&desc[idx]); idx 1307 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1308 drivers/crypto/ccree/cc_hash.c idx++; idx 1310 drivers/crypto/ccree/cc_hash.c cc_set_desc(state, ctx, DIN_AES_DOUT, desc, false, &idx); idx 1312 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1313 drivers/crypto/ccree/cc_hash.c set_din_const(&desc[idx], 0x00, CC_AES_BLOCK_SIZE); idx 1314 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], DIN_AES_DOUT); idx 1315 drivers/crypto/ccree/cc_hash.c idx++; idx 1319 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1321 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_result_dma_addr, idx 1323 drivers/crypto/ccree/cc_hash.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 1324 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_AES_to_DOUT); idx 1325 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 1326 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 1327 drivers/crypto/ccree/cc_hash.c idx++; idx 1329 drivers/crypto/ccree/cc_hash.c rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base); idx 1347 drivers/crypto/ccree/cc_hash.c int idx = 0; idx 1383 drivers/crypto/ccree/cc_hash.c cc_setup_xcbc(req, desc, &idx); idx 1386 drivers/crypto/ccree/cc_hash.c cc_setup_cmac(req, desc, &idx); idx 1390 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1391 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 1392 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], key_len); idx 1393 drivers/crypto/ccree/cc_hash.c set_cmac_size0_mode(&desc[idx]); idx 1394 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1395 drivers/crypto/ccree/cc_hash.c idx++; idx 1397 drivers/crypto/ccree/cc_hash.c cc_set_desc(state, ctx, DIN_AES_DOUT, desc, false, &idx); idx 1401 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1403 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_result_dma_addr, idx 1405 drivers/crypto/ccree/cc_hash.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 1406 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_AES_to_DOUT); idx 1407 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 1408 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 1409 drivers/crypto/ccree/cc_hash.c idx++; idx 1411 drivers/crypto/ccree/cc_hash.c rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base); idx 1431 drivers/crypto/ccree/cc_hash.c unsigned int idx = 0; idx 1462 drivers/crypto/ccree/cc_hash.c cc_setup_xcbc(req, desc, &idx); idx 1465 drivers/crypto/ccree/cc_hash.c cc_setup_cmac(req, desc, &idx); idx 1469 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1470 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 1471 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], key_len); idx 1472 drivers/crypto/ccree/cc_hash.c set_cmac_size0_mode(&desc[idx]); idx 1473 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 1474 drivers/crypto/ccree/cc_hash.c idx++; idx 1476 drivers/crypto/ccree/cc_hash.c cc_set_desc(state, ctx, DIN_AES_DOUT, desc, false, &idx); idx 1480 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 1481 drivers/crypto/ccree/cc_hash.c set_dout_dlli(&desc[idx], state->digest_result_dma_addr, idx 1483 drivers/crypto/ccree/cc_hash.c set_queue_last_ind(ctx->drvdata, &desc[idx]); idx 1484 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_AES_to_DOUT); idx 1485 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_WRITE_STATE0); idx 1486 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 1487 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], ctx->hw_mode); idx 1488 drivers/crypto/ccree/cc_hash.c idx++; idx 1490 drivers/crypto/ccree/cc_hash.c rc = cc_send_request(ctx->drvdata, &cc_req, desc, idx, &req->base); idx 2127 drivers/crypto/ccree/cc_hash.c unsigned int idx = *seq_size; idx 2133 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2134 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, (ctx->opad_tmp_keys_dma_addr + idx 2137 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 2138 drivers/crypto/ccree/cc_hash.c set_hash_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC, ctx->hash_mode); idx 2139 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 2140 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 2141 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 2142 drivers/crypto/ccree/cc_hash.c idx++; idx 2145 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2146 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 2149 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE1); idx 2150 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 2151 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 2152 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 2153 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 2154 drivers/crypto/ccree/cc_hash.c idx++; idx 2157 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2158 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 2161 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE2); idx 2162 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 2163 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 2164 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 2165 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 2166 drivers/crypto/ccree/cc_hash.c idx++; idx 2169 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2170 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->digest_buff_dma_addr, idx 2172 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 2173 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], DRV_CIPHER_XCBC_MAC); idx 2174 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 2175 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], CC_AES_128_BIT_KEY_SIZE); idx 2176 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 2177 drivers/crypto/ccree/cc_hash.c idx++; idx 2178 drivers/crypto/ccree/cc_hash.c *seq_size = idx; idx 2184 drivers/crypto/ccree/cc_hash.c unsigned int idx = *seq_size; idx 2190 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2191 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, ctx->opad_tmp_keys_dma_addr, idx 2194 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_KEY0); idx 2195 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], DRV_CIPHER_CMAC); idx 2196 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 2197 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], ctx->key_params.keylen); idx 2198 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 2199 drivers/crypto/ccree/cc_hash.c idx++; idx 2202 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2203 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, state->digest_buff_dma_addr, idx 2205 drivers/crypto/ccree/cc_hash.c set_setup_mode(&desc[idx], SETUP_LOAD_STATE0); idx 2206 drivers/crypto/ccree/cc_hash.c set_cipher_mode(&desc[idx], DRV_CIPHER_CMAC); idx 2207 drivers/crypto/ccree/cc_hash.c set_cipher_config0(&desc[idx], DESC_DIRECTION_ENCRYPT_ENCRYPT); idx 2208 drivers/crypto/ccree/cc_hash.c set_key_size_aes(&desc[idx], ctx->key_params.keylen); idx 2209 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], S_DIN_to_AES); idx 2210 drivers/crypto/ccree/cc_hash.c idx++; idx 2211 drivers/crypto/ccree/cc_hash.c *seq_size = idx; idx 2219 drivers/crypto/ccree/cc_hash.c unsigned int idx = *seq_size; idx 2223 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2224 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 2227 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], flow_mode); idx 2228 drivers/crypto/ccree/cc_hash.c idx++; idx 2236 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2237 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_DLLI, idx 2240 drivers/crypto/ccree/cc_hash.c set_dout_sram(&desc[idx], ctx->drvdata->mlli_sram_addr, idx 2242 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], BYPASS); idx 2243 drivers/crypto/ccree/cc_hash.c idx++; idx 2245 drivers/crypto/ccree/cc_hash.c hw_desc_init(&desc[idx]); idx 2246 drivers/crypto/ccree/cc_hash.c set_din_type(&desc[idx], DMA_MLLI, idx 2249 drivers/crypto/ccree/cc_hash.c set_flow_mode(&desc[idx], flow_mode); idx 2250 drivers/crypto/ccree/cc_hash.c idx++; idx 2253 drivers/crypto/ccree/cc_hash.c set_din_not_last_indication(&desc[(idx - 1)]); idx 2255 drivers/crypto/ccree/cc_hash.c *seq_size = idx; idx 109 drivers/crypto/ccree/cc_sram_mgr.c unsigned int idx = *seq_len; idx 111 drivers/crypto/ccree/cc_sram_mgr.c for (i = 0; i < nelement; i++, idx++) { idx 112 drivers/crypto/ccree/cc_sram_mgr.c hw_desc_init(&seq[idx]); idx 113 drivers/crypto/ccree/cc_sram_mgr.c set_din_const(&seq[idx], src[i], sizeof(u32)); idx 114 drivers/crypto/ccree/cc_sram_mgr.c set_dout_sram(&seq[idx], dst + (i * sizeof(u32)), sizeof(u32)); idx 115 drivers/crypto/ccree/cc_sram_mgr.c set_flow_mode(&seq[idx], BYPASS); idx 118 drivers/crypto/ccree/cc_sram_mgr.c *seq_len = idx; idx 580 drivers/crypto/chelsio/chcr_algo.c static int cxgb4_is_crypto_q_full(struct net_device *dev, unsigned int idx) idx 589 drivers/crypto/chelsio/chcr_algo.c txq = &txq_info->uldtxq[idx]; idx 950 drivers/crypto/chelsio/chtls/chtls_cm.c L2T_IDX_V(csk->l2t_entry->idx) | idx 1234 drivers/crypto/hifn_795x.c int idx; idx 1239 drivers/crypto/hifn_795x.c idx = dma->srci; idx 1241 drivers/crypto/hifn_795x.c dma->srcr[idx].p = __cpu_to_le32(addr); idx 1242 drivers/crypto/hifn_795x.c dma->srcr[idx].l = __cpu_to_le32(size | HIFN_D_VALID | idx 1245 drivers/crypto/hifn_795x.c if (++idx == HIFN_D_SRC_RSIZE) { idx 1246 drivers/crypto/hifn_795x.c dma->srcr[idx].l = __cpu_to_le32(HIFN_D_VALID | idx 1249 drivers/crypto/hifn_795x.c idx = 0; idx 1252 drivers/crypto/hifn_795x.c dma->srci = idx; idx 1292 drivers/crypto/hifn_795x.c int idx; idx 1297 drivers/crypto/hifn_795x.c idx = dma->dsti; idx 1298 drivers/crypto/hifn_795x.c dma->dstr[idx].p = __cpu_to_le32(addr); idx 1299 drivers/crypto/hifn_795x.c dma->dstr[idx].l = __cpu_to_le32(size | HIFN_D_VALID | idx 1302 drivers/crypto/hifn_795x.c if (++idx == HIFN_D_DST_RSIZE) { idx 1303 drivers/crypto/hifn_795x.c dma->dstr[idx].l = __cpu_to_le32(HIFN_D_VALID | idx 1306 drivers/crypto/hifn_795x.c idx = 0; idx 1308 drivers/crypto/hifn_795x.c dma->dsti = idx; idx 1411 drivers/crypto/hifn_795x.c int idx = 0; idx 1427 drivers/crypto/hifn_795x.c idx++; idx 1433 drivers/crypto/hifn_795x.c return idx; idx 1441 drivers/crypto/hifn_795x.c int idx, tidx, err; idx 1443 drivers/crypto/hifn_795x.c tidx = idx = 0; idx 1446 drivers/crypto/hifn_795x.c if (idx >= w->num && (w->flags & ASYNC_FLAGS_MISALIGNED)) idx 1449 drivers/crypto/hifn_795x.c dst = &req->dst[idx]; idx 1460 drivers/crypto/hifn_795x.c t = &w->cache[idx]; idx 1466 drivers/crypto/hifn_795x.c idx += err; idx 1499 drivers/crypto/hifn_795x.c dst = &req->dst[idx]; idx 1505 drivers/crypto/hifn_795x.c idx += err; idx 1512 drivers/crypto/hifn_795x.c idx++; idx 1527 drivers/crypto/hifn_795x.c unsigned int nbytes = req->nbytes, idx = 0; idx 1537 drivers/crypto/hifn_795x.c dst = &req->dst[idx]; idx 1545 drivers/crypto/hifn_795x.c idx++; idx 1549 drivers/crypto/hifn_795x.c err = hifn_cipher_walk_init(&rctx->walk, idx, GFP_ATOMIC); idx 1618 drivers/crypto/hifn_795x.c int idx = 0; idx 1640 drivers/crypto/hifn_795x.c idx++; idx 1646 drivers/crypto/hifn_795x.c return idx; idx 1669 drivers/crypto/hifn_795x.c int idx = 0, err; idx 1674 drivers/crypto/hifn_795x.c t = &rctx->walk.cache[idx]; idx 1675 drivers/crypto/hifn_795x.c dst = &req->dst[idx]; idx 1685 drivers/crypto/hifn_795x.c idx++; idx 1698 drivers/crypto/hifn_795x.c idx += err; idx 271 drivers/crypto/ixp4xx_crypto.c static int idx = 0; idx 282 drivers/crypto/ixp4xx_crypto.c i = idx; idx 284 drivers/crypto/ixp4xx_crypto.c if (++idx >= NPE_QLEN) idx 285 drivers/crypto/ixp4xx_crypto.c idx = 0; idx 299 drivers/crypto/ixp4xx_crypto.c static int idx = NPE_QLEN; idx 310 drivers/crypto/ixp4xx_crypto.c i = idx; idx 312 drivers/crypto/ixp4xx_crypto.c if (++idx >= NPE_QLEN_TOTAL) idx 313 drivers/crypto/ixp4xx_crypto.c idx = NPE_QLEN; idx 374 drivers/crypto/marvell/cesa.c static int mv_cesa_get_sram(struct platform_device *pdev, int idx) idx 377 drivers/crypto/marvell/cesa.c struct mv_cesa_engine *engine = &cesa->engines[idx]; idx 382 drivers/crypto/marvell/cesa.c "marvell,crypto-srams", idx); idx 395 drivers/crypto/marvell/cesa.c if (!idx) idx 419 drivers/crypto/marvell/cesa.c static void mv_cesa_put_sram(struct platform_device *pdev, int idx) idx 422 drivers/crypto/marvell/cesa.c struct mv_cesa_engine *engine = &cesa->engines[idx]; idx 447 drivers/crypto/sahara.c int idx = 0; idx 455 drivers/crypto/sahara.c dev->hw_desc[idx]->len1 = AES_BLOCK_SIZE; idx 456 drivers/crypto/sahara.c dev->hw_desc[idx]->p1 = dev->iv_phys_base; idx 458 drivers/crypto/sahara.c dev->hw_desc[idx]->len1 = 0; idx 459 drivers/crypto/sahara.c dev->hw_desc[idx]->p1 = 0; idx 461 drivers/crypto/sahara.c dev->hw_desc[idx]->len2 = ctx->keylen; idx 462 drivers/crypto/sahara.c dev->hw_desc[idx]->p2 = dev->key_phys_base; idx 463 drivers/crypto/sahara.c dev->hw_desc[idx]->next = dev->hw_phys_desc[1]; idx 465 drivers/crypto/sahara.c dev->hw_desc[idx]->hdr = sahara_aes_key_hdr(dev); idx 467 drivers/crypto/sahara.c idx++; idx 500 drivers/crypto/sahara.c dev->hw_desc[idx]->p1 = dev->hw_phys_link[0]; idx 514 drivers/crypto/sahara.c dev->hw_desc[idx]->p2 = dev->hw_phys_link[i]; idx 528 drivers/crypto/sahara.c dev->hw_desc[idx]->hdr = sahara_aes_data_link_hdr(dev); idx 529 drivers/crypto/sahara.c dev->hw_desc[idx]->len1 = dev->total; idx 530 drivers/crypto/sahara.c dev->hw_desc[idx]->len2 = dev->total; idx 531 drivers/crypto/sahara.c dev->hw_desc[idx]->next = 0; idx 526 drivers/dma-buf/dma-fence.c uint32_t *idx) idx 533 drivers/dma-buf/dma-fence.c if (idx) idx 534 drivers/dma-buf/dma-fence.c *idx = i; idx 563 drivers/dma-buf/dma-fence.c bool intr, signed long timeout, uint32_t *idx) idx 575 drivers/dma-buf/dma-fence.c if (idx) idx 576 drivers/dma-buf/dma-fence.c *idx = i; idx 596 drivers/dma-buf/dma-fence.c if (idx) idx 597 drivers/dma-buf/dma-fence.c *idx = i; idx 608 drivers/dma-buf/dma-fence.c if (dma_fence_test_signaled_any(fences, count, idx)) idx 325 drivers/dma/ioat/dma.c ioat_alloc_ring_ent(struct dma_chan *chan, int idx, gfp_t flags) idx 335 drivers/dma/ioat/dma.c chunk = idx / IOAT_DESCS_PER_2M; idx 336 drivers/dma/ioat/dma.c idx &= (IOAT_DESCS_PER_2M - 1); idx 337 drivers/dma/ioat/dma.c offs = idx * IOAT_DESC_SZ; idx 381 drivers/dma/ioat/dma.c int idx; idx 383 drivers/dma/ioat/dma.c for (idx = 0; idx < i; idx++) { idx 384 drivers/dma/ioat/dma.c descs = &ioat_chan->descs[idx]; idx 400 drivers/dma/ioat/dma.c int idx; idx 405 drivers/dma/ioat/dma.c for (idx = 0; idx < ioat_chan->desc_chunks; idx++) { idx 408 drivers/dma/ioat/dma.c ioat_chan->descs[idx].virt, idx 409 drivers/dma/ioat/dma.c ioat_chan->descs[idx].hw); idx 410 drivers/dma/ioat/dma.c ioat_chan->descs[idx].virt = NULL; idx 411 drivers/dma/ioat/dma.c ioat_chan->descs[idx].hw = 0; idx 581 drivers/dma/ioat/dma.c int idx = ioat_chan->tail, i; idx 601 drivers/dma/ioat/dma.c prefetch(ioat_get_ring_ent(ioat_chan, idx + i + 1)); idx 602 drivers/dma/ioat/dma.c desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 636 drivers/dma/ioat/dma.c ioat_chan->tail = idx + i; idx 711 drivers/dma/ioat/dma.c int idx = ioat_chan->tail, i; idx 724 drivers/dma/ioat/dma.c prefetch(ioat_get_ring_ent(ioat_chan, idx + i + 1)); idx 725 drivers/dma/ioat/dma.c desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 753 drivers/dma/ioat/dma.c ioat_chan->tail = idx + active; idx 73 drivers/dma/ioat/dma.h struct ioatdma_chan *idx[IOAT_MAX_CHANS]; idx 235 drivers/dma/ioat/dma.h return ioat_dma->idx[index]; idx 342 drivers/dma/ioat/dma.h ioat_get_ring_ent(struct ioatdma_chan *ioat_chan, u16 idx) idx 344 drivers/dma/ioat/dma.h return ioat_chan->ring[idx & (ioat_ring_size(ioat_chan) - 1)]; idx 121 drivers/dma/ioat/init.c struct ioatdma_chan *ioat_chan, int idx); idx 580 drivers/dma/ioat/init.c if (dma->chancnt > ARRAY_SIZE(ioat_dma->idx)) { idx 582 drivers/dma/ioat/init.c dma->chancnt, ARRAY_SIZE(ioat_dma->idx)); idx 583 drivers/dma/ioat/init.c dma->chancnt = ARRAY_SIZE(ioat_dma->idx); idx 767 drivers/dma/ioat/init.c struct ioatdma_chan *ioat_chan, int idx) idx 774 drivers/dma/ioat/init.c ioat_chan->reg_base = ioat_dma->reg_base + (0x80 * (idx + 1)); idx 779 drivers/dma/ioat/init.c ioat_dma->idx[idx] = ioat_chan; idx 1223 drivers/dma/ioat/init.c ioat_chan = ioat_dma->idx[i]; idx 1252 drivers/dma/ioat/init.c ioat_chan = ioat_dma->idx[i]; idx 32 drivers/dma/ioat/prep.c dma_addr_t addr, u32 offset, int idx) idx 34 drivers/dma/ioat/prep.c struct ioat_raw_descriptor *raw = descs[xor_idx_to_desc >> idx & 1]; idx 36 drivers/dma/ioat/prep.c raw->field[xor_idx_to_field[idx]] = addr + offset; idx 39 drivers/dma/ioat/prep.c static dma_addr_t pq_get_src(struct ioat_raw_descriptor *descs[2], int idx) idx 41 drivers/dma/ioat/prep.c struct ioat_raw_descriptor *raw = descs[pq_idx_to_desc >> idx & 1]; idx 43 drivers/dma/ioat/prep.c return raw->field[pq_idx_to_field[idx]]; idx 46 drivers/dma/ioat/prep.c static dma_addr_t pq16_get_src(struct ioat_raw_descriptor *desc[3], int idx) idx 48 drivers/dma/ioat/prep.c struct ioat_raw_descriptor *raw = desc[pq16_idx_to_desc[idx]]; idx 50 drivers/dma/ioat/prep.c return raw->field[pq16_idx_to_field[idx]]; idx 54 drivers/dma/ioat/prep.c dma_addr_t addr, u32 offset, u8 coef, int idx) idx 57 drivers/dma/ioat/prep.c struct ioat_raw_descriptor *raw = descs[pq_idx_to_desc >> idx & 1]; idx 59 drivers/dma/ioat/prep.c raw->field[pq_idx_to_field[idx]] = addr + offset; idx 60 drivers/dma/ioat/prep.c pq->coef[idx] = coef; idx 64 drivers/dma/ioat/prep.c dma_addr_t addr, u32 offset, u8 coef, unsigned idx) idx 69 drivers/dma/ioat/prep.c struct ioat_raw_descriptor *raw = desc[pq16_idx_to_desc[idx]]; idx 71 drivers/dma/ioat/prep.c raw->field[pq16_idx_to_field[idx]] = addr + offset; idx 73 drivers/dma/ioat/prep.c if (idx < 8) idx 74 drivers/dma/ioat/prep.c pq->coef[idx] = coef; idx 76 drivers/dma/ioat/prep.c pq16->coef[idx - 8] = coef; idx 110 drivers/dma/ioat/prep.c int num_descs, idx, i; idx 118 drivers/dma/ioat/prep.c idx = ioat_chan->head; idx 125 drivers/dma/ioat/prep.c desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 164 drivers/dma/ioat/prep.c int num_descs, with_ext, idx, i; idx 187 drivers/dma/ioat/prep.c idx = ioat_chan->head; idx 197 drivers/dma/ioat/prep.c desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 204 drivers/dma/ioat/prep.c ext = ioat_get_ring_ent(ioat_chan, idx + i + 1); idx 230 drivers/dma/ioat/prep.c compl_desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 355 drivers/dma/ioat/prep.c int i, s, idx, with_ext, num_descs; idx 383 drivers/dma/ioat/prep.c idx = ioat_chan->head; idx 392 drivers/dma/ioat/prep.c desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 399 drivers/dma/ioat/prep.c ext = ioat_get_ring_ent(ioat_chan, idx + i + with_ext); idx 446 drivers/dma/ioat/prep.c compl_desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 475 drivers/dma/ioat/prep.c int i, s, idx, num_descs; idx 489 drivers/dma/ioat/prep.c idx = ioat_chan->head; idx 500 drivers/dma/ioat/prep.c desc = ioat_get_ring_ent(ioat_chan, idx + i); idx 96 drivers/dma/iop-adma.c desc->idx, desc->slots_per_op); idx 120 drivers/dma/iop-adma.c iter->async_tx.cookie, iter->idx, busy, idx 177 drivers/dma/iop-adma.c grp_iter->idx, zero_sum_result); idx 310 drivers/dma/iop-adma.c iter->idx, iter->hw_desc, idx 399 drivers/dma/iop-adma.c __func__, sw_desc->async_tx.cookie, sw_desc->idx); idx 420 drivers/dma/iop-adma.c int idx; idx 430 drivers/dma/iop-adma.c idx = iop_chan->slots_allocated; idx 431 drivers/dma/iop-adma.c if (idx == num_descs_in_pool) idx 437 drivers/dma/iop-adma.c " %d descriptor slots", idx); idx 441 drivers/dma/iop-adma.c slot->hw_desc = (void *) &hw_desc[idx * IOP_ADMA_SLOT_SIZE]; idx 450 drivers/dma/iop-adma.c (dma_addr_t) &hw_desc[idx * IOP_ADMA_SLOT_SIZE]; idx 451 drivers/dma/iop-adma.c slot->idx = idx; idx 459 drivers/dma/iop-adma.c if (idx && !iop_chan->last_used) idx 480 drivers/dma/iop-adma.c return (idx > 0) ? idx : -ENOMEM; idx 288 drivers/dma/iop-adma.h return (desc->idx & (num_slots - 1)) ? 0 : 1; idx 92 drivers/dma/k3dma.c u32 idx; idx 155 drivers/dma/k3dma.c val = 0x1 << phy->idx; idx 175 drivers/dma/k3dma.c cnt = readl_relaxed(d->base + CX_CUR_CNT + phy->idx * 0x10); idx 273 drivers/dma/k3dma.c if (BIT(c->phy->idx) & k3_dma_get_chan_stat(d)) idx 314 drivers/dma/k3dma.c dev_dbg(d->slave.dev, "pchan %u: free\n", p->idx); idx 913 drivers/dma/k3dma.c p->idx = i; idx 260 drivers/dma/mic_x100_dma.c u32 idx = mic_dma_hw_ring_dec(ch->head); idx 261 drivers/dma/mic_x100_dma.c struct dma_async_tx_descriptor *tx = &ch->tx_array[idx]; idx 118 drivers/dma/mmp_pdma.c int idx; idx 147 drivers/dma/mmp_pdma.c u32 reg = (phy->idx << 4) + DDADR; idx 160 drivers/dma/mmp_pdma.c writel(DRCMR_MAPVLD | phy->idx, phy->base + reg); idx 164 drivers/dma/mmp_pdma.c dalgn |= 1 << phy->idx; idx 166 drivers/dma/mmp_pdma.c dalgn &= ~(1 << phy->idx); idx 169 drivers/dma/mmp_pdma.c reg = (phy->idx << 2) + DCSR; idx 180 drivers/dma/mmp_pdma.c reg = (phy->idx << 2) + DCSR; idx 188 drivers/dma/mmp_pdma.c u32 reg = (phy->idx << 2) + DCSR; idx 190 drivers/dma/mmp_pdma.c if (!(dint & BIT(phy->idx))) idx 785 drivers/dma/mmp_pdma.c curr = readl(chan->phy->base + DTADR(chan->phy->idx)); idx 787 drivers/dma/mmp_pdma.c curr = readl(chan->phy->base + DSADR(chan->phy->idx)); idx 969 drivers/dma/mmp_pdma.c static int mmp_pdma_chan_init(struct mmp_pdma_device *pdev, int idx, int irq) idx 971 drivers/dma/mmp_pdma.c struct mmp_pdma_phy *phy = &pdev->phy[idx]; idx 979 drivers/dma/mmp_pdma.c phy->idx = idx; idx 117 drivers/dma/mmp_tdma.c int idx; idx 306 drivers/dma/mmp_tdma.c if (tdmac->idx == 0) { idx 309 drivers/dma/mmp_tdma.c } else if (tdmac->idx == 1) { idx 554 drivers/dma/mmp_tdma.c int idx, int irq, idx 559 drivers/dma/mmp_tdma.c if (idx >= TDMA_CHANNEL_NUM) { idx 573 drivers/dma/mmp_tdma.c tdmac->idx = idx; idx 575 drivers/dma/mmp_tdma.c tdmac->reg_base = tdev->base + idx * 4; idx 578 drivers/dma/mmp_tdma.c tdev->tdmac[tdmac->idx] = tdmac; idx 404 drivers/dma/moxart-dma.c static void moxart_dma_start_sg(struct moxart_chan *ch, unsigned int idx) idx 407 drivers/dma/moxart-dma.c struct moxart_sg *sg = ch->desc->sg + idx; idx 112 drivers/dma/mv_xor.c val |= XOR_INTR_MASK_VALUE << (chan->idx * 16); idx 119 drivers/dma/mv_xor.c intr_cause = (intr_cause >> (chan->idx * 16)) & 0xFFFF; idx 128 drivers/dma/mv_xor.c val = ~(val << (chan->idx * 16)); idx 135 drivers/dma/mv_xor.c u32 val = 0xFFFF0000 >> (chan->idx * 16); idx 438 drivers/dma/mv_xor.c int idx; idx 444 drivers/dma/mv_xor.c idx = mv_chan->slots_allocated; idx 445 drivers/dma/mv_xor.c while (idx < num_descs_in_pool) { idx 450 drivers/dma/mv_xor.c idx); idx 454 drivers/dma/mv_xor.c slot->hw_desc = virt_desc + idx * MV_XOR_SLOT_SIZE; idx 461 drivers/dma/mv_xor.c slot->async_tx.phys = dma_desc + idx * MV_XOR_SLOT_SIZE; idx 462 drivers/dma/mv_xor.c slot->idx = idx++; idx 465 drivers/dma/mv_xor.c mv_chan->slots_allocated = idx; idx 724 drivers/dma/mv_xor.c chan->idx, intr_cause); idx 1036 drivers/dma/mv_xor.c int idx, dma_cap_mask_t cap_mask, int irq) idx 1046 drivers/dma/mv_xor.c mv_chan->idx = idx; idx 36 drivers/dma/mv_xor.h #define XOR_CURR_DESC(chan) (chan->mmr_high_base + 0x10 + (chan->idx * 4)) idx 37 drivers/dma/mv_xor.h #define XOR_NEXT_DESC(chan) (chan->mmr_high_base + 0x00 + (chan->idx * 4)) idx 38 drivers/dma/mv_xor.h #define XOR_BYTE_COUNT(chan) (chan->mmr_high_base + 0x20 + (chan->idx * 4)) idx 39 drivers/dma/mv_xor.h #define XOR_DEST_POINTER(chan) (chan->mmr_high_base + 0xB0 + (chan->idx * 4)) idx 40 drivers/dma/mv_xor.h #define XOR_BLOCK_SIZE(chan) (chan->mmr_high_base + 0xC0 + (chan->idx * 4)) idx 44 drivers/dma/mv_xor.h #define XOR_CONFIG(chan) (chan->mmr_base + 0x10 + (chan->idx * 4)) idx 45 drivers/dma/mv_xor.h #define XOR_ACTIVATION(chan) (chan->mmr_base + 0x20 + (chan->idx * 4)) idx 109 drivers/dma/mv_xor.h unsigned int idx; idx 146 drivers/dma/mv_xor.h u16 idx; idx 190 drivers/dma/mv_xor.h #define mv_hw_desc_slot_idx(hw_desc, idx) \ idx 191 drivers/dma/mv_xor.h ((void *)(((unsigned long)hw_desc) + ((idx) << 5))) idx 178 drivers/dma/mv_xor_v2.c int idx; idx 397 drivers/dma/mv_xor_v2.c hw_descriptor->desc_id = sw_desc->idx; idx 453 drivers/dma/mv_xor_v2.c hw_descriptor->desc_id = sw_desc->idx; idx 501 drivers/dma/mv_xor_v2.c hw_descriptor->desc_id = sw_desc->idx; idx 818 drivers/dma/mv_xor_v2.c sw_desc->idx = i; idx 505 drivers/dma/mxs-dma.c int idx = 0; idx 508 drivers/dma/mxs-dma.c idx = mxs_chan->desc_count; idx 510 drivers/dma/mxs-dma.c if (sg_len + idx > NUM_CCW) { idx 524 drivers/dma/mxs-dma.c if (idx) { idx 525 drivers/dma/mxs-dma.c BUG_ON(idx < 1); idx 526 drivers/dma/mxs-dma.c ccw = &mxs_chan->ccw[idx - 1]; idx 527 drivers/dma/mxs-dma.c ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx; idx 532 drivers/dma/mxs-dma.c idx = 0; idx 536 drivers/dma/mxs-dma.c ccw = &mxs_chan->ccw[idx++]; idx 561 drivers/dma/mxs-dma.c ccw = &mxs_chan->ccw[idx++]; idx 563 drivers/dma/mxs-dma.c ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx; idx 584 drivers/dma/mxs-dma.c mxs_chan->desc_count = idx; idx 1001 drivers/dma/pl330.c int idx; idx 1007 drivers/dma/pl330.c idx = 1 - thrd->lstenq; idx 1008 drivers/dma/pl330.c if (thrd->req[idx].desc != NULL) { idx 1009 drivers/dma/pl330.c req = &thrd->req[idx]; idx 1011 drivers/dma/pl330.c idx = thrd->lstenq; idx 1012 drivers/dma/pl330.c if (thrd->req[idx].desc != NULL) idx 1013 drivers/dma/pl330.c req = &thrd->req[idx]; idx 1023 drivers/dma/pl330.c if (idx == thrd->req_running) idx 1046 drivers/dma/pl330.c thrd->req_running = idx; idx 1466 drivers/dma/pl330.c unsigned idx; idx 1515 drivers/dma/pl330.c idx = thrd->req[0].desc == NULL ? 0 : 1; idx 1521 drivers/dma/pl330.c ret = _setup_req(pl330, 1, thrd, idx, &xs); idx 1533 drivers/dma/pl330.c thrd->lstenq = idx; idx 1534 drivers/dma/pl330.c thrd->req[idx].desc = desc; idx 1535 drivers/dma/pl330.c _setup_req(pl330, 0, thrd, idx, &xs); idx 786 drivers/dma/ppc4xx/adma.c return (desc->idx & (num_slots - 1)) ? 0 : 1; idx 1256 drivers/dma/ppc4xx/adma.c int idx = 0; idx 1267 drivers/dma/ppc4xx/adma.c ppc440spe_rxor_srcs[idx++] = srcs[i]; idx 1269 drivers/dma/ppc4xx/adma.c src_cnt = idx; idx 1521 drivers/dma/ppc4xx/adma.c desc->phys, desc->idx, desc->slots_per_op); idx 1561 drivers/dma/ppc4xx/adma.c iter->async_tx.cookie, iter->idx, busy, iter->phys, idx 1802 drivers/dma/ppc4xx/adma.c slot->idx = i; idx 1870 drivers/dma/ppc4xx/adma.c u8 xor_arg_no, u8 idx, u8 mult) idx 1874 drivers/dma/ppc4xx/adma.c xcb->ops[xor_arg_no].h |= mult << (DMA_CUED_MULT1_OFF + idx * 8); idx 1938 drivers/dma/ppc4xx/adma.c sw_desc->async_tx.cookie, sw_desc->idx, sw_desc); idx 3028 drivers/dma/ppc4xx/adma.c int idx; idx 3035 drivers/dma/ppc4xx/adma.c idx = (paddr && qaddr) ? 2 : 1; idx 3039 drivers/dma/ppc4xx/adma.c if (!(--idx)) idx 3043 drivers/dma/ppc4xx/adma.c idx = (paddr && qaddr) ? 2 : 1; idx 3044 drivers/dma/ppc4xx/adma.c iter = ppc440spe_get_group_entry(sw_desc, idx); idx 4015 drivers/dma/ppc4xx/adma.c const u32 *idx; idx 4029 drivers/dma/ppc4xx/adma.c idx = of_get_property(np, "cell-index", &len); idx 4030 drivers/dma/ppc4xx/adma.c if (!idx || (len != sizeof(u32))) { idx 4036 drivers/dma/ppc4xx/adma.c id = *idx; idx 157 drivers/dma/ppc4xx/adma.h u16 idx; idx 98 drivers/dma/pxa_dma.c int idx; idx 146 drivers/dma/pxa_dma.c readl_relaxed((phy)->base + _reg((phy)->idx)) idx 150 drivers/dma/pxa_dma.c _v = readl_relaxed((phy)->base + _reg((phy)->idx)); \ idx 158 drivers/dma/pxa_dma.c writel((val), (phy)->base + _reg((phy)->idx)); \ idx 165 drivers/dma/pxa_dma.c writel_relaxed((val), (phy)->base + _reg((phy)->idx)); \ idx 194 drivers/dma/pxa_dma.c seq_printf(s, "DMA channel %d requester :\n", phy->idx); idx 197 drivers/dma/pxa_dma.c if ((drcmr & DRCMR_CHLNUM) == phy->idx) idx 229 drivers/dma/pxa_dma.c seq_printf(s, "DMA channel %d descriptors :\n", phy->idx); idx 276 drivers/dma/pxa_dma.c seq_printf(s, "DMA channel %d\n", phy->idx); idx 278 drivers/dma/pxa_dma.c str_prio[(phy->idx & 0xf) / 4]); idx 280 drivers/dma/pxa_dma.c _phy_readl_relaxed(phy, DALGN) & BIT(phy->idx) ? idx 402 drivers/dma/pxa_dma.c found ? found->idx : -1); idx 447 drivers/dma/pxa_dma.c return dalgn & (BIT(chan->phy->idx)); idx 460 drivers/dma/pxa_dma.c phy, phy->idx, misaligned); idx 465 drivers/dma/pxa_dma.c writel_relaxed(DRCMR_MAPVLD | phy->idx, phy->base + reg); idx 470 drivers/dma/pxa_dma.c dalgn |= BIT(phy->idx); idx 472 drivers/dma/pxa_dma.c dalgn &= ~BIT(phy->idx); idx 488 drivers/dma/pxa_dma.c "%s(): phy=%p(%d)\n", __func__, phy, phy->idx); idx 589 drivers/dma/pxa_dma.c if (!(dint & BIT(phy->idx))) idx 1265 drivers/dma/pxa_dma.c phy->idx = i; idx 32 drivers/dma/qcom/hidma.h u32 idx; /* index of the tre */ idx 35 drivers/dma/qcom/hidma_dbg.c seq_printf(s, "idx = 0x%x\n", tre->idx); idx 158 drivers/dma/qcom/hidma_ll.c tre->idx = i; idx 715 drivers/dma/ste_dma40.c int idx = d40c->phy_chan->num * D40_LCLA_LINK_PER_EVENT_GRP + i; idx 717 drivers/dma/ste_dma40.c if (!d40c->base->lcla_pool.alloc_map[idx]) { idx 718 drivers/dma/ste_dma40.c d40c->base->lcla_pool.alloc_map[idx] = d40d; idx 743 drivers/dma/ste_dma40.c int idx = d40c->phy_chan->num * D40_LCLA_LINK_PER_EVENT_GRP + i; idx 745 drivers/dma/ste_dma40.c if (d40c->base->lcla_pool.alloc_map[idx] == d40d) { idx 746 drivers/dma/ste_dma40.c d40c->base->lcla_pool.alloc_map[idx] = NULL; idx 1642 drivers/dma/ste_dma40.c u32 idx; idx 1668 drivers/dma/ste_dma40.c idx = chan & (BITS_PER_LONG - 1); idx 1671 drivers/dma/ste_dma40.c d40c = base->lookup_phy_chans[idx]; idx 1673 drivers/dma/ste_dma40.c d40c = base->lookup_log_chans[il[row].offset + idx]; idx 1684 drivers/dma/ste_dma40.c writel(BIT(idx), base->virtbase + il[row].clr); idx 1692 drivers/dma/ste_dma40.c chan, il[row].offset, idx); idx 2959 drivers/dma/ste_dma40.c int idx; idx 2965 drivers/dma/ste_dma40.c idx = i * ARRAY_SIZE(d40_backup_regs_chan); idx 2967 drivers/dma/ste_dma40.c dma40_backup(addr, &base->reg_val_backup_chan[idx], idx 168 drivers/dma/sun6i-dma.c u32 idx; idx 255 drivers/dma/sun6i-dma.c pchan->idx, ®, idx 450 drivers/dma/sun6i-dma.c irq_reg = pchan->idx / DMA_IRQ_CHAN_NR; idx 451 drivers/dma/sun6i-dma.c irq_offset = pchan->idx % DMA_IRQ_CHAN_NR; idx 489 drivers/dma/sun6i-dma.c pchan->idx); idx 517 drivers/dma/sun6i-dma.c pchan->idx, &vchan->vc); idx 1351 drivers/dma/sun6i-dma.c pchan->idx = i; idx 855 drivers/dma/ti/cppi41.c u32 idx; idx 865 drivers/dma/ti/cppi41.c idx = 0; idx 868 drivers/dma/ti/cppi41.c reg = idx << QMGR_MEMCTRL_IDX_SH; idx 881 drivers/dma/ti/cppi41.c idx += ALLOC_DECS_NUM; idx 450 drivers/dma/ti/edma.c int idx = EDMA_REG_ARRAY_INDEX(channel); idx 454 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_ICR, idx, ch_bit); idx 455 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_IESR, idx, ch_bit); idx 457 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_IECR, idx, ch_bit); idx 595 drivers/dma/ti/edma.c int idx = EDMA_REG_ARRAY_INDEX(channel); idx 600 drivers/dma/ti/edma.c dev_dbg(ecc->dev, "ESR%d %08x\n", idx, idx 601 drivers/dma/ti/edma.c edma_shadow0_read_array(ecc, SH_ESR, idx)); idx 602 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_ESR, idx, ch_bit); idx 605 drivers/dma/ti/edma.c dev_dbg(ecc->dev, "ER%d %08x\n", idx, idx 606 drivers/dma/ti/edma.c edma_shadow0_read_array(ecc, SH_ER, idx)); idx 608 drivers/dma/ti/edma.c edma_write_array(ecc, EDMA_ECR, idx, ch_bit); idx 609 drivers/dma/ti/edma.c edma_write_array(ecc, EDMA_EMCR, idx, ch_bit); idx 611 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_SECR, idx, ch_bit); idx 612 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_EESR, idx, ch_bit); idx 613 drivers/dma/ti/edma.c dev_dbg(ecc->dev, "EER%d %08x\n", idx, idx 614 drivers/dma/ti/edma.c edma_shadow0_read_array(ecc, SH_EER, idx)); idx 622 drivers/dma/ti/edma.c int idx = EDMA_REG_ARRAY_INDEX(channel); idx 625 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_EECR, idx, ch_bit); idx 626 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_ECR, idx, ch_bit); idx 627 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_SECR, idx, ch_bit); idx 628 drivers/dma/ti/edma.c edma_write_array(ecc, EDMA_EMCR, idx, ch_bit); idx 631 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_ICR, idx, ch_bit); idx 633 drivers/dma/ti/edma.c dev_dbg(ecc->dev, "EER%d %08x\n", idx, idx 634 drivers/dma/ti/edma.c edma_shadow0_read_array(ecc, SH_EER, idx)); idx 668 drivers/dma/ti/edma.c int idx = EDMA_REG_ARRAY_INDEX(channel); idx 671 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_ESR, idx, ch_bit); idx 673 drivers/dma/ti/edma.c dev_dbg(ecc->dev, "ESR%d %08x\n", idx, idx 674 drivers/dma/ti/edma.c edma_shadow0_read_array(ecc, SH_ESR, idx)); idx 681 drivers/dma/ti/edma.c int idx = EDMA_REG_ARRAY_INDEX(channel); idx 684 drivers/dma/ti/edma.c dev_dbg(ecc->dev, "EMR%d %08x\n", idx, idx 685 drivers/dma/ti/edma.c edma_read_array(ecc, EDMA_EMR, idx)); idx 686 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_ECR, idx, ch_bit); idx 688 drivers/dma/ti/edma.c edma_write_array(ecc, EDMA_EMCR, idx, ch_bit); idx 690 drivers/dma/ti/edma.c edma_shadow0_write_array(ecc, SH_SECR, idx, ch_bit); idx 1743 drivers/dma/ti/edma.c int idx = EDMA_REG_ARRAY_INDEX(channel); idx 1768 drivers/dma/ti/edma.c while (edma_shadow0_read_array(echan->ecc, event_reg, idx) & ch_bit) { idx 244 drivers/dma/ti/omap-dma.c static void omap_dma_fill_type2_desc(struct omap_desc *d, int idx, idx 247 drivers/dma/ti/omap-dma.c struct omap_sg *sg = &d->sg[idx]; idx 250 drivers/dma/ti/omap-dma.c if (idx) idx 251 drivers/dma/ti/omap-dma.c d->sg[idx - 1].t2_desc->next_desc = sg->t2_desc_paddr; idx 382 drivers/dma/xgene-dma.c static __le64 *xgene_dma_lookup_ext8(struct xgene_dma_desc_hw *desc, int idx) idx 384 drivers/dma/xgene-dma.c switch (idx) { idx 110 drivers/dma/zx_dma.c u32 idx; idx 147 drivers/dma/zx_dma.c val = 0x1 << phy->idx; idx 195 drivers/dma/zx_dma.c if (BIT(c->phy->idx) & zx_dma_get_chan_stat(d)) idx 231 drivers/dma/zx_dma.c dev_dbg(d->slave.dev, "pchan %u: free\n", p->idx); idx 806 drivers/dma/zx_dma.c p->idx = i; idx 221 drivers/edac/amd64_edac_inj.c struct attribute *attr, int idx) idx 609 drivers/edac/edac_device_sysfs.c int idx) idx 616 drivers/edac/edac_device_sysfs.c instance = &edac_dev->instances[idx]; idx 643 drivers/edac/edac_device_sysfs.c instance->nr_blocks, idx); idx 660 drivers/edac/edac_device_sysfs.c idx, instance->name); idx 677 drivers/edac/edac_device_sysfs.c int idx) idx 682 drivers/edac/edac_device_sysfs.c instance = &edac_dev->instances[idx]; idx 673 drivers/edac/edac_mc.c struct mem_ctl_info *edac_mc_find(int idx) idx 682 drivers/edac/edac_mc.c if (mci->mc_idx == idx) idx 176 drivers/edac/edac_mc.h extern struct mem_ctl_info *edac_mc_find(int idx); idx 353 drivers/edac/edac_mc_sysfs.c struct attribute *attr, int idx) idx 358 drivers/edac/edac_mc_sysfs.c if (idx >= csrow->nr_channels) idx 361 drivers/edac/edac_mc_sysfs.c if (idx >= ARRAY_SIZE(dynamic_csrow_ce_count_attr) - 1) { idx 362 drivers/edac/edac_mc_sysfs.c WARN_ONCE(1, "idx: %d\n", idx); idx 367 drivers/edac/edac_mc_sysfs.c if (!csrow->channels[idx]->dimm->nr_pages) idx 875 drivers/edac/edac_mc_sysfs.c struct attribute *attr, int idx) idx 156 drivers/edac/edac_pci_sysfs.c static int edac_pci_create_instance_kobj(struct edac_pci_ctl_info *pci, int idx) idx 175 drivers/edac/edac_pci_sysfs.c edac_pci_top_main_kobj, "pci%d", idx); idx 177 drivers/edac/edac_pci_sysfs.c edac_dbg(2, "failed to register instance pci%d\n", idx); idx 183 drivers/edac/edac_pci_sysfs.c edac_dbg(1, "Register instance 'pci%d' kobject\n", idx); idx 483 drivers/edac/ghes_edac.c int idx = -1; idx 487 drivers/edac/ghes_edac.c idx = acpi_match_platform_list(plat_list); idx 488 drivers/edac/ghes_edac.c if (!force_load && idx < 0) idx 491 drivers/edac/ghes_edac.c idx = 0; idx 539 drivers/edac/ghes_edac.c } else if (idx < 0) { idx 900 drivers/edac/pnd2_edac.c static int bank_hash(u64 pmiaddr, int idx, int shft) idx 904 drivers/edac/pnd2_edac.c switch (idx) { idx 934 drivers/edac/pnd2_edac.c int i, idx, type, skiprs = 0; idx 945 drivers/edac/pnd2_edac.c idx = d->bits[i + skiprs] & 0xf; idx 954 drivers/edac/pnd2_edac.c idx = d->bits[i + skiprs] & 0xf; idx 959 drivers/edac/pnd2_edac.c column |= (bit << idx); idx 962 drivers/edac/pnd2_edac.c bank |= (bit << idx); idx 964 drivers/edac/pnd2_edac.c bank ^= bank_hash(pmiaddr, idx, d->addrdec); idx 967 drivers/edac/pnd2_edac.c row |= (bit << idx); idx 1095 drivers/edac/sb_edac.c static int haswell_chan_hash(int idx, u64 addr) idx 1104 drivers/edac/sb_edac.c idx ^= (addr >> i) & 3; idx 1106 drivers/edac/sb_edac.c return idx; idx 1938 drivers/edac/sb_edac.c int sad_interl, idx, base_ch; idx 2013 drivers/edac/sb_edac.c idx = ((addr >> 6) ^ (addr >> 16)) & 7; idx 2015 drivers/edac/sb_edac.c idx = (addr >> 6) & 7; idx 2018 drivers/edac/sb_edac.c idx = 0; idx 2021 drivers/edac/sb_edac.c idx = idx & 1; idx 2024 drivers/edac/sb_edac.c idx = idx & 3; idx 2032 drivers/edac/sb_edac.c *socket = sad_interleave[idx]; idx 2034 drivers/edac/sb_edac.c idx, sad_way, *socket); idx 2047 drivers/edac/sb_edac.c idx = GET_BITFIELD(addr, 16, 18); idx 2048 drivers/edac/sb_edac.c idx ^= bits; idx 2050 drivers/edac/sb_edac.c idx = bits; idx 2052 drivers/edac/sb_edac.c pkg = sad_pkg(pvt->info.interleave_pkg, reg, idx); idx 2063 drivers/edac/sb_edac.c idx, *socket, sad_ha, shiftup); idx 2066 drivers/edac/sb_edac.c idx = (addr >> 6) & 7; idx 2067 drivers/edac/sb_edac.c pkg = sad_pkg(pvt->info.interleave_pkg, reg, idx); idx 2071 drivers/edac/sb_edac.c idx, *socket, sad_ha); idx 2114 drivers/edac/sb_edac.c idx = addr >> 6; idx 2116 drivers/edac/sb_edac.c idx = (addr >> (6 + sck_way + shiftup)) & 0x3; idx 2118 drivers/edac/sb_edac.c idx = haswell_chan_hash(idx, addr); idx 2120 drivers/edac/sb_edac.c idx = idx % ch_way; idx 2125 drivers/edac/sb_edac.c switch (idx) { idx 2177 drivers/edac/sb_edac.c idx, idx 2223 drivers/edac/sb_edac.c idx = (ch_addr >> 6); idx 2225 drivers/edac/sb_edac.c idx = (ch_addr >> 13); /* FIXME: Datasheet says to shift by 15 */ idx 2226 drivers/edac/sb_edac.c idx %= 1 << rir_way; idx 2228 drivers/edac/sb_edac.c pci_read_config_dword(pvt->pci_tad[base_ch], rir_offset[n_rir][idx], ®); idx 2236 drivers/edac/sb_edac.c idx); idx 36 drivers/edac/skx_base.c static struct skx_dev *get_skx_dev(struct pci_bus *bus, u8 idx) idx 41 drivers/edac/skx_base.c if (d->seg == pci_domain_nr(bus) && d->bus[idx] == bus->number) idx 219 drivers/edac/skx_base.c int i, idx, tgt, lchan, shift; idx 249 drivers/edac/skx_base.c idx = GET_BITFIELD(addr, 6, 8); idx 252 drivers/edac/skx_base.c idx = GET_BITFIELD(addr, 8, 10); idx 255 drivers/edac/skx_base.c idx = GET_BITFIELD(addr, 12, 14); idx 258 drivers/edac/skx_base.c idx = GET_BITFIELD(addr, 30, 32); idx 262 drivers/edac/skx_base.c tgt = GET_BITFIELD(ilv, 4 * idx, 4 * idx + 3); idx 404 drivers/edac/skx_base.c #define SKX_GET_RIRILV(d, mc, ch, idx, i, reg) \ idx 406 drivers/edac/skx_base.c 0x120 + 16 * (idx) + 4 * (i), &(reg)) idx 416 drivers/edac/skx_base.c int i, idx, chan_rank; idx 446 drivers/edac/skx_base.c idx = (res->chan_addr >> shift) % SKX_RIR_WAYS(rirway); idx 448 drivers/edac/skx_base.c SKX_GET_RIRILV(res->dev, res->imc, res->channel, idx, i, rirlv); idx 1348 drivers/edac/thunderx_edac.c int idx; idx 1365 drivers/edac/thunderx_edac.c idx = edac_device_alloc_index(); idx 1366 drivers/edac/thunderx_edac.c snprintf(name, sizeof(name), "OCX%d", idx); idx 1369 drivers/edac/thunderx_edac.c 0, NULL, 0, idx); idx 1956 drivers/edac/thunderx_edac.c int idx; idx 2003 drivers/edac/thunderx_edac.c idx = edac_device_alloc_index(); idx 2004 drivers/edac/thunderx_edac.c snprintf(name, sizeof(name), fmt, idx); idx 2008 drivers/edac/thunderx_edac.c NULL, 0, idx); idx 903 drivers/extcon/extcon.c int ret, idx = -EINVAL; idx 908 drivers/extcon/extcon.c idx = find_cable_index_by_id(edev, id); idx 909 drivers/extcon/extcon.c if (idx < 0) idx 910 drivers/extcon/extcon.c return idx; idx 913 drivers/extcon/extcon.c ret = raw_notifier_chain_register(&edev->nh[idx], nb); idx 932 drivers/extcon/extcon.c int ret, idx; idx 937 drivers/extcon/extcon.c idx = find_cable_index_by_id(edev, id); idx 938 drivers/extcon/extcon.c if (idx < 0) idx 939 drivers/extcon/extcon.c return idx; idx 942 drivers/extcon/extcon.c ret = raw_notifier_chain_unregister(&edev->nh[idx], nb); idx 716 drivers/firmware/arm_scmi/driver.c static int scmi_mailbox_check(struct device_node *np, int idx) idx 719 drivers/firmware/arm_scmi/driver.c idx, NULL); idx 725 drivers/firmware/arm_scmi/driver.c int ret, idx; idx 735 drivers/firmware/arm_scmi/driver.c idx = tx ? 0 : 1; idx 738 drivers/firmware/arm_scmi/driver.c if (scmi_mailbox_check(np, idx)) { idx 758 drivers/firmware/arm_scmi/driver.c shmem = of_parse_phandle(np, "shmem", idx); idx 773 drivers/firmware/arm_scmi/driver.c cinfo->chan = mbox_request_channel(cl, idx); idx 598 drivers/firmware/arm_scmi/perf.c int idx, ret, domain; idx 610 drivers/firmware/arm_scmi/perf.c for (opp = dom->opp, idx = 0; idx < dom->opp_count; idx++, opp++) { idx 617 drivers/firmware/arm_scmi/perf.c while (idx-- > 0) { idx 673 drivers/firmware/arm_scmi/perf.c int idx, ret = -EINVAL; idx 680 drivers/firmware/arm_scmi/perf.c for (opp = dom->opp, idx = 0; idx < dom->opp_count; idx++, opp++) { idx 470 drivers/firmware/arm_scpi.c static int scpi_send_message(u8 idx, void *tx_buf, unsigned int tx_len, idx 479 drivers/firmware/arm_scpi.c if (scpi_info->commands[idx] < 0) idx 482 drivers/firmware/arm_scpi.c cmd = scpi_info->commands[idx]; idx 685 drivers/firmware/arm_scpi.c int idx, ret; idx 695 drivers/firmware/arm_scpi.c for (opp = info->opps, idx = 0; idx < info->count; idx++, opp++) { idx 700 drivers/firmware/arm_scpi.c while (idx-- > 0) idx 902 drivers/firmware/arm_scpi.c int count, idx, ret; idx 931 drivers/firmware/arm_scpi.c int idx = scpi_info->num_chans; idx 932 drivers/firmware/arm_scpi.c struct scpi_chan *pchan = scpi_info->channels + idx; idx 934 drivers/firmware/arm_scpi.c struct device_node *shmem = of_parse_phandle(np, "shmem", idx); idx 965 drivers/firmware/arm_scpi.c pchan->chan = mbox_request_channel(cl, idx); idx 971 drivers/firmware/arm_scpi.c idx, ret); idx 986 drivers/firmware/arm_scpi.c for (idx = 0; idx < ARRAY_SIZE(legacy_hpriority_cmds); idx++) idx 987 drivers/firmware/arm_scpi.c set_bit(legacy_hpriority_cmds[idx], idx 25 drivers/firmware/imx/imx-dsp.c int imx_dsp_ring_doorbell(struct imx_dsp_ipc *ipc, unsigned int idx) idx 30 drivers/firmware/imx/imx-dsp.c if (idx >= DSP_MU_CHAN_NUM) idx 33 drivers/firmware/imx/imx-dsp.c dsp_chan = &ipc->chans[idx]; idx 55 drivers/firmware/imx/imx-dsp.c if (chan->idx == 0) { idx 96 drivers/firmware/imx/imx-dsp.c dsp_chan->idx = i % 2; idx 31 drivers/firmware/imx/imx-scu.c int idx; idx 123 drivers/firmware/imx/imx-scu.c sc_chan->idx, *data); idx 140 drivers/firmware/imx/imx-scu.c if (sc_chan->idx == 0) { idx 149 drivers/firmware/imx/imx-scu.c sc_ipc->msg[sc_chan->idx] = *data; idx 152 drivers/firmware/imx/imx-scu.c dev_dbg(sc_ipc->dev, "mu %u msg %u 0x%x\n", sc_chan->idx, idx 296 drivers/firmware/imx/imx-scu.c sc_chan->idx = i % (num_channel / 2); idx 232 drivers/firmware/imx/scu-pd.c imx_scu_add_pm_domain(struct device *dev, int idx, idx 242 drivers/firmware/imx/scu-pd.c sc_pd->rsrc = pd_ranges->rsrc + idx; idx 248 drivers/firmware/imx/scu-pd.c "%s%i", pd_ranges->name, pd_ranges->start_from + idx); idx 308 drivers/firmware/qcom_scm.c unsigned long idx) idx 310 drivers/firmware/qcom_scm.c if (idx != 0) idx 317 drivers/firmware/qcom_scm.c unsigned long idx) idx 319 drivers/firmware/qcom_scm.c if (idx != 0) idx 527 drivers/fsi/fsi-core.c slave->master->idx, slave->link, idx 1009 drivers/fsi/fsi-core.c cfam_id, master->idx, link, id); idx 1273 drivers/fsi/fsi-core.c master->idx = ida_simple_get(&master_ida, 0, INT_MAX, GFP_KERNEL); idx 1274 drivers/fsi/fsi-core.c dev_set_name(&master->dev, "fsi%d", master->idx); idx 1278 drivers/fsi/fsi-core.c ida_simple_remove(&master_ida, master->idx); idx 1285 drivers/fsi/fsi-core.c ida_simple_remove(&master_ida, master->idx); idx 1292 drivers/fsi/fsi-core.c ida_simple_remove(&master_ida, master->idx); idx 1309 drivers/fsi/fsi-core.c if (master->idx >= 0) { idx 1310 drivers/fsi/fsi-core.c ida_simple_remove(&master_ida, master->idx); idx 1311 drivers/fsi/fsi-core.c master->idx = -1; idx 129 drivers/fsi/fsi-master-hub.c int idx, bit; idx 133 drivers/fsi/fsi-master-hub.c idx = link / 32; idx 138 drivers/fsi/fsi-master-hub.c rc = fsi_device_write(hub->upstream, FSI_MSENP0 + (4 * idx), ®, 4); idx 142 drivers/fsi/fsi-master-hub.c fsi_device_read(hub->upstream, FSI_MENP0 + (4 * idx), ®, 4); idx 52 drivers/fsi/fsi-master.h int idx; idx 44 drivers/fsi/fsi-occ.c int idx; idx 528 drivers/fsi/fsi-occ.c occ->idx = ida_simple_get(&occ_ida, reg, reg + 1, idx 530 drivers/fsi/fsi-occ.c if (occ->idx < 0) idx 531 drivers/fsi/fsi-occ.c occ->idx = ida_simple_get(&occ_ida, 1, INT_MAX, idx 534 drivers/fsi/fsi-occ.c occ->idx = ida_simple_get(&occ_ida, 1, INT_MAX, idx 538 drivers/fsi/fsi-occ.c occ->idx = ida_simple_get(&occ_ida, 1, INT_MAX, GFP_KERNEL); idx 543 drivers/fsi/fsi-occ.c snprintf(occ->name, sizeof(occ->name), "occ%d", occ->idx); idx 552 drivers/fsi/fsi-occ.c ida_simple_remove(&occ_ida, occ->idx); idx 556 drivers/fsi/fsi-occ.c hwmon_dev_info.id = occ->idx; idx 572 drivers/fsi/fsi-occ.c ida_simple_remove(&occ_ida, occ->idx); idx 75 drivers/gpio/gpio-74x164.c unsigned int i, idx, shift; idx 81 drivers/gpio/gpio-74x164.c idx = i / sizeof(*mask); idx 83 drivers/gpio/gpio-74x164.c bankmask = mask[idx] >> shift; idx 88 drivers/gpio/gpio-74x164.c chip->buffer[bank] |= bankmask & (bits[idx] >> shift); idx 89 drivers/gpio/gpio-dwapb.c unsigned int idx; idx 161 drivers/gpio/gpio-dwapb.c if (port->idx == offs / 32) idx 500 drivers/gpio/gpio-dwapb.c port->idx = pp->idx; idx 508 drivers/gpio/gpio-dwapb.c dat = gpio->regs + GPIO_EXT_PORTA + (pp->idx * GPIO_EXT_PORT_STRIDE); idx 509 drivers/gpio/gpio-dwapb.c set = gpio->regs + GPIO_SWPORTA_DR + (pp->idx * GPIO_SWPORT_DR_STRIDE); idx 511 drivers/gpio/gpio-dwapb.c (pp->idx * GPIO_SWPORT_DDR_STRIDE); idx 518 drivers/gpio/gpio-dwapb.c port->idx); idx 529 drivers/gpio/gpio-dwapb.c if (pp->idx == 0) idx 538 drivers/gpio/gpio-dwapb.c port->idx); idx 588 drivers/gpio/gpio-dwapb.c if (fwnode_property_read_u32(fwnode, "reg", &pp->idx) || idx 589 drivers/gpio/gpio-dwapb.c pp->idx >= DWAPB_MAX_PORTS) { idx 611 drivers/gpio/gpio-dwapb.c if (pp->idx != 0) idx 632 drivers/gpio/gpio-dwapb.c dev_warn(dev, "no irq for port%d\n", pp->idx); idx 755 drivers/gpio/gpio-dwapb.c unsigned int idx = gpio->ports[i].idx; idx 760 drivers/gpio/gpio-dwapb.c offset = GPIO_SWPORTA_DDR + idx * GPIO_SWPORT_DDR_STRIDE; idx 763 drivers/gpio/gpio-dwapb.c offset = GPIO_SWPORTA_DR + idx * GPIO_SWPORT_DR_STRIDE; idx 766 drivers/gpio/gpio-dwapb.c offset = GPIO_EXT_PORTA + idx * GPIO_EXT_PORT_STRIDE; idx 770 drivers/gpio/gpio-dwapb.c if (idx == 0) { idx 802 drivers/gpio/gpio-dwapb.c unsigned int idx = gpio->ports[i].idx; idx 807 drivers/gpio/gpio-dwapb.c offset = GPIO_SWPORTA_DR + idx * GPIO_SWPORT_DR_STRIDE; idx 810 drivers/gpio/gpio-dwapb.c offset = GPIO_SWPORTA_DDR + idx * GPIO_SWPORT_DDR_STRIDE; idx 813 drivers/gpio/gpio-dwapb.c offset = GPIO_EXT_PORTA + idx * GPIO_EXT_PORT_STRIDE; idx 817 drivers/gpio/gpio-dwapb.c if (idx == 0) { idx 83 drivers/gpio/gpio-tps65910.c int idx; idx 93 drivers/gpio/gpio-tps65910.c for (idx = 0; idx < ngpio; idx++) idx 94 drivers/gpio/gpio-tps65910.c tps65910_board->en_gpio_sleep[idx] = (prop_array[idx] != 0); idx 521 drivers/gpio/gpio-winbond.c static bool winbond_gpio_configure_port(unsigned long base, unsigned int idx) idx 523 drivers/gpio/gpio-winbond.c const struct winbond_gpio_info *info = &winbond_gpio_infos[idx]; idx 535 drivers/gpio/gpio-winbond.c idx + 1, conflict->name); idx 538 drivers/gpio/gpio-winbond.c idx + 1, conflict->name); idx 545 drivers/gpio/gpio-winbond.c if (idx == 0) idx 547 drivers/gpio/gpio-winbond.c else if (idx == 1) idx 554 drivers/gpio/gpio-winbond.c if (params.ppgpios & BIT(idx)) idx 557 drivers/gpio/gpio-winbond.c else if (params.odgpios & BIT(idx)) idx 561 drivers/gpio/gpio-winbond.c pr_notice("GPIO%u pins are %s\n", idx + 1, idx 814 drivers/gpio/gpiolib-acpi.c unsigned int idx, idx 834 drivers/gpio/gpiolib-acpi.c desc = acpi_get_gpiod_by_index(adev, propname, idx, &info); idx 846 drivers/gpio/gpiolib-acpi.c desc = acpi_get_gpiod_by_index(adev, NULL, idx, &info); idx 922 drivers/gpio/gpiolib-acpi.c int idx, i; idx 926 drivers/gpio/gpiolib-acpi.c for (i = 0, idx = 0; idx <= index; i++) { idx 936 drivers/gpio/gpiolib-acpi.c if (info.gpioint && idx++ == index) { idx 47 drivers/gpio/gpiolib-acpi.h unsigned int idx, idx 79 drivers/gpio/gpiolib-acpi.h unsigned int idx, enum gpiod_flags *dflags, idx 95 drivers/gpio/gpiolib-devres.c unsigned int idx, idx 101 drivers/gpio/gpiolib-devres.c desc = gpiod_get_index(dev, con_id, idx, flags); idx 408 drivers/gpio/gpiolib-of.c unsigned int idx, idx 429 drivers/gpio/gpiolib-of.c return of_find_gpio(dev, NULL, idx, flags); idx 478 drivers/gpio/gpiolib-of.c unsigned int idx, unsigned long *flags) idx 494 drivers/gpio/gpiolib-of.c desc = of_get_named_gpiod_flags(dev->of_node, prop_name, idx, idx 508 drivers/gpio/gpiolib-of.c desc = of_find_spi_cs_gpio(dev, con_id, idx, flags); idx 560 drivers/gpio/gpiolib-of.c unsigned int idx, const char **name, idx 588 drivers/gpio/gpiolib-of.c ret = of_property_read_u32_index(np, "gpios", idx * tmp + i, idx 12 drivers/gpio/gpiolib-of.h unsigned int idx, idx 21 drivers/gpio/gpiolib-of.h unsigned int idx, idx 4320 drivers/gpio/gpiolib.c unsigned int idx, unsigned long *flags) idx 4334 drivers/gpio/gpiolib.c if (p->idx != idx) idx 4359 drivers/gpio/gpiolib.c idx, p->chip_hwnum, chip->ngpio - 1, idx 4536 drivers/gpio/gpiolib.c unsigned int idx, idx 4551 drivers/gpio/gpiolib.c desc = of_find_gpio(dev, con_id, idx, &lookupflags); idx 4554 drivers/gpio/gpiolib.c desc = acpi_find_gpio(dev, con_id, idx, &flags, &lookupflags); idx 4564 drivers/gpio/gpiolib.c desc = gpiod_find(dev, con_id, idx, &lookupflags); idx 490 drivers/gpu/drm/amd/amdgpu/amdgpu.h uint32_t ib_idx, int idx) idx 492 drivers/gpu/drm/amd/amdgpu/amdgpu.h return p->job->ibs[ib_idx].ptr[idx]; idx 496 drivers/gpu/drm/amd/amdgpu/amdgpu.h uint32_t ib_idx, int idx, idx 499 drivers/gpu/drm/amd/amdgpu/amdgpu.h p->job->ibs[ib_idx].ptr[idx] = value; idx 473 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c unsigned idx = 0; idx 475 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c idx = seq & (amdgpu_sched_jobs - 1); idx 476 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c other = centity->fences[idx]; idx 483 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c centity->fences[idx] = fence; idx 545 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c unsigned idx; idx 549 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c idx = centity->sequence & (amdgpu_sched_jobs - 1); idx 550 drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c other = dma_fence_get(centity->fences[idx]); idx 562 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c int idx, x, outsize, r, valuesize; idx 572 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c idx = *pos >> 2; idx 575 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c r = amdgpu_dpm_read_sensor(adev, idx, &values[0], &valuesize); idx 1055 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c DRM_WARN("failed to preempt ring %d\n", ring->idx); idx 1063 drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c DRM_INFO("ring %d was preempted\n", ring->idx); idx 897 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c amdgpu_get_vce_clock_state(void *handle, u32 idx) idx 901 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c if (idx < adev->pm.dpm.num_of_vce_states) idx 902 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c return &adev->pm.dpm.vce_states[idx]; idx 517 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.h amdgpu_get_vce_clock_state(void *handle, u32 idx); idx 153 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c adev->fence_context + ring->idx, idx 654 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c DMA_FENCE_TRACE(&fence->base, "armed on ring %i!\n", ring->idx); idx 283 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault = &gmc->fault_ring[gmc->fault_hash[hash].idx]; idx 304 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c fault->next = gmc->fault_hash[hash].idx; idx 305 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c gmc->fault_hash[hash].idx = gmc->last_fault++; idx 174 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h uint64_t idx:AMDGPU_GMC_FAULT_RING_ORDER; idx 226 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c u64 fence_context = adev->vm_manager.fence_context + ring->idx; idx 227 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c unsigned seqno = ++adev->vm_manager.seqno[ring->idx]; idx 274 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c uint64_t fence_context = adev->fence_context + ring->idx; idx 340 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c uint64_t fence_context = adev->fence_context + ring->idx; idx 1177 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c int idx = amdgpu_display_crtc_idx_to_irq_type(adev, pipe); idx 1179 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c return amdgpu_irq_get(adev, &adev->crtc_irq, idx); idx 1193 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c int idx = amdgpu_display_crtc_idx_to_irq_type(adev, pipe); idx 1195 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c amdgpu_irq_put(adev, &adev->crtc_irq, idx); idx 487 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c unsigned long idx; idx 498 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c ret = kstrtoul(buf, 0, &idx); idx 499 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c if (ret || idx >= ARRAY_SIZE(data.states)) { idx 503 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c idx = array_index_nospec(idx, ARRAY_SIZE(data.states)); idx 506 drivers/gpu/drm/amd/amdgpu/amdgpu_pm.c state = data.states[idx]; idx 260 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c ring->idx = adev->num_rings++; idx 261 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c adev->rings[ring->idx] = ring; idx 357 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c if (!(ring->adev) || !(ring->adev->rings[ring->idx])) idx 376 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c ring->adev->rings[ring->idx] = NULL; idx 195 drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h u32 idx; idx 261 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c uint32_t idx = best_bo->fence->context; idx 263 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c idx %= AMDGPU_SA_NUM_FENCE_LISTS; idx 264 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c ++tries[idx]; idx 358 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c uint32_t idx; idx 361 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c idx = fence->context % AMDGPU_SA_NUM_FENCE_LISTS; idx 362 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c list_add_tail(&(*sa_bo)->flist, &sa_manager->flist[idx]); idx 154 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->ring = to_amdgpu_ring(p->entity->rq->sched)->idx; idx 91 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c unsigned idx; idx 492 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c cmd = amdgpu_get_ib_value(ctx->parser, ctx->ib_idx, ctx->idx) >> 1; idx 833 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c cmd = amdgpu_get_ib_value(ctx->parser, ctx->ib_idx, ctx->idx) >> 1; idx 896 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ctx->idx++; idx 900 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c if (ctx->idx >= ib->length_dw) { idx 907 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ctx->data0 = ctx->idx; idx 910 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ctx->data1 = ctx->idx; idx 924 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ctx->idx++; idx 943 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c for (ctx->idx = 0 ; ctx->idx < ib->length_dw; ) { idx 944 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c uint32_t cmd = amdgpu_get_ib_value(ctx->parser, ctx->ib_idx, ctx->idx); idx 955 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c ++ctx->idx; idx 719 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c unsigned idx; idx 725 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c for (idx = 0; idx < ib->length_dw;) { idx 726 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c uint32_t len = amdgpu_get_ib_value(p, ib_idx, idx); idx 727 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c uint32_t cmd = amdgpu_get_ib_value(p, ib_idx, idx + 1); idx 737 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c fb_idx = amdgpu_get_ib_value(p, ib_idx, idx + 6); idx 738 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c bs_idx = amdgpu_get_ib_value(p, ib_idx, idx + 7); idx 742 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_validate_bo(p, ib_idx, idx + 10, idx 743 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx + 9, 0, 0); idx 747 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_validate_bo(p, ib_idx, idx + 12, idx 748 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx + 11, 0, 0); idx 754 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_validate_bo(p, ib_idx, idx + 3, idx 755 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx + 2, 0, 0); idx 761 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c tmp = amdgpu_get_ib_value(p, ib_idx, idx + 4); idx 762 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_validate_bo(p, ib_idx, idx + 3, idx + 2, idx 769 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_validate_bo(p, ib_idx, idx + 3, idx + 2, idx 776 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_validate_bo(p, ib_idx, idx + 3, idx 777 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx + 2, 0, 0); idx 781 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_validate_bo(p, ib_idx, idx + 8, idx 782 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx + 7, 0, 0); idx 788 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx += len / 4; idx 791 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c for (idx = 0; idx < ib->length_dw;) { idx 792 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c uint32_t len = amdgpu_get_ib_value(p, ib_idx, idx); idx 793 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c uint32_t cmd = amdgpu_get_ib_value(p, ib_idx, idx + 1); idx 797 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c handle = amdgpu_get_ib_value(p, ib_idx, idx + 2); idx 808 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c fb_idx = amdgpu_get_ib_value(p, ib_idx, idx + 6); idx 809 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c bs_idx = amdgpu_get_ib_value(p, ib_idx, idx + 7); idx 824 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c *size = amdgpu_get_ib_value(p, ib_idx, idx + 8) * idx 825 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c amdgpu_get_ib_value(p, ib_idx, idx + 10) * idx 854 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_cs_reloc(p, ib_idx, idx + 10, idx + 9, idx 859 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_cs_reloc(p, ib_idx, idx + 12, idx + 11, idx 870 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_cs_reloc(p, ib_idx, idx + 3, idx + 2, idx 877 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c tmp = amdgpu_get_ib_value(p, ib_idx, idx + 4); idx 878 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_cs_reloc(p, ib_idx, idx + 3, idx + 2, idx 885 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_cs_reloc(p, ib_idx, idx + 3, idx + 2, idx 892 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_cs_reloc(p, ib_idx, idx + 3, idx 893 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx + 2, *size, 0); idx 897 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c r = amdgpu_vce_cs_reloc(p, ib_idx, idx + 8, idx 898 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx + 7, *size / 12, 0); idx 915 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx += len / 4; idx 953 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c int i, r = 0, idx = 0; idx 955 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c while (idx < ib->length_dw) { idx 956 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c uint32_t len = amdgpu_get_ib_value(p, ib_idx, idx); idx 957 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c uint32_t cmd = amdgpu_get_ib_value(p, ib_idx, idx + 1); idx 967 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c handle = amdgpu_get_ib_value(p, ib_idx, idx + 2); idx 1004 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c idx += len / 4; idx 387 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c unsigned mask, shift, idx; idx 397 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c idx = (cursor->pfn >> shift) & mask; idx 399 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c cursor->entry = &cursor->entry->entries[idx]; idx 1630 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c uint64_t idx = pfn + count; idx 1632 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (pages_addr[idx] != idx 1633 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c (pages_addr[idx - 1] + PAGE_SIZE)) idx 179 drivers/gpu/drm/amd/amdgpu/atom.c uint32_t idx, val = 0xCDCDCDCD, align, arg; idx 185 drivers/gpu/drm/amd/amdgpu/atom.c idx = U16(*ptr); idx 188 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("REG[0x%04X]", idx); idx 189 drivers/gpu/drm/amd/amdgpu/atom.c idx += gctx->reg_block; idx 192 drivers/gpu/drm/amd/amdgpu/atom.c val = gctx->card->reg_read(gctx->card, idx); idx 213 drivers/gpu/drm/amd/amdgpu/atom.c idx, 0); idx 217 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 221 drivers/gpu/drm/amd/amdgpu/atom.c val = get_unaligned_le32((u32 *)&ctx->ps[idx]); idx 223 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("PS[0x%02X,0x%04X]", idx, val); idx 226 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 229 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("WS[0x%02X]", idx); idx 230 drivers/gpu/drm/amd/amdgpu/atom.c switch (idx) { idx 259 drivers/gpu/drm/amd/amdgpu/atom.c val = ctx->ws[idx]; idx 263 drivers/gpu/drm/amd/amdgpu/atom.c idx = U16(*ptr); idx 267 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("ID[0x%04X+%04X]", idx, gctx->data_block); idx 269 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("ID[0x%04X]", idx); idx 271 drivers/gpu/drm/amd/amdgpu/atom.c val = U32(idx + gctx->data_block); idx 274 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 276 drivers/gpu/drm/amd/amdgpu/atom.c if ((gctx->fb_base + (idx * 4)) > gctx->scratch_size_bytes) { idx 278 drivers/gpu/drm/amd/amdgpu/atom.c gctx->fb_base + (idx * 4), gctx->scratch_size_bytes); idx 281 drivers/gpu/drm/amd/amdgpu/atom.c val = gctx->scratch[(gctx->fb_base / 4) + idx]; idx 283 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("FB[0x%02X]", idx); idx 313 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 316 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("PLL[0x%02X]", idx); idx 317 drivers/gpu/drm/amd/amdgpu/atom.c val = gctx->card->pll_read(gctx->card, idx); idx 320 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 323 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("MC[0x%02X]", idx); idx 324 drivers/gpu/drm/amd/amdgpu/atom.c val = gctx->card->mc_read(gctx->card, idx); idx 449 drivers/gpu/drm/amd/amdgpu/atom.c val, idx; idx 458 drivers/gpu/drm/amd/amdgpu/atom.c idx = U16(*ptr); idx 460 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("REG[0x%04X]", idx); idx 461 drivers/gpu/drm/amd/amdgpu/atom.c idx += gctx->reg_block; idx 464 drivers/gpu/drm/amd/amdgpu/atom.c if (idx == 0) idx 465 drivers/gpu/drm/amd/amdgpu/atom.c gctx->card->reg_write(gctx->card, idx, idx 468 drivers/gpu/drm/amd/amdgpu/atom.c gctx->card->reg_write(gctx->card, idx, val); idx 487 drivers/gpu/drm/amd/amdgpu/atom.c idx, val); idx 491 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 493 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("PS[0x%02X]", idx); idx 494 drivers/gpu/drm/amd/amdgpu/atom.c ctx->ps[idx] = cpu_to_le32(val); idx 497 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 499 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("WS[0x%02X]", idx); idx 500 drivers/gpu/drm/amd/amdgpu/atom.c switch (idx) { idx 526 drivers/gpu/drm/amd/amdgpu/atom.c ctx->ws[idx] = val; idx 530 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 532 drivers/gpu/drm/amd/amdgpu/atom.c if ((gctx->fb_base + (idx * 4)) > gctx->scratch_size_bytes) { idx 534 drivers/gpu/drm/amd/amdgpu/atom.c gctx->fb_base + (idx * 4), gctx->scratch_size_bytes); idx 536 drivers/gpu/drm/amd/amdgpu/atom.c gctx->scratch[(gctx->fb_base / 4) + idx] = val; idx 537 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("FB[0x%02X]", idx); idx 540 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 542 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("PLL[0x%02X]", idx); idx 543 drivers/gpu/drm/amd/amdgpu/atom.c gctx->card->pll_write(gctx->card, idx, val); idx 546 drivers/gpu/drm/amd/amdgpu/atom.c idx = U8(*ptr); idx 548 drivers/gpu/drm/amd/amdgpu/atom.c DEBUG("MC[0x%02X]", idx); idx 549 drivers/gpu/drm/amd/amdgpu/atom.c gctx->card->mc_write(gctx->card, idx, val); idx 615 drivers/gpu/drm/amd/amdgpu/atom.c int idx = U8((*ptr)++); idx 618 drivers/gpu/drm/amd/amdgpu/atom.c if (idx < ATOM_TABLE_NAMES_CNT) idx 619 drivers/gpu/drm/amd/amdgpu/atom.c SDEBUG(" table: %d (%s)\n", idx, atom_table_names[idx]); idx 621 drivers/gpu/drm/amd/amdgpu/atom.c SDEBUG(" table: %d\n", idx); idx 622 drivers/gpu/drm/amd/amdgpu/atom.c if (U16(ctx->ctx->cmd_table + 4 + 2 * idx)) idx 623 drivers/gpu/drm/amd/amdgpu/atom.c r = amdgpu_atom_execute_table_locked(ctx->ctx, idx, ctx->ps + ctx->ps_shift); idx 864 drivers/gpu/drm/amd/amdgpu/atom.c int idx = U8(*ptr); idx 866 drivers/gpu/drm/amd/amdgpu/atom.c SDEBUG(" block: %d\n", idx); idx 867 drivers/gpu/drm/amd/amdgpu/atom.c if (!idx) idx 869 drivers/gpu/drm/amd/amdgpu/atom.c else if (idx == 255) idx 872 drivers/gpu/drm/amd/amdgpu/atom.c ctx->ctx->data_block = U16(ctx->ctx->data_table + 4 + 2 * idx); idx 1305 drivers/gpu/drm/amd/amdgpu/atom.c u16 idx; idx 1343 drivers/gpu/drm/amd/amdgpu/atom.c idx = CU16(ATOM_ROM_PART_NUMBER_PTR); idx 1344 drivers/gpu/drm/amd/amdgpu/atom.c if (idx == 0) idx 1345 drivers/gpu/drm/amd/amdgpu/atom.c idx = 0x80; idx 1347 drivers/gpu/drm/amd/amdgpu/atom.c str = CSTR(idx); idx 1392 drivers/gpu/drm/amd/amdgpu/atom.c int idx = CU16(ctx->data_table + offset); idx 1399 drivers/gpu/drm/amd/amdgpu/atom.c *size = CU16(idx); idx 1401 drivers/gpu/drm/amd/amdgpu/atom.c *frev = CU8(idx + 2); idx 1403 drivers/gpu/drm/amd/amdgpu/atom.c *crev = CU8(idx + 3); idx 1404 drivers/gpu/drm/amd/amdgpu/atom.c *data_start = idx; idx 1412 drivers/gpu/drm/amd/amdgpu/atom.c int idx = CU16(ctx->cmd_table + offset); idx 1419 drivers/gpu/drm/amd/amdgpu/atom.c *frev = CU8(idx + 2); idx 1421 drivers/gpu/drm/amd/amdgpu/atom.c *crev = CU8(idx + 3); idx 1058 drivers/gpu/drm/amd/amdgpu/cik.c unsigned idx; idx 1097 drivers/gpu/drm/amd/amdgpu/cik.c idx = (reg_offset - mmGB_TILE_MODE0); idx 1098 drivers/gpu/drm/amd/amdgpu/cik.c return adev->gfx.config.tile_mode_array[idx]; idx 1115 drivers/gpu/drm/amd/amdgpu/cik.c idx = (reg_offset - mmGB_MACROTILE_MODE0); idx 1116 drivers/gpu/drm/amd/amdgpu/cik.c return adev->gfx.config.macrotile_mode_array[idx]; idx 452 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c ring->idx, r); idx 474 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c ring->idx, i); idx 477 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c ring->idx, i); idx 480 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c ring->idx, scratch, tmp); idx 533 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c DRM_INFO("ib test on ring %d succeeded\n", ring->idx); idx 4712 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c DRM_ERROR("ring %d failed to preempt ib\n", ring->idx); idx 1397 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c int idx = (se / 2) * 2; idx 1399 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c if ((num_se > 1) && (!se_mask[idx] || !se_mask[idx + 1])) { idx 1402 drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c if (!se_mask[idx]) idx 1707 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c int idx = (se / 2) * 2; idx 1709 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c if ((num_se > 1) && (!se_mask[idx] || !se_mask[idx + 1])) { idx 1712 drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c if (!se_mask[idx]) { idx 3571 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c int idx = (se / 2) * 2; idx 3573 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c if ((num_se > 1) && (!se_mask[idx] || !se_mask[idx + 1])) { idx 3576 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c if (!se_mask[idx]) { idx 2617 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c int idx; idx 2628 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c for (idx = 0; idx < unique_indirect_reg_count; idx++) { idx 2629 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c if (unique_indirect_regs[idx] == idx 2631 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c !unique_indirect_regs[idx]) idx 2635 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c BUG_ON(idx >= unique_indirect_reg_count); idx 2637 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c if (!unique_indirect_regs[idx]) idx 2638 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c unique_indirect_regs[idx] = register_list_format[indirect_offset]; idx 563 drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c ring->idx, ring->name, ring->vm_inv_eng, idx 2040 drivers/gpu/drm/amd/amdgpu/kv_dpm.c int idx = pi->sys_info.sclk_voltage_mapping_table.num_max_dpm_entries - 1; idx 2042 drivers/gpu/drm/amd/amdgpu/kv_dpm.c pi->sys_info.sclk_voltage_mapping_table.entries[idx].sclk_frequency; idx 2045 drivers/gpu/drm/amd/amdgpu/kv_dpm.c pi->sys_info.sclk_voltage_mapping_table.entries[idx].vid_2bit); idx 2742 drivers/gpu/drm/amd/amdgpu/kv_dpm.c u8 *idx; idx 2754 drivers/gpu/drm/amd/amdgpu/kv_dpm.c idx = (u8 *)&power_state->v2.clockInfoIndex[0]; idx 2756 drivers/gpu/drm/amd/amdgpu/kv_dpm.c clock_array_index = idx[j]; idx 3280 drivers/gpu/drm/amd/amdgpu/kv_dpm.c static int kv_dpm_read_sensor(void *handle, int idx, idx 3295 drivers/gpu/drm/amd/amdgpu/kv_dpm.c switch (idx) { idx 906 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c DRM_ERROR("amdgpu: dma failed to lock ring %d (%d).\n", ring->idx, r); idx 931 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c DRM_INFO("ring test on %d succeeded in %d msecs\n", ring->idx, i); idx 933 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c DRM_INFO("ring test on %d succeeded in %d usecs\n", ring->idx, i); idx 936 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c ring->idx, tmp); idx 1004 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c DRM_INFO("ib test on ring %d succeeded\n", ring->idx); idx 1396 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c DRM_ERROR("ring %d failed to be preempted\n", ring->idx); idx 1046 drivers/gpu/drm/amd/amdgpu/si.c unsigned idx; idx 1085 drivers/gpu/drm/amd/amdgpu/si.c idx = (reg_offset - mmGB_TILE_MODE0); idx 1086 drivers/gpu/drm/amd/amdgpu/si.c return adev->gfx.config.tile_mode_array[idx]; idx 7254 drivers/gpu/drm/amd/amdgpu/si_dpm.c u8 *idx; idx 7269 drivers/gpu/drm/amd/amdgpu/si_dpm.c idx = (u8 *)&power_state->v2.clockInfoIndex[0]; idx 7271 drivers/gpu/drm/amd/amdgpu/si_dpm.c clock_array_index = idx[j]; idx 7983 drivers/gpu/drm/amd/amdgpu/si_dpm.c static int si_dpm_read_sensor(void *handle, int idx, idx 7999 drivers/gpu/drm/amd/amdgpu/si_dpm.c switch (idx) { idx 30 drivers/gpu/drm/amd/amdgpu/soc15_common.h #define WREG32_FIELD15(ip, idx, reg, field, val) \ idx 31 drivers/gpu/drm/amd/amdgpu/soc15_common.h WREG32(adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg, \ idx 32 drivers/gpu/drm/amd/amdgpu/soc15_common.h (RREG32(adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg) \ idx 122 drivers/gpu/drm/amd/amdgpu/soc15_common.h #define WREG32_FIELD15_RLC(ip, idx, reg, field, val) \ idx 123 drivers/gpu/drm/amd/amdgpu/soc15_common.h WREG32_RLC((adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg), \ idx 124 drivers/gpu/drm/amd/amdgpu/soc15_common.h (RREG32(adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg) \ idx 64 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c static void vce_v3_0_mc_resume(struct amdgpu_device *adev, int idx); idx 268 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c int idx, r; idx 271 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c for (idx = 0; idx < 2; ++idx) { idx 272 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c if (adev->vce.harvest_config & (1 << idx)) idx 275 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c WREG32(mmGRBM_GFX_INDEX, GET_VCE_INSTANCE(idx)); idx 279 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c if (idx != 1 || adev->vce.harvest_config == AMDGPU_VCE_HARVEST_VCE0) { idx 302 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c vce_v3_0_mc_resume(adev, idx); idx 333 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c int idx; idx 336 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c for (idx = 0; idx < 2; ++idx) { idx 337 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c if (adev->vce.harvest_config & (1 << idx)) idx 340 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c WREG32(mmGRBM_GFX_INDEX, GET_VCE_INSTANCE(idx)); idx 524 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c static void vce_v3_0_mc_resume(struct amdgpu_device *adev, int idx) idx 551 drivers/gpu/drm/amd/amdgpu/vce_v3_0.c if (idx == 0) { idx 582 drivers/gpu/drm/amd/amdgpu/vi.c unsigned idx; idx 621 drivers/gpu/drm/amd/amdgpu/vi.c idx = (reg_offset - mmGB_TILE_MODE0); idx 622 drivers/gpu/drm/amd/amdgpu/vi.c return adev->gfx.config.tile_mode_array[idx]; idx 639 drivers/gpu/drm/amd/amdgpu/vi.c idx = (reg_offset - mmGB_MACROTILE_MODE0); idx 640 drivers/gpu/drm/amd/amdgpu/vi.c return adev->gfx.config.macrotile_mode_array[idx]; idx 1018 drivers/gpu/drm/amd/amdkfd/kfd_events.c uint32_t id, idx; idx 1034 drivers/gpu/drm/amd/amdkfd/kfd_events.c idx = srcu_read_lock(&kfd_processes_srcu); idx 1051 drivers/gpu/drm/amd/amdkfd/kfd_events.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 224 drivers/gpu/drm/amd/amdkfd/kfd_iommu.c int idx = srcu_read_lock(&kfd_processes_srcu); idx 248 drivers/gpu/drm/amd/amdkfd/kfd_iommu.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 264 drivers/gpu/drm/amd/amdkfd/kfd_iommu.c int idx = srcu_read_lock(&kfd_processes_srcu); idx 280 drivers/gpu/drm/amd/amdkfd/kfd_iommu.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 852 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int kfd_topology_enum_kfd_devices(uint8_t idx, struct kfd_dev **kdev); idx 365 drivers/gpu/drm/amd/amdkfd/kfd_process.c int idx; idx 367 drivers/gpu/drm/amd/amdkfd/kfd_process.c idx = srcu_read_lock(&kfd_processes_srcu); idx 369 drivers/gpu/drm/amd/amdkfd/kfd_process.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 910 drivers/gpu/drm/amd/amdkfd/kfd_process.c int idx = srcu_read_lock(&kfd_processes_srcu); idx 920 drivers/gpu/drm/amd/amdkfd/kfd_process.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 930 drivers/gpu/drm/amd/amdkfd/kfd_process.c int idx = srcu_read_lock(&kfd_processes_srcu); idx 936 drivers/gpu/drm/amd/amdkfd/kfd_process.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 1084 drivers/gpu/drm/amd/amdkfd/kfd_process.c int idx = srcu_read_lock(&kfd_processes_srcu); idx 1096 drivers/gpu/drm/amd/amdkfd/kfd_process.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 1103 drivers/gpu/drm/amd/amdkfd/kfd_process.c int ret = 0, idx = srcu_read_lock(&kfd_processes_srcu); idx 1112 drivers/gpu/drm/amd/amdkfd/kfd_process.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 1171 drivers/gpu/drm/amd/amdkfd/kfd_process.c int idx = srcu_read_lock(&kfd_processes_srcu); idx 1185 drivers/gpu/drm/amd/amdkfd/kfd_process.c srcu_read_unlock(&kfd_processes_srcu, idx); idx 1408 drivers/gpu/drm/amd/amdkfd/kfd_topology.c int kfd_topology_enum_kfd_devices(uint8_t idx, struct kfd_dev **kdev) idx 1418 drivers/gpu/drm/amd/amdkfd/kfd_topology.c if (device_idx == idx) { idx 361 drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c int idx = 0; idx 366 drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c available_slices[idx++] = 1; idx 369 drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c available_slices[idx++] = 2; idx 372 drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c available_slices[idx++] = 4; idx 375 drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c available_slices[idx++] = 8; idx 377 drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c return idx; idx 652 drivers/gpu/drm/amd/display/modules/freesync/freesync.c unsigned int idx = 0; idx 660 drivers/gpu/drm/amd/display/modules/freesync/freesync.c for (idx = 1; idx <= *payload_size; idx++) idx 661 drivers/gpu/drm/amd/display/modules/freesync/freesync.c checksum += infopacket->sb[idx]; idx 253 drivers/gpu/drm/amd/include/kgd_pp_interface.h int (*read_sensor)(void *handle, int idx, void *value, int *size); idx 264 drivers/gpu/drm/amd/include/kgd_pp_interface.h struct amd_vce_state *(*get_vce_clock_state)(void *handle, u32 idx); idx 830 drivers/gpu/drm/amd/powerplay/amd_powerplay.c static int pp_dpm_read_sensor(void *handle, int idx, idx 839 drivers/gpu/drm/amd/powerplay/amd_powerplay.c switch (idx) { idx 854 drivers/gpu/drm/amd/powerplay/amd_powerplay.c ret = hwmgr->hwmgr_func->read_sensor(hwmgr, idx, value, size); idx 861 drivers/gpu/drm/amd/powerplay/amd_powerplay.c pp_dpm_get_vce_clock_state(void *handle, unsigned idx) idx 868 drivers/gpu/drm/amd/powerplay/amd_powerplay.c if (idx < hwmgr->num_vce_state_tables) idx 869 drivers/gpu/drm/amd/powerplay/amd_powerplay.c return &hwmgr->vce_states[idx]; idx 58 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c uint16_t idx; idx 60 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx = GetIndexIntoMasterDataTable(voltageobject_info); idx 62 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx, NULL, NULL, NULL); idx 166 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c uint16_t idx; idx 168 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx = GetIndexIntoMasterDataTable(gpio_pin_lut); idx 170 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx, NULL, NULL, NULL); idx 253 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c uint32_t idx; idx 258 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx = GetIndexIntoMasterCmdTable(computegpuclockparam); idx 261 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c adev->mode_info.atom_context, idx, (uint32_t *)&pll_parameters)) idx 279 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c uint16_t idx; idx 285 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx = GetIndexIntoMasterDataTable(asic_profiling_info); idx 288 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx, NULL, NULL, NULL); idx 467 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c uint16_t idx; idx 469 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx = GetIndexIntoMasterDataTable(smu_info); idx 472 drivers/gpu/drm/amd/powerplay/hwmgr/ppatomfwctrl.c idx, NULL, NULL, NULL); idx 1116 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.c static int smu10_read_sensor(struct pp_hwmgr *hwmgr, int idx, idx 1123 drivers/gpu/drm/amd/powerplay/hwmgr/smu10_hwmgr.c switch (idx) { idx 3525 drivers/gpu/drm/amd/powerplay/hwmgr/smu7_hwmgr.c static int smu7_read_sensor(struct pp_hwmgr *hwmgr, int idx, idx 3536 drivers/gpu/drm/amd/powerplay/hwmgr/smu7_hwmgr.c switch (idx) { idx 3553 drivers/gpu/drm/amd/powerplay/hwmgr/smu7_hwmgr.c (idx == AMDGPU_PP_SENSOR_GPU_LOAD) ? idx 1669 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.c static int smu8_read_sensor(struct pp_hwmgr *hwmgr, int idx, idx 1699 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.c switch (idx) { idx 3766 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c static int vega10_read_sensor(struct pp_hwmgr *hwmgr, int idx, idx 3776 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c switch (idx) { idx 3919 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c uint32_t idx; idx 3955 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c idx = vega10_get_uclk_index(hwmgr, mclk_table, min_clocks.memoryClock); idx 3956 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c smum_send_msg_to_smc_with_parameter(hwmgr, PPSMC_MSG_SetSoftMinUclkByIndex, idx); idx 3957 drivers/gpu/drm/amd/powerplay/hwmgr/vega10_hwmgr.c data->dpm_table.mem_table.dpm_state.soft_min_level= idx; idx 1321 drivers/gpu/drm/amd/powerplay/hwmgr/vega12_hwmgr.c int idx, idx 1331 drivers/gpu/drm/amd/powerplay/hwmgr/vega12_hwmgr.c switch (idx) { idx 1346 drivers/gpu/drm/amd/powerplay/hwmgr/vega12_hwmgr.c static int vega12_read_sensor(struct pp_hwmgr *hwmgr, int idx, idx 1353 drivers/gpu/drm/amd/powerplay/hwmgr/vega12_hwmgr.c switch (idx) { idx 1366 drivers/gpu/drm/amd/powerplay/hwmgr/vega12_hwmgr.c ret = vega12_get_current_activity_percent(hwmgr, idx, (uint32_t *)value); idx 2132 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_hwmgr.c int idx, idx 2142 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_hwmgr.c switch (idx) { idx 2157 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_hwmgr.c static int vega20_read_sensor(struct pp_hwmgr *hwmgr, int idx, idx 2166 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_hwmgr.c switch (idx) { idx 2184 drivers/gpu/drm/amd/powerplay/hwmgr/vega20_hwmgr.c ret = vega20_get_current_activity_percent(hwmgr, idx, (uint32_t *)value); idx 321 drivers/gpu/drm/amd/powerplay/inc/hwmgr.h int (*read_sensor)(struct pp_hwmgr *hwmgr, int idx, void *value, int *size); idx 157 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c static inline u32 to_d71_input_id(struct komeda_component_state *st, int idx) idx 159 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct komeda_component_output *input = &st->inputs[idx]; idx 162 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c if (has_bit(idx, st->active_inputs)) idx 168 drivers/gpu/drm/arm/display/komeda/komeda_pipeline.c int idx, *num = NULL; idx 181 drivers/gpu/drm/arm/display/komeda/komeda_pipeline.c idx = id - KOMEDA_COMPONENT_LAYER0; idx 183 drivers/gpu/drm/arm/display/komeda/komeda_pipeline.c if (idx != pipe->n_layers) { idx 188 drivers/gpu/drm/arm/display/komeda/komeda_pipeline.c idx = id - KOMEDA_COMPONENT_SCALER0; idx 190 drivers/gpu/drm/arm/display/komeda/komeda_pipeline.c if (idx != pipe->n_scalers) { idx 186 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c int idx) idx 190 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c WARN_ON((idx < 0 || idx >= c->max_active_inputs)); idx 197 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c if (!has_bit(idx, state->affected_inputs) || idx 198 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c memcmp(&state->inputs[idx], input, sizeof(*input))) { idx 199 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c memcpy(&state->inputs[idx], input, sizeof(*input)); idx 200 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c state->changed_active_inputs |= BIT(idx); idx 202 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c state->active_inputs |= BIT(idx); idx 203 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c state->affected_inputs |= BIT(idx); idx 209 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c int idx) idx 213 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c if ((idx < 0) || (idx >= c->max_active_inputs)) { idx 215 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c input->component->name, c->name, idx); idx 219 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c if (has_bit(idx, state->active_inputs)) { idx 221 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c input->component->name, c->name, idx); idx 659 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c int idx = dflow->blending_zorder; idx 677 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c if (komeda_component_check_input(c_st, &dflow->input, idx)) idx 680 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c cin = &(to_compiz_st(c_st)->cins[idx]); idx 693 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c if (memcmp(&(to_compiz_st(old_st)->cins[idx]), cin, sizeof(*cin))) idx 694 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c c_st->changed_active_inputs |= BIT(idx); idx 696 drivers/gpu/drm/arm/display/komeda/komeda_pipeline_state.c komeda_component_add_input(c_st, &dflow->input, idx); idx 23 drivers/gpu/drm/armada/armada_510.c int idx; idx 38 drivers/gpu/drm/armada/armada_510.c idx = 0; idx 40 drivers/gpu/drm/armada/armada_510.c idx = 1; idx 42 drivers/gpu/drm/armada/armada_510.c idx = 2; idx 44 drivers/gpu/drm/armada/armada_510.c idx = 3; idx 52 drivers/gpu/drm/armada/armada_510.c v->clks[idx] = clk; idx 106 drivers/gpu/drm/armada/armada_510.c int ret, idx; idx 108 drivers/gpu/drm/armada/armada_510.c idx = armada_crtc_select_clock(dcrtc, &res, &armada510_clocking, idx 111 drivers/gpu/drm/armada/armada_510.c if (idx < 0) idx 112 drivers/gpu/drm/armada/armada_510.c return idx; idx 121 drivers/gpu/drm/armada/armada_510.c *sclk = res.div | armada510_clk_sels[idx]; idx 74 drivers/gpu/drm/armada/armada_overlay.c unsigned int idx; idx 91 drivers/gpu/drm/armada/armada_overlay.c idx = 0; idx 93 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_mod(regs, idx, idx 98 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_DMA_HPXL_VLN); idx 101 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_DMA_OVSA_HPXL_VLN); idx 104 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_DZM_HPXL_VLN); idx 113 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, armada_addr(state, 0, 0), idx 115 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, armada_addr(state, 0, 1), idx 117 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, armada_addr(state, 0, 2), idx 119 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, armada_addr(state, 1, 0), idx 121 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, armada_addr(state, 1, 1), idx 123 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, armada_addr(state, 1, 2), idx 127 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_DMA_PITCH_YC); idx 129 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_DMA_PITCH_UV); idx 168 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_mod(regs, idx, cfg, cfg_mask, idx 174 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_CONTRAST); idx 178 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_SATURATION); idx 180 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, 0x00002000, LCD_SPU_CBSH_HUE); idx 184 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_mod(regs, idx, val, CFG_CSC_MASK, idx 189 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_COLORKEY_Y); idx 193 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_COLORKEY_U); idx 197 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_set(regs, idx, val, LCD_SPU_COLORKEY_V); idx 201 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_mod(regs, idx, val, CFG_CKMODE_MASK | idx 208 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_mod(regs, idx, val, ADV_GRACOLORKEY | idx 211 drivers/gpu/drm/armada/armada_overlay.c dcrtc->regs_idx += idx; idx 219 drivers/gpu/drm/armada/armada_overlay.c unsigned int idx = 0; idx 235 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_mod(regs, idx, 0, CFG_DMA_ENA, LCD_SPU_DMA_CTRL0); idx 236 drivers/gpu/drm/armada/armada_overlay.c armada_reg_queue_mod(regs, idx, CFG_PDWN16x66 | CFG_PDWN32x66, 0, idx 239 drivers/gpu/drm/armada/armada_overlay.c dcrtc->regs_idx += idx; idx 165 drivers/gpu/drm/armada/armada_plane.c unsigned int idx; idx 181 drivers/gpu/drm/armada/armada_plane.c idx = 0; idx 186 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_mod(regs, idx, 0, val, LCD_SPU_SRAM_PARA1); idx 190 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_set(regs, idx, val, LCD_SPU_GRA_HPXL_VLN); idx 193 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_set(regs, idx, val, LCD_SPU_GRA_OVSA_HPXL_VLN); idx 196 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_set(regs, idx, val, LCD_SPU_GZM_HPXL_VLN); idx 201 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_set(regs, idx, armada_addr(state, 0, 0), idx 203 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_set(regs, idx, armada_addr(state, 1, 0), idx 205 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_mod(regs, idx, armada_pitch(state, 0), 0xffff, idx 238 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_mod(regs, idx, cfg, cfg_mask, idx 241 drivers/gpu/drm/armada/armada_plane.c dcrtc->regs_idx += idx; idx 249 drivers/gpu/drm/armada/armada_plane.c unsigned int idx = 0; idx 265 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_mod(regs, idx, 0, CFG_GRA_ENA, LCD_SPU_DMA_CTRL0); idx 266 drivers/gpu/drm/armada/armada_plane.c armada_reg_queue_mod(regs, idx, CFG_PDWN256x32 | CFG_PDWN256x24 | idx 270 drivers/gpu/drm/armada/armada_plane.c dcrtc->regs_idx += idx; idx 419 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c int idx; idx 429 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c for (idx = 0; idx < ATMEL_HLCDC_CLUT_SIZE; idx++, lut++) { idx 434 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c atmel_hlcdc_layer_write_clut(&plane->layer, idx, val); idx 466 drivers/gpu/drm/drm_bufs.c int idx; idx 473 drivers/gpu/drm/drm_bufs.c idx = map->offset; idx 474 drivers/gpu/drm/drm_bufs.c if (idx < 0) idx 480 drivers/gpu/drm/drm_bufs.c if (i == idx) { idx 804 drivers/gpu/drm/drm_bufs.c buf->idx = dma->buf_count + entry->buf_count; idx 1004 drivers/gpu/drm/drm_bufs.c buf->idx = dma->buf_count + entry->buf_count; idx 1166 drivers/gpu/drm/drm_bufs.c buf->idx = dma->buf_count + entry->buf_count; idx 1434 drivers/gpu/drm/drm_bufs.c int idx; idx 1448 drivers/gpu/drm/drm_bufs.c if (copy_from_user(&idx, &request->list[i], sizeof(idx))) idx 1450 drivers/gpu/drm/drm_bufs.c if (idx < 0 || idx >= dma->buf_count) { idx 1452 drivers/gpu/drm/drm_bufs.c idx, dma->buf_count - 1); idx 1455 drivers/gpu/drm/drm_bufs.c idx = array_index_nospec(idx, dma->buf_count); idx 1456 drivers/gpu/drm/drm_bufs.c buf = dma->buflist[idx]; idx 1551 drivers/gpu/drm/drm_bufs.c static int map_one_buf(void *data, int idx, unsigned long virtual, idx 1557 drivers/gpu/drm/drm_bufs.c if (copy_to_user(&request->list[idx].idx, &buf->idx, idx 1558 drivers/gpu/drm/drm_bufs.c sizeof(request->list[0].idx))) idx 1560 drivers/gpu/drm/drm_bufs.c if (copy_to_user(&request->list[idx].total, &buf->total, idx 1563 drivers/gpu/drm/drm_bufs.c if (clear_user(&request->list[idx].used, sizeof(int))) idx 1565 drivers/gpu/drm/drm_bufs.c if (copy_to_user(&request->list[idx].address, &address, idx 312 drivers/gpu/drm/drm_client_modeset.c int idx, idx 335 drivers/gpu/drm/drm_client_modeset.c offsets[idx].x = hoffset; idx 336 drivers/gpu/drm/drm_client_modeset.c offsets[idx].y = voffset; idx 84 drivers/gpu/drm/drm_crtc.c struct drm_crtc *drm_crtc_from_index(struct drm_device *dev, int idx) idx 89 drivers/gpu/drm/drm_crtc.c if (idx == crtc->index) idx 212 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 0; idx 215 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = ((hdr->lct & 0xf) << 4) | (hdr->lcr & 0xf); idx 217 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = hdr->rad[i]; idx 218 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = (hdr->broadcast << 7) | (hdr->path_msg << 6) | idx 220 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = (hdr->somt << 7) | (hdr->eomt << 6) | (hdr->seqno << 4); idx 222 drivers/gpu/drm/drm_dp_mst_topology.c crc4 = drm_dp_msg_header_crc4(buf, (idx * 2) - 1); idx 223 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx - 1] |= (crc4 & 0xf); idx 225 drivers/gpu/drm/drm_dp_mst_topology.c *len = idx; idx 234 drivers/gpu/drm/drm_dp_mst_topology.c u8 idx; idx 250 drivers/gpu/drm/drm_dp_mst_topology.c idx = 1; idx 252 drivers/gpu/drm/drm_dp_mst_topology.c hdr->rad[i] = buf[idx++]; idx 253 drivers/gpu/drm/drm_dp_mst_topology.c hdr->broadcast = (buf[idx] >> 7) & 0x1; idx 254 drivers/gpu/drm/drm_dp_mst_topology.c hdr->path_msg = (buf[idx] >> 6) & 0x1; idx 255 drivers/gpu/drm/drm_dp_mst_topology.c hdr->msg_len = buf[idx] & 0x3f; idx 256 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 257 drivers/gpu/drm/drm_dp_mst_topology.c hdr->somt = (buf[idx] >> 7) & 0x1; idx 258 drivers/gpu/drm/drm_dp_mst_topology.c hdr->eomt = (buf[idx] >> 6) & 0x1; idx 259 drivers/gpu/drm/drm_dp_mst_topology.c hdr->seqno = (buf[idx] >> 4) & 0x1; idx 260 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 261 drivers/gpu/drm/drm_dp_mst_topology.c *hdrlen = idx; idx 268 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 0; idx 271 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = req->req_type & 0x7f; idx 275 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.port_num.port_number & 0xf) << 4; idx 276 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 279 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.allocate_payload.port_number & 0xf) << 4 | idx 281 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 282 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.allocate_payload.vcpi & 0x7f); idx 283 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 284 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.allocate_payload.pbn >> 8); idx 285 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 286 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.allocate_payload.pbn & 0xff); idx 287 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 289 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = ((req->u.allocate_payload.sdp_stream_sink[i * 2] & 0xf) << 4) | idx 291 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 295 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.allocate_payload.sdp_stream_sink[i] & 0xf) << 4; idx 296 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 300 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.query_payload.port_number & 0xf) << 4; idx 301 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 302 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.query_payload.vcpi & 0x7f); idx 303 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 306 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_read.port_number & 0xf) << 4; idx 307 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] |= ((req->u.dpcd_read.dpcd_address & 0xf0000) >> 16) & 0xf; idx 308 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 309 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_read.dpcd_address & 0xff00) >> 8; idx 310 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 311 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_read.dpcd_address & 0xff); idx 312 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 313 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_read.num_bytes); idx 314 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 318 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_write.port_number & 0xf) << 4; idx 319 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] |= ((req->u.dpcd_write.dpcd_address & 0xf0000) >> 16) & 0xf; idx 320 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 321 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_write.dpcd_address & 0xff00) >> 8; idx 322 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 323 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_write.dpcd_address & 0xff); idx 324 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 325 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.dpcd_write.num_bytes); idx 326 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 327 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(&buf[idx], req->u.dpcd_write.bytes, req->u.dpcd_write.num_bytes); idx 328 drivers/gpu/drm/drm_dp_mst_topology.c idx += req->u.dpcd_write.num_bytes; idx 331 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.i2c_read.port_number & 0xf) << 4; idx 332 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] |= (req->u.i2c_read.num_transactions & 0x3); idx 333 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 335 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = req->u.i2c_read.transactions[i].i2c_dev_id & 0x7f; idx 336 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 337 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = req->u.i2c_read.transactions[i].num_bytes; idx 338 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 339 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(&buf[idx], req->u.i2c_read.transactions[i].bytes, req->u.i2c_read.transactions[i].num_bytes); idx 340 drivers/gpu/drm/drm_dp_mst_topology.c idx += req->u.i2c_read.transactions[i].num_bytes; idx 342 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.i2c_read.transactions[i].no_stop_bit & 0x1) << 4; idx 343 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] |= (req->u.i2c_read.transactions[i].i2c_transaction_delay & 0xf); idx 344 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 346 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.i2c_read.read_i2c_device_id) & 0x7f; idx 347 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 348 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.i2c_read.num_bytes_read); idx 349 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 353 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.i2c_write.port_number & 0xf) << 4; idx 354 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 355 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.i2c_write.write_i2c_device_id) & 0x7f; idx 356 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 357 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.i2c_write.num_bytes); idx 358 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 359 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(&buf[idx], req->u.i2c_write.bytes, req->u.i2c_write.num_bytes); idx 360 drivers/gpu/drm/drm_dp_mst_topology.c idx += req->u.i2c_write.num_bytes; idx 365 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx] = (req->u.port_num.port_number & 0xf) << 4; idx 366 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 369 drivers/gpu/drm/drm_dp_mst_topology.c raw->cur_len = idx; idx 382 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 0; idx 385 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = (rep->reply_type & 0x1) << 7 | (rep->req_type & 0x7f); idx 387 drivers/gpu/drm/drm_dp_mst_topology.c raw->cur_len = idx; idx 449 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 451 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(repmsg->u.link_addr.guid, &raw->msg[idx], 16); idx 452 drivers/gpu/drm/drm_dp_mst_topology.c idx += 16; idx 453 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.nports = raw->msg[idx] & 0xf; idx 454 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 455 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 458 drivers/gpu/drm/drm_dp_mst_topology.c if (raw->msg[idx] & 0x80) idx 461 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].peer_device_type = (raw->msg[idx] >> 4) & 0x7; idx 462 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].port_number = (raw->msg[idx] & 0xf); idx 464 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 465 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 467 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].mcs = (raw->msg[idx] >> 7) & 0x1; idx 468 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].ddps = (raw->msg[idx] >> 6) & 0x1; idx 470 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].legacy_device_plug_status = (raw->msg[idx] >> 5) & 0x1; idx 471 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 472 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 475 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].dpcd_revision = (raw->msg[idx]); idx 476 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 477 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 479 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(repmsg->u.link_addr.ports[i].peer_guid, &raw->msg[idx], 16); idx 480 drivers/gpu/drm/drm_dp_mst_topology.c idx += 16; idx 481 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 483 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].num_sdp_streams = (raw->msg[idx] >> 4) & 0xf; idx 484 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.link_addr.ports[i].num_sdp_stream_sinks = (raw->msg[idx] & 0xf); idx 485 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 488 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 494 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("link address reply parse length fail %d %d\n", idx, raw->curlen); idx 501 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 502 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.remote_dpcd_read_ack.port_number = raw->msg[idx] & 0xf; idx 503 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 504 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 506 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.remote_dpcd_read_ack.num_bytes = raw->msg[idx]; idx 507 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 508 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 511 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(repmsg->u.remote_dpcd_read_ack.bytes, &raw->msg[idx], repmsg->u.remote_dpcd_read_ack.num_bytes); idx 514 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("link address reply parse length fail %d %d\n", idx, raw->curlen); idx 521 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 522 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.remote_dpcd_write_ack.port_number = raw->msg[idx] & 0xf; idx 523 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 524 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 528 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("parse length fail %d %d\n", idx, raw->curlen); idx 535 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 537 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.remote_i2c_read_ack.port_number = (raw->msg[idx] & 0xf); idx 538 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 539 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 541 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.remote_i2c_read_ack.num_bytes = raw->msg[idx]; idx 542 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 544 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(repmsg->u.remote_i2c_read_ack.bytes, &raw->msg[idx], repmsg->u.remote_i2c_read_ack.num_bytes); idx 547 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("remote i2c reply parse length fail %d %d\n", idx, raw->curlen); idx 554 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 555 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.path_resources.port_number = (raw->msg[idx] >> 4) & 0xf; idx 556 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 557 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 559 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.path_resources.full_payload_bw_number = (raw->msg[idx] << 8) | (raw->msg[idx+1]); idx 560 drivers/gpu/drm/drm_dp_mst_topology.c idx += 2; idx 561 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 563 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.path_resources.avail_payload_bw_number = (raw->msg[idx] << 8) | (raw->msg[idx+1]); idx 564 drivers/gpu/drm/drm_dp_mst_topology.c idx += 2; idx 565 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 569 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("enum resource parse length fail %d %d\n", idx, raw->curlen); idx 576 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 577 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.allocate_payload.port_number = (raw->msg[idx] >> 4) & 0xf; idx 578 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 579 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 581 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.allocate_payload.vcpi = raw->msg[idx]; idx 582 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 583 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 585 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.allocate_payload.allocated_pbn = (raw->msg[idx] << 8) | (raw->msg[idx+1]); idx 586 drivers/gpu/drm/drm_dp_mst_topology.c idx += 2; idx 587 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 591 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("allocate payload parse length fail %d %d\n", idx, raw->curlen); idx 598 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 599 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.query_payload.port_number = (raw->msg[idx] >> 4) & 0xf; idx 600 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 601 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 603 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.query_payload.allocated_pbn = (raw->msg[idx] << 8) | (raw->msg[idx + 1]); idx 604 drivers/gpu/drm/drm_dp_mst_topology.c idx += 2; idx 605 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 609 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("query payload parse length fail %d %d\n", idx, raw->curlen); idx 616 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 618 drivers/gpu/drm/drm_dp_mst_topology.c repmsg->u.port_number.port_number = (raw->msg[idx] >> 4) & 0xf; idx 619 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 620 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) { idx 622 drivers/gpu/drm/drm_dp_mst_topology.c idx, raw->curlen); idx 670 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 672 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.conn_stat.port_number = (raw->msg[idx] & 0xf0) >> 4; idx 673 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 674 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 677 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(msg->u.conn_stat.guid, &raw->msg[idx], 16); idx 678 drivers/gpu/drm/drm_dp_mst_topology.c idx += 16; idx 679 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 682 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.conn_stat.legacy_device_plug_status = (raw->msg[idx] >> 6) & 0x1; idx 683 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.conn_stat.displayport_device_plug_status = (raw->msg[idx] >> 5) & 0x1; idx 684 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.conn_stat.message_capability_status = (raw->msg[idx] >> 4) & 0x1; idx 685 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.conn_stat.input_port = (raw->msg[idx] >> 3) & 0x1; idx 686 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.conn_stat.peer_device_type = (raw->msg[idx] & 0x7); idx 687 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 690 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("connection status reply parse length fail %d %d\n", idx, raw->curlen); idx 697 drivers/gpu/drm/drm_dp_mst_topology.c int idx = 1; idx 699 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.resource_stat.port_number = (raw->msg[idx] & 0xf0) >> 4; idx 700 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 701 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 704 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(msg->u.resource_stat.guid, &raw->msg[idx], 16); idx 705 drivers/gpu/drm/drm_dp_mst_topology.c idx += 16; idx 706 drivers/gpu/drm/drm_dp_mst_topology.c if (idx > raw->curlen) idx 709 drivers/gpu/drm/drm_dp_mst_topology.c msg->u.resource_stat.available_pbn = (raw->msg[idx] << 8) | (raw->msg[idx + 1]); idx 710 drivers/gpu/drm/drm_dp_mst_topology.c idx++; idx 713 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("resource status reply parse length fail %d %d\n", idx, raw->curlen); idx 1446 drivers/gpu/drm/drm_dp_mst_topology.c int idx = (parent_lct - 1) / 2; idx 1448 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(rad, port->parent->rad, idx + 1); idx 1453 drivers/gpu/drm/drm_dp_mst_topology.c rad[idx] |= port->port_num << shift; idx 2001 drivers/gpu/drm/drm_dp_mst_topology.c int len, space, idx, tosend; idx 2031 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_encode_sideband_msg_hdr(&hdr, chunk, &idx); idx 2032 drivers/gpu/drm/drm_dp_mst_topology.c memcpy(&chunk[idx], &txmsg->msg[txmsg->cur_offset], tosend); idx 2034 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_crc_sideband_chunk_req(&chunk[idx], tosend); idx 2035 drivers/gpu/drm/drm_dp_mst_topology.c idx += tosend + 1; idx 2037 drivers/gpu/drm/drm_dp_mst_topology.c ret = drm_dp_send_sideband_msg(mgr, up, chunk, idx); idx 465 drivers/gpu/drm/drm_drv.c bool drm_dev_enter(struct drm_device *dev, int *idx) idx 467 drivers/gpu/drm/drm_drv.c *idx = srcu_read_lock(&drm_unplug_srcu); idx 470 drivers/gpu/drm/drm_drv.c srcu_read_unlock(&drm_unplug_srcu, *idx); idx 485 drivers/gpu/drm/drm_drv.c void drm_dev_exit(int idx) idx 487 drivers/gpu/drm/drm_drv.c srcu_read_unlock(&drm_unplug_srcu, idx); idx 1349 drivers/gpu/drm/drm_edid.c static int validate_displayid(u8 *displayid, int length, int idx); idx 2943 drivers/gpu/drm/drm_edid.c int idx = 1; idx 2959 drivers/gpu/drm/drm_edid.c ret = validate_displayid(displayid, length, idx); idx 2963 drivers/gpu/drm/drm_edid.c idx += sizeof(struct displayid_hdr); idx 2964 drivers/gpu/drm/drm_edid.c for_each_displayid_db(displayid, block, idx, length) { idx 4718 drivers/gpu/drm/drm_edid.c static int validate_displayid(u8 *displayid, int length, int idx) idx 4724 drivers/gpu/drm/drm_edid.c base = (struct displayid_hdr *)&displayid[idx]; idx 4729 drivers/gpu/drm/drm_edid.c if (base->bytes + 5 > length - idx) idx 4731 drivers/gpu/drm/drm_edid.c for (i = idx; i <= base->bytes + 5; i++) { idx 4817 drivers/gpu/drm/drm_edid.c int idx = 1; idx 4826 drivers/gpu/drm/drm_edid.c ret = validate_displayid(displayid, length, idx); idx 4830 drivers/gpu/drm/drm_edid.c idx += sizeof(struct displayid_hdr); idx 4831 drivers/gpu/drm/drm_edid.c for_each_displayid_db(displayid, block, idx, length) { idx 5436 drivers/gpu/drm/drm_edid.c int idx = 0; idx 5441 drivers/gpu/drm/drm_edid.c idx = 1; idx 5443 drivers/gpu/drm/drm_edid.c ret = validate_displayid(displayid, length, idx); idx 5447 drivers/gpu/drm/drm_edid.c idx += sizeof(struct displayid_hdr); idx 5448 drivers/gpu/drm/drm_edid.c for_each_displayid_db(displayid, block, idx, length) { idx 249 drivers/gpu/drm/drm_ioc32.c int idx; /* Which client desired? */ idx 268 drivers/gpu/drm/drm_ioc32.c client.idx = c32.idx; idx 274 drivers/gpu/drm/drm_ioc32.c c32.idx = client.idx; idx 418 drivers/gpu/drm/drm_ioc32.c int idx; /**< Index into the master buffer list */ idx 430 drivers/gpu/drm/drm_ioc32.c static int map_one_buf32(void *data, int idx, unsigned long virtual, idx 434 drivers/gpu/drm/drm_ioc32.c drm_buf_pub32_t __user *to = compat_ptr(request->list) + idx; idx 437 drivers/gpu/drm/drm_ioc32.c v.idx = buf->idx; idx 198 drivers/gpu/drm/drm_ioctl.c if (client->idx == 0) { idx 250 drivers/gpu/drm/drm_mipi_dbi.c int idx, ret = 0; idx 257 drivers/gpu/drm/drm_mipi_dbi.c if (!drm_dev_enter(fb->dev, &idx)) idx 287 drivers/gpu/drm/drm_mipi_dbi.c drm_dev_exit(idx); idx 342 drivers/gpu/drm/drm_mipi_dbi.c int idx; idx 344 drivers/gpu/drm/drm_mipi_dbi.c if (!drm_dev_enter(&dbidev->drm, &idx)) idx 351 drivers/gpu/drm/drm_mipi_dbi.c drm_dev_exit(idx); idx 362 drivers/gpu/drm/drm_mipi_dbi.c int idx; idx 364 drivers/gpu/drm/drm_mipi_dbi.c if (!drm_dev_enter(drm, &idx)) idx 376 drivers/gpu/drm/drm_mipi_dbi.c drm_dev_exit(idx); idx 1190 drivers/gpu/drm/drm_mipi_dbi.c int i, ret, idx; idx 1192 drivers/gpu/drm/drm_mipi_dbi.c if (!drm_dev_enter(&dbidev->drm, &idx)) idx 1236 drivers/gpu/drm/drm_mipi_dbi.c drm_dev_exit(idx); idx 1246 drivers/gpu/drm/drm_mipi_dbi.c int ret, idx; idx 1249 drivers/gpu/drm/drm_mipi_dbi.c if (!drm_dev_enter(&dbidev->drm, &idx)) idx 1281 drivers/gpu/drm/drm_mipi_dbi.c drm_dev_exit(idx); idx 394 drivers/gpu/drm/drm_plane.c drm_plane_from_index(struct drm_device *dev, int idx) idx 399 drivers/gpu/drm/drm_plane.c if (idx == plane->index) idx 874 drivers/gpu/drm/drm_syncobj.c uint32_t *idx) idx 928 drivers/gpu/drm/drm_syncobj.c if (signaled_count == 0 && idx) idx 929 drivers/gpu/drm/drm_syncobj.c *idx = i; idx 970 drivers/gpu/drm/drm_syncobj.c if (idx) idx 971 drivers/gpu/drm/drm_syncobj.c *idx = i; idx 257 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c static int submit_bo(struct etnaviv_gem_submit *submit, u32 idx, idx 260 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c if (idx >= submit->nr_bos) { idx 262 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c idx, submit->nr_bos); idx 266 drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c *bo = &submit->bos[idx]; idx 318 drivers/gpu/drm/exynos/exynos_drm_dsi.c static inline void exynos_dsi_write(struct exynos_dsi *dsi, enum reg_idx idx, idx 322 drivers/gpu/drm/exynos/exynos_drm_dsi.c writel(val, dsi->reg_base + dsi->driver_data->reg_ofs[idx]); idx 325 drivers/gpu/drm/exynos/exynos_drm_dsi.c static inline u32 exynos_dsi_read(struct exynos_dsi *dsi, enum reg_idx idx) idx 327 drivers/gpu/drm/exynos/exynos_drm_dsi.c return readl(dsi->reg_base + dsi->driver_data->reg_ofs[idx]); idx 87 drivers/gpu/drm/i810/i810_dma.c DRM_ERROR("Freeing buffer thats not in use : %d\n", buf->idx); idx 197 drivers/gpu/drm/i810/i810_dma.c d->request_idx = buf->idx; idx 942 drivers/gpu/drm/i810/i810_dma.c vertex->idx, vertex->used, vertex->discard); idx 944 drivers/gpu/drm/i810/i810_dma.c if (vertex->idx < 0 || vertex->idx >= dma->buf_count) idx 948 drivers/gpu/drm/i810/i810_dma.c dma->buflist[vertex->idx], idx 1104 drivers/gpu/drm/i810/i810_dma.c if (mc->idx >= dma->buf_count || mc->idx < 0) idx 1107 drivers/gpu/drm/i810/i810_dma.c i810_dma_dispatch_mc(dev, dma->buflist[mc->idx], mc->used, idx 229 drivers/gpu/drm/i915/display/intel_atomic.c const char *name, int idx, idx 249 drivers/gpu/drm/i915/display/intel_atomic.c if (WARN(*scaler_id < 0, "Cannot find scaler for %s:%d\n", name, idx)) idx 290 drivers/gpu/drm/i915/display/intel_atomic.c intel_crtc->pipe, *scaler_id, name, idx); idx 350 drivers/gpu/drm/i915/display/intel_atomic.c int idx; idx 358 drivers/gpu/drm/i915/display/intel_atomic.c idx = intel_crtc->base.base.id; idx 392 drivers/gpu/drm/i915/display/intel_atomic.c idx = plane->base.id; idx 404 drivers/gpu/drm/i915/display/intel_atomic.c intel_crtc, name, idx, idx 9253 drivers/gpu/drm/i915/display/intel_display.c int idx = BEND_IDX(steps); idx 9258 drivers/gpu/drm/i915/display/intel_display.c if (WARN_ON(idx >= ARRAY_SIZE(sscdivintphase))) idx 9271 drivers/gpu/drm/i915/display/intel_display.c tmp |= sscdivintphase[idx]; idx 319 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 352 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 388 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 432 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 450 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 485 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 509 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 618 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 898 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->hsw.idx; idx 1059 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->vlv.idx; idx 1107 drivers/gpu/drm/i915/display/intel_display_power.c int pw_idx = power_well->desc->vlv.idx; idx 2728 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = HSW_PW_CTL_IDX_GLOBAL, idx 2749 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = HSW_PW_CTL_IDX_GLOBAL, idx 2791 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DISP2D, idx 2803 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_B_LANES_01, idx 2815 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_B_LANES_23, idx 2827 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_C_LANES_01, idx 2839 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DPIO_TX_C_LANES_23, idx 2848 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DPIO_CMN_BC, idx 2878 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DPIO_CMN_BC, idx 2887 drivers/gpu/drm/i915/display/intel_display_power.c .vlv.idx = PUNIT_PWGT_IDX_DPIO_CMN_D, idx 2921 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_1, idx 2934 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_MISC_IO, idx 2950 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_2, idx 2963 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_A_E, idx 2973 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_B, idx 2983 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_C, idx 2993 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_D, idx 3015 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_1, idx 3032 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_2, idx 3075 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_1, idx 3092 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_2, idx 3132 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_AUX_A, idx 3142 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_AUX_B, idx 3152 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_AUX_C, idx 3162 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_DDI_A, idx 3172 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_B, idx 3182 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_C, idx 3204 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_1, idx 3215 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_AUX_A, idx 3225 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_AUX_B, idx 3235 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_AUX_C, idx 3245 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = CNL_PW_CTL_IDX_AUX_D, idx 3261 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_PW_2, idx 3274 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = GLK_PW_CTL_IDX_DDI_A, idx 3284 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_B, idx 3294 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_C, idx 3304 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = SKL_PW_CTL_IDX_DDI_D, idx 3314 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = CNL_PW_CTL_IDX_DDI_F, idx 3324 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = CNL_PW_CTL_IDX_AUX_F, idx 3372 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_1, idx 3389 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_2, idx 3400 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_3, idx 3413 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_A, idx 3423 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_B, idx 3433 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_C, idx 3443 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_D, idx 3453 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_E, idx 3463 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_F, idx 3473 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_A, idx 3483 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_B, idx 3493 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_C, idx 3504 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_D, idx 3515 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_E, idx 3526 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_F, idx 3537 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT1, idx 3548 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT2, idx 3559 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT3, idx 3570 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_TBT4, idx 3581 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_4, idx 3605 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_1, idx 3622 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_2, idx 3633 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_3, idx 3646 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_A, idx 3656 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_B, idx 3666 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_DDI_C, idx 3676 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_DDI_TC1, idx 3686 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_DDI_TC2, idx 3696 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_DDI_TC3, idx 3706 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_DDI_TC4, idx 3716 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_DDI_TC5, idx 3726 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_DDI_TC6, idx 3736 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_A, idx 3746 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_B, idx 3756 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_AUX_C, idx 3766 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TC1, idx 3777 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TC2, idx 3788 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TC3, idx 3799 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TC4, idx 3810 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TC5, idx 3821 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TC6, idx 3832 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TBT1, idx 3843 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TBT2, idx 3854 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TBT3, idx 3865 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TBT4, idx 3876 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TBT5, idx 3887 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_AUX_TBT6, idx 3898 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = ICL_PW_CTL_IDX_PW_4, idx 3910 drivers/gpu/drm/i915/display/intel_display_power.c .hsw.idx = TGL_PW_CTL_IDX_PW_5, idx 176 drivers/gpu/drm/i915/display/intel_display_power.h u8 idx; idx 187 drivers/gpu/drm/i915/display/intel_display_power.h u8 idx; idx 148 drivers/gpu/drm/i915/gem/i915_gem_context.c int idx; idx 162 drivers/gpu/drm/i915/gem/i915_gem_context.c idx = engine->legacy_idx; idx 164 drivers/gpu/drm/i915/gem/i915_gem_context.c idx = ci->engine_instance; idx 167 drivers/gpu/drm/i915/gem/i915_gem_context.c return i915_gem_context_get_engine(ctx, idx); idx 1371 drivers/gpu/drm/i915/gem/i915_gem_context.c u16 num_siblings, idx; idx 1381 drivers/gpu/drm/i915/gem/i915_gem_context.c if (get_user(idx, &ext->engine_index)) idx 1384 drivers/gpu/drm/i915/gem/i915_gem_context.c if (idx >= set->engines->num_engines) { idx 1386 drivers/gpu/drm/i915/gem/i915_gem_context.c idx, set->engines->num_engines); idx 1390 drivers/gpu/drm/i915/gem/i915_gem_context.c idx = array_index_nospec(idx, set->engines->num_engines); idx 1391 drivers/gpu/drm/i915/gem/i915_gem_context.c if (set->engines->engines[idx]) { idx 1392 drivers/gpu/drm/i915/gem/i915_gem_context.c DRM_DEBUG("Invalid placement[%d], already occupied\n", idx); idx 1441 drivers/gpu/drm/i915/gem/i915_gem_context.c if (cmpxchg(&set->engines->engines[idx], NULL, ce)) { idx 1463 drivers/gpu/drm/i915/gem/i915_gem_context.c u16 idx, num_bonds; idx 1466 drivers/gpu/drm/i915/gem/i915_gem_context.c if (get_user(idx, &ext->virtual_index)) idx 1469 drivers/gpu/drm/i915/gem/i915_gem_context.c if (idx >= set->engines->num_engines) { idx 1471 drivers/gpu/drm/i915/gem/i915_gem_context.c idx, set->engines->num_engines); idx 1475 drivers/gpu/drm/i915/gem/i915_gem_context.c idx = array_index_nospec(idx, set->engines->num_engines); idx 1476 drivers/gpu/drm/i915/gem/i915_gem_context.c if (!set->engines->engines[idx]) { idx 1477 drivers/gpu/drm/i915/gem/i915_gem_context.c DRM_DEBUG("Invalid engine at %d\n", idx); idx 1480 drivers/gpu/drm/i915/gem/i915_gem_context.c virtual = set->engines->engines[idx]->engine; idx 2374 drivers/gpu/drm/i915/gem/i915_gem_context.c if (it->idx >= e->num_engines) idx 2377 drivers/gpu/drm/i915/gem/i915_gem_context.c ctx = e->engines[it->idx++]; idx 199 drivers/gpu/drm/i915/gem/i915_gem_context.h i915_gem_context_get_engine(struct i915_gem_context *ctx, unsigned int idx) idx 205 drivers/gpu/drm/i915/gem/i915_gem_context.h if (likely(idx < e->num_engines && e->engines[idx])) idx 206 drivers/gpu/drm/i915/gem/i915_gem_context.h ce = intel_context_get(e->engines[idx]); idx 218 drivers/gpu/drm/i915/gem/i915_gem_context.h it->idx = 0; idx 39 drivers/gpu/drm/i915/gem/i915_gem_context_types.h unsigned int idx; idx 2329 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c unsigned int idx; idx 2333 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c idx = args->flags & I915_EXEC_RING_MASK; idx 2335 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c idx = eb_select_legacy_ring(eb, file, args); idx 2337 drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c ce = i915_gem_context_get_engine(eb->gem_context, idx); idx 380 drivers/gpu/drm/i915/gem/i915_gem_pages.c unsigned int idx, count; idx 406 drivers/gpu/drm/i915/gem/i915_gem_pages.c idx = iter->sg_idx; idx 409 drivers/gpu/drm/i915/gem/i915_gem_pages.c while (idx + count <= n) { idx 422 drivers/gpu/drm/i915/gem/i915_gem_pages.c ret = radix_tree_insert(&iter->radix, idx, sg); idx 426 drivers/gpu/drm/i915/gem/i915_gem_pages.c entry = xa_mk_value(idx); idx 428 drivers/gpu/drm/i915/gem/i915_gem_pages.c ret = radix_tree_insert(&iter->radix, idx + i, entry); idx 433 drivers/gpu/drm/i915/gem/i915_gem_pages.c idx += count; idx 440 drivers/gpu/drm/i915/gem/i915_gem_pages.c iter->sg_idx = idx; idx 444 drivers/gpu/drm/i915/gem/i915_gem_pages.c if (unlikely(n < idx)) /* insertion completed by another thread */ idx 450 drivers/gpu/drm/i915/gem/i915_gem_pages.c while (idx + count <= n) { idx 451 drivers/gpu/drm/i915/gem/i915_gem_pages.c idx += count; idx 456 drivers/gpu/drm/i915/gem/i915_gem_pages.c *offset = n - idx; idx 243 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c unsigned int idx, unsigned int max) idx 262 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c __builtin_return_address(0), idx, idx 273 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c __builtin_return_address(0), idx, n, m, idx 1046 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c unsigned long idx, ndwords, dw; idx 1123 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c idx = 0; idx 1133 drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c err = cpu_check(obj, idx++, num_writes); idx 1254 drivers/gpu/drm/i915/gt/intel_engine_cs.c unsigned int idx; idx 1277 drivers/gpu/drm/i915/gt/intel_engine_cs.c idx = ++read % num_entries; idx 1279 drivers/gpu/drm/i915/gt/intel_engine_cs.c idx, hws[idx * 2], hws[idx * 2 + 1]); idx 174 drivers/gpu/drm/i915/gt/intel_engine_user.c int idx; idx 182 drivers/gpu/drm/i915/gt/intel_engine_user.c idx = legacy_ring_idx(ring); idx 183 drivers/gpu/drm/i915/gt/intel_engine_user.c if (unlikely(idx == -1)) idx 186 drivers/gpu/drm/i915/gt/intel_engine_user.c GEM_BUG_ON(idx >= ARRAY_SIZE(ring->gt->engine)); idx 187 drivers/gpu/drm/i915/gt/intel_engine_user.c ring->gt->engine[idx] = engine; idx 190 drivers/gpu/drm/i915/gt/intel_engine_user.c engine->legacy_idx = idx; idx 574 drivers/gpu/drm/i915/gt/intel_lrc.c execlists_schedule_in(struct i915_request *rq, int idx) idx 580 drivers/gpu/drm/i915/gt/intel_lrc.c trace_i915_request_in(rq, idx); idx 740 drivers/gpu/drm/i915/gt/selftest_hangcheck.c unsigned int idx = count++ & (ARRAY_SIZE(rq) - 1); idx 741 drivers/gpu/drm/i915/gt/selftest_hangcheck.c struct i915_request *old = rq[idx]; idx 745 drivers/gpu/drm/i915/gt/selftest_hangcheck.c new = igt_request_alloc(ctx[idx], engine); idx 753 drivers/gpu/drm/i915/gt/selftest_hangcheck.c ctx[idx]->sched.priority = idx 756 drivers/gpu/drm/i915/gt/selftest_hangcheck.c rq[idx] = i915_request_get(new); idx 83 drivers/gpu/drm/i915/gt/selftest_lrc.c emit_semaphore_chain(struct i915_request *rq, struct i915_vma *vma, int idx) idx 98 drivers/gpu/drm/i915/gt/selftest_lrc.c *cs++ = i915_ggtt_offset(vma) + 4 * idx; idx 101 drivers/gpu/drm/i915/gt/selftest_lrc.c if (idx > 0) { idx 103 drivers/gpu/drm/i915/gt/selftest_lrc.c *cs++ = i915_ggtt_offset(vma) + 4 * (idx - 1); idx 120 drivers/gpu/drm/i915/gt/selftest_lrc.c semaphore_queue(struct intel_engine_cs *engine, struct i915_vma *vma, int idx) idx 134 drivers/gpu/drm/i915/gt/selftest_lrc.c err = emit_semaphore_chain(rq, vma, idx); idx 147 drivers/gpu/drm/i915/gt/selftest_lrc.c int idx) idx 166 drivers/gpu/drm/i915/gt/selftest_lrc.c *cs++ = i915_ggtt_offset(vma) + 4 * (idx - 1); idx 49 drivers/gpu/drm/i915/gt/selftest_timeline.c unsigned int idx, idx 52 drivers/gpu/drm/i915/gt/selftest_timeline.c tl = xchg(&state->history[idx], tl); idx 64 drivers/gpu/drm/i915/gt/selftest_timeline.c unsigned int idx; idx 85 drivers/gpu/drm/i915/gt/selftest_timeline.c idx = state->count++ % state->max; idx 86 drivers/gpu/drm/i915/gt/selftest_timeline.c __mock_hwsp_record(state, idx, tl); idx 98 drivers/gpu/drm/i915/gt/selftest_timeline.c idx = --state->count % state->max; idx 99 drivers/gpu/drm/i915/gt/selftest_timeline.c __mock_hwsp_record(state, idx, NULL); idx 487 drivers/gpu/drm/i915/gt/selftest_workarounds.c int idx; idx 515 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx = 1; idx 525 drivers/gpu/drm/i915/gt/selftest_workarounds.c *cs++ = lower_32_bits(addr + sizeof(u32) * idx); idx 526 drivers/gpu/drm/i915/gt/selftest_workarounds.c *cs++ = upper_32_bits(addr + sizeof(u32) * idx); idx 527 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx++; idx 538 drivers/gpu/drm/i915/gt/selftest_workarounds.c *cs++ = lower_32_bits(addr + sizeof(u32) * idx); idx 539 drivers/gpu/drm/i915/gt/selftest_workarounds.c *cs++ = upper_32_bits(addr + sizeof(u32) * idx); idx 540 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx++; idx 542 drivers/gpu/drm/i915/gt/selftest_workarounds.c GEM_BUG_ON(idx * sizeof(u32) > scratch->size); idx 606 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx = 1; idx 613 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (results[idx] != expect) idx 615 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx++; idx 623 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (results[idx] != expect) idx 625 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx++; idx 639 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx = 1; idx 648 drivers/gpu/drm/i915/gt/selftest_workarounds.c w, results[idx], expect); idx 649 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx++; idx 659 drivers/gpu/drm/i915/gt/selftest_workarounds.c w, results[idx], expect); idx 660 drivers/gpu/drm/i915/gt/selftest_workarounds.c idx++; idx 514 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c static struct i915_request *schedule_in(struct i915_request *rq, int idx) idx 516 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c trace_i915_request_in(rq, idx); idx 613 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c int idx = port - execlists->inflight; idx 614 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c int rem = ARRAY_SIZE(execlists->inflight) - idx; idx 1005 drivers/gpu/drm/i915/gvt/handlers.c int idx, i, ret = 0; idx 1028 drivers/gpu/drm/i915/gvt/handlers.c for (idx = 1; idx <= 5; idx++) { idx 1030 drivers/gpu/drm/i915/gvt/handlers.c vgpu_vreg(vgpu, offset + 4 * idx) = 0; idx 1639 drivers/gpu/drm/i915/gvt/kvmgt.c int idx; idx 1647 drivers/gpu/drm/i915/gvt/kvmgt.c idx = srcu_read_lock(&kvm->srcu); idx 1650 drivers/gpu/drm/i915/gvt/kvmgt.c srcu_read_unlock(&kvm->srcu, idx); idx 1664 drivers/gpu/drm/i915/gvt/kvmgt.c srcu_read_unlock(&kvm->srcu, idx); idx 1673 drivers/gpu/drm/i915/gvt/kvmgt.c int idx; idx 1681 drivers/gpu/drm/i915/gvt/kvmgt.c idx = srcu_read_lock(&kvm->srcu); idx 1684 drivers/gpu/drm/i915/gvt/kvmgt.c srcu_read_unlock(&kvm->srcu, idx); idx 1698 drivers/gpu/drm/i915/gvt/kvmgt.c srcu_read_unlock(&kvm->srcu, idx); idx 1967 drivers/gpu/drm/i915/gvt/kvmgt.c int idx, ret; idx 1982 drivers/gpu/drm/i915/gvt/kvmgt.c idx = srcu_read_lock(&kvm->srcu); idx 1985 drivers/gpu/drm/i915/gvt/kvmgt.c srcu_read_unlock(&kvm->srcu, idx); idx 2016 drivers/gpu/drm/i915/gvt/kvmgt.c int idx; idx 2025 drivers/gpu/drm/i915/gvt/kvmgt.c idx = srcu_read_lock(&kvm->srcu); idx 2027 drivers/gpu/drm/i915/gvt/kvmgt.c srcu_read_unlock(&kvm->srcu, idx); idx 186 drivers/gpu/drm/i915/i915_active.c u64 idx = tl->fence_context; idx 196 drivers/gpu/drm/i915/i915_active.c if (node && node->timeline == idx) idx 213 drivers/gpu/drm/i915/i915_active.c if (node->timeline == idx) { idx 218 drivers/gpu/drm/i915/i915_active.c if (node->timeline < idx) idx 227 drivers/gpu/drm/i915/i915_active.c node->timeline = idx; idx 495 drivers/gpu/drm/i915/i915_active.c static inline bool is_idle_barrier(struct active_node *node, u64 idx) idx 497 drivers/gpu/drm/i915/i915_active.c return node->timeline == idx && !i915_active_request_isset(&node->base); idx 500 drivers/gpu/drm/i915/i915_active.c static struct active_node *reuse_idle_barrier(struct i915_active *ref, u64 idx) idx 517 drivers/gpu/drm/i915/i915_active.c if (ref->cache && is_idle_barrier(ref->cache, idx)) { idx 528 drivers/gpu/drm/i915/i915_active.c if (is_idle_barrier(node, idx)) idx 532 drivers/gpu/drm/i915/i915_active.c if (node->timeline < idx) idx 549 drivers/gpu/drm/i915/i915_active.c if (node->timeline > idx) idx 552 drivers/gpu/drm/i915/i915_active.c if (node->timeline < idx) idx 555 drivers/gpu/drm/i915/i915_active.c if (is_idle_barrier(node, idx)) idx 602 drivers/gpu/drm/i915/i915_active.c u64 idx = engine->kernel_context->timeline->fence_context; idx 605 drivers/gpu/drm/i915/i915_active.c node = reuse_idle_barrier(ref, idx); idx 619 drivers/gpu/drm/i915/i915_active.c node->timeline = idx; idx 586 drivers/gpu/drm/i915/i915_cmd_parser.c #define REG64_IDX(_reg, idx) \ idx 587 drivers/gpu/drm/i915/i915_cmd_parser.c { .addr = _reg(idx) }, \ idx 588 drivers/gpu/drm/i915/i915_cmd_parser.c { .addr = _reg ## _UDW(idx) } idx 267 drivers/gpu/drm/i915/i915_gem.c unsigned int idx, offset; idx 285 drivers/gpu/drm/i915/i915_gem.c for (idx = args->offset >> PAGE_SHIFT; remain; idx++) { idx 286 drivers/gpu/drm/i915/i915_gem.c struct page *page = i915_gem_object_get_page(obj, idx); idx 685 drivers/gpu/drm/i915/i915_gem.c unsigned int offset, idx; idx 711 drivers/gpu/drm/i915/i915_gem.c for (idx = args->offset >> PAGE_SHIFT; remain; idx++) { idx 712 drivers/gpu/drm/i915/i915_gem.c struct page *page = i915_gem_object_get_page(obj, idx); idx 760 drivers/gpu/drm/i915/i915_gem_gtt.c const unsigned short idx, idx 765 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr[idx] = encoded_entry; idx 771 drivers/gpu/drm/i915/i915_gem_gtt.c const unsigned short idx, idx 779 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[idx] = to; idx 780 drivers/gpu/drm/i915/i915_gem_gtt.c write_dma_entry(px_base(pd), idx, encode(to->daddr, I915_CACHE_LLC)); idx 783 drivers/gpu/drm/i915/i915_gem_gtt.c #define set_pd_entry(pd, idx, to) \ idx 784 drivers/gpu/drm/i915/i915_gem_gtt.c __set_pd_entry((pd), (idx), px_base(to), gen8_pde_encode) idx 788 drivers/gpu/drm/i915/i915_gem_gtt.c const unsigned short idx, idx 793 drivers/gpu/drm/i915/i915_gem_gtt.c write_dma_entry(px_base(pd), idx, scratch->encode); idx 794 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[idx] = NULL; idx 800 drivers/gpu/drm/i915/i915_gem_gtt.c const unsigned short idx, idx 811 drivers/gpu/drm/i915/i915_gem_gtt.c clear_pd_entry(pd, idx, scratch); idx 879 drivers/gpu/drm/i915/i915_gem_gtt.c gen8_pd_range(u64 start, u64 end, int lvl, unsigned int *idx) idx 887 drivers/gpu/drm/i915/i915_gem_gtt.c *idx = i915_pde_index(start, shift); idx 889 drivers/gpu/drm/i915/i915_gem_gtt.c return GEN8_PDES - *idx; idx 891 drivers/gpu/drm/i915/i915_gem_gtt.c return i915_pde_index(end, shift) - *idx; idx 918 drivers/gpu/drm/i915/i915_gem_gtt.c gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) idx 925 drivers/gpu/drm/i915/i915_gem_gtt.c return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm->top)); idx 968 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int idx, len; idx 972 drivers/gpu/drm/i915/i915_gem_gtt.c len = gen8_pd_range(start, end, lvl--, &idx); idx 975 drivers/gpu/drm/i915/i915_gem_gtt.c idx, len, atomic_read(px_used(pd))); idx 979 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_table *pt = pd->entry[idx]; idx 984 drivers/gpu/drm/i915/i915_gem_gtt.c __func__, vm, lvl + 1, idx, start, end); idx 985 drivers/gpu/drm/i915/i915_gem_gtt.c clear_pd_entry(pd, idx, scratch); idx 1015 drivers/gpu/drm/i915/i915_gem_gtt.c if (release_pd_entry(pd, idx, pt, scratch)) idx 1017 drivers/gpu/drm/i915/i915_gem_gtt.c } while (idx++, --len); idx 1043 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int idx, len; idx 1048 drivers/gpu/drm/i915/i915_gem_gtt.c len = gen8_pd_range(*start, end, lvl--, &idx); idx 1051 drivers/gpu/drm/i915/i915_gem_gtt.c idx, len, atomic_read(px_used(pd))); idx 1052 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(!len || (idx + len - 1) >> gen8_pd_shift(1)); idx 1057 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_table *pt = pd->entry[idx]; idx 1063 drivers/gpu/drm/i915/i915_gem_gtt.c __func__, vm, lvl + 1, idx); idx 1091 drivers/gpu/drm/i915/i915_gem_gtt.c if (likely(!pd->entry[idx])) idx 1092 drivers/gpu/drm/i915/i915_gem_gtt.c set_pd_entry(pd, idx, pt); idx 1094 drivers/gpu/drm/i915/i915_gem_gtt.c alloc = pt, pt = pd->entry[idx]; idx 1104 drivers/gpu/drm/i915/i915_gem_gtt.c if (release_pd_entry(pd, idx, pt, scratch)) idx 1125 drivers/gpu/drm/i915/i915_gem_gtt.c } while (idx++, --len); idx 1170 drivers/gpu/drm/i915/i915_gem_gtt.c u64 idx, idx 1178 drivers/gpu/drm/i915/i915_gem_gtt.c pd = i915_pd_entry(pdp, gen8_pd_index(idx, 2)); idx 1179 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(i915_pt_entry(pd, gen8_pd_index(idx, 1))); idx 1182 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr[gen8_pd_index(idx, 0)] = pte_encode | iter->dma; idx 1188 drivers/gpu/drm/i915/i915_gem_gtt.c idx = 0; idx 1196 drivers/gpu/drm/i915/i915_gem_gtt.c if (gen8_pd_index(++idx, 0) == 0) { idx 1197 drivers/gpu/drm/i915/i915_gem_gtt.c if (gen8_pd_index(idx, 1) == 0) { idx 1199 drivers/gpu/drm/i915/i915_gem_gtt.c if (gen8_pd_index(idx, 2) == 0) idx 1202 drivers/gpu/drm/i915/i915_gem_gtt.c pd = pdp->entry[gen8_pd_index(idx, 2)]; idx 1206 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(i915_pt_entry(pd, gen8_pd_index(idx, 1))); idx 1211 drivers/gpu/drm/i915/i915_gem_gtt.c return idx; idx 1345 drivers/gpu/drm/i915/i915_gem_gtt.c u64 idx = vma->node.start >> GEN8_PTE_SHIFT; idx 1349 drivers/gpu/drm/i915/i915_gem_gtt.c gen8_pdp_for_page_index(vm, idx); idx 1351 drivers/gpu/drm/i915/i915_gem_gtt.c idx = gen8_ppgtt_insert_pte(ppgtt, pdp, &iter, idx, idx 1353 drivers/gpu/drm/i915/i915_gem_gtt.c } while (idx); idx 1410 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int idx; idx 1415 drivers/gpu/drm/i915/i915_gem_gtt.c for (idx = 0; idx < GEN8_3LVL_PDPES; idx++) { idx 1423 drivers/gpu/drm/i915/i915_gem_gtt.c set_pd_entry(pd, idx, pde); idx 1720 drivers/gpu/drm/i915/i915_reg.h #define BXT_PORT_PLL(phy, ch, idx) _MMIO(_PORT_PLL_BASE(phy, ch) + \ idx 1721 drivers/gpu/drm/i915/i915_reg.h (idx) * 4) idx 78 drivers/gpu/drm/i915/i915_scheduler.c int idx, i; idx 84 drivers/gpu/drm/i915/i915_scheduler.c idx = I915_PRIORITY_COUNT - (prio & I915_PRIORITY_MASK) - 1; idx 135 drivers/gpu/drm/i915/i915_scheduler.c p->used |= BIT(idx); idx 136 drivers/gpu/drm/i915/i915_scheduler.c return &p->requests[idx]; idx 16 drivers/gpu/drm/i915/i915_scheduler.h #define priolist_for_each_request(it, plist, idx) \ idx 17 drivers/gpu/drm/i915/i915_scheduler.h for (idx = 0; idx < ARRAY_SIZE((plist)->requests); idx++) \ idx 18 drivers/gpu/drm/i915/i915_scheduler.h list_for_each_entry(it, &(plist)->requests[idx], sched.link) idx 20 drivers/gpu/drm/i915/i915_scheduler.h #define priolist_for_each_request_consume(it, n, plist, idx) \ idx 22 drivers/gpu/drm/i915/i915_scheduler.h (plist)->used ? (idx = __ffs((plist)->used)), 1 : 0; \ idx 23 drivers/gpu/drm/i915/i915_scheduler.h (plist)->used &= ~BIT(idx)) \ idx 25 drivers/gpu/drm/i915/i915_scheduler.h &(plist)->requests[idx], \ idx 157 drivers/gpu/drm/i915/i915_syncmap.c unsigned int idx; idx 191 drivers/gpu/drm/i915/i915_syncmap.c idx = __sync_leaf_idx(p, id); idx 192 drivers/gpu/drm/i915/i915_syncmap.c if (!(p->bitmap & BIT(idx))) idx 195 drivers/gpu/drm/i915/i915_syncmap.c return seqno_later(__sync_seqno(p)[idx], seqno); idx 216 drivers/gpu/drm/i915/i915_syncmap.c unsigned int idx = __sync_leaf_idx(p, id); idx 218 drivers/gpu/drm/i915/i915_syncmap.c p->bitmap |= BIT(idx); idx 219 drivers/gpu/drm/i915/i915_syncmap.c __sync_seqno(p)[idx] = seqno; idx 223 drivers/gpu/drm/i915/i915_syncmap.c unsigned int idx, idx 226 drivers/gpu/drm/i915/i915_syncmap.c p->bitmap |= BIT(idx); idx 227 drivers/gpu/drm/i915/i915_syncmap.c __sync_child(p)[idx] = child; idx 233 drivers/gpu/drm/i915/i915_syncmap.c unsigned int idx; idx 298 drivers/gpu/drm/i915/i915_syncmap.c idx = __sync_branch_idx(p->parent, id); idx 299 drivers/gpu/drm/i915/i915_syncmap.c __sync_child(p->parent)[idx] = next; idx 300 drivers/gpu/drm/i915/i915_syncmap.c GEM_BUG_ON(!(p->parent->bitmap & BIT(idx))); idx 305 drivers/gpu/drm/i915/i915_syncmap.c idx = p->prefix >> (above - SHIFT) & MASK; idx 306 drivers/gpu/drm/i915/i915_syncmap.c __sync_set_child(next, idx, p); idx 318 drivers/gpu/drm/i915/i915_syncmap.c idx = __sync_branch_idx(p, id); idx 319 drivers/gpu/drm/i915/i915_syncmap.c next = __sync_child(p)[idx]; idx 325 drivers/gpu/drm/i915/i915_syncmap.c __sync_set_child(p, idx, next); idx 33 drivers/gpu/drm/i915/selftests/i915_syncmap.c unsigned int idx) idx 49 drivers/gpu/drm/i915/selftests/i915_syncmap.c len = scnprintf(buf, *sz, "%x-> ", idx); idx 146 drivers/gpu/drm/i915/selftests/i915_syncmap.c static int check_seqno(struct i915_syncmap *leaf, unsigned int idx, u32 seqno) idx 154 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (__sync_seqno(leaf)[idx] != seqno) { idx 156 drivers/gpu/drm/i915/selftests/i915_syncmap.c __func__, idx, __sync_seqno(leaf)[idx], seqno); idx 335 drivers/gpu/drm/i915/selftests/i915_syncmap.c unsigned int step, order, idx; idx 372 drivers/gpu/drm/i915/selftests/i915_syncmap.c for (idx = 1; idx < KSYNCMAP; idx++) { idx 373 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (i915_syncmap_is_later(&sync, context + idx, 0)) { idx 375 drivers/gpu/drm/i915/selftests/i915_syncmap.c context + idx, order, step); idx 415 drivers/gpu/drm/i915/selftests/i915_syncmap.c unsigned int idx; idx 420 drivers/gpu/drm/i915/selftests/i915_syncmap.c for (idx = 0; idx < KSYNCMAP; idx++) { idx 421 drivers/gpu/drm/i915/selftests/i915_syncmap.c err = i915_syncmap_set(&sync, context + idx, 0); idx 432 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (sync->bitmap != BIT(idx + 1) - 1) { idx 434 drivers/gpu/drm/i915/selftests/i915_syncmap.c context, idx, idx 436 drivers/gpu/drm/i915/selftests/i915_syncmap.c BIT(idx + 1) - 1, idx + 1); idx 449 drivers/gpu/drm/i915/selftests/i915_syncmap.c unsigned int idx, order; idx 468 drivers/gpu/drm/i915/selftests/i915_syncmap.c for (idx = 0; idx < KSYNCMAP; idx++) { idx 469 drivers/gpu/drm/i915/selftests/i915_syncmap.c u64 context = idx * BIT_ULL(order) + idx; idx 477 drivers/gpu/drm/i915/selftests/i915_syncmap.c context, order, idx, idx 507 drivers/gpu/drm/i915/selftests/i915_syncmap.c for (idx = 0; idx < KSYNCMAP; idx++) { idx 508 drivers/gpu/drm/i915/selftests/i915_syncmap.c struct i915_syncmap *leaf = __sync_child(sync)[idx]; idx 511 drivers/gpu/drm/i915/selftests/i915_syncmap.c pr_err("Child %d is a not leaf!\n", idx); idx 518 drivers/gpu/drm/i915/selftests/i915_syncmap.c idx); idx 525 drivers/gpu/drm/i915/selftests/i915_syncmap.c idx, leaf->bitmap, hweight32(leaf->bitmap)); idx 530 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (leaf->bitmap != BIT(idx)) { idx 532 drivers/gpu/drm/i915/selftests/i915_syncmap.c idx, ilog2(leaf->bitmap), idx); idx 270 drivers/gpu/drm/lima/lima_vm.c int idx = (i << LIMA_VM_NUM_PT_PER_BT_SHIFT) + j; idx 272 drivers/gpu/drm/lima/lima_vm.c printk(KERN_INFO "lima vm pd %03x:%08x\n", idx, pd[idx]); idx 132 drivers/gpu/drm/mediatek/mtk_disp_ovl.c static void mtk_ovl_layer_on(struct mtk_ddp_comp *comp, unsigned int idx) idx 136 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel(0x1, comp->regs + DISP_REG_OVL_RDMA_CTRL(idx)); idx 137 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel(OVL_RDMA_MEM_GMC, comp->regs + DISP_REG_OVL_RDMA_GMC(idx)); idx 140 drivers/gpu/drm/mediatek/mtk_disp_ovl.c reg = reg | BIT(idx); idx 144 drivers/gpu/drm/mediatek/mtk_disp_ovl.c static void mtk_ovl_layer_off(struct mtk_ddp_comp *comp, unsigned int idx) idx 149 drivers/gpu/drm/mediatek/mtk_disp_ovl.c reg = reg & ~BIT(idx); idx 152 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel(0x0, comp->regs + DISP_REG_OVL_RDMA_CTRL(idx)); idx 191 drivers/gpu/drm/mediatek/mtk_disp_ovl.c static void mtk_ovl_layer_config(struct mtk_ddp_comp *comp, unsigned int idx, idx 204 drivers/gpu/drm/mediatek/mtk_disp_ovl.c mtk_ovl_layer_off(comp, idx); idx 207 drivers/gpu/drm/mediatek/mtk_disp_ovl.c if (idx != 0) idx 210 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel_relaxed(con, comp->regs + DISP_REG_OVL_CON(idx)); idx 211 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel_relaxed(pitch, comp->regs + DISP_REG_OVL_PITCH(idx)); idx 212 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel_relaxed(src_size, comp->regs + DISP_REG_OVL_SRC_SIZE(idx)); idx 213 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel_relaxed(offset, comp->regs + DISP_REG_OVL_OFFSET(idx)); idx 214 drivers/gpu/drm/mediatek/mtk_disp_ovl.c writel_relaxed(addr, comp->regs + DISP_REG_OVL_ADDR(ovl, idx)); idx 217 drivers/gpu/drm/mediatek/mtk_disp_ovl.c mtk_ovl_layer_on(comp, idx); idx 192 drivers/gpu/drm/mediatek/mtk_disp_rdma.c static void mtk_rdma_layer_config(struct mtk_ddp_comp *comp, unsigned int idx, idx 74 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h void (*layer_on)(struct mtk_ddp_comp *comp, unsigned int idx); idx 75 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h void (*layer_off)(struct mtk_ddp_comp *comp, unsigned int idx); idx 76 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h void (*layer_config)(struct mtk_ddp_comp *comp, unsigned int idx, idx 133 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h unsigned int idx) idx 136 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h comp->funcs->layer_on(comp, idx); idx 140 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h unsigned int idx) idx 143 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h comp->funcs->layer_off(comp, idx); idx 147 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h unsigned int idx, idx 151 drivers/gpu/drm/mediatek/mtk_drm_ddp_comp.h comp->funcs->layer_config(comp, idx, state); idx 237 drivers/gpu/drm/mga/mga_dma.c entry, entry->buf->idx, entry->age.head, idx 1098 drivers/gpu/drm/mga/mga_dma.c &buf->idx, sizeof(buf->idx))) idx 631 drivers/gpu/drm/mga/mga_state.c DRM_DEBUG("buf=%d used=%d\n", buf->idx, buf->used); idx 678 drivers/gpu/drm/mga/mga_state.c DRM_DEBUG("buf=%d start=%d end=%d\n", buf->idx, start, end); idx 728 drivers/gpu/drm/mga/mga_state.c DRM_DEBUG("buf=%d used=%d\n", buf->idx, buf->used); idx 880 drivers/gpu/drm/mga/mga_state.c if (vertex->idx < 0 || vertex->idx > dma->buf_count) idx 882 drivers/gpu/drm/mga/mga_state.c buf = dma->buflist[vertex->idx]; idx 915 drivers/gpu/drm/mga/mga_state.c if (indices->idx < 0 || indices->idx > dma->buf_count) idx 918 drivers/gpu/drm/mga/mga_state.c buf = dma->buflist[indices->idx]; idx 958 drivers/gpu/drm/mga/mga_state.c if (iload->idx < 0 || iload->idx > dma->buf_count) idx 961 drivers/gpu/drm/mga/mga_state.c buf = dma->buflist[iload->idx]; idx 65 drivers/gpu/drm/msm/adreno/a5xx_gpu.c obj = submit->bos[submit->cmd[i].idx].obj; idx 223 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c mixer[i].hw_lm->idx); idx 229 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c mixer[i].hw_lm->idx - LM_0, idx 231 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c ctl->idx - CTL_0, idx 234 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c ctl->ops.setup_blendstage(ctl, mixer[i].hw_lm->idx, idx 1113 drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c m->hw_lm->idx - LM_0, m->lm_ctl->idx - CTL_0, idx 38 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c (p) ? ((p)->hw_pp ? (p)->hw_pp->idx - PINGPONG_0 : -1) : -1, \ idx 44 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c (p) ? ((p)->hw_pp ? (p)->hw_pp->idx - PINGPONG_0 : -1) : -1, \ idx 217 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c phys_enc->hw_pp->idx - PINGPONG_0, intr_idx); idx 261 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c irq->irq_idx, phys_enc->hw_pp->idx - PINGPONG_0, idx 279 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c phys_enc->hw_pp->idx - PINGPONG_0, idx 291 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c phys_enc->hw_pp->idx - PINGPONG_0, idx 298 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c phys_enc->hw_pp->idx - PINGPONG_0, idx 662 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c vsync_cfg.ppnumber[i] = dpu_enc->hw_pp[i]->idx; idx 1066 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c if (hw_intf->idx == phys->intf_idx) idx 1454 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c pending_kickoff_cnt, ctl->idx, idx 1490 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c trace_dpu_enc_trigger_start(DRMID(phys_enc->parent), ctl->idx); idx 1536 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c ctl->idx); idx 1540 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c DPU_ERROR_ENC(dpu_enc, "ctl %d reset failure\n", ctl->idx); idx 96 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 160 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->hw_idx = phys_enc->hw_ctl->idx; idx 164 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->hw_idx = phys_enc->hw_pp->idx; idx 168 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c irq->hw_idx = phys_enc->hw_pp->idx; idx 215 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 224 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 225 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_ctl->idx - CTL_0, idx 296 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 309 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, ret, idx 327 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 367 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c DPU_DEBUG_CMDENC(cmd_enc, "pp %d\n", phys_enc->hw_pp->idx - PINGPONG_0); idx 418 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, vsync_hz, idx 422 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, tc_enable, tc_cfg.start_pos, idx 426 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, tc_cfg.hw_vsync_mode, idx 430 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, tc_cfg.sync_cfg_height, idx 450 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0); idx 500 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c DPU_DEBUG_CMDENC(cmd_enc, "pp %d\n", phys_enc->hw_pp->idx - PINGPONG_0); idx 556 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 600 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 613 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0); idx 617 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c phys_enc->hw_pp->idx - PINGPONG_0, idx 16 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c (e)->hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__) idx 22 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c (e)->hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__) idx 274 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c intf_cfg.intf = phys_enc->hw_intf->idx; idx 428 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c phys_enc->hw_intf->idx - INTF_0, ret, enable, idx 446 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c dpu_encoder_helper_split_config(phys_enc, phys_enc->hw_intf->idx); idx 459 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c ctl->ops.get_bitmask_intf(ctl, &flush_mask, phys_enc->hw_intf->idx); idx 465 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c ctl->idx - CTL_0, flush_mask); idx 562 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c ctl->idx, rc); idx 614 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c phys_enc->hw_intf->idx - INTF_0, ret); idx 632 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c phys_enc->hw_intf->idx - INTF_0); idx 649 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c phys_enc->hw_intf->idx - INTF_0, idx 251 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c pr_debug("issuing hw ctl reset for ctl:%d\n", ctx->idx); idx 269 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c pr_debug("hw ctl reset is set for ctl:%d\n", ctx->idx); idx 271 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c pr_err("hw recovery is not complete for ctl:%d\n", ctx->idx); idx 477 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c struct dpu_hw_ctl *dpu_hw_ctl_init(enum dpu_ctl idx, idx 488 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c cfg = _ctl_offset(idx, m, addr, &c->hw); idx 491 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c pr_err("failed to create dpu_hw_ctl %d\n", idx); idx 497 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c c->idx = idx; idx 501 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c dpu_hw_blk_init(&c->base, DPU_HW_BLK_CTL, idx, &dpu_hw_ops); idx 169 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h int idx; idx 196 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h struct dpu_hw_ctl *dpu_hw_ctl_init(enum dpu_ctl idx, idx 261 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c struct dpu_hw_intf *dpu_hw_intf_init(enum dpu_intf idx, idx 272 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c cfg = _intf_offset(idx, m, addr, &c->hw); idx 275 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c pr_err("failed to create dpu_hw_intf %d\n", idx); idx 282 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c c->idx = idx; idx 287 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.c dpu_hw_blk_init(&c->base, DPU_HW_BLK_INTF, idx, &dpu_hw_ops); idx 78 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.h enum dpu_intf idx; idx 93 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_intf.h struct dpu_hw_intf *dpu_hw_intf_init(enum dpu_intf idx, idx 165 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c struct dpu_hw_mixer *dpu_hw_lm_init(enum dpu_lm idx, idx 176 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c cfg = _lm_offset(idx, m, addr, &c->hw); idx 183 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c c->idx = idx; idx 187 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.c dpu_hw_blk_init(&c->base, DPU_HW_BLK_LM, idx, &dpu_hw_ops); idx 63 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.h enum dpu_lm idx; idx 92 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_lm.h struct dpu_hw_mixer *dpu_hw_lm_init(enum dpu_lm idx, idx 129 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c trace_dpu_pp_connect_ext_te(pp->idx - PINGPONG_0, cfg); idx 196 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c struct dpu_hw_pingpong *dpu_hw_pingpong_init(enum dpu_pingpong idx, idx 207 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c cfg = _pingpong_offset(idx, m, addr, &c->hw); idx 213 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c c->idx = idx; idx 217 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.c dpu_hw_blk_init(&c->base, DPU_HW_BLK_PINGPONG, idx, &dpu_hw_ops); idx 92 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.h enum dpu_pingpong idx; idx 107 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_pingpong.h struct dpu_hw_pingpong *dpu_hw_pingpong_init(enum dpu_pingpong idx, idx 137 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 *idx) idx 147 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c *idx = sblk->src_blk.base; idx 152 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c *idx = sblk->scaler_blk.base; idx 156 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c *idx = sblk->csc_blk.base; idx 170 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 172 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) idx 183 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c mode_mask = DPU_REG_READ(&ctx->hw, SSPP_MULTIRECT_OPMODE + idx); idx 191 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_MULTIRECT_OPMODE + idx, mode_mask); idx 197 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 201 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c _sspp_subblk_offset(ctx, DPU_SSPP_SCALER_QSEED2, &idx) || idx 205 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c opmode = DPU_REG_READ(&ctx->hw, SSPP_VIG_OP_MODE + idx); idx 212 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_VIG_OP_MODE + idx, opmode); idx 218 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 221 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_CSC_10BIT, &idx)) idx 224 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c opmode = DPU_REG_READ(&ctx->hw, SSPP_VIG_CSC_10_OP_MODE + idx); idx 230 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_VIG_CSC_10_OP_MODE + idx, opmode); idx 245 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 247 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx) || !fmt) idx 261 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c opmode = DPU_REG_READ(c, op_mode_off + idx); idx 331 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, format_off + idx, src_format); idx 332 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, unpack_pat_off + idx, unpack); idx 333 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, op_mode_off + idx, opmode); idx 336 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_UBWC_ERROR_STATUS + idx, BIT(31)); idx 347 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 349 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx) || !pe_ext) idx 379 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C0_LR + idx, lr_pe[0]); idx 380 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C0_TB + idx, tb_pe[0]); idx 381 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C0_REQ_PIXELS + idx, idx 385 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C1C2_LR + idx, lr_pe[1]); idx 386 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C1C2_TB + idx, tb_pe[1]); idx 387 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C1C2_REQ_PIXELS + idx, idx 391 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C3_LR + idx, lr_pe[3]); idx 392 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C3_TB + idx, lr_pe[3]); idx 393 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SW_PIX_EXT_C3_REQ_PIXELS + idx, idx 402 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 406 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SCALER_QSEED3, &idx) || !sspp idx 410 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c dpu_hw_setup_scaler3(&ctx->hw, scaler3_cfg, idx, idx 417 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 419 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (!ctx || _sspp_subblk_offset(ctx, DPU_SSPP_SCALER_QSEED3, &idx)) idx 422 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c return dpu_hw_get_scaler3_ver(&ctx->hw, idx); idx 435 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 437 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx) || !cfg) idx 469 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c ystride0 = DPU_REG_READ(c, SSPP_SRC_YSTRIDE0 + idx); idx 470 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c ystride1 = DPU_REG_READ(c, SSPP_SRC_YSTRIDE1 + idx); idx 488 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, src_size_off + idx, src_size); idx 489 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, src_xy_off + idx, src_xy); idx 490 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, out_size_off + idx, dst_size); idx 491 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, out_xy_off + idx, dst_xy); idx 493 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SRC_YSTRIDE0 + idx, ystride0); idx 494 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(c, SSPP_SRC_YSTRIDE1 + idx, ystride1); idx 502 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 504 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) idx 509 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SRC0_ADDR + idx + i * 0x4, idx 512 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SRC0_ADDR + idx, idx 514 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SRC2_ADDR + idx, idx 517 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SRC1_ADDR + idx, idx 519 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SRC3_ADDR + idx, idx 527 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 530 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_CSC, &idx) || !data) idx 534 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c idx += CSC_10BIT_OFFSET; idx 538 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c dpu_hw_csc_setup(&ctx->hw, idx, data, csc10); idx 544 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 546 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) idx 550 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SRC_CONSTANT_COLOR + idx, color); idx 552 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SRC_CONSTANT_COLOR_REC1 + idx, idx 559 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 561 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) idx 564 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_DANGER_LUT + idx, cfg->danger_lut); idx 565 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_SAFE_LUT + idx, cfg->safe_lut); idx 571 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 573 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) idx 577 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_CREQ_LUT_0 + idx, cfg->creq_lut); idx 578 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_CREQ_LUT_1 + idx, idx 581 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_CREQ_LUT + idx, cfg->creq_lut); idx 588 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 591 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) idx 607 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c DPU_REG_WRITE(&ctx->hw, SSPP_QOS_CTRL + idx, qos_ctrl); idx 613 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c u32 idx; idx 619 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c if (_sspp_subblk_offset(ctx, DPU_SSPP_SRC, &idx)) idx 694 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c struct dpu_hw_pipe *dpu_hw_sspp_init(enum dpu_sspp idx, idx 708 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c cfg = _sspp_offset(idx, addr, catalog, &hw_pipe->hw); idx 717 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c hw_pipe->idx = idx; idx 721 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.c dpu_hw_blk_init(&hw_pipe->base, DPU_HW_BLK_SSPP, idx, &dpu_hw_ops); idx 379 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.h enum dpu_sspp idx; idx 394 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_sspp.h struct dpu_hw_pipe *dpu_hw_sspp_init(enum dpu_sspp idx, idx 318 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c struct dpu_hw_mdp *dpu_hw_mdptop_init(enum dpu_mdp idx, idx 332 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c cfg = _top_offset(idx, m, addr, &mdp->hw); idx 341 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c mdp->idx = idx; idx 345 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.c dpu_hw_blk_init(&mdp->base, DPU_HW_BLK_TOP, idx, &dpu_hw_ops); idx 148 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.h enum dpu_mdp idx; idx 161 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_top.h struct dpu_hw_mdp *dpu_hw_mdptop_init(enum dpu_mdp idx, idx 234 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c struct dpu_hw_vbif *dpu_hw_vbif_init(enum dpu_vbif idx, idx 245 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c cfg = _top_offset(idx, m, addr, &c->hw); idx 254 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c c->idx = idx; idx 101 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h enum dpu_vbif idx; idx 114 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h struct dpu_hw_vbif *dpu_hw_vbif_init(enum dpu_vbif idx, idx 378 drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c ot_params.num = pdpu->pipe_hw->idx - SSPP_NONE; idx 404 drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c qos_params.num = pdpu->pipe_hw->idx - SSPP_VIG0; idx 432 drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c trace_dpu_plane_set_scanout(pdpu->pipe_hw->idx, idx 419 drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h TP_PROTO(uint32_t drm_id, unsigned int idx, idx 421 drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h TP_ARGS(drm_id, idx, frame_busy_mask), idx 424 drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h __field( unsigned int, idx ) idx 429 drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h __entry->idx = idx; idx 433 drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h __entry->idx, __entry->frame_busy_mask) idx 46 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c vbif->idx - VBIF_0, xin_id); idx 50 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c vbif->idx - VBIF_0, xin_id); idx 91 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c vbif->idx - VBIF_0, params->xin_id, idx 137 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c vbif->idx - VBIF_0, params->xin_id, ot_lim); idx 165 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c dpu_kms->hw_vbif[i]->idx == params->vbif_idx) idx 200 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c trace_dpu_vbif_wait_xin_halt_fail(vbif->idx, params->xin_id); idx 225 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c dpu_kms->hw_vbif[i]->idx == params->vbif_idx) { idx 274 drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c vbif->idx - VBIF_0, pnd, src); idx 310 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h static inline uint32_t __offset_OVLP(uint32_t idx) idx 312 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h switch (idx) { idx 316 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h default: return INVALID_IDX(idx); idx 343 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h static inline uint32_t __offset_STAGE(uint32_t idx) idx 345 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h switch (idx) { idx 350 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h default: return INVALID_IDX(idx); idx 387 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h static inline uint32_t __offset_STAGE_CO3(uint32_t idx) idx 389 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h switch (idx) { idx 394 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h default: return INVALID_IDX(idx); idx 447 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h static inline uint32_t __offset_DMA(enum mdp4_dma idx) idx 449 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h switch (idx) { idx 453 drivers/gpu/drm/msm/disp/mdp4/mdp4.xml.h default: return INVALID_IDX(idx); idx 167 drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c int idx = idxs[pipe_id]; idx 169 drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c pipe_id, stages[idx]); idx 191 drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c int idx = idxs[pipe_id]; idx 192 drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c if (idx > 0) { idx 195 drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c alpha[idx-1] = format->alpha_enable; idx 59 drivers/gpu/drm/msm/disp/mdp4/mdp4_dtv_encoder.c static void bs_set(struct mdp4_dtv_encoder *mdp4_dtv_encoder, int idx) idx 62 drivers/gpu/drm/msm/disp/mdp4/mdp4_dtv_encoder.c DBG("set bus scaling: %d", idx); idx 63 drivers/gpu/drm/msm/disp/mdp4/mdp4_dtv_encoder.c msm_bus_scale_client_update_request(mdp4_dtv_encoder->bsc, idx); idx 69 drivers/gpu/drm/msm/disp/mdp4/mdp4_dtv_encoder.c static void bs_set(struct mdp4_dtv_encoder *mdp4_dtv_encoder, int idx) {} idx 60 drivers/gpu/drm/msm/disp/mdp4/mdp4_lcdc_encoder.c static void bs_set(struct mdp4_lcdc_encoder *mdp4_lcdc_encoder, int idx) idx 63 drivers/gpu/drm/msm/disp/mdp4/mdp4_lcdc_encoder.c DBG("set bus scaling: %d", idx); idx 64 drivers/gpu/drm/msm/disp/mdp4/mdp4_lcdc_encoder.c msm_bus_scale_client_update_request(mdp4_lcdc_encoder->bsc, idx); idx 70 drivers/gpu/drm/msm/disp/mdp4/mdp4_lcdc_encoder.c static void bs_set(struct mdp4_lcdc_encoder *mdp4_lcdc_encoder, int idx) {} idx 309 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_IGC(enum mdp5_igc_type idx) idx 311 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 316 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 349 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_CTL(uint32_t idx) idx 351 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 357 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 362 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_LAYER(uint32_t idx) idx 364 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 371 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 497 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_LAYER_EXT(uint32_t idx) idx 499 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 506 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 535 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_PIPE(enum mdp5_pipe idx) idx 537 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 538 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h case SSPP_NONE: return (INVALID_IDX(idx)); idx 551 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 934 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_SW_PIX_EXT(enum mdp_component_type idx) idx 936 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 940 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 1063 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_LM(uint32_t idx) idx 1065 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 1072 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 1105 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_BLEND(uint32_t idx) idx 1107 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 1115 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 1258 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_DSPP(uint32_t idx) idx 1260 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 1265 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 1303 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_PP(uint32_t idx) idx 1305 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 1310 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 1391 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_WB(uint32_t idx) idx 1393 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 1401 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 1743 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_INTF(uint32_t idx) idx 1745 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 1751 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 1889 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h static inline uint32_t __offset_AD(uint32_t idx) idx 1891 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h switch (idx) { idx 1894 drivers/gpu/drm/msm/disp/mdp5/mdp5.xml.h default: return INVALID_IDX(idx); idx 22 drivers/gpu/drm/msm/disp/mdp5/mdp5_cmd_encoder.c static void bs_set(struct mdp5_encoder *mdp5_cmd_enc, int idx) idx 25 drivers/gpu/drm/msm/disp/mdp5/mdp5_cmd_encoder.c DBG("set bus scaling: %d", idx); idx 30 drivers/gpu/drm/msm/disp/mdp5/mdp5_cmd_encoder.c idx = 1; idx 31 drivers/gpu/drm/msm/disp/mdp5/mdp5_cmd_encoder.c msm_bus_scale_client_update_request(mdp5_cmd_enc->bsc, idx); idx 35 drivers/gpu/drm/msm/disp/mdp5/mdp5_cmd_encoder.c static void bs_set(struct mdp5_encoder *mdp5_cmd_enc, int idx) {} idx 63 drivers/gpu/drm/msm/disp/mdp5/mdp5_encoder.c static void bs_set(struct mdp5_encoder *mdp5_encoder, int idx) idx 66 drivers/gpu/drm/msm/disp/mdp5/mdp5_encoder.c DBG("set bus scaling: %d", idx); idx 71 drivers/gpu/drm/msm/disp/mdp5/mdp5_encoder.c idx = 1; idx 72 drivers/gpu/drm/msm/disp/mdp5/mdp5_encoder.c msm_bus_scale_client_update_request(mdp5_encoder->bsc, idx); idx 78 drivers/gpu/drm/msm/disp/mdp5/mdp5_encoder.c static void bs_set(struct mdp5_encoder *mdp5_encoder, int idx) {} idx 818 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.c hwpipe->idx = mdp5_kms->num_hwpipes; idx 891 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.c mixer->idx = mdp5_kms->num_hwmixers; idx 923 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.c intf->idx = mdp5_kms->num_intfs; idx 153 drivers/gpu/drm/msm/disp/mdp5/mdp5_kms.h int idx; idx 32 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c return mixer->idx; idx 63 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c if (new_state->hwmixer_to_crtc[cur->idx] && idx 64 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c new_state->hwmixer_to_crtc[cur->idx] != crtc) idx 109 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c new_state->hwmixer_to_crtc[(*mixer)->idx] = crtc; idx 113 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c new_state->hwmixer_to_crtc[(*r_mixer)->idx] = crtc; idx 127 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c if (WARN_ON(!new_state->hwmixer_to_crtc[mixer->idx])) idx 131 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c new_state->hwmixer_to_crtc[mixer->idx]->name); idx 133 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c new_state->hwmixer_to_crtc[mixer->idx] = NULL; idx 11 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.h int idx; idx 40 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (new_state->hwpipe_to_plane[cur->idx] || idx 41 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c old_state->hwpipe_to_plane[cur->idx]) idx 111 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c new_state->hwpipe_to_plane[(*hwpipe)->idx] = plane; idx 116 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c new_state->hwpipe_to_plane[(*r_hwpipe)->idx] = plane; idx 132 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (WARN_ON(!new_state->hwpipe_to_plane[hwpipe->idx])) idx 136 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c new_state->hwpipe_to_plane[hwpipe->idx]->name); idx 143 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c new_state->hwpipe_to_plane[hwpipe->idx] = NULL; idx 15 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.h int idx; idx 228 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c int idx = blk / 3; idx 231 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c val = smp->alloc_w[idx]; idx 248 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c smp->alloc_w[idx] = val; idx 249 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c smp->alloc_r[idx] = val; idx 352 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c struct drm_plane *plane = hwpstate->hwpipe_to_plane[hwpipe->idx]; idx 56 drivers/gpu/drm/msm/dsi/mmss_cc.xml.h static inline uint32_t __offset_CLK(enum mmss_cc_clk idx) idx 58 drivers/gpu/drm/msm/dsi/mmss_cc.xml.h switch (idx) { idx 61 drivers/gpu/drm/msm/dsi/mmss_cc.xml.h default: return INVALID_IDX(idx); idx 161 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c u32 resp, phy_addr, idx = 0; idx 173 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c scm_buf[i].addr = phy_addr + preg[idx]; idx 174 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c scm_buf[i].val = pdata[idx]; idx 175 drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c idx++; idx 438 drivers/gpu/drm/msm/msm_drv.h #define INVALID_IDX(idx) ({BUG(); 0;}) idx 151 drivers/gpu/drm/msm/msm_gem.h uint32_t idx; /* cmdstream buffer idx in bos[] */ idx 282 drivers/gpu/drm/msm/msm_gem_submit.c static int submit_bo(struct msm_gem_submit *submit, uint32_t idx, idx 285 drivers/gpu/drm/msm/msm_gem_submit.c if (idx >= submit->nr_bos) { idx 287 drivers/gpu/drm/msm/msm_gem_submit.c idx, submit->nr_bos); idx 292 drivers/gpu/drm/msm/msm_gem_submit.c *obj = submit->bos[idx].obj; idx 294 drivers/gpu/drm/msm/msm_gem_submit.c *iova = submit->bos[idx].iova; idx 296 drivers/gpu/drm/msm/msm_gem_submit.c *valid = !!(submit->bos[idx].flags & BO_VALID); idx 552 drivers/gpu/drm/msm/msm_gem_submit.c submit->cmd[i].idx = submit_cmd.submit_idx; idx 364 drivers/gpu/drm/msm/msm_gpu.c int idx = submit->cmd[i].idx; idx 366 drivers/gpu/drm/msm/msm_gpu.c msm_gpu_crashstate_get_bo(state, submit->bos[idx].obj, idx 367 drivers/gpu/drm/msm/msm_gpu.c submit->bos[idx].iova, submit->bos[idx].flags); idx 39 drivers/gpu/drm/msm/msm_gpummu.c unsigned idx = (iova - GPUMMU_VA_START) / GPUMMU_PAGE_SIZE; idx 51 drivers/gpu/drm/msm/msm_gpummu.c for (j = 0; j < sg->length / GPUMMU_PAGE_SIZE; j++, idx++) { idx 52 drivers/gpu/drm/msm/msm_gpummu.c gpummu->table[idx] = addr | prot_bits; idx 67 drivers/gpu/drm/msm/msm_gpummu.c unsigned idx = (iova - GPUMMU_VA_START) / GPUMMU_PAGE_SIZE; idx 70 drivers/gpu/drm/msm/msm_gpummu.c for (i = 0; i < len / GPUMMU_PAGE_SIZE; i++, idx++) idx 71 drivers/gpu/drm/msm/msm_gpummu.c gpummu->table[idx] = 0; idx 300 drivers/gpu/drm/msm/msm_rd.c struct msm_gem_submit *submit, int idx, idx 303 drivers/gpu/drm/msm/msm_rd.c struct msm_gem_object *obj = submit->bos[idx].obj; idx 308 drivers/gpu/drm/msm/msm_rd.c offset = iova - submit->bos[idx].iova; idx 310 drivers/gpu/drm/msm/msm_rd.c iova = submit->bos[idx].iova; idx 322 drivers/gpu/drm/msm/msm_rd.c if (!(submit->bos[idx].flags & MSM_SUBMIT_BO_READ)) idx 337 drivers/gpu/drm/msm/msm_rd.c should_dump(struct msm_gem_submit *submit, int idx) idx 339 drivers/gpu/drm/msm/msm_rd.c return rd_full || (submit->bos[idx].flags & MSM_SUBMIT_BO_DUMP); idx 393 drivers/gpu/drm/msm/msm_rd.c snapshot_buf(rd, submit, submit->cmd[i].idx, idx 205 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*bar )(struct nvkm_device *, int idx, struct nvkm_bar **); idx 206 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*bios )(struct nvkm_device *, int idx, struct nvkm_bios **); idx 207 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*bus )(struct nvkm_device *, int idx, struct nvkm_bus **); idx 208 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*clk )(struct nvkm_device *, int idx, struct nvkm_clk **); idx 209 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*devinit )(struct nvkm_device *, int idx, struct nvkm_devinit **); idx 210 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*fault )(struct nvkm_device *, int idx, struct nvkm_fault **); idx 211 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*fb )(struct nvkm_device *, int idx, struct nvkm_fb **); idx 212 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*fuse )(struct nvkm_device *, int idx, struct nvkm_fuse **); idx 213 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*gpio )(struct nvkm_device *, int idx, struct nvkm_gpio **); idx 214 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*gsp )(struct nvkm_device *, int idx, struct nvkm_gsp **); idx 215 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*i2c )(struct nvkm_device *, int idx, struct nvkm_i2c **); idx 216 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*ibus )(struct nvkm_device *, int idx, struct nvkm_subdev **); idx 217 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*iccsense)(struct nvkm_device *, int idx, struct nvkm_iccsense **); idx 218 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*imem )(struct nvkm_device *, int idx, struct nvkm_instmem **); idx 219 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*ltc )(struct nvkm_device *, int idx, struct nvkm_ltc **); idx 220 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*mc )(struct nvkm_device *, int idx, struct nvkm_mc **); idx 221 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*mmu )(struct nvkm_device *, int idx, struct nvkm_mmu **); idx 222 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*mxm )(struct nvkm_device *, int idx, struct nvkm_subdev **); idx 223 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*pci )(struct nvkm_device *, int idx, struct nvkm_pci **); idx 224 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*pmu )(struct nvkm_device *, int idx, struct nvkm_pmu **); idx 225 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*secboot )(struct nvkm_device *, int idx, struct nvkm_secboot **); idx 226 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*therm )(struct nvkm_device *, int idx, struct nvkm_therm **); idx 227 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*timer )(struct nvkm_device *, int idx, struct nvkm_timer **); idx 228 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*top )(struct nvkm_device *, int idx, struct nvkm_top **); idx 229 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*volt )(struct nvkm_device *, int idx, struct nvkm_volt **); idx 231 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*bsp )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 232 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*ce[9] )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 233 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*cipher )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 234 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*disp )(struct nvkm_device *, int idx, struct nvkm_disp **); idx 235 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*dma )(struct nvkm_device *, int idx, struct nvkm_dma **); idx 236 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*fifo )(struct nvkm_device *, int idx, struct nvkm_fifo **); idx 237 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*gr )(struct nvkm_device *, int idx, struct nvkm_gr **); idx 238 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*ifb )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 239 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*me )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 240 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*mpeg )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 241 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*msenc )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 242 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*mspdec )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 243 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*msppp )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 244 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*msvld )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 245 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*nvenc[3])(struct nvkm_device *, int idx, struct nvkm_engine **); idx 246 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*nvdec[3])(struct nvkm_device *, int idx, struct nvkm_nvdec **); idx 247 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*pm )(struct nvkm_device *, int idx, struct nvkm_pm **); idx 248 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*sec )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 249 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*sec2 )(struct nvkm_device *, int idx, struct nvkm_sec2 **); idx 250 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*sw )(struct nvkm_device *, int idx, struct nvkm_sw **); idx 251 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*vic )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 252 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*vp )(struct nvkm_device *, int idx, struct nvkm_engine **); idx 28 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0203.h u32 nvbios_M0203Ee(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr); idx 29 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0203.h u32 nvbios_M0203Ep(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr, idx 18 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0205.h u32 nvbios_M0205Ee(struct nvkm_bios *, int idx, idx 20 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0205.h u32 nvbios_M0205Ep(struct nvkm_bios *, int idx, idx 27 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0205.h u32 nvbios_M0205Se(struct nvkm_bios *, int ent, int idx, u8 *ver, u8 *hdr); idx 28 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0205.h u32 nvbios_M0205Sp(struct nvkm_bios *, int ent, int idx, u8 *ver, u8 *hdr, idx 16 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0209.h u32 nvbios_M0209Ee(struct nvkm_bios *, int idx, idx 18 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0209.h u32 nvbios_M0209Ep(struct nvkm_bios *, int idx, idx 25 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0209.h u32 nvbios_M0209Se(struct nvkm_bios *, int ent, int idx, u8 *ver, u8 *hdr); idx 26 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0209.h u32 nvbios_M0209Sp(struct nvkm_bios *, int ent, int idx, u8 *ver, u8 *hdr, idx 11 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/P0260.h u32 nvbios_P0260Ee(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr); idx 12 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/P0260.h u32 nvbios_P0260Ep(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr, idx 19 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/P0260.h u32 nvbios_P0260Xe(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr); idx 20 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/P0260.h u32 nvbios_P0260Xp(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr, idx 12 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/boost.h u32 nvbios_boostEe(struct nvkm_bios *, int idx, u8 *, u8 *, u8 *, u8 *); idx 13 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/boost.h u32 nvbios_boostEp(struct nvkm_bios *, int idx, u8 *, u8 *, u8 *, u8 *, idx 44 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/conn.h u32 nvbios_connEe(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *hdr); idx 45 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/conn.h u32 nvbios_connEp(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *hdr, idx 12 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/cstep.h u32 nvbios_cstepEe(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr); idx 13 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/cstep.h u32 nvbios_cstepEp(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr, idx 24 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/cstep.h u32 nvbios_cstepXe(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr); idx 25 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/cstep.h u32 nvbios_cstepXp(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr, idx 60 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/dcb.h u16 dcb_outp(struct nvkm_bios *, u8 idx, u8 *ver, u8 *len); idx 61 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/dcb.h u16 dcb_outp_parse(struct nvkm_bios *, u8 idx, u8 *, u8 *, idx 11 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/disp.h u16 nvbios_disp_entry(struct nvkm_bios *, u8 idx, u8 *ver, u8 *hdr, u8 *sub); idx 12 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/disp.h u16 nvbios_disp_parse(struct nvkm_bios *, u8 idx, u8 *ver, u8 *hdr, u8 *sub, idx 21 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/disp.h u16 nvbios_outp_entry(struct nvkm_bios *, u8 idx, idx 23 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/disp.h u16 nvbios_outp_parse(struct nvkm_bios *, u8 idx, idx 34 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/disp.h u16 nvbios_ocfg_entry(struct nvkm_bios *, u16 outp, u8 idx, idx 36 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/disp.h u16 nvbios_ocfg_parse(struct nvkm_bios *, u16 outp, u8 idx, idx 16 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/dp.h u16 nvbios_dpout_parse(struct nvkm_bios *, u8 idx, idx 31 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/dp.h nvbios_dpcfg_parse(struct nvkm_bios *, u16 outp, u8 idx, idx 47 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/gpio.h u16 dcb_gpio_entry(struct nvkm_bios *, int idx, int ent, u8 *ver, u8 *len); idx 48 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/gpio.h u16 dcb_gpio_parse(struct nvkm_bios *, int idx, int ent, u8 *ver, u8 *len, idx 50 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/gpio.h u16 dcb_gpio_match(struct nvkm_bios *, int idx, u8 func, u8 line, idx 21 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/perf.h u32 nvbios_perf_entry(struct nvkm_bios *, int idx, idx 23 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/perf.h u32 nvbios_perfEp(struct nvkm_bios *, int idx, idx 34 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/perf.h u32 nvbios_perfSe(struct nvkm_bios *, u32 data, int idx, idx 36 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/perf.h u32 nvbios_perfSp(struct nvkm_bios *, u32 data, int idx, idx 14 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/pmu.h u32 nvbios_pmuEe(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr); idx 15 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/pmu.h u32 nvbios_pmuEp(struct nvkm_bios *, int idx, u8 *ver, u8 *hdr, idx 25 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/power_budget.h u8 idx, struct nvbios_power_budget_entry *); idx 9 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/rammap.h u32 nvbios_rammapEe(struct nvkm_bios *, int idx, idx 13 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/rammap.h u32 nvbios_rammapEp(struct nvkm_bios *, int idx, idx 19 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/rammap.h u8 ever, u8 ehdr, u8 ecnt, u8 elen, int idx, idx 21 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/rammap.h u32 nvbios_rammapSp_from_perf(struct nvkm_bios *bios, u32 data, u8 size, int idx, idx 24 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/rammap.h u8 ever, u8 ehdr, u8 ecnt, u8 elen, int idx, idx 8 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/timing.h u32 nvbios_timingEe(struct nvkm_bios *, int idx, idx 10 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/timing.h u32 nvbios_timingEp(struct nvkm_bios *, int idx, idx 22 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/vmap.h u32 nvbios_vmap_entry(struct nvkm_bios *, int idx, u8 *ver, u8 *len); idx 23 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/vmap.h u32 nvbios_vmap_entry_parse(struct nvkm_bios *, int idx, u8 *ver, u8 *len, idx 35 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/volt.h u32 nvbios_volt_entry(struct nvkm_bios *, int idx, u8 *ver, u8 *len); idx 36 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/volt.h u32 nvbios_volt_entry_parse(struct nvkm_bios *, int idx, u8 *ver, u8 *len, idx 24 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/vpstate.h u8 idx, struct nvbios_vpstate_entry *); idx 15 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/xpio.h u16 dcb_xpio_table(struct nvkm_bios *, u8 idx, idx 17 drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/xpio.h u16 dcb_xpio_parse(struct nvkm_bios *, u8 idx, idx 30 drivers/gpu/drm/nouveau/include/nvkm/subdev/gpio.h int nvkm_gpio_find(struct nvkm_gpio *, int idx, u8 tag, u8 line, idx 32 drivers/gpu/drm/nouveau/include/nvkm/subdev/gpio.h int nvkm_gpio_set(struct nvkm_gpio *, int idx, u8 tag, u8 line, int state); idx 33 drivers/gpu/drm/nouveau/include/nvkm/subdev/gpio.h int nvkm_gpio_get(struct nvkm_gpio *, int idx, u8 tag, u8 line); idx 1302 drivers/gpu/drm/nouveau/nouveau_bios.c olddcb_outp(struct drm_device *dev, u8 idx) idx 1306 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx < dcb[2]) idx 1307 drivers/gpu/drm/nouveau/nouveau_bios.c return dcb + dcb[1] + (idx * dcb[3]); idx 1311 drivers/gpu/drm/nouveau/nouveau_bios.c u8 *ent = dcb + 8 + (idx * 8); idx 1317 drivers/gpu/drm/nouveau/nouveau_bios.c u8 *ent = dcb + 4 + (idx * 10); idx 1327 drivers/gpu/drm/nouveau/nouveau_bios.c int (*exec)(struct drm_device *, void *, int idx, u8 *outp)) idx 1329 drivers/gpu/drm/nouveau/nouveau_bios.c int ret, idx = -1; idx 1331 drivers/gpu/drm/nouveau/nouveau_bios.c while ((outp = olddcb_outp(dev, ++idx))) { idx 1342 drivers/gpu/drm/nouveau/nouveau_bios.c ret = exec(dev, data, idx, outp); idx 1363 drivers/gpu/drm/nouveau/nouveau_bios.c olddcb_conn(struct drm_device *dev, u8 idx) idx 1366 drivers/gpu/drm/nouveau/nouveau_bios.c if (conntab && idx < conntab[2]) idx 1367 drivers/gpu/drm/nouveau/nouveau_bios.c return conntab + conntab[1] + (idx * conntab[3]); idx 1646 drivers/gpu/drm/nouveau/nouveau_bios.c apply_dcb_encoder_quirks(struct drm_device *dev, int idx, u32 *conn, u32 *conf) idx 1685 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 0) { idx 1689 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 1) { idx 1693 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 2) { idx 1697 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 3) { idx 1716 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 0) { idx 1720 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 1) { idx 1724 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 2) { idx 1728 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 3) { idx 1739 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 0 && *conn == 0x02000300) idx 1742 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 1 && *conn == 0x04011310) idx 1745 drivers/gpu/drm/nouveau/nouveau_bios.c if (idx == 2 && *conn == 0x02011312) idx 1784 drivers/gpu/drm/nouveau/nouveau_bios.c parse_dcb_entry(struct drm_device *dev, void *data, int idx, u8 *outp) idx 1792 drivers/gpu/drm/nouveau/nouveau_bios.c if (apply_dcb_encoder_quirks(dev, idx, &conn, &conf)) { idx 1795 drivers/gpu/drm/nouveau/nouveau_bios.c NV_INFO(drm, "DCB outp %02d: %08x %08x\n", idx, conn, conf); idx 1821 drivers/gpu/drm/nouveau/nouveau_bios.c int i, idx = 0; idx 1844 drivers/gpu/drm/nouveau/nouveau_bios.c dcbt->entry[i].connector = idx++; idx 1847 drivers/gpu/drm/nouveau/nouveau_bios.c map[i2c] = ++idx; idx 1868 drivers/gpu/drm/nouveau/nouveau_bios.c int idx; idx 1894 drivers/gpu/drm/nouveau/nouveau_bios.c idx = -1; idx 1895 drivers/gpu/drm/nouveau/nouveau_bios.c while ((conn = olddcb_conn(dev, ++idx))) { idx 1899 drivers/gpu/drm/nouveau/nouveau_bios.c idx, ROM16(conn[0])); idx 1902 drivers/gpu/drm/nouveau/nouveau_bios.c idx, ROM32(conn[0])); idx 157 drivers/gpu/drm/nouveau/nouveau_bios.h void *olddcb_outp(struct drm_device *, u8 idx); idx 159 drivers/gpu/drm/nouveau/nouveau_bios.h int (*)(struct drm_device *, void *, int idx, u8 *outp)); idx 161 drivers/gpu/drm/nouveau/nouveau_bios.h u8 *olddcb_conn(struct drm_device *, u8 idx); idx 487 drivers/gpu/drm/nouveau/nouveau_connector.c int idx = name ? name[strlen(name) - 1] - 'A' : 0; idx 489 drivers/gpu/drm/nouveau/nouveau_connector.c if (nv_encoder->dcb->i2c_index == idx && edid) { idx 90 drivers/gpu/drm/nouveau/nouveau_dmem.c unsigned long idx = page_to_pfn(page) - chunk->pfn_first; idx 92 drivers/gpu/drm/nouveau/nouveau_dmem.c return (idx << PAGE_SHIFT) + chunk->bo->bo.offset; idx 98 drivers/gpu/drm/nouveau/nouveau_dmem.c unsigned long idx = page_to_pfn(page) - chunk->pfn_first; idx 108 drivers/gpu/drm/nouveau/nouveau_dmem.c clear_bit(idx, chunk->bitmap); idx 72 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c nvbios_M0203Ee(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr) idx 76 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c if (data && idx < cnt) { idx 77 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c data = data + *hdr + idx * len; idx 85 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c nvbios_M0203Ep(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr, idx 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c u32 data = nvbios_M0203Ee(bios, idx, ver, hdr); idx 108 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c u8 cnt, len, idx = 0xff; idx 116 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c while ((data = nvbios_M0203Ep(bios, ++idx, ver, hdr, info))) { idx 75 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c nvbios_M0205Ee(struct nvkm_bios *bios, int idx, idx 80 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c if (data && idx < *cnt) { idx 81 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c data = data + *hdr + idx * (*len + (snr * ssz)); idx 91 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c nvbios_M0205Ep(struct nvkm_bios *bios, int idx, idx 95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c u32 data = nvbios_M0205Ee(bios, idx, ver, hdr, cnt, len); idx 108 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c nvbios_M0205Se(struct nvkm_bios *bios, int ent, int idx, u8 *ver, u8 *hdr) idx 113 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c if (data && idx < cnt) { idx 114 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c data = data + *hdr + idx * len; idx 122 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c nvbios_M0205Sp(struct nvkm_bios *bios, int ent, int idx, u8 *ver, u8 *hdr, idx 125 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0205.c u32 data = nvbios_M0205Se(bios, ent, idx, ver, hdr); idx 58 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c nvbios_M0209Ee(struct nvkm_bios *bios, int idx, idx 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c if (data && idx < *cnt) { idx 64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c data = data + *hdr + idx * (*len + (snr * ssz)); idx 74 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c nvbios_M0209Ep(struct nvkm_bios *bios, int idx, idx 77 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c u32 data = nvbios_M0209Ee(bios, idx, ver, hdr, cnt, len); idx 95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c nvbios_M0209Se(struct nvkm_bios *bios, int ent, int idx, u8 *ver, u8 *hdr) idx 100 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c if (data && idx < cnt) { idx 101 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c data = data + *hdr + idx * len; idx 109 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c nvbios_M0209Sp(struct nvkm_bios *bios, int ent, int idx, u8 *ver, u8 *hdr, idx 116 drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0209.c u32 i, data = nvbios_M0209Se(bios, ent, idx, ver, hdr); idx 58 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c nvbios_P0260Ee(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len) idx 62 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c if (data && idx < cnt) idx 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c return data + hdr + (idx * *len); idx 68 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c nvbios_P0260Ep(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len, idx 71 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c u32 data = nvbios_P0260Ee(bios, idx, ver, len); idx 84 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c nvbios_P0260Xe(struct nvkm_bios *bios, int idx, u8 *ver, u8 *xsz) idx 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c if (data && idx < xnr) idx 89 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c return data + hdr + (cnt * len) + (idx * *xsz); idx 94 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c nvbios_P0260Xp(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr, idx 97 drivers/gpu/drm/nouveau/nvkm/subdev/bios/P0260.c u32 data = nvbios_P0260Xe(bios, idx, ver, hdr); idx 148 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c int ret, idx = 0; idx 161 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c if (nvbios_image(bios, idx++, &image)) { idx 163 drivers/gpu/drm/nouveau/nvkm/subdev/bios/base.c while (nvbios_image(bios, idx++, &image)) { idx 59 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c nvbios_boostEe(struct nvkm_bios *bios, int idx, idx 64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c if (data && idx < *cnt) { idx 65 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c data = data + *hdr + (idx * (*len + (snr * ssz))); idx 75 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c nvbios_boostEp(struct nvkm_bios *bios, int idx, idx 78 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c u32 data = nvbios_boostEe(bios, idx, ver, hdr, cnt, len); idx 92 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c u32 data, idx = 0; idx 93 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c while ((data = nvbios_boostEp(bios, idx++, ver, hdr, cnt, len, info))) { idx 101 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c nvbios_boostSe(struct nvkm_bios *bios, int idx, idx 104 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c if (data && idx < cnt) { idx 105 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c data = data + *hdr + (idx * len); idx 113 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c nvbios_boostSp(struct nvkm_bios *bios, int idx, idx 117 drivers/gpu/drm/nouveau/nvkm/subdev/bios/boost.c data = nvbios_boostSe(bios, idx, data, ver, hdr, cnt, len); idx 62 drivers/gpu/drm/nouveau/nvkm/subdev/bios/conn.c nvbios_connEe(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *len) idx 66 drivers/gpu/drm/nouveau/nvkm/subdev/bios/conn.c if (data && idx < cnt) idx 67 drivers/gpu/drm/nouveau/nvkm/subdev/bios/conn.c return data + hdr + (idx * *len); idx 72 drivers/gpu/drm/nouveau/nvkm/subdev/bios/conn.c nvbios_connEp(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *len, idx 75 drivers/gpu/drm/nouveau/nvkm/subdev/bios/conn.c u32 data = nvbios_connEe(bios, idx, ver, len); idx 59 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c nvbios_cstepEe(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr) idx 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c if (data && idx < cnt) { idx 64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c data = data + *hdr + (idx * len); idx 72 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c nvbios_cstepEp(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr, idx 75 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c u32 data = nvbios_cstepEe(bios, idx, ver, hdr); idx 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c u32 data, idx = 0; idx 89 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c while ((data = nvbios_cstepEp(bios, idx++, ver, hdr, info))) { idx 97 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c nvbios_cstepXe(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr) idx 101 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c if (data && idx < xnr) { idx 102 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c data = data + *hdr + (cnt * len) + (idx * xsz); idx 110 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c nvbios_cstepXp(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr, idx 113 drivers/gpu/drm/nouveau/nvkm/subdev/bios/cstep.c u32 data = nvbios_cstepXe(bios, idx, ver, hdr); idx 99 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c dcb_outp(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *len) idx 103 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c if (dcb && idx < cnt) idx 104 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c return dcb + hdr + (idx * *len); idx 121 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c dcb_outp_parse(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *len, idx 124 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c u16 dcb = dcb_outp(bios, idx, ver, len); idx 200 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c u16 dcb, idx = 0; idx 201 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c while ((dcb = dcb_outp_parse(bios, idx++, ver, len, outp))) { idx 214 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c int ret, idx = -1; idx 218 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c while ((outp = dcb_outp(bios, ++idx, &ver, &len))) { idx 229 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dcb.c ret = exec(bios, data, idx, outp); idx 59 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c nvbios_disp_entry(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *len, u8 *sub) idx 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c if (data && idx < cnt) idx 64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c return data + hdr + (idx * *len); idx 70 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c nvbios_disp_parse(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *len, u8 *sub, idx 73 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c u16 data = nvbios_disp_entry(bios, idx, ver, len, sub); idx 82 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c nvbios_outp_entry(struct nvkm_bios *bios, u8 idx, idx 86 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c u16 data = nvbios_disp_parse(bios, idx, ver, len, hdr, &info); idx 96 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c nvbios_outp_parse(struct nvkm_bios *bios, u8 idx, idx 99 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c u16 data = nvbios_outp_entry(bios, idx, ver, hdr, cnt, len); idx 119 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c u16 data, idx = 0; idx 120 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c while ((data = nvbios_outp_parse(bios, idx++, ver, hdr, cnt, len, info)) || *ver) { idx 130 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c nvbios_ocfg_entry(struct nvkm_bios *bios, u16 outp, u8 idx, idx 133 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c if (idx < *cnt) idx 134 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c return outp + *hdr + (idx * *len); idx 139 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c nvbios_ocfg_parse(struct nvkm_bios *bios, u16 outp, u8 idx, idx 142 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c u16 data = nvbios_ocfg_entry(bios, outp, idx, ver, hdr, cnt, len); idx 156 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c u16 data, idx = 0; idx 157 drivers/gpu/drm/nouveau/nvkm/subdev/bios/disp.c while ((data = nvbios_ocfg_parse(bios, outp, idx++, ver, hdr, cnt, len, info))) { idx 60 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c nvbios_dpout_entry(struct nvkm_bios *bios, u8 idx, idx 64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c if (data && idx < *cnt) { idx 65 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c u16 outp = nvbios_rd16(bios, data + *hdr + idx * *len); idx 91 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c nvbios_dpout_parse(struct nvkm_bios *bios, u8 idx, idx 95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c u16 data = nvbios_dpout_entry(bios, idx, ver, hdr, cnt, len); idx 142 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c u16 data, idx = 0; idx 143 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c while ((data = nvbios_dpout_parse(bios, idx++, ver, hdr, cnt, len, info)) || *ver) { idx 153 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c nvbios_dpcfg_entry(struct nvkm_bios *bios, u16 outp, u8 idx, idx 164 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c if (idx < *cnt) idx 165 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c return outp + *hdr + (idx * *len); idx 171 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c nvbios_dpcfg_parse(struct nvkm_bios *bios, u16 outp, u8 idx, idx 175 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c u16 data = nvbios_dpcfg_entry(bios, outp, idx, ver, hdr, cnt, len); idx 211 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c u8 idx = 0xff; idx 216 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c idx = (pc * 10) + vsoff[vs] + pe; idx 218 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c idx += nvbios_rd08(bios, outp + 0x11) * 40; idx 221 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c idx += nvbios_rd08(bios, outp + 0x11) * 10; idx 223 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c while ((data = nvbios_dpcfg_entry(bios, outp, ++idx, idx 231 drivers/gpu/drm/nouveau/nvkm/subdev/bios/dp.c return nvbios_dpcfg_parse(bios, outp, idx, ver, hdr, cnt, len, info); idx 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c nvbios_extdev_entry(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len) idx 67 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c if (extdev && idx < cnt) idx 68 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c return extdev + idx * *len; idx 82 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c nvbios_extdev_parse(struct nvkm_bios *bios, int idx, idx 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/extdev.c if (!(entry = nvbios_extdev_entry(bios, idx, &ver, &len))) idx 56 drivers/gpu/drm/nouveau/nvkm/subdev/bios/fan.c nvbios_fan_entry(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr, idx 60 drivers/gpu/drm/nouveau/nvkm/subdev/bios/fan.c if (data && idx < *cnt) idx 61 drivers/gpu/drm/nouveau/nvkm/subdev/bios/fan.c return data + *hdr + (idx * (*len)); idx 61 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c dcb_gpio_entry(struct nvkm_bios *bios, int idx, int ent, u8 *ver, u8 *len) idx 66 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c if (!idx--) idx 69 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c gpio = dcb_xpio_table(bios, idx, &xver, &hdr, &cnt, len); idx 78 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c dcb_gpio_parse(struct nvkm_bios *bios, int idx, int ent, u8 *ver, u8 *len, idx 81 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c u16 data = dcb_gpio_entry(bios, idx, ent, ver, len); idx 119 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c dcb_gpio_match(struct nvkm_bios *bios, int idx, u8 func, u8 line, idx 125 drivers/gpu/drm/nouveau/nvkm/subdev/bios/gpio.c while ((data = dcb_gpio_parse(bios, idx, i++, ver, len, gpio))) { idx 61 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c dcb_i2c_entry(struct nvkm_bios *bios, u8 idx, u8 *ver, u8 *len) idx 65 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c if (i2c && idx < cnt) idx 66 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c return i2c + hdr + (idx * *len); idx 71 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c dcb_i2c_parse(struct nvkm_bios *bios, u8 idx, struct dcb_i2c_entry *info) idx 75 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c u16 ent = dcb_i2c_entry(bios, idx, &ver, &len); idx 136 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c if (bios->bmp_offset && idx < 2) { idx 145 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c if (idx == 0) { idx 151 drivers/gpu/drm/nouveau/nvkm/subdev/bios/i2c.c if (idx == 1) { idx 69 drivers/gpu/drm/nouveau/nvkm/subdev/bios/image.c nvbios_image(struct nvkm_bios *bios, int idx, struct nvbios_image *image) idx 80 drivers/gpu/drm/nouveau/nvkm/subdev/bios/image.c } while(idx--); idx 2197 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c u8 idx = 0, ver, len; idx 2207 drivers/gpu/drm/nouveau/nvkm/subdev/bios/init.c while ((data = dcb_gpio_parse(bios, 0, idx++, &ver, &len, &func))) { idx 79 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c nvbios_perf_entry(struct nvkm_bios *bios, int idx, idx 84 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c if (perf && idx < *cnt) { idx 85 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c perf = perf + *hdr + (idx * (*len + (snr * ssz))); idx 95 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c nvbios_perfEp(struct nvkm_bios *bios, int idx, idx 98 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c u32 perf = nvbios_perf_entry(bios, idx, ver, hdr, cnt, len); idx 173 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c nvbios_perfSe(struct nvkm_bios *bios, u32 perfE, int idx, idx 177 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c if (idx < cnt) { idx 178 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c data = perfE + *hdr + (idx * len); idx 185 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c nvbios_perfSp(struct nvkm_bios *bios, u32 perfE, int idx, idx 189 drivers/gpu/drm/nouveau/nvkm/subdev/bios/perf.c u32 data = nvbios_perfSe(bios, perfE, idx, ver, hdr, cnt, len); idx 50 drivers/gpu/drm/nouveau/nvkm/subdev/bios/pmu.c nvbios_pmuEe(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr) idx 54 drivers/gpu/drm/nouveau/nvkm/subdev/bios/pmu.c if (data && idx < cnt) { idx 55 drivers/gpu/drm/nouveau/nvkm/subdev/bios/pmu.c data = data + *hdr + (idx * len); idx 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/pmu.c nvbios_pmuEp(struct nvkm_bios *bios, int idx, u8 *ver, u8 *hdr, idx 66 drivers/gpu/drm/nouveau/nvkm/subdev/bios/pmu.c u32 data = nvbios_pmuEe(bios, idx, ver, hdr); idx 81 drivers/gpu/drm/nouveau/nvkm/subdev/bios/pmu.c u8 ver, hdr, idx = 0; idx 84 drivers/gpu/drm/nouveau/nvkm/subdev/bios/pmu.c while ((data = nvbios_pmuEp(bios, idx++, &ver, &hdr, &pmuE))) { idx 105 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c u8 idx, struct nvbios_power_budget_entry *entry) idx 109 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c if (!bios || !budget || !budget->offset || idx >= budget->ecount idx 113 drivers/gpu/drm/nouveau/nvkm/subdev/bios/power_budget.c entry_offset = budget->offset + budget->hlen + idx * budget->elen; idx 60 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c nvbios_rammapEe(struct nvkm_bios *bios, int idx, idx 65 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c if (rammap && idx < *cnt) { idx 66 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c rammap = rammap + *hdr + (idx * (*len + (snr * ssz))); idx 91 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c nvbios_rammapEp(struct nvkm_bios *bios, int idx, idx 94 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c u32 data = nvbios_rammapEe(bios, idx, ver, hdr, cnt, len), temp; idx 136 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c int idx = 0; idx 138 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c while ((data = nvbios_rammapEp(bios, idx++, ver, hdr, cnt, len, info))) { idx 147 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c u8 ever, u8 ehdr, u8 ecnt, u8 elen, int idx, u8 *ver, u8 *hdr) idx 149 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c if (idx < ecnt) { idx 150 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c data = data + ehdr + (idx * elen); idx 159 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c nvbios_rammapSp_from_perf(struct nvkm_bios *bios, u32 data, u8 size, int idx, idx 162 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c data += (idx * size); idx 191 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c u8 ever, u8 ehdr, u8 ecnt, u8 elen, int idx, idx 194 drivers/gpu/drm/nouveau/nvkm/subdev/bios/rammap.c data = nvbios_rammapSe(bios, data, ever, ehdr, ecnt, elen, idx, ver, hdr); idx 52 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c shadow_image(struct nvkm_bios *bios, int idx, u32 offset, struct shadow *mthd) idx 70 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c if (!nvbios_image(bios, idx, &image)) { idx 71 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c nvkm_debug(subdev, "image %d invalid\n", idx); idx 105 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c score += shadow_image(bios, idx + 1, offset + image.size, mthd); idx 57 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c nvbios_therm_entry(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len) idx 61 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c if (therm && idx < cnt) idx 62 drivers/gpu/drm/nouveau/nvkm/subdev/bios/therm.c return therm + idx * *len; idx 69 drivers/gpu/drm/nouveau/nvkm/subdev/bios/timing.c nvbios_timingEe(struct nvkm_bios *bios, int idx, idx 74 drivers/gpu/drm/nouveau/nvkm/subdev/bios/timing.c if (timing && idx < *cnt) { idx 75 drivers/gpu/drm/nouveau/nvkm/subdev/bios/timing.c timing += *hdr + idx * (*len + (snr * ssz)); idx 85 drivers/gpu/drm/nouveau/nvkm/subdev/bios/timing.c nvbios_timingEp(struct nvkm_bios *bios, int idx, idx 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/timing.c u32 data = nvbios_timingEe(bios, idx, ver, hdr, cnt, len), temp; idx 81 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c nvbios_vmap_entry(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len) idx 85 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c if (vmap && idx < cnt) { idx 86 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c vmap = vmap + hdr + (idx * *len); idx 93 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c nvbios_vmap_entry_parse(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len, idx 96 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vmap.c u32 vmap = nvbios_vmap_entry(bios, idx, ver, len); idx 125 drivers/gpu/drm/nouveau/nvkm/subdev/bios/volt.c nvbios_volt_entry(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len) idx 129 drivers/gpu/drm/nouveau/nvkm/subdev/bios/volt.c if (volt && idx < cnt) { idx 130 drivers/gpu/drm/nouveau/nvkm/subdev/bios/volt.c volt = volt + hdr + (idx * *len); idx 137 drivers/gpu/drm/nouveau/nvkm/subdev/bios/volt.c nvbios_volt_entry_parse(struct nvkm_bios *bios, int idx, u8 *ver, u8 *len, idx 140 drivers/gpu/drm/nouveau/nvkm/subdev/bios/volt.c u32 volt = nvbios_volt_entry(bios, idx, ver, len); idx 77 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c u8 idx, struct nvbios_vpstate_entry *e) idx 81 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c if (!e || !h || idx > h->ecount) idx 84 drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c offset = h->offset + h->hlen + idx * (h->elen + (h->slen * h->scount)); idx 46 drivers/gpu/drm/nouveau/nvkm/subdev/bios/xpio.c dcb_xpio_table(struct nvkm_bios *bios, u8 idx, idx 50 drivers/gpu/drm/nouveau/nvkm/subdev/bios/xpio.c if (data && idx < *cnt) { idx 51 drivers/gpu/drm/nouveau/nvkm/subdev/bios/xpio.c u16 xpio = nvbios_rd16(bios, data + *hdr + (idx * *len)); idx 64 drivers/gpu/drm/nouveau/nvkm/subdev/bios/xpio.c dcb_xpio_parse(struct nvkm_bios *bios, u8 idx, idx 67 drivers/gpu/drm/nouveau/nvkm/subdev/bios/xpio.c u16 data = dcb_xpio_table(bios, idx, ver, hdr, cnt, len); idx 52 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c u8 idx = 0, sver, shdr; idx 60 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c subd = nvbios_boostSp(bios, idx++, data, &sver, &shdr, idx 223 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c nvkm_cstate_new(struct nvkm_clk *clk, int idx, struct nvkm_pstate *pstate) idx 233 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c data = nvbios_cstepXp(bios, idx, &ver, &hdr, &cstepX); idx 246 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c cstate->id = idx; idx 271 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c int ret, idx = 0; idx 274 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c if (idx++ == pstatei) idx 401 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c nvkm_pstate_new(struct nvkm_clk *clk, int idx) idx 412 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c data = nvbios_perfEp(bios, idx, &ver, &hdr, &cnt, &len, &perfE); idx 456 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c int idx = cstepE.index; idx 458 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c nvkm_cstate_new(clk, idx, pstate); idx 459 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c } while(idx--); idx 656 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c int ret, idx, arglen; idx 683 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c idx = 0; idx 685 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c ret = nvkm_pstate_new(clk, idx++); idx 688 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c for (idx = 0; idx < func->nr_pstates; idx++) idx 689 drivers/gpu/drm/nouveau/nvkm/subdev/clk/base.c list_add_tail(&func->pstates[idx].head, &clk->states); idx 133 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c read_clk(struct gf100_clk *clk, int idx) idx 136 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c u32 sctl = nvkm_rd32(device, 0x137250 + (idx * 4)); idx 140 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (ssel & (1 << idx)) { idx 141 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (idx < 7) idx 142 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c sclk = read_pll(clk, 0x137000 + (idx * 0x20)); idx 147 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c sclk = read_div(clk, idx, 0x137160, 0x1371d0); idx 210 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c calc_div(struct gf100_clk *clk, int idx, u32 ref, u32 freq, u32 *ddiv) idx 221 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c calc_src(struct gf100_clk *clk, int idx, u32 freq, u32 *dsrc, u32 *ddiv) idx 243 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c sclk = read_vco(clk, 0x137160 + (idx * 4)); idx 244 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (idx < 7) idx 245 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c sclk = calc_div(clk, idx, sclk, freq, ddiv); idx 250 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c calc_pll(struct gf100_clk *clk, int idx, u32 freq, u32 *coef) idx 257 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c ret = nvbios_pll_parse(bios, 0x137000 + (idx * 0x20), &limits); idx 261 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c limits.refclk = read_div(clk, idx, 0x137120, 0x137140); idx 274 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c calc_clk(struct gf100_clk *clk, struct nvkm_cstate *cstate, int idx, int dom) idx 276 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c struct gf100_clk_info *info = &clk->eng[idx]; idx 286 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c clk0 = calc_src(clk, idx, freq, &src0, &div0); idx 287 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c clk0 = calc_div(clk, idx, clk0, freq, &div1D); idx 290 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (clk0 != freq && (0x00004387 & (1 << idx))) { idx 291 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (idx <= 7) idx 292 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c clk1 = calc_pll(clk, idx, freq, &info->coef); idx 295 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c clk1 = calc_div(clk, idx, clk1, freq, &div1P); idx 317 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c info->ssel = (1 << idx); idx 344 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c gf100_clk_prog_0(struct gf100_clk *clk, int idx) idx 346 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c struct gf100_clk_info *info = &clk->eng[idx]; idx 348 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (idx < 7 && !info->ssel) { idx 349 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c nvkm_mask(device, 0x1371d0 + (idx * 0x04), 0x80003f3f, info->ddiv); idx 350 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c nvkm_wr32(device, 0x137160 + (idx * 0x04), info->dsrc); idx 355 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c gf100_clk_prog_1(struct gf100_clk *clk, int idx) idx 358 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c nvkm_mask(device, 0x137100, (1 << idx), 0x00000000); idx 360 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (!(nvkm_rd32(device, 0x137100) & (1 << idx))) idx 366 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c gf100_clk_prog_2(struct gf100_clk *clk, int idx) idx 368 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c struct gf100_clk_info *info = &clk->eng[idx]; idx 370 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c const u32 addr = 0x137000 + (idx * 0x20); idx 371 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c if (idx <= 7) { idx 393 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c gf100_clk_prog_3(struct gf100_clk *clk, int idx) idx 395 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c struct gf100_clk_info *info = &clk->eng[idx]; idx 398 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c nvkm_mask(device, 0x137100, (1 << idx), info->ssel); idx 400 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c u32 tmp = nvkm_rd32(device, 0x137100) & (1 << idx); idx 408 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c gf100_clk_prog_4(struct gf100_clk *clk, int idx) idx 410 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c struct gf100_clk_info *info = &clk->eng[idx]; idx 412 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gf100.c nvkm_mask(device, 0x137250 + (idx * 0x04), 0x00003f3f, info->mdiv); idx 145 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c read_clk(struct gk104_clk *clk, int idx) idx 148 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c u32 sctl = nvkm_rd32(device, 0x137250 + (idx * 4)); idx 151 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c if (idx < 7) { idx 153 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c if (ssel & (1 << idx)) { idx 154 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c sclk = read_pll(clk, 0x137000 + (idx * 0x20)); idx 157 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c sclk = read_div(clk, idx, 0x137160, 0x1371d0); idx 161 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c u32 ssrc = nvkm_rd32(device, 0x137160 + (idx * 0x04)); idx 163 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c sclk = read_div(clk, idx, 0x137160, 0x1371d0); idx 172 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c sclk = read_div(clk, idx, 0x137160, 0x1371d0); idx 223 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c calc_div(struct gk104_clk *clk, int idx, u32 ref, u32 freq, u32 *ddiv) idx 234 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c calc_src(struct gk104_clk *clk, int idx, u32 freq, u32 *dsrc, u32 *ddiv) idx 256 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c sclk = read_vco(clk, 0x137160 + (idx * 4)); idx 257 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c if (idx < 7) idx 258 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c sclk = calc_div(clk, idx, sclk, freq, ddiv); idx 263 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c calc_pll(struct gk104_clk *clk, int idx, u32 freq, u32 *coef) idx 270 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c ret = nvbios_pll_parse(bios, 0x137000 + (idx * 0x20), &limits); idx 274 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c limits.refclk = read_div(clk, idx, 0x137120, 0x137140); idx 288 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c struct nvkm_cstate *cstate, int idx, int dom) idx 290 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c struct gk104_clk_info *info = &clk->eng[idx]; idx 300 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c clk0 = calc_src(clk, idx, freq, &src0, &div0); idx 301 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c clk0 = calc_div(clk, idx, clk0, freq, &div1D); idx 304 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c if (clk0 != freq && (0x0000ff87 & (1 << idx))) { idx 305 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c if (idx <= 7) idx 306 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c clk1 = calc_pll(clk, idx, freq, &info->coef); idx 309 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c clk1 = calc_div(clk, idx, clk1, freq, &div1P); idx 330 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c info->ssel = (1 << idx); idx 357 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c gk104_clk_prog_0(struct gk104_clk *clk, int idx) idx 359 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c struct gk104_clk_info *info = &clk->eng[idx]; idx 362 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x1371d0 + (idx * 0x04), 0x8000003f, info->ddiv); idx 363 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_wr32(device, 0x137160 + (idx * 0x04), info->dsrc); idx 368 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c gk104_clk_prog_1_0(struct gk104_clk *clk, int idx) idx 371 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x137100, (1 << idx), 0x00000000); idx 373 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c if (!(nvkm_rd32(device, 0x137100) & (1 << idx))) idx 379 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c gk104_clk_prog_1_1(struct gk104_clk *clk, int idx) idx 382 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x137160 + (idx * 0x04), 0x00000100, 0x00000000); idx 386 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c gk104_clk_prog_2(struct gk104_clk *clk, int idx) idx 388 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c struct gk104_clk_info *info = &clk->eng[idx]; idx 390 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c const u32 addr = 0x137000 + (idx * 0x20); idx 411 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c gk104_clk_prog_3(struct gk104_clk *clk, int idx) idx 413 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c struct gk104_clk_info *info = &clk->eng[idx]; idx 416 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x137250 + (idx * 0x04), 0x00003f00, info->mdiv); idx 418 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x137250 + (idx * 0x04), 0x0000003f, info->mdiv); idx 422 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c gk104_clk_prog_4_0(struct gk104_clk *clk, int idx) idx 424 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c struct gk104_clk_info *info = &clk->eng[idx]; idx 427 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x137100, (1 << idx), info->ssel); idx 429 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c u32 tmp = nvkm_rd32(device, 0x137100) & (1 << idx); idx 437 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c gk104_clk_prog_4_1(struct gk104_clk *clk, int idx) idx 439 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c struct gk104_clk_info *info = &clk->eng[idx]; idx 442 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x137160 + (idx * 0x04), 0x40000000, 0x40000000); idx 443 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gk104.c nvkm_mask(device, 0x137160 + (idx * 0x04), 0x00000100, 0x00000100); idx 43 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c read_vco(struct gt215_clk *clk, int idx) idx 46 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c u32 sctl = nvkm_rd32(device, 0x4120 + (idx * 4)); idx 61 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c read_clk(struct gt215_clk *clk, int idx, bool ignore_en) idx 67 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c if (idx >= 0x40) { idx 76 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c sctl = nvkm_rd32(device, 0x4120 + (idx * 4)); idx 99 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c sclk = read_vco(clk, idx); idx 108 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c read_pll(struct gt215_clk *clk, int idx, u32 pll) idx 128 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c sclk = read_clk(clk, 0x00 + idx, false); idx 131 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c sclk = read_clk(clk, 0x10 + idx, false); idx 187 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c gt215_clk_info(struct nvkm_clk *base, int idx, u32 khz, idx 207 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c sclk = read_vco(clk, idx); idx 235 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c gt215_pll_info(struct nvkm_clk *base, int idx, u32 pll, u32 khz, idx 248 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c ret = gt215_clk_info(&clk->base, idx, khz, info); idx 259 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c ret = gt215_clk_info(&clk->base, idx - 0x10, limits.refclk, info); idx 275 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c int idx, u32 pll, int dom) idx 277 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c int ret = gt215_pll_info(&clk->base, idx, pll, cstate->domain[dom], idx 363 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c prog_pll(struct gt215_clk *clk, int idx, u32 pll, int dom) idx 367 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c const u32 src0 = 0x004120 + (idx * 4); idx 368 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c const u32 src1 = 0x004160 + (idx * 4); idx 407 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c prog_clk(struct gt215_clk *clk, int idx, int dom) idx 411 drivers/gpu/drm/nouveau/nvkm/subdev/clk/gt215.c nvkm_mask(device, 0x004120 + (idx * 4), 0x003f3141, 0x00000101 | info->clk); idx 325 drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c calc_pll(struct nv50_clk *clk, u32 reg, u32 idx, int *N, int *M, int *P) idx 340 drivers/gpu/drm/nouveau/nvkm/subdev/clk/nv50.c return nv04_pll_calc(subdev, &pll, idx, N, M, NULL, NULL, P); idx 143 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv50.c u32 trap[6], idx, inst; idx 148 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv50.c idx = nvkm_rd32(device, 0x100c90); idx 149 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv50.c if (!(idx & 0x80000000)) idx 151 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv50.c idx &= 0x00ffffff; idx 154 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv50.c nvkm_wr32(device, 0x100c90, idx | i << 24); idx 157 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv50.c nvkm_wr32(device, 0x100c90, idx | 0x80000000); idx 30 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c nvkm_gpio_drive(struct nvkm_gpio *gpio, int idx, int line, int dir, int out) idx 36 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c nvkm_gpio_sense(struct nvkm_gpio *gpio, int idx, int line) idx 49 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c nvkm_gpio_find(struct nvkm_gpio *gpio, int idx, u8 tag, u8 line, idx 60 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c data = dcb_gpio_match(bios, idx, tag, line, &ver, &len, func); idx 81 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c nvkm_gpio_set(struct nvkm_gpio *gpio, int idx, u8 tag, u8 line, int state) idx 86 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c ret = nvkm_gpio_find(gpio, idx, tag, line, &func); idx 90 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c ret = nvkm_gpio_drive(gpio, idx, func.line, dir, out); idx 97 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c nvkm_gpio_get(struct nvkm_gpio *gpio, int idx, u8 tag, u8 line) idx 102 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c ret = nvkm_gpio_find(gpio, idx, tag, line, &func); idx 104 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c ret = nvkm_gpio_sense(gpio, idx, func.line); idx 94 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c 1 + (rail->idx * 2), 3, idx 95 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c 2 + (rail->idx * 2), 3, rail->mohm, idx 288 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/base.c rail->idx = r; idx 21 drivers/gpu/drm/nouveau/nvkm/subdev/iccsense/priv.h u8 idx; idx 95 drivers/gpu/drm/nouveau/nvkm/subdev/mxm/nv50.c mxm_dcb_sanitise_entry(struct nvkm_bios *bios, void *data, int idx, u16 pdcb) idx 107 drivers/gpu/drm/nouveau/nvkm/subdev/mxm/nv50.c idx, ctx.outp[0], ctx.outp[1]); idx 50 drivers/gpu/drm/omapdrm/dss/dispc.c #define REG_GET(dispc, idx, start, end) \ idx 51 drivers/gpu/drm/omapdrm/dss/dispc.c FLD_GET(dispc_read_reg(dispc, idx), start, end) idx 53 drivers/gpu/drm/omapdrm/dss/dispc.c #define REG_FLD_MOD(dispc, idx, val, start, end) \ idx 54 drivers/gpu/drm/omapdrm/dss/dispc.c dispc_write_reg(dispc, idx, \ idx 55 drivers/gpu/drm/omapdrm/dss/dispc.c FLD_MOD(dispc_read_reg(dispc, idx), val, start, end)) idx 358 drivers/gpu/drm/omapdrm/dss/dispc.c static inline void dispc_write_reg(struct dispc_device *dispc, u16 idx, u32 val) idx 360 drivers/gpu/drm/omapdrm/dss/dispc.c __raw_writel(val, dispc->base + idx); idx 363 drivers/gpu/drm/omapdrm/dss/dispc.c static inline u32 dispc_read_reg(struct dispc_device *dispc, u16 idx) idx 365 drivers/gpu/drm/omapdrm/dss/dispc.c return __raw_readl(dispc->base + idx); idx 1744 drivers/gpu/drm/omapdrm/dss/dispc.c int idx; idx 1781 drivers/gpu/drm/omapdrm/dss/dispc.c idx = 0; idx 1784 drivers/gpu/drm/omapdrm/dss/dispc.c idx = 3; idx 1787 drivers/gpu/drm/omapdrm/dss/dispc.c idx = 2; idx 1790 drivers/gpu/drm/omapdrm/dss/dispc.c idx = 1; idx 1810 drivers/gpu/drm/omapdrm/dss/dispc.c accu_val = &accu_table[idx]; idx 43 drivers/gpu/drm/omapdrm/dss/dsi.c struct dsi_reg { u16 module; u16 idx; }; idx 45 drivers/gpu/drm/omapdrm/dss/dsi.c #define DSI_REG(mod, idx) ((const struct dsi_reg) { mod, idx }) idx 111 drivers/gpu/drm/omapdrm/dss/dsi.c #define REG_GET(dsi, idx, start, end) \ idx 112 drivers/gpu/drm/omapdrm/dss/dsi.c FLD_GET(dsi_read_reg(dsi, idx), start, end) idx 114 drivers/gpu/drm/omapdrm/dss/dsi.c #define REG_FLD_MOD(dsi, idx, val, start, end) \ idx 115 drivers/gpu/drm/omapdrm/dss/dsi.c dsi_write_reg(dsi, idx, FLD_MOD(dsi_read_reg(dsi, idx), val, start, end)) idx 436 drivers/gpu/drm/omapdrm/dss/dsi.c const struct dsi_reg idx, u32 val) idx 440 drivers/gpu/drm/omapdrm/dss/dsi.c switch(idx.module) { idx 447 drivers/gpu/drm/omapdrm/dss/dsi.c __raw_writel(val, base + idx.idx); idx 450 drivers/gpu/drm/omapdrm/dss/dsi.c static inline u32 dsi_read_reg(struct dsi_data *dsi, const struct dsi_reg idx) idx 454 drivers/gpu/drm/omapdrm/dss/dsi.c switch(idx.module) { idx 461 drivers/gpu/drm/omapdrm/dss/dsi.c return __raw_readl(base + idx.idx); idx 489 drivers/gpu/drm/omapdrm/dss/dsi.c const struct dsi_reg idx, idx 499 drivers/gpu/drm/omapdrm/dss/dsi.c if (REG_GET(dsi, idx, bitnum, bitnum) == value) idx 506 drivers/gpu/drm/omapdrm/dss/dsi.c if (REG_GET(dsi, idx, bitnum, bitnum) == value) idx 41 drivers/gpu/drm/omapdrm/dss/dss.c u16 idx; idx 44 drivers/gpu/drm/omapdrm/dss/dss.c #define DSS_REG(idx) ((const struct dss_reg) { idx }) idx 54 drivers/gpu/drm/omapdrm/dss/dss.c #define REG_GET(dss, idx, start, end) \ idx 55 drivers/gpu/drm/omapdrm/dss/dss.c FLD_GET(dss_read_reg(dss, idx), start, end) idx 57 drivers/gpu/drm/omapdrm/dss/dss.c #define REG_FLD_MOD(dss, idx, val, start, end) \ idx 58 drivers/gpu/drm/omapdrm/dss/dss.c dss_write_reg(dss, idx, \ idx 59 drivers/gpu/drm/omapdrm/dss/dss.c FLD_MOD(dss_read_reg(dss, idx), val, start, end)) idx 95 drivers/gpu/drm/omapdrm/dss/dss.c const struct dss_reg idx, u32 val) idx 97 drivers/gpu/drm/omapdrm/dss/dss.c __raw_writel(val, dss->base + idx.idx); idx 100 drivers/gpu/drm/omapdrm/dss/dss.c static inline u32 dss_read_reg(struct dss_device *dss, const struct dss_reg idx) idx 102 drivers/gpu/drm/omapdrm/dss/dss.c return __raw_readl(dss->base + idx.idx); idx 106 drivers/gpu/drm/omapdrm/dss/dss.c dss->ctx[(DSS_##reg).idx / sizeof(u32)] = dss_read_reg(dss, DSS_##reg) idx 108 drivers/gpu/drm/omapdrm/dss/dss.c dss_write_reg(dss, DSS_##reg, dss->ctx[(DSS_##reg).idx / sizeof(u32)]) idx 558 drivers/gpu/drm/omapdrm/dss/dss.c int idx = dss_get_channel_index(channel); idx 563 drivers/gpu/drm/omapdrm/dss/dss.c dss->lcd_clk_source[idx] = clk_src; idx 571 drivers/gpu/drm/omapdrm/dss/dss.c dss->lcd_clk_source[idx] = clk_src; idx 589 drivers/gpu/drm/omapdrm/dss/dss.c int idx = dss_get_channel_index(channel); idx 590 drivers/gpu/drm/omapdrm/dss/dss.c return dss->lcd_clk_source[idx]; idx 265 drivers/gpu/drm/omapdrm/dss/hdmi.h static inline void hdmi_write_reg(void __iomem *base_addr, const u32 idx, idx 268 drivers/gpu/drm/omapdrm/dss/hdmi.h __raw_writel(val, base_addr + idx); idx 271 drivers/gpu/drm/omapdrm/dss/hdmi.h static inline u32 hdmi_read_reg(void __iomem *base_addr, const u32 idx) idx 273 drivers/gpu/drm/omapdrm/dss/hdmi.h return __raw_readl(base_addr + idx); idx 276 drivers/gpu/drm/omapdrm/dss/hdmi.h #define REG_FLD_MOD(base, idx, val, start, end) \ idx 277 drivers/gpu/drm/omapdrm/dss/hdmi.h hdmi_write_reg(base, idx, FLD_MOD(hdmi_read_reg(base, idx),\ idx 279 drivers/gpu/drm/omapdrm/dss/hdmi.h #define REG_GET(base, idx, start, end) \ idx 280 drivers/gpu/drm/omapdrm/dss/hdmi.h FLD_GET(hdmi_read_reg(base, idx), start, end) idx 283 drivers/gpu/drm/omapdrm/dss/hdmi.h const u32 idx, int b2, int b1, u32 val) idx 286 drivers/gpu/drm/omapdrm/dss/hdmi.h while (val != (v = REG_GET(base_addr, idx, b2, b1))) { idx 310 drivers/gpu/drm/omapdrm/dss/venc.c static inline void venc_write_reg(struct venc_device *venc, int idx, u32 val) idx 312 drivers/gpu/drm/omapdrm/dss/venc.c __raw_writel(val, venc->base + idx); idx 315 drivers/gpu/drm/omapdrm/dss/venc.c static inline u32 venc_read_reg(struct venc_device *venc, int idx) idx 317 drivers/gpu/drm/omapdrm/dss/venc.c u32 l = __raw_readl(venc->base + idx); idx 251 drivers/gpu/drm/omapdrm/omap_plane.c int idx, enum drm_plane_type type, idx 263 drivers/gpu/drm/omapdrm/omap_plane.c if (WARN_ON(idx >= ARRAY_SIZE(plane_idx_to_id))) idx 266 drivers/gpu/drm/omapdrm/omap_plane.c id = plane_idx_to_id[idx]; idx 21 drivers/gpu/drm/omapdrm/omap_plane.h int idx, enum drm_plane_type type, idx 117 drivers/gpu/drm/qxl/qxl_cmd.c int idx, ret; idx 143 drivers/gpu/drm/qxl/qxl_cmd.c idx = header->prod & (ring->n_elements - 1); idx 144 drivers/gpu/drm/qxl/qxl_cmd.c elt = ring->ring->elements + idx * ring->element_size; idx 164 drivers/gpu/drm/qxl/qxl_cmd.c int idx; idx 174 drivers/gpu/drm/qxl/qxl_cmd.c idx = header->cons & (ring->n_elements - 1); idx 175 drivers/gpu/drm/qxl/qxl_cmd.c ring_elt = ring->ring->elements + idx * ring->element_size; idx 375 drivers/gpu/drm/qxl/qxl_ioctl.c int byte, idx; idx 378 drivers/gpu/drm/qxl/qxl_ioctl.c idx = param->index % 8; idx 386 drivers/gpu/drm/qxl/qxl_ioctl.c if (qdev->rom->client_capabilities[byte] & (1 << idx)) idx 903 drivers/gpu/drm/r128/r128_cce.c if (copy_to_user(&d->request_indices[i], &buf->idx, idx 904 drivers/gpu/drm/r128/r128_cce.c sizeof(buf->idx))) idx 582 drivers/gpu/drm/r128/r128_state.c DRM_DEBUG("buf=%d nbox=%d\n", buf->idx, sarea_priv->nbox); idx 646 drivers/gpu/drm/r128/r128_state.c DRM_DEBUG("indirect: buf=%d s=0x%x e=0x%x\n", buf->idx, start, end); idx 829 drivers/gpu/drm/r128/r128_state.c buf = dma->buflist[blit->idx]; idx 838 drivers/gpu/drm/r128/r128_state.c DRM_ERROR("sending pending buffer %d\n", blit->idx); idx 1327 drivers/gpu/drm/r128/r128_state.c task_pid_nr(current), vertex->idx, vertex->count, vertex->discard); idx 1329 drivers/gpu/drm/r128/r128_state.c if (vertex->idx < 0 || vertex->idx >= dma->buf_count) { idx 1331 drivers/gpu/drm/r128/r128_state.c vertex->idx, dma->buf_count - 1); idx 1343 drivers/gpu/drm/r128/r128_state.c buf = dma->buflist[vertex->idx]; idx 1352 drivers/gpu/drm/r128/r128_state.c DRM_ERROR("sending pending buffer %d\n", vertex->idx); idx 1380 drivers/gpu/drm/r128/r128_state.c elts->idx, elts->start, elts->end, elts->discard); idx 1382 drivers/gpu/drm/r128/r128_state.c if (elts->idx < 0 || elts->idx >= dma->buf_count) { idx 1384 drivers/gpu/drm/r128/r128_state.c elts->idx, dma->buf_count - 1); idx 1396 drivers/gpu/drm/r128/r128_state.c buf = dma->buflist[elts->idx]; idx 1405 drivers/gpu/drm/r128/r128_state.c DRM_ERROR("sending pending buffer %d\n", elts->idx); idx 1442 drivers/gpu/drm/r128/r128_state.c DRM_DEBUG("pid=%d index=%d\n", task_pid_nr(current), blit->idx); idx 1444 drivers/gpu/drm/r128/r128_state.c if (blit->idx < 0 || blit->idx >= dma->buf_count) { idx 1446 drivers/gpu/drm/r128/r128_state.c blit->idx, dma->buf_count - 1); idx 1528 drivers/gpu/drm/r128/r128_state.c indirect->idx, indirect->start, indirect->end, idx 1531 drivers/gpu/drm/r128/r128_state.c if (indirect->idx < 0 || indirect->idx >= dma->buf_count) { idx 1533 drivers/gpu/drm/r128/r128_state.c indirect->idx, dma->buf_count - 1); idx 1537 drivers/gpu/drm/r128/r128_state.c buf = dma->buflist[indirect->idx]; idx 1546 drivers/gpu/drm/r128/r128_state.c DRM_ERROR("sending pending buffer %d\n", indirect->idx); idx 185 drivers/gpu/drm/radeon/atom.c uint32_t idx, val = 0xCDCDCDCD, align, arg; idx 191 drivers/gpu/drm/radeon/atom.c idx = U16(*ptr); idx 194 drivers/gpu/drm/radeon/atom.c DEBUG("REG[0x%04X]", idx); idx 195 drivers/gpu/drm/radeon/atom.c idx += gctx->reg_block; idx 198 drivers/gpu/drm/radeon/atom.c val = gctx->card->reg_read(gctx->card, idx); idx 219 drivers/gpu/drm/radeon/atom.c idx, 0); idx 223 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 227 drivers/gpu/drm/radeon/atom.c val = get_unaligned_le32((u32 *)&ctx->ps[idx]); idx 229 drivers/gpu/drm/radeon/atom.c DEBUG("PS[0x%02X,0x%04X]", idx, val); idx 232 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 235 drivers/gpu/drm/radeon/atom.c DEBUG("WS[0x%02X]", idx); idx 236 drivers/gpu/drm/radeon/atom.c switch (idx) { idx 265 drivers/gpu/drm/radeon/atom.c val = ctx->ws[idx]; idx 269 drivers/gpu/drm/radeon/atom.c idx = U16(*ptr); idx 273 drivers/gpu/drm/radeon/atom.c DEBUG("ID[0x%04X+%04X]", idx, gctx->data_block); idx 275 drivers/gpu/drm/radeon/atom.c DEBUG("ID[0x%04X]", idx); idx 277 drivers/gpu/drm/radeon/atom.c val = U32(idx + gctx->data_block); idx 280 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 282 drivers/gpu/drm/radeon/atom.c if ((gctx->fb_base + (idx * 4)) > gctx->scratch_size_bytes) { idx 284 drivers/gpu/drm/radeon/atom.c gctx->fb_base + (idx * 4), gctx->scratch_size_bytes); idx 287 drivers/gpu/drm/radeon/atom.c val = gctx->scratch[(gctx->fb_base / 4) + idx]; idx 289 drivers/gpu/drm/radeon/atom.c DEBUG("FB[0x%02X]", idx); idx 319 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 322 drivers/gpu/drm/radeon/atom.c DEBUG("PLL[0x%02X]", idx); idx 323 drivers/gpu/drm/radeon/atom.c val = gctx->card->pll_read(gctx->card, idx); idx 326 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 329 drivers/gpu/drm/radeon/atom.c DEBUG("MC[0x%02X]", idx); idx 330 drivers/gpu/drm/radeon/atom.c val = gctx->card->mc_read(gctx->card, idx); idx 455 drivers/gpu/drm/radeon/atom.c val, idx; idx 464 drivers/gpu/drm/radeon/atom.c idx = U16(*ptr); idx 466 drivers/gpu/drm/radeon/atom.c DEBUG("REG[0x%04X]", idx); idx 467 drivers/gpu/drm/radeon/atom.c idx += gctx->reg_block; idx 470 drivers/gpu/drm/radeon/atom.c if (idx == 0) idx 471 drivers/gpu/drm/radeon/atom.c gctx->card->reg_write(gctx->card, idx, idx 474 drivers/gpu/drm/radeon/atom.c gctx->card->reg_write(gctx->card, idx, val); idx 493 drivers/gpu/drm/radeon/atom.c idx, val); idx 497 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 499 drivers/gpu/drm/radeon/atom.c DEBUG("PS[0x%02X]", idx); idx 500 drivers/gpu/drm/radeon/atom.c ctx->ps[idx] = cpu_to_le32(val); idx 503 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 505 drivers/gpu/drm/radeon/atom.c DEBUG("WS[0x%02X]", idx); idx 506 drivers/gpu/drm/radeon/atom.c switch (idx) { idx 532 drivers/gpu/drm/radeon/atom.c ctx->ws[idx] = val; idx 536 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 538 drivers/gpu/drm/radeon/atom.c if ((gctx->fb_base + (idx * 4)) > gctx->scratch_size_bytes) { idx 540 drivers/gpu/drm/radeon/atom.c gctx->fb_base + (idx * 4), gctx->scratch_size_bytes); idx 542 drivers/gpu/drm/radeon/atom.c gctx->scratch[(gctx->fb_base / 4) + idx] = val; idx 543 drivers/gpu/drm/radeon/atom.c DEBUG("FB[0x%02X]", idx); idx 546 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 548 drivers/gpu/drm/radeon/atom.c DEBUG("PLL[0x%02X]", idx); idx 549 drivers/gpu/drm/radeon/atom.c gctx->card->pll_write(gctx->card, idx, val); idx 552 drivers/gpu/drm/radeon/atom.c idx = U8(*ptr); idx 554 drivers/gpu/drm/radeon/atom.c DEBUG("MC[0x%02X]", idx); idx 555 drivers/gpu/drm/radeon/atom.c gctx->card->mc_write(gctx->card, idx, val); idx 621 drivers/gpu/drm/radeon/atom.c int idx = U8((*ptr)++); idx 624 drivers/gpu/drm/radeon/atom.c if (idx < ATOM_TABLE_NAMES_CNT) idx 625 drivers/gpu/drm/radeon/atom.c SDEBUG(" table: %d (%s)\n", idx, atom_table_names[idx]); idx 627 drivers/gpu/drm/radeon/atom.c SDEBUG(" table: %d\n", idx); idx 628 drivers/gpu/drm/radeon/atom.c if (U16(ctx->ctx->cmd_table + 4 + 2 * idx)) idx 629 drivers/gpu/drm/radeon/atom.c r = atom_execute_table_locked(ctx->ctx, idx, ctx->ps + ctx->ps_shift); idx 835 drivers/gpu/drm/radeon/atom.c int idx = U8(*ptr); idx 837 drivers/gpu/drm/radeon/atom.c SDEBUG(" block: %d\n", idx); idx 838 drivers/gpu/drm/radeon/atom.c if (!idx) idx 840 drivers/gpu/drm/radeon/atom.c else if (idx == 255) idx 843 drivers/gpu/drm/radeon/atom.c ctx->ctx->data_block = U16(ctx->ctx->data_table + 4 + 2 * idx); idx 1366 drivers/gpu/drm/radeon/atom.c int idx = CU16(ctx->data_table + offset); idx 1373 drivers/gpu/drm/radeon/atom.c *size = CU16(idx); idx 1375 drivers/gpu/drm/radeon/atom.c *frev = CU8(idx + 2); idx 1377 drivers/gpu/drm/radeon/atom.c *crev = CU8(idx + 3); idx 1378 drivers/gpu/drm/radeon/atom.c *data_start = idx; idx 1386 drivers/gpu/drm/radeon/atom.c int idx = CU16(ctx->cmd_table + offset); idx 1393 drivers/gpu/drm/radeon/atom.c *frev = CU8(idx + 2); idx 1395 drivers/gpu/drm/radeon/atom.c *crev = CU8(idx + 3); idx 1697 drivers/gpu/drm/radeon/btc_dpm.c int idx = 0; idx 1700 drivers/gpu/drm/radeon/btc_dpm.c idx = 1; idx 1702 drivers/gpu/drm/radeon/btc_dpm.c if ((idx == 1) && !eg_pi->smu_uvd_hs) { idx 1708 drivers/gpu/drm/radeon/btc_dpm.c pi->rlp = eg_pi->ats[idx].rlp; idx 1709 drivers/gpu/drm/radeon/btc_dpm.c pi->rmp = eg_pi->ats[idx].rmp; idx 1710 drivers/gpu/drm/radeon/btc_dpm.c pi->lhp = eg_pi->ats[idx].lhp; idx 1711 drivers/gpu/drm/radeon/btc_dpm.c pi->lmp = eg_pi->ats[idx].lmp; idx 2332 drivers/gpu/drm/radeon/ci_dpm.c u16 v_index, idx; idx 2346 drivers/gpu/drm/radeon/ci_dpm.c idx = v_index; idx 2348 drivers/gpu/drm/radeon/ci_dpm.c idx = rdev->pm.dpm.dyn_state.cac_leakage_table.count - 1; idx 2350 drivers/gpu/drm/radeon/ci_dpm.c rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].vddc * VOLTAGE_SCALE; idx 2352 drivers/gpu/drm/radeon/ci_dpm.c rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].leakage * VOLTAGE_SCALE; idx 2363 drivers/gpu/drm/radeon/ci_dpm.c idx = v_index; idx 2365 drivers/gpu/drm/radeon/ci_dpm.c idx = rdev->pm.dpm.dyn_state.cac_leakage_table.count - 1; idx 2367 drivers/gpu/drm/radeon/ci_dpm.c rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].vddc * VOLTAGE_SCALE; idx 2369 drivers/gpu/drm/radeon/ci_dpm.c rdev->pm.dpm.dyn_state.cac_leakage_table.entries[idx].leakage * VOLTAGE_SCALE; idx 5582 drivers/gpu/drm/radeon/ci_dpm.c u8 *idx; idx 5599 drivers/gpu/drm/radeon/ci_dpm.c idx = (u8 *)&power_state->v2.clockInfoIndex[0]; idx 5601 drivers/gpu/drm/radeon/ci_dpm.c clock_array_index = idx[j]; idx 3474 drivers/gpu/drm/radeon/cik.c DRM_ERROR("radeon: cp failed to lock ring %d (%d).\n", ring->idx, r); idx 3490 drivers/gpu/drm/radeon/cik.c DRM_INFO("ring test on %d succeeded in %d usecs\n", ring->idx, i); idx 3493 drivers/gpu/drm/radeon/cik.c ring->idx, scratch, tmp); idx 3514 drivers/gpu/drm/radeon/cik.c switch (ring->idx) { idx 3637 drivers/gpu/drm/radeon/cik.c if (emit_wait && ring->idx == RADEON_RING_TYPE_GFX_INDEX) { idx 3684 drivers/gpu/drm/radeon/cik.c radeon_sync_rings(rdev, &sync, ring->idx); idx 3705 drivers/gpu/drm/radeon/cik.c r = radeon_fence_emit(rdev, &fence, ring->idx); idx 3797 drivers/gpu/drm/radeon/cik.c r = radeon_ib_get(rdev, ring->idx, &ib, NULL, 256); idx 4348 drivers/gpu/drm/radeon/cik.c int i, idx, r; idx 4354 drivers/gpu/drm/radeon/cik.c idx = CAYMAN_RING_TYPE_CP1_INDEX; idx 4356 drivers/gpu/drm/radeon/cik.c idx = CAYMAN_RING_TYPE_CP2_INDEX; idx 4358 drivers/gpu/drm/radeon/cik.c if (rdev->ring[idx].mqd_obj) { idx 4359 drivers/gpu/drm/radeon/cik.c r = radeon_bo_reserve(rdev->ring[idx].mqd_obj, false); idx 4363 drivers/gpu/drm/radeon/cik.c radeon_bo_unpin(rdev->ring[idx].mqd_obj); idx 4364 drivers/gpu/drm/radeon/cik.c radeon_bo_unreserve(rdev->ring[idx].mqd_obj); idx 4366 drivers/gpu/drm/radeon/cik.c radeon_bo_unref(&rdev->ring[idx].mqd_obj); idx 4367 drivers/gpu/drm/radeon/cik.c rdev->ring[idx].mqd_obj = NULL; idx 4524 drivers/gpu/drm/radeon/cik.c int r, i, j, idx; idx 4573 drivers/gpu/drm/radeon/cik.c idx = CAYMAN_RING_TYPE_CP1_INDEX; idx 4575 drivers/gpu/drm/radeon/cik.c idx = CAYMAN_RING_TYPE_CP2_INDEX; idx 4577 drivers/gpu/drm/radeon/cik.c if (rdev->ring[idx].mqd_obj == NULL) { idx 4582 drivers/gpu/drm/radeon/cik.c NULL, &rdev->ring[idx].mqd_obj); idx 4589 drivers/gpu/drm/radeon/cik.c r = radeon_bo_reserve(rdev->ring[idx].mqd_obj, false); idx 4594 drivers/gpu/drm/radeon/cik.c r = radeon_bo_pin(rdev->ring[idx].mqd_obj, RADEON_GEM_DOMAIN_GTT, idx 4601 drivers/gpu/drm/radeon/cik.c r = radeon_bo_kmap(rdev->ring[idx].mqd_obj, (void **)&buf); idx 4619 drivers/gpu/drm/radeon/cik.c cik_srbm_select(rdev, rdev->ring[idx].me, idx 4620 drivers/gpu/drm/radeon/cik.c rdev->ring[idx].pipe, idx 4621 drivers/gpu/drm/radeon/cik.c rdev->ring[idx].queue, 0); idx 4665 drivers/gpu/drm/radeon/cik.c hqd_gpu_addr = rdev->ring[idx].gpu_addr >> 8; idx 4677 drivers/gpu/drm/radeon/cik.c order_base_2(rdev->ring[idx].ring_size / 8); idx 4719 drivers/gpu/drm/radeon/cik.c DOORBELL_OFFSET(rdev->ring[idx].doorbell_index); idx 4731 drivers/gpu/drm/radeon/cik.c rdev->ring[idx].wptr = 0; idx 4732 drivers/gpu/drm/radeon/cik.c mqd->queue_state.cp_hqd_pq_wptr = rdev->ring[idx].wptr; idx 4747 drivers/gpu/drm/radeon/cik.c radeon_bo_kunmap(rdev->ring[idx].mqd_obj); idx 4748 drivers/gpu/drm/radeon/cik.c radeon_bo_unreserve(rdev->ring[idx].mqd_obj); idx 4750 drivers/gpu/drm/radeon/cik.c rdev->ring[idx].ready = true; idx 4751 drivers/gpu/drm/radeon/cik.c r = radeon_ring_test(rdev, idx, &rdev->ring[idx]); idx 4753 drivers/gpu/drm/radeon/cik.c rdev->ring[idx].ready = false; idx 5695 drivers/gpu/drm/radeon/cik.c int usepfp = (ring->idx == RADEON_RING_TYPE_GFX_INDEX); idx 5737 drivers/gpu/drm/radeon/cik.c cik_hdp_flush_cp_ring_emit(rdev, ring->idx); idx 71 drivers/gpu/drm/radeon/cik_sdma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 95 drivers/gpu/drm/radeon/cik_sdma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 116 drivers/gpu/drm/radeon/cik_sdma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 426 drivers/gpu/drm/radeon/cik_sdma.c r = radeon_ring_test(rdev, ring->idx, ring); idx 604 drivers/gpu/drm/radeon/cik_sdma.c radeon_sync_rings(rdev, &sync, ring->idx); idx 622 drivers/gpu/drm/radeon/cik_sdma.c r = radeon_fence_emit(rdev, &fence, ring->idx); idx 654 drivers/gpu/drm/radeon/cik_sdma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 666 drivers/gpu/drm/radeon/cik_sdma.c DRM_ERROR("radeon: dma failed to lock ring %d (%d).\n", ring->idx, r); idx 684 drivers/gpu/drm/radeon/cik_sdma.c DRM_INFO("ring test on %d succeeded in %d usecs\n", ring->idx, i); idx 687 drivers/gpu/drm/radeon/cik_sdma.c ring->idx, tmp); idx 711 drivers/gpu/drm/radeon/cik_sdma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 721 drivers/gpu/drm/radeon/cik_sdma.c r = radeon_ib_get(rdev, ring->idx, &ib, NULL, 256); idx 780 drivers/gpu/drm/radeon/cik_sdma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 987 drivers/gpu/drm/radeon/cik_sdma.c cik_sdma_hdp_flush_ring_emit(rdev, ring->idx); idx 1509 drivers/gpu/drm/radeon/evergreen.c int idx; idx 1519 drivers/gpu/drm/radeon/evergreen.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 0); idx 1521 drivers/gpu/drm/radeon/evergreen.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0); idx 1523 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_ps_idx = idx; idx 1524 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_ps_idx = idx; idx 1528 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_ps_idx = idx; idx 1529 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_ps_idx = idx; idx 1533 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_ps_idx = idx; idx 1534 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_ps_idx = idx; idx 1538 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_ps_idx = idx; idx 1539 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_ps_idx = idx; idx 1544 drivers/gpu/drm/radeon/evergreen.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0); idx 1545 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_ps_idx = idx; idx 1546 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_ps_idx = idx; idx 1549 drivers/gpu/drm/radeon/evergreen.c rdev->pm.power_state[idx].num_clock_modes - 1; idx 1551 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_ps_idx = idx; idx 1552 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_ps_idx = idx; idx 1555 drivers/gpu/drm/radeon/evergreen.c rdev->pm.power_state[idx].num_clock_modes - 1; idx 1569 drivers/gpu/drm/radeon/evergreen.c int idx; idx 1581 drivers/gpu/drm/radeon/evergreen.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 0); idx 1583 drivers/gpu/drm/radeon/evergreen.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0); idx 1585 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_ps_idx = idx; idx 1586 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_ps_idx = idx; idx 1590 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_ps_idx = idx; idx 1591 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_ps_idx = idx; idx 1595 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_ps_idx = idx; idx 1596 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_ps_idx = idx; idx 1600 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_ps_idx = idx; idx 1601 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_ps_idx = idx; idx 1605 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_ps_idx = idx; idx 1606 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_ps_idx = idx; idx 1610 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_ps_idx = idx; idx 1611 drivers/gpu/drm/radeon/evergreen.c rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_ps_idx = idx; idx 756 drivers/gpu/drm/radeon/evergreen_cs.c unsigned idx) idx 764 drivers/gpu/drm/radeon/evergreen_cs.c texdw[0] = radeon_get_ib_value(p, idx + 0); idx 765 drivers/gpu/drm/radeon/evergreen_cs.c texdw[1] = radeon_get_ib_value(p, idx + 1); idx 766 drivers/gpu/drm/radeon/evergreen_cs.c texdw[2] = radeon_get_ib_value(p, idx + 2); idx 767 drivers/gpu/drm/radeon/evergreen_cs.c texdw[3] = radeon_get_ib_value(p, idx + 3); idx 768 drivers/gpu/drm/radeon/evergreen_cs.c texdw[4] = radeon_get_ib_value(p, idx + 4); idx 769 drivers/gpu/drm/radeon/evergreen_cs.c texdw[5] = radeon_get_ib_value(p, idx + 5); idx 770 drivers/gpu/drm/radeon/evergreen_cs.c texdw[6] = radeon_get_ib_value(p, idx + 6); idx 771 drivers/gpu/drm/radeon/evergreen_cs.c texdw[7] = radeon_get_ib_value(p, idx + 7); idx 1050 drivers/gpu/drm/radeon/evergreen_cs.c unsigned idx, unsigned reg) idx 1059 drivers/gpu/drm/radeon/evergreen_cs.c idx, reg); idx 1064 drivers/gpu/drm/radeon/evergreen_cs.c pr_err("Forbidden register 0x%04X in cs at %d\n", reg, idx); idx 1074 drivers/gpu/drm/radeon/evergreen_cs.c unsigned idx; idx 1077 drivers/gpu/drm/radeon/evergreen_cs.c idx = pkt->idx + 1; idx 1079 drivers/gpu/drm/radeon/evergreen_cs.c for (i = 0; i <= pkt->count; i++, idx++, reg += 4) { idx 1080 drivers/gpu/drm/radeon/evergreen_cs.c r = evergreen_packet0_check(p, pkt, idx, reg); idx 1094 drivers/gpu/drm/radeon/evergreen_cs.c static int evergreen_cs_handle_reg(struct radeon_cs_parser *p, u32 reg, u32 idx) idx 1149 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1152 drivers/gpu/drm/radeon/evergreen_cs.c track->db_depth_control = radeon_get_ib_value(p, idx); idx 1170 drivers/gpu/drm/radeon/evergreen_cs.c track->db_z_info = radeon_get_ib_value(p, idx); idx 1178 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] &= ~Z_ARRAY_MODE(0xf); idx 1180 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= Z_ARRAY_MODE(evergreen_cs_get_aray_mode(reloc->tiling_flags)); idx 1188 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= DB_NUM_BANKS(evergreen_cs_get_num_banks(track->nbanks)); idx 1189 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= DB_TILE_SPLIT(tile_split) | idx 1198 drivers/gpu/drm/radeon/evergreen_cs.c track->db_s_info = radeon_get_ib_value(p, idx); idx 1202 drivers/gpu/drm/radeon/evergreen_cs.c track->db_depth_view = radeon_get_ib_value(p, idx); idx 1206 drivers/gpu/drm/radeon/evergreen_cs.c track->db_depth_size = radeon_get_ib_value(p, idx); idx 1210 drivers/gpu/drm/radeon/evergreen_cs.c track->db_depth_slice = radeon_get_ib_value(p, idx); idx 1220 drivers/gpu/drm/radeon/evergreen_cs.c track->db_z_read_offset = radeon_get_ib_value(p, idx); idx 1221 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1232 drivers/gpu/drm/radeon/evergreen_cs.c track->db_z_write_offset = radeon_get_ib_value(p, idx); idx 1233 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1244 drivers/gpu/drm/radeon/evergreen_cs.c track->db_s_read_offset = radeon_get_ib_value(p, idx); idx 1245 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1256 drivers/gpu/drm/radeon/evergreen_cs.c track->db_s_write_offset = radeon_get_ib_value(p, idx); idx 1257 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1262 drivers/gpu/drm/radeon/evergreen_cs.c track->vgt_strmout_config = radeon_get_ib_value(p, idx); idx 1266 drivers/gpu/drm/radeon/evergreen_cs.c track->vgt_strmout_buffer_config = radeon_get_ib_value(p, idx); idx 1280 drivers/gpu/drm/radeon/evergreen_cs.c track->vgt_strmout_bo_offset[tmp] = radeon_get_ib_value(p, idx) << 8; idx 1281 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1291 drivers/gpu/drm/radeon/evergreen_cs.c track->vgt_strmout_size[tmp] = radeon_get_ib_value(p, idx) * 4; idx 1301 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1304 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_target_mask = radeon_get_ib_value(p, idx); idx 1308 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_shader_mask = radeon_get_ib_value(p, idx); idx 1317 drivers/gpu/drm/radeon/evergreen_cs.c tmp = radeon_get_ib_value(p, idx) & MSAA_NUM_SAMPLES_MASK; idx 1326 drivers/gpu/drm/radeon/evergreen_cs.c tmp = radeon_get_ib_value(p, idx) & CAYMAN_MSAA_NUM_SAMPLES_MASK; idx 1338 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_view[tmp] = radeon_get_ib_value(p, idx); idx 1346 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_view[tmp] = radeon_get_ib_value(p, idx); idx 1358 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_info[tmp] = radeon_get_ib_value(p, idx); idx 1366 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= CB_ARRAY_MODE(evergreen_cs_get_aray_mode(reloc->tiling_flags)); idx 1376 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_info[tmp] = radeon_get_ib_value(p, idx); idx 1384 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= CB_ARRAY_MODE(evergreen_cs_get_aray_mode(reloc->tiling_flags)); idx 1398 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_pitch[tmp] = radeon_get_ib_value(p, idx); idx 1406 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_pitch[tmp] = radeon_get_ib_value(p, idx); idx 1418 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_slice[tmp] = radeon_get_ib_value(p, idx); idx 1419 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_slice_idx[tmp] = idx; idx 1427 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_slice[tmp] = radeon_get_ib_value(p, idx); idx 1428 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_slice_idx[tmp] = idx; idx 1452 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= CB_NUM_BANKS(evergreen_cs_get_num_banks(track->nbanks)); idx 1453 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= CB_TILE_SPLIT(tile_split) | idx 1460 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_attrib[tmp] = ib[idx]; idx 1480 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= CB_NUM_BANKS(evergreen_cs_get_num_banks(track->nbanks)); idx 1481 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= CB_TILE_SPLIT(tile_split) | idx 1488 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_attrib[tmp] = ib[idx]; idx 1505 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1522 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1534 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_fmask_slice[tmp] = radeon_get_ib_value(p, idx); idx 1545 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_cmask_slice[tmp] = radeon_get_ib_value(p, idx); idx 1562 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_bo_offset[tmp] = radeon_get_ib_value(p, idx); idx 1563 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1578 drivers/gpu/drm/radeon/evergreen_cs.c track->cb_color_bo_offset[tmp] = radeon_get_ib_value(p, idx); idx 1579 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1590 drivers/gpu/drm/radeon/evergreen_cs.c track->htile_offset = radeon_get_ib_value(p, idx); idx 1591 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1597 drivers/gpu/drm/radeon/evergreen_cs.c track->htile_surface = radeon_get_ib_value(p, idx); idx 1599 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] |= 3; idx 1708 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1722 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1736 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1739 drivers/gpu/drm/radeon/evergreen_cs.c track->sx_misc_kill_all_prims = (radeon_get_ib_value(p, idx) & 0x1) != 0; idx 1742 drivers/gpu/drm/radeon/evergreen_cs.c dev_warn(p->dev, "forbidden register 0x%08x at %d\n", reg, idx); idx 1778 drivers/gpu/drm/radeon/evergreen_cs.c unsigned idx; idx 1786 drivers/gpu/drm/radeon/evergreen_cs.c idx = pkt->idx + 1; idx 1787 drivers/gpu/drm/radeon/evergreen_cs.c idx_value = radeon_get_ib_value(p, idx); idx 1801 drivers/gpu/drm/radeon/evergreen_cs.c tmp = radeon_get_ib_value(p, idx + 1); idx 1823 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx + 0] = offset; idx 1824 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx + 1] = (tmp & 0xffffff00) | (upper_32_bits(offset) & 0xff); idx 1867 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+1) & 0xff) << 32); idx 1869 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+0] = offset; idx 1870 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = upper_32_bits(offset) & 0xff; idx 1902 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+1) & 0xff) << 32); idx 1904 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+0] = offset; idx 1905 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = upper_32_bits(offset) & 0xff; idx 1929 drivers/gpu/drm/radeon/evergreen_cs.c radeon_get_ib_value(p, idx+1) + idx 1930 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 1932 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = offset; idx 1933 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 1949 drivers/gpu/drm/radeon/evergreen_cs.c dev_warn(p->dev, "%s:%d invalid cmd stream %d\n", __func__, __LINE__, idx); idx 1960 drivers/gpu/drm/radeon/evergreen_cs.c dev_warn(p->dev, "%s:%d invalid cmd stream %d\n", __func__, __LINE__, idx); idx 2025 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = reloc->gpu_offset; idx 2026 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = upper_32_bits(reloc->gpu_offset) & 0xff; idx 2065 drivers/gpu/drm/radeon/evergreen_cs.c dev_warn(p->dev, "%s:%d invalid cmd stream %d\n", __func__, __LINE__, idx); idx 2079 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+0] = idx_value + (u32)(reloc->gpu_offset & 0xffffffff); idx 2102 drivers/gpu/drm/radeon/evergreen_cs.c (radeon_get_ib_value(p, idx+1) & 0xfffffffc) + idx 2103 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 2105 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = (ib[idx+1] & 0x3) | (offset & 0xfffffffc); idx 2106 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 2120 drivers/gpu/drm/radeon/evergreen_cs.c command = radeon_get_ib_value(p, idx+4); idx 2122 drivers/gpu/drm/radeon/evergreen_cs.c info = radeon_get_ib_value(p, idx+1); idx 2155 drivers/gpu/drm/radeon/evergreen_cs.c tmp = radeon_get_ib_value(p, idx) + idx 2156 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+1) & 0xff) << 32); idx 2166 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx] = offset; idx 2167 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = (ib[idx+1] & 0xffffff00) | (upper_32_bits(offset) & 0xff); idx 2193 drivers/gpu/drm/radeon/evergreen_cs.c tmp = radeon_get_ib_value(p, idx+2) + idx 2194 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+3) & 0xff) << 32); idx 2204 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = offset; idx 2205 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+3] = upper_32_bits(offset) & 0xff; idx 2225 drivers/gpu/drm/radeon/evergreen_cs.c if (radeon_get_ib_value(p, idx + 1) != 0xffffffff || idx 2226 drivers/gpu/drm/radeon/evergreen_cs.c radeon_get_ib_value(p, idx + 2) != 0) { idx 2232 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 2249 drivers/gpu/drm/radeon/evergreen_cs.c (radeon_get_ib_value(p, idx+1) & 0xfffffff8) + idx 2250 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 2252 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = offset & 0xfffffff8; idx 2253 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 2271 drivers/gpu/drm/radeon/evergreen_cs.c (radeon_get_ib_value(p, idx+1) & 0xfffffffc) + idx 2272 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 2274 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = offset & 0xfffffffc; idx 2275 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = (ib[idx+2] & 0xffffff00) | (upper_32_bits(offset) & 0xff); idx 2293 drivers/gpu/drm/radeon/evergreen_cs.c (radeon_get_ib_value(p, idx+1) & 0xfffffffc) + idx 2294 drivers/gpu/drm/radeon/evergreen_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 2296 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = offset & 0xfffffffc; idx 2297 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = (ib[idx+2] & 0xffffff00) | (upper_32_bits(offset) & 0xff); idx 2309 drivers/gpu/drm/radeon/evergreen_cs.c for (reg = start_reg, idx++; reg <= end_reg; reg += 4, idx++) { idx 2312 drivers/gpu/drm/radeon/evergreen_cs.c r = evergreen_cs_handle_reg(p, reg, idx); idx 2326 drivers/gpu/drm/radeon/evergreen_cs.c for (reg = start_reg, idx++; reg <= end_reg; reg += 4, idx++) { idx 2329 drivers/gpu/drm/radeon/evergreen_cs.c r = evergreen_cs_handle_reg(p, reg, idx); idx 2352 drivers/gpu/drm/radeon/evergreen_cs.c switch (G__SQ_CONSTANT_TYPE(radeon_get_ib_value(p, idx+1+(i*8)+7))) { idx 2361 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+1] |= idx 2369 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+6] |= TEX_TILE_SPLIT(tile_split); idx 2370 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+7] |= idx 2381 drivers/gpu/drm/radeon/evergreen_cs.c tex_dim = ib[idx+1+(i*8)+0] & 0x7; idx 2382 drivers/gpu/drm/radeon/evergreen_cs.c mip_address = ib[idx+1+(i*8)+3]; idx 2401 drivers/gpu/drm/radeon/evergreen_cs.c r = evergreen_cs_track_validate_texture(p, texture, mipmap, idx+1+(i*8)); idx 2404 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+2] += toffset; idx 2405 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+3] += moffset; idx 2416 drivers/gpu/drm/radeon/evergreen_cs.c offset = radeon_get_ib_value(p, idx+1+(i*8)+0); idx 2417 drivers/gpu/drm/radeon/evergreen_cs.c size = radeon_get_ib_value(p, idx+1+(i*8)+1); idx 2421 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+1] = radeon_bo_size(reloc->robj) - offset; idx 2425 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+0] = offset64; idx 2426 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1+(i*8)+2] = (ib[idx+1+(i*8)+2] & 0xffffff00) | idx 2498 drivers/gpu/drm/radeon/evergreen_cs.c offset = radeon_get_ib_value(p, idx+1); idx 2499 drivers/gpu/drm/radeon/evergreen_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+2) & 0xff)) << 32; idx 2506 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = offset; idx 2507 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 2517 drivers/gpu/drm/radeon/evergreen_cs.c offset = radeon_get_ib_value(p, idx+3); idx 2518 drivers/gpu/drm/radeon/evergreen_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2525 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+3] = offset; idx 2526 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+4] = upper_32_bits(offset) & 0xff; idx 2542 drivers/gpu/drm/radeon/evergreen_cs.c offset = radeon_get_ib_value(p, idx+0); idx 2543 drivers/gpu/drm/radeon/evergreen_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+1) & 0xff)) << 32UL; idx 2554 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+0] = offset; idx 2555 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = upper_32_bits(offset) & 0xff; idx 2571 drivers/gpu/drm/radeon/evergreen_cs.c offset = radeon_get_ib_value(p, idx+1); idx 2572 drivers/gpu/drm/radeon/evergreen_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+2) & 0xff)) << 32; idx 2579 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = offset; idx 2580 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 2583 drivers/gpu/drm/radeon/evergreen_cs.c reg = radeon_get_ib_value(p, idx+1) << 2; idx 2586 drivers/gpu/drm/radeon/evergreen_cs.c reg, idx + 1); idx 2598 drivers/gpu/drm/radeon/evergreen_cs.c offset = radeon_get_ib_value(p, idx+3); idx 2599 drivers/gpu/drm/radeon/evergreen_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2606 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+3] = offset; idx 2607 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+4] = upper_32_bits(offset) & 0xff; idx 2610 drivers/gpu/drm/radeon/evergreen_cs.c reg = radeon_get_ib_value(p, idx+3) << 2; idx 2613 drivers/gpu/drm/radeon/evergreen_cs.c reg, idx + 3); idx 2635 drivers/gpu/drm/radeon/evergreen_cs.c areg, idx); idx 2648 drivers/gpu/drm/radeon/evergreen_cs.c offset = radeon_get_ib_value(p, idx + 1); idx 2652 drivers/gpu/drm/radeon/evergreen_cs.c offset += ((u64)(radeon_get_ib_value(p, idx + 2) & 0xff)) << 32; idx 2655 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] = (offset & 0xfffffffc) | swap; idx 2656 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 2749 drivers/gpu/drm/radeon/evergreen_cs.c r = radeon_cs_packet_parse(p, &pkt, p->idx); idx 2755 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += pkt.count + 2; idx 2776 drivers/gpu/drm/radeon/evergreen_cs.c } while (p->idx < p->chunk_ib->length_dw); idx 2803 drivers/gpu/drm/radeon/evergreen_cs.c u32 idx; idx 2808 drivers/gpu/drm/radeon/evergreen_cs.c if (p->idx >= ib_chunk->length_dw) { idx 2810 drivers/gpu/drm/radeon/evergreen_cs.c p->idx, ib_chunk->length_dw); idx 2813 drivers/gpu/drm/radeon/evergreen_cs.c idx = p->idx; idx 2814 drivers/gpu/drm/radeon/evergreen_cs.c header = radeon_get_ib_value(p, idx); idx 2829 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2832 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 2833 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += count + 7; idx 2837 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2838 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+2) & 0xff)) << 32; idx 2840 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2841 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2842 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += count + 3; idx 2845 drivers/gpu/drm/radeon/evergreen_cs.c DRM_ERROR("bad DMA_PACKET_WRITE [%6d] 0x%08x sub cmd is not 0 or 8\n", idx, header); idx 2869 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+2); idx 2870 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2871 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2872 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32; idx 2883 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2884 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 2885 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+3] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2886 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+4] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2887 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 5; idx 2892 drivers/gpu/drm/radeon/evergreen_cs.c if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) { idx 2894 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+1); idx 2896 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(src_reloc->gpu_offset >> 8); idx 2898 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx + 7); idx 2899 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32; idx 2900 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+7] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2901 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2904 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+7); idx 2905 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32; idx 2906 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+7] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 2907 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2909 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2911 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 2923 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 9; idx 2928 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+2); idx 2929 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2930 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2931 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32; idx 2942 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xffffffff); idx 2943 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += (u32)(src_reloc->gpu_offset & 0xffffffff); idx 2944 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+3] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2945 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+4] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2946 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 5; idx 2955 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(src_reloc->gpu_offset & 0xffffffff); idx 2956 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2957 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+4] += (u32)(dst_reloc->gpu_offset & 0xffffffff); idx 2958 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+5] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2960 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 9; idx 2970 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2971 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2972 drivers/gpu/drm/radeon/evergreen_cs.c dst2_offset = radeon_get_ib_value(p, idx+2); idx 2973 drivers/gpu/drm/radeon/evergreen_cs.c dst2_offset |= ((u64)(radeon_get_ib_value(p, idx+5) & 0xff)) << 32; idx 2974 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+3); idx 2975 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+6) & 0xff)) << 32; idx 2991 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2992 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += (u32)(dst2_reloc->gpu_offset & 0xfffffffc); idx 2993 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+3] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 2994 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+4] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2995 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+5] += upper_32_bits(dst2_reloc->gpu_offset) & 0xff; idx 2996 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+6] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2997 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 7; idx 3001 drivers/gpu/drm/radeon/evergreen_cs.c if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) { idx 3010 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 3012 drivers/gpu/drm/radeon/evergreen_cs.c dst2_offset = radeon_get_ib_value(p, idx+2); idx 3014 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+8); idx 3015 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32; idx 3031 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 3032 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += (u32)(dst2_reloc->gpu_offset >> 8); idx 3033 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 3034 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+9] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 3035 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 10; idx 3045 drivers/gpu/drm/radeon/evergreen_cs.c if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) { idx 3047 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(src_reloc->gpu_offset >> 8); idx 3049 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+7] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 3050 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 3053 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+7] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 3054 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 3056 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 3058 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 12; idx 3063 drivers/gpu/drm/radeon/evergreen_cs.c if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) { idx 3072 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 3074 drivers/gpu/drm/radeon/evergreen_cs.c dst2_offset = radeon_get_ib_value(p, idx+2); idx 3076 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+8); idx 3077 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32; idx 3093 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 3094 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += (u32)(dst2_reloc->gpu_offset >> 8); idx 3095 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 3096 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+9] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 3097 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 10; idx 3103 drivers/gpu/drm/radeon/evergreen_cs.c if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) { idx 3105 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+1); idx 3107 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(src_reloc->gpu_offset >> 8); idx 3109 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+7); idx 3110 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32; idx 3111 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+7] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 3112 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 3115 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+7); idx 3116 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+8) & 0xff)) << 32; idx 3117 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+7] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 3118 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 3120 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 3122 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 3134 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 9; idx 3143 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(src_reloc->gpu_offset >> 8); idx 3144 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+4] += (u32)(dst_reloc->gpu_offset >> 8); idx 3145 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 13; idx 3150 drivers/gpu/drm/radeon/evergreen_cs.c if (radeon_get_ib_value(p, idx + 2) & (1 << 31)) { idx 3159 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 3161 drivers/gpu/drm/radeon/evergreen_cs.c dst2_offset = radeon_get_ib_value(p, idx+2); idx 3163 drivers/gpu/drm/radeon/evergreen_cs.c src_offset = radeon_get_ib_value(p, idx+8); idx 3164 drivers/gpu/drm/radeon/evergreen_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+9) & 0xff)) << 32; idx 3180 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 3181 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+2] += (u32)(dst2_reloc->gpu_offset >> 8); idx 3182 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+8] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 3183 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+9] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 3184 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 10; idx 3187 drivers/gpu/drm/radeon/evergreen_cs.c DRM_ERROR("bad DMA_PACKET_COPY [%6d] 0x%08x invalid sub cmd\n", idx, header); idx 3197 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 3198 drivers/gpu/drm/radeon/evergreen_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0x00ff0000)) << 16; idx 3204 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 3205 drivers/gpu/drm/radeon/evergreen_cs.c ib[idx+3] += (upper_32_bits(dst_reloc->gpu_offset) << 16) & 0x00ff0000; idx 3206 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 4; idx 3209 drivers/gpu/drm/radeon/evergreen_cs.c p->idx += 1; idx 3212 drivers/gpu/drm/radeon/evergreen_cs.c DRM_ERROR("Unknown packet type %d at %d !\n", cmd, idx); idx 3215 drivers/gpu/drm/radeon/evergreen_cs.c } while (p->idx < p->chunk_ib->length_dw); idx 3351 drivers/gpu/drm/radeon/evergreen_cs.c u32 idx = pkt->idx + 1; idx 3352 drivers/gpu/drm/radeon/evergreen_cs.c u32 idx_value = ib[idx]; idx 3409 drivers/gpu/drm/radeon/evergreen_cs.c reg = ib[idx + 5] * 4; idx 3416 drivers/gpu/drm/radeon/evergreen_cs.c reg = ib[idx + 3] * 4; idx 3437 drivers/gpu/drm/radeon/evergreen_cs.c command = ib[idx + 4]; idx 3438 drivers/gpu/drm/radeon/evergreen_cs.c info = ib[idx + 1]; idx 3475 drivers/gpu/drm/radeon/evergreen_cs.c start_reg = ib[idx + 2]; idx 3510 drivers/gpu/drm/radeon/evergreen_cs.c areg, idx); idx 3524 drivers/gpu/drm/radeon/evergreen_cs.c u32 idx = 0; idx 3528 drivers/gpu/drm/radeon/evergreen_cs.c pkt.idx = idx; idx 3529 drivers/gpu/drm/radeon/evergreen_cs.c pkt.type = RADEON_CP_PACKET_GET_TYPE(ib->ptr[idx]); idx 3530 drivers/gpu/drm/radeon/evergreen_cs.c pkt.count = RADEON_CP_PACKET_GET_COUNT(ib->ptr[idx]); idx 3538 drivers/gpu/drm/radeon/evergreen_cs.c idx += 1; idx 3541 drivers/gpu/drm/radeon/evergreen_cs.c pkt.opcode = RADEON_CP_PACKET3_GET_OPCODE(ib->ptr[idx]); idx 3543 drivers/gpu/drm/radeon/evergreen_cs.c idx += pkt.count + 2; idx 3552 drivers/gpu/drm/radeon/evergreen_cs.c } while (idx < ib->length_dw); idx 3568 drivers/gpu/drm/radeon/evergreen_cs.c u32 idx = 0; idx 3572 drivers/gpu/drm/radeon/evergreen_cs.c header = ib->ptr[idx]; idx 3582 drivers/gpu/drm/radeon/evergreen_cs.c idx += count + 7; idx 3586 drivers/gpu/drm/radeon/evergreen_cs.c idx += count + 3; idx 3589 drivers/gpu/drm/radeon/evergreen_cs.c DRM_ERROR("bad DMA_PACKET_WRITE [%6d] 0x%08x sub cmd is not 0 or 8\n", idx, ib->ptr[idx]); idx 3597 drivers/gpu/drm/radeon/evergreen_cs.c idx += 5; idx 3601 drivers/gpu/drm/radeon/evergreen_cs.c idx += 9; idx 3605 drivers/gpu/drm/radeon/evergreen_cs.c idx += 5; idx 3609 drivers/gpu/drm/radeon/evergreen_cs.c idx += 9; idx 3613 drivers/gpu/drm/radeon/evergreen_cs.c idx += 7; idx 3617 drivers/gpu/drm/radeon/evergreen_cs.c idx += 10; idx 3621 drivers/gpu/drm/radeon/evergreen_cs.c idx += 12; idx 3625 drivers/gpu/drm/radeon/evergreen_cs.c idx += 10; idx 3629 drivers/gpu/drm/radeon/evergreen_cs.c idx += 9; idx 3633 drivers/gpu/drm/radeon/evergreen_cs.c idx += 13; idx 3637 drivers/gpu/drm/radeon/evergreen_cs.c idx += 10; idx 3640 drivers/gpu/drm/radeon/evergreen_cs.c DRM_ERROR("bad DMA_PACKET_COPY [%6d] 0x%08x invalid sub cmd\n", idx, ib->ptr[idx]); idx 3645 drivers/gpu/drm/radeon/evergreen_cs.c idx += 4; idx 3648 drivers/gpu/drm/radeon/evergreen_cs.c idx += 1; idx 3651 drivers/gpu/drm/radeon/evergreen_cs.c DRM_ERROR("Unknown packet type %d at %d !\n", cmd, idx); idx 3654 drivers/gpu/drm/radeon/evergreen_cs.c } while (idx < ib->length_dw); idx 133 drivers/gpu/drm/radeon/evergreen_dma.c radeon_sync_rings(rdev, &sync, ring->idx); idx 149 drivers/gpu/drm/radeon/evergreen_dma.c r = radeon_fence_emit(rdev, &fence, ring->idx); idx 1974 drivers/gpu/drm/radeon/kv_dpm.c int idx = pi->sys_info.sclk_voltage_mapping_table.num_max_dpm_entries - 1; idx 1976 drivers/gpu/drm/radeon/kv_dpm.c pi->sys_info.sclk_voltage_mapping_table.entries[idx].sclk_frequency; idx 1979 drivers/gpu/drm/radeon/kv_dpm.c pi->sys_info.sclk_voltage_mapping_table.entries[idx].vid_2bit); idx 2672 drivers/gpu/drm/radeon/kv_dpm.c u8 *idx; idx 2686 drivers/gpu/drm/radeon/kv_dpm.c idx = (u8 *)&power_state->v2.clockInfoIndex[0]; idx 2688 drivers/gpu/drm/radeon/kv_dpm.c clock_array_index = idx[j]; idx 1482 drivers/gpu/drm/radeon/ni.c if (ring->idx == RADEON_RING_TYPE_GFX_INDEX) idx 1484 drivers/gpu/drm/radeon/ni.c else if (ring->idx == CAYMAN_RING_TYPE_CP1_INDEX) idx 1498 drivers/gpu/drm/radeon/ni.c if (ring->idx == RADEON_RING_TYPE_GFX_INDEX) idx 1500 drivers/gpu/drm/radeon/ni.c else if (ring->idx == CAYMAN_RING_TYPE_CP1_INDEX) idx 1511 drivers/gpu/drm/radeon/ni.c if (ring->idx == RADEON_RING_TYPE_GFX_INDEX) { idx 1514 drivers/gpu/drm/radeon/ni.c } else if (ring->idx == CAYMAN_RING_TYPE_CP1_INDEX) { idx 61 drivers/gpu/drm/radeon/ni_dma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 85 drivers/gpu/drm/radeon/ni_dma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 106 drivers/gpu/drm/radeon/ni_dma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 250 drivers/gpu/drm/radeon/ni_dma.c r = radeon_ring_test(rdev, ring->idx, ring); idx 292 drivers/gpu/drm/radeon/ni_dma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 4020 drivers/gpu/drm/radeon/ni_dpm.c u8 *idx; idx 4030 drivers/gpu/drm/radeon/ni_dpm.c idx = (u8 *)&power_state->v1.ucClockStateIndices[0]; idx 4035 drivers/gpu/drm/radeon/ni_dpm.c (idx[j] * power_info->pplib.ucClockInfoSize)); idx 1261 drivers/gpu/drm/radeon/r100.c unsigned idx, idx 1273 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1278 drivers/gpu/drm/radeon/r100.c value = radeon_get_ib_value(p, idx); idx 1295 drivers/gpu/drm/radeon/r100.c p->ib.ptr[idx] = (value & 0x3fc00000) | tmp; idx 1297 drivers/gpu/drm/radeon/r100.c p->ib.ptr[idx] = (value & 0xffc00000) | tmp; idx 1303 drivers/gpu/drm/radeon/r100.c int idx) idx 1314 drivers/gpu/drm/radeon/r100.c c = radeon_get_ib_value(p, idx++) & 0x1F; idx 1322 drivers/gpu/drm/radeon/r100.c for (i = 0; i < (c - 1); i+=2, idx+=3) { idx 1330 drivers/gpu/drm/radeon/r100.c idx_value = radeon_get_ib_value(p, idx); idx 1331 drivers/gpu/drm/radeon/r100.c ib[idx+1] = radeon_get_ib_value(p, idx + 1) + ((u32)reloc->gpu_offset); idx 1343 drivers/gpu/drm/radeon/r100.c ib[idx+2] = radeon_get_ib_value(p, idx + 2) + ((u32)reloc->gpu_offset); idx 1356 drivers/gpu/drm/radeon/r100.c idx_value = radeon_get_ib_value(p, idx); idx 1357 drivers/gpu/drm/radeon/r100.c ib[idx+1] = radeon_get_ib_value(p, idx + 1) + ((u32)reloc->gpu_offset); idx 1372 drivers/gpu/drm/radeon/r100.c unsigned idx; idx 1375 drivers/gpu/drm/radeon/r100.c idx = pkt->idx + 1; idx 1390 drivers/gpu/drm/radeon/r100.c for (i = 0; i <= pkt->count; i++, idx++) { idx 1394 drivers/gpu/drm/radeon/r100.c r = check(p, pkt, idx, reg); idx 1437 drivers/gpu/drm/radeon/r100.c r = radeon_cs_packet_parse(p, &waitreloc, p->idx); idx 1448 drivers/gpu/drm/radeon/r100.c if (radeon_get_ib_value(p, waitreloc.idx + 1) != RADEON_WAIT_CRTC_VLINE) { idx 1454 drivers/gpu/drm/radeon/r100.c r = radeon_cs_packet_parse(p, &p3reloc, p->idx + waitreloc.count + 2); idx 1458 drivers/gpu/drm/radeon/r100.c h_idx = p->idx - 2; idx 1459 drivers/gpu/drm/radeon/r100.c p->idx += waitreloc.count + 2; idx 1460 drivers/gpu/drm/radeon/r100.c p->idx += p3reloc.count + 2; idx 1553 drivers/gpu/drm/radeon/r100.c unsigned idx, unsigned reg) idx 1567 drivers/gpu/drm/radeon/r100.c idx_value = radeon_get_ib_value(p, idx); idx 1574 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1583 drivers/gpu/drm/radeon/r100.c r = r100_reloc_pitch_offset(p, pkt, idx, reg); idx 1591 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1598 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1604 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1611 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1620 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1632 drivers/gpu/drm/radeon/r100.c ib[idx] = tmp + ((u32)reloc->gpu_offset); idx 1634 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1647 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1652 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1665 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1670 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1683 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1688 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1701 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1713 drivers/gpu/drm/radeon/r100.c ib[idx] = tmp; idx 1715 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value; idx 1772 drivers/gpu/drm/radeon/r100.c idx, reg); idx 1776 drivers/gpu/drm/radeon/r100.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1886 drivers/gpu/drm/radeon/r100.c pr_err("Forbidden register 0x%04X in cs at %d\n", reg, idx); idx 1896 drivers/gpu/drm/radeon/r100.c unsigned idx; idx 1898 drivers/gpu/drm/radeon/r100.c idx = pkt->idx + 1; idx 1899 drivers/gpu/drm/radeon/r100.c value = radeon_get_ib_value(p, idx + 2); idx 1915 drivers/gpu/drm/radeon/r100.c unsigned idx; idx 1920 drivers/gpu/drm/radeon/r100.c idx = pkt->idx + 1; idx 1924 drivers/gpu/drm/radeon/r100.c r = r100_packet3_load_vbpntr(p, pkt, idx); idx 1935 drivers/gpu/drm/radeon/r100.c ib[idx+1] = radeon_get_ib_value(p, idx+1) + ((u32)reloc->gpu_offset); idx 1949 drivers/gpu/drm/radeon/r100.c ib[idx] = radeon_get_ib_value(p, idx) + ((u32)reloc->gpu_offset); idx 1951 drivers/gpu/drm/radeon/r100.c track->vtx_size = r100_get_vtx_size(radeon_get_ib_value(p, idx + 2)); idx 1956 drivers/gpu/drm/radeon/r100.c track->max_indx = radeon_get_ib_value(p, idx+1); idx 1958 drivers/gpu/drm/radeon/r100.c track->vap_vf_cntl = radeon_get_ib_value(p, idx+3); idx 1965 drivers/gpu/drm/radeon/r100.c if (((radeon_get_ib_value(p, idx + 1) >> 4) & 0x3) != 3) { idx 1969 drivers/gpu/drm/radeon/r100.c track->vtx_size = r100_get_vtx_size(radeon_get_ib_value(p, idx + 0)); idx 1970 drivers/gpu/drm/radeon/r100.c track->vap_vf_cntl = radeon_get_ib_value(p, idx + 1); idx 1978 drivers/gpu/drm/radeon/r100.c if (((radeon_get_ib_value(p, idx) >> 4) & 0x3) != 3) { idx 1982 drivers/gpu/drm/radeon/r100.c track->vap_vf_cntl = radeon_get_ib_value(p, idx); idx 1990 drivers/gpu/drm/radeon/r100.c track->vap_vf_cntl = radeon_get_ib_value(p, idx); idx 1997 drivers/gpu/drm/radeon/r100.c track->vap_vf_cntl = radeon_get_ib_value(p, idx); idx 2004 drivers/gpu/drm/radeon/r100.c track->vap_vf_cntl = radeon_get_ib_value(p, idx + 1); idx 2011 drivers/gpu/drm/radeon/r100.c track->vap_vf_cntl = radeon_get_ib_value(p, idx + 1); idx 2043 drivers/gpu/drm/radeon/r100.c r = radeon_cs_packet_parse(p, &pkt, p->idx); idx 2047 drivers/gpu/drm/radeon/r100.c p->idx += pkt.count + 2; idx 2073 drivers/gpu/drm/radeon/r100.c } while (p->idx < p->chunk_ib->length_dw); idx 2125 drivers/gpu/drm/radeon/r100.c struct r100_cs_track *track, unsigned idx) idx 2130 drivers/gpu/drm/radeon/r100.c unsigned compress_format = track->textures[idx].compress_format; idx 2133 drivers/gpu/drm/radeon/r100.c cube_robj = track->textures[idx].cube_info[face].robj; idx 2134 drivers/gpu/drm/radeon/r100.c w = track->textures[idx].cube_info[face].width; idx 2135 drivers/gpu/drm/radeon/r100.c h = track->textures[idx].cube_info[face].height; idx 2141 drivers/gpu/drm/radeon/r100.c size *= track->textures[idx].cpp; idx 2143 drivers/gpu/drm/radeon/r100.c size += track->textures[idx].cube_info[face].offset; idx 2148 drivers/gpu/drm/radeon/r100.c r100_cs_track_texture_print(&track->textures[idx]); idx 92 drivers/gpu/drm/radeon/r100_track.h unsigned idx, unsigned reg); idx 96 drivers/gpu/drm/radeon/r100_track.h unsigned idx, idx 100 drivers/gpu/drm/radeon/r100_track.h int idx); idx 147 drivers/gpu/drm/radeon/r200.c unsigned idx, unsigned reg) idx 161 drivers/gpu/drm/radeon/r200.c idx_value = radeon_get_ib_value(p, idx); idx 167 drivers/gpu/drm/radeon/r200.c idx, reg); idx 176 drivers/gpu/drm/radeon/r200.c r = r100_reloc_pitch_offset(p, pkt, idx, reg); idx 184 drivers/gpu/drm/radeon/r200.c idx, reg); idx 191 drivers/gpu/drm/radeon/r200.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 197 drivers/gpu/drm/radeon/r200.c idx, reg); idx 204 drivers/gpu/drm/radeon/r200.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 216 drivers/gpu/drm/radeon/r200.c idx, reg); idx 228 drivers/gpu/drm/radeon/r200.c ib[idx] = tmp + ((u32)reloc->gpu_offset); idx 230 drivers/gpu/drm/radeon/r200.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 269 drivers/gpu/drm/radeon/r200.c idx, reg); idx 274 drivers/gpu/drm/radeon/r200.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 287 drivers/gpu/drm/radeon/r200.c idx, reg); idx 300 drivers/gpu/drm/radeon/r200.c ib[idx] = tmp; idx 302 drivers/gpu/drm/radeon/r200.c ib[idx] = idx_value; idx 364 drivers/gpu/drm/radeon/r200.c idx, reg); idx 368 drivers/gpu/drm/radeon/r200.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 540 drivers/gpu/drm/radeon/r200.c pr_err("Forbidden register 0x%04X in cs at %d\n", reg, idx); idx 633 drivers/gpu/drm/radeon/r300.c unsigned idx, unsigned reg) idx 645 drivers/gpu/drm/radeon/r300.c idx_value = radeon_get_ib_value(p, idx); idx 653 drivers/gpu/drm/radeon/r300.c idx, reg); idx 660 drivers/gpu/drm/radeon/r300.c r = r100_reloc_pitch_offset(p, pkt, idx, reg); idx 672 drivers/gpu/drm/radeon/r300.c idx, reg); idx 679 drivers/gpu/drm/radeon/r300.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 685 drivers/gpu/drm/radeon/r300.c idx, reg); idx 692 drivers/gpu/drm/radeon/r300.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 714 drivers/gpu/drm/radeon/r300.c idx, reg); idx 720 drivers/gpu/drm/radeon/r300.c ib[idx] = (idx_value & 31) | /* keep the 1st 5 bits */ idx 732 drivers/gpu/drm/radeon/r300.c ib[idx] = tmp; idx 787 drivers/gpu/drm/radeon/r300.c idx, reg); idx 801 drivers/gpu/drm/radeon/r300.c ib[idx] = tmp; idx 872 drivers/gpu/drm/radeon/r300.c idx, reg); idx 886 drivers/gpu/drm/radeon/r300.c ib[idx] = tmp; idx 1087 drivers/gpu/drm/radeon/r300.c idx, reg); idx 1091 drivers/gpu/drm/radeon/r300.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1104 drivers/gpu/drm/radeon/r300.c ib[idx] = idx_value & ~1; idx 1129 drivers/gpu/drm/radeon/r300.c idx, reg); idx 1136 drivers/gpu/drm/radeon/r300.c ib[idx] = idx_value + ((u32)reloc->gpu_offset); idx 1172 drivers/gpu/drm/radeon/r300.c reg, idx, idx_value); idx 1182 drivers/gpu/drm/radeon/r300.c unsigned idx; idx 1186 drivers/gpu/drm/radeon/r300.c idx = pkt->idx + 1; idx 1190 drivers/gpu/drm/radeon/r300.c r = r100_packet3_load_vbpntr(p, pkt, idx); idx 1201 drivers/gpu/drm/radeon/r300.c ib[idx+1] = radeon_get_ib_value(p, idx + 1) + ((u32)reloc->gpu_offset); idx 1212 drivers/gpu/drm/radeon/r300.c if (((radeon_get_ib_value(p, idx + 1) >> 4) & 0x3) != 3) { idx 1216 drivers/gpu/drm/radeon/r300.c track->vap_vf_cntl = radeon_get_ib_value(p, idx + 1); idx 1227 drivers/gpu/drm/radeon/r300.c if (((radeon_get_ib_value(p, idx) >> 4) & 0x3) != 3) { idx 1231 drivers/gpu/drm/radeon/r300.c track->vap_vf_cntl = radeon_get_ib_value(p, idx); idx 1239 drivers/gpu/drm/radeon/r300.c track->vap_vf_cntl = radeon_get_ib_value(p, idx + 1); idx 1246 drivers/gpu/drm/radeon/r300.c track->vap_vf_cntl = radeon_get_ib_value(p, idx); idx 1253 drivers/gpu/drm/radeon/r300.c track->vap_vf_cntl = radeon_get_ib_value(p, idx + 1); idx 1260 drivers/gpu/drm/radeon/r300.c track->vap_vf_cntl = radeon_get_ib_value(p, idx); idx 1296 drivers/gpu/drm/radeon/r300.c r = radeon_cs_packet_parse(p, &pkt, p->idx); idx 1300 drivers/gpu/drm/radeon/r300.c p->idx += pkt.count + 2; idx 1320 drivers/gpu/drm/radeon/r300.c } while (p->idx < p->chunk_ib->length_dw); idx 650 drivers/gpu/drm/radeon/r600.c int idx; idx 734 drivers/gpu/drm/radeon/r600.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 0); idx 736 drivers/gpu/drm/radeon/r600.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0); idx 737 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_off_ps_idx = idx; idx 738 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_LOW_SH_IDX].dpms_on_ps_idx = idx; idx 742 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_off_ps_idx = idx; idx 743 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_MID_SH_IDX].dpms_on_ps_idx = idx; idx 747 drivers/gpu/drm/radeon/r600.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 0); idx 748 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_off_ps_idx = idx; idx 749 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_HIGH_SH_IDX].dpms_on_ps_idx = idx; idx 754 drivers/gpu/drm/radeon/r600.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_BATTERY, 1); idx 756 drivers/gpu/drm/radeon/r600.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 1); idx 757 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_off_ps_idx = idx; idx 758 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_LOW_MH_IDX].dpms_on_ps_idx = idx; idx 762 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_off_ps_idx = idx; idx 763 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_MID_MH_IDX].dpms_on_ps_idx = idx; idx 767 drivers/gpu/drm/radeon/r600.c idx = radeon_pm_get_type_index(rdev, POWER_STATE_TYPE_PERFORMANCE, 1); idx 768 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_off_ps_idx = idx; idx 769 drivers/gpu/drm/radeon/r600.c rdev->pm.profiles[PM_PROFILE_HIGH_MH_IDX].dpms_on_ps_idx = idx; idx 2837 drivers/gpu/drm/radeon/r600.c DRM_ERROR("radeon: cp failed to lock ring %d (%d).\n", ring->idx, r); idx 2852 drivers/gpu/drm/radeon/r600.c DRM_INFO("ring test on %d succeeded in %d usecs\n", ring->idx, i); idx 2855 drivers/gpu/drm/radeon/r600.c ring->idx, scratch, tmp); idx 2988 drivers/gpu/drm/radeon/r600.c radeon_sync_rings(rdev, &sync, ring->idx); idx 3014 drivers/gpu/drm/radeon/r600.c r = radeon_fence_emit(rdev, &fence, ring->idx); idx 3413 drivers/gpu/drm/radeon/r600.c r = radeon_ib_get(rdev, ring->idx, &ib, NULL, 256); idx 840 drivers/gpu/drm/radeon/r600_cs.c r = radeon_cs_packet_parse(p, &wait_reg_mem, p->idx); idx 851 drivers/gpu/drm/radeon/r600_cs.c wait_reg_mem_info = radeon_get_ib_value(p, wait_reg_mem.idx + 1); idx 867 drivers/gpu/drm/radeon/r600_cs.c if ((radeon_get_ib_value(p, wait_reg_mem.idx + 2) << 2) != vline_status[0]) { idx 872 drivers/gpu/drm/radeon/r600_cs.c if (radeon_get_ib_value(p, wait_reg_mem.idx + 5) != RADEON_VLINE_STAT) { idx 878 drivers/gpu/drm/radeon/r600_cs.c r = radeon_cs_packet_parse(p, &p3reloc, p->idx + wait_reg_mem.count + 2); idx 882 drivers/gpu/drm/radeon/r600_cs.c h_idx = p->idx - 2; idx 883 drivers/gpu/drm/radeon/r600_cs.c p->idx += wait_reg_mem.count + 2; idx 884 drivers/gpu/drm/radeon/r600_cs.c p->idx += p3reloc.count + 2; idx 921 drivers/gpu/drm/radeon/r600_cs.c unsigned idx, unsigned reg) idx 930 drivers/gpu/drm/radeon/r600_cs.c idx, reg); idx 935 drivers/gpu/drm/radeon/r600_cs.c pr_err("Forbidden register 0x%04X in cs at %d\n", reg, idx); idx 945 drivers/gpu/drm/radeon/r600_cs.c unsigned idx; idx 948 drivers/gpu/drm/radeon/r600_cs.c idx = pkt->idx + 1; idx 950 drivers/gpu/drm/radeon/r600_cs.c for (i = 0; i <= pkt->count; i++, idx++, reg += 4) { idx 951 drivers/gpu/drm/radeon/r600_cs.c r = r600_packet0_check(p, pkt, idx, reg); idx 969 drivers/gpu/drm/radeon/r600_cs.c static int r600_cs_check_reg(struct radeon_cs_parser *p, u32 reg, u32 idx) idx 978 drivers/gpu/drm/radeon/r600_cs.c dev_warn(p->dev, "forbidden register 0x%08x at %d\n", reg, idx); idx 1023 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1026 drivers/gpu/drm/radeon/r600_cs.c track->sq_config = radeon_get_ib_value(p, idx); idx 1029 drivers/gpu/drm/radeon/r600_cs.c track->db_depth_control = radeon_get_ib_value(p, idx); idx 1041 drivers/gpu/drm/radeon/r600_cs.c track->db_depth_info = radeon_get_ib_value(p, idx); idx 1042 drivers/gpu/drm/radeon/r600_cs.c ib[idx] &= C_028010_ARRAY_MODE; idx 1045 drivers/gpu/drm/radeon/r600_cs.c ib[idx] |= S_028010_ARRAY_MODE(V_028010_ARRAY_2D_TILED_THIN1); idx 1048 drivers/gpu/drm/radeon/r600_cs.c ib[idx] |= S_028010_ARRAY_MODE(V_028010_ARRAY_1D_TILED_THIN1); idx 1052 drivers/gpu/drm/radeon/r600_cs.c track->db_depth_info = radeon_get_ib_value(p, idx); idx 1057 drivers/gpu/drm/radeon/r600_cs.c track->db_depth_view = radeon_get_ib_value(p, idx); idx 1061 drivers/gpu/drm/radeon/r600_cs.c track->db_depth_size = radeon_get_ib_value(p, idx); idx 1062 drivers/gpu/drm/radeon/r600_cs.c track->db_depth_size_idx = idx; idx 1066 drivers/gpu/drm/radeon/r600_cs.c track->vgt_strmout_en = radeon_get_ib_value(p, idx); idx 1070 drivers/gpu/drm/radeon/r600_cs.c track->vgt_strmout_buffer_en = radeon_get_ib_value(p, idx); idx 1084 drivers/gpu/drm/radeon/r600_cs.c track->vgt_strmout_bo_offset[tmp] = radeon_get_ib_value(p, idx) << 8; idx 1085 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1096 drivers/gpu/drm/radeon/r600_cs.c track->vgt_strmout_size[tmp] = radeon_get_ib_value(p, idx) * 4; idx 1106 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1109 drivers/gpu/drm/radeon/r600_cs.c track->cb_target_mask = radeon_get_ib_value(p, idx); idx 1113 drivers/gpu/drm/radeon/r600_cs.c track->cb_shader_mask = radeon_get_ib_value(p, idx); idx 1116 drivers/gpu/drm/radeon/r600_cs.c tmp = G_028C04_MSAA_NUM_SAMPLES(radeon_get_ib_value(p, idx)); idx 1122 drivers/gpu/drm/radeon/r600_cs.c tmp = G_028808_SPECIAL_OP(radeon_get_ib_value(p, idx)); idx 1142 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_info[tmp] = radeon_get_ib_value(p, idx); idx 1144 drivers/gpu/drm/radeon/r600_cs.c ib[idx] |= S_0280A0_ARRAY_MODE(V_0280A0_ARRAY_2D_TILED_THIN1); idx 1147 drivers/gpu/drm/radeon/r600_cs.c ib[idx] |= S_0280A0_ARRAY_MODE(V_0280A0_ARRAY_1D_TILED_THIN1); idx 1152 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_info[tmp] = radeon_get_ib_value(p, idx); idx 1165 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_view[tmp] = radeon_get_ib_value(p, idx); idx 1177 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_size[tmp] = radeon_get_ib_value(p, idx); idx 1178 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_size_idx[tmp] = idx; idx 1206 drivers/gpu/drm/radeon/r600_cs.c ib[idx] = track->cb_color_base_last[tmp]; idx 1214 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_frag_offset[tmp] = (u64)ib[idx] << 8; idx 1215 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1237 drivers/gpu/drm/radeon/r600_cs.c ib[idx] = track->cb_color_base_last[tmp]; idx 1245 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_tile_offset[tmp] = (u64)ib[idx] << 8; idx 1246 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1261 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_mask[tmp] = radeon_get_ib_value(p, idx); idx 1281 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_bo_offset[tmp] = radeon_get_ib_value(p, idx) << 8; idx 1282 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1283 drivers/gpu/drm/radeon/r600_cs.c track->cb_color_base_last[tmp] = ib[idx]; idx 1295 drivers/gpu/drm/radeon/r600_cs.c track->db_offset = radeon_get_ib_value(p, idx) << 8; idx 1296 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1308 drivers/gpu/drm/radeon/r600_cs.c track->htile_offset = radeon_get_ib_value(p, idx) << 8; idx 1309 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1314 drivers/gpu/drm/radeon/r600_cs.c track->htile_surface = radeon_get_ib_value(p, idx); idx 1316 drivers/gpu/drm/radeon/r600_cs.c ib[idx] |= 3; idx 1378 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1387 drivers/gpu/drm/radeon/r600_cs.c ib[idx] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1390 drivers/gpu/drm/radeon/r600_cs.c track->sx_misc_kill_all_prims = (radeon_get_ib_value(p, idx) & 0x1) != 0; idx 1393 drivers/gpu/drm/radeon/r600_cs.c dev_warn(p->dev, "forbidden register 0x%08x at %d\n", reg, idx); idx 1469 drivers/gpu/drm/radeon/r600_cs.c static int r600_check_texture_resource(struct radeon_cs_parser *p, u32 idx, idx 1494 drivers/gpu/drm/radeon/r600_cs.c word0 = radeon_get_ib_value(p, idx + 0); idx 1501 drivers/gpu/drm/radeon/r600_cs.c word1 = radeon_get_ib_value(p, idx + 1); idx 1502 drivers/gpu/drm/radeon/r600_cs.c word2 = radeon_get_ib_value(p, idx + 2) << 8; idx 1503 drivers/gpu/drm/radeon/r600_cs.c word3 = radeon_get_ib_value(p, idx + 3) << 8; idx 1504 drivers/gpu/drm/radeon/r600_cs.c word4 = radeon_get_ib_value(p, idx + 4); idx 1505 drivers/gpu/drm/radeon/r600_cs.c word5 = radeon_get_ib_value(p, idx + 5); idx 1610 drivers/gpu/drm/radeon/r600_cs.c static bool r600_is_safe_reg(struct radeon_cs_parser *p, u32 reg, u32 idx) idx 1616 drivers/gpu/drm/radeon/r600_cs.c dev_warn(p->dev, "forbidden register 0x%08x at %d\n", reg, idx); idx 1622 drivers/gpu/drm/radeon/r600_cs.c dev_warn(p->dev, "forbidden register 0x%08x at %d\n", reg, idx); idx 1632 drivers/gpu/drm/radeon/r600_cs.c unsigned idx; idx 1640 drivers/gpu/drm/radeon/r600_cs.c idx = pkt->idx + 1; idx 1641 drivers/gpu/drm/radeon/r600_cs.c idx_value = radeon_get_ib_value(p, idx); idx 1655 drivers/gpu/drm/radeon/r600_cs.c tmp = radeon_get_ib_value(p, idx + 1); idx 1677 drivers/gpu/drm/radeon/r600_cs.c ib[idx + 0] = offset; idx 1678 drivers/gpu/drm/radeon/r600_cs.c ib[idx + 1] = (tmp & 0xffffff00) | (upper_32_bits(offset) & 0xff); idx 1716 drivers/gpu/drm/radeon/r600_cs.c ((u64)(radeon_get_ib_value(p, idx+1) & 0xff) << 32); idx 1718 drivers/gpu/drm/radeon/r600_cs.c ib[idx+0] = offset; idx 1719 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = upper_32_bits(offset) & 0xff; idx 1735 drivers/gpu/drm/radeon/r600_cs.c dev_warn(p->dev, "%s:%d invalid cmd stream %d\n", __func__, __LINE__, idx); idx 1767 drivers/gpu/drm/radeon/r600_cs.c (radeon_get_ib_value(p, idx+1) & 0xfffffff0) + idx 1768 drivers/gpu/drm/radeon/r600_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 1770 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = (ib[idx+1] & 0x3) | (offset & 0xfffffff0); idx 1771 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 1785 drivers/gpu/drm/radeon/r600_cs.c command = radeon_get_ib_value(p, idx+4); idx 1803 drivers/gpu/drm/radeon/r600_cs.c tmp = radeon_get_ib_value(p, idx) + idx 1804 drivers/gpu/drm/radeon/r600_cs.c ((u64)(radeon_get_ib_value(p, idx+1) & 0xff) << 32); idx 1814 drivers/gpu/drm/radeon/r600_cs.c ib[idx] = offset; idx 1815 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = (ib[idx+1] & 0xffffff00) | (upper_32_bits(offset) & 0xff); idx 1833 drivers/gpu/drm/radeon/r600_cs.c tmp = radeon_get_ib_value(p, idx+2) + idx 1834 drivers/gpu/drm/radeon/r600_cs.c ((u64)(radeon_get_ib_value(p, idx+3) & 0xff) << 32); idx 1844 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] = offset; idx 1845 drivers/gpu/drm/radeon/r600_cs.c ib[idx+3] = upper_32_bits(offset) & 0xff; idx 1855 drivers/gpu/drm/radeon/r600_cs.c if (radeon_get_ib_value(p, idx + 1) != 0xffffffff || idx 1856 drivers/gpu/drm/radeon/r600_cs.c radeon_get_ib_value(p, idx + 2) != 0) { idx 1862 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 1879 drivers/gpu/drm/radeon/r600_cs.c (radeon_get_ib_value(p, idx+1) & 0xfffffff8) + idx 1880 drivers/gpu/drm/radeon/r600_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 1882 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = offset & 0xfffffff8; idx 1883 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 1901 drivers/gpu/drm/radeon/r600_cs.c (radeon_get_ib_value(p, idx+1) & 0xfffffffc) + idx 1902 drivers/gpu/drm/radeon/r600_cs.c ((u64)(radeon_get_ib_value(p, idx+2) & 0xff) << 32); idx 1904 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = offset & 0xfffffffc; idx 1905 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] = (ib[idx+2] & 0xffffff00) | (upper_32_bits(offset) & 0xff); idx 1919 drivers/gpu/drm/radeon/r600_cs.c r = r600_cs_check_reg(p, reg, idx+1+i); idx 1935 drivers/gpu/drm/radeon/r600_cs.c r = r600_cs_check_reg(p, reg, idx+1+i); idx 1957 drivers/gpu/drm/radeon/r600_cs.c switch (G__SQ_VTX_CONSTANT_TYPE(radeon_get_ib_value(p, idx+(i*7)+6+1))) { idx 1968 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1+(i*7)+0] |= S_038000_TILE_MODE(V_038000_ARRAY_2D_TILED_THIN1); idx 1970 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1+(i*7)+0] |= S_038000_TILE_MODE(V_038000_ARRAY_1D_TILED_THIN1); idx 1981 drivers/gpu/drm/radeon/r600_cs.c r = r600_check_texture_resource(p, idx+(i*7)+1, idx 1983 drivers/gpu/drm/radeon/r600_cs.c base_offset + radeon_get_ib_value(p, idx+1+(i*7)+2), idx 1984 drivers/gpu/drm/radeon/r600_cs.c mip_offset + radeon_get_ib_value(p, idx+1+(i*7)+3), idx 1988 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1+(i*7)+2] += base_offset; idx 1989 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1+(i*7)+3] += mip_offset; idx 2000 drivers/gpu/drm/radeon/r600_cs.c offset = radeon_get_ib_value(p, idx+1+(i*7)+0); idx 2001 drivers/gpu/drm/radeon/r600_cs.c size = radeon_get_ib_value(p, idx+1+(i*7)+1) + 1; idx 2006 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1+(i*7)+1] = radeon_bo_size(reloc->robj) - offset; idx 2010 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1+(i*8)+0] = offset64; idx 2011 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1+(i*8)+2] = (ib[idx+1+(i*8)+2] & 0xffffff00) | idx 2107 drivers/gpu/drm/radeon/r600_cs.c offset = radeon_get_ib_value(p, idx+1) << 8; idx 2119 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)((reloc->gpu_offset >> 8) & 0xffffffff); idx 2145 drivers/gpu/drm/radeon/r600_cs.c offset = radeon_get_ib_value(p, idx+1); idx 2146 drivers/gpu/drm/radeon/r600_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+2) & 0xff)) << 32; idx 2153 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = offset; idx 2154 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 2164 drivers/gpu/drm/radeon/r600_cs.c offset = radeon_get_ib_value(p, idx+3); idx 2165 drivers/gpu/drm/radeon/r600_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2172 drivers/gpu/drm/radeon/r600_cs.c ib[idx+3] = offset; idx 2173 drivers/gpu/drm/radeon/r600_cs.c ib[idx+4] = upper_32_bits(offset) & 0xff; idx 2189 drivers/gpu/drm/radeon/r600_cs.c offset = radeon_get_ib_value(p, idx+0); idx 2190 drivers/gpu/drm/radeon/r600_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+1) & 0xff)) << 32UL; idx 2201 drivers/gpu/drm/radeon/r600_cs.c ib[idx+0] = offset; idx 2202 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = upper_32_bits(offset) & 0xff; idx 2218 drivers/gpu/drm/radeon/r600_cs.c offset = radeon_get_ib_value(p, idx+1); idx 2219 drivers/gpu/drm/radeon/r600_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+2) & 0xff)) << 32; idx 2226 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] = offset; idx 2227 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] = upper_32_bits(offset) & 0xff; idx 2230 drivers/gpu/drm/radeon/r600_cs.c reg = radeon_get_ib_value(p, idx+1) << 2; idx 2231 drivers/gpu/drm/radeon/r600_cs.c if (!r600_is_safe_reg(p, reg, idx+1)) idx 2242 drivers/gpu/drm/radeon/r600_cs.c offset = radeon_get_ib_value(p, idx+3); idx 2243 drivers/gpu/drm/radeon/r600_cs.c offset += ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2250 drivers/gpu/drm/radeon/r600_cs.c ib[idx+3] = offset; idx 2251 drivers/gpu/drm/radeon/r600_cs.c ib[idx+4] = upper_32_bits(offset) & 0xff; idx 2254 drivers/gpu/drm/radeon/r600_cs.c reg = radeon_get_ib_value(p, idx+3) << 2; idx 2255 drivers/gpu/drm/radeon/r600_cs.c if (!r600_is_safe_reg(p, reg, idx+3)) idx 2292 drivers/gpu/drm/radeon/r600_cs.c r = radeon_cs_packet_parse(p, &pkt, p->idx); idx 2298 drivers/gpu/drm/radeon/r600_cs.c p->idx += pkt.count + 2; idx 2319 drivers/gpu/drm/radeon/r600_cs.c } while (p->idx < p->chunk_ib->length_dw); idx 2346 drivers/gpu/drm/radeon/r600_cs.c unsigned idx; idx 2354 drivers/gpu/drm/radeon/r600_cs.c idx = p->dma_reloc_idx; idx 2355 drivers/gpu/drm/radeon/r600_cs.c if (idx >= p->nrelocs) { idx 2357 drivers/gpu/drm/radeon/r600_cs.c idx, p->nrelocs); idx 2360 drivers/gpu/drm/radeon/r600_cs.c *cs_reloc = &p->relocs[idx]; idx 2384 drivers/gpu/drm/radeon/r600_cs.c u32 idx, idx_value; idx 2389 drivers/gpu/drm/radeon/r600_cs.c if (p->idx >= ib_chunk->length_dw) { idx 2391 drivers/gpu/drm/radeon/r600_cs.c p->idx, ib_chunk->length_dw); idx 2394 drivers/gpu/drm/radeon/r600_cs.c idx = p->idx; idx 2395 drivers/gpu/drm/radeon/r600_cs.c header = radeon_get_ib_value(p, idx); idx 2408 drivers/gpu/drm/radeon/r600_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2411 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 2412 drivers/gpu/drm/radeon/r600_cs.c p->idx += count + 5; idx 2414 drivers/gpu/drm/radeon/r600_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2415 drivers/gpu/drm/radeon/r600_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+2) & 0xff)) << 32; idx 2417 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2418 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2419 drivers/gpu/drm/radeon/r600_cs.c p->idx += count + 3; idx 2439 drivers/gpu/drm/radeon/r600_cs.c idx_value = radeon_get_ib_value(p, idx + 2); idx 2443 drivers/gpu/drm/radeon/r600_cs.c src_offset = radeon_get_ib_value(p, idx+1); idx 2445 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)(src_reloc->gpu_offset >> 8); idx 2447 drivers/gpu/drm/radeon/r600_cs.c dst_offset = radeon_get_ib_value(p, idx+5); idx 2448 drivers/gpu/drm/radeon/r600_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+6) & 0xff)) << 32; idx 2449 drivers/gpu/drm/radeon/r600_cs.c ib[idx+5] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2450 drivers/gpu/drm/radeon/r600_cs.c ib[idx+6] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2453 drivers/gpu/drm/radeon/r600_cs.c src_offset = radeon_get_ib_value(p, idx+5); idx 2454 drivers/gpu/drm/radeon/r600_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+6) & 0xff)) << 32; idx 2455 drivers/gpu/drm/radeon/r600_cs.c ib[idx+5] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 2456 drivers/gpu/drm/radeon/r600_cs.c ib[idx+6] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2458 drivers/gpu/drm/radeon/r600_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2460 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset >> 8); idx 2462 drivers/gpu/drm/radeon/r600_cs.c p->idx += 7; idx 2465 drivers/gpu/drm/radeon/r600_cs.c src_offset = radeon_get_ib_value(p, idx+2); idx 2466 drivers/gpu/drm/radeon/r600_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+4) & 0xff)) << 32; idx 2467 drivers/gpu/drm/radeon/r600_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2468 drivers/gpu/drm/radeon/r600_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32; idx 2470 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2471 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 2472 drivers/gpu/drm/radeon/r600_cs.c ib[idx+3] += upper_32_bits(dst_reloc->gpu_offset) & 0xff; idx 2473 drivers/gpu/drm/radeon/r600_cs.c ib[idx+4] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2474 drivers/gpu/drm/radeon/r600_cs.c p->idx += 5; idx 2476 drivers/gpu/drm/radeon/r600_cs.c src_offset = radeon_get_ib_value(p, idx+2); idx 2477 drivers/gpu/drm/radeon/r600_cs.c src_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff)) << 32; idx 2478 drivers/gpu/drm/radeon/r600_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2479 drivers/gpu/drm/radeon/r600_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0xff0000)) << 16; idx 2481 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2482 drivers/gpu/drm/radeon/r600_cs.c ib[idx+2] += (u32)(src_reloc->gpu_offset & 0xfffffffc); idx 2483 drivers/gpu/drm/radeon/r600_cs.c ib[idx+3] += upper_32_bits(src_reloc->gpu_offset) & 0xff; idx 2484 drivers/gpu/drm/radeon/r600_cs.c ib[idx+3] += (upper_32_bits(dst_reloc->gpu_offset) & 0xff) << 16; idx 2485 drivers/gpu/drm/radeon/r600_cs.c p->idx += 4; idx 2509 drivers/gpu/drm/radeon/r600_cs.c dst_offset = radeon_get_ib_value(p, idx+1); idx 2510 drivers/gpu/drm/radeon/r600_cs.c dst_offset |= ((u64)(radeon_get_ib_value(p, idx+3) & 0x00ff0000)) << 16; idx 2516 drivers/gpu/drm/radeon/r600_cs.c ib[idx+1] += (u32)(dst_reloc->gpu_offset & 0xfffffffc); idx 2517 drivers/gpu/drm/radeon/r600_cs.c ib[idx+3] += (upper_32_bits(dst_reloc->gpu_offset) << 16) & 0x00ff0000; idx 2518 drivers/gpu/drm/radeon/r600_cs.c p->idx += 4; idx 2521 drivers/gpu/drm/radeon/r600_cs.c p->idx += 1; idx 2524 drivers/gpu/drm/radeon/r600_cs.c DRM_ERROR("Unknown packet type %d at %d !\n", cmd, idx); idx 2527 drivers/gpu/drm/radeon/r600_cs.c } while (p->idx < p->chunk_ib->length_dw); idx 239 drivers/gpu/drm/radeon/r600_dma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 251 drivers/gpu/drm/radeon/r600_dma.c DRM_ERROR("radeon: dma failed to lock ring %d (%d).\n", ring->idx, r); idx 268 drivers/gpu/drm/radeon/r600_dma.c DRM_INFO("ring test on %d succeeded in %d usecs\n", ring->idx, i); idx 271 drivers/gpu/drm/radeon/r600_dma.c ring->idx, tmp); idx 346 drivers/gpu/drm/radeon/r600_dma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 353 drivers/gpu/drm/radeon/r600_dma.c r = radeon_ib_get(rdev, ring->idx, &ib, NULL, 256); idx 469 drivers/gpu/drm/radeon/r600_dma.c radeon_sync_rings(rdev, &sync, ring->idx); idx 485 drivers/gpu/drm/radeon/r600_dma.c r = radeon_fence_emit(rdev, &fence, ring->idx); idx 852 drivers/gpu/drm/radeon/radeon.h u32 idx; idx 1066 drivers/gpu/drm/radeon/radeon.h unsigned idx; idx 1089 drivers/gpu/drm/radeon/radeon.h static inline u32 radeon_get_ib_value(struct radeon_cs_parser *p, int idx) idx 1094 drivers/gpu/drm/radeon/radeon.h return ibc->kdata[idx]; idx 1095 drivers/gpu/drm/radeon/radeon.h return p->ib.ptr[idx]; idx 1100 drivers/gpu/drm/radeon/radeon.h unsigned idx; idx 1110 drivers/gpu/drm/radeon/radeon.h unsigned idx, unsigned reg); idx 2721 drivers/gpu/drm/radeon/radeon.h #define radeon_ring_vm_flush(rdev, r, vm_id, pd_addr) (rdev)->asic->ring[(r)->idx]->vm_flush((rdev), (r), (vm_id), (pd_addr)) idx 2722 drivers/gpu/drm/radeon/radeon.h #define radeon_ring_get_rptr(rdev, r) (rdev)->asic->ring[(r)->idx]->get_rptr((rdev), (r)) idx 2723 drivers/gpu/drm/radeon/radeon.h #define radeon_ring_get_wptr(rdev, r) (rdev)->asic->ring[(r)->idx]->get_wptr((rdev), (r)) idx 2724 drivers/gpu/drm/radeon/radeon.h #define radeon_ring_set_wptr(rdev, r) (rdev)->asic->ring[(r)->idx]->set_wptr((rdev), (r)) idx 2943 drivers/gpu/drm/radeon/radeon.h unsigned idx); idx 130 drivers/gpu/drm/radeon/radeon_asic.h unsigned idx); idx 285 drivers/gpu/drm/radeon/radeon_cs.c p->idx = 0; idx 734 drivers/gpu/drm/radeon/radeon_cs.c unsigned idx) idx 741 drivers/gpu/drm/radeon/radeon_cs.c if (idx >= ib_chunk->length_dw) { idx 743 drivers/gpu/drm/radeon/radeon_cs.c idx, ib_chunk->length_dw); idx 746 drivers/gpu/drm/radeon/radeon_cs.c header = radeon_get_ib_value(p, idx); idx 747 drivers/gpu/drm/radeon/radeon_cs.c pkt->idx = idx; idx 767 drivers/gpu/drm/radeon/radeon_cs.c DRM_ERROR("Unknown packet type %d at %d !\n", pkt->type, idx); idx 771 drivers/gpu/drm/radeon/radeon_cs.c if ((pkt->count + 1 + pkt->idx) >= ib_chunk->length_dw) { idx 773 drivers/gpu/drm/radeon/radeon_cs.c pkt->idx, pkt->type, pkt->count, ib_chunk->length_dw); idx 781 drivers/gpu/drm/radeon/radeon_cs.c if (i == idx) idx 800 drivers/gpu/drm/radeon/radeon_cs.c r = radeon_cs_packet_parse(p, &p3reloc, p->idx); idx 822 drivers/gpu/drm/radeon/radeon_cs.c unsigned idx; idx 825 drivers/gpu/drm/radeon/radeon_cs.c idx = pkt->idx; idx 826 drivers/gpu/drm/radeon/radeon_cs.c for (i = 0; i <= (pkt->count + 1); i++, idx++) idx 827 drivers/gpu/drm/radeon/radeon_cs.c DRM_INFO("ib[%d]=0x%08X\n", idx, ib[idx]); idx 847 drivers/gpu/drm/radeon/radeon_cs.c unsigned idx; idx 856 drivers/gpu/drm/radeon/radeon_cs.c r = radeon_cs_packet_parse(p, &p3reloc, p->idx); idx 859 drivers/gpu/drm/radeon/radeon_cs.c p->idx += p3reloc.count + 2; idx 863 drivers/gpu/drm/radeon/radeon_cs.c p3reloc.idx); idx 867 drivers/gpu/drm/radeon/radeon_cs.c idx = radeon_get_ib_value(p, p3reloc.idx + 1); idx 868 drivers/gpu/drm/radeon/radeon_cs.c if (idx >= relocs_chunk->length_dw) { idx 870 drivers/gpu/drm/radeon/radeon_cs.c idx, relocs_chunk->length_dw); idx 878 drivers/gpu/drm/radeon/radeon_cs.c (u64)relocs_chunk->kdata[idx + 3] << 32; idx 879 drivers/gpu/drm/radeon/radeon_cs.c (*cs_reloc)->gpu_offset |= relocs_chunk->kdata[idx + 0]; idx 881 drivers/gpu/drm/radeon/radeon_cs.c *cs_reloc = &p->relocs[(idx / 4)]; idx 1308 drivers/gpu/drm/radeon/radeon_device.c rdev->ring[i].idx = i; idx 119 drivers/gpu/drm/radeon/radeon_dp_mst.c int idx = 0; idx 141 drivers/gpu/drm/radeon/radeon_dp_mst.c new_attribs[idx].fe = mst_enc->fe; idx 142 drivers/gpu/drm/radeon/radeon_dp_mst.c new_attribs[idx].slots = drm_dp_mst_get_vcpi_slots(&mst_conn->mst_mgr, mst_enc->port); idx 143 drivers/gpu/drm/radeon/radeon_dp_mst.c idx++; idx 146 drivers/gpu/drm/radeon/radeon_dp_mst.c for (i = 0; i < idx; i++) { idx 155 drivers/gpu/drm/radeon/radeon_dp_mst.c for (i = idx; i < mst_conn->enabled_attribs; i++) { idx 160 drivers/gpu/drm/radeon/radeon_dp_mst.c mst_conn->enabled_attribs = idx; idx 64 drivers/gpu/drm/radeon/radeon_ring.c switch (ring->idx) { idx 124 drivers/gpu/drm/radeon/radeon_ring.c r = radeon_fence_wait_next(rdev, ring->idx); idx 174 drivers/gpu/drm/radeon/radeon_ring.c if (hdp_flush && rdev->asic->ring[ring->idx]->hdp_flush) idx 175 drivers/gpu/drm/radeon/radeon_ring.c rdev->asic->ring[ring->idx]->hdp_flush(rdev, ring); idx 266 drivers/gpu/drm/radeon/radeon_ring.c ring->idx, elapsed); idx 296 drivers/gpu/drm/radeon/radeon_ring.c if (!radeon_fence_count_emitted(rdev, ring->idx)) { idx 417 drivers/gpu/drm/radeon/radeon_ring.c u32 index = RADEON_WB_RING0_NEXT_RPTR + (ring->idx * 4); idx 266 drivers/gpu/drm/radeon/radeon_test.c uint32_t handle = ring->idx ^ 0xdeafbeef; idx 269 drivers/gpu/drm/radeon/radeon_test.c if (ring->idx == R600_RING_TYPE_UVD_INDEX) { idx 270 drivers/gpu/drm/radeon/radeon_test.c r = radeon_uvd_get_create_msg(rdev, ring->idx, handle, NULL); idx 276 drivers/gpu/drm/radeon/radeon_test.c r = radeon_uvd_get_destroy_msg(rdev, ring->idx, handle, fence); idx 282 drivers/gpu/drm/radeon/radeon_test.c } else if (ring->idx == TN_RING_TYPE_VCE1_INDEX || idx 283 drivers/gpu/drm/radeon/radeon_test.c ring->idx == TN_RING_TYPE_VCE2_INDEX) { idx 284 drivers/gpu/drm/radeon/radeon_test.c r = radeon_vce_get_create_msg(rdev, ring->idx, handle, NULL); idx 290 drivers/gpu/drm/radeon/radeon_test.c r = radeon_vce_get_destroy_msg(rdev, ring->idx, handle, fence); idx 299 drivers/gpu/drm/radeon/radeon_test.c DRM_ERROR("Failed to lock ring A %d\n", ring->idx); idx 302 drivers/gpu/drm/radeon/radeon_test.c r = radeon_fence_emit(rdev, fence, ring->idx); idx 329 drivers/gpu/drm/radeon/radeon_test.c DRM_ERROR("Failed to lock ring A %d\n", ringA->idx); idx 332 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_wait(rdev, ringA->idx, semaphore); idx 341 drivers/gpu/drm/radeon/radeon_test.c DRM_ERROR("Failed to lock ring A %d\n", ringA->idx); idx 344 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_wait(rdev, ringA->idx, semaphore); idx 363 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_signal(rdev, ringB->idx, semaphore); idx 384 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_signal(rdev, ringB->idx, semaphore); idx 424 drivers/gpu/drm/radeon/radeon_test.c DRM_ERROR("Failed to lock ring A %d\n", ringA->idx); idx 427 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_wait(rdev, ringA->idx, semaphore); idx 436 drivers/gpu/drm/radeon/radeon_test.c DRM_ERROR("Failed to lock ring B %d\n", ringB->idx); idx 439 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_wait(rdev, ringB->idx, semaphore); idx 461 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_signal(rdev, ringC->idx, semaphore); idx 487 drivers/gpu/drm/radeon/radeon_test.c radeon_semaphore_emit_signal(rdev, ringC->idx, semaphore); idx 519 drivers/gpu/drm/radeon/radeon_test.c if (ringA->idx == TN_RING_TYPE_VCE2_INDEX && idx 520 drivers/gpu/drm/radeon/radeon_test.c ringB->idx == TN_RING_TYPE_VCE1_INDEX) idx 578 drivers/gpu/drm/radeon/radeon_uvd.c unsigned idx, cmd, offset; idx 584 drivers/gpu/drm/radeon/radeon_uvd.c idx = radeon_get_ib_value(p, data1); idx 585 drivers/gpu/drm/radeon/radeon_uvd.c if (idx >= relocs_chunk->length_dw) { idx 587 drivers/gpu/drm/radeon/radeon_uvd.c idx, relocs_chunk->length_dw); idx 591 drivers/gpu/drm/radeon/radeon_uvd.c reloc = &p->relocs[(idx / 4)]; idx 599 drivers/gpu/drm/radeon/radeon_uvd.c cmd = radeon_get_ib_value(p, p->idx) >> 1; idx 656 drivers/gpu/drm/radeon/radeon_uvd.c p->idx++; idx 660 drivers/gpu/drm/radeon/radeon_uvd.c *data0 = p->idx; idx 663 drivers/gpu/drm/radeon/radeon_uvd.c *data1 = p->idx; idx 679 drivers/gpu/drm/radeon/radeon_uvd.c p->idx++; idx 713 drivers/gpu/drm/radeon/radeon_uvd.c r = radeon_cs_packet_parse(p, &pkt, p->idx); idx 724 drivers/gpu/drm/radeon/radeon_uvd.c p->idx += pkt.count + 2; idx 730 drivers/gpu/drm/radeon/radeon_uvd.c } while (p->idx < p->chunk_ib->length_dw); idx 476 drivers/gpu/drm/radeon/radeon_vce.c unsigned idx; idx 480 drivers/gpu/drm/radeon/radeon_vce.c idx = radeon_get_ib_value(p, hi); idx 482 drivers/gpu/drm/radeon/radeon_vce.c if (idx >= relocs_chunk->length_dw) { idx 484 drivers/gpu/drm/radeon/radeon_vce.c idx, relocs_chunk->length_dw); idx 488 drivers/gpu/drm/radeon/radeon_vce.c reloc = &p->relocs[(idx / 4)]; idx 565 drivers/gpu/drm/radeon/radeon_vce.c while (p->idx < p->chunk_ib->length_dw) { idx 566 drivers/gpu/drm/radeon/radeon_vce.c uint32_t len = radeon_get_ib_value(p, p->idx); idx 567 drivers/gpu/drm/radeon/radeon_vce.c uint32_t cmd = radeon_get_ib_value(p, p->idx + 1); idx 583 drivers/gpu/drm/radeon/radeon_vce.c handle = radeon_get_ib_value(p, p->idx + 2); idx 602 drivers/gpu/drm/radeon/radeon_vce.c *size = radeon_get_ib_value(p, p->idx + 8) * idx 603 drivers/gpu/drm/radeon/radeon_vce.c radeon_get_ib_value(p, p->idx + 10) * idx 616 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_vce_cs_reloc(p, p->idx + 10, p->idx + 9, idx 621 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_vce_cs_reloc(p, p->idx + 12, p->idx + 11, idx 632 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_vce_cs_reloc(p, p->idx + 3, p->idx + 2, idx 639 drivers/gpu/drm/radeon/radeon_vce.c tmp = radeon_get_ib_value(p, p->idx + 4); idx 640 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_vce_cs_reloc(p, p->idx + 3, p->idx + 2, idx 647 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_vce_cs_reloc(p, p->idx + 3, p->idx + 2, idx 665 drivers/gpu/drm/radeon/radeon_vce.c p->idx += len / 4; idx 765 drivers/gpu/drm/radeon/radeon_vce.c ring->idx, r); idx 779 drivers/gpu/drm/radeon/radeon_vce.c ring->idx, i); idx 782 drivers/gpu/drm/radeon/radeon_vce.c ring->idx); idx 801 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_vce_get_create_msg(rdev, ring->idx, 1, NULL); idx 807 drivers/gpu/drm/radeon/radeon_vce.c r = radeon_vce_get_destroy_msg(rdev, ring->idx, 1, &fence); idx 821 drivers/gpu/drm/radeon/radeon_vce.c DRM_INFO("ib test on ring %d succeeded\n", ring->idx); idx 133 drivers/gpu/drm/radeon/radeon_vm.c unsigned i, idx; idx 149 drivers/gpu/drm/radeon/radeon_vm.c for (i = 0, idx = 1; i <= vm->max_pde_used; i++) { idx 153 drivers/gpu/drm/radeon/radeon_vm.c list[idx].robj = vm->page_tables[i].bo; idx 154 drivers/gpu/drm/radeon/radeon_vm.c list[idx].preferred_domains = RADEON_GEM_DOMAIN_VRAM; idx 155 drivers/gpu/drm/radeon/radeon_vm.c list[idx].allowed_domains = RADEON_GEM_DOMAIN_VRAM; idx 156 drivers/gpu/drm/radeon/radeon_vm.c list[idx].tv.bo = &list[idx].robj->tbo; idx 157 drivers/gpu/drm/radeon/radeon_vm.c list[idx].tv.num_shared = 1; idx 158 drivers/gpu/drm/radeon/radeon_vm.c list[idx].tiling_flags = 0; idx 159 drivers/gpu/drm/radeon/radeon_vm.c list_add(&list[idx++].tv.head, head); idx 1907 drivers/gpu/drm/radeon/rv6xx_dpm.c u8 *idx; idx 1916 drivers/gpu/drm/radeon/rv6xx_dpm.c idx = (u8 *)&power_state->v1.ucClockStateIndices[0]; idx 1921 drivers/gpu/drm/radeon/rv6xx_dpm.c (idx[j] * power_info->pplib.ucClockInfoSize)); idx 67 drivers/gpu/drm/radeon/rv770_dma.c radeon_sync_rings(rdev, &sync, ring->idx); idx 83 drivers/gpu/drm/radeon/rv770_dma.c r = radeon_fence_emit(rdev, &fence, ring->idx); idx 2301 drivers/gpu/drm/radeon/rv770_dpm.c u8 *idx; idx 2311 drivers/gpu/drm/radeon/rv770_dpm.c idx = (u8 *)&power_state->v1.ucClockStateIndices[0]; idx 2316 drivers/gpu/drm/radeon/rv770_dpm.c (idx[j] * power_info->pplib.ucClockInfoSize)); idx 4479 drivers/gpu/drm/radeon/si.c static int si_vm_packet3_cp_dma_check(u32 *ib, u32 idx) idx 4482 drivers/gpu/drm/radeon/si.c u32 command = ib[idx + 4]; idx 4483 drivers/gpu/drm/radeon/si.c u32 info = ib[idx + 1]; idx 4484 drivers/gpu/drm/radeon/si.c u32 idx_value = ib[idx]; idx 4509 drivers/gpu/drm/radeon/si.c start_reg = ib[idx + 2]; idx 4534 drivers/gpu/drm/radeon/si.c u32 idx = pkt->idx + 1; idx 4535 drivers/gpu/drm/radeon/si.c u32 idx_value = ib[idx]; idx 4587 drivers/gpu/drm/radeon/si.c reg = ib[idx + 3] * 4; idx 4594 drivers/gpu/drm/radeon/si.c start_reg = ib[idx + 1] * 4; idx 4609 drivers/gpu/drm/radeon/si.c reg = ib[idx + 5] * 4; idx 4616 drivers/gpu/drm/radeon/si.c reg = ib[idx + 3] * 4; idx 4637 drivers/gpu/drm/radeon/si.c r = si_vm_packet3_cp_dma_check(ib, idx); idx 4652 drivers/gpu/drm/radeon/si.c u32 idx = pkt->idx + 1; idx 4653 drivers/gpu/drm/radeon/si.c u32 idx_value = ib[idx]; idx 4690 drivers/gpu/drm/radeon/si.c reg = ib[idx + 3] * 4; idx 4697 drivers/gpu/drm/radeon/si.c start_reg = ib[idx + 1] * 4; idx 4712 drivers/gpu/drm/radeon/si.c reg = ib[idx + 5] * 4; idx 4719 drivers/gpu/drm/radeon/si.c reg = ib[idx + 3] * 4; idx 4725 drivers/gpu/drm/radeon/si.c r = si_vm_packet3_cp_dma_check(ib, idx); idx 4739 drivers/gpu/drm/radeon/si.c u32 idx = 0, i; idx 4743 drivers/gpu/drm/radeon/si.c pkt.idx = idx; idx 4744 drivers/gpu/drm/radeon/si.c pkt.type = RADEON_CP_PACKET_GET_TYPE(ib->ptr[idx]); idx 4745 drivers/gpu/drm/radeon/si.c pkt.count = RADEON_CP_PACKET_GET_COUNT(ib->ptr[idx]); idx 4753 drivers/gpu/drm/radeon/si.c idx += 1; idx 4756 drivers/gpu/drm/radeon/si.c pkt.opcode = RADEON_CP_PACKET3_GET_OPCODE(ib->ptr[idx]); idx 4774 drivers/gpu/drm/radeon/si.c idx += pkt.count + 2; idx 4783 drivers/gpu/drm/radeon/si.c if (i == idx) idx 4790 drivers/gpu/drm/radeon/si.c } while (idx < ib->length_dw); idx 46 drivers/gpu/drm/radeon/si_dma.c if (ring->idx == R600_RING_TYPE_DMA_INDEX) idx 256 drivers/gpu/drm/radeon/si_dma.c radeon_sync_rings(rdev, &sync, ring->idx); idx 272 drivers/gpu/drm/radeon/si_dma.c r = radeon_fence_emit(rdev, &fence, ring->idx); idx 6847 drivers/gpu/drm/radeon/si_dpm.c u8 *idx; idx 6864 drivers/gpu/drm/radeon/si_dpm.c idx = (u8 *)&power_state->v2.clockInfoIndex[0]; idx 6866 drivers/gpu/drm/radeon/si_dpm.c clock_array_index = idx[j]; idx 1491 drivers/gpu/drm/radeon/sumo_dpm.c u8 *idx; idx 1505 drivers/gpu/drm/radeon/sumo_dpm.c idx = (u8 *)&power_state->v2.clockInfoIndex[0]; idx 1507 drivers/gpu/drm/radeon/sumo_dpm.c clock_array_index = idx[j]; idx 1769 drivers/gpu/drm/radeon/trinity_dpm.c u8 *idx; idx 1783 drivers/gpu/drm/radeon/trinity_dpm.c idx = (u8 *)&power_state->v2.clockInfoIndex[0]; idx 1785 drivers/gpu/drm/radeon/trinity_dpm.c clock_array_index = idx[j]; idx 431 drivers/gpu/drm/radeon/uvd_v1_0.c ring->idx, r); idx 446 drivers/gpu/drm/radeon/uvd_v1_0.c ring->idx, i); idx 449 drivers/gpu/drm/radeon/uvd_v1_0.c ring->idx, tmp); idx 514 drivers/gpu/drm/radeon/uvd_v1_0.c r = radeon_uvd_get_create_msg(rdev, ring->idx, 1, NULL); idx 520 drivers/gpu/drm/radeon/uvd_v1_0.c r = radeon_uvd_get_destroy_msg(rdev, ring->idx, 1, &fence); idx 537 drivers/gpu/drm/radeon/uvd_v1_0.c DRM_INFO("ib test on ring %d succeeded\n", ring->idx); idx 62 drivers/gpu/drm/radeon/vce_v1_0.c if (ring->idx == TN_RING_TYPE_VCE1_INDEX) idx 79 drivers/gpu/drm/radeon/vce_v1_0.c if (ring->idx == TN_RING_TYPE_VCE1_INDEX) idx 96 drivers/gpu/drm/radeon/vce_v1_0.c if (ring->idx == TN_RING_TYPE_VCE1_INDEX) idx 258 drivers/gpu/drm/rcar-du/rcar_du_plane.c int idx; idx 286 drivers/gpu/drm/rcar-du/rcar_du_plane.c idx = rcar_du_plane_hwalloc(plane, new_plane_state, idx 288 drivers/gpu/drm/rcar-du/rcar_du_plane.c if (idx < 0) idx 289 drivers/gpu/drm/rcar-du/rcar_du_plane.c idx = rcar_du_plane_hwalloc(plane, new_plane_state, idx 291 drivers/gpu/drm/rcar-du/rcar_du_plane.c if (idx < 0) { idx 294 drivers/gpu/drm/rcar-du/rcar_du_plane.c return idx; idx 298 drivers/gpu/drm/rcar-du/rcar_du_plane.c __func__, new_plane_state->format->planes, idx); idx 300 drivers/gpu/drm/rcar-du/rcar_du_plane.c new_plane_state->hwindex = idx; idx 1003 drivers/gpu/drm/savage/savage_bci.c &buf->idx, sizeof(buf->idx))) idx 547 drivers/gpu/drm/savage/savage_state.c const uint16_t *idx, idx 551 drivers/gpu/drm/savage/savage_state.c unsigned int prim = cmd_header->idx.prim; idx 552 drivers/gpu/drm/savage/savage_state.c unsigned int skip = cmd_header->idx.skip; idx 553 drivers/gpu/drm/savage/savage_state.c unsigned int n = cmd_header->idx.count; idx 638 drivers/gpu/drm/savage/savage_state.c if (idx[i] > dmabuf->total / 32) { idx 640 drivers/gpu/drm/savage/savage_state.c i, idx[i], dmabuf->total / 32); idx 652 drivers/gpu/drm/savage/savage_state.c BCI_DRAW_INDICES_S3D(count, prim, idx[2]); idx 655 drivers/gpu/drm/savage/savage_state.c BCI_WRITE(idx[i + reorder[i % 3]] | idx 656 drivers/gpu/drm/savage/savage_state.c (idx[i + 1 + idx 659 drivers/gpu/drm/savage/savage_state.c BCI_WRITE(idx[i + reorder[i % 3]]); idx 662 drivers/gpu/drm/savage/savage_state.c BCI_DRAW_INDICES_S3D(count, prim, idx[0]); idx 665 drivers/gpu/drm/savage/savage_state.c BCI_WRITE(idx[i] | (idx[i + 1] << 16)); idx 667 drivers/gpu/drm/savage/savage_state.c BCI_WRITE(idx[i]); idx 673 drivers/gpu/drm/savage/savage_state.c BCI_WRITE(idx[i] | (idx[i + 1] << 16)); idx 675 drivers/gpu/drm/savage/savage_state.c BCI_WRITE(idx[i]); idx 678 drivers/gpu/drm/savage/savage_state.c idx += count; idx 689 drivers/gpu/drm/savage/savage_state.c const uint16_t *idx, idx 694 drivers/gpu/drm/savage/savage_state.c unsigned int prim = cmd_header->idx.prim; idx 695 drivers/gpu/drm/savage/savage_state.c unsigned int skip = cmd_header->idx.skip; idx 696 drivers/gpu/drm/savage/savage_state.c unsigned int n = cmd_header->idx.count; idx 759 drivers/gpu/drm/savage/savage_state.c if (idx[i] > vb_size / (vb_stride * 4)) { idx 761 drivers/gpu/drm/savage/savage_state.c i, idx[i], vb_size / (vb_stride * 4)); idx 776 drivers/gpu/drm/savage/savage_state.c unsigned int j = idx[i + reorder[i % 3]]; idx 786 drivers/gpu/drm/savage/savage_state.c unsigned int j = idx[i]; idx 793 drivers/gpu/drm/savage/savage_state.c idx += count; idx 935 drivers/gpu/drm/savage/savage_state.c j = (cmd_header.idx.count + 3) / 4; idx 943 drivers/gpu/drm/savage/savage_state.c j = (cmd_header.idx.count + 3) / 4; idx 1061 drivers/gpu/drm/savage/savage_state.c j = (cmd_header.idx.count + 3) / 4; idx 281 drivers/gpu/drm/sti/sti_hda.c static bool hda_get_mode_idx(struct drm_display_mode mode, int *idx) idx 287 drivers/gpu/drm/sti/sti_hda.c *idx = i; idx 610 drivers/gpu/drm/sti/sti_hda.c int idx; idx 615 drivers/gpu/drm/sti/sti_hda.c if (!hda_get_mode_idx(*mode, &idx)) { idx 53 drivers/gpu/drm/tiny/hx8357d.c int ret, idx; idx 55 drivers/gpu/drm/tiny/hx8357d.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 179 drivers/gpu/drm/tiny/hx8357d.c drm_dev_exit(idx); idx 88 drivers/gpu/drm/tiny/ili9225.c int idx, ret = 0; idx 95 drivers/gpu/drm/tiny/ili9225.c if (!drm_dev_enter(fb->dev, &idx)) idx 161 drivers/gpu/drm/tiny/ili9225.c drm_dev_exit(idx); idx 196 drivers/gpu/drm/tiny/ili9225.c int ret, idx; idx 199 drivers/gpu/drm/tiny/ili9225.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 289 drivers/gpu/drm/tiny/ili9225.c drm_dev_exit(idx); idx 59 drivers/gpu/drm/tiny/ili9341.c int ret, idx; idx 61 drivers/gpu/drm/tiny/ili9341.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 135 drivers/gpu/drm/tiny/ili9341.c drm_dev_exit(idx); idx 57 drivers/gpu/drm/tiny/mi0283qt.c int ret, idx; idx 59 drivers/gpu/drm/tiny/mi0283qt.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 139 drivers/gpu/drm/tiny/mi0283qt.c drm_dev_exit(idx); idx 537 drivers/gpu/drm/tiny/repaper.c int idx, ret = 0; idx 543 drivers/gpu/drm/tiny/repaper.c if (!drm_dev_enter(fb->dev, &idx)) idx 631 drivers/gpu/drm/tiny/repaper.c drm_dev_exit(idx); idx 661 drivers/gpu/drm/tiny/repaper.c int i, ret, idx; idx 663 drivers/gpu/drm/tiny/repaper.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 791 drivers/gpu/drm/tiny/repaper.c drm_dev_exit(idx); idx 119 drivers/gpu/drm/tiny/st7586.c int start, end, idx, ret = 0; idx 124 drivers/gpu/drm/tiny/st7586.c if (!drm_dev_enter(fb->dev, &idx)) idx 155 drivers/gpu/drm/tiny/st7586.c drm_dev_exit(idx); idx 189 drivers/gpu/drm/tiny/st7586.c int idx, ret; idx 192 drivers/gpu/drm/tiny/st7586.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 253 drivers/gpu/drm/tiny/st7586.c drm_dev_exit(idx); idx 47 drivers/gpu/drm/tiny/st7735r.c int ret, idx; idx 50 drivers/gpu/drm/tiny/st7735r.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 109 drivers/gpu/drm/tiny/st7735r.c drm_dev_exit(idx); idx 1096 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c unsigned idx = 0; idx 1118 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c if (++idx < pool_offset) idx 287 drivers/gpu/drm/virtio/virtgpu_drv.h int virtio_gpu_cmd_get_capset_info(struct virtio_gpu_device *vgdev, int idx); idx 289 drivers/gpu/drm/virtio/virtgpu_drv.h int idx, int version, idx 669 drivers/gpu/drm/virtio/virtgpu_vq.c int virtio_gpu_cmd_get_capset_info(struct virtio_gpu_device *vgdev, int idx) idx 687 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->capset_index = cpu_to_le32(idx); idx 693 drivers/gpu/drm/virtio/virtgpu_vq.c int idx, int version, idx 705 drivers/gpu/drm/virtio/virtgpu_vq.c if (idx >= vgdev->num_capsets) idx 708 drivers/gpu/drm/virtio/virtgpu_vq.c if (version > vgdev->capsets[idx].max_version) idx 715 drivers/gpu/drm/virtio/virtgpu_vq.c max_size = vgdev->capsets[idx].max_size; idx 731 drivers/gpu/drm/virtio/virtgpu_vq.c cache_ent->id = vgdev->capsets[idx].id; idx 737 drivers/gpu/drm/virtio/virtgpu_vq.c if (search_ent->id == vgdev->capsets[idx].id && idx 760 drivers/gpu/drm/virtio/virtgpu_vq.c cmd_p->capset_id = cpu_to_le32(vgdev->capsets[idx].id); idx 439 drivers/gpu/drm/xen/xen_drm_front.c int idx; idx 441 drivers/gpu/drm/xen/xen_drm_front.c if (drm_dev_enter(obj->dev, &idx)) { idx 444 drivers/gpu/drm/xen/xen_drm_front.c drm_dev_exit(idx); idx 39 drivers/gpu/drm/xen/xen_drm_front_kms.c int idx; idx 41 drivers/gpu/drm/xen/xen_drm_front_kms.c if (drm_dev_enter(fb->dev, &idx)) { idx 44 drivers/gpu/drm/xen/xen_drm_front_kms.c drm_dev_exit(idx); idx 119 drivers/gpu/drm/xen/xen_drm_front_kms.c int ret, idx; idx 121 drivers/gpu/drm/xen/xen_drm_front_kms.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) idx 134 drivers/gpu/drm/xen/xen_drm_front_kms.c drm_dev_exit(idx); idx 141 drivers/gpu/drm/xen/xen_drm_front_kms.c int ret = 0, idx; idx 143 drivers/gpu/drm/xen/xen_drm_front_kms.c if (drm_dev_enter(pipe->crtc.dev, &idx)) { idx 146 drivers/gpu/drm/xen/xen_drm_front_kms.c drm_dev_exit(idx); idx 237 drivers/gpu/drm/xen/xen_drm_front_kms.c int idx; idx 253 drivers/gpu/drm/xen/xen_drm_front_kms.c if (!drm_dev_enter(pipe->crtc.dev, &idx)) { idx 269 drivers/gpu/drm/xen/xen_drm_front_kms.c drm_dev_exit(idx); idx 1343 drivers/hid/hid-core.c unsigned int idx = offset / 8; idx 1351 drivers/hid/hid-core.c value |= ((u32)report[idx] >> bit_shift) << bit_nr; idx 1356 drivers/hid/hid-core.c idx++; idx 1386 drivers/hid/hid-core.c unsigned int idx = offset / 8; idx 1391 drivers/hid/hid-core.c report[idx] &= ~(0xff << bit_shift); idx 1392 drivers/hid/hid-core.c report[idx] |= value << bit_shift; idx 1397 drivers/hid/hid-core.c idx++; idx 1403 drivers/hid/hid-core.c report[idx] &= ~(bit_mask << bit_shift); idx 1404 drivers/hid/hid-core.c report[idx] |= value << bit_shift; idx 77 drivers/hid/hid-input.c unsigned int cur_idx, unsigned int idx) idx 79 drivers/hid/hid-input.c return cur_idx == idx; idx 141 drivers/hid/hid-magicmouse.c int idx = msc->tracking_ids[ii]; idx 142 drivers/hid/hid-magicmouse.c if (msc->touches[idx].size < 8) { idx 148 drivers/hid/hid-magicmouse.c touch = idx; idx 20 drivers/hid/hid-primax.c int idx = size; idx 41 drivers/hid/hid-primax.c while (--idx > 1) { idx 42 drivers/hid/hid-primax.c if (data[idx] < 0xE0 || data[idx] > 0xE7) idx 44 drivers/hid/hid-primax.c data[0] |= (1 << (data[idx] - 0xE0)); idx 45 drivers/hid/hid-primax.c data[idx] = 0; idx 65 drivers/hid/hid-uclogic-params.c __u8 idx, size_t len) idx 85 drivers/hid/hid-uclogic-params.c (USB_DT_STRING << 8) + idx, idx 89 drivers/hid/hid-uclogic-params.c hid_dbg(hdev, "string descriptor #%hhu not found\n", idx); idx 94 drivers/hid/hid-uclogic-params.c idx, rc); idx 566 drivers/hid/usbhid/hiddev.c int idx, len; idx 569 drivers/hid/usbhid/hiddev.c if (get_user(idx, (int __user *)user_arg)) idx 575 drivers/hid/usbhid/hiddev.c if ((len = usb_string(dev, idx, buf, HID_STRING_SIZE-1)) < 0) { idx 734 drivers/hid/wacom_wac.c int idx = (features->type == INTUOS) ? (data[1] & 0x01) : 0; idx 745 drivers/hid/wacom_wac.c if (wacom->tool[idx] >= BTN_TOOL_MOUSE) { idx 762 drivers/hid/wacom_wac.c input_report_key(input, wacom->tool[idx], 0); idx 764 drivers/hid/wacom_wac.c input_event(input, EV_MSC, MSC_SERIAL, wacom->serial[idx]); idx 765 drivers/hid/wacom_wac.c wacom->id[idx] = 0; idx 773 drivers/hid/wacom_wac.c int idx = (features->type == INTUOS) ? (data[1] & 0x01) : 0; idx 783 drivers/hid/wacom_wac.c wacom->serial[idx] = ((data[3] & 0x0f) << 28) + idx 787 drivers/hid/wacom_wac.c wacom->id[idx] = (data[2] << 4) | (data[3] >> 4) | idx 790 drivers/hid/wacom_wac.c wacom->tool[idx] = wacom_intuos_get_tool_type(wacom->id[idx]); idx 817 drivers/hid/wacom_wac.c if (!wacom->id[idx]) idx 842 drivers/hid/wacom_wac.c int idx = (features->type == INTUOS) ? (data[1] & 0x01) : 0; idx 854 drivers/hid/wacom_wac.c if (!wacom->id[idx]) { idx 864 drivers/hid/wacom_wac.c if ((!((wacom->id[idx] >> 16) & 0x01) && idx 867 drivers/hid/wacom_wac.c (wacom->tool[idx] == BTN_TOOL_LENS && idx 967 drivers/hid/wacom_wac.c if (wacom->tool[idx] == BTN_TOOL_MOUSE) { idx 981 drivers/hid/wacom_wac.c else if (wacom->tool[idx] == BTN_TOOL_LENS) { idx 1003 drivers/hid/wacom_wac.c wacom_intuos_id_mangle(wacom->id[idx])); /* report tool id */ idx 1004 drivers/hid/wacom_wac.c input_report_key(input, wacom->tool[idx], 1); idx 1005 drivers/hid/wacom_wac.c input_event(input, EV_MSC, MSC_SERIAL, wacom->serial[idx]); idx 190 drivers/hv/hv_utils_transport.c if (hvt->cn_id.idx == msg->id.idx && idx 230 drivers/hv/hv_utils_transport.c cn_msg->id.idx = hvt->cn_id.idx; idx 281 drivers/hv/hv_utils_transport.c hvt->cn_id.idx = cn_idx; idx 311 drivers/hv/hv_utils_transport.c if (hvt->cn_id.idx > 0 && hvt->cn_id.val > 0 && idx 343 drivers/hv/hv_utils_transport.c if (hvt->cn_id.idx > 0 && hvt->cn_id.val > 0) idx 657 drivers/hv/vmbus_drv.c struct attribute *attr, int idx) idx 1734 drivers/hv/vmbus_drv.c struct attribute *attr, int idx) idx 256 drivers/hwmon/adm1031.c #define GET_FAN_AUTO_BITFIELD(data, idx) \ idx 257 drivers/hwmon/adm1031.c (*(data)->chan_select_table)[FAN_CHAN_FROM_REG((data)->conf1)][idx % 2] idx 135 drivers/hwmon/adt7475.c #define TACH_REG(idx) (REG_TACH_BASE + ((idx) * 2)) idx 136 drivers/hwmon/adt7475.c #define TACH_MIN_REG(idx) (REG_TACH_MIN_BASE + ((idx) * 2)) idx 138 drivers/hwmon/adt7475.c #define PWM_REG(idx) (REG_PWM_BASE + (idx)) idx 139 drivers/hwmon/adt7475.c #define PWM_MAX_REG(idx) (REG_PWM_MAX_BASE + (idx)) idx 140 drivers/hwmon/adt7475.c #define PWM_MIN_REG(idx) (REG_PWM_MIN_BASE + (idx)) idx 141 drivers/hwmon/adt7475.c #define PWM_CONFIG_REG(idx) (REG_PWM_CONFIG_BASE + (idx)) idx 143 drivers/hwmon/adt7475.c #define VOLTAGE_REG(idx) (REG_VOLTAGE_BASE + (idx)) idx 144 drivers/hwmon/adt7475.c #define VOLTAGE_MIN_REG(idx) (REG_VOLTAGE_MIN_BASE + ((idx) * 2)) idx 145 drivers/hwmon/adt7475.c #define VOLTAGE_MAX_REG(idx) (REG_VOLTAGE_MAX_BASE + ((idx) * 2)) idx 147 drivers/hwmon/adt7475.c #define TEMP_REG(idx) (REG_TEMP_BASE + (idx)) idx 148 drivers/hwmon/adt7475.c #define TEMP_MIN_REG(idx) (REG_TEMP_MIN_BASE + ((idx) * 2)) idx 149 drivers/hwmon/adt7475.c #define TEMP_MAX_REG(idx) (REG_TEMP_MAX_BASE + ((idx) * 2)) idx 150 drivers/hwmon/adt7475.c #define TEMP_TMIN_REG(idx) (REG_TEMP_TMIN_BASE + (idx)) idx 151 drivers/hwmon/adt7475.c #define TEMP_THERM_REG(idx) (REG_TEMP_THERM_BASE + (idx)) idx 152 drivers/hwmon/adt7475.c #define TEMP_OFFSET_REG(idx) (REG_TEMP_OFFSET_BASE + (idx)) idx 153 drivers/hwmon/adt7475.c #define TEMP_TRANGE_REG(idx) (REG_TEMP_TRANGE_BASE + (idx)) idx 577 drivers/hwmon/adt7475.c int shift, idx; idx 587 drivers/hwmon/adt7475.c idx = 0; idx 592 drivers/hwmon/adt7475.c idx = 1; idx 598 drivers/hwmon/adt7475.c idx = 1; idx 610 drivers/hwmon/adt7475.c data->enh_acoustics[idx] &= ~(0xf << shift); idx 611 drivers/hwmon/adt7475.c data->enh_acoustics[idx] |= (val << shift); idx 613 drivers/hwmon/adt7475.c i2c_smbus_write_byte_data(client, reg, data->enh_acoustics[idx]); idx 971 drivers/hwmon/adt7475.c int idx; idx 975 drivers/hwmon/adt7475.c idx = clamp_val(data->range[sattr->index] & 0xf, 0, idx 978 drivers/hwmon/adt7475.c return sprintf(buf, "%d\n", pwmfreq_table[idx]); idx 558 drivers/hwmon/emc2103.c int status, idx = 0; idx 604 drivers/hwmon/emc2103.c data->groups[idx++] = &emc2103_group; idx 606 drivers/hwmon/emc2103.c data->groups[idx++] = &emc2103_temp3_group; idx 608 drivers/hwmon/emc2103.c data->groups[idx++] = &emc2103_temp4_group; idx 2232 drivers/hwmon/f71882fg.c struct platform_device *pdev, int idx) idx 2241 drivers/hwmon/f71882fg.c if (((data->pwm_enable >> (idx * 2)) & 3) == 3) idx 2245 drivers/hwmon/f71882fg.c if (((data->pwm_enable >> (idx * 2)) & 1) != 1) idx 2249 drivers/hwmon/f71882fg.c if (idx == 2) idx 2259 drivers/hwmon/f71882fg.c (data->pwm_enable >> (idx * 2)) & 3, idx + 1); idx 2263 drivers/hwmon/f71882fg.c err = f71882fg_create_sysfs_files(pdev, &fxxxx_fan_attr[idx][0], idx 2270 drivers/hwmon/f71882fg.c &fxxxx_fan_beep_attr[idx], idx 2276 drivers/hwmon/f71882fg.c dev_info(&pdev->dev, "Fan: %d is in %s mode\n", idx + 1, idx 2277 drivers/hwmon/f71882fg.c (data->pwm_enable & (1 << (2 * idx))) ? "duty-cycle" : "RPM"); idx 2288 drivers/hwmon/f71882fg.c data->pwm_auto_point_mapping[idx] = idx 2289 drivers/hwmon/f71882fg.c f71882fg_read8(data, F71882FG_REG_POINT_MAPPING(idx)); idx 2290 drivers/hwmon/f71882fg.c if ((data->pwm_auto_point_mapping[idx] & 0x80) || idx 2291 drivers/hwmon/f71882fg.c (data->pwm_auto_point_mapping[idx] & 3) == 0) { idx 2295 drivers/hwmon/f71882fg.c "sysfs attributes for fan %d\n", idx + 1); idx 2306 drivers/hwmon/f71882fg.c &f71862fg_auto_pwm_attr[idx][0], idx 2312 drivers/hwmon/f71882fg.c &f71869_auto_pwm_attr[idx][0], idx 2317 drivers/hwmon/f71882fg.c &f8000_auto_pwm_attr[idx][0], idx 2322 drivers/hwmon/f71882fg.c &fxxxx_auto_pwm_attr[idx][0], idx 86 drivers/hwmon/ftsteutates.c #define FTS_REG_FAN_INPUT(idx) ((idx) + 0x20) idx 87 drivers/hwmon/ftsteutates.c #define FTS_REG_FAN_SOURCE(idx) ((idx) + 0x30) idx 88 drivers/hwmon/ftsteutates.c #define FTS_REG_FAN_CONTROL(idx) (((idx) << 16) + 0x4881) idx 90 drivers/hwmon/ftsteutates.c #define FTS_REG_TEMP_INPUT(idx) ((idx) + 0x40) idx 91 drivers/hwmon/ftsteutates.c #define FTS_REG_TEMP_CONTROL(idx) (((idx) << 16) + 0x0681) idx 93 drivers/hwmon/ftsteutates.c #define FTS_REG_VOLT(idx) ((idx) + 0x18) idx 326 drivers/hwmon/ina3221.c int ret, idx; idx 331 drivers/hwmon/ina3221.c idx = find_closest(val, ina3221_avg_samples, idx 335 drivers/hwmon/ina3221.c (idx << INA3221_CONFIG_AVG_SHIFT); idx 345 drivers/hwmon/ina3221.c idx = find_closest(tmp, ina3221_conv_time, idx 351 drivers/hwmon/ina3221.c (idx << INA3221_CONFIG_VBUS_CT_SHIFT) | idx 352 drivers/hwmon/ina3221.c (idx << INA3221_CONFIG_VSH_CT_SHIFT); idx 427 drivers/hwmon/lineage-pem.c int ret, idx = 0; idx 458 drivers/hwmon/lineage-pem.c data->groups[idx++] = &pem_group; idx 483 drivers/hwmon/lineage-pem.c data->groups[idx++] = &pem_input_group; idx 497 drivers/hwmon/lineage-pem.c data->groups[idx++] = &pem_fan_group; idx 1552 drivers/hwmon/lm85.c int idx = 0; idx 1593 drivers/hwmon/lm85.c data->groups[idx++] = &lm85_group; idx 1597 drivers/hwmon/lm85.c data->groups[idx++] = &lm85_group_minctl; idx 1598 drivers/hwmon/lm85.c data->groups[idx++] = &lm85_group_temp_off; idx 1612 drivers/hwmon/lm85.c data->groups[idx++] = &lm85_group_in4; idx 1616 drivers/hwmon/lm85.c data->groups[idx++] = &lm85_group_in567; idx 31 drivers/hwmon/menf21bmc_hwmon.c #define IDX_TO_VOLT_MIN_CMD(idx) (0x40 + idx) idx 32 drivers/hwmon/menf21bmc_hwmon.c #define IDX_TO_VOLT_MAX_CMD(idx) (0x50 + idx) idx 33 drivers/hwmon/menf21bmc_hwmon.c #define IDX_TO_VOLT_INP_CMD(idx) (0x60 + idx) idx 140 drivers/hwmon/pmbus/ibm-cffps.c int idx = *idxp; idx 141 drivers/hwmon/pmbus/ibm-cffps.c struct ibm_cffps *psu = to_psu(idxp, idx); idx 146 drivers/hwmon/pmbus/ibm-cffps.c switch (idx) { idx 1038 drivers/hwmon/pmbus/pmbus_core.c int idx) idx 1041 drivers/hwmon/pmbus/pmbus_core.c a->index = idx; idx 2289 drivers/hwmon/pmbus/pmbus_core.c int i, idx = 0; idx 2318 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2319 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2322 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2327 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2328 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2329 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_VOUT; idx 2332 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2337 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2338 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2339 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_IOUT; idx 2342 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2347 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2348 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2349 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_INPUT; idx 2352 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2357 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2358 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2359 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_TEMPERATURE; idx 2362 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2367 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2368 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2369 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_CML; idx 2372 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2377 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2378 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2379 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_OTHER; idx 2382 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2388 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2389 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2390 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_MFR_SPECIFIC; idx 2393 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2398 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2399 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2400 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_FAN_12; idx 2403 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 2408 drivers/hwmon/pmbus/pmbus_core.c entries[idx].client = client; idx 2409 drivers/hwmon/pmbus/pmbus_core.c entries[idx].page = i; idx 2410 drivers/hwmon/pmbus/pmbus_core.c entries[idx].reg = PMBUS_STATUS_FAN_34; idx 2413 drivers/hwmon/pmbus/pmbus_core.c &entries[idx++], idx 161 drivers/hwmon/scmi-hwmon.c int i, idx; idx 248 drivers/hwmon/scmi-hwmon.c idx = --nr_count[type]; idx 249 drivers/hwmon/scmi-hwmon.c *(scmi_sensors->info[type] + idx) = sensor; idx 136 drivers/hwmon/scpi-hwmon.c int idx, ret; idx 172 drivers/hwmon/scpi-hwmon.c for (i = 0, idx = 0; i < nr_sensors; i++) { idx 173 drivers/hwmon/scpi-hwmon.c struct sensor_data *sensor = &scpi_sensors->data[idx]; idx 229 drivers/hwmon/scpi-hwmon.c scpi_sensors->attrs[idx << 1] = &sensor->dev_attr_input.attr; idx 230 drivers/hwmon/scpi-hwmon.c scpi_sensors->attrs[(idx << 1) + 1] = &sensor->dev_attr_label.attr; idx 232 drivers/hwmon/scpi-hwmon.c sysfs_attr_init(scpi_sensors->attrs[idx << 1]); idx 233 drivers/hwmon/scpi-hwmon.c sysfs_attr_init(scpi_sensors->attrs[(idx << 1) + 1]); idx 234 drivers/hwmon/scpi-hwmon.c idx++; idx 597 drivers/hwmon/stts751.c int idx; idx 604 drivers/hwmon/stts751.c idx = find_closest_descending(val, stts751_intervals, idx 608 drivers/hwmon/stts751.c val, idx, stts751_intervals[idx]); idx 611 drivers/hwmon/stts751.c if (priv->interval == idx) idx 623 drivers/hwmon/stts751.c if (priv->interval < idx) { idx 625 drivers/hwmon/stts751.c priv->interval = idx; idx 631 drivers/hwmon/stts751.c ret = i2c_smbus_write_byte_data(priv->client, STTS751_REG_RATE, idx); idx 635 drivers/hwmon/stts751.c if (priv->interval != idx) { idx 637 drivers/hwmon/stts751.c priv->interval = idx; idx 33 drivers/hwmon/tc654.c #define TC654_REG_RPM(idx) (TC654_REG_RPM1 + (idx)) idx 34 drivers/hwmon/tc654.c #define TC654_REG_FAN_FAULT(idx) (TC654_REG_FAN_FAULT1 + (idx)) idx 386 drivers/hwmon/thmc50.c int idx = 0; idx 399 drivers/hwmon/thmc50.c data->groups[idx++] = &thmc50_group; idx 403 drivers/hwmon/thmc50.c data->groups[idx++] = &temp3_group; idx 59 drivers/hwspinlock/sirf_hwspinlock.c int idx, ret; idx 76 drivers/hwspinlock/sirf_hwspinlock.c for (idx = 0; idx < HW_SPINLOCK_NUMBER; idx++) { idx 77 drivers/hwspinlock/sirf_hwspinlock.c hwlock = &hwspin->bank.lock[idx]; idx 78 drivers/hwspinlock/sirf_hwspinlock.c hwlock->priv = hwspin->io_base + HW_SPINLOCK_OFFSET(idx); idx 312 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 318 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 319 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 320 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_SINGLE)) { idx 325 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c val = config->addr_val[idx]; idx 335 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 346 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 347 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 348 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_SINGLE)) { idx 353 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_val[idx] = val; idx 354 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_SINGLE; idx 364 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 370 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 371 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (idx % 2 != 0) { idx 375 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!((config->addr_type[idx] == ETM_ADDR_TYPE_NONE && idx 376 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_NONE) || idx 377 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c (config->addr_type[idx] == ETM_ADDR_TYPE_RANGE && idx 378 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_RANGE))) { idx 383 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c val1 = config->addr_val[idx]; idx 384 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c val2 = config->addr_val[idx + 1]; idx 394 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 406 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 407 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (idx % 2 != 0) { idx 411 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!((config->addr_type[idx] == ETM_ADDR_TYPE_NONE && idx 412 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_NONE) || idx 413 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c (config->addr_type[idx] == ETM_ADDR_TYPE_RANGE && idx 414 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_RANGE))) { idx 419 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_val[idx] = val1; idx 420 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_RANGE; idx 421 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_val[idx + 1] = val2; idx 422 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx + 1] = ETM_ADDR_TYPE_RANGE; idx 423 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->enable_ctrl1 |= (1 << (idx/2)); idx 433 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 439 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 440 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 441 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_START)) { idx 446 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c val = config->addr_val[idx]; idx 456 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 467 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 468 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 469 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_START)) { idx 474 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_val[idx] = val; idx 475 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_START; idx 476 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->startstop_ctrl |= (1 << idx); idx 487 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 493 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 494 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 495 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_STOP)) { idx 500 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c val = config->addr_val[idx]; idx 510 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c u8 idx; idx 521 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c idx = config->addr_idx; idx 522 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 523 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_STOP)) { idx 528 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_val[idx] = val; idx 529 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_STOP; idx 530 drivers/hwtracing/coresight/coresight-etm3x-sysfs.c config->startstop_ctrl |= (1 << (idx + 16)); idx 15 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 18 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 24 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (BMVAL(config->addr_acc[idx], 0, 1) == ETM_INSTR_ADDR) { idx 25 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (idx % 2 != 0) idx 33 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (config->addr_type[idx] != ETM_ADDR_TYPE_RANGE || idx 34 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx + 1] != ETM_ADDR_TYPE_RANGE) idx 42 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->viiectlr |= BIT(idx / 2 + 16); idx 43 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->viiectlr &= ~BIT(idx / 2); idx 49 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->viiectlr |= BIT(idx / 2); idx 50 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->viiectlr &= ~BIT(idx / 2 + 16); idx 848 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 val, idx; idx 853 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 854 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = BMVAL(config->addr_acc[idx], 0, 1); idx 868 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 879 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 882 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] &= ~(BIT(0) | BIT(1)); idx 893 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 898 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 900 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 901 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_SINGLE)) { idx 905 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = (unsigned long)config->addr_val[idx]; idx 914 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 923 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 924 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 925 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_SINGLE)) { idx 930 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_val[idx] = (u64)val; idx 931 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_SINGLE; idx 941 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 947 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 948 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (idx % 2 != 0) { idx 952 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!((config->addr_type[idx] == ETM_ADDR_TYPE_NONE && idx 953 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_NONE) || idx 954 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c (config->addr_type[idx] == ETM_ADDR_TYPE_RANGE && idx 955 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_RANGE))) { idx 960 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val1 = (unsigned long)config->addr_val[idx]; idx 961 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val2 = (unsigned long)config->addr_val[idx + 1]; idx 970 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 982 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 983 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (idx % 2 != 0) { idx 988 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!((config->addr_type[idx] == ETM_ADDR_TYPE_NONE && idx 989 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_NONE) || idx 990 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c (config->addr_type[idx] == ETM_ADDR_TYPE_RANGE && idx 991 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx + 1] == ETM_ADDR_TYPE_RANGE))) { idx 996 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_val[idx] = (u64)val1; idx 997 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_RANGE; idx 998 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_val[idx + 1] = (u64)val2; idx 999 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx + 1] = ETM_ADDR_TYPE_RANGE; idx 1016 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1022 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1024 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 1025 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_START)) { idx 1030 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = (unsigned long)config->addr_val[idx]; idx 1039 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1048 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1053 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 1054 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_START)) { idx 1059 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_val[idx] = (u64)val; idx 1060 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_START; idx 1061 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->vissctlr |= BIT(idx); idx 1073 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1079 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1081 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 1082 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_STOP)) { idx 1087 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = (unsigned long)config->addr_val[idx]; idx 1096 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1105 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1110 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (!(config->addr_type[idx] == ETM_ADDR_TYPE_NONE || idx 1111 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] == ETM_ADDR_TYPE_STOP)) { idx 1116 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_val[idx] = (u64)val; idx 1117 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_type[idx] = ETM_ADDR_TYPE_STOP; idx 1118 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->vissctlr |= BIT(idx + 16); idx 1131 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx, val; idx 1136 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1138 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = BMVAL(config->addr_acc[idx], 2, 3); idx 1150 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1161 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1164 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] &= ~(BIT(2) | BIT(3)); idx 1168 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] |= BIT(2); idx 1169 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] &= ~BIT(3); idx 1174 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] &= ~BIT(2); idx 1175 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] |= BIT(3); idx 1183 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] |= BIT(2); idx 1185 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] |= BIT(3); idx 1196 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1202 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1204 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = BMVAL(config->addr_acc[idx], 4, 6); idx 1213 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1227 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->addr_idx; idx 1229 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] &= ~(BIT(4) | BIT(5) | BIT(6)); idx 1230 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->addr_acc[idx] |= (val << 4); idx 1306 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1312 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->seq_idx; idx 1313 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = config->seq_ctrl[idx]; idx 1322 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1331 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->seq_idx; idx 1333 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->seq_ctrl[idx] = val & 0xFFFF; idx 1409 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1415 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->cntr_idx; idx 1416 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = config->cntrldvr[idx]; idx 1425 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1436 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->cntr_idx; idx 1437 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->cntrldvr[idx] = val; idx 1447 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1453 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->cntr_idx; idx 1454 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = config->cntr_val[idx]; idx 1463 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1474 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->cntr_idx; idx 1475 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->cntr_val[idx] = val; idx 1485 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1491 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->cntr_idx; idx 1492 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = config->cntr_ctrl[idx]; idx 1501 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1510 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->cntr_idx; idx 1511 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->cntr_ctrl[idx] = val; idx 1558 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1564 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->res_idx; idx 1565 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = config->res_ctrl[idx]; idx 1574 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1583 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->res_idx; idx 1585 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c if (idx % 2 != 0) idx 1588 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->res_ctrl[idx] = val & GENMASK(21, 0); idx 1634 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1647 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->ctxid_idx; idx 1648 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c val = (unsigned long)config->ctxid_pid[idx]; idx 1657 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c u8 idx; idx 1685 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c idx = config->ctxid_idx; idx 1686 drivers/hwtracing/coresight/coresight-etm4x-sysfs.c config->ctxid_pid[idx] = (u64)pid; idx 50 drivers/hwtracing/coresight/coresight-tmc-etr.c #define PERF_IDX2OFF(idx, buf) ((idx) % ((buf)->nr_pages << PAGE_SHIFT)) idx 479 drivers/hwtracing/coresight/coresight-tmc-etr.c #define INC_IDX_ROUND(idx, size) ((idx) = ((idx) + 1) % (size)) idx 1304 drivers/hwtracing/coresight/coresight.c int idx; idx 1310 drivers/hwtracing/coresight/coresight.c idx = coresight_search_device_idx(dict, dev_fwnode(dev)); idx 1311 drivers/hwtracing/coresight/coresight.c if (idx < 0) { idx 1313 drivers/hwtracing/coresight/coresight.c idx = dict->nr_idx; idx 1315 drivers/hwtracing/coresight/coresight.c (idx + 1) * sizeof(*dict->fwnode_list), idx 1318 drivers/hwtracing/coresight/coresight.c idx = -ENOMEM; idx 1322 drivers/hwtracing/coresight/coresight.c list[idx] = dev_fwnode(dev); idx 1324 drivers/hwtracing/coresight/coresight.c dict->nr_idx = idx + 1; idx 1327 drivers/hwtracing/coresight/coresight.c name = devm_kasprintf(dev, GFP_KERNEL, "%s%d", dict->pfx, idx); idx 429 drivers/hwtracing/intel_th/gth.c unsigned int idx = i * nparms + j; idx 437 drivers/hwtracing/intel_th/gth.c out_attrs[idx].attr.attr.name = name; idx 440 drivers/hwtracing/intel_th/gth.c out_attrs[idx].attr.attr.mode |= S_IRUGO; idx 441 drivers/hwtracing/intel_th/gth.c out_attrs[idx].attr.show = output_attr_show; idx 445 drivers/hwtracing/intel_th/gth.c out_attrs[idx].attr.attr.mode |= S_IWUSR; idx 446 drivers/hwtracing/intel_th/gth.c out_attrs[idx].attr.store = output_attr_store; idx 449 drivers/hwtracing/intel_th/gth.c sysfs_attr_init(&out_attrs[idx].attr.attr); idx 450 drivers/hwtracing/intel_th/gth.c attrs[idx] = &out_attrs[idx].attr.attr; idx 452 drivers/hwtracing/intel_th/gth.c out_attrs[idx].gth = gth; idx 453 drivers/hwtracing/intel_th/gth.c out_attrs[idx].port = i; idx 454 drivers/hwtracing/intel_th/gth.c out_attrs[idx].parm = j; idx 151 drivers/hwtracing/stm/core.c stm_master(struct stm_device *stm, unsigned int idx) idx 153 drivers/hwtracing/stm/core.c if (idx < stm->data->sw_start || idx > stm->data->sw_end) idx 156 drivers/hwtracing/stm/core.c return __stm_master(stm, idx); idx 159 drivers/hwtracing/stm/core.c static int stp_master_alloc(struct stm_device *stm, unsigned int idx) idx 170 drivers/hwtracing/stm/core.c __stm_master(stm, idx) = master; idx 175 drivers/hwtracing/stm/core.c static void stp_master_free(struct stm_device *stm, unsigned int idx) idx 177 drivers/hwtracing/stm/core.c struct stp_master *master = stm_master(stm, idx); idx 182 drivers/hwtracing/stm/core.c __stm_master(stm, idx) = NULL; idx 1134 drivers/hwtracing/stm/core.c int idx, ret; idx 1137 drivers/hwtracing/stm/core.c idx = srcu_read_lock(&stm_source_srcu); idx 1152 drivers/hwtracing/stm/core.c srcu_read_unlock(&stm_source_srcu, idx); idx 1165 drivers/hwtracing/stm/core.c int idx, ret; idx 1167 drivers/hwtracing/stm/core.c idx = srcu_read_lock(&stm_source_srcu); idx 1171 drivers/hwtracing/stm/core.c srcu_read_unlock(&stm_source_srcu, idx); idx 1299 drivers/hwtracing/stm/core.c int idx; idx 1307 drivers/hwtracing/stm/core.c idx = srcu_read_lock(&stm_source_srcu); idx 1315 drivers/hwtracing/stm/core.c srcu_read_unlock(&stm_source_srcu, idx); idx 425 drivers/i2c/busses/i2c-bcm-iproc.c unsigned int idx = iproc_i2c->tx_bytes + i; idx 427 drivers/i2c/busses/i2c-bcm-iproc.c val = msg->buf[idx]; idx 430 drivers/i2c/busses/i2c-bcm-iproc.c if (idx == msg->len - 1) { idx 83 drivers/i2c/busses/i2c-ibm_iic.c printk(KERN_DEBUG "ibm-iic%d: %s\n", dev->idx, header); idx 136 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: init\n", dev->idx); idx 183 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: soft reset\n", dev->idx); idx 192 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: trying to regain bus control\n", dev->idx); idx 248 drivers/i2c/busses/i2c-ibm_iic.c dev->idx); idx 252 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: smbus_quick(0x%02x)\n", dev->idx, p->addr); idx 305 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: smbus_quick -> %s\n", dev->idx, res ? "NACK" : "ACK"); idx 315 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: smbus_quick - bus is stuck\n", dev->idx); idx 329 drivers/i2c/busses/i2c-ibm_iic.c dev->idx, in_8(&iic->sts), in_8(&iic->extsts)); idx 347 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: xfer error, EXTSTS = 0x%02x\n", dev->idx, idx 363 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: bus is stuck, resetting\n", dev->idx); idx 380 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: iic_abort_xfer\n", dev->idx); idx 391 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: abort timeout, resetting...\n", dev->idx); idx 418 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: wait interrupted\n", dev->idx); idx 420 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: wait timeout\n", dev->idx); idx 430 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: poll timeout\n", dev->idx); idx 436 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: poll interrupted\n", dev->idx); idx 449 drivers/i2c/busses/i2c-ibm_iic.c DBG2("%d: iic_wait_for_tc -> %d\n", dev->idx, ret); idx 483 drivers/i2c/busses/i2c-ibm_iic.c DBG2("%d: xfer_bytes, %d, CNTL = 0x%02x\n", dev->idx, count, cmd); idx 495 drivers/i2c/busses/i2c-ibm_iic.c dev->idx, count, ret); idx 521 drivers/i2c/busses/i2c-ibm_iic.c DBG2("%d: iic_address, 0x%03x (%d-bit)\n", dev->idx, idx 557 drivers/i2c/busses/i2c-ibm_iic.c DBG2("%d: iic_xfer, %d msg(s)\n", dev->idx, num); idx 563 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: invalid address 0x%03x (%d-bit)\n", dev->idx, idx 576 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: invalid len %d in msg[%d]\n", dev->idx, idx 581 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: invalid addr in msg[%d]\n", dev->idx, i); idx 588 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: iic_xfer, bus is not free\n", dev->idx); idx 601 drivers/i2c/busses/i2c-ibm_iic.c DBG("%d: iic_xfer, bus is still not free\n", dev->idx); idx 44 drivers/i2c/busses/i2c-ibm_iic.h int idx; idx 519 drivers/i2c/busses/i2c-jz4780.c int idx) idx 578 drivers/i2c/busses/i2c-jz4780.c int cnt, int idx) idx 588 drivers/i2c/busses/i2c-jz4780.c if (idx < (cnt - 1)) idx 272 drivers/i2c/busses/i2c-mpc.c u32 idx; idx 282 drivers/i2c/busses/i2c-mpc.c idx = (*pval & 0xff) / 0x20; idx 283 drivers/i2c/busses/i2c-mpc.c setbits32(ctrl, 1 << (24 + idx * 2)); idx 249 drivers/i2c/busses/i2c-owl.c int ret, idx; idx 310 drivers/i2c/busses/i2c-owl.c for (idx = 0; idx < msgs[0].len; idx++) idx 311 drivers/i2c/busses/i2c-owl.c writel(msgs[0].buf[idx], idx 332 drivers/i2c/busses/i2c-owl.c for (idx = 0; idx < msg->len; idx++) { idx 338 drivers/i2c/busses/i2c-owl.c writel(msg->buf[idx], idx 342 drivers/i2c/busses/i2c-owl.c i2c_dev->msg_ptr = idx; idx 604 drivers/i2c/busses/i2c-piix4.c static uint8_t piix4_imc_read(uint8_t idx) idx 606 drivers/i2c/busses/i2c-piix4.c outb_p(idx, KERNCZ_IMC_IDX); idx 610 drivers/i2c/busses/i2c-piix4.c static void piix4_imc_write(uint8_t idx, uint8_t value) idx 612 drivers/i2c/busses/i2c-piix4.c outb_p(idx, KERNCZ_IMC_IDX); idx 467 drivers/i2c/busses/i2c-qup.c int idx; idx 472 drivers/i2c/busses/i2c-qup.c idx = 1; idx 476 drivers/i2c/busses/i2c-qup.c idx = 0; idx 485 drivers/i2c/busses/i2c-qup.c if (idx & 1) idx 491 drivers/i2c/busses/i2c-qup.c if (idx & 1 || qup->pos == msg->len - 1) idx 495 drivers/i2c/busses/i2c-qup.c idx++; idx 845 drivers/i2c/busses/i2c-qup.c int idx = 0; idx 870 drivers/i2c/busses/i2c-qup.c for (idx = 0; idx < num; idx++) { idx 871 drivers/i2c/busses/i2c-qup.c qup->msg = msg + idx; idx 872 drivers/i2c/busses/i2c-qup.c qup->is_last = idx == (num - 1); idx 926 drivers/i2c/busses/i2c-qup.c int idx = 0; idx 929 drivers/i2c/busses/i2c-qup.c if ((idx & 1) == 0) { idx 936 drivers/i2c/busses/i2c-qup.c idx++; idx 1073 drivers/i2c/busses/i2c-qup.c int ret, idx; idx 1090 drivers/i2c/busses/i2c-qup.c for (idx = 0; idx < num; idx++) { idx 1096 drivers/i2c/busses/i2c-qup.c if (qup_i2c_check_msg_len(&msgs[idx])) { idx 1101 drivers/i2c/busses/i2c-qup.c qup->msg = &msgs[idx]; idx 1102 drivers/i2c/busses/i2c-qup.c if (msgs[idx].flags & I2C_M_RD) idx 1512 drivers/i2c/busses/i2c-qup.c int idx; idx 1517 drivers/i2c/busses/i2c-qup.c for (idx = 0; idx < num; idx++) { idx 1518 drivers/i2c/busses/i2c-qup.c if (msgs[idx].flags & I2C_M_RD) idx 1520 drivers/i2c/busses/i2c-qup.c msgs[idx].len); idx 1523 drivers/i2c/busses/i2c-qup.c msgs[idx].len); idx 1525 drivers/i2c/busses/i2c-qup.c if (is_vmalloc_addr(msgs[idx].buf)) idx 1528 drivers/i2c/busses/i2c-qup.c total_len += msgs[idx].len; idx 1549 drivers/i2c/busses/i2c-qup.c int ret, idx = 0; idx 1583 drivers/i2c/busses/i2c-qup.c for (idx = 0; idx < num; idx++) { idx 1584 drivers/i2c/busses/i2c-qup.c qup->msg = &msgs[idx]; idx 1585 drivers/i2c/busses/i2c-qup.c qup->is_last = idx == (num - 1); idx 1587 drivers/i2c/busses/i2c-qup.c ret = qup_i2c_xfer_v2_msg(qup, idx, idx 1588 drivers/i2c/busses/i2c-qup.c !!(msgs[idx].flags & I2C_M_RD)); idx 190 drivers/i3c/master/dw-i3c-master.c #define DEV_ADDR_TABLE_LOC(start, idx) ((start) + ((idx) << 2)) idx 259 drivers/ide/aec62xx.c u8 idx = id->driver_data; idx 272 drivers/ide/aec62xx.c d = aec62xx_chipsets[idx]; idx 274 drivers/ide/aec62xx.c if (idx == 3 || idx == 4) { idx 279 drivers/ide/aec62xx.c "\n", pci_name(dev), (idx == 4) ? "R" : ""); idx 538 drivers/ide/alim15x3.c u8 rev = dev->revision, idx = id->driver_data; idx 564 drivers/ide/alim15x3.c if (idx == 0) idx 237 drivers/ide/amd74xx.c u8 idx = id->driver_data; idx 239 drivers/ide/amd74xx.c d = amd74xx_chipsets[idx]; idx 244 drivers/ide/amd74xx.c if (idx == 1) { idx 248 drivers/ide/amd74xx.c } else if (idx == 3) { idx 383 drivers/ide/cmd64x.c u8 idx = id->driver_data; idx 385 drivers/ide/cmd64x.c d = cmd64x_chipsets[idx]; idx 387 drivers/ide/cmd64x.c if (idx == 1) { idx 1408 drivers/ide/hpt366.c u8 idx = id->driver_data; idx 1412 drivers/ide/hpt366.c if ((idx == 0 || idx == 4) && (PCI_FUNC(dev->devfn) & 1)) idx 1415 drivers/ide/hpt366.c switch (idx) { idx 1426 drivers/ide/hpt366.c idx++; idx 1449 drivers/ide/hpt366.c d = hpt366_chipsets[min_t(u8, idx, 1)]; idx 1265 drivers/ide/ide-probe.c int idx = -ENOENT; idx 1282 drivers/ide/ide-probe.c idx = ffz(ide_indexes | i); idx 1285 drivers/ide/ide-probe.c idx = ffz(ide_indexes | 3); idx 1287 drivers/ide/ide-probe.c idx = ffz(ide_indexes); idx 1289 drivers/ide/ide-probe.c if (idx >= 0) idx 1290 drivers/ide/ide-probe.c ide_indexes |= (1 << idx); idx 1293 drivers/ide/ide-probe.c return idx; idx 1296 drivers/ide/ide-probe.c static void ide_free_port_slot(int idx) idx 1299 drivers/ide/ide-probe.c ide_indexes &= ~(1 << idx); idx 1361 drivers/ide/ide-probe.c int idx; idx 1375 drivers/ide/ide-probe.c idx = ide_find_port_slot(d); idx 1376 drivers/ide/ide-probe.c if (idx < 0) { idx 1384 drivers/ide/ide-probe.c ide_init_port_data(hwif, idx); idx 304 drivers/ide/pdc202xx_old.c u8 idx = id->driver_data; idx 306 drivers/ide/pdc202xx_old.c d = &pdc202xx_chipsets[idx]; idx 308 drivers/ide/pdc202xx_old.c if (idx < 2) idx 404 drivers/ide/serverworks.c u8 idx = id->driver_data; idx 406 drivers/ide/serverworks.c d = serverworks_chipsets[idx]; idx 408 drivers/ide/serverworks.c if (idx == 1) idx 410 drivers/ide/serverworks.c else if (idx == 2 || idx == 3) { idx 743 drivers/ide/siimage.c u8 idx = id->driver_data; idx 746 drivers/ide/siimage.c d = siimage_chipsets[idx]; idx 748 drivers/ide/siimage.c if (idx) { idx 247 drivers/ide/sis5513.c u8 drive_pci = sis_ata133_get_base(drive), clk, idx; idx 255 drivers/ide/sis5513.c idx = mode - XFER_MW_DMA_0 + 5; idx 257 drivers/ide/sis5513.c idx = mode - XFER_PIO_0; idx 258 drivers/ide/sis5513.c t1 |= ini_time_value[clk][idx] << 12; idx 259 drivers/ide/sis5513.c t1 |= act_time_value[clk][idx] << 16; idx 260 drivers/ide/sis5513.c t1 |= rco_time_value[clk][idx] << 24; idx 303 drivers/ide/sis5513.c u8 drive_pci = sis_ata133_get_base(drive), clk, idx; idx 311 drivers/ide/sis5513.c idx = mode - XFER_UDMA_0; idx 312 drivers/ide/sis5513.c regdw |= cycle_time_value[clk][idx] << 4; idx 313 drivers/ide/sis5513.c regdw |= cvs_time_value[clk][idx] << 8; idx 421 drivers/ide/via82cxxx.c u8 idx = id->driver_data; idx 459 drivers/ide/via82cxxx.c if (idx == 1) idx 464 drivers/ide/via82cxxx.c if (idx == VIA_IDFLAG_SINGLE) idx 77 drivers/iio/accel/cros_ec_accel_legacy.c int idx = chan->scan_index; idx 83 drivers/iio/accel/cros_ec_accel_legacy.c ret = st->read_ec_sensors_data(indio_dev, 1 << idx, &data); idx 422 drivers/iio/accel/kxcjk-1013.c int idx = data->chipset; idx 424 drivers/iio/accel/kxcjk-1013.c for (i = 0; i < ARRAY_SIZE(odr_start_up_times[idx]); ++i) { idx 425 drivers/iio/accel/kxcjk-1013.c if (odr_start_up_times[idx][i].odr_bits == data->odr_bits) idx 426 drivers/iio/accel/kxcjk-1013.c return odr_start_up_times[idx][i].usec; idx 201 drivers/iio/accel/mma7455_core.c #define MMA7455_CHANNEL(axis, idx) { \ idx 209 drivers/iio/accel/mma7455_core.c .scan_index = idx, \ idx 1207 drivers/iio/accel/mma8452.c #define MMA8452_CHANNEL(axis, idx, bits) { \ idx 1217 drivers/iio/accel/mma8452.c .scan_index = idx, \ idx 1229 drivers/iio/accel/mma8452.c #define MMA8652_CHANNEL(axis, idx, bits) { \ idx 1238 drivers/iio/accel/mma8452.c .scan_index = idx, \ idx 162 drivers/iio/adc/ad7124.c int i, idx; idx 166 drivers/iio/adc/ad7124.c idx = 0; idx 172 drivers/iio/adc/ad7124.c idx = i; idx 176 drivers/iio/adc/ad7124.c return idx; idx 287 drivers/iio/adc/ad7124.c int idx, ret; idx 305 drivers/iio/adc/ad7124.c idx = st->channel_config[chan->address].pga_bits; idx 308 drivers/iio/adc/ad7124.c *val2 = chan->scan_type.realbits - 1 + idx; idx 310 drivers/iio/adc/ad7124.c *val2 = chan->scan_type.realbits + idx; idx 295 drivers/iio/adc/ad7768-1.c unsigned int diff_new, diff_old, pwr_mode, i, idx; idx 299 drivers/iio/adc/ad7768-1.c idx = 0; idx 308 drivers/iio/adc/ad7768-1.c idx = i; idx 316 drivers/iio/adc/ad7768-1.c pwr_mode = AD7768_PWR_MCLK_DIV(ad7768_clk_config[idx].mclk_div) | idx 317 drivers/iio/adc/ad7768-1.c AD7768_PWR_PWRMODE(ad7768_clk_config[idx].pwrmode); idx 322 drivers/iio/adc/ad7768-1.c ret = ad7768_set_dig_fil(st, ad7768_clk_config[idx].dec_rate); idx 327 drivers/iio/adc/ad7768-1.c ad7768_clk_config[idx].clk_div); idx 470 drivers/iio/adc/at91_adc.c int bit, idx = 0; idx 494 drivers/iio/adc/at91_adc.c struct iio_chan_spec *chan = chan_array + idx; idx 499 drivers/iio/adc/at91_adc.c chan->scan_index = idx; idx 505 drivers/iio/adc/at91_adc.c idx++; idx 507 drivers/iio/adc/at91_adc.c timestamp = chan_array + idx; idx 511 drivers/iio/adc/at91_adc.c timestamp->scan_index = idx; idx 269 drivers/iio/adc/cc10001_adc.c unsigned int bit, idx = 0; idx 281 drivers/iio/adc/cc10001_adc.c struct iio_chan_spec *chan = &chan_array[idx]; idx 286 drivers/iio/adc/cc10001_adc.c chan->scan_index = idx; idx 292 drivers/iio/adc/cc10001_adc.c idx++; idx 295 drivers/iio/adc/cc10001_adc.c timestamp = &chan_array[idx]; idx 298 drivers/iio/adc/cc10001_adc.c timestamp->scan_index = idx; idx 451 drivers/iio/adc/dln2-adc.c #define DLN2_ADC_CHAN(lval, idx) { \ idx 453 drivers/iio/adc/dln2-adc.c lval.channel = idx; \ idx 458 drivers/iio/adc/dln2-adc.c lval.scan_index = idx; \ idx 184 drivers/iio/adc/mcp3911.c #define MCP3911_CHAN(idx) { \ idx 187 drivers/iio/adc/mcp3911.c .channel = idx, \ idx 29 drivers/iio/adc/men_z188_adc.c #define Z188_ADC_CHANNEL(idx) { \ idx 32 drivers/iio/adc/men_z188_adc.c .channel = (idx), \ idx 59 drivers/iio/adc/mt6577_auxadc.c #define MT6577_AUXADC_CHANNEL(idx) { \ idx 62 drivers/iio/adc/mt6577_auxadc.c .channel = (idx), \ idx 578 drivers/iio/adc/mxs-lradc-adc.c #define MXS_ADC_CHAN(idx, chan_type, name) { \ idx 581 drivers/iio/adc/mxs-lradc-adc.c .scan_index = (idx), \ idx 584 drivers/iio/adc/mxs-lradc-adc.c .channel = (idx), \ idx 585 drivers/iio/adc/mxs-lradc-adc.c .address = (idx), \ idx 206 drivers/iio/adc/spear_adc.c #define SPEAR_ADC_CHAN(idx) { \ idx 212 drivers/iio/adc/spear_adc.c .channel = idx, \ idx 443 drivers/iio/adc/ti-ads1015.c int ret, idx; idx 483 drivers/iio/adc/ti-ads1015.c idx = data->channel_data[chan->address].pga; idx 484 drivers/iio/adc/ti-ads1015.c *val = ads1015_fullscale_range[idx]; idx 489 drivers/iio/adc/ti-ads1015.c idx = data->channel_data[chan->address].data_rate; idx 490 drivers/iio/adc/ti-ads1015.c *val = data->data_rate[idx]; idx 45 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c int idx = chan->scan_index; idx 51 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c ret = st->core.read_ec_sensors_data(indio_dev, 1 << idx, &data); idx 70 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c *val = st->core.calib[idx].offset; idx 92 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c *val = st->core.calib[idx].scale >> 15; idx 93 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c *val2 = ((st->core.calib[idx].scale & 0x7FFF) * 1000000LL) / idx 157 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c int idx = chan->scan_index; idx 163 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c st->core.calib[idx].offset = val; idx 178 drivers/iio/common/cros_ec_sensors/cros_ec_sensors.c st->core.calib[idx].scale = val; idx 261 drivers/iio/common/cros_ec_sensors/cros_ec_sensors_core.c unsigned int idx) idx 269 drivers/iio/common/cros_ec_sensors/cros_ec_sensors_core.c (1 + idx + st->param.info.sensor_num * idx 272 drivers/iio/common/cros_ec_sensors/cros_ec_sensors_core.c return EC_MEMMAP_GYRO_DATA + sizeof(u16) * idx; idx 260 drivers/iio/common/ssp_sensors/ssp_spi.c static int ssp_handle_big_data(struct ssp_data *data, char *dataframe, int *idx) idx 263 drivers/iio/common/ssp_sensors/ssp_spi.c *idx += 8; idx 269 drivers/iio/common/ssp_sensors/ssp_spi.c int idx, sd; idx 273 drivers/iio/common/ssp_sensors/ssp_spi.c for (idx = 0; idx < len;) { idx 274 drivers/iio/common/ssp_sensors/ssp_spi.c switch (dataframe[idx++]) { idx 276 drivers/iio/common/ssp_sensors/ssp_spi.c sd = dataframe[idx++]; idx 287 drivers/iio/common/ssp_sensors/ssp_spi.c &dataframe[idx], idx 293 drivers/iio/common/ssp_sensors/ssp_spi.c idx += ssp_offset_map[sd]; idx 296 drivers/iio/common/ssp_sensors/ssp_spi.c sd = ssp_print_mcu_debug(dataframe, &idx, len); idx 304 drivers/iio/common/ssp_sensors/ssp_spi.c idx += len; idx 307 drivers/iio/common/ssp_sensors/ssp_spi.c ssp_handle_big_data(data, dataframe, &idx); idx 249 drivers/iio/health/max30100.c int idx; idx 257 drivers/iio/health/max30100.c for (idx = 0; idx < ARRAY_SIZE(max30100_led_current_mapping); idx++) { idx 258 drivers/iio/health/max30100.c if (max30100_led_current_mapping[idx] == val) { idx 259 drivers/iio/health/max30100.c *reg = idx + 1; idx 453 drivers/iio/humidity/hts221_core.c u8 idx; idx 459 drivers/iio/humidity/hts221_core.c idx = hw->sensors[HTS221_SENSOR_H].cur_avg_idx; idx 460 drivers/iio/humidity/hts221_core.c *val = avg->avg_avl[idx]; idx 465 drivers/iio/humidity/hts221_core.c idx = hw->sensors[HTS221_SENSOR_T].cur_avg_idx; idx 466 drivers/iio/humidity/hts221_core.c *val = avg->avg_avl[idx]; idx 364 drivers/iio/light/apds9960.c int idx; idx 366 drivers/iio/light/apds9960.c for (idx = 0; idx < ARRAY_SIZE(apds9960_int_time); idx++) { idx 367 drivers/iio/light/apds9960.c if (apds9960_int_time[idx][0] == val2) { idx 370 drivers/iio/light/apds9960.c apds9960_int_time[idx][1]); idx 384 drivers/iio/light/apds9960.c int idx; idx 386 drivers/iio/light/apds9960.c for (idx = 0; idx < ARRAY_SIZE(apds9960_pxs_gain_map); idx++) { idx 387 drivers/iio/light/apds9960.c if (apds9960_pxs_gain_map[idx] == val) { idx 393 drivers/iio/light/apds9960.c idx << APDS9960_REG_CONTROL_PGAIN_MASK_SHIFT); idx 402 drivers/iio/light/apds9960.c idx << APDS9960_REG_CONFIG_2_GGAIN_MASK_SHIFT); idx 404 drivers/iio/light/apds9960.c data->pxs_gain = idx; idx 416 drivers/iio/light/apds9960.c int idx; idx 418 drivers/iio/light/apds9960.c for (idx = 0; idx < ARRAY_SIZE(apds9960_als_gain_map); idx++) { idx 419 drivers/iio/light/apds9960.c if (apds9960_als_gain_map[idx] == val) { idx 423 drivers/iio/light/apds9960.c APDS9960_REG_CONTROL_AGAIN_MASK, idx); idx 425 drivers/iio/light/apds9960.c data->als_gain = idx; idx 47 drivers/iio/light/cros_ec_light_prox.c int idx = chan->scan_index; idx 54 drivers/iio/light/cros_ec_light_prox.c ret = cros_ec_sensors_read_cmd(indio_dev, 1 << idx, idx 66 drivers/iio/light/cros_ec_light_prox.c ret = cros_ec_sensors_read_cmd(indio_dev, 1 << idx, idx 93 drivers/iio/light/cros_ec_light_prox.c *val = st->core.calib[idx].offset; idx 131 drivers/iio/light/cros_ec_light_prox.c int idx = chan->scan_index; idx 137 drivers/iio/light/cros_ec_light_prox.c st->core.calib[idx].offset = val; idx 581 drivers/iio/light/rpr0521.c int ret, reg, idx; idx 587 drivers/iio/light/rpr0521.c idx = (rpr0521_gain[chan].mask & reg) >> rpr0521_gain[chan].shift; idx 588 drivers/iio/light/rpr0521.c *val = rpr0521_gain[chan].gain[idx].scale; idx 589 drivers/iio/light/rpr0521.c *val2 = rpr0521_gain[chan].gain[idx].uscale; idx 597 drivers/iio/light/rpr0521.c int i, idx = -EINVAL; idx 603 drivers/iio/light/rpr0521.c idx = i; idx 607 drivers/iio/light/rpr0521.c if (idx < 0) idx 608 drivers/iio/light/rpr0521.c return idx; idx 612 drivers/iio/light/rpr0521.c idx << rpr0521_gain[chan].shift); idx 349 drivers/iio/light/zopt2201.c static int zopt2201_write_scale_als_by_idx(struct zopt2201_data *data, int idx) idx 354 drivers/iio/light/zopt2201.c ret = zopt2201_set_resolution(data, zopt2201_scale_als[idx].res); idx 358 drivers/iio/light/zopt2201.c ret = zopt2201_set_gain(data, zopt2201_scale_als[idx].gain); idx 379 drivers/iio/light/zopt2201.c static int zopt2201_write_scale_uvb_by_idx(struct zopt2201_data *data, int idx) idx 384 drivers/iio/light/zopt2201.c ret = zopt2201_set_resolution(data, zopt2201_scale_als[idx].res); idx 388 drivers/iio/light/zopt2201.c ret = zopt2201_set_gain(data, zopt2201_scale_als[idx].gain); idx 173 drivers/iio/magnetometer/hmc5843_core.c int idx, int *val) idx 190 drivers/iio/magnetometer/hmc5843_core.c *val = sign_extend32(be16_to_cpu(values[idx]), 15); idx 464 drivers/iio/magnetometer/hmc5843_core.c #define HMC5843_CHANNEL(axis, idx) \ idx 472 drivers/iio/magnetometer/hmc5843_core.c .scan_index = idx, \ idx 482 drivers/iio/magnetometer/hmc5843_core.c #define HMC5983_CHANNEL(axis, idx) \ idx 490 drivers/iio/magnetometer/hmc5843_core.c .scan_index = idx, \ idx 413 drivers/iio/magnetometer/mag3110.c #define MAG3110_CHANNEL(axis, idx) { \ idx 421 drivers/iio/magnetometer/mag3110.c .scan_index = idx, \ idx 206 drivers/iio/magnetometer/rm3100-core.c static int rm3100_read_mag(struct rm3100_data *data, int idx, int *val) idx 213 drivers/iio/magnetometer/rm3100-core.c ret = regmap_write(regmap, RM3100_REG_POLL, BIT(4 + idx)); idx 221 drivers/iio/magnetometer/rm3100-core.c ret = regmap_bulk_read(regmap, RM3100_REG_MX2 + 3 * idx, buffer, 3); idx 236 drivers/iio/magnetometer/rm3100-core.c #define RM3100_CHANNEL(axis, idx) \ idx 244 drivers/iio/magnetometer/rm3100-core.c .scan_index = idx, \ idx 38 drivers/iio/multiplexer/iio-mux.c static int iio_mux_select(struct mux *mux, int idx) idx 40 drivers/iio/multiplexer/iio-mux.c struct mux_child *child = &mux->child[idx]; idx 41 drivers/iio/multiplexer/iio-mux.c struct iio_chan_spec const *chan = &mux->chan[idx]; idx 90 drivers/iio/multiplexer/iio-mux.c int idx = chan - mux->chan; idx 93 drivers/iio/multiplexer/iio-mux.c ret = iio_mux_select(mux, idx); idx 121 drivers/iio/multiplexer/iio-mux.c int idx = chan - mux->chan; idx 124 drivers/iio/multiplexer/iio-mux.c ret = iio_mux_select(mux, idx); idx 148 drivers/iio/multiplexer/iio-mux.c int idx = chan - mux->chan; idx 151 drivers/iio/multiplexer/iio-mux.c ret = iio_mux_select(mux, idx); idx 179 drivers/iio/multiplexer/iio-mux.c int idx = chan - mux->chan; idx 182 drivers/iio/multiplexer/iio-mux.c ret = iio_mux_select(mux, idx); idx 201 drivers/iio/multiplexer/iio-mux.c int idx = chan - mux->chan; idx 208 drivers/iio/multiplexer/iio-mux.c ret = iio_mux_select(mux, idx); idx 229 drivers/iio/multiplexer/iio-mux.c devm_kfree(dev, mux->child[idx].ext_info_cache[private].data); idx 230 drivers/iio/multiplexer/iio-mux.c mux->child[idx].ext_info_cache[private].data = new; idx 231 drivers/iio/multiplexer/iio-mux.c mux->child[idx].ext_info_cache[private].size = len; idx 239 drivers/iio/multiplexer/iio-mux.c u32 state, const char *label, int idx) idx 241 drivers/iio/multiplexer/iio-mux.c struct mux_child *child = &mux->child[idx]; idx 242 drivers/iio/multiplexer/iio-mux.c struct iio_chan_spec *chan = &mux->chan[idx]; idx 44 drivers/iio/pressure/cros_ec_baro.c int idx = chan->scan_index; idx 50 drivers/iio/pressure/cros_ec_baro.c ret = cros_ec_sensors_read_cmd(indio_dev, 1 << idx, idx 108 drivers/iio/proximity/sx9500.c #define SX9500_CHANNEL(idx) \ idx 114 drivers/iio/proximity/sx9500.c .channel = idx, \ idx 117 drivers/iio/proximity/sx9500.c .scan_index = idx, \ idx 107 drivers/infiniband/core/core_priv.h unsigned int idx); idx 2320 drivers/infiniband/core/device.c unsigned int idx = 0; idx 2328 drivers/infiniband/core/device.c ret = nldev_cb(dev, skb, cb, idx); idx 2331 drivers/infiniband/core/device.c idx++; idx 293 drivers/infiniband/core/netlink.c int idx; idx 295 drivers/infiniband/core/netlink.c for (idx = 0; idx < RDMA_NL_NUM_CLIENTS; idx++) idx 296 drivers/infiniband/core/netlink.c init_rwsem(&rdma_nl_types[idx].sem); idx 301 drivers/infiniband/core/netlink.c int idx; idx 303 drivers/infiniband/core/netlink.c for (idx = 0; idx < RDMA_NL_NUM_CLIENTS; idx++) idx 304 drivers/infiniband/core/netlink.c WARN(rdma_nl_types[idx].cb_table, idx 306 drivers/infiniband/core/netlink.c idx, KBUILD_MODNAME); idx 899 drivers/infiniband/core/nldev.c unsigned int idx) idx 904 drivers/infiniband/core/nldev.c if (idx < start) idx 918 drivers/infiniband/core/nldev.c idx++; idx 920 drivers/infiniband/core/nldev.c out: cb->args[0] = idx; idx 994 drivers/infiniband/core/nldev.c u32 idx = 0; idx 1020 drivers/infiniband/core/nldev.c if (idx < start) { idx 1021 drivers/infiniband/core/nldev.c idx++; idx 1035 drivers/infiniband/core/nldev.c idx++; idx 1041 drivers/infiniband/core/nldev.c cb->args[0] = idx; idx 1092 drivers/infiniband/core/nldev.c unsigned int idx) idx 1097 drivers/infiniband/core/nldev.c if (idx < start) idx 1110 drivers/infiniband/core/nldev.c idx++; idx 1113 drivers/infiniband/core/nldev.c cb->args[0] = idx; idx 1276 drivers/infiniband/core/nldev.c int err, ret = 0, idx = 0; idx 1344 drivers/infiniband/core/nldev.c if (idx < start || !rdma_restrack_get(res)) idx 1371 drivers/infiniband/core/nldev.c next: idx++; idx 1378 drivers/infiniband/core/nldev.c cb->args[0] = idx; idx 108 drivers/infiniband/core/rdma_core.h int uverbs_output_written(const struct uverbs_attr_bundle *bundle, size_t idx); idx 720 drivers/infiniband/core/umem_odp.c int idx; idx 734 drivers/infiniband/core/umem_odp.c idx = (addr - ib_umem_start(umem_odp)) >> umem_odp->page_shift; idx 735 drivers/infiniband/core/umem_odp.c if (umem_odp->page_list[idx]) { idx 736 drivers/infiniband/core/umem_odp.c struct page *page = umem_odp->page_list[idx]; idx 737 drivers/infiniband/core/umem_odp.c dma_addr_t dma = umem_odp->dma_list[idx]; idx 758 drivers/infiniband/core/umem_odp.c umem_odp->page_list[idx] = NULL; idx 759 drivers/infiniband/core/umem_odp.c umem_odp->dma_list[idx] = 0; idx 650 drivers/infiniband/core/uverbs_ioctl.c size_t idx, u64 allowed_bits) idx 655 drivers/infiniband/core/uverbs_ioctl.c attr = uverbs_attr_get(attrs_bundle, idx); idx 683 drivers/infiniband/core/uverbs_ioctl.c size_t idx, u64 allowed_bits) idx 688 drivers/infiniband/core/uverbs_ioctl.c ret = uverbs_get_flags64(&flags, attrs_bundle, idx, allowed_bits); idx 738 drivers/infiniband/core/uverbs_ioctl.c int uverbs_copy_to(const struct uverbs_attr_bundle *bundle, size_t idx, idx 741 drivers/infiniband/core/uverbs_ioctl.c const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx); idx 760 drivers/infiniband/core/uverbs_ioctl.c int uverbs_output_written(const struct uverbs_attr_bundle *bundle, size_t idx) idx 762 drivers/infiniband/core/uverbs_ioctl.c const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx); idx 771 drivers/infiniband/core/uverbs_ioctl.c size_t idx, s64 lower_bound, u64 upper_bound, idx 776 drivers/infiniband/core/uverbs_ioctl.c attr = uverbs_attr_get(attrs_bundle, idx); idx 794 drivers/infiniband/core/uverbs_ioctl.c size_t idx, const void *from, size_t size) idx 796 drivers/infiniband/core/uverbs_ioctl.c const struct uverbs_attr *attr = uverbs_attr_get(bundle, idx); idx 803 drivers/infiniband/core/uverbs_ioctl.c return uverbs_copy_to(bundle, idx, from, size); idx 319 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (ctx->idx >= sgid_tbl->max) idx 321 drivers/infiniband/hw/bnxt_re/ib_verbs.c gid_to_del = &sgid_tbl->tbl[ctx->idx].gid; idx 322 drivers/infiniband/hw/bnxt_re/ib_verbs.c vlan_id = sgid_tbl->tbl[ctx->idx].vlan_id; idx 331 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (ctx->idx == 0 && idx 347 drivers/infiniband/hw/bnxt_re/ib_verbs.c ctx_tbl[ctx->idx] = NULL; idx 389 drivers/infiniband/hw/bnxt_re/ib_verbs.c ctx->idx = tbl_idx; idx 43 drivers/infiniband/hw/bnxt_re/ib_verbs.h u32 idx; idx 540 drivers/infiniband/hw/bnxt_re/qplib_fp.c int rc, idx; idx 593 drivers/infiniband/hw/bnxt_re/qplib_fp.c for (idx = 0; idx < srq->hwq.max_elements; idx++) idx 594 drivers/infiniband/hw/bnxt_re/qplib_fp.c srq->swq[idx].next_idx = idx + 1; idx 49 drivers/infiniband/hw/bnxt_re/qplib_res.h #define HWQ_CMP(idx, hwq) ((idx) & ((hwq)->max_elements - 1)) idx 53 drivers/infiniband/hw/cxgb3/cxio_resource.c u32 i, j, entry = 0, idx; idx 73 drivers/infiniband/hw/cxgb3/cxio_resource.c idx = (random_bytes >> (j * 2)) & 0xF; idx 75 drivers/infiniband/hw/cxgb3/cxio_resource.c (unsigned char *) &rarray[idx], idx 77 drivers/infiniband/hw/cxgb3/cxio_resource.c rarray[idx] = i; idx 449 drivers/infiniband/hw/cxgb3/iwch_cm.c V_L2T_IDX(ep->l2t->idx) | V_TX_CHANNEL(ep->l2t->smt_idx); idx 1287 drivers/infiniband/hw/cxgb3/iwch_cm.c V_L2T_IDX(ep->l2t->idx) | V_TX_CHANNEL(ep->l2t->smt_idx); idx 359 drivers/infiniband/hw/cxgb3/iwch_qp.c u32 idx; idx 385 drivers/infiniband/hw/cxgb3/iwch_qp.c idx = Q_PTR2IDX(qhp->wq.wptr, qhp->wq.size_log2); idx 386 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe = (union t3_wr *) (qhp->wq.queue + idx); idx 448 drivers/infiniband/hw/cxgb3/iwch_qp.c __func__, (unsigned long long)wr->wr_id, idx, idx 471 drivers/infiniband/hw/cxgb3/iwch_qp.c u32 idx; idx 495 drivers/infiniband/hw/cxgb3/iwch_qp.c idx = Q_PTR2IDX(qhp->wq.wptr, qhp->wq.size_log2); idx 496 drivers/infiniband/hw/cxgb3/iwch_qp.c wqe = (union t3_wr *) (qhp->wq.queue + idx); idx 513 drivers/infiniband/hw/cxgb3/iwch_qp.c idx, qhp->wq.rq_wptr, qhp->wq.rq_rptr, wqe); idx 793 drivers/infiniband/hw/cxgb4/cm.c L2T_IDX_V(ep->l2t->idx) | idx 2006 drivers/infiniband/hw/cxgb4/cm.c L2T_IDX_V(ep->l2t->idx) | idx 2233 drivers/infiniband/hw/cxgb4/cm.c ep->l2t->idx); idx 2442 drivers/infiniband/hw/cxgb4/cm.c L2T_IDX_V(ep->l2t->idx) | idx 3437 drivers/infiniband/hw/cxgb4/cm.c ep->l2t->idx); idx 230 drivers/infiniband/hw/cxgb4/cq.c CQE_WRID_SQ_IDX(&cqe) = swcqe->idx; idx 244 drivers/infiniband/hw/cxgb4/cq.c int idx; idx 249 drivers/infiniband/hw/cxgb4/cq.c idx = wq->sq.flush_cidx; idx 250 drivers/infiniband/hw/cxgb4/cq.c while (idx != wq->sq.pidx) { idx 251 drivers/infiniband/hw/cxgb4/cq.c swsqe = &wq->sq.sw_sq[idx]; idx 258 drivers/infiniband/hw/cxgb4/cq.c if (++idx == wq->sq.size) idx 259 drivers/infiniband/hw/cxgb4/cq.c idx = 0; idx 303 drivers/infiniband/hw/cxgb4/cq.c read_cqe->u.scqe.cidx = wq->sq.oldest_read->idx; idx 465 drivers/infiniband/hw/cxgb4/cq.c u16 idx = 0; idx 481 drivers/infiniband/hw/cxgb4/cq.c idx += DIV_ROUND_UP(pwr->len16 * 16, T4_EQ_ENTRY_SIZE); idx 484 drivers/infiniband/hw/cxgb4/cq.c if (idx) { idx 485 drivers/infiniband/hw/cxgb4/cq.c t4_ring_srq_db(srq, idx, pwr->len16, &pwr->wqe); idx 700 drivers/infiniband/hw/cxgb4/cq.c int idx = CQE_WRID_SQ_IDX(hw_cqe); idx 710 drivers/infiniband/hw/cxgb4/cq.c if (idx < wq->sq.cidx) idx 711 drivers/infiniband/hw/cxgb4/cq.c wq->sq.in_use -= wq->sq.size + idx - wq->sq.cidx; idx 713 drivers/infiniband/hw/cxgb4/cq.c wq->sq.in_use -= idx - wq->sq.cidx; idx 715 drivers/infiniband/hw/cxgb4/cq.c wq->sq.cidx = (uint16_t)idx; idx 95 drivers/infiniband/hw/cxgb4/device.c int idx; idx 100 drivers/infiniband/hw/cxgb4/device.c idx = (atomic_inc_return(&wq->rdev->wr_log_idx) - 1) & idx 119 drivers/infiniband/hw/cxgb4/device.c wq->rdev->wr_log[idx] = le; idx 128 drivers/infiniband/hw/cxgb4/device.c int idx, end; idx 132 drivers/infiniband/hw/cxgb4/device.c idx = atomic_read(&dev->rdev.wr_log_idx) & idx 134 drivers/infiniband/hw/cxgb4/device.c end = idx - 1; idx 137 drivers/infiniband/hw/cxgb4/device.c lep = &dev->rdev.wr_log[idx]; idx 138 drivers/infiniband/hw/cxgb4/device.c while (idx != end) { idx 149 drivers/infiniband/hw/cxgb4/device.c idx, idx 164 drivers/infiniband/hw/cxgb4/device.c idx++; idx 165 drivers/infiniband/hw/cxgb4/device.c if (idx > (dev->rdev.wr_log_size - 1)) idx 166 drivers/infiniband/hw/cxgb4/device.c idx = 0; idx 167 drivers/infiniband/hw/cxgb4/device.c lep = &dev->rdev.wr_log[idx]; idx 272 drivers/infiniband/hw/cxgb4/device.c qp->srq ? qp->srq->idx : qp->wq.rq.qid, idx 1355 drivers/infiniband/hw/cxgb4/device.c unsigned idx; idx 1361 drivers/infiniband/hw/cxgb4/device.c int idx; idx 1363 drivers/infiniband/hw/cxgb4/device.c for (idx = 0; idx < qp_list->idx; idx++) idx 1364 drivers/infiniband/hw/cxgb4/device.c c4iw_qp_rem_ref(&qp_list->qps[idx]->ibqp); idx 1369 drivers/infiniband/hw/cxgb4/device.c int idx; idx 1372 drivers/infiniband/hw/cxgb4/device.c for (idx = 0; idx < qp_list->idx; idx++) { idx 1373 drivers/infiniband/hw/cxgb4/device.c struct c4iw_qp *qp = qp_list->qps[idx]; idx 1446 drivers/infiniband/hw/cxgb4/device.c qp_list.idx = 0; idx 1451 drivers/infiniband/hw/cxgb4/device.c qp_list.qps[qp_list.idx++] = qp; idx 515 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int idx; idx 1042 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_free_srq_idx(struct c4iw_rdev *rdev, int idx); idx 699 drivers/infiniband/hw/cxgb4/qp.c u16 idx; idx 712 drivers/infiniband/hw/cxgb4/qp.c swsqe->idx = qhp->wq.sq.pidx; idx 736 drivers/infiniband/hw/cxgb4/qp.c swsqe->idx = qhp->wq.sq.pidx; idx 753 drivers/infiniband/hw/cxgb4/qp.c idx = DIV_ROUND_UP(len16 * 16, T4_EQ_ENTRY_SIZE); idx 755 drivers/infiniband/hw/cxgb4/qp.c t4_ring_sq_db(&qhp->wq, idx, wqe); idx 1089 drivers/infiniband/hw/cxgb4/qp.c u16 idx = 0; idx 1228 drivers/infiniband/hw/cxgb4/qp.c swsqe->idx = qhp->wq.sq.pidx; idx 1248 drivers/infiniband/hw/cxgb4/qp.c idx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); idx 1251 drivers/infiniband/hw/cxgb4/qp.c t4_ring_sq_db(&qhp->wq, idx, wqe); idx 1255 drivers/infiniband/hw/cxgb4/qp.c ring_kernel_sq_db(qhp, idx); idx 1269 drivers/infiniband/hw/cxgb4/qp.c u16 idx = 0; idx 1326 drivers/infiniband/hw/cxgb4/qp.c idx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); idx 1331 drivers/infiniband/hw/cxgb4/qp.c t4_ring_rq_db(&qhp->wq, idx, wqe); idx 1335 drivers/infiniband/hw/cxgb4/qp.c ring_kernel_rq_db(qhp, idx); idx 1363 drivers/infiniband/hw/cxgb4/qp.c u16 idx = 0; idx 1412 drivers/infiniband/hw/cxgb4/qp.c idx += DIV_ROUND_UP(len16 * 16, T4_EQ_ENTRY_SIZE); idx 1417 drivers/infiniband/hw/cxgb4/qp.c if (idx) idx 1418 drivers/infiniband/hw/cxgb4/qp.c t4_ring_srq_db(&srq->wq, idx, len16, wqe); idx 1804 drivers/infiniband/hw/cxgb4/qp.c qhp->srq->idx); idx 2503 drivers/infiniband/hw/cxgb4/qp.c res->u.srq.srqid = cpu_to_be32(srq->idx); idx 2615 drivers/infiniband/hw/cxgb4/qp.c res->u.srq.srqid = cpu_to_be32(srq->idx); idx 2629 drivers/infiniband/hw/cxgb4/qp.c __func__, srq->idx, wq->qid, srq->pdid, wq->queue, idx 2708 drivers/infiniband/hw/cxgb4/qp.c srq->idx = c4iw_alloc_srq_idx(&rhp->rdev); idx 2709 drivers/infiniband/hw/cxgb4/qp.c if (srq->idx < 0) { idx 2778 drivers/infiniband/hw/cxgb4/qp.c __func__, srq->wq.qid, srq->idx, srq->wq.size, idx 2794 drivers/infiniband/hw/cxgb4/qp.c c4iw_free_srq_idx(&rhp->rdev, srq->idx); idx 2814 drivers/infiniband/hw/cxgb4/qp.c c4iw_free_srq_idx(&rhp->rdev, srq->idx); idx 429 drivers/infiniband/hw/cxgb4/resource.c int idx; idx 431 drivers/infiniband/hw/cxgb4/resource.c idx = c4iw_id_alloc(&rdev->resource.srq_table); idx 433 drivers/infiniband/hw/cxgb4/resource.c if (idx == -1) { idx 442 drivers/infiniband/hw/cxgb4/resource.c return idx; idx 445 drivers/infiniband/hw/cxgb4/resource.c void c4iw_free_srq_idx(struct c4iw_rdev *rdev, int idx) idx 447 drivers/infiniband/hw/cxgb4/resource.c c4iw_id_free(&rdev->resource.srq_table, idx); idx 95 drivers/infiniband/hw/cxgb4/restrack.c static int fill_swsqe(struct sk_buff *msg, struct t4_sq *sq, u16 idx, idx 98 drivers/infiniband/hw/cxgb4/restrack.c if (rdma_nl_put_driver_u32(msg, "idx", idx)) idx 310 drivers/infiniband/hw/cxgb4/restrack.c static int fill_cqe(struct sk_buff *msg, struct t4_cqe *cqe, u16 idx, idx 313 drivers/infiniband/hw/cxgb4/restrack.c if (rdma_nl_put_driver_u32(msg, qstr, idx)) idx 339 drivers/infiniband/hw/cxgb4/restrack.c u16 idx; idx 341 drivers/infiniband/hw/cxgb4/restrack.c idx = (cq->cidx > 0) ? cq->cidx - 1 : cq->size - 1; idx 342 drivers/infiniband/hw/cxgb4/restrack.c if (fill_cqe(msg, cqes, idx, "hwcq_idx")) idx 344 drivers/infiniband/hw/cxgb4/restrack.c idx = cq->cidx; idx 345 drivers/infiniband/hw/cxgb4/restrack.c if (fill_cqe(msg, cqes + 1, idx, "hwcq_idx")) idx 356 drivers/infiniband/hw/cxgb4/restrack.c u16 idx; idx 361 drivers/infiniband/hw/cxgb4/restrack.c idx = cq->sw_cidx; idx 362 drivers/infiniband/hw/cxgb4/restrack.c if (fill_cqe(msg, cqes, idx, "swcq_idx")) idx 366 drivers/infiniband/hw/cxgb4/restrack.c idx = (cq->sw_pidx > 0) ? cq->sw_pidx - 1 : cq->size - 1; idx 367 drivers/infiniband/hw/cxgb4/restrack.c if (fill_cqe(msg, cqes + 1, idx, "swcq_idx")) idx 384 drivers/infiniband/hw/cxgb4/restrack.c u16 idx; idx 401 drivers/infiniband/hw/cxgb4/restrack.c idx = (cq.cidx > 0) ? cq.cidx - 1 : cq.size - 1; idx 402 drivers/infiniband/hw/cxgb4/restrack.c hwcqes[0] = chp->cq.queue[idx]; idx 404 drivers/infiniband/hw/cxgb4/restrack.c idx = cq.cidx; idx 405 drivers/infiniband/hw/cxgb4/restrack.c hwcqes[1] = chp->cq.queue[idx]; idx 411 drivers/infiniband/hw/cxgb4/restrack.c idx = (cq.sw_pidx > 0) ? cq.sw_pidx - 1 : cq.size - 1; idx 412 drivers/infiniband/hw/cxgb4/restrack.c swcqes[1] = chp->cq.sw_queue[idx]; idx 313 drivers/infiniband/hw/cxgb4/t4.h u16 idx; idx 1411 drivers/infiniband/hw/hfi1/chip.c void *context, int idx, int mode, u64 data) idx 1415 drivers/infiniband/hw/hfi1/chip.c if (dd->per_sdma && idx < dd->num_sdma) idx 1416 drivers/infiniband/hw/hfi1/chip.c return dd->per_sdma[idx].err_cnt; idx 1421 drivers/infiniband/hw/hfi1/chip.c void *context, int idx, int mode, u64 data) idx 1425 drivers/infiniband/hw/hfi1/chip.c if (dd->per_sdma && idx < dd->num_sdma) idx 1426 drivers/infiniband/hw/hfi1/chip.c return dd->per_sdma[idx].sdma_int_cnt; idx 1431 drivers/infiniband/hw/hfi1/chip.c void *context, int idx, int mode, u64 data) idx 1435 drivers/infiniband/hw/hfi1/chip.c if (dd->per_sdma && idx < dd->num_sdma) idx 1436 drivers/infiniband/hw/hfi1/chip.c return dd->per_sdma[idx].idle_int_cnt; idx 1441 drivers/infiniband/hw/hfi1/chip.c void *context, int idx, int mode, idx 1446 drivers/infiniband/hw/hfi1/chip.c if (dd->per_sdma && idx < dd->num_sdma) idx 1447 drivers/infiniband/hw/hfi1/chip.c return dd->per_sdma[idx].progress_int_cnt; idx 11043 drivers/infiniband/hw/hfi1/chip.c vl_arb_lock_cache(struct hfi1_pportdata *ppd, int idx) idx 11045 drivers/infiniband/hw/hfi1/chip.c if (idx != LO_PRIO_TABLE && idx != HI_PRIO_TABLE) idx 11047 drivers/infiniband/hw/hfi1/chip.c spin_lock(&ppd->vl_arb_cache[idx].lock); idx 11048 drivers/infiniband/hw/hfi1/chip.c return &ppd->vl_arb_cache[idx]; idx 11051 drivers/infiniband/hw/hfi1/chip.c static inline void vl_arb_unlock_cache(struct hfi1_pportdata *ppd, int idx) idx 11053 drivers/infiniband/hw/hfi1/chip.c spin_unlock(&ppd->vl_arb_cache[idx].lock); idx 11252 drivers/infiniband/hw/hfi1/chip.c static void nonzero_msg(struct hfi1_devdata *dd, int idx, const char *what, idx 11257 drivers/infiniband/hw/hfi1/chip.c what, (int)limit, idx); idx 11399 drivers/infiniband/hw/hfi1/chip.c #define valid_vl(idx) ((idx) < TXE_NUM_DATA_VL || (idx) == 15) idx 13068 drivers/infiniband/hw/hfi1/chip.c u16 idx = src / BITS_PER_REGISTER; idx 13071 drivers/infiniband/hw/hfi1/chip.c reg = read_csr(dd, CCE_INT_MASK + (8 * idx)); idx 13076 drivers/infiniband/hw/hfi1/chip.c write_csr(dd, CCE_INT_MASK + (8 * idx), reg); idx 14052 drivers/infiniband/hw/hfi1/chip.c u8 hfi1_get_qp_map(struct hfi1_devdata *dd, u8 idx) idx 14054 drivers/infiniband/hw/hfi1/chip.c u64 reg = read_csr(dd, RCV_QP_MAP_TABLE + (idx / 8) * 8); idx 14056 drivers/infiniband/hw/hfi1/chip.c reg >>= (idx % 8) * 8; idx 14276 drivers/infiniband/hw/hfi1/chip.c unsigned idx, regoff, regidx; idx 14279 drivers/infiniband/hw/hfi1/chip.c idx = rmt->used + ((qpn << n) ^ i); idx 14280 drivers/infiniband/hw/hfi1/chip.c regoff = (idx % 8) * 8; idx 14281 drivers/infiniband/hw/hfi1/chip.c regidx = idx / 8; idx 14326 drivers/infiniband/hw/hfi1/chip.c int i, idx, regoff, regidx, start; idx 14356 drivers/infiniband/hw/hfi1/chip.c for (i = start, idx = rmt->used; i < dd->num_rcv_contexts; idx 14357 drivers/infiniband/hw/hfi1/chip.c i++, idx++) { idx 14359 drivers/infiniband/hw/hfi1/chip.c regoff = (idx % 8) * 8; idx 14360 drivers/infiniband/hw/hfi1/chip.c regidx = idx / 8; idx 848 drivers/infiniband/hw/hfi1/chip.h static inline int vl_from_idx(int idx) idx 850 drivers/infiniband/hw/hfi1/chip.h return (idx == C_VL_15 ? 15 : idx); idx 1451 drivers/infiniband/hw/hfi1/chip.h u8 hfi1_get_qp_map(struct hfi1_devdata *dd, u8 idx); idx 202 drivers/infiniband/hw/hfi1/driver.c u32 idx = rhf_egr_index(rhf), offset = rhf_egr_buf_offset(rhf); idx 204 drivers/infiniband/hw/hfi1/driver.c *update |= !(idx & (rcd->egrbufs.threshold - 1)) && !offset; idx 205 drivers/infiniband/hw/hfi1/driver.c return (void *)(((u64)(rcd->egrbufs.rcvtids[idx].addr)) + idx 203 drivers/infiniband/hw/hfi1/exp_rcv.h hfi1_idx_to_tid_group(struct hfi1_ctxtdata *rcd, u16 idx) idx 205 drivers/infiniband/hw/hfi1/exp_rcv.h return &rcd->groups[idx]; idx 309 drivers/infiniband/hw/hfi1/file_ops.c int idx; idx 311 drivers/infiniband/hw/hfi1/file_ops.c idx = srcu_read_lock(&fd->pq_srcu); idx 314 drivers/infiniband/hw/hfi1/file_ops.c srcu_read_unlock(&fd->pq_srcu, idx); idx 319 drivers/infiniband/hw/hfi1/file_ops.c srcu_read_unlock(&fd->pq_srcu, idx); idx 326 drivers/infiniband/hw/hfi1/file_ops.c srcu_read_unlock(&fd->pq_srcu, idx); idx 346 drivers/infiniband/hw/hfi1/file_ops.c srcu_read_unlock(&fd->pq_srcu, idx); idx 1731 drivers/infiniband/hw/hfi1/hfi.h u8 sc5, u8 idx, u32 slid, bool force) idx 1745 drivers/infiniband/hw/hfi1/hfi.h if (ingress_pkey_matches_entry(pkey, ppd->pkeys[idx])) idx 1929 drivers/infiniband/hw/hfi1/init.c u16 order, idx = 0; idx 1967 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].addr = idx 1970 drivers/infiniband/hw/hfi1/init.c &rcd->egrbufs.buffers[idx].dma, idx 1972 drivers/infiniband/hw/hfi1/init.c if (rcd->egrbufs.buffers[idx].addr) { idx 1973 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].len = idx 1976 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].addr; idx 1978 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].dma; idx 1981 drivers/infiniband/hw/hfi1/init.c idx++; idx 2007 drivers/infiniband/hw/hfi1/init.c if (idx == 0) { idx 2017 drivers/infiniband/hw/hfi1/init.c for (i = 0, j = 0, offset = 0; j < idx; i++) { idx 2038 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.numbufs = idx; idx 2077 drivers/infiniband/hw/hfi1/init.c for (idx = 0; idx < rcd->egrbufs.alloced; idx++) { idx 2078 drivers/infiniband/hw/hfi1/init.c hfi1_put_tid(dd, rcd->eager_base + idx, PT_EAGER, idx 2079 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.rcvtids[idx].dma, order); idx 2086 drivers/infiniband/hw/hfi1/init.c for (idx = 0; idx < rcd->egrbufs.alloced && idx 2087 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].addr; idx 2088 drivers/infiniband/hw/hfi1/init.c idx++) { idx 2090 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].len, idx 2091 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].addr, idx 2092 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].dma); idx 2093 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].addr = NULL; idx 2094 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].dma = 0; idx 2095 drivers/infiniband/hw/hfi1/init.c rcd->egrbufs.buffers[idx].len = 0; idx 116 drivers/infiniband/hw/hfi1/iowait.c uint idx, uint top_idx) idx 125 drivers/infiniband/hw/hfi1/iowait.c return idx; idx 442 drivers/infiniband/hw/hfi1/iowait.h uint idx, uint top_idx); idx 118 drivers/infiniband/hw/hfi1/msix.c u32 idx, enum irq_type type) idx 155 drivers/infiniband/hw/hfi1/msix.c dd->unit, idx); idx 160 drivers/infiniband/hw/hfi1/msix.c dd->unit, idx); idx 174 drivers/infiniband/hw/hfi1/msix.c err_info, irq, idx, ret); idx 601 drivers/infiniband/hw/hfi1/sysfs.c #define temp2str(temp, buf, size, idx) \ idx 602 drivers/infiniband/hw/hfi1/sysfs.c scnprintf((buf) + (idx), (size) - (idx), "%u.%02u ", \ idx 619 drivers/infiniband/hw/hfi1/sysfs.c int idx = 0; idx 621 drivers/infiniband/hw/hfi1/sysfs.c idx += temp2str(temp.curr, buf, PAGE_SIZE, idx); idx 622 drivers/infiniband/hw/hfi1/sysfs.c idx += temp2str(temp.lo_lim, buf, PAGE_SIZE, idx); idx 623 drivers/infiniband/hw/hfi1/sysfs.c idx += temp2str(temp.hi_lim, buf, PAGE_SIZE, idx); idx 624 drivers/infiniband/hw/hfi1/sysfs.c idx += temp2str(temp.crit_lim, buf, PAGE_SIZE, idx); idx 625 drivers/infiniband/hw/hfi1/sysfs.c idx += scnprintf(buf + idx, PAGE_SIZE - idx, idx 628 drivers/infiniband/hw/hfi1/sysfs.c ret = idx; idx 927 drivers/infiniband/hw/hfi1/tid_rdma.c list[setcount].idx = pageidx; idx 930 drivers/infiniband/hw/hfi1/tid_rdma.c list[setcount].idx, idx 971 drivers/infiniband/hw/hfi1/tid_rdma.c u32 *idx, u32 pages, u32 sets) idx 980 drivers/infiniband/hw/hfi1/tid_rdma.c list[sets].idx = *idx; idx 982 drivers/infiniband/hw/hfi1/tid_rdma.c *idx += maxpages; idx 1018 drivers/infiniband/hw/hfi1/tid_rdma.c u32 idx, sets = 0, i; idx 1024 drivers/infiniband/hw/hfi1/tid_rdma.c for (idx = 0, i = 0, vm1 = NULL; i < npages; i += 2) { idx 1034 drivers/infiniband/hw/hfi1/tid_rdma.c sets = tid_flush_pages(list, &idx, pagecnt, sets); idx 1036 drivers/infiniband/hw/hfi1/tid_rdma.c list[sets].idx = idx++; idx 1040 drivers/infiniband/hw/hfi1/tid_rdma.c list[sets++].idx = idx++; idx 1051 drivers/infiniband/hw/hfi1/tid_rdma.c sets = tid_flush_pages(list, &idx, pagecnt, sets); idx 1061 drivers/infiniband/hw/hfi1/tid_rdma.c sets = tid_flush_pages(list, &idx, npages - idx, sets); idx 1144 drivers/infiniband/hw/hfi1/tid_rdma.c pages[pset->idx], idx 1513 drivers/infiniband/hw/hfi1/tid_rdma.c flow->idx = qpriv->flow_state.index; idx 1752 drivers/infiniband/hw/hfi1/tid_rdma.c ((flow->idx & TID_RDMA_DESTQP_FLOW_MASK) << idx 1935 drivers/infiniband/hw/hfi1/tid_rdma.c flow->idx = (flow->tid_qpn >> TID_RDMA_DESTQP_FLOW_SHIFT) & idx 2792 drivers/infiniband/hw/hfi1/tid_rdma.c flow->idx); idx 2953 drivers/infiniband/hw/hfi1/tid_rdma.c flow->idx); idx 3913 drivers/infiniband/hw/hfi1/tid_rdma.c ((flow->idx & TID_RDMA_DESTQP_FLOW_MASK) << idx 4100 drivers/infiniband/hw/hfi1/tid_rdma.c flow->idx = (flow->tid_qpn >> TID_RDMA_DESTQP_FLOW_SHIFT) & idx 4336 drivers/infiniband/hw/hfi1/tid_rdma.c rcd->flows[flow->idx].psn = psn & HFI1_KDETH_BTH_SEQ_MASK; idx 4447 drivers/infiniband/hw/hfi1/tid_rdma.c ((flow->idx & TID_RDMA_DESTQP_FLOW_MASK) << idx 4867 drivers/infiniband/hw/hfi1/tid_rdma.c u32 psn, generation, idx, gen_next; idx 4915 drivers/infiniband/hw/hfi1/tid_rdma.c for (idx = qpriv->r_tid_tail; ; idx++) { idx 4918 drivers/infiniband/hw/hfi1/tid_rdma.c if (idx > rvt_size_atomic(&dev->rdi)) idx 4919 drivers/infiniband/hw/hfi1/tid_rdma.c idx = 0; idx 4920 drivers/infiniband/hw/hfi1/tid_rdma.c e = &qp->s_ack_queue[idx]; idx 4951 drivers/infiniband/hw/hfi1/tid_rdma.c if (idx == qp->s_tail_ack_queue) idx 5514 drivers/infiniband/hw/hfi1/tid_rdma.c read_r_next_psn(dd, rcd->ctxt, flow->idx); idx 152 drivers/infiniband/hw/hfi1/tid_rdma.h u8 idx: 8; idx 187 drivers/infiniband/hw/hfi1/tid_rdma.h u8 idx; idx 457 drivers/infiniband/hw/hfi1/trace_tid.h TP_PROTO(struct rvt_qp *qp, u32 index, u16 idx, u16 count), idx 458 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(qp, index, idx, count), idx 463 drivers/infiniband/hw/hfi1/trace_tid.h __field(u16, idx) idx 470 drivers/infiniband/hw/hfi1/trace_tid.h __entry->idx = idx; idx 478 drivers/infiniband/hw/hfi1/trace_tid.h __entry->idx, idx 485 drivers/infiniband/hw/hfi1/trace_tid.h TP_PROTO(struct rvt_qp *qp, u32 index, u16 idx, u16 count), idx 486 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(qp, index, idx, count) idx 497 drivers/infiniband/hw/hfi1/trace_tid.h __field(int, idx) idx 517 drivers/infiniband/hw/hfi1/trace_tid.h __entry->idx = flow->idx; idx 540 drivers/infiniband/hw/hfi1/trace_tid.h __entry->idx, idx 690 drivers/infiniband/hw/hfi1/trace_tid.h __field(u16, idx) idx 698 drivers/infiniband/hw/hfi1/trace_tid.h __entry->idx = hfi1_trace_get_tid_idx(ent); idx 706 drivers/infiniband/hw/hfi1/trace_tid.h __entry->idx, idx 161 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 166 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = sde->this_idx; idx 173 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 189 drivers/infiniband/hw/hfi1/trace_tx.h TP_PROTO(struct hfi1_devdata *dd, u32 sel, u8 vl, u8 idx), idx 190 drivers/infiniband/hw/hfi1/trace_tx.h TP_ARGS(dd, sel, vl, idx), idx 194 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 199 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = idx; idx 203 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 348 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 352 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = sde->this_idx; idx 356 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 376 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 379 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = sde->this_idx; idx 384 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 412 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 420 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = sde->this_idx; idx 426 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 448 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 456 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = sde->this_idx; idx 461 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 476 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 480 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = sde->this_idx; idx 484 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 631 drivers/infiniband/hw/hfi1/trace_tx.h TP_PROTO(struct hfi1_devdata *dd, u16 ctxt, u8 subctxt, u16 idx, idx 633 drivers/infiniband/hw/hfi1/trace_tx.h TP_ARGS(dd, ctxt, subctxt, idx, state, code), idx 638 drivers/infiniband/hw/hfi1/trace_tx.h __field(u16, idx) idx 646 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = idx; idx 652 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, show_usdma_complete_state(__entry->state), idx 669 drivers/infiniband/hw/hfi1/trace_tx.h __field(u8, idx) idx 680 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = ahgidx; idx 691 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 780 drivers/infiniband/hw/hfi1/trace_tx.h TP_PROTO(struct rvt_qp *qp, struct rvt_swqe *wqe, u32 idx), idx 781 drivers/infiniband/hw/hfi1/trace_tx.h TP_ARGS(qp, wqe, idx), idx 789 drivers/infiniband/hw/hfi1/trace_tx.h __field(u32, idx) idx 801 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx = idx; idx 812 drivers/infiniband/hw/hfi1/trace_tx.h __entry->idx, idx 194 drivers/infiniband/hw/hfi1/user_exp_rcv.c unsigned int idx, idx 204 drivers/infiniband/hw/hfi1/user_exp_rcv.c pages = &node->pages[idx]; idx 206 drivers/infiniband/hw/hfi1/user_exp_rcv.c pages = &tidbuf->pages[idx]; idx 637 drivers/infiniband/hw/hfi1/user_exp_rcv.c list[setcount].idx = pageidx; idx 688 drivers/infiniband/hw/hfi1/user_exp_rcv.c u16 idx; idx 697 drivers/infiniband/hw/hfi1/user_exp_rcv.c for (idx = 0; idx < grp->size; idx++) { idx 698 drivers/infiniband/hw/hfi1/user_exp_rcv.c if (!(grp->map & (1 << idx))) { idx 699 drivers/infiniband/hw/hfi1/user_exp_rcv.c useidx = idx; idx 702 drivers/infiniband/hw/hfi1/user_exp_rcv.c rcv_array_wc_fill(dd, grp->base + idx); idx 705 drivers/infiniband/hw/hfi1/user_exp_rcv.c idx = 0; idx 706 drivers/infiniband/hw/hfi1/user_exp_rcv.c while (idx < count) { idx 707 drivers/infiniband/hw/hfi1/user_exp_rcv.c u16 npages, pageidx, setidx = start + idx; idx 724 drivers/infiniband/hw/hfi1/user_exp_rcv.c pageidx = tbuf->psets[setidx].idx; idx 738 drivers/infiniband/hw/hfi1/user_exp_rcv.c idx++; idx 745 drivers/infiniband/hw/hfi1/user_exp_rcv.c return idx; idx 54 drivers/infiniband/hw/hfi1/user_exp_rcv.h u16 idx; idx 96 drivers/infiniband/hw/hfi1/user_sdma.c u16 idx, enum hfi1_sdma_comp_state state, idx 352 drivers/infiniband/hw/hfi1/user_sdma.c unsigned long idx = 0; idx 362 drivers/infiniband/hw/hfi1/user_sdma.c if (iovec[idx].iov_len < sizeof(info) + sizeof(req->hdr)) { idx 367 drivers/infiniband/hw/hfi1/user_sdma.c iovec[idx].iov_len, sizeof(info) + sizeof(req->hdr)); idx 370 drivers/infiniband/hw/hfi1/user_sdma.c ret = copy_from_user(&info, iovec[idx].iov_base, sizeof(info)); idx 455 drivers/infiniband/hw/hfi1/user_sdma.c ret = copy_from_user(&req->hdr, iovec[idx].iov_base + sizeof(info), idx 519 drivers/infiniband/hw/hfi1/user_sdma.c idx++; idx 526 drivers/infiniband/hw/hfi1/user_sdma.c iovec + idx++, idx 548 drivers/infiniband/hw/hfi1/user_sdma.c u16 ntids = iovec[idx].iov_len / sizeof(*req->tids); idx 562 drivers/infiniband/hw/hfi1/user_sdma.c tmp = memdup_user(iovec[idx].iov_base, idx 573 drivers/infiniband/hw/hfi1/user_sdma.c idx++; idx 612 drivers/infiniband/hw/hfi1/user_sdma.c *count += idx; idx 1274 drivers/infiniband/hw/hfi1/user_sdma.c int idx = 0; idx 1284 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set(ahg, idx, array_size, 0, 0, 12, idx 1286 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1287 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1289 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set(ahg, idx, array_size, 3, 0, 16, idx 1291 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1292 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1303 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set(ahg, idx, array_size, 6, 0, 16, idx 1305 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1306 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1307 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set(ahg, idx, array_size, 6, 16, 16, idx 1309 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1310 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1312 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set(ahg, idx, array_size, 15, 0, 16, idx 1314 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1315 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1316 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set(ahg, idx, array_size, 15, 16, 16, idx 1318 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1319 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1346 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set( idx 1347 drivers/infiniband/hw/hfi1/user_sdma.c ahg, idx, array_size, 7, 0, 16, idx 1351 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1352 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1369 drivers/infiniband/hw/hfi1/user_sdma.c idx = ahg_header_set(ahg, idx, array_size, idx 1371 drivers/infiniband/hw/hfi1/user_sdma.c if (idx < 0) idx 1372 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1377 drivers/infiniband/hw/hfi1/user_sdma.c req->ahg_idx, ahg, idx, tidval); idx 1380 drivers/infiniband/hw/hfi1/user_sdma.c datalen, req->ahg_idx, idx, idx 1384 drivers/infiniband/hw/hfi1/user_sdma.c return idx; idx 1475 drivers/infiniband/hw/hfi1/user_sdma.c u16 idx, enum hfi1_sdma_comp_state state, idx 1479 drivers/infiniband/hw/hfi1/user_sdma.c cq->comps[idx].errcode = -ret; idx 1481 drivers/infiniband/hw/hfi1/user_sdma.c cq->comps[idx].status = state; idx 1483 drivers/infiniband/hw/hfi1/user_sdma.c idx, state, ret); idx 95 drivers/infiniband/hw/hfi1/user_sdma.h static inline int ahg_header_set(u32 *arr, int idx, size_t array_size, idx 98 drivers/infiniband/hw/hfi1/user_sdma.h if ((size_t)idx >= array_size) idx 100 drivers/infiniband/hw/hfi1/user_sdma.h arr[idx++] = sdma_build_ahg_descriptor(value, dw, bit, width); idx 101 drivers/infiniband/hw/hfi1/user_sdma.h return idx; idx 103 drivers/infiniband/hw/hfi1/vnic.h u8 idx; idx 449 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[rxq->idx].rx_oversize++; idx 451 drivers/infiniband/hw/hfi1/vnic_main.c vinfo->stats[rxq->idx].rx_runt++; idx 492 drivers/infiniband/hw/hfi1/vnic_main.c hfi1_vnic_update_rx_counters(vinfo, rxq->idx, skb, rc); idx 514 drivers/infiniband/hw/hfi1/vnic_main.c v_dbg("napi %d budget %d\n", rxq->idx, budget); idx 517 drivers/infiniband/hw/hfi1/vnic_main.c v_dbg("napi %d work_done %d\n", rxq->idx, work_done); idx 841 drivers/infiniband/hw/hfi1/vnic_main.c rxq->idx = i; idx 271 drivers/infiniband/hw/hns/hns_roce_alloc.c int idx = 0; idx 283 drivers/infiniband/hw/hns/hns_roce_alloc.c if (idx >= start) { idx 288 drivers/infiniband/hw/hns/hns_roce_alloc.c idx++; idx 5313 drivers/infiniband/hw/hns/hns_roce_hw_v2.c u64 idx; idx 5351 drivers/infiniband/hw/hns/hns_roce_hw_v2.c idx = i * (bt_chk_sz / BA_BYTE_LEN) + j; idx 5355 drivers/infiniband/hw/hns/hns_roce_hw_v2.c * idx; idx 5359 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf[idx], idx 5360 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); idx 5363 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, buf_chk_sz, eq->buf[idx], idx 5364 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); idx 5555 drivers/infiniband/hw/hns/hns_roce_hw_v2.c u64 idx; idx 5650 drivers/infiniband/hw/hns/hns_roce_hw_v2.c idx = i * bt_chk_sz / BA_BYTE_LEN + j; idx 5655 drivers/infiniband/hw/hns/hns_roce_hw_v2.c * idx; idx 5659 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf[idx] = dma_alloc_coherent(dev, size, idx 5660 drivers/infiniband/hw/hns/hns_roce_hw_v2.c &(eq->buf_dma[idx]), idx 5662 drivers/infiniband/hw/hns/hns_roce_hw_v2.c if (!eq->buf[idx]) idx 5665 drivers/infiniband/hw/hns/hns_roce_hw_v2.c *(eq->bt_l1[i] + j) = eq->buf_dma[idx]; idx 5697 drivers/infiniband/hw/hns/hns_roce_hw_v2.c idx = i * bt_chk_sz / BA_BYTE_LEN + j; idx 5698 drivers/infiniband/hw/hns/hns_roce_hw_v2.c dma_free_coherent(dev, buf_chk_sz, eq->buf[idx], idx 5699 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); idx 5724 drivers/infiniband/hw/hns/hns_roce_hw_v2.c idx = i * bt_chk_sz / BA_BYTE_LEN + j; idx 5726 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf[idx], idx 5727 drivers/infiniband/hw/hns/hns_roce_hw_v2.c eq->buf_dma[idx]); idx 206 drivers/infiniband/hw/i40iw/i40iw.h u32 idx; idx 60 drivers/infiniband/hw/i40iw/i40iw_hmc.c u32 idx, idx 68 drivers/infiniband/hw/i40iw/i40iw_hmc.c hmc_info->hmc_obj[type].size * idx; idx 90 drivers/infiniband/hw/i40iw/i40iw_hmc.c u32 idx, idx 98 drivers/infiniband/hw/i40iw/i40iw_hmc.c hmc_info->hmc_obj[type].size * idx; idx 113 drivers/infiniband/hw/i40iw/i40iw_hmc.c u32 idx, idx 121 drivers/infiniband/hw/i40iw/i40iw_hmc.c entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15)); idx 130 drivers/infiniband/hw/i40iw/i40iw_hmc.c static inline void i40iw_clr_sd_entry(u32 idx, enum i40iw_sd_entry_type type, idx 137 drivers/infiniband/hw/i40iw/i40iw_hmc.c entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15)); idx 231 drivers/infiniband/hw/i40iw/i40iw_hmc.c u16 idx; idx 233 drivers/infiniband/hw/i40iw/i40iw_hmc.c for (idx = 0; idx < I40IW_MAX_PE_ENABLED_VF_COUNT; idx++) { idx 234 drivers/infiniband/hw/i40iw/i40iw_hmc.c if (dev->vf_dev[idx] && idx 235 drivers/infiniband/hw/i40iw/i40iw_hmc.c ((u8)dev->vf_dev[idx]->pmf_index == hmc_fn_id)) { idx 236 drivers/infiniband/hw/i40iw/i40iw_hmc.c vf_dev = dev->vf_dev[idx]; idx 252 drivers/infiniband/hw/i40iw/i40iw_hmc.c u16 idx; idx 254 drivers/infiniband/hw/i40iw/i40iw_hmc.c for (idx = 0; idx < I40IW_MAX_PE_ENABLED_VF_COUNT; idx++) { idx 255 drivers/infiniband/hw/i40iw/i40iw_hmc.c if (dev->vf_dev[idx] && idx 256 drivers/infiniband/hw/i40iw/i40iw_hmc.c ((u8)dev->vf_dev[idx]->pmf_index == hmc_fn_id)) { idx 257 drivers/infiniband/hw/i40iw/i40iw_hmc.c hmc_info = &dev->vf_dev[idx]->hmc_info; idx 694 drivers/infiniband/hw/i40iw/i40iw_hmc.c u32 idx, idx 704 drivers/infiniband/hw/i40iw/i40iw_hmc.c sd_idx = idx / I40IW_HMC_PD_CNT_IN_SD; idx 705 drivers/infiniband/hw/i40iw/i40iw_hmc.c rel_pd_idx = idx % I40IW_HMC_PD_CNT_IN_SD; idx 725 drivers/infiniband/hw/i40iw/i40iw_hmc.c I40IW_INVALIDATE_PF_HMC_PD(hw, sd_idx, idx); idx 727 drivers/infiniband/hw/i40iw/i40iw_hmc.c I40IW_INVALIDATE_VF_HMC_PD(hw, sd_idx, idx, idx 747 drivers/infiniband/hw/i40iw/i40iw_hmc.c enum i40iw_status_code i40iw_prep_remove_sd_bp(struct i40iw_hmc_info *hmc_info, u32 idx) idx 751 drivers/infiniband/hw/i40iw/i40iw_hmc.c sd_entry = &hmc_info->sd_table.sd_entry[idx]; idx 768 drivers/infiniband/hw/i40iw/i40iw_hmc.c u32 idx) idx 772 drivers/infiniband/hw/i40iw/i40iw_hmc.c sd_entry = &hmc_info->sd_table.sd_entry[idx]; idx 234 drivers/infiniband/hw/i40iw/i40iw_hmc.h struct i40iw_hmc_info *hmc_info, u32 idx, bool is_pf); idx 235 drivers/infiniband/hw/i40iw/i40iw_hmc.h enum i40iw_status_code i40iw_prep_remove_sd_bp(struct i40iw_hmc_info *hmc_info, u32 idx); idx 236 drivers/infiniband/hw/i40iw/i40iw_hmc.h enum i40iw_status_code i40iw_prep_remove_pd_page(struct i40iw_hmc_info *hmc_info, u32 idx); idx 202 drivers/infiniband/hw/i40iw/i40iw_main.c i40iw_enable_intr(&iwdev->sc_dev, iwdev->iw_msixtbl[0].idx); idx 270 drivers/infiniband/hw/i40iw/i40iw_main.c i40iw_wr32(dev->hw, I40E_PFINT_DYN_CTLN(msix_vec->idx - 1), 0); idx 272 drivers/infiniband/hw/i40iw/i40iw_main.c i40iw_wr32(dev->hw, I40E_VFINT_DYN_CTLN1(msix_vec->idx - 1), 0); idx 810 drivers/infiniband/hw/i40iw/i40iw_main.c iwceq->msix_idx = msix_vec->idx; idx 816 drivers/infiniband/hw/i40iw/i40iw_main.c i40iw_enable_intr(&iwdev->sc_dev, msix_vec->idx); idx 919 drivers/infiniband/hw/i40iw/i40iw_main.c i40iw_enable_intr(dev, iwdev->iw_msixtbl[0].idx); idx 1044 drivers/infiniband/hw/i40iw/i40iw_main.c static void i40iw_del_macip_entry(struct i40iw_device *iwdev, u8 idx) idx 1061 drivers/infiniband/hw/i40iw/i40iw_main.c cqp_info->in.u.del_local_mac_ipaddr_entry.entry_idx = idx; idx 1076 drivers/infiniband/hw/i40iw/i40iw_main.c u8 idx) idx 1095 drivers/infiniband/hw/i40iw/i40iw_main.c info->entry_idx = idx; idx 1457 drivers/infiniband/hw/i40iw/i40iw_main.c iwdev->iw_msixtbl[i].idx = ldev->msix_entries[i].entry; idx 1471 drivers/infiniband/hw/i40iw/i40iw_main.c iw_qvinfo->v_idx = iwdev->iw_msixtbl[i].idx; idx 119 drivers/infiniband/hw/i40iw/i40iw_pble.c struct sd_pd_idx *idx) idx 121 drivers/infiniband/hw/i40iw/i40iw_pble.c idx->sd_idx = (u32)(pble_rsrc->next_fpm_addr) / I40IW_HMC_DIRECT_BP_SIZE; idx 122 drivers/infiniband/hw/i40iw/i40iw_pble.c idx->pd_idx = (u32)(pble_rsrc->next_fpm_addr) / I40IW_HMC_PAGED_BP_SIZE; idx 123 drivers/infiniband/hw/i40iw/i40iw_pble.c idx->rel_pd_idx = (idx->pd_idx % I40IW_HMC_PD_CNT_IN_SD); idx 137 drivers/infiniband/hw/i40iw/i40iw_pble.c struct sd_pd_idx *idx = &info->idx; idx 146 drivers/infiniband/hw/i40iw/i40iw_pble.c info->idx.sd_idx, idx 154 drivers/infiniband/hw/i40iw/i40iw_pble.c offset = idx->rel_pd_idx << I40IW_HMC_PAGED_BP_SHIFT; idx 262 drivers/infiniband/hw/i40iw/i40iw_pble.c u32 rel_pd_idx = info->idx.rel_pd_idx; idx 263 drivers/infiniband/hw/i40iw/i40iw_pble.c u32 pd_idx = info->idx.pd_idx; idx 270 drivers/infiniband/hw/i40iw/i40iw_pble.c info->idx.sd_idx, I40IW_SD_TYPE_PAGED, idx 300 drivers/infiniband/hw/i40iw/i40iw_pble.c vf_pble_info.first_pd_index = info->idx.rel_pd_idx; idx 304 drivers/infiniband/hw/i40iw/i40iw_pble.c vf_pble_info.sd_index = info->idx.sd_idx; idx 331 drivers/infiniband/hw/i40iw/i40iw_pble.c struct sd_pd_idx *idx = &info.idx; idx 348 drivers/infiniband/hw/i40iw/i40iw_pble.c get_sd_pd_idx(pble_rsrc, idx); idx 349 drivers/infiniband/hw/i40iw/i40iw_pble.c sd_entry = &hmc_info->sd_table.sd_entry[idx->sd_idx]; idx 350 drivers/infiniband/hw/i40iw/i40iw_pble.c pages = (idx->rel_pd_idx) ? (I40IW_HMC_PD_CNT_IN_SD - idx 351 drivers/infiniband/hw/i40iw/i40iw_pble.c idx->rel_pd_idx) : I40IW_HMC_PD_CNT_IN_SD; idx 358 drivers/infiniband/hw/i40iw/i40iw_pble.c sd_entry_type = (!idx->rel_pd_idx && idx 402 drivers/infiniband/hw/i40iw/i40iw_pble.c sd_reg_val, idx->sd_idx, idx 481 drivers/infiniband/hw/i40iw/i40iw_pble.c root->idx = fpm_to_idx(pble_rsrc, idx 490 drivers/infiniband/hw/i40iw/i40iw_pble.c leaf->idx = fpm_to_idx(pble_rsrc, (u64)gen_pool_virt_to_phys(pool, leaf->addr)); idx 493 drivers/infiniband/hw/i40iw/i40iw_pble.c *addr = (u64)leaf->idx; idx 526 drivers/infiniband/hw/i40iw/i40iw_pble.c lvl1->idx = fpm_to_idx(pble_rsrc, (u64)gen_pool_virt_to_phys(pool, idx 57 drivers/infiniband/hw/i40iw/i40iw_pble.h u32 idx; idx 86 drivers/infiniband/hw/i40iw/i40iw_pble.h struct sd_pd_idx idx; idx 442 drivers/infiniband/hw/i40iw/i40iw_verbs.c init_info->sq_pa = qpmr->sq_pbl.idx; idx 443 drivers/infiniband/hw/i40iw/i40iw_verbs.c init_info->rq_pa = qpmr->rq_pbl.idx; idx 1153 drivers/infiniband/hw/i40iw/i40iw_verbs.c info.first_pm_pbl_idx = cqmr->cq_pbl.idx; idx 1294 drivers/infiniband/hw/i40iw/i40iw_verbs.c u32 *idx) idx 1296 drivers/infiniband/hw/i40iw/i40iw_verbs.c *idx += 1; idx 1297 drivers/infiniband/hw/i40iw/i40iw_verbs.c if ((!(*pinfo)) || (*idx != (*pinfo)->cnt)) idx 1299 drivers/infiniband/hw/i40iw/i40iw_verbs.c *idx = 0; idx 1319 drivers/infiniband/hw/i40iw/i40iw_verbs.c u32 idx = 0; idx 1329 drivers/infiniband/hw/i40iw/i40iw_verbs.c pbl = i40iw_next_pbl_addr(pbl, &pinfo, &idx); idx 1475 drivers/infiniband/hw/i40iw/i40iw_verbs.c hmc_p->idx = palloc->level1.idx; idx 1477 drivers/infiniband/hw/i40iw/i40iw_verbs.c hmc_p->idx = palloc->level1.idx + req->sq_pages; idx 1491 drivers/infiniband/hw/i40iw/i40iw_verbs.c hmc_p->idx = palloc->level1.idx; idx 1703 drivers/infiniband/hw/i40iw/i40iw_verbs.c stag_info->first_pm_pbl_index = palloc->level1.idx; idx 1706 drivers/infiniband/hw/i40iw/i40iw_verbs.c stag_info->first_pm_pbl_index = palloc->level2.root.idx; idx 2252 drivers/infiniband/hw/i40iw/i40iw_verbs.c info.first_pm_pbl_index = palloc->level1.idx; idx 56 drivers/infiniband/hw/i40iw/i40iw_verbs.h u32 idx; idx 461 drivers/infiniband/hw/mlx4/sysfs.c int idx; idx 469 drivers/infiniband/hw/mlx4/sysfs.c idx = p->dev->dev->phys_caps.pkey_phys_table_len[p->port_num] - 1; idx 470 drivers/infiniband/hw/mlx4/sysfs.c else if (sscanf(buf, "%i", &idx) != 1 || idx 471 drivers/infiniband/hw/mlx4/sysfs.c idx >= p->dev->dev->caps.pkey_table_len[p->port_num] || idx 472 drivers/infiniband/hw/mlx4/sysfs.c idx < 0) idx 476 drivers/infiniband/hw/mlx4/sysfs.c [tab_attr->index] = idx; idx 478 drivers/infiniband/hw/mlx4/sysfs.c tab_attr->index, idx); idx 482 drivers/infiniband/hw/mlx4/sysfs.c " port %d, index %d\n", p->slave, p->port_num, idx); idx 98 drivers/infiniband/hw/mlx5/cq.c static enum ib_wc_opcode get_umr_comp(struct mlx5_ib_wq *wq, int idx) idx 100 drivers/infiniband/hw/mlx5/cq.c switch (wq->wr_data[idx]) { idx 117 drivers/infiniband/hw/mlx5/cq.c struct mlx5_ib_wq *wq, int idx) idx 155 drivers/infiniband/hw/mlx5/cq.c wc->opcode = get_umr_comp(wq, idx); idx 336 drivers/infiniband/hw/mlx5/cq.c u16 idx; idx 339 drivers/infiniband/hw/mlx5/cq.c idx = tail & (qp->sq.wqe_cnt - 1); idx 340 drivers/infiniband/hw/mlx5/cq.c if (idx == head) idx 343 drivers/infiniband/hw/mlx5/cq.c tail = qp->sq.w_list[idx].next; idx 345 drivers/infiniband/hw/mlx5/cq.c tail = qp->sq.w_list[idx].next; idx 402 drivers/infiniband/hw/mlx5/cq.c unsigned int idx; idx 404 drivers/infiniband/hw/mlx5/cq.c idx = (is_send) ? wq->last_poll : wq->tail; idx 405 drivers/infiniband/hw/mlx5/cq.c idx &= (wq->wqe_cnt - 1); idx 406 drivers/infiniband/hw/mlx5/cq.c wc->wr_id = wq->wrid[idx]; idx 411 drivers/infiniband/hw/mlx5/cq.c wq->last_poll = wq->w_list[idx].next; idx 455 drivers/infiniband/hw/mlx5/cq.c int idx; idx 499 drivers/infiniband/hw/mlx5/cq.c idx = wqe_ctr & (wq->wqe_cnt - 1); idx 500 drivers/infiniband/hw/mlx5/cq.c handle_good_req(wc, cqe64, wq, idx); idx 501 drivers/infiniband/hw/mlx5/cq.c handle_atomics(*cur_qp, cqe64, wq->last_poll, idx); idx 502 drivers/infiniband/hw/mlx5/cq.c wc->wr_id = wq->wrid[idx]; idx 503 drivers/infiniband/hw/mlx5/cq.c wq->tail = wq->wqe_head[idx] + 1; idx 527 drivers/infiniband/hw/mlx5/cq.c idx = wqe_ctr & (wq->wqe_cnt - 1); idx 528 drivers/infiniband/hw/mlx5/cq.c wc->wr_id = wq->wrid[idx]; idx 529 drivers/infiniband/hw/mlx5/cq.c wq->tail = wq->wqe_head[idx] + 1; idx 2086 drivers/infiniband/hw/mlx5/main.c unsigned long idx; idx 2099 drivers/infiniband/hw/mlx5/main.c idx = get_extended_index(vma->vm_pgoff) + bfregi->num_static_sys_pages; idx 2101 drivers/infiniband/hw/mlx5/main.c idx = get_index(vma->vm_pgoff); idx 2103 drivers/infiniband/hw/mlx5/main.c if (idx >= max_valid_idx) { idx 2105 drivers/infiniband/hw/mlx5/main.c idx, max_valid_idx); idx 2135 drivers/infiniband/hw/mlx5/main.c bfreg_dyn_idx = idx * (uars_per_page * MLX5_NON_FP_BFREGS_PER_UAR); idx 2147 drivers/infiniband/hw/mlx5/main.c mlx5_ib_warn(dev, "wrong offset, idx %lu is busy, bfregn=%u\n", idx, bfreg_dyn_idx); idx 2161 drivers/infiniband/hw/mlx5/main.c uar_index = bfregi->sys_pages[idx]; idx 2165 drivers/infiniband/hw/mlx5/main.c mlx5_ib_dbg(dev, "uar idx 0x%lx, pfn %pa\n", idx, &pfn); idx 2177 drivers/infiniband/hw/mlx5/main.c bfregi->sys_pages[idx] = uar_index; idx 2184 drivers/infiniband/hw/mlx5/main.c mlx5_cmd_free_uar(dev->mdev, idx); idx 134 drivers/infiniband/hw/mlx5/mem.c int i, k, idx; idx 177 drivers/infiniband/hw/mlx5/mem.c idx = (i >> shift) - offset; idx 179 drivers/infiniband/hw/mlx5/mem.c pas[idx] = cpu_to_be64(cur); idx 181 drivers/infiniband/hw/mlx5/mem.c i >> shift, be64_to_cpu(pas[idx])); idx 1160 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_update_xlt(struct mlx5_ib_mr *mr, u64 idx, int npages, idx 885 drivers/infiniband/hw/mlx5/mr.c static inline int populate_xlt(struct mlx5_ib_mr *mr, int idx, int npages, idx 895 drivers/infiniband/hw/mlx5/mr.c mlx5_odp_populate_klm(xlt, idx, npages, mr, flags); idx 899 drivers/infiniband/hw/mlx5/mr.c npages = min_t(size_t, npages, ib_umem_num_pages(umem) - idx); idx 903 drivers/infiniband/hw/mlx5/mr.c idx, npages, xlt, idx 919 drivers/infiniband/hw/mlx5/mr.c int mlx5_ib_update_xlt(struct mlx5_ib_mr *mr, u64 idx, int npages, idx 948 drivers/infiniband/hw/mlx5/mr.c if (idx & page_mask) { idx 949 drivers/infiniband/hw/mlx5/mr.c npages += idx & page_mask; idx 950 drivers/infiniband/hw/mlx5/mr.c idx &= ~page_mask; idx 1004 drivers/infiniband/hw/mlx5/mr.c pages_mapped += pages_iter, idx += pages_iter) { idx 1007 drivers/infiniband/hw/mlx5/mr.c npages = populate_xlt(mr, idx, npages, xlt, idx 1030 drivers/infiniband/hw/mlx5/mr.c wr.offset = idx * desc_size; idx 226 drivers/infiniband/hw/mlx5/odp.c int idx = ib_umem_start(odp) >> MLX5_IMR_MTT_SHIFT; idx 237 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_update_xlt(imr, idx, 1, 0, idx 256 drivers/infiniband/hw/mlx5/odp.c u64 idx = 0, blk_start_idx = 0; idx 281 drivers/infiniband/hw/mlx5/odp.c idx = (addr - ib_umem_start(umem_odp)) >> umem_odp->page_shift; idx 288 drivers/infiniband/hw/mlx5/odp.c if (umem_odp->dma_list[idx] & idx 291 drivers/infiniband/hw/mlx5/odp.c blk_start_idx = idx; idx 295 drivers/infiniband/hw/mlx5/odp.c u64 umr_offset = idx & umr_block_mask; idx 299 drivers/infiniband/hw/mlx5/odp.c idx - blk_start_idx, 0, idx 308 drivers/infiniband/hw/mlx5/odp.c idx - blk_start_idx + 1, 0, idx 1019 drivers/infiniband/hw/mlx5/qp.c static void *get_sq_edge(struct mlx5_ib_wq *sq, u32 idx) idx 1025 drivers/infiniband/hw/mlx5/qp.c mlx5_frag_buf_get_idx_last_contig_stride(&sq->fbc, idx)); idx 4017 drivers/infiniband/hw/mlx5/qp.c u32 idx; idx 4019 drivers/infiniband/hw/mlx5/qp.c idx = (sq->cur_post + (wqe_sz >> 2)) & (sq->wqe_cnt - 1); idx 4020 drivers/infiniband/hw/mlx5/qp.c *cur_edge = get_sq_edge(sq, idx); idx 4022 drivers/infiniband/hw/mlx5/qp.c *seg = mlx5_frag_buf_get_wqe(&sq->fbc, idx); idx 4877 drivers/infiniband/hw/mlx5/qp.c static void dump_wqe(struct mlx5_ib_qp *qp, u32 idx, int size_16) idx 4882 drivers/infiniband/hw/mlx5/qp.c pr_debug("dump WQE index %u:\n", idx); idx 4885 drivers/infiniband/hw/mlx5/qp.c p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, idx); idx 4888 drivers/infiniband/hw/mlx5/qp.c idx = (idx + 1) & (qp->sq.wqe_cnt - 1); idx 4898 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr, unsigned int *idx, idx 4905 drivers/infiniband/hw/mlx5/qp.c *idx = qp->sq.cur_post & (qp->sq.wqe_cnt - 1); idx 4906 drivers/infiniband/hw/mlx5/qp.c *seg = mlx5_frag_buf_get_wqe(&qp->sq.fbc, *idx); idx 4923 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr, unsigned *idx, idx 4926 drivers/infiniband/hw/mlx5/qp.c return __begin_wqe(qp, seg, ctrl, wr, idx, size, cur_edge, nreq, idx 4934 drivers/infiniband/hw/mlx5/qp.c unsigned int idx, u64 wr_id, int nreq, u8 fence, idx 4946 drivers/infiniband/hw/mlx5/qp.c qp->sq.wrid[idx] = wr_id; idx 4947 drivers/infiniband/hw/mlx5/qp.c qp->sq.w_list[idx].opcode = mlx5_opcode; idx 4948 drivers/infiniband/hw/mlx5/qp.c qp->sq.wqe_head[idx] = qp->sq.head + nreq; idx 4950 drivers/infiniband/hw/mlx5/qp.c qp->sq.w_list[idx].next = qp->sq.cur_post; idx 4979 drivers/infiniband/hw/mlx5/qp.c unsigned idx; idx 5018 drivers/infiniband/hw/mlx5/qp.c err = begin_wqe(qp, &seg, &ctrl, wr, &idx, &size, &cur_edge, idx 5068 drivers/infiniband/hw/mlx5/qp.c qp->sq.wr_data[idx] = IB_WR_LOCAL_INV; idx 5075 drivers/infiniband/hw/mlx5/qp.c qp->sq.wr_data[idx] = IB_WR_REG_MR; idx 5087 drivers/infiniband/hw/mlx5/qp.c qp->sq.wr_data[idx] = IB_WR_REG_MR_INTEGRITY; idx 5110 drivers/infiniband/hw/mlx5/qp.c cur_edge, idx, wr->wr_id, idx 5115 drivers/infiniband/hw/mlx5/qp.c &idx, &size, &cur_edge, idx 5153 drivers/infiniband/hw/mlx5/qp.c finish_wqe(qp, ctrl, seg, size, cur_edge, idx, idx 5162 drivers/infiniband/hw/mlx5/qp.c err = __begin_wqe(qp, &seg, &ctrl, wr, &idx, idx 5179 drivers/infiniband/hw/mlx5/qp.c finish_wqe(qp, ctrl, seg, size, cur_edge, idx, idx 5183 drivers/infiniband/hw/mlx5/qp.c err = __begin_wqe(qp, &seg, &ctrl, wr, &idx, idx 5200 drivers/infiniband/hw/mlx5/qp.c finish_wqe(qp, ctrl, seg, size, cur_edge, idx, idx 5269 drivers/infiniband/hw/mlx5/qp.c qp->sq.wr_data[idx] = MLX5_IB_WR_UMR; idx 5309 drivers/infiniband/hw/mlx5/qp.c finish_wqe(qp, ctrl, seg, size, cur_edge, idx, wr->wr_id, nreq, idx 5313 drivers/infiniband/hw/mlx5/qp.c dump_wqe(qp, idx, size); idx 280 drivers/infiniband/hw/mthca/mthca_memfree.c int idx, offset, dma_offset, i; idx 290 drivers/infiniband/hw/mthca/mthca_memfree.c idx = (obj & (table->num_obj - 1)) * table->obj_size; idx 291 drivers/infiniband/hw/mthca/mthca_memfree.c icm = table->icm[idx / MTHCA_TABLE_CHUNK_SIZE]; idx 292 drivers/infiniband/hw/mthca/mthca_memfree.c dma_offset = offset = idx % MTHCA_TABLE_CHUNK_SIZE; idx 573 drivers/infiniband/hw/mthca/mthca_mr.c u32 key, idx; idx 593 drivers/infiniband/hw/mthca/mthca_mr.c idx = key & (dev->limits.num_mpts - 1); idx 605 drivers/infiniband/hw/mthca/mthca_mr.c sizeof *(mr->mem.tavor.mpt) * idx; idx 772 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c u32 idx = 0; idx 777 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pbl_size = OCRDMA_MIN_HPAGE_SIZE * (1 << idx); idx 784 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c idx++; idx 1556 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c static void ocrdma_srq_toggle_bit(struct ocrdma_srq *srq, unsigned int idx) idx 1558 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c unsigned int i = idx / 32; idx 1559 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c u32 mask = (1U << (idx % 32)); idx 1585 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c u32 idx) idx 1587 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c return q->va + (idx * q->entry_size); idx 517 drivers/infiniband/hw/qedr/main.c u16 idx; idx 522 drivers/infiniband/hw/qedr/main.c idx = i * dev->num_hwfns + dev->affin_hwfn_idx; idx 523 drivers/infiniband/hw/qedr/main.c vector = dev->int_info.msix[idx].vector; idx 535 drivers/infiniband/hw/qedr/main.c u16 idx; idx 545 drivers/infiniband/hw/qedr/main.c idx = i * dev->num_hwfns + dev->affin_hwfn_idx; idx 546 drivers/infiniband/hw/qedr/main.c rc = request_irq(dev->int_info.msix[idx].vector, idx 282 drivers/infiniband/hw/qib/qib_driver.c const u32 idx = etail & ((u32)rcd->rcvegrbufs_perchunk - 1); idx 284 drivers/infiniband/hw/qib/qib_driver.c return rcd->rcvegrbuf[chunk] + (idx << rcd->dd->rcvegrbufsize_shift); idx 3866 drivers/infiniband/hw/qib/qib_iba7220.c unsigned idx = offs / sizeof(u64); idx 3869 drivers/infiniband/hw/qib/qib_iba7220.c if (idx != kr_sendctrl) { idx 3888 drivers/infiniband/hw/qib/qib_iba7220.c local_data = (u64)qib_read_kreg32(dd, idx); idx 3890 drivers/infiniband/hw/qib/qib_iba7220.c local_data = qib_read_kreg64(dd, idx); idx 1396 drivers/infiniband/hw/qib/qib_iba7322.c int idx = -1; idx 1399 drivers/infiniband/hw/qib/qib_iba7322.c ++idx; idx 1402 drivers/infiniband/hw/qib/qib_iba7322.c took = scnprintf(msg, len, "_%d", idx); idx 5842 drivers/infiniband/hw/qib/qib_iba7322.c unsigned idx; idx 5872 drivers/infiniband/hw/qib/qib_iba7322.c idx = offs / sizeof(u64); idx 5888 drivers/infiniband/hw/qib/qib_iba7322.c local_data = (u64)qib_read_kreg32(dd, idx); idx 5890 drivers/infiniband/hw/qib/qib_iba7322.c local_data = qib_read_kreg64(dd, idx); idx 5911 drivers/infiniband/hw/qib/qib_iba7322.c qib_write_kreg(dd, idx, tval); idx 7622 drivers/infiniband/hw/qib/qib_iba7322.c int idx; idx 7625 drivers/infiniband/hw/qib/qib_iba7322.c for (idx = 0; !override && idx < ARRAY_SIZE(vendor_txdds); ++idx) { idx 7626 drivers/infiniband/hw/qib/qib_iba7322.c const struct vendor_txdds_ent *v = vendor_txdds + idx; idx 7660 drivers/infiniband/hw/qib/qib_iba7322.c idx = ppd->cpspec->no_eep; idx 7661 drivers/infiniband/hw/qib/qib_iba7322.c *sdr_dds = &txdds_sdr[idx]; idx 7662 drivers/infiniband/hw/qib/qib_iba7322.c *ddr_dds = &txdds_ddr[idx]; idx 7663 drivers/infiniband/hw/qib/qib_iba7322.c *qdr_dds = &txdds_qdr[idx]; idx 7666 drivers/infiniband/hw/qib/qib_iba7322.c idx = ppd->cpspec->no_eep - TXDDS_TABLE_SZ; idx 7667 drivers/infiniband/hw/qib/qib_iba7322.c *sdr_dds = &txdds_extra_sdr[idx]; idx 7668 drivers/infiniband/hw/qib/qib_iba7322.c *ddr_dds = &txdds_extra_ddr[idx]; idx 7669 drivers/infiniband/hw/qib/qib_iba7322.c *qdr_dds = &txdds_extra_qdr[idx]; idx 7673 drivers/infiniband/hw/qib/qib_iba7322.c idx = ppd->cpspec->no_eep - (TXDDS_TABLE_SZ + TXDDS_EXTRA_SZ); idx 7675 drivers/infiniband/hw/qib/qib_iba7322.c ppd->dd->unit, ppd->port, idx); idx 7676 drivers/infiniband/hw/qib/qib_iba7322.c *sdr_dds = &txdds_extra_mfg[idx]; idx 7677 drivers/infiniband/hw/qib/qib_iba7322.c *ddr_dds = &txdds_extra_mfg[idx]; idx 7678 drivers/infiniband/hw/qib/qib_iba7322.c *qdr_dds = &txdds_extra_mfg[idx]; idx 7691 drivers/infiniband/hw/qib/qib_iba7322.c int idx; idx 7714 drivers/infiniband/hw/qib/qib_iba7322.c for (idx = 1; idx < ARRAY_SIZE(txdds_sdr); ++idx) { idx 7715 drivers/infiniband/hw/qib/qib_iba7322.c set_txdds(ppd, idx, single_ent ? sdr_dds : txdds_sdr + idx); idx 7716 drivers/infiniband/hw/qib/qib_iba7322.c set_txdds(ppd, idx + TXDDS_TABLE_SZ, idx 7717 drivers/infiniband/hw/qib/qib_iba7322.c single_ent ? ddr_dds : txdds_ddr + idx); idx 7718 drivers/infiniband/hw/qib/qib_iba7322.c set_txdds(ppd, idx + 2 * TXDDS_TABLE_SZ, idx 7719 drivers/infiniband/hw/qib/qib_iba7322.c single_ent ? qdr_dds : txdds_qdr + idx); idx 276 drivers/infiniband/hw/qib/qib_qsfp.c int idx; idx 342 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_VEND_LEN; ++idx) idx 343 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->vendor[idx]; idx 353 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_VOUI_LEN; ++idx) idx 354 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->oui[idx]; idx 359 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_PN_LEN; ++idx) idx 360 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->partnum[idx]; idx 365 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_REV_LEN; ++idx) idx 366 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->rev[idx]; idx 371 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_ATTEN_LEN; ++idx) idx 372 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->atten[idx]; idx 397 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_SN_LEN; ++idx) idx 398 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->serial[idx]; idx 403 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_DATE_LEN; ++idx) idx 404 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->date[idx]; idx 409 drivers/infiniband/hw/qib/qib_qsfp.c for (idx = 0; idx < QSFP_LOT_LEN; ++idx) idx 410 drivers/infiniband/hw/qib/qib_qsfp.c cks += cp->lot[idx]; idx 886 drivers/infiniband/hw/qib/qib_sd7220.c int cnt, sofar, req, idx, errors; idx 902 drivers/infiniband/hw/qib/qib_sd7220.c for (idx = 0; idx < cnt; ++idx) { idx 903 drivers/infiniband/hw/qib/qib_sd7220.c if (readback[idx] != img[idx+sofar]) idx 1044 drivers/infiniband/hw/qib/qib_sd7220.c int idx, midx; idx 1068 drivers/infiniband/hw/qib/qib_sd7220.c for (idx = 0; idx < NUM_DDS_REGS; ++idx) { idx 1070 drivers/infiniband/hw/qib/qib_sd7220.c writeq(data, iaddr + idx); idx 1074 drivers/infiniband/hw/qib/qib_sd7220.c u64 __iomem *daddr = taddr + ((midx << 4) + idx); idx 1076 drivers/infiniband/hw/qib/qib_sd7220.c data = dds_init_vals[midx].reg_vals[idx]; idx 1088 drivers/infiniband/hw/qib/qib_sd7220.c min_idx = idx; /* RXEQ indices pick up where DDS left off */ idx 1091 drivers/infiniband/hw/qib/qib_sd7220.c for (idx = 0; idx < RXEQ_ROWS; ++idx) { idx 1096 drivers/infiniband/hw/qib/qib_sd7220.c didx = idx + min_idx; idx 1098 drivers/infiniband/hw/qib/qib_sd7220.c writeq(rxeq_init_vals[idx].rdesc, iaddr + didx); idx 1102 drivers/infiniband/hw/qib/qib_sd7220.c data = rxeq_init_vals[idx].rdata[vidx]; idx 1103 drivers/infiniband/hw/qib/qib_sd7220.c writeq(data, taddr + (vidx << 6) + idx); idx 1194 drivers/infiniband/hw/qib/qib_sd7220.c int idx, reg, data; idx 1198 drivers/infiniband/hw/qib/qib_sd7220.c for (idx = 0; idx < NUM_DDS_REGS; ++idx) { idx 1201 drivers/infiniband/hw/qib/qib_sd7220.c data = ddi->reg_vals[idx]; idx 108 drivers/infiniband/hw/qib/qib_sdma.c unsigned idx; idx 110 drivers/infiniband/hw/qib/qib_sdma.c idx = txp->start_idx; idx 111 drivers/infiniband/hw/qib/qib_sdma.c while (idx != txp->next_descq_idx) { idx 112 drivers/infiniband/hw/qib/qib_sdma.c unmap_desc(ppd, idx); idx 113 drivers/infiniband/hw/qib/qib_sdma.c if (++idx == ppd->sdma_descq_cnt) idx 114 drivers/infiniband/hw/qib/qib_sdma.c idx = 0; idx 331 drivers/infiniband/hw/qib/qib_sdma.c u16 idx = 0; idx 344 drivers/infiniband/hw/qib/qib_sdma.c idx = txp->start_idx; idx 350 drivers/infiniband/hw/qib/qib_sdma.c (idx == ppd->sdma_descq_head)) { idx 352 drivers/infiniband/hw/qib/qib_sdma.c if (++idx == ppd->sdma_descq_cnt) idx 353 drivers/infiniband/hw/qib/qib_sdma.c idx = 0; idx 376 drivers/infiniband/hw/qib/qib_sdma.c idx = txp->start_idx; idx 693 drivers/infiniband/hw/qib/qib_sysfs.c int idx; idx 697 drivers/infiniband/hw/qib/qib_sysfs.c for (idx = 0; idx < 8; ++idx) { idx 698 drivers/infiniband/hw/qib/qib_sysfs.c if (idx == 6) idx 700 drivers/infiniband/hw/qib/qib_sysfs.c ret = dd->f_tempsense_rd(dd, idx); idx 703 drivers/infiniband/hw/qib/qib_sysfs.c regvals[idx] = ret; idx 705 drivers/infiniband/hw/qib/qib_sysfs.c if (idx == 8) idx 722 drivers/infiniband/hw/qib/qib_user_sdma.c unsigned long idx; idx 724 drivers/infiniband/hw/qib/qib_user_sdma.c for (idx = 0; idx < niov; idx++) { idx 725 drivers/infiniband/hw/qib/qib_user_sdma.c const int npages = qib_user_sdma_num_pages(iov + idx); idx 726 drivers/infiniband/hw/qib/qib_user_sdma.c const unsigned long addr = (unsigned long) iov[idx].iov_base; idx 729 drivers/infiniband/hw/qib/qib_user_sdma.c iov[idx].iov_len, npages); idx 738 drivers/infiniband/hw/qib/qib_user_sdma.c for (idx = 1; idx < pkt->naddr; idx++) idx 739 drivers/infiniband/hw/qib/qib_user_sdma.c qib_user_sdma_free_pkt_frag(&dd->pcidev->dev, pq, pkt, idx); idx 810 drivers/infiniband/hw/qib/qib_user_sdma.c unsigned long idx = 0; idx 821 drivers/infiniband/hw/qib/qib_user_sdma.c while (idx < niov && npkts < *maxpkts) { idx 822 drivers/infiniband/hw/qib/qib_user_sdma.c const unsigned long addr = (unsigned long) iov[idx].iov_base; idx 823 drivers/infiniband/hw/qib/qib_user_sdma.c const unsigned long idx_save = idx; idx 832 drivers/infiniband/hw/qib/qib_user_sdma.c len = iov[idx].iov_len; idx 847 drivers/infiniband/hw/qib/qib_user_sdma.c cfur = copy_from_user(pbc, iov[idx].iov_base, len); idx 875 drivers/infiniband/hw/qib/qib_user_sdma.c idx++; idx 876 drivers/infiniband/hw/qib/qib_user_sdma.c while (pktnwc < pktnw && idx < niov) { idx 877 drivers/infiniband/hw/qib/qib_user_sdma.c const size_t slen = iov[idx].iov_len; idx 879 drivers/infiniband/hw/qib/qib_user_sdma.c (unsigned long) iov[idx].iov_base; idx 886 drivers/infiniband/hw/qib/qib_user_sdma.c npages += qib_user_sdma_num_pages(&iov[idx]); idx 890 drivers/infiniband/hw/qib/qib_user_sdma.c idx++; idx 922 drivers/infiniband/hw/qib/qib_user_sdma.c tidsmsize = iov[idx].iov_len; idx 939 drivers/infiniband/hw/qib/qib_user_sdma.c iov[idx].iov_base, tidsmsize); idx 949 drivers/infiniband/hw/qib/qib_user_sdma.c idx++; idx 1021 drivers/infiniband/hw/qib/qib_user_sdma.c ret = idx; idx 1209 drivers/infiniband/hw/qib/qib_user_sdma.c struct qib_user_sdma_pkt *pkt, int idx, idx 1212 drivers/infiniband/hw/qib/qib_user_sdma.c const u64 addr = (u64) pkt->addr[idx].addr + idx 1213 drivers/infiniband/hw/qib/qib_user_sdma.c (u64) pkt->addr[idx].offset; idx 1214 drivers/infiniband/hw/qib/qib_user_sdma.c const u64 dwlen = (u64) pkt->addr[idx].length / 4; idx 1221 drivers/infiniband/hw/qib/qib_user_sdma.c if (pkt->addr[idx].first_desc) idx 1223 drivers/infiniband/hw/qib/qib_user_sdma.c if (pkt->addr[idx].last_desc) { idx 540 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h int pvrdma_page_dir_insert_dma(struct pvrdma_page_dir *pdir, u64 idx, idx 544 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h dma_addr_t pvrdma_page_dir_get_dma(struct pvrdma_page_dir *pdir, u64 idx); idx 111 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c static u64 *pvrdma_page_dir_table(struct pvrdma_page_dir *pdir, u64 idx) idx 113 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c return pdir->tables[PVRDMA_PAGE_DIR_TABLE(idx)]; idx 116 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c dma_addr_t pvrdma_page_dir_get_dma(struct pvrdma_page_dir *pdir, u64 idx) idx 118 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c return pvrdma_page_dir_table(pdir, idx)[PVRDMA_PAGE_DIR_PAGE(idx)]; idx 168 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c int pvrdma_page_dir_insert_dma(struct pvrdma_page_dir *pdir, u64 idx, idx 173 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c if (idx >= pdir->npages) idx 176 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c table = pvrdma_page_dir_table(pdir, idx); idx 177 drivers/infiniband/hw/vmw_pvrdma/pvrdma_misc.c table[PVRDMA_PAGE_DIR_PAGE(idx)] = daddr; idx 63 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h static inline int pvrdma_idx_valid(__u32 idx, __u32 max_elems) idx 66 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h return (idx & ~((max_elems << 1) - 1)) == 0; idx 71 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h const unsigned int idx = atomic_read(var); idx 73 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h if (pvrdma_idx_valid(idx, max_elems)) idx 74 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h return idx & (max_elems - 1); idx 80 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h __u32 idx = atomic_read(var) + 1; /* Increment. */ idx 82 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h idx &= (max_elems << 1) - 1; /* Modulo size, flip gen. */ idx 83 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h atomic_set(var, idx); idx 112 drivers/infiniband/sw/rdmavt/trace_cq.h TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), idx 113 drivers/infiniband/sw/rdmavt/trace_cq.h TP_ARGS(cq, wc, idx), idx 121 drivers/infiniband/sw/rdmavt/trace_cq.h __field(u32, idx) idx 132 drivers/infiniband/sw/rdmavt/trace_cq.h __entry->idx = idx; idx 139 drivers/infiniband/sw/rdmavt/trace_cq.h __entry->idx, idx 152 drivers/infiniband/sw/rdmavt/trace_cq.h TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), idx 153 drivers/infiniband/sw/rdmavt/trace_cq.h TP_ARGS(cq, wc, idx)); idx 157 drivers/infiniband/sw/rdmavt/trace_cq.h TP_PROTO(struct rvt_cq *cq, struct ib_wc *wc, u32 idx), idx 158 drivers/infiniband/sw/rdmavt/trace_cq.h TP_ARGS(cq, wc, idx)); idx 158 drivers/infiniband/sw/rdmavt/trace_tx.h TP_PROTO(struct rvt_qp *qp, struct rvt_swqe *wqe, u32 idx), idx 159 drivers/infiniband/sw/rdmavt/trace_tx.h TP_ARGS(qp, wqe, idx), idx 167 drivers/infiniband/sw/rdmavt/trace_tx.h __field(u32, idx) idx 179 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->idx = idx; idx 190 drivers/infiniband/sw/rdmavt/trace_tx.h __entry->idx, idx 325 drivers/infiniband/sw/siw/siw_mem.c dma_addr_t siw_pbl_get_buffer(struct siw_pbl *pbl, u64 off, int *len, int *idx) idx 327 drivers/infiniband/sw/siw/siw_mem.c int i = idx ? *idx : 0; idx 337 drivers/infiniband/sw/siw/siw_mem.c if (idx) idx 338 drivers/infiniband/sw/siw/siw_mem.c *idx = i; idx 12 drivers/infiniband/sw/siw/siw_mem.h dma_addr_t siw_pbl_get_buffer(struct siw_pbl *pbl, u64 off, int *len, int *idx); idx 1056 drivers/infiniband/sw/siw/siw_qp.c u32 idx; idx 1061 drivers/infiniband/sw/siw/siw_qp.c idx = cq->cq_put % cq->num_cqe; idx 1062 drivers/infiniband/sw/siw/siw_qp.c cqe = &cq->queue[idx]; idx 1113 drivers/infiniband/sw/siw/siw_qp.c u32 idx; idx 1118 drivers/infiniband/sw/siw/siw_qp.c idx = cq->cq_put % cq->num_cqe; idx 1119 drivers/infiniband/sw/siw/siw_qp.c cqe = &cq->queue[idx]; idx 25 drivers/infiniband/sw/siw/siw_qp_tx.c static struct page *siw_get_pblpage(struct siw_mem *mem, u64 addr, int *idx) idx 29 drivers/infiniband/sw/siw/siw_qp_tx.c dma_addr_t paddr = siw_pbl_get_buffer(pbl, offset, NULL, idx); idx 246 drivers/infiniband/sw/siw/siw_verbs.c int siw_query_pkey(struct ib_device *base_dev, u8 port, u16 idx, u16 *pkey) idx 253 drivers/infiniband/sw/siw/siw_verbs.c int siw_query_gid(struct ib_device *base_dev, u8 port, int idx, idx 799 drivers/infiniband/sw/siw/siw_verbs.c u32 idx = qp->sq_put % qp->attrs.sq_size; idx 800 drivers/infiniband/sw/siw/siw_verbs.c struct siw_sqe *sqe = &qp->sendq[idx]; idx 1045 drivers/infiniband/sw/siw/siw_verbs.c u32 idx = qp->rq_put % qp->attrs.rq_size; idx 1046 drivers/infiniband/sw/siw/siw_verbs.c struct siw_rqe *rqe = &qp->recvq[idx]; idx 1765 drivers/infiniband/sw/siw/siw_verbs.c u32 idx = srq->rq_put % srq->num_rqe; idx 1766 drivers/infiniband/sw/siw/siw_verbs.c struct siw_rqe *rqe = &srq->recvq[idx]; idx 49 drivers/infiniband/sw/siw/siw_verbs.h int siw_query_pkey(struct ib_device *base_dev, u8 port, u16 idx, u16 *pkey); idx 50 drivers/infiniband/sw/siw/siw_verbs.h int siw_query_gid(struct ib_device *base_dev, u8 port, int idx, idx 317 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c u16 start_idx, num_macs, idx = 0, count = 0; idx 325 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c if (start_idx > idx++) idx 349 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c u16 start_idx, tot_macs, num_macs, idx = 0, count = 0, em_macs = 0; idx 365 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c if (start_idx > idx++) idx 377 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema_iface.c if (start_idx > idx++) idx 2350 drivers/infiniband/ulp/srp/ib_srp.c u16 idx; idx 2360 drivers/infiniband/ulp/srp/ib_srp.c idx = blk_mq_unique_tag_to_tag(tag); idx 2361 drivers/infiniband/ulp/srp/ib_srp.c WARN_ONCE(idx >= target->req_ring_size, "%s: tag %#x: idx %d >= %d\n", idx 2362 drivers/infiniband/ulp/srp/ib_srp.c dev_name(&shost->shost_gendev), tag, idx, idx 2372 drivers/infiniband/ulp/srp/ib_srp.c req = &ch->req_ring[idx]; idx 207 drivers/input/gameport/lightning.c int i, idx; idx 211 drivers/input/gameport/lightning.c idx = card_no * 4 + i; idx 212 drivers/input/gameport/lightning.c l4 = &l4_ports[idx]; idx 222 drivers/input/gameport/lightning.c l4->port = idx; idx 230 drivers/input/gameport/lightning.c gameport_set_phys(port, "isa%04x/gameport%d", L4_PORT, idx); idx 232 drivers/input/gameport/lightning.c if (idx == 0) idx 96 drivers/input/joystick/adi.c int idx; idx 170 drivers/input/joystick/adi.c adi[0].idx = adi[1].idx = 0; idx 190 drivers/input/joystick/adi.c if ((adi->idx += count) > adi->ret) return 0; idx 192 drivers/input/joystick/adi.c bits |= ((adi->data[adi->idx - i] >> 5) & 1) << i; idx 84 drivers/input/joystick/gamecon.c unsigned int idx; idx 257 drivers/input/joystick/gamecon.c unsigned char target = 1 << sdev->idx; /* select desired pin */ idx 301 drivers/input/joystick/gamecon.c sdev->idx = i; idx 800 drivers/input/joystick/gamecon.c static int gc_setup_pad(struct gc *gc, int idx, int pad_type) idx 802 drivers/input/joystick/gamecon.c struct gc_pad *pad = &gc->pads[idx]; idx 821 drivers/input/joystick/gamecon.c "%s/input%d", gc->pd->port->name, idx); idx 856 drivers/input/joystick/gamecon.c err = gc_n64_init_ff(input_dev, idx); idx 859 drivers/input/joystick/gamecon.c idx); idx 16 drivers/input/joystick/iforce/iforce-serio.c int idx, pkt, len, id; idx 157 drivers/input/joystick/iforce/iforce-serio.c if (iforce_serio->idx < iforce_serio->len) { idx 158 drivers/input/joystick/iforce/iforce-serio.c iforce_serio->data_in[iforce_serio->idx++] = data; idx 163 drivers/input/joystick/iforce/iforce-serio.c if (iforce_serio->idx == iforce_serio->len) { idx 182 drivers/input/joystick/iforce/iforce-serio.c iforce_serio->idx = 0; idx 40 drivers/input/joystick/magellan.c int idx; idx 72 drivers/input/joystick/magellan.c if (!magellan->idx) return; idx 77 drivers/input/joystick/magellan.c if (magellan->idx != 25) return; idx 86 drivers/input/joystick/magellan.c if (magellan->idx != 4) return; idx 103 drivers/input/joystick/magellan.c magellan->idx = 0; idx 105 drivers/input/joystick/magellan.c if (magellan->idx < MAGELLAN_MAX_LENGTH) idx 106 drivers/input/joystick/magellan.c magellan->data[magellan->idx++] = data; idx 55 drivers/input/joystick/spaceball.c int idx; idx 72 drivers/input/joystick/spaceball.c if (spaceball->idx < 2) return; idx 77 drivers/input/joystick/spaceball.c if (spaceball->idx != 15) return; idx 84 drivers/input/joystick/spaceball.c if (spaceball->idx != 3) return; idx 96 drivers/input/joystick/spaceball.c if (spaceball->idx != 3) return; idx 113 drivers/input/joystick/spaceball.c spaceball->data[spaceball->idx - 1] = 0; idx 118 drivers/input/joystick/spaceball.c spaceball->data[spaceball->idx - 1] = 0; idx 140 drivers/input/joystick/spaceball.c spaceball->idx = 0; idx 161 drivers/input/joystick/spaceball.c if (spaceball->idx < SPACEBALL_MAX_LENGTH) idx 162 drivers/input/joystick/spaceball.c spaceball->data[spaceball->idx++] = data; idx 43 drivers/input/joystick/spaceorb.c int idx; idx 66 drivers/input/joystick/spaceorb.c if (spaceorb->idx < 2) return; idx 67 drivers/input/joystick/spaceorb.c for (i = 0; i < spaceorb->idx; i++) c ^= data[i]; idx 73 drivers/input/joystick/spaceorb.c spaceorb->data[spaceorb->idx - 1] = 0; idx 74 drivers/input/joystick/spaceorb.c for (i = 1; i < spaceorb->idx && spaceorb->data[i] == ' '; i++); idx 80 drivers/input/joystick/spaceorb.c if (spaceorb->idx != 12) return; idx 95 drivers/input/joystick/spaceorb.c if (spaceorb->idx != 5) return; idx 102 drivers/input/joystick/spaceorb.c if (spaceorb->idx != 4) return; idx 118 drivers/input/joystick/spaceorb.c if (spaceorb->idx) spaceorb_process_packet(spaceorb); idx 119 drivers/input/joystick/spaceorb.c spaceorb->idx = 0; idx 121 drivers/input/joystick/spaceorb.c if (spaceorb->idx < SPACEORB_MAX_LENGTH) idx 122 drivers/input/joystick/spaceorb.c spaceorb->data[spaceorb->idx++] = data & 0x7f; idx 38 drivers/input/joystick/stinger.c int idx; idx 53 drivers/input/joystick/stinger.c if (!stinger->idx) return; idx 87 drivers/input/joystick/stinger.c if (stinger->idx < STINGER_MAX_LENGTH) idx 88 drivers/input/joystick/stinger.c stinger->data[stinger->idx++] = data; idx 90 drivers/input/joystick/stinger.c if (stinger->idx == 4) { idx 92 drivers/input/joystick/stinger.c stinger->idx = 0; idx 262 drivers/input/joystick/tmdc.c static int tmdc_setup_port(struct tmdc *tmdc, int idx, unsigned char *data) idx 270 drivers/input/joystick/tmdc.c tmdc->port[idx] = port = kzalloc(sizeof (struct tmdc_port), GFP_KERNEL); idx 80 drivers/input/joystick/twidjoy.c int idx; idx 134 drivers/input/joystick/twidjoy.c twidjoy->idx = 0; /* this byte starts a new packet */ idx 135 drivers/input/joystick/twidjoy.c else if (twidjoy->idx == 0) idx 138 drivers/input/joystick/twidjoy.c if (twidjoy->idx < TWIDJOY_MAX_LENGTH) idx 139 drivers/input/joystick/twidjoy.c twidjoy->data[twidjoy->idx++] = data; idx 141 drivers/input/joystick/twidjoy.c if (twidjoy->idx == TWIDJOY_MAX_LENGTH) { idx 143 drivers/input/joystick/twidjoy.c twidjoy->idx = 0; idx 38 drivers/input/joystick/warrior.c int idx, len; idx 53 drivers/input/joystick/warrior.c if (!warrior->idx) return; idx 88 drivers/input/joystick/warrior.c if (warrior->idx) warrior_process_packet(warrior); idx 89 drivers/input/joystick/warrior.c warrior->idx = 0; idx 93 drivers/input/joystick/warrior.c if (warrior->idx < warrior->len) idx 94 drivers/input/joystick/warrior.c warrior->data[warrior->idx++] = data; idx 96 drivers/input/joystick/warrior.c if (warrior->idx == warrior->len) { idx 97 drivers/input/joystick/warrior.c if (warrior->idx) warrior_process_packet(warrior); idx 98 drivers/input/joystick/warrior.c warrior->idx = 0; idx 58 drivers/input/joystick/zhenhua.c int idx; idx 96 drivers/input/joystick/zhenhua.c zhenhua->idx = 0; /* this byte starts a new packet */ idx 97 drivers/input/joystick/zhenhua.c else if (zhenhua->idx == 0) idx 100 drivers/input/joystick/zhenhua.c if (zhenhua->idx < ZHENHUA_MAX_LENGTH) idx 101 drivers/input/joystick/zhenhua.c zhenhua->data[zhenhua->idx++] = bitrev8(data); idx 103 drivers/input/joystick/zhenhua.c if (zhenhua->idx == ZHENHUA_MAX_LENGTH) { idx 105 drivers/input/joystick/zhenhua.c zhenhua->idx = 0; idx 481 drivers/input/keyboard/gpio_keys.c int idx, idx 486 drivers/input/keyboard/gpio_keys.c struct gpio_button_data *bdata = &ddata->data[idx]; idx 612 drivers/input/keyboard/gpio_keys.c bdata->code = &ddata->keymap[idx]; idx 106 drivers/input/keyboard/hil_kbd.c int i, idx; idx 108 drivers/input/keyboard/hil_kbd.c idx = dev->idx4 / 4; idx 109 drivers/input/keyboard/hil_kbd.c p = dev->data[idx - 1]; idx 138 drivers/input/keyboard/hil_kbd.c for (i = 0; i < idx; i++) idx 149 drivers/input/keyboard/hil_kbd.c int idx = kbd->idx4 / 4; idx 157 drivers/input/keyboard/hil_kbd.c for (i = 1; i < idx - 1; i++) idx 164 drivers/input/keyboard/hil_kbd.c for (i = 1; i < idx - 1; i++) idx 169 drivers/input/keyboard/hil_kbd.c for (i = 1; i < idx - 1; i++) { idx 180 drivers/input/keyboard/hil_kbd.c for (i = 1; i < idx - 1; i++) { idx 191 drivers/input/keyboard/hil_kbd.c for (i = 1; i < idx - 1; i++) { idx 208 drivers/input/keyboard/hil_kbd.c int idx = ptr->idx4 / 4; idx 209 drivers/input/keyboard/hil_kbd.c hil_packet p = ptr->data[idx - 1]; idx 213 drivers/input/keyboard/hil_kbd.c if ((p & HIL_CMDCT_POL) != idx - 1) { idx 215 drivers/input/keyboard/hil_kbd.c "Malformed poll packet %x (idx = %i)\n", p, idx); idx 250 drivers/input/keyboard/hil_kbd.c while (cnt < idx - 1) { idx 279 drivers/input/keyboard/hil_kbd.c int idx; idx 289 drivers/input/keyboard/hil_kbd.c idx = dev->idx4 / 4; idx 291 drivers/input/keyboard/hil_kbd.c dev->data[idx] = 0; idx 292 drivers/input/keyboard/hil_kbd.c packet = dev->data[idx]; idx 294 drivers/input/keyboard/hil_kbd.c dev->data[idx] = packet; idx 110 drivers/input/keyboard/qt2160.c int error, idx = 0; idx 119 drivers/input/keyboard/qt2160.c error = i2c_smbus_write_byte(client, inireg + idx); idx 137 drivers/input/keyboard/qt2160.c error = i2c_smbus_write_byte(client, inireg + idx); idx 151 drivers/input/keyboard/qt2160.c buffer[idx++] = data; idx 213 drivers/input/misc/ad714x.c static void ad714x_button_state_machine(struct ad714x_chip *ad714x, int idx) idx 215 drivers/input/misc/ad714x.c struct ad714x_button_plat *hw = &ad714x->hw->button[idx]; idx 216 drivers/input/misc/ad714x.c struct ad714x_button_drv *sw = &ad714x->sw->button[idx]; idx 222 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "button %d touched\n", idx); idx 232 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "button %d released\n", idx); idx 248 drivers/input/misc/ad714x.c static void ad714x_slider_cal_sensor_val(struct ad714x_chip *ad714x, int idx) idx 250 drivers/input/misc/ad714x.c struct ad714x_slider_plat *hw = &ad714x->hw->slider[idx]; idx 266 drivers/input/misc/ad714x.c static void ad714x_slider_cal_highest_stage(struct ad714x_chip *ad714x, int idx) idx 268 drivers/input/misc/ad714x.c struct ad714x_slider_plat *hw = &ad714x->hw->slider[idx]; idx 269 drivers/input/misc/ad714x.c struct ad714x_slider_drv *sw = &ad714x->sw->slider[idx]; idx 274 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "slider %d highest_stage:%d\n", idx, idx 291 drivers/input/misc/ad714x.c static void ad714x_slider_cal_abs_pos(struct ad714x_chip *ad714x, int idx) idx 293 drivers/input/misc/ad714x.c struct ad714x_slider_plat *hw = &ad714x->hw->slider[idx]; idx 294 drivers/input/misc/ad714x.c struct ad714x_slider_drv *sw = &ad714x->sw->slider[idx]; idx 299 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "slider %d absolute position:%d\n", idx, idx 313 drivers/input/misc/ad714x.c static void ad714x_slider_cal_flt_pos(struct ad714x_chip *ad714x, int idx) idx 315 drivers/input/misc/ad714x.c struct ad714x_slider_drv *sw = &ad714x->sw->slider[idx]; idx 320 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "slider %d filter position:%d\n", idx, idx 324 drivers/input/misc/ad714x.c static void ad714x_slider_use_com_int(struct ad714x_chip *ad714x, int idx) idx 326 drivers/input/misc/ad714x.c struct ad714x_slider_plat *hw = &ad714x->hw->slider[idx]; idx 331 drivers/input/misc/ad714x.c static void ad714x_slider_use_thr_int(struct ad714x_chip *ad714x, int idx) idx 333 drivers/input/misc/ad714x.c struct ad714x_slider_plat *hw = &ad714x->hw->slider[idx]; idx 338 drivers/input/misc/ad714x.c static void ad714x_slider_state_machine(struct ad714x_chip *ad714x, int idx) idx 340 drivers/input/misc/ad714x.c struct ad714x_slider_plat *hw = &ad714x->hw->slider[idx]; idx 341 drivers/input/misc/ad714x.c struct ad714x_slider_drv *sw = &ad714x->sw->slider[idx]; idx 357 drivers/input/misc/ad714x.c ad714x_slider_use_com_int(ad714x, idx); idx 358 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "slider %d touched\n", idx); idx 364 drivers/input/misc/ad714x.c ad714x_slider_cal_sensor_val(ad714x, idx); idx 365 drivers/input/misc/ad714x.c ad714x_slider_cal_highest_stage(ad714x, idx); idx 366 drivers/input/misc/ad714x.c ad714x_slider_cal_abs_pos(ad714x, idx); idx 375 drivers/input/misc/ad714x.c ad714x_slider_cal_sensor_val(ad714x, idx); idx 376 drivers/input/misc/ad714x.c ad714x_slider_cal_highest_stage(ad714x, idx); idx 377 drivers/input/misc/ad714x.c ad714x_slider_cal_abs_pos(ad714x, idx); idx 378 drivers/input/misc/ad714x.c ad714x_slider_cal_flt_pos(ad714x, idx); idx 385 drivers/input/misc/ad714x.c ad714x_slider_use_thr_int(ad714x, idx); idx 389 drivers/input/misc/ad714x.c idx); idx 407 drivers/input/misc/ad714x.c static void ad714x_wheel_cal_highest_stage(struct ad714x_chip *ad714x, int idx) idx 409 drivers/input/misc/ad714x.c struct ad714x_wheel_plat *hw = &ad714x->hw->wheel[idx]; idx 410 drivers/input/misc/ad714x.c struct ad714x_wheel_drv *sw = &ad714x->sw->wheel[idx]; idx 416 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "wheel %d highest_stage:%d\n", idx, idx 420 drivers/input/misc/ad714x.c static void ad714x_wheel_cal_sensor_val(struct ad714x_chip *ad714x, int idx) idx 422 drivers/input/misc/ad714x.c struct ad714x_wheel_plat *hw = &ad714x->hw->wheel[idx]; idx 449 drivers/input/misc/ad714x.c static void ad714x_wheel_cal_abs_pos(struct ad714x_chip *ad714x, int idx) idx 451 drivers/input/misc/ad714x.c struct ad714x_wheel_plat *hw = &ad714x->hw->wheel[idx]; idx 452 drivers/input/misc/ad714x.c struct ad714x_wheel_drv *sw = &ad714x->sw->wheel[idx]; idx 480 drivers/input/misc/ad714x.c static void ad714x_wheel_cal_flt_pos(struct ad714x_chip *ad714x, int idx) idx 482 drivers/input/misc/ad714x.c struct ad714x_wheel_plat *hw = &ad714x->hw->wheel[idx]; idx 483 drivers/input/misc/ad714x.c struct ad714x_wheel_drv *sw = &ad714x->sw->wheel[idx]; idx 496 drivers/input/misc/ad714x.c static void ad714x_wheel_use_com_int(struct ad714x_chip *ad714x, int idx) idx 498 drivers/input/misc/ad714x.c struct ad714x_wheel_plat *hw = &ad714x->hw->wheel[idx]; idx 503 drivers/input/misc/ad714x.c static void ad714x_wheel_use_thr_int(struct ad714x_chip *ad714x, int idx) idx 505 drivers/input/misc/ad714x.c struct ad714x_wheel_plat *hw = &ad714x->hw->wheel[idx]; idx 510 drivers/input/misc/ad714x.c static void ad714x_wheel_state_machine(struct ad714x_chip *ad714x, int idx) idx 512 drivers/input/misc/ad714x.c struct ad714x_wheel_plat *hw = &ad714x->hw->wheel[idx]; idx 513 drivers/input/misc/ad714x.c struct ad714x_wheel_drv *sw = &ad714x->sw->wheel[idx]; idx 529 drivers/input/misc/ad714x.c ad714x_wheel_use_com_int(ad714x, idx); idx 530 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "wheel %d touched\n", idx); idx 536 drivers/input/misc/ad714x.c ad714x_wheel_cal_sensor_val(ad714x, idx); idx 537 drivers/input/misc/ad714x.c ad714x_wheel_cal_highest_stage(ad714x, idx); idx 538 drivers/input/misc/ad714x.c ad714x_wheel_cal_abs_pos(ad714x, idx); idx 547 drivers/input/misc/ad714x.c ad714x_wheel_cal_sensor_val(ad714x, idx); idx 548 drivers/input/misc/ad714x.c ad714x_wheel_cal_highest_stage(ad714x, idx); idx 549 drivers/input/misc/ad714x.c ad714x_wheel_cal_abs_pos(ad714x, idx); idx 550 drivers/input/misc/ad714x.c ad714x_wheel_cal_flt_pos(ad714x, idx); idx 558 drivers/input/misc/ad714x.c ad714x_wheel_use_thr_int(ad714x, idx); idx 563 drivers/input/misc/ad714x.c idx); idx 574 drivers/input/misc/ad714x.c static void touchpad_cal_sensor_val(struct ad714x_chip *ad714x, int idx) idx 576 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 594 drivers/input/misc/ad714x.c static void touchpad_cal_highest_stage(struct ad714x_chip *ad714x, int idx) idx 596 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 597 drivers/input/misc/ad714x.c struct ad714x_touchpad_drv *sw = &ad714x->sw->touchpad[idx]; idx 606 drivers/input/misc/ad714x.c idx, sw->x_highest_stage, sw->y_highest_stage); idx 615 drivers/input/misc/ad714x.c static int touchpad_check_second_peak(struct ad714x_chip *ad714x, int idx) idx 617 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 618 drivers/input/misc/ad714x.c struct ad714x_touchpad_drv *sw = &ad714x->sw->touchpad[idx]; idx 654 drivers/input/misc/ad714x.c static void touchpad_cal_abs_pos(struct ad714x_chip *ad714x, int idx) idx 656 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 657 drivers/input/misc/ad714x.c struct ad714x_touchpad_drv *sw = &ad714x->sw->touchpad[idx]; idx 664 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "touchpad %d absolute position:(%d, %d)\n", idx, idx 668 drivers/input/misc/ad714x.c static void touchpad_cal_flt_pos(struct ad714x_chip *ad714x, int idx) idx 670 drivers/input/misc/ad714x.c struct ad714x_touchpad_drv *sw = &ad714x->sw->touchpad[idx]; idx 678 drivers/input/misc/ad714x.c idx, sw->x_flt_pos, sw->y_flt_pos); idx 699 drivers/input/misc/ad714x.c static int touchpad_check_endpoint(struct ad714x_chip *ad714x, int idx) idx 701 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 702 drivers/input/misc/ad714x.c struct ad714x_touchpad_drv *sw = &ad714x->sw->touchpad[idx]; idx 776 drivers/input/misc/ad714x.c static void touchpad_use_com_int(struct ad714x_chip *ad714x, int idx) idx 778 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 783 drivers/input/misc/ad714x.c static void touchpad_use_thr_int(struct ad714x_chip *ad714x, int idx) idx 785 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 791 drivers/input/misc/ad714x.c static void ad714x_touchpad_state_machine(struct ad714x_chip *ad714x, int idx) idx 793 drivers/input/misc/ad714x.c struct ad714x_touchpad_plat *hw = &ad714x->hw->touchpad[idx]; idx 794 drivers/input/misc/ad714x.c struct ad714x_touchpad_drv *sw = &ad714x->sw->touchpad[idx]; idx 813 drivers/input/misc/ad714x.c touchpad_use_com_int(ad714x, idx); idx 814 drivers/input/misc/ad714x.c dev_dbg(ad714x->dev, "touchpad %d touched\n", idx); idx 820 drivers/input/misc/ad714x.c touchpad_cal_sensor_val(ad714x, idx); idx 821 drivers/input/misc/ad714x.c touchpad_cal_highest_stage(ad714x, idx); idx 822 drivers/input/misc/ad714x.c if ((!touchpad_check_second_peak(ad714x, idx)) && idx 823 drivers/input/misc/ad714x.c (!touchpad_check_endpoint(ad714x, idx))) { idx 826 drivers/input/misc/ad714x.c idx); idx 827 drivers/input/misc/ad714x.c touchpad_cal_abs_pos(ad714x, idx); idx 838 drivers/input/misc/ad714x.c touchpad_cal_sensor_val(ad714x, idx); idx 839 drivers/input/misc/ad714x.c touchpad_cal_highest_stage(ad714x, idx); idx 840 drivers/input/misc/ad714x.c if ((!touchpad_check_second_peak(ad714x, idx)) idx 841 drivers/input/misc/ad714x.c && (!touchpad_check_endpoint(ad714x, idx))) { idx 842 drivers/input/misc/ad714x.c touchpad_cal_abs_pos(ad714x, idx); idx 843 drivers/input/misc/ad714x.c touchpad_cal_flt_pos(ad714x, idx); idx 855 drivers/input/misc/ad714x.c touchpad_use_thr_int(ad714x, idx); idx 859 drivers/input/misc/ad714x.c idx); idx 49 drivers/input/misc/axp20x-pek.c unsigned int idx; idx 53 drivers/input/misc/axp20x-pek.c { .time = 128, .idx = 0 }, idx 54 drivers/input/misc/axp20x-pek.c { .time = 1000, .idx = 2 }, idx 55 drivers/input/misc/axp20x-pek.c { .time = 3000, .idx = 1 }, idx 56 drivers/input/misc/axp20x-pek.c { .time = 2000, .idx = 3 }, idx 60 drivers/input/misc/axp20x-pek.c { .time = 128, .idx = 0 }, idx 61 drivers/input/misc/axp20x-pek.c { .time = 1000, .idx = 1 }, idx 62 drivers/input/misc/axp20x-pek.c { .time = 2000, .idx = 2 }, idx 63 drivers/input/misc/axp20x-pek.c { .time = 3000, .idx = 3 }, idx 67 drivers/input/misc/axp20x-pek.c { .time = 4000, .idx = 0 }, idx 68 drivers/input/misc/axp20x-pek.c { .time = 6000, .idx = 1 }, idx 69 drivers/input/misc/axp20x-pek.c { .time = 8000, .idx = 2 }, idx 70 drivers/input/misc/axp20x-pek.c { .time = 10000, .idx = 3 }, idx 103 drivers/input/misc/axp20x-pek.c if (val == time[i].idx) idx 138 drivers/input/misc/axp20x-pek.c unsigned int val, idx = 0; idx 158 drivers/input/misc/axp20x-pek.c idx = time[i].idx; idx 165 drivers/input/misc/axp20x-pek.c idx <<= ffs(mask) - 1; idx 167 drivers/input/misc/axp20x-pek.c mask, idx); idx 31 drivers/input/mouse/cyapa.h #define SMBUS_ENCODE_IDX(cmd, idx) ((cmd) | (((idx) & 0x03) << 1)) idx 77 drivers/input/mouse/elan_i2c.h const u8 *page, u16 checksum, int idx); idx 434 drivers/input/mouse/elan_i2c_core.c const u8 *page, u16 checksum, int idx) idx 441 drivers/input/mouse/elan_i2c_core.c page, checksum, idx); idx 446 drivers/input/mouse/elan_i2c_core.c "IAP retrying page %d (error: %d)\n", idx, error); idx 576 drivers/input/mouse/elan_i2c_i2c.c const u8 *page, u16 checksum, int idx) idx 593 drivers/input/mouse/elan_i2c_i2c.c dev_err(dev, "Failed to write page %d: %d\n", idx, error); idx 418 drivers/input/mouse/elan_i2c_smbus.c const u8 *page, u16 checksum, int idx) idx 436 drivers/input/mouse/elan_i2c_smbus.c idx, 1, error); idx 446 drivers/input/mouse/elan_i2c_smbus.c idx, 2, error); idx 1791 drivers/input/mouse/elantech.c unsigned int idx = 0; idx 1795 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("touchscreen-size-x", idx 1797 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("touchscreen-size-y", idx 1799 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("touchscreen-min-x", idx 1801 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("touchscreen-min-y", idx 1804 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("touchscreen-x-mm", idx 1807 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("touchscreen-y-mm", idx 1811 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_BOOL("elan,trackpoint"); idx 1814 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_BOOL("elan,middle-button"); idx 1817 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("elan,x_traces", idx 1820 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_U32("elan,y_traces", idx 1824 drivers/input/mouse/elantech.c i2c_props[idx++] = PROPERTY_ENTRY_BOOL("elan,clickpad"); idx 1157 drivers/input/mouse/synaptics.c int idx, enum synaptics_pkt_type pkt_type) idx 1166 drivers/input/mouse/synaptics.c if (idx < 0 || idx > 4) idx 1173 drivers/input/mouse/synaptics.c return (packet[idx] & newabs_rel_mask[idx]) == newabs_rslt[idx]; idx 1176 drivers/input/mouse/synaptics.c return (packet[idx] & newabs_mask[idx]) == newabs_rslt[idx]; idx 1179 drivers/input/mouse/synaptics.c return (packet[idx] & oldabs_mask[idx]) == oldabs_rslt[idx]; idx 100 drivers/input/serio/hil_mlc.c static void hil_mlc_copy_di_scratch(hil_mlc *mlc, int idx) idx 102 drivers/input/serio/hil_mlc.c memcpy(&mlc->di[idx], &mlc->di_scratch, sizeof(mlc->di_scratch)); idx 107 drivers/input/serio/hil_mlc.c int idx; idx 109 drivers/input/serio/hil_mlc.c for (idx = 0; idx < HIL_MLC_DEVMEM; idx++) { idx 114 drivers/input/serio/hil_mlc.c if (mlc->di_map[j] == idx) idx 120 drivers/input/serio/hil_mlc.c if (!memcmp(mlc->di + idx, &mlc->di_scratch, idx 124 drivers/input/serio/hil_mlc.c return idx >= HIL_MLC_DEVMEM ? -1 : idx; idx 129 drivers/input/serio/hil_mlc.c int idx; idx 134 drivers/input/serio/hil_mlc.c for (idx = 0; idx < HIL_MLC_DEVMEM; idx++) { idx 138 drivers/input/serio/hil_mlc.c if (mlc->di_map[j] == idx) idx 145 drivers/input/serio/hil_mlc.c return idx; /* Note: It is guaranteed at least one above will match */ idx 150 drivers/input/serio/hil_mlc.c int idx; idx 152 drivers/input/serio/hil_mlc.c for (idx = 0; idx < HIL_MLC_DEVMEM; idx++) { idx 156 drivers/input/serio/hil_mlc.c if (mlc->di_map[j] == idx) idx 160 drivers/input/serio/hil_mlc.c mlc->serio_map[idx].di_revmap = -1; idx 797 drivers/input/serio/hil_mlc.c uint8_t *idx, *last; idx 814 drivers/input/serio/hil_mlc.c idx = mlc->di[map->didx].idd; idx 817 drivers/input/serio/hil_mlc.c idx = mlc->di[map->didx].rsc; idx 820 drivers/input/serio/hil_mlc.c idx = mlc->di[map->didx].exd; idx 823 drivers/input/serio/hil_mlc.c idx = mlc->di[map->didx].rnm; idx 838 drivers/input/serio/hil_mlc.c last = idx + 15; idx 839 drivers/input/serio/hil_mlc.c while ((last != idx) && (*last == 0)) idx 842 drivers/input/serio/hil_mlc.c while (idx != last) { idx 846 drivers/input/serio/hil_mlc.c drv->interrupt(serio, *idx, 0); idx 847 drivers/input/serio/hil_mlc.c idx++; idx 852 drivers/input/serio/hil_mlc.c drv->interrupt(serio, *idx, 0); idx 193 drivers/input/serio/hp_sdc.c curr->seq[curr->idx++] = status; idx 194 drivers/input/serio/hp_sdc.c curr->seq[curr->idx++] = data; idx 208 drivers/input/serio/hp_sdc.c curr->actidx = curr->idx; idx 209 drivers/input/serio/hp_sdc.c curr->idx++; idx 323 drivers/input/serio/hp_sdc.c curr->idx += hp_sdc.rqty; idx 339 drivers/input/serio/hp_sdc.c curr->actidx = curr->idx; idx 340 drivers/input/serio/hp_sdc.c curr->idx++; idx 352 drivers/input/serio/hp_sdc.c int idx, curridx; idx 413 drivers/input/serio/hp_sdc.c idx = curr->actidx; idx 424 drivers/input/serio/hp_sdc.c act = curr->seq[idx]; idx 425 drivers/input/serio/hp_sdc.c idx++; idx 427 drivers/input/serio/hp_sdc.c if (curr->idx >= curr->endidx) { idx 439 drivers/input/serio/hp_sdc.c if (curr->idx != idx) { idx 440 drivers/input/serio/hp_sdc.c idx++; idx 444 drivers/input/serio/hp_sdc.c hp_sdc_status_out8(curr->seq[idx]); idx 445 drivers/input/serio/hp_sdc.c curr->idx++; idx 451 drivers/input/serio/hp_sdc.c curr->idx++; idx 457 drivers/input/serio/hp_sdc.c qty = curr->seq[idx]; idx 458 drivers/input/serio/hp_sdc.c idx++; idx 459 drivers/input/serio/hp_sdc.c if (curr->idx - idx < qty) { idx 460 drivers/input/serio/hp_sdc.c hp_sdc_data_out8(curr->seq[curr->idx]); idx 461 drivers/input/serio/hp_sdc.c curr->idx++; idx 463 drivers/input/serio/hp_sdc.c if (curr->idx - idx >= qty && idx 468 drivers/input/serio/hp_sdc.c idx += qty; idx 475 drivers/input/serio/hp_sdc.c mask = curr->seq[idx]; idx 476 drivers/input/serio/hp_sdc.c if (idx != curr->idx) { idx 477 drivers/input/serio/hp_sdc.c idx++; idx 478 drivers/input/serio/hp_sdc.c idx += !!(mask & 1); idx 479 drivers/input/serio/hp_sdc.c idx += !!(mask & 2); idx 480 drivers/input/serio/hp_sdc.c idx += !!(mask & 4); idx 481 drivers/input/serio/hp_sdc.c idx += !!(mask & 8); idx 486 drivers/input/serio/hp_sdc.c w7[0] = (mask & 1) ? curr->seq[++idx] : hp_sdc.r7[0]; idx 487 drivers/input/serio/hp_sdc.c w7[1] = (mask & 2) ? curr->seq[++idx] : hp_sdc.r7[1]; idx 488 drivers/input/serio/hp_sdc.c w7[2] = (mask & 4) ? curr->seq[++idx] : hp_sdc.r7[2]; idx 489 drivers/input/serio/hp_sdc.c w7[3] = (mask & 8) ? curr->seq[++idx] : hp_sdc.r7[3]; idx 505 drivers/input/serio/hp_sdc.c idx++; idx 509 drivers/input/serio/hp_sdc.c curr->idx = idx; idx 523 drivers/input/serio/hp_sdc.c curr->idx = idx + 1; idx 545 drivers/input/serio/hp_sdc.c postcmd = curr->seq[idx]; idx 546 drivers/input/serio/hp_sdc.c curr->idx++; idx 550 drivers/input/serio/hp_sdc.c hp_sdc.rqty = curr->seq[curr->idx]; idx 552 drivers/input/serio/hp_sdc.c curr->idx++; idx 570 drivers/input/serio/hp_sdc.c if (curr->idx >= curr->endidx) { /* This transaction is over. */ idx 575 drivers/input/serio/hp_sdc.c curr->actidx = idx + 1; idx 576 drivers/input/serio/hp_sdc.c curr->idx = idx + 2; idx 613 drivers/input/serio/hp_sdc.c this->idx = 1; idx 897 drivers/input/serio/hp_sdc.c t_sync.idx = 1; idx 1036 drivers/input/serio/hp_sdc.c tq_init.idx = 1; idx 1064 drivers/input/serio/hp_sdc.c tq_init.idx = 1; idx 1088 drivers/input/serio/hp_sdc.c tq_init.idx = 1; idx 64 drivers/input/serio/hp_sdc_mlc.c int idx; idx 73 drivers/input/serio/hp_sdc_mlc.c idx = 15 - mlc->icount; idx 75 drivers/input/serio/hp_sdc_mlc.c mlc->ipacket[idx] |= data | HIL_ERR_INT; idx 77 drivers/input/serio/hp_sdc_mlc.c if (hp_sdc_mlc_priv.got5x || !idx) idx 79 drivers/input/serio/hp_sdc_mlc.c if ((mlc->ipacket[idx - 1] & HIL_PKT_ADDR_MASK) != idx 80 drivers/input/serio/hp_sdc_mlc.c (mlc->ipacket[idx] & HIL_PKT_ADDR_MASK)) { idx 81 drivers/input/serio/hp_sdc_mlc.c mlc->ipacket[idx] &= ~HIL_PKT_ADDR_MASK; idx 82 drivers/input/serio/hp_sdc_mlc.c mlc->ipacket[idx] |= (mlc->ipacket[idx - 1] idx 90 drivers/input/serio/hp_sdc_mlc.c mlc->ipacket[idx] = idx 100 drivers/input/serio/hp_sdc_mlc.c && (mlc->ipacket[idx] == (mlc->imatch | idx))) idx 102 drivers/input/serio/hp_sdc_mlc.c if (mlc->ipacket[idx] == mlc->imatch) idx 115 drivers/input/serio/hp_sdc_mlc.c mlc->ipacket[idx] |= HIL_ERR_INT | HIL_ERR_PERR | idx 120 drivers/input/serio/hp_sdc_mlc.c mlc->ipacket[idx] |= HIL_ERR_INT | HIL_ERR_LERR; idx 205 drivers/input/serio/hp_sdc_mlc.c priv->trans.idx = 1; idx 244 drivers/input/serio/hp_sdc_mlc.c priv->trans.idx = 1; idx 285 drivers/input/serio/hp_sdc_mlc.c priv->trans.idx = 1; idx 1345 drivers/input/serio/i8042.c static int __init i8042_create_aux_port(int idx) idx 1348 drivers/input/serio/i8042.c int port_no = idx < 0 ? I8042_AUX_PORT_NO : I8042_MUX_PORT_NO + idx; idx 1362 drivers/input/serio/i8042.c if (idx < 0) { idx 1369 drivers/input/serio/i8042.c snprintf(serio->name, sizeof(serio->name), "i8042 AUX%d port", idx); idx 1370 drivers/input/serio/i8042.c snprintf(serio->phys, sizeof(serio->phys), I8042_MUX_PHYS_DESC, idx + 1); idx 1376 drivers/input/serio/i8042.c port->mux = idx; idx 116 drivers/input/serio/maceps2.c static struct serio *maceps2_allocate_port(int idx) idx 126 drivers/input/serio/maceps2.c snprintf(serio->name, sizeof(serio->name), "MACE PS/2 port%d", idx); idx 127 drivers/input/serio/maceps2.c snprintf(serio->phys, sizeof(serio->phys), "mace/serio%d", idx); idx 128 drivers/input/serio/maceps2.c serio->port_data = &port_data[idx]; idx 26 drivers/input/sparse-keymap.c unsigned int idx = 0; idx 32 drivers/input/sparse-keymap.c idx++; idx 36 drivers/input/sparse-keymap.c return idx; idx 174 drivers/input/tablet/wacom_serial4.c unsigned int idx; idx 369 drivers/input/tablet/wacom_serial4.c wacom->idx = 0; idx 378 drivers/input/tablet/wacom_serial4.c wacom->idx = 0; idx 396 drivers/input/tablet/wacom_serial4.c if (wacom->idx > (DATA_SIZE - 2)) { idx 398 drivers/input/tablet/wacom_serial4.c "throwing away %d bytes of garbage\n", wacom->idx); idx 401 drivers/input/tablet/wacom_serial4.c wacom->data[wacom->idx++] = data; idx 403 drivers/input/tablet/wacom_serial4.c if (wacom->idx == PACKET_LENGTH && (wacom->data[0] & 0x80)) { idx 532 drivers/input/tablet/wacom_serial4.c wacom->tool = wacom->idx = 0; idx 282 drivers/input/touchscreen/cyttsp_core.c int idx) idx 284 drivers/input/touchscreen/cyttsp_core.c switch (idx) { idx 53 drivers/input/touchscreen/dynapro.c int idx; idx 62 drivers/input/touchscreen/dynapro.c if (DYNAPRO_FORMAT_LENGTH == ++pdynapro->idx) { idx 69 drivers/input/touchscreen/dynapro.c pdynapro->idx = 0; idx 78 drivers/input/touchscreen/dynapro.c pdynapro->data[pdynapro->idx] = data; idx 43 drivers/input/touchscreen/egalax_ts_serial.c int idx; idx 74 drivers/input/touchscreen/egalax_ts_serial.c egalax->data[egalax->idx++] = data; idx 78 drivers/input/touchscreen/egalax_ts_serial.c if (pkt_len == egalax->idx) { idx 80 drivers/input/touchscreen/egalax_ts_serial.c egalax->idx = 0; idx 85 drivers/input/touchscreen/egalax_ts_serial.c egalax->idx = 0; idx 57 drivers/input/touchscreen/elo.c int idx; idx 69 drivers/input/touchscreen/elo.c elo->data[elo->idx] = data; idx 71 drivers/input/touchscreen/elo.c switch (elo->idx++) { idx 77 drivers/input/touchscreen/elo.c elo->idx = 0; idx 82 drivers/input/touchscreen/elo.c elo->idx = 0; idx 121 drivers/input/touchscreen/elo.c elo->data[elo->idx] = data; idx 123 drivers/input/touchscreen/elo.c switch (elo->idx++) { idx 127 drivers/input/touchscreen/elo.c elo->idx = 0; idx 132 drivers/input/touchscreen/elo.c elo->idx = 0; idx 137 drivers/input/touchscreen/elo.c elo->idx = 0; idx 142 drivers/input/touchscreen/elo.c elo->idx = 0; idx 152 drivers/input/touchscreen/elo.c elo->idx = 0; idx 160 drivers/input/touchscreen/elo.c elo->idx = 0; idx 170 drivers/input/touchscreen/elo.c elo->idx = 0; idx 179 drivers/input/touchscreen/elo.c elo->data[elo->idx] = data; idx 181 drivers/input/touchscreen/elo.c switch (elo->idx++) { idx 185 drivers/input/touchscreen/elo.c elo->idx = 0; idx 192 drivers/input/touchscreen/elo.c elo->idx = 0; idx 30 drivers/input/touchscreen/fujitsu_ts.c int idx; idx 49 drivers/input/touchscreen/fujitsu_ts.c if (fujitsu->idx == 0) { idx 56 drivers/input/touchscreen/fujitsu_ts.c fujitsu->idx = 0; idx 61 drivers/input/touchscreen/fujitsu_ts.c fujitsu->data[fujitsu->idx++] = data; idx 62 drivers/input/touchscreen/fujitsu_ts.c if (fujitsu->idx == FUJITSU_LENGTH) { idx 70 drivers/input/touchscreen/fujitsu_ts.c fujitsu->idx = 0; idx 39 drivers/input/touchscreen/gunze.c int idx; idx 48 drivers/input/touchscreen/gunze.c if (gunze->idx != GUNZE_MAX_LENGTH || gunze->data[5] != ',' || idx 67 drivers/input/touchscreen/gunze.c gunze->idx = 0; idx 69 drivers/input/touchscreen/gunze.c if (gunze->idx < GUNZE_MAX_LENGTH) idx 70 drivers/input/touchscreen/gunze.c gunze->data[gunze->idx++] = data; idx 52 drivers/input/touchscreen/hampshire.c int idx; idx 61 drivers/input/touchscreen/hampshire.c if (HAMPSHIRE_FORMAT_LENGTH == ++phampshire->idx) { idx 68 drivers/input/touchscreen/hampshire.c phampshire->idx = 0; idx 77 drivers/input/touchscreen/hampshire.c phampshire->data[phampshire->idx] = data; idx 55 drivers/input/touchscreen/inexio.c int idx; idx 64 drivers/input/touchscreen/inexio.c if (INEXIO_FORMAT_LENGTH == ++pinexio->idx) { idx 70 drivers/input/touchscreen/inexio.c pinexio->idx = 0; idx 79 drivers/input/touchscreen/inexio.c pinexio->data[pinexio->idx] = data; idx 81 drivers/input/touchscreen/lpc32xx_ts.c int idx; idx 98 drivers/input/touchscreen/lpc32xx_ts.c idx = 0; idx 99 drivers/input/touchscreen/lpc32xx_ts.c while (idx < 4 && idx 103 drivers/input/touchscreen/lpc32xx_ts.c xs[idx] = LPC32XX_TSC_ADCDAT_VALUE_MASK - idx 105 drivers/input/touchscreen/lpc32xx_ts.c ys[idx] = LPC32XX_TSC_ADCDAT_VALUE_MASK - idx 107 drivers/input/touchscreen/lpc32xx_ts.c rv[idx] = tmp; idx 108 drivers/input/touchscreen/lpc32xx_ts.c idx++; idx 112 drivers/input/touchscreen/lpc32xx_ts.c if (!(rv[3] & LPC32XX_TSC_FIFO_TS_P_LEVEL) && idx == 4) { idx 174 drivers/input/touchscreen/mainstone-wm97xx.c int idx = 0, ret = 0; idx 181 drivers/input/touchscreen/mainstone-wm97xx.c for (idx = 0; idx < ARRAY_SIZE(cinfo); idx++) { idx 182 drivers/input/touchscreen/mainstone-wm97xx.c if (wm->id != cinfo[idx].id) idx 184 drivers/input/touchscreen/mainstone-wm97xx.c sp_idx = idx; idx 185 drivers/input/touchscreen/mainstone-wm97xx.c if (cont_rate <= cinfo[idx].speed) idx 56 drivers/input/touchscreen/mtouch.c int idx; idx 65 drivers/input/touchscreen/mtouch.c if (MTOUCH_FORMAT_TABLET_LENGTH == ++mtouch->idx) { idx 71 drivers/input/touchscreen/mtouch.c mtouch->idx = 0; idx 77 drivers/input/touchscreen/mtouch.c if (MTOUCH_RESPONSE_END_BYTE == mtouch->data[mtouch->idx++]) { idx 79 drivers/input/touchscreen/mtouch.c mtouch->idx = 0; idx 80 drivers/input/touchscreen/mtouch.c } else if (MTOUCH_MAX_LENGTH == mtouch->idx) { idx 82 drivers/input/touchscreen/mtouch.c mtouch->idx = 0; idx 91 drivers/input/touchscreen/mtouch.c mtouch->data[mtouch->idx] = data; idx 53 drivers/input/touchscreen/penmount.c int idx; idx 103 drivers/input/touchscreen/penmount.c if ((pm->data[0] & 0x80) && pm->packetsize == ++pm->idx) { idx 108 drivers/input/touchscreen/penmount.c pm->idx = 0; idx 116 drivers/input/touchscreen/penmount.c if ((pm->data[0] & 0xbf) == 0x30 && pm->packetsize == ++pm->idx) { idx 125 drivers/input/touchscreen/penmount.c pm->idx = 0; idx 133 drivers/input/touchscreen/penmount.c if ((pm->data[0] & 0xce) == 0x40 && pm->packetsize == ++pm->idx) { idx 141 drivers/input/touchscreen/penmount.c pm->idx = 0; idx 149 drivers/input/touchscreen/penmount.c if ((pm->data[0] & 0xb0) == 0x30 && pm->packetsize == ++pm->idx) { idx 157 drivers/input/touchscreen/penmount.c pm->idx = 0; idx 166 drivers/input/touchscreen/penmount.c pm->data[pm->idx] = data; idx 76 drivers/input/touchscreen/touchit213.c int idx; idx 88 drivers/input/touchscreen/touchit213.c touchit213->data[touchit213->idx] = data; idx 90 drivers/input/touchscreen/touchit213.c switch (touchit213->idx++) { idx 95 drivers/input/touchscreen/touchit213.c touchit213->idx = 0; idx 100 drivers/input/touchscreen/touchit213.c touchit213->idx = 0; idx 48 drivers/input/touchscreen/touchright.c int idx; idx 59 drivers/input/touchscreen/touchright.c tr->data[tr->idx] = data; idx 62 drivers/input/touchscreen/touchright.c if (++tr->idx == TR_LENGTH) { idx 70 drivers/input/touchscreen/touchright.c tr->idx = 0; idx 51 drivers/input/touchscreen/touchwin.c int idx; idx 65 drivers/input/touchscreen/touchwin.c tw->data[tw->idx++] = data; idx 67 drivers/input/touchscreen/touchwin.c if (tw->idx == TW_LENGTH && tw->data[1] == tw->data[2]) { idx 72 drivers/input/touchscreen/touchwin.c tw->idx = 0; idx 77 drivers/input/touchscreen/touchwin.c tw->idx = 0; idx 19 drivers/input/touchscreen/tsc40.c u32 idx; idx 47 drivers/input/touchscreen/tsc40.c ptsc->data[ptsc->idx] = data; idx 48 drivers/input/touchscreen/tsc40.c switch (ptsc->idx++) { idx 53 drivers/input/touchscreen/tsc40.c ptsc->idx = 0; idx 57 drivers/input/touchscreen/tsc40.c ptsc->idx = 0; idx 66 drivers/input/touchscreen/tsc40.c data, ptsc->idx - 1); idx 67 drivers/input/touchscreen/tsc40.c ptsc->idx = 0; idx 73 drivers/input/touchscreen/tsc40.c ptsc->idx = 0; idx 90 drivers/input/touchscreen/wacom_w8001.c int idx; idx 288 drivers/input/touchscreen/wacom_w8001.c w8001->data[w8001->idx] = data; idx 289 drivers/input/touchscreen/wacom_w8001.c switch (w8001->idx++) { idx 293 drivers/input/touchscreen/wacom_w8001.c w8001->idx = 0; idx 303 drivers/input/touchscreen/wacom_w8001.c if (w8001->pktlen == w8001->idx) { idx 304 drivers/input/touchscreen/wacom_w8001.c w8001->idx = 0; idx 323 drivers/input/touchscreen/wacom_w8001.c w8001->idx = 0; idx 334 drivers/input/touchscreen/wacom_w8001.c w8001->idx = 0; idx 342 drivers/input/touchscreen/wacom_w8001.c w8001->idx = 0; idx 352 drivers/input/touchscreen/wacom_w8001.c if (!w8001->touch_dev && w8001->idx > W8001_PKTLEN_TPCPEN - 1) idx 353 drivers/input/touchscreen/wacom_w8001.c w8001->idx = 0; idx 141 drivers/input/touchscreen/zylonite-wm97xx.c int idx; idx 148 drivers/input/touchscreen/zylonite-wm97xx.c for (idx = 0; idx < ARRAY_SIZE(cinfo); idx++) { idx 149 drivers/input/touchscreen/zylonite-wm97xx.c if (wm->id != cinfo[idx].id) idx 151 drivers/input/touchscreen/zylonite-wm97xx.c sp_idx = idx; idx 152 drivers/input/touchscreen/zylonite-wm97xx.c if (cont_rate <= cinfo[idx].speed) idx 261 drivers/interconnect/core.c unsigned int idx = spec->args[0]; idx 263 drivers/interconnect/core.c if (idx >= icc_data->num_nodes) { idx 264 drivers/interconnect/core.c pr_err("%s: invalid index %u\n", __func__, idx); idx 268 drivers/interconnect/core.c return icc_data->nodes[idx]; idx 323 drivers/interconnect/core.c int idx = 0; idx 344 drivers/interconnect/core.c idx = of_property_match_string(np, "interconnect-names", name); idx 345 drivers/interconnect/core.c if (idx < 0) idx 346 drivers/interconnect/core.c return ERR_PTR(idx); idx 350 drivers/interconnect/core.c "#interconnect-cells", idx * 2, idx 358 drivers/interconnect/core.c "#interconnect-cells", idx * 2 + 1, idx 563 drivers/interconnect/qcom/sdm845.c size_t idx = 0, batch = 0, cur_vcd_size = 0; idx 575 drivers/interconnect/qcom/sdm845.c tcs_cmd_gen(&tcs_list[idx], bcm->vote_x[bucket], idx 577 drivers/interconnect/qcom/sdm845.c idx++; idx 2451 drivers/iommu/amd_iommu_init.c int idx; idx 2464 drivers/iommu/amd_iommu_init.c for (idx = 0; idx < nr_ioapics; idx++) { idx 2465 drivers/iommu/amd_iommu_init.c int devid, id = mpc_ioapic_id(idx); idx 3085 drivers/iommu/amd_iommu_init.c struct amd_iommu *get_amd_iommu(unsigned int idx) idx 3091 drivers/iommu/amd_iommu_init.c if (i++ == idx) idx 3104 drivers/iommu/amd_iommu_init.c u8 amd_iommu_pc_get_max_banks(unsigned int idx) idx 3106 drivers/iommu/amd_iommu_init.c struct amd_iommu *iommu = get_amd_iommu(idx); idx 3121 drivers/iommu/amd_iommu_init.c u8 amd_iommu_pc_get_max_counters(unsigned int idx) idx 3123 drivers/iommu/amd_iommu_init.c struct amd_iommu *iommu = get_amd_iommu(idx); idx 949 drivers/iommu/arm-smmu-v3.c u32 idx = FIELD_GET(CMDQ_CONS_ERR, cons); idx 955 drivers/iommu/arm-smmu-v3.c idx < ARRAY_SIZE(cerror_str) ? cerror_str[idx] : "Unknown"); idx 957 drivers/iommu/arm-smmu-v3.c switch (idx) { idx 2108 drivers/iommu/arm-smmu-v3.c int idx, size = 1 << span; idx 2111 drivers/iommu/arm-smmu-v3.c idx = find_first_zero_bit(map, size); idx 2112 drivers/iommu/arm-smmu-v3.c if (idx == size) idx 2114 drivers/iommu/arm-smmu-v3.c } while (test_and_set_bit(idx, map)); idx 2116 drivers/iommu/arm-smmu-v3.c return idx; idx 2119 drivers/iommu/arm-smmu-v3.c static void arm_smmu_bitmap_free(unsigned long *map, int idx) idx 2121 drivers/iommu/arm-smmu-v3.c clear_bit(idx, map); idx 2281 drivers/iommu/arm-smmu-v3.c int idx; idx 2284 drivers/iommu/arm-smmu-v3.c idx = (sid >> STRTAB_SPLIT) * STRTAB_L1_DESC_DWORDS; idx 2285 drivers/iommu/arm-smmu-v3.c l1_desc = &cfg->l1_desc[idx]; idx 2286 drivers/iommu/arm-smmu-v3.c idx = (sid & ((1 << STRTAB_SPLIT) - 1)) * STRTAB_STE_DWORDS; idx 2287 drivers/iommu/arm-smmu-v3.c step = &l1_desc->l2ptr[idx]; idx 109 drivers/iommu/arm-smmu.c #define for_each_cfg_sme(fw, i, idx) \ idx 110 drivers/iommu/arm-smmu.c for (i = 0; idx = fwspec_smendx(fw, i), i < fw->num_ids; ++i) idx 224 drivers/iommu/arm-smmu.c int idx; idx 227 drivers/iommu/arm-smmu.c idx = find_next_zero_bit(map, end, start); idx 228 drivers/iommu/arm-smmu.c if (idx == end) idx 230 drivers/iommu/arm-smmu.c } while (test_and_set_bit(idx, map)); idx 232 drivers/iommu/arm-smmu.c return idx; idx 235 drivers/iommu/arm-smmu.c static void __arm_smmu_free_bitmap(unsigned long *map, int idx) idx 237 drivers/iommu/arm-smmu.c clear_bit(idx, map); idx 320 drivers/iommu/arm-smmu.c int reg, idx = cfg->cbndx; idx 331 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, reg, iova); idx 338 drivers/iommu/arm-smmu.c arm_smmu_cb_writeq(smmu, idx, reg, iova); idx 349 drivers/iommu/arm-smmu.c int reg, idx = smmu_domain->cfg.cbndx; idx 358 drivers/iommu/arm-smmu.c arm_smmu_cb_writeq(smmu, idx, reg, iova); idx 360 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, reg, iova); idx 453 drivers/iommu/arm-smmu.c int idx = smmu_domain->cfg.cbndx; idx 455 drivers/iommu/arm-smmu.c fsr = arm_smmu_cb_read(smmu, idx, ARM_SMMU_CB_FSR); idx 459 drivers/iommu/arm-smmu.c fsynr = arm_smmu_cb_read(smmu, idx, ARM_SMMU_CB_FSYNR0); idx 460 drivers/iommu/arm-smmu.c iova = arm_smmu_cb_readq(smmu, idx, ARM_SMMU_CB_FAR); idx 461 drivers/iommu/arm-smmu.c cbfrsynra = arm_smmu_gr1_read(smmu, ARM_SMMU_GR1_CBFRSYNRA(idx)); idx 465 drivers/iommu/arm-smmu.c fsr, iova, fsynr, cbfrsynra, idx); idx 467 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_FSR, fsr); idx 545 drivers/iommu/arm-smmu.c static void arm_smmu_write_context_bank(struct arm_smmu_device *smmu, int idx) idx 549 drivers/iommu/arm-smmu.c struct arm_smmu_cb *cb = &smmu->cbs[idx]; idx 554 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_SCTLR, 0); idx 570 drivers/iommu/arm-smmu.c arm_smmu_gr1_write(smmu, ARM_SMMU_GR1_CBA2R(idx), reg); idx 589 drivers/iommu/arm-smmu.c arm_smmu_gr1_write(smmu, ARM_SMMU_GR1_CBAR(idx), reg); idx 597 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TCR2, cb->tcr[1]); idx 598 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TCR, cb->tcr[0]); idx 602 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_CONTEXTIDR, cfg->asid); idx 603 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TTBR0, cb->ttbr[0]); idx 604 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_TTBR1, cb->ttbr[1]); idx 606 drivers/iommu/arm-smmu.c arm_smmu_cb_writeq(smmu, idx, ARM_SMMU_CB_TTBR0, cb->ttbr[0]); idx 608 drivers/iommu/arm-smmu.c arm_smmu_cb_writeq(smmu, idx, ARM_SMMU_CB_TTBR1, idx 614 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_S1_MAIR0, cb->mair[0]); idx 615 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_S1_MAIR1, cb->mair[1]); idx 625 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_SCTLR, reg); idx 896 drivers/iommu/arm-smmu.c static void arm_smmu_write_smr(struct arm_smmu_device *smmu, int idx) idx 898 drivers/iommu/arm-smmu.c struct arm_smmu_smr *smr = smmu->smrs + idx; idx 903 drivers/iommu/arm-smmu.c arm_smmu_gr0_write(smmu, ARM_SMMU_GR0_SMR(idx), reg); idx 906 drivers/iommu/arm-smmu.c static void arm_smmu_write_s2cr(struct arm_smmu_device *smmu, int idx) idx 908 drivers/iommu/arm-smmu.c struct arm_smmu_s2cr *s2cr = smmu->s2crs + idx; idx 914 drivers/iommu/arm-smmu.c smmu->smrs[idx].valid) idx 916 drivers/iommu/arm-smmu.c arm_smmu_gr0_write(smmu, ARM_SMMU_GR0_S2CR(idx), reg); idx 919 drivers/iommu/arm-smmu.c static void arm_smmu_write_sme(struct arm_smmu_device *smmu, int idx) idx 921 drivers/iommu/arm-smmu.c arm_smmu_write_s2cr(smmu, idx); idx 923 drivers/iommu/arm-smmu.c arm_smmu_write_smr(smmu, idx); idx 995 drivers/iommu/arm-smmu.c static bool arm_smmu_free_sme(struct arm_smmu_device *smmu, int idx) idx 997 drivers/iommu/arm-smmu.c if (--smmu->s2crs[idx].count) idx 1000 drivers/iommu/arm-smmu.c smmu->s2crs[idx] = s2cr_init_val; idx 1002 drivers/iommu/arm-smmu.c smmu->smrs[idx].valid = false; idx 1014 drivers/iommu/arm-smmu.c int i, idx, ret; idx 1018 drivers/iommu/arm-smmu.c for_each_cfg_sme(fwspec, i, idx) { idx 1022 drivers/iommu/arm-smmu.c if (idx != INVALID_SMENDX) { idx 1031 drivers/iommu/arm-smmu.c idx = ret; idx 1032 drivers/iommu/arm-smmu.c if (smrs && smmu->s2crs[idx].count == 0) { idx 1033 drivers/iommu/arm-smmu.c smrs[idx].id = sid; idx 1034 drivers/iommu/arm-smmu.c smrs[idx].mask = mask; idx 1035 drivers/iommu/arm-smmu.c smrs[idx].valid = true; idx 1037 drivers/iommu/arm-smmu.c smmu->s2crs[idx].count++; idx 1038 drivers/iommu/arm-smmu.c cfg->smendx[i] = (s16)idx; idx 1051 drivers/iommu/arm-smmu.c for_each_cfg_sme(fwspec, i, idx) { idx 1052 drivers/iommu/arm-smmu.c arm_smmu_write_sme(smmu, idx); idx 1053 drivers/iommu/arm-smmu.c smmu->s2crs[idx].group = group; idx 1072 drivers/iommu/arm-smmu.c int i, idx; idx 1075 drivers/iommu/arm-smmu.c for_each_cfg_sme(fwspec, i, idx) { idx 1076 drivers/iommu/arm-smmu.c if (arm_smmu_free_sme(smmu, idx)) idx 1077 drivers/iommu/arm-smmu.c arm_smmu_write_sme(smmu, idx); idx 1090 drivers/iommu/arm-smmu.c int i, idx; idx 1097 drivers/iommu/arm-smmu.c for_each_cfg_sme(fwspec, i, idx) { idx 1098 drivers/iommu/arm-smmu.c if (type == s2cr[idx].type && cbndx == s2cr[idx].cbndx) idx 1101 drivers/iommu/arm-smmu.c s2cr[idx].type = type; idx 1102 drivers/iommu/arm-smmu.c s2cr[idx].privcfg = S2CR_PRIVCFG_DEFAULT; idx 1103 drivers/iommu/arm-smmu.c s2cr[idx].cbndx = cbndx; idx 1104 drivers/iommu/arm-smmu.c arm_smmu_write_s2cr(smmu, idx); idx 1233 drivers/iommu/arm-smmu.c int ret, idx = cfg->cbndx; idx 1242 drivers/iommu/arm-smmu.c arm_smmu_cb_writeq(smmu, idx, ARM_SMMU_CB_ATS1PR, va); idx 1244 drivers/iommu/arm-smmu.c arm_smmu_cb_write(smmu, idx, ARM_SMMU_CB_ATS1PR, va); idx 1246 drivers/iommu/arm-smmu.c reg = arm_smmu_page(smmu, ARM_SMMU_CB(smmu, idx)) + ARM_SMMU_CB_ATSR; idx 1255 drivers/iommu/arm-smmu.c phys = arm_smmu_cb_readq(smmu, idx, ARM_SMMU_CB_PAR); idx 1420 drivers/iommu/arm-smmu.c int i, idx; idx 1422 drivers/iommu/arm-smmu.c for_each_cfg_sme(fwspec, i, idx) { idx 1423 drivers/iommu/arm-smmu.c if (group && smmu->s2crs[idx].group && idx 1424 drivers/iommu/arm-smmu.c group != smmu->s2crs[idx].group) idx 1427 drivers/iommu/arm-smmu.c group = smmu->s2crs[idx].group; idx 309 drivers/iommu/intel-iommu-debugfs.c int idx; idx 314 drivers/iommu/intel-iommu-debugfs.c for (idx = 0; idx < INTR_REMAP_TABLE_ENTRIES; idx++) { idx 315 drivers/iommu/intel-iommu-debugfs.c ri_entry = &iommu->ir_table->base[idx]; idx 320 drivers/iommu/intel-iommu-debugfs.c idx, PCI_BUS_NUM(ri_entry->sid), idx 333 drivers/iommu/intel-iommu-debugfs.c int idx; idx 338 drivers/iommu/intel-iommu-debugfs.c for (idx = 0; idx < INTR_REMAP_TABLE_ENTRIES; idx++) { idx 339 drivers/iommu/intel-iommu-debugfs.c pi_entry = &iommu->ir_table->base[idx]; idx 344 drivers/iommu/intel-iommu-debugfs.c idx, PCI_BUS_NUM(pi_entry->sid), idx 310 drivers/iommu/intel-iommu.c #define for_each_domain_iommu(idx, domain) \ idx 311 drivers/iommu/intel-iommu.c for (idx = 0; idx < g_num_of_iommus; idx++) \ idx 312 drivers/iommu/intel-iommu.c if (domain->iommu_refcnt[idx]) idx 487 drivers/iommu/intel-iommu.c int idx = did >> 8; idx 489 drivers/iommu/intel-iommu.c domains = iommu->domains[idx]; idx 500 drivers/iommu/intel-iommu.c int idx = did >> 8; idx 502 drivers/iommu/intel-iommu.c if (!iommu->domains[idx]) { idx 504 drivers/iommu/intel-iommu.c iommu->domains[idx] = kzalloc(size, GFP_ATOMIC); idx 507 drivers/iommu/intel-iommu.c domains = iommu->domains[idx]; idx 1535 drivers/iommu/intel-iommu.c int idx; idx 1539 drivers/iommu/intel-iommu.c for_each_domain_iommu(idx, domain) { idx 1540 drivers/iommu/intel-iommu.c struct intel_iommu *iommu = g_iommus[idx]; idx 2995 drivers/iommu/intel-iommu.c int tbl_idx, pos = 0, idx, devfn, ret = 0, did; idx 3006 drivers/iommu/intel-iommu.c idx = (ext ? devfn * 2 : devfn) % 256; idx 3008 drivers/iommu/intel-iommu.c if (idx == 0) { idx 3050 drivers/iommu/intel-iommu.c memcpy(&ce, old_ce + idx, sizeof(ce)); idx 3078 drivers/iommu/intel-iommu.c new_ce[idx] = ce; idx 3145 drivers/iommu/intel-iommu.c int idx = ext ? bus * 2 : bus; idx 3148 drivers/iommu/intel-iommu.c if (ctxt_tbls[idx]) { idx 3149 drivers/iommu/intel-iommu.c val = virt_to_phys(ctxt_tbls[idx]) | 1; idx 3153 drivers/iommu/intel-iommu.c if (!ext || !ctxt_tbls[idx + 1]) idx 3156 drivers/iommu/intel-iommu.c val = virt_to_phys(ctxt_tbls[idx + 1]) | 1; idx 568 drivers/iommu/io-pgtable-arm-v7s.c unsigned long iova, int idx, int lvl, idx 581 drivers/iommu/io-pgtable-arm-v7s.c ptep -= idx & (ARM_V7S_CONT_PAGES - 1); idx 645 drivers/iommu/io-pgtable-arm-v7s.c int idx, i = 0, num_entries = size >> ARM_V7S_LVL_SHIFT(lvl); idx 651 drivers/iommu/io-pgtable-arm-v7s.c idx = ARM_V7S_LVL_IDX(iova, lvl); idx 652 drivers/iommu/io-pgtable-arm-v7s.c ptep += idx; idx 673 drivers/iommu/io-pgtable-arm-v7s.c pte[0] = arm_v7s_split_cont(data, iova, idx, lvl, ptep); idx 470 drivers/iommu/iova.c unsigned idx = fq->tail; idx 474 drivers/iommu/iova.c fq->tail = (idx + 1) % IOVA_FQ_SIZE; idx 476 drivers/iommu/iova.c return idx; idx 482 drivers/iommu/iova.c unsigned idx; idx 486 drivers/iommu/iova.c fq_ring_for_each(idx, fq) { idx 488 drivers/iommu/iova.c if (fq->entries[idx].counter >= counter) idx 492 drivers/iommu/iova.c iovad->entry_dtor(fq->entries[idx].data); idx 495 drivers/iommu/iova.c fq->entries[idx].iova_pfn, idx 496 drivers/iommu/iova.c fq->entries[idx].pages); idx 523 drivers/iommu/iova.c int idx; idx 525 drivers/iommu/iova.c fq_ring_for_each(idx, fq) idx 526 drivers/iommu/iova.c iovad->entry_dtor(fq->entries[idx].data); idx 555 drivers/iommu/iova.c unsigned idx; idx 571 drivers/iommu/iova.c idx = fq_ring_add(fq); idx 573 drivers/iommu/iova.c fq->entries[idx].iova_pfn = pfn; idx 574 drivers/iommu/iova.c fq->entries[idx].pages = pages; idx 575 drivers/iommu/iova.c fq->entries[idx].data = data; idx 576 drivers/iommu/iova.c fq->entries[idx].counter = atomic64_read(&iovad->fq_flush_start_cnt); idx 198 drivers/iommu/msm_iommu.c int idx; idx 201 drivers/iommu/msm_iommu.c idx = find_next_zero_bit(map, end, start); idx 202 drivers/iommu/msm_iommu.c if (idx == end) idx 204 drivers/iommu/msm_iommu.c } while (test_and_set_bit(idx, map)); idx 206 drivers/iommu/msm_iommu.c return idx; idx 209 drivers/iommu/msm_iommu.c static void msm_iommu_free_ctx(unsigned long *map, int idx) idx 211 drivers/iommu/msm_iommu.c clear_bit(idx, map); idx 186 drivers/iommu/of_iommu.c int idx = 0; idx 190 drivers/iommu/of_iommu.c idx, &iommu_spec)) { idx 193 drivers/iommu/of_iommu.c idx++; idx 181 drivers/irqchip/irq-atmel-aic.c int idx; idx 192 drivers/irqchip/irq-atmel-aic.c idx = intspec[0] / dgc->irqs_per_chip; idx 193 drivers/irqchip/irq-atmel-aic.c if (idx >= dgc->num_chips) idx 196 drivers/irqchip/irq-atmel-aic.c gc = dgc->gc[idx]; idx 121 drivers/irqchip/irq-bcm6345-l1.c unsigned int idx; idx 131 drivers/irqchip/irq-bcm6345-l1.c for (idx = 0; idx < intc->n_words; idx++) { idx 132 drivers/irqchip/irq-bcm6345-l1.c int base = idx * IRQS_PER_WORD; idx 137 drivers/irqchip/irq-bcm6345-l1.c pending = __raw_readl(cpu->map_base + reg_status(intc, idx)); idx 138 drivers/irqchip/irq-bcm6345-l1.c pending &= __raw_readl(cpu->map_base + reg_enable(intc, idx)); idx 237 drivers/irqchip/irq-bcm6345-l1.c unsigned int idx, idx 245 drivers/irqchip/irq-bcm6345-l1.c if (of_address_to_resource(dn, idx, &res)) idx 255 drivers/irqchip/irq-bcm6345-l1.c cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32), idx 269 drivers/irqchip/irq-bcm6345-l1.c cpu->parent_irq = irq_of_parse_and_map(dn, idx); idx 306 drivers/irqchip/irq-bcm6345-l1.c unsigned int idx; idx 313 drivers/irqchip/irq-bcm6345-l1.c for_each_possible_cpu(idx) { idx 314 drivers/irqchip/irq-bcm6345-l1.c ret = bcm6345_l1_init_one(dn, idx, intc); idx 317 drivers/irqchip/irq-bcm6345-l1.c idx, ret); idx 319 drivers/irqchip/irq-bcm6345-l1.c cpumask_set_cpu(idx, &intc->cpumask); idx 339 drivers/irqchip/irq-bcm6345-l1.c for_each_cpu(idx, &intc->cpumask) { idx 340 drivers/irqchip/irq-bcm6345-l1.c struct bcm6345_l1_cpu *cpu = intc->cpus[idx]; idx 342 drivers/irqchip/irq-bcm6345-l1.c pr_info(" CPU%u at MMIO 0x%p (irq = %d)\n", idx, idx 349 drivers/irqchip/irq-bcm6345-l1.c for_each_possible_cpu(idx) { idx 350 drivers/irqchip/irq-bcm6345-l1.c struct bcm6345_l1_cpu *cpu = intc->cpus[idx]; idx 119 drivers/irqchip/irq-bcm7038-l1.c unsigned int idx; idx 129 drivers/irqchip/irq-bcm7038-l1.c for (idx = 0; idx < intc->n_words; idx++) { idx 130 drivers/irqchip/irq-bcm7038-l1.c int base = idx * IRQS_PER_WORD; idx 135 drivers/irqchip/irq-bcm7038-l1.c pending = l1_readl(cpu->map_base + reg_status(intc, idx)) & idx 136 drivers/irqchip/irq-bcm7038-l1.c ~cpu->mask_cache[idx]; idx 245 drivers/irqchip/irq-bcm7038-l1.c unsigned int idx, idx 253 drivers/irqchip/irq-bcm7038-l1.c if (of_address_to_resource(dn, idx, &res)) idx 265 drivers/irqchip/irq-bcm7038-l1.c cpu = intc->cpus[idx] = kzalloc(sizeof(*cpu) + n_words * sizeof(u32), idx 279 drivers/irqchip/irq-bcm7038-l1.c parent_irq = irq_of_parse_and_map(dn, idx); idx 322 drivers/irqchip/irq-bcm7038-l1.c int idx, ret; idx 329 drivers/irqchip/irq-bcm7038-l1.c for_each_possible_cpu(idx) { idx 330 drivers/irqchip/irq-bcm7038-l1.c ret = bcm7038_l1_init_one(dn, idx, intc); idx 332 drivers/irqchip/irq-bcm7038-l1.c if (idx) idx 353 drivers/irqchip/irq-bcm7038-l1.c for_each_possible_cpu(idx) { idx 354 drivers/irqchip/irq-bcm7038-l1.c struct bcm7038_l1_cpu *cpu = intc->cpus[idx]; idx 60 drivers/irqchip/irq-bcm7120-l2.c unsigned int idx; idx 64 drivers/irqchip/irq-bcm7120-l2.c for (idx = 0; idx < b->n_words; idx++) { idx 65 drivers/irqchip/irq-bcm7120-l2.c int base = idx * IRQS_PER_WORD; idx 72 drivers/irqchip/irq-bcm7120-l2.c pending = irq_reg_readl(gc, b->stat_offset[idx]) & idx 74 drivers/irqchip/irq-bcm7120-l2.c data->irq_map_mask[idx]; idx 114 drivers/irqchip/irq-bcm7120-l2.c unsigned int idx; idx 131 drivers/irqchip/irq-bcm7120-l2.c for (idx = 0; idx < data->n_words; idx++) { idx 133 drivers/irqchip/irq-bcm7120-l2.c l1_data->irq_map_mask[idx] |= idx 135 drivers/irqchip/irq-bcm7120-l2.c irq * data->n_words + idx); idx 137 drivers/irqchip/irq-bcm7120-l2.c l1_data->irq_map_mask[idx] = 0xffffffff; idx 139 drivers/irqchip/irq-bcm7120-l2.c valid_mask[idx] |= l1_data->irq_map_mask[idx]; idx 225 drivers/irqchip/irq-bcm7120-l2.c unsigned int idx, irq, flags; idx 280 drivers/irqchip/irq-bcm7120-l2.c for (idx = 0; idx < data->n_words; idx++) { idx 281 drivers/irqchip/irq-bcm7120-l2.c irq = idx * IRQS_PER_WORD; idx 284 drivers/irqchip/irq-bcm7120-l2.c gc->unused = 0xffffffff & ~valid_mask[idx]; idx 288 drivers/irqchip/irq-bcm7120-l2.c gc->reg_base = data->pair_base[idx]; idx 289 drivers/irqchip/irq-bcm7120-l2.c ct->regs.mask = data->en_offset[idx]; idx 292 drivers/irqchip/irq-bcm7120-l2.c irq_reg_writel(gc, data->irq_fwd_mask[idx], idx 293 drivers/irqchip/irq-bcm7120-l2.c data->en_offset[idx]); idx 328 drivers/irqchip/irq-bcm7120-l2.c for (idx = 0; idx < MAX_MAPPINGS; idx++) { idx 329 drivers/irqchip/irq-bcm7120-l2.c if (data->map_base[idx]) idx 330 drivers/irqchip/irq-bcm7120-l2.c iounmap(data->map_base[idx]); idx 75 drivers/irqchip/irq-csky-apb-intc.c static inline u32 build_channel_val(u32 idx, u32 magic) idx 82 drivers/irqchip/irq-csky-apb-intc.c res = idx | (idx << 8) | (idx << 16) | (idx << 24); idx 1740 drivers/irqchip/irq-gic-v3-its.c u32 idx = baser - its->tables; idx 1742 drivers/irqchip/irq-gic-v3-its.c return gits_read_baser(its->base + GITS_BASER + (idx << 3)); idx 1748 drivers/irqchip/irq-gic-v3-its.c u32 idx = baser - its->tables; idx 1750 drivers/irqchip/irq-gic-v3-its.c gits_write_baser(val, its->base + GITS_BASER + (idx << 3)); idx 2317 drivers/irqchip/irq-gic-v3-its.c u32 esz, idx; idx 2326 drivers/irqchip/irq-gic-v3-its.c idx = id >> ilog2(baser->psz / esz); idx 2327 drivers/irqchip/irq-gic-v3-its.c if (idx >= (PAGE_ORDER_TO_SIZE(baser->order) / GITS_LVL1_ENTRY_SIZE)) idx 2333 drivers/irqchip/irq-gic-v3-its.c if (!table[idx]) { idx 2343 drivers/irqchip/irq-gic-v3-its.c table[idx] = cpu_to_le64(page_to_phys(page) | GITS_BASER_VALID); idx 2347 drivers/irqchip/irq-gic-v3-its.c gic_flush_dcache_to_poc(table + idx, GITS_LVL1_ENTRY_SIZE); idx 2480 drivers/irqchip/irq-gic-v3-its.c int idx; idx 2483 drivers/irqchip/irq-gic-v3-its.c idx = bitmap_find_free_region(dev->event_map.lpi_map, idx 2486 drivers/irqchip/irq-gic-v3-its.c if (idx < 0) idx 2489 drivers/irqchip/irq-gic-v3-its.c *hwirq = dev->event_map.lpi_base + idx; idx 458 drivers/irqchip/irq-gic-v3.c u32 idx = gic_get_ppi_index(d); idx 461 drivers/irqchip/irq-gic-v3.c if (!refcount_inc_not_zero(&ppi_nmi_refs[idx])) { idx 462 drivers/irqchip/irq-gic-v3.c refcount_set(&ppi_nmi_refs[idx], 1); idx 495 drivers/irqchip/irq-gic-v3.c u32 idx = gic_get_ppi_index(d); idx 498 drivers/irqchip/irq-gic-v3.c if (refcount_dec_and_test(&ppi_nmi_refs[idx])) idx 76 drivers/irqchip/irq-imx-gpcv2.c unsigned int idx = d->hwirq / 32; idx 82 drivers/irqchip/irq-imx-gpcv2.c val = cd->wakeup_sources[idx]; idx 84 drivers/irqchip/irq-imx-gpcv2.c cd->wakeup_sources[idx] = on ? (val & ~mask) : (val | mask); idx 48 drivers/irqchip/irq-imx-irqsteer.c int idx = imx_irqsteer_get_reg_index(data, d->hwirq); idx 53 drivers/irqchip/irq-imx-irqsteer.c val = readl_relaxed(data->regs + CHANMASK(idx, data->reg_num)); idx 55 drivers/irqchip/irq-imx-irqsteer.c writel_relaxed(val, data->regs + CHANMASK(idx, data->reg_num)); idx 62 drivers/irqchip/irq-imx-irqsteer.c int idx = imx_irqsteer_get_reg_index(data, d->hwirq); idx 67 drivers/irqchip/irq-imx-irqsteer.c val = readl_relaxed(data->regs + CHANMASK(idx, data->reg_num)); idx 69 drivers/irqchip/irq-imx-irqsteer.c writel_relaxed(val, data->regs + CHANMASK(idx, data->reg_num)); idx 123 drivers/irqchip/irq-imx-irqsteer.c int idx = imx_irqsteer_get_reg_index(data, hwirq); idx 131 drivers/irqchip/irq-imx-irqsteer.c CHANSTATUS(idx, data->reg_num)); idx 113 drivers/irqchip/irq-meson-gpio.c unsigned int reg, idx; idx 118 drivers/irqchip/irq-meson-gpio.c idx = find_first_zero_bit(ctl->channel_map, NUM_CHANNEL); idx 119 drivers/irqchip/irq-meson-gpio.c if (idx >= NUM_CHANNEL) { idx 126 drivers/irqchip/irq-meson-gpio.c set_bit(idx, ctl->channel_map); idx 132 drivers/irqchip/irq-meson-gpio.c reg = meson_gpio_irq_channel_to_reg(idx); idx 134 drivers/irqchip/irq-meson-gpio.c 0xff << REG_PIN_SEL_SHIFT(idx), idx 135 drivers/irqchip/irq-meson-gpio.c hwirq << REG_PIN_SEL_SHIFT(idx)); idx 143 drivers/irqchip/irq-meson-gpio.c *channel_hwirq = &(ctl->channel_irqs[idx]); idx 148 drivers/irqchip/irq-meson-gpio.c hwirq, idx, **channel_hwirq); idx 164 drivers/irqchip/irq-meson-gpio.c unsigned int idx; idx 166 drivers/irqchip/irq-meson-gpio.c idx = meson_gpio_irq_get_channel_idx(ctl, channel_hwirq); idx 167 drivers/irqchip/irq-meson-gpio.c clear_bit(idx, ctl->channel_map); idx 175 drivers/irqchip/irq-meson-gpio.c unsigned int idx; idx 177 drivers/irqchip/irq-meson-gpio.c idx = meson_gpio_irq_get_channel_idx(ctl, channel_hwirq); idx 196 drivers/irqchip/irq-meson-gpio.c val |= REG_BOTH_EDGE(idx); idx 199 drivers/irqchip/irq-meson-gpio.c val |= REG_EDGE_POL_EDGE(idx); idx 202 drivers/irqchip/irq-meson-gpio.c val |= REG_EDGE_POL_LOW(idx); idx 208 drivers/irqchip/irq-meson-gpio.c REG_EDGE_POL_MASK(idx), val); idx 40 drivers/irqchip/irq-mvebu-gicp.c static int gicp_idx_to_spi(struct mvebu_gicp *gicp, int idx) idx 47 drivers/irqchip/irq-mvebu-gicp.c if (idx < r->count) idx 48 drivers/irqchip/irq-mvebu-gicp.c return r->start + idx; idx 50 drivers/irqchip/irq-mvebu-gicp.c idx -= r->count; idx 18 drivers/irqchip/irq-mvebu-sei.c #define GICP_SECR(idx) (0x0 + ((idx) * 0x4)) idx 20 drivers/irqchip/irq-mvebu-sei.c #define GICP_SEMR(idx) (0x20 + ((idx) * 0x4)) idx 329 drivers/irqchip/irq-mvebu-sei.c u32 idx; idx 333 drivers/irqchip/irq-mvebu-sei.c for (idx = 0; idx < SEI_IRQ_REG_COUNT; idx++) { idx 337 drivers/irqchip/irq-mvebu-sei.c irqmap = readl_relaxed(sei->base + GICP_SECR(idx)); idx 342 drivers/irqchip/irq-mvebu-sei.c hwirq = idx * SEI_IRQ_COUNT_PER_REG + bit; idx 173 drivers/isdn/hardware/mISDN/avmfritz.c u8 idx = (offset > 0x2f) ? AVM_ISAC_REG_HIGH : AVM_ISAC_REG_LOW; idx 175 drivers/isdn/hardware/mISDN/avmfritz.c outb(idx, fc->addr + CHIP_INDEX); idx 183 drivers/isdn/hardware/mISDN/avmfritz.c u8 idx = (offset > 0x2f) ? AVM_ISAC_REG_HIGH : AVM_ISAC_REG_LOW; idx 185 drivers/isdn/hardware/mISDN/avmfritz.c outb(idx, fc->addr + CHIP_INDEX); idx 262 drivers/isdn/hardware/mISDN/avmfritz.c u32 idx = channel == 2 ? AVM_HDLC_2 : AVM_HDLC_1; idx 264 drivers/isdn/hardware/mISDN/avmfritz.c outl(idx, fc->addr + CHIP_INDEX); idx 440 drivers/isdn/hardware/mISDN/avmfritz.c int count, fs, cnt = 0, idx; idx 445 drivers/isdn/hardware/mISDN/avmfritz.c idx = (bch->nr - 1) & 1; idx 446 drivers/isdn/hardware/mISDN/avmfritz.c hdlc = &fc->hdlc[idx]; idx 1917 drivers/isdn/hardware/mISDN/hfcsusb.c int ifnum = iface->desc.bInterfaceNumber, i, idx, alt_idx, idx 1968 drivers/isdn/hardware/mISDN/hfcsusb.c idx = ((ep_addr & 0x7f) - 1) * 2; idx 1969 drivers/isdn/hardware/mISDN/hfcsusb.c if (idx > 15) idx 1973 drivers/isdn/hardware/mISDN/hfcsusb.c idx++; idx 1976 drivers/isdn/hardware/mISDN/hfcsusb.c if (cmptbl[idx] != EP_NOP) { idx 1977 drivers/isdn/hardware/mISDN/hfcsusb.c if (cmptbl[idx] == EP_NUL) idx 1980 drivers/isdn/hardware/mISDN/hfcsusb.c && cmptbl[idx] == EP_INT) idx 1981 drivers/isdn/hardware/mISDN/hfcsusb.c cmptbl[idx] = EP_NUL; idx 1983 drivers/isdn/hardware/mISDN/hfcsusb.c && cmptbl[idx] == EP_BLK) idx 1984 drivers/isdn/hardware/mISDN/hfcsusb.c cmptbl[idx] = EP_NUL; idx 1986 drivers/isdn/hardware/mISDN/hfcsusb.c && cmptbl[idx] == EP_ISO) idx 1987 drivers/isdn/hardware/mISDN/hfcsusb.c cmptbl[idx] = EP_NUL; idx 2031 drivers/isdn/hardware/mISDN/hfcsusb.c idx = ((ep_addr & 0x7f) - 1) * 2; idx 2033 drivers/isdn/hardware/mISDN/hfcsusb.c idx++; idx 2034 drivers/isdn/hardware/mISDN/hfcsusb.c f = &hw->fifos[idx & 7]; idx 2037 drivers/isdn/hardware/mISDN/hfcsusb.c if (vcf[idx] == EP_NOP || vcf[idx] == EP_NUL) { idx 2073 drivers/isdn/hardware/mISDN/hfcsusb.c f->fifonum = idx & 7; idx 32 drivers/isdn/hardware/mISDN/netjet.c int idx; idx 44 drivers/isdn/hardware/mISDN/netjet.c int idx; idx 175 drivers/isdn/hardware/mISDN/netjet.c fill_mem(struct tiger_ch *bc, u32 idx, u32 cnt, u32 fill) idx 181 drivers/isdn/hardware/mISDN/netjet.c bc->bch.nr, fill, cnt, idx, card->send.idx); idx 188 drivers/isdn/hardware/mISDN/netjet.c val = card->send.start[idx]; idx 191 drivers/isdn/hardware/mISDN/netjet.c card->send.start[idx++] = val; idx 192 drivers/isdn/hardware/mISDN/netjet.c if (idx >= card->send.size) idx 193 drivers/isdn/hardware/mISDN/netjet.c idx = 0; idx 226 drivers/isdn/hardware/mISDN/netjet.c bc->idx = 0; idx 240 drivers/isdn/hardware/mISDN/netjet.c bc->idx = 0; idx 260 drivers/isdn/hardware/mISDN/netjet.c card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; idx 261 drivers/isdn/hardware/mISDN/netjet.c card->recv.idx = (card->recv.dmacur - card->recv.dmastart) >> 2; idx 267 drivers/isdn/hardware/mISDN/netjet.c card->send.idx, idx 268 drivers/isdn/hardware/mISDN/netjet.c card->recv.idx); idx 363 drivers/isdn/hardware/mISDN/netjet.c read_dma(struct tiger_ch *bc, u32 idx, int cnt) idx 370 drivers/isdn/hardware/mISDN/netjet.c if (bc->lastrx == idx) { idx 373 drivers/isdn/hardware/mISDN/netjet.c bc->bch.nr, idx); idx 375 drivers/isdn/hardware/mISDN/netjet.c bc->lastrx = idx; idx 393 drivers/isdn/hardware/mISDN/netjet.c val = card->recv.start[idx++]; idx 396 drivers/isdn/hardware/mISDN/netjet.c if (idx >= card->recv.size) idx 397 drivers/isdn/hardware/mISDN/netjet.c idx = 0; idx 445 drivers/isdn/hardware/mISDN/netjet.c u32 idx; idx 453 drivers/isdn/hardware/mISDN/netjet.c idx = cnt - 1; idx 455 drivers/isdn/hardware/mISDN/netjet.c idx = card->recv.size - 1; idx 458 drivers/isdn/hardware/mISDN/netjet.c read_dma(&card->bc[0], idx, cnt); idx 460 drivers/isdn/hardware/mISDN/netjet.c read_dma(&card->bc[1], idx, cnt); idx 468 drivers/isdn/hardware/mISDN/netjet.c card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; idx 475 drivers/isdn/hardware/mISDN/netjet.c if (card->send.idx < ((card->send.size / 2) - 1)) idx 476 drivers/isdn/hardware/mISDN/netjet.c bc->idx = (card->recv.size / 2) - 1; idx 478 drivers/isdn/hardware/mISDN/netjet.c bc->idx = card->recv.size - 1; idx 481 drivers/isdn/hardware/mISDN/netjet.c __func__, bc->bch.nr, bc->free, bc->idx, card->send.idx); idx 498 drivers/isdn/hardware/mISDN/netjet.c bc->idx, card->send.idx); idx 509 drivers/isdn/hardware/mISDN/netjet.c if (bc->idx >= card->send.size) idx 510 drivers/isdn/hardware/mISDN/netjet.c bc->idx = 0; idx 511 drivers/isdn/hardware/mISDN/netjet.c v = card->send.start[bc->idx]; idx 514 drivers/isdn/hardware/mISDN/netjet.c card->send.start[bc->idx++] = v; idx 546 drivers/isdn/hardware/mISDN/netjet.c bc->idx, card->send.idx); idx 572 drivers/isdn/hardware/mISDN/netjet.c if (bc->idx >= card->send.size) idx 573 drivers/isdn/hardware/mISDN/netjet.c bc->idx = 0; idx 574 drivers/isdn/hardware/mISDN/netjet.c v = card->send.start[bc->idx]; idx 577 drivers/isdn/hardware/mISDN/netjet.c card->send.start[bc->idx++] = v; idx 581 drivers/isdn/hardware/mISDN/netjet.c if (bc->idx >= card->send.size) idx 582 drivers/isdn/hardware/mISDN/netjet.c bc->idx = 0; idx 583 drivers/isdn/hardware/mISDN/netjet.c v = card->send.start[bc->idx]; idx 587 drivers/isdn/hardware/mISDN/netjet.c card->send.start[bc->idx++] = v; idx 645 drivers/isdn/hardware/mISDN/netjet.c bc->bch.nr, bc->free, bc->idx, card->send.idx); idx 647 drivers/isdn/hardware/mISDN/netjet.c fill_mem(bc, bc->idx, bc->free, 0xff); idx 704 drivers/isdn/hardware/mISDN/netjet.c card->recv.idx = (card->recv.dmacur - card->recv.dmastart) >> 2; idx 711 drivers/isdn/hardware/mISDN/netjet.c card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; idx 719 drivers/isdn/hardware/mISDN/netjet.c card->recv.idx, card->send.idx); idx 53 drivers/isdn/mISDN/l1oip.h int idx; /* card index */ idx 1293 drivers/isdn/mISDN/l1oip_core.c hc->idx = l1oip_cnt; idx 210 drivers/leds/leds-gpio.c static struct gpio_desc *gpio_led_get_gpiod(struct device *dev, int idx, idx 223 drivers/leds/leds-gpio.c gpiod = devm_gpiod_get_index(dev, NULL, idx, flags); idx 123 drivers/leds/leds-lp5521.c enum lp55xx_engine_index idx = chip->engine_idx; idx 136 drivers/leds/leds-lp5521.c lp55xx_update_bits(chip, LP5521_REG_OP_MODE, mask[idx], val[idx]); idx 149 drivers/leds/leds-lp5521.c enum lp55xx_engine_index idx = chip->engine_idx; idx 156 drivers/leds/leds-lp5521.c lp55xx_update_bits(chip, LP5521_REG_OP_MODE, mask[idx], 0); idx 214 drivers/leds/leds-lp5521.c enum lp55xx_engine_index idx = chip->engine_idx; idx 248 drivers/leds/leds-lp5521.c ret = lp55xx_write(chip, addr[idx] + i, pattern[i]); idx 157 drivers/leds/leds-lp5523.c enum lp55xx_engine_index idx = chip->engine_idx; idx 170 drivers/leds/leds-lp5523.c lp55xx_update_bits(chip, LP5523_REG_OP_MODE, mask[idx], val[idx]); idx 177 drivers/leds/leds-lp5523.c enum lp55xx_engine_index idx = chip->engine_idx; idx 186 drivers/leds/leds-lp5523.c lp55xx_write(chip, LP5523_REG_PROG_PAGE_SEL, page_sel[idx]); idx 197 drivers/leds/leds-lp5523.c enum lp55xx_engine_index idx = chip->engine_idx; idx 204 drivers/leds/leds-lp5523.c lp55xx_update_bits(chip, LP5523_REG_OP_MODE, mask[idx], 0); idx 129 drivers/leds/leds-lp5562.c enum lp55xx_engine_index idx = chip->engine_idx; idx 142 drivers/leds/leds-lp5562.c lp55xx_update_bits(chip, LP5562_REG_OP_MODE, mask[idx], val[idx]); idx 209 drivers/leds/leds-lp5562.c enum lp55xx_engine_index idx = chip->engine_idx; idx 226 drivers/leds/leds-lp5562.c lp55xx_write(chip, addr[idx] + i, 0); idx 250 drivers/leds/leds-lp5562.c lp55xx_write(chip, addr[idx] + i, pattern[i]); idx 453 drivers/leds/leds-lp5562.c enum lp55xx_engine_index idx = chip->engine_idx; idx 456 drivers/leds/leds-lp5562.c switch (idx) { idx 197 drivers/leds/leds-lp55xx-common.c enum lp55xx_engine_index idx = chip->engine_idx; idx 207 drivers/leds/leds-lp55xx-common.c chip->engines[idx - 1].mode = LP55XX_ENGINE_LOAD; idx 116 drivers/leds/leds-lp8501.c enum lp55xx_engine_index idx = chip->engine_idx; idx 135 drivers/leds/leds-lp8501.c lp55xx_update_bits(chip, LP8501_REG_OP_MODE, mask[idx], val[idx]); idx 139 drivers/leds/leds-lp8501.c lp55xx_write(chip, LP8501_REG_PROG_PAGE_SEL, page_sel[idx]); idx 1363 drivers/macintosh/via-pmu.c int idx; idx 1392 drivers/macintosh/via-pmu.c idx = ffs(ints) - 1; idx 1393 drivers/macintosh/via-pmu.c ints &= ~BIT(idx); idx 1395 drivers/macintosh/via-pmu.c pmu_irq_stats[idx]++; idx 1401 drivers/macintosh/via-pmu.c switch (BIT(idx)) { idx 48 drivers/mailbox/imx-mailbox.c unsigned int idx; idx 119 drivers/mailbox/imx-mailbox.c val &= IMX_MU_xSR_TEn(cp->idx) & idx 120 drivers/mailbox/imx-mailbox.c (ctrl & IMX_MU_xCR_TIEn(cp->idx)); idx 123 drivers/mailbox/imx-mailbox.c val &= IMX_MU_xSR_RFn(cp->idx) & idx 124 drivers/mailbox/imx-mailbox.c (ctrl & IMX_MU_xCR_RIEn(cp->idx)); idx 127 drivers/mailbox/imx-mailbox.c val &= IMX_MU_xSR_GIPn(cp->idx) & idx 128 drivers/mailbox/imx-mailbox.c (ctrl & IMX_MU_xCR_GIEn(cp->idx)); idx 137 drivers/mailbox/imx-mailbox.c if (val == IMX_MU_xSR_TEn(cp->idx)) { idx 138 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, 0, IMX_MU_xCR_TIEn(cp->idx)); idx 140 drivers/mailbox/imx-mailbox.c } else if (val == IMX_MU_xSR_RFn(cp->idx)) { idx 141 drivers/mailbox/imx-mailbox.c dat = imx_mu_read(priv, IMX_MU_xRRn(cp->idx)); idx 143 drivers/mailbox/imx-mailbox.c } else if (val == IMX_MU_xSR_GIPn(cp->idx)) { idx 144 drivers/mailbox/imx-mailbox.c imx_mu_write(priv, IMX_MU_xSR_GIPn(cp->idx), IMX_MU_xSR); idx 162 drivers/mailbox/imx-mailbox.c imx_mu_write(priv, *arg, IMX_MU_xTRn(cp->idx)); idx 163 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, IMX_MU_xCR_TIEn(cp->idx), 0); idx 166 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, IMX_MU_xCR_GIRn(cp->idx), 0); idx 200 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, IMX_MU_xCR_RIEn(cp->idx), 0); idx 203 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, IMX_MU_xCR_GIEn(cp->idx), 0); idx 224 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, 0, IMX_MU_xCR_TIEn(cp->idx)); idx 227 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, 0, IMX_MU_xCR_RIEn(cp->idx)); idx 230 drivers/mailbox/imx-mailbox.c imx_mu_xcr_rmw(priv, 0, IMX_MU_xCR_GIEn(cp->idx)); idx 248 drivers/mailbox/imx-mailbox.c u32 type, idx, chan; idx 256 drivers/mailbox/imx-mailbox.c idx = sp->args[1]; /* index */ idx 257 drivers/mailbox/imx-mailbox.c chan = type * 4 + idx; idx 260 drivers/mailbox/imx-mailbox.c dev_err(mbox->dev, "Not supported channel number: %d. (type: %d, idx: %d)\n", chan, type, idx); idx 315 drivers/mailbox/imx-mailbox.c cp->idx = i % 4; idx 320 drivers/mailbox/imx-mailbox.c "imx_mu_chan[%i-%i]", cp->type, cp->idx); idx 28 drivers/mailbox/mailbox.c int idx; idx 39 drivers/mailbox/mailbox.c idx = chan->msg_free; idx 40 drivers/mailbox/mailbox.c chan->msg_data[idx] = mssg; idx 43 drivers/mailbox/mailbox.c if (idx == MBOX_TX_QUEUE_LEN - 1) idx 50 drivers/mailbox/mailbox.c return idx; idx 55 drivers/mailbox/mailbox.c unsigned count, idx; idx 66 drivers/mailbox/mailbox.c idx = chan->msg_free; idx 67 drivers/mailbox/mailbox.c if (idx >= count) idx 68 drivers/mailbox/mailbox.c idx -= count; idx 70 drivers/mailbox/mailbox.c idx += MBOX_TX_QUEUE_LEN - count; idx 72 drivers/mailbox/mailbox.c data = chan->msg_data[idx]; idx 39 drivers/mailbox/qcom-apcs-ipc-mailbox.c unsigned long idx = (unsigned long)chan->con_priv; idx 41 drivers/mailbox/qcom-apcs-ipc-mailbox.c return regmap_write(apcs->regmap, apcs->offset, BIT(idx)); idx 35 drivers/mailbox/rockchip-mailbox.c int idx; idx 68 drivers/mailbox/rockchip-mailbox.c chans->idx, msg->cmd); idx 70 drivers/mailbox/rockchip-mailbox.c mb->chans[chans->idx].msg = msg; idx 72 drivers/mailbox/rockchip-mailbox.c writel_relaxed(msg->cmd, mb->mbox_base + MAILBOX_A2B_CMD(chans->idx)); idx 74 drivers/mailbox/rockchip-mailbox.c MAILBOX_A2B_DAT(chans->idx)); idx 98 drivers/mailbox/rockchip-mailbox.c mb->chans[chans->idx].msg = NULL; idx 109 drivers/mailbox/rockchip-mailbox.c int idx; idx 113 drivers/mailbox/rockchip-mailbox.c for (idx = 0; idx < mb->mbox.num_chans; idx++) { idx 114 drivers/mailbox/rockchip-mailbox.c if ((status & (1 << idx)) && (irq == mb->chans[idx].irq)) { idx 116 drivers/mailbox/rockchip-mailbox.c writel_relaxed(1 << idx, idx 127 drivers/mailbox/rockchip-mailbox.c int idx; idx 131 drivers/mailbox/rockchip-mailbox.c for (idx = 0; idx < mb->mbox.num_chans; idx++) { idx 132 drivers/mailbox/rockchip-mailbox.c if (irq != mb->chans[idx].irq) idx 135 drivers/mailbox/rockchip-mailbox.c msg = mb->chans[idx].msg; idx 138 drivers/mailbox/rockchip-mailbox.c "Chan[%d]: B2A message is NULL\n", idx); idx 142 drivers/mailbox/rockchip-mailbox.c mbox_chan_received_data(&mb->mbox.chans[idx], msg); idx 143 drivers/mailbox/rockchip-mailbox.c mb->chans[idx].msg = NULL; idx 146 drivers/mailbox/rockchip-mailbox.c idx, msg->cmd); idx 236 drivers/mailbox/rockchip-mailbox.c mb->chans[i].idx = i; idx 581 drivers/mailbox/ti-msgmgr.c static int ti_msgmgr_queue_setup(int idx, struct device *dev, idx 596 drivers/mailbox/ti-msgmgr.c idx, qinst->queue_id, d->queue_count); idx 640 drivers/mailbox/ti-msgmgr.c idx, qinst->queue_id, qinst->proxy_id, qinst->irq, idx 403 drivers/md/bcache/bset.h static inline struct bkey *bset_bkey_idx(struct bset *i, unsigned int idx) idx 405 drivers/md/bcache/bset.h return bkey_idx(i->start, idx); idx 772 drivers/md/bcache/super.c static inline int idx_to_first_minor(int idx) idx 774 drivers/md/bcache/super.c return (idx * BCACHE_MINORS); idx 817 drivers/md/bcache/super.c int idx; idx 840 drivers/md/bcache/super.c idx = ida_simple_get(&bcache_device_idx, 0, idx 842 drivers/md/bcache/super.c if (idx < 0) idx 843 drivers/md/bcache/super.c return idx; idx 854 drivers/md/bcache/super.c snprintf(d->disk->disk_name, DISK_NAME_LEN, "bcache%i", idx); idx 857 drivers/md/bcache/super.c d->disk->first_minor = idx_to_first_minor(idx); idx 887 drivers/md/bcache/super.c ida_simple_remove(&bcache_device_idx, idx); idx 314 drivers/md/dm-integrity.c unsigned idx; idx 2499 drivers/md/dm-integrity.c bbs->idx * (BITMAP_BLOCK_SIZE >> SECTOR_SHIFT), idx 3205 drivers/md/dm-integrity.c unsigned idx; idx 3221 drivers/md/dm-integrity.c for (idx = start_index; idx <= end_index; idx++) { idx 3222 drivers/md/dm-integrity.c char *va = lowmem_page_address(pl[idx].page); idx 3224 drivers/md/dm-integrity.c if (idx == start_index) idx 3226 drivers/md/dm-integrity.c if (idx == end_index) idx 3228 drivers/md/dm-integrity.c sg_set_buf(&s[idx - start_index], va + start, end - start); idx 4055 drivers/md/dm-integrity.c bbs->idx = i; idx 36 drivers/md/dm-log-userspace-transfer.c .idx = CN_IDX_DM, idx 63 drivers/md/dm-log-userspace-transfer.c msg->id.idx = ulog_cn_id.idx; idx 515 drivers/md/dm-stats.c int idx, sector_t len, idx 547 drivers/md/dm-stats.c atomic_inc(&shared->in_flight[idx]); idx 551 drivers/md/dm-stats.c atomic_dec(&shared->in_flight[idx]); idx 552 drivers/md/dm-stats.c p->sectors[idx] += len; idx 553 drivers/md/dm-stats.c p->ios[idx] += 1; idx 554 drivers/md/dm-stats.c p->merges[idx] += stats_aux->merged; idx 556 drivers/md/dm-stats.c p->ticks[idx] += duration_jiffies; idx 559 drivers/md/dm-stats.c p->ticks[idx] += stats_aux->duration_ns; idx 195 drivers/md/dm-verity-target.c unsigned idx; idx 202 drivers/md/dm-verity-target.c idx = position & ((1 << v->hash_per_block_bits) - 1); idx 204 drivers/md/dm-verity-target.c *offset = idx * v->digest_size; idx 206 drivers/md/dm-verity-target.c *offset = idx << (v->hash_dev_block_bits - v->hash_per_block_bits); idx 256 drivers/md/persistent-data/dm-space-map-common.c struct disk_index_entry idx; idx 262 drivers/md/persistent-data/dm-space-map-common.c idx.blocknr = cpu_to_le64(dm_block_location(b)); idx 266 drivers/md/persistent-data/dm-space-map-common.c idx.nr_free = cpu_to_le32(ll->entries_per_block); idx 267 drivers/md/persistent-data/dm-space-map-common.c idx.none_free_before = 0; idx 269 drivers/md/persistent-data/dm-space-map-common.c r = ll->save_ie(ll, i, &idx); idx 78 drivers/md/raid1-10.c unsigned idx) idx 80 drivers/md/raid1-10.c if (WARN_ON_ONCE(idx >= RESYNC_PAGES)) idx 82 drivers/md/raid1-10.c return rp->pages[idx]; idx 98 drivers/md/raid1-10.c int idx = 0; idx 102 drivers/md/raid1-10.c struct page *page = resync_fetch_page(rp, idx); idx 111 drivers/md/raid1-10.c } while (idx++ < RESYNC_PAGES && size > 0); idx 262 drivers/md/raid1.c int idx; idx 264 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); idx 267 drivers/md/raid1.c atomic_inc(&conf->nr_queued[idx]); idx 883 drivers/md/raid1.c int idx = sector_to_idx(sector_nr); idx 889 drivers/md/raid1.c !atomic_read(&conf->nr_waiting[idx]), idx 893 drivers/md/raid1.c atomic_inc(&conf->barrier[idx]); idx 913 drivers/md/raid1.c !atomic_read(&conf->nr_pending[idx]) && idx 914 drivers/md/raid1.c atomic_read(&conf->barrier[idx]) < RESYNC_DEPTH) || idx 919 drivers/md/raid1.c atomic_dec(&conf->barrier[idx]); idx 933 drivers/md/raid1.c int idx = sector_to_idx(sector_nr); idx 935 drivers/md/raid1.c BUG_ON(atomic_read(&conf->barrier[idx]) <= 0); idx 937 drivers/md/raid1.c atomic_dec(&conf->barrier[idx]); idx 942 drivers/md/raid1.c static void _wait_barrier(struct r1conf *conf, int idx) idx 952 drivers/md/raid1.c atomic_inc(&conf->nr_pending[idx]); idx 973 drivers/md/raid1.c !atomic_read(&conf->barrier[idx])) idx 984 drivers/md/raid1.c atomic_inc(&conf->nr_waiting[idx]); idx 985 drivers/md/raid1.c atomic_dec(&conf->nr_pending[idx]); idx 994 drivers/md/raid1.c !atomic_read(&conf->barrier[idx]), idx 996 drivers/md/raid1.c atomic_inc(&conf->nr_pending[idx]); idx 997 drivers/md/raid1.c atomic_dec(&conf->nr_waiting[idx]); idx 1003 drivers/md/raid1.c int idx = sector_to_idx(sector_nr); idx 1012 drivers/md/raid1.c atomic_inc(&conf->nr_pending[idx]); idx 1018 drivers/md/raid1.c atomic_inc(&conf->nr_waiting[idx]); idx 1019 drivers/md/raid1.c atomic_dec(&conf->nr_pending[idx]); idx 1029 drivers/md/raid1.c atomic_inc(&conf->nr_pending[idx]); idx 1030 drivers/md/raid1.c atomic_dec(&conf->nr_waiting[idx]); idx 1036 drivers/md/raid1.c int idx = sector_to_idx(sector_nr); idx 1038 drivers/md/raid1.c _wait_barrier(conf, idx); idx 1041 drivers/md/raid1.c static void _allow_barrier(struct r1conf *conf, int idx) idx 1043 drivers/md/raid1.c atomic_dec(&conf->nr_pending[idx]); idx 1049 drivers/md/raid1.c int idx = sector_to_idx(sector_nr); idx 1051 drivers/md/raid1.c _allow_barrier(conf, idx); idx 1057 drivers/md/raid1.c int idx, ret; idx 1060 drivers/md/raid1.c for (idx = 0; idx < BARRIER_BUCKETS_NR; idx++) idx 1061 drivers/md/raid1.c ret += atomic_read(&conf->nr_pending[idx]) - idx 1062 drivers/md/raid1.c atomic_read(&conf->nr_queued[idx]); idx 1680 drivers/md/raid1.c int idx; idx 1682 drivers/md/raid1.c for (idx = 0; idx < BARRIER_BUCKETS_NR; idx++) { idx 1683 drivers/md/raid1.c _wait_barrier(conf, idx); idx 1684 drivers/md/raid1.c _allow_barrier(conf, idx); idx 1991 drivers/md/raid1.c int idx = 0; idx 2022 drivers/md/raid1.c pages[idx], idx 2062 drivers/md/raid1.c idx++; idx 2076 drivers/md/raid1.c pages[idx], idx 2091 drivers/md/raid1.c pages[idx], idx 2097 drivers/md/raid1.c idx ++; idx 2435 drivers/md/raid1.c int m, idx; idx 2463 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); idx 2464 drivers/md/raid1.c atomic_inc(&conf->nr_queued[idx]); idx 2529 drivers/md/raid1.c int idx; idx 2544 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); idx 2545 drivers/md/raid1.c atomic_dec(&conf->nr_queued[idx]); idx 2566 drivers/md/raid1.c idx = sector_to_idx(r1_bio->sector); idx 2567 drivers/md/raid1.c atomic_dec(&conf->nr_queued[idx]); idx 2646 drivers/md/raid1.c int idx = sector_to_idx(sector_nr); idx 2697 drivers/md/raid1.c if (atomic_read(&conf->nr_waiting[idx])) idx 2155 drivers/md/raid10.c int idx = 0; idx 2174 drivers/md/raid10.c pages[idx], idx 2182 drivers/md/raid10.c pages[idx], idx 2220 drivers/md/raid10.c idx++; idx 4772 drivers/md/raid10.c int idx = 0; idx 4805 drivers/md/raid10.c addr = r10b->devs[slot].addr + idx * PAGE_SIZE; idx 4811 drivers/md/raid10.c pages[idx], idx 4833 drivers/md/raid10.c idx++; idx 134 drivers/md/raid5.c static int raid6_idx_to_slot(int idx, struct stripe_head *sh, idx 141 drivers/md/raid5.c if (idx == sh->pd_idx) idx 143 drivers/md/raid5.c if (idx == sh->qd_idx) idx 1242 drivers/media/cec/cec-adap.c unsigned int idx, idx 1305 drivers/media/cec/cec-adap.c las->log_addr[idx] = log_addr; idx 1831 drivers/media/cec/cec-adap.c unsigned int idx; idx 1841 drivers/media/cec/cec-adap.c for (idx = 0; idx < ARRAY_SIZE(las->features[0]); idx++) { idx 1842 drivers/media/cec/cec-adap.c msg->msg[msg->len++] = features[idx]; idx 1843 drivers/media/cec/cec-adap.c if ((features[idx] & CEC_OP_FEAT_EXT) == 0) { idx 499 drivers/media/cec/cec-pin.c unsigned int idx = (pin->tx_bit / 10); idx 500 drivers/media/cec/cec-pin.c u8 val = idx; idx 502 drivers/media/cec/cec-pin.c if (idx < pin->tx_msg.len) idx 503 drivers/media/cec/cec-pin.c val = pin->tx_msg.msg[idx]; idx 1072 drivers/media/cec/cec-pin.c unsigned int idx = pin->work_pin_events_rd; idx 1073 drivers/media/cec/cec-pin.c u8 v = pin->work_pin_events[idx]; idx 1078 drivers/media/cec/cec-pin.c pin->work_pin_ts[idx]); idx 1079 drivers/media/cec/cec-pin.c pin->work_pin_events_rd = (idx + 1) % CEC_NUM_PIN_EVENTS; idx 78 drivers/media/common/v4l2-tpg/v4l2-tpg-core.c #define cos(idx) sin[((idx) + 64) % sizeof(sin)] idx 1540 drivers/media/dvb-core/dvb_ca_en50221.c int idx; idx 1556 drivers/media/dvb-core/dvb_ca_en50221.c idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, -1, &fraglen); idx 1557 drivers/media/dvb-core/dvb_ca_en50221.c while (idx != -1) { idx 1558 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_read(&sl->rx_buffer, idx, 0, hdr, 2); idx 1568 drivers/media/dvb-core/dvb_ca_en50221.c idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, idx, idx 1602 drivers/media/dvb-core/dvb_ca_en50221.c size_t idx, idx2; idx 1636 drivers/media/dvb-core/dvb_ca_en50221.c idx = dvb_ringbuffer_pkt_next(&sl->rx_buffer, -1, &fraglen); idx 1639 drivers/media/dvb-core/dvb_ca_en50221.c if (idx == -1) { idx 1646 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_read(&sl->rx_buffer, idx, 0, hdr, 2); idx 1658 drivers/media/dvb-core/dvb_ca_en50221.c idx, 2, idx 1672 drivers/media/dvb-core/dvb_ca_en50221.c idx2 = dvb_ringbuffer_pkt_next(&sl->rx_buffer, idx, &fraglen); idx 1674 drivers/media/dvb-core/dvb_ca_en50221.c dvb_ringbuffer_pkt_dispose(&sl->rx_buffer, idx); idx 1675 drivers/media/dvb-core/dvb_ca_en50221.c idx = idx2; idx 252 drivers/media/dvb-core/dvb_ringbuffer.c ssize_t dvb_ringbuffer_pkt_read_user(struct dvb_ringbuffer *rbuf, size_t idx, idx 259 drivers/media/dvb-core/dvb_ringbuffer.c pktlen = rbuf->data[idx] << 8; idx 260 drivers/media/dvb-core/dvb_ringbuffer.c pktlen |= rbuf->data[(idx + 1) % rbuf->size]; idx 264 drivers/media/dvb-core/dvb_ringbuffer.c idx = (idx + DVB_RINGBUFFER_PKTHDRSIZE + offset) % rbuf->size; idx 266 drivers/media/dvb-core/dvb_ringbuffer.c split = ((idx + len) > rbuf->size) ? rbuf->size - idx : 0; idx 268 drivers/media/dvb-core/dvb_ringbuffer.c if (copy_to_user(buf, rbuf->data+idx, split)) idx 272 drivers/media/dvb-core/dvb_ringbuffer.c idx = 0; idx 274 drivers/media/dvb-core/dvb_ringbuffer.c if (copy_to_user(buf, rbuf->data+idx, todo)) idx 280 drivers/media/dvb-core/dvb_ringbuffer.c ssize_t dvb_ringbuffer_pkt_read(struct dvb_ringbuffer *rbuf, size_t idx, idx 287 drivers/media/dvb-core/dvb_ringbuffer.c pktlen = rbuf->data[idx] << 8; idx 288 drivers/media/dvb-core/dvb_ringbuffer.c pktlen |= rbuf->data[(idx + 1) % rbuf->size]; idx 292 drivers/media/dvb-core/dvb_ringbuffer.c idx = (idx + DVB_RINGBUFFER_PKTHDRSIZE + offset) % rbuf->size; idx 294 drivers/media/dvb-core/dvb_ringbuffer.c split = ((idx + len) > rbuf->size) ? rbuf->size - idx : 0; idx 296 drivers/media/dvb-core/dvb_ringbuffer.c memcpy(buf, rbuf->data+idx, split); idx 299 drivers/media/dvb-core/dvb_ringbuffer.c idx = 0; idx 301 drivers/media/dvb-core/dvb_ringbuffer.c memcpy(buf, rbuf->data+idx, todo); idx 305 drivers/media/dvb-core/dvb_ringbuffer.c void dvb_ringbuffer_pkt_dispose(struct dvb_ringbuffer *rbuf, size_t idx) idx 309 drivers/media/dvb-core/dvb_ringbuffer.c rbuf->data[(idx + 2) % rbuf->size] = PKT_DISPOSED; idx 324 drivers/media/dvb-core/dvb_ringbuffer.c ssize_t dvb_ringbuffer_pkt_next(struct dvb_ringbuffer *rbuf, size_t idx, size_t* pktlen) idx 330 drivers/media/dvb-core/dvb_ringbuffer.c if (idx == -1) { idx 331 drivers/media/dvb-core/dvb_ringbuffer.c idx = rbuf->pread; idx 333 drivers/media/dvb-core/dvb_ringbuffer.c curpktlen = rbuf->data[idx] << 8; idx 334 drivers/media/dvb-core/dvb_ringbuffer.c curpktlen |= rbuf->data[(idx + 1) % rbuf->size]; idx 335 drivers/media/dvb-core/dvb_ringbuffer.c idx = (idx + curpktlen + DVB_RINGBUFFER_PKTHDRSIZE) % rbuf->size; idx 338 drivers/media/dvb-core/dvb_ringbuffer.c consumed = (idx - rbuf->pread) % rbuf->size; idx 342 drivers/media/dvb-core/dvb_ringbuffer.c curpktlen = rbuf->data[idx] << 8; idx 343 drivers/media/dvb-core/dvb_ringbuffer.c curpktlen |= rbuf->data[(idx + 1) % rbuf->size]; idx 344 drivers/media/dvb-core/dvb_ringbuffer.c curpktstatus = rbuf->data[(idx + 2) % rbuf->size]; idx 348 drivers/media/dvb-core/dvb_ringbuffer.c return idx; idx 352 drivers/media/dvb-core/dvb_ringbuffer.c idx = (idx + curpktlen + DVB_RINGBUFFER_PKTHDRSIZE) % rbuf->size; idx 160 drivers/media/dvb-frontends/as102_fe_types.h uint8_t idx; /* index in filtering table */ idx 1559 drivers/media/dvb-frontends/cx24117.c int ret, idx; idx 1589 drivers/media/dvb-frontends/cx24117.c idx = st & 0x3f; idx 1591 drivers/media/dvb-frontends/cx24117.c if (idx > 11) idx 1592 drivers/media/dvb-frontends/cx24117.c idx += 9; idx 1594 drivers/media/dvb-frontends/cx24117.c idx += 7; idx 1597 drivers/media/dvb-frontends/cx24117.c c->modulation = cx24117_modfec_modes[idx].modulation; idx 1598 drivers/media/dvb-frontends/cx24117.c c->fec_inner = cx24117_modfec_modes[idx].fec; idx 798 drivers/media/dvb-frontends/cx24120.c int idx; idx 807 drivers/media/dvb-frontends/cx24120.c for (idx = 0; idx < ARRAY_SIZE(modfec_lookup_table); idx++) { idx 808 drivers/media/dvb-frontends/cx24120.c if (modfec_lookup_table[idx].delsys != state->dcur.delsys) idx 810 drivers/media/dvb-frontends/cx24120.c if (modfec_lookup_table[idx].val != fec) idx 816 drivers/media/dvb-frontends/cx24120.c if (idx >= ARRAY_SIZE(modfec_lookup_table)) { idx 822 drivers/media/dvb-frontends/cx24120.c c->modulation = modfec_lookup_table[idx].mod; idx 823 drivers/media/dvb-frontends/cx24120.c c->fec_inner = modfec_lookup_table[idx].fec; idx 918 drivers/media/dvb-frontends/cx24120.c int ret, idx; idx 926 drivers/media/dvb-frontends/cx24120.c for (idx = 0; idx < ARRAY_SIZE(clock_ratios_table); idx++) { idx 927 drivers/media/dvb-frontends/cx24120.c if (clock_ratios_table[idx].delsys != state->dcur.delsys) idx 929 drivers/media/dvb-frontends/cx24120.c if (clock_ratios_table[idx].mod != c->modulation) idx 931 drivers/media/dvb-frontends/cx24120.c if (clock_ratios_table[idx].fec != c->fec_inner) idx 933 drivers/media/dvb-frontends/cx24120.c if (clock_ratios_table[idx].pilot != c->pilot) idx 939 drivers/media/dvb-frontends/cx24120.c if (idx >= ARRAY_SIZE(clock_ratios_table)) { idx 956 drivers/media/dvb-frontends/cx24120.c idx, idx 957 drivers/media/dvb-frontends/cx24120.c clock_ratios_table[idx].m_rat, idx 958 drivers/media/dvb-frontends/cx24120.c clock_ratios_table[idx].n_rat, idx 959 drivers/media/dvb-frontends/cx24120.c clock_ratios_table[idx].rate); idx 966 drivers/media/dvb-frontends/cx24120.c cmd.arg[2] = (clock_ratios_table[idx].m_rat >> 16) & 0xff; idx 967 drivers/media/dvb-frontends/cx24120.c cmd.arg[3] = (clock_ratios_table[idx].m_rat >> 8) & 0xff; idx 968 drivers/media/dvb-frontends/cx24120.c cmd.arg[4] = (clock_ratios_table[idx].m_rat >> 0) & 0xff; idx 969 drivers/media/dvb-frontends/cx24120.c cmd.arg[5] = (clock_ratios_table[idx].n_rat >> 16) & 0xff; idx 970 drivers/media/dvb-frontends/cx24120.c cmd.arg[6] = (clock_ratios_table[idx].n_rat >> 8) & 0xff; idx 971 drivers/media/dvb-frontends/cx24120.c cmd.arg[7] = (clock_ratios_table[idx].n_rat >> 0) & 0xff; idx 972 drivers/media/dvb-frontends/cx24120.c cmd.arg[8] = (clock_ratios_table[idx].rate >> 8) & 0xff; idx 973 drivers/media/dvb-frontends/cx24120.c cmd.arg[9] = (clock_ratios_table[idx].rate >> 0) & 0xff; idx 978 drivers/media/dvb-frontends/cx24120.c cx24120_calculate_ber_window(state, clock_ratios_table[idx].rate); idx 1044 drivers/media/dvb-frontends/cx24120.c int idx; idx 1051 drivers/media/dvb-frontends/cx24120.c for (idx = 0; idx < ARRAY_SIZE(modfec_table); idx++) { idx 1052 drivers/media/dvb-frontends/cx24120.c if (modfec_table[idx].delsys != state->dnxt.delsys) idx 1054 drivers/media/dvb-frontends/cx24120.c if (modfec_table[idx].mod != mod) idx 1056 drivers/media/dvb-frontends/cx24120.c if (modfec_table[idx].fec != fec) idx 1061 drivers/media/dvb-frontends/cx24120.c state->dnxt.fec_val = modfec_table[idx].val; idx 627 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c u8 idx = 0; idx 629 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c idx = (addr - 0x10) / 6; idx 631 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c tnr_dmd->lna_thrs_tbl_air->thrs[idx].off_on; idx 633 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c tnr_dmd->lna_thrs_tbl_air->thrs[idx].on_off; idx 659 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c u8 idx = 0; idx 661 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c idx = (addr - 0x10) / 6; idx 663 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c tnr_dmd->lna_thrs_tbl_cable->thrs[idx].off_on; idx 665 drivers/media/dvb-frontends/cxd2880/cxd2880_tnrdmd.c tnr_dmd->lna_thrs_tbl_cable->thrs[idx].on_off; idx 608 drivers/media/i2c/cx25840/cx25840-core.c #define CX25840_VCONFIG_SET_BIT(state, opt_msk, voc, idx, bit, oneval) \ idx 613 drivers/media/i2c/cx25840/cx25840-core.c (voc)[idx] |= BIT(bit); \ idx 615 drivers/media/i2c/cx25840/cx25840-core.c (voc)[idx] &= ~BIT(bit); \ idx 42 drivers/media/i2c/max2175.c u8 idx; /* Register index */ idx 316 drivers/media/i2c/max2175.c static int max2175_read(struct max2175 *ctx, u8 idx, u8 *val) idx 321 drivers/media/i2c/max2175.c ret = regmap_read(ctx->regmap, idx, ®val); idx 323 drivers/media/i2c/max2175.c mxm_err(ctx, "read ret(%d): idx 0x%02x\n", ret, idx); idx 330 drivers/media/i2c/max2175.c static int max2175_write(struct max2175 *ctx, u8 idx, u8 val) idx 334 drivers/media/i2c/max2175.c ret = regmap_write(ctx->regmap, idx, val); idx 337 drivers/media/i2c/max2175.c ret, idx, val); idx 342 drivers/media/i2c/max2175.c static u8 max2175_read_bits(struct max2175 *ctx, u8 idx, u8 msb, u8 lsb) idx 346 drivers/media/i2c/max2175.c if (max2175_read(ctx, idx, &val)) idx 352 drivers/media/i2c/max2175.c static int max2175_write_bits(struct max2175 *ctx, u8 idx, idx 355 drivers/media/i2c/max2175.c int ret = regmap_update_bits(ctx->regmap, idx, GENMASK(msb, lsb), idx 359 drivers/media/i2c/max2175.c mxm_err(ctx, "wbits ret(%d): idx 0x%02x\n", ret, idx); idx 364 drivers/media/i2c/max2175.c static int max2175_write_bit(struct max2175 *ctx, u8 idx, u8 bit, u8 newval) idx 366 drivers/media/i2c/max2175.c return max2175_write_bits(ctx, idx, bit, bit, newval); idx 370 drivers/media/i2c/max2175.c static int max2175_poll_timeout(struct max2175 *ctx, u8 idx, u8 msb, u8 lsb, idx 375 drivers/media/i2c/max2175.c return regmap_read_poll_timeout(ctx->regmap, idx, val, idx 453 drivers/media/i2c/max2175.c max2175_write(ctx, fmeu1p2_map[i].idx, fmeu1p2_map[i].val); idx 468 drivers/media/i2c/max2175.c max2175_write(ctx, dab12_map[i].idx, dab12_map[i].val); idx 481 drivers/media/i2c/max2175.c max2175_write(ctx, fmna1p0_map[i].idx, fmna1p0_map[i].val); idx 489 drivers/media/i2c/max2175.c max2175_write(ctx, fmna2p0_map[i].idx, fmna2p0_map[i].val); idx 460 drivers/media/i2c/mt9v111.c unsigned int idx; idx 468 drivers/media/i2c/mt9v111.c for (i = 0, idx = 0; i < ARRAY_SIZE(mt9v111_frame_intervals); i++) { idx 471 drivers/media/i2c/mt9v111.c idx = i; idx 475 drivers/media/i2c/mt9v111.c fps = mt9v111_frame_intervals[idx]; idx 887 drivers/media/i2c/mt9v111.c unsigned int idx = 0; idx 918 drivers/media/i2c/mt9v111.c idx = i; idx 924 drivers/media/i2c/mt9v111.c new_fmt.width = mt9v111_frame_sizes[idx].width; idx 925 drivers/media/i2c/mt9v111.c new_fmt.height = mt9v111_frame_sizes[idx].height; idx 609 drivers/media/i2c/ov772x.c unsigned int idx; idx 614 drivers/media/i2c/ov772x.c for (i = 0, idx = 0; i < ARRAY_SIZE(ov772x_frame_intervals); i++) { idx 617 drivers/media/i2c/ov772x.c idx = i; idx 622 drivers/media/i2c/ov772x.c return ov772x_frame_intervals[idx]; idx 797 drivers/media/i2c/s5c73m3/s5c73m3-core.c enum s5c73m3_resolution_types idx) idx 804 drivers/media/i2c/s5c73m3/s5c73m3-core.c fs = s5c73m3_resolutions[idx]; idx 806 drivers/media/i2c/s5c73m3/s5c73m3-core.c for (i = 0; i < s5c73m3_resolutions_len[idx]; ++i) { idx 1220 drivers/media/i2c/s5c73m3/s5c73m3-core.c int idx; idx 1225 drivers/media/i2c/s5c73m3/s5c73m3-core.c idx = RES_ISP; idx 1229 drivers/media/i2c/s5c73m3/s5c73m3-core.c idx = RES_JPEG; idx 1232 drivers/media/i2c/s5c73m3/s5c73m3-core.c if (fse->index >= s5c73m3_resolutions_len[idx]) idx 1235 drivers/media/i2c/s5c73m3/s5c73m3-core.c fse->min_width = s5c73m3_resolutions[idx][fse->index].width; idx 1237 drivers/media/i2c/s5c73m3/s5c73m3-core.c fse->max_height = s5c73m3_resolutions[idx][fse->index].height; idx 1248 drivers/media/i2c/s5c73m3/s5c73m3-core.c int idx; idx 1287 drivers/media/i2c/s5c73m3/s5c73m3-core.c idx = RES_JPEG; idx 1289 drivers/media/i2c/s5c73m3/s5c73m3-core.c idx = RES_ISP; idx 1291 drivers/media/i2c/s5c73m3/s5c73m3-core.c if (fse->index >= s5c73m3_resolutions_len[idx]) idx 1294 drivers/media/i2c/s5c73m3/s5c73m3-core.c fse->min_width = s5c73m3_resolutions[idx][fse->index].width; idx 1296 drivers/media/i2c/s5c73m3/s5c73m3-core.c fse->max_height = s5c73m3_resolutions[idx][fse->index].height; idx 735 drivers/media/i2c/s5k6aa.c int idx = preset->index; idx 746 drivers/media/i2c/s5k6aa.c ret = s5k6aa_write(client, REG_P_MAX_OUT_RATE(idx), idx 749 drivers/media/i2c/s5k6aa.c ret = s5k6aa_write(client, REG_P_MIN_OUT_RATE(idx), idx 752 drivers/media/i2c/s5k6aa.c ret = s5k6aa_write(client, REG_P_CLK_INDEX(idx), idx 755 drivers/media/i2c/s5k6aa.c ret = s5k6aa_write(client, REG_P_FR_RATE_TYPE(idx), idx 758 drivers/media/i2c/s5k6aa.c ret = s5k6aa_write(client, REG_P_FR_RATE_Q_TYPE(idx), idx 761 drivers/media/i2c/s5k6aa.c ret = s5k6aa_write(client, REG_P_MAX_FR_TIME(idx), idx 764 drivers/media/i2c/s5k6aa.c ret = s5k6aa_write(client, REG_P_MIN_FR_TIME(idx), idx 767 drivers/media/i2c/s5k6aa.c ret = s5k6aa_new_config_sync(client, 250, idx); idx 1255 drivers/media/i2c/s5k6aa.c int idx, err = 0; idx 1267 drivers/media/i2c/s5k6aa.c idx = s5k6aa->preset->index; idx 1310 drivers/media/i2c/s5k6aa.c err = s5k6aa_write(client, REG_P_COLORTEMP(idx), ctrl->val); idx 1576 drivers/media/i2c/smiapp/smiapp-core.c int idx = -1; idx 1595 drivers/media/i2c/smiapp/smiapp-core.c idx++; idx 1597 drivers/media/i2c/smiapp/smiapp-core.c if (idx == code->index) { idx 190 drivers/media/i2c/video-i2c.c unsigned int n, idx; idx 198 drivers/media/i2c/video-i2c.c idx = data->chip->num_frame_intervals - n - 1; idx 202 drivers/media/i2c/video-i2c.c idx << MLX90640_REG_CTL1_MASK_SHIFT); idx 79 drivers/media/pci/cobalt/cobalt-alsa-pcm.c unsigned idx = 0; idx 82 drivers/media/pci/cobalt/cobalt-alsa-pcm.c unsigned offset = map[idx] * 4; idx 93 drivers/media/pci/cobalt/cobalt-alsa-pcm.c idx++; idx 325 drivers/media/pci/cobalt/cobalt-alsa-pcm.c unsigned idx = 0; idx 328 drivers/media/pci/cobalt/cobalt-alsa-pcm.c unsigned offset = map[idx] * 4; idx 341 drivers/media/pci/cobalt/cobalt-alsa-pcm.c idx++; idx 440 drivers/media/pci/cobalt/cobalt-driver.c unsigned idx = i - COBALT_AUDIO_IN_STREAM; idx 442 drivers/media/pci/cobalt/cobalt-driver.c s->dma_channel = 6 + idx; idx 444 drivers/media/pci/cobalt/cobalt-driver.c s->video_channel = idx; idx 82 drivers/media/pci/cobalt/cobalt-i2c.c cobalt_i2c_regs(struct cobalt *cobalt, unsigned idx) idx 84 drivers/media/pci/cobalt/cobalt-i2c.c switch (idx) { idx 92 drivers/media/pci/cx18/cx18-controls.c static int cx18_s_audio_sampling_freq(struct cx2341x_handler *cxhdl, u32 idx) idx 99 drivers/media/pci/cx18/cx18-controls.c if (idx < ARRAY_SIZE(freqs)) idx 100 drivers/media/pci/cx18/cx18-controls.c cx18_call_all(cx, audio, s_clock_freq, freqs[idx]); idx 248 drivers/media/pci/cx18/cx18-fileops.c int idx = cx->vbi.inserted_frame % CX18_VBI_FRAMES; idx 250 drivers/media/pci/cx18/cx18-fileops.c buf->buf = cx->vbi.sliced_mpeg_data[idx]; idx 251 drivers/media/pci/cx18/cx18-fileops.c buf->bytesused = cx->vbi.sliced_mpeg_size[idx]; idx 255 drivers/media/pci/cx18/cx18-fileops.c mdl->bytesused = cx->vbi.sliced_mpeg_size[idx]; idx 454 drivers/media/pci/cx18/cx18-fileops.c int idx = cx->vbi.inserted_frame % CX18_VBI_FRAMES; idx 456 drivers/media/pci/cx18/cx18-fileops.c cx->vbi.sliced_mpeg_size[idx] = 0; idx 95 drivers/media/pci/cx18/cx18-i2c.c int cx18_i2c_register(struct cx18 *cx, unsigned idx) idx 98 drivers/media/pci/cx18/cx18-i2c.c int bus = hw_bus[idx]; idx 100 drivers/media/pci/cx18/cx18-i2c.c const char *type = hw_devicenames[idx]; idx 101 drivers/media/pci/cx18/cx18-i2c.c u32 hw = 1 << idx; idx 121 drivers/media/pci/cx18/cx18-i2c.c return cx18_i2c_new_ir(cx, adap, hw, type, hw_addrs[idx]); idx 124 drivers/media/pci/cx18/cx18-i2c.c if (!hw_addrs[idx]) idx 128 drivers/media/pci/cx18/cx18-i2c.c sd = v4l2_i2c_new_subdev(&cx->v4l2_dev, adap, type, hw_addrs[idx], idx 10 drivers/media/pci/cx18/cx18-i2c.h int cx18_i2c_register(struct cx18 *cx, unsigned idx); idx 660 drivers/media/pci/cx18/cx18-ioctl.c struct v4l2_enc_idx *idx) idx 679 drivers/media/pci/cx18/cx18-ioctl.c e_idx = &idx->entry[idx->entries]; idx 683 drivers/media/pci/cx18/cx18-ioctl.c idx->entries < V4L2_ENC_IDX_ENTRIES) { idx 698 drivers/media/pci/cx18/cx18-ioctl.c idx->entries++; idx 699 drivers/media/pci/cx18/cx18-ioctl.c e_idx = &idx->entry[idx->entries]; idx 715 drivers/media/pci/cx18/cx18-ioctl.c struct v4l2_enc_idx *idx) idx 740 drivers/media/pci/cx18/cx18-ioctl.c mdl->readpos += _cx18_process_idx_data(mdl->curr_buf, idx); idx 743 drivers/media/pci/cx18/cx18-ioctl.c if (idx->entries >= V4L2_ENC_IDX_ENTRIES || idx 752 drivers/media/pci/cx18/cx18-ioctl.c struct v4l2_enc_idx *idx) idx 770 drivers/media/pci/cx18/cx18-ioctl.c idx->entries = 0; idx 771 drivers/media/pci/cx18/cx18-ioctl.c idx->entries_cap = tmp; idx 772 drivers/media/pci/cx18/cx18-ioctl.c memset(idx->reserved, 0, sizeof(idx->reserved)); idx 781 drivers/media/pci/cx18/cx18-ioctl.c cx18_process_idx_data(s, mdl, idx); idx 791 drivers/media/pci/cx18/cx18-ioctl.c } while (idx->entries < V4L2_ENC_IDX_ENTRIES); idx 45 drivers/media/pci/cx18/cx18-vbi.c int idx = cx->vbi.frame % CX18_VBI_FRAMES; idx 46 drivers/media/pci/cx18/cx18-vbi.c u8 *dst = &cx->vbi.sliced_mpeg_data[idx][0]; idx 89 drivers/media/pci/cx18/cx18-vbi.c cx->vbi.sliced_mpeg_size[idx] = sd + size; idx 581 drivers/media/pci/ddbridge/ddbridge-core.c u32 idx, off, stat = output->dma->stat; idx 584 drivers/media/pci/ddbridge/ddbridge-core.c idx = (stat >> 11) & 0x1f; idx 587 drivers/media/pci/ddbridge/ddbridge-core.c if (output->dma->cbuf != idx) { idx 588 drivers/media/pci/ddbridge/ddbridge-core.c if ((((output->dma->cbuf + 1) % output->dma->num) == idx) && idx 603 drivers/media/pci/ddbridge/ddbridge-core.c u32 idx, off, stat = output->dma->stat; idx 606 drivers/media/pci/ddbridge/ddbridge-core.c idx = (stat >> 11) & 0x1f; idx 611 drivers/media/pci/ddbridge/ddbridge-core.c if ((((output->dma->cbuf + 1) % output->dma->num) == idx) && idx 617 drivers/media/pci/ddbridge/ddbridge-core.c if (output->dma->cbuf == idx) { idx 656 drivers/media/pci/ddbridge/ddbridge-core.c u32 idx, off, stat = input->dma->stat; idx 659 drivers/media/pci/ddbridge/ddbridge-core.c idx = (stat >> 11) & 0x1f; idx 663 drivers/media/pci/ddbridge/ddbridge-core.c dev_err(dev->dev, "IA %d %d %08x\n", idx, off, ctrl); idx 667 drivers/media/pci/ddbridge/ddbridge-core.c if (input->dma->cbuf != idx) idx 677 drivers/media/pci/ddbridge/ddbridge-core.c u32 idx, free, stat = input->dma->stat; idx 680 drivers/media/pci/ddbridge/ddbridge-core.c idx = (stat >> 11) & 0x1f; idx 683 drivers/media/pci/ddbridge/ddbridge-core.c if (input->dma->cbuf == idx) idx 122 drivers/media/pci/ivtv/ivtv-alsa-main.c int ret, idx; idx 131 drivers/media/pci/ivtv/ivtv-alsa-main.c idx = index[itv->instance] == -1 ? SNDRV_DEFAULT_IDX1 : index[itv->instance]; idx 133 drivers/media/pci/ivtv/ivtv-alsa-main.c idx, idx 67 drivers/media/pci/ivtv/ivtv-controls.c static int ivtv_s_audio_sampling_freq(struct cx2341x_handler *cxhdl, u32 idx) idx 74 drivers/media/pci/ivtv/ivtv-controls.c if (idx < ARRAY_SIZE(freqs)) idx 75 drivers/media/pci/ivtv/ivtv-controls.c ivtv_call_all(itv, audio, s_clock_freq, freqs[idx]); idx 173 drivers/media/pci/ivtv/ivtv-fileops.c int idx = (itv->pgm_info_write_idx + i) % itv->pgm_info_num; idx 174 drivers/media/pci/ivtv/ivtv-fileops.c struct v4l2_enc_idx_entry *e = itv->pgm_info + idx; idx 175 drivers/media/pci/ivtv/ivtv-fileops.c u32 addr = itv->pgm_info_offset + 4 + idx * 24; idx 279 drivers/media/pci/ivtv/ivtv-fileops.c int idx = itv->vbi.inserted_frame % IVTV_VBI_FRAMES; idx 281 drivers/media/pci/ivtv/ivtv-fileops.c itv->vbi.sliced_mpeg_buf.buf = itv->vbi.sliced_mpeg_data[idx]; idx 282 drivers/media/pci/ivtv/ivtv-fileops.c itv->vbi.sliced_mpeg_buf.bytesused = itv->vbi.sliced_mpeg_size[idx]; idx 388 drivers/media/pci/ivtv/ivtv-fileops.c int idx = itv->vbi.inserted_frame % IVTV_VBI_FRAMES; idx 389 drivers/media/pci/ivtv/ivtv-fileops.c itv->vbi.sliced_mpeg_size[idx] = 0; idx 241 drivers/media/pci/ivtv/ivtv-i2c.c int ivtv_i2c_register(struct ivtv *itv, unsigned idx) idx 245 drivers/media/pci/ivtv/ivtv-i2c.c const char *type = hw_devicenames[idx]; idx 246 drivers/media/pci/ivtv/ivtv-i2c.c u32 hw = 1 << idx; idx 253 drivers/media/pci/ivtv/ivtv-i2c.c sd->grp_id = 1 << idx; idx 257 drivers/media/pci/ivtv/ivtv-i2c.c sd->grp_id = 1 << idx; idx 261 drivers/media/pci/ivtv/ivtv-i2c.c sd->grp_id = 1 << idx; idx 266 drivers/media/pci/ivtv/ivtv-i2c.c return ivtv_i2c_new_ir(itv, hw, type, hw_addrs[idx]); idx 269 drivers/media/pci/ivtv/ivtv-i2c.c if (!hw_addrs[idx]) idx 275 drivers/media/pci/ivtv/ivtv-i2c.c adap, type, 0, I2C_ADDRS(hw_addrs[idx])); idx 280 drivers/media/pci/ivtv/ivtv-i2c.c .addr = hw_addrs[idx], idx 290 drivers/media/pci/ivtv/ivtv-i2c.c adap, type, hw_addrs[idx], NULL); idx 293 drivers/media/pci/ivtv/ivtv-i2c.c sd->grp_id = 1 << idx; idx 13 drivers/media/pci/ivtv/ivtv-i2c.h int ivtv_i2c_register(struct ivtv *itv, unsigned idx); idx 1254 drivers/media/pci/ivtv/ivtv-ioctl.c static int ivtv_g_enc_index(struct file *file, void *fh, struct v4l2_enc_idx *idx) idx 1257 drivers/media/pci/ivtv/ivtv-ioctl.c struct v4l2_enc_idx_entry *e = idx->entry; idx 1265 drivers/media/pci/ivtv/ivtv-ioctl.c idx->entries = 0; idx 1266 drivers/media/pci/ivtv/ivtv-ioctl.c idx->entries_cap = IVTV_MAX_PGM_INDEX; idx 1272 drivers/media/pci/ivtv/ivtv-ioctl.c idx->entries++; idx 1276 drivers/media/pci/ivtv/ivtv-ioctl.c itv->pgm_info_read_idx = (itv->pgm_info_read_idx + idx->entries) % IVTV_MAX_PGM_INDEX; idx 119 drivers/media/pci/ivtv/ivtv-irq.c int idx = s->sg_pending_size; idx 222 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].dst = buf->dma_handle; idx 223 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].src = offset; idx 224 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].size = s->buf_size; idx 240 drivers/media/pci/ivtv/ivtv-irq.c idx++; idx 242 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending_size = idx; idx 348 drivers/media/pci/ivtv/ivtv-irq.c int idx = 0; idx 355 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].src = yi->blanking_dmaptr; idx 356 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].dst = offset; idx 357 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].size = 720 * 16; idx 360 drivers/media/pci/ivtv/ivtv-irq.c idx++; idx 367 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].src = buf->dma_handle; idx 368 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].dst = offset; idx 369 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].size = y_size - bytes_written; idx 371 drivers/media/pci/ivtv/ivtv-irq.c if (s->sg_pending[idx].size != buf->bytesused) { idx 372 drivers/media/pci/ivtv/ivtv-irq.c idx++; idx 373 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].src = idx 374 drivers/media/pci/ivtv/ivtv-irq.c buf->dma_handle + s->sg_pending[idx - 1].size; idx 375 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].dst = offset; idx 376 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].size = idx 377 drivers/media/pci/ivtv/ivtv-irq.c buf->bytesused - s->sg_pending[idx - 1].size; idx 378 drivers/media/pci/ivtv/ivtv-irq.c offset += s->sg_pending[idx].size; idx 382 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].src = buf->dma_handle; idx 383 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].dst = offset; idx 384 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending[idx].size = buf->bytesused; idx 391 drivers/media/pci/ivtv/ivtv-irq.c idx++; idx 393 drivers/media/pci/ivtv/ivtv-irq.c s->sg_pending_size = idx; idx 1021 drivers/media/pci/ivtv/ivtv-irq.c int idx = (i + itv->irq_rr_idx) % IVTV_MAX_STREAMS; idx 1022 drivers/media/pci/ivtv/ivtv-irq.c struct ivtv_stream *s = &itv->streams[idx]; idx 1041 drivers/media/pci/ivtv/ivtv-irq.c int idx = (i + itv->irq_rr_idx) % IVTV_MAX_STREAMS; idx 1042 drivers/media/pci/ivtv/ivtv-irq.c struct ivtv_stream *s = &itv->streams[idx]; idx 187 drivers/media/pci/ivtv/ivtv-vbi.c int idx = itv->vbi.frame % IVTV_VBI_FRAMES; idx 188 drivers/media/pci/ivtv/ivtv-vbi.c u8 *dst = &itv->vbi.sliced_mpeg_data[idx][0]; idx 231 drivers/media/pci/ivtv/ivtv-vbi.c itv->vbi.sliced_mpeg_size[idx] = sd + size; idx 1423 drivers/media/pci/meye/meye.c long idx = (long)vma->vm_private_data; idx 1424 drivers/media/pci/meye/meye.c meye.vma_use_count[idx]++; idx 1429 drivers/media/pci/meye/meye.c long idx = (long)vma->vm_private_data; idx 1430 drivers/media/pci/meye/meye.c meye.vma_use_count[idx]--; idx 16 drivers/media/pci/pt3/pt3_dma.c static u32 get_dma_base(int idx) idx 20 drivers/media/pci/pt3/pt3_dma.c i = (idx == 1 || idx == 2) ? 3 - idx : idx; idx 61 drivers/media/pci/pt3/pt3_dma.c static u8 *next_unit(struct pt3_adapter *adap, int *idx, int *ofs) idx 66 drivers/media/pci/pt3/pt3_dma.c (*idx)++; idx 67 drivers/media/pci/pt3/pt3_dma.c if (*idx == adap->num_bufs) idx 68 drivers/media/pci/pt3/pt3_dma.c *idx = 0; idx 70 drivers/media/pci/pt3/pt3_dma.c return &adap->buffer[*idx].data[*ofs]; idx 75 drivers/media/pci/pt3/pt3_dma.c int idx, ofs; idx 77 drivers/media/pci/pt3/pt3_dma.c idx = adap->buf_idx; idx 80 drivers/media/pci/pt3/pt3_dma.c if (adap->buffer[idx].data[ofs] == PT3_BUF_CANARY) idx 83 drivers/media/pci/pt3/pt3_dma.c while (*next_unit(adap, &idx, &ofs) != PT3_BUF_CANARY) { idx 93 drivers/media/pci/pt3/pt3_dma.c adap->buffer[idx].data, ofs / TS_PACKET_SZ); idx 99 drivers/media/pci/pt3/pt3_dma.c adap->buf_idx = idx; idx 107 drivers/media/pci/pt3/pt3_dma.c int idx, ofs; idx 110 drivers/media/pci/pt3/pt3_dma.c idx = 0; idx 114 drivers/media/pci/pt3/pt3_dma.c while (idx < adap->num_bufs) { idx 119 drivers/media/pci/pt3/pt3_dma.c idx++; idx 120 drivers/media/pci/pt3/pt3_dma.c p = adap->buffer[idx].data; idx 150 drivers/media/pci/pt3/pt3_dma.c int idx, ofs; idx 169 drivers/media/pci/pt3/pt3_dma.c idx = 0; idx 188 drivers/media/pci/pt3/pt3_dma.c data_addr = adap->buffer[idx].b_addr + ofs; idx 201 drivers/media/pci/pt3/pt3_dma.c idx++; idx 202 drivers/media/pci/pt3/pt3_dma.c if (idx >= adap->num_bufs) { idx 1065 drivers/media/pci/saa7134/saa7134-alsa.c unsigned int idx; idx 1070 drivers/media/pci/saa7134/saa7134-alsa.c for (idx = 0; idx < ARRAY_SIZE(snd_saa7134_volume_controls); idx++) { idx 1071 drivers/media/pci/saa7134/saa7134-alsa.c kcontrol = snd_ctl_new1(&snd_saa7134_volume_controls[idx], idx 1078 drivers/media/pci/saa7134/saa7134-alsa.c for (idx = 0; idx < ARRAY_SIZE(snd_saa7134_capture_controls); idx++) { idx 1079 drivers/media/pci/saa7134/saa7134-alsa.c kcontrol = snd_ctl_new1(&snd_saa7134_capture_controls[idx], idx 1081 drivers/media/pci/saa7134/saa7134-alsa.c addr = snd_saa7134_capture_controls[idx].private_value; idx 1247 drivers/media/pci/saa7134/saa7134-alsa.c int idx; idx 1249 drivers/media/pci/saa7134/saa7134-alsa.c for (idx = 0; idx < SNDRV_CARDS; idx++) { idx 1250 drivers/media/pci/saa7134/saa7134-alsa.c if (snd_saa7134_cards[idx]) idx 1251 drivers/media/pci/saa7134/saa7134-alsa.c snd_card_free(snd_saa7134_cards[idx]); idx 182 drivers/media/pci/saa7146/mxb.c static inline void tea6420_route(struct mxb *mxb, int idx) idx 185 drivers/media/pci/saa7146/mxb.c TEA6420_cd[idx][0].input, TEA6420_cd[idx][0].output, 0); idx 187 drivers/media/pci/saa7146/mxb.c TEA6420_cd[idx][1].input, TEA6420_cd[idx][1].output, 0); idx 189 drivers/media/pci/saa7146/mxb.c TEA6420_line[idx][0].input, TEA6420_line[idx][0].output, 0); idx 191 drivers/media/pci/saa7146/mxb.c TEA6420_line[idx][1].input, TEA6420_line[idx][1].output, 0); idx 920 drivers/media/pci/saa7164/saa7164-api.c u32 idx, next_offset; idx 941 drivers/media/pci/saa7164/saa7164-api.c for (idx = 0; idx < (len - sizeof(struct tmComResDescrHeader));) { idx 943 drivers/media/pci/saa7164/saa7164-api.c hdr = (struct tmComResDescrHeader *)(buf + idx); idx 948 drivers/media/pci/saa7164/saa7164-api.c dprintk(DBGLVL_API, "@ 0x%x =\n", idx); idx 955 drivers/media/pci/saa7164/saa7164-api.c pathhdr = (struct tmComResPathDescrHeader *)(buf + idx); idx 963 drivers/media/pci/saa7164/saa7164-api.c (struct tmComResAntTermDescrHeader *)(buf + idx); idx 1006 drivers/media/pci/saa7164/saa7164-api.c (struct tmComResDMATermDescrHeader *)(buf + idx); idx 1061 drivers/media/pci/saa7164/saa7164-api.c ((struct tmComResDMATermDescrHeader *)(buf + idx)); idx 1062 drivers/media/pci/saa7164/saa7164-api.c next_offset = idx + (vcoutputtermhdr->len); idx 1129 drivers/media/pci/saa7164/saa7164-api.c (struct tmComResTunerDescrHeader *)(buf + idx); idx 1156 drivers/media/pci/saa7164/saa7164-api.c psel = (struct tmComResSelDescrHeader *)(buf + idx); idx 1166 drivers/media/pci/saa7164/saa7164-api.c pdh = (struct tmComResProcDescrHeader *)(buf + idx); idx 1186 drivers/media/pci/saa7164/saa7164-api.c afd = (struct tmComResAFeatureDescrHeader *)(buf + idx); idx 1204 drivers/media/pci/saa7164/saa7164-api.c edh = (struct tmComResEncoderDescrHeader *)(buf + idx); idx 1227 drivers/media/pci/saa7164/saa7164-api.c exthdr = (struct tmComResExtDevDescrHeader *)(buf + idx); idx 1306 drivers/media/pci/saa7164/saa7164-api.c idx += hdr->len; idx 61 drivers/media/pci/saa7164/saa7164-buffer.c __func__, buf, buf->idx); idx 95 drivers/media/pci/saa7164/saa7164-buffer.c buf->idx = -1; idx 196 drivers/media/pci/saa7164/saa7164-buffer.c buf->idx = i; /* Note of which buffer list index position we occupy */ idx 206 drivers/media/pci/saa7164/saa7164-buffer.c buf->idx, idx 213 drivers/media/pci/saa7164/saa7164-buffer.c buf->idx); idx 177 drivers/media/pci/saa7164/saa7164-cmd.c u16 size, idx; idx 190 drivers/media/pci/saa7164/saa7164-cmd.c idx = 0; idx 198 drivers/media/pci/saa7164/saa7164-cmd.c for (idx = 0; idx < cmds; idx++) { idx 202 drivers/media/pci/saa7164/saa7164-cmd.c tmp = buf + idx * bus->m_wMaxReqSize; idx 219 drivers/media/pci/saa7164/saa7164-cmd.c if (idx != 0) idx 222 drivers/media/pci/saa7164/saa7164-cmd.c msg->size = size - idx * bus->m_wMaxReqSize; idx 224 drivers/media/pci/saa7164/saa7164-cmd.c ret = saa7164_bus_set(dev, msg, buf + idx * bus->m_wMaxReqSize); idx 266 drivers/media/pci/saa7164/saa7164-core.c if (buf->idx == bufnr) { idx 596 drivers/media/pci/saa7164/saa7164-core.c if (buf->idx == rp) { idx 297 drivers/media/pci/saa7164/saa7164.h int idx; idx 299 drivers/media/pci/smipcie/smipcie-main.c port->idx); idx 322 drivers/media/pci/smipcie/smipcie-main.c port->idx); idx 368 drivers/media/pci/smipcie/smipcie-main.c "%s, port %d, dmaused %d\n", __func__, port->idx, dmaChanUsed); idx 370 drivers/media/pci/smipcie/smipcie-main.c if (port->idx == 0) { idx 407 drivers/media/pci/smipcie/smipcie-main.c port->idx); idx 419 drivers/media/pci/smipcie/smipcie-main.c port->idx); idx 532 drivers/media/pci/smipcie/smipcie-main.c i2c = (port->idx == 0) ? &dev->i2c_bus[0] : &dev->i2c_bus[1]; idx 588 drivers/media/pci/smipcie/smipcie-main.c i2c = (port->idx == 0) ? &dev->i2c_bus[0] : &dev->i2c_bus[1]; idx 632 drivers/media/pci/smipcie/smipcie-main.c i2c = (port->idx == 0) ? &dev->i2c_bus[0] : &dev->i2c_bus[1]; idx 682 drivers/media/pci/smipcie/smipcie-main.c __func__, port->idx, port->fe_type); idx 711 drivers/media/pci/smipcie/smipcie-main.c port->idx, mac_ee + (port->idx)*8); idx 712 drivers/media/pci/smipcie/smipcie-main.c memcpy(adap->proposed_mac, mac_ee + (port->idx)*8, 6); idx 849 drivers/media/pci/smipcie/smipcie-main.c "%s, port %d\n", __func__, port->idx); idx 907 drivers/media/pci/smipcie/smipcie-main.c port->idx = index; idx 242 drivers/media/pci/smipcie/smipcie.h int idx; idx 176 drivers/media/pci/solo6x10/solo6x10-enc.c unsigned int idx, reg; idx 185 drivers/media/pci/solo6x10/solo6x10-enc.c idx = 0; idx 189 drivers/media/pci/solo6x10/solo6x10-enc.c idx = 1; idx 196 drivers/media/pci/solo6x10/solo6x10-enc.c solo_dev->jpeg_qp[idx] &= ~(3 << ch); idx 197 drivers/media/pci/solo6x10/solo6x10-enc.c solo_dev->jpeg_qp[idx] |= (qp & 3) << ch; idx 199 drivers/media/pci/solo6x10/solo6x10-enc.c solo_reg_write(solo_dev, reg, solo_dev->jpeg_qp[idx]); idx 206 drivers/media/pci/solo6x10/solo6x10-enc.c int idx; idx 215 drivers/media/pci/solo6x10/solo6x10-enc.c idx = 0; idx 218 drivers/media/pci/solo6x10/solo6x10-enc.c idx = 1; idx 222 drivers/media/pci/solo6x10/solo6x10-enc.c return (solo_dev->jpeg_qp[idx] >> ch) & 3; idx 212 drivers/media/pci/solo6x10/solo6x10-g723.c snd_pcm_uframes_t idx = solo_reg_read(solo_dev, SOLO_AUDIO_STA) & 0x1f; idx 214 drivers/media/pci/solo6x10/solo6x10-g723.c return idx * G723_FRAMES_PER_PAGE; idx 80 drivers/media/pci/solo6x10/solo6x10-v4l2.c static int solo_v4l2_ch_ext_4up(struct solo_dev *solo_dev, u8 idx, int on) idx 82 drivers/media/pci/solo6x10/solo6x10-v4l2.c u8 ch = idx * 4; idx 377 drivers/media/pci/tw686x/tw686x-video.c unsigned int idx, real_fps; idx 381 drivers/media/pci/tw686x/tw686x-video.c idx = (12 + 15 * fps) / max_fps; idx 384 drivers/media/pci/tw686x/tw686x-video.c if (!idx) idx 388 drivers/media/pci/tw686x/tw686x-video.c real_fps = tw686x_real_fps(idx, max_fps); idx 391 drivers/media/pci/tw686x/tw686x-video.c idx++; idx 393 drivers/media/pci/tw686x/tw686x-video.c idx--; idx 396 drivers/media/pci/tw686x/tw686x-video.c if (idx >= 15) idx 399 drivers/media/pci/tw686x/tw686x-video.c return idx; idx 131 drivers/media/platform/cadence/cdns-csi2rx.c unsigned int idx = find_first_zero_bit(&lanes_used, idx 133 drivers/media/platform/cadence/cdns-csi2rx.c set_bit(idx, &lanes_used); idx 87 drivers/media/platform/coda/coda-bit.c coda_write(dev, ctx->idx, CODA_REG_BIT_RUN_INDEX); idx 114 drivers/media/platform/coda/coda-bit.c unsigned int idx; idx 122 drivers/media/platform/coda/coda-bit.c idx = coda_read(dev, CODA_REG_BIT_RUN_INDEX); idx 143 drivers/media/platform/coda/coda-bit.c coda_write(dev, idx, CODA_REG_BIT_RUN_INDEX); idx 442 drivers/media/platform/coda/coda-bit.c (ctx->idx == coda_read(dev, CODA_REG_BIT_RUN_INDEX))) { idx 2407 drivers/media/platform/coda/coda-common.c int idx; idx 2415 drivers/media/platform/coda/coda-common.c idx = ida_alloc_max(&dev->ida, max, GFP_KERNEL); idx 2416 drivers/media/platform/coda/coda-common.c if (idx < 0) { idx 2417 drivers/media/platform/coda/coda-common.c ret = idx; idx 2421 drivers/media/platform/coda/coda-common.c name = kasprintf(GFP_KERNEL, "context%d", idx); idx 2444 drivers/media/platform/coda/coda-common.c ctx->idx = idx; idx 2464 drivers/media/platform/coda/coda-common.c ctx->reg_idx = idx; idx 2528 drivers/media/platform/coda/coda-common.c ida_free(&dev->ida, ctx->idx); idx 2565 drivers/media/platform/coda/coda-common.c ida_free(&dev->ida, ctx->idx); idx 256 drivers/media/platform/coda/coda.h int idx; idx 283 drivers/media/platform/coda/coda.h "%u: " fmt, (ctx)->idx, ##arg); \ idx 26 drivers/media/platform/coda/trace.h __entry->ctx = ctx->idx; idx 46 drivers/media/platform/coda/trace.h __entry->ctx = ctx->idx; idx 66 drivers/media/platform/coda/trace.h __entry->ctx = ctx->idx; idx 102 drivers/media/platform/coda/trace.h __entry->ctx = ctx->idx; idx 134 drivers/media/platform/coda/trace.h __entry->ctx = ctx->idx; idx 135 drivers/media/platform/exynos-gsc/gsc-core.h int idx; idx 501 drivers/media/platform/exynos4-is/fimc-isp.c unsigned int idx, iso; idx 507 drivers/media/platform/exynos4-is/fimc-isp.c idx = is->isp.ctrls.iso->val; idx 508 drivers/media/platform/exynos4-is/fimc-isp.c if (idx >= ARRAY_SIZE(iso_qmenu)) idx 511 drivers/media/platform/exynos4-is/fimc-isp.c iso = iso_qmenu[idx]; idx 168 drivers/media/platform/exynos4-is/media-dev.c unsigned int idx = seq[on][i]; idx 170 drivers/media/platform/exynos4-is/media-dev.c ret = __subdev_set_power(p->subdevs[idx], on); idx 179 drivers/media/platform/exynos4-is/media-dev.c unsigned int idx = seq[on][i]; idx 180 drivers/media/platform/exynos4-is/media-dev.c __subdev_set_power(p->subdevs[idx], !on); idx 331 drivers/media/platform/exynos4-is/media-dev.c unsigned int idx = seq[on][i]; idx 333 drivers/media/platform/exynos4-is/media-dev.c ret = v4l2_subdev_call(p->subdevs[idx], video, s_stream, on); idx 343 drivers/media/platform/exynos4-is/media-dev.c unsigned int idx = seq[on][i]; idx 344 drivers/media/platform/exynos4-is/media-dev.c v4l2_subdev_call(p->subdevs[idx], video, s_stream, !on); idx 286 drivers/media/platform/mtk-jpeg/mtk_jpeg_hw.c static void mtk_jpeg_dec_set_pause_mcu_idx(void __iomem *base, u32 idx) idx 288 drivers/media/platform/mtk-jpeg/mtk_jpeg_hw.c writel(idx & 0x0003FFFFFF, base + JPGDEC_REG_PAUSE_MCU_NUM); idx 71 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c unsigned int idx; idx 257 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c static void vp9_ref_cnt_fb(struct vdec_vp9_inst *inst, int *idx, idx 261 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c int ref_idx = *idx; idx 280 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c *idx = new_idx; idx 307 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c int idx; idx 312 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c for (idx = 0; idx 313 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c idx < ARRAY_SIZE(vsi->sf_ref_fb); idx 314 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c idx++) { idx 315 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (vsi->sf_ref_fb[idx].fb.base_y.va && idx 316 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c vsi->sf_ref_fb[idx].used == 0) { idx 317 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c return idx; idx 321 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c for (idx = 0; idx 322 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c idx < ARRAY_SIZE(vsi->sf_ref_fb); idx 323 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c idx++) { idx 324 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (vsi->sf_ref_fb[idx].fb.base_y.va == NULL) idx 328 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c if (idx == ARRAY_SIZE(vsi->sf_ref_fb)) { idx 333 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c mem_basy_y = &vsi->sf_ref_fb[idx].fb.base_y; idx 342 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c mem_basy_c = &vsi->sf_ref_fb[idx].fb.base_c; idx 350 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c vsi->sf_ref_fb[idx].used = 0; idx 352 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c return idx; idx 866 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c unsigned int idx = vsi->sf_frm_idx; idx 870 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c vsi->sf_frm_offset[idx]), idx 871 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c vsi->sf_frm_sz[idx]); idx 930 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c unsigned int idx = vsi->frm_refs[i].idx; idx 932 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c vsi->frm_refs[i].buf = &vsi->frm_bufs[idx].buf; idx 288 drivers/media/platform/omap3isp/isp.c unsigned int idx = clkspec->args[0]; idx 291 drivers/media/platform/omap3isp/isp.c if (idx >= ARRAY_SIZE(isp->xclks)) idx 294 drivers/media/platform/omap3isp/isp.c return isp->xclks[idx].clk; idx 1718 drivers/media/platform/pxa_camera.c unsigned int idx, idx 1725 drivers/media/platform/pxa_camera.c .index = idx, idx 1737 drivers/media/platform/pxa_camera.c "Invalid format code #%u: %d\n", idx, code.code); idx 1847 drivers/media/platform/pxa_camera.c unsigned int idx; idx 1849 drivers/media/platform/pxa_camera.c for (idx = 0; pcdev->user_formats[idx].code; idx++); idx 1850 drivers/media/platform/pxa_camera.c if (f->index >= idx) idx 983 drivers/media/platform/qcom/venus/helpers.c void venus_helper_release_buf_ref(struct venus_inst *inst, unsigned int idx) idx 988 drivers/media/platform/qcom/venus/helpers.c if (buf->vb.vb2_buf.index == idx) { idx 1019 drivers/media/platform/qcom/venus/helpers.c venus_helper_find_buf(struct venus_inst *inst, unsigned int type, u32 idx) idx 1024 drivers/media/platform/qcom/venus/helpers.c return v4l2_m2m_src_buf_remove_by_idx(m2m_ctx, idx); idx 1026 drivers/media/platform/qcom/venus/helpers.c return v4l2_m2m_dst_buf_remove_by_idx(m2m_ctx, idx); idx 16 drivers/media/platform/qcom/venus/helpers.h unsigned int type, u32 idx); idx 49 drivers/media/platform/qcom/venus/helpers.h void venus_helper_release_buf_ref(struct venus_inst *inst, unsigned int idx); idx 337 drivers/media/platform/qcom/venus/hfi_msgs.c unsigned int idx = 0; idx 350 drivers/media/platform/qcom/venus/hfi_msgs.c memcpy(&bufreq[idx], buf_req, sizeof(*bufreq)); idx 351 drivers/media/platform/qcom/venus/hfi_msgs.c idx++; idx 353 drivers/media/platform/qcom/venus/hfi_msgs.c if (idx > HFI_BUFFER_TYPE_MAX) idx 131 drivers/media/platform/rcar_drif.c #define to_rcar_drif_buf_pair(sdr, ch_num, idx) \ idx 132 drivers/media/platform/rcar_drif.c (&((sdr)->ch[!(ch_num)]->buf[(idx)])) idx 501 drivers/media/platform/rcar_drif.c static void rcar_drif_channel_complete(struct rcar_drif *ch, u32 idx) idx 505 drivers/media/platform/rcar_drif.c ch->buf[idx].status |= RCAR_DRIF_BUF_DONE; idx 514 drivers/media/platform/rcar_drif.c ch->buf[idx].status |= RCAR_DRIF_BUF_OVERFLOW; idx 526 drivers/media/platform/rcar_drif.c u32 idx, produced; idx 537 drivers/media/platform/rcar_drif.c idx = sdr->produced % RCAR_DRIF_NUM_HWBUFS; idx 538 drivers/media/platform/rcar_drif.c rcar_drif_channel_complete(ch, idx); idx 541 drivers/media/platform/rcar_drif.c buf[0] = ch->num ? to_rcar_drif_buf_pair(sdr, ch->num, idx) : idx 542 drivers/media/platform/rcar_drif.c &ch->buf[idx]; idx 543 drivers/media/platform/rcar_drif.c buf[1] = ch->num ? &ch->buf[idx] : idx 544 drivers/media/platform/rcar_drif.c to_rcar_drif_buf_pair(sdr, ch->num, idx); idx 561 drivers/media/platform/rcar_drif.c buf[0] = &ch->buf[idx]; idx 1099 drivers/media/platform/s3c-camif/camif-capture.c int s3c_camif_register_video_node(struct camif_dev *camif, int idx) idx 1101 drivers/media/platform/s3c-camif/camif-capture.c struct camif_vp *vp = &camif->vp[idx]; idx 1178 drivers/media/platform/s3c-camif/camif-capture.c void s3c_camif_unregister_video_node(struct camif_dev *camif, int idx) idx 1180 drivers/media/platform/s3c-camif/camif-capture.c struct video_device *vfd = &camif->vp[idx].vdev; idx 327 drivers/media/platform/s3c-camif/camif-core.h int s3c_camif_register_video_node(struct camif_dev *camif, int idx); idx 328 drivers/media/platform/s3c-camif/camif-core.h void s3c_camif_unregister_video_node(struct camif_dev *camif, int idx); idx 1077 drivers/media/platform/s5p-mfc/s5p_mfc.c const char *name, unsigned int idx) idx 1102 drivers/media/platform/s5p-mfc/s5p_mfc.c idx); idx 510 drivers/media/platform/sh_vou.c int i, idx = 0; idx 536 drivers/media/platform/sh_vou.c idx = i; idx 544 drivers/media/platform/sh_vou.c geo->scale_idx_h = idx; idx 561 drivers/media/platform/sh_vou.c idx = i; idx 569 drivers/media/platform/sh_vou.c geo->scale_idx_v = idx; idx 65 drivers/media/platform/ti-vpe/sc.c int idx; idx 71 drivers/media/platform/ti-vpe/sc.c idx = HS_UP_SCALE; idx 79 drivers/media/platform/ti-vpe/sc.c idx = HS_LE_16_16_SCALE; idx 84 drivers/media/platform/ti-vpe/sc.c idx = HS_LT_9_16_SCALE + sixteenths - 8; idx 88 drivers/media/platform/ti-vpe/sc.c cp = scaler_hs_coeffs[idx]; idx 114 drivers/media/platform/ti-vpe/sc.c int idx; idx 120 drivers/media/platform/ti-vpe/sc.c idx = VS_UP_SCALE; idx 122 drivers/media/platform/ti-vpe/sc.c idx = VS_1_TO_1_SCALE; idx 127 drivers/media/platform/ti-vpe/sc.c idx = VS_LT_9_16_SCALE + sixteenths - 8; idx 130 drivers/media/platform/ti-vpe/sc.c cp = scaler_vs_coeffs[idx]; idx 552 drivers/media/platform/ti-vpe/vpdma.c void *write_dtd, int drop, int idx) idx 558 drivers/media/platform/ti-vpe/vpdma.c dtd += idx; idx 226 drivers/media/platform/ti-vpe/vpdma.h void *write_dtd, int drop, int idx); idx 87 drivers/media/platform/vicodec/codec-v4l2-fwht.c const struct v4l2_fwht_pixfmt_info *v4l2_fwht_get_pixfmt(u32 idx) idx 89 drivers/media/platform/vicodec/codec-v4l2-fwht.c if (idx >= ARRAY_SIZE(v4l2_fwht_pixfmts)) idx 91 drivers/media/platform/vicodec/codec-v4l2-fwht.c return v4l2_fwht_pixfmts + idx; idx 51 drivers/media/platform/vicodec/codec-v4l2-fwht.h const struct v4l2_fwht_pixfmt_info *v4l2_fwht_get_pixfmt(u32 idx); idx 276 drivers/media/platform/vivid/vivid-cec.c unsigned int idx, idx 284 drivers/media/platform/vivid/vivid-cec.c idx); idx 10 drivers/media/platform/vivid/vivid-cec.h unsigned int idx, idx 46 drivers/media/platform/vivid/vivid-rds-gen.c unsigned idx; idx 65 drivers/media/platform/vivid/vivid-rds-gen.c idx = (grp % 22) % 4; idx 67 drivers/media/platform/vivid/vivid-rds-gen.c data[1].lsb |= vivid_get_di(rds, idx); idx 72 drivers/media/platform/vivid/vivid-rds-gen.c data[3].lsb = rds->psname[2 * idx + 1]; idx 73 drivers/media/platform/vivid/vivid-rds-gen.c data[3].msb = rds->psname[2 * idx]; idx 77 drivers/media/platform/vivid/vivid-rds-gen.c idx = ((grp - 4) % 22) % 16; idx 78 drivers/media/platform/vivid/vivid-rds-gen.c data[1].lsb |= idx; idx 80 drivers/media/platform/vivid/vivid-rds-gen.c data[2].msb = rds->radiotext[4 * idx]; idx 81 drivers/media/platform/vivid/vivid-rds-gen.c data[2].lsb = rds->radiotext[4 * idx + 1]; idx 83 drivers/media/platform/vivid/vivid-rds-gen.c data[3].msb = rds->radiotext[4 * idx + 2]; idx 84 drivers/media/platform/vivid/vivid-rds-gen.c data[3].lsb = rds->radiotext[4 * idx + 3]; idx 112 drivers/media/platform/vivid/vivid-vbi-gen.c unsigned idx; idx 114 drivers/media/platform/vivid/vivid-vbi-gen.c for (idx = 0; idx < 25; idx++) { idx 115 drivers/media/platform/vivid/vivid-vbi-gen.c const struct v4l2_sliced_vbi_data *data = vbi->data + idx; idx 561 drivers/media/platform/vsp1/vsp1_entity.c #define VSP1_ENTITY_ROUTE_RPF(idx) \ idx 562 drivers/media/platform/vsp1/vsp1_entity.c { VSP1_ENTITY_RPF, idx, VI6_DPR_RPF_ROUTE(idx), \ idx 563 drivers/media/platform/vsp1/vsp1_entity.c { 0, }, VI6_DPR_NODE_RPF(idx) } idx 565 drivers/media/platform/vsp1/vsp1_entity.c #define VSP1_ENTITY_ROUTE_UDS(idx) \ idx 566 drivers/media/platform/vsp1/vsp1_entity.c { VSP1_ENTITY_UDS, idx, VI6_DPR_UDS_ROUTE(idx), \ idx 567 drivers/media/platform/vsp1/vsp1_entity.c { VI6_DPR_NODE_UDS(idx) }, VI6_DPR_NODE_UDS(idx) } idx 569 drivers/media/platform/vsp1/vsp1_entity.c #define VSP1_ENTITY_ROUTE_UIF(idx) \ idx 570 drivers/media/platform/vsp1/vsp1_entity.c { VSP1_ENTITY_UIF, idx, VI6_DPR_UIF_ROUTE(idx), \ idx 571 drivers/media/platform/vsp1/vsp1_entity.c { VI6_DPR_NODE_UIF(idx) }, VI6_DPR_NODE_UIF(idx) } idx 573 drivers/media/platform/vsp1/vsp1_entity.c #define VSP1_ENTITY_ROUTE_WPF(idx) \ idx 574 drivers/media/platform/vsp1/vsp1_entity.c { VSP1_ENTITY_WPF, idx, 0, \ idx 575 drivers/media/platform/vsp1/vsp1_entity.c { VI6_DPR_NODE_WPF(idx) }, VI6_DPR_NODE_WPF(idx) } idx 123 drivers/media/radio/radio-si476x.c si476x_phase_diversity_idx_to_mode(enum phase_diversity_modes_idx idx) idx 133 drivers/media/radio/radio-si476x.c return idx_to_value[idx]; idx 1405 drivers/media/radio/radio-si476x.c enum si476x_ctrl_idx idx) idx 1411 drivers/media/radio/radio-si476x.c &si476x_ctrls[idx], idx 1417 drivers/media/radio/radio-si476x.c si476x_ctrls[idx].name, rval); idx 1490 drivers/media/tuners/mt2063.c u32 idx; /* index loop */ idx 1496 drivers/media/tuners/mt2063.c for (idx = 0; idx < 31; ++idx) { idx 1497 drivers/media/tuners/mt2063.c if (state->CTFiltMax[idx] >= f_in) { idx 1498 drivers/media/tuners/mt2063.c RFBand = idx; idx 178 drivers/media/tuners/tda18271-common.c static int __tda18271_write_regs(struct dvb_frontend *fe, int idx, int len, idx 188 drivers/media/tuners/tda18271-common.c BUG_ON((len == 0) || (idx + len > sizeof(buf))); idx 222 drivers/media/tuners/tda18271-common.c buf[0] = idx; idx 224 drivers/media/tuners/tda18271-common.c buf[i] = regs[idx - 1 + i]; idx 233 drivers/media/tuners/tda18271-common.c idx += max; idx 243 drivers/media/tuners/tda18271-common.c idx, max, ret); idx 248 drivers/media/tuners/tda18271-common.c int tda18271_write_regs(struct dvb_frontend *fe, int idx, int len) idx 250 drivers/media/tuners/tda18271-common.c return __tda18271_write_regs(fe, idx, len, true); idx 200 drivers/media/tuners/tda18271-priv.h extern int tda18271_write_regs(struct dvb_frontend *fe, int idx, int len); idx 118 drivers/media/usb/as102/as102_drv.c filter.idx = 0xFF; idx 124 drivers/media/usb/as102/as102_drv.c index, filter.idx, filter.pid, ret); idx 188 drivers/media/usb/as102/as10x_cmd.h uint8_t idx; idx 37 drivers/media/usb/as102/as10x_cmd_stream.c if (filter->idx < 16) idx 38 drivers/media/usb/as102/as10x_cmd_stream.c pcmd->body.add_pid_filter.req.idx = filter->idx; idx 40 drivers/media/usb/as102/as10x_cmd_stream.c pcmd->body.add_pid_filter.req.idx = 0xFF; idx 61 drivers/media/usb/as102/as10x_cmd_stream.c filter->idx = prsp->body.add_pid_filter.rsp.filter_id; idx 1934 drivers/media/usb/cx231xx/cx231xx-417.c static int cx231xx_s_audio_sampling_freq(struct cx2341x_handler *cxhdl, u32 idx) idx 1941 drivers/media/usb/cx231xx/cx231xx-417.c if (idx < ARRAY_SIZE(freqs)) idx 1942 drivers/media/usb/cx231xx/cx231xx-417.c call_all(dev, audio, s_clock_freq, freqs[idx]); idx 1555 drivers/media/usb/cx231xx/cx231xx-cards.c int i, idx; idx 1560 drivers/media/usb/cx231xx/cx231xx-cards.c idx = dev->current_pcb_config.hs_config_info[0].interface_info.video_index + 1; idx 1561 drivers/media/usb/cx231xx/cx231xx-cards.c if (idx >= dev->max_iad_interface_count) { idx 1563 drivers/media/usb/cx231xx/cx231xx-cards.c "Video PCB interface #%d doesn't exist\n", idx); idx 1567 drivers/media/usb/cx231xx/cx231xx-cards.c uif = udev->actconfig->interface[idx]; idx 1599 drivers/media/usb/cx231xx/cx231xx-cards.c idx = dev->current_pcb_config.hs_config_info[0].interface_info.vanc_index + 1; idx 1600 drivers/media/usb/cx231xx/cx231xx-cards.c if (idx >= dev->max_iad_interface_count) { idx 1602 drivers/media/usb/cx231xx/cx231xx-cards.c "VBI PCB interface #%d doesn't exist\n", idx); idx 1605 drivers/media/usb/cx231xx/cx231xx-cards.c uif = udev->actconfig->interface[idx]; idx 1643 drivers/media/usb/cx231xx/cx231xx-cards.c idx = dev->current_pcb_config.hs_config_info[0].interface_info.hanc_index + 1; idx 1644 drivers/media/usb/cx231xx/cx231xx-cards.c if (idx >= dev->max_iad_interface_count) { idx 1646 drivers/media/usb/cx231xx/cx231xx-cards.c "Sliced CC PCB interface #%d doesn't exist\n", idx); idx 1649 drivers/media/usb/cx231xx/cx231xx-cards.c uif = udev->actconfig->interface[idx]; idx 1700 drivers/media/usb/cx231xx/cx231xx-cards.c u8 idx; idx 1833 drivers/media/usb/cx231xx/cx231xx-cards.c idx = dev->current_pcb_config.hs_config_info[0].interface_info.ts1_index + 1; idx 1834 drivers/media/usb/cx231xx/cx231xx-cards.c if (idx >= dev->max_iad_interface_count) { idx 1836 drivers/media/usb/cx231xx/cx231xx-cards.c idx); idx 1840 drivers/media/usb/cx231xx/cx231xx-cards.c uif = udev->actconfig->interface[idx]; idx 329 drivers/media/usb/dvb-usb/cxusb-analog.c unsigned int idx; idx 336 drivers/media/usb/dvb-usb/cxusb-analog.c for (idx = 0; idx <= tocheck - 3; idx++) idx 337 drivers/media/usb/dvb-usb/cxusb-analog.c if (memcmp(buf + idx, CXUSB_BT656_PREAMBLE, 3) == 0) { idx 338 drivers/media/usb/dvb-usb/cxusb-analog.c bt656->pos += (1 + idx); idx 29 drivers/media/usb/dvb-usb/m920x.c static int m920x_set_filter(struct dvb_usb_device *d, int type, int idx, int pid); idx 317 drivers/media/usb/dvb-usb/m920x.c static int m920x_set_filter(struct dvb_usb_device *d, int type, int idx, int pid) idx 326 drivers/media/usb/dvb-usb/m920x.c if ((ret = m920x_write(d->udev, M9206_FILTER, pid, (type << 8) | (idx * 4) )) != 0) idx 329 drivers/media/usb/dvb-usb/m920x.c if ((ret = m920x_write(d->udev, M9206_FILTER, 0, (type << 8) | (idx * 4) )) != 0) idx 3042 drivers/media/usb/em28xx/em28xx-cards.c int i, j, idx; idx 3207 drivers/media/usb/em28xx/em28xx-cards.c idx = 0; idx 3214 drivers/media/usb/em28xx/em28xx-cards.c for (j = 0; j < idx; j++) { idx 3223 drivers/media/usb/em28xx/em28xx-cards.c dev->amux_map[idx++] = INPUT(i)->amux; idx 3225 drivers/media/usb/em28xx/em28xx-cards.c for (; idx < MAX_EM28XX_INPUT; idx++) idx 3226 drivers/media/usb/em28xx/em28xx-cards.c dev->amux_map[idx] = EM28XX_AMUX_UNUSED; idx 1719 drivers/media/usb/em28xx/em28xx-video.c unsigned int idx = dev->amux_map[index]; idx 1728 drivers/media/usb/em28xx/em28xx-video.c if (dev->has_msp34xx && idx != EM28XX_AMUX_UNUSED) idx 1729 drivers/media/usb/em28xx/em28xx-video.c idx = EM28XX_AMUX_LINE_IN; idx 1731 drivers/media/usb/em28xx/em28xx-video.c switch (idx) { idx 1796 drivers/media/usb/em28xx/em28xx-video.c int idx, i; idx 1801 drivers/media/usb/em28xx/em28xx-video.c idx = dev->amux_map[a->index]; idx 1803 drivers/media/usb/em28xx/em28xx-video.c if (idx == EM28XX_AMUX_UNUSED) idx 1806 drivers/media/usb/em28xx/em28xx-video.c dev->ctl_ainput = idx; idx 1816 drivers/media/usb/em28xx/em28xx-video.c if (idx == dev->amux_map[i]) idx 340 drivers/media/usb/go7007/go7007-driver.c int idx = y * go->width / 16 + x; idx 342 drivers/media/usb/go7007/go7007-driver.c go->modet[go->modet_map[idx]].enable = 1; idx 585 drivers/media/usb/gspca/gl860/gl860.c if (tbl[n].idx != 0xffff) idx 587 drivers/media/usb/gspca/gl860/gl860.c tbl[n].idx, 0, NULL); idx 600 drivers/media/usb/gspca/gl860/gl860.c if (tbl[n].idx != 0xffff) idx 601 drivers/media/usb/gspca/gl860/gl860.c ctrl_out(gspca_dev, 0x40, 1, tbl[n].val, tbl[n].idx, idx 617 drivers/media/usb/gspca/gl860/gl860.c ctrl_out(gspca_dev, 0x40, 3, 0x7a00, tbl[n].idx, idx 620 drivers/media/usb/gspca/gl860/gl860.c msleep(tbl[n].idx); idx 71 drivers/media/usb/gspca/gl860/gl860.h u16 idx; idx 75 drivers/media/usb/gspca/gl860/gl860.h u8 idx; idx 485 drivers/media/usb/gspca/sq930x.c u16 val, idx; idx 495 drivers/media/usb/gspca/sq930x.c idx = (cmd->val & 0xff00) | cmd->reg; idx 510 drivers/media/usb/gspca/sq930x.c val, idx, gspca_dev->usb_buf[0], buf[-1]); idx 515 drivers/media/usb/gspca/sq930x.c val, idx, idx 530 drivers/media/usb/gspca/sq930x.c u16 val, idx; idx 549 drivers/media/usb/gspca/sq930x.c idx = (cmd->bw_data << 8) | (cmd->bw_addr >> 8); idx 560 drivers/media/usb/gspca/sq930x.c val, idx, idx 564 drivers/media/usb/gspca/sq930x.c val, idx); idx 569 drivers/media/usb/gspca/sq930x.c val, idx, idx 108 drivers/media/usb/gspca/sunplus.c u16 idx; idx 318 drivers/media/usb/gspca/sunplus.c reg_w_riv(gspca_dev, data->req, data->idx, data->val); idx 338 drivers/media/usb/gspca/sunplus.c u8 req, u16 idx, u16 val) idx 340 drivers/media/usb/gspca/sunplus.c reg_w_riv(gspca_dev, req, idx, val); idx 344 drivers/media/usb/gspca/sunplus.c reg_w_riv(gspca_dev, req, idx, val); idx 372 drivers/media/usb/gspca/sunplus.c u16 idx, u16 val, u8 endcode, u8 count) idx 376 drivers/media/usb/gspca/sunplus.c reg_w_riv(gspca_dev, req, idx, val); idx 684 drivers/media/usb/gspca/t613.c int idx = 0; idx 690 drivers/media/usb/gspca/t613.c idx = 2; idx 693 drivers/media/usb/gspca/t613.c idx = 3; idx 696 drivers/media/usb/gspca/t613.c idx = 4; idx 699 drivers/media/usb/gspca/t613.c idx = 6; idx 705 drivers/media/usb/gspca/t613.c reg_w_buf(gspca_dev, effects_table[idx], idx 127 drivers/media/usb/gspca/zc3xx.c u16 idx; idx 5616 drivers/media/usb/gspca/zc3xx.c reg_w(gspca_dev, action->val, action->idx); idx 5619 drivers/media/usb/gspca/zc3xx.c reg_r(gspca_dev, action->idx); idx 5624 drivers/media/usb/gspca/zc3xx.c action->idx & 0xff, /* valL */ idx 5625 drivers/media/usb/gspca/zc3xx.c action->idx >> 8); /* valH */ idx 5629 drivers/media/usb/gspca/zc3xx.c action->idx >> 8, /* reg */ idx 5630 drivers/media/usb/gspca/zc3xx.c action->idx & 0xff, /* valL */ idx 5635 drivers/media/usb/gspca/zc3xx.c msleep(action->idx); idx 73 drivers/media/usb/hdpvr/hdpvr-core.c uint i, idx; idx 75 drivers/media/usb/hdpvr/hdpvr-core.c for (idx = 0; idx < 32; ++idx) { idx 77 drivers/media/usb/hdpvr/hdpvr-core.c if (idx & 0x3) idx 78 drivers/media/usb/hdpvr/hdpvr-core.c bytes[(idx >> 3) + 3] = bytes[(idx >> 2) & 0x3]; idx 80 drivers/media/usb/hdpvr/hdpvr-core.c switch (idx & 0x3) { idx 83 drivers/media/usb/hdpvr/hdpvr-core.c bytes[4] += bytes[(idx & 0x1) * 2] * 9 + 9; idx 87 drivers/media/usb/hdpvr/hdpvr-core.c bytes[0] += 7*idx + 4; idx 91 drivers/media/usb/hdpvr/hdpvr-core.c bytes[3 - (idx >> 3)] = bytes[idx >> 2]; idx 125 drivers/media/usb/pulse8-cec/pulse8-cec.c unsigned int idx; idx 182 drivers/media/usb/pulse8-cec/pulse8-cec.c pulse8->idx, pulse8->buf); idx 216 drivers/media/usb/pulse8-cec/pulse8-cec.c if (pulse8->idx == 0) idx 218 drivers/media/usb/pulse8-cec/pulse8-cec.c memcpy(pulse8->data, pulse8->buf, pulse8->idx); idx 219 drivers/media/usb/pulse8-cec/pulse8-cec.c pulse8->len = pulse8->idx; idx 223 drivers/media/usb/pulse8-cec/pulse8-cec.c pulse8->idx = 0; idx 227 drivers/media/usb/pulse8-cec/pulse8-cec.c pulse8->idx = 0; idx 232 drivers/media/usb/pulse8-cec/pulse8-cec.c if (pulse8->idx >= DATA_SIZE) { idx 234 drivers/media/usb/pulse8-cec/pulse8-cec.c "throwing away %d bytes of garbage\n", pulse8->idx); idx 235 drivers/media/usb/pulse8-cec/pulse8-cec.c pulse8->idx = 0; idx 237 drivers/media/usb/pulse8-cec/pulse8-cec.c pulse8->buf[pulse8->idx++] = data; idx 206 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c unsigned int idx; idx 210 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c for (idx = 0, msk = 1; val; idx++, msk <<= 1) { idx 213 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c names[idx]); idx 295 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c unsigned int idx; idx 304 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c for (idx = 0, sm = 1; msk; idx++, sm <<= 1) { idx 307 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c idStr = names[idx]; idx 360 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c unsigned int idx; idx 365 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c for (idx = 0; idx < namecnt; idx++) { idx 366 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c if (!names[idx]) continue; idx 367 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c slen = strlen(names[idx]); idx 369 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c if (memcmp(names[idx],ptr,slen)) continue; idx 370 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c *valptr = idx; idx 394 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c unsigned int idx; idx 398 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c for (idx = 0, msk = 1; valid_bits; idx++, msk <<= 1) { idx 401 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c if (!names[idx]) continue; idx 402 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c slen = strlen(names[idx]); idx 404 drivers/media/usb/pvrusb2/pvrusb2-ctrl.c if (memcmp(names[idx],ptr,slen)) continue; idx 98 drivers/media/usb/pvrusb2/pvrusb2-dvb.c unsigned int idx; idx 120 drivers/media/usb/pvrusb2/pvrusb2-dvb.c for (idx = 0; idx < PVR2_DVB_BUFFER_COUNT; idx++) { idx 121 drivers/media/usb/pvrusb2/pvrusb2-dvb.c if (!(adap->buffer_storage[idx])) continue; idx 122 drivers/media/usb/pvrusb2/pvrusb2-dvb.c kfree(adap->buffer_storage[idx]); idx 123 drivers/media/usb/pvrusb2/pvrusb2-dvb.c adap->buffer_storage[idx] = NULL; idx 132 drivers/media/usb/pvrusb2/pvrusb2-dvb.c unsigned int idx; idx 145 drivers/media/usb/pvrusb2/pvrusb2-dvb.c for (idx = 0; idx < PVR2_DVB_BUFFER_COUNT; idx++) { idx 146 drivers/media/usb/pvrusb2/pvrusb2-dvb.c adap->buffer_storage[idx] = kmalloc(PVR2_DVB_BUFFER_SIZE, idx 148 drivers/media/usb/pvrusb2/pvrusb2-dvb.c if (!(adap->buffer_storage[idx])) return -ENOMEM; idx 157 drivers/media/usb/pvrusb2/pvrusb2-dvb.c for (idx = 0; idx < PVR2_DVB_BUFFER_COUNT; idx++) { idx 158 drivers/media/usb/pvrusb2/pvrusb2-dvb.c bp = pvr2_stream_get_buffer(stream, idx); idx 160 drivers/media/usb/pvrusb2/pvrusb2-dvb.c adap->buffer_storage[idx], idx 30 drivers/media/usb/pvrusb2/pvrusb2-encoder.c unsigned int idx,addr; idx 51 drivers/media/usb/pvrusb2/pvrusb2-encoder.c for (idx = 0; idx < chunkCnt; idx++) { idx 52 drivers/media/usb/pvrusb2/pvrusb2-encoder.c addr = idx + offs; idx 56 drivers/media/usb/pvrusb2/pvrusb2-encoder.c PVR2_DECOMPOSE_LE(hdw->cmd_buffer, bAddr,data[idx]); idx 76 drivers/media/usb/pvrusb2/pvrusb2-encoder.c unsigned int idx; idx 110 drivers/media/usb/pvrusb2/pvrusb2-encoder.c for (idx = 0; idx < chunkCnt; idx++) { idx 111 drivers/media/usb/pvrusb2/pvrusb2-encoder.c data[idx] = PVR2_COMPOSE_LE(hdw->cmd_buffer,idx*4); idx 137 drivers/media/usb/pvrusb2/pvrusb2-encoder.c unsigned int idx; idx 205 drivers/media/usb/pvrusb2/pvrusb2-encoder.c for (idx = 0; idx < arg_cnt_send; idx++) { idx 206 drivers/media/usb/pvrusb2/pvrusb2-encoder.c wrData[idx+4] = argp[idx]; idx 208 drivers/media/usb/pvrusb2/pvrusb2-encoder.c for (; idx < ARRAY_SIZE(wrData) - 4; idx++) { idx 209 drivers/media/usb/pvrusb2/pvrusb2-encoder.c wrData[idx+4] = 0; idx 212 drivers/media/usb/pvrusb2/pvrusb2-encoder.c ret = pvr2_encoder_write_words(hdw,MBOX_BASE,wrData,idx); idx 243 drivers/media/usb/pvrusb2/pvrusb2-encoder.c for (idx = 4; idx < arg_cnt_send; idx++) { idx 247 drivers/media/usb/pvrusb2/pvrusb2-encoder.c idx-3,wrData[idx]); idx 280 drivers/media/usb/pvrusb2/pvrusb2-encoder.c for (idx = 0; idx < arg_cnt_recv; idx++) { idx 281 drivers/media/usb/pvrusb2/pvrusb2-encoder.c argp[idx] = rdData[idx+4]; idx 298 drivers/media/usb/pvrusb2/pvrusb2-encoder.c unsigned int idx; idx 310 drivers/media/usb/pvrusb2/pvrusb2-encoder.c for (idx = 0; idx < args; idx++) { idx 311 drivers/media/usb/pvrusb2/pvrusb2-encoder.c data[idx] = va_arg(vl, u32); idx 1357 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 1359 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < fwcount; idx++) { idx 1361 drivers/media/usb/pvrusb2/pvrusb2-hdw.c fwnames[idx], idx 1366 drivers/media/usb/pvrusb2/pvrusb2-hdw.c fwnames[idx]); idx 1367 drivers/media/usb/pvrusb2/pvrusb2-hdw.c return idx; idx 1387 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < fwcount; idx++) { idx 1390 drivers/media/usb/pvrusb2/pvrusb2-hdw.c fwnames[idx]); idx 1907 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 1908 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < ARRAY_SIZE(std_eeprom_maps); idx++) { idx 1909 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (std_eeprom_maps[idx].msk ? idx 1910 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ((std_eeprom_maps[idx].pat ^ idx 1912 drivers/media/usb/pvrusb2/pvrusb2-hdw.c std_eeprom_maps[idx].msk) : idx 1913 drivers/media/usb/pvrusb2/pvrusb2-hdw.c (std_eeprom_maps[idx].pat != idx 1916 drivers/media/usb/pvrusb2/pvrusb2-hdw.c std_eeprom_maps[idx].std); idx 1920 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->std_mask_cur = std_eeprom_maps[idx].std; idx 2063 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 2069 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < cm->cnt; idx++) { idx 2070 drivers/media/usb/pvrusb2/pvrusb2-hdw.c request_module(cm->lst[idx]); idx 2074 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < ct->cnt; idx++) { idx 2075 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (pvr2_hdw_load_subdev(hdw, &ct->lst[idx]) < 0) okFl = 0; idx 2087 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 2161 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < CTRLDEF_COUNT; idx++) { idx 2162 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 2200 drivers/media/usb/pvrusb2/pvrusb2-hdw.c idx = scnprintf(hdw->identifier, sizeof(hdw->identifier) - 1, idx 2203 drivers/media/usb/pvrusb2/pvrusb2-hdw.c idx = scnprintf(hdw->identifier, sizeof(hdw->identifier) - 1, idx 2207 drivers/media/usb/pvrusb2/pvrusb2-hdw.c idx = scnprintf(hdw->identifier, sizeof(hdw->identifier) - 1, idx 2210 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->identifier[idx] = 0; idx 2237 drivers/media/usb/pvrusb2/pvrusb2-hdw.c idx = get_default_error_tolerance(hdw); idx 2238 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (idx) { idx 2241 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->vid_stream,idx); idx 2244 drivers/media/usb/pvrusb2/pvrusb2-hdw.c PVR2_VID_ENDPOINT,idx); idx 2350 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx,cnt1,cnt2,m; idx 2433 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < hdw->control_cnt; idx++) { idx 2434 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 2437 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < 32; idx++) { idx 2438 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->std_mask_ptrs[idx] = hdw->std_mask_names[idx]; idx 2440 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < CTRLDEF_COUNT; idx++) { idx 2441 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 2442 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr->info = control_defs+idx; idx 2447 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (m) for (idx = 0; idx < (sizeof(m) << 3); idx++) { idx 2448 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (!((1UL << idx) & m)) continue; idx 2449 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->input_val = idx; idx 2458 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < MPEGDEF_COUNT; idx++) { idx 2459 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx + CTRLDEF_COUNT; idx 2460 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ciptr = &(hdw->mpeg_ctrl_info[idx].info); idx 2461 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ciptr->desc = hdw->mpeg_ctrl_info[idx].desc; idx 2462 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ciptr->name = mpeg_ids[idx].strid; idx 2463 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ciptr->v4l_id = mpeg_ids[idx].id; idx 2468 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (!idx) ciptr->clear_dirty = ctrl_cx2341x_clear_dirty; idx 2474 drivers/media/usb/pvrusb2/pvrusb2-hdw.c strscpy(hdw->mpeg_ctrl_info[idx].desc, qctrl.name, idx 2475 drivers/media/usb/pvrusb2/pvrusb2-hdw.c sizeof(hdw->mpeg_ctrl_info[idx].desc)); idx 2503 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < 32; idx++) { idx 2504 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (!(valid_std_mask & (1UL << idx))) continue; idx 2506 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->std_mask_names[idx], idx 2507 drivers/media/usb/pvrusb2/pvrusb2-hdw.c sizeof(hdw->std_mask_names[idx])-1, idx 2508 drivers/media/usb/pvrusb2/pvrusb2-hdw.c 1UL << idx); idx 2509 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->std_mask_names[idx][cnt1] = 0; idx 2564 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < PVR_NUM; idx++) { idx 2565 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (unit_pointers[idx]) continue; idx 2566 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->unit_number = idx; idx 2567 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unit_pointers[idx] = hdw; idx 2725 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx) idx 2727 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (idx >= hdw->control_cnt) return NULL; idx 2728 drivers/media/usb/pvrusb2/pvrusb2-hdw.c return hdw->controls + idx; idx 2737 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 2741 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < hdw->control_cnt; idx++) { idx 2742 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 2754 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 2758 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < hdw->control_cnt; idx++) { idx 2759 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 2773 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 2778 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < hdw->control_cnt; idx++) { idx 2779 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 2972 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 2979 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < hdw->control_cnt; idx++) { idx 2980 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 3021 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 3156 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < hdw->control_cnt; idx++) { idx 3157 drivers/media/usb/pvrusb2/pvrusb2-hdw.c cptr = hdw->controls + idx; idx 3579 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 3647 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < write_len; idx++) { idx 3648 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->ctl_write_buffer[idx] = idx 3649 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ((unsigned char *)write_data)[idx]; idx 3789 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < read_len; idx++) { idx 3790 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ((unsigned char *)read_data)[idx] = idx 3791 drivers/media/usb/pvrusb2/pvrusb2-hdw.c hdw->ctl_read_buffer[idx]; idx 3834 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx; idx 3843 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < ARRAY_SIZE(pvr2_fx2cmd_desc); idx++) { idx 3844 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (pvr2_fx2cmd_desc[idx].id == cmdcode) { idx 3848 drivers/media/usb/pvrusb2/pvrusb2-hdw.c pvr2_fx2cmd_desc[idx].desc); idx 4672 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx,ccnt; idx 4674 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < ARRAY_SIZE(control_values_input); idx++) { idx 4675 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (!((1UL << idx) & msk)) continue; idx 4680 drivers/media/usb/pvrusb2/pvrusb2-hdw.c control_values_input[idx]); idx 4851 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int bcnt,ccnt,idx; idx 4854 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; ; idx++) { idx 4855 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ccnt = pvr2_hdw_report_unlocked(hdw,idx,buf,acnt); idx 4872 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int idx, ccnt; idx 4875 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; ; idx++) { idx 4876 drivers/media/usb/pvrusb2/pvrusb2-hdw.c ccnt = pvr2_hdw_report_unlocked(hdw,idx,buf,sizeof(buf)); idx 5090 drivers/media/usb/pvrusb2/pvrusb2-hdw.c unsigned int nv,m,idx; idx 5114 drivers/media/usb/pvrusb2/pvrusb2-hdw.c for (idx = 0; idx < (sizeof(m) << 3); idx++) { idx 5115 drivers/media/usb/pvrusb2/pvrusb2-hdw.c if (!((1UL << idx) & m)) continue; idx 5116 drivers/media/usb/pvrusb2/pvrusb2-hdw.c pvr2_hdw_set_input(hdw,idx); idx 468 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c unsigned int idx,offs,cnt; idx 469 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c for (idx = 0; idx < num; idx++) { idx 470 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c cnt = msgs[idx].len; idx 472 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c idx+1,num, idx 473 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c msgs[idx].addr, idx 475 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c (msgs[idx].flags & I2C_M_RD ? idx 477 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c if ((ret > 0) || !(msgs[idx].flags & I2C_M_RD)) { idx 482 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c pr_cont("%02x", msgs[idx].buf[offs]); idx 487 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c if (idx+1 == num) { idx 593 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c unsigned int idx; idx 597 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c for (idx = 0; idx < PVR2_I2C_FUNC_CNT; idx++) { idx 598 drivers/media/usb/pvrusb2/pvrusb2-i2c-core.c hdw->i2c_func[idx] = pvr2_i2c_basic_op; idx 593 drivers/media/usb/pvrusb2/pvrusb2-io.c unsigned int idx; idx 609 drivers/media/usb/pvrusb2/pvrusb2-io.c for (idx = 0; idx < (bp->max_count) / 4; idx++) { idx 611 drivers/media/usb/pvrusb2/pvrusb2-io.c val |= idx; idx 612 drivers/media/usb/pvrusb2/pvrusb2-io.c ((unsigned int *)(bp->ptr))[idx] = val; idx 41 drivers/media/usb/pvrusb2/pvrusb2-ioread.c unsigned int idx; idx 46 drivers/media/usb/pvrusb2/pvrusb2-ioread.c for (idx = 0; idx < BUFFER_COUNT; idx++) { idx 47 drivers/media/usb/pvrusb2/pvrusb2-ioread.c cp->buffer_storage[idx] = kmalloc(BUFFER_SIZE,GFP_KERNEL); idx 48 drivers/media/usb/pvrusb2/pvrusb2-ioread.c if (!(cp->buffer_storage[idx])) break; idx 51 drivers/media/usb/pvrusb2/pvrusb2-ioread.c if (idx < BUFFER_COUNT) { idx 53 drivers/media/usb/pvrusb2/pvrusb2-ioread.c for (idx = 0; idx < BUFFER_COUNT; idx++) { idx 54 drivers/media/usb/pvrusb2/pvrusb2-ioread.c if (!(cp->buffer_storage[idx])) continue; idx 55 drivers/media/usb/pvrusb2/pvrusb2-ioread.c kfree(cp->buffer_storage[idx]); idx 64 drivers/media/usb/pvrusb2/pvrusb2-ioread.c unsigned int idx; idx 67 drivers/media/usb/pvrusb2/pvrusb2-ioread.c for (idx = 0; idx < BUFFER_COUNT; idx++) { idx 68 drivers/media/usb/pvrusb2/pvrusb2-ioread.c if (!(cp->buffer_storage[idx])) continue; idx 69 drivers/media/usb/pvrusb2/pvrusb2-ioread.c kfree(cp->buffer_storage[idx]); idx 189 drivers/media/usb/pvrusb2/pvrusb2-ioread.c unsigned int idx; idx 215 drivers/media/usb/pvrusb2/pvrusb2-ioread.c for (idx = 0; idx < BUFFER_COUNT; idx++) { idx 216 drivers/media/usb/pvrusb2/pvrusb2-ioread.c bp = pvr2_stream_get_buffer(sp,idx); idx 218 drivers/media/usb/pvrusb2/pvrusb2-ioread.c cp->buffer_storage[idx], idx 295 drivers/media/usb/pvrusb2/pvrusb2-ioread.c unsigned int idx; idx 310 drivers/media/usb/pvrusb2/pvrusb2-ioread.c for (idx = cp->c_data_offs; idx < cp->c_data_len; idx++) { idx 312 drivers/media/usb/pvrusb2/pvrusb2-ioread.c if (cp->c_data_ptr[idx] == idx 323 drivers/media/usb/pvrusb2/pvrusb2-ioread.c cp->c_data_offs += idx; idx 324 drivers/media/usb/pvrusb2/pvrusb2-ioread.c cp->sync_trashed_count += idx; idx 109 drivers/media/usb/pvrusb2/pvrusb2-std.c unsigned int idx; idx 111 drivers/media/usb/pvrusb2/pvrusb2-std.c for (idx = 0; idx < arrSize; idx++) { idx 112 drivers/media/usb/pvrusb2/pvrusb2-std.c p = arrPtr + idx; idx 269 drivers/media/usb/pvrusb2/pvrusb2-std.c unsigned int idx; idx 270 drivers/media/usb/pvrusb2/pvrusb2-std.c for (idx = 0; idx < ARRAY_SIZE(generic_standards); idx++) { idx 271 drivers/media/usb/pvrusb2/pvrusb2-std.c if (generic_standards[idx].id & id) { idx 272 drivers/media/usb/pvrusb2/pvrusb2-std.c return generic_standards + idx; idx 281 drivers/media/usb/pvrusb2/pvrusb2-std.c int idx; idx 285 drivers/media/usb/pvrusb2/pvrusb2-std.c idx = std->index; idx 287 drivers/media/usb/pvrusb2/pvrusb2-std.c std->index = idx; idx 309 drivers/media/usb/pvrusb2/pvrusb2-std.c unsigned int idx,bcnt,idx2; idx 359 drivers/media/usb/pvrusb2/pvrusb2-std.c for (idx = 0; idx < std_cnt; idx++) idx 360 drivers/media/usb/pvrusb2/pvrusb2-std.c stddefs[idx].index = idx; idx 362 drivers/media/usb/pvrusb2/pvrusb2-std.c idx = 0; idx 365 drivers/media/usb/pvrusb2/pvrusb2-std.c for (idx2 = 0; (idx2 < ARRAY_SIZE(std_mixes)) && (idx < std_cnt); idx 368 drivers/media/usb/pvrusb2/pvrusb2-std.c if (pvr2_std_fill(stddefs+idx,std_mixes[idx2])) idx++; idx 371 drivers/media/usb/pvrusb2/pvrusb2-std.c for (idmsk = 1, cmsk = id; cmsk && (idx < std_cnt); idmsk <<= 1) { idx 374 drivers/media/usb/pvrusb2/pvrusb2-std.c if (!pvr2_std_fill(stddefs+idx,idmsk)) continue; idx 375 drivers/media/usb/pvrusb2/pvrusb2-std.c idx++; idx 468 drivers/media/usb/pvrusb2/pvrusb2-sysfs.c unsigned int idx,cnt; idx 470 drivers/media/usb/pvrusb2/pvrusb2-sysfs.c for (idx = 0; idx < cnt; idx++) { idx 471 drivers/media/usb/pvrusb2/pvrusb2-sysfs.c pvr2_sysfs_add_control(sfp,idx); idx 214 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c unsigned int idx; idx 223 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c for (idx = 0; idx < fh->input_cnt; idx++) { idx 224 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c if (fh->input_map[idx] == val) { idx 225 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c *i = idx; idx 606 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c unsigned int idx; idx 611 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c for (idx = 0; idx < ctls->count; idx++) { idx 612 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c ctrl = ctls->controls + idx; idx 623 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c ctls->error_idx = idx; idx 640 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c unsigned int idx; idx 648 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c for (idx = 0; idx < ctls->count; idx++) { idx 649 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c ctrl = ctls->controls + idx; idx 654 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c ctls->error_idx = idx; idx 670 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c unsigned int idx; idx 674 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c for (idx = 0; idx < ctls->count; idx++) { idx 675 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c ctrl = ctls->controls + idx; idx 678 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c ctls->error_idx = idx; idx 957 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c unsigned int input_cnt,idx; idx 1010 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c for (idx = 0; idx < (sizeof(input_mask) << 3); idx++) { idx 1011 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c if (input_mask & (1UL << idx)) input_cnt++; idx 1025 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c for (idx = 0; idx < (sizeof(input_mask) << 3); idx++) { idx 1026 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c if (!(input_mask & (1UL << idx))) continue; idx 1027 drivers/media/usb/pvrusb2/pvrusb2-v4l2.c fhp->input_map[input_cnt++] = idx; idx 197 drivers/media/usb/s2255/s2255drv.c u32 idx; idx 240 drivers/media/usb/s2255/s2255drv.c int idx; /* channel number on device, 0-3 */ idx 993 drivers/media/usb/s2255/s2255drv.c chn_rev = G_chnmap[vc->idx]; idx 994 drivers/media/usb/s2255/s2255drv.c dprintk(dev, 3, "%s channel: %d\n", __func__, vc->idx); idx 1028 drivers/media/usb/s2255/s2255drv.c dprintk(dev, 1, "%s chn %d, result: %d\n", __func__, vc->idx, res); idx 1041 drivers/media/usb/s2255/s2255drv.c chn_rev = G_chnmap[vc->idx]; idx 1042 drivers/media/usb/s2255/s2255drv.c dprintk(dev, 4, "%s chan %d\n", __func__, vc->idx); idx 1179 drivers/media/usb/s2255/s2255drv.c strscpy(inp->name, (vc->idx < 2) ? "Composite" : "S-Video", idx 1617 drivers/media/usb/s2255/s2255drv.c (dev->pid != 0x2257 || vc->idx <= 1)) idx 1700 drivers/media/usb/s2255/s2255drv.c s32 idx = -1; idx 1706 drivers/media/usb/s2255/s2255drv.c idx = vc->cur_frame; idx 1707 drivers/media/usb/s2255/s2255drv.c frm = &vc->buffer.frame[idx]; idx 1792 drivers/media/usb/s2255/s2255drv.c idx = vc->cur_frame; idx 1793 drivers/media/usb/s2255/s2255drv.c frm = &vc->buffer.frame[idx]; idx 1812 drivers/media/usb/s2255/s2255drv.c frm, dev, dev->cc, idx); idx 1831 drivers/media/usb/s2255/s2255drv.c dev->cc, idx); idx 2124 drivers/media/usb/s2255/s2255drv.c chn_rev = G_chnmap[vc->idx]; idx 2141 drivers/media/usb/s2255/s2255drv.c dprintk(dev, 2, "start acquire exit[%d] %d\n", vc->idx, res); idx 2154 drivers/media/usb/s2255/s2255drv.c chn_rev = G_chnmap[vc->idx]; idx 2164 drivers/media/usb/s2255/s2255drv.c dprintk(dev, 4, "%s: chn %d, res %d\n", __func__, vc->idx, res); idx 2262 drivers/media/usb/s2255/s2255drv.c vc->idx = i; idx 919 drivers/media/usb/stkwebcam/stk-webcam.c struct v4l2_format *fmtd, int *idx) idx 941 drivers/media/usb/stkwebcam/stk-webcam.c if (idx) idx 942 drivers/media/usb/stkwebcam/stk-webcam.c *idx = i - 1; idx 946 drivers/media/usb/stkwebcam/stk-webcam.c if (idx) idx 947 drivers/media/usb/stkwebcam/stk-webcam.c *idx = i; idx 1007 drivers/media/usb/stkwebcam/stk-webcam.c int idx; idx 1018 drivers/media/usb/stkwebcam/stk-webcam.c ret = stk_try_fmt_vid_cap(filp, fmtd, &idx); idx 1025 drivers/media/usb/stkwebcam/stk-webcam.c dev->vsettings.mode = stk_sizes[idx].m; idx 585 drivers/media/usb/usbvision/usbvision-core.c int rest_pixel, idx, pos, extra_pos, block_len, block_type_pos, block_type_len; idx 600 drivers/media/usb/usbvision/usbvision-core.c for (idx = 0; idx < len; idx++) { idx 615 drivers/media/usb/usbvision/usbvision-core.c idx += 23; idx 617 drivers/media/usb/usbvision/usbvision-core.c integrator = decompressed[idx]; idx 619 drivers/media/usb/usbvision/usbvision-core.c idx += rest_pixel - 1; idx 641 drivers/media/usb/usbvision/usbvision-core.c integrator = decompressed[idx]; idx 654 drivers/media/usb/usbvision/usbvision-core.c decompressed[idx] = integrator; idx 661 drivers/media/usb/usbvision/usbvision-core.c return idx; idx 684 drivers/media/usb/usbvision/usbvision-core.c int idx, idx_end, strip_len, strip_ptr, startblock_pos, block_pos, block_type_pos; idx 777 drivers/media/usb/usbvision/usbvision-core.c for (idx = 0; idx < idx_end; idx++) { idx 779 drivers/media/usb/usbvision/usbvision-core.c *f++ = Y[idx]; idx 780 drivers/media/usb/usbvision/usbvision-core.c *f++ = idx & 0x01 ? U[idx / 2] : V[idx / 2]; idx 782 drivers/media/usb/usbvision/usbvision-core.c *f++ = Y[idx]; idx 783 drivers/media/usb/usbvision/usbvision-core.c if (idx & 0x01) idx 784 drivers/media/usb/usbvision/usbvision-core.c *u++ = U[idx >> 1]; idx 786 drivers/media/usb/usbvision/usbvision-core.c *v++ = V[idx >> 1]; idx 788 drivers/media/usb/usbvision/usbvision-core.c *f++ = Y[idx]; idx 789 drivers/media/usb/usbvision/usbvision-core.c if (!((idx & 0x01) | (frame->curline & 0x01))) { idx 792 drivers/media/usb/usbvision/usbvision-core.c *u++ = U[idx >> 1]; idx 793 drivers/media/usb/usbvision/usbvision-core.c *v++ = V[idx >> 1]; idx 796 drivers/media/usb/usbvision/usbvision-core.c YUV_TO_RGB_BY_THE_BOOK(Y[idx], U[idx / 2], V[idx / 2], rv, gv, bv); idx 799 drivers/media/usb/usbvision/usbvision-core.c *f++ = Y[idx]; idx 322 drivers/media/usb/usbvision/usbvision-i2c.c int idx; idx 324 drivers/media/usb/usbvision/usbvision-i2c.c for (idx = 0; idx < len; idx++) idx 325 drivers/media/usb/usbvision/usbvision-i2c.c PDEBUG(DBG_I2C, "read %x from address %x", (unsigned char)buf[idx], addr); idx 383 drivers/media/usb/usbvision/usbvision-i2c.c int idx; idx 385 drivers/media/usb/usbvision/usbvision-i2c.c for (idx = 0; idx < len; idx++) idx 386 drivers/media/usb/usbvision/usbvision-i2c.c PDEBUG(DBG_I2C, "wrote %x at address %x", (unsigned char)buf[idx], addr); idx 140 drivers/media/usb/zr364xx/zr364xx.c u32 idx; idx 559 drivers/media/usb/zr364xx/zr364xx.c s32 idx = -1; idx 565 drivers/media/usb/zr364xx/zr364xx.c idx = cam->cur_frame; idx 566 drivers/media/usb/zr364xx/zr364xx.c frm = &cam->buffer.frame[idx]; idx 635 drivers/media/usb/zr364xx/zr364xx.c _DBG("****************Buffer[%d]full*************\n", idx); idx 1500 drivers/media/v4l2-core/v4l2-ctrls.c static bool std_equal(const struct v4l2_ctrl *ctrl, u32 idx, idx 1508 drivers/media/v4l2-core/v4l2-ctrls.c idx *= ctrl->elem_size; idx 1510 drivers/media/v4l2-core/v4l2-ctrls.c return !strcmp(ptr1.p_char + idx, ptr2.p_char + idx); idx 1512 drivers/media/v4l2-core/v4l2-ctrls.c return ptr1.p_s64[idx] == ptr2.p_s64[idx]; idx 1514 drivers/media/v4l2-core/v4l2-ctrls.c return ptr1.p_u8[idx] == ptr2.p_u8[idx]; idx 1516 drivers/media/v4l2-core/v4l2-ctrls.c return ptr1.p_u16[idx] == ptr2.p_u16[idx]; idx 1518 drivers/media/v4l2-core/v4l2-ctrls.c return ptr1.p_u32[idx] == ptr2.p_u32[idx]; idx 1521 drivers/media/v4l2-core/v4l2-ctrls.c return ptr1.p_s32[idx] == ptr2.p_s32[idx]; idx 1522 drivers/media/v4l2-core/v4l2-ctrls.c idx *= ctrl->elem_size; idx 1523 drivers/media/v4l2-core/v4l2-ctrls.c return !memcmp(ptr1.p + idx, ptr2.p + idx, ctrl->elem_size); idx 1527 drivers/media/v4l2-core/v4l2-ctrls.c static void std_init_compound(const struct v4l2_ctrl *ctrl, u32 idx, idx 1531 drivers/media/v4l2-core/v4l2-ctrls.c void *p = ptr.p + idx * ctrl->elem_size; idx 1553 drivers/media/v4l2-core/v4l2-ctrls.c static void std_init(const struct v4l2_ctrl *ctrl, u32 idx, idx 1558 drivers/media/v4l2-core/v4l2-ctrls.c idx *= ctrl->elem_size; idx 1559 drivers/media/v4l2-core/v4l2-ctrls.c memset(ptr.p_char + idx, ' ', ctrl->minimum); idx 1560 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_char[idx + ctrl->minimum] = '\0'; idx 1563 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_s64[idx] = ctrl->default_value; idx 1570 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_s32[idx] = ctrl->default_value; idx 1574 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_s32[idx] = 0; idx 1577 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_u8[idx] = ctrl->default_value; idx 1580 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_u16[idx] = ctrl->default_value; idx 1583 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_u32[idx] = ctrl->default_value; idx 1586 drivers/media/v4l2-core/v4l2-ctrls.c std_init_compound(ctrl, idx, ptr); idx 1670 drivers/media/v4l2-core/v4l2-ctrls.c static int std_validate_compound(const struct v4l2_ctrl *ctrl, u32 idx, idx 1675 drivers/media/v4l2-core/v4l2-ctrls.c void *p = ptr.p + idx * ctrl->elem_size; idx 1758 drivers/media/v4l2-core/v4l2-ctrls.c static int std_validate(const struct v4l2_ctrl *ctrl, u32 idx, idx 1767 drivers/media/v4l2-core/v4l2-ctrls.c return ROUND_TO_RANGE(ptr.p_s32[idx], u32, ctrl); idx 1773 drivers/media/v4l2-core/v4l2-ctrls.c val = ptr.p_s64[idx]; idx 1781 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_s64[idx] = ctrl->minimum + offset * ctrl->step; idx 1784 drivers/media/v4l2-core/v4l2-ctrls.c return ROUND_TO_RANGE(ptr.p_u8[idx], u8, ctrl); idx 1786 drivers/media/v4l2-core/v4l2-ctrls.c return ROUND_TO_RANGE(ptr.p_u16[idx], u16, ctrl); idx 1788 drivers/media/v4l2-core/v4l2-ctrls.c return ROUND_TO_RANGE(ptr.p_u32[idx], u32, ctrl); idx 1791 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_s32[idx] = !!ptr.p_s32[idx]; idx 1796 drivers/media/v4l2-core/v4l2-ctrls.c if (ptr.p_s32[idx] < ctrl->minimum || ptr.p_s32[idx] > ctrl->maximum) idx 1798 drivers/media/v4l2-core/v4l2-ctrls.c if (ctrl->menu_skip_mask & (1ULL << ptr.p_s32[idx])) idx 1801 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->qmenu[ptr.p_s32[idx]][0] == '\0') idx 1806 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_s32[idx] &= ctrl->maximum; idx 1811 drivers/media/v4l2-core/v4l2-ctrls.c ptr.p_s32[idx] = 0; idx 1815 drivers/media/v4l2-core/v4l2-ctrls.c idx *= ctrl->elem_size; idx 1816 drivers/media/v4l2-core/v4l2-ctrls.c len = strlen(ptr.p_char + idx); idx 1824 drivers/media/v4l2-core/v4l2-ctrls.c return std_validate_compound(ctrl, idx, ptr); idx 1889 drivers/media/v4l2-core/v4l2-ctrls.c int idx; idx 1891 drivers/media/v4l2-core/v4l2-ctrls.c for (idx = 0; idx < ctrl->elems; idx++) idx 1892 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->type_ops->init(ctrl, idx, ctrl->p_new); idx 1907 drivers/media/v4l2-core/v4l2-ctrls.c unsigned idx; idx 1912 drivers/media/v4l2-core/v4l2-ctrls.c for (idx = c->size / ctrl->elem_size; idx < ctrl->elems; idx++) idx 1913 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->type_ops->init(ctrl, idx, ptr); idx 2030 drivers/media/v4l2-core/v4l2-ctrls.c unsigned idx; idx 2052 drivers/media/v4l2-core/v4l2-ctrls.c for (idx = 0; !ctrl_changed && idx < ctrl->elems; idx++) idx 2053 drivers/media/v4l2-core/v4l2-ctrls.c ctrl_changed = !ctrl->type_ops->equal(ctrl, idx, idx 2103 drivers/media/v4l2-core/v4l2-ctrls.c unsigned idx; idx 2106 drivers/media/v4l2-core/v4l2-ctrls.c for (idx = 0; !err && idx < ctrl->elems; idx++) idx 2107 drivers/media/v4l2-core/v4l2-ctrls.c err = ctrl->type_ops->validate(ctrl, idx, p_new); idx 2365 drivers/media/v4l2-core/v4l2-ctrls.c unsigned idx; idx 2506 drivers/media/v4l2-core/v4l2-ctrls.c for (idx = 0; idx < elems; idx++) { idx 2507 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->type_ops->init(ctrl, idx, ctrl->p_cur); idx 2508 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->type_ops->init(ctrl, idx, ctrl->p_new); idx 3483 drivers/media/v4l2-core/v4l2-ctrls.c u32 idx = i; idx 3486 drivers/media/v4l2-core/v4l2-ctrls.c if (helpers[idx].ref->req) idx 3487 drivers/media/v4l2-core/v4l2-ctrls.c ret = req_to_user(cs->controls + idx, idx 3488 drivers/media/v4l2-core/v4l2-ctrls.c helpers[idx].ref->req); idx 3490 drivers/media/v4l2-core/v4l2-ctrls.c ret = ctrl_to_user(cs->controls + idx, idx 3491 drivers/media/v4l2-core/v4l2-ctrls.c helpers[idx].ref->ctrl); idx 3492 drivers/media/v4l2-core/v4l2-ctrls.c idx = helpers[idx].next; idx 3493 drivers/media/v4l2-core/v4l2-ctrls.c } while (!ret && idx); idx 3828 drivers/media/v4l2-core/v4l2-ctrls.c u32 idx = i; idx 3852 drivers/media/v4l2-core/v4l2-ctrls.c u32 tmp_idx = idx; idx 3870 drivers/media/v4l2-core/v4l2-ctrls.c struct v4l2_ctrl *ctrl = helpers[idx].ref->ctrl; idx 3872 drivers/media/v4l2-core/v4l2-ctrls.c ret = user_to_new(cs->controls + idx, ctrl); idx 3875 drivers/media/v4l2-core/v4l2-ctrls.c idx = helpers[idx].next; idx 3876 drivers/media/v4l2-core/v4l2-ctrls.c } while (!ret && idx); idx 3892 drivers/media/v4l2-core/v4l2-ctrls.c idx = i; idx 3894 drivers/media/v4l2-core/v4l2-ctrls.c ret = new_to_user(cs->controls + idx, idx 3895 drivers/media/v4l2-core/v4l2-ctrls.c helpers[idx].ref->ctrl); idx 3896 drivers/media/v4l2-core/v4l2-ctrls.c idx = helpers[idx].next; idx 3897 drivers/media/v4l2-core/v4l2-ctrls.c } while (!ret && idx); idx 114 drivers/media/v4l2-core/v4l2-dev.c int idx = (vfl_type > VFL_TYPE_RADIO) ? VFL_TYPE_MAX - 1 : vfl_type; idx 116 drivers/media/v4l2-core/v4l2-dev.c return devnode_nums[idx]; idx 173 drivers/media/v4l2-core/v4l2-dv-timings.c u32 i, idx; idx 176 drivers/media/v4l2-core/v4l2-dv-timings.c for (i = idx = 0; v4l2_dv_timings_presets[i].bt.width; i++) { idx 179 drivers/media/v4l2-core/v4l2-dv-timings.c idx++ == t->index) { idx 21 drivers/media/v4l2-core/v4l2-event.c static unsigned sev_pos(const struct v4l2_subscribed_event *sev, unsigned idx) idx 23 drivers/media/v4l2-core/v4l2-event.c idx += sev->first; idx 24 drivers/media/v4l2-core/v4l2-event.c return idx >= sev->elems ? idx - sev->elems : idx; idx 2435 drivers/media/v4l2-core/v4l2-ioctl.c int idx = 0; idx 2443 drivers/media/v4l2-core/v4l2-ioctl.c if (p->match.addr == idx++) idx 2463 drivers/media/v4l2-core/v4l2-ioctl.c int idx = 0; idx 2471 drivers/media/v4l2-core/v4l2-ioctl.c if (p->match.addr == idx++) idx 2491 drivers/media/v4l2-core/v4l2-ioctl.c int idx = 0; idx 2510 drivers/media/v4l2-core/v4l2-ioctl.c if (p->match.addr != idx++) idx 200 drivers/media/v4l2-core/v4l2-mem2mem.c v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx) idx 209 drivers/media/v4l2-core/v4l2-mem2mem.c if (b->vb.vb2_buf.index == idx) { idx 255 drivers/memory/omap-gpmc.c static void gpmc_write_reg(int idx, u32 val) idx 257 drivers/memory/omap-gpmc.c writel_relaxed(val, gpmc_base + idx); idx 260 drivers/memory/omap-gpmc.c static u32 gpmc_read_reg(int idx) idx 262 drivers/memory/omap-gpmc.c return readl_relaxed(gpmc_base + idx); idx 265 drivers/memory/omap-gpmc.c void gpmc_cs_write_reg(int cs, int idx, u32 val) idx 269 drivers/memory/omap-gpmc.c reg_addr = gpmc_base + GPMC_CS0_OFFSET + (cs * GPMC_CS_SIZE) + idx; idx 273 drivers/memory/omap-gpmc.c static u32 gpmc_cs_read_reg(int cs, int idx) idx 277 drivers/memory/omap-gpmc.c reg_addr = gpmc_base + GPMC_CS0_OFFSET + (cs * GPMC_CS_SIZE) + idx; idx 7748 drivers/message/fusion/mptbase.c int idx; idx 7750 drivers/message/fusion/mptbase.c idx = ioc->eventContext % MPTCTL_EVENT_LOG_SIZE; idx 7752 drivers/message/fusion/mptbase.c ioc->events[idx].event = event; idx 7753 drivers/message/fusion/mptbase.c ioc->events[idx].eventContext = ioc->eventContext; idx 7757 drivers/message/fusion/mptbase.c ioc->events[idx].data[ii] = le32_to_cpu(pEventReply->Data[ii]); idx 7759 drivers/message/fusion/mptbase.c ioc->events[idx].data[ii] = 0; idx 841 drivers/message/fusion/mptbase.h #define MPT_INDEX_2_MFPTR(ioc,idx) \ idx 842 drivers/message/fusion/mptbase.h (MPT_FRAME_HDR*)( (u8*)(ioc)->req_frames + (ioc)->req_sz * (idx) ) idx 847 drivers/message/fusion/mptbase.h #define MPT_INDEX_2_RFPTR(ioc,idx) \ idx 848 drivers/message/fusion/mptbase.h (MPT_FRAME_HDR*)( (u8*)(ioc)->reply_frames + (ioc)->req_sz * (idx) ) idx 2431 drivers/message/fusion/mptscsih.c int idx; idx 2433 drivers/message/fusion/mptscsih.c idx = ioc->eventContext % MPTCTL_EVENT_LOG_SIZE; idx 2434 drivers/message/fusion/mptscsih.c ioc->events[idx].event = MPI_EVENT_SCSI_DEVICE_STATUS_CHANGE; idx 2435 drivers/message/fusion/mptscsih.c ioc->events[idx].eventContext = ioc->eventContext; idx 2437 drivers/message/fusion/mptscsih.c ioc->events[idx].data[0] = (pReq->LUN[1] << 24) | idx 2441 drivers/message/fusion/mptscsih.c ioc->events[idx].data[1] = (sense_data[13] << 8) | sense_data[12]; idx 215 drivers/mfd/intel_quark_i2c_gpio.c pdata->properties->idx = 0; idx 54 drivers/mfd/max8997-irq.c #define DECLARE_IRQ(idx, _group, _mask) \ idx 55 drivers/mfd/max8997-irq.c [(idx)] = { .group = (_group), .mask = (_mask) } idx 85 drivers/mfd/pcf50633-core.c int n, n1, idx = 0; idx 98 drivers/mfd/pcf50633-core.c if (n == address_no_read[idx]) { idx 99 drivers/mfd/pcf50633-core.c idx++; idx 53 drivers/mfd/pcf50633-irq.c int idx; idx 55 drivers/mfd/pcf50633-irq.c idx = irq >> 3; idx 56 drivers/mfd/pcf50633-irq.c reg = PCF50633_REG_INT1M + idx; idx 64 drivers/mfd/pcf50633-irq.c pcf->mask_regs[idx] |= bit; idx 66 drivers/mfd/pcf50633-irq.c pcf->mask_regs[idx] &= ~bit; idx 1155 drivers/mfd/sm501.c lookup->table[0].idx = 0; idx 1163 drivers/mfd/sm501.c lookup->table[1].idx = 1; idx 57 drivers/misc/cardreader/rtsx_pcr.h static inline u8 map_sd_drive(int idx) idx 66 drivers/misc/cardreader/rtsx_pcr.h return sd_drive[idx]; idx 520 drivers/misc/cxl/guest.c u32 pid, idx; idx 575 drivers/misc/cxl/guest.c idx = ctx->irqs.offset[r] + i - adapter->guest->irq_base_offset; idx 576 drivers/misc/cxl/guest.c elem->applicationVirtualIsnBitmap[idx / 8] |= 0x80 >> (idx % 8); idx 284 drivers/misc/cxl/trace.h TP_PROTO(struct cxl_context *ctx, unsigned int idx, u64 e, u64 v), idx 286 drivers/misc/cxl/trace.h TP_ARGS(ctx, idx, e, v), idx 292 drivers/misc/cxl/trace.h __field(unsigned int, idx) idx 301 drivers/misc/cxl/trace.h __entry->idx = idx; idx 310 drivers/misc/cxl/trace.h __entry->idx, idx 294 drivers/misc/eeprom/idt_89hpesx.c int idx; idx 297 drivers/misc/eeprom/idt_89hpesx.c for (idx = 0; idx < seq->bytecnt; idx++) { idx 300 drivers/misc/eeprom/idt_89hpesx.c if (idx == 0) idx 302 drivers/misc/eeprom/idt_89hpesx.c if (idx == seq->bytecnt - 1) idx 307 drivers/misc/eeprom/idt_89hpesx.c seq->data[idx]); idx 326 drivers/misc/eeprom/idt_89hpesx.c int idx; idx 329 drivers/misc/eeprom/idt_89hpesx.c for (idx = 0; idx < seq->bytecnt; idx++) { idx 332 drivers/misc/eeprom/idt_89hpesx.c if (idx == 0) idx 334 drivers/misc/eeprom/idt_89hpesx.c if (idx == seq->bytecnt - 1) idx 342 drivers/misc/eeprom/idt_89hpesx.c seq->data[idx] = (u8)sts; idx 359 drivers/misc/eeprom/idt_89hpesx.c int idx, evencnt; idx 365 drivers/misc/eeprom/idt_89hpesx.c for (idx = 0; idx < evencnt; idx += 2) { idx 368 drivers/misc/eeprom/idt_89hpesx.c if (idx == 0) idx 370 drivers/misc/eeprom/idt_89hpesx.c if (idx == evencnt - 2) idx 375 drivers/misc/eeprom/idt_89hpesx.c *(u16 *)&seq->data[idx]); idx 384 drivers/misc/eeprom/idt_89hpesx.c if (idx == 0) idx 389 drivers/misc/eeprom/idt_89hpesx.c seq->data[idx]); idx 408 drivers/misc/eeprom/idt_89hpesx.c int idx, evencnt; idx 414 drivers/misc/eeprom/idt_89hpesx.c for (idx = 0; idx < evencnt; idx += 2) { idx 417 drivers/misc/eeprom/idt_89hpesx.c if (idx == 0) idx 419 drivers/misc/eeprom/idt_89hpesx.c if (idx == evencnt - 2) idx 427 drivers/misc/eeprom/idt_89hpesx.c *(u16 *)&seq->data[idx] = (u16)sts; idx 434 drivers/misc/eeprom/idt_89hpesx.c if (idx == 0) idx 442 drivers/misc/eeprom/idt_89hpesx.c seq->data[idx] = (u8)sts; idx 654 drivers/misc/eeprom/idt_89hpesx.c u16 idx; idx 661 drivers/misc/eeprom/idt_89hpesx.c for (idx = 0; idx < len; idx++, memaddr++) { idx 670 drivers/misc/eeprom/idt_89hpesx.c eeseq.data = data[idx]; idx 675 drivers/misc/eeprom/idt_89hpesx.c memaddr, data[idx]); idx 683 drivers/misc/eeprom/idt_89hpesx.c eeseq.data = ~data[idx]; idx 689 drivers/misc/eeprom/idt_89hpesx.c if (eeseq.data != data[idx]) { idx 691 drivers/misc/eeprom/idt_89hpesx.c eeseq.data, data[idx]); idx 717 drivers/misc/eeprom/idt_89hpesx.c u16 idx; idx 720 drivers/misc/eeprom/idt_89hpesx.c for (idx = 0; idx < len; idx++, memaddr++) { idx 725 drivers/misc/eeprom/idt_89hpesx.c ret = idt_eeprom_read_byte(pdev, memaddr, &buf[idx]); idx 77 drivers/misc/genwqe/card_base.h u32 idx; idx 39 drivers/misc/genwqe/card_debugfs.c regs[i].addr, regs[i].idx, v_hi, v_lo); idx 769 drivers/misc/genwqe/card_utils.c unsigned int *i, unsigned int m, u32 addr, u32 idx, idx 776 drivers/misc/genwqe/card_utils.c r[*i].idx = idx; idx 791 drivers/misc/genwqe/card_utils.c unsigned int i, j, idx = 0; idx 797 drivers/misc/genwqe/card_utils.c set_reg(cd, regs, &idx, max_regs, IO_SLC_CFGREG_GFIR, gfir); idx 801 drivers/misc/genwqe/card_utils.c set_reg(cd, regs, &idx, max_regs, IO_SLU_UNITCFG, sluid); idx 805 drivers/misc/genwqe/card_utils.c set_reg(cd, regs, &idx, max_regs, IO_APP_UNITCFG, appid); idx 813 drivers/misc/genwqe/card_utils.c set_reg(cd, regs, &idx, max_regs, ufir_addr, ufir); idx 818 drivers/misc/genwqe/card_utils.c set_reg(cd, regs, &idx, max_regs, ufec_addr, ufec); idx 827 drivers/misc/genwqe/card_utils.c set_reg(cd, regs, &idx, max_regs, sfir_addr, sfir); idx 831 drivers/misc/genwqe/card_utils.c set_reg(cd, regs, &idx, max_regs, sfec_addr, sfec); idx 836 drivers/misc/genwqe/card_utils.c for (i = idx; i < max_regs; i++) { idx 840 drivers/misc/genwqe/card_utils.c return idx; idx 903 drivers/misc/genwqe/card_utils.c unsigned int idx = 0; idx 925 drivers/misc/genwqe/card_utils.c set_reg_idx(cd, regs, &idx, max_regs, idx 932 drivers/misc/genwqe/card_utils.c set_reg_idx(cd, regs, &idx, max_regs, idx 973 drivers/misc/genwqe/card_utils.c set_reg_idx(cd, regs, &idx, max_regs, addr, idx 172 drivers/misc/kgdbts.c int idx; idx 358 drivers/misc/kgdbts.c ts.idx -= 2; idx 363 drivers/misc/kgdbts.c ts.idx -= 4; idx 392 drivers/misc/kgdbts.c ts.idx++; idx 397 drivers/misc/kgdbts.c ts.idx -= go_back; idx 399 drivers/misc/kgdbts.c fill_get_buf(ts.tst[ts.idx].get); idx 423 drivers/misc/kgdbts.c ts.idx--; idx 433 drivers/misc/kgdbts.c ts.idx = -1; idx 521 drivers/misc/kgdbts.c ts.idx--; idx 707 drivers/misc/kgdbts.c v2printk("get%i: %s\n", ts.idx, get_buf); idx 714 drivers/misc/kgdbts.c if (ts.tst[ts.idx].put_handler) idx 715 drivers/misc/kgdbts.c return ts.tst[ts.idx].put_handler(put_str, idx 716 drivers/misc/kgdbts.c ts.tst[ts.idx].put); idx 718 drivers/misc/kgdbts.c chk_str = ts.tst[ts.idx].put; idx 755 drivers/misc/kgdbts.c if (ts.tst[ts.idx].get_handler) idx 756 drivers/misc/kgdbts.c ts.tst[ts.idx].get_handler(ts.tst[ts.idx].get); idx 758 drivers/misc/kgdbts.c fill_get_buf(ts.tst[ts.idx].get); idx 763 drivers/misc/kgdbts.c ts.name, ts.idx); idx 775 drivers/misc/kgdbts.c if (ts.tst[ts.idx].get[0] == '\0' && ts.tst[ts.idx].put[0] == '\0' && idx 776 drivers/misc/kgdbts.c !ts.tst[ts.idx].get_handler) { idx 778 drivers/misc/kgdbts.c " '%s' line %i\n", ts.name, ts.idx); idx 784 drivers/misc/kgdbts.c " '%s' line %i\n", ts.name, ts.idx); idx 799 drivers/misc/kgdbts.c " '%s' line %i\n", ts.name, ts.idx); idx 804 drivers/misc/kgdbts.c v2printk("put%i: %s\n", ts.idx, put_buf); idx 809 drivers/misc/kgdbts.c ts.name, ts.idx, ts.tst[ts.idx].put, put_buf); idx 811 drivers/misc/kgdbts.c ts.idx++; idx 1451 drivers/misc/mei/hw-me.c const struct mei_cfg *mei_me_get_cfg(kernel_ulong_t idx) idx 1455 drivers/misc/mei/hw-me.c if (idx >= MEI_ME_NUM_CFG) idx 1458 drivers/misc/mei/hw-me.c return mei_cfg_list[idx]; idx 92 drivers/misc/mei/hw-me.h const struct mei_cfg *mei_me_get_cfg(kernel_ulong_t idx); idx 486 drivers/misc/mei/hw-txe.c unsigned long idx, u32 value) idx 491 drivers/misc/mei/hw-txe.c (idx * sizeof(u32)), value); idx 504 drivers/misc/mei/hw-txe.c unsigned long idx) idx 509 drivers/misc/mei/hw-txe.c BRIDGE_IPC_OUTPUT_PAYLOAD_REG + (idx * sizeof(u32))); idx 124 drivers/misc/mic/card/mic_device.h u32 mic_read_spad(struct mic_device *mdev, unsigned int idx); idx 36 drivers/misc/mic/card/mic_x100.c u32 mic_read_spad(struct mic_device *mdev, unsigned int idx) idx 40 drivers/misc/mic/card/mic_x100.c MIC_X100_SBOX_SPAD0 + idx * 4); idx 112 drivers/misc/mic/host/mic_device.h u32 (*read_spad)(struct mic_device *mdev, unsigned int idx); idx 113 drivers/misc/mic/host/mic_device.h void (*write_spad)(struct mic_device *mdev, unsigned int idx, u32 val); idx 108 drivers/misc/mic/host/mic_intr.c u8 idx, irq_handler_t handler, irq_handler_t thread_fn, idx 131 drivers/misc/mic/host/mic_intr.c list_add_tail(&intr_cb->list, &mdev->irq_info.cb_list[idx]); idx 150 drivers/misc/mic/host/mic_intr.c static u8 mic_unregister_intr_callback(struct mic_device *mdev, u32 idx) idx 162 drivers/misc/mic/host/mic_intr.c if (intr_cb->cb_id == idx) { idx 121 drivers/misc/mic/host/mic_intr.h int idx, int intr_src, bool set); idx 123 drivers/misc/mic/host/mic_intr.h int idx); idx 32 drivers/misc/mic/host/mic_x100.c mic_x100_write_spad(struct mic_device *mdev, unsigned int idx, u32 val) idx 35 drivers/misc/mic/host/mic_x100.c val, idx); idx 38 drivers/misc/mic/host/mic_x100.c MIC_X100_SBOX_SPAD0 + idx * 4); idx 51 drivers/misc/mic/host/mic_x100.c mic_x100_read_spad(struct mic_device *mdev, unsigned int idx) idx 55 drivers/misc/mic/host/mic_x100.c MIC_X100_SBOX_SPAD0 + idx * 4); idx 58 drivers/misc/mic/host/mic_x100.c "Reading 0x%x from scratch pad index %d\n", val, idx); idx 218 drivers/misc/mic/host/mic_x100.c mic_x100_read_msi_to_src_map(struct mic_device *mdev, int idx) idx 222 drivers/misc/mic/host/mic_x100.c MIC_X100_SBOX_MXAR0 + idx * 4); idx 237 drivers/misc/mic/host/mic_x100.c int idx, int offset, bool set) idx 242 drivers/misc/mic/host/mic_x100.c MIC_X100_SBOX_MXAR0 + idx * 4; idx 142 drivers/misc/mic/vop/vop_debugfs.c avail->idx) & (num - 1)); idx 145 drivers/misc/mic/vop/vop_debugfs.c vringh16_to_cpu(vrh, avail->idx)); idx 152 drivers/misc/mic/vop/vop_debugfs.c vringh16_to_cpu(vrh, used->idx) & (num - 1)); idx 155 drivers/misc/mic/vop/vop_debugfs.c vringh16_to_cpu(vrh, used->idx)); idx 265 drivers/misc/mic/vop/vop_main.c int idx = 0; idx 270 drivers/misc/mic/vop/vop_main.c vop_del_vq(vq, idx++); idx 54 drivers/misc/ocxl/config.c u8 idx; idx 66 drivers/misc/ocxl/config.c &idx); idx 67 drivers/misc/ocxl/config.c if (idx == afu_idx) idx 415 drivers/misc/ocxl/config.c pos = find_dvsec_afu_ctrl(dev, afu->idx); idx 418 drivers/misc/ocxl/config.c afu->idx); idx 585 drivers/misc/ocxl/config.c afu->idx = afu_idx; idx 588 drivers/misc/ocxl/config.c afu->idx); idx 134 drivers/misc/ocxl/core.c int rc, idx; idx 139 drivers/misc/ocxl/core.c idx = bar >> 1; idx 140 drivers/misc/ocxl/core.c if (fn->bar_used[idx]++ == 0) { idx 151 drivers/misc/ocxl/core.c int idx; idx 156 drivers/misc/ocxl/core.c idx = bar >> 1; idx 157 drivers/misc/ocxl/core.c if (--fn->bar_used[idx] == 0) idx 159 drivers/misc/ocxl/core.c WARN_ON(fn->bar_used[idx] < 0); idx 526 drivers/misc/ocxl/core.c if (afu->config.idx == afu_idx) idx 539 drivers/misc/ocxl/file.c afu->config.name, dev_name(&pci_dev->dev), afu->config.idx); idx 37 drivers/misc/ocxl/pci.c afu->config.name, afu->config.idx); idx 455 drivers/misc/pti.c int idx = tty->index; idx 464 drivers/misc/pti.c if (idx == PTITTY_MINOR_START) idx 155 drivers/misc/sgi-gru/grumain.c char *idx) idx 166 drivers/misc/sgi-gru/grumain.c if (idx) idx 167 drivers/misc/sgi-gru/grumain.c *idx++ = i; idx 750 drivers/misc/vmw_balloon.c static unsigned long vmballoon_status_page(struct vmballoon *b, int idx, idx 755 drivers/misc/vmw_balloon.c *p = pfn_to_page(b->batch_page[idx].pfn); idx 756 drivers/misc/vmw_balloon.c return b->batch_page[idx].status; idx 828 drivers/misc/vmw_balloon.c static void vmballoon_add_page(struct vmballoon *b, unsigned int idx, idx 834 drivers/misc/vmw_balloon.c b->batch_page[idx] = (struct vmballoon_batch_entry) idx 39 drivers/misc/vmw_vmci/vmci_doorbell.c u32 idx; idx 118 drivers/misc/vmw_vmci/vmci_doorbell.c static struct dbell_entry *dbell_index_table_find(u32 idx) idx 120 drivers/misc/vmw_vmci/vmci_doorbell.c u32 bucket = VMCI_DOORBELL_HASH(idx); idx 125 drivers/misc/vmw_vmci/vmci_doorbell.c if (idx == dbell->idx) idx 186 drivers/misc/vmw_vmci/vmci_doorbell.c entry->idx = new_notify_idx; idx 187 drivers/misc/vmw_vmci/vmci_doorbell.c bucket = VMCI_DOORBELL_HASH(entry->idx); idx 204 drivers/misc/vmw_vmci/vmci_doorbell.c if (entry->idx == max_notify_idx - 1) { idx 217 drivers/misc/vmw_vmci/vmci_doorbell.c last_notify_idx_released = entry->idx; idx 361 drivers/misc/vmw_vmci/vmci_doorbell.c if (dbell->idx == notify_idx && idx 382 drivers/misc/vmw_vmci/vmci_doorbell.c u32 idx; idx 384 drivers/misc/vmw_vmci/vmci_doorbell.c for (idx = 0; idx < max_notify_idx; idx++) { idx 385 drivers/misc/vmw_vmci/vmci_doorbell.c if (bitmap[idx] & 0x1) { idx 386 drivers/misc/vmw_vmci/vmci_doorbell.c bitmap[idx] &= ~1; idx 387 drivers/misc/vmw_vmci/vmci_doorbell.c dbell_fire_entries(idx); idx 465 drivers/misc/vmw_vmci/vmci_doorbell.c entry->idx = 0; idx 486 drivers/misc/vmw_vmci/vmci_doorbell.c result = dbell_link(new_handle, entry->idx); idx 42 drivers/misc/vmw_vmci/vmci_resource.c unsigned int idx = vmci_resource_hash(handle); idx 46 drivers/misc/vmw_vmci/vmci_resource.c &vmci_resource_table.entries[idx], node) { idx 104 drivers/misc/vmw_vmci/vmci_resource.c unsigned int idx; idx 127 drivers/misc/vmw_vmci/vmci_resource.c idx = vmci_resource_hash(resource->handle); idx 128 drivers/misc/vmw_vmci/vmci_resource.c hlist_add_head_rcu(&resource->node, &vmci_resource_table.entries[idx]); idx 140 drivers/misc/vmw_vmci/vmci_resource.c unsigned int idx = vmci_resource_hash(handle); idx 146 drivers/misc/vmw_vmci/vmci_resource.c hlist_for_each_entry(r, &vmci_resource_table.entries[idx], node) { idx 2579 drivers/mmc/core/block.c int idx, ret; idx 2584 drivers/mmc/core/block.c for (idx = 0; idx < card->nr_parts; idx++) { idx 2585 drivers/mmc/core/block.c if (card->part[idx].area_type & MMC_BLK_DATA_AREA_RPMB) { idx 2593 drivers/mmc/core/block.c card->part[idx].part_cfg, idx 2594 drivers/mmc/core/block.c card->part[idx].size >> 9, idx 2595 drivers/mmc/core/block.c card->part[idx].name); idx 2598 drivers/mmc/core/block.c } else if (card->part[idx].size) { idx 2600 drivers/mmc/core/block.c card->part[idx].part_cfg, idx 2601 drivers/mmc/core/block.c card->part[idx].size >> 9, idx 2602 drivers/mmc/core/block.c card->part[idx].force_ro, idx 2603 drivers/mmc/core/block.c card->part[idx].name, idx 2604 drivers/mmc/core/block.c card->part[idx].area_type); idx 301 drivers/mmc/core/mmc.c unsigned int part_cfg, char *name, int idx, bool ro, idx 306 drivers/mmc/core/mmc.c sprintf(card->part[card->nr_parts].name, name, idx); idx 314 drivers/mmc/core/mmc.c int idx; idx 330 drivers/mmc/core/mmc.c for (idx = 0; idx < MMC_NUM_GP_PARTITION; idx++) { idx 331 drivers/mmc/core/mmc.c if (!ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3] && idx 332 drivers/mmc/core/mmc.c !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1] && idx 333 drivers/mmc/core/mmc.c !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2]) idx 341 drivers/mmc/core/mmc.c (ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2] idx 343 drivers/mmc/core/mmc.c (ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1] idx 345 drivers/mmc/core/mmc.c ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3]; idx 349 drivers/mmc/core/mmc.c EXT_CSD_PART_CONFIG_ACC_GP0 + idx, idx 350 drivers/mmc/core/mmc.c "gp%d", idx, false, idx 364 drivers/mmc/core/mmc.c int err = 0, idx; idx 450 drivers/mmc/core/mmc.c for (idx = 0; idx < MMC_NUM_BOOT_PARTITION; idx++) { idx 453 drivers/mmc/core/mmc.c EXT_CSD_PART_CONFIG_ACC_BOOT0 + idx, idx 454 drivers/mmc/core/mmc.c "boot%d", idx, true, idx 994 drivers/mmc/core/mmc.c unsigned idx, bus_width = 0; idx 1001 drivers/mmc/core/mmc.c idx = (host->caps & MMC_CAP_8_BIT_DATA) ? 0 : 1; idx 1009 drivers/mmc/core/mmc.c for (; idx < ARRAY_SIZE(bus_widths); idx++) { idx 1019 drivers/mmc/core/mmc.c ext_csd_bits[idx], idx 1024 drivers/mmc/core/mmc.c bus_width = bus_widths[idx]; idx 728 drivers/mmc/core/sdio_uart.c int idx = tty->index; idx 729 drivers/mmc/core/sdio_uart.c struct sdio_uart_port *port = sdio_uart_port_get(idx); idx 180 drivers/mmc/core/slot-gpio.c unsigned int idx, bool override_active_level, idx 187 drivers/mmc/core/slot-gpio.c desc = devm_gpiod_get_index(host->parent, con_id, idx, GPIOD_IN); idx 227 drivers/mmc/core/slot-gpio.c unsigned int idx, idx 234 drivers/mmc/core/slot-gpio.c desc = devm_gpiod_get_index(host->parent, con_id, idx, GPIOD_IN); idx 57 drivers/mmc/host/au1xmmc.c #define DBG(fmt, idx, args...) \ idx 58 drivers/mmc/host/au1xmmc.c pr_debug("au1xmmc(%d): DEBUG: " fmt, idx, ##args) idx 60 drivers/mmc/host/au1xmmc.c #define DBG(fmt, idx, args...) do {} while (0) idx 342 drivers/mmc/host/dw_mmc-exynos.c int idx; idx 349 drivers/mmc/host/dw_mmc-exynos.c for (idx = 0; idx < ARRAY_SIZE(exynos_compat); idx++) { idx 350 drivers/mmc/host/dw_mmc-exynos.c if (of_device_is_compatible(np, exynos_compat[idx].compatible)) idx 351 drivers/mmc/host/dw_mmc-exynos.c priv->ctrl_type = exynos_compat[idx].ctrl_type; idx 989 drivers/mmc/host/dw_mmc.c int idx = ARRAY_SIZE(mszs) - 1; idx 1006 drivers/mmc/host/dw_mmc.c if (!((blksz_depth % mszs[idx]) || idx 1007 drivers/mmc/host/dw_mmc.c (tx_wmark_invers % mszs[idx]))) { idx 1008 drivers/mmc/host/dw_mmc.c msize = idx; idx 1009 drivers/mmc/host/dw_mmc.c rx_wmark = mszs[idx] - 1; idx 1012 drivers/mmc/host/dw_mmc.c } while (--idx > 0); idx 611 drivers/mmc/host/rtsx_usb_sdmmc.c static inline u32 get_phase_point(u32 phase_map, unsigned int idx) idx 613 drivers/mmc/host/rtsx_usb_sdmmc.c idx &= MAX_PHASE; idx 614 drivers/mmc/host/rtsx_usb_sdmmc.c return phase_map & (1 << idx); idx 617 drivers/mmc/host/rtsx_usb_sdmmc.c static int get_phase_len(u32 phase_map, unsigned int idx) idx 622 drivers/mmc/host/rtsx_usb_sdmmc.c if (get_phase_point(phase_map, idx + i) == 0) idx 402 drivers/mmc/host/wbsd.c int i, idx, fsr, fifo; idx 411 drivers/mmc/host/wbsd.c idx = 0; idx 430 drivers/mmc/host/wbsd.c buffer[idx++] = inb(host->base + WBSD_DFR); idx 448 drivers/mmc/host/wbsd.c idx = 0; idx 467 drivers/mmc/host/wbsd.c int i, idx, fsr, fifo; idx 477 drivers/mmc/host/wbsd.c idx = 0; idx 496 drivers/mmc/host/wbsd.c outb(buffer[idx], host->base + WBSD_DFR); idx 514 drivers/mmc/host/wbsd.c idx = 0; idx 106 drivers/mtd/devices/docg3.h #define DOC_BCH_HW_ECC(idx) (0x1048 + idx) idx 222 drivers/mtd/mtdswap.c static void mtdswap_rb_add(struct mtdswap_dev *d, struct swap_eb *eb, int idx) idx 226 drivers/mtd/mtdswap.c if (eb->root == &d->trees[idx].root) idx 230 drivers/mtd/mtdswap.c root = &d->trees[idx].root; idx 233 drivers/mtd/mtdswap.c d->trees[idx].count++; idx 236 drivers/mtd/mtdswap.c static struct rb_node *mtdswap_rb_index(struct rb_root *root, unsigned int idx) idx 243 drivers/mtd/mtdswap.c while (i < idx && p) { idx 458 drivers/mtd/mtdswap.c unsigned int i, idx; idx 474 drivers/mtd/mtdswap.c idx = MTDSWAP_CLEAN; idx 478 drivers/mtd/mtdswap.c idx = MTDSWAP_DIRTY; idx 481 drivers/mtd/mtdswap.c idx = MTDSWAP_FAILING; idx 484 drivers/mtd/mtdswap.c eb->flags |= (idx << EBLOCK_IDX_SHIFT); idx 495 drivers/mtd/mtdswap.c idx = eb->flags >> EBLOCK_IDX_SHIFT; idx 496 drivers/mtd/mtdswap.c mtdswap_rb_add(d, eb, idx); idx 759 drivers/mtd/mtdswap.c int idx, stopat; idx 766 drivers/mtd/mtdswap.c for (idx = MTDSWAP_BITFLIP; idx >= stopat; idx--) idx 767 drivers/mtd/mtdswap.c if (d->trees[idx].root.rb_node != NULL) idx 768 drivers/mtd/mtdswap.c return idx; idx 804 drivers/mtd/mtdswap.c unsigned int i, idx = -1, wear, max; idx 816 drivers/mtd/mtdswap.c idx = i; idx 822 drivers/mtd/mtdswap.c return idx; idx 832 drivers/mtd/mtdswap.c int idx; idx 838 drivers/mtd/mtdswap.c idx = mtdswap_choose_wl_tree(d); idx 839 drivers/mtd/mtdswap.c if (idx >= MTDSWAP_CLEAN) idx 840 drivers/mtd/mtdswap.c return idx; idx 850 drivers/mtd/mtdswap.c int idx; idx 856 drivers/mtd/mtdswap.c idx = mtdswap_choose_gc_tree(d, background); idx 857 drivers/mtd/mtdswap.c if (idx < 0) idx 860 drivers/mtd/mtdswap.c rp = &d->trees[idx].root; idx 865 drivers/mtd/mtdswap.c d->trees[idx].count--; idx 1209 drivers/mtd/nand/raw/nandsim.c int i, j, idx = 0; idx 1229 drivers/mtd/nand/raw/nandsim.c && (ns->options & ops[idx].reqopts)) { idx 1235 drivers/mtd/nand/raw/nandsim.c idx = i; idx 1242 drivers/mtd/nand/raw/nandsim.c ns->op = &ops[idx].states[0]; idx 1259 drivers/mtd/nand/raw/nandsim.c idx, get_state_name(ns->state), get_state_name(ns->nxstate)); idx 155 drivers/mtd/nand/raw/tango_nand.c static void tango_select_chip(struct nand_chip *chip, int idx) idx 160 drivers/mtd/nand/raw/tango_nand.c if (idx < 0) idx 450 drivers/mtd/nand/raw/tango_nand.c static int oob_ecc(struct mtd_info *mtd, int idx, struct mtd_oob_region *res) idx 455 drivers/mtd/nand/raw/tango_nand.c if (idx >= ecc->steps) idx 458 drivers/mtd/nand/raw/tango_nand.c res->offset = BBM_SIZE + METADATA_SIZE + ecc->bytes * idx; idx 464 drivers/mtd/nand/raw/tango_nand.c static int oob_free(struct mtd_info *mtd, int idx, struct mtd_oob_region *res) idx 157 drivers/mtd/spi-nor/mtk-quadspi.c int i, ret, idx; idx 165 drivers/mtd/spi-nor/mtk-quadspi.c idx = MTK_NOR_MAX_RX_TX_SHIFT - 1; idx 168 drivers/mtd/spi-nor/mtk-quadspi.c writeb(op, mtk_nor->base + MTK_NOR_PRG_REG(idx)); idx 169 drivers/mtd/spi-nor/mtk-quadspi.c idx--; idx 172 drivers/mtd/spi-nor/mtk-quadspi.c for (i = 0; i < txlen; i++, idx--) idx 173 drivers/mtd/spi-nor/mtk-quadspi.c writeb(tx[i], mtk_nor->base + MTK_NOR_PRG_REG(idx)); idx 176 drivers/mtd/spi-nor/mtk-quadspi.c while (idx >= 0) { idx 177 drivers/mtd/spi-nor/mtk-quadspi.c writeb(0, mtk_nor->base + MTK_NOR_PRG_REG(idx)); idx 178 drivers/mtd/spi-nor/mtk-quadspi.c idx--; idx 186 drivers/mtd/spi-nor/mtk-quadspi.c idx = rxlen - 1; idx 189 drivers/mtd/spi-nor/mtk-quadspi.c for (i = 0; i < rxlen; i++, idx--) idx 190 drivers/mtd/spi-nor/mtk-quadspi.c rx[i] = readb(mtk_nor->base + MTK_NOR_SHREG(idx)); idx 3213 drivers/mtd/spi-nor/spi-nor.c erase->idx = i; idx 3257 drivers/mtd/spi-nor/spi-nor.c if (erase_type[i].size && erase_mask & BIT(erase_type[i].idx)) idx 3793 drivers/mtd/spi-nor/spi-nor.c if (!(regions_erase_type & BIT(erase[i].idx))) idx 4020 drivers/mtd/spi-nor/spi-nor.c erase_type[i].idx * 8) & 0xFF; idx 126 drivers/mtd/ubi/debug.c void ubi_dump_vtbl_record(const struct ubi_vtbl_record *r, int idx) idx 130 drivers/mtd/ubi/debug.c pr_err("Volume table record %d dump:\n", idx); idx 44 drivers/mtd/ubi/debug.h void ubi_dump_vtbl_record(const struct ubi_vtbl_record *r, int idx); idx 906 drivers/mtd/ubi/eba.c int err, idx = vol_id2idx(ubi, vol_id), tries; idx 907 drivers/mtd/ubi/eba.c struct ubi_volume *vol = ubi->volumes[idx]; idx 1305 drivers/mtd/ubi/eba.c int err, vol_id, lnum, data_size, aldata_size, idx; idx 1324 drivers/mtd/ubi/eba.c idx = vol_id2idx(ubi, vol_id); idx 1332 drivers/mtd/ubi/eba.c vol = ubi->volumes[idx]; idx 837 drivers/mtd/ubi/ubi.h int ubi_change_vtbl_record(struct ubi_device *ubi, int idx, idx 1180 drivers/mtd/ubi/ubi.h static inline int idx2vol_id(const struct ubi_device *ubi, int idx) idx 1182 drivers/mtd/ubi/ubi.h if (idx >= ubi->vtbl_slots) idx 1183 drivers/mtd/ubi/ubi.h return idx - ubi->vtbl_slots + UBI_INTERNAL_VOL_START; idx 1185 drivers/mtd/ubi/ubi.h return idx; idx 637 drivers/mtd/ubi/vmt.c int idx = vol_id2idx(ubi, vol_id); idx 645 drivers/mtd/ubi/vmt.c vol = ubi->volumes[idx]; idx 90 drivers/mtd/ubi/vtbl.c int ubi_change_vtbl_record(struct ubi_device *ubi, int idx, idx 96 drivers/mtd/ubi/vtbl.c ubi_assert(idx >= 0 && idx < ubi->vtbl_slots); idx 105 drivers/mtd/ubi/vtbl.c memcpy(&ubi->vtbl[idx], vtbl_rec, sizeof(struct ubi_vtbl_record)); idx 4070 drivers/net/bonding/bond_main.c int idx; idx 4081 drivers/net/bonding/bond_main.c for (idx = 0; old_arr != NULL && idx < old_arr->count; idx++) { idx 4082 drivers/net/bonding/bond_main.c if (skipslave == old_arr->arr[idx]) { idx 4083 drivers/net/bonding/bond_main.c old_arr->arr[idx] = idx 306 drivers/net/can/c_can/c_can.c struct can_frame *frame, int idx) idx 328 drivers/net/can/c_can/c_can.c if (rtr != (bool)test_bit(idx, &priv->tx_dir)) { idx 329 drivers/net/can/c_can/c_can.c u32 obj = idx + C_CAN_MSG_OBJ_TX_FIRST; idx 332 drivers/net/can/c_can/c_can.c change_bit(idx, &priv->tx_dir); idx 467 drivers/net/can/c_can/c_can.c u32 idx, obj; idx 475 drivers/net/can/c_can/c_can.c idx = fls(atomic_read(&priv->tx_active)); idx 476 drivers/net/can/c_can/c_can.c obj = idx + C_CAN_MSG_OBJ_TX_FIRST; idx 479 drivers/net/can/c_can/c_can.c if (idx == C_CAN_MSG_OBJ_TX_NUM - 1) idx 486 drivers/net/can/c_can/c_can.c c_can_setup_tx_object(dev, IF_TX, frame, idx); idx 487 drivers/net/can/c_can/c_can.c priv->dlc[idx] = frame->can_dlc; idx 488 drivers/net/can/c_can/c_can.c can_put_echo_skb(skb, dev, idx); idx 491 drivers/net/can/c_can/c_can.c atomic_add((1 << idx), &priv->tx_active); idx 736 drivers/net/can/c_can/c_can.c u32 idx, obj, pkts = 0, bytes = 0, pend, clr; idx 740 drivers/net/can/c_can/c_can.c while ((idx = ffs(pend))) { idx 741 drivers/net/can/c_can/c_can.c idx--; idx 742 drivers/net/can/c_can/c_can.c pend &= ~(1 << idx); idx 743 drivers/net/can/c_can/c_can.c obj = idx + C_CAN_MSG_OBJ_TX_FIRST; idx 745 drivers/net/can/c_can/c_can.c can_get_echo_skb(dev, idx); idx 746 drivers/net/can/c_can/c_can.c bytes += priv->dlc[idx]; idx 167 drivers/net/can/cc770/cc770_isa.c int idx = pdev->id; idx 172 drivers/net/can/cc770/cc770_isa.c idx, port[idx], mem[idx], irq[idx]); idx 173 drivers/net/can/cc770/cc770_isa.c if (mem[idx]) { idx 174 drivers/net/can/cc770/cc770_isa.c if (!request_mem_region(mem[idx], iosize, KBUILD_MODNAME)) { idx 178 drivers/net/can/cc770/cc770_isa.c base = ioremap_nocache(mem[idx], iosize); idx 184 drivers/net/can/cc770/cc770_isa.c if (indirect[idx] > 0 || idx 185 drivers/net/can/cc770/cc770_isa.c (indirect[idx] == -1 && indirect[0] > 0)) idx 187 drivers/net/can/cc770/cc770_isa.c if (!request_region(port[idx], iosize, KBUILD_MODNAME)) { idx 200 drivers/net/can/cc770/cc770_isa.c dev->irq = irq[idx]; idx 202 drivers/net/can/cc770/cc770_isa.c if (mem[idx]) { idx 204 drivers/net/can/cc770/cc770_isa.c dev->base_addr = mem[idx]; idx 208 drivers/net/can/cc770/cc770_isa.c priv->reg_base = (void __iomem *)port[idx]; idx 209 drivers/net/can/cc770/cc770_isa.c dev->base_addr = port[idx]; idx 220 drivers/net/can/cc770/cc770_isa.c if (clk[idx]) idx 221 drivers/net/can/cc770/cc770_isa.c clktmp = clk[idx]; idx 228 drivers/net/can/cc770/cc770_isa.c if (cir[idx] != 0xff) { idx 229 drivers/net/can/cc770/cc770_isa.c priv->cpu_interface = cir[idx]; idx 246 drivers/net/can/cc770/cc770_isa.c if (bcr[idx] != 0xff) idx 247 drivers/net/can/cc770/cc770_isa.c priv->bus_config = bcr[idx]; idx 253 drivers/net/can/cc770/cc770_isa.c if (cor[idx] != 0xff) idx 254 drivers/net/can/cc770/cc770_isa.c priv->clkout = cor[idx]; idx 275 drivers/net/can/cc770/cc770_isa.c if (mem[idx]) idx 278 drivers/net/can/cc770/cc770_isa.c if (mem[idx]) idx 279 drivers/net/can/cc770/cc770_isa.c release_mem_region(mem[idx], iosize); idx 281 drivers/net/can/cc770/cc770_isa.c release_region(port[idx], iosize); idx 290 drivers/net/can/cc770/cc770_isa.c int idx = pdev->id; idx 294 drivers/net/can/cc770/cc770_isa.c if (mem[idx]) { idx 296 drivers/net/can/cc770/cc770_isa.c release_mem_region(mem[idx], CC770_IOSIZE); idx 299 drivers/net/can/cc770/cc770_isa.c release_region(port[idx], CC770_IOSIZE_INDIRECT); idx 301 drivers/net/can/cc770/cc770_isa.c release_region(port[idx], CC770_IOSIZE); idx 318 drivers/net/can/cc770/cc770_isa.c int idx, err; idx 320 drivers/net/can/cc770/cc770_isa.c for (idx = 0; idx < ARRAY_SIZE(cc770_isa_devs); idx++) { idx 321 drivers/net/can/cc770/cc770_isa.c if ((port[idx] || mem[idx]) && irq[idx]) { idx 322 drivers/net/can/cc770/cc770_isa.c cc770_isa_devs[idx] = idx 323 drivers/net/can/cc770/cc770_isa.c platform_device_alloc(KBUILD_MODNAME, idx); idx 324 drivers/net/can/cc770/cc770_isa.c if (!cc770_isa_devs[idx]) { idx 328 drivers/net/can/cc770/cc770_isa.c err = platform_device_add(cc770_isa_devs[idx]); idx 330 drivers/net/can/cc770/cc770_isa.c platform_device_put(cc770_isa_devs[idx]); idx 335 drivers/net/can/cc770/cc770_isa.c idx, port[idx], mem[idx], irq[idx]); idx 336 drivers/net/can/cc770/cc770_isa.c } else if (idx == 0 || port[idx] || mem[idx]) { idx 352 drivers/net/can/cc770/cc770_isa.c while (--idx >= 0) { idx 353 drivers/net/can/cc770/cc770_isa.c if (cc770_isa_devs[idx]) idx 354 drivers/net/can/cc770/cc770_isa.c platform_device_unregister(cc770_isa_devs[idx]); idx 363 drivers/net/can/cc770/cc770_isa.c int idx; idx 366 drivers/net/can/cc770/cc770_isa.c for (idx = 0; idx < ARRAY_SIZE(cc770_isa_devs); idx++) { idx 367 drivers/net/can/cc770/cc770_isa.c if (cc770_isa_devs[idx]) idx 368 drivers/net/can/cc770/cc770_isa.c platform_device_unregister(cc770_isa_devs[idx]); idx 438 drivers/net/can/dev.c unsigned int idx) idx 442 drivers/net/can/dev.c BUG_ON(idx >= priv->echo_skb_max); idx 452 drivers/net/can/dev.c if (!priv->echo_skb[idx]) { idx 463 drivers/net/can/dev.c priv->echo_skb[idx] = skb; idx 473 drivers/net/can/dev.c __can_get_echo_skb(struct net_device *dev, unsigned int idx, u8 *len_ptr) idx 477 drivers/net/can/dev.c if (idx >= priv->echo_skb_max) { idx 479 drivers/net/can/dev.c __func__, idx, priv->echo_skb_max); idx 483 drivers/net/can/dev.c if (priv->echo_skb[idx]) { idx 487 drivers/net/can/dev.c struct sk_buff *skb = priv->echo_skb[idx]; idx 492 drivers/net/can/dev.c priv->echo_skb[idx] = NULL; idx 506 drivers/net/can/dev.c unsigned int can_get_echo_skb(struct net_device *dev, unsigned int idx) idx 511 drivers/net/can/dev.c skb = __can_get_echo_skb(dev, idx, &len); idx 525 drivers/net/can/dev.c void can_free_echo_skb(struct net_device *dev, unsigned int idx) idx 529 drivers/net/can/dev.c BUG_ON(idx >= priv->echo_skb_max); idx 531 drivers/net/can/dev.c if (priv->echo_skb[idx]) { idx 532 drivers/net/can/dev.c dev_kfree_skb_any(priv->echo_skb[idx]); idx 533 drivers/net/can/dev.c priv->echo_skb[idx] = NULL; idx 174 drivers/net/can/peak_canfd/peak_canfd.c cmd->idx = row; idx 298 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_CFCC(ch, idx) (0x0118 + (0x0c * (ch)) + \ idx 299 drivers/net/can/rcar/rcar_canfd.c (0x04 * (idx))) idx 301 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_CFSTS(ch, idx) (0x0178 + (0x0c * (ch)) + \ idx 302 drivers/net/can/rcar/rcar_canfd.c (0x04 * (idx))) idx 304 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_CFPCTR(ch, idx) (0x01d8 + (0x0c * (ch)) + \ idx 305 drivers/net/can/rcar/rcar_canfd.c (0x04 * (idx))) idx 393 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_C_CFID(ch, idx) (RCANFD_C_CFOFFSET + (0x30 * (ch)) + \ idx 394 drivers/net/can/rcar/rcar_canfd.c (0x10 * (idx))) idx 395 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_C_CFPTR(ch, idx) (RCANFD_C_CFOFFSET + 0x04 + \ idx 396 drivers/net/can/rcar/rcar_canfd.c (0x30 * (ch)) + (0x10 * (idx))) idx 397 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_C_CFDF(ch, idx, df) (RCANFD_C_CFOFFSET + 0x08 + \ idx 398 drivers/net/can/rcar/rcar_canfd.c (0x30 * (ch)) + (0x10 * (idx)) + \ idx 442 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_F_CFID(ch, idx) (RCANFD_F_CFOFFSET + (0x180 * (ch)) + \ idx 443 drivers/net/can/rcar/rcar_canfd.c (0x80 * (idx))) idx 444 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_F_CFPTR(ch, idx) (RCANFD_F_CFOFFSET + 0x04 + \ idx 445 drivers/net/can/rcar/rcar_canfd.c (0x180 * (ch)) + (0x80 * (idx))) idx 446 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_F_CFFDCSTS(ch, idx) (RCANFD_F_CFOFFSET + 0x08 + \ idx 447 drivers/net/can/rcar/rcar_canfd.c (0x180 * (ch)) + (0x80 * (idx))) idx 448 drivers/net/can/rcar/rcar_canfd.c #define RCANFD_F_CFDF(ch, idx, df) (RCANFD_F_CFOFFSET + 0x0c + \ idx 449 drivers/net/can/rcar/rcar_canfd.c (0x180 * (ch)) + (0x80 * (idx)) + \ idx 293 drivers/net/can/rx-offload.c unsigned int idx, u32 timestamp) idx 301 drivers/net/can/rx-offload.c skb = __can_get_echo_skb(dev, idx, &len); idx 122 drivers/net/can/sja1000/sja1000_isa.c int idx = pdev->id; idx 126 drivers/net/can/sja1000/sja1000_isa.c idx, port[idx], mem[idx], irq[idx]); idx 128 drivers/net/can/sja1000/sja1000_isa.c if (mem[idx]) { idx 129 drivers/net/can/sja1000/sja1000_isa.c if (!request_mem_region(mem[idx], iosize, DRV_NAME)) { idx 133 drivers/net/can/sja1000/sja1000_isa.c base = ioremap_nocache(mem[idx], iosize); idx 139 drivers/net/can/sja1000/sja1000_isa.c if (indirect[idx] > 0 || idx 140 drivers/net/can/sja1000/sja1000_isa.c (indirect[idx] == -1 && indirect[0] > 0)) idx 142 drivers/net/can/sja1000/sja1000_isa.c if (!request_region(port[idx], iosize, DRV_NAME)) { idx 155 drivers/net/can/sja1000/sja1000_isa.c dev->irq = irq[idx]; idx 157 drivers/net/can/sja1000/sja1000_isa.c if (mem[idx]) { idx 159 drivers/net/can/sja1000/sja1000_isa.c dev->base_addr = mem[idx]; idx 163 drivers/net/can/sja1000/sja1000_isa.c priv->reg_base = (void __iomem *)port[idx]; idx 164 drivers/net/can/sja1000/sja1000_isa.c dev->base_addr = port[idx]; idx 169 drivers/net/can/sja1000/sja1000_isa.c spin_lock_init(&indirect_lock[idx]); idx 176 drivers/net/can/sja1000/sja1000_isa.c if (clk[idx]) idx 177 drivers/net/can/sja1000/sja1000_isa.c priv->can.clock.freq = clk[idx] / 2; idx 183 drivers/net/can/sja1000/sja1000_isa.c if (ocr[idx] != 0xff) idx 184 drivers/net/can/sja1000/sja1000_isa.c priv->ocr = ocr[idx]; idx 190 drivers/net/can/sja1000/sja1000_isa.c if (cdr[idx] != 0xff) idx 191 drivers/net/can/sja1000/sja1000_isa.c priv->cdr = cdr[idx]; idx 199 drivers/net/can/sja1000/sja1000_isa.c dev->dev_id = idx; idx 213 drivers/net/can/sja1000/sja1000_isa.c if (mem[idx]) idx 216 drivers/net/can/sja1000/sja1000_isa.c if (mem[idx]) idx 217 drivers/net/can/sja1000/sja1000_isa.c release_mem_region(mem[idx], iosize); idx 219 drivers/net/can/sja1000/sja1000_isa.c release_region(port[idx], iosize); idx 228 drivers/net/can/sja1000/sja1000_isa.c int idx = pdev->id; idx 232 drivers/net/can/sja1000/sja1000_isa.c if (mem[idx]) { idx 234 drivers/net/can/sja1000/sja1000_isa.c release_mem_region(mem[idx], SJA1000_IOSIZE); idx 237 drivers/net/can/sja1000/sja1000_isa.c release_region(port[idx], SJA1000_IOSIZE_INDIRECT); idx 239 drivers/net/can/sja1000/sja1000_isa.c release_region(port[idx], SJA1000_IOSIZE); idx 256 drivers/net/can/sja1000/sja1000_isa.c int idx, err; idx 258 drivers/net/can/sja1000/sja1000_isa.c for (idx = 0; idx < MAXDEV; idx++) { idx 259 drivers/net/can/sja1000/sja1000_isa.c if ((port[idx] || mem[idx]) && irq[idx]) { idx 260 drivers/net/can/sja1000/sja1000_isa.c sja1000_isa_devs[idx] = idx 261 drivers/net/can/sja1000/sja1000_isa.c platform_device_alloc(DRV_NAME, idx); idx 262 drivers/net/can/sja1000/sja1000_isa.c if (!sja1000_isa_devs[idx]) { idx 266 drivers/net/can/sja1000/sja1000_isa.c err = platform_device_add(sja1000_isa_devs[idx]); idx 268 drivers/net/can/sja1000/sja1000_isa.c platform_device_put(sja1000_isa_devs[idx]); idx 273 drivers/net/can/sja1000/sja1000_isa.c DRV_NAME, idx, port[idx], mem[idx], irq[idx]); idx 274 drivers/net/can/sja1000/sja1000_isa.c } else if (idx == 0 || port[idx] || mem[idx]) { idx 292 drivers/net/can/sja1000/sja1000_isa.c while (--idx >= 0) { idx 293 drivers/net/can/sja1000/sja1000_isa.c if (sja1000_isa_devs[idx]) idx 294 drivers/net/can/sja1000/sja1000_isa.c platform_device_unregister(sja1000_isa_devs[idx]); idx 302 drivers/net/can/sja1000/sja1000_isa.c int idx; idx 305 drivers/net/can/sja1000/sja1000_isa.c for (idx = 0; idx < MAXDEV; idx++) { idx 306 drivers/net/can/sja1000/sja1000_isa.c if (sja1000_isa_devs[idx]) idx 307 drivers/net/can/sja1000/sja1000_isa.c platform_device_unregister(sja1000_isa_devs[idx]); idx 471 drivers/net/can/usb/gs_usb.c unsigned int idx; idx 494 drivers/net/can/usb/gs_usb.c idx = txc->echo_id; idx 496 drivers/net/can/usb/gs_usb.c if (idx >= GS_MAX_TX_URBS) { idx 497 drivers/net/can/usb/gs_usb.c netdev_err(netdev, "Invalid tx context %d\n", idx); idx 501 drivers/net/can/usb/gs_usb.c hf->echo_id = idx; idx 520 drivers/net/can/usb/gs_usb.c can_put_echo_skb(skb, netdev, idx); idx 528 drivers/net/can/usb/gs_usb.c can_free_echo_skb(netdev, idx); idx 312 drivers/net/can/usb/peak_usb/pcan_usb_fd.c static int pcan_usb_fd_set_filter_std(struct peak_usb_device *dev, int idx, idx 319 drivers/net/can/usb/peak_usb/pcan_usb_fd.c if ((idx < 0) || (idx >= (1 << PUCAN_FLTSTD_ROW_IDX_BITS))) { idx 321 drivers/net/can/usb/peak_usb/pcan_usb_fd.c idx = 0; idx 325 drivers/net/can/usb/peak_usb/pcan_usb_fd.c n = idx + 1; idx 328 drivers/net/can/usb/peak_usb/pcan_usb_fd.c for (i = idx; i < n; i++, cmd++) { idx 331 drivers/net/can/usb/peak_usb/pcan_usb_fd.c cmd->idx = cpu_to_le16(i); idx 1454 drivers/net/dsa/b53/b53_common.c u16 vid, struct b53_arl_entry *ent, u8 *idx, idx 1487 drivers/net/dsa/b53/b53_common.c *idx = i; idx 1494 drivers/net/dsa/b53/b53_common.c *idx = find_first_bit(free_bins, dev->num_arl_entries); idx 1505 drivers/net/dsa/b53/b53_common.c u8 idx = 0; idx 1520 drivers/net/dsa/b53/b53_common.c ret = b53_arl_read(dev, mac, vid, &ent, &idx, is_valid); idx 1533 drivers/net/dsa/b53/b53_common.c addr, vid, idx); idx 1538 drivers/net/dsa/b53/b53_common.c addr, vid, idx); idx 1550 drivers/net/dsa/b53/b53_common.c B53_ARLTBL_MAC_VID_ENTRY(idx), mac_vid); idx 1552 drivers/net/dsa/b53/b53_common.c B53_ARLTBL_DATA_ENTRY(idx), fwd_entry); idx 1600 drivers/net/dsa/b53/b53_common.c static void b53_arl_search_rd(struct b53_device *dev, u8 idx, idx 1607 drivers/net/dsa/b53/b53_common.c B53_ARL_SRCH_RSTL_MACVID(idx), &mac_vid); idx 1609 drivers/net/dsa/b53/b53_common.c B53_ARL_SRCH_RSTL(idx), &fwd_entry); idx 855 drivers/net/dsa/lantiq_gswip.c int idx = -1; idx 862 drivers/net/dsa/lantiq_gswip.c idx = i; idx 867 drivers/net/dsa/lantiq_gswip.c if (idx == -1) idx 871 drivers/net/dsa/lantiq_gswip.c fid = idx; idx 873 drivers/net/dsa/lantiq_gswip.c vlan_active.index = idx; idx 885 drivers/net/dsa/lantiq_gswip.c priv->vlans[idx].bridge = bridge; idx 886 drivers/net/dsa/lantiq_gswip.c priv->vlans[idx].vid = vid; idx 887 drivers/net/dsa/lantiq_gswip.c priv->vlans[idx].fid = fid; idx 889 drivers/net/dsa/lantiq_gswip.c return idx; idx 892 drivers/net/dsa/lantiq_gswip.c static int gswip_vlan_active_remove(struct gswip_priv *priv, int idx) idx 897 drivers/net/dsa/lantiq_gswip.c vlan_active.index = idx; idx 903 drivers/net/dsa/lantiq_gswip.c priv->vlans[idx].bridge = NULL; idx 915 drivers/net/dsa/lantiq_gswip.c int idx = -1; idx 922 drivers/net/dsa/lantiq_gswip.c idx = i; idx 930 drivers/net/dsa/lantiq_gswip.c if (idx == -1) { idx 931 drivers/net/dsa/lantiq_gswip.c idx = gswip_vlan_active_create(priv, bridge, -1, 0); idx 932 drivers/net/dsa/lantiq_gswip.c if (idx < 0) idx 933 drivers/net/dsa/lantiq_gswip.c return idx; idx 936 drivers/net/dsa/lantiq_gswip.c vlan_mapping.index = idx; idx 942 drivers/net/dsa/lantiq_gswip.c vlan_mapping.index = idx; idx 960 drivers/net/dsa/lantiq_gswip.c gswip_vlan_active_remove(priv, idx); idx 977 drivers/net/dsa/lantiq_gswip.c int idx = -1; idx 989 drivers/net/dsa/lantiq_gswip.c idx = i; idx 998 drivers/net/dsa/lantiq_gswip.c if (idx == -1) { idx 999 drivers/net/dsa/lantiq_gswip.c idx = gswip_vlan_active_create(priv, bridge, fid, vid); idx 1000 drivers/net/dsa/lantiq_gswip.c if (idx < 0) idx 1001 drivers/net/dsa/lantiq_gswip.c return idx; idx 1004 drivers/net/dsa/lantiq_gswip.c vlan_mapping.index = idx; idx 1010 drivers/net/dsa/lantiq_gswip.c vlan_mapping.index = idx; idx 1034 drivers/net/dsa/lantiq_gswip.c gswip_vlan_active_remove(priv, idx); idx 1039 drivers/net/dsa/lantiq_gswip.c gswip_switch_w(priv, idx, GSWIP_PCE_DEFPVID(port)); idx 1051 drivers/net/dsa/lantiq_gswip.c int idx = -1; idx 1059 drivers/net/dsa/lantiq_gswip.c idx = i; idx 1064 drivers/net/dsa/lantiq_gswip.c if (idx == -1) { idx 1069 drivers/net/dsa/lantiq_gswip.c vlan_mapping.index = idx; idx 1087 drivers/net/dsa/lantiq_gswip.c err = gswip_vlan_active_remove(priv, idx); idx 1151 drivers/net/dsa/lantiq_gswip.c int idx = -1; idx 1157 drivers/net/dsa/lantiq_gswip.c idx = i; idx 1166 drivers/net/dsa/lantiq_gswip.c if (idx == -1) { idx 1170 drivers/net/dsa/lantiq_gswip.c idx = pos; idx 1176 drivers/net/dsa/lantiq_gswip.c if (idx == -1) idx 1560 drivers/net/ethernet/3com/typhoon.c typhoon_recycle_rx_skb(struct typhoon *tp, u32 idx) idx 1563 drivers/net/ethernet/3com/typhoon.c struct rxbuff_ent *rxb = &tp->rxbuffers[idx]; idx 1578 drivers/net/ethernet/3com/typhoon.c r->virtAddr = idx; idx 1587 drivers/net/ethernet/3com/typhoon.c typhoon_alloc_rx_skb(struct typhoon *tp, u32 idx) idx 1590 drivers/net/ethernet/3com/typhoon.c struct rxbuff_ent *rxb = &tp->rxbuffers[idx]; idx 1621 drivers/net/ethernet/3com/typhoon.c r->virtAddr = idx; idx 1643 drivers/net/ethernet/3com/typhoon.c u32 idx; idx 1652 drivers/net/ethernet/3com/typhoon.c idx = rx->addr; idx 1653 drivers/net/ethernet/3com/typhoon.c rxb = &tp->rxbuffers[idx]; idx 1660 drivers/net/ethernet/3com/typhoon.c typhoon_recycle_rx_skb(tp, idx); idx 1677 drivers/net/ethernet/3com/typhoon.c typhoon_recycle_rx_skb(tp, idx); idx 1683 drivers/net/ethernet/3com/typhoon.c typhoon_alloc_rx_skb(tp, idx); idx 65 drivers/net/ethernet/alacritech/slicoss.c static inline int slic_next_queue_idx(unsigned int idx, unsigned int qlen) idx 67 drivers/net/ethernet/alacritech/slicoss.c return (idx + 1) & (qlen - 1); idx 85 drivers/net/ethernet/alacritech/slicoss.c unsigned int idx; idx 93 drivers/net/ethernet/alacritech/slicoss.c idx = (le32_to_cpu(stat->hnd) & 0xffff) - 1; idx 111 drivers/net/ethernet/alacritech/slicoss.c return idx; idx 350 drivers/net/ethernet/alacritech/slicoss.c unsigned int idx; idx 356 drivers/net/ethernet/alacritech/slicoss.c idx = slic_next_compl_idx(sdev); idx 357 drivers/net/ethernet/alacritech/slicoss.c if (idx == SLIC_INVALID_STAT_DESC_IDX) idx 360 drivers/net/ethernet/alacritech/slicoss.c txq->done_idx = idx; idx 361 drivers/net/ethernet/alacritech/slicoss.c buff = &txq->txbuffs[idx]; idx 365 drivers/net/ethernet/alacritech/slicoss.c "no skb found for desc idx %i\n", idx); idx 1029 drivers/net/ethernet/alacritech/slicoss.c int idx = *offset; idx 1033 drivers/net/ethernet/alacritech/slicoss.c idx += 4; idx 1034 drivers/net/ethernet/alacritech/slicoss.c *offset = idx; idx 1047 drivers/net/ethernet/alacritech/slicoss.c int idx = 0; idx 1071 drivers/net/ethernet/alacritech/slicoss.c codelen = slic_read_dword_from_firmware(fw, &idx); idx 1088 drivers/net/ethernet/alacritech/slicoss.c instr = slic_read_dword_from_firmware(fw, &idx); idx 1092 drivers/net/ethernet/alacritech/slicoss.c val = (__le32)fw->data[idx]; idx 1094 drivers/net/ethernet/alacritech/slicoss.c idx++; idx 1120 drivers/net/ethernet/alacritech/slicoss.c int idx = 0; idx 1145 drivers/net/ethernet/alacritech/slicoss.c numsects = slic_read_dword_from_firmware(fw, &idx); idx 1155 drivers/net/ethernet/alacritech/slicoss.c sectsize[i] = slic_read_dword_from_firmware(fw, &idx); idx 1169 drivers/net/ethernet/alacritech/slicoss.c sectstart[i] = slic_read_dword_from_firmware(fw, &idx); idx 1171 drivers/net/ethernet/alacritech/slicoss.c code_start = idx; idx 1172 drivers/net/ethernet/alacritech/slicoss.c instr = slic_read_dword_from_firmware(fw, &idx); idx 1184 drivers/net/ethernet/alacritech/slicoss.c instr = slic_read_dword_from_firmware(fw, &idx); idx 1187 drivers/net/ethernet/alacritech/slicoss.c instr = slic_read_dword_from_firmware(fw, &idx); idx 1191 drivers/net/ethernet/alacritech/slicoss.c idx = code_start; idx 1196 drivers/net/ethernet/alacritech/slicoss.c instr = slic_read_dword_from_firmware(fw, &idx); idx 1207 drivers/net/ethernet/alacritech/slicoss.c instr = slic_read_dword_from_firmware(fw, &idx); idx 1210 drivers/net/ethernet/alacritech/slicoss.c instr = slic_read_dword_from_firmware(fw, &idx); idx 850 drivers/net/ethernet/alteon/acenic.c u32 idx; idx 852 drivers/net/ethernet/alteon/acenic.c idx = readl(®s->CmdPrd); idx 854 drivers/net/ethernet/alteon/acenic.c writel(*(u32 *)(cmd), ®s->CmdRng[idx]); idx 855 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % CMD_RING_ENTRIES; idx 857 drivers/net/ethernet/alteon/acenic.c writel(idx, ®s->CmdPrd); idx 1633 drivers/net/ethernet/alteon/acenic.c short i, idx; idx 1638 drivers/net/ethernet/alteon/acenic.c idx = ap->rx_std_skbprd; idx 1653 drivers/net/ethernet/alteon/acenic.c ap->skb->rx_std_skbuff[idx].skb = skb; idx 1654 drivers/net/ethernet/alteon/acenic.c dma_unmap_addr_set(&ap->skb->rx_std_skbuff[idx], idx 1657 drivers/net/ethernet/alteon/acenic.c rd = &ap->rx_std_ring[idx]; idx 1660 drivers/net/ethernet/alteon/acenic.c rd->idx = idx; idx 1661 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % RX_STD_RING_ENTRIES; idx 1668 drivers/net/ethernet/alteon/acenic.c ap->rx_std_skbprd = idx; idx 1674 drivers/net/ethernet/alteon/acenic.c cmd.idx = ap->rx_std_skbprd; idx 1677 drivers/net/ethernet/alteon/acenic.c writel(idx, ®s->RxStdPrd); idx 1696 drivers/net/ethernet/alteon/acenic.c short i, idx; idx 1700 drivers/net/ethernet/alteon/acenic.c idx = ap->rx_mini_skbprd; idx 1714 drivers/net/ethernet/alteon/acenic.c ap->skb->rx_mini_skbuff[idx].skb = skb; idx 1715 drivers/net/ethernet/alteon/acenic.c dma_unmap_addr_set(&ap->skb->rx_mini_skbuff[idx], idx 1718 drivers/net/ethernet/alteon/acenic.c rd = &ap->rx_mini_ring[idx]; idx 1721 drivers/net/ethernet/alteon/acenic.c rd->idx = idx; idx 1722 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % RX_MINI_RING_ENTRIES; idx 1730 drivers/net/ethernet/alteon/acenic.c ap->rx_mini_skbprd = idx; idx 1732 drivers/net/ethernet/alteon/acenic.c writel(idx, ®s->RxMiniPrd); idx 1753 drivers/net/ethernet/alteon/acenic.c short i, idx; idx 1755 drivers/net/ethernet/alteon/acenic.c idx = ap->rx_jumbo_skbprd; idx 1770 drivers/net/ethernet/alteon/acenic.c ap->skb->rx_jumbo_skbuff[idx].skb = skb; idx 1771 drivers/net/ethernet/alteon/acenic.c dma_unmap_addr_set(&ap->skb->rx_jumbo_skbuff[idx], idx 1774 drivers/net/ethernet/alteon/acenic.c rd = &ap->rx_jumbo_ring[idx]; idx 1777 drivers/net/ethernet/alteon/acenic.c rd->idx = idx; idx 1778 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % RX_JUMBO_RING_ENTRIES; idx 1785 drivers/net/ethernet/alteon/acenic.c ap->rx_jumbo_skbprd = idx; idx 1791 drivers/net/ethernet/alteon/acenic.c cmd.idx = ap->rx_jumbo_skbprd; idx 1794 drivers/net/ethernet/alteon/acenic.c writel(idx, ®s->RxJumboPrd); idx 1894 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 1922 drivers/net/ethernet/alteon/acenic.c u32 idx; idx 1925 drivers/net/ethernet/alteon/acenic.c idx = rxretcsm; idx 1930 drivers/net/ethernet/alteon/acenic.c while (idx != rxretprd) { idx 1940 drivers/net/ethernet/alteon/acenic.c if (idx == rxretcsm) idx 1943 drivers/net/ethernet/alteon/acenic.c retdesc = &ap->rx_return_ring[idx]; idx 1944 drivers/net/ethernet/alteon/acenic.c skbidx = retdesc->idx; idx 2012 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % RX_RETURN_RING_ENTRIES; idx 2025 drivers/net/ethernet/alteon/acenic.c writel(idx, &ap->regs->RxRetCsm); idx 2027 drivers/net/ethernet/alteon/acenic.c ap->cur_rx = idx; idx 2031 drivers/net/ethernet/alteon/acenic.c idx = rxretprd; idx 2037 drivers/net/ethernet/alteon/acenic.c u32 txcsm, u32 idx) idx 2045 drivers/net/ethernet/alteon/acenic.c info = ap->skb->tx_skbuff + idx; idx 2062 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % ACE_TX_RING_ENTRIES(ap); idx 2063 drivers/net/ethernet/alteon/acenic.c } while (idx != txcsm); idx 2106 drivers/net/ethernet/alteon/acenic.c u32 idx; idx 2143 drivers/net/ethernet/alteon/acenic.c idx = ap->tx_ret_csm; idx 2145 drivers/net/ethernet/alteon/acenic.c if (txcsm != idx) { idx 2154 drivers/net/ethernet/alteon/acenic.c ace_tx_int(dev, txcsm, idx); idx 2243 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2248 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2258 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2269 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2302 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2309 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2354 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2367 drivers/net/ethernet/alteon/acenic.c struct sk_buff *tail, u32 idx) idx 2376 drivers/net/ethernet/alteon/acenic.c info = ap->skb->tx_skbuff + idx; idx 2413 drivers/net/ethernet/alteon/acenic.c u32 idx, flagsize; idx 2417 drivers/net/ethernet/alteon/acenic.c idx = ap->tx_prd; idx 2419 drivers/net/ethernet/alteon/acenic.c if (tx_ring_full(ap, ap->tx_ret_csm, idx)) idx 2426 drivers/net/ethernet/alteon/acenic.c mapping = ace_map_tx_skb(ap, skb, skb, idx); idx 2434 drivers/net/ethernet/alteon/acenic.c desc = ap->tx_ring + idx; idx 2435 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % ACE_TX_RING_ENTRIES(ap); idx 2438 drivers/net/ethernet/alteon/acenic.c if (tx_ring_full(ap, ap->tx_ret_csm, idx)) idx 2447 drivers/net/ethernet/alteon/acenic.c mapping = ace_map_tx_skb(ap, skb, NULL, idx); idx 2456 drivers/net/ethernet/alteon/acenic.c ace_load_tx_bd(ap, ap->tx_ring + idx, mapping, flagsize, vlan_tag); idx 2458 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % ACE_TX_RING_ENTRIES(ap); idx 2465 drivers/net/ethernet/alteon/acenic.c info = ap->skb->tx_skbuff + idx; idx 2466 drivers/net/ethernet/alteon/acenic.c desc = ap->tx_ring + idx; idx 2475 drivers/net/ethernet/alteon/acenic.c idx = (idx + 1) % ACE_TX_RING_ENTRIES(ap); idx 2479 drivers/net/ethernet/alteon/acenic.c if (tx_ring_full(ap, ap->tx_ret_csm, idx)) idx 2497 drivers/net/ethernet/alteon/acenic.c ap->tx_prd = idx; idx 2498 drivers/net/ethernet/alteon/acenic.c ace_set_txprd(regs, ap, idx); idx 2509 drivers/net/ethernet/alteon/acenic.c if (!tx_ring_full(ap, ap->tx_ret_csm, idx)) idx 2570 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2690 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2736 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2752 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2758 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2766 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2772 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2786 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 2791 drivers/net/ethernet/alteon/acenic.c cmd.idx = 0; idx 318 drivers/net/ethernet/alteon/acenic.h u32 idx:12; idx 324 drivers/net/ethernet/alteon/acenic.h u32 idx:12; idx 364 drivers/net/ethernet/alteon/acenic.h u32 idx:12; idx 370 drivers/net/ethernet/alteon/acenic.h u32 idx:12; idx 493 drivers/net/ethernet/alteon/acenic.h u16 idx; idx 495 drivers/net/ethernet/alteon/acenic.h u16 idx; idx 896 drivers/net/ethernet/amazon/ena/ena_com.c destroy_cmd.sq.sq_idx = io_sq->idx; idx 1252 drivers/net/ethernet/amazon/ena/ena_com.c io_sq->idx = cmd_completion.sq_idx; idx 1266 drivers/net/ethernet/amazon/ena/ena_com.c pr_debug("created sq[%u], depth[%u]\n", io_sq->idx, io_sq->q_depth); idx 1288 drivers/net/ethernet/amazon/ena/ena_com.c rss->rss_ind_tbl[i].cq_idx = io_sq->idx; idx 1397 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->idx = cmd_completion.cq_idx; idx 1412 drivers/net/ethernet/amazon/ena/ena_com.c pr_debug("created cq[%u], depth[%u]\n", io_cq->idx, io_cq->q_depth); idx 1477 drivers/net/ethernet/amazon/ena/ena_com.c destroy_cmd.cq_idx = io_cq->idx; idx 1866 drivers/net/ethernet/amazon/ena/ena_com.c ret = ena_com_create_io_sq(ena_dev, io_sq, io_cq->idx); idx 159 drivers/net/ethernet/amazon/ena/ena_com.h u16 idx; idx 177 drivers/net/ethernet/amazon/ena/ena_com.h u16 idx; idx 199 drivers/net/ethernet/amazon/ena/ena_com.h u16 idx; idx 158 drivers/net/ethernet/amazon/ena/ena_eth_com.c sq_desc = bounce_buffer + pkt_ctrl->idx * io_sq->desc_entry_size; idx 159 drivers/net/ethernet/amazon/ena/ena_eth_com.c pkt_ctrl->idx++; idx 175 drivers/net/ethernet/amazon/ena/ena_eth_com.c if (pkt_ctrl->idx) { idx 187 drivers/net/ethernet/amazon/ena/ena_eth_com.c pkt_ctrl->idx = 0; idx 217 drivers/net/ethernet/amazon/ena/ena_eth_com.c pkt_ctrl->idx = 0; idx 243 drivers/net/ethernet/amazon/ena/ena_eth_com.c ena_com_rx_cdesc_idx_to_ptr(struct ena_com_io_cq *io_cq, u16 idx) idx 245 drivers/net/ethernet/amazon/ena/ena_eth_com.c idx &= (io_cq->q_depth - 1); idx 248 drivers/net/ethernet/amazon/ena/ena_eth_com.c idx * io_cq->cdesc_entry_size_in_bytes); idx 122 drivers/net/ethernet/amazon/ena/ena_netdev.h #define ENA_TX_RING_IDX_NEXT(idx, ring_size) (((idx) + 1) & ((ring_size) - 1)) idx 124 drivers/net/ethernet/amazon/ena/ena_netdev.h #define ENA_RX_RING_IDX_NEXT(idx, ring_size) (((idx) + 1) & ((ring_size) - 1)) idx 125 drivers/net/ethernet/amazon/ena/ena_netdev.h #define ENA_RX_RING_IDX_ADD(idx, n, ring_size) \ idx 126 drivers/net/ethernet/amazon/ena/ena_netdev.h (((idx) + (n)) & ((ring_size) - 1)) idx 198 drivers/net/ethernet/amd/am79c961a.c int idx, bit; idx 203 drivers/net/ethernet/amd/am79c961a.c idx = crc >> 30; idx 206 drivers/net/ethernet/amd/am79c961a.c hash[idx] |= 1 << bit; idx 2943 drivers/net/ethernet/amd/xgbe/xgbe-drv.c unsigned int idx, unsigned int count, unsigned int flag) idx 2949 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata = XGBE_GET_DESC_DATA(ring, idx); idx 2952 drivers/net/ethernet/amd/xgbe/xgbe-drv.c "TX_NORMAL_DESC[%d %s] = %08x:%08x:%08x:%08x\n", idx, idx 2958 drivers/net/ethernet/amd/xgbe/xgbe-drv.c idx++; idx 2963 drivers/net/ethernet/amd/xgbe/xgbe-drv.c unsigned int idx) idx 2968 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata = XGBE_GET_DESC_DATA(ring, idx); idx 2972 drivers/net/ethernet/amd/xgbe/xgbe-drv.c idx, le32_to_cpu(rdesc->desc0), le32_to_cpu(rdesc->desc1), idx 646 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c u32 idx = CLE_PKTRAM_SIZE / sizeof(u32); idx 663 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c ret = xgene_cle_dram_wr(cle, &sband, 1, idx, PKT_RAM, CLE_CMD_WR); idx 679 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c ret = xgene_cle_dram_wr(cle, &sband, 1, idx + 1, PKT_RAM, CLE_CMD_WR); idx 699 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c u32 fpsel, dstqid, nfpsel, idt_reg, idx; idx 704 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c idx = i % pdata->rxq_cnt; idx 705 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c pool_id = pdata->rx_ring[idx]->buf_pool->id; idx 707 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c dstqid = xgene_enet_dst_ring_num(pdata->rx_ring[idx]); idx 709 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c if (pdata->rx_ring[idx]->page_pool) { idx 710 drivers/net/ethernet/apm/xgene/xgene_enet_cle.c pool_id = pdata->rx_ring[idx]->page_pool->id; idx 384 drivers/net/ethernet/apm/xgene/xgene_enet_main.c static void xgene_set_addr_len(__le64 *desc, u32 idx, dma_addr_t addr, u32 len) idx 386 drivers/net/ethernet/apm/xgene/xgene_enet_main.c desc[idx ^ 1] = cpu_to_le64(SET_VAL(DATAADDR, addr) | idx 418 drivers/net/ethernet/apm/xgene/xgene_enet_main.c u8 ll = 0, nv = 0, idx = 0; idx 502 drivers/net/ethernet/apm/xgene/xgene_enet_main.c xgene_set_addr_len(exp_bufs, idx, dma_addr, idx 504 drivers/net/ethernet/apm/xgene/xgene_enet_main.c idx++; idx 512 drivers/net/ethernet/apm/xgene/xgene_enet_main.c xgene_set_addr_len(exp_bufs, idx, dma_addr, hw_len); idx 513 drivers/net/ethernet/apm/xgene/xgene_enet_main.c idx++; idx 523 drivers/net/ethernet/apm/xgene/xgene_enet_main.c if (idx) { idx 535 drivers/net/ethernet/apm/xgene/xgene_enet_main.c SET_VAL(LL_LEN, idx)); idx 301 drivers/net/ethernet/aquantia/atlantic/aq_nic.c void aq_nic_set_tx_ring(struct aq_nic_s *self, unsigned int idx, idx 304 drivers/net/ethernet/aquantia/atlantic/aq_nic.c self->aq_ring_tx[idx] = ring; idx 593 drivers/net/ethernet/aquantia/atlantic/aq_nic.c if (__netif_subqueue_stopped(self->ndev, ring->idx)) { idx 121 drivers/net/ethernet/aquantia/atlantic/aq_nic.h void aq_nic_set_tx_ring(struct aq_nic_s *self, unsigned int idx, idx 125 drivers/net/ethernet/aquantia/atlantic/aq_ring.c unsigned int idx, idx 131 drivers/net/ethernet/aquantia/atlantic/aq_ring.c self->idx = idx; idx 151 drivers/net/ethernet/aquantia/atlantic/aq_ring.c unsigned int idx, idx 157 drivers/net/ethernet/aquantia/atlantic/aq_ring.c self->idx = idx; idx 206 drivers/net/ethernet/aquantia/atlantic/aq_ring.c if (__netif_subqueue_stopped(ndev, ring->idx)) { idx 207 drivers/net/ethernet/aquantia/atlantic/aq_ring.c netif_wake_subqueue(ndev, ring->idx); idx 216 drivers/net/ethernet/aquantia/atlantic/aq_ring.c if (!__netif_subqueue_stopped(ndev, ring->idx)) idx 217 drivers/net/ethernet/aquantia/atlantic/aq_ring.c netif_stop_subqueue(ndev, ring->idx); idx 428 drivers/net/ethernet/aquantia/atlantic/aq_ring.c skb_record_rx_queue(skb, self->idx); idx 115 drivers/net/ethernet/aquantia/atlantic/aq_ring.h unsigned int idx; /* for HW layer registers operations */ idx 158 drivers/net/ethernet/aquantia/atlantic/aq_ring.h unsigned int idx, idx 162 drivers/net/ethernet/aquantia/atlantic/aq_ring.h unsigned int idx, idx 103 drivers/net/ethernet/aquantia/atlantic/aq_vec.c struct aq_vec_s *aq_vec_alloc(struct aq_nic_s *aq_nic, unsigned int idx, idx 118 drivers/net/ethernet/aquantia/atlantic/aq_vec.c self->aq_ring_param.vec_idx = idx; idx 120 drivers/net/ethernet/aquantia/atlantic/aq_vec.c idx + aq_nic_cfg->aq_rss.base_cpu_number; idx 26 drivers/net/ethernet/aquantia/atlantic/aq_vec.h struct aq_vec_s *aq_vec_alloc(struct aq_nic_s *aq_nic, unsigned int idx, idx 406 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_tdm_tx_desc_en_set(self, 1, ring->idx); idx 413 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_en_set(self, 1, ring->idx); idx 427 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_reg_tx_dma_desc_tail_ptr_set(self, ring->sw_tail, ring->idx); idx 513 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_en_set(self, false, aq_ring->idx); idx 515 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_head_splitting_set(self, 0U, aq_ring->idx); idx 518 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c aq_ring->idx); idx 522 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c aq_ring->idx); idx 524 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_len_set(self, aq_ring->size / 8U, aq_ring->idx); idx 528 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c aq_ring->idx); idx 530 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_head_buff_size_set(self, 0U, aq_ring->idx); idx 531 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_head_splitting_set(self, 0U, aq_ring->idx); idx 532 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rpo_rx_desc_vlan_stripping_set(self, 0U, aq_ring->idx); idx 537 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_itr_irq_map_rx_set(self, aq_ring_param->vec_idx, aq_ring->idx); idx 538 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_itr_irq_map_en_rx_set(self, true, aq_ring->idx); idx 540 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_cpu_id_set(self, aq_ring_param->cpu, aq_ring->idx); idx 541 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_dca_en_set(self, 0U, aq_ring->idx); idx 542 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_head_dca_en_set(self, 0U, aq_ring->idx); idx 543 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_pld_dca_en_set(self, 0U, aq_ring->idx); idx 556 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c aq_ring->idx); idx 559 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c aq_ring->idx); idx 561 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_tdm_tx_desc_len_set(self, aq_ring->size / 8U, aq_ring->idx); idx 566 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_tdm_tx_desc_wr_wb_threshold_set(self, 0U, aq_ring->idx); idx 569 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_itr_irq_map_tx_set(self, aq_ring_param->vec_idx, aq_ring->idx); idx 570 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_itr_irq_map_en_tx_set(self, true, aq_ring->idx); idx 572 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_tdm_cpu_id_set(self, aq_ring_param->cpu, aq_ring->idx); idx 573 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_tdm_tx_desc_dca_en_set(self, 0U, aq_ring->idx); idx 594 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_reg_rx_dma_desc_tail_ptr_set(self, sw_tail_old, ring->idx); idx 603 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c unsigned int hw_head = hw_atl_tdm_tx_desc_head_ptr_get(self, ring->idx); idx 631 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_reg_rx_dma_desc_status_get(self, ring->idx)) { idx 632 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_en_set(self, false, ring->idx); idx 633 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_res_set(self, true, ring->idx); idx 634 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_res_set(self, false, ring->idx); idx 635 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_en_set(self, true, ring->idx); idx 640 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c ring->idx) < 2U)) { idx 868 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_tdm_tx_desc_en_set(self, 0U, ring->idx); idx 875 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c hw_atl_rdm_rx_desc_en_set(self, 0U, ring->idx); idx 461 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_tdm_tx_desc_en_set(self, 1, ring->idx); idx 468 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_en_set(self, 1, ring->idx); idx 482 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_reg_tx_dma_desc_tail_ptr_set(self, ring->sw_tail, ring->idx); idx 581 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_en_set(self, false, aq_ring->idx); idx 583 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_head_splitting_set(self, 0U, aq_ring->idx); idx 586 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c aq_ring->idx); idx 589 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c dma_desc_addr_msw, aq_ring->idx); idx 591 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_len_set(self, aq_ring->size / 8U, aq_ring->idx); idx 595 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c aq_ring->idx); idx 597 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_head_buff_size_set(self, 0U, aq_ring->idx); idx 598 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_head_splitting_set(self, 0U, aq_ring->idx); idx 600 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c aq_ring->idx); idx 605 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_itr_irq_map_rx_set(self, aq_ring_param->vec_idx, aq_ring->idx); idx 606 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_itr_irq_map_en_rx_set(self, true, aq_ring->idx); idx 608 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_cpu_id_set(self, aq_ring_param->cpu, aq_ring->idx); idx 609 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_dca_en_set(self, 0U, aq_ring->idx); idx 610 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_head_dca_en_set(self, 0U, aq_ring->idx); idx 611 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_pld_dca_en_set(self, 0U, aq_ring->idx); idx 624 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c aq_ring->idx); idx 627 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c aq_ring->idx); idx 629 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_tdm_tx_desc_len_set(self, aq_ring->size / 8U, aq_ring->idx); idx 634 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_tdm_tx_desc_wr_wb_threshold_set(self, 0U, aq_ring->idx); idx 637 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_itr_irq_map_tx_set(self, aq_ring_param->vec_idx, aq_ring->idx); idx 638 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_itr_irq_map_en_tx_set(self, true, aq_ring->idx); idx 640 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_tdm_cpu_id_set(self, aq_ring_param->cpu, aq_ring->idx); idx 641 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_tdm_tx_desc_dca_en_set(self, 0U, aq_ring->idx); idx 662 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_reg_rx_dma_desc_tail_ptr_set(self, sw_tail_old, ring->idx); idx 671 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c unsigned int hw_head_ = hw_atl_tdm_tx_desc_head_ptr_get(self, ring->idx); idx 997 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_tdm_tx_desc_en_set(self, 0U, ring->idx); idx 1004 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c hw_atl_rdm_rx_desc_en_set(self, 0U, ring->idx); idx 337 drivers/net/ethernet/atheros/ag71xx.c static struct ag71xx_desc *ag71xx_ring_desc(struct ag71xx_ring *ring, int idx) idx 339 drivers/net/ethernet/atheros/ag71xx.c return (struct ag71xx_desc *)&ring->descs_cpu[idx * AG71XX_DESC_SIZE]; idx 811 drivers/net/ethernet/atheros/alx/main.c int i, vector, idx, shift; idx 816 drivers/net/ethernet/atheros/alx/main.c idx = txq_vec_mapping_shift[i * 2]; idx 818 drivers/net/ethernet/atheros/alx/main.c tbl[idx] |= vector << shift; idx 3767 drivers/net/ethernet/broadcom/bnx2.c rv2p_fw_fixup(u32 rv2p_proc, int idx, u32 loc, u32 rv2p_code) idx 3769 drivers/net/ethernet/broadcom/bnx2.c switch (idx) { idx 450 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define BIT_VEC64_SET_BIT(vec64, idx) \ idx 451 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h __BIT_VEC64_SET_BIT((vec64)[(idx) >> BIT_VEC64_ELEM_SHIFT], \ idx 452 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h (idx) & BIT_VEC64_ELEM_MASK) idx 454 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define BIT_VEC64_CLEAR_BIT(vec64, idx) \ idx 455 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h __BIT_VEC64_CLEAR_BIT((vec64)[(idx) >> BIT_VEC64_ELEM_SHIFT], \ idx 456 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h (idx) & BIT_VEC64_ELEM_MASK) idx 458 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define BIT_VEC64_TEST_BIT(vec64, idx) \ idx 459 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h (((vec64)[(idx) >> BIT_VEC64_ELEM_SHIFT] >> \ idx 460 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h ((idx) & BIT_VEC64_ELEM_MASK)) & 0x1) idx 464 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define BIT_VEC64_ONES_MASK(idx) \ idx 465 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h (((u64)0x1 << (((idx) & BIT_VEC64_ELEM_MASK) + 1)) - 1) idx 637 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define IS_FCOE_IDX(idx) ((idx) == FCOE_IDX(bp)) idx 1967 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define skip_rx_queue(bp, idx) (NO_FCOE(bp) && IS_FCOE_IDX(idx)) idx 1972 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define skip_tx_queue(bp, idx) (NO_FCOE(bp) && IS_FCOE_IDX(idx)) idx 1974 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h #define skip_queue(bp, idx) (NO_FCOE(bp) && IS_FCOE_IDX(idx)) idx 2036 drivers/net/ethernet/broadcom/bnx2x/bnx2x.h void bnx2x_post_dmae(struct bnx2x *bp, struct dmae_command *dmae, int idx); idx 195 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c u16 idx, unsigned int *pkts_compl, idx 198 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c struct sw_tx_bd *tx_buf = &txdata->tx_buf_ring[idx]; idx 210 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c txdata->txq_index, idx, tx_buf, skb); idx 346 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c u16 idx) idx 350 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (SUB_S16(idx, last_max) > 0) idx 351 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c fp->last_max_sge = idx; idx 499 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h int bnx2x_set_vf_spoofchk(struct net_device *dev, int idx, bool val); idx 863 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h int idx = RX_SGE_CNT * i - 1; idx 866 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h BIT_VEC64_CLEAR_BIT(fp->sge_mask, idx); idx 867 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h idx--; idx 741 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c #define IS_REG_IN_PRESET(presets, idx) \ idx 742 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c ((presets & (1 << (idx-1))) == (1 << (idx-1))) idx 2232 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c int idx, i, rc = -ENODEV; idx 2341 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c for (idx = 0; idx < 2; idx++) { idx 2343 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c switch (idx) { idx 693 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h static inline u32 bnx2x_parity_reg_mask(struct bnx2x *bp, int idx) idx 696 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h return bnx2x_blocks_parity_data[idx].reg_mask.e1; idx 698 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h return bnx2x_blocks_parity_data[idx].reg_mask.e1h; idx 700 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h return bnx2x_blocks_parity_data[idx].reg_mask.e2; idx 702 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h return bnx2x_blocks_parity_data[idx].reg_mask.e3; idx 690 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h struct bnx2x_ilt *ilt, int idx, u8 initop) idx 693 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h int abs_idx = ilt->start_line + idx; idx 700 drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h bnx2x_ilt_line_wr(bp, abs_idx, ilt->lines[idx].page_mapping); idx 3056 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c int idx; idx 3076 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < I2C_SWITCH_WIDTH; idx++) idx 3077 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c bnx2x_set_cfg_pin(bp, i2c_pins[idx], i2c_val[idx]); idx 8200 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c int idx, cfg_idx = 0; idx 8202 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = INT_PHY; idx < MAX_PHYS; idx++) { idx 8203 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c if (params->phy[idx].type == phy->type) { idx 8204 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c cfg_idx = LINK_CONFIG_IDX(idx); idx 9974 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c int idx; idx 9985 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < PHY848xx_CMDHDLR_WAIT; idx++) { idx 9993 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c if (idx >= PHY848xx_CMDHDLR_WAIT) { idx 10002 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < argc; idx++) { idx 10004 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c MDIO_848xx_CMD_HDLR_DATA1 + idx, idx 10005 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c cmd_args[idx]); idx 10019 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < PHY848xx_CMDHDLR_WAIT; idx++) { idx 10027 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c if ((idx >= PHY848xx_CMDHDLR_WAIT) || idx 10037 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < argc; idx++) { idx 10039 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c MDIO_848xx_CMD_HDLR_DATA1 + idx, idx 10040 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c &cmd_args[idx]); idx 10050 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c int idx; idx 10062 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < PHY848xx_CMDHDLR_WAIT; idx++) { idx 10069 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c if (idx >= PHY848xx_CMDHDLR_WAIT) { idx 10085 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < argc; idx++) { idx 10087 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c MDIO_848xx_CMD_HDLR_DATA1 + idx, idx 10088 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c cmd_args[idx]); idx 10094 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < PHY848xx_CMDHDLR_WAIT; idx++) { idx 10102 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c if ((idx >= PHY848xx_CMDHDLR_WAIT) || idx 10109 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < argc; idx++) { idx 10111 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c MDIO_848xx_CMD_HDLR_DATA1 + idx, idx 10112 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c &cmd_args[idx]); idx 10180 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c u32 idx; idx 10184 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < 2; idx++) { idx 10186 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] = REG_RD(bp, shmem_base_path[idx] + idx 10189 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] = (reset_pin[idx] & idx 10192 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] -= PIN_CFG_GPIO0_P0; idx 10193 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] = (1 << reset_pin[idx]); idx 10198 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c for (idx = 0; idx < 2; idx++) { idx 10199 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] = REG_RD(bp, shmem_base_path[idx] + idx 10202 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] &= PORT_HW_CFG_EXT_PHY_GPIO_RST_MASK; idx 10203 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] -= PORT_HW_CFG_EXT_PHY_GPIO_RST_GPIO0_P0; idx 10204 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] >>= PORT_HW_CFG_EXT_PHY_GPIO_RST_SHIFT; idx 10205 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c reset_pin[idx] = (1 << reset_pin[idx]); idx 456 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c void bnx2x_post_dmae(struct bnx2x *bp, struct dmae_command *dmae, int idx) idx 461 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c cmd_offset = (DMAE_REG_CMD_MEM + sizeof(struct dmae_command) * idx); idx 465 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c REG_WR(bp, dmae_reg_go_c[idx], 1); idx 3668 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int idx = BP_FW_MB_IDX(bp); idx 3701 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c SHMEM2_WR(bp, func_os_drv_ver[idx].versions[DRV_PERS_ETHERNET], ethver); idx 3702 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c SHMEM2_WR(bp, func_os_drv_ver[idx].versions[DRV_PERS_ISCSI], iscsiver); idx 3703 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c SHMEM2_WR(bp, func_os_drv_ver[idx].versions[DRV_PERS_FCOE], fcoever); idx 4576 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c static void _print_next_block(int idx, const char *blk) idx 4578 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c pr_cont("%s%s", idx ? ", " : "", blk); idx 11252 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int cfg_size = 0, idx, port = BP_PORT(bp); idx 11312 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c for (idx = 0; idx < cfg_size; idx++) { idx 11313 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11315 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~SUPPORTED_10baseT_Half; idx 11317 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11319 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~SUPPORTED_10baseT_Full; idx 11321 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11323 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~SUPPORTED_100baseT_Half; idx 11325 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11327 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~SUPPORTED_100baseT_Full; idx 11329 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11331 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~(SUPPORTED_1000baseT_Half | idx 11334 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11336 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~SUPPORTED_2500baseX_Full; idx 11338 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11340 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~SUPPORTED_10000baseT_Full; idx 11342 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->link_params.speed_cap_mask[idx] & idx 11344 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx] &= ~SUPPORTED_20000baseKR2_Full; idx 11353 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c u32 link_config, idx, cfg_size = 0; idx 11365 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c for (idx = 0; idx < cfg_size; idx++) { idx 11366 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_duplex[idx] = DUPLEX_FULL; idx 11367 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c link_config = bp->port.link_config[idx]; idx 11370 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & SUPPORTED_Autoneg) { idx 11371 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11373 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11374 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx]; idx 11377 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11382 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11384 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11392 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & SUPPORTED_10baseT_Full) { idx 11393 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11395 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11401 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.speed_cap_mask[idx]); idx 11407 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & SUPPORTED_10baseT_Half) { idx 11408 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11410 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_duplex[idx] = idx 11412 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11418 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.speed_cap_mask[idx]); idx 11424 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & idx 11426 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11428 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11434 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.speed_cap_mask[idx]); idx 11440 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & idx 11442 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11444 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_duplex[idx] = idx 11446 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11452 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.speed_cap_mask[idx]); idx 11458 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & idx 11460 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11462 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11465 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c } else if (bp->port.supported[idx] & idx 11467 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11469 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11474 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.speed_cap_mask[idx]); idx 11480 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & idx 11482 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11484 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11490 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.speed_cap_mask[idx]); idx 11496 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->port.supported[idx] & idx 11498 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11500 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11503 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c } else if (bp->port.supported[idx] & idx 11505 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11507 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] |= idx 11513 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.speed_cap_mask[idx]); idx 11518 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = SPEED_20000; idx 11524 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx] = idx 11526 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx] = idx 11527 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.supported[idx]; idx 11531 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_flow_ctrl[idx] = (link_config & idx 11533 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (bp->link_params.req_flow_ctrl[idx] == idx 11535 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c if (!(bp->port.supported[idx] & SUPPORTED_Autoneg)) idx 11536 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_flow_ctrl[idx] = idx 11543 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_line_speed[idx], idx 11544 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_duplex[idx], idx 11545 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->link_params.req_flow_ctrl[idx], idx 11546 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c bp->port.advertising[idx]); idx 14872 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int idx = BP_FW_MB_IDX(bp); idx 14873 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c u32 cap = SHMEM2_RD(bp, drv_capabilities_flag[idx]); idx 14885 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c SHMEM2_WR(bp, drv_capabilities_flag[idx], cap); idx 14913 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c int idx = BP_FW_MB_IDX(bp); idx 14916 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c cap = SHMEM2_RD(bp, drv_capabilities_flag[idx]); idx 14921 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c SHMEM2_WR(bp, drv_capabilities_flag[idx], cap); idx 1785 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c int rc, idx = 0; idx 1826 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c o->set_one_rule(bp, o, elem, idx, idx 1831 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c idx += 2; idx 1833 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c idx++; idx 2829 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c struct bnx2x_mcast_obj *o, int idx, idx 2844 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->rules[idx].cmd_general_data |= rx_tx_add_flag; idx 2886 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->rules[idx].bin_id = (u8)bin; idx 2887 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->rules[idx].func_id = func_id; idx 2888 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->rules[idx].engine_id = o->engine_id; idx 3657 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c struct bnx2x_mcast_obj *o, int idx, idx 3667 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c bnx2x_set_fw_mac_addr(&data->config_table[idx].msb_mac_addr, idx 3668 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c &data->config_table[idx].middle_mac_addr, idx 3669 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c &data->config_table[idx].lsb_mac_addr, idx 3672 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->config_table[idx].vlan_id = 0; idx 3673 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->config_table[idx].pf_id = r->func_id; idx 3674 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c data->config_table[idx].clients_bit_vector = idx 3677 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c SET_FLAG(data->config_table[idx].flags, idx 4253 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c int idx, vec, i; idx 4265 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c for (idx = vec * BIT_VEC64_ELEM_SZ, i = 0; idx 4266 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c i < BIT_VEC64_ELEM_SZ; idx++, i++) idx 4268 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c if (BIT_VEC64_TEST_BIT(o->pool_mirror, idx)) { idx 4270 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c BIT_VEC64_CLEAR_BIT(o->pool_mirror, idx); idx 4271 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c *offset = o->base_pool_offset + idx; idx 616 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h struct bnx2x_mcast_obj *o, int idx, idx 63 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int idx; idx 65 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c for_each_vf(bp, idx) idx 66 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (bnx2x_vf(bp, idx, abs_vfid) == abs_vfid) idx 68 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c return idx; idx 74 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c u16 idx = (u16)bnx2x_vf_idx_by_abs_fid(bp, abs_vfid); idx 75 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c return (idx < BNX2X_NR_VIRTFN(bp)) ? BP_VF(bp, idx) : NULL; idx 1457 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int bnx2x_iov_link_update_vf(struct bnx2x *bp, int idx) idx 1466 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_op_prep(bp, idx, &vf, &bulletin, false); idx 1504 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c "vf %d mode %u speed %d flags %x\n", idx, idx 1508 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_post_vf_bulletin(bp, idx); idx 1510 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c BNX2X_ERR("failed to update VF[%d] bulletin\n", idx); idx 1520 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int bnx2x_set_vf_link_state(struct net_device *dev, int idx, int link_state) idx 1523 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct bnx2x_virtf *vf = BP_VF(bp, idx); idx 1533 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c return bnx2x_iov_link_update_vf(bp, idx); idx 2963 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int bnx2x_set_vf_spoofchk(struct net_device *dev, int idx, bool val) idx 2969 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c vf = BP_VF(bp, idx); idx 2980 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c val ? "enabling" : "disabling", idx); idx 3021 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c val ? "enable" : "disable", idx, i); idx 3029 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c idx); idx 321 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h #define BP_VF(bp, idx) ((BP_VFDB(bp) && (bp)->vfdb->vfs) ? \ idx 322 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h &((bp)->vfdb->vfs[idx]) : NULL) idx 323 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h #define bnx2x_vf(bp, idx, var) ((bp)->vfdb->vfs[idx].var) idx 552 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h int bnx2x_iov_link_update_vf(struct bnx2x *bp, int idx); idx 622 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h static inline int bnx2x_iov_link_update_vf(struct bnx2x *bp, int idx) {return 0; } idx 1698 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c static void bnx2x_vf_mbx_dp_q_filter(struct bnx2x *bp, int msglvl, int idx, idx 1701 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c DP(msglvl, "MAC-VLAN[%d] -- flags=0x%x\n", idx, filter->flags); idx 263 drivers/net/ethernet/broadcom/bnxt/bnxt.c static bool bnxt_vf_pciid(enum board_idx idx) idx 265 drivers/net/ethernet/broadcom/bnxt/bnxt.c return (idx == NETXTREME_C_VF || idx == NETXTREME_E_VF || idx 266 drivers/net/ethernet/broadcom/bnxt/bnxt.c idx == NETXTREME_S_VF || idx == NETXTREME_E_P5_VF); idx 276 drivers/net/ethernet/broadcom/bnxt/bnxt.c #define BNXT_DB_CQ(db, idx) \ idx 277 drivers/net/ethernet/broadcom/bnxt/bnxt.c writel(DB_CP_FLAGS | RING_CMP(idx), (db)->doorbell) idx 279 drivers/net/ethernet/broadcom/bnxt/bnxt.c #define BNXT_DB_NQ_P5(db, idx) \ idx 280 drivers/net/ethernet/broadcom/bnxt/bnxt.c writeq((db)->db_key64 | DBR_TYPE_NQ | RING_CMP(idx), (db)->doorbell) idx 282 drivers/net/ethernet/broadcom/bnxt/bnxt.c #define BNXT_DB_CQ_ARM(db, idx) \ idx 283 drivers/net/ethernet/broadcom/bnxt/bnxt.c writel(DB_CP_REARM_FLAGS | RING_CMP(idx), (db)->doorbell) idx 285 drivers/net/ethernet/broadcom/bnxt/bnxt.c #define BNXT_DB_NQ_ARM_P5(db, idx) \ idx 286 drivers/net/ethernet/broadcom/bnxt/bnxt.c writeq((db)->db_key64 | DBR_TYPE_NQ_ARM | RING_CMP(idx), (db)->doorbell) idx 288 drivers/net/ethernet/broadcom/bnxt/bnxt.c static void bnxt_db_nq(struct bnxt *bp, struct bnxt_db_info *db, u32 idx) idx 291 drivers/net/ethernet/broadcom/bnxt/bnxt.c BNXT_DB_NQ_P5(db, idx); idx 293 drivers/net/ethernet/broadcom/bnxt/bnxt.c BNXT_DB_CQ(db, idx); idx 296 drivers/net/ethernet/broadcom/bnxt/bnxt.c static void bnxt_db_nq_arm(struct bnxt *bp, struct bnxt_db_info *db, u32 idx) idx 299 drivers/net/ethernet/broadcom/bnxt/bnxt.c BNXT_DB_NQ_ARM_P5(db, idx); idx 301 drivers/net/ethernet/broadcom/bnxt/bnxt.c BNXT_DB_CQ_ARM(db, idx); idx 304 drivers/net/ethernet/broadcom/bnxt/bnxt.c static void bnxt_db_cq(struct bnxt *bp, struct bnxt_db_info *db, u32 idx) idx 307 drivers/net/ethernet/broadcom/bnxt/bnxt.c writeq(db->db_key64 | DBR_TYPE_CQ_ARMALL | RING_CMP(idx), idx 310 drivers/net/ethernet/broadcom/bnxt/bnxt.c BNXT_DB_CQ(db, idx); idx 778 drivers/net/ethernet/broadcom/bnxt/bnxt.c static inline u16 bnxt_find_next_agg_idx(struct bnxt_rx_ring_info *rxr, u16 idx) idx 782 drivers/net/ethernet/broadcom/bnxt/bnxt.c next = find_next_zero_bit(rxr->rx_agg_bmap, max, idx); idx 866 drivers/net/ethernet/broadcom/bnxt/bnxt.c static void bnxt_reuse_rx_agg_bufs(struct bnxt_cp_ring_info *cpr, u16 idx, idx 888 drivers/net/ethernet/broadcom/bnxt/bnxt.c agg = bnxt_get_tpa_agg_p5(bp, rxr, idx, start + i); idx 890 drivers/net/ethernet/broadcom/bnxt/bnxt.c agg = bnxt_get_agg(bp, cpr, idx, start + i); idx 1001 drivers/net/ethernet/broadcom/bnxt/bnxt.c struct sk_buff *skb, u16 idx, idx 1022 drivers/net/ethernet/broadcom/bnxt/bnxt.c agg = bnxt_get_tpa_agg_p5(bp, rxr, idx, i); idx 1024 drivers/net/ethernet/broadcom/bnxt/bnxt.c agg = bnxt_get_agg(bp, cpr, idx, i); idx 1058 drivers/net/ethernet/broadcom/bnxt/bnxt.c bnxt_reuse_rx_agg_bufs(cpr, idx, i, agg_bufs - i, tpa); idx 1181 drivers/net/ethernet/broadcom/bnxt/bnxt.c u16 idx = agg_id & MAX_TPA_P5_MASK; idx 1183 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (test_bit(idx, map->agg_idx_bmap)) idx 1184 drivers/net/ethernet/broadcom/bnxt/bnxt.c idx = find_first_zero_bit(map->agg_idx_bmap, idx 1186 drivers/net/ethernet/broadcom/bnxt/bnxt.c __set_bit(idx, map->agg_idx_bmap); idx 1187 drivers/net/ethernet/broadcom/bnxt/bnxt.c map->agg_id_tbl[agg_id] = idx; idx 1188 drivers/net/ethernet/broadcom/bnxt/bnxt.c return idx; idx 1191 drivers/net/ethernet/broadcom/bnxt/bnxt.c static void bnxt_free_agg_idx(struct bnxt_rx_ring_info *rxr, u16 idx) idx 1195 drivers/net/ethernet/broadcom/bnxt/bnxt.c __clear_bit(idx, map->agg_idx_bmap); idx 1288 drivers/net/ethernet/broadcom/bnxt/bnxt.c static void bnxt_abort_tpa(struct bnxt_cp_ring_info *cpr, u16 idx, u32 agg_bufs) idx 1291 drivers/net/ethernet/broadcom/bnxt/bnxt.c bnxt_reuse_rx_agg_bufs(cpr, idx, 0, agg_bufs, true); idx 1515 drivers/net/ethernet/broadcom/bnxt/bnxt.c u16 idx = 0, agg_id; idx 1540 drivers/net/ethernet/broadcom/bnxt/bnxt.c idx = agg_id; idx 1546 drivers/net/ethernet/broadcom/bnxt/bnxt.c idx = RING_CMP(*raw_cons); idx 1552 drivers/net/ethernet/broadcom/bnxt/bnxt.c idx = NEXT_CMP(idx); idx 1563 drivers/net/ethernet/broadcom/bnxt/bnxt.c bnxt_abort_tpa(cpr, idx, agg_bufs); idx 1573 drivers/net/ethernet/broadcom/bnxt/bnxt.c bnxt_abort_tpa(cpr, idx, agg_bufs); idx 1582 drivers/net/ethernet/broadcom/bnxt/bnxt.c bnxt_abort_tpa(cpr, idx, agg_bufs); idx 1597 drivers/net/ethernet/broadcom/bnxt/bnxt.c bnxt_abort_tpa(cpr, idx, agg_bufs); idx 1605 drivers/net/ethernet/broadcom/bnxt/bnxt.c skb = bnxt_rx_pages(bp, cpr, skb, idx, agg_bufs, true); idx 2454 drivers/net/ethernet/broadcom/bnxt/bnxt.c u32 idx = le32_to_cpu(nqcmp->cq_handle_low); idx 2457 drivers/net/ethernet/broadcom/bnxt/bnxt.c cpr2 = cpr->cp_ring_arr[idx]; idx 4465 drivers/net/ethernet/broadcom/bnxt/bnxt.c unsigned int bit, idx; idx 4467 drivers/net/ethernet/broadcom/bnxt/bnxt.c idx = cmd / 32; idx 4469 drivers/net/ethernet/broadcom/bnxt/bnxt.c data[idx] |= 1 << bit; idx 4676 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_hwrm_set_vnic_filter(struct bnxt *bp, u16 vnic_id, u16 idx, idx 4704 drivers/net/ethernet/broadcom/bnxt/bnxt.c bp->vnic_info[vnic_id].fw_l2_filter_id[idx] = idx 5332 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_hwrm_set_async_event_cr(struct bnxt *bp, int idx) idx 5342 drivers/net/ethernet/broadcom/bnxt/bnxt.c req.async_event_cr = cpu_to_le16(idx); idx 5350 drivers/net/ethernet/broadcom/bnxt/bnxt.c req.async_event_cr = cpu_to_le16(idx); idx 11041 drivers/net/ethernet/broadcom/bnxt/bnxt.c int rc = 0, idx, bit_id, l2_idx = 0; idx 11091 drivers/net/ethernet/broadcom/bnxt/bnxt.c idx = skb_get_hash_raw(skb) & BNXT_NTP_FLTR_HASH_MASK; idx 11092 drivers/net/ethernet/broadcom/bnxt/bnxt.c head = &bp->ntp_fltr_hash_tbl[idx]; idx 634 drivers/net/ethernet/broadcom/bnxt/bnxt.h #define NEXT_RX(idx) (((idx) + 1) & bp->rx_ring_mask) idx 636 drivers/net/ethernet/broadcom/bnxt/bnxt.h #define NEXT_RX_AGG(idx) (((idx) + 1) & bp->rx_agg_ring_mask) idx 638 drivers/net/ethernet/broadcom/bnxt/bnxt.h #define NEXT_TX(idx) (((idx) + 1) & bp->tx_ring_mask) idx 640 drivers/net/ethernet/broadcom/bnxt/bnxt.h #define ADV_RAW_CMP(idx, n) ((idx) + (n)) idx 641 drivers/net/ethernet/broadcom/bnxt/bnxt.h #define NEXT_RAW_CMP(idx) ADV_RAW_CMP(idx, 1) idx 642 drivers/net/ethernet/broadcom/bnxt/bnxt.h #define RING_CMP(idx) ((idx) & bp->cp_ring_mask) idx 643 drivers/net/ethernet/broadcom/bnxt/bnxt.h #define NEXT_CMP(idx) RING_CMP(ADV_RAW_CMP(idx, 1)) idx 1872 drivers/net/ethernet/broadcom/bnxt/bnxt.h struct bnxt_db_info *db, u32 idx) idx 1875 drivers/net/ethernet/broadcom/bnxt/bnxt.h writeq_relaxed(db->db_key64 | idx, db->doorbell); idx 1877 drivers/net/ethernet/broadcom/bnxt/bnxt.h u32 db_val = db->db_key32 | idx; idx 1887 drivers/net/ethernet/broadcom/bnxt/bnxt.h u32 idx) idx 1890 drivers/net/ethernet/broadcom/bnxt/bnxt.h writeq(db->db_key64 | idx, db->doorbell); idx 1892 drivers/net/ethernet/broadcom/bnxt/bnxt.h u32 db_val = db->db_key32 | idx; idx 299 drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c int idx = 0, rc, i; idx 316 drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c idx = bp->pf.port_id; idx 318 drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c idx = bp->pf.fw_fid - BNXT_FIRST_PF_FID; idx 328 drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c req->index_0 = cpu_to_le16(idx); idx 329 drivers/net/ethernet/broadcom/bnxt/bnxt_devlink.c if (idx) idx 100 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c int num_msix, idx, i; idx 103 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c idx = edev->ulp_tbl[BNXT_ROCE_ULP].msix_base; idx 105 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c ent[i].vector = bp->irq_tbl[idx + i].vector; idx 106 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c ent[i].ring_idx = idx + i; idx 107 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c ent[i].db_offset = (idx + i) * 0x80; idx 118 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c int avail_msix, idx; idx 140 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c idx = bp->cp_nr_rings; idx 143 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c idx = max_idx - avail_msix; idx 145 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c edev->ulp_tbl[ulp_id].msix_base = idx; idx 148 drivers/net/ethernet/broadcom/bnxt/bnxt_ulp.c total_vecs = idx + avail_msix; idx 958 drivers/net/ethernet/broadcom/cnic.c static u16 cnic_bnx2_next_idx(u16 idx) idx 960 drivers/net/ethernet/broadcom/cnic.c return idx + 1; idx 963 drivers/net/ethernet/broadcom/cnic.c static u16 cnic_bnx2_hw_idx(u16 idx) idx 965 drivers/net/ethernet/broadcom/cnic.c return idx; idx 968 drivers/net/ethernet/broadcom/cnic.c static u16 cnic_bnx2x_next_idx(u16 idx) idx 970 drivers/net/ethernet/broadcom/cnic.c idx++; idx 971 drivers/net/ethernet/broadcom/cnic.c if ((idx & MAX_KCQE_CNT) == MAX_KCQE_CNT) idx 972 drivers/net/ethernet/broadcom/cnic.c idx++; idx 974 drivers/net/ethernet/broadcom/cnic.c return idx; idx 977 drivers/net/ethernet/broadcom/cnic.c static u16 cnic_bnx2x_hw_idx(u16 idx) idx 979 drivers/net/ethernet/broadcom/cnic.c if ((idx & MAX_KCQE_CNT) == MAX_KCQE_CNT) idx 980 drivers/net/ethernet/broadcom/cnic.c idx++; idx 981 drivers/net/ethernet/broadcom/cnic.c return idx; idx 3101 drivers/net/ethernet/broadcom/cnic.c static void cnic_arm_bnx2x_msix(struct cnic_dev *dev, u32 idx) idx 3105 drivers/net/ethernet/broadcom/cnic.c cnic_ack_bnx2x_int(dev, cp->bnx2x_igu_sb_id, CSTORM_ID, idx, idx 3109 drivers/net/ethernet/broadcom/cnic.c static void cnic_arm_bnx2x_e2_msix(struct cnic_dev *dev, u32 idx) idx 3113 drivers/net/ethernet/broadcom/cnic.c cnic_ack_igu_sb(dev, cp->bnx2x_igu_sb_id, IGU_SEG_ACCESS_DEF, idx, idx 4393 drivers/net/ethernet/broadcom/cnic.c u32 idx = cp->ctx_arr[i].cid / cp->cids_per_blk; idx 4402 drivers/net/ethernet/broadcom/cnic.c CNIC_WR(dev, BNX2_CTX_HOST_PAGE_TBL_CTRL, idx | idx 425 drivers/net/ethernet/broadcom/sb1250-mac.c int idx; idx 477 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < 16; idx++) { idx 978 drivers/net/ethernet/broadcom/sb1250-mac.c int idx; idx 981 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < d->sbdma_maxdescr; idx++) { idx 982 drivers/net/ethernet/broadcom/sb1250-mac.c sb = d->sbdma_ctxtable[idx]; idx 985 drivers/net/ethernet/broadcom/sb1250-mac.c d->sbdma_ctxtable[idx] = NULL; idx 1007 drivers/net/ethernet/broadcom/sb1250-mac.c int idx; idx 1009 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < SBMAC_MAX_RXDESCR - 1; idx++) { idx 1389 drivers/net/ethernet/broadcom/sb1250-mac.c int idx, th_value; idx 1450 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < MAC_HASH_COUNT; idx++) { idx 1460 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < MAC_ADDR_COUNT; idx++) { idx 1470 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < MAC_CHMAP_COUNT; idx++) { idx 1477 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < MAC_CHMAP_COUNT; idx++) { idx 2062 drivers/net/ethernet/broadcom/sb1250-mac.c int idx; idx 2072 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 1; idx < MAC_ADDR_COUNT; idx++) { idx 2073 drivers/net/ethernet/broadcom/sb1250-mac.c port = sc->sbm_base + R_MAC_ADDR_BASE+(idx*sizeof(uint64_t)); idx 2077 drivers/net/ethernet/broadcom/sb1250-mac.c for (idx = 0; idx < MAC_HASH_COUNT; idx++) { idx 2078 drivers/net/ethernet/broadcom/sb1250-mac.c port = sc->sbm_base + R_MAC_HASH_BASE+(idx*sizeof(uint64_t)); idx 2111 drivers/net/ethernet/broadcom/sb1250-mac.c idx = 1; /* skip station address */ idx 2113 drivers/net/ethernet/broadcom/sb1250-mac.c if (idx == MAC_ADDR_COUNT) idx 2116 drivers/net/ethernet/broadcom/sb1250-mac.c port = sc->sbm_base + R_MAC_ADDR_BASE+(idx * sizeof(uint64_t)); idx 2118 drivers/net/ethernet/broadcom/sb1250-mac.c idx++; idx 2126 drivers/net/ethernet/broadcom/sb1250-mac.c if (idx > 1) { idx 2162 drivers/net/ethernet/broadcom/sb1250-mac.c int idx = pldev->id; idx 2170 drivers/net/ethernet/broadcom/sb1250-mac.c sc->sbe_idx = idx; idx 2210 drivers/net/ethernet/broadcom/sb1250-mac.c dev->irq = UNIT_INT(idx); idx 2223 drivers/net/ethernet/broadcom/sb1250-mac.c pldev->name, idx); idx 2243 drivers/net/ethernet/broadcom/sb1250-mac.c sbmac_string, idx); idx 2247 drivers/net/ethernet/broadcom/sb1250-mac.c pr_info("%s.%d: registered as %s\n", sbmac_string, idx, dev->name); idx 1055 drivers/net/ethernet/broadcom/tg3.c if (sblk->idx[0].tx_consumer != tnapi->tx_cons) idx 6497 drivers/net/ethernet/broadcom/tg3.c tnapi->hw_status->idx[0].rx_producer, idx 6498 drivers/net/ethernet/broadcom/tg3.c tnapi->hw_status->idx[0].tx_consumer); idx 6548 drivers/net/ethernet/broadcom/tg3.c u32 hw_idx = tnapi->hw_status->idx[0].tx_consumer; idx 7178 drivers/net/ethernet/broadcom/tg3.c if (tnapi->hw_status->idx[0].tx_consumer != tnapi->tx_cons) { idx 7260 drivers/net/ethernet/broadcom/tg3.c if (likely(sblk->idx[0].tx_consumer == tnapi->tx_cons && idx 8799 drivers/net/ethernet/broadcom/tg3.c prodptr = &sblk->idx[0].rx_producer; idx 8813 drivers/net/ethernet/broadcom/tg3.c tnapi->rx_rcb_prod_idx = &sblk->idx[0].rx_producer; idx 13533 drivers/net/ethernet/broadcom/tg3.c rx_start_idx = rnapi->hw_status->idx[0].rx_producer; idx 13560 drivers/net/ethernet/broadcom/tg3.c tx_idx = tnapi->hw_status->idx[0].tx_consumer; idx 13561 drivers/net/ethernet/broadcom/tg3.c rx_idx = rnapi->hw_status->idx[0].rx_producer; idx 2709 drivers/net/ethernet/broadcom/tg3.h } idx[16]; idx 193 drivers/net/ethernet/brocade/bna/bfa_msgq.c dbell->idx.cmdq_pi = htons(cmdq->producer_index); idx 234 drivers/net/ethernet/brocade/bna/bfa_msgq.c cmdq->consumer_index = ntohs(dbell->idx.cmdq_ci); idx 451 drivers/net/ethernet/brocade/bna/bfa_msgq.c dbell->idx.rspq_ci = htons(rspq->consumer_index); idx 468 drivers/net/ethernet/brocade/bna/bfa_msgq.c rspq->producer_index = ntohs(dbell->idx.rspq_pi); idx 437 drivers/net/ethernet/brocade/bna/bfi.h } __packed idx; idx 446 drivers/net/ethernet/brocade/bna/bfi.h } __packed idx; idx 185 drivers/net/ethernet/cadence/macb.h #define GEM_IP4SRC_CMP(idx) (idx * 3) idx 186 drivers/net/ethernet/cadence/macb.h #define GEM_IP4DST_CMP(idx) (idx * 3 + 1) idx 187 drivers/net/ethernet/cadence/macb.h #define GEM_PORT_CMP(idx) (idx * 3 + 2) idx 691 drivers/net/ethernet/cadence/macb.h #define gem_readl_n(port, reg, idx) (port)->macb_reg_readl((port), GEM_##reg + idx * 4) idx 692 drivers/net/ethernet/cadence/macb.h #define gem_writel_n(port, reg, idx, value) (port)->macb_reg_writel((port), GEM_##reg + idx * 4, (value)) idx 2507 drivers/net/ethernet/cadence/macb_main.c unsigned int i, q, idx; idx 2527 drivers/net/ethernet/cadence/macb_main.c idx = GEM_STATS_LEN; idx 2530 drivers/net/ethernet/cadence/macb_main.c bp->ethtool_stats[idx++] = *stat; idx 1043 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c u32 idx, int valid) idx 1050 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c oct, CN23XX_PEM_BAR1_INDEX_REG(oct->pcie_port, idx)); idx 1053 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c CN23XX_PEM_BAR1_INDEX_REG(oct->pcie_port, idx)); idx 1055 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c oct, CN23XX_PEM_BAR1_INDEX_REG(oct->pcie_port, idx)); idx 1064 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c CN23XX_PEM_BAR1_INDEX_REG(oct->pcie_port, idx)); idx 1067 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c oct, CN23XX_PEM_BAR1_INDEX_REG(oct->pcie_port, idx))); idx 1070 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c static void cn23xx_bar1_idx_write(struct octeon_device *oct, u32 idx, u32 mask) idx 1073 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c CN23XX_PEM_BAR1_INDEX_REG(oct->pcie_port, idx)); idx 1076 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c static u32 cn23xx_bar1_idx_read(struct octeon_device *oct, u32 idx) idx 1079 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_device.c oct, CN23XX_PEM_BAR1_INDEX_REG(oct->pcie_port, idx)); idx 338 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h #define CN23XX_SLI_PKT_PF_VF_MBOX_SIG(q, idx) \ idx 341 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h (idx) * CN23XX_SLI_MBOX_SIG_IDX_OFFSET)) idx 391 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h #define CN23XX_MSIX_TABLE_ADDR(idx) \ idx 392 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h (CN23XX_MSIX_TABLE_ADDR_START + ((idx) * CN23XX_MSIX_TABLE_SIZE)) idx 394 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h #define CN23XX_MSIX_TABLE_DATA(idx) \ idx 395 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h (CN23XX_MSIX_TABLE_DATA_START + ((idx) * CN23XX_MSIX_TABLE_SIZE)) idx 523 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h #define CN23XX_PEM_BAR1_INDEX_REG(port, idx) \ idx 525 drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h ((idx) << CN23XX_BAR1_INDEX_OFFSET)) idx 217 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_regs.h #define CN23XX_SLI_PKT_PF_VF_MBOX_SIG(q, idx) \ idx 220 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_regs.h (idx) * CN23XX_SLI_MBOX_SIG_IDX_OFFSET)) idx 268 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_regs.h #define CN23XX_MSIX_TABLE_ADDR(idx) \ idx 269 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_regs.h (CN23XX_MSIX_TABLE_ADDR_START + ((idx) * CN23XX_MSIX_TABLE_SIZE)) idx 271 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_regs.h #define CN23XX_MSIX_TABLE_DATA(idx) \ idx 272 drivers/net/ethernet/cavium/liquidio/cn23xx_vf_regs.h (CN23XX_MSIX_TABLE_DATA_START + ((idx) * CN23XX_MSIX_TABLE_SIZE)) idx 414 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c u32 idx, idx 420 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c bar1 = lio_pci_readq(oct, CN6XXX_BAR1_REG(idx, oct->pcie_port)); idx 422 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c CN6XXX_BAR1_REG(idx, oct->pcie_port)); idx 423 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c bar1 = lio_pci_readq(oct, CN6XXX_BAR1_REG(idx, oct->pcie_port)); idx 431 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c CN6XXX_BAR1_REG(idx, oct->pcie_port)); idx 433 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c bar1 = lio_pci_readq(oct, CN6XXX_BAR1_REG(idx, oct->pcie_port)); idx 437 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c u32 idx, idx 440 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c lio_pci_writeq(oct, mask, CN6XXX_BAR1_REG(idx, oct->pcie_port)); idx 443 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c u32 lio_cn6xxx_bar1_idx_read(struct octeon_device *oct, u32 idx) idx 445 drivers/net/ethernet/cavium/liquidio/cn66xx_device.c return (u32)lio_pci_readq(oct, CN6XXX_BAR1_REG(idx, oct->pcie_port)); idx 82 drivers/net/ethernet/cavium/liquidio/cn66xx_device.h u32 idx, int valid); idx 83 drivers/net/ethernet/cavium/liquidio/cn66xx_device.h void lio_cn6xxx_bar1_idx_write(struct octeon_device *oct, u32 idx, u32 mask); idx 84 drivers/net/ethernet/cavium/liquidio/cn66xx_device.h u32 lio_cn6xxx_bar1_idx_read(struct octeon_device *oct, u32 idx); idx 455 drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h #define CN6XXX_BAR1_REG(idx, port) \ idx 457 drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h (CN6XXX_PCI_BAR1_OFFSET * (idx))) idx 1264 drivers/net/ethernet/cavium/liquidio/lio_core.c int idx; idx 1269 drivers/net/ethernet/cavium/liquidio/lio_core.c for (idx = 0; idx < MAX_OCTEON_OUTPUT_QUEUES(oct); idx++) { idx 1270 drivers/net/ethernet/cavium/liquidio/lio_core.c if (!(oct->io_qmask.oq & BIT_ULL(idx))) idx 1273 drivers/net/ethernet/cavium/liquidio/lio_core.c atomic_read(&oct->droq[idx]->pkts_pending); idx 146 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_MAX_TXQS_NIC_IF(cfg, idx) \ idx 147 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].max_txqs) idx 148 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_NUM_TXQS_NIC_IF(cfg, idx) \ idx 149 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].num_txqs) idx 150 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_MAX_RXQS_NIC_IF(cfg, idx) \ idx 151 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].max_rxqs) idx 152 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_NUM_RXQS_NIC_IF(cfg, idx) \ idx 153 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].num_rxqs) idx 154 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_NUM_RX_DESCS_NIC_IF(cfg, idx) \ idx 155 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].num_rx_descs) idx 156 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_NUM_TX_DESCS_NIC_IF(cfg, idx) \ idx 157 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].num_tx_descs) idx 158 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_NUM_RX_BUF_SIZE_NIC_IF(cfg, idx) \ idx 159 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].rx_buf_size) idx 160 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_BASE_QUE_NIC_IF(cfg, idx) \ idx 161 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].base_queue) idx 162 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_GET_GMXID_NIC_IF(cfg, idx) \ idx 163 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].gmx_port_id) idx 172 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_SET_NUM_RX_DESCS_NIC_IF(cfg, idx, value) \ idx 173 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].num_rx_descs = value) idx 174 drivers/net/ethernet/cavium/liquidio/octeon_config.h #define CFG_SET_NUM_TX_DESCS_NIC_IF(cfg, idx, value) \ idx 175 drivers/net/ethernet/cavium/liquidio/octeon_config.h ((cfg)->nic_if_cfg[idx].num_tx_descs = value) idx 763 drivers/net/ethernet/cavium/liquidio/octeon_device.c int idx, refcount; idx 777 drivers/net/ethernet/cavium/liquidio/octeon_device.c for (idx = (int)oct->octeon_id - 1; idx >= 0; idx--) { idx 778 drivers/net/ethernet/cavium/liquidio/octeon_device.c if (!octeon_device[idx]) { idx 789 drivers/net/ethernet/cavium/liquidio/octeon_device.c if ((octeon_device[idx]->loc.bus == bus) && idx 790 drivers/net/ethernet/cavium/liquidio/octeon_device.c (octeon_device[idx]->loc.dev == dev)) { idx 792 drivers/net/ethernet/cavium/liquidio/octeon_device.c octeon_device[idx]->adapter_refcount; idx 794 drivers/net/ethernet/cavium/liquidio/octeon_device.c octeon_device[idx]->adapter_fw_state; idx 1067 drivers/net/ethernet/cavium/liquidio/octeon_device.c u32 idx; idx 1072 drivers/net/ethernet/cavium/liquidio/octeon_device.c idx = combined_opcode & OCTEON_OPCODE_MASK; idx 1081 drivers/net/ethernet/cavium/liquidio/octeon_device.c if (!(octeon_dev->dispatch.dlist[idx].opcode)) { idx 1086 drivers/net/ethernet/cavium/liquidio/octeon_device.c if (octeon_dev->dispatch.dlist[idx].opcode == combined_opcode) { idx 1087 drivers/net/ethernet/cavium/liquidio/octeon_device.c fn = octeon_dev->dispatch.dlist[idx].dispatch_fn; idx 1090 drivers/net/ethernet/cavium/liquidio/octeon_device.c &octeon_dev->dispatch.dlist[idx].list) { idx 1127 drivers/net/ethernet/cavium/liquidio/octeon_device.c u32 idx; idx 1131 drivers/net/ethernet/cavium/liquidio/octeon_device.c idx = combined_opcode & OCTEON_OPCODE_MASK; idx 1135 drivers/net/ethernet/cavium/liquidio/octeon_device.c if (oct->dispatch.dlist[idx].opcode == 0) { idx 1136 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->dispatch.dlist[idx].opcode = combined_opcode; idx 1137 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->dispatch.dlist[idx].dispatch_fn = fn; idx 1138 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->dispatch.dlist[idx].arg = fn_arg; idx 1170 drivers/net/ethernet/cavium/liquidio/octeon_device.c list_add(&dispatch->list, &oct->dispatch.dlist[idx].list); idx 58 drivers/net/ethernet/cavium/liquidio/octeon_droq.c int idx; idx 63 drivers/net/ethernet/cavium/liquidio/octeon_droq.c idx = combined_opcode & OCTEON_OPCODE_MASK; idx 72 drivers/net/ethernet/cavium/liquidio/octeon_droq.c if (octeon_dev->dispatch.dlist[idx].opcode == combined_opcode) { idx 73 drivers/net/ethernet/cavium/liquidio/octeon_droq.c fn_arg = octeon_dev->dispatch.dlist[idx].arg; idx 76 drivers/net/ethernet/cavium/liquidio/octeon_droq.c &octeon_dev->dispatch.dlist[idx].list) { idx 339 drivers/net/ethernet/cavium/liquidio/octeon_droq.c u32 idx) idx 347 drivers/net/ethernet/cavium/liquidio/octeon_droq.c info = (struct octeon_droq_info *)droq->recv_buf_list[idx].data; idx 364 drivers/net/ethernet/cavium/liquidio/octeon_droq.c pg_info = &droq->recv_buf_list[idx].pg_info; idx 376 drivers/net/ethernet/cavium/liquidio/octeon_droq.c recv_pkt->buffer_ptr[i] = droq->recv_buf_list[idx].buffer; idx 377 drivers/net/ethernet/cavium/liquidio/octeon_droq.c droq->recv_buf_list[idx].buffer = NULL; idx 379 drivers/net/ethernet/cavium/liquidio/octeon_droq.c idx = incr_index(idx, 1, droq->max_count); idx 662 drivers/net/ethernet/cavium/liquidio/octeon_droq.c int cpy_len, idx = droq->read_idx; idx 673 drivers/net/ethernet/cavium/liquidio/octeon_droq.c idx); idx 675 drivers/net/ethernet/cavium/liquidio/octeon_droq.c idx].buffer; idx 677 drivers/net/ethernet/cavium/liquidio/octeon_droq.c droq->recv_buf_list[idx].buffer idx 30 drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c octeon_toggle_bar1_swapmode(struct octeon_device *oct, u32 idx) idx 34 drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c mask = oct->fn_list.bar1_idx_read(oct, idx); idx 36 drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c oct->fn_list.bar1_idx_write(oct, idx, mask); idx 39 drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c #define octeon_toggle_bar1_swapmode(oct, idx) idx 493 drivers/net/ethernet/cavium/liquidio/octeon_network.h int idx) idx 495 drivers/net/ethernet/cavium/liquidio/octeon_network.h skb_put_data(nicbuf, get_rbd(droq->recv_buf_list[idx].buffer), idx 364 drivers/net/ethernet/cavium/liquidio/request_manager.c int idx, void *buf, int reqtype) idx 366 drivers/net/ethernet/cavium/liquidio/request_manager.c iq->request_list[idx].buf = buf; idx 367 drivers/net/ethernet/cavium/liquidio/request_manager.c iq->request_list[idx].reqtype = reqtype; idx 504 drivers/net/ethernet/cavium/thunder/nic.h u8 idx; idx 242 drivers/net/ethernet/cavium/thunder/nic_main.c mbx.bgx_stats.idx = bgx->idx; idx 245 drivers/net/ethernet/cavium/thunder/nic_main.c lmac, bgx->idx); idx 248 drivers/net/ethernet/cavium/thunder/nic_main.c lmac, bgx->idx); idx 567 drivers/net/ethernet/cavium/thunder/nic_main.c u8 qset, idx = 0; idx 576 drivers/net/ethernet/cavium/thunder/nic_main.c u8 svf = cfg->ind_tbl[idx] >> 3; idx 583 drivers/net/ethernet/cavium/thunder/nic_main.c (qset << 3) | (cfg->ind_tbl[idx] & 0x7)); idx 584 drivers/net/ethernet/cavium/thunder/nic_main.c idx++; idx 738 drivers/net/ethernet/cavium/thunder/nic_main.c int idx, alloc_qs = 0; idx 744 drivers/net/ethernet/cavium/thunder/nic_main.c for (idx = 0; idx < sqs->qs_count; idx++) { idx 748 drivers/net/ethernet/cavium/thunder/nic_main.c nic->vf_sqs[sqs->vf_id][idx] = sqs_id; idx 656 drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c int idx; idx 659 drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c for (idx = 0; idx < rss->rss_size; idx++) idx 660 drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c indir[idx] = rss->ind_tbl[idx]; idx 677 drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c int idx; idx 689 drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c for (idx = 0; idx < rss->rss_size; idx++) idx 690 drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c rss->ind_tbl[idx] = indir[idx]; idx 196 drivers/net/ethernet/cavium/thunder/nicvf_main.c nic->bgx_stats.rx_stats[bgx->idx] = bgx->stats; idx 198 drivers/net/ethernet/cavium/thunder/nicvf_main.c nic->bgx_stats.tx_stats[bgx->idx] = bgx->stats; idx 364 drivers/net/ethernet/cavium/thunder/nicvf_main.c int idx; idx 366 drivers/net/ethernet/cavium/thunder/nicvf_main.c for (idx = 0; idx < RSS_HASH_KEY_SIZE; idx++) { idx 367 drivers/net/ethernet/cavium/thunder/nicvf_main.c nicvf_reg_write(nic, key_addr, rss->key[idx]); idx 375 drivers/net/ethernet/cavium/thunder/nicvf_main.c int idx; idx 395 drivers/net/ethernet/cavium/thunder/nicvf_main.c for (idx = 0; idx < rss->rss_size; idx++) idx 396 drivers/net/ethernet/cavium/thunder/nicvf_main.c rss->ind_tbl[idx] = ethtool_rxfh_indir_default(idx, idx 1644 drivers/net/ethernet/cavium/thunder/nicvf_main.c mbx.bgx_stats.idx = stat; idx 1655 drivers/net/ethernet/cavium/thunder/nicvf_main.c mbx.bgx_stats.idx = stat; idx 1993 drivers/net/ethernet/cavium/thunder/nicvf_main.c int idx; idx 2019 drivers/net/ethernet/cavium/thunder/nicvf_main.c for (idx = 0; idx < mc_addrs->count; idx++) { idx 2021 drivers/net/ethernet/cavium/thunder/nicvf_main.c mbx.xcast.mac = mc_addrs->mc[idx]; idx 255 drivers/net/ethernet/cavium/thunder/nicvf_queues.c int idx; idx 299 drivers/net/ethernet/cavium/thunder/nicvf_queues.c for (idx = 0; idx < ring_len; idx++) { idx 304 drivers/net/ethernet/cavium/thunder/nicvf_queues.c rbdr->tail = idx - 1; idx 308 drivers/net/ethernet/cavium/thunder/nicvf_queues.c desc = GET_RBDR_DESC(rbdr, idx); idx 556 drivers/net/ethernet/cavium/thunder/nicvf_queues.c u8 idx; idx 560 drivers/net/ethernet/cavium/thunder/nicvf_queues.c for (idx = 0; idx < subdesc_cnt; idx++) { idx 305 drivers/net/ethernet/cavium/thunder/nicvf_queues.h #define GET_RBDR_DESC(RING, idx)\ idx 306 drivers/net/ethernet/cavium/thunder/nicvf_queues.h (&(((struct rbdr_entry_t *)((RING)->desc))[idx])) idx 307 drivers/net/ethernet/cavium/thunder/nicvf_queues.h #define GET_SQ_DESC(RING, idx)\ idx 308 drivers/net/ethernet/cavium/thunder/nicvf_queues.h (&(((struct sq_hdr_subdesc *)((RING)->desc))[idx])) idx 309 drivers/net/ethernet/cavium/thunder/nicvf_queues.h #define GET_CQ_DESC(RING, idx)\ idx 310 drivers/net/ethernet/cavium/thunder/nicvf_queues.h (&(((union cq_desc_t *)((RING)->desc))[idx])) idx 171 drivers/net/ethernet/cavium/thunder/thunder_bgx.c int idx = (node * max_bgx_per_node) + bgx_idx; idx 173 drivers/net/ethernet/cavium/thunder/thunder_bgx.c return bgx_vnic[idx]; idx 248 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u8 idx = 0; idx 252 drivers/net/ethernet/cavium/thunder/thunder_bgx.c for (idx = 0; idx < lmac->dmacs_count; idx++) idx 254 drivers/net/ethernet/cavium/thunder/thunder_bgx.c ((lmacid * lmac->dmacs_count) + idx) * idx 312 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u64 cam_dmac, u8 idx) idx 327 drivers/net/ethernet/cavium/thunder/thunder_bgx.c ((lmacid * lmac->dmacs_count) + idx) * sizeof(u64), cfg); idx 630 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u64 bgx_get_rx_stats(int node, int bgx_idx, int lmac, int idx) idx 638 drivers/net/ethernet/cavium/thunder/thunder_bgx.c if (idx > 8) idx 640 drivers/net/ethernet/cavium/thunder/thunder_bgx.c return bgx_reg_read(bgx, lmac, BGX_CMRX_RX_STAT0 + (idx * 8)); idx 644 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u64 bgx_get_tx_stats(int node, int bgx_idx, int lmac, int idx) idx 652 drivers/net/ethernet/cavium/thunder/thunder_bgx.c return bgx_reg_read(bgx, lmac, BGX_CMRX_TX_STAT0 + (idx * 8)); idx 1324 drivers/net/ethernet/cavium/thunder/thunder_bgx.c static void bgx_set_lmac_config(struct bgx *bgx, u8 idx) idx 1331 drivers/net/ethernet/cavium/thunder/thunder_bgx.c lmac = &bgx->lmac[idx]; idx 1351 drivers/net/ethernet/cavium/thunder/thunder_bgx.c cmr_cfg = bgx_reg_read(bgx, idx, BGX_CMRX_CFG); idx 1366 drivers/net/ethernet/cavium/thunder/thunder_bgx.c u8 idx; idx 1369 drivers/net/ethernet/cavium/thunder/thunder_bgx.c for (idx = 0; idx < bgx->max_lmac; idx++) { idx 1370 drivers/net/ethernet/cavium/thunder/thunder_bgx.c lmac = &bgx->lmac[idx]; idx 1371 drivers/net/ethernet/cavium/thunder/thunder_bgx.c lmac->lmacid = idx; idx 1381 drivers/net/ethernet/cavium/thunder/thunder_bgx.c for (idx = 0; idx < bgx->lmac_count; idx++) { idx 1382 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx_set_lmac_config(bgx, idx); idx 1383 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx_print_qlm_mode(bgx, idx); idx 239 drivers/net/ethernet/cavium/thunder/thunder_bgx.h u64 bgx_get_rx_stats(int node, int bgx_idx, int lmac, int idx); idx 240 drivers/net/ethernet/cavium/thunder/thunder_bgx.h u64 bgx_get_tx_stats(int node, int bgx_idx, int lmac, int idx); idx 60 drivers/net/ethernet/chelsio/cxgb/common.h #define CH_DEVICE(devid, ssid, idx) \ idx 61 drivers/net/ethernet/chelsio/cxgb/common.h { PCI_VENDOR_ID_CHELSIO, devid, PCI_ANY_ID, ssid, 0, 0, idx } idx 215 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_ADDR(idx, reg) (MAC_REG_BASE + (idx) * 128 + (reg)) idx 217 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_IDLO(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_LO) idx 218 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_IDHI(idx) MAC_REG_ADDR(idx, A_GMAC_MACID_HI) idx 219 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_CSR(idx) MAC_REG_ADDR(idx, A_GMAC_CSR) idx 220 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_IFS(idx) MAC_REG_ADDR(idx, A_GMAC_IFS) idx 221 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_LARGEFRAMELENGTH(idx) MAC_REG_ADDR(idx, A_GMAC_JUMBO_FRAME_LEN) idx 222 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_LINKDLY(idx) MAC_REG_ADDR(idx, A_GMAC_LNK_DLY) idx 223 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_PAUSETIME(idx) MAC_REG_ADDR(idx, A_GMAC_PAUSETIME) idx 224 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_CASTLO(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_LO) idx 225 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_MCASTHI(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_HI) idx 226 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_CASTMASKLO(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_MASK_LO) idx 227 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_MCASTMASKHI(idx) MAC_REG_ADDR(idx, A_GMAC_MCAST_MASK_HI) idx 228 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_RMCNT(idx) MAC_REG_ADDR(idx, A_GMAC_RMT_CNT) idx 229 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_RMDATA(idx) MAC_REG_ADDR(idx, A_GMAC_RMT_DATA) idx 230 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_GMRANDBACKOFFSEED(idx) MAC_REG_ADDR(idx, A_GMAC_BACKOFF_SEED) idx 231 drivers/net/ethernet/chelsio/cxgb/fpga_defs.h #define MAC_REG_TXFTHRESHOLDS(idx) MAC_REG_ADDR(idx, A_GMAC_TXF_THRES) idx 1002 drivers/net/ethernet/chelsio/cxgb/sge.c static void recycle_fl_buf(struct freelQ *fl, int idx) idx 1004 drivers/net/ethernet/chelsio/cxgb/sge.c struct freelQ_e *from = &fl->entries[idx]; idx 1007 drivers/net/ethernet/chelsio/cxgb/sge.c fl->centries[fl->pidx] = fl->centries[idx]; idx 284 drivers/net/ethernet/chelsio/cxgb3/adapter.h static inline struct port_info *adap2pinfo(struct adapter *adap, int idx) idx 286 drivers/net/ethernet/chelsio/cxgb3/adapter.h return netdev_priv(adap->port[idx]); idx 598 drivers/net/ethernet/chelsio/cxgb3/common.h #define XGM_REG(reg_addr, idx) \ idx 599 drivers/net/ethernet/chelsio/cxgb3/common.h ((reg_addr) + (idx) * (XGMAC0_1_BASE_ADDR - XGMAC0_0_BASE_ADDR)) idx 668 drivers/net/ethernet/chelsio/cxgb3/common.h void t3_xgm_intr_enable(struct adapter *adapter, int idx); idx 669 drivers/net/ethernet/chelsio/cxgb3/common.h void t3_xgm_intr_disable(struct adapter *adapter, int idx); idx 670 drivers/net/ethernet/chelsio/cxgb3/common.h void t3_port_intr_enable(struct adapter *adapter, int idx); idx 671 drivers/net/ethernet/chelsio/cxgb3/common.h void t3_port_intr_disable(struct adapter *adapter, int idx); idx 713 drivers/net/ethernet/chelsio/cxgb3/common.h int t3_mac_set_address(struct cmac *mac, unsigned int idx, u8 addr[6]); idx 71 drivers/net/ethernet/chelsio/cxgb3/cxgb3_ioctl.h uint32_t idx; idx 102 drivers/net/ethernet/chelsio/cxgb3/cxgb3_ioctl.h uint8_t idx; idx 85 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c #define CH_DEVICE(devid, idx) \ idx 86 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c { PCI_VENDOR_ID_CHELSIO, devid, PCI_ANY_ID, PCI_ANY_ID, 0, 0, idx } idx 903 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c static int write_smt_entry(struct adapter *adapter, int idx) idx 906 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct port_info *pi = netdev_priv(adapter->port[idx]); idx 914 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SMT_WRITE_REQ, idx)); idx 916 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c req->iff = idx; idx 917 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c memcpy(req->src_mac0, adapter->port[idx]->dev_addr, ETH_ALEN); idx 959 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c req->idx = qidx; idx 1654 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c struct port_info *p, int idx) idx 1660 drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c tot += adapter->sge.qs[i].port_stats[idx]; idx 1106 drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c req->val = cpu_to_be64(V_TCB_L2T_IX(e->idx)); idx 101 drivers/net/ethernet/chelsio/cxgb3/l2t.c OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_L2T_WRITE_REQ, e->idx)); idx 102 drivers/net/ethernet/chelsio/cxgb3/l2t.c req->params = htonl(V_L2T_W_IDX(e->idx) | V_L2T_W_IFF(e->smt_idx) | idx 458 drivers/net/ethernet/chelsio/cxgb3/l2t.c d->l2tab[i].idx = i; idx 59 drivers/net/ethernet/chelsio/cxgb3/l2t.h u16 idx; /* entry index */ idx 579 drivers/net/ethernet/chelsio/cxgb3/sge.c unsigned int idx) idx 581 drivers/net/ethernet/chelsio/cxgb3/sge.c struct rx_desc *from = &q->desc[idx]; idx 584 drivers/net/ethernet/chelsio/cxgb3/sge.c q->sdesc[q->pidx] = q->sdesc[idx]; idx 890 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h __u8 idx; idx 37 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static void t3_port_intr_clear(struct adapter *adapter, int idx); idx 1844 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static int mac_intr_handler(struct adapter *adap, unsigned int idx) idx 1846 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c struct cmac *mac = &adap2pinfo(adap, idx)->mac; idx 1858 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c CH_ALERT(adap, "port%d: MAC TX FIFO parity error\n", idx); idx 1862 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c CH_ALERT(adap, "port%d: MAC RX FIFO parity error\n", idx); idx 1880 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_os_link_fault_handler(adap, idx); idx 2092 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c void t3_xgm_intr_enable(struct adapter *adapter, int idx) idx 2094 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c struct port_info *pi = adap2pinfo(adapter, idx); idx 2100 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c void t3_xgm_intr_disable(struct adapter *adapter, int idx) idx 2102 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c struct port_info *pi = adap2pinfo(adapter, idx); idx 2116 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c void t3_port_intr_enable(struct adapter *adapter, int idx) idx 2118 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c struct cphy *phy = &adap2pinfo(adapter, idx)->phy; idx 2120 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_write_reg(adapter, XGM_REG(A_XGM_INT_ENABLE, idx), XGM_INTR_MASK); idx 2121 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_read_reg(adapter, XGM_REG(A_XGM_INT_ENABLE, idx)); /* flush */ idx 2133 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c void t3_port_intr_disable(struct adapter *adapter, int idx) idx 2135 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c struct cphy *phy = &adap2pinfo(adapter, idx)->phy; idx 2137 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_write_reg(adapter, XGM_REG(A_XGM_INT_ENABLE, idx), 0); idx 2138 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_read_reg(adapter, XGM_REG(A_XGM_INT_ENABLE, idx)); /* flush */ idx 2150 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c static void t3_port_intr_clear(struct adapter *adapter, int idx) idx 2152 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c struct cphy *phy = &adap2pinfo(adapter, idx)->phy; idx 2154 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_write_reg(adapter, XGM_REG(A_XGM_INT_CAUSE, idx), 0xffffffff); idx 2155 drivers/net/ethernet/chelsio/cxgb3/t3_hw.c t3_read_reg(adapter, XGM_REG(A_XGM_INT_CAUSE, idx)); /* flush */ idx 154 drivers/net/ethernet/chelsio/cxgb3/xgmac.c int idx = macidx(mac); idx 171 drivers/net/ethernet/chelsio/cxgb3/xgmac.c t3_write_reg(adap, A_TP_PIO_ADDR, A_TP_TX_DROP_CFG_CH0 + idx); idx 172 drivers/net/ethernet/chelsio/cxgb3/xgmac.c store = t3_read_reg(adap, A_TP_TX_DROP_CFG_CH0 + idx); idx 177 drivers/net/ethernet/chelsio/cxgb3/xgmac.c t3_write_reg(adap, A_TP_PIO_ADDR, A_TP_TX_DROP_CFG_CH0 + idx); idx 210 drivers/net/ethernet/chelsio/cxgb3/xgmac.c t3_write_reg(adap, A_TP_PIO_ADDR, A_TP_TX_DROP_CFG_CH0 + idx); idx 213 drivers/net/ethernet/chelsio/cxgb3/xgmac.c if (!idx) idx 230 drivers/net/ethernet/chelsio/cxgb3/xgmac.c static void set_addr_filter(struct cmac *mac, int idx, const u8 * addr) idx 233 drivers/net/ethernet/chelsio/cxgb3/xgmac.c unsigned int oft = mac->offset + idx * 8; idx 243 drivers/net/ethernet/chelsio/cxgb3/xgmac.c int t3_mac_set_address(struct cmac *mac, unsigned int idx, u8 addr[6]) idx 245 drivers/net/ethernet/chelsio/cxgb3/xgmac.c if (idx >= mac->nucast) idx 247 drivers/net/ethernet/chelsio/cxgb3/xgmac.c set_addr_filter(mac, idx, addr); idx 469 drivers/net/ethernet/chelsio/cxgb3/xgmac.c int idx = macidx(mac); idx 475 drivers/net/ethernet/chelsio/cxgb3/xgmac.c t3_write_reg(adap, A_TP_PIO_ADDR, A_TP_TX_DROP_CFG_CH0 + idx); idx 480 drivers/net/ethernet/chelsio/cxgb3/xgmac.c t3_set_reg_field(adap, A_TP_PIO_DATA, 1 << idx, idx 481 drivers/net/ethernet/chelsio/cxgb3/xgmac.c adap->params.rev == T3_REV_C ? 0 : 1 << idx); idx 485 drivers/net/ethernet/chelsio/cxgb3/xgmac.c t3_write_reg(adap, A_TP_PIO_ADDR, A_TP_TX_DROP_CNT_CH0 + idx); idx 108 drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h u32 idx; idx 207 drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h u32 idx; idx 145 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c meminfo_buff->mem[i].idx = i; idx 158 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c meminfo_buff->avail[i].idx = 0; idx 169 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c meminfo_buff->avail[i].idx = 1; idx 181 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c meminfo_buff->avail[i].idx = 3; idx 192 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c meminfo_buff->avail[i].idx = 4; idx 203 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c meminfo_buff->avail[i].idx = 2; idx 214 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c meminfo_buff->avail[i].idx = 5; idx 260 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c md->idx = ARRAY_SIZE(cudbg_region); /* hide it */ idx 279 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c md->idx = ARRAY_SIZE(cudbg_region); idx 312 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c md->idx = ARRAY_SIZE(cudbg_region); /* hide it */ idx 737 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c u8 mem_type, u8 *idx) idx 763 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c if (mem_info->avail[i].idx == flag) { idx 764 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c *idx = i; idx 779 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c u32 i, idx = 0; idx 789 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c idx = i; idx 798 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c if (meminfo->mem[i].idx >= ARRAY_SIZE(cudbg_region)) idx 806 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c if (meminfo->mem[i].idx == idx) { idx 2121 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c struct cudbg_mps_tcam *tcam, u32 idx) idx 2137 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c if (idx < 256) idx 2138 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c ctl |= CTLTCAMINDEX_V(idx) | CTLTCAMSEL_V(0); idx 2140 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c ctl |= CTLTCAMINDEX_V(idx - 256) | CTLTCAMSEL_V(1); idx 2178 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c tcamy = t4_read_reg64(padap, MPS_CLS_TCAM_Y_L(idx)); idx 2179 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c tcamx = t4_read_reg64(padap, MPS_CLS_TCAM_X_L(idx)); idx 2186 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c tcam->cls_lo = t4_read_reg(padap, MPS_CLS_SRAM_L(idx)); idx 2187 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c tcam->cls_hi = t4_read_reg(padap, MPS_CLS_SRAM_H(idx)); idx 2206 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c FW_LDST_CMD_IDX_V(idx)); idx 2237 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c tcam->idx = idx; idx 684 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h u8 idx; /* queue index within its group */ idx 883 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h unsigned int idx; idx 921 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h u16 idx; idx 1364 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h static inline struct port_info *adap2pinfo(struct adapter *adap, int idx) idx 1366 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h return netdev_priv(adap->port[idx]); idx 1395 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h unsigned int idx = q->intr_params >> 1; idx 1397 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0; idx 1695 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_write_rss_key(struct adapter *adap, const u32 *key, int idx, idx 1723 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_get_port_stats(struct adapter *adap, int idx, struct port_stats *p); idx 1724 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_get_port_stats_offset(struct adapter *adap, int idx, idx 1727 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_get_lb_stats(struct adapter *adap, int idx, struct lb_port_stats *p); idx 1743 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_get_fcoe_stats(struct adapter *adap, unsigned int idx, idx 1797 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h const u8 *addr, const u8 *mask, unsigned int idx, idx 1799 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_free_encap_mac_filt(struct adapter *adap, unsigned int viid, int idx, idx 1806 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h const u8 *addr, const u8 *mask, unsigned int idx, idx 1810 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h const u8 **addr, u16 *idx, u64 *hash, bool sleep_ok); idx 1815 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int idx, const u8 *addr, bool persist, u8 *smt_idx); idx 1912 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h cpumask_var_t *aff_mask, int idx); idx 1921 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h const u8 **addr, u16 *idx, idx 1932 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int idx, bool sleep_ok); idx 1937 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h unsigned int idx, idx 1945 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h unsigned int idx, idx 363 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c int idx = fwap->idx; idx 364 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c struct app_priority *ap = &dcb->app_priority[idx]; idx 803 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c pcmd.u.dcb.app_priority.idx = i; idx 854 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c pcmd.u.dcb.app_priority.idx = i; idx 887 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c pcmd.u.dcb.app_priority.idx = i; idx 1127 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c pcmd.u.dcb.app_priority.idx = *app_count; idx 1157 drivers/net/ethernet/chelsio/cxgb4/cxgb4_dcb.c pcmd.u.dcb.app_priority.idx = i; idx 122 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int cim_la_show(struct seq_file *seq, void *v, int idx) idx 139 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int cim_la_show_3in1(struct seq_file *seq, void *v, int idx) idx 157 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int cim_la_show_t6(struct seq_file *seq, void *v, int idx) idx 176 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int cim_la_show_pc_t6(struct seq_file *seq, void *v, int idx) idx 236 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int cim_pif_la_show(struct seq_file *seq, void *v, int idx) idx 242 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c } else if (idx < CIM_PIFLA_SIZE) { idx 247 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (idx == CIM_PIFLA_SIZE) idx 279 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int cim_ma_la_show(struct seq_file *seq, void *v, int idx) idx 285 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c } else if (idx < CIM_MALA_SIZE) { idx 289 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (idx == CIM_MALA_SIZE) idx 382 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int cimq_show(struct seq_file *seq, void *v, int idx) idx 386 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c seq_printf(seq, "%#06x: %08x %08x %08x %08x\n", idx * 16, p[0], p[1], idx 512 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int tp_la_show(struct seq_file *seq, void *v, int idx) idx 520 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int tp_la_show2(struct seq_file *seq, void *v, int idx) idx 524 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (idx) idx 527 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (idx < (TPLA_SIZE / 2 - 1) || p[1] != ~0ULL) idx 532 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int tp_la_show3(struct seq_file *seq, void *v, int idx) idx 630 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (idx) idx 633 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (idx < (TPLA_SIZE / 2 - 1) || p[1] != ~0ULL) idx 694 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int ulprx_la_show(struct seq_file *seq, void *v, int idx) idx 1698 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c unsigned int idx = (uintptr_t)v - 2; idx 1713 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (idx < 256) idx 1714 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c ctl |= CTLTCAMINDEX_V(idx) | CTLTCAMSEL_V(0); idx 1716 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c ctl |= CTLTCAMINDEX_V(idx - 256) | idx 1753 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c tcamy = t4_read_reg64(adap, MPS_CLS_TCAM_Y_L(idx)); idx 1754 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c tcamx = t4_read_reg64(adap, MPS_CLS_TCAM_X_L(idx)); idx 1757 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c cls_lo = t4_read_reg(adap, MPS_CLS_SRAM_L(idx)); idx 1758 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c cls_hi = t4_read_reg(adap, MPS_CLS_SRAM_H(idx)); idx 1761 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c seq_printf(seq, "%3u -\n", idx); idx 1788 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c FW_LDST_CMD_IDX_V(idx)); idx 1794 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c idx, -ret); idx 1818 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c "%3c %#x%4u%4d", idx, addr[0], idx 1834 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c idx, addr[0], addr[1], addr[2], idx 1855 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c idx, addr[0], addr[1], addr[2], addr[3], idx 1986 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int rss_show(struct seq_file *seq, void *v, int idx) idx 1991 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c idx * 8, entry[0], entry[1], entry[2], entry[3], entry[4], idx 2251 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int rss_pf_config_show(struct seq_file *seq, void *v, int idx) idx 2272 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c idx, idx 2276 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c G_PFnLKPIDX(pfconf->rss_pf_map, idx), idx 2277 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c G_PFnMSKSIZE(pfconf->rss_pf_mask, idx), idx 2333 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c static int rss_vf_config_show(struct seq_file *seq, void *v, int idx) idx 2343 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c idx, idx 3202 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c unsigned int idx, unsigned int size_mb) idx 3205 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c (void *)adap + idx, &mem_debugfs_fops, idx 3284 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c mem_region_show(seq, memory[meminfo.avail[i].idx], idx 3290 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (meminfo.mem[i].idx >= ARRAY_SIZE(cudbg_region)) idx 3296 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c mem_region_show(seq, cudbg_region[meminfo.mem[i].idx], idx 48 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.h int (*show)(struct seq_file *seq, void *v, int idx); idx 154 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c TCB_SMAC_SEL_V(f->smt->idx), 1); idx 618 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c FW_FILTER_WR_L2TIX_V(f->l2t ? f->l2t->idx : 0)); idx 1029 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c L2T_IDX_V(f->l2t ? f->l2t->idx : 0) | idx 1068 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c L2T_IDX_V(f->l2t ? f->l2t->idx : 0) | idx 1768 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c int idx; idx 1774 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c idx = tid - adap->tids.ftid_base; idx 1775 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c if (idx >= max_fidx) idx 1777 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c f = &adap->tids.ftid_tab[idx]; idx 1811 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c ctx->tid = idx; idx 1823 drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c idx, ret); idx 374 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c u16 idx[1] = {}; idx 381 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c idx, ucast ? &uhash : &mhash, false); idx 712 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c cpumask_var_t *aff_mask, int idx) idx 721 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c cpumask_set_cpu(cpumask_local_spread(idx, dev_to_node(adap->pdev_dev)), idx 1013 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c q->rspq.idx = j; idx 1634 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c unsigned int *idx) idx 1640 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c if (idx) idx 1641 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c *idx = i; idx 3127 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c qe.class = e->idx; idx 5347 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c adap->msix_info_ulds[j].idx = i; idx 31 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c static int cxgb4_mps_ref_dec(struct adapter *adap, u16 idx) idx 38 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c if (mps_entry->idx == idx) { idx 54 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c u16 idx, const u8 *mask) idx 62 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c if (mps_entry->idx == idx) { idx 74 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c mps_entry->idx = idx; idx 102 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c u16 *idx, u64 *hash, bool sleep_ok) idx 107 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c naddr, addr, idx, hash, sleep_ok); idx 112 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c if (idx[i] != 0xffff) { idx 113 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c if (cxgb4_mps_ref_inc(adap, addr[i], idx[i], NULL)) { idx 148 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c unsigned int idx, idx 155 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c if (!cxgb4_mps_ref_dec(adap, idx)) idx 157 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c mask, idx, lookup_type, idx 167 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c unsigned int idx, idx 175 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c mask, idx, lookup_type, idx 183 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c mask, idx, lookup_type, idx 191 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c int idx, bool sleep_ok) idx 195 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c if (!cxgb4_mps_ref_dec(adap, idx)) idx 196 drivers/net/ethernet/chelsio/cxgb4/cxgb4_mps.c ret = t4_free_encap_mac_filt(adap, viid, idx, sleep_ok); idx 167 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c msi_idx = adap->msix_info_ulds[bmap_idx].idx; idx 360 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c unsigned int idx, bmap_idx; idx 362 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c for_each_uldrxq(rxq_info, idx) { idx 363 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c bmap_idx = rxq_info->msix_tbl[idx]; idx 368 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c &rxq_info->uldrxq[idx].rspq); idx 373 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c &minfo->aff_mask, idx); idx 378 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c while (idx-- > 0) { idx 379 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c bmap_idx = rxq_info->msix_tbl[idx]; idx 383 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c free_irq(minfo->vec, &rxq_info->uldrxq[idx].rspq); idx 393 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c unsigned int idx, bmap_idx; idx 395 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c for_each_uldrxq(rxq_info, idx) { idx 396 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c bmap_idx = rxq_info->msix_tbl[idx]; idx 401 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c free_irq(minfo->vec, &rxq_info->uldrxq[idx].rspq); idx 409 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c unsigned int idx, bmap_idx; idx 411 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c for_each_uldrxq(rxq_info, idx) { idx 412 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c bmap_idx = rxq_info->msix_tbl[idx]; idx 415 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c adap->port[0]->name, rxq_info->name, idx); idx 442 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c int idx; idx 444 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c for_each_uldrxq(rxq_info, idx) idx 445 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c enable_rx(adap, &rxq_info->uldrxq[idx].rspq); idx 451 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c int idx; idx 453 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c for_each_uldrxq(rxq_info, idx) idx 454 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c quiesce_rx(adap, &rxq_info->uldrxq[idx].rspq); idx 392 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h int cxgb4_immdata_send(struct net_device *dev, unsigned int idx, idx 402 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h unsigned int *idx); idx 141 drivers/net/ethernet/chelsio/cxgb4/l2t.c unsigned int l2t_idx = e->idx + d->l2t_start; idx 365 drivers/net/ethernet/chelsio/cxgb4/l2t.c d = container_of(e, struct l2t_data, l2tab[e->idx]); idx 386 drivers/net/ethernet/chelsio/cxgb4/l2t.c d = container_of(e, struct l2t_data, l2tab[e->idx]); idx 661 drivers/net/ethernet/chelsio/cxgb4/l2t.c d->l2tab[i].idx = i; idx 723 drivers/net/ethernet/chelsio/cxgb4/l2t.c e->idx + d->l2t_start, ip, e->dmac, idx 75 drivers/net/ethernet/chelsio/cxgb4/l2t.h u16 idx; /* entry index within in-memory table */ idx 57 drivers/net/ethernet/chelsio/cxgb4/sched.c p->u.params.channel, e->idx, idx 431 drivers/net/ethernet/chelsio/cxgb4/sched.c np.u.params.class = e->idx; idx 488 drivers/net/ethernet/chelsio/cxgb4/sched.c s->tab[i].idx = i; idx 69 drivers/net/ethernet/chelsio/cxgb4/sched.h u8 idx; idx 2432 drivers/net/ethernet/chelsio/cxgb4/sge.c unsigned int idx = skb_txq(skb); idx 2437 drivers/net/ethernet/chelsio/cxgb4/sge.c idx = 0; idx 2438 drivers/net/ethernet/chelsio/cxgb4/sge.c return ctrl_xmit(&adap->sge.ctrlq[idx], skb); idx 2447 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &txq_info->uldtxq[idx]; idx 2556 drivers/net/ethernet/chelsio/cxgb4/sge.c int cxgb4_immdata_send(struct net_device *dev, unsigned int idx, idx 2573 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &txq_info->uldtxq[idx]; idx 2762 drivers/net/ethernet/chelsio/cxgb4/sge.c skb_record_rx_queue(skb, rxq->rspq.idx); idx 2914 drivers/net/ethernet/chelsio/cxgb4/sge.c txq = &s->ethtxq[pi->first_qset + rspq->idx]; idx 3001 drivers/net/ethernet/chelsio/cxgb4/sge.c skb_record_rx_queue(skb, q->idx); idx 56 drivers/net/ethernet/chelsio/cxgb4/smt.c s->smtab[i].idx = i; idx 160 drivers/net/ethernet/chelsio/cxgb4/smt.c row = (e->idx >> 1); idx 161 drivers/net/ethernet/chelsio/cxgb4/smt.c if (e->idx & 1) { idx 169 drivers/net/ethernet/chelsio/cxgb4/smt.c memcpy(req->src_mac0, s->smtab[e->idx - 1].src_mac, idx 179 drivers/net/ethernet/chelsio/cxgb4/smt.c memcpy(req->src_mac1, s->smtab[e->idx + 1].src_mac, idx 194 drivers/net/ethernet/chelsio/cxgb4/smt.c memcpy(req->src_mac0, s->smtab[e->idx].src_mac, ETH_ALEN); idx 195 drivers/net/ethernet/chelsio/cxgb4/smt.c row = e->idx; idx 199 drivers/net/ethernet/chelsio/cxgb4/smt.c htonl(MK_OPCODE_TID(CPL_SMT_WRITE_REQ, e->idx | idx 59 drivers/net/ethernet/chelsio/cxgb4/smt.h u16 idx; idx 92 drivers/net/ethernet/chelsio/cxgb4/srq.c req->idx = srq_idx; idx 105 drivers/net/ethernet/chelsio/cxgb4/srq.c WARN_ON_ONCE(entryp->idx != srq_idx); idx 115 drivers/net/ethernet/chelsio/cxgb4/srq.c unsigned int idx = TID_TID_G(GET_TID(rpl)); idx 122 drivers/net/ethernet/chelsio/cxgb4/srq.c rpl->status, idx); idx 129 drivers/net/ethernet/chelsio/cxgb4/srq.c e->idx = idx; idx 45 drivers/net/ethernet/chelsio/cxgb4/srq.h u8 idx; idx 449 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c static int t4_edc_err_read(struct adapter *adap, int idx) idx 458 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (idx != 0 && idx != 1) { idx 459 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c CH_WARN(adap, "%s: idx %d NOT supported.\n", __func__, idx); idx 463 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c edc_ecc_err_addr_reg = EDC_T5_REG(EDC_H_ECC_ERR_ADDR_A, idx); idx 464 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c rdata_reg = EDC_T5_REG(EDC_H_BIST_STATUS_RDATA_A, idx); idx 468 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx, edc_ecc_err_addr_reg, idx 3934 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int i, j, idx; idx 3942 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx = 8 * i + j; idx 3943 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c t4_write_reg(adap, CIM_DEBUGCFG_A, POLADBGRDPTR_V(idx) | idx 3944 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c PILADBGRDPTR_V(idx)); idx 4848 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c static void mem_intr_handler(struct adapter *adapter, int idx) idx 4854 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (idx <= MEM_EDC1) { idx 4855 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c addr = EDC_REG(EDC_INT_CAUSE_A, idx); idx 4856 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c cnt_addr = EDC_REG(EDC_ECC_STATUS_A, idx); idx 4857 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c } else if (idx == MEM_MC) { idx 4873 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c name[idx]); idx 4877 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c t4_edc_err_read(adapter, idx); idx 4883 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c cnt, name[idx], cnt > 1 ? "s" : ""); idx 4887 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c "%s uncorrectable ECC data error\n", name[idx]); idx 5482 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_write_rss_key(struct adapter *adap, const u32 *key, int idx, idx 5498 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (idx >= 0 && idx < rss_key_addr_cnt) { idx 5501 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c KEYWRADDRX_V(idx >> 4) | idx 5502 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c T6_VFWRADDR_V(idx) | KEYWREN_F); idx 5505 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c KEYWRADDR_V(idx) | KEYWREN_F); idx 5707 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_get_fcoe_stats(struct adapter *adap, unsigned int idx, idx 5712 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c t4_tp_mib_read(adap, &st->frames_ddp, 1, TP_MIB_FCOE_DDP_0_A + idx, idx 5716 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c TP_MIB_FCOE_DROP_0_A + idx, sleep_ok); idx 5718 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c t4_tp_mib_read(adap, val, 2, TP_MIB_FCOE_BYTE_0_HI_A + 2 * idx, idx 5953 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int idx, int enable) idx 5955 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int i, ofst = idx * 4; idx 5976 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (tp->snap_len > 9600 || idx) idx 5988 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx *= (MPS_TRC_FILTER1_MATCH_A - MPS_TRC_FILTER0_MATCH_A); idx 5989 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c data_reg = MPS_TRC_FILTER0_MATCH_A + idx; idx 5990 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c mask_reg = MPS_TRC_FILTER0_DONT_CARE_A + idx; idx 6018 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_get_trace_filter(struct adapter *adap, struct trace_params *tp, int idx, idx 6022 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int i, ofst = idx * 4; idx 6042 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c ofst = (MPS_TRC_FILTER1_MATCH_A - MPS_TRC_FILTER0_MATCH_A) * idx; idx 6335 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_get_port_stats_offset(struct adapter *adap, int idx, idx 6342 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c t4_get_port_stats(adap, idx, stats); idx 6357 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_get_port_stats(struct adapter *adap, int idx, struct port_stats *p) idx 6359 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c u32 bgmap = t4_get_mps_bg_map(adap, idx); idx 6364 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c (is_t4(adap->params.chip) ? PORT_REG(idx, MPS_PORT_STAT_##name##_L) : \ idx 6365 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c T5_PORT_REG(idx, MPS_PORT_STAT_##name##_L))) idx 6454 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_get_lb_stats(struct adapter *adap, int idx, struct lb_port_stats *p) idx 6456 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c u32 bgmap = t4_get_mps_bg_map(adap, idx); idx 6461 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c PORT_REG(idx, MPS_PORT_STAT_LB_PORT_##name##_L) : \ idx 6462 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c T5_PORT_REG(idx, MPS_PORT_STAT_LB_PORT_##name##_L))) idx 7762 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int idx, bool sleep_ok) idx 7781 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c FW_VI_MAC_CMD_IDX_V(idx)); idx 7803 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c const u8 *addr, const u8 *mask, unsigned int idx, idx 7820 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c p->raw_idx_pkd = cpu_to_be32(FW_VI_MAC_CMD_RAW_IDX_V(idx) | idx 7902 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c const u8 *addr, const u8 *mask, unsigned int idx, idx 7919 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c p->raw_idx_pkd = cpu_to_be32(FW_VI_MAC_CMD_RAW_IDX_V(idx)); idx 7935 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (ret != idx) idx 7966 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c const u8 **addr, u16 *idx, u64 *hash, bool sleep_ok) idx 8016 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (idx) idx 8017 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx[offset + i] = (index >= max_naddr ? idx 8131 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int idx, const u8 *addr, bool persist, u8 *smt_idx) idx 8138 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (idx < 0) /* new allocation */ idx 8139 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx = persist ? FW_VI_MAC_ADD_PERSIST_MAC : FW_VI_MAC_ADD_MAC; idx 8149 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c FW_VI_MAC_CMD_IDX_V(idx)); idx 9902 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c unsigned int cfg, val, idx; idx 9918 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx = UPDBGLAWRPTR_G(val); idx 9920 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c *wrptr = idx; idx 9924 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c UPDBGLARDPTR_V(idx) | UPDBGLARDEN_F); idx 9941 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (is_t6(adap->params.chip) && (idx & 0xf) >= 9) idx 9942 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx = (idx & 0xff0) + 0x10; idx 9944 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx++; idx 9946 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx &= UPDBGLARDPTR_M; idx 9971 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c unsigned int i, cfg, val, idx; idx 9979 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx = DBGLAWPTR_G(val); idx 9982 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx = (idx + 1) & DBGLARPTR_M; idx 9984 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c *wrptr = idx; idx 9991 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c t4_write_reg(adap, TP_DBG_LA_CONFIG_A, DBGLARPTR_V(idx) | val); idx 9993 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c idx = (idx + 1) & DBGLARPTR_M; idx 2183 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h __u8 idx; idx 2194 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h __u8 idx; idx 45 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define PF_BASE(idx) (PF0_BASE + (idx) * PF_STRIDE) idx 46 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define PF_REG(idx, reg) (PF_BASE(idx) + (reg)) idx 58 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define PORT_BASE(idx) (PORT0_BASE + (idx) * PORT_STRIDE) idx 59 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define PORT_REG(idx, reg) (PORT_BASE(idx) + (reg)) idx 62 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define EDC_REG(reg, idx) (reg + EDC_STRIDE * idx) idx 64 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define PCIE_MEM_ACCESS_REG(reg_addr, idx) ((reg_addr) + (idx) * 8) idx 65 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define PCIE_MAILBOX_REG(reg_addr, idx) ((reg_addr) + (idx) * 8) idx 66 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define MC_BIST_STATUS_REG(reg_addr, idx) ((reg_addr) + (idx) * 4) idx 67 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define EDC_BIST_STATUS_REG(reg_addr, idx) ((reg_addr) + (idx) * 4) idx 69 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define PCIE_FW_REG(reg_addr, idx) ((reg_addr) + (idx) * 4) idx 2662 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define MPS_CLS_TCAM_Y_L(idx) (MPS_CLS_TCAM_Y_L_A + (idx) * 16) idx 2665 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define MPS_CLS_TCAM_X_L(idx) (MPS_CLS_TCAM_X_L_A + (idx) * 16) idx 2710 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define MPS_CLS_SRAM_L(idx) (MPS_CLS_SRAM_L_A + (idx) * 8) idx 2713 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define MPS_CLS_SRAM_H(idx) (MPS_CLS_SRAM_H_A + (idx) * 8) idx 3175 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define T5_PORT_BASE(idx) (T5_PORT0_BASE + (idx) * T5_PORT_STRIDE) idx 3176 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define T5_PORT_REG(idx, reg) (T5_PORT_BASE(idx) + (reg)) idx 3181 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define MC_REG(reg, idx) (reg + MC_STRIDE * idx) idx 3201 drivers/net/ethernet/chelsio/cxgb4/t4_regs.h #define EDC_T5_REG(reg, idx) (reg + EDC_T5_STRIDE * idx) idx 2812 drivers/net/ethernet/chelsio/cxgb4/t4fw_api.h __u8 idx; idx 182 drivers/net/ethernet/chelsio/cxgb4vf/adapter.h u8 idx; /* queue index within its group */ idx 651 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c rxq->rspq.idx = qs; idx 1588 drivers/net/ethernet/chelsio/cxgb4vf/sge.c skb_record_rx_queue(skb, rxq->rspq.idx); idx 1648 drivers/net/ethernet/chelsio/cxgb4vf/sge.c skb_record_rx_queue(skb, rspq->idx); idx 1528 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c unsigned int naddr, const u8 **addr, u16 *idx, idx 1576 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c if (idx) idx 1577 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c idx[offset+i] = idx 1691 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c int idx, const u8 *addr, bool persist) idx 1704 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c if (idx < 0) idx 1705 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c idx = persist ? FW_VI_MAC_ADD_PERSIST_MAC : FW_VI_MAC_ADD_MAC; idx 1714 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c FW_VI_MAC_CMD_IDX_V(idx)); idx 66 drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.h unsigned int *idx, int use_ts, int ipv6) idx 76 drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.h cxgb4_best_aligned_mtu(mtus, hdr_size, data_size, 8, idx); idx 227 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c void cxgbi_ppm_ppod_release(struct cxgbi_ppm *ppm, u32 idx) idx 231 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (idx >= ppm->ppmax) { idx 232 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pr_warn("ippm: idx too big %u > %u.\n", idx, ppm->ppmax); idx 236 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pdata = ppm->ppod_data + idx; idx 238 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pr_warn("ippm: idx %u, npods 0.\n", idx); idx 242 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pr_debug("release idx %u, npods %u.\n", idx, pdata->npods); idx 243 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c ppm_unmark_entries(ppm, idx, pdata->npods); idx 253 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c int idx = -1; idx 265 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c idx = ppm_get_cpu_entries(ppm, npods, caller_data); idx 267 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (idx < 0) idx 268 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c idx = ppm_get_entries(ppm, npods, caller_data); idx 269 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (idx < 0) { idx 272 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c return idx; idx 275 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pdata = ppm->ppod_data + idx; idx 276 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c hwidx = ppm->base_idx + idx; idx 283 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c *ppod_idx = idx; idx 287 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c nr_pages, tag, idx, npods, caller_data); idx 90 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h unsigned int idx; idx 262 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h u32 idx = cxgbi_ppm_ddp_tag_get_idx(ppm, ddp_tag); idx 264 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h return ppm->ppod_data[idx].caller_data; idx 322 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h void cxgbi_ppm_ppod_release(struct cxgbi_ppm *, u32 idx); idx 1862 drivers/net/ethernet/dec/tulip/de2104x.c unsigned idx; idx 1869 drivers/net/ethernet/dec/tulip/de2104x.c idx = DE_MEDIA_TP; idx 1874 drivers/net/ethernet/dec/tulip/de2104x.c idx = DE_MEDIA_BNC; idx 1878 drivers/net/ethernet/dec/tulip/de2104x.c idx = DE_MEDIA_AUI; idx 1883 drivers/net/ethernet/dec/tulip/de2104x.c idx = DE_MEDIA_TP_FD; idx 1890 drivers/net/ethernet/dec/tulip/de2104x.c de->media[idx].type = idx; idx 1895 drivers/net/ethernet/dec/tulip/de2104x.c media_name[de->media[idx].type]); idx 1900 drivers/net/ethernet/dec/tulip/de2104x.c de->media[idx].csr13 = get_unaligned(&ib->csr13); idx 1901 drivers/net/ethernet/dec/tulip/de2104x.c de->media[idx].csr14 = get_unaligned(&ib->csr14); idx 1902 drivers/net/ethernet/dec/tulip/de2104x.c de->media[idx].csr15 = get_unaligned(&ib->csr15); idx 1908 drivers/net/ethernet/dec/tulip/de2104x.c de->media[idx].csr13, idx 1909 drivers/net/ethernet/dec/tulip/de2104x.c de->media[idx].csr14, idx 1910 drivers/net/ethernet/dec/tulip/de2104x.c de->media[idx].csr15); idx 188 drivers/net/ethernet/emulex/benet/be.h u8 idx; /* array index */ idx 813 drivers/net/ethernet/emulex/benet/be.h for (i = eqo->idx, rxo = &adapter->rx_obj[i]; i < adapter->num_rx_qs;\ idx 817 drivers/net/ethernet/emulex/benet/be.h for (i = eqo->idx, txo = &adapter->tx_obj[i]; i < adapter->num_tx_qs;\ idx 820 drivers/net/ethernet/emulex/benet/be.h #define is_mcc_eqo(eqo) (eqo->idx == 0) idx 1031 drivers/net/ethernet/emulex/benet/be_cmds.c (ver == 2) ? le16_to_cpu(resp->msix_idx) : eqo->idx; idx 2149 drivers/net/ethernet/emulex/benet/be_main.c aic = &adapter->aic_obj[eqo->idx]; idx 2202 drivers/net/ethernet/emulex/benet/be_main.c struct be_aic_obj *aic = &adapter->aic_obj[eqo->idx]; idx 2237 drivers/net/ethernet/emulex/benet/be_main.c aic = &adapter->aic_obj[eqo->idx]; idx 2969 drivers/net/ethernet/emulex/benet/be_main.c eqo->idx = i; idx 3096 drivers/net/ethernet/emulex/benet/be_main.c eqo->idx); idx 3267 drivers/net/ethernet/emulex/benet/be_main.c int idx) idx 3283 drivers/net/ethernet/emulex/benet/be_main.c if (__netif_subqueue_stopped(adapter->netdev, idx) && idx 3285 drivers/net/ethernet/emulex/benet/be_main.c netif_wake_subqueue(adapter->netdev, idx); idx 618 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c static inline void dpaa_assign_wq(struct dpaa_fq *fq, int idx) idx 634 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c switch (idx / DPAA_TC_TXQ_NUM) { idx 1105 drivers/net/ethernet/freescale/enetc/enetc.c int idx = tx_ring->index; idx 1108 drivers/net/ethernet/freescale/enetc/enetc.c enetc_txbdr_wr(hw, idx, ENETC_TBBAR0, idx 1111 drivers/net/ethernet/freescale/enetc/enetc.c enetc_txbdr_wr(hw, idx, ENETC_TBBAR1, idx 1115 drivers/net/ethernet/freescale/enetc/enetc.c enetc_txbdr_wr(hw, idx, ENETC_TBLENR, idx 1119 drivers/net/ethernet/freescale/enetc/enetc.c tx_ring->next_to_use = enetc_txbdr_rd(hw, idx, ENETC_TBPIR); idx 1120 drivers/net/ethernet/freescale/enetc/enetc.c tx_ring->next_to_clean = enetc_txbdr_rd(hw, idx, ENETC_TBCIR); idx 1123 drivers/net/ethernet/freescale/enetc/enetc.c enetc_txbdr_wr(hw, idx, ENETC_TBICIR0, ENETC_TBICIR0_ICEN | 0x1); idx 1130 drivers/net/ethernet/freescale/enetc/enetc.c enetc_txbdr_wr(hw, idx, ENETC_TBMR, tbmr); idx 1132 drivers/net/ethernet/freescale/enetc/enetc.c tx_ring->tpir = hw->reg + ENETC_BDR(TX, idx, ENETC_TBPIR); idx 1133 drivers/net/ethernet/freescale/enetc/enetc.c tx_ring->tcir = hw->reg + ENETC_BDR(TX, idx, ENETC_TBCIR); idx 1139 drivers/net/ethernet/freescale/enetc/enetc.c int idx = rx_ring->index; idx 1142 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBBAR0, idx 1145 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBBAR1, idx 1149 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBLENR, idx 1152 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBBSR, ENETC_RXB_DMA_SIZE); idx 1154 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBPIR, 0); idx 1157 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBICIR0, ENETC_RBICIR0_ICEN | 0x1); idx 1166 drivers/net/ethernet/freescale/enetc/enetc.c rx_ring->rcir = hw->reg + ENETC_BDR(RX, idx, ENETC_RBCIR); idx 1172 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBMR, rbmr); idx 1188 drivers/net/ethernet/freescale/enetc/enetc.c int idx = rx_ring->index; idx 1191 drivers/net/ethernet/freescale/enetc/enetc.c enetc_rxbdr_wr(hw, idx, ENETC_RBMR, 0); idx 1197 drivers/net/ethernet/freescale/enetc/enetc.c int idx = tx_ring->index; idx 1200 drivers/net/ethernet/freescale/enetc/enetc.c enetc_txbdr_wr(hw, idx, ENETC_TBMR, 0); idx 1204 drivers/net/ethernet/freescale/enetc/enetc.c enetc_txbdr_rd(hw, idx, ENETC_TBSR) & ENETC_TBSR_BUSY) { idx 1211 drivers/net/ethernet/freescale/enetc/enetc.c idx); idx 1253 drivers/net/ethernet/freescale/enetc/enetc.c int idx = v->tx_ring[j].index; idx 1255 drivers/net/ethernet/freescale/enetc/enetc.c enetc_wr(hw, ENETC_SIMSITRV(idx), entry); idx 1644 drivers/net/ethernet/freescale/enetc/enetc.c int idx; idx 1648 drivers/net/ethernet/freescale/enetc/enetc.c idx = 2 * j + i; /* 2 CPUs */ idx 1650 drivers/net/ethernet/freescale/enetc/enetc.c idx = j + i * v_tx_rings; /* default */ idx 1652 drivers/net/ethernet/freescale/enetc/enetc.c __set_bit(idx, &v->tx_rings_map); idx 1654 drivers/net/ethernet/freescale/enetc/enetc.c bdr->index = idx; idx 1658 drivers/net/ethernet/freescale/enetc/enetc.c priv->tx_ring[idx] = bdr; idx 64 drivers/net/ethernet/freescale/enetc/enetc_msg.c static int enetc_msg_alloc_mbx(struct enetc_si *si, int idx) idx 72 drivers/net/ethernet/freescale/enetc/enetc_msg.c msg = &pf->rxmsg[idx]; idx 86 drivers/net/ethernet/freescale/enetc/enetc_msg.c enetc_wr(hw, ENETC_PSIVMSGRCVAR0(idx), val); idx 88 drivers/net/ethernet/freescale/enetc/enetc_msg.c enetc_wr(hw, ENETC_PSIVMSGRCVAR1(idx), val); idx 93 drivers/net/ethernet/freescale/enetc/enetc_msg.c static void enetc_msg_free_mbx(struct enetc_si *si, int idx) idx 99 drivers/net/ethernet/freescale/enetc/enetc_msg.c msg = &pf->rxmsg[idx]; idx 103 drivers/net/ethernet/freescale/enetc/enetc_msg.c enetc_wr(hw, ENETC_PSIVMSGRCVAR0(idx), 0); idx 104 drivers/net/ethernet/freescale/enetc/enetc_msg.c enetc_wr(hw, ENETC_PSIVMSGRCVAR1(idx), 0); idx 131 drivers/net/ethernet/freescale/enetc/enetc_pf.c int idx = enetc_mac_addr_hash_idx(addr); idx 134 drivers/net/ethernet/freescale/enetc/enetc_pf.c __set_bit(idx, filter->mac_hash_table); idx 313 drivers/net/ethernet/freescale/enetc/enetc_pf.c int idx; idx 320 drivers/net/ethernet/freescale/enetc/enetc_pf.c idx = enetc_vid_hash_idx(vid); idx 321 drivers/net/ethernet/freescale/enetc/enetc_pf.c if (!__test_and_set_bit(idx, pf->vlan_ht_filter)) idx 160 drivers/net/ethernet/freescale/fs_enet/fs_enet.h int idx; /* FEC1 = 0, FEC2 = 1 */ idx 166 drivers/net/ethernet/freescale/fs_enet/fs_enet.h int idx; /* FCC1-3 = 0-2 */ idx 175 drivers/net/ethernet/freescale/fs_enet/fs_enet.h int idx; /* FEC1 = 0, FEC2 = 1 */ idx 3054 drivers/net/ethernet/freescale/gianfar.c int idx; idx 3057 drivers/net/ethernet/freescale/gianfar.c for (idx = 1; idx < GFAR_EM_NUM + 1; idx++) idx 3058 drivers/net/ethernet/freescale/gianfar.c gfar_set_mac_for_addr(dev, idx, zero_arr); idx 3105 drivers/net/ethernet/freescale/gianfar.c int idx; idx 3132 drivers/net/ethernet/freescale/gianfar.c idx = 1; idx 3134 drivers/net/ethernet/freescale/gianfar.c idx = 0; idx 3143 drivers/net/ethernet/freescale/gianfar.c if (idx < em_num) { idx 3144 drivers/net/ethernet/freescale/gianfar.c gfar_set_mac_for_addr(dev, idx, ha->addr); idx 3145 drivers/net/ethernet/freescale/gianfar.c idx++; idx 794 drivers/net/ethernet/google/gve/gve_main.c int idx; idx 803 drivers/net/ethernet/google/gve/gve_main.c for (idx = 0; idx < priv->tx_cfg.num_queues; idx++) { idx 804 drivers/net/ethernet/google/gve/gve_main.c int ntfy_idx = gve_tx_idx_to_ntfy(priv, idx); idx 809 drivers/net/ethernet/google/gve/gve_main.c for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { idx 810 drivers/net/ethernet/google/gve/gve_main.c int ntfy_idx = gve_rx_idx_to_ntfy(priv, idx); idx 824 drivers/net/ethernet/google/gve/gve_main.c int idx; idx 830 drivers/net/ethernet/google/gve/gve_main.c for (idx = 0; idx < priv->tx_cfg.num_queues; idx++) { idx 831 drivers/net/ethernet/google/gve/gve_main.c int ntfy_idx = gve_tx_idx_to_ntfy(priv, idx); idx 837 drivers/net/ethernet/google/gve/gve_main.c for (idx = 0; idx < priv->rx_cfg.num_queues; idx++) { idx 838 drivers/net/ethernet/google/gve/gve_main.c int ntfy_idx = gve_rx_idx_to_ntfy(priv, idx); idx 19 drivers/net/ethernet/google/gve/gve_rx.c static void gve_rx_free_ring(struct gve_priv *priv, int idx) idx 21 drivers/net/ethernet/google/gve/gve_rx.c struct gve_rx_ring *rx = &priv->rx[idx]; idx 26 drivers/net/ethernet/google/gve/gve_rx.c gve_rx_remove_from_block(priv, idx); idx 45 drivers/net/ethernet/google/gve/gve_rx.c netif_dbg(priv, drv, priv->dev, "freed rx ring %d\n", idx); idx 97 drivers/net/ethernet/google/gve/gve_rx.c static int gve_rx_alloc_ring(struct gve_priv *priv, int idx) idx 99 drivers/net/ethernet/google/gve/gve_rx.c struct gve_rx_ring *rx = &priv->rx[idx]; idx 111 drivers/net/ethernet/google/gve/gve_rx.c rx->q_num = idx; idx 142 drivers/net/ethernet/google/gve/gve_rx.c netif_dbg(priv, drv, priv->dev, "rx[%d]->data.data_bus=%lx\n", idx, idx 162 drivers/net/ethernet/google/gve/gve_rx.c gve_rx_add_to_block(priv, idx); idx 276 drivers/net/ethernet/google/gve/gve_rx.c netdev_features_t feat, u32 idx) idx 291 drivers/net/ethernet/google/gve/gve_rx.c page_info = &rx->data.page_info[idx]; idx 292 drivers/net/ethernet/google/gve/gve_rx.c dma_sync_single_for_cpu(&priv->pdev->dev, rx->data.qpl->page_buses[idx], idx 322 drivers/net/ethernet/google/gve/gve_rx.c gve_rx_flip_buff(page_info, &rx->data.data_ring[idx]); idx 387 drivers/net/ethernet/google/gve/gve_rx.c u32 idx = cnt & rx->mask; idx 391 drivers/net/ethernet/google/gve/gve_rx.c desc = rx->desc.desc_ring + idx; idx 396 drivers/net/ethernet/google/gve/gve_rx.c rx->q_num, idx, desc, desc->flags_seq); idx 402 drivers/net/ethernet/google/gve/gve_rx.c if (!gve_rx(rx, desc, feat, idx)) idx 405 drivers/net/ethernet/google/gve/gve_rx.c idx = cnt & rx->mask; idx 406 drivers/net/ethernet/google/gve/gve_rx.c desc = rx->desc.desc_ring + idx; idx 145 drivers/net/ethernet/google/gve/gve_tx.c static void gve_tx_free_ring(struct gve_priv *priv, int idx) idx 147 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx = &priv->tx[idx]; idx 152 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_remove_from_block(priv, idx); idx 172 drivers/net/ethernet/google/gve/gve_tx.c netif_dbg(priv, drv, priv->dev, "freed tx queue %d\n", idx); idx 185 drivers/net/ethernet/google/gve/gve_tx.c static int gve_tx_alloc_ring(struct gve_priv *priv, int idx) idx 187 drivers/net/ethernet/google/gve/gve_tx.c struct gve_tx_ring *tx = &priv->tx[idx]; idx 194 drivers/net/ethernet/google/gve/gve_tx.c tx->q_num = idx; idx 223 drivers/net/ethernet/google/gve/gve_tx.c netif_dbg(priv, drv, priv->dev, "tx[%d]->bus=%lx\n", idx, idx 225 drivers/net/ethernet/google/gve/gve_tx.c tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx); idx 226 drivers/net/ethernet/google/gve/gve_tx.c gve_tx_add_to_block(priv, idx); idx 414 drivers/net/ethernet/google/gve/gve_tx.c u32 idx = tx->req & tx->mask; idx 420 drivers/net/ethernet/google/gve/gve_tx.c info = &tx->info[idx]; idx 421 drivers/net/ethernet/google/gve/gve_tx.c pkt_desc = &tx->desc[idx]; idx 518 drivers/net/ethernet/google/gve/gve_tx.c u32 idx; idx 521 drivers/net/ethernet/google/gve/gve_tx.c idx = tx->done & tx->mask; idx 524 drivers/net/ethernet/google/gve/gve_tx.c tx->q_num, __func__, idx, tx->req, tx->done); idx 525 drivers/net/ethernet/google/gve/gve_tx.c info = &tx->info[idx]; idx 310 drivers/net/ethernet/hisilicon/hns/hnae.h #define assert_is_ring_idx(ring, idx) \ idx 311 drivers/net/ethernet/hisilicon/hns/hnae.h assert((idx) >= 0 && (idx) < (ring)->desc_num) idx 645 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c int idx; idx 662 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c for (idx = 0; idx < handle->q_num; idx++) { idx 663 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c queue = handle->qs[idx]; idx 726 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c int idx; idx 741 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c for (idx = 0; idx < handle->q_num; idx++) { idx 742 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c hns_rcb_get_stats(handle->qs[idx], p); idx 760 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c int idx; idx 774 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c for (idx = 0; idx < handle->q_num; idx++) { idx 775 drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c hns_rcb_get_strings(stringset, p, idx); idx 1168 drivers/net/ethernet/hisilicon/hns/hns_enet.c static int hns_nic_ring_open(struct net_device *netdev, int idx) idx 1173 drivers/net/ethernet/hisilicon/hns/hns_enet.c napi_enable(&priv->ring_data[idx].napi); idx 1175 drivers/net/ethernet/hisilicon/hns/hns_enet.c enable_irq(priv->ring_data[idx].ring->irq); idx 1176 drivers/net/ethernet/hisilicon/hns/hns_enet.c h->dev->ops->toggle_ring_irq(priv->ring_data[idx].ring, 0); idx 1222 drivers/net/ethernet/hisilicon/hns/hns_enet.c static void hns_nic_ring_close(struct net_device *netdev, int idx) idx 1227 drivers/net/ethernet/hisilicon/hns/hns_enet.c h->dev->ops->toggle_ring_irq(priv->ring_data[idx].ring, 1); idx 1228 drivers/net/ethernet/hisilicon/hns/hns_enet.c disable_irq(priv->ring_data[idx].ring->irq); idx 1230 drivers/net/ethernet/hisilicon/hns/hns_enet.c napi_disable(&priv->ring_data[idx].napi); idx 1897 drivers/net/ethernet/hisilicon/hns/hns_enet.c int idx = 0; idx 1905 drivers/net/ethernet/hisilicon/hns/hns_enet.c for (idx = 0; idx < h->q_num; idx++) { idx 1906 drivers/net/ethernet/hisilicon/hns/hns_enet.c tx_bytes += h->qs[idx]->tx_ring.stats.tx_bytes; idx 1907 drivers/net/ethernet/hisilicon/hns/hns_enet.c tx_pkts += h->qs[idx]->tx_ring.stats.tx_pkts; idx 1908 drivers/net/ethernet/hisilicon/hns/hns_enet.c rx_bytes += h->qs[idx]->rx_ring.stats.rx_bytes; idx 1909 drivers/net/ethernet/hisilicon/hns/hns_enet.c rx_pkts += h->qs[idx]->rx_ring.stats.rx_pkts; idx 83 drivers/net/ethernet/hisilicon/hns/hns_enet.h #define tx_ring_data(priv, idx) ((priv)->ring_data[idx]) idx 84 drivers/net/ethernet/hisilicon/hns/hns_enet.h #define rx_ring_data(priv, idx) \ idx 85 drivers/net/ethernet/hisilicon/hns/hns_enet.h ((priv)->ring_data[(priv)->ae_handle->q_num + (idx)]) idx 1499 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c unsigned int idx; idx 1512 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c for (idx = 0; idx < queue_num; idx++) { idx 1514 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ring = priv->ring_data[idx].ring; idx 1532 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c ring = priv->ring_data[idx + queue_num].ring; idx 3418 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c tqp_vector->idx = i; idx 483 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h u16 idx; /* index in the TQP vector array per handle. */ idx 619 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h #define tx_ring_data(priv, idx) ((priv)->ring_data[idx]) idx 2118 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c u32 idx = i * HCLGE_TC_NUM_ONE_DESC + j; idx 2120 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c priv = &buf_alloc->priv_buf[idx]; idx 10114 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c unsigned int idx = 0; idx 10116 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c for (; idx < size; idx++) { idx 10117 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c supported[idx] = hdev->hw.mac.supported[idx]; idx 10118 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c advertising[idx] = hdev->hw.mac.advertising[idx]; idx 259 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c u8 idx; idx 290 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c idx = (u8)msg_q[1]; idx 291 drivers/net/ethernet/hisilicon/hns3/hns3vf/hclgevf_mbx.c if (idx) idx 2079 drivers/net/ethernet/hp/hp100.c int i, idx; idx 2094 drivers/net/ethernet/hp/hp100.c for (i = idx = 0; i < 6; i++) { idx 2095 drivers/net/ethernet/hp/hp100.c idx ^= *addrs++ & 0x3f; idx 2096 drivers/net/ethernet/hp/hp100.c printk(":%02x:", idx); idx 2099 drivers/net/ethernet/hp/hp100.c printk("idx = %i\n", idx); idx 2101 drivers/net/ethernet/hp/hp100.c lp->hash_bytes[idx >> 3] |= (1 << (idx & 7)); idx 49 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c #define MASKED_IDX(chain, idx) ((idx) & ((chain)->num_cells - 1)) idx 78 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c int idx; idx 83 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c for (idx = 0; idx < 7; idx++) idx 84 drivers/net/ethernet/huawei/hinic/hinic_hw_api_cmd.c checksum ^= val[idx]; idx 21 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_DMA_ATTR_ADDR(idx) \ idx 22 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_DMA_ATTR_BASE + (idx) * HINIC_DMA_ATTR_STRIDE) idx 27 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_PPF_ELECTION_ADDR(idx) \ idx 28 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_ELECTION_BASE + (idx) * HINIC_PPF_ELECTION_STRIDE) idx 35 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_CHAIN_HEAD_HI_ADDR(idx) \ idx 36 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x0 + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 38 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_CHAIN_HEAD_LO_ADDR(idx) \ idx 39 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x4 + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 41 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_STATUS_HI_ADDR(idx) \ idx 42 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x8 + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 44 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_STATUS_LO_ADDR(idx) \ idx 45 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0xC + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 47 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_CHAIN_NUM_CELLS_ADDR(idx) \ idx 48 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x10 + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 50 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_CHAIN_CTRL_ADDR(idx) \ idx 51 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x14 + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 53 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_CHAIN_PI_ADDR(idx) \ idx 54 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x1C + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 56 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_CHAIN_REQ_ADDR(idx) \ idx 57 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x20 + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 59 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_API_CMD_STATUS_ADDR(idx) \ idx 60 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_API_CMD_BASE + 0x30 + (idx) * HINIC_CSR_API_CMD_STRIDE) idx 68 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_MSIX_CTRL_ADDR(idx) \ idx 69 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_MSIX_CTRL_BASE + (idx) * HINIC_CSR_MSIX_STRIDE) idx 71 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_MSIX_CNT_ADDR(idx) \ idx 72 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CSR_MSIX_CNT_BASE + (idx) * HINIC_CSR_MSIX_STRIDE) idx 116 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_AEQ_CTRL_0_ADDR(idx) \ idx 117 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_AEQ_CTRL_0_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 119 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_AEQ_CTRL_1_ADDR(idx) \ idx 120 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_AEQ_CTRL_1_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 122 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_AEQ_CONS_IDX_ADDR(idx) \ idx 123 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_AEQ_CONS_IDX_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 125 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_AEQ_PROD_IDX_ADDR(idx) \ idx 126 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_AEQ_PROD_IDX_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 128 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_CEQ_CTRL_0_ADDR(idx) \ idx 129 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CEQ_CTRL_0_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 131 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_CEQ_CTRL_1_ADDR(idx) \ idx 132 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CEQ_CTRL_1_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 134 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_CEQ_CONS_IDX_ADDR(idx) \ idx 135 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CEQ_CONS_IDX_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 137 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h #define HINIC_CSR_CEQ_PROD_IDX_ADDR(idx) \ idx 138 drivers/net/ethernet/huawei/hinic/hinic_hw_csr.h (HINIC_CEQ_PROD_IDX_ADDR_BASE + (idx) * HINIC_EQ_OFF_STRIDE) idx 47 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_EQ_ELEMENT(eq, idx) \ idx 48 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c ((eq)->virt_addr[(idx) / (eq)->num_elem_in_pg] + \ idx 49 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c (((idx) & ((eq)->num_elem_in_pg - 1)) * (eq)->elem_size)) idx 51 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_AEQ_ELEM(eq, idx) ((struct hinic_aeq_elem *) \ idx 52 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c GET_EQ_ELEMENT(eq, idx)) idx 54 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c #define GET_CEQ_ELEM(eq, idx) ((u32 *) \ idx 55 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c GET_EQ_ELEMENT(eq, idx)) idx 179 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c int idx; idx 181 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c for (idx = 0; idx < 32; idx += 4) idx 182 drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c checksum ^= ((val >> idx) & 0xF); idx 60 drivers/net/ethernet/huawei/hinic/hinic_hw_io.c int pos, idx; idx 75 drivers/net/ethernet/huawei/hinic/hinic_hw_io.c idx = free_db_area->db_idx[pos]; idx 81 drivers/net/ethernet/huawei/hinic/hinic_hw_io.c return func_to_io->db_base + idx * HINIC_DB_PAGE_SIZE; idx 88 drivers/net/ethernet/huawei/hinic/hinic_hw_io.c int pos, idx = DB_IDX(db_base, func_to_io->db_base); idx 95 drivers/net/ethernet/huawei/hinic/hinic_hw_io.c free_db_area->db_idx[pos] = idx; idx 61 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c #define SQ_MASKED_IDX(sq, idx) ((idx) & (sq)->wq->mask) idx 62 drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c #define RQ_MASKED_IDX(rq, idx) ((idx) & (rq)->wq->mask) idx 68 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c #define WQ_PAGE_ADDR(wq, idx) \ idx 69 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c ((wq)->shadow_block_vaddr[WQE_PAGE_NUM(wq, idx)]) idx 71 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c #define MASKED_WQE_IDX(wq, idx) ((idx) & (wq)->mask) idx 81 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c static inline int WQE_PAGE_OFF(struct hinic_wq *wq, u16 idx) idx 83 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c return (((idx) & ((wq)->num_wqebbs_per_page - 1)) idx 87 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c static inline int WQE_PAGE_NUM(struct hinic_wq *wq, u16 idx) idx 89 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c return (((idx) >> ((wq)->wqebbs_per_page_shift)) idx 702 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c int num_wqebbs, u16 idx) idx 707 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c for (i = 0; i < num_wqebbs; i++, idx++) { idx 708 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c idx = MASKED_WQE_IDX(wq, idx); idx 709 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c wqebb_addr = WQ_PAGE_ADDR(wq, idx) + idx 710 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c WQE_PAGE_OFF(wq, idx); idx 719 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c int num_wqebbs, u16 idx) idx 724 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c for (i = 0; i < num_wqebbs; i++, idx++) { idx 725 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c idx = MASKED_WQE_IDX(wq, idx); idx 726 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c wqebb_addr = WQ_PAGE_ADDR(wq, idx) + idx 727 drivers/net/ethernet/huawei/hinic/hinic_hw_wq.c WQE_PAGE_OFF(wq, idx); idx 572 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int top, dir, idx; idx 589 drivers/net/ethernet/ibm/ehea/ehea_qmr.c for (idx = 0; idx < EHEA_MAP_ENTRIES; idx++) { idx 590 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (!ehea_dir->ent[idx]) idx 593 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ehea_dir->ent[idx] = vaddr; idx 628 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int idx = i & EHEA_INDEX_MASK; idx 645 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ehea_bmap->top[top]->dir[dir]->ent[idx] = flag; idx 762 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int top, dir, idx; idx 777 drivers/net/ethernet/ibm/ehea/ehea_qmr.c idx = index & EHEA_INDEX_MASK; idx 778 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (!ehea_bmap->top[top]->dir[dir]->ent[idx]) idx 782 drivers/net/ethernet/ibm/ehea/ehea_qmr.c return ehea_bmap->top[top]->dir[dir]->ent[idx] | offset; idx 785 drivers/net/ethernet/ibm/ehea/ehea_qmr.c static inline void *ehea_calc_sectbase(int top, int dir, int idx) idx 787 drivers/net/ethernet/ibm/ehea/ehea_qmr.c unsigned long ret = idx; idx 793 drivers/net/ethernet/ibm/ehea/ehea_qmr.c static u64 ehea_reg_mr_section(int top, int dir, int idx, u64 *pt, idx 802 drivers/net/ethernet/ibm/ehea/ehea_qmr.c void *sectbase = ehea_calc_sectbase(top, dir, idx); idx 829 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int idx; idx 831 drivers/net/ethernet/ibm/ehea/ehea_qmr.c for (idx = 0; idx < EHEA_MAP_ENTRIES; idx++) { idx 832 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (!ehea_bmap->top[top]->dir[dir]->ent[idx]) idx 835 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hret = ehea_reg_mr_section(top, dir, idx, pt, adapter, mr); idx 3332 drivers/net/ethernet/ibm/emac/core.c const u32 *idx; idx 3338 drivers/net/ethernet/ibm/emac/core.c idx = of_get_property(np, "cell-index", NULL); idx 3339 drivers/net/ethernet/ibm/emac/core.c if (idx == NULL) idx 3341 drivers/net/ethernet/ibm/emac/core.c cell_indices[i] = *idx; idx 34 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_FER_MASK(idx) (0x7 << ((idx) * 4)) idx 35 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_FER_RTBI(idx) (0x4 << ((idx) * 4)) idx 36 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_FER_RGMII(idx) (0x5 << ((idx) * 4)) idx 37 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_FER_TBI(idx) (0x6 << ((idx) * 4)) idx 38 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_FER_GMII(idx) (0x7 << ((idx) * 4)) idx 39 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_FER_MII(idx) RGMII_FER_GMII(idx) idx 42 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_SSR_MASK(idx) (0x7 << ((idx) * 8)) idx 43 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_SSR_10(idx) (0x1 << ((idx) * 8)) idx 44 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_SSR_100(idx) (0x2 << ((idx) * 8)) idx 45 drivers/net/ethernet/ibm/emac/rgmii.c #define RGMII_SSR_1000(idx) (0x4 << ((idx) * 8)) idx 29 drivers/net/ethernet/ibm/emac/zmii.c #define ZMII_FER_MDI(idx) (0x80000000 >> ((idx) * 4)) idx 33 drivers/net/ethernet/ibm/emac/zmii.c #define ZMII_FER_SMII(idx) (0x40000000 >> ((idx) * 4)) idx 34 drivers/net/ethernet/ibm/emac/zmii.c #define ZMII_FER_RMII(idx) (0x20000000 >> ((idx) * 4)) idx 35 drivers/net/ethernet/ibm/emac/zmii.c #define ZMII_FER_MII(idx) (0x10000000 >> ((idx) * 4)) idx 38 drivers/net/ethernet/ibm/emac/zmii.c #define ZMII_SSR_SCI(idx) (0x40000000 >> ((idx) * 4)) idx 39 drivers/net/ethernet/ibm/emac/zmii.c #define ZMII_SSR_FSS(idx) (0x20000000 >> ((idx) * 4)) idx 40 drivers/net/ethernet/ibm/emac/zmii.c #define ZMII_SSR_SP(idx) (0x10000000 >> ((idx) * 4)) idx 330 drivers/net/ethernet/intel/fm10k/fm10k_common.c u32 idx) idx 336 drivers/net/ethernet/intel/fm10k/fm10k_common.c id_tx = fm10k_read_reg(hw, FM10K_TXQCTL(idx)); idx 340 drivers/net/ethernet/intel/fm10k/fm10k_common.c tx_packets = fm10k_read_hw_stats_32b(hw, FM10K_QPTC(idx), idx 345 drivers/net/ethernet/intel/fm10k/fm10k_common.c FM10K_QBTC_L(idx), idx 350 drivers/net/ethernet/intel/fm10k/fm10k_common.c id_tx = fm10k_read_reg(hw, FM10K_TXQCTL(idx)); idx 381 drivers/net/ethernet/intel/fm10k/fm10k_common.c u32 idx) idx 387 drivers/net/ethernet/intel/fm10k/fm10k_common.c id_rx = fm10k_read_reg(hw, FM10K_RXQCTL(idx)); idx 391 drivers/net/ethernet/intel/fm10k/fm10k_common.c rx_drops = fm10k_read_hw_stats_32b(hw, FM10K_QPRDC(idx), idx 394 drivers/net/ethernet/intel/fm10k/fm10k_common.c rx_packets = fm10k_read_hw_stats_32b(hw, FM10K_QPRC(idx), idx 399 drivers/net/ethernet/intel/fm10k/fm10k_common.c FM10K_QBRC_L(idx), idx 404 drivers/net/ethernet/intel/fm10k/fm10k_common.c id_rx = fm10k_read_reg(hw, FM10K_RXQCTL(idx)); idx 437 drivers/net/ethernet/intel/fm10k/fm10k_common.c u32 idx, u32 count) idx 441 drivers/net/ethernet/intel/fm10k/fm10k_common.c for (i = 0; i < count; i++, idx++, q++) { idx 442 drivers/net/ethernet/intel/fm10k/fm10k_common.c fm10k_update_hw_stats_tx_q(hw, q, idx); idx 443 drivers/net/ethernet/intel/fm10k/fm10k_common.c fm10k_update_hw_stats_rx_q(hw, q, idx); idx 456 drivers/net/ethernet/intel/fm10k/fm10k_common.c void fm10k_unbind_hw_stats_q(struct fm10k_hw_stats_q *q, u32 idx, u32 count) idx 460 drivers/net/ethernet/intel/fm10k/fm10k_common.c for (i = 0; i < count; i++, idx++, q++) { idx 44 drivers/net/ethernet/intel/fm10k/fm10k_common.h u32 idx, u32 count); idx 46 drivers/net/ethernet/intel/fm10k/fm10k_common.h void fm10k_unbind_hw_stats_q(struct fm10k_hw_stats_q *q, u32 idx, u32 count); idx 300 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c int idx = 0; idx 302 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RDBAL(i)); idx 303 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RDBAH(i)); idx 304 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RDLEN(i)); idx 305 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TPH_RXCTRL(i)); idx 306 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RDH(i)); idx 307 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RDT(i)); idx 308 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RXQCTL(i)); idx 309 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RXDCTL(i)); idx 310 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RXINT(i)); idx 311 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_SRRCTL(i)); idx 312 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_QPRC(i)); idx 313 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_QPRDC(i)); idx 314 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_QBRC_L(i)); idx 315 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_QBRC_H(i)); idx 316 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TDBAL(i)); idx 317 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TDBAH(i)); idx 318 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TDLEN(i)); idx 319 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TPH_TXCTRL(i)); idx 320 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TDH(i)); idx 321 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TDT(i)); idx 322 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TXDCTL(i)); idx 323 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TXQCTL(i)); idx 324 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TXINT(i)); idx 325 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_QPTC(i)); idx 326 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_QBTC_L(i)); idx 327 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_QBTC_H(i)); idx 328 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TQDLOC(i)); idx 329 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_TX_SGLORT(i)); idx 330 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_PFVTCTL(i)); idx 332 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c BUG_ON(idx != FM10K_REGS_LEN_Q); idx 340 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c int idx = 0, j; idx 342 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_MRQC(i)); idx 344 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RSSRK(i, j)); idx 346 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c buff[idx++] = fm10k_read_reg(hw, FM10K_RETA(i, j)); idx 348 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c BUG_ON(idx != FM10K_REGS_LEN_VSI); idx 352 drivers/net/ethernet/intel/fm10k/fm10k_iov.c dglort.idx = fm10k_dglort_vf_rss; idx 423 drivers/net/ethernet/intel/fm10k/fm10k_main.c u16 idx = le16_to_cpu(FM10K_CB(skb)->fi.w.dglort) - 1; idx 425 drivers/net/ethernet/intel/fm10k/fm10k_main.c idx -= l2_accel->dglort; idx 426 drivers/net/ethernet/intel/fm10k/fm10k_main.c if (idx < l2_accel->size && l2_accel->macvlan[idx]) idx 427 drivers/net/ethernet/intel/fm10k/fm10k_main.c dev = l2_accel->macvlan[idx]; idx 1526 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c dglort.idx = fm10k_dglort_pf_rss; idx 1605 drivers/net/ethernet/intel/fm10k/fm10k_netdev.c dglort.idx = fm10k_dglort_pf_rss; idx 1131 drivers/net/ethernet/intel/fm10k/fm10k_pci.c dglort.idx = fm10k_dglort_pf_queue; idx 1143 drivers/net/ethernet/intel/fm10k/fm10k_pci.c dglort.idx = fm10k_dglort_pf_rss; idx 510 drivers/net/ethernet/intel/fm10k/fm10k_pf.c if ((dglort->idx > 7) || (dglort->rss_l > 7) || (dglort->pc_l > 3) || idx 563 drivers/net/ethernet/intel/fm10k/fm10k_pf.c dglortmap = (dglort->idx == fm10k_dglort_default) ? idx 569 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fm10k_write_reg(hw, FM10K_DGLORTDEC(dglort->idx), dglortdec); idx 570 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fm10k_write_reg(hw, FM10K_DGLORTMAP(dglort->idx), dglortmap); idx 1137 drivers/net/ethernet/intel/fm10k/fm10k_pf.c u32 idx, qpp; idx 1141 drivers/net/ethernet/intel/fm10k/fm10k_pf.c idx = fm10k_vf_queue_index(hw, vf_idx); idx 1142 drivers/net/ethernet/intel/fm10k/fm10k_pf.c fm10k_update_hw_stats_q(hw, q, idx, qpp); idx 458 drivers/net/ethernet/intel/fm10k/fm10k_type.h u8 idx; /* index of DGLORTDEC entry */ idx 373 drivers/net/ethernet/intel/i40e/i40e.h #define I40E_ORT_SET_IDX(idx) (((idx) << \ idx 385 drivers/net/ethernet/intel/i40e/i40e.h #define I40E_ORT_PREP_VAL(idx, count, payload) (I40E_ORT_SET_IDX(idx) | \ idx 708 drivers/net/ethernet/intel/i40e/i40e.h u16 idx; idx 818 drivers/net/ethernet/intel/i40e/i40e.h u16 idx; /* index in pf->vsi[] */ idx 2889 drivers/net/ethernet/intel/i40e/i40e_adminq_cmd.h __le32 idx; idx 1439 drivers/net/ethernet/intel/i40e/i40e_common.c static u32 i40e_led_is_mine(struct i40e_hw *hw, int idx) idx 1444 drivers/net/ethernet/intel/i40e/i40e_common.c if (!hw->func_caps.led[idx]) idx 1447 drivers/net/ethernet/intel/i40e/i40e_common.c gpio_val = rd32(hw, I40E_GLGEN_GPIO_CTL(idx)); idx 4554 drivers/net/ethernet/intel/i40e/i40e_common.c cmd->idx = cpu_to_le32(start_index); idx 4565 drivers/net/ethernet/intel/i40e/i40e_common.c *ret_next_index = le32_to_cpu(resp->idx); idx 419 drivers/net/ethernet/intel/i40e/i40e_debugfs.c dev_info(&pf->pdev->dev, " idx = %d\n", vsi->idx); idx 628 drivers/net/ethernet/intel/i40e/i40e_debugfs.c veb->idx, veb->veb_idx, veb->stats_idx, veb->seid, idx 190 drivers/net/ethernet/intel/i40e/i40e_hmc.c u32 idx) idx 200 drivers/net/ethernet/intel/i40e/i40e_hmc.c sd_idx = idx / I40E_HMC_PD_CNT_IN_SD; idx 201 drivers/net/ethernet/intel/i40e/i40e_hmc.c rel_pd_idx = idx % I40E_HMC_PD_CNT_IN_SD; idx 226 drivers/net/ethernet/intel/i40e/i40e_hmc.c I40E_INVALIDATE_PF_HMC_PD(hw, sd_idx, idx); idx 245 drivers/net/ethernet/intel/i40e/i40e_hmc.c u32 idx) idx 251 drivers/net/ethernet/intel/i40e/i40e_hmc.c sd_entry = &hmc_info->sd_table.sd_entry[idx]; idx 274 drivers/net/ethernet/intel/i40e/i40e_hmc.c u32 idx, bool is_pf) idx 282 drivers/net/ethernet/intel/i40e/i40e_hmc.c sd_entry = &hmc_info->sd_table.sd_entry[idx]; idx 283 drivers/net/ethernet/intel/i40e/i40e_hmc.c I40E_CLEAR_PF_SD_ENTRY(hw, idx, I40E_SD_TYPE_DIRECT); idx 294 drivers/net/ethernet/intel/i40e/i40e_hmc.c u32 idx) idx 299 drivers/net/ethernet/intel/i40e/i40e_hmc.c sd_entry = &hmc_info->sd_table.sd_entry[idx]; idx 323 drivers/net/ethernet/intel/i40e/i40e_hmc.c u32 idx, bool is_pf) idx 330 drivers/net/ethernet/intel/i40e/i40e_hmc.c sd_entry = &hmc_info->sd_table.sd_entry[idx]; idx 331 drivers/net/ethernet/intel/i40e/i40e_hmc.c I40E_CLEAR_PF_SD_ENTRY(hw, idx, I40E_SD_TYPE_PAGED); idx 180 drivers/net/ethernet/intel/i40e/i40e_hmc.h #define I40E_FIND_PD_INDEX_LIMIT(hmc_info, type, idx, cnt, pd_index, pd_limit)\ idx 184 drivers/net/ethernet/intel/i40e/i40e_hmc.h (hmc_info)->hmc_obj[(type)].size * (idx); \ idx 203 drivers/net/ethernet/intel/i40e/i40e_hmc.h u32 idx); idx 205 drivers/net/ethernet/intel/i40e/i40e_hmc.h u32 idx); idx 208 drivers/net/ethernet/intel/i40e/i40e_hmc.h u32 idx, bool is_pf); idx 210 drivers/net/ethernet/intel/i40e/i40e_hmc.h u32 idx); idx 213 drivers/net/ethernet/intel/i40e/i40e_hmc.h u32 idx, bool is_pf); idx 234 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c u32 idx) idx 238 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c if (!i40e_prep_remove_pd_page(hmc_info, idx)) idx 239 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c ret_code = i40e_remove_pd_page_new(hw, hmc_info, idx, true); idx 261 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c u32 idx) idx 265 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c if (!i40e_prep_remove_sd_bp(hmc_info, idx)) idx 266 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c ret_code = i40e_remove_sd_bp_new(hw, hmc_info, idx, true); idx 693 drivers/net/ethernet/intel/i40e/i40e_main.c int i, idx = 0; idx 695 drivers/net/ethernet/intel/i40e/i40e_main.c idx = veb->stats_idx; idx 702 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update32(hw, I40E_GLSW_TDPC(idx), idx 706 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update32(hw, I40E_GLSW_RUPP(idx), idx 710 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_GORCH(idx), I40E_GLSW_GORCL(idx), idx 713 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_UPRCH(idx), I40E_GLSW_UPRCL(idx), idx 716 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_MPRCH(idx), I40E_GLSW_MPRCL(idx), idx 719 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_BPRCH(idx), I40E_GLSW_BPRCL(idx), idx 723 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_GOTCH(idx), I40E_GLSW_GOTCL(idx), idx 726 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_UPTCH(idx), I40E_GLSW_UPTCL(idx), idx 729 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_MPTCH(idx), I40E_GLSW_MPTCL(idx), idx 732 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLSW_BPTCH(idx), I40E_GLSW_BPTCL(idx), idx 736 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLVEBTC_RPCH(i, idx), idx 737 drivers/net/ethernet/intel/i40e/i40e_main.c I40E_GLVEBTC_RPCL(i, idx), idx 741 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLVEBTC_RBCH(i, idx), idx 742 drivers/net/ethernet/intel/i40e/i40e_main.c I40E_GLVEBTC_RBCL(i, idx), idx 746 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLVEBTC_TPCH(i, idx), idx 747 drivers/net/ethernet/intel/i40e/i40e_main.c I40E_GLVEBTC_TPCL(i, idx), idx 751 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_stat_update48(hw, I40E_GLVEBTC_TBCH(i, idx), idx 752 drivers/net/ethernet/intel/i40e/i40e_main.c I40E_GLVEBTC_TBCL(i, idx), idx 9480 drivers/net/ethernet/intel/i40e/i40e_main.c pf->vsi[v]->veb_idx == veb->idx && idx 9488 drivers/net/ethernet/intel/i40e/i40e_main.c "missing owner VSI for veb_idx %d\n", veb->idx); idx 9498 drivers/net/ethernet/intel/i40e/i40e_main.c veb->idx, ret); idx 9519 drivers/net/ethernet/intel/i40e/i40e_main.c if (pf->vsi[v]->veb_idx == veb->idx) { idx 9536 drivers/net/ethernet/intel/i40e/i40e_main.c if (pf->veb[veb_idx] && pf->veb[veb_idx]->veb_idx == veb->idx) { idx 10678 drivers/net/ethernet/intel/i40e/i40e_main.c vsi->idx = vsi_idx; idx 10773 drivers/net/ethernet/intel/i40e/i40e_main.c if (!pf->vsi[vsi->idx]) { idx 10775 drivers/net/ethernet/intel/i40e/i40e_main.c vsi->idx, vsi->idx, vsi->type); idx 10779 drivers/net/ethernet/intel/i40e/i40e_main.c if (pf->vsi[vsi->idx] != vsi) { idx 10782 drivers/net/ethernet/intel/i40e/i40e_main.c pf->vsi[vsi->idx]->idx, idx 10783 drivers/net/ethernet/intel/i40e/i40e_main.c pf->vsi[vsi->idx]->type, idx 10784 drivers/net/ethernet/intel/i40e/i40e_main.c vsi->idx, vsi->type); idx 10789 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_put_lump(pf->qp_pile, vsi->base_queue, vsi->idx); idx 10790 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_put_lump(pf->irq_pile, vsi->base_vector, vsi->idx); idx 10796 drivers/net/ethernet/intel/i40e/i40e_main.c pf->vsi[vsi->idx] = NULL; idx 10797 drivers/net/ethernet/intel/i40e/i40e_main.c if (vsi->idx < pf->next_vsi) idx 10798 drivers/net/ethernet/intel/i40e/i40e_main.c pf->next_vsi = vsi->idx; idx 12169 drivers/net/ethernet/intel/i40e/i40e_main.c u8 idx; idx 12171 drivers/net/ethernet/intel/i40e/i40e_main.c idx = i40e_get_udp_port_idx(pf, port); idx 12174 drivers/net/ethernet/intel/i40e/i40e_main.c if (idx < I40E_MAX_PF_UDP_OFFLOAD_PORTS) { idx 12220 drivers/net/ethernet/intel/i40e/i40e_main.c u8 idx; idx 12222 drivers/net/ethernet/intel/i40e/i40e_main.c idx = i40e_get_udp_port_idx(pf, port); idx 12225 drivers/net/ethernet/intel/i40e/i40e_main.c if (idx >= I40E_MAX_PF_UDP_OFFLOAD_PORTS) idx 12230 drivers/net/ethernet/intel/i40e/i40e_main.c if (pf->udp_ports[idx].type != I40E_AQC_TUNNEL_TYPE_VXLAN) idx 12234 drivers/net/ethernet/intel/i40e/i40e_main.c if (pf->udp_ports[idx].type != I40E_AQC_TUNNEL_TYPE_NGE) idx 12244 drivers/net/ethernet/intel/i40e/i40e_main.c pf->udp_ports[idx].port = 0; idx 12250 drivers/net/ethernet/intel/i40e/i40e_main.c pf->pending_udp_bitmap ^= BIT_ULL(idx); idx 13458 drivers/net/ethernet/intel/i40e/i40e_main.c vsi->num_q_vectors, vsi->idx); idx 13493 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_put_lump(pf->qp_pile, vsi->base_queue, vsi->idx); idx 13505 drivers/net/ethernet/intel/i40e/i40e_main.c ret = i40e_get_lump(pf, pf->qp_pile, alloc_queue_pairs, vsi->idx); idx 13646 drivers/net/ethernet/intel/i40e/i40e_main.c vsi->veb_idx = (veb ? veb->idx : I40E_NO_VEB); idx 13656 drivers/net/ethernet/intel/i40e/i40e_main.c ret = i40e_get_lump(pf, pf->qp_pile, alloc_queue_pairs, vsi->idx); idx 13822 drivers/net/ethernet/intel/i40e/i40e_main.c veb->idx = i; idx 13843 drivers/net/ethernet/intel/i40e/i40e_main.c u16 veb_idx = branch->idx; idx 13890 drivers/net/ethernet/intel/i40e/i40e_main.c if (pf->veb[veb->idx] == veb) idx 13891 drivers/net/ethernet/intel/i40e/i40e_main.c pf->veb[veb->idx] = NULL; idx 13987 drivers/net/ethernet/intel/i40e/i40e_main.c vsi->veb_idx = veb->idx; idx 14058 drivers/net/ethernet/intel/i40e/i40e_main.c veb->veb_idx = (uplink_veb ? uplink_veb->idx : I40E_NO_VEB); idx 14066 drivers/net/ethernet/intel/i40e/i40e_main.c pf->lan_veb = veb->idx; idx 709 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c static int i40e_alloc_vsi_res(struct i40e_vf *vf, u8 idx) idx 728 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c if (!idx) { idx 732 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c vf->lan_vsi_idx = vsi->idx; idx 768 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c vf->ch[idx].vsi_idx = vsi->idx; idx 769 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c vf->ch[idx].vsi_id = vsi->id; idx 775 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c } else if (vf->ch[idx].max_tx_rate) { idx 776 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c max_tx_rate = vf->ch[idx].max_tx_rate; idx 1019 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c int ret, idx; idx 1038 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c for (idx = 1; idx < vf->num_tc; idx++) { idx 1039 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c ret = i40e_alloc_vsi_res(vf, idx); idx 2097 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c int i, j = 0, idx = 0; idx 2146 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c if (idx >= ARRAY_SIZE(vf->ch)) { idx 2150 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c vsi_id = vf->ch[idx].vsi_id; idx 2167 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c if (idx >= ARRAY_SIZE(vf->ch)) { idx 2171 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c if (j == (vf->ch[idx].num_qps - 1)) { idx 2172 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c idx++; idx 234 drivers/net/ethernet/intel/ice/ice.h u16 idx; /* software index in pf->vsi[] */ idx 10 drivers/net/ethernet/intel/ice/ice_common.c #define ICE_PROG_FLEX_ENTRY(hw, rxdid, mdid, idx) \ idx 11 drivers/net/ethernet/intel/ice/ice_common.c wr32((hw), GLFLXP_RXDID_FLX_WRD_##idx(rxdid), \ idx 13 drivers/net/ethernet/intel/ice/ice_common.c GLFLXP_RXDID_FLX_WRD_##idx##_RXDID_OPCODE_S) & \ idx 14 drivers/net/ethernet/intel/ice/ice_common.c GLFLXP_RXDID_FLX_WRD_##idx##_RXDID_OPCODE_M) | \ idx 15 drivers/net/ethernet/intel/ice/ice_common.c (((mdid) << GLFLXP_RXDID_FLX_WRD_##idx##_PROT_MDID_S) & \ idx 16 drivers/net/ethernet/intel/ice/ice_common.c GLFLXP_RXDID_FLX_WRD_##idx##_PROT_MDID_M)) idx 18 drivers/net/ethernet/intel/ice/ice_common.c #define ICE_PROG_FLG_ENTRY(hw, rxdid, flg_0, flg_1, flg_2, flg_3, idx) \ idx 19 drivers/net/ethernet/intel/ice/ice_common.c wr32((hw), GLFLXP_RXDID_FLAGS(rxdid, idx), \ idx 359 drivers/net/ethernet/intel/ice/ice_common.c u8 idx = 0; idx 376 drivers/net/ethernet/intel/ice/ice_common.c ICE_FLG_FIN, idx++); idx 381 drivers/net/ethernet/intel/ice/ice_common.c ICE_FLG_PKT_DSI, ICE_FLG_PKT_DSI, idx++); idx 384 drivers/net/ethernet/intel/ice/ice_common.c ICE_FLG_EVLAN_x9100, idx++); idx 387 drivers/net/ethernet/intel/ice/ice_common.c ICE_FLG_TNL0, idx++); idx 389 drivers/net/ethernet/intel/ice/ice_common.c ICE_FLG_PKT_DSI, ICE_FLG_PKT_DSI, idx); idx 168 drivers/net/ethernet/intel/ice/ice_dcb_lib.c pf->vsi[v]->idx); idx 2355 drivers/net/ethernet/intel/ice/ice_ethtool.c u16 adv_link_speed, curr_link_speed, idx; idx 2372 drivers/net/ethernet/intel/ice/ice_ethtool.c ice_for_each_vsi(pf, idx) idx 2373 drivers/net/ethernet/intel/ice/ice_ethtool.c if (pf->vsi[idx]->type == ICE_VSI_PF) { idx 2374 drivers/net/ethernet/intel/ice/ice_ethtool.c if (np->vsi != pf->vsi[idx]) idx 991 drivers/net/ethernet/intel/ice/ice_flex_pipe.c u16 idx = vsig & ICE_VSIG_IDX_M; idx 993 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt2.vsig_tbl[idx].in_use) { idx 994 drivers/net/ethernet/intel/ice/ice_flex_pipe.c INIT_LIST_HEAD(&hw->blk[blk].xlt2.vsig_tbl[idx].prop_lst); idx 995 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsig_tbl[idx].in_use = true; idx 998 drivers/net/ethernet/intel/ice/ice_flex_pipe.c return ICE_VSIG_VALUE(idx, hw->pf_id); idx 1015 drivers/net/ethernet/intel/ice/ice_flex_pipe.c u16 idx; idx 1017 drivers/net/ethernet/intel/ice/ice_flex_pipe.c idx = vsig & ICE_VSIG_IDX_M; idx 1019 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (vsi >= ICE_MAX_VSI || idx >= ICE_MAX_VSIGS) idx 1022 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt2.vsig_tbl[idx].in_use) idx 1026 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (idx == ICE_DEFAULT_VSIG) idx 1029 drivers/net/ethernet/intel/ice/ice_flex_pipe.c vsi_head = &hw->blk[blk].xlt2.vsig_tbl[idx].first_vsi; idx 1074 drivers/net/ethernet/intel/ice/ice_flex_pipe.c u16 orig_vsig, idx; idx 1076 drivers/net/ethernet/intel/ice/ice_flex_pipe.c idx = vsig & ICE_VSIG_IDX_M; idx 1078 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (vsi >= ICE_MAX_VSI || idx >= ICE_MAX_VSIGS) idx 1084 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt2.vsig_tbl[idx].in_use && idx 1103 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (idx == ICE_DEFAULT_VSIG) idx 1111 drivers/net/ethernet/intel/ice/ice_flex_pipe.c tmp = hw->blk[blk].xlt2.vsig_tbl[idx].first_vsi; idx 1112 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsig_tbl[idx].first_vsi = idx 155 drivers/net/ethernet/intel/ice/ice_lib.c tlan_ctx->src_vsi = ice_get_hw_vsi_num(hw, vsi->idx); idx 228 drivers/net/ethernet/intel/ice/ice_lib.c vsi->idx, pf_q, (ena ? "en" : "dis")); idx 434 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_free_vsi(&pf->hw, vsi->idx, ctxt, false, NULL); idx 494 drivers/net/ethernet/intel/ice/ice_lib.c if (!pf->vsi[vsi->idx] || pf->vsi[vsi->idx] != vsi) { idx 496 drivers/net/ethernet/intel/ice/ice_lib.c vsi->idx); idx 503 drivers/net/ethernet/intel/ice/ice_lib.c pf->vsi[vsi->idx] = NULL; idx 504 drivers/net/ethernet/intel/ice/ice_lib.c if (vsi->idx < pf->next_vsi) idx 505 drivers/net/ethernet/intel/ice/ice_lib.c pf->next_vsi = vsi->idx; idx 564 drivers/net/ethernet/intel/ice/ice_lib.c vsi->idx = pf->next_vsi; idx 1083 drivers/net/ethernet/intel/ice/ice_lib.c ret = ice_add_vsi(hw, vsi->idx, ctxt, NULL); idx 1251 drivers/net/ethernet/intel/ice/ice_lib.c vsi->idx); idx 1458 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_aq_set_rss_lut(&pf->hw, vsi->idx, vsi->rss_lut_type, lut, idx 1482 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_aq_set_rss_key(&pf->hw, vsi->idx, key); idx 1520 drivers/net/ethernet/intel/ice/ice_lib.c tmp->fltr_info.vsi_handle = vsi->idx; idx 1614 drivers/net/ethernet/intel/ice/ice_lib.c tmp->fltr_info.vsi_handle = vsi->idx; idx 1651 drivers/net/ethernet/intel/ice/ice_lib.c list->fltr_info.vsi_handle = vsi->idx; idx 1750 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_ena_vsi_txq(vsi->port_info, vsi->idx, tc, ring->q_handle, idx 2053 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_update_vsi(hw, vsi->idx, ctxt, NULL); idx 2100 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_update_vsi(hw, vsi->idx, ctxt, NULL); idx 2235 drivers/net/ethernet/intel/ice/ice_lib.c txq_meta->vsi_idx = vsi->idx; idx 2339 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_update_vsi(&pf->hw, vsi->idx, ctxt, NULL); idx 2342 drivers/net/ethernet/intel/ice/ice_lib.c ena ? "En" : "Dis", vsi->idx, vsi->vsi_num, status, idx 2432 drivers/net/ethernet/intel/ice/ice_lib.c list->fltr_info.vsi_handle = vsi->idx; idx 2469 drivers/net/ethernet/intel/ice/ice_lib.c list->fltr_info.vsi_handle = vsi->idx; idx 2543 drivers/net/ethernet/intel/ice/ice_lib.c vsi->idx); idx 2624 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_cfg_vsi_lan(vsi->port_info, vsi->idx, vsi->tc_cfg.ena_tc, idx 2660 drivers/net/ethernet/intel/ice/ice_lib.c ice_free_res(pf->irq_tracker, vsi->base_vector, vsi->idx); idx 2990 drivers/net/ethernet/intel/ice/ice_lib.c ice_free_res(pf->irq_tracker, vsi->base_vector, vsi->idx); idx 3006 drivers/net/ethernet/intel/ice/ice_lib.c ice_remove_vsi_fltr(&pf->hw, vsi->idx); idx 3007 drivers/net/ethernet/intel/ice/ice_lib.c ice_rm_vsi_lan_cfg(vsi->port_info, vsi->idx); idx 3052 drivers/net/ethernet/intel/ice/ice_lib.c ice_rm_vsi_lan_cfg(vsi->port_info, vsi->idx); idx 3062 drivers/net/ethernet/intel/ice/ice_lib.c ice_free_res(pf->irq_tracker, vsi->base_vector, vsi->idx); idx 3137 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_cfg_vsi_lan(vsi->port_info, vsi->idx, vsi->tc_cfg.ena_tc, idx 3227 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_update_vsi(&pf->hw, vsi->idx, ctx, NULL); idx 3234 drivers/net/ethernet/intel/ice/ice_lib.c status = ice_cfg_vsi_lan(vsi->port_info, vsi->idx, vsi->tc_cfg.ena_tc, idx 244 drivers/net/ethernet/intel/ice/ice_main.c status = ice_set_vlan_vsi_promisc(hw, vsi->idx, promisc_m, idx 248 drivers/net/ethernet/intel/ice/ice_main.c status = ice_set_vsi_promisc(hw, vsi->idx, promisc_m, idx 251 drivers/net/ethernet/intel/ice/ice_main.c status = ice_clear_vsi_promisc(hw, vsi->idx, promisc_m, idx 379 drivers/net/ethernet/intel/ice/ice_main.c status = ice_cfg_dflt_vsi(hw, vsi->idx, true, idx 390 drivers/net/ethernet/intel/ice/ice_main.c status = ice_cfg_dflt_vsi(hw, vsi->idx, false, idx 4172 drivers/net/ethernet/intel/ice/ice_main.c err, vsi->idx, type); idx 4177 drivers/net/ethernet/intel/ice/ice_main.c status = ice_replay_vsi(&pf->hw, vsi->idx); idx 4181 drivers/net/ethernet/intel/ice/ice_main.c status, vsi->idx, type); idx 4188 drivers/net/ethernet/intel/ice/ice_main.c vsi->vsi_num = ice_get_hw_vsi_num(&pf->hw, vsi->idx); idx 4195 drivers/net/ethernet/intel/ice/ice_main.c err, vsi->idx, type); idx 4200 drivers/net/ethernet/intel/ice/ice_main.c vsi->idx, type); idx 4427 drivers/net/ethernet/intel/ice/ice_main.c status = ice_aq_set_rss_key(hw, vsi->idx, buf); idx 4438 drivers/net/ethernet/intel/ice/ice_main.c status = ice_aq_set_rss_lut(hw, vsi->idx, vsi->rss_lut_type, idx 4470 drivers/net/ethernet/intel/ice/ice_main.c status = ice_aq_get_rss_key(hw, vsi->idx, buf); idx 4480 drivers/net/ethernet/intel/ice/ice_main.c status = ice_aq_get_rss_lut(hw, vsi->idx, vsi->rss_lut_type, idx 4551 drivers/net/ethernet/intel/ice/ice_main.c status = ice_update_vsi(hw, vsi->idx, ctxt, NULL); idx 477 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_update_vsi(hw, vsi->idx, ctxt, NULL); idx 546 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c vf->lan_vsi_idx = vsi->idx; idx 994 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_set_vlan_vsi_promisc(hw, vsi->idx, promisc_m, idx 998 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_clear_vsi_promisc(hw, vsi->idx, promisc_m, idx 1001 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_set_vsi_promisc(hw, vsi->idx, promisc_m, idx 1005 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_clear_vsi_promisc(hw, vsi->idx, promisc_m, idx 1008 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_set_vsi_promisc(hw, vsi->idx, promisc_m, idx 1093 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c ice_dis_vsi_txq(vsi->port_info, vsi->idx, 0, 0, NULL, NULL, idx 1194 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c ice_dis_vsi_txq(vsi->port_info, vsi->idx, 0, 0, NULL, NULL, idx 2755 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_set_vsi_promisc(hw, vsi->idx, idx 2797 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c ice_clear_vsi_promisc(hw, vsi->idx, idx 3113 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c status = ice_update_vsi(&pf->hw, vsi->idx, ctx, NULL); idx 6897 drivers/net/ethernet/intel/igb/igb_main.c int idx; idx 6904 drivers/net/ethernet/intel/igb/igb_main.c for (idx = E1000_VLVF_ARRAY_SIZE; --idx;) { idx 6905 drivers/net/ethernet/intel/igb/igb_main.c vlvf = rd32(E1000_VLVF(idx)); idx 6910 drivers/net/ethernet/intel/igb/igb_main.c return idx; idx 6917 drivers/net/ethernet/intel/igb/igb_main.c int idx; idx 6919 drivers/net/ethernet/intel/igb/igb_main.c idx = igb_find_vlvf_entry(hw, vid); idx 6920 drivers/net/ethernet/intel/igb/igb_main.c if (!idx) idx 6928 drivers/net/ethernet/intel/igb/igb_main.c bits &= rd32(E1000_VLVF(idx)); idx 6933 drivers/net/ethernet/intel/igb/igb_main.c wr32(E1000_VLVF(idx), BIT(pf_id)); idx 6935 drivers/net/ethernet/intel/igb/igb_main.c wr32(E1000_VLVF(idx), 0); idx 21 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c static void ixgbe_ipsec_set_tx_sa(struct ixgbe_hw *hw, u16 idx, idx 35 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c reg |= idx << IXGBE_RXTXIDX_IDX_SHIFT | IXGBE_RXTXIDX_WRITE; idx 49 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c static void ixgbe_ipsec_set_rx_item(struct ixgbe_hw *hw, u16 idx, idx 57 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c idx << IXGBE_RXTXIDX_IDX_SHIFT | idx 73 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c static void ixgbe_ipsec_set_rx_sa(struct ixgbe_hw *hw, u16 idx, __be32 spi, idx 84 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_rx_item(hw, idx, ips_rx_spi_tbl); idx 94 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_rx_item(hw, idx, ips_rx_key_tbl); idx 103 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c static void ixgbe_ipsec_set_rx_ip(struct ixgbe_hw *hw, u16 idx, __be32 addr[]) idx 113 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_rx_item(hw, idx, ips_rx_ip_tbl); idx 124 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c u16 idx; idx 131 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c for (idx = 0; idx < IXGBE_IPSEC_MAX_RX_IP_COUNT; idx++) { idx 132 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_tx_sa(hw, idx, buf, 0); idx 133 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_rx_sa(hw, idx, 0, buf, 0, 0, 0); idx 134 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_rx_ip(hw, idx, (__be32 *)buf); idx 136 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c for (; idx < IXGBE_IPSEC_MAX_SA_COUNT; idx++) { idx 137 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_tx_sa(hw, idx, buf, 0); idx 138 drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c ixgbe_ipsec_set_rx_sa(hw, idx, 0, buf, 0, 0, 0); idx 4441 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c int idx; idx 4448 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c for (idx = IXGBE_VLVF_ENTRIES; --idx;) { idx 4449 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c vlvf = IXGBE_READ_REG(hw, IXGBE_VLVF(idx)); idx 4454 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c return idx; idx 4461 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c int idx; idx 4463 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c idx = ixgbe_find_vlvf_entry(hw, vid); idx 4464 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!idx) idx 4470 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c word = idx * 2 + (VMDQ_P(0) / 32); idx 4478 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c IXGBE_WRITE_REG(hw, IXGBE_VLVF(idx), 0); idx 2268 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c int err, idx = 0; idx 2272 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c while (api[idx] != ixgbe_mbox_api_unknown) { idx 2273 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c err = hw->mac.ops.negotiate_api_version(hw, api[idx]); idx 2276 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c idx++; idx 996 drivers/net/ethernet/jme.c jme_alloc_and_feed_skb(struct jme_adapter *jme, int idx) idx 1004 drivers/net/ethernet/jme.c rxdesc += idx; idx 1005 drivers/net/ethernet/jme.c rxbi += idx; idx 1013 drivers/net/ethernet/jme.c if (unlikely(jme_make_new_rx_buf(jme, idx))) { idx 1049 drivers/net/ethernet/jme.c jme_set_clean_rxdesc(jme, idx); idx 1950 drivers/net/ethernet/jme.c int idx, nr_alloc, mask = jme->tx_ring_mask; idx 1952 drivers/net/ethernet/jme.c idx = txring->next_to_use; idx 1962 drivers/net/ethernet/jme.c return idx; idx 2024 drivers/net/ethernet/jme.c jme_map_tx_skb(struct jme_adapter *jme, struct sk_buff *skb, int idx) idx 2038 drivers/net/ethernet/jme.c ctxdesc = txdesc + ((idx + i + 2) & (mask)); idx 2039 drivers/net/ethernet/jme.c ctxbi = txbi + ((idx + i + 2) & (mask)); idx 2045 drivers/net/ethernet/jme.c jme_drop_tx_map(jme, idx, i); idx 2051 drivers/net/ethernet/jme.c ctxdesc = txdesc + ((idx + 1) & (mask)); idx 2052 drivers/net/ethernet/jme.c ctxbi = txbi + ((idx + 1) & (mask)); idx 2056 drivers/net/ethernet/jme.c jme_drop_tx_map(jme, idx, i); idx 2136 drivers/net/ethernet/jme.c jme_fill_tx_desc(struct jme_adapter *jme, struct sk_buff *skb, int idx) idx 2144 drivers/net/ethernet/jme.c txdesc = (struct txdesc *)txring->desc + idx; idx 2145 drivers/net/ethernet/jme.c txbi = txring->bufinf + idx; idx 2167 drivers/net/ethernet/jme.c ret = jme_map_tx_skb(jme, skb, idx); idx 2192 drivers/net/ethernet/jme.c int idx = atomic_read(&txring->next_to_clean); idx 2194 drivers/net/ethernet/jme.c txbi += idx; idx 2213 drivers/net/ethernet/jme.c "TX Queue Stopped %d@%lu\n", idx, jiffies); idx 2225 drivers/net/ethernet/jme.c int idx; idx 2233 drivers/net/ethernet/jme.c idx = jme_alloc_txdesc(jme, skb); idx 2235 drivers/net/ethernet/jme.c if (unlikely(idx < 0)) { idx 2243 drivers/net/ethernet/jme.c if (jme_fill_tx_desc(jme, skb, idx)) idx 2252 drivers/net/ethernet/jme.c idx, skb_shinfo(skb)->nr_frags + 2, jiffies); idx 80 drivers/net/ethernet/lantiq_etop.c int idx; idx 175 drivers/net/ethernet/lantiq_etop.c netdev_get_tx_queue(ch->netdev, ch->idx >> 1); idx 215 drivers/net/ethernet/lantiq_etop.c if (IS_RX(ch->idx)) { idx 268 drivers/net/ethernet/lantiq_etop.c ch->idx = ch->dma.nr = i; idx 62 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP2_PRS_TCAM_DATA_REG(idx) (0x1104 + (idx) * 4) idx 65 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP2_PRS_SRAM_DATA_REG(idx) (0x1204 + (idx) * 4) idx 74 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP22_RSS_INDEX_TABLE_ENTRY(idx) (idx) idx 75 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP22_RSS_INDEX_TABLE(idx) ((idx) << 8) idx 76 drivers/net/ethernet/marvell/mvpp2/mvpp2.h #define MVPP22_RSS_INDEX_QUEUE(idx) ((idx) << 16) idx 284 drivers/net/ethernet/marvell/octeontx2/af/cgx.c int cgx_get_rx_stats(void *cgxd, int lmac_id, int idx, u64 *rx_stat) idx 290 drivers/net/ethernet/marvell/octeontx2/af/cgx.c *rx_stat = cgx_read(cgx, lmac_id, CGXX_CMRX_RX_STAT0 + (idx * 8)); idx 295 drivers/net/ethernet/marvell/octeontx2/af/cgx.c int cgx_get_tx_stats(void *cgxd, int lmac_id, int idx, u64 *tx_stat) idx 301 drivers/net/ethernet/marvell/octeontx2/af/cgx.c *tx_stat = cgx_read(cgx, lmac_id, CGXX_CMRX_TX_STAT0 + (idx * 8)); idx 104 drivers/net/ethernet/marvell/octeontx2/af/cgx.h int cgx_get_tx_stats(void *cgxd, int lmac_id, int idx, u64 *tx_stat); idx 105 drivers/net/ethernet/marvell/octeontx2/af/cgx.h int cgx_get_rx_stats(void *cgxd, int lmac_id, int idx, u64 *rx_stat); idx 1905 drivers/net/ethernet/marvell/octeontx2/af/rvu.c static void rvu_me_handle_vfset(struct rvu *rvu, int idx, u64 intr) idx 1915 drivers/net/ethernet/marvell/octeontx2/af/rvu.c rvupf_write64(rvu, RVU_PF_VFTRPENDX(idx), BIT_ULL(vf)); idx 1917 drivers/net/ethernet/marvell/octeontx2/af/rvu.c rvupf_write64(rvu, RVU_PF_VFME_INTX(idx), BIT_ULL(vf)); idx 67 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c u16 idx; idx 103 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int idx; idx 108 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c idx = mcast->next_free_mce; idx 110 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c return idx; idx 310 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c u64 cfg, idx, fidx = 0; idx 329 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c idx = NIX_LSO_FORMAT_IDX_TSOV4; idx 330 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c nix_setup_lso_tso_l3(rvu, blkaddr, idx, true, &fidx); idx 331 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c nix_setup_lso_tso_l4(rvu, blkaddr, idx, &fidx); idx 336 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c NIX_AF_LSO_FORMATX_FIELDX(idx, fidx), 0x0ULL); idx 341 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c idx = NIX_LSO_FORMAT_IDX_TSOV6; idx 343 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c nix_setup_lso_tso_l3(rvu, blkaddr, idx, false, &fidx); idx 344 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c nix_setup_lso_tso_l4(rvu, blkaddr, idx, &fidx); idx 349 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c NIX_AF_LSO_FORMATX_FIELDX(idx, fidx), 0x0ULL); idx 1137 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int lvl, idx, req_schq; idx 1213 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx < req->schq_contig[lvl]; idx++) { idx 1217 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c rsp->schq_contig_list[lvl][idx] = schq; idx 1223 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx < req->schq[lvl]; idx++) { idx 1228 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c rsp->schq_list[lvl][idx] = schq; idx 1438 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int blkaddr, idx, err = 0; idx 1461 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx < schq_cnt; idx++) { idx 1462 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c schq = schq_list[idx]; idx 1499 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int blkaddr, idx, err; idx 1530 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx < req->num_regs; idx++) { idx 1531 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c reg = req->reg[idx]; idx 1532 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c regval = req->regval[idx]; idx 1653 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c u16 pcifunc, int idx, bool add) idx 1682 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c mce->idx = idx; idx 1694 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int err = 0, idx, next_idx, count; idx 1718 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c idx = pfvf->bcast_mce_idx + (pcifunc & RVU_PFVF_FUNC_MASK); idx 1721 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c if (idx > (pfvf->bcast_mce_idx + mce_list->max)) { idx 1724 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c __func__, idx, mce_list->max, idx 1731 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c err = nix_update_mce_list(mce_list, pcifunc, idx, add); idx 1748 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c next_idx = next_mce->idx; idx 1751 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c err = nix_setup_mce(rvu, mce->idx, idx 1766 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int err, pf, numvfs, idx; idx 1786 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx < (numvfs + 1); idx++) { idx 1789 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c pcifunc |= idx; idx 1795 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c err = nix_setup_mce(rvu, pfvf->bcast_mce_idx + idx, idx 1992 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int idx, nr_field, key_off, field_marker, keyoff_marker; idx 2026 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx 2027 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c idx < max_bit_pos && nr_field < FIELDS_PER_ALG && idx 2028 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c key_off < MAX_KEY_OFF; idx++) { idx 2029 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c key_type = BIT(idx); idx 2115 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c if (idx == max_bit_pos && key_off <= MAX_KEY_OFF) idx 2615 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int idx, err; idx 2631 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx < rvu->cgx_cnt_max; idx++) { idx 2633 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c if (!rvu_cgx_pdata(idx, rvu) || idx 2634 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c (status & (BIT_ULL(16 + idx)))) idx 2637 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c "CGX%d didn't respond to NIX X2P calibration\n", idx); idx 2917 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c int blkaddr, idx, f; idx 2930 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c for (idx = 0; idx < nix_hw->lso.in_use; idx++) { idx 2933 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c NIX_AF_LSO_FORMATX_FIELDX(idx, f)); idx 2942 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c if (idx < nix_hw->lso.in_use) { idx 2944 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c rsp->lso_format_idx = idx; idx 123 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c static void npc_get_keyword(struct mcam_entry *entry, int idx, idx 135 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c switch (idx) { idx 952 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c int num_pkinds, num_kpus, idx; idx 959 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (idx = 0; idx < hw->npc_kpus; idx++) { idx 961 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NPC_AF_KPUX_ENTRY_DISX(idx, 0), ~0ULL); idx 963 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c NPC_AF_KPUX_ENTRY_DISX(idx, 1), ~0ULL); idx 964 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c rvu_write64(rvu, blkaddr, NPC_AF_KPUX_CFG(idx), 0x00); idx 975 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (idx = 0; idx < num_pkinds; idx++) idx 977 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c &ikpu_action_entries[idx], 0, idx, true); idx 983 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c for (idx = 0; idx < num_kpus; idx++) idx 985 drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c idx, &npc_kpu_profiles[idx]); idx 44 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.c int idx; idx 65 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.c for (idx = 0; idx < map->num_ranges; idx++) { idx 66 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.c if (reg >= map->range[idx].start && idx 67 drivers/net/ethernet/marvell/octeontx2/af/rvu_reg.c reg < map->range[idx].end) idx 1125 drivers/net/ethernet/marvell/sky2.c static inline void sky2_put_idx(struct sky2_hw *hw, unsigned q, u16 idx) idx 1129 drivers/net/ethernet/marvell/sky2.c sky2_write16(hw, Y2_QADDR(q, PREF_UNIT_PUT_IDX), idx); idx 2003 drivers/net/ethernet/marvell/sky2.c u16 idx; idx 2008 drivers/net/ethernet/marvell/sky2.c for (idx = sky2->tx_cons; idx != done; idx 2009 drivers/net/ethernet/marvell/sky2.c idx = RING_NEXT(idx, sky2->tx_ring_size)) { idx 2010 drivers/net/ethernet/marvell/sky2.c struct tx_ring_info *re = sky2->tx_ring + idx; idx 2017 drivers/net/ethernet/marvell/sky2.c "tx done %u\n", idx); idx 2025 drivers/net/ethernet/marvell/sky2.c sky2->tx_next = RING_NEXT(idx, sky2->tx_ring_size); idx 2029 drivers/net/ethernet/marvell/sky2.c sky2->tx_cons = idx; idx 2702 drivers/net/ethernet/marvell/sky2.c static int sky2_status_intr(struct sky2_hw *hw, int to_do, u16 idx) idx 2790 drivers/net/ethernet/marvell/sky2.c } while (hw->st_idx != idx); idx 2918 drivers/net/ethernet/marvell/sky2.c u16 idx = sky2_read16(hw, Y2_QADDR(q, PREF_UNIT_GET_IDX)); idx 2921 drivers/net/ethernet/marvell/sky2.c dev->name, (unsigned) q, (unsigned) idx, idx 3025 drivers/net/ethernet/marvell/sky2.c u16 idx; idx 3039 drivers/net/ethernet/marvell/sky2.c while ((idx = sky2_read16(hw, STAT_PUT_IDX)) != hw->st_idx) { idx 3040 drivers/net/ethernet/marvell/sky2.c work_done += sky2_status_intr(hw, work_limit - work_done, idx); idx 4519 drivers/net/ethernet/marvell/sky2.c unsigned idx, last; idx 4542 drivers/net/ethernet/marvell/sky2.c for (idx = hw->st_idx; idx != last && idx < hw->st_size; idx 4543 drivers/net/ethernet/marvell/sky2.c idx = RING_NEXT(idx, hw->st_size)) { idx 4544 drivers/net/ethernet/marvell/sky2.c const struct sky2_status_le *le = hw->st_le + idx; idx 4546 drivers/net/ethernet/marvell/sky2.c idx, le->opcode, le->length, le->status); idx 4558 drivers/net/ethernet/marvell/sky2.c for (idx = sky2->tx_next; idx != sky2->tx_prod && idx < sky2->tx_ring_size; idx 4559 drivers/net/ethernet/marvell/sky2.c idx = RING_NEXT(idx, sky2->tx_ring_size)) { idx 4560 drivers/net/ethernet/marvell/sky2.c const struct sky2_tx_le *le = sky2->tx_le + idx; idx 4564 drivers/net/ethernet/marvell/sky2.c seq_printf(seq, "%u:", idx); idx 831 drivers/net/ethernet/mediatek/mtk_eth_soc.c int idx = txd - ring->dma; idx 833 drivers/net/ethernet/mediatek/mtk_eth_soc.c return &ring->buf[idx]; idx 886 drivers/net/ethernet/mediatek/mtk_eth_soc.c size_t size, int idx) idx 892 drivers/net/ethernet/mediatek/mtk_eth_soc.c if (idx & 1) { idx 1189 drivers/net/ethernet/mediatek/mtk_eth_soc.c int idx; idx 1196 drivers/net/ethernet/mediatek/mtk_eth_soc.c idx = NEXT_DESP_IDX(ring->calc_idx, ring->dma_size); idx 1197 drivers/net/ethernet/mediatek/mtk_eth_soc.c if (ring->dma[idx].rxd2 & RX_DMA_DONE) { idx 1229 drivers/net/ethernet/mediatek/mtk_eth_soc.c int idx; idx 1245 drivers/net/ethernet/mediatek/mtk_eth_soc.c idx = NEXT_DESP_IDX(ring->calc_idx, ring->dma_size); idx 1246 drivers/net/ethernet/mediatek/mtk_eth_soc.c rxd = &ring->dma[idx]; idx 1247 drivers/net/ethernet/mediatek/mtk_eth_soc.c data = ring->data[idx]; idx 1315 drivers/net/ethernet/mediatek/mtk_eth_soc.c ring->data[idx] = new_data; idx 1324 drivers/net/ethernet/mediatek/mtk_eth_soc.c ring->calc_idx = idx; idx 1826 drivers/net/ethernet/mediatek/mtk_eth_soc.c static void mtk_hwlro_val_ipaddr(struct mtk_eth *eth, int idx, __be32 ip) idx 1830 drivers/net/ethernet/mediatek/mtk_eth_soc.c reg_val = mtk_r32(eth, MTK_LRO_CTRL_DW2_CFG(idx)); idx 1833 drivers/net/ethernet/mediatek/mtk_eth_soc.c mtk_w32(eth, (reg_val & ~MTK_RING_MYIP_VLD), MTK_LRO_CTRL_DW2_CFG(idx)); idx 1835 drivers/net/ethernet/mediatek/mtk_eth_soc.c mtk_w32(eth, ip, MTK_LRO_DIP_DW0_CFG(idx)); idx 1838 drivers/net/ethernet/mediatek/mtk_eth_soc.c mtk_w32(eth, (reg_val | MTK_RING_MYIP_VLD), MTK_LRO_CTRL_DW2_CFG(idx)); idx 1841 drivers/net/ethernet/mediatek/mtk_eth_soc.c static void mtk_hwlro_inval_ipaddr(struct mtk_eth *eth, int idx) idx 1845 drivers/net/ethernet/mediatek/mtk_eth_soc.c reg_val = mtk_r32(eth, MTK_LRO_CTRL_DW2_CFG(idx)); idx 1848 drivers/net/ethernet/mediatek/mtk_eth_soc.c mtk_w32(eth, (reg_val & ~MTK_RING_MYIP_VLD), MTK_LRO_CTRL_DW2_CFG(idx)); idx 1850 drivers/net/ethernet/mediatek/mtk_eth_soc.c mtk_w32(eth, 0, MTK_LRO_DIP_DW0_CFG(idx)); idx 477 drivers/net/ethernet/mellanox/mlx4/en_netdev.c int idx; idx 492 drivers/net/ethernet/mellanox/mlx4/en_netdev.c err = mlx4_register_vlan(mdev->dev, priv->port, vid, &idx); idx 318 drivers/net/ethernet/mellanox/mlx4/icm.c u64 idx; idx 328 drivers/net/ethernet/mellanox/mlx4/icm.c idx = (u64) (obj & (table->num_obj - 1)) * table->obj_size; idx 329 drivers/net/ethernet/mellanox/mlx4/icm.c icm = table->icm[idx / MLX4_TABLE_CHUNK_SIZE]; idx 330 drivers/net/ethernet/mellanox/mlx4/icm.c dma_offset = offset = idx % MLX4_TABLE_CHUNK_SIZE; idx 2543 drivers/net/ethernet/mellanox/mlx4/main.c u32 idx; idx 2549 drivers/net/ethernet/mellanox/mlx4/main.c err = mlx4_counter_alloc(dev, &idx, MLX4_RES_USAGE_DRIVER); idx 2552 drivers/net/ethernet/mellanox/mlx4/main.c priv->def_counter[port] = idx; idx 2576 drivers/net/ethernet/mellanox/mlx4/main.c int __mlx4_counter_alloc(struct mlx4_dev *dev, u32 *idx) idx 2583 drivers/net/ethernet/mellanox/mlx4/main.c *idx = mlx4_bitmap_alloc(&priv->counters_bitmap); idx 2584 drivers/net/ethernet/mellanox/mlx4/main.c if (*idx == -1) { idx 2585 drivers/net/ethernet/mellanox/mlx4/main.c *idx = MLX4_SINK_COUNTER_INDEX(dev); idx 2592 drivers/net/ethernet/mellanox/mlx4/main.c int mlx4_counter_alloc(struct mlx4_dev *dev, u32 *idx, u8 usage) idx 2603 drivers/net/ethernet/mellanox/mlx4/main.c *idx = get_param_l(&out_param); idx 2608 drivers/net/ethernet/mellanox/mlx4/main.c return __mlx4_counter_alloc(dev, idx); idx 2631 drivers/net/ethernet/mellanox/mlx4/main.c void __mlx4_counter_free(struct mlx4_dev *dev, u32 idx) idx 2636 drivers/net/ethernet/mellanox/mlx4/main.c if (idx == MLX4_SINK_COUNTER_INDEX(dev)) idx 2639 drivers/net/ethernet/mellanox/mlx4/main.c __mlx4_clear_if_stat(dev, idx); idx 2641 drivers/net/ethernet/mellanox/mlx4/main.c mlx4_bitmap_free(&mlx4_priv(dev)->counters_bitmap, idx, MLX4_USE_RR); idx 2645 drivers/net/ethernet/mellanox/mlx4/main.c void mlx4_counter_free(struct mlx4_dev *dev, u32 idx) idx 2650 drivers/net/ethernet/mellanox/mlx4/main.c set_param_l(&in_param, idx); idx 2656 drivers/net/ethernet/mellanox/mlx4/main.c __mlx4_counter_free(dev, idx); idx 1033 drivers/net/ethernet/mellanox/mlx4/mlx4.h int __mlx4_counter_alloc(struct mlx4_dev *dev, u32 *idx); idx 1034 drivers/net/ethernet/mellanox/mlx4/mlx4.h void __mlx4_counter_free(struct mlx4_dev *dev, u32 idx); idx 660 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h static inline struct mlx4_cqe *mlx4_en_get_cqe(void *buf, int idx, int cqe_sz) idx 662 drivers/net/ethernet/mellanox/mlx4/mlx4_en.h return buf + idx * cqe_sz; idx 177 drivers/net/ethernet/mellanox/mlx4/pd.c int idx; idx 219 drivers/net/ethernet/mellanox/mlx4/pd.c idx = ffz(uar->free_bf_bmap); idx 220 drivers/net/ethernet/mellanox/mlx4/pd.c uar->free_bf_bmap |= 1 << idx; idx 224 drivers/net/ethernet/mellanox/mlx4/pd.c bf->reg = uar->bf_map + idx * dev->caps.bf_reg_size; idx 249 drivers/net/ethernet/mellanox/mlx4/pd.c int idx; idx 255 drivers/net/ethernet/mellanox/mlx4/pd.c idx = (bf->reg - bf->uar->bf_map) / dev->caps.bf_reg_size; idx 256 drivers/net/ethernet/mellanox/mlx4/pd.c bf->uar->free_bf_bmap &= ~(1 << idx); idx 150 drivers/net/ethernet/mellanox/mlx4/port.c int mlx4_find_cached_mac(struct mlx4_dev *dev, u8 port, u64 mac, int *idx) idx 161 drivers/net/ethernet/mellanox/mlx4/port.c *idx = i; idx 558 drivers/net/ethernet/mellanox/mlx4/port.c int mlx4_find_cached_vlan(struct mlx4_dev *dev, u8 port, u16 vid, int *idx) idx 568 drivers/net/ethernet/mellanox/mlx4/port.c *idx = i; idx 126 drivers/net/ethernet/mellanox/mlx5/core/cmd.c static void free_ent(struct mlx5_cmd *cmd, int idx) idx 131 drivers/net/ethernet/mellanox/mlx5/core/cmd.c set_bit(idx, &cmd->bitmask); idx 135 drivers/net/ethernet/mellanox/mlx5/core/cmd.c static struct mlx5_cmd_layout *get_inst(struct mlx5_cmd *cmd, int idx) idx 137 drivers/net/ethernet/mellanox/mlx5/core/cmd.c return cmd->cmd_buf + (idx << cmd->log_stride); idx 844 drivers/net/ethernet/mellanox/mlx5/core/cmd.c mlx5_cmd_comp_handler(dev, 1UL << ent->idx, true); idx 883 drivers/net/ethernet/mellanox/mlx5/core/cmd.c ent->idx = alloc_ret; idx 885 drivers/net/ethernet/mellanox/mlx5/core/cmd.c ent->idx = cmd->max_reg_cmds; idx 887 drivers/net/ethernet/mellanox/mlx5/core/cmd.c clear_bit(ent->idx, &cmd->bitmask); idx 891 drivers/net/ethernet/mellanox/mlx5/core/cmd.c cmd->ent_arr[ent->idx] = ent; idx 892 drivers/net/ethernet/mellanox/mlx5/core/cmd.c lay = get_inst(cmd, ent->idx); idx 925 drivers/net/ethernet/mellanox/mlx5/core/cmd.c mlx5_cmd_comp_handler(dev, 1UL << ent->idx, true); idx 927 drivers/net/ethernet/mellanox/mlx5/core/cmd.c free_ent(cmd, ent->idx); idx 934 drivers/net/ethernet/mellanox/mlx5/core/cmd.c mlx5_core_dbg(dev, "writing 0x%x to command doorbell\n", 1 << ent->idx); idx 936 drivers/net/ethernet/mellanox/mlx5/core/cmd.c iowrite32be(1 << ent->idx, &dev->iseg->cmd_dbell); idx 942 drivers/net/ethernet/mellanox/mlx5/core/cmd.c mlx5_cmd_comp_handler(dev, 1UL << ent->idx, (ent->ret == -ETIMEDOUT)); idx 991 drivers/net/ethernet/mellanox/mlx5/core/cmd.c mlx5_cmd_comp_handler(dev, 1UL << ent->idx, true); idx 1492 drivers/net/ethernet/mellanox/mlx5/core/cmd.c ent->idx); idx 1493 drivers/net/ethernet/mellanox/mlx5/core/cmd.c free_ent(cmd, ent->idx); idx 1524 drivers/net/ethernet/mellanox/mlx5/core/cmd.c free_ent(cmd, ent->idx); idx 88 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_stats.c unsigned int i, idx = 0; idx 94 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 98 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 119 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_stats.c int i, idx = 0; idx 125 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_stats.c data[idx++] = MLX5E_READ_CTR64_CPU(&priv->ipsec->stats, idx 129 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_stats.c data[idx++] = MLX5E_READ_CTR_ATOMIC64(&priv->ipsec->sw_stats, idx 64 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_stats.c unsigned int i, idx = 0; idx 70 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 78 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_stats.c int i, idx = 0; idx 84 drivers/net/ethernet/mellanox/mlx5/core/en_accel/tls_stats.c data[idx++] = idx 76 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c unsigned int i, bit, idx; \ idx 84 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c idx = modes[i] / 64; \ idx 85 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c __set_bit(bit, &cfg->supported[idx]); \ idx 86 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c __set_bit(bit, &cfg->advertised[idx]); \ idx 247 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c int i, idx = 0; idx 250 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c idx = mlx5e_stats_grps[i].fill_strings(priv, data, idx); idx 286 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c int i, idx = 0; idx 293 drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c idx = mlx5e_stats_grps[i].fill_stats(priv, data, idx); idx 758 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c int idx = 0; idx 762 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c while ((!err || err == -ENOENT) && idx < info->rule_cnt) { idx 765 drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c rule_locs[idx++] = location; idx 154 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c static int mlx5e_grp_sw_fill_strings(struct mlx5e_priv *priv, u8 *data, int idx) idx 159 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, sw_stats_desc[i].format); idx 160 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 163 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c static int mlx5e_grp_sw_fill_stats(struct mlx5e_priv *priv, u64 *data, int idx) idx 168 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_CPU(&priv->stats.sw, sw_stats_desc, i); idx 169 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 331 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c static int mlx5e_grp_q_fill_strings(struct mlx5e_priv *priv, u8 *data, int idx) idx 336 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 340 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 343 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 346 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c static int mlx5e_grp_q_fill_stats(struct mlx5e_priv *priv, u64 *data, int idx) idx 351 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR32_CPU(&priv->stats.qcnt, idx 354 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR32_CPU(&priv->stats.qcnt, idx 356 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 401 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 406 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 410 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 412 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 416 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 421 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_BE(priv->stats.vnic.query_vnic_env_out, idx 425 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR32_BE(priv->stats.vnic.query_vnic_env_out, idx 427 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 499 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 504 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, vport_stats_desc[i].format); idx 505 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 509 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 514 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_BE(priv->stats.vport.query_vport_out, idx 516 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 564 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 569 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, pport_802_3_stats_desc[i].format); idx 570 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 574 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 579 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_BE(&priv->stats.pport.IEEE_802_3_counters, idx 581 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 621 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 626 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, pport_2863_stats_desc[i].format); idx 627 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 631 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 636 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_BE(&priv->stats.pport.RFC_2863_counters, idx 638 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 682 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 687 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, pport_2819_stats_desc[i].format); idx 688 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 692 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 697 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_BE(&priv->stats.pport.RFC_2819_counters, idx 699 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 758 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 763 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, "link_down_events_phy"); idx 766 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 769 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 774 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 777 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 780 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c static int mlx5e_grp_phy_fill_stats(struct mlx5e_priv *priv, u64 *data, int idx) idx 786 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5_GET(ppcnt_reg, priv->stats.pport.phy_counters, idx 790 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 793 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 799 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 803 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 845 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 851 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 853 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 857 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 863 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 866 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 927 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 933 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 938 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 943 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 945 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 949 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 955 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 961 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 967 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 970 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1022 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c u8 *data, int idx) idx 1028 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1032 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1035 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1039 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1043 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c u64 *data, int idx) idx 1050 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1054 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1058 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1063 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1149 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1155 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1159 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1164 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1170 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1175 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1237 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1248 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1255 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1261 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, idx 1264 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1269 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1277 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1285 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1292 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_BE(&priv->stats.pport.per_prio_counters[0], idx 1295 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1305 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1307 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c idx = mlx5e_grp_per_prio_traffic_fill_strings(priv, data, idx); idx 1308 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c idx = mlx5e_grp_per_prio_pfc_fill_strings(priv, data, idx); idx 1309 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1313 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1315 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c idx = mlx5e_grp_per_prio_traffic_fill_stats(priv, data, idx); idx 1316 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c idx = mlx5e_grp_per_prio_pfc_fill_stats(priv, data, idx); idx 1317 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1361 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1366 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, mlx5e_pme_status_desc[i].format); idx 1369 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c strcpy(data + (idx++) * ETH_GSTRING_LEN, mlx5e_pme_error_desc[i].format); idx 1371 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1375 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1383 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_CPU(pme_stats.status_counters, idx 1387 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = MLX5E_READ_CTR64_CPU(pme_stats.error_counters, idx 1390 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1399 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1401 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx + mlx5e_ipsec_get_strings(priv, idx 1402 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data + idx * ETH_GSTRING_LEN); idx 1406 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1408 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx + mlx5e_ipsec_get_stats(priv, data + idx); idx 1422 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1424 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx + mlx5e_tls_get_strings(priv, data + idx * ETH_GSTRING_LEN); idx 1427 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c static int mlx5e_grp_tls_fill_stats(struct mlx5e_priv *priv, u64 *data, int idx) idx 1429 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx + mlx5e_tls_get_stats(priv, data + idx); idx 1579 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1587 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1592 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1595 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1598 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1605 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1611 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1614 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c sprintf(data + (idx++) * ETH_GSTRING_LEN, idx 1618 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 1622 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c int idx) idx 1630 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1636 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1640 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1644 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1652 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1658 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1662 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c data[idx++] = idx 1667 drivers/net/ethernet/mellanox/mlx5/core/en_stats.c return idx; idx 333 drivers/net/ethernet/mellanox/mlx5/core/en_stats.h int (*fill_strings)(struct mlx5e_priv *priv, u8 *data, int idx); idx 334 drivers/net/ethernet/mellanox/mlx5/core/en_stats.h int (*fill_stats)(struct mlx5e_priv *priv, u64 *data, int idx); idx 67 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c u16 idx; idx 72 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c idx = mlx5_eswitch_vport_num_to_index(esw, vport_num); idx 74 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c if (idx > esw->total_vports - 1) { idx 76 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c vport_num, idx); idx 80 drivers/net/ethernet/mellanox/mlx5/core/eswitch.c return &esw->vports[idx]; idx 59 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c int idx = mlx5_eswitch_vport_num_to_index(esw, vport_num); idx 61 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c WARN_ON(idx > esw->total_vports - 1); idx 62 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c return &esw->offloads.vport_reps[idx]; idx 381 drivers/net/ethernet/mellanox/mlx5/core/lag.c int idx; idx 396 drivers/net/ethernet/mellanox/mlx5/core/lag.c idx = mlx5_lag_dev_get_netdev_idx(ldev, ndev_tmp); idx 397 drivers/net/ethernet/mellanox/mlx5/core/lag.c if (idx > -1) idx 398 drivers/net/ethernet/mellanox/mlx5/core/lag.c bond_status |= (1 << idx); idx 435 drivers/net/ethernet/mellanox/mlx5/core/lag.c int idx; idx 440 drivers/net/ethernet/mellanox/mlx5/core/lag.c idx = mlx5_lag_dev_get_netdev_idx(ldev, ndev); idx 441 drivers/net/ethernet/mellanox/mlx5/core/lag.c if (idx == -1) idx 451 drivers/net/ethernet/mellanox/mlx5/core/lag.c tracker->netdev_state[idx] = *lag_lower_info; idx 33 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c int idx; idx 53 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c u8 idx) idx 62 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c devcom->idx = idx; idx 73 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c int idx, i; idx 82 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c idx = -1; idx 87 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c idx = i; idx 90 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c if (idx == -1) idx 106 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c idx = 0; idx 110 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c priv->devs[idx] = dev; idx 111 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c devcom = mlx5_devcom_alloc(priv, idx); idx 133 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c priv->devs[devcom->idx] = NULL; idx 163 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c comp->device[devcom->idx].data = data; idx 177 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c comp->device[devcom->idx].data = NULL; idx 195 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c if (i != devcom->idx && comp->device[i].data) { idx 243 drivers/net/ethernet/mellanox/mlx5/core/lib/devcom.c if (i != devcom->idx) idx 179 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c int idx = 0; idx 221 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_general_purpose(&sb[idx++], &mask, inner, rx); idx 224 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_register_0(&sb[idx++], &mask, inner, rx); idx 227 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_register_1(&sb[idx++], &mask, inner, rx); idx 232 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = mlx5dr_ste_build_src_gvmi_qpn(&sb[idx++], &mask, idx 240 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = mlx5dr_ste_build_eth_l2_src_des(&sb[idx++], &mask, idx 247 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l2_src(&sb[idx++], &mask, inner, rx); idx 250 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l2_dst(&sb[idx++], &mask, inner, rx); idx 254 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv6_dst(&sb[idx++], &mask, idx 258 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv6_src(&sb[idx++], &mask, idx 262 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_ipv6_l3_l4(&sb[idx++], &mask, idx 266 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv4_5_tuple(&sb[idx++], &mask, idx 270 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv4_misc(&sb[idx++], &mask, idx 276 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_flex_parser_tnl(&sb[idx++], &mask, idx 280 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l4_misc(&sb[idx++], &mask, inner, rx); idx 283 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_mpls(&sb[idx++], &mask, inner, rx); idx 286 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_flex_parser_0(&sb[idx++], &mask, idx 294 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = mlx5dr_ste_build_flex_parser_1(&sb[idx++], idx 301 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_gre(&sb[idx++], &mask, inner, rx); idx 312 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l2_tnl(&sb[idx++], &mask, inner, rx); idx 316 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c ret = mlx5dr_ste_build_eth_l2_src_des(&sb[idx++], idx 323 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l2_src(&sb[idx++], &mask, inner, rx); idx 326 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l2_dst(&sb[idx++], &mask, inner, rx); idx 330 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv6_dst(&sb[idx++], &mask, idx 334 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv6_src(&sb[idx++], &mask, idx 338 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_ipv6_l3_l4(&sb[idx++], &mask, idx 342 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv4_5_tuple(&sb[idx++], &mask, idx 346 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l3_ipv4_misc(&sb[idx++], &mask, idx 351 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_eth_l4_misc(&sb[idx++], &mask, inner, rx); idx 354 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_mpls(&sb[idx++], &mask, inner, rx); idx 357 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_flex_parser_0(&sb[idx++], &mask, inner, rx); idx 361 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c mlx5dr_ste_build_empty_always_hit(&sb[idx++], rx); idx 363 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c if (idx == 0) { idx 376 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c *num_of_builders = idx; idx 52 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c unsigned int idx; idx 57 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c idx = be16_to_cpu(cqe64->wqe_counter) & idx 59 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c dr_cq->qp->sq.cc = dr_cq->qp->sq.wqe_head[idx] + 1; idx 63 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c idx = be16_to_cpu(cqe64->wqe_counter) & idx 65 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c dr_cq->qp->sq.cc = dr_cq->qp->sq.wqe_head[idx] + 1; idx 232 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c unsigned int idx; idx 237 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c idx = dr_qp->sq.pc & (dr_qp->sq.wqe_cnt - 1); idx 239 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c wq_ctrl = mlx5_wq_cyc_get_wqe(&dr_qp->wq.sq, idx); idx 256 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c dr_qp->sq.wqe_head[idx] = dr_qp->sq.pc++; idx 99 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c u16 idx; idx 102 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mlxfw_mfa2_tlv_multi_foreach(mfa2_file, tlv, idx, multi) { idx 228 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c u16 idx; idx 233 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mlxfw_mfa2_tlv_foreach(mfa2_file, tlv, idx, mfa2_file->first_dev, idx 241 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c if (!mlxfw_mfa2_file_dev_validate(mfa2_file, tlv, idx)) idx 246 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mlxfw_mfa2_tlv_foreach(mfa2_file, tlv, idx, mfa2_file->first_component, idx 254 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c if (!mlxfw_mfa2_file_comp_validate(mfa2_file, tlv, idx)) idx 337 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c u32 idx; idx 340 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mlxfw_mfa2_tlv_foreach(mfa2_file, dev_tlv, idx, mfa2_file->first_dev, idx 46 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c u16 idx; idx 48 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c mlxfw_mfa2_tlv_foreach(mfa2_file, tlv, idx, from_tlv, count) idx 61 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c u16 idx; idx 63 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c mlxfw_mfa2_tlv_multi_foreach(mfa2_file, tlv, idx, multi) { idx 82 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c u16 idx; idx 84 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.c mlxfw_mfa2_tlv_multi_foreach(mfa2_file, tlv, idx, multi) { idx 33 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.h #define mlxfw_mfa2_tlv_foreach(mfa2_file, tlv, idx, from_tlv, count) \ idx 34 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.h for (idx = 0, tlv = from_tlv; idx < (count); \ idx 35 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.h idx++, tlv = mlxfw_mfa2_tlv_next(mfa2_file, tlv)) idx 37 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.h #define mlxfw_mfa2_tlv_multi_foreach(mfa2_file, tlv, idx, multi) \ idx 38 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2_tlv_multi.h mlxfw_mfa2_tlv_foreach(mfa2_file, tlv, idx, \ idx 677 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c int err, idx; idx 680 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c idx = mlxsw_get_cooling_device_idx(thermal, cdev); idx 681 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c if (idx < 0) idx 682 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c return idx; idx 684 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c mlxsw_reg_mfsc_pack(mfsc_pl, idx, 0); idx 704 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c int idx; idx 708 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c idx = mlxsw_get_cooling_device_idx(thermal, cdev); idx 709 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c if (idx < 0) idx 710 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c return idx; idx 726 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c mlxsw_reg_mfsc_pack(mfsc_pl, idx, 0); idx 748 drivers/net/ethernet/mellanox/mlxsw/core_thermal.c mlxsw_reg_mfsc_pack(mfsc_pl, idx, mlxsw_state_to_duty(state)); idx 808 drivers/net/ethernet/mscc/ocelot.c int idx; idx 819 drivers/net/ethernet/mscc/ocelot.c if (dump->idx < dump->cb->args[2]) idx 845 drivers/net/ethernet/mscc/ocelot.c dump->idx++; idx 903 drivers/net/ethernet/mscc/ocelot.c struct net_device *filter_dev, int *idx) idx 911 drivers/net/ethernet/mscc/ocelot.c .idx = *idx, idx 937 drivers/net/ethernet/mscc/ocelot.c *idx = dump.idx; idx 1120 drivers/net/ethernet/mscc/ocelot.c unsigned int idx = i * ocelot->num_stats + j; idx 1125 drivers/net/ethernet/mscc/ocelot.c if (val < (ocelot->stats[idx] & U32_MAX)) idx 1126 drivers/net/ethernet/mscc/ocelot.c ocelot->stats[idx] += (u64)1 << 32; idx 1128 drivers/net/ethernet/mscc/ocelot.c ocelot->stats[idx] = (ocelot->stats[idx] & idx 160 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx; idx 1055 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->rx_done.idx = 0; idx 1195 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx; idx 1205 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = rx->fill_cnt & rx->mask; idx 1235 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx->info[idx].page = rx->page; idx 1236 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx->info[idx].page_offset = rx->page_offset; idx 1239 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr_set(&rx->info[idx], bus, rx->bus); idx 1240 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx->shadow[idx].addr_low = idx 1242 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx->shadow[idx].addr_high = idx 1257 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if ((idx & 7) == 7) { idx 1258 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_submit_8rx(&rx->lanai[idx - 7], idx 1259 drivers/net/ethernet/myricom/myri10ge/myri10ge.c &rx->shadow[idx - 7]); idx 1323 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int i, idx, remainder, bytes; idx 1337 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = rx->cnt & rx->mask; idx 1338 drivers/net/ethernet/myricom/myri10ge/myri10ge.c va = page_address(rx->info[idx].page) + rx->info[idx].page_offset; idx 1345 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_rx_page(pdev, &rx->info[idx], bytes); idx 1346 drivers/net/ethernet/myricom/myri10ge/myri10ge.c put_page(rx->info[idx].page); idx 1348 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = rx->cnt & rx->mask; idx 1356 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_rx_page(pdev, &rx->info[idx], bytes); idx 1357 drivers/net/ethernet/myricom/myri10ge/myri10ge.c skb_fill_page_desc(skb, i, rx->info[idx].page, idx 1358 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx->info[idx].page_offset, idx 1362 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = rx->cnt & rx->mask; idx 1393 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx, len; idx 1396 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->done & tx->mask; idx 1397 drivers/net/ethernet/myricom/myri10ge/myri10ge.c skb = tx->info[idx].skb; idx 1400 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = NULL; idx 1401 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (tx->info[idx].last) { idx 1403 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].last = 0; idx 1406 drivers/net/ethernet/myricom/myri10ge/myri10ge.c len = dma_unmap_len(&tx->info[idx], len); idx 1407 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, 0); idx 1414 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], idx 1420 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], idx 1463 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx = rx_done->idx; idx 1469 drivers/net/ethernet/myricom/myri10ge/myri10ge.c while (rx_done->entry[idx].length != 0 && work_done < budget) { idx 1470 drivers/net/ethernet/myricom/myri10ge/myri10ge.c length = ntohs(rx_done->entry[idx].length); idx 1471 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx_done->entry[idx].length = 0; idx 1472 drivers/net/ethernet/myricom/myri10ge/myri10ge.c checksum = csum_unfold(rx_done->entry[idx].checksum); idx 1477 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = cnt & (mgp->max_intr_slots - 1); idx 1480 drivers/net/ethernet/myricom/myri10ge/myri10ge.c rx_done->idx = idx; idx 2040 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx = i & ss->rx_big.mask; idx 2041 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_rx_page(mgp->pdev, &ss->rx_big.info[idx], idx 2043 drivers/net/ethernet/myricom/myri10ge/myri10ge.c put_page(ss->rx_big.info[idx].page); idx 2050 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx = i & ss->rx_small.mask; idx 2051 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_rx_page(mgp->pdev, &ss->rx_small.info[idx], idx 2053 drivers/net/ethernet/myricom/myri10ge/myri10ge.c put_page(ss->rx_small.info[idx].page); idx 2084 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int i, len, idx; idx 2091 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = i & ss->rx_big.mask; idx 2093 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->rx_big.info[idx].page_offset = MYRI10GE_ALLOC_SIZE; idx 2094 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_rx_page(mgp->pdev, &ss->rx_big.info[idx], idx 2096 drivers/net/ethernet/myricom/myri10ge/myri10ge.c put_page(ss->rx_big.info[idx].page); idx 2102 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = i & ss->rx_small.mask; idx 2104 drivers/net/ethernet/myricom/myri10ge/myri10ge.c ss->rx_small.info[idx].page_offset = idx 2106 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_rx_page(mgp->pdev, &ss->rx_small.info[idx], idx 2108 drivers/net/ethernet/myricom/myri10ge/myri10ge.c put_page(ss->rx_small.info[idx].page); idx 2112 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->done & tx->mask; idx 2113 drivers/net/ethernet/myricom/myri10ge/myri10ge.c skb = tx->info[idx].skb; idx 2116 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = NULL; idx 2118 drivers/net/ethernet/myricom/myri10ge/myri10ge.c len = dma_unmap_len(&tx->info[idx], len); idx 2119 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, 0); idx 2125 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], idx 2131 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], idx 2525 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx, starting_slot; idx 2529 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = (starting_slot + cnt) & tx->mask; idx 2530 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_pio_copy(&tx->lanai[idx], &src[cnt], sizeof(*src)); idx 2546 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx, i; idx 2551 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->req & tx->mask; idx 2556 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dst = dstp = &tx->lanai[idx]; idx 2559 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if ((idx + cnt) < tx->mask) { idx 2586 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_tx_buf *tx, int idx) idx 2592 drivers/net/ethernet/myricom/myri10ge/myri10ge.c last_idx = (idx + 1) & tx->mask; idx 2593 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->req & tx->mask; idx 2595 drivers/net/ethernet/myricom/myri10ge/myri10ge.c len = dma_unmap_len(&tx->info[idx], len); idx 2597 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (tx->info[idx].skb != NULL) idx 2599 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], idx 2604 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr(&tx->info[idx], idx 2607 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, 0); idx 2608 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = NULL; idx 2610 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = (idx + 1) & tx->mask; idx 2611 drivers/net/ethernet/myricom/myri10ge/myri10ge.c } while (idx != last_idx); idx 2637 drivers/net/ethernet/myricom/myri10ge/myri10ge.c int idx, avail, frag_cnt, frag_idx, count, mss, max_segments; idx 2733 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = tx->req & tx->mask; idx 2734 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].skb = skb; idx 2735 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr_set(&tx->info[idx], bus, bus); idx 2736 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, len); idx 2839 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_tx_dma(mgp, tx, idx); idx 2842 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = (count + tx->req) & tx->mask; idx 2843 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_addr_set(&tx->info[idx], bus, bus); idx 2844 drivers/net/ethernet/myricom/myri10ge/myri10ge.c dma_unmap_len_set(&tx->info[idx], len, len); idx 2854 drivers/net/ethernet/myricom/myri10ge/myri10ge.c idx = ((count - 1) + tx->req) & tx->mask; idx 2855 drivers/net/ethernet/myricom/myri10ge/myri10ge.c tx->info[idx].last = 1; idx 2872 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_unmap_tx_dma(mgp, tx, idx); idx 1025 drivers/net/ethernet/netronome/nfp/bpf/jit.c u32 idx, src_byte; idx 1034 drivers/net/ethernet/netronome/nfp/bpf/jit.c idx = off / 4; idx 1039 drivers/net/ethernet/netronome/nfp/bpf/jit.c should_inc ? reg_lm_inc(3) : reg_lm(lm3 ? 3 : 0, idx)); idx 1043 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (WARN_ON_ONCE(lm3 && idx > RE_REG_LM_IDX_MAX)) idx 1067 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (idx <= RE_REG_LM_IDX_MAX) { idx 1068 drivers/net/ethernet/netronome/nfp/bpf/jit.c reg = reg_lm(lm3 ? 3 : 0, idx); idx 1077 drivers/net/ethernet/netronome/nfp/bpf/jit.c wrp_mov(nfp_prog, reg, reg_lm(0, idx)); idx 1094 drivers/net/ethernet/netronome/nfp/bpf/jit.c u32 idx, dst_byte; idx 1103 drivers/net/ethernet/netronome/nfp/bpf/jit.c idx = off / 4; idx 1108 drivers/net/ethernet/netronome/nfp/bpf/jit.c should_inc ? reg_lm_inc(3) : reg_lm(lm3 ? 3 : 0, idx), idx 1113 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (WARN_ON_ONCE(lm3 && idx > RE_REG_LM_IDX_MAX)) idx 1137 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (idx <= RE_REG_LM_IDX_MAX) { idx 1138 drivers/net/ethernet/netronome/nfp/bpf/jit.c reg = reg_lm(lm3 ? 3 : 0, idx); idx 1145 drivers/net/ethernet/netronome/nfp/bpf/jit.c wrp_mov(nfp_prog, reg, reg_lm(0, idx)); idx 1151 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (idx > RE_REG_LM_IDX_MAX) idx 1152 drivers/net/ethernet/netronome/nfp/bpf/jit.c wrp_mov(nfp_prog, reg_lm(0, idx), reg); idx 2765 drivers/net/ethernet/netronome/nfp/bpf/jit.c u8 idx = insn_off / REG_WIDTH; idx 2770 drivers/net/ethernet/netronome/nfp/bpf/jit.c src_lo = reg_xfer(idx); idx 2790 drivers/net/ethernet/netronome/nfp/bpf/jit.c src_mid = reg_xfer(idx + 1); idx 2796 drivers/net/ethernet/netronome/nfp/bpf/jit.c swreg src_hi = reg_xfer(idx + 2); idx 2815 drivers/net/ethernet/netronome/nfp/bpf/jit.c u8 dst_gpr, idx; idx 2817 drivers/net/ethernet/netronome/nfp/bpf/jit.c idx = (meta->insn.off - meta->pkt_cache.range_start) / REG_WIDTH; idx 2821 drivers/net/ethernet/netronome/nfp/bpf/jit.c src_lo = reg_xfer(idx); idx 2830 drivers/net/ethernet/netronome/nfp/bpf/jit.c swreg src_hi = reg_xfer(idx + 1); idx 3551 drivers/net/ethernet/netronome/nfp/bpf/jit.c u32 idx, br_idx; idx 3616 drivers/net/ethernet/netronome/nfp/bpf/jit.c for (idx = meta->off; idx <= br_idx; idx++) { idx 3617 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (!nfp_is_br(nfp_prog->prog[idx])) idx 3619 drivers/net/ethernet/netronome/nfp/bpf/jit.c br_set_offset(&nfp_prog->prog[idx], jmp_dst->off); idx 3746 drivers/net/ethernet/netronome/nfp/bpf/jit.c unsigned int idx; idx 3748 drivers/net/ethernet/netronome/nfp/bpf/jit.c for (idx = 1; idx < nfp_prog->subprog_cnt; idx++) idx 3749 drivers/net/ethernet/netronome/nfp/bpf/jit.c if (nfp_prog->subprog[idx].needs_reg_push) idx 3765 drivers/net/ethernet/netronome/nfp/bpf/jit.c u8 idx = (reg - BPF_REG_6) * 2; idx 3770 drivers/net/ethernet/netronome/nfp/bpf/jit.c wrp_mov(nfp_prog, reg_lm(0, 1 + idx), reg_b(adj)); idx 3776 drivers/net/ethernet/netronome/nfp/bpf/jit.c wrp_mov(nfp_prog, reg_lm(0, 1 + idx + 1), reg_b(adj + 1)); idx 3790 drivers/net/ethernet/netronome/nfp/bpf/jit.c u8 idx = (reg - BPF_REG_6) * 2; idx 3795 drivers/net/ethernet/netronome/nfp/bpf/jit.c wrp_mov(nfp_prog, reg_both(adj), reg_lm(0, 1 + idx)); idx 3801 drivers/net/ethernet/netronome/nfp/bpf/jit.c wrp_mov(nfp_prog, reg_both(adj + 1), reg_lm(0, 1 + idx + 1)); idx 700 drivers/net/ethernet/netronome/nfp/bpf/verifier.c unsigned short idx = meta->subprog_idx; idx 709 drivers/net/ethernet/netronome/nfp/bpf/verifier.c frame_depths[frame] = nfp_prog->subprog[idx].stack_depth; idx 715 drivers/net/ethernet/netronome/nfp/bpf/verifier.c for (; meta != nfp_prog_last_meta(nfp_prog) && meta->subprog_idx == idx; idx 724 drivers/net/ethernet/netronome/nfp/bpf/verifier.c ret_prog[frame] = idx; idx 729 drivers/net/ethernet/netronome/nfp/bpf/verifier.c idx = meta->subprog_idx; idx 743 drivers/net/ethernet/netronome/nfp/bpf/verifier.c idx = ret_prog[frame]; idx 62 drivers/net/ethernet/netronome/nfp/flower/cmsg.c nfp_flower_cmsg_mac_repr_add(struct sk_buff *skb, unsigned int idx, idx 69 drivers/net/ethernet/netronome/nfp/flower/cmsg.c msg->ports[idx].idx = idx; idx 70 drivers/net/ethernet/netronome/nfp/flower/cmsg.c msg->ports[idx].info = nbi & NFP_FLOWER_CMSG_MAC_REPR_NBI; idx 71 drivers/net/ethernet/netronome/nfp/flower/cmsg.c msg->ports[idx].nbi_port = nbi_port; idx 72 drivers/net/ethernet/netronome/nfp/flower/cmsg.c msg->ports[idx].phys_port = phys_port; idx 496 drivers/net/ethernet/netronome/nfp/flower/cmsg.h u8 idx; idx 626 drivers/net/ethernet/netronome/nfp/flower/cmsg.h nfp_flower_cmsg_mac_repr_add(struct sk_buff *skb, unsigned int idx, idx 506 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c __nfp_tunnel_offload_mac(struct nfp_app *app, u8 *mac, u16 idx, bool del) idx 517 drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c payload.index = cpu_to_be16(idx); idx 40 drivers/net/ethernet/netronome/nfp/nfp_devlink.c nfp_devlink_set_lanes(struct nfp_pf *pf, unsigned int idx, unsigned int lanes) idx 45 drivers/net/ethernet/netronome/nfp/nfp_devlink.c nsp = nfp_eth_config_start(pf->cpp, idx); idx 113 drivers/net/ethernet/netronome/nfp/nfp_net.h #define D_IDX(ring, idx) ((idx) & ((ring)->cnt - 1)) idx 212 drivers/net/ethernet/netronome/nfp/nfp_net.h u32 idx; idx 338 drivers/net/ethernet/netronome/nfp/nfp_net.h u32 idx; idx 573 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_r_vector *r_vec, unsigned int idx, idx 578 drivers/net/ethernet/netronome/nfp/nfp_net_common.c tx_ring->idx = idx; idx 583 drivers/net/ethernet/netronome/nfp/nfp_net_common.c tx_ring->qcidx = tx_ring->idx * nn->stride_tx; idx 595 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_r_vector *r_vec, unsigned int idx) idx 599 drivers/net/ethernet/netronome/nfp/nfp_net_common.c rx_ring->idx = idx; idx 603 drivers/net/ethernet/netronome/nfp/nfp_net_common.c rx_ring->fl_qcidx = rx_ring->idx * nn->stride_rx; idx 1091 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nd_q = netdev_get_tx_queue(dp->netdev, tx_ring->idx); idx 1162 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx; idx 1164 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx = D_IDX(tx_ring, tx_ring->rd_p++); idx 1165 drivers/net/ethernet/netronome/nfp/nfp_net_common.c tx_buf = &tx_ring->txbufs[idx]; idx 1207 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nd_q = netdev_get_tx_queue(dp->netdev, tx_ring->idx); idx 1227 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx, todo; idx 1245 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx = D_IDX(tx_ring, tx_ring->rd_p); idx 1248 drivers/net/ethernet/netronome/nfp/nfp_net_common.c done_bytes += tx_ring->txbufs[idx].real_len; idx 1279 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx, nr_frags; idx 1281 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx = D_IDX(tx_ring, tx_ring->rd_p); idx 1282 drivers/net/ethernet/netronome/nfp/nfp_net_common.c tx_buf = &tx_ring->txbufs[idx]; idx 1284 drivers/net/ethernet/netronome/nfp/nfp_net_common.c skb = tx_ring->txbufs[idx].skb; idx 1319 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nd_q = netdev_get_tx_queue(dp->netdev, tx_ring->idx); idx 1813 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx; idx 1832 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx = D_IDX(rx_ring, rx_ring->rd_p); idx 1834 drivers/net/ethernet/netronome/nfp/nfp_net_common.c rxd = &rx_ring->rxds[idx]; idx 1848 drivers/net/ethernet/netronome/nfp/nfp_net_common.c rxbuf = &rx_ring->rxbufs[idx]; idx 1995 drivers/net/ethernet/netronome/nfp/nfp_net_common.c skb_record_rx_queue(skb, rx_ring->idx); idx 2206 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx; idx 2208 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx = D_IDX(rx_ring, rx_ring->rd_p); idx 2210 drivers/net/ethernet/netronome/nfp/nfp_net_common.c rxd = &rx_ring->rxds[idx]; idx 2221 drivers/net/ethernet/netronome/nfp/nfp_net_common.c rxbuf = &rx_ring->rxbufs[idx]; idx 2396 drivers/net/ethernet/netronome/nfp/nfp_net_common.c tx_ring->idx); idx 2531 drivers/net/ethernet/netronome/nfp/nfp_net_common.c rx_ring->idx); idx 2604 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_r_vector *r_vec, int idx) idx 2606 drivers/net/ethernet/netronome/nfp/nfp_net_common.c r_vec->rx_ring = idx < dp->num_rx_rings ? &dp->rx_rings[idx] : NULL; idx 2608 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx < dp->num_stack_tx_rings ? &dp->tx_rings[idx] : NULL; idx 2610 drivers/net/ethernet/netronome/nfp/nfp_net_common.c r_vec->xdp_ring = idx < dp->num_tx_rings - dp->num_stack_tx_rings ? idx 2611 drivers/net/ethernet/netronome/nfp/nfp_net_common.c &dp->tx_rings[dp->num_stack_tx_rings + idx] : NULL; idx 2616 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx) idx 2628 drivers/net/ethernet/netronome/nfp/nfp_net_common.c "%s-rxtx-%d", nfp_net_name(nn), idx); idx 2644 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_dbg(nn, "RV%02d: irq=%03d/%03d\n", idx, r_vec->irq_vector, idx 2732 drivers/net/ethernet/netronome/nfp/nfp_net_common.c static void nfp_net_vec_clear_ring_data(struct nfp_net *nn, unsigned int idx) idx 2734 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeq(nn, NFP_NET_CFG_RXR_ADDR(idx), 0); idx 2735 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_RXR_SZ(idx), 0); idx 2736 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_RXR_VEC(idx), 0); idx 2738 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeq(nn, NFP_NET_CFG_TXR_ADDR(idx), 0); idx 2739 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_TXR_SZ(idx), 0); idx 2740 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_TXR_VEC(idx), 0); idx 2784 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_rx_ring *rx_ring, unsigned int idx) idx 2787 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeq(nn, NFP_NET_CFG_RXR_ADDR(idx), rx_ring->dma); idx 2788 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_RXR_SZ(idx), ilog2(rx_ring->cnt)); idx 2789 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_RXR_VEC(idx), rx_ring->r_vec->irq_entry); idx 2794 drivers/net/ethernet/netronome/nfp/nfp_net_common.c struct nfp_net_tx_ring *tx_ring, unsigned int idx) idx 2796 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeq(nn, NFP_NET_CFG_TXR_ADDR(idx), tx_ring->dma); idx 2797 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_TXR_SZ(idx), ilog2(tx_ring->cnt)); idx 2798 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn_writeb(nn, NFP_NET_CFG_TXR_VEC(idx), tx_ring->r_vec->irq_entry); idx 3569 drivers/net/ethernet/netronome/nfp/nfp_net_common.c static void nfp_net_set_vxlan_port(struct nfp_net *nn, int idx, __be16 port) idx 3573 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nn->vxlan_ports[idx] = port; idx 3614 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx; idx 3619 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx = nfp_net_find_vxlan_idx(nn, ti->port); idx 3620 drivers/net/ethernet/netronome/nfp/nfp_net_common.c if (idx == -ENOSPC) idx 3623 drivers/net/ethernet/netronome/nfp/nfp_net_common.c if (!nn->vxlan_usecnt[idx]++) idx 3624 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_net_set_vxlan_port(nn, idx, ti->port); idx 3631 drivers/net/ethernet/netronome/nfp/nfp_net_common.c int idx; idx 3636 drivers/net/ethernet/netronome/nfp/nfp_net_common.c idx = nfp_net_find_vxlan_idx(nn, ti->port); idx 3637 drivers/net/ethernet/netronome/nfp/nfp_net_common.c if (idx == -ENOSPC || !nn->vxlan_usecnt[idx]) idx 3640 drivers/net/ethernet/netronome/nfp/nfp_net_common.c if (!--nn->vxlan_usecnt[idx]) idx 3641 drivers/net/ethernet/netronome/nfp/nfp_net_common.c nfp_net_set_vxlan_port(nn, idx, 0); idx 36 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c rx_ring->idx, rx_ring->fl_qcidx, idx 100 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c tx_ring->idx, tx_ring->qcidx, idx 35 drivers/net/ethernet/netronome/nfp/nfpcore/nfp.h unsigned int idx); idx 69 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nffw.h const struct nfp_rtsym *nfp_rtsym_get(struct nfp_rtsym_table *rtbl, int idx); idx 136 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c unsigned int idx; idx 199 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c return state->idx; idx 203 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c nfp_nsp_config_set_state(struct nfp_nsp *state, void *entries, unsigned int idx) idx 206 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c state->idx = idx; idx 212 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c state->idx = 0; idx 193 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.h int nfp_eth_set_mod_enable(struct nfp_cpp *cpp, unsigned int idx, bool enable); idx 194 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.h int nfp_eth_set_configured(struct nfp_cpp *cpp, unsigned int idx, idx 197 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.h nfp_eth_set_fec(struct nfp_cpp *cpp, unsigned int idx, enum nfp_eth_fec mode); idx 210 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.h struct nfp_nsp *nfp_eth_config_start(struct nfp_cpp *cpp, unsigned int idx); idx 298 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c struct nfp_nsp *nfp_eth_config_start(struct nfp_cpp *cpp, unsigned int idx) idx 320 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c if (!(entries[idx].port & NSP_ETH_PORT_LANES_MASK)) { idx 322 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c idx); idx 326 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c nfp_nsp_config_set_state(nsp, entries, idx); idx 388 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c int nfp_eth_set_mod_enable(struct nfp_cpp *cpp, unsigned int idx, bool enable) idx 394 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c nsp = nfp_eth_config_start(cpp, idx); idx 401 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c reg = le64_to_cpu(entries[idx].state); idx 403 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c reg = le64_to_cpu(entries[idx].control); idx 406 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entries[idx].control = cpu_to_le64(reg); idx 427 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c int nfp_eth_set_configured(struct nfp_cpp *cpp, unsigned int idx, bool configed) idx 433 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c nsp = nfp_eth_config_start(cpp, idx); idx 448 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c reg = le64_to_cpu(entries[idx].state); idx 450 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c reg = le64_to_cpu(entries[idx].control); idx 453 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entries[idx].control = cpu_to_le64(reg); idx 467 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c unsigned int idx = nfp_nsp_config_idx(nsp); idx 480 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c reg = le64_to_cpu(entries[idx].raw[raw_idx]); idx 486 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entries[idx].raw[raw_idx] = cpu_to_le64(reg); idx 488 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c entries[idx].control |= cpu_to_le64(ctrl_bit); idx 548 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c nfp_eth_set_fec(struct nfp_cpp *cpp, unsigned int idx, enum nfp_eth_fec mode) idx 553 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp_eth.c nsp = nfp_eth_config_start(cpp, idx); idx 176 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c const struct nfp_rtsym *nfp_rtsym_get(struct nfp_rtsym_table *rtbl, int idx) idx 180 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c if (idx >= rtbl->num) idx 183 drivers/net/ethernet/netronome/nfp/nfpcore/nfp_rtsym.c return &rtbl->symtab[idx]; idx 617 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c __le32 idx = cpu_to_le32(*((int *)&unirom[adapter->file_prd_off] + idx 629 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c (cpu_to_le32(tab_desc->entry_size) * (idx + 1)); idx 635 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c (cpu_to_le32(tab_desc->entry_size) * (idx)); idx 652 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c __le32 idx = cpu_to_le32(*((int *)&unirom[adapter->file_prd_off] + idx 664 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c (cpu_to_le32(tab_desc->entry_size) * (idx + 1)); idx 670 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c (cpu_to_le32(tab_desc->entry_size) * (idx)); idx 768 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c int idx = cpu_to_le32(*((int *)&unirom[adapter->file_prd_off] + idx 779 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c (cpu_to_le32(tab_desc->entry_size) * idx); idx 11776 drivers/net/ethernet/qlogic/qed/qed_hsi.h #define SECTION_ADDR(_offsize, idx) (MCP_REG_SCRATCH + \ idx 11778 drivers/net/ethernet/qlogic/qed/qed_hsi.h (QED_SECTION_SIZE(_offsize) * idx)) idx 59 drivers/net/ethernet/qlogic/qed/qed_hw.c unsigned int idx; idx 80 drivers/net/ethernet/qlogic/qed/qed_hw.c p_pool->ptts[i].idx = i; idx 129 drivers/net/ethernet/qlogic/qed/qed_hw.c "allocated ptt %d\n", p_ptt->idx); idx 157 drivers/net/ethernet/qlogic/qed/qed_hw.c p_ptt->idx * sizeof(struct pxp_ptt_entry); idx 163 drivers/net/ethernet/qlogic/qed/qed_hw.c p_ptt->idx * PXP_EXTERNAL_BAR_PF_WINDOW_SINGLE_SIZE; idx 179 drivers/net/ethernet/qlogic/qed/qed_hw.c p_ptt->idx, new_hw_addr); idx 201 drivers/net/ethernet/qlogic/qed/qed_hw.c p_ptt->idx, p_ptt->hwfn_id, p_hwfn->my_id); idx 473 drivers/net/ethernet/qlogic/qed/qed_hw.c u32 qed_dmae_idx_to_go_cmd(u8 idx) idx 476 drivers/net/ethernet/qlogic/qed/qed_hw.c return DMAE_REG_GO_C0 + (idx << 2); idx 274 drivers/net/ethernet/qlogic/qed/qed_hw.h u32 qed_dmae_idx_to_go_cmd(u8 idx); idx 1782 drivers/net/ethernet/qlogic/qed/qed_iwarp.c u32 idx; idx 1784 drivers/net/ethernet/qlogic/qed/qed_iwarp.c idx = cid - qed_cxt_get_proto_cid_start(p_hwfn, PROTOCOLID_IWARP); idx 1785 drivers/net/ethernet/qlogic/qed/qed_iwarp.c if (idx >= iwarp_info->max_num_partial_fpdus) { idx 1791 drivers/net/ethernet/qlogic/qed/qed_iwarp.c partial_fpdu = &iwarp_info->partial_fpdus[idx]; idx 265 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c int idx = p_tlv->tlv_type - DRV_TLV_LOCAL_ADMIN_ADDR; idx 267 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->mac_set[idx]) { idx 268 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = p_drv_buf->mac[idx]; idx 445 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c u8 idx; idx 633 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - idx 636 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->crc_err_src_fcid_set[idx]) { idx 637 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->crc_err_src_fcid[idx]; idx 638 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->crc_err_src_fcid[idx]); idx 646 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - DRV_TLV_CRC_ERROR_1_TIMESTAMP) / 2; idx 648 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return qed_mfw_get_tlv_time_value(&p_drv_buf->crc_err[idx], idx 684 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = p_tlv->tlv_type - idx 686 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->flogi_param_set[idx]) { idx 687 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->flogi_param[idx]; idx 688 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->flogi_param[idx]); idx 698 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = p_tlv->tlv_type - idx 701 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->flogi_acc_param_set[idx]) { idx 702 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->flogi_acc_param[idx]; idx 703 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->flogi_acc_param[idx]); idx 759 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - idx 762 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->plogi_dst_fcid_set[idx]) { idx 763 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->plogi_dst_fcid[idx]; idx 764 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->plogi_dst_fcid[idx]); idx 772 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - DRV_TLV_PLOGI_1_TIMESTAMP) / 2; idx 774 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return qed_mfw_get_tlv_time_value(&p_drv_buf->plogi_tstamp[idx], idx 781 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - idx 784 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->plogi_acc_src_fcid_set[idx]) { idx 785 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->plogi_acc_src_fcid[idx]; idx 786 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->plogi_acc_src_fcid[idx]); idx 794 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - DRV_TLV_PLOGI_1_ACC_TIMESTAMP) / 2; idx 795 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_time = &p_drv_buf->plogi_acc_tstamp[idx]; idx 821 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - DRV_TLV_LOGO_1_RECEIVED_SOURCE_FC_ID) / idx 824 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->plogo_src_fcid_set[idx]) { idx 825 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->plogo_src_fcid[idx]; idx 826 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->plogo_src_fcid[idx]); idx 834 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - DRV_TLV_LOGO_1_TIMESTAMP) / 2; idx 836 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return qed_mfw_get_tlv_time_value(&p_drv_buf->plogo_tstamp[idx], idx 885 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - idx 888 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->abts_dst_fcid_set[idx]) { idx 889 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->abts_dst_fcid[idx]; idx 890 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->abts_dst_fcid[idx]); idx 898 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - DRV_TLV_ABTS_1_TIMESTAMP) / 2; idx 900 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return qed_mfw_get_tlv_time_value(&p_drv_buf->abts_tstamp[idx], idx 912 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = p_tlv->tlv_type - DRV_TLV_LAST_RSCN_RECEIVED_N_PORT_1; idx 914 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->rx_rscn_nport_set[idx]) { idx 915 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->rx_rscn_nport[idx]; idx 916 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->rx_rscn_nport[idx]); idx 1050 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - idx 1053 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c if (p_drv_buf->scsi_rx_chk_set[idx]) { idx 1054 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_buf->p_val = &p_drv_buf->scsi_rx_chk[idx]; idx 1055 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c return sizeof(p_drv_buf->scsi_rx_chk[idx]); idx 1063 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c idx = (p_tlv->tlv_type - DRV_TLV_SCSI_CHECK_1_TIMESTAMP) / 2; idx 1064 drivers/net/ethernet/qlogic/qed/qed_mng_tlv.c p_time = &p_drv_buf->scsi_chk_tstamp[idx]; idx 52 drivers/net/ethernet/qlogic/qed/qed_ooo.c u32 idx = (cid & 0xffff) - p_ooo_info->cid_base; idx 55 drivers/net/ethernet/qlogic/qed/qed_ooo.c if (idx >= p_ooo_info->max_num_archipelagos) idx 58 drivers/net/ethernet/qlogic/qed/qed_ooo.c p_archipelago = &p_ooo_info->p_archipelagos_mem[idx]; idx 419 drivers/net/ethernet/qlogic/qed/qed_ooo.c u32 idx = (cid & 0xffff) - p_ooo_info->cid_base; idx 421 drivers/net/ethernet/qlogic/qed/qed_ooo.c p_archipelago = &p_ooo_info->p_archipelagos_mem[idx]; idx 435 drivers/net/ethernet/qlogic/qed/qed_sriov.c u8 idx = 0; idx 451 drivers/net/ethernet/qlogic/qed/qed_sriov.c for (idx = 0; idx < p_iov->total_vfs; idx++) { idx 452 drivers/net/ethernet/qlogic/qed/qed_sriov.c struct qed_vf_info *vf = &p_iov_info->vfs_array[idx]; idx 455 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->vf_mbx.req_virt = p_req_virt_addr + idx; idx 456 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->vf_mbx.req_phys = req_p + idx * sizeof(union vfpf_tlvs); idx 457 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->vf_mbx.reply_virt = p_reply_virt_addr + idx; idx 458 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->vf_mbx.reply_phys = rply_p + idx * sizeof(union pfvf_tlvs); idx 463 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->bulletin.phys = idx * idx 466 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->bulletin.p_virt = p_bulletin_virt + idx; idx 469 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->relative_vf_id = idx; idx 470 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->abs_vf_id = idx + p_iov->first_vf_in_pf; idx 475 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->vport_id = idx + 1; idx 937 drivers/net/ethernet/qlogic/qed/qed_sriov.c int idx, igu_id; idx 941 drivers/net/ethernet/qlogic/qed/qed_sriov.c for (idx = 0; idx < vf->num_sbs; idx++) { idx 942 drivers/net/ethernet/qlogic/qed/qed_sriov.c igu_id = vf->igu_sbs[idx]; idx 417 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_TXQ_IDX_TO_XDP(edev, idx) ((idx) + QEDE_MAX_TSS_CNT(edev)) idx 418 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_NDEV_TXQ_ID_TO_FP_ID(edev, idx) ((edev)->fp_num_rx + \ idx 419 drivers/net/ethernet/qlogic/qede/qede.h ((idx) % QEDE_TSS_COUNT(edev))) idx 420 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_NDEV_TXQ_ID_TO_TXQ_COS(edev, idx) ((idx) / QEDE_TSS_COUNT(edev)) idx 423 drivers/net/ethernet/qlogic/qede/qede.h #define QEDE_NDEV_TXQ_ID_TO_TXQ(edev, idx) \ idx 424 drivers/net/ethernet/qlogic/qede/qede.h (&((edev)->fp_array[QEDE_NDEV_TXQ_ID_TO_FP_ID(edev, idx)].txq \ idx 425 drivers/net/ethernet/qlogic/qede/qede.h [QEDE_NDEV_TXQ_ID_TO_TXQ_COS(edev, idx)])) idx 1496 drivers/net/ethernet/qlogic/qede/qede_ethtool.c int i, idx; idx 1514 drivers/net/ethernet/qlogic/qede/qede_ethtool.c idx = txq->sw_tx_prod; idx 1515 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->sw_tx_ring.skbs[idx].skb = skb; idx 1562 drivers/net/ethernet/qlogic/qede/qede_ethtool.c txq->sw_tx_ring.skbs[idx].skb = NULL; idx 98 drivers/net/ethernet/qlogic/qede/qede_filter.c #define QEDE_ARFS_BUCKET_HEAD(edev, idx) (&(edev)->arfs->arfs_hl_head[idx]) idx 619 drivers/net/ethernet/qlogic/qede/qede_filter.c u16 idx = QEDE_RX_QUEUE_IDX(edev, edev->rss_ind_table[i]); idx 621 drivers/net/ethernet/qlogic/qede/qede_filter.c rss->rss_ind_table[i] = edev->fp_array[idx].rxq->handle; idx 102 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 idx = txq->sw_tx_cons; idx 103 drivers/net/ethernet/qlogic/qede/qede_fp.c struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; idx 108 drivers/net/ethernet/qlogic/qede/qede_fp.c bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; idx 114 drivers/net/ethernet/qlogic/qede/qede_fp.c idx, txq->sw_tx_cons, txq->sw_tx_prod); idx 148 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].skb = NULL; idx 149 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].flags = 0; idx 159 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 idx = txq->sw_tx_prod; idx 160 drivers/net/ethernet/qlogic/qede/qede_fp.c struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; idx 196 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].skb = NULL; idx 197 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].flags = 0; idx 335 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 idx = txq->sw_tx_prod; idx 365 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.xdp[idx].page = metadata->data; idx 366 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.xdp[idx].mapping = metadata->mapping; idx 390 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 hw_bd_cons, idx; idx 397 drivers/net/ethernet/qlogic/qede/qede_fp.c idx = txq->sw_tx_cons; idx 400 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.xdp[idx].mapping, idx 402 drivers/net/ethernet/qlogic/qede/qede_fp.c __free_page(txq->sw_tx_ring.xdp[idx].page); idx 1443 drivers/net/ethernet/qlogic/qede/qede_fp.c u16 idx; idx 1469 drivers/net/ethernet/qlogic/qede/qede_fp.c idx = txq->sw_tx_prod; idx 1470 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].skb = skb; idx 1592 drivers/net/ethernet/qlogic/qede/qede_fp.c txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD; idx 2126 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h void *tmpl_hdr, int idx, u32 value) idx 2128 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h adapter->ahw->hw_ops->set_sys_info(tmpl_hdr, idx, value); idx 566 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c char *buf, u8 idx) idx 711 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c qlcnic_dcb_fill_cee_app_params(struct qlcnic_adapter *adapter, u8 idx, idx 738 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c if (idx == QLC_DCB_OPER_IDX && adapter->netdev->dcbnl_ops) idx 743 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c static void qlcnic_dcb_map_cee_params(struct qlcnic_adapter *adapter, u8 idx) idx 746 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c struct qlcnic_dcb_param *each = &mbx->type[idx]; idx 748 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c struct qlcnic_dcb_cee *type = &cfg->type[idx]; idx 767 drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c qlcnic_dcb_fill_cee_app_params(adapter, idx, each, type); idx 796 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c u32 offs, tab_size, data_size, idx; idx 802 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c idx = le32_to_cpu(temp); idx 809 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c le32_to_cpu(tab_desc->entry_size) * (idx + 1); idx 815 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c le32_to_cpu(tab_desc->entry_size) * idx; idx 832 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c u32 offs, tab_size, data_size, idx; idx 837 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c idx = le32_to_cpu(temp); idx 844 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c le32_to_cpu(tab_desc->entry_size) * (idx + 1); idx 850 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c le32_to_cpu(tab_desc->entry_size) * idx; idx 946 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c u32 offs, idx; idx 950 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c idx = le32_to_cpu(temp); idx 958 drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c le32_to_cpu(tab_desc->entry_size) * idx; idx 433 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c struct net_device *filter_dev, int *idx) idx 439 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c return ndo_dflt_fdb_dump(skb, ncb, netdev, filter_dev, idx); idx 443 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c err = ndo_dflt_fdb_dump(skb, ncb, netdev, filter_dev, idx); idx 309 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c void qlcnic_82xx_set_sys_info(void *t_hdr, int idx, u32 value) idx 313 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c hdr->sys_info[idx] = value; idx 364 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c void qlcnic_83xx_set_sys_info(void *t_hdr, int idx, u32 value) idx 368 drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c hdr->sys_info[idx] = value; idx 141 drivers/net/ethernet/sfc/ef10.c static u32 efx_ef10_make_filter_id(unsigned int pri, u16 idx) idx 143 drivers/net/ethernet/sfc/ef10.c return pri * HUNT_FILTER_TBL_ROWS * 2 + idx; idx 606 drivers/net/ethernet/socionext/netsec.c struct netsec_desc_ring *dring, u16 idx, idx 609 drivers/net/ethernet/socionext/netsec.c struct netsec_de *de = dring->vaddr + DESC_SZ * idx; idx 614 drivers/net/ethernet/socionext/netsec.c if (idx == DESC_NUM - 1) idx 623 drivers/net/ethernet/socionext/netsec.c dring->desc[idx].dma_addr = desc->dma_addr; idx 624 drivers/net/ethernet/socionext/netsec.c dring->desc[idx].addr = desc->addr; idx 625 drivers/net/ethernet/socionext/netsec.c dring->desc[idx].len = desc->len; idx 747 drivers/net/ethernet/socionext/netsec.c u16 idx = from; idx 750 drivers/net/ethernet/socionext/netsec.c netsec_set_rx_de(priv, dring, idx, &dring->desc[idx]); idx 751 drivers/net/ethernet/socionext/netsec.c idx++; idx 752 drivers/net/ethernet/socionext/netsec.c if (idx >= DESC_NUM) idx 753 drivers/net/ethernet/socionext/netsec.c idx = 0; idx 779 drivers/net/ethernet/socionext/netsec.c int idx = dring->head; idx 783 drivers/net/ethernet/socionext/netsec.c de = dring->vaddr + (DESC_SZ * idx); idx 793 drivers/net/ethernet/socionext/netsec.c if (idx == DESC_NUM - 1) idx 801 drivers/net/ethernet/socionext/netsec.c dring->desc[idx] = *desc; idx 803 drivers/net/ethernet/socionext/netsec.c dring->desc[idx].skb = buf; idx 806 drivers/net/ethernet/socionext/netsec.c dring->desc[idx].xdpf = buf; idx 940 drivers/net/ethernet/socionext/netsec.c u16 idx = dring->tail; idx 941 drivers/net/ethernet/socionext/netsec.c struct netsec_de *de = dring->vaddr + (DESC_SZ * idx); idx 942 drivers/net/ethernet/socionext/netsec.c struct netsec_desc *desc = &dring->desc[idx]; idx 975 drivers/net/ethernet/socionext/netsec.c netsec_rx_fill(priv, idx, 1); idx 1043 drivers/net/ethernet/socionext/netsec.c netsec_rx_fill(priv, idx, 1); idx 1182 drivers/net/ethernet/socionext/netsec.c u16 idx; idx 1186 drivers/net/ethernet/socionext/netsec.c for (idx = 0; idx < DESC_NUM; idx++) { idx 1187 drivers/net/ethernet/socionext/netsec.c desc = &dring->desc[idx]; idx 508 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c static int dwxgmac2_rss_write_reg(void __iomem *ioaddr, bool is_key, int idx, idx 514 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c ctrl |= idx << XGMAC_RSSIA_SHIFT; idx 137 drivers/net/ethernet/stmicro/stmmac/stmmac.h int idx; idx 257 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c priv->flow_entries[i].idx = i; idx 429 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, true, idx 437 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, true, idx 477 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, true, idx 485 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, true, idx 564 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l4_filter(priv, priv->hw, entry->idx, false, idx 567 drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c ret = stmmac_config_l3_filter(priv, priv->hw, entry->idx, false, idx 2718 drivers/net/ethernet/sun/niu.c static int niu_set_alt_mac_rdc_table(struct niu *np, int idx, idx 2721 drivers/net/ethernet/sun/niu.c if (idx >= niu_num_alt_addr(np)) idx 2723 drivers/net/ethernet/sun/niu.c return __set_rdc_table_num(np, idx, idx + 1, table_num, mac_pref); idx 3240 drivers/net/ethernet/sun/niu.c static u16 tcam_get_index(struct niu *np, u16 idx) idx 3243 drivers/net/ethernet/sun/niu.c if (idx >= (np->clas.tcam_sz - 1)) idx 3244 drivers/net/ethernet/sun/niu.c idx = 0; idx 3245 drivers/net/ethernet/sun/niu.c return np->clas.tcam_top + ((idx+1) * np->parent->num_ports); idx 3540 drivers/net/ethernet/sun/niu.c static int release_tx_packet(struct niu *np, struct tx_ring_info *rp, int idx) idx 3542 drivers/net/ethernet/sun/niu.c struct tx_buff_info *tb = &rp->tx_buffs[idx]; idx 3559 drivers/net/ethernet/sun/niu.c if (le64_to_cpu(rp->descr[idx]) & TX_DESC_MARK) idx 3564 drivers/net/ethernet/sun/niu.c idx = NEXT_TX(rp, idx); idx 3569 drivers/net/ethernet/sun/niu.c tb = &rp->tx_buffs[idx]; idx 3574 drivers/net/ethernet/sun/niu.c idx = NEXT_TX(rp, idx); idx 3579 drivers/net/ethernet/sun/niu.c return idx; idx 7158 drivers/net/ethernet/sun/niu.c u16 idx; idx 7162 drivers/net/ethernet/sun/niu.c idx = tcam_get_index(np, (u16)nfc->fs.location); idx 7164 drivers/net/ethernet/sun/niu.c tp = &parent->tcam[idx]; idx 7167 drivers/net/ethernet/sun/niu.c parent->index, (u16)nfc->fs.location, idx); idx 7237 drivers/net/ethernet/sun/niu.c int i, idx, cnt; idx 7246 drivers/net/ethernet/sun/niu.c idx = tcam_get_index(np, i); idx 7247 drivers/net/ethernet/sun/niu.c tp = &parent->tcam[idx]; idx 7419 drivers/net/ethernet/sun/niu.c u16 idx; idx 7426 drivers/net/ethernet/sun/niu.c idx = nfc->fs.location; idx 7427 drivers/net/ethernet/sun/niu.c if (idx >= tcam_get_size(np)) idx 7500 drivers/net/ethernet/sun/niu.c idx = tcam_get_index(np, idx); idx 7501 drivers/net/ethernet/sun/niu.c tp = &parent->tcam[idx]; idx 7550 drivers/net/ethernet/sun/niu.c err = tcam_write(np, idx, tp->key, tp->key_mask); idx 7555 drivers/net/ethernet/sun/niu.c err = tcam_assoc_write(np, idx, tp->assoc_data); idx 7574 drivers/net/ethernet/sun/niu.c u16 idx; idx 7584 drivers/net/ethernet/sun/niu.c idx = tcam_get_index(np, loc); idx 7585 drivers/net/ethernet/sun/niu.c tp = &parent->tcam[idx]; idx 7618 drivers/net/ethernet/sun/niu.c ret = tcam_flush(np, idx); idx 8545 drivers/net/ethernet/sun/niu.c u8 idx; idx 8571 drivers/net/ethernet/sun/niu.c idx = p->cur[type]; idx 8572 drivers/net/ethernet/sun/niu.c p->phy_id[type][idx] = id; idx 8573 drivers/net/ethernet/sun/niu.c p->phy_port[type][idx] = phy_port; idx 8574 drivers/net/ethernet/sun/niu.c p->cur[type] = idx + 1; idx 672 drivers/net/ethernet/sun/sunvnet_common.c u32 idx = dr->cons; idx 675 drivers/net/ethernet/sun/sunvnet_common.c while (idx != dr->prod) { idx 676 drivers/net/ethernet/sun/sunvnet_common.c if (idx == end) { idx 680 drivers/net/ethernet/sun/sunvnet_common.c idx = vio_dring_next(dr, idx); idx 286 drivers/net/ethernet/synopsys/dwc-xlgmac-common.c unsigned int idx, idx 294 drivers/net/ethernet/synopsys/dwc-xlgmac-common.c desc_data = XLGMAC_GET_DESC_DATA(ring, idx); idx 300 drivers/net/ethernet/synopsys/dwc-xlgmac-common.c "TX_NORMAL_DESC[%d %s] = %08x:%08x:%08x:%08x\n", idx, idx 307 drivers/net/ethernet/synopsys/dwc-xlgmac-common.c idx++; idx 313 drivers/net/ethernet/synopsys/dwc-xlgmac-common.c unsigned int idx) idx 318 drivers/net/ethernet/synopsys/dwc-xlgmac-common.c desc_data = XLGMAC_GET_DESC_DATA(ring, idx); idx 325 drivers/net/ethernet/synopsys/dwc-xlgmac-common.c idx, idx 94 drivers/net/ethernet/synopsys/dwc-xlgmac.h #define XLGMAC_GET_DESC_DATA(ring, idx) ({ \ idx 97 drivers/net/ethernet/synopsys/dwc-xlgmac.h ((idx) & ((_ring)->dma_desc_count - 1))); \ idx 638 drivers/net/ethernet/synopsys/dwc-xlgmac.h unsigned int idx, idx 643 drivers/net/ethernet/synopsys/dwc-xlgmac.h unsigned int idx); idx 1086 drivers/net/ethernet/tehuti/tehuti.c int dno, delta, idx; idx 1098 drivers/net/ethernet/tehuti/tehuti.c idx = bdx_rxdb_alloc_elem(db); idx 1099 drivers/net/ethernet/tehuti/tehuti.c dm = bdx_rxdb_addr_elem(db, idx); idx 1106 drivers/net/ethernet/tehuti/tehuti.c rxfd->va_lo = idx; idx 63 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 65 drivers/net/ethernet/ti/cpsw_ale.c idx = start / 32; idx 66 drivers/net/ethernet/ti/cpsw_ale.c start -= idx * 32; idx 67 drivers/net/ethernet/ti/cpsw_ale.c idx = 2 - idx; /* flip */ idx 68 drivers/net/ethernet/ti/cpsw_ale.c return (ale_entry[idx] >> start) & BITMASK(bits); idx 74 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 77 drivers/net/ethernet/ti/cpsw_ale.c idx = start / 32; idx 78 drivers/net/ethernet/ti/cpsw_ale.c start -= idx * 32; idx 79 drivers/net/ethernet/ti/cpsw_ale.c idx = 2 - idx; /* flip */ idx 80 drivers/net/ethernet/ti/cpsw_ale.c ale_entry[idx] &= ~(BITMASK(bits) << start); idx 81 drivers/net/ethernet/ti/cpsw_ale.c ale_entry[idx] |= (value << start); idx 140 drivers/net/ethernet/ti/cpsw_ale.c static int cpsw_ale_read(struct cpsw_ale *ale, int idx, u32 *ale_entry) idx 144 drivers/net/ethernet/ti/cpsw_ale.c WARN_ON(idx > ale->params.ale_entries); idx 146 drivers/net/ethernet/ti/cpsw_ale.c writel_relaxed(idx, ale->params.ale_regs + ALE_TABLE_CONTROL); idx 152 drivers/net/ethernet/ti/cpsw_ale.c return idx; idx 155 drivers/net/ethernet/ti/cpsw_ale.c static int cpsw_ale_write(struct cpsw_ale *ale, int idx, u32 *ale_entry) idx 159 drivers/net/ethernet/ti/cpsw_ale.c WARN_ON(idx > ale->params.ale_entries); idx 165 drivers/net/ethernet/ti/cpsw_ale.c writel_relaxed(idx | ALE_TABLE_WRITE, ale->params.ale_regs + idx 168 drivers/net/ethernet/ti/cpsw_ale.c return idx; idx 174 drivers/net/ethernet/ti/cpsw_ale.c int type, idx; idx 176 drivers/net/ethernet/ti/cpsw_ale.c for (idx = 0; idx < ale->params.ale_entries; idx++) { idx 179 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 187 drivers/net/ethernet/ti/cpsw_ale.c return idx; idx 195 drivers/net/ethernet/ti/cpsw_ale.c int type, idx; idx 197 drivers/net/ethernet/ti/cpsw_ale.c for (idx = 0; idx < ale->params.ale_entries; idx++) { idx 198 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 203 drivers/net/ethernet/ti/cpsw_ale.c return idx; idx 211 drivers/net/ethernet/ti/cpsw_ale.c int type, idx; idx 213 drivers/net/ethernet/ti/cpsw_ale.c for (idx = 0; idx < ale->params.ale_entries; idx++) { idx 214 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 217 drivers/net/ethernet/ti/cpsw_ale.c return idx; idx 225 drivers/net/ethernet/ti/cpsw_ale.c int type, idx; idx 227 drivers/net/ethernet/ti/cpsw_ale.c for (idx = 0; idx < ale->params.ale_entries; idx++) { idx 228 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 237 drivers/net/ethernet/ti/cpsw_ale.c return idx; idx 264 drivers/net/ethernet/ti/cpsw_ale.c int ret, idx; idx 266 drivers/net/ethernet/ti/cpsw_ale.c for (idx = 0; idx < ale->params.ale_entries; idx++) { idx 267 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 291 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 311 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 321 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_addr(ale, addr, (flags & ALE_VLAN) ? vid : 0); idx 322 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 323 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_free(ale); idx 324 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 325 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_find_ageable(ale); idx 326 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 329 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 337 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 339 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_addr(ale, addr, (flags & ALE_VLAN) ? vid : 0); idx 340 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 344 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 352 drivers/net/ethernet/ti/cpsw_ale.c int idx, mask; idx 354 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_addr(ale, addr, (flags & ALE_VLAN) ? vid : 0); idx 355 drivers/net/ethernet/ti/cpsw_ale.c if (idx >= 0) idx 356 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 370 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 371 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_free(ale); idx 372 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 373 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_find_ageable(ale); idx 374 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 377 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 385 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 387 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_addr(ale, addr, (flags & ALE_VLAN) ? vid : 0); idx 388 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 391 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 399 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 407 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 410 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_get_vlan_reg_mcast_idx(ale_entry); idx 411 drivers/net/ethernet/ti/cpsw_ale.c writel(reg_mcast, ale->params.ale_regs + ALE_VLAN_MASK_MUX(idx)); idx 414 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_get_vlan_unreg_mcast_idx(ale_entry); idx 415 drivers/net/ethernet/ti/cpsw_ale.c writel(unreg_mcast, ale->params.ale_regs + ALE_VLAN_MASK_MUX(idx)); idx 422 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 424 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_vlan(ale, vid); idx 425 drivers/net/ethernet/ti/cpsw_ale.c if (idx >= 0) idx 426 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 442 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 443 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_free(ale); idx 444 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 445 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_find_ageable(ale); idx 446 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 449 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 456 drivers/net/ethernet/ti/cpsw_ale.c int idx; idx 458 drivers/net/ethernet/ti/cpsw_ale.c idx = cpsw_ale_match_vlan(ale, vid); idx 459 drivers/net/ethernet/ti/cpsw_ale.c if (idx < 0) idx 462 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 470 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 478 drivers/net/ethernet/ti/cpsw_ale.c int type, idx; idx 480 drivers/net/ethernet/ti/cpsw_ale.c for (idx = 0; idx < ale->params.ale_entries; idx++) { idx 483 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_read(ale, idx, ale_entry); idx 503 drivers/net/ethernet/ti/cpsw_ale.c cpsw_ale_write(ale, idx, ale_entry); idx 1406 drivers/net/ethernet/via/via-rhine.c static void rhine_set_cam(void __iomem *ioaddr, int idx, u8 *addr) idx 1414 drivers/net/ethernet/via/via-rhine.c idx &= (MCAM_SIZE - 1); idx 1416 drivers/net/ethernet/via/via-rhine.c iowrite8((u8) idx, ioaddr + CamAddr); idx 1437 drivers/net/ethernet/via/via-rhine.c static void rhine_set_vlan_cam(void __iomem *ioaddr, int idx, u8 *addr) idx 1443 drivers/net/ethernet/via/via-rhine.c idx &= (VCAM_SIZE - 1); idx 1445 drivers/net/ethernet/via/via-rhine.c iowrite8((u8) idx, ioaddr + CamAddr); idx 174 drivers/net/ethernet/via/via-velocity.c static void mac_set_cam(struct mac_regs __iomem *regs, int idx, const u8 *addr) idx 181 drivers/net/ethernet/via/via-velocity.c idx &= (64 - 1); idx 183 drivers/net/ethernet/via/via-velocity.c writeb(CAMADDR_CAMEN | idx, ®s->CAMADDR); idx 198 drivers/net/ethernet/via/via-velocity.c static void mac_set_vlan_cam(struct mac_regs __iomem *regs, int idx, idx 205 drivers/net/ethernet/via/via-velocity.c idx &= (64 - 1); idx 207 drivers/net/ethernet/via/via-velocity.c writeb(CAMADDR_CAMEN | CAMADDR_VCAMSL | idx, ®s->CAMADDR); idx 1514 drivers/net/ethernet/via/via-velocity.c static int velocity_alloc_rx_buf(struct velocity_info *vptr, int idx) idx 1516 drivers/net/ethernet/via/via-velocity.c struct rx_desc *rd = &(vptr->rx.ring[idx]); idx 1517 drivers/net/ethernet/via/via-velocity.c struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]); idx 1893 drivers/net/ethernet/via/via-velocity.c int idx; idx 1899 drivers/net/ethernet/via/via-velocity.c for (idx = vptr->tx.tail[qnum]; vptr->tx.used[qnum] > 0; idx 1900 drivers/net/ethernet/via/via-velocity.c idx = (idx + 1) % vptr->options.numtx) { idx 1905 drivers/net/ethernet/via/via-velocity.c td = &(vptr->tx.rings[qnum][idx]); idx 1906 drivers/net/ethernet/via/via-velocity.c tdinfo = &(vptr->tx.infos[qnum][idx]); idx 1932 drivers/net/ethernet/via/via-velocity.c vptr->tx.tail[qnum] = idx; idx 2029 drivers/net/ethernet/via/via-velocity.c static int velocity_receive_frame(struct velocity_info *vptr, int idx) idx 2032 drivers/net/ethernet/via/via-velocity.c struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]); idx 2033 drivers/net/ethernet/via/via-velocity.c struct rx_desc *rd = &(vptr->rx.ring[idx]); idx 1425 drivers/net/ethernet/via/via-velocity.h #define GET_RD_BY_IDX(vptr, idx) (vptr->rd_ring[idx]) idx 482 drivers/net/fjes/fjes_hw.c int i, idx; idx 497 drivers/net/fjes/fjes_hw.c idx = 0; idx 498 drivers/net/fjes/fjes_hw.c req_buf->share_buffer.buffer[idx++] = buf_pair->tx.size; idx 503 drivers/net/fjes/fjes_hw.c req_buf->share_buffer.buffer[idx++] = idx 508 drivers/net/fjes/fjes_hw.c req_buf->share_buffer.buffer[idx++] = buf_pair->rx.size; idx 513 drivers/net/fjes/fjes_hw.c req_buf->share_buffer.buffer[idx++] = idx 820 drivers/net/fjes/fjes_main.c int idx, epidx; idx 822 drivers/net/fjes/fjes_main.c for (idx = 0; fjes_support_mtu[idx] != 0; idx++) { idx 823 drivers/net/fjes/fjes_main.c if (new_mtu <= fjes_support_mtu[idx]) { idx 824 drivers/net/fjes/fjes_main.c new_mtu = fjes_support_mtu[idx]; idx 251 drivers/net/hippi/rrunner.c u32 idx; idx 266 drivers/net/hippi/rrunner.c idx = rrpriv->info->cmd_ctrl.pi; idx 268 drivers/net/hippi/rrunner.c writel(*(u32*)(cmd), ®s->CmdRing[idx]); idx 271 drivers/net/hippi/rrunner.c idx = (idx - 1) % CMD_RING_ENTRIES; idx 272 drivers/net/hippi/rrunner.c rrpriv->info->cmd_ctrl.pi = idx; idx 70 drivers/net/ieee802154/mac802154_hwsim.c u32 idx; idx 185 drivers/net/ieee802154/mac802154_hwsim.c s64 idx = -1; idx 190 drivers/net/ieee802154/mac802154_hwsim.c idx = nla_get_u32(info->attrs[MAC802154_HWSIM_ATTR_RADIO_ID]); idx 194 drivers/net/ieee802154/mac802154_hwsim.c if (idx == phy->idx) { idx 212 drivers/net/ieee802154/mac802154_hwsim.c ret = nla_put_u32(skb, MAC802154_HWSIM_ATTR_RADIO_ID, phy->idx); idx 239 drivers/net/ieee802154/mac802154_hwsim.c e->endpoint->idx); idx 297 drivers/net/ieee802154/mac802154_hwsim.c int idx, res = -ENODEV; idx 301 drivers/net/ieee802154/mac802154_hwsim.c idx = nla_get_u32(info->attrs[MAC802154_HWSIM_ATTR_RADIO_ID]); idx 305 drivers/net/ieee802154/mac802154_hwsim.c if (phy->idx != idx) idx 334 drivers/net/ieee802154/mac802154_hwsim.c int idx = cb->args[0]; idx 340 drivers/net/ieee802154/mac802154_hwsim.c if (idx == hwsim_radio_idx) idx 344 drivers/net/ieee802154/mac802154_hwsim.c if (phy->idx < idx) idx 352 drivers/net/ieee802154/mac802154_hwsim.c idx = phy->idx + 1; idx 355 drivers/net/ieee802154/mac802154_hwsim.c cb->args[0] = idx; idx 363 drivers/net/ieee802154/mac802154_hwsim.c static struct hwsim_phy *hwsim_get_radio_by_id(uint32_t idx) idx 368 drivers/net/ieee802154/mac802154_hwsim.c if (phy->idx == idx) idx 452 drivers/net/ieee802154/mac802154_hwsim.c if (e->endpoint->idx == v1) { idx 505 drivers/net/ieee802154/mac802154_hwsim.c if (e->endpoint->idx == v1) { idx 561 drivers/net/ieee802154/mac802154_hwsim.c if (e->endpoint->idx == v1) { idx 684 drivers/net/ieee802154/mac802154_hwsim.c if (e->endpoint->idx == phy->idx) { idx 736 drivers/net/ieee802154/mac802154_hwsim.c int idx; idx 739 drivers/net/ieee802154/mac802154_hwsim.c idx = hwsim_radio_idx++; idx 790 drivers/net/ieee802154/mac802154_hwsim.c phy->idx = idx; idx 813 drivers/net/ieee802154/mac802154_hwsim.c return idx; idx 62 drivers/net/ipvlan/ipvlan_main.c int err, idx; idx 72 drivers/net/ipvlan/ipvlan_main.c for (idx = 0; idx < IPVLAN_HASH_SIZE; idx++) idx 73 drivers/net/ipvlan/ipvlan_main.c INIT_HLIST_HEAD(&port->hlhead[idx]); idx 277 drivers/net/ipvlan/ipvlan_main.c int idx; idx 279 drivers/net/ipvlan/ipvlan_main.c for_each_possible_cpu(idx) { idx 280 drivers/net/ipvlan/ipvlan_main.c pcptr = per_cpu_ptr(ipvlan->pcpu_stats, idx); idx 124 drivers/net/macvlan.c u32 idx = macvlan_eth_hash(addr); idx 126 drivers/net/macvlan.c hlist_for_each_entry_rcu(vlan, &port->vlan_hash[idx], hlist) { idx 138 drivers/net/macvlan.c u32 idx = macvlan_eth_hash(addr); idx 139 drivers/net/macvlan.c struct hlist_head *h = &vlan->port->vlan_source_hash[idx]; idx 177 drivers/net/macvlan.c u32 idx = macvlan_eth_hash(addr); idx 179 drivers/net/macvlan.c hlist_add_head_rcu(&vlan->hlist, &port->vlan_hash[idx]); idx 428 drivers/net/macvlan.c u32 idx = macvlan_eth_hash(addr); idx 429 drivers/net/macvlan.c struct hlist_head *h = &port->vlan_source_hash[idx]; idx 346 drivers/net/netdevsim/bpf.c nsim_map_alloc_elem(struct bpf_offloaded_map *offmap, unsigned int idx) idx 350 drivers/net/netdevsim/bpf.c nmap->entry[idx].key = kmalloc(offmap->map.key_size, GFP_USER); idx 351 drivers/net/netdevsim/bpf.c if (!nmap->entry[idx].key) idx 353 drivers/net/netdevsim/bpf.c nmap->entry[idx].value = kmalloc(offmap->map.value_size, GFP_USER); idx 354 drivers/net/netdevsim/bpf.c if (!nmap->entry[idx].value) { idx 355 drivers/net/netdevsim/bpf.c kfree(nmap->entry[idx].key); idx 356 drivers/net/netdevsim/bpf.c nmap->entry[idx].key = NULL; idx 368 drivers/net/netdevsim/bpf.c int idx = -ENOENT; idx 373 drivers/net/netdevsim/bpf.c idx = nsim_map_key_find(offmap, key); idx 374 drivers/net/netdevsim/bpf.c if (idx == -ENOENT) idx 375 drivers/net/netdevsim/bpf.c idx = 0; idx 377 drivers/net/netdevsim/bpf.c idx++; idx 379 drivers/net/netdevsim/bpf.c for (; idx < ARRAY_SIZE(nmap->entry); idx++) { idx 380 drivers/net/netdevsim/bpf.c if (nmap->entry[idx].key) { idx 381 drivers/net/netdevsim/bpf.c memcpy(next_key, nmap->entry[idx].key, idx 389 drivers/net/netdevsim/bpf.c if (idx == ARRAY_SIZE(nmap->entry)) idx 398 drivers/net/netdevsim/bpf.c int idx; idx 402 drivers/net/netdevsim/bpf.c idx = nsim_map_key_find(offmap, key); idx 403 drivers/net/netdevsim/bpf.c if (idx >= 0) idx 404 drivers/net/netdevsim/bpf.c memcpy(value, nmap->entry[idx].value, offmap->map.value_size); idx 408 drivers/net/netdevsim/bpf.c return idx < 0 ? idx : 0; idx 416 drivers/net/netdevsim/bpf.c int idx, err = 0; idx 420 drivers/net/netdevsim/bpf.c idx = nsim_map_key_find(offmap, key); idx 421 drivers/net/netdevsim/bpf.c if (idx < 0 && flags == BPF_EXIST) { idx 422 drivers/net/netdevsim/bpf.c err = idx; idx 425 drivers/net/netdevsim/bpf.c if (idx >= 0 && flags == BPF_NOEXIST) { idx 430 drivers/net/netdevsim/bpf.c if (idx < 0) { idx 431 drivers/net/netdevsim/bpf.c for (idx = 0; idx < ARRAY_SIZE(nmap->entry); idx++) idx 432 drivers/net/netdevsim/bpf.c if (!nmap->entry[idx].key) idx 434 drivers/net/netdevsim/bpf.c if (idx == ARRAY_SIZE(nmap->entry)) { idx 439 drivers/net/netdevsim/bpf.c err = nsim_map_alloc_elem(offmap, idx); idx 444 drivers/net/netdevsim/bpf.c memcpy(nmap->entry[idx].key, key, offmap->map.key_size); idx 445 drivers/net/netdevsim/bpf.c memcpy(nmap->entry[idx].value, value, offmap->map.value_size); idx 455 drivers/net/netdevsim/bpf.c int idx; idx 462 drivers/net/netdevsim/bpf.c idx = nsim_map_key_find(offmap, key); idx 463 drivers/net/netdevsim/bpf.c if (idx >= 0) { idx 464 drivers/net/netdevsim/bpf.c kfree(nmap->entry[idx].key); idx 465 drivers/net/netdevsim/bpf.c kfree(nmap->entry[idx].value); idx 466 drivers/net/netdevsim/bpf.c memset(&nmap->entry[idx], 0, sizeof(nmap->entry[idx])); idx 471 drivers/net/netdevsim/bpf.c return idx < 0 ? idx : 0; idx 523 drivers/net/ppp/bsd_comp.c #define dict_ptrx(p,idx) &(p->dict[idx]) idx 524 drivers/net/ppp/bsd_comp.c #define lens_ptrx(p,idx) &(p->lens[idx]) idx 527 drivers/net/ppp/bsd_comp.c static unsigned short *lens_ptr(struct bsd_db *db, int idx) idx 529 drivers/net/ppp/bsd_comp.c if ((unsigned int) idx > (unsigned int) db->maxmaxcode) idx 531 drivers/net/ppp/bsd_comp.c printk ("<9>ppp: lens_ptr(%d) > max\n", idx); idx 532 drivers/net/ppp/bsd_comp.c idx = 0; idx 534 drivers/net/ppp/bsd_comp.c return lens_ptrx (db, idx); idx 537 drivers/net/ppp/bsd_comp.c static struct bsd_dict *dict_ptr(struct bsd_db *db, int idx) idx 539 drivers/net/ppp/bsd_comp.c if ((unsigned int) idx >= (unsigned int) db->hsize) idx 541 drivers/net/ppp/bsd_comp.c printk ("<9>ppp: dict_ptr(%d) > max\n", idx); idx 542 drivers/net/ppp/bsd_comp.c idx = 0; idx 544 drivers/net/ppp/bsd_comp.c return dict_ptrx (db, idx); idx 548 drivers/net/ppp/bsd_comp.c #define lens_ptr(db,idx) lens_ptrx(db,idx) idx 549 drivers/net/ppp/bsd_comp.c #define dict_ptr(db,idx) dict_ptrx(db,idx) idx 519 drivers/net/usb/hso.c int idx; idx 521 drivers/net/usb/hso.c for (idx = 0; idx < serial->num_rx_urbs; idx++) idx 522 drivers/net/usb/hso.c if (serial->rx_urb[idx] == urb) idx 523 drivers/net/usb/hso.c return idx; idx 1726 drivers/net/usb/lan78xx.c static int lan78xx_mdiobus_read(struct mii_bus *bus, int phy_id, int idx) idx 1744 drivers/net/usb/lan78xx.c addr = mii_access(phy_id, idx, MII_READ); idx 1762 drivers/net/usb/lan78xx.c static int lan78xx_mdiobus_write(struct mii_bus *bus, int phy_id, int idx, idx 1784 drivers/net/usb/lan78xx.c addr = mii_access(phy_id, idx, MII_WRITE); idx 179 drivers/net/usb/smsc75xx.c static int __smsc75xx_mdio_read(struct net_device *netdev, int phy_id, int idx, idx 197 drivers/net/usb/smsc75xx.c idx &= dev->mii.reg_num_mask; idx 199 drivers/net/usb/smsc75xx.c | ((idx << MII_ACCESS_REG_ADDR_SHIFT) & MII_ACCESS_REG_ADDR) idx 209 drivers/net/usb/smsc75xx.c netdev_warn(dev->net, "Timed out reading MII reg %02X\n", idx); idx 227 drivers/net/usb/smsc75xx.c int idx, int regval, int in_pm) idx 251 drivers/net/usb/smsc75xx.c idx &= dev->mii.reg_num_mask; idx 253 drivers/net/usb/smsc75xx.c | ((idx << MII_ACCESS_REG_ADDR_SHIFT) & MII_ACCESS_REG_ADDR) idx 263 drivers/net/usb/smsc75xx.c netdev_warn(dev->net, "Timed out writing MII reg %02X\n", idx); idx 272 drivers/net/usb/smsc75xx.c int idx) idx 274 drivers/net/usb/smsc75xx.c return __smsc75xx_mdio_read(netdev, phy_id, idx, 1); idx 278 drivers/net/usb/smsc75xx.c int idx, int regval) idx 280 drivers/net/usb/smsc75xx.c __smsc75xx_mdio_write(netdev, phy_id, idx, regval, 1); idx 283 drivers/net/usb/smsc75xx.c static int smsc75xx_mdio_read(struct net_device *netdev, int phy_id, int idx) idx 285 drivers/net/usb/smsc75xx.c return __smsc75xx_mdio_read(netdev, phy_id, idx, 0); idx 288 drivers/net/usb/smsc75xx.c static void smsc75xx_mdio_write(struct net_device *netdev, int phy_id, int idx, idx 291 drivers/net/usb/smsc75xx.c __smsc75xx_mdio_write(netdev, phy_id, idx, regval, 0); idx 176 drivers/net/usb/smsc95xx.c static int __smsc95xx_mdio_read(struct net_device *netdev, int phy_id, int idx, idx 194 drivers/net/usb/smsc95xx.c idx &= dev->mii.reg_num_mask; idx 195 drivers/net/usb/smsc95xx.c addr = (phy_id << 11) | (idx << 6) | MII_READ_ | MII_BUSY_; idx 204 drivers/net/usb/smsc95xx.c netdev_warn(dev->net, "Timed out reading MII reg %02X\n", idx); idx 222 drivers/net/usb/smsc95xx.c int idx, int regval, int in_pm) idx 246 drivers/net/usb/smsc95xx.c idx &= dev->mii.reg_num_mask; idx 247 drivers/net/usb/smsc95xx.c addr = (phy_id << 11) | (idx << 6) | MII_WRITE_ | MII_BUSY_; idx 256 drivers/net/usb/smsc95xx.c netdev_warn(dev->net, "Timed out writing MII reg %02X\n", idx); idx 265 drivers/net/usb/smsc95xx.c int idx) idx 267 drivers/net/usb/smsc95xx.c return __smsc95xx_mdio_read(netdev, phy_id, idx, 1); idx 271 drivers/net/usb/smsc95xx.c int idx, int regval) idx 273 drivers/net/usb/smsc95xx.c __smsc95xx_mdio_write(netdev, phy_id, idx, regval, 1); idx 276 drivers/net/usb/smsc95xx.c static int smsc95xx_mdio_read(struct net_device *netdev, int phy_id, int idx) idx 278 drivers/net/usb/smsc95xx.c return __smsc95xx_mdio_read(netdev, phy_id, idx, 0); idx 281 drivers/net/usb/smsc95xx.c static void smsc95xx_mdio_write(struct net_device *netdev, int phy_id, int idx, idx 284 drivers/net/usb/smsc95xx.c __smsc95xx_mdio_write(netdev, phy_id, idx, regval, 0); idx 152 drivers/net/veth.c int i, j, idx; idx 155 drivers/net/veth.c idx = 1; idx 166 drivers/net/veth.c data[idx + j] = *(u64 *)(stats_base + offset); idx 169 drivers/net/veth.c idx += VETH_RQ_STATS_LEN; idx 2125 drivers/net/virtio_net.c unsigned int idx = 0, start, i, j; idx 2137 drivers/net/virtio_net.c data[idx + j] = *(u64 *)(stats_base + offset); idx 2140 drivers/net/virtio_net.c idx += VIRTNET_RQ_STATS_LEN; idx 2151 drivers/net/virtio_net.c data[idx + j] = *(u64 *)(stats_base + offset); idx 2154 drivers/net/virtio_net.c idx += VIRTNET_SQ_STATS_LEN; idx 735 drivers/net/vmxnet3/vmxnet3_defs.h #define VMXNET3_INC_RING_IDX_ONLY(idx, ring_size) \ idx 737 drivers/net/vmxnet3/vmxnet3_defs.h (idx)++;\ idx 738 drivers/net/vmxnet3/vmxnet3_defs.h if (unlikely((idx) == (ring_size))) {\ idx 739 drivers/net/vmxnet3/vmxnet3_defs.h (idx) = 0;\ idx 1302 drivers/net/vmxnet3/vmxnet3_drv.c u32 idx, ring_idx; idx 1318 drivers/net/vmxnet3/vmxnet3_drv.c idx = rcd->rxdIdx; idx 1321 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_getRxDesc(rxd, &rq->rx_ring[ring_idx].base[idx].rxd, idx 1323 drivers/net/vmxnet3/vmxnet3_drv.c rbi = rq->buf_info[ring_idx] + idx; idx 1349 drivers/net/vmxnet3/vmxnet3_drv.c ring_idx, idx); idx 1543 drivers/net/vmxnet3/vmxnet3_drv.c ring->next2comp = idx; idx 1219 drivers/net/vxlan.c struct net_device *filter_dev, int *idx) idx 1232 drivers/net/vxlan.c if (*idx < cb->args[2]) idx 1243 drivers/net/vxlan.c *idx += 1; idx 344 drivers/net/wireless/ath/ath10k/ce.h #define CE_SRC_RING_TO_DESC(baddr, idx) \ idx 345 drivers/net/wireless/ath/ath10k/ce.h (&(((struct ce_desc *)baddr)[idx])) idx 347 drivers/net/wireless/ath/ath10k/ce.h #define CE_DEST_RING_TO_DESC(baddr, idx) \ idx 348 drivers/net/wireless/ath/ath10k/ce.h (&(((struct ce_desc *)baddr)[idx])) idx 350 drivers/net/wireless/ath/ath10k/ce.h #define CE_SRC_RING_TO_DESC_64(baddr, idx) \ idx 351 drivers/net/wireless/ath/ath10k/ce.h (&(((struct ce_desc_64 *)baddr)[idx])) idx 353 drivers/net/wireless/ath/ath10k/ce.h #define CE_DEST_RING_TO_DESC_64(baddr, idx) \ idx 354 drivers/net/wireless/ath/ath10k/ce.h (&(((struct ce_desc_64 *)baddr)[idx])) idx 360 drivers/net/wireless/ath/ath10k/ce.h #define CE_RING_IDX_INCR(nentries_mask, idx) (((idx) + 1) & (nentries_mask)) idx 361 drivers/net/wireless/ath/ath10k/ce.h #define CE_RING_IDX_ADD(nentries_mask, idx, num) \ idx 362 drivers/net/wireless/ath/ath10k/ce.h (((idx) + (num)) & (nentries_mask)) idx 247 drivers/net/wireless/ath/ath10k/htc.h #define SVC(group, idx) \ idx 248 drivers/net/wireless/ath/ath10k/htc.h (int)(((int)(group) << 8) | (int)(idx)) idx 2106 drivers/net/wireless/ath/ath10k/htt.h int idx); idx 2108 drivers/net/wireless/ath/ath10k/htt.h void (*htt_reset_paddrs_ring)(struct ath10k_htt *htt, int idx); idx 2131 drivers/net/wireless/ath/ath10k/htt.h int idx) idx 2134 drivers/net/wireless/ath/ath10k/htt.h htt->rx_ops->htt_set_paddrs_ring(htt, paddr, idx); idx 2145 drivers/net/wireless/ath/ath10k/htt.h static inline void ath10k_htt_reset_paddrs_ring(struct ath10k_htt *htt, int idx) idx 2148 drivers/net/wireless/ath/ath10k/htt.h htt->rx_ops->htt_reset_paddrs_ring(htt, idx); idx 98 drivers/net/wireless/ath/ath10k/htt_rx.c dma_addr_t paddr, int idx) idx 100 drivers/net/wireless/ath/ath10k/htt_rx.c htt->rx_ring.paddrs_ring_32[idx] = __cpu_to_le32(paddr); idx 104 drivers/net/wireless/ath/ath10k/htt_rx.c dma_addr_t paddr, int idx) idx 106 drivers/net/wireless/ath/ath10k/htt_rx.c htt->rx_ring.paddrs_ring_64[idx] = __cpu_to_le64(paddr); idx 109 drivers/net/wireless/ath/ath10k/htt_rx.c static void ath10k_htt_reset_paddrs_ring_32(struct ath10k_htt *htt, int idx) idx 111 drivers/net/wireless/ath/ath10k/htt_rx.c htt->rx_ring.paddrs_ring_32[idx] = 0; idx 114 drivers/net/wireless/ath/ath10k/htt_rx.c static void ath10k_htt_reset_paddrs_ring_64(struct ath10k_htt *htt, int idx) idx 116 drivers/net/wireless/ath/ath10k/htt_rx.c htt->rx_ring.paddrs_ring_64[idx] = 0; idx 135 drivers/net/wireless/ath/ath10k/htt_rx.c int ret = 0, idx; idx 144 drivers/net/wireless/ath/ath10k/htt_rx.c idx = __le32_to_cpu(*htt->rx_ring.alloc_idx.vaddr); idx 173 drivers/net/wireless/ath/ath10k/htt_rx.c htt->rx_ring.netbufs_ring[idx] = skb; idx 174 drivers/net/wireless/ath/ath10k/htt_rx.c ath10k_htt_set_paddrs_ring(htt, paddr, idx); idx 184 drivers/net/wireless/ath/ath10k/htt_rx.c idx++; idx 185 drivers/net/wireless/ath/ath10k/htt_rx.c idx &= htt->rx_ring.size_mask; idx 194 drivers/net/wireless/ath/ath10k/htt_rx.c *htt->rx_ring.alloc_idx.vaddr = __cpu_to_le32(idx); idx 303 drivers/net/wireless/ath/ath10k/htt_rx.c int idx; idx 313 drivers/net/wireless/ath/ath10k/htt_rx.c idx = htt->rx_ring.sw_rd_idx.msdu_payld; idx 314 drivers/net/wireless/ath/ath10k/htt_rx.c msdu = htt->rx_ring.netbufs_ring[idx]; idx 315 drivers/net/wireless/ath/ath10k/htt_rx.c htt->rx_ring.netbufs_ring[idx] = NULL; idx 316 drivers/net/wireless/ath/ath10k/htt_rx.c ath10k_htt_reset_paddrs_ring(htt, idx); idx 318 drivers/net/wireless/ath/ath10k/htt_rx.c idx++; idx 319 drivers/net/wireless/ath/ath10k/htt_rx.c idx &= htt->rx_ring.size_mask; idx 320 drivers/net/wireless/ath/ath10k/htt_rx.c htt->rx_ring.sw_rd_idx.msdu_payld = idx; idx 3336 drivers/net/wireless/ath/ath10k/htt_rx.c int idx, ht_idx, gi, mcs, bw, nss; idx 3349 drivers/net/wireless/ath/ath10k/htt_rx.c idx = mcs * 8 + 8 * 10 * (nss - 1); idx 3350 drivers/net/wireless/ath/ath10k/htt_rx.c idx += bw * 2 + gi; idx 3399 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(AMPDU).rate_table[0][idx] += idx 3407 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(AMPDU).rate_table[1][idx] += idx 3439 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(SUCC).rate_table[0][idx] += pstats->succ_bytes; idx 3440 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(SUCC).rate_table[1][idx] += pstats->succ_pkts; idx 3441 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(FAIL).rate_table[0][idx] += pstats->failed_bytes; idx 3442 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(FAIL).rate_table[1][idx] += pstats->failed_pkts; idx 3443 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(RETRY).rate_table[0][idx] += pstats->retry_bytes; idx 3444 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(RETRY).rate_table[1][idx] += pstats->retry_pkts; idx 3514 drivers/net/wireless/ath/ath10k/htt_rx.c arsta->tx_info.status.rates[0].idx = rate_idx - 4; idx 3517 drivers/net/wireless/ath/ath10k/htt_rx.c arsta->tx_info.status.rates[0].idx = rate_idx; idx 3524 drivers/net/wireless/ath/ath10k/htt_rx.c arsta->tx_info.status.rates[0].idx = idx 45 drivers/net/wireless/ath/ath10k/htt_tx.c int idx; idx 68 drivers/net/wireless/ath/ath10k/htt_tx.c idx = peer_id / 32; idx 81 drivers/net/wireless/ath/ath10k/htt_tx.c ar->htt.tx_q_state.vaddr->map[tid][idx] &= ~bit; idx 82 drivers/net/wireless/ath/ath10k/htt_tx.c ar->htt.tx_q_state.vaddr->map[tid][idx] |= count ? bit : 0; idx 7158 drivers/net/wireless/ath/ath10k/mac.c static int ath10k_get_survey(struct ieee80211_hw *hw, int idx, idx 7163 drivers/net/wireless/ath/ath10k/mac.c struct survey_info *ar_survey = &ar->survey[idx]; idx 7169 drivers/net/wireless/ath/ath10k/mac.c if (sband && idx >= sband->n_channels) { idx 7170 drivers/net/wireless/ath/ath10k/mac.c idx -= sband->n_channels; idx 7177 drivers/net/wireless/ath/ath10k/mac.c if (!sband || idx >= sband->n_channels) { idx 7182 drivers/net/wireless/ath/ath10k/mac.c ath10k_mac_update_bss_chan_survey(ar, &sband->channels[idx]); idx 7188 drivers/net/wireless/ath/ath10k/mac.c survey->channel = &sband->channels[idx]; idx 98 drivers/net/wireless/ath/ath10k/txrx.c info->status.rates[0].idx = -1; idx 2562 drivers/net/wireless/ath/ath10k/wmi.c int band, ch, idx = 0; idx 2569 drivers/net/wireless/ath/ath10k/wmi.c for (ch = 0; ch < sband->n_channels; ch++, idx++) idx 2575 drivers/net/wireless/ath/ath10k/wmi.c return idx; idx 2628 drivers/net/wireless/ath/ath10k/wmi.c int idx; idx 2635 drivers/net/wireless/ath/ath10k/wmi.c idx = freq_to_idx(ar, params->freq); idx 2636 drivers/net/wireless/ath/ath10k/wmi.c if (idx >= ARRAY_SIZE(ar->survey)) { idx 2638 drivers/net/wireless/ath/ath10k/wmi.c params->freq, idx); idx 2642 drivers/net/wireless/ath/ath10k/wmi.c survey = &ar->survey[idx]; idx 2664 drivers/net/wireless/ath/ath10k/wmi.c int idx; idx 2666 drivers/net/wireless/ath/ath10k/wmi.c idx = freq_to_idx(ar, params->freq); idx 2667 drivers/net/wireless/ath/ath10k/wmi.c if (idx >= ARRAY_SIZE(ar->survey)) { idx 2669 drivers/net/wireless/ath/ath10k/wmi.c params->freq, idx); idx 2675 drivers/net/wireless/ath/ath10k/wmi.c survey = &ar->survey[idx]; idx 5250 drivers/net/wireless/ath/ath10k/wmi.c int idx = ar->wmi.num_mem_chunks; idx 5259 drivers/net/wireless/ath/ath10k/wmi.c ar->wmi.mem_chunks[idx].vaddr = vaddr; idx 5260 drivers/net/wireless/ath/ath10k/wmi.c ar->wmi.mem_chunks[idx].paddr = paddr; idx 5261 drivers/net/wireless/ath/ath10k/wmi.c ar->wmi.mem_chunks[idx].len = pool_size; idx 5262 drivers/net/wireless/ath/ath10k/wmi.c ar->wmi.mem_chunks[idx].req_id = req_id; idx 5673 drivers/net/wireless/ath/ath10k/wmi.c int idx; idx 5692 drivers/net/wireless/ath/ath10k/wmi.c idx = freq_to_idx(ar, freq); idx 5693 drivers/net/wireless/ath/ath10k/wmi.c if (idx >= ARRAY_SIZE(ar->survey)) { idx 5695 drivers/net/wireless/ath/ath10k/wmi.c freq, idx); idx 5699 drivers/net/wireless/ath/ath10k/wmi.c survey = &ar->survey[idx]; idx 698 drivers/net/wireless/ath/ath5k/base.c struct ath5k_buf *bf, int idx) idx 704 drivers/net/wireless/ath/ath5k/base.c if (bf->rates[idx].idx < 0) { idx 708 drivers/net/wireless/ath/ath5k/base.c return &hw->wiphy->bands[info->band]->bitrates[ bf->rates[idx].idx ]; idx 714 drivers/net/wireless/ath/ath5k/base.c struct ath5k_buf *bf, int idx) idx 720 drivers/net/wireless/ath/ath5k/base.c rate = ath5k_get_rate(hw, info, bf, idx); idx 724 drivers/net/wireless/ath/ath5k/base.c rc_flags = bf->rates[idx].flags; idx 1696 drivers/net/wireless/ath/ath5k/base.c info->status.rates[ts->ts_final_idx + 1].idx = -1; idx 720 drivers/net/wireless/ath/ath5k/eeprom.c u8 pier, point, idx; idx 741 drivers/net/wireless/ath/ath5k/eeprom.c for (idx = 0; idx < AR5K_EEPROM_N_PD_CURVES; idx++) { idx 743 drivers/net/wireless/ath/ath5k/eeprom.c if (!((ee->ee_x_gain[mode] >> idx) & 0x1)) { idx 744 drivers/net/wireless/ath/ath5k/eeprom.c pdgain_idx[0] = idx; idx 751 drivers/net/wireless/ath/ath5k/eeprom.c pd = &chinfo[pier].pd_curves[idx]; idx 927 drivers/net/wireless/ath/ath5k/eeprom.c u8 idx = pdgain_idx[pdg]; idx 929 drivers/net/wireless/ath/ath5k/eeprom.c &chinfo[pier].pd_curves[idx]; idx 1219 drivers/net/wireless/ath/ath5k/eeprom.c u8 idx = pdgain_idx[pdg]; idx 1221 drivers/net/wireless/ath/ath5k/eeprom.c &chinfo[pier].pd_curves[idx]; idx 1286 drivers/net/wireless/ath/ath5k/eeprom.c int idx, i; idx 1295 drivers/net/wireless/ath/ath5k/eeprom.c for (idx = AR5K_EEPROM_N_PD_CURVES - 1; idx >= 0; idx--) { idx 1297 drivers/net/wireless/ath/ath5k/eeprom.c if ((ee->ee_x_gain[mode] >> idx) & 0x1) idx 1298 drivers/net/wireless/ath/ath5k/eeprom.c pdgain_idx[pd_gains++] = idx; idx 647 drivers/net/wireless/ath/ath5k/mac80211-ops.c ath5k_get_survey(struct ieee80211_hw *hw, int idx, struct survey_info *survey) idx 655 drivers/net/wireless/ath/ath5k/mac80211-ops.c if (idx != 0) idx 2526 drivers/net/wireless/ath/ath5k/phy.c u8 idx[2] = { 0, 1 }; idx 2551 drivers/net/wireless/ath/ath5k/phy.c if ((pwr_i > pwr[idx[1]]) && (idx[1] < num_points - 1)) { idx 2552 drivers/net/wireless/ath/ath5k/phy.c idx[0]++; idx 2553 drivers/net/wireless/ath/ath5k/phy.c idx[1]++; idx 2557 drivers/net/wireless/ath/ath5k/phy.c pwr[idx[0]], pwr[idx[1]], idx 2558 drivers/net/wireless/ath/ath5k/phy.c vpd[idx[0]], vpd[idx[1]]); idx 3299 drivers/net/wireless/ath/ath5k/phy.c u8 idx = pdg_curve_to_idx[pdg]; idx 3302 drivers/net/wireless/ath/ath5k/phy.c pdg_L = &pcinfo_L->pd_curves[idx]; idx 3303 drivers/net/wireless/ath/ath5k/phy.c pdg_R = &pcinfo_R->pd_curves[idx]; idx 452 drivers/net/wireless/ath/ath6kl/init.c static int ath6kl_target_config_wlan_params(struct ath6kl *ar, int idx) idx 461 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_set_rx_frame_format_cmd(ar->wmi, idx, idx 469 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_pmparams_cmd(ar->wmi, idx, 0, 1, 0, 0, 1, idx 479 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_set_lpreamble_cmd(ar->wmi, idx, 0, idx 488 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_set_keepalive_cmd(ar->wmi, idx, idx 495 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_disctimeout_cmd(ar->wmi, idx, idx 503 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_set_wmm_txop(ar->wmi, idx, WMI_TXOP_DISABLED); idx 510 drivers/net/wireless/ath/ath6kl/init.c if (ar->p2p && (ar->vif_max == 1 || idx)) { idx 511 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_info_req_cmd(ar->wmi, idx, idx 523 drivers/net/wireless/ath/ath6kl/init.c if (ar->p2p && (ar->vif_max == 1 || idx)) { idx 525 drivers/net/wireless/ath/ath6kl/init.c ret = ath6kl_wmi_probe_report_req_cmd(ar->wmi, idx, true); idx 1052 drivers/net/wireless/ath/ath6kl/txrx.c u16 idx, idx_end, seq_end; idx 1059 drivers/net/wireless/ath/ath6kl/txrx.c idx = AGGR_WIN_IDX(rxtid->seq_next, rxtid->hold_q_sz); idx 1078 drivers/net/wireless/ath/ath6kl/txrx.c node = &rxtid->hold_q[idx]; idx 1094 drivers/net/wireless/ath/ath6kl/txrx.c idx = AGGR_WIN_IDX(rxtid->seq_next, rxtid->hold_q_sz); idx 1095 drivers/net/wireless/ath/ath6kl/txrx.c } while (idx != idx_end); idx 1113 drivers/net/wireless/ath/ath6kl/txrx.c u16 idx, st, cur, end; idx 1173 drivers/net/wireless/ath/ath6kl/txrx.c idx = AGGR_WIN_IDX(seq_no, rxtid->hold_q_sz); idx 1175 drivers/net/wireless/ath/ath6kl/txrx.c node = &rxtid->hold_q[idx]; idx 1212 drivers/net/wireless/ath/ath6kl/txrx.c for (idx = 0; idx < rxtid->hold_q_sz; idx++) { idx 1213 drivers/net/wireless/ath/ath6kl/txrx.c if (rxtid->hold_q[idx].skb) { idx 1459 drivers/net/wireless/ath/ath6kl/txrx.c u8 idx; idx 1470 drivers/net/wireless/ath/ath6kl/txrx.c idx = vif->fw_vif_idx; idx 1473 drivers/net/wireless/ath/ath6kl/txrx.c idx, idx 1613 drivers/net/wireless/ath/ath6kl/wmi.c int ath6kl_wmi_set_txe_notify(struct wmi *wmi, u8 idx, idx 1628 drivers/net/wireless/ath/ath6kl/wmi.c return ath6kl_wmi_cmd_send(wmi, idx, skb, WMI_SET_TXE_NOTIFY_CMDID, idx 2663 drivers/net/wireless/ath/ath6kl/wmi.h int ath6kl_wmi_set_txe_notify(struct wmi *wmi, u8 idx, idx 1274 drivers/net/wireless/ath/ath9k/ar5008_phy.c int i, idx = 0; idx 1277 drivers/net/wireless/ath/ath9k/ar5008_phy.c ah->tx_power[i] = rate_array[idx]; idx 1278 drivers/net/wireless/ath/ath9k/ar5008_phy.c idx++; idx 1100 drivers/net/wireless/ath/ath9k/ar9003_calib.c u32 idx = 2 * j, offset = 4 * (3 * im + j); idx 1108 drivers/net/wireless/ath/ath9k/ar9003_calib.c iq_res[idx] = REG_READ(ah, idx 1118 drivers/net/wireless/ath/ath9k/ar9003_calib.c iq_res[idx + 1] = 0xffff & REG_READ(ah, idx 1123 drivers/net/wireless/ath/ath9k/ar9003_calib.c idx, iq_res[idx], idx + 1, idx 1124 drivers/net/wireless/ath/ath9k/ar9003_calib.c iq_res[idx + 1]); idx 5141 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c int idx, idx 5149 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c return CTL_EDGE_TPOWER(ctl_2g[idx].ctlEdges[edge]); idx 5151 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c return CTL_EDGE_TPOWER(ctl_5g[idx].ctlEdges[edge]); idx 5155 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c int idx, idx 5164 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c &eep->ctl_freqbin_2G[idx][0] : idx 5165 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c &eep->ctl_freqbin_5G[idx][0]; idx 5169 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c CTL_EDGE_FLAGS(ctl_2g[idx].ctlEdges[edge - 1])) idx 5170 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c return CTL_EDGE_TPOWER(ctl_2g[idx].ctlEdges[edge - 1]); idx 5173 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c CTL_EDGE_FLAGS(ctl_5g[idx].ctlEdges[edge - 1])) idx 5174 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c return CTL_EDGE_TPOWER(ctl_5g[idx].ctlEdges[edge - 1]); idx 5184 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c u16 freq, int idx, bool is2GHz) idx 5188 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c &eep->ctl_freqbin_2G[idx][0] : idx 5189 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c &eep->ctl_freqbin_5G[idx][0]; idx 5204 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c ar9003_hw_get_direct_edge_power(eep, idx, idx 5211 drivers/net/wireless/ath/ath9k/ar9003_eeprom.c ar9003_hw_get_indirect_edge_power(eep, idx, idx 826 drivers/net/wireless/ath/ath9k/ar9003_hw.c int idx = ar9003_hw_get_tx_gain_idx(ah); idx 828 drivers/net/wireless/ath/ath9k/ar9003_hw.c if (idx >= ARRAY_SIZE(modes)) idx 829 drivers/net/wireless/ath/ath9k/ar9003_hw.c idx = 0; idx 831 drivers/net/wireless/ath/ath9k/ar9003_hw.c modes[idx](ah); idx 187 drivers/net/wireless/ath/ath9k/beacon.c ath9k_beacon_setup(sc, vif, bf, info->control.rates[0].idx); idx 329 drivers/net/wireless/ath/ath9k/channel.c int idx = ctx - &sc->chanctx[0]; idx 331 drivers/net/wireless/ath/ath9k/channel.c return &sc->chanctx[!idx]; idx 150 drivers/net/wireless/ath/ath9k/common-spectral.h u8 idx; idx 152 drivers/net/wireless/ath/ath9k/common-spectral.h idx = spectral_max_index(bins, SPECTRAL_HT20_40_NUM_BINS); idx 157 drivers/net/wireless/ath/ath9k/common-spectral.h return idx % (SPECTRAL_HT20_40_NUM_BINS / 2); idx 258 drivers/net/wireless/ath/ath9k/dynack.c rate = &common->sbands[info->band].bitrates[rates[ridx].idx]; idx 482 drivers/net/wireless/ath/ath9k/hif_usb.c static inline bool check_index(struct sk_buff *skb, u8 idx) idx 489 drivers/net/wireless/ath/ath9k/hif_usb.c (tx_ctl->sta_idx == idx)) idx 495 drivers/net/wireless/ath/ath9k/hif_usb.c static void hif_usb_sta_drain(void *hif_handle, u8 idx) idx 504 drivers/net/wireless/ath/ath9k/hif_usb.c if (check_index(skb, idx)) { idx 487 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c rate->idx = MS(txs->ts_rate, ATH9K_HTC_TXSTAT_RATE); idx 498 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c rate->idx += 4; /* No CCK rates */ idx 317 drivers/net/wireless/ath/ath9k/htc_hst.c void htc_sta_drain(struct htc_target *target, u8 idx) idx 319 drivers/net/wireless/ath/ath9k/htc_hst.c target->hif->sta_drain(target->hif_dev, idx); idx 38 drivers/net/wireless/ath/ath9k/htc_hst.h void (*sta_drain) (void *hif_handle, u8 idx); idx 214 drivers/net/wireless/ath/ath9k/htc_hst.h void htc_sta_drain(struct htc_target *target, u8 idx); idx 187 drivers/net/wireless/ath/ath9k/link.c tx_info->control.rates[0].idx = 0; idx 190 drivers/net/wireless/ath/ath9k/link.c tx_info->control.rates[1].idx = -1; idx 1955 drivers/net/wireless/ath/ath9k/main.c static int ath9k_get_survey(struct ieee80211_hw *hw, int idx, idx 1969 drivers/net/wireless/ath/ath9k/main.c if (idx == 0) idx 1973 drivers/net/wireless/ath/ath9k/main.c if (sband && idx >= sband->n_channels) { idx 1974 drivers/net/wireless/ath/ath9k/main.c idx -= sband->n_channels; idx 1981 drivers/net/wireless/ath/ath9k/main.c if (!sband || idx >= sband->n_channels) { idx 1986 drivers/net/wireless/ath/ath9k/main.c chan = &sband->channels[idx]; idx 742 drivers/net/wireless/ath/ath9k/xmit.c if (!rates[i].count || rates[i].idx < 0) idx 792 drivers/net/wireless/ath/ath9k/xmit.c frmlen = sc->tx.max_aggr_framelen[q][modeidx][rates[i].idx]; idx 868 drivers/net/wireless/ath/ath9k/xmit.c rix = bf->rates[0].idx; idx 1232 drivers/net/wireless/ath/ath9k/xmit.c if (!rates[i].count || (rates[i].idx < 0)) idx 1235 drivers/net/wireless/ath/ath9k/xmit.c rix = rates[i].idx; idx 1281 drivers/net/wireless/ath/ath9k/xmit.c rate = &common->sbands[tx_info->band].bitrates[rates[i].idx]; idx 1846 drivers/net/wireless/ath/ath9k/xmit.c int idx = txq->txq_tailidx; idx 1848 drivers/net/wireless/ath/ath9k/xmit.c while (!list_empty(&txq->txq_fifo[idx])) { idx 1849 drivers/net/wireless/ath/ath9k/xmit.c ath_drain_txq_list(sc, txq, &txq->txq_fifo[idx]); idx 1851 drivers/net/wireless/ath/ath9k/xmit.c INCR(idx, ATH_TXFIFO_DEPTH); idx 1853 drivers/net/wireless/ath/ath9k/xmit.c txq->txq_tailidx = idx; idx 2544 drivers/net/wireless/ath/ath9k/xmit.c tx_info->status.rates[i].idx = -1; idx 1644 drivers/net/wireless/ath/carl9170/main.c static int carl9170_op_get_survey(struct ieee80211_hw *hw, int idx, idx 1656 drivers/net/wireless/ath/carl9170/main.c if (idx == chan->hw_value) { idx 1671 drivers/net/wireless/ath/carl9170/main.c if (band->channels[i].hw_value == idx) { idx 1680 drivers/net/wireless/ath/carl9170/main.c memcpy(survey, &ar->survey[idx], sizeof(*survey)); idx 1057 drivers/net/wireless/ath/carl9170/phy.c int idx = nfreqs - 2; idx 1059 drivers/net/wireless/ath/carl9170/phy.c while (idx >= 0) { idx 1060 drivers/net/wireless/ath/carl9170/phy.c if (f >= freqs[idx]) idx 1061 drivers/net/wireless/ath/carl9170/phy.c return idx; idx 1062 drivers/net/wireless/ath/carl9170/phy.c idx--; idx 1123 drivers/net/wireless/ath/carl9170/phy.c int chain, idx, i; idx 1151 drivers/net/wireless/ath/carl9170/phy.c idx = carl9170_find_freq_idx(i, cal_freq_pier, f); idx 1163 drivers/net/wireless/ath/carl9170/phy.c cal_pier_data_2G[chain][idx]; idx 1168 drivers/net/wireless/ath/carl9170/phy.c cal_pier_data_5G[chain][idx]; idx 1177 drivers/net/wireless/ath/carl9170/phy.c cal_freq_pier[idx], idx 1179 drivers/net/wireless/ath/carl9170/phy.c cal_freq_pier[idx + 1], idx 1183 drivers/net/wireless/ath/carl9170/phy.c cal_freq_pier[idx], idx 1185 drivers/net/wireless/ath/carl9170/phy.c cal_freq_pier[idx + 1], idx 1435 drivers/net/wireless/ath/carl9170/phy.c int idx, i, n; idx 1476 drivers/net/wireless/ath/carl9170/phy.c idx = carl9170_find_freq_idx(ntargets, pwr_freqs, f); idx 1479 drivers/net/wireless/ath/carl9170/phy.c ctpl[idx + 0].freq, ctpl[idx + 0].power[n], idx 1480 drivers/net/wireless/ath/carl9170/phy.c ctpl[idx + 1].freq, ctpl[idx + 1].power[n]); idx 1516 drivers/net/wireless/ath/carl9170/phy.c idx = carl9170_find_freq_idx(ntargets, pwr_freqs, f); idx 1519 drivers/net/wireless/ath/carl9170/phy.c ctph[idx + 0].freq, ctph[idx + 0].power[n], idx 1520 drivers/net/wireless/ath/carl9170/phy.c ctph[idx + 1].freq, ctph[idx + 1].power[n]); idx 542 drivers/net/wireless/ath/carl9170/tx.c if (txinfo->status.rates[i].idx < 0) idx 553 drivers/net/wireless/ath/carl9170/tx.c txinfo->status.rates[i].idx = -1; idx 712 drivers/net/wireless/ath/carl9170/tx.c unsigned int idx; idx 714 drivers/net/wireless/ath/carl9170/tx.c idx = txrate->idx; idx 734 drivers/net/wireless/ath/carl9170/tx.c *phyrate = txrate->idx; idx 735 drivers/net/wireless/ath/carl9170/tx.c *tpc += txpower[idx & 7]; idx 738 drivers/net/wireless/ath/carl9170/tx.c if (idx < 4) idx 744 drivers/net/wireless/ath/carl9170/tx.c idx += 4; idx 747 drivers/net/wireless/ath/carl9170/tx.c rate = &__carl9170_ratetable[idx]; idx 785 drivers/net/wireless/ath/carl9170/tx.c SET_VAL(AR9170_TX_PHY_MCS, phyrate, txrate->idx); idx 788 drivers/net/wireless/ath/carl9170/tx.c tmp |= cpu_to_le32((txrate->idx & 0x7) << idx 801 drivers/net/wireless/ath/carl9170/tx.c if (txrate->idx <= AR9170_TX_PHY_RATE_CCK_11M) idx 917 drivers/net/wireless/ath/carl9170/tx.c if (txrate->idx < 0) idx 482 drivers/net/wireless/ath/key.c int idx; idx 514 drivers/net/wireless/ath/key.c idx = ath_reserve_key_cache_slot(common, key->cipher); idx 518 drivers/net/wireless/ath/key.c idx = key->keyidx; idx 524 drivers/net/wireless/ath/key.c idx = ath_reserve_key_cache_slot(common, key->cipher); idx 527 drivers/net/wireless/ath/key.c idx = key->keyidx; idx 538 drivers/net/wireless/ath/key.c idx = key->keyidx; idx 546 drivers/net/wireless/ath/key.c idx = ath_reserve_key_cache_slot(common, key->cipher); idx 549 drivers/net/wireless/ath/key.c if (idx < 0) idx 553 drivers/net/wireless/ath/key.c ret = ath_setkey_tkip(common, idx, key->key, &hk, mac, idx 556 drivers/net/wireless/ath/key.c ret = ath_hw_set_keycache_entry(common, idx, &hk, mac); idx 561 drivers/net/wireless/ath/key.c set_bit(idx, common->keymap); idx 563 drivers/net/wireless/ath/key.c set_bit(idx, common->ccmp_keymap); idx 566 drivers/net/wireless/ath/key.c set_bit(idx + 64, common->keymap); idx 567 drivers/net/wireless/ath/key.c set_bit(idx, common->tkip_keymap); idx 568 drivers/net/wireless/ath/key.c set_bit(idx + 64, common->tkip_keymap); idx 570 drivers/net/wireless/ath/key.c set_bit(idx + 32, common->keymap); idx 571 drivers/net/wireless/ath/key.c set_bit(idx + 64 + 32, common->keymap); idx 572 drivers/net/wireless/ath/key.c set_bit(idx + 32, common->tkip_keymap); idx 573 drivers/net/wireless/ath/key.c set_bit(idx + 64 + 32, common->tkip_keymap); idx 577 drivers/net/wireless/ath/key.c return idx; idx 547 drivers/net/wireless/ath/wil6210/cfg80211.c int wil_find_cid_by_idx(struct wil6210_priv *wil, u8 mid, int idx) idx 556 drivers/net/wireless/ath/wil6210/cfg80211.c if (idx == 0) idx 558 drivers/net/wireless/ath/wil6210/cfg80211.c idx--; idx 565 drivers/net/wireless/ath/wil6210/cfg80211.c struct net_device *dev, int idx, idx 571 drivers/net/wireless/ath/wil6210/cfg80211.c int cid = wil_find_cid_by_idx(wil, vif->mid, idx); idx 55 drivers/net/wireless/ath/wil6210/debugfs.c char _s, char _h, int idx) idx 63 drivers/net/wireless/ath/wil6210/debugfs.c &ring->va[idx].rx.enhanced; idx 73 drivers/net/wireless/ath/wil6210/debugfs.c &ring->va[idx].tx.enhanced; idx 76 drivers/net/wireless/ath/wil6210/debugfs.c has_skb = ring->ctx && ring->ctx[idx].skb; idx 335 drivers/net/wireless/ath/wil6210/pmc.c unsigned long long idx; idx 356 drivers/net/wireless/ath/wil6210/pmc.c idx = *f_pos; idx 357 drivers/net/wireless/ath/wil6210/pmc.c do_div(idx, pmc->descriptor_size); idx 358 drivers/net/wireless/ath/wil6210/pmc.c offset = *f_pos - (idx * pmc->descriptor_size); idx 370 drivers/net/wireless/ath/wil6210/pmc.c *f_pos, idx, offset, count); idx 376 drivers/net/wireless/ath/wil6210/pmc.c pmc->descriptors[idx].va, idx 1240 drivers/net/wireless/ath/wil6210/wil6210.h int wil_find_cid_by_idx(struct wil6210_priv *wil, u8 mid, int idx); idx 1836 drivers/net/wireless/atmel/atmel.c int idx, key_len, alg = ext->alg, set_key = 1; idx 1839 drivers/net/wireless/atmel/atmel.c idx = encoding->flags & IW_ENCODE_INDEX; idx 1840 drivers/net/wireless/atmel/atmel.c if (idx) { idx 1841 drivers/net/wireless/atmel/atmel.c if (idx < 1 || idx > 4) idx 1843 drivers/net/wireless/atmel/atmel.c idx--; idx 1845 drivers/net/wireless/atmel/atmel.c idx = priv->default_key; idx 1851 drivers/net/wireless/atmel/atmel.c priv->default_key = idx; idx 1865 drivers/net/wireless/atmel/atmel.c priv->wep_key_len[idx] = 13; idx 1869 drivers/net/wireless/atmel/atmel.c priv->wep_key_len[idx] = 5; idx 1876 drivers/net/wireless/atmel/atmel.c memset(priv->wep_keys[idx], 0, 13); idx 1877 drivers/net/wireless/atmel/atmel.c key_len = min ((int)ext->key_len, priv->wep_key_len[idx]); idx 1878 drivers/net/wireless/atmel/atmel.c memcpy(priv->wep_keys[idx], ext->key, key_len); idx 1896 drivers/net/wireless/atmel/atmel.c int idx, max_key_len; idx 1902 drivers/net/wireless/atmel/atmel.c idx = encoding->flags & IW_ENCODE_INDEX; idx 1903 drivers/net/wireless/atmel/atmel.c if (idx) { idx 1904 drivers/net/wireless/atmel/atmel.c if (idx < 1 || idx > 4) idx 1906 drivers/net/wireless/atmel/atmel.c idx--; idx 1908 drivers/net/wireless/atmel/atmel.c idx = priv->default_key; idx 1910 drivers/net/wireless/atmel/atmel.c encoding->flags = idx + 1; idx 1923 drivers/net/wireless/atmel/atmel.c ext->key_len = priv->wep_key_len[idx]; idx 1924 drivers/net/wireless/atmel/atmel.c memcpy(ext->key, priv->wep_keys[idx], ext->key_len); idx 354 drivers/net/wireless/broadcom/b43/debugfs.c int i, idx; idx 366 drivers/net/wireless/broadcom/b43/debugfs.c idx = 0; idx 377 drivers/net/wireless/broadcom/b43/debugfs.c idx, idx 383 drivers/net/wireless/broadcom/b43/debugfs.c idx++; idx 831 drivers/net/wireless/broadcom/b43/lo.c int idx; idx 855 drivers/net/wireless/broadcom/b43/lo.c idx = i / 2; idx 859 drivers/net/wireless/broadcom/b43/lo.c lo->dc_lt[idx] = (lo->dc_lt[idx] & 0x00FF) idx 863 drivers/net/wireless/broadcom/b43/lo.c lo->dc_lt[idx] = (lo->dc_lt[idx] & 0xFF00) idx 5153 drivers/net/wireless/broadcom/b43/main.c static int b43_op_get_survey(struct ieee80211_hw *hw, int idx, idx 5160 drivers/net/wireless/broadcom/b43/main.c if (idx != 0) idx 959 drivers/net/wireless/broadcom/b43/phy_lp.c static void lpphy_set_rx_gain_by_index(struct b43_wldev *dev, u16 idx) idx 961 drivers/net/wireless/broadcom/b43/phy_lp.c u32 gain = b43_lptab_read(dev, B43_LPTAB16(12, idx)); idx 3986 drivers/net/wireless/broadcom/b43/phy_n.c u8 idx, delta; idx 4011 drivers/net/wireless/broadcom/b43/phy_n.c idx = 68; idx 4014 drivers/net/wireless/broadcom/b43/phy_n.c idx = b43_is_40mhz(dev) ? 52 : 4; idx 4018 drivers/net/wireless/broadcom/b43/phy_n.c idx = b43_is_40mhz(dev) ? 76 : 28; idx 4021 drivers/net/wireless/broadcom/b43/phy_n.c idx = b43_is_40mhz(dev) ? 84 : 36; idx 4024 drivers/net/wireless/broadcom/b43/phy_n.c idx = b43_is_40mhz(dev) ? 92 : 44; idx 4030 drivers/net/wireless/broadcom/b43/phy_n.c nphy->tx_power_offset[idx]; idx 4032 drivers/net/wireless/broadcom/b43/phy_n.c idx += delta; idx 4034 drivers/net/wireless/broadcom/b43/phy_n.c idx += 1 - delta; idx 4037 drivers/net/wireless/broadcom/b43/phy_n.c idx += 1; idx 903 drivers/net/wireless/broadcom/b43/xmit.c report->status.rates[1].idx = -1; idx 130 drivers/net/wireless/broadcom/b43legacy/debugfs.c int i, idx; idx 143 drivers/net/wireless/broadcom/b43legacy/debugfs.c idx = 0; idx 154 drivers/net/wireless/broadcom/b43legacy/debugfs.c idx, idx 160 drivers/net/wireless/broadcom/b43legacy/debugfs.c idx++; idx 1196 drivers/net/wireless/broadcom/b43legacy/dma.c info->status.rates[1].idx = -1; idx 3515 drivers/net/wireless/broadcom/b43legacy/main.c static int b43legacy_op_get_survey(struct ieee80211_hw *hw, int idx, idx 3522 drivers/net/wireless/broadcom/b43legacy/main.c if (idx != 0) idx 1975 drivers/net/wireless/broadcom/b43legacy/phy.c u8 idx; idx 2005 drivers/net/wireless/broadcom/b43legacy/phy.c for (idx = 0; idx < 64; idx++) idx 2006 drivers/net/wireless/broadcom/b43legacy/phy.c if (b43legacy_tssi2dbm_entry(dyn_tssi2dbm, idx, pab0, idx 522 drivers/net/wireless/broadcom/b43legacy/pio.c info->status.rates[1].idx = -1; idx 59 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c #define BCDC_SET_IF_IDX(hdr, idx) \ idx 61 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c ((idx) << BCDC_FLAG2_IF_SHIFT))) idx 2723 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c int idx, u8 *mac, struct station_info *sinfo) idx 2730 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c brcmf_dbg(TRACE, "Enter, idx %d\n", idx); idx 2732 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c if (idx == 0) { idx 2744 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c if (idx < le32_to_cpu(cfg->assoclist.count)) { idx 2745 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c memcpy(mac, cfg->assoclist.mac[idx], ETH_ALEN); idx 492 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c int idx = 1; idx 496 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c idx++, core->pub.id, core->pub.rev, core->pub.base, idx 541 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c static bool brcmf_chip_socram_banksize(struct brcmf_core_priv *core, u8 idx, idx 547 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c bankidx |= idx; idx 622 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c u32 idx; idx 632 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c for (idx = 0; idx < nb; idx++) { idx 633 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c brcmf_chip_socram_banksize(sysmem, idx, &banksize); idx 649 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c u32 idx; idx 657 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c for (idx = 0; idx < totb; idx++) { idx 658 drivers/net/wireless/broadcom/brcm80211/brcmfmac/chip.c brcmf_chip_core_write32(cr4, ARMCR4_BANKIDX, idx); idx 917 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c int idx, i, ret; idx 922 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c for (idx = 0; idx < BRCMF_MAX_IFS; idx++) { idx 923 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c ifp = drvr->iflist[idx]; idx 926 drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c if (idx == BRCMF_MAX_IFS - 1) idx 320 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c dma_addr_t *physaddr, u32 *idx) idx 335 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c *idx = pktids->last_allocated_idx; idx 339 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c (*idx)++; idx 340 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c if (*idx == pktids->array_size) idx 341 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c *idx = 0; idx 342 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c if (array[*idx].allocated.counter == 0) idx 343 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c if (atomic_cmpxchg(&array[*idx].allocated, 0, 1) == 0) idx 351 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c array[*idx].data_offset = data_offset; idx 352 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c array[*idx].physaddr = *physaddr; idx 353 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c array[*idx].skb = skb; idx 355 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c pktids->last_allocated_idx = *idx; idx 363 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c u32 idx) idx 368 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c if (idx < 0 || idx >= pktids->array_size) { idx 369 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c brcmf_err("Invalid packet id %d (max %d)\n", idx, idx 373 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c if (pktids->array[idx].allocated.counter) { idx 374 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c pktid = &pktids->array[idx]; idx 382 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c brcmf_err("Invalid packet id %d (not in use)\n", idx); idx 872 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c u32 idx; idx 877 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c idx = le32_to_cpu(tx_status->msg.request_id) - 1; idx 881 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c msgbuf->tx_pktids, idx); idx 1098 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c u32 idx; idx 1104 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c idx = le32_to_cpu(event->msg.request_id); idx 1112 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c msgbuf->rx_pktids, idx); idx 1146 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c u32 idx; idx 1154 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c idx = le32_to_cpu(rx_complete->msg.request_id); idx 1158 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c msgbuf->rx_pktids, idx); idx 58 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c __le32 idx; idx 2814 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c u32 n, idx, addr; idx 2836 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c idx = le32_to_cpu(c->log_le.idx); idx 2839 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (idx > c->bufsize) idx 2844 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (idx == c->last) idx 2853 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c while (c->last != idx) { idx 2855 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c if (c->last == idx) { idx 2957 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c addr = sh->console_addr + offsetof(struct rte_console, log_le.idx); idx 831 drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c txrate[i].idx = -1; idx 182 drivers/net/wireless/broadcom/brcm80211/brcmsmac/antsel.c u8 idx = BRCMS_ANTIDX_11N(BRCMS_ANTSEL_11N(ant_cfg)); idx 187 drivers/net/wireless/broadcom/brcm80211/brcmsmac/antsel.c mimo_antsel = (mimo_2x4_div_antselpat_tbl[idx] & 0xf); idx 192 drivers/net/wireless/broadcom/brcm80211/brcmsmac/antsel.c mimo_antsel = (mimo_2x3_div_antselpat_tbl[idx] & 0xf); idx 78 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c __le32 idx; idx 1594 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c int brcms_ucode_init_buf(struct brcms_info *wl, void **pbuf, u32 idx) idx 1604 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c if (le32_to_cpu(hdr->idx) == idx) { idx 1616 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c "ERROR: ucode buf tag:%d can not be found!\n", idx); idx 1626 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c int brcms_ucode_init_uint(struct brcms_info *wl, size_t *n_bytes, u32 idx) idx 1635 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c if (le32_to_cpu(hdr->idx) == idx) { idx 1649 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c "ERROR: ucode tag:%d can not be found!\n", idx); idx 706 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c u8 idx; idx 712 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c for (idx = 0; idx < MHFMAX; idx++) idx 713 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_b_write_shm(wlc_hw, addr[idx], mhfs[idx]); idx 961 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if ((tx_frame_count > fbl) && (txrate[1].idx >= 0)) { idx 978 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c txrate[1].idx = -1; idx 984 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c txrate[i].idx = -1; idx 1329 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_b_mhf(struct brcms_hardware *wlc_hw, u8 idx, u16 mask, u16 val, idx 1339 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if ((val & ~mask) || idx >= MHFMAX) idx 1361 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c save = band->mhfs[idx]; idx 1362 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c band->mhfs[idx] = (band->mhfs[idx] & ~mask) | val; idx 1367 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if (wlc_hw->clk && (band->mhfs[idx] != save) idx 1369 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_b_write_shm(wlc_hw, addr[idx], idx 1370 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c (u16) band->mhfs[idx]); idx 1374 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c wlc_hw->bandstate[0]->mhfs[idx] = idx 1375 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c (wlc_hw->bandstate[0]->mhfs[idx] & ~mask) | val; idx 1376 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c wlc_hw->bandstate[1]->mhfs[idx] = idx 1377 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c (wlc_hw->bandstate[1]->mhfs[idx] & ~mask) | val; idx 2340 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c uint intstatus, idx; idx 2345 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c for (idx = 0; idx < NFIFO; idx++) { idx 2349 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c D11REGOFFS(intctrlregs[idx].intstatus)) & idx 2355 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c unit, idx, intstatus); idx 2359 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c "overflow\n", unit, idx); idx 2365 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c unit, idx); idx 2371 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c idx); idx 2377 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c "error\n", unit, idx); idx 2383 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c "underflow\n", idx, unit); idx 2387 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c "underflow\n", idx, unit); idx 2396 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c D11REGOFFS(intctrlregs[idx].intstatus), idx 3991 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c void brcms_c_protection_upd(struct brcms_c_info *wlc, uint idx, int val) idx 3997 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c BCMMSG(wlc->wiphy, "idx %d, val %d\n", idx, val); idx 3999 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c switch (idx) { idx 6298 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if (txrate[1]->idx < 0) idx 6304 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c if ((txrate[k]->idx >= 0) idx 6305 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c && (txrate[k]->idx < idx 6309 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c bitrates[txrate[k]->idx].hw_value; idx 6319 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c NRATE_MCS_INUSE | txrate[k]->idx); idx 643 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.h void brcms_b_mhf(struct brcms_hardware *wlc_hw, u8 idx, u16 mask, u16 val, idx 862 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c uint idx; idx 872 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c for (idx = 0; idx < ptbl_info->tbl_len; idx++) { idx 880 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c (tbl_id << 10) | (tbl_offset + idx)); idx 885 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c (u16) (ptbl_32b[idx] >> 16)); idx 886 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c write_phy_reg(pi, tblDataLo, (u16) ptbl_32b[idx]); idx 888 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c write_phy_reg(pi, tblDataLo, ptbl_16b[idx]); idx 890 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c write_phy_reg(pi, tblDataLo, ptbl_8b[idx]); idx 899 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c uint idx; idx 909 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c for (idx = 0; idx < ptbl_info->tbl_len; idx++) { idx 916 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c (tbl_id << 10) | (tbl_offset + idx)); idx 920 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c ptbl_32b[idx] = read_phy_reg(pi, tblDataLo); idx 921 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c ptbl_32b[idx] |= (read_phy_reg(pi, tblDataHi) << 16); idx 923 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c ptbl_16b[idx] = read_phy_reg(pi, tblDataLo); idx 925 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c ptbl_8b[idx] = (u8) read_phy_reg(pi, tblDataLo); idx 2238 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c u8 idx, core; idx 2243 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c for (idx = 0, core = 0; core < pi->pubpi.phy_corenum; idx += 2, idx 2245 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c lo = wlapi_bmac_read_shm(pi->sh->physhim, M_PWRIND_MAP(idx)); idx 2247 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_cmn.c M_PWRIND_MAP(idx + 1)); idx 919 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c #define wlc_lcnphy_set_start_tx_pwr_idx(pi, idx) \ idx 922 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c (u16)(idx) << 0) idx 3902 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c uint idx, SAVE_txpwrindex = 0xFF; idx 3976 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c for (idx = 0; idx < 128; idx++) { idx 3977 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c tab.tbl_offset = LCNPHY_TX_PWR_CTRL_IQ_OFFSET + idx; idx 3985 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c tab.tbl_offset = LCNPHY_TX_PWR_CTRL_LO_OFFSET + idx; idx 4609 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c uint idx; idx 4614 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c for (idx = 0; idx < dot11lcnphytbl_info_sz_rev0; idx++) idx 4615 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c wlc_lcnphy_write_table(pi, &dot11lcnphytbl_info_rev0[idx]); idx 4670 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c for (idx = 0; idx < l; idx++) idx 4671 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c wlc_lcnphy_write_table(pi, &tb[idx]); idx 14180 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c uint idx; idx 14183 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < mimophytbl_info_sz_rev16; idx++) idx 14185 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c &mimophytbl_info_rev16[idx]); idx 14187 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < mimophytbl_info_sz_rev7; idx++) idx 14189 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c &mimophytbl_info_rev7[idx]); idx 14191 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < mimophytbl_info_sz_rev3; idx++) idx 14193 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c &mimophytbl_info_rev3[idx]); idx 14195 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < mimophytbl_info_sz_rev0; idx++) idx 14197 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c &mimophytbl_info_rev0[idx]); idx 14203 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c uint idx = 0; idx 14278 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < mimophytbl_info_sz_rev3_volatile; idx++) { idx 14280 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c if (idx == ANT_SWCTRL_TBL_REV3_IDX) { idx 14290 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c [idx]); idx 14296 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c [idx]); idx 14302 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c [idx]); idx 14308 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c [idx]); idx 14316 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c &mimophytbl_info_rev3_volatile[idx]); idx 14320 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < mimophytbl_info_sz_rev0_volatile; idx++) idx 14323 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c [idx]); idx 17475 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c u8 idx, idx2, i, delta_ind; idx 17477 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = TXP_FIRST_CCK; idx <= TXP_LAST_CCK; idx++) idx 17478 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->adj_pwr_tbl_nphy[idx] = pi->tx_power_offset[idx]; idx 17490 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = TXP_FIRST_MCS_40_SISO; idx 17492 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = (CHSPEC_IS40(pi->radio_chanspec)) ? idx 17500 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = (CHSPEC_IS40(pi->radio_chanspec)) ? idx 17506 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = (CHSPEC_IS40(pi->radio_chanspec)) ? idx 17512 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = (CHSPEC_IS40(pi->radio_chanspec)) ? idx 17518 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17519 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = idx + delta_ind; idx 17521 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17523 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17525 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx++]; idx 17528 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx++]; idx 17530 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17532 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17534 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx++]; idx 17537 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx++]; idx 17539 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17541 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17543 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx++]; idx 17546 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17548 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx++]; idx 17550 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17551 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = idx + 1 - delta_ind; idx 17553 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17556 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17558 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17560 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17562 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pi->tx_power_offset[idx]; idx 17568 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c u32 idx; idx 17747 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < tbl_len; idx++) { idx 17749 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c (16 * b0[tbl_id - 26] + b1[tbl_id - 26] * idx); idx 17750 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c den = 32768 + a1[tbl_id - 26] * idx; idx 17753 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c if (idx <= idx 17760 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c regval[idx] = (u32) pwr_est; idx 18795 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c u32 idx; idx 18819 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < tbl_len; idx++) idx 18820 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c regval[idx] = iqcomp; idx 18833 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < tbl_len; idx++) { idx 18839 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c nphy_tpc_loscale[idx] + idx 18843 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c nphy_tpc_loscale[idx] + idx 18848 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c regval[idx] = curr_locomp; idx 19371 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c u16 idx; idx 19421 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < 128; idx++) { idx 19422 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pga_gn = (tx_pwrctrl_tbl[idx] >> 24) & 0xf; idx 19423 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pad_gn = (tx_pwrctrl_tbl[idx] >> 19) & 0x1f; idx 19429 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c 1, 576 + idx, 32, idx 19434 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c 1, 576 + idx, 32, idx 19439 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c for (idx = 0; idx < 128; idx++) { idx 19440 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c pga_gn = (tx_pwrctrl_tbl[idx] >> 24) & 0xf; idx 19453 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c 1, 576 + idx, 32, idx 19458 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c 1, 576 + idx, 32, idx 23406 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c int idx; idx 23439 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = -1; idx 23443 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_n.c idx = k; idx 110 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c wlapi_bmac_mhf(struct phy_shim_info *physhim, u8 idx, u16 mask, idx 113 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.c brcms_b_mhf(physhim->wlc_hw, idx, mask, val, bands); idx 145 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy_shim.h void wlapi_bmac_mhf(struct phy_shim_info *physhim, u8 idx, u16 mask, u16 val, idx 289 drivers/net/wireless/broadcom/brcm80211/brcmsmac/pub.h void brcms_c_protection_upd(struct brcms_c_info *wlc, uint idx, int val); idx 274 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c uint idx; idx 283 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c for (idx = 0; idx < rs->count; idx++) { idx 284 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c if (rs->rates[idx] & BRCMS_RATE_FLAG) idx 50 drivers/net/wireless/broadcom/brcm80211/brcmsmac/ucode_loader.h int brcms_ucode_init_buf(struct brcms_info *wl, void **pbuf, unsigned int idx); idx 52 drivers/net/wireless/broadcom/brcm80211/brcmsmac/ucode_loader.h unsigned int idx); idx 6472 drivers/net/wireless/cisco/airo.c int idx, key_len, alg = ext->alg, set_key = 1, rc; idx 6481 drivers/net/wireless/cisco/airo.c idx = encoding->flags & IW_ENCODE_INDEX; idx 6482 drivers/net/wireless/cisco/airo.c if (idx) { idx 6483 drivers/net/wireless/cisco/airo.c if (!valid_index(local, idx - 1)) idx 6485 drivers/net/wireless/cisco/airo.c idx--; idx 6487 drivers/net/wireless/cisco/airo.c idx = get_wep_tx_idx(local); idx 6488 drivers/net/wireless/cisco/airo.c if (idx < 0) idx 6489 drivers/net/wireless/cisco/airo.c idx = 0; idx 6499 drivers/net/wireless/cisco/airo.c rc = set_wep_tx_idx(local, idx, perm, 1); idx 6503 drivers/net/wireless/cisco/airo.c idx, rc); idx 6531 drivers/net/wireless/cisco/airo.c rc = set_wep_tx_idx(local, idx, perm, 1); idx 6535 drivers/net/wireless/cisco/airo.c idx, rc); idx 6539 drivers/net/wireless/cisco/airo.c rc = set_wep_key(local, idx, key.key, key.len, perm, 1); idx 6543 drivers/net/wireless/cisco/airo.c idx, rc); idx 6576 drivers/net/wireless/cisco/airo.c int idx, max_key_len, wep_key_len; idx 6588 drivers/net/wireless/cisco/airo.c idx = encoding->flags & IW_ENCODE_INDEX; idx 6589 drivers/net/wireless/cisco/airo.c if (idx) { idx 6590 drivers/net/wireless/cisco/airo.c if (!valid_index(local, idx - 1)) idx 6592 drivers/net/wireless/cisco/airo.c idx--; idx 6594 drivers/net/wireless/cisco/airo.c idx = get_wep_tx_idx(local); idx 6595 drivers/net/wireless/cisco/airo.c if (idx < 0) idx 6596 drivers/net/wireless/cisco/airo.c idx = 0; idx 6599 drivers/net/wireless/cisco/airo.c encoding->flags = idx + 1; idx 6620 drivers/net/wireless/cisco/airo.c wep_key_len = get_wep_key(local, idx, &buf[0], sizeof(buf)); idx 5350 drivers/net/wireless/intel/ipw2x00/ipw2100.c u8 idx; idx 5377 drivers/net/wireless/intel/ipw2x00/ipw2100.c int idx, char *key, int len, int batch_mode) idx 5389 drivers/net/wireless/intel/ipw2x00/ipw2100.c idx, keylen, len); idx 5395 drivers/net/wireless/intel/ipw2x00/ipw2100.c wep_key->idx = idx; idx 5406 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->net_dev->name, wep_key->idx); idx 5409 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->net_dev->name, wep_key->idx, wep_key->len, idx 5414 drivers/net/wireless/intel/ipw2x00/ipw2100.c priv->net_dev->name, wep_key->idx, wep_key->len, idx 5440 drivers/net/wireless/intel/ipw2x00/ipw2100.c int idx, int batch_mode) idx 5446 drivers/net/wireless/intel/ipw2x00/ipw2100.c .host_command_parameters = {idx}, idx 5450 drivers/net/wireless/intel/ipw2x00/ipw2100.c IPW_DEBUG_HC("WEP_KEY_INDEX: index = %d\n", idx); idx 5452 drivers/net/wireless/intel/ipw2x00/ipw2100.c if (idx < 0 || idx > 3) idx 509 drivers/net/wireless/intel/ipw2x00/libipw_wx.c int i, idx, ret = 0; idx 519 drivers/net/wireless/intel/ipw2x00/libipw_wx.c idx = encoding->flags & IW_ENCODE_INDEX; idx 520 drivers/net/wireless/intel/ipw2x00/libipw_wx.c if (idx) { idx 521 drivers/net/wireless/intel/ipw2x00/libipw_wx.c if (idx < 1 || idx > WEP_KEYS) idx 523 drivers/net/wireless/intel/ipw2x00/libipw_wx.c idx--; idx 525 drivers/net/wireless/intel/ipw2x00/libipw_wx.c idx = ieee->crypt_info.tx_keyidx; idx 528 drivers/net/wireless/intel/ipw2x00/libipw_wx.c crypt = &ieee->crypt_info.crypt[idx]; idx 532 drivers/net/wireless/intel/ipw2x00/libipw_wx.c if (idx != 0 && ext->alg != IW_ENCODE_ALG_WEP) idx 535 drivers/net/wireless/intel/ipw2x00/libipw_wx.c crypt = &ieee->crypt_info.crypt[idx]; idx 611 drivers/net/wireless/intel/ipw2x00/libipw_wx.c new_crypt->priv = new_crypt->ops->init(idx); idx 630 drivers/net/wireless/intel/ipw2x00/libipw_wx.c ieee->crypt_info.tx_keyidx = idx; idx 631 drivers/net/wireless/intel/ipw2x00/libipw_wx.c sec.active_key = idx; idx 636 drivers/net/wireless/intel/ipw2x00/libipw_wx.c memcpy(sec.keys[idx], ext->key, ext->key_len); idx 637 drivers/net/wireless/intel/ipw2x00/libipw_wx.c sec.key_sizes[idx] = ext->key_len; idx 638 drivers/net/wireless/intel/ipw2x00/libipw_wx.c sec.flags |= (1 << idx); idx 640 drivers/net/wireless/intel/ipw2x00/libipw_wx.c sec.encode_alg[idx] = SEC_ALG_WEP; idx 644 drivers/net/wireless/intel/ipw2x00/libipw_wx.c sec.encode_alg[idx] = SEC_ALG_TKIP; idx 648 drivers/net/wireless/intel/ipw2x00/libipw_wx.c sec.encode_alg[idx] = SEC_ALG_CCMP; idx 670 drivers/net/wireless/intel/ipw2x00/libipw_wx.c int idx, max_key_len; idx 676 drivers/net/wireless/intel/ipw2x00/libipw_wx.c idx = encoding->flags & IW_ENCODE_INDEX; idx 677 drivers/net/wireless/intel/ipw2x00/libipw_wx.c if (idx) { idx 678 drivers/net/wireless/intel/ipw2x00/libipw_wx.c if (idx < 1 || idx > WEP_KEYS) idx 680 drivers/net/wireless/intel/ipw2x00/libipw_wx.c idx--; idx 682 drivers/net/wireless/intel/ipw2x00/libipw_wx.c idx = ieee->crypt_info.tx_keyidx; idx 686 drivers/net/wireless/intel/ipw2x00/libipw_wx.c if (idx != 0 || ieee->iw_mode != IW_MODE_INFRA) idx 689 drivers/net/wireless/intel/ipw2x00/libipw_wx.c encoding->flags = idx + 1; idx 697 drivers/net/wireless/intel/ipw2x00/libipw_wx.c if (sec->encode_alg[idx] == SEC_ALG_WEP) idx 699 drivers/net/wireless/intel/ipw2x00/libipw_wx.c else if (sec->encode_alg[idx] == SEC_ALG_TKIP) idx 701 drivers/net/wireless/intel/ipw2x00/libipw_wx.c else if (sec->encode_alg[idx] == SEC_ALG_CCMP) idx 706 drivers/net/wireless/intel/ipw2x00/libipw_wx.c ext->key_len = sec->key_sizes[idx]; idx 707 drivers/net/wireless/intel/ipw2x00/libipw_wx.c memcpy(ext->key, sec->keys[idx], ext->key_len); idx 461 drivers/net/wireless/intel/iwlegacy/3945-mac.c u16 len, idx, hdr_len; idx 521 drivers/net/wireless/intel/iwlegacy/3945-mac.c idx = il_get_cmd_idx(q, q->write_ptr, 0); idx 526 drivers/net/wireless/intel/iwlegacy/3945-mac.c out_cmd = txq->cmd[idx]; idx 527 drivers/net/wireless/intel/iwlegacy/3945-mac.c out_meta = &txq->meta[idx]; idx 1825 drivers/net/wireless/intel/iwlegacy/3945-mac.c int ret = -EINVAL, idx; idx 1838 drivers/net/wireless/intel/iwlegacy/3945-mac.c for (idx = api_max; idx >= api_min; idx--) { idx 1839 drivers/net/wireless/intel/iwlegacy/3945-mac.c sprintf(buf, "%s%u%s", name_pre, idx, ".ucode"); idx 1848 drivers/net/wireless/intel/iwlegacy/3945-mac.c if (idx < api_max) idx 46 drivers/net/wireless/intel/iwlegacy/3945-rs.c u8 idx; idx 87 drivers/net/wireless/intel/iwlegacy/3945-rs.c u32 idx = 0; idx 108 drivers/net/wireless/intel/iwlegacy/3945-rs.c while (idx < table_size && rssi < tpt_table[idx].min_rssi) idx 109 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx++; idx 111 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx = min(idx, table_size - 1); idx 113 drivers/net/wireless/intel/iwlegacy/3945-rs.c return tpt_table[idx].idx; idx 242 drivers/net/wireless/intel/iwlegacy/3945-rs.c int retries, int idx) idx 304 drivers/net/wireless/intel/iwlegacy/3945-rs.c ((win->success_ratio * rs_sta->expected_tpt[idx] + idx 444 drivers/net/wireless/intel/iwlegacy/3945-rs.c first_idx = sband->bitrates[info->status.rates[0].idx].hw_value; idx 524 drivers/net/wireless/intel/iwlegacy/3945-rs.c il3945_get_adjacent_rate(struct il3945_rs_sta *rs_sta, u8 idx, u16 rate_mask, idx 538 drivers/net/wireless/intel/iwlegacy/3945-rs.c i = idx - 1; idx 547 drivers/net/wireless/intel/iwlegacy/3945-rs.c i = idx + 1; idx 558 drivers/net/wireless/intel/iwlegacy/3945-rs.c low = idx; idx 571 drivers/net/wireless/intel/iwlegacy/3945-rs.c high = idx; idx 612 drivers/net/wireless/intel/iwlegacy/3945-rs.c int idx; idx 643 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx = min(rs_sta->last_txrate_idx & 0xffff, RATE_COUNT_3945 - 1); idx 654 drivers/net/wireless/intel/iwlegacy/3945-rs.c if (rs_sta->start_rate < idx && idx 656 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx = rs_sta->start_rate; idx 661 drivers/net/wireless/intel/iwlegacy/3945-rs.c if (max_rate_idx != -1 && max_rate_idx < idx) { idx 663 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx = max_rate_idx; idx 666 drivers/net/wireless/intel/iwlegacy/3945-rs.c win = &(rs_sta->win[idx]); idx 676 drivers/net/wireless/intel/iwlegacy/3945-rs.c "expected_tpt is %sNULL\n", idx, win->counter, idx 689 drivers/net/wireless/intel/iwlegacy/3945-rs.c il3945_get_adjacent_rate(rs_sta, idx, rate_mask, sband->band); idx 768 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx = low; idx 773 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx = high; idx 782 drivers/net/wireless/intel/iwlegacy/3945-rs.c D_RATE("Selected %d (action %d) - low %d high %d\n", idx, scale_action, idx 788 drivers/net/wireless/intel/iwlegacy/3945-rs.c if (WARN_ON_ONCE(idx < IL_FIRST_OFDM_RATE)) idx 789 drivers/net/wireless/intel/iwlegacy/3945-rs.c idx = IL_FIRST_OFDM_RATE; idx 790 drivers/net/wireless/intel/iwlegacy/3945-rs.c rs_sta->last_txrate_idx = idx; idx 791 drivers/net/wireless/intel/iwlegacy/3945-rs.c info->control.rates[0].idx = idx - IL_FIRST_OFDM_RATE; idx 793 drivers/net/wireless/intel/iwlegacy/3945-rs.c rs_sta->last_txrate_idx = idx; idx 794 drivers/net/wireless/intel/iwlegacy/3945-rs.c info->control.rates[0].idx = rs_sta->last_txrate_idx; idx 798 drivers/net/wireless/intel/iwlegacy/3945-rs.c D_RATE("leave: %d\n", idx); idx 187 drivers/net/wireless/intel/iwlegacy/3945.c int idx; idx 189 drivers/net/wireless/intel/iwlegacy/3945.c for (idx = 0; idx < RATE_COUNT_3945; idx++) idx 190 drivers/net/wireless/intel/iwlegacy/3945.c if (il3945_rates[idx].plcp == plcp) idx 191 drivers/net/wireless/intel/iwlegacy/3945.c return idx; idx 272 drivers/net/wireless/intel/iwlegacy/3945.c il3945_tx_queue_reclaim(struct il_priv *il, int txq_id, int idx) idx 280 drivers/net/wireless/intel/iwlegacy/3945.c for (idx = il_queue_inc_wrap(idx, q->n_bd); q->read_ptr != idx; idx 303 drivers/net/wireless/intel/iwlegacy/3945.c int idx = SEQ_TO_IDX(sequence); idx 311 drivers/net/wireless/intel/iwlegacy/3945.c if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) { idx 313 drivers/net/wireless/intel/iwlegacy/3945.c "is out of range [0-%d] %d %d\n", txq_id, idx, idx 342 drivers/net/wireless/intel/iwlegacy/3945.c info->status.rates[0].idx = rate_idx; idx 354 drivers/net/wireless/intel/iwlegacy/3945.c D_TX_REPLY("Tx queue reclaim %d\n", idx); idx 355 drivers/net/wireless/intel/iwlegacy/3945.c il3945_tx_queue_reclaim(il, txq_id, idx); idx 639 drivers/net/wireless/intel/iwlegacy/3945.c int idx = txq->q.read_ptr; idx 640 drivers/net/wireless/intel/iwlegacy/3945.c struct il3945_tfd *tfd = &tfd_tmp[idx]; idx 655 drivers/net/wireless/intel/iwlegacy/3945.c pci_unmap_single(dev, dma_unmap_addr(&txq->meta[idx], mapping), idx 656 drivers/net/wireless/intel/iwlegacy/3945.c dma_unmap_len(&txq->meta[idx], len), idx 1306 drivers/net/wireless/intel/iwlegacy/3945.c il3945_hw_reg_fix_power_idx(int idx) idx 1308 drivers/net/wireless/intel/iwlegacy/3945.c if (idx < 0) idx 1310 drivers/net/wireless/intel/iwlegacy/3945.c if (idx >= IL_MAX_GAIN_ENTRIES) idx 1312 drivers/net/wireless/intel/iwlegacy/3945.c return (u8) idx; idx 2314 drivers/net/wireless/intel/iwlegacy/3945.c int rc, i, idx, prev_idx; idx 2321 drivers/net/wireless/intel/iwlegacy/3945.c idx = il3945_rates[i].table_rs_idx; idx 2323 drivers/net/wireless/intel/iwlegacy/3945.c table[idx].rate_n_flags = cpu_to_le16(il3945_rates[i].plcp); idx 2324 drivers/net/wireless/intel/iwlegacy/3945.c table[idx].try_cnt = il->retry_rate; idx 2326 drivers/net/wireless/intel/iwlegacy/3945.c table[idx].next_rate_idx = il3945_rates[prev_idx].table_rs_idx; idx 2354 drivers/net/wireless/intel/iwlegacy/3945.c idx = IL_FIRST_CCK_RATE; idx 2357 drivers/net/wireless/intel/iwlegacy/3945.c il3945_rates[idx].table_rs_idx; idx 2359 drivers/net/wireless/intel/iwlegacy/3945.c idx = RATE_11M_IDX_TBL; idx 2361 drivers/net/wireless/intel/iwlegacy/3945.c table[idx].next_rate_idx = RATE_5M_IDX_TBL; idx 446 drivers/net/wireless/intel/iwlegacy/4965-mac.c int idx = 0; idx 451 drivers/net/wireless/intel/iwlegacy/4965-mac.c idx = (rate_n_flags & 0xff); idx 452 drivers/net/wireless/intel/iwlegacy/4965-mac.c return idx; idx 457 drivers/net/wireless/intel/iwlegacy/4965-mac.c for (idx = band_offset; idx < RATE_COUNT_LEGACY; idx++) idx 458 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (il_rates[idx].plcp == (rate_n_flags & 0xFF)) idx 459 drivers/net/wireless/intel/iwlegacy/4965-mac.c return idx - band_offset; idx 1571 drivers/net/wireless/intel/iwlegacy/4965-mac.c rate_idx = info->control.rates[0].idx; idx 2455 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tx_queue_reclaim(struct il_priv *il, int txq_id, int idx) idx 2463 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (idx >= q->n_bd || il_queue_used(q, idx) == 0) { idx 2465 drivers/net/wireless/intel/iwlegacy/4965-mac.c "is out of range [0-%d] %d %d.\n", txq_id, idx, q->n_bd, idx 2470 drivers/net/wireless/intel/iwlegacy/4965-mac.c for (idx = il_queue_inc_wrap(idx, q->n_bd); q->read_ptr != idx; idx 2656 drivers/net/wireless/intel/iwlegacy/4965-mac.c int i, sh, idx; idx 2670 drivers/net/wireless/intel/iwlegacy/4965-mac.c idx = start_idx; idx 2673 drivers/net/wireless/intel/iwlegacy/4965-mac.c agg->frame_count, agg->start_idx, idx); idx 2675 drivers/net/wireless/intel/iwlegacy/4965-mac.c info = IEEE80211_SKB_CB(il->txq[txq_id].skbs[idx]); idx 2697 drivers/net/wireless/intel/iwlegacy/4965-mac.c idx = SEQ_TO_IDX(seq); idx 2706 drivers/net/wireless/intel/iwlegacy/4965-mac.c agg->frame_count, txq_id, idx); idx 2708 drivers/net/wireless/intel/iwlegacy/4965-mac.c skb = il->txq[txq_id].skbs[idx]; idx 2714 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (idx != (IEEE80211_SEQ_TO_SN(sc) & 0xff)) { idx 2716 drivers/net/wireless/intel/iwlegacy/4965-mac.c " idx=%d, seq_idx=%d, seq=%d\n", idx, idx 2721 drivers/net/wireless/intel/iwlegacy/4965-mac.c D_TX_REPLY("AGG Frame i=%d idx %d seq=%d\n", i, idx, idx 2724 drivers/net/wireless/intel/iwlegacy/4965-mac.c sh = idx - start; idx 2726 drivers/net/wireless/intel/iwlegacy/4965-mac.c sh = (start - idx) + 0xff; idx 2729 drivers/net/wireless/intel/iwlegacy/4965-mac.c start = idx; idx 2731 drivers/net/wireless/intel/iwlegacy/4965-mac.c sh = 0xff - (start - idx); idx 2733 drivers/net/wireless/intel/iwlegacy/4965-mac.c sh = start - idx; idx 2734 drivers/net/wireless/intel/iwlegacy/4965-mac.c start = idx; idx 2764 drivers/net/wireless/intel/iwlegacy/4965-mac.c int idx = SEQ_TO_IDX(sequence); idx 2777 drivers/net/wireless/intel/iwlegacy/4965-mac.c if (idx >= txq->q.n_bd || il_queue_used(&txq->q, idx) == 0) { idx 2779 drivers/net/wireless/intel/iwlegacy/4965-mac.c "is out of range [0-%d] %d %d\n", txq_id, idx, idx 2823 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tx_status_reply_tx(il, agg, tx_resp, txq_id, idx); idx 2831 drivers/net/wireless/intel/iwlegacy/4965-mac.c idx = il_queue_dec_wrap(scd_ssn & 0xff, txq->q.n_bd); idx 2833 drivers/net/wireless/intel/iwlegacy/4965-mac.c "%d idx %d\n", scd_ssn, idx); idx 2834 drivers/net/wireless/intel/iwlegacy/4965-mac.c freed = il4965_tx_queue_reclaim(il, txq_id, idx); idx 2857 drivers/net/wireless/intel/iwlegacy/4965-mac.c freed = il4965_tx_queue_reclaim(il, txq_id, idx); idx 2896 drivers/net/wireless/intel/iwlegacy/4965-mac.c r->idx = il4965_hwrate_to_mac80211_idx(rate_n_flags, info->band); idx 2912 drivers/net/wireless/intel/iwlegacy/4965-mac.c int idx; idx 2946 drivers/net/wireless/intel/iwlegacy/4965-mac.c idx = il_queue_dec_wrap(ba_resp_scd_ssn & 0xff, txq->q.n_bd); idx 2968 drivers/net/wireless/intel/iwlegacy/4965-mac.c int freed = il4965_tx_queue_reclaim(il, scd_flow, idx); idx 3181 drivers/net/wireless/intel/iwlegacy/4965-mac.c int idx = keyconf->keyidx; idx 3185 drivers/net/wireless/intel/iwlegacy/4965-mac.c D_WEP("Removing default WEP key: idx=%d\n", idx); idx 3187 drivers/net/wireless/intel/iwlegacy/4965-mac.c memset(&il->_4965.wep_keys[idx], 0, sizeof(struct il_wep_key)); idx 3194 drivers/net/wireless/intel/iwlegacy/4965-mac.c D_WEP("Remove default WEP key: idx=%d ret=%d\n", idx, ret); idx 3205 drivers/net/wireless/intel/iwlegacy/4965-mac.c int idx = keyconf->keyidx; idx 3218 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->_4965.wep_keys[idx].key_size = len; idx 3219 drivers/net/wireless/intel/iwlegacy/4965-mac.c memcpy(&il->_4965.wep_keys[idx].key, &keyconf->key, len); idx 3223 drivers/net/wireless/intel/iwlegacy/4965-mac.c D_WEP("Set default WEP key: len=%d idx=%d ret=%d\n", len, idx, ret); idx 3864 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tfd_tb_get_addr(struct il_tfd *tfd, u8 idx) idx 3866 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tfd_tb *tb = &tfd->tbs[idx]; idx 3878 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tfd_tb_get_len(struct il_tfd *tfd, u8 idx) idx 3880 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tfd_tb *tb = &tfd->tbs[idx]; idx 3886 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_tfd_set_tb(struct il_tfd *tfd, u8 idx, dma_addr_t addr, u16 len) idx 3888 drivers/net/wireless/intel/iwlegacy/4965-mac.c struct il_tfd_tb *tb = &tfd->tbs[idx]; idx 3897 drivers/net/wireless/intel/iwlegacy/4965-mac.c tfd->num_tbs = idx + 1; idx 3920 drivers/net/wireless/intel/iwlegacy/4965-mac.c int idx = txq->q.read_ptr; idx 3924 drivers/net/wireless/intel/iwlegacy/4965-mac.c tfd = &tfd_tmp[idx]; idx 3937 drivers/net/wireless/intel/iwlegacy/4965-mac.c pci_unmap_single(dev, dma_unmap_addr(&txq->meta[idx], mapping), idx 3938 drivers/net/wireless/intel/iwlegacy/4965-mac.c dma_unmap_len(&txq->meta[idx], len), idx 6284 drivers/net/wireless/intel/iwlegacy/4965-mac.c il4965_set_wr_ptrs(struct il_priv *il, int txq_id, u32 idx) idx 6286 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_wr(il, HBUS_TARG_WRPTR, (idx & 0xff) | (txq_id << 8)); idx 6287 drivers/net/wireless/intel/iwlegacy/4965-mac.c il_wr_prph(il, IL49_SCD_QUEUE_RDPTR(txq_id), idx); idx 100 drivers/net/wireless/intel/iwlegacy/4965-rs.c int idx = 0; idx 104 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = (rate_n_flags & 0xff); idx 106 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (idx >= RATE_MIMO2_6M_PLCP) idx 107 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = idx - RATE_MIMO2_6M_PLCP; idx 109 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx += IL_FIRST_OFDM_RATE; idx 111 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (idx >= RATE_9M_IDX) idx 112 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx += 1; idx 113 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (idx >= IL_FIRST_OFDM_RATE && idx <= IL_LAST_OFDM_RATE) idx 114 drivers/net/wireless/intel/iwlegacy/4965-rs.c return idx; idx 118 drivers/net/wireless/intel/iwlegacy/4965-rs.c for (idx = 0; idx < ARRAY_SIZE(il_rates); idx++) idx 119 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (il_rates[idx].plcp == (rate_n_flags & 0xFF)) idx 120 drivers/net/wireless/intel/iwlegacy/4965-rs.c return idx; idx 137 drivers/net/wireless/intel/iwlegacy/4965-rs.c u32 *rate_n_flags, int idx); idx 140 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_dbgfs_set_mcs(struct il_lq_sta *lq_sta, u32 * rate_n_flags, int idx) idx 260 drivers/net/wireless/intel/iwlegacy/4965-rs.c s32 idx; idx 288 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = time_diff / TID_QUEUE_CELL_SPACING; idx 292 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (idx >= TID_QUEUE_MAX_SIZE) idx 295 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = (tl->head + idx) % TID_QUEUE_MAX_SIZE; idx 296 drivers/net/wireless/intel/iwlegacy/4965-rs.c tl->packet_count[idx] = tl->packet_count[idx] + 1; idx 299 drivers/net/wireless/intel/iwlegacy/4965-rs.c if ((idx + 1) > tl->queue_count) idx 300 drivers/net/wireless/intel/iwlegacy/4965-rs.c tl->queue_count = idx + 1; idx 313 drivers/net/wireless/intel/iwlegacy/4965-rs.c s32 idx; idx 327 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = time_diff / TID_QUEUE_CELL_SPACING; idx 331 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (idx >= TID_QUEUE_MAX_SIZE) idx 483 drivers/net/wireless/intel/iwlegacy/4965-rs.c int idx, u8 use_green) idx 488 drivers/net/wireless/intel/iwlegacy/4965-rs.c rate_n_flags = il_rates[idx].plcp; idx 489 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (idx >= IL_FIRST_CCK_RATE && idx <= IL_LAST_CCK_RATE) idx 493 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (idx > IL_LAST_OFDM_RATE) { idx 494 drivers/net/wireless/intel/iwlegacy/4965-rs.c IL_ERR("Invalid HT rate idx %d\n", idx); idx 495 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = IL_LAST_OFDM_RATE; idx 500 drivers/net/wireless/intel/iwlegacy/4965-rs.c rate_n_flags |= il_rates[idx].plcp_siso; idx 502 drivers/net/wireless/intel/iwlegacy/4965-rs.c rate_n_flags |= il_rates[idx].plcp_mimo2; idx 657 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_get_adjacent_rate(struct il_priv *il, u8 idx, u16 rate_mask, idx 670 drivers/net/wireless/intel/iwlegacy/4965-rs.c i = idx - 1; idx 679 drivers/net/wireless/intel/iwlegacy/4965-rs.c i = idx + 1; idx 690 drivers/net/wireless/intel/iwlegacy/4965-rs.c low = idx; idx 700 drivers/net/wireless/intel/iwlegacy/4965-rs.c high = idx; idx 842 drivers/net/wireless/intel/iwlegacy/4965-rs.c mac_idx = info->status.rates[0].idx; idx 1066 drivers/net/wireless/intel/iwlegacy/4965-rs.c u16 rate_mask, s8 idx) idx 1071 drivers/net/wireless/intel/iwlegacy/4965-rs.c s32 active_sr = active_tbl->win[idx].success_ratio; idx 1072 drivers/net/wireless/intel/iwlegacy/4965-rs.c s32 active_tpt = active_tbl->expected_tpt[idx]; idx 1079 drivers/net/wireless/intel/iwlegacy/4965-rs.c s8 rate = idx; idx 1161 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct il_scale_tbl_info *tbl, int idx) idx 1192 drivers/net/wireless/intel/iwlegacy/4965-rs.c rate = il4965_rs_get_best_rate(il, lq_sta, tbl, rate_mask, idx); idx 1214 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct il_scale_tbl_info *tbl, int idx) idx 1240 drivers/net/wireless/intel/iwlegacy/4965-rs.c rate = il4965_rs_get_best_rate(il, lq_sta, tbl, rate_mask, idx); idx 1261 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct ieee80211_sta *sta, int idx) idx 1266 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct il_rate_scale_data *win = &(tbl->win[idx]); idx 1316 drivers/net/wireless/intel/iwlegacy/4965-rs.c search_tbl, idx); idx 1345 drivers/net/wireless/intel/iwlegacy/4965-rs.c search_tbl, idx); idx 1380 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct ieee80211_sta *sta, int idx) idx 1386 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct il_rate_scale_data *win = &(tbl->win[idx]); idx 1442 drivers/net/wireless/intel/iwlegacy/4965-rs.c search_tbl, idx); idx 1467 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (tpt >= search_tbl->expected_tpt[idx]) idx 1471 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rate_n_flags_from_tbl(il, search_tbl, idx, idx 1503 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct ieee80211_sta *sta, int idx) idx 1509 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct il_rate_scale_data *win = &(tbl->win[idx]); idx 1563 drivers/net/wireless/intel/iwlegacy/4965-rs.c search_tbl, idx); idx 1591 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (tpt >= search_tbl->expected_tpt[idx]) idx 1595 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rate_n_flags_from_tbl(il, search_tbl, idx, idx 1711 drivers/net/wireless/intel/iwlegacy/4965-rs.c struct il_scale_tbl_info *tbl, int idx, u8 is_green) idx 1716 drivers/net/wireless/intel/iwlegacy/4965-rs.c rate = il4965_rate_n_flags_from_tbl(il, tbl, idx, is_green); idx 1735 drivers/net/wireless/intel/iwlegacy/4965-rs.c int idx; idx 1793 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = lq_sta->last_txrate_idx; idx 1795 drivers/net/wireless/intel/iwlegacy/4965-rs.c D_RATE("Rate scale idx %d for type %d\n", idx, tbl->lq_type); idx 1819 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (!((1 << idx) & rate_scale_idx_msk)) { idx 1827 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = il4965_hwrate_to_plcp_idx(tbl->current_rate); idx 1828 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_update_rate_tbl(il, lq_sta, tbl, idx, idx 1841 drivers/net/wireless/intel/iwlegacy/4965-rs.c if (lq_sta->max_rate_idx != -1 && lq_sta->max_rate_idx < idx) { idx 1842 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = lq_sta->max_rate_idx; idx 1844 drivers/net/wireless/intel/iwlegacy/4965-rs.c win = &(tbl->win[idx]); idx 1848 drivers/net/wireless/intel/iwlegacy/4965-rs.c win = &(tbl->win[idx]); idx 1861 drivers/net/wireless/intel/iwlegacy/4965-rs.c win->success_counter, win->counter, idx); idx 1875 drivers/net/wireless/intel/iwlegacy/4965-rs.c ((win->success_ratio * tbl->expected_tpt[idx] + 64) / 128)) { idx 1878 drivers/net/wireless/intel/iwlegacy/4965-rs.c ((win->success_ratio * tbl->expected_tpt[idx] + 64) / 128); idx 1916 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = il4965_hwrate_to_plcp_idx(tbl->current_rate); idx 1933 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_get_adjacent_rate(il, idx, rate_scale_idx_msk, idx 2007 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = low; idx 2015 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = high; idx 2026 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx, scale_action, low, high, tbl->lq_type); idx 2031 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_update_rate_tbl(il, lq_sta, tbl, idx, is_green); idx 2050 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_move_legacy_other(il, lq_sta, conf, sta, idx); idx 2053 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx); idx 2056 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx); idx 2066 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = il4965_hwrate_to_plcp_idx(tbl->current_rate); idx 2069 drivers/net/wireless/intel/iwlegacy/4965-rs.c tbl->current_rate, idx); idx 2112 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rate_n_flags_from_tbl(il, tbl, idx, is_green); idx 2113 drivers/net/wireless/intel/iwlegacy/4965-rs.c i = idx; idx 2248 drivers/net/wireless/intel/iwlegacy/4965-rs.c info->control.rates[0].idx = rate_idx; idx 2360 drivers/net/wireless/intel/iwlegacy/4965-rs.c int idx = 0; idx 2369 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_dbgfs_set_mcs(lq_sta, &new_rate, idx); idx 2387 drivers/net/wireless/intel/iwlegacy/4965-rs.c lq_cmd->rs_table[idx].rate_n_flags = cpu_to_le32(new_rate); idx 2396 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx++; idx 2402 drivers/net/wireless/intel/iwlegacy/4965-rs.c while (idx < LINK_QUAL_MAX_RETRY_NUM) { idx 2406 drivers/net/wireless/intel/iwlegacy/4965-rs.c while (repeat_rate > 0 && idx < LINK_QUAL_MAX_RETRY_NUM) { idx 2418 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_dbgfs_set_mcs(lq_sta, &new_rate, idx); idx 2421 drivers/net/wireless/intel/iwlegacy/4965-rs.c lq_cmd->rs_table[idx].rate_n_flags = idx 2424 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx++; idx 2434 drivers/net/wireless/intel/iwlegacy/4965-rs.c lq_cmd->general_params.mimo_delimiter = idx; idx 2460 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_dbgfs_set_mcs(lq_sta, &new_rate, idx); idx 2463 drivers/net/wireless/intel/iwlegacy/4965-rs.c lq_cmd->rs_table[idx].rate_n_flags = cpu_to_le32(new_rate); idx 2465 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx++; idx 2501 drivers/net/wireless/intel/iwlegacy/4965-rs.c il4965_rs_dbgfs_set_mcs(struct il_lq_sta *lq_sta, u32 * rate_n_flags, int idx) idx 2573 drivers/net/wireless/intel/iwlegacy/4965-rs.c int idx = 0; idx 2641 drivers/net/wireless/intel/iwlegacy/4965-rs.c idx = idx 2650 drivers/net/wireless/intel/iwlegacy/4965-rs.c il_rate_mcs[idx].mbps); idx 2657 drivers/net/wireless/intel/iwlegacy/4965-rs.c il_rate_mcs[idx].mbps, idx 2658 drivers/net/wireless/intel/iwlegacy/4965-rs.c il_rate_mcs[idx].mcs); idx 71 drivers/net/wireless/intel/iwlegacy/4965.h int il4965_tx_queue_reclaim(struct il_priv *il, int txq_id, int idx); idx 81 drivers/net/wireless/intel/iwlegacy/4965.h void il4965_set_wr_ptrs(struct il_priv *il, int txq_id, u32 idx); idx 1867 drivers/net/wireless/intel/iwlegacy/common.c il_set_ht_add_station(struct il_priv *il, u8 idx, struct ieee80211_sta *sta) idx 1880 drivers/net/wireless/intel/iwlegacy/common.c sta_flags = il->stations[idx].sta.station_flags; idx 1911 drivers/net/wireless/intel/iwlegacy/common.c il->stations[idx].sta.station_flags = sta_flags; idx 3127 drivers/net/wireless/intel/iwlegacy/common.c u32 idx; idx 3158 drivers/net/wireless/intel/iwlegacy/common.c idx = il_get_cmd_idx(q, q->write_ptr, cmd->flags & CMD_SIZE_HUGE); idx 3159 drivers/net/wireless/intel/iwlegacy/common.c out_cmd = txq->cmd[idx]; idx 3160 drivers/net/wireless/intel/iwlegacy/common.c out_meta = &txq->meta[idx]; idx 3186 drivers/net/wireless/intel/iwlegacy/common.c if (idx == TFD_CMD_SLOTS) idx 3197 drivers/net/wireless/intel/iwlegacy/common.c q->write_ptr, idx, il->cmd_queue); idx 3204 drivers/net/wireless/intel/iwlegacy/common.c idx, il->cmd_queue); idx 3212 drivers/net/wireless/intel/iwlegacy/common.c idx = -ENOMEM; idx 3233 drivers/net/wireless/intel/iwlegacy/common.c return idx; idx 3244 drivers/net/wireless/intel/iwlegacy/common.c il_hcmd_queue_reclaim(struct il_priv *il, int txq_id, int idx, int cmd_idx) idx 3250 drivers/net/wireless/intel/iwlegacy/common.c if (idx >= q->n_bd || il_queue_used(q, idx) == 0) { idx 3252 drivers/net/wireless/intel/iwlegacy/common.c "is out of range [0-%d] %d %d.\n", txq_id, idx, q->n_bd, idx 3257 drivers/net/wireless/intel/iwlegacy/common.c for (idx = il_queue_inc_wrap(idx, q->n_bd); q->read_ptr != idx; idx 3261 drivers/net/wireless/intel/iwlegacy/common.c IL_ERR("HCMD skipped: idx (%d) %d %d\n", idx, idx 3283 drivers/net/wireless/intel/iwlegacy/common.c int idx = SEQ_TO_IDX(sequence); idx 3303 drivers/net/wireless/intel/iwlegacy/common.c cmd_idx = il_get_cmd_idx(&txq->q, idx, huge); idx 3321 drivers/net/wireless/intel/iwlegacy/common.c il_hcmd_queue_reclaim(il, txq_id, idx, cmd_idx); idx 862 drivers/net/wireless/intel/iwlegacy/common.h il_get_cmd_idx(struct il_queue *q, u32 idx, int is_huge) idx 873 drivers/net/wireless/intel/iwlegacy/common.h return idx & (q->n_win - 1); idx 2196 drivers/net/wireless/intel/iwlegacy/common.h il_queue_inc_wrap(int idx, int n_bd) idx 2198 drivers/net/wireless/intel/iwlegacy/common.h return ++idx & (n_bd - 1); idx 2207 drivers/net/wireless/intel/iwlegacy/common.h il_queue_dec_wrap(int idx, int n_bd) idx 2209 drivers/net/wireless/intel/iwlegacy/common.h return --idx & (n_bd - 1); idx 81 drivers/net/wireless/intel/iwlwifi/dvm/lib.c int idx = 0; idx 86 drivers/net/wireless/intel/iwlwifi/dvm/lib.c idx = (rate_n_flags & 0xff); idx 87 drivers/net/wireless/intel/iwlwifi/dvm/lib.c return idx; idx 92 drivers/net/wireless/intel/iwlwifi/dvm/lib.c for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++) idx 93 drivers/net/wireless/intel/iwlwifi/dvm/lib.c if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF)) idx 94 drivers/net/wireless/intel/iwlwifi/dvm/lib.c return idx - band_offset; idx 240 drivers/net/wireless/intel/iwlwifi/dvm/main.c if (info->control.rates[0].idx < 0 || idx 244 drivers/net/wireless/intel/iwlwifi/dvm/main.c rate = info->control.rates[0].idx; idx 106 drivers/net/wireless/intel/iwlwifi/dvm/rs.c int idx = 0; idx 110 drivers/net/wireless/intel/iwlwifi/dvm/rs.c idx = rs_extract_rate(rate_n_flags); idx 112 drivers/net/wireless/intel/iwlwifi/dvm/rs.c if (idx >= IWL_RATE_MIMO3_6M_PLCP) idx 113 drivers/net/wireless/intel/iwlwifi/dvm/rs.c idx = idx - IWL_RATE_MIMO3_6M_PLCP; idx 114 drivers/net/wireless/intel/iwlwifi/dvm/rs.c else if (idx >= IWL_RATE_MIMO2_6M_PLCP) idx 115 drivers/net/wireless/intel/iwlwifi/dvm/rs.c idx = idx - IWL_RATE_MIMO2_6M_PLCP; idx 117 drivers/net/wireless/intel/iwlwifi/dvm/rs.c idx += IWL_FIRST_OFDM_RATE; idx 119 drivers/net/wireless/intel/iwlwifi/dvm/rs.c if (idx >= IWL_RATE_9M_INDEX) idx 120 drivers/net/wireless/intel/iwlwifi/dvm/rs.c idx += 1; idx 121 drivers/net/wireless/intel/iwlwifi/dvm/rs.c if ((idx >= IWL_FIRST_OFDM_RATE) && (idx <= IWL_LAST_OFDM_RATE)) idx 122 drivers/net/wireless/intel/iwlwifi/dvm/rs.c return idx; idx 126 drivers/net/wireless/intel/iwlwifi/dvm/rs.c for (idx = 0; idx < ARRAY_SIZE(iwl_rates); idx++) idx 127 drivers/net/wireless/intel/iwlwifi/dvm/rs.c if (iwl_rates[idx].plcp == idx 129 drivers/net/wireless/intel/iwlwifi/dvm/rs.c return idx; idx 930 drivers/net/wireless/intel/iwlwifi/dvm/rs.c mac_index = info->status.rates[0].idx; idx 2755 drivers/net/wireless/intel/iwlwifi/dvm/rs.c info->control.rates[0].idx = rate_idx; idx 157 drivers/net/wireless/intel/iwlwifi/dvm/tx.c rate_idx = info->control.rates[0].idx; idx 826 drivers/net/wireless/intel/iwlwifi/dvm/tx.c r->idx = iwlagn_hwrate_to_mac80211_idx(rate_n_flags, info->band); idx 108 drivers/net/wireless/intel/iwlwifi/fw/api/mac-cfg.h u8 idx; idx 1060 drivers/net/wireless/intel/iwlwifi/fw/dbg.c void *range_ptr, int idx) idx 1065 drivers/net/wireless/intel/iwlwifi/fw/dbg.c u32 addr = le32_to_cpu(reg->start_addr[idx]) + le32_to_cpu(reg->offset); idx 1082 drivers/net/wireless/intel/iwlwifi/fw/dbg.c void *range_ptr, int idx) idx 1086 drivers/net/wireless/intel/iwlwifi/fw/dbg.c u32 addr = le32_to_cpu(reg->start_addr[idx]) + le32_to_cpu(reg->offset); idx 1099 drivers/net/wireless/intel/iwlwifi/fw/dbg.c void *range_ptr, int idx) idx 1102 drivers/net/wireless/intel/iwlwifi/fw/dbg.c u32 addr = le32_to_cpu(reg->start_addr[idx]) + le32_to_cpu(reg->offset); idx 1114 drivers/net/wireless/intel/iwlwifi/fw/dbg.c void *range_ptr, int idx) idx 1119 drivers/net/wireless/intel/iwlwifi/fw/dbg.c struct page *page = fwrt->fw_paging_db[++idx].fw_paging_block; idx 1121 drivers/net/wireless/intel/iwlwifi/fw/dbg.c dma_addr_t addr = fwrt->fw_paging_db[idx].fw_paging_phys; idx 1122 drivers/net/wireless/intel/iwlwifi/fw/dbg.c u32 page_size = fwrt->fw_paging_db[idx].fw_paging_size; idx 1124 drivers/net/wireless/intel/iwlwifi/fw/dbg.c range->page_num = cpu_to_le32(idx); idx 1137 drivers/net/wireless/intel/iwlwifi/fw/dbg.c void *range_ptr, int idx) idx 1143 drivers/net/wireless/intel/iwlwifi/fw/dbg.c return _iwl_dump_ini_paging_iter(fwrt, reg, range_ptr, idx); idx 1146 drivers/net/wireless/intel/iwlwifi/fw/dbg.c page_size = fwrt->trans->init_dram.paging[idx].size; idx 1148 drivers/net/wireless/intel/iwlwifi/fw/dbg.c range->page_num = cpu_to_le32(idx); idx 1150 drivers/net/wireless/intel/iwlwifi/fw/dbg.c memcpy(range->data, fwrt->trans->init_dram.paging[idx].block, idx 1159 drivers/net/wireless/intel/iwlwifi/fw/dbg.c int idx) idx 1169 drivers/net/wireless/intel/iwlwifi/fw/dbg.c range->range_data_size = cpu_to_le32(fwrt->trans->dbg.fw_mon[idx].size); idx 1171 drivers/net/wireless/intel/iwlwifi/fw/dbg.c memcpy(range->data, fwrt->trans->dbg.fw_mon[idx].block, idx 1172 drivers/net/wireless/intel/iwlwifi/fw/dbg.c fwrt->trans->dbg.fw_mon[idx].size); idx 1178 drivers/net/wireless/intel/iwlwifi/fw/dbg.c struct iwl_fw_ini_region_cfg *reg, int idx) idx 1186 drivers/net/wireless/intel/iwlwifi/fw/dbg.c if (!idx) { idx 1228 drivers/net/wireless/intel/iwlwifi/fw/dbg.c void *range_ptr, int idx) idx 1240 drivers/net/wireless/intel/iwlwifi/fw/dbg.c if (!iwl_ini_txf_iter(fwrt, reg, idx)) idx 1334 drivers/net/wireless/intel/iwlwifi/fw/dbg.c void *range_ptr, int idx) idx 1634 drivers/net/wireless/intel/iwlwifi/fw/dbg.c int idx); idx 2051 drivers/net/wireless/intel/iwlwifi/fw/dbg.c if (test_and_set_bit(fwrt->dump.wks[0].idx, &fwrt->dump.active_wks)) idx 2142 drivers/net/wireless/intel/iwlwifi/fw/dbg.c unsigned long idx; idx 2174 drivers/net/wireless/intel/iwlwifi/fw/dbg.c idx = ffz(fwrt->dump.active_wks); idx 2176 drivers/net/wireless/intel/iwlwifi/fw/dbg.c if (idx >= IWL_FW_RUNTIME_DUMP_WK_NUM || idx 2177 drivers/net/wireless/intel/iwlwifi/fw/dbg.c test_and_set_bit(fwrt->dump.wks[idx].idx, &fwrt->dump.active_wks)) idx 2180 drivers/net/wireless/intel/iwlwifi/fw/dbg.c fwrt->dump.wks[idx].ini_trig_id = id; idx 2184 drivers/net/wireless/intel/iwlwifi/fw/dbg.c schedule_delayed_work(&fwrt->dump.wks[idx].wk, usecs_to_jiffies(delay)); idx 2335 drivers/net/wireless/intel/iwlwifi/fw/dbg.c fwrt = container_of(wks, struct iwl_fw_runtime, dump.wks[wks->idx]); idx 2344 drivers/net/wireless/intel/iwlwifi/fw/dbg.c iwl_fw_dbg_collect_sync(fwrt, wks->idx); idx 80 drivers/net/wireless/intel/iwlwifi/fw/init.c fwrt->dump.wks[i].idx = i; idx 168 drivers/net/wireless/intel/iwlwifi/fw/paging.c int sec_idx, idx, ret; idx 228 drivers/net/wireless/intel/iwlwifi/fw/paging.c for (idx = 1; idx < fwrt->num_of_paging_blk + 1; idx++) { idx 229 drivers/net/wireless/intel/iwlwifi/fw/paging.c struct iwl_fw_paging *block = &fwrt->fw_paging_db[idx]; idx 237 drivers/net/wireless/intel/iwlwifi/fw/paging.c if (idx == fwrt->num_of_paging_blk) { idx 250 drivers/net/wireless/intel/iwlwifi/fw/paging.c idx, remaining); idx 264 drivers/net/wireless/intel/iwlwifi/fw/paging.c len, idx); idx 145 drivers/net/wireless/intel/iwlwifi/fw/runtime.h u8 idx; idx 418 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c int idx, entries; idx 432 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c for (idx = 0; idx < entries; idx++) { idx 433 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c txp = &txp_array[idx]; idx 711 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c int n = 0, idx = 0; idx 713 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c while (idx < n_channels && chan->band != band) idx 714 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c chan = &data->channels[++idx]; idx 716 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c sband->channels = &data->channels[idx]; idx 718 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c while (idx < n_channels && chan->band == band) { idx 719 drivers/net/wireless/intel/iwlwifi/iwl-eeprom-parse.c chan = &data->channels[++idx]; idx 67 drivers/net/wireless/intel/iwlwifi/mvm/binding.c int idx; idx 106 drivers/net/wireless/intel/iwlwifi/mvm/binding.c for (i = 0; i < data->idx; i++) idx 139 drivers/net/wireless/intel/iwlwifi/mvm/binding.c if (WARN_ON_ONCE(data->idx >= MAX_MACS_IN_BINDING)) idx 142 drivers/net/wireless/intel/iwlwifi/mvm/binding.c data->ids[data->idx] = mvmvif->id; idx 143 drivers/net/wireless/intel/iwlwifi/mvm/binding.c data->colors[data->idx] = mvmvif->color; idx 144 drivers/net/wireless/intel/iwlwifi/mvm/binding.c data->idx++; idx 170 drivers/net/wireless/intel/iwlwifi/mvm/binding.c if (data.idx == 0) { idx 178 drivers/net/wireless/intel/iwlwifi/mvm/binding.c if (WARN_ON_ONCE(data.idx >= MAX_MACS_IN_BINDING)) idx 181 drivers/net/wireless/intel/iwlwifi/mvm/binding.c data.ids[data.idx] = mvmvif->id; idx 182 drivers/net/wireless/intel/iwlwifi/mvm/binding.c data.colors[data.idx] = mvmvif->color; idx 183 drivers/net/wireless/intel/iwlwifi/mvm/binding.c data.idx++; idx 104 drivers/net/wireless/intel/iwlwifi/mvm/d3.c int idx = 0; idx 110 drivers/net/wireless/intel/iwlwifi/mvm/d3.c mvmvif->target_ipv6_addrs[idx] = ifa->addr; idx 112 drivers/net/wireless/intel/iwlwifi/mvm/d3.c __set_bit(idx, mvmvif->tentative_addrs); idx 113 drivers/net/wireless/intel/iwlwifi/mvm/d3.c idx++; idx 114 drivers/net/wireless/intel/iwlwifi/mvm/d3.c if (idx >= IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX) idx 119 drivers/net/wireless/intel/iwlwifi/mvm/d3.c mvmvif->num_target_ipv6_addrs = idx; idx 124 drivers/net/wireless/intel/iwlwifi/mvm/d3.c struct ieee80211_vif *vif, int idx) idx 128 drivers/net/wireless/intel/iwlwifi/mvm/d3.c mvmvif->tx_key_idx = idx; idx 1749 drivers/net/wireless/intel/iwlwifi/mvm/d3.c int idx) idx 1759 drivers/net/wireless/intel/iwlwifi/mvm/d3.c n_chans += hweight8(matches[idx].matching_channels[i]); idx 1765 drivers/net/wireless/intel/iwlwifi/mvm/d3.c n_chans += hweight8(matches[idx].matching_channels[i]); idx 1774 drivers/net/wireless/intel/iwlwifi/mvm/d3.c int idx) idx 1784 drivers/net/wireless/intel/iwlwifi/mvm/d3.c if (matches[idx].matching_channels[i / 8] & (BIT(i % 8))) idx 1792 drivers/net/wireless/intel/iwlwifi/mvm/d3.c if (matches[idx].matching_channels[i / 8] & (BIT(i % 8))) idx 1845 drivers/net/wireless/intel/iwlwifi/mvm/d3.c int idx, n_channels = 0; idx 1859 drivers/net/wireless/intel/iwlwifi/mvm/d3.c idx = mvm->n_nd_match_sets - i - 1; idx 1860 drivers/net/wireless/intel/iwlwifi/mvm/d3.c match->ssid.ssid_len = mvm->nd_match_sets[idx].ssid.ssid_len; idx 1861 drivers/net/wireless/intel/iwlwifi/mvm/d3.c memcpy(match->ssid.ssid, mvm->nd_match_sets[idx].ssid.ssid, idx 965 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c int idx, i; idx 1006 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c idx = stats->last_frame_idx - 1; idx 1008 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c idx = (idx + 1) % ARRAY_SIZE(stats->last_rates); idx 1009 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c if (stats->last_rates[idx] == 0) idx 1014 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c stats->last_rates[idx]); idx 803 drivers/net/wireless/intel/iwlwifi/mvm/fw.c int idx = 1; idx 826 drivers/net/wireless/intel/iwlwifi/mvm/fw.c entry = &wifi_pkg->package.elements[idx++]; idx 848 drivers/net/wireless/intel/iwlwifi/mvm/fw.c int i, j, idx; idx 894 drivers/net/wireless/intel/iwlwifi/mvm/fw.c idx = (i * ACPI_SAR_NUM_SUB_BANDS) + j; idx 896 drivers/net/wireless/intel/iwlwifi/mvm/fw.c cpu_to_le16(prof->table[idx]); idx 898 drivers/net/wireless/intel/iwlwifi/mvm/fw.c j, prof->table[idx]); idx 1037 drivers/net/wireless/intel/iwlwifi/mvm/fw.c int idx = 2; idx 1079 drivers/net/wireless/intel/iwlwifi/mvm/fw.c ent = &wifi_pkg->package.elements[idx++]; idx 4693 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c static int iwl_mvm_mac_get_survey(struct ieee80211_hw *hw, int idx, idx 4702 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c if (idx != 0) idx 1813 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h struct ieee80211_vif *vif, int idx); idx 188 drivers/net/wireless/intel/iwlwifi/mvm/quota.c int i, idx, err, num_active_macs, quota, quota_rem, n_non_lowlat; idx 268 drivers/net/wireless/intel/iwlwifi/mvm/quota.c for (idx = 0, i = 0; i < MAX_BINDINGS; i++) { idx 272 drivers/net/wireless/intel/iwlwifi/mvm/quota.c qdata = iwl_mvm_quota_cmd_get_quota(mvm, &cmd, idx); idx 299 drivers/net/wireless/intel/iwlwifi/mvm/quota.c idx, le32_to_cpu(qdata->quota), QUOTA_100); idx 303 drivers/net/wireless/intel/iwlwifi/mvm/quota.c idx++; idx 344 drivers/net/wireless/intel/iwlwifi/mvm/rs.c int idx = 0; idx 347 drivers/net/wireless/intel/iwlwifi/mvm/rs.c idx = rate_n_flags & RATE_HT_MCS_RATE_CODE_MSK; idx 348 drivers/net/wireless/intel/iwlwifi/mvm/rs.c idx += IWL_RATE_MCS_0_INDEX; idx 351 drivers/net/wireless/intel/iwlwifi/mvm/rs.c if (idx >= IWL_RATE_9M_INDEX) idx 352 drivers/net/wireless/intel/iwlwifi/mvm/rs.c idx += 1; idx 353 drivers/net/wireless/intel/iwlwifi/mvm/rs.c if ((idx >= IWL_FIRST_HT_RATE) && (idx <= IWL_LAST_HT_RATE)) idx 354 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return idx; idx 357 drivers/net/wireless/intel/iwlwifi/mvm/rs.c idx = rate_n_flags & RATE_VHT_MCS_RATE_CODE_MSK; idx 358 drivers/net/wireless/intel/iwlwifi/mvm/rs.c idx += IWL_RATE_MCS_0_INDEX; idx 361 drivers/net/wireless/intel/iwlwifi/mvm/rs.c if (idx >= IWL_RATE_9M_INDEX) idx 362 drivers/net/wireless/intel/iwlwifi/mvm/rs.c idx++; idx 363 drivers/net/wireless/intel/iwlwifi/mvm/rs.c if ((idx >= IWL_FIRST_VHT_RATE) && (idx <= IWL_LAST_VHT_RATE)) idx 364 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return idx; idx 366 drivers/net/wireless/intel/iwlwifi/mvm/rs.c (idx <= IWL_LAST_HE_RATE)) idx 367 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return idx; idx 372 drivers/net/wireless/intel/iwlwifi/mvm/rs.c for (idx = 0; idx < ARRAY_SIZE(iwl_rates); idx++) idx 373 drivers/net/wireless/intel/iwlwifi/mvm/rs.c if (iwl_rates[idx].plcp == legacy_rate) idx 374 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return idx; idx 565 drivers/net/wireless/intel/iwlwifi/mvm/tt.c int i, j, idx = 0; idx 579 drivers/net/wireless/intel/iwlwifi/mvm/tt.c cmd.thresholds[idx++] = idx 583 drivers/net/wireless/intel/iwlwifi/mvm/tt.c cmd.num_temps = cpu_to_le32(idx); idx 585 drivers/net/wireless/intel/iwlwifi/mvm/tt.c if (!idx) idx 589 drivers/net/wireless/intel/iwlwifi/mvm/tt.c sort(cmd.thresholds, idx, sizeof(s16), compare_temps, NULL); idx 594 drivers/net/wireless/intel/iwlwifi/mvm/tt.c for (i = 0; i < idx; i++) { idx 336 drivers/net/wireless/intel/iwlwifi/mvm/tx.c info->control.rates[0].idx); idx 338 drivers/net/wireless/intel/iwlwifi/mvm/tx.c rate_idx = info->control.rates[0].idx; idx 1366 drivers/net/wireless/intel/iwlwifi/mvm/tx.c r->idx = rate_n_flags & RATE_HT_MCS_INDEX_MSK; idx 1374 drivers/net/wireless/intel/iwlwifi/mvm/tx.c r->idx = iwl_mvm_legacy_rate_to_mac80211_idx(rate_n_flags, idx 216 drivers/net/wireless/intel/iwlwifi/mvm/utils.c int idx; idx 222 drivers/net/wireless/intel/iwlwifi/mvm/utils.c for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++) idx 223 drivers/net/wireless/intel/iwlwifi/mvm/utils.c if (fw_rate_idx_to_plcp[idx] == rate) idx 224 drivers/net/wireless/intel/iwlwifi/mvm/utils.c return idx - band_offset; idx 390 drivers/net/wireless/intel/iwlwifi/pcie/internal.h iwl_pcie_get_first_tb_dma(struct iwl_txq *txq, int idx) idx 393 drivers/net/wireless/intel/iwlwifi/pcie/internal.h sizeof(struct iwl_pcie_first_tb_buf) * idx; idx 696 drivers/net/wireless/intel/iwlwifi/pcie/internal.h void iwl_pcie_cmdq_reclaim(struct iwl_trans *trans, int txq_id, int idx); idx 710 drivers/net/wireless/intel/iwlwifi/pcie/internal.h u8 idx) idx 714 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_tfh_tb *tb = &tfd->tbs[idx]; idx 719 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_tfd_tb *tb = &tfd->tbs[idx]; idx 914 drivers/net/wireless/intel/iwlwifi/pcie/internal.h struct iwl_txq *txq, int idx) idx 919 drivers/net/wireless/intel/iwlwifi/pcie/internal.h idx = iwl_pcie_get_cmd_index(txq, idx); idx 921 drivers/net/wireless/intel/iwlwifi/pcie/internal.h return txq->tfds + trans_pcie->tfd_size * idx; idx 1163 drivers/net/wireless/intel/iwlwifi/pcie/trans.c u32 val, idx; idx 1172 drivers/net/wireless/intel/iwlwifi/pcie/trans.c for (idx = 1; idx < trans->num_rx_queues; idx++) { idx 1173 drivers/net/wireless/intel/iwlwifi/pcie/trans.c iwl_write8(trans, CSR_MSIX_RX_IVAR(idx), idx 1174 drivers/net/wireless/intel/iwlwifi/pcie/trans.c MSIX_FH_INT_CAUSES_Q(idx - offset)); idx 1175 drivers/net/wireless/intel/iwlwifi/pcie/trans.c val |= BIT(MSIX_FH_INT_CAUSES_Q(idx)); idx 3305 drivers/net/wireless/intel/iwlwifi/pcie/trans.c u8 idx = iwl_pcie_get_cmd_index(cmdq, ptr); idx 3310 drivers/net/wireless/intel/iwlwifi/pcie/trans.c tfdidx = idx; idx 3323 drivers/net/wireless/intel/iwlwifi/pcie/trans.c memcpy(txcmd->data, cmdq->entries[idx].cmd, idx 96 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); idx 101 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (WARN(idx >= txq->n_window, "%d >= %d\n", idx, txq->n_window)) idx 121 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c scd_bc_tbl_gen3->tfd_offset[idx] = bc_ent; idx 128 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c scd_bc_tbl->tfd_offset[idx] = bc_ent; idx 192 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->read_ptr); idx 196 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_tfd_unmap(trans, &txq->entries[idx].meta, idx 197 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_get_tfd(trans, txq, idx)); idx 203 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c skb = txq->entries[idx].skb; idx 211 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].skb = NULL; idx 221 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_gen2_get_num_tbs(trans, tfd); idx 224 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (WARN_ON(idx >= IWL_TFH_NUM_TBS)) idx 226 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tb = &tfd->tbs[idx]; idx 238 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tfd->num_tbs = cpu_to_le16(idx + 1); idx 240 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c return idx; idx 381 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); idx 382 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx); idx 387 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tb_phys = iwl_pcie_get_first_tb_dma(txq, idx); idx 419 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); idx 471 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); idx 472 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx); idx 478 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c tb_phys = iwl_pcie_get_first_tb_dma(txq, idx); idx 481 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c memcpy(&txq->first_tb_bufs[idx], dev_cmd, IWL_FIRST_TB_SIZE); idx 557 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); idx 558 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct iwl_tfh_tfd *tfd = iwl_pcie_get_tfd(trans, txq, idx); idx 598 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx; idx 629 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); idx 632 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].skb = skb; idx 633 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].cmd = dev_cmd; idx 637 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c INDEX_TO_SEQ(idx))); idx 640 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c out_meta = &txq->entries[idx].meta; idx 701 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int i, cmd_pos, idx; idx 733 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = -EINVAL; idx 745 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = -EINVAL; idx 756 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = -EINVAL; idx 772 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = -EINVAL; idx 778 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); idx 787 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = -ENOSPC; idx 791 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c out_cmd = txq->entries[idx].cmd; idx 792 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c out_meta = &txq->entries[idx].meta; idx 855 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c cmd_size, txq->write_ptr, idx, trans_pcie->cmd_queue); idx 859 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c memcpy(&txq->first_tb_bufs[idx], out_cmd, tb0_size); idx 860 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c iwl_pcie_gen2_set_tb(trans, tfd, iwl_pcie_get_first_tb_dma(txq, idx), idx 870 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = -ENOMEM; idx 892 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c idx = -ENOMEM; idx 901 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (WARN_ON_ONCE(txq->entries[idx].free_buf)) idx 902 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c kzfree(txq->entries[idx].free_buf); idx 903 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->entries[idx].free_buf = dup_buf; idx 920 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (idx < 0) idx 922 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c return idx; idx 1059 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int idx = iwl_pcie_get_cmd_index(txq, txq->read_ptr); idx 1060 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c struct sk_buff *skb = txq->entries[idx].skb; idx 346 drivers/net/wireless/intel/iwlwifi/pcie/tx.c void *_tfd, u8 idx) idx 351 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_tfh_tb *tb = &tfd->tbs[idx]; idx 356 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_tfd_tb *tb = &tfd->tbs[idx]; idx 375 drivers/net/wireless/intel/iwlwifi/pcie/tx.c u8 idx, dma_addr_t addr, u16 len) idx 378 drivers/net/wireless/intel/iwlwifi/pcie/tx.c struct iwl_tfd_tb *tb = &tfd_fh->tbs[idx]; idx 387 drivers/net/wireless/intel/iwlwifi/pcie/tx.c tfd_fh->num_tbs = idx + 1; idx 466 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int idx = iwl_pcie_get_cmd_index(txq, rd_ptr); idx 473 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_tfd_unmap(trans, &txq->entries[idx].meta, txq, rd_ptr); idx 479 drivers/net/wireless/intel/iwlwifi/pcie/tx.c skb = txq->entries[idx].skb; idx 487 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[idx].skb = NULL; idx 1284 drivers/net/wireless/intel/iwlwifi/pcie/tx.c void iwl_pcie_cmdq_reclaim(struct iwl_trans *trans, int txq_id, int idx) idx 1294 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = iwl_pcie_get_cmd_index(txq, idx); idx 1297 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (idx >= trans->trans_cfg->base_params->max_tfd_queue_size || idx 1298 drivers/net/wireless/intel/iwlwifi/pcie/tx.c (!iwl_queue_used(txq, idx))) { idx 1301 drivers/net/wireless/intel/iwlwifi/pcie/tx.c __func__, txq_id, idx, idx 1307 drivers/net/wireless/intel/iwlwifi/pcie/tx.c for (idx = iwl_queue_inc_wrap(trans, idx); r != idx; idx 1313 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx, txq->write_ptr, r); idx 1534 drivers/net/wireless/intel/iwlwifi/pcie/tx.c int idx; idx 1580 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = -EINVAL; idx 1592 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = -EINVAL; idx 1603 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = -EINVAL; idx 1621 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = -EINVAL; idx 1632 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = -ENOSPC; idx 1636 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = iwl_pcie_get_cmd_index(txq, txq->write_ptr); idx 1637 drivers/net/wireless/intel/iwlwifi/pcie/tx.c out_cmd = txq->entries[idx].cmd; idx 1638 drivers/net/wireless/intel/iwlwifi/pcie/tx.c out_meta = &txq->entries[idx].meta; idx 1713 drivers/net/wireless/intel/iwlwifi/pcie/tx.c cmd_size, txq->write_ptr, idx, trans_pcie->cmd_queue); idx 1717 drivers/net/wireless/intel/iwlwifi/pcie/tx.c memcpy(&txq->first_tb_bufs[idx], &out_cmd->hdr, tb0_size); idx 1719 drivers/net/wireless/intel/iwlwifi/pcie/tx.c iwl_pcie_get_first_tb_dma(txq, idx), idx 1731 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = -ENOMEM; idx 1755 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = -ENOMEM; idx 1764 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (WARN_ON_ONCE(txq->entries[idx].free_buf)) idx 1765 drivers/net/wireless/intel/iwlwifi/pcie/tx.c kzfree(txq->entries[idx].free_buf); idx 1766 drivers/net/wireless/intel/iwlwifi/pcie/tx.c txq->entries[idx].free_buf = dup_buf; idx 1777 drivers/net/wireless/intel/iwlwifi/pcie/tx.c idx = ret; idx 1791 drivers/net/wireless/intel/iwlwifi/pcie/tx.c if (idx < 0) idx 1793 drivers/net/wireless/intel/iwlwifi/pcie/tx.c return idx; idx 22 drivers/net/wireless/intersil/hostap/hostap.h int hostap_tx_callback_unregister(local_info_t *local, u16 idx); idx 781 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c int idx = 0; idx 783 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c idx = skb->data[hdrlen + 3] >> 6; idx 784 drivers/net/wireless/intersil/hostap/hostap_80211_rx.c crypt = local->crypt_info.crypt[idx]; idx 1318 drivers/net/wireless/intersil/hostap/hostap_ap.c int idx = 0; idx 1320 drivers/net/wireless/intersil/hostap/hostap_ap.c idx = skb->data[hdrlen + 3] >> 6; idx 1321 drivers/net/wireless/intersil/hostap/hostap_ap.c crypt = local->crypt_info.crypt[idx]; idx 378 drivers/net/wireless/intersil/hostap/hostap_common.h u8 idx; idx 1656 drivers/net/wireless/intersil/hostap/hostap_hw.c int idx, end; idx 1660 drivers/net/wireless/intersil/hostap/hostap_hw.c end = idx = local->next_txfid; idx 1662 drivers/net/wireless/intersil/hostap/hostap_hw.c if (local->intransmitfid[idx] == PRISM2_TXFID_EMPTY) { idx 1663 drivers/net/wireless/intersil/hostap/hostap_hw.c local->intransmitfid[idx] = PRISM2_TXFID_RESERVED; idx 1665 drivers/net/wireless/intersil/hostap/hostap_hw.c return idx; idx 1667 drivers/net/wireless/intersil/hostap/hostap_hw.c idx++; idx 1668 drivers/net/wireless/intersil/hostap/hostap_hw.c if (idx >= PRISM2_TXFID_COUNT) idx 1669 drivers/net/wireless/intersil/hostap/hostap_hw.c idx = 0; idx 1670 drivers/net/wireless/intersil/hostap/hostap_hw.c } while (idx != end); idx 1687 drivers/net/wireless/intersil/hostap/hostap_hw.c int idx = (int) context; idx 1698 drivers/net/wireless/intersil/hostap/hostap_hw.c if (idx < 0 || idx >= PRISM2_TXFID_COUNT) { idx 1700 drivers/net/wireless/intersil/hostap/hostap_hw.c "idx=%d\n", dev->name, idx); idx 1719 drivers/net/wireless/intersil/hostap/hostap_hw.c local->intransmitfid[idx] = resp0; idx 1723 drivers/net/wireless/intersil/hostap/hostap_hw.c dev->name, idx, local->txfid[idx], idx 1726 drivers/net/wireless/intersil/hostap/hostap_hw.c idx++; idx 1727 drivers/net/wireless/intersil/hostap/hostap_hw.c if (idx >= PRISM2_TXFID_COUNT) idx 1728 drivers/net/wireless/intersil/hostap/hostap_hw.c idx = 0; idx 1729 drivers/net/wireless/intersil/hostap/hostap_hw.c local->next_txfid = idx; idx 1733 drivers/net/wireless/intersil/hostap/hostap_hw.c if (local->intransmitfid[idx] == PRISM2_TXFID_EMPTY) { idx 1737 drivers/net/wireless/intersil/hostap/hostap_hw.c idx++; idx 1738 drivers/net/wireless/intersil/hostap/hostap_hw.c if (idx >= PRISM2_TXFID_COUNT) idx 1739 drivers/net/wireless/intersil/hostap/hostap_hw.c idx = 0; idx 1740 drivers/net/wireless/intersil/hostap/hostap_hw.c } while (idx != local->next_txfid); idx 1750 drivers/net/wireless/intersil/hostap/hostap_hw.c static int prism2_transmit(struct net_device *dev, int idx) idx 1776 drivers/net/wireless/intersil/hostap/hostap_hw.c local->txfid[idx], idx 1777 drivers/net/wireless/intersil/hostap/hostap_hw.c prism2_transmit_cb, (long) idx); idx 1805 drivers/net/wireless/intersil/hostap/hostap_hw.c int hdr_len, data_len, idx, res, ret = -1; idx 1851 drivers/net/wireless/intersil/hostap/hostap_hw.c idx = prism2_get_txfid_idx(local); idx 1852 drivers/net/wireless/intersil/hostap/hostap_hw.c if (idx < 0) idx 1859 drivers/net/wireless/intersil/hostap/hostap_hw.c res = hfa384x_setup_bap(dev, BAP0, local->txfid[idx], 0); idx 1869 drivers/net/wireless/intersil/hostap/hostap_hw.c res = prism2_transmit(dev, idx); idx 1873 drivers/net/wireless/intersil/hostap/hostap_hw.c local->intransmitfid[idx] = PRISM2_TXFID_EMPTY; idx 2101 drivers/net/wireless/intersil/hostap/hostap_hw.c int idx; idx 2112 drivers/net/wireless/intersil/hostap/hostap_hw.c idx = local->next_alloc; idx 2115 drivers/net/wireless/intersil/hostap/hostap_hw.c if (local->txfid[idx] == fid) { idx 2117 drivers/net/wireless/intersil/hostap/hostap_hw.c idx); idx 2120 drivers/net/wireless/intersil/hostap/hostap_hw.c if (local->intransmitfid[idx] == PRISM2_TXFID_EMPTY) idx 2122 drivers/net/wireless/intersil/hostap/hostap_hw.c "%d\n", idx); idx 2123 drivers/net/wireless/intersil/hostap/hostap_hw.c if (local->intransmitfid[idx] == PRISM2_TXFID_RESERVED) idx 2125 drivers/net/wireless/intersil/hostap/hostap_hw.c "%d\n", idx); idx 2127 drivers/net/wireless/intersil/hostap/hostap_hw.c local->intransmitfid[idx] = PRISM2_TXFID_EMPTY; idx 2128 drivers/net/wireless/intersil/hostap/hostap_hw.c idx++; idx 2129 drivers/net/wireless/intersil/hostap/hostap_hw.c local->next_alloc = idx >= PRISM2_TXFID_COUNT ? 0 : idx 2130 drivers/net/wireless/intersil/hostap/hostap_hw.c idx; idx 2140 drivers/net/wireless/intersil/hostap/hostap_hw.c idx++; idx 2141 drivers/net/wireless/intersil/hostap/hostap_hw.c if (idx >= PRISM2_TXFID_COUNT) idx 2142 drivers/net/wireless/intersil/hostap/hostap_hw.c idx = 0; idx 2143 drivers/net/wireless/intersil/hostap/hostap_hw.c } while (idx != local->next_alloc); idx 2149 drivers/net/wireless/intersil/hostap/hostap_hw.c for (idx = 0; idx < PRISM2_TXFID_COUNT; idx++) idx 2150 drivers/net/wireless/intersil/hostap/hostap_hw.c printk(" %04x[%04x]", local->txfid[idx], idx 2151 drivers/net/wireless/intersil/hostap/hostap_hw.c local->intransmitfid[idx]); idx 2182 drivers/net/wireless/intersil/hostap/hostap_hw.c while (cb != NULL && cb->idx != sw_support) idx 3458 drivers/net/wireless/intersil/hostap/hostap_ioctl.c if (param->u.crypt.idx >= WEP_KEYS) idx 3461 drivers/net/wireless/intersil/hostap/hostap_ioctl.c crypt = &local->crypt_info.crypt[param->u.crypt.idx]; idx 3463 drivers/net/wireless/intersil/hostap/hostap_ioctl.c if (param->u.crypt.idx) idx 3517 drivers/net/wireless/intersil/hostap/hostap_ioctl.c new_crypt->priv = new_crypt->ops->init(param->u.crypt.idx); idx 3543 drivers/net/wireless/intersil/hostap/hostap_ioctl.c local->crypt_info.tx_keyidx = param->u.crypt.idx; idx 3544 drivers/net/wireless/intersil/hostap/hostap_ioctl.c else if (param->u.crypt.idx) { idx 3592 drivers/net/wireless/intersil/hostap/hostap_ioctl.c if (param->u.crypt.idx >= WEP_KEYS) idx 3593 drivers/net/wireless/intersil/hostap/hostap_ioctl.c param->u.crypt.idx = local->crypt_info.tx_keyidx; idx 3594 drivers/net/wireless/intersil/hostap/hostap_ioctl.c crypt = &local->crypt_info.crypt[param->u.crypt.idx]; idx 3596 drivers/net/wireless/intersil/hostap/hostap_ioctl.c param->u.crypt.idx = 0; idx 3609 drivers/net/wireless/intersil/hostap/hostap_ioctl.c param->u.crypt.idx = 0xff; idx 249 drivers/net/wireless/intersil/hostap/hostap_main.c entry->idx = local->tx_callback ? local->tx_callback->idx + 1 : 1; idx 254 drivers/net/wireless/intersil/hostap/hostap_main.c return entry->idx; idx 258 drivers/net/wireless/intersil/hostap/hostap_main.c int hostap_tx_callback_unregister(local_info_t *local, u16 idx) idx 265 drivers/net/wireless/intersil/hostap/hostap_main.c while (cb != NULL && cb->idx != idx) { idx 329 drivers/net/wireless/intersil/hostap/hostap_main.c int i, keylen, len, idx; idx 333 drivers/net/wireless/intersil/hostap/hostap_main.c idx = local->crypt_info.tx_keyidx; idx 334 drivers/net/wireless/intersil/hostap/hostap_main.c if (local->crypt_info.crypt[idx] == NULL || idx 335 drivers/net/wireless/intersil/hostap/hostap_main.c local->crypt_info.crypt[idx]->ops == NULL) idx 337 drivers/net/wireless/intersil/hostap/hostap_main.c else if (strcmp(local->crypt_info.crypt[idx]->ops->name, "WEP") == 0) idx 385 drivers/net/wireless/intersil/hostap/hostap_main.c len = local->crypt_info.crypt[idx]->ops->get_key(keybuf, sizeof(keybuf), NULL, idx 386 drivers/net/wireless/intersil/hostap/hostap_main.c local->crypt_info.crypt[idx]->priv); idx 387 drivers/net/wireless/intersil/hostap/hostap_main.c if (idx >= 0 && idx < WEP_KEYS && len > 5) idx 405 drivers/net/wireless/intersil/hostap/hostap_main.c if (hostap_set_word(local->dev, HFA384X_RID_CNFWEPDEFAULTKEYID, idx)) { idx 406 drivers/net/wireless/intersil/hostap/hostap_main.c printk(KERN_DEBUG "Could not set default keyid %d\n", idx); idx 21 drivers/net/wireless/intersil/hostap/hostap_wlan.h #define GET_INT_PARM(var,idx) var[var[idx] < 0 ? 0 : idx] idx 528 drivers/net/wireless/intersil/hostap/hostap_wlan.h u16 idx; idx 995 drivers/net/wireless/intersil/orinoco/hw.c __le16 idx; idx 1013 drivers/net/wireless/intersil/orinoco/hw.c buf.idx = cpu_to_le16(key_idx); idx 727 drivers/net/wireless/intersil/orinoco/wext.c int idx, alg = ext->alg, set_key = 1; idx 735 drivers/net/wireless/intersil/orinoco/wext.c idx = encoding->flags & IW_ENCODE_INDEX; idx 736 drivers/net/wireless/intersil/orinoco/wext.c if (idx) { idx 737 drivers/net/wireless/intersil/orinoco/wext.c if ((idx < 1) || (idx > 4)) idx 739 drivers/net/wireless/intersil/orinoco/wext.c idx--; idx 741 drivers/net/wireless/intersil/orinoco/wext.c idx = priv->tx_key; idx 752 drivers/net/wireless/intersil/orinoco/wext.c priv->tx_key = idx; idx 762 drivers/net/wireless/intersil/orinoco/wext.c err = orinoco_set_key(priv, idx, ORINOCO_ALG_NONE, idx 771 drivers/net/wireless/intersil/orinoco/wext.c err = orinoco_set_key(priv, idx, ORINOCO_ALG_WEP, idx 788 drivers/net/wireless/intersil/orinoco/wext.c err = orinoco_set_key(priv, idx, ORINOCO_ALG_TKIP, idx 792 drivers/net/wireless/intersil/orinoco/wext.c err = __orinoco_hw_set_tkip_key(priv, idx, idx 794 drivers/net/wireless/intersil/orinoco/wext.c priv->keys[idx].key, idx 821 drivers/net/wireless/intersil/orinoco/wext.c int idx, max_key_len; idx 833 drivers/net/wireless/intersil/orinoco/wext.c idx = encoding->flags & IW_ENCODE_INDEX; idx 834 drivers/net/wireless/intersil/orinoco/wext.c if (idx) { idx 835 drivers/net/wireless/intersil/orinoco/wext.c if ((idx < 1) || (idx > 4)) idx 837 drivers/net/wireless/intersil/orinoco/wext.c idx--; idx 839 drivers/net/wireless/intersil/orinoco/wext.c idx = priv->tx_key; idx 841 drivers/net/wireless/intersil/orinoco/wext.c encoding->flags = idx + 1; idx 852 drivers/net/wireless/intersil/orinoco/wext.c ext->key_len = min(priv->keys[idx].key_len, max_key_len); idx 853 drivers/net/wireless/intersil/orinoco/wext.c memcpy(ext->key, priv->keys[idx].key, ext->key_len); idx 858 drivers/net/wireless/intersil/orinoco/wext.c ext->key_len = min(priv->keys[idx].key_len, max_key_len); idx 859 drivers/net/wireless/intersil/orinoco/wext.c memcpy(ext->key, priv->keys[idx].key, ext->key_len); idx 652 drivers/net/wireless/intersil/p54/fwio.c int p54_upload_key(struct p54_common *priv, u8 algo, int slot, u8 idx, u8 len, idx 665 drivers/net/wireless/intersil/p54/fwio.c rxkey->key_id = idx; idx 549 drivers/net/wireless/intersil/p54/lmac.h u8 idx, u8 len, u8 *addr, u8* key); idx 603 drivers/net/wireless/intersil/p54/main.c static int p54_get_survey(struct ieee80211_hw *dev, int idx, idx 611 drivers/net/wireless/intersil/p54/main.c if (idx >= priv->chan_num) idx 617 drivers/net/wireless/intersil/p54/main.c if (chan && chan->hw_value == idx) { idx 627 drivers/net/wireless/intersil/p54/main.c memcpy(survey, &priv->survey[idx], sizeof(*survey)); idx 138 drivers/net/wireless/intersil/p54/p54pci.c u32 limit, idx, i; idx 140 drivers/net/wireless/intersil/p54/p54pci.c idx = le32_to_cpu(ring_control->host_idx[ring_index]); idx 141 drivers/net/wireless/intersil/p54/p54pci.c limit = idx; idx 145 drivers/net/wireless/intersil/p54/p54pci.c i = idx % ring_limit; idx 176 drivers/net/wireless/intersil/p54/p54pci.c idx++; idx 181 drivers/net/wireless/intersil/p54/p54pci.c ring_control->host_idx[ring_index] = cpu_to_le32(idx); idx 191 drivers/net/wireless/intersil/p54/p54pci.c u32 idx, i; idx 194 drivers/net/wireless/intersil/p54/p54pci.c (*index) = idx = le32_to_cpu(ring_control->device_idx[ring_index]); idx 195 drivers/net/wireless/intersil/p54/p54pci.c idx %= ring_limit; idx 196 drivers/net/wireless/intersil/p54/p54pci.c while (i != idx) { idx 249 drivers/net/wireless/intersil/p54/p54pci.c u32 idx, i; idx 252 drivers/net/wireless/intersil/p54/p54pci.c (*index) = idx = le32_to_cpu(ring_control->device_idx[ring_index]); idx 253 drivers/net/wireless/intersil/p54/p54pci.c idx %= ring_limit; idx 255 drivers/net/wireless/intersil/p54/p54pci.c while (i != idx) { idx 331 drivers/net/wireless/intersil/p54/p54pci.c u32 idx, i; idx 334 drivers/net/wireless/intersil/p54/p54pci.c idx = le32_to_cpu(ring_control->host_idx[1]); idx 335 drivers/net/wireless/intersil/p54/p54pci.c i = idx % ARRAY_SIZE(ring_control->tx_data); idx 354 drivers/net/wireless/intersil/p54/p54pci.c ring_control->host_idx[1] = cpu_to_le32(idx + 1); idx 408 drivers/net/wireless/intersil/p54/txrx.c int count, idx; idx 445 drivers/net/wireless/intersil/p54/txrx.c for (idx = 0; idx < 4; idx++) { idx 446 drivers/net/wireless/intersil/p54/txrx.c if (count >= info->status.rates[idx].count) { idx 447 drivers/net/wireless/intersil/p54/txrx.c count -= info->status.rates[idx].count; idx 449 drivers/net/wireless/intersil/p54/txrx.c info->status.rates[idx].count = count; idx 452 drivers/net/wireless/intersil/p54/txrx.c info->status.rates[idx].idx = -1; idx 453 drivers/net/wireless/intersil/p54/txrx.c info->status.rates[idx].count = 0; idx 842 drivers/net/wireless/intersil/p54/txrx.c if (info->control.rates[i].idx < 0) idx 875 drivers/net/wireless/intersil/p54/txrx.c rate = info->control.rates[i].idx; idx 1545 drivers/net/wireless/intersil/prism54/isl_ioctl.c int idx, alg = ext->alg, set_key = 1; idx 1554 drivers/net/wireless/intersil/prism54/isl_ioctl.c idx = (encoding->flags & IW_ENCODE_INDEX) - 1; idx 1555 drivers/net/wireless/intersil/prism54/isl_ioctl.c if (idx) { idx 1556 drivers/net/wireless/intersil/prism54/isl_ioctl.c if (idx < 0 || idx > 3) idx 1562 drivers/net/wireless/intersil/prism54/isl_ioctl.c idx = r.u; idx 1572 drivers/net/wireless/intersil/prism54/isl_ioctl.c ret = mgt_set_request(priv, DOT11_OID_DEFKEYID, 0, &idx); idx 1606 drivers/net/wireless/intersil/prism54/isl_ioctl.c ret = mgt_set_request(priv, DOT11_OID_DEFKEYX, idx, idx 1654 drivers/net/wireless/intersil/prism54/isl_ioctl.c int idx, max_key_len; idx 1676 drivers/net/wireless/intersil/prism54/isl_ioctl.c idx = (encoding->flags & IW_ENCODE_INDEX) - 1; idx 1677 drivers/net/wireless/intersil/prism54/isl_ioctl.c if (idx) { idx 1678 drivers/net/wireless/intersil/prism54/isl_ioctl.c if (idx < 0 || idx > 3) idx 1684 drivers/net/wireless/intersil/prism54/isl_ioctl.c idx = r.u; idx 1687 drivers/net/wireless/intersil/prism54/isl_ioctl.c encoding->flags = idx + 1; idx 1713 drivers/net/wireless/intersil/prism54/isl_ioctl.c ret = mgt_get_request(priv, DOT11_OID_DEFKEYX, idx, NULL, &r); idx 495 drivers/net/wireless/mac80211_hwsim.c int channels, idx; idx 1097 drivers/net/wireless/mac80211_hwsim.c tx_attempts[i].idx = info->status.rates[i].idx; idx 1098 drivers/net/wireless/mac80211_hwsim.c tx_attempts_flags[i].idx = info->status.rates[i].idx; idx 1238 drivers/net/wireless/mac80211_hwsim.c rx_status.rate_idx = info->control.rates[0].idx; idx 1441 drivers/net/wireless/mac80211_hwsim.c txi->control.rates[1].idx = -1; idx 1625 drivers/net/wireless/mac80211_hwsim.c int idx; idx 1652 drivers/net/wireless/mac80211_hwsim.c for (idx = 0; idx < ARRAY_SIZE(data->survey_data); idx++) { idx 1653 drivers/net/wireless/mac80211_hwsim.c if (data->survey_data[idx].channel == data->channel) { idx 1654 drivers/net/wireless/mac80211_hwsim.c data->survey_data[idx].start = idx 1655 drivers/net/wireless/mac80211_hwsim.c data->survey_data[idx].next_start; idx 1656 drivers/net/wireless/mac80211_hwsim.c data->survey_data[idx].end = jiffies; idx 1663 drivers/net/wireless/mac80211_hwsim.c for (idx = 0; idx < ARRAY_SIZE(data->survey_data); idx++) { idx 1664 drivers/net/wireless/mac80211_hwsim.c if (data->survey_data[idx].channel && idx 1665 drivers/net/wireless/mac80211_hwsim.c data->survey_data[idx].channel != data->channel) idx 1667 drivers/net/wireless/mac80211_hwsim.c data->survey_data[idx].channel = data->channel; idx 1668 drivers/net/wireless/mac80211_hwsim.c data->survey_data[idx].next_start = jiffies; idx 1861 drivers/net/wireless/mac80211_hwsim.c static int mac80211_hwsim_get_survey(struct ieee80211_hw *hw, int idx, idx 1866 drivers/net/wireless/mac80211_hwsim.c if (idx < 0 || idx >= ARRAY_SIZE(hwsim->survey_data)) idx 1870 drivers/net/wireless/mac80211_hwsim.c survey->channel = hwsim->survey_data[idx].channel; idx 1887 drivers/net/wireless/mac80211_hwsim.c jiffies_to_msecs(hwsim->survey_data[idx].end - idx 1888 drivers/net/wireless/mac80211_hwsim.c hwsim->survey_data[idx].start); idx 2738 drivers/net/wireless/mac80211_hwsim.c int idx, i; idx 2745 drivers/net/wireless/mac80211_hwsim.c idx = hwsim_radio_idx++; idx 2769 drivers/net/wireless/mac80211_hwsim.c data->dev = device_create(hwsim_class, NULL, 0, hw, "hwsim%d", idx); idx 2791 drivers/net/wireless/mac80211_hwsim.c addr[3] = idx >> 8; idx 2792 drivers/net/wireless/mac80211_hwsim.c addr[4] = idx; idx 2810 drivers/net/wireless/mac80211_hwsim.c data->idx = idx; idx 3072 drivers/net/wireless/mac80211_hwsim.c hwsim_mcast_new_radio(idx, info, param); idx 3074 drivers/net/wireless/mac80211_hwsim.c return idx; idx 3128 drivers/net/wireless/mac80211_hwsim.c hwsim_mcast_del_radio(data->idx, hwname, info); idx 3165 drivers/net/wireless/mac80211_hwsim.c res = append_radio_msg(skb, data->idx, ¶m); idx 3299 drivers/net/wireless/mac80211_hwsim.c txi->status.rates[i].idx = tx_attempts[i].idx; idx 3504 drivers/net/wireless/mac80211_hwsim.c u32 idx = nla_get_u32(info->attrs[HWSIM_ATTR_REG_CUSTOM_REG]); idx 3506 drivers/net/wireless/mac80211_hwsim.c if (idx >= ARRAY_SIZE(hwsim_world_regdom_custom)) idx 3509 drivers/net/wireless/mac80211_hwsim.c idx = array_index_nospec(idx, idx 3511 drivers/net/wireless/mac80211_hwsim.c param.regd = hwsim_world_regdom_custom[idx]; idx 3594 drivers/net/wireless/mac80211_hwsim.c s64 idx = -1; idx 3598 drivers/net/wireless/mac80211_hwsim.c idx = nla_get_u32(info->attrs[HWSIM_ATTR_RADIO_ID]); idx 3610 drivers/net/wireless/mac80211_hwsim.c if (idx >= 0) { idx 3611 drivers/net/wireless/mac80211_hwsim.c if (data->idx != idx) idx 3642 drivers/net/wireless/mac80211_hwsim.c int idx, res = -ENODEV; idx 3646 drivers/net/wireless/mac80211_hwsim.c idx = nla_get_u32(info->attrs[HWSIM_ATTR_RADIO_ID]); idx 3650 drivers/net/wireless/mac80211_hwsim.c if (data->idx != idx) idx 3694 drivers/net/wireless/mac80211_hwsim.c if (data->idx <= last_idx) idx 3707 drivers/net/wireless/mac80211_hwsim.c last_idx = data->idx; idx 182 drivers/net/wireless/mac80211_hwsim.h s8 idx; idx 245 drivers/net/wireless/mac80211_hwsim.h s8 idx; idx 1460 drivers/net/wireless/marvell/libertas/cfg.c u8 idx, bool pairwise, const u8 *mac_addr, idx 1474 drivers/net/wireless/marvell/libertas/cfg.c idx, params->key_len); idx 1488 drivers/net/wireless/marvell/libertas/cfg.c if ((priv->wep_key_len[idx] != params->key_len) || idx 1489 drivers/net/wireless/marvell/libertas/cfg.c memcmp(priv->wep_key[idx], idx 1491 drivers/net/wireless/marvell/libertas/cfg.c priv->wep_key_len[idx] = params->key_len; idx 1492 drivers/net/wireless/marvell/libertas/cfg.c memcpy(priv->wep_key[idx], idx 1499 drivers/net/wireless/marvell/libertas/cfg.c key_info = KEY_INFO_WPA_ENABLED | ((idx == 0) idx 76 drivers/net/wireless/marvell/libertas/main.c u32 lbs_fw_index_to_data_rate(u8 idx) idx 78 drivers/net/wireless/marvell/libertas/main.c if (idx >= sizeof(fw_data_rates)) idx 79 drivers/net/wireless/marvell/libertas/main.c idx = 0; idx 80 drivers/net/wireless/marvell/libertas/main.c return fw_data_rates[idx]; idx 459 drivers/net/wireless/marvell/libertas_tf/main.c static int lbtf_op_get_survey(struct ieee80211_hw *hw, int idx, idx 465 drivers/net/wireless/marvell/libertas_tf/main.c if (idx != 0) idx 1496 drivers/net/wireless/marvell/mwifiex/cfg80211.c int idx, u8 *mac, struct station_info *sinfo) idx 1502 drivers/net/wireless/marvell/mwifiex/cfg80211.c priv->media_connected && idx == 0) { idx 1526 drivers/net/wireless/marvell/mwifiex/cfg80211.c int idx, struct survey_info *survey) idx 1532 drivers/net/wireless/marvell/mwifiex/cfg80211.c mwifiex_dbg(priv->adapter, DUMP, "dump_survey idx=%d\n", idx); idx 1537 drivers/net/wireless/marvell/mwifiex/cfg80211.c priv->media_connected && idx == 0) { idx 1552 drivers/net/wireless/marvell/mwifiex/cfg80211.c if (idx >= priv->adapter->num_in_chan_stats) idx 1555 drivers/net/wireless/marvell/mwifiex/cfg80211.c if (!pchan_stats[idx].cca_scan_dur) idx 1558 drivers/net/wireless/marvell/mwifiex/cfg80211.c band = pchan_stats[idx].bandcfg; idx 1560 drivers/net/wireless/marvell/mwifiex/cfg80211.c ieee80211_channel_to_frequency(pchan_stats[idx].chan_num, band)); idx 1564 drivers/net/wireless/marvell/mwifiex/cfg80211.c survey->noise = pchan_stats[idx].noise; idx 1565 drivers/net/wireless/marvell/mwifiex/cfg80211.c survey->time = pchan_stats[idx].cca_scan_dur; idx 1566 drivers/net/wireless/marvell/mwifiex/cfg80211.c survey->time_busy = pchan_stats[idx].cca_busy_dur; idx 28 drivers/net/wireless/marvell/mwifiex/ie.c mwifiex_ie_index_used_by_other_intf(struct mwifiex_private *priv, u16 idx) idx 36 drivers/net/wireless/marvell/mwifiex/ie.c ie = &adapter->priv[i]->mgmt_ie[idx]; idx 1098 drivers/net/wireless/marvell/mwifiex/main.c int i, idx; idx 1158 drivers/net/wireless/marvell/mwifiex/main.c for (idx = 0; idx < priv->netdev->num_tx_queues; idx++) { idx 1159 drivers/net/wireless/marvell/mwifiex/main.c txq = netdev_get_tx_queue(priv->netdev, idx); idx 1160 drivers/net/wireless/marvell/mwifiex/main.c p += sprintf(p, "tx queue %d:%s ", idx, idx 1199 drivers/net/wireless/marvell/mwifiex/main.c u8 idx; idx 1203 drivers/net/wireless/marvell/mwifiex/main.c for (idx = 0; idx < adapter->num_mem_types; idx++) { idx 1205 drivers/net/wireless/marvell/mwifiex/main.c &adapter->mem_type_mapping_tbl[idx]; idx 1235 drivers/net/wireless/marvell/mwifiex/main.c for (idx = 0; idx < adapter->num_mem_types; idx++) { idx 1237 drivers/net/wireless/marvell/mwifiex/main.c &adapter->mem_type_mapping_tbl[idx]; idx 1259 drivers/net/wireless/marvell/mwifiex/main.c for (idx = 0; idx < adapter->num_mem_types; idx++) { idx 1261 drivers/net/wireless/marvell/mwifiex/main.c &adapter->mem_type_mapping_tbl[idx]; idx 2664 drivers/net/wireless/marvell/mwifiex/pcie.c u8 idx, i, read_reg, doneflag = 0; idx 2672 drivers/net/wireless/marvell/mwifiex/pcie.c for (idx = 0; idx < adapter->num_mem_types; idx++) { idx 2674 drivers/net/wireless/marvell/mwifiex/pcie.c &adapter->mem_type_mapping_tbl[idx]; idx 2700 drivers/net/wireless/marvell/mwifiex/pcie.c for (idx = 0; idx < dump_num; idx++) { idx 2702 drivers/net/wireless/marvell/mwifiex/pcie.c &adapter->mem_type_mapping_tbl[idx]; idx 1658 drivers/net/wireless/marvell/mwifiex/scan.c int idx; idx 1660 drivers/net/wireless/marvell/mwifiex/scan.c for (idx = 0; idx < ssid->ssid_len; idx++) { idx 1661 drivers/net/wireless/marvell/mwifiex/scan.c if (ssid->ssid[idx]) idx 2127 drivers/net/wireless/marvell/mwifiex/scan.c u32 idx; idx 2213 drivers/net/wireless/marvell/mwifiex/scan.c for (idx = 0; idx < scan_rsp->number_of_sets && bytes_left; idx++) { idx 2221 drivers/net/wireless/marvell/mwifiex/scan.c memcpy(&fw_tsf, &tsf_tlv->tsf_data[idx * TSF_DATA_SIZE], idx 2225 drivers/net/wireless/marvell/mwifiex/scan.c chan_band = &chan_band_tlv->chan_band_param[idx]; idx 2232 drivers/net/wireless/marvell/mwifiex/scan.c adapter->nd_info->matches[idx] = idx 2236 drivers/net/wireless/marvell/mwifiex/scan.c pmatch = adapter->nd_info->matches[idx]; idx 2626 drivers/net/wireless/marvell/mwifiex/scan.c u32 bytes_left, bytes_left_for_tlv, idx; idx 2657 drivers/net/wireless/marvell/mwifiex/scan.c for (idx = 0; idx < num_of_set && bytes_left; idx++) { idx 2305 drivers/net/wireless/marvell/mwifiex/sdio.c u8 *dbg_ptr, *end_ptr, dump_num, idx, i, read_reg, doneflag = 0; idx 2312 drivers/net/wireless/marvell/mwifiex/sdio.c for (idx = 0; idx < ARRAY_SIZE(mem_type_mapping_tbl); idx++) { idx 2313 drivers/net/wireless/marvell/mwifiex/sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; idx 2340 drivers/net/wireless/marvell/mwifiex/sdio.c for (idx = 0; idx < dump_num; idx++) { idx 2341 drivers/net/wireless/marvell/mwifiex/sdio.c struct memory_type_mapping *entry = &mem_type_mapping_tbl[idx]; idx 1640 drivers/net/wireless/marvell/mwifiex/sta_cmd.c u16 cnt, idx, length; idx 1660 drivers/net/wireless/marvell/mwifiex/sta_cmd.c for (idx = 0; idx < cfg->rule[cnt].num_of_fields; idx++) { idx 1661 drivers/net/wireless/marvell/mwifiex/sta_cmd.c param->operation = cfg->rule[cnt].params[idx].operation; idx 1663 drivers/net/wireless/marvell/mwifiex/sta_cmd.c cfg->rule[cnt].params[idx].operand_len; idx 1665 drivers/net/wireless/marvell/mwifiex/sta_cmd.c cpu_to_le16(cfg->rule[cnt].params[idx].offset); idx 1667 drivers/net/wireless/marvell/mwifiex/sta_cmd.c cfg->rule[cnt].params[idx].operand_byte_stream, idx 1027 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c int idx, freq, prev_freq = 0; idx 1038 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c for (idx = 0; idx < num_chan; idx++) { idx 1069 drivers/net/wireless/marvell/mwifiex/sta_cmdresp.c if (idx == 0 || prev_chflags != chflags || prev_bw != bw || idx 783 drivers/net/wireless/marvell/mwifiex/usb.c int idx; idx 785 drivers/net/wireless/marvell/mwifiex/usb.c for (idx = 0; idx < MWIFIEX_TX_DATA_PORT; idx++) { idx 786 drivers/net/wireless/marvell/mwifiex/usb.c if (priv->usb_port == card->port[idx].tx_data_ep) idx 787 drivers/net/wireless/marvell/mwifiex/usb.c return !card->port[idx].block_status; idx 1171 drivers/net/wireless/marvell/mwifiex/usb.c int idx, ret; idx 1190 drivers/net/wireless/marvell/mwifiex/usb.c for (idx = 0; idx < MWIFIEX_TX_DATA_PORT; idx++) { idx 1191 drivers/net/wireless/marvell/mwifiex/usb.c if (ep == card->port[idx].tx_data_ep) { idx 1192 drivers/net/wireless/marvell/mwifiex/usb.c port = &card->port[idx]; idx 1347 drivers/net/wireless/marvell/mwifiex/usb.c int idx; idx 1349 drivers/net/wireless/marvell/mwifiex/usb.c for (idx = 0; idx < MWIFIEX_TX_DATA_PORT; idx++) { idx 1350 drivers/net/wireless/marvell/mwifiex/usb.c port = &card->port[idx]; idx 178 drivers/net/wireless/marvell/mwl8k.c u8 idx; idx 1749 drivers/net/wireless/marvell/mwl8k.c info->status.rates[0].idx = -1; idx 1797 drivers/net/wireless/marvell/mwl8k.c stream->idx = i; idx 1992 drivers/net/wireless/marvell/mwl8k.c txpriority = (BA_QUEUE + stream->idx) % idx 1994 drivers/net/wireless/marvell/mwl8k.c if (stream->idx <= 1) idx 1995 drivers/net/wireless/marvell/mwl8k.c index = stream->idx + idx 3081 drivers/net/wireless/marvell/mwl8k.c int band, ch, idx = 0; idx 3088 drivers/net/wireless/marvell/mwl8k.c for (ch = 0; ch < sband->n_channels; ch++, idx++) idx 3094 drivers/net/wireless/marvell/mwl8k.c return idx; idx 3101 drivers/net/wireless/marvell/mwl8k.c s8 nf = 0, idx; idx 3104 drivers/net/wireless/marvell/mwl8k.c idx = freq_to_idx(priv, priv->acs_chan->center_freq); idx 3105 drivers/net/wireless/marvell/mwl8k.c if (idx >= MWL8K_NUM_CHANS) { idx 3110 drivers/net/wireless/marvell/mwl8k.c survey = &priv->survey[idx]; idx 3811 drivers/net/wireless/marvell/mwl8k.c u8 idx); idx 3982 drivers/net/wireless/marvell/mwl8k.c cmd->create_params.queue_id = stream->idx; idx 4017 drivers/net/wireless/marvell/mwl8k.c cmd->create_params.queue_id = stream->idx; idx 4044 drivers/net/wireless/marvell/mwl8k.c u8 idx) idx 4056 drivers/net/wireless/marvell/mwl8k.c cmd->destroy_params.ba_context = cpu_to_le32(idx); idx 4059 drivers/net/wireless/marvell/mwl8k.c wiphy_debug(hw->wiphy, "Deleted BA stream index %d\n", idx); idx 4158 drivers/net/wireless/marvell/mwl8k.c u8 idx; idx 4168 drivers/net/wireless/marvell/mwl8k.c idx = s->idx; idx 4170 drivers/net/wireless/marvell/mwl8k.c mwl8k_destroy_ba(hw, idx); idx 4332 drivers/net/wireless/marvell/mwl8k.c u8 idx; idx 4343 drivers/net/wireless/marvell/mwl8k.c idx = key->keyidx; idx 4353 drivers/net/wireless/marvell/mwl8k.c if (!mwl8k_vif->wep_key_conf[idx].enabled) { idx 4354 drivers/net/wireless/marvell/mwl8k.c memcpy(mwl8k_vif->wep_key_conf[idx].key, key, idx 4356 drivers/net/wireless/marvell/mwl8k.c mwl8k_vif->wep_key_conf[idx].enabled = 1; idx 5048 drivers/net/wireless/marvell/mwl8k.c int idx; idx 5053 drivers/net/wireless/marvell/mwl8k.c idx = ffs(vif->bss_conf.basic_rates); idx 5054 drivers/net/wireless/marvell/mwl8k.c if (idx) idx 5055 drivers/net/wireless/marvell/mwl8k.c idx--; idx 5059 drivers/net/wireless/marvell/mwl8k.c rate = mwl8k_rates_24[idx].hw_value; idx 5061 drivers/net/wireless/marvell/mwl8k.c rate = mwl8k_rates_50[idx].hw_value; idx 5120 drivers/net/wireless/marvell/mwl8k.c int idx; idx 5128 drivers/net/wireless/marvell/mwl8k.c idx = ffs(vif->bss_conf.basic_rates); idx 5129 drivers/net/wireless/marvell/mwl8k.c if (idx) idx 5130 drivers/net/wireless/marvell/mwl8k.c idx--; idx 5133 drivers/net/wireless/marvell/mwl8k.c rate = mwl8k_rates_24[idx].hw_value; idx 5135 drivers/net/wireless/marvell/mwl8k.c rate = mwl8k_rates_50[idx].hw_value; idx 5396 drivers/net/wireless/marvell/mwl8k.c static int mwl8k_get_survey(struct ieee80211_hw *hw, int idx, idx 5406 drivers/net/wireless/marvell/mwl8k.c if (sband && idx >= sband->n_channels) { idx 5407 drivers/net/wireless/marvell/mwl8k.c idx -= sband->n_channels; idx 5414 drivers/net/wireless/marvell/mwl8k.c if (!sband || idx >= sband->n_channels) idx 5417 drivers/net/wireless/marvell/mwl8k.c memcpy(survey, &priv->survey[idx], sizeof(*survey)); idx 5418 drivers/net/wireless/marvell/mwl8k.c survey->channel = &sband->channels[idx]; idx 5423 drivers/net/wireless/marvell/mwl8k.c if (idx != 0) idx 5447 drivers/net/wireless/marvell/mwl8k.c u8 *addr = sta->addr, idx; idx 5530 drivers/net/wireless/marvell/mwl8k.c idx = stream->idx; idx 5532 drivers/net/wireless/marvell/mwl8k.c mwl8k_destroy_ba(hw, idx); idx 5548 drivers/net/wireless/marvell/mwl8k.c idx = stream->idx; idx 5550 drivers/net/wireless/marvell/mwl8k.c mwl8k_destroy_ba(hw, idx); idx 10 drivers/net/wireless/mediatek/mt76/agg-rx.c mt76_aggr_release(struct mt76_rx_tid *tid, struct sk_buff_head *frames, int idx) idx 16 drivers/net/wireless/mediatek/mt76/agg-rx.c skb = tid->reorder_buf[idx]; idx 20 drivers/net/wireless/mediatek/mt76/agg-rx.c tid->reorder_buf[idx] = NULL; idx 30 drivers/net/wireless/mediatek/mt76/agg-rx.c int idx; idx 33 drivers/net/wireless/mediatek/mt76/agg-rx.c idx = tid->head % tid->size; idx 34 drivers/net/wireless/mediatek/mt76/agg-rx.c mt76_aggr_release(tid, frames, idx); idx 41 drivers/net/wireless/mediatek/mt76/agg-rx.c int idx = tid->head % tid->size; idx 43 drivers/net/wireless/mediatek/mt76/agg-rx.c while (tid->reorder_buf[idx]) { idx 44 drivers/net/wireless/mediatek/mt76/agg-rx.c mt76_aggr_release(tid, frames, idx); idx 45 drivers/net/wireless/mediatek/mt76/agg-rx.c idx = tid->head % tid->size; idx 54 drivers/net/wireless/mediatek/mt76/agg-rx.c int start, idx, nframes; idx 64 drivers/net/wireless/mediatek/mt76/agg-rx.c for (idx = (tid->head + 1) % tid->size; idx 65 drivers/net/wireless/mediatek/mt76/agg-rx.c idx != start && nframes; idx 66 drivers/net/wireless/mediatek/mt76/agg-rx.c idx = (idx + 1) % tid->size) { idx 67 drivers/net/wireless/mediatek/mt76/agg-rx.c skb = tid->reorder_buf[idx]; idx 147 drivers/net/wireless/mediatek/mt76/agg-rx.c u8 ackp, idx; idx 212 drivers/net/wireless/mediatek/mt76/agg-rx.c idx = seqno % size; idx 215 drivers/net/wireless/mediatek/mt76/agg-rx.c if (tid->reorder_buf[idx]) { idx 221 drivers/net/wireless/mediatek/mt76/agg-rx.c tid->reorder_buf[idx] = skb; idx 12 drivers/net/wireless/mediatek/mt76/dma.c int idx, int n_desc, int bufsize, idx 20 drivers/net/wireless/mediatek/mt76/dma.c q->regs = dev->mmio.regs + ring_base + idx * MT_RING_SIZE; idx 23 drivers/net/wireless/mediatek/mt76/dma.c q->hw_idx = idx; idx 54 drivers/net/wireless/mediatek/mt76/dma.c int i, idx = -1; idx 75 drivers/net/wireless/mediatek/mt76/dma.c idx = q->head; idx 78 drivers/net/wireless/mediatek/mt76/dma.c desc = &q->desc[idx]; idx 88 drivers/net/wireless/mediatek/mt76/dma.c q->entry[idx].txwi = txwi; idx 89 drivers/net/wireless/mediatek/mt76/dma.c q->entry[idx].skb = skb; idx 91 drivers/net/wireless/mediatek/mt76/dma.c return idx; idx 95 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_cleanup_idx(struct mt76_dev *dev, struct mt76_queue *q, int idx, idx 98 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue_entry *e = &q->entry[idx]; idx 99 drivers/net/wireless/mediatek/mt76/dma.c __le32 __ctrl = READ_ONCE(q->desc[idx].ctrl); idx 103 drivers/net/wireless/mediatek/mt76/dma.c __le32 addr = READ_ONCE(q->desc[idx].buf0); idx 111 drivers/net/wireless/mediatek/mt76/dma.c __le32 addr = READ_ONCE(q->desc[idx].buf1); idx 206 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_get_buf(struct mt76_dev *dev, struct mt76_queue *q, int idx, idx 209 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue_entry *e = &q->entry[idx]; idx 210 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_desc *desc = &q->desc[idx]; idx 235 drivers/net/wireless/mediatek/mt76/dma.c int idx = q->tail; idx 241 drivers/net/wireless/mediatek/mt76/dma.c if (!flush && !(q->desc[idx].ctrl & cpu_to_le32(MT_DMA_CTL_DMA_DONE))) idx 247 drivers/net/wireless/mediatek/mt76/dma.c return mt76_dma_get_buf(dev, q, idx, len, info, more); idx 368 drivers/net/wireless/mediatek/mt76/dma.c int idx; idx 387 drivers/net/wireless/mediatek/mt76/dma.c idx = mt76_dma_add_buf(dev, q, &qbuf, 1, 0, buf, NULL); idx 426 drivers/net/wireless/mediatek/mt76/mac80211.c int mt76_get_survey(struct ieee80211_hw *hw, int idx, idx 435 drivers/net/wireless/mediatek/mt76/mac80211.c if (idx == 0 && dev->drv->update_survey) idx 439 drivers/net/wireless/mediatek/mt76/mac80211.c if (idx >= sband->sband.n_channels) { idx 440 drivers/net/wireless/mediatek/mt76/mac80211.c idx -= sband->sband.n_channels; idx 444 drivers/net/wireless/mediatek/mt76/mac80211.c if (idx >= sband->sband.n_channels) idx 447 drivers/net/wireless/mediatek/mt76/mac80211.c chan = &sband->sband.channels[idx]; idx 688 drivers/net/wireless/mediatek/mt76/mac80211.c rcu_assign_pointer(dev->wcid[wcid->idx], wcid); idx 700 drivers/net/wireless/mediatek/mt76/mac80211.c int i, idx = wcid->idx; idx 702 drivers/net/wireless/mediatek/mt76/mac80211.c rcu_assign_pointer(dev->wcid[idx], NULL); idx 714 drivers/net/wireless/mediatek/mt76/mac80211.c mt76_wcid_free(dev->wcid_mask, idx); idx 851 drivers/net/wireless/mediatek/mt76/mac80211.c int idx, bool cck) idx 859 drivers/net/wireless/mediatek/mt76/mac80211.c idx &= ~BIT(2); /* short preamble */ idx 865 drivers/net/wireless/mediatek/mt76/mac80211.c if ((sband->bitrates[i].hw_value & GENMASK(7, 0)) == idx) idx 152 drivers/net/wireless/mediatek/mt76/mt76.h int idx, int n_desc, int bufsize, idx 201 drivers/net/wireless/mediatek/mt76/mt76.h u8 idx; idx 609 drivers/net/wireless/mediatek/mt76/mt76.h int idx; idx 616 drivers/net/wireless/mediatek/mt76/mt76.h idx = c - &msband->sband.channels[0]; idx 617 drivers/net/wireless/mediatek/mt76/mt76.h return &msband->chan[idx]; idx 721 drivers/net/wireless/mediatek/mt76/mt76.h int mt76_get_survey(struct ieee80211_hw *hw, int idx, idx 766 drivers/net/wireless/mediatek/mt76/mt76.h int idx, bool cck); idx 19 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c if (!(dev->mt76.beacon_mask & BIT(mvif->idx))) idx 30 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c FIELD_PREP(MT_DMA_FQCR0_TARGET_WCID, mvif->sta.wcid.idx) | idx 51 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c if (!(dev->mt76.beacon_mask & BIT(mvif->idx))) idx 63 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c data->tail[mvif->idx] = skb; idx 64 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c data->count[mvif->idx]++; idx 144 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c void mt7603_beacon_set_timer(struct mt7603_dev *dev, int idx, int intval) idx 148 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c if (idx >= 0) { idx 150 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c dev->mt76.beacon_mask |= BIT(idx); idx 152 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c dev->mt76.beacon_mask &= ~BIT(idx); idx 155 drivers/net/wireless/mediatek/mt76/mt7603/beacon.c if (!dev->mt76.beacon_mask || (!intval && idx < 0)) { idx 9 drivers/net/wireless/mediatek/mt76/mt7603/dma.c int idx, int n_desc) idx 18 drivers/net/wireless/mediatek/mt76/mt7603/dma.c err = mt76_queue_alloc(dev, hwq, idx, n_desc, 0, MT_TX_RING_BASE); idx 25 drivers/net/wireless/mediatek/mt76/mt7603/dma.c mt7603_irq_enable(dev, MT_INT_TX_DONE(idx)); idx 39 drivers/net/wireless/mediatek/mt76/mt7603/dma.c int idx; idx 47 drivers/net/wireless/mediatek/mt76/mt7603/dma.c idx = FIELD_GET(MT_TXD1_WLAN_IDX, val); idx 50 drivers/net/wireless/mediatek/mt76/mt7603/dma.c if (idx >= MT7603_WTBL_STA - 1) idx 53 drivers/net/wireless/mediatek/mt76/mt7603/dma.c wcid = rcu_dereference(dev->mt76.wcid[idx]); idx 124 drivers/net/wireless/mediatek/mt76/mt7603/dma.c int idx, int n_desc, int bufsize) idx 128 drivers/net/wireless/mediatek/mt76/mt7603/dma.c err = mt76_queue_alloc(dev, q, idx, n_desc, bufsize, idx 133 drivers/net/wireless/mediatek/mt76/mt7603/dma.c mt7603_irq_enable(dev, MT_INT_RX_DONE(idx)); idx 223 drivers/net/wireless/mediatek/mt76/mt7603/init.c dev->global_sta.wcid.idx = MT7603_WTBL_RESERVED; idx 75 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_update(struct mt7603_dev *dev, int idx, u32 mask) idx 78 drivers/net/wireless/mediatek/mt76/mt7603/mac.c FIELD_PREP(MT_WTBL_UPDATE_WLAN_IDX, idx) | mask); idx 84 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl1_addr(int idx) idx 86 drivers/net/wireless/mediatek/mt76/mt7603/mac.c return MT_WTBL1_BASE + idx * MT_WTBL1_SIZE; idx 90 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl2_addr(int idx) idx 93 drivers/net/wireless/mediatek/mt76/mt7603/mac.c return MT_PCIE_REMAP_BASE_1 + idx * MT_WTBL2_SIZE; idx 97 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl3_addr(int idx) idx 101 drivers/net/wireless/mediatek/mt76/mt7603/mac.c return base + idx * MT_WTBL3_SIZE; idx 105 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl4_addr(int idx) idx 109 drivers/net/wireless/mediatek/mt76/mt7603/mac.c return base + idx * MT_WTBL4_SIZE; idx 112 drivers/net/wireless/mediatek/mt76/mt7603/mac.c void mt7603_wtbl_init(struct mt7603_dev *dev, int idx, int vif, idx 116 drivers/net/wireless/mediatek/mt76/mt7603/mac.c u32 addr = mt7603_wtbl1_addr(idx); idx 140 drivers/net/wireless/mediatek/mt76/mt7603/mac.c addr = mt7603_wtbl2_addr(idx); idx 143 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_update(dev, idx, MT_WTBL_UPDATE_WTBL2); idx 146 drivers/net/wireless/mediatek/mt76/mt7603/mac.c addr = mt7603_wtbl3_addr(idx); idx 150 drivers/net/wireless/mediatek/mt76/mt7603/mac.c addr = mt7603_wtbl4_addr(idx); idx 156 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_set_skip_tx(struct mt7603_dev *dev, int idx, bool enabled) idx 158 drivers/net/wireless/mediatek/mt76/mt7603/mac.c u32 addr = mt7603_wtbl1_addr(idx); idx 167 drivers/net/wireless/mediatek/mt76/mt7603/mac.c void mt7603_filter_tx(struct mt7603_dev *dev, int idx, bool abort) idx 179 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_set_skip_tx(dev, idx, true); idx 182 drivers/net/wireless/mediatek/mt76/mt7603/mac.c FIELD_PREP(MT_TX_ABORT_WCID, idx)); idx 186 drivers/net/wireless/mediatek/mt76/mt7603/mac.c FIELD_PREP(MT_DMA_FQCR0_TARGET_WCID, idx) | idx 197 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_set_skip_tx(dev, idx, false); idx 203 drivers/net/wireless/mediatek/mt76/mt7603/mac.c u32 addr = mt7603_wtbl1_addr(sta->wcid.idx); idx 215 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int idx = sta->wcid.idx; idx 224 drivers/net/wireless/mediatek/mt76/mt7603/mac.c FIELD_PREP(MT_PSE_RTA_TAG_ID, idx) | idx 233 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_filter_tx(dev, idx, false); idx 235 drivers/net/wireless/mediatek/mt76/mt7603/mac.c addr = mt7603_wtbl1_addr(idx); idx 246 drivers/net/wireless/mediatek/mt76/mt7603/mac.c void mt7603_wtbl_clear(struct mt7603_dev *dev, int idx) idx 249 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int wtbl2_frame = idx / wtbl2_frame_size; idx 250 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int wtbl2_entry = idx % wtbl2_frame_size; idx 254 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int wtbl3_frame = wtbl3_base_frame + idx / wtbl3_frame_size; idx 255 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int wtbl3_entry = (idx % wtbl3_frame_size) * 2; idx 259 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int wtbl4_frame = wtbl4_base_frame + idx / wtbl4_frame_size; idx 260 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int wtbl4_entry = idx % wtbl4_frame_size; idx 262 drivers/net/wireless/mediatek/mt76/mt7603/mac.c u32 addr = MT_WTBL1_BASE + idx * MT_WTBL1_SIZE; idx 288 drivers/net/wireless/mediatek/mt76/mt7603/mac.c addr = mt7603_wtbl2_addr(idx); idx 296 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_update(dev, idx, MT_WTBL_UPDATE_WTBL2); idx 299 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_update(dev, idx, MT_WTBL_UPDATE_RX_COUNT_CLEAR); idx 300 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_update(dev, idx, MT_WTBL_UPDATE_TX_COUNT_CLEAR); idx 301 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_wtbl_update(dev, idx, MT_WTBL_UPDATE_ADM_COUNT_CLEAR); idx 307 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int idx = msta->wcid.idx; idx 311 drivers/net/wireless/mediatek/mt76/mt7603/mac.c addr = mt7603_wtbl1_addr(idx); idx 326 drivers/net/wireless/mediatek/mt76/mt7603/mac.c addr = mt7603_wtbl2_addr(idx); idx 374 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_rx_get_wcid(struct mt7603_dev *dev, u8 idx, bool unicast) idx 379 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (idx >= ARRAY_SIZE(dev->mt76.wcid)) idx 382 drivers/net/wireless/mediatek/mt76/mt7603/mac.c wcid = rcu_dereference(dev->mt76.wcid[idx]); idx 409 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int idx; idx 418 drivers/net/wireless/mediatek/mt76/mt7603/mac.c idx = FIELD_GET(MT_RXD2_NORMAL_WLAN_IDX, rxd2); idx 419 drivers/net/wireless/mediatek/mt76/mt7603/mac.c status->wcid = mt7603_rx_get_wcid(dev, idx, unicast); idx 558 drivers/net/wireless/mediatek/mt76/mt7603/mac.c rate_idx = rate->idx; idx 559 drivers/net/wireless/mediatek/mt76/mt7603/mac.c nss = 1 + (rate->idx >> 3); idx 571 drivers/net/wireless/mediatek/mt76/mt7603/mac.c r = &mt76_hw(dev)->wiphy->bands[band]->bitrates[rate->idx]; idx 595 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int wcid = sta->wcid.idx; idx 619 drivers/net/wireless/mediatek/mt76/mt7603/mac.c sta->rateset[rateset].probe_rate.idx = -1; idx 636 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (rates[i].idx != rates[k].idx) idx 642 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (!rates[i].idx) idx 645 drivers/net/wireless/mediatek/mt76/mt7603/mac.c rates[i].idx--; idx 802 drivers/net/wireless/mediatek/mt76/mt7603/mac.c vif_idx = mvif->idx; idx 814 drivers/net/wireless/mediatek/mt76/mt7603/mac.c wlan_idx = wcid->idx; idx 855 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (rate->idx >= 0 && rate->count && idx 964 drivers/net/wireless/mediatek/mt76/mt7603/mac.c int idx; idx 1009 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (!first_idx && rs->probe_rate.idx >= 0) { idx 1024 drivers/net/wireless/mediatek/mt76/mt7603/mac.c for (i = 0, idx = first_idx; count && idx <= last_idx; idx++) { idx 1028 drivers/net/wireless/mediatek/mt76/mt7603/mac.c cur_rate = &rs->rates[idx / 2]; idx 1032 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (idx && (cur_rate->idx != info->status.rates[i].idx || idx 1075 drivers/net/wireless/mediatek/mt76/mt7603/mac.c info->status.rates[i].idx = final_rate; idx 1099 drivers/net/wireless/mediatek/mt76/mt7603/mac.c info->status.rates[0].idx = -1; idx 41 drivers/net/wireless/mediatek/mt76/mt7603/main.c int idx; idx 46 drivers/net/wireless/mediatek/mt76/mt7603/main.c mvif->idx = ffs(~dev->vif_mask) - 1; idx 47 drivers/net/wireless/mediatek/mt76/mt7603/main.c if (mvif->idx >= MT7603_MAX_INTERFACES) { idx 52 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_MAC_ADDR0(mvif->idx), idx 54 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_MAC_ADDR1(mvif->idx), idx 59 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID0(mvif->idx), idx 61 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID1(mvif->idx), idx 66 drivers/net/wireless/mediatek/mt76/mt7603/main.c idx = MT7603_WTBL_RESERVED - 1 - mvif->idx; idx 67 drivers/net/wireless/mediatek/mt76/mt7603/main.c dev->vif_mask |= BIT(mvif->idx); idx 68 drivers/net/wireless/mediatek/mt76/mt7603/main.c mvif->sta.wcid.idx = idx; idx 72 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_wtbl_init(dev, idx, mvif->idx, bc_addr); idx 77 drivers/net/wireless/mediatek/mt76/mt7603/main.c rcu_assign_pointer(dev->mt76.wcid[idx], &mvif->sta.wcid); idx 90 drivers/net/wireless/mediatek/mt76/mt7603/main.c int idx = mvif->sta.wcid.idx; idx 92 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_MAC_ADDR0(mvif->idx), 0); idx 93 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_MAC_ADDR1(mvif->idx), 0); idx 94 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID0(mvif->idx), 0); idx 95 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID1(mvif->idx), 0); idx 96 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_beacon_set_timer(dev, mvif->idx, 0); idx 98 drivers/net/wireless/mediatek/mt76/mt7603/main.c rcu_assign_pointer(dev->mt76.wcid[idx], NULL); idx 102 drivers/net/wireless/mediatek/mt76/mt7603/main.c dev->vif_mask &= ~BIT(mvif->idx); idx 130 drivers/net/wireless/mediatek/mt76/mt7603/main.c int idx, ret; idx 156 drivers/net/wireless/mediatek/mt76/mt7603/main.c idx = 1; idx 159 drivers/net/wireless/mediatek/mt76/mt7603/main.c idx = 0; idx 165 drivers/net/wireless/mediatek/mt76/mt7603/main.c idx |= (def->chan - idx 167 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_WF_RMAC_CH_FREQ, idx); idx 282 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID0(mvif->idx), idx 284 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID1(mvif->idx), idx 288 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID0(mvif->idx), 0); idx 289 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt76_wr(dev, MT_BSSID1(mvif->idx), 0); idx 306 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_beacon_set_timer(dev, mvif->idx, beacon_int); idx 320 drivers/net/wireless/mediatek/mt76/mt7603/main.c int idx; idx 323 drivers/net/wireless/mediatek/mt76/mt7603/main.c idx = mt76_wcid_alloc(dev->mt76.wcid_mask, MT7603_WTBL_STA - 1); idx 324 drivers/net/wireless/mediatek/mt76/mt7603/main.c if (idx < 0) idx 331 drivers/net/wireless/mediatek/mt76/mt7603/main.c msta->wcid.idx = idx; idx 332 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_wtbl_init(dev, idx, mvif->idx, sta->addr); idx 360 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_filter_tx(dev, wcid->idx, true); idx 363 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_wtbl_clear(dev, wcid->idx); idx 458 drivers/net/wireless/mediatek/mt76/mt7603/main.c int idx = key->keyidx; idx 481 drivers/net/wireless/mediatek/mt76/mt7603/main.c key->hw_key_idx = wcid->idx; idx 482 drivers/net/wireless/mediatek/mt76/mt7603/main.c wcid->hw_key_idx = idx; idx 484 drivers/net/wireless/mediatek/mt76/mt7603/main.c if (idx == wcid->hw_key_idx) idx 491 drivers/net/wireless/mediatek/mt76/mt7603/main.c return mt7603_wtbl_set_key(dev, wcid->idx, key); idx 576 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_mac_tx_ba_reset(dev, msta->wcid.idx, tid, ba_size); idx 581 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_mac_tx_ba_reset(dev, msta->wcid.idx, tid, -1); idx 589 drivers/net/wireless/mediatek/mt76/mt7603/main.c mt7603_mac_tx_ba_reset(dev, msta->wcid.idx, tid, -1); idx 608 drivers/net/wireless/mediatek/mt76/mt7603/main.c msta->rates[i].idx = sta_rates->rate[i].idx; idx 612 drivers/net/wireless/mediatek/mt76/mt7603/main.c if (msta->rates[i].idx < 0 || !msta->rates[i].count) idx 83 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h u8 idx; idx 199 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h void mt7603_beacon_set_timer(struct mt7603_dev *dev, int idx, int intval); idx 212 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h void mt7603_wtbl_init(struct mt7603_dev *dev, int idx, int vif, idx 214 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h void mt7603_wtbl_clear(struct mt7603_dev *dev, int idx); idx 225 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h void mt7603_filter_tx(struct mt7603_dev *dev, int idx, bool abort); idx 425 drivers/net/wireless/mediatek/mt76/mt7603/regs.h #define MT_BSSID0(idx) MT_WF_RMAC(0x004 + (idx) * 8) idx 426 drivers/net/wireless/mediatek/mt76/mt7603/regs.h #define MT_BSSID1(idx) MT_WF_RMAC(0x008 + (idx) * 8) idx 429 drivers/net/wireless/mediatek/mt76/mt7603/regs.h #define MT_MAC_ADDR0(idx) MT_WF_RMAC(0x024 + (idx) * 8) idx 430 drivers/net/wireless/mediatek/mt76/mt7603/regs.h #define MT_MAC_ADDR1(idx) MT_WF_RMAC(0x028 + (idx) * 8) idx 40 drivers/net/wireless/mediatek/mt76/mt7615/dma.c int idx, int n_desc) idx 49 drivers/net/wireless/mediatek/mt76/mt7615/dma.c err = mt76_queue_alloc(dev, hwq, idx, n_desc, 0, MT_TX_RING_BASE); idx 92 drivers/net/wireless/mediatek/mt76/mt7615/init.c int ret, idx; idx 120 drivers/net/wireless/mediatek/mt76/mt7615/init.c idx = mt76_wcid_alloc(dev->mt76.wcid_mask, MT7615_WTBL_STA - 1); idx 121 drivers/net/wireless/mediatek/mt76/mt7615/init.c if (idx) idx 124 drivers/net/wireless/mediatek/mt76/mt7615/init.c dev->mt76.global_wcid.idx = idx; idx 126 drivers/net/wireless/mediatek/mt76/mt7615/init.c rcu_assign_pointer(dev->mt76.wcid[idx], &dev->mt76.global_wcid); idx 22 drivers/net/wireless/mediatek/mt76/mt7615/mac.c u8 idx, bool unicast) idx 27 drivers/net/wireless/mediatek/mt76/mt7615/mac.c if (idx >= ARRAY_SIZE(dev->mt76.wcid)) idx 30 drivers/net/wireless/mediatek/mt76/mt7615/mac.c wcid = rcu_dereference(dev->mt76.wcid[idx]); idx 54 drivers/net/wireless/mediatek/mt76/mt7615/mac.c int i, idx; idx 62 drivers/net/wireless/mediatek/mt76/mt7615/mac.c idx = FIELD_GET(MT_RXD2_NORMAL_WLAN_IDX, rxd2); idx 63 drivers/net/wireless/mediatek/mt76/mt7615/mac.c status->wcid = mt7615_rx_get_wcid(dev, idx, unicast); idx 270 drivers/net/wireless/mediatek/mt76/mt7615/mac.c rate_idx = rate->idx; idx 271 drivers/net/wireless/mediatek/mt76/mt7615/mac.c nss = 1 + (rate->idx >> 3); idx 283 drivers/net/wireless/mediatek/mt76/mt7615/mac.c r = &mt76_hw(dev)->wiphy->bands[band]->bitrates[rate->idx]; idx 355 drivers/net/wireless/mediatek/mt76/mt7615/mac.c FIELD_PREP(MT_TXD1_WLAN_IDX, wcid->idx) | idx 387 drivers/net/wireless/mediatek/mt76/mt7615/mac.c if (rate->idx >= 0 && rate->count && idx 468 drivers/net/wireless/mediatek/mt76/mt7615/mac.c int wcid = sta->wcid.idx; idx 492 drivers/net/wireless/mediatek/mt76/mt7615/mac.c sta->rateset[rateset].probe_rate.idx = -1; idx 509 drivers/net/wireless/mediatek/mt76/mt7615/mac.c if (rates[i].idx != rates[k].idx) idx 517 drivers/net/wireless/mediatek/mt76/mt7615/mac.c if (!rates[i].idx) idx 520 drivers/net/wireless/mediatek/mt76/mt7615/mac.c rates[i].idx--; idx 633 drivers/net/wireless/mediatek/mt76/mt7615/mac.c u32 addr = mt7615_mac_wtbl_addr(wcid->idx) + 30 * 4; idx 671 drivers/net/wireless/mediatek/mt76/mt7615/mac.c u32 addr = mt7615_mac_wtbl_addr(wcid->idx), w0, w1; idx 696 drivers/net/wireless/mediatek/mt76/mt7615/mac.c FIELD_PREP(MT_WTBL_UPDATE_WLAN_IDX, wcid->idx) | idx 710 drivers/net/wireless/mediatek/mt76/mt7615/mac.c u32 addr = mt7615_mac_wtbl_addr(wcid->idx); idx 816 drivers/net/wireless/mediatek/mt76/mt7615/mac.c txp->bss_idx = mvif->idx; idx 841 drivers/net/wireless/mediatek/mt76/mt7615/mac.c int i, idx, count; idx 892 drivers/net/wireless/mediatek/mt76/mt7615/mac.c if (!first_idx && rs->probe_rate.idx >= 0) { idx 906 drivers/net/wireless/mediatek/mt76/mt7615/mac.c for (i = 0, idx = first_idx; count && idx <= last_idx; idx++) { idx 910 drivers/net/wireless/mediatek/mt76/mt7615/mac.c cur_rate = &rs->rates[idx / 2]; idx 914 drivers/net/wireless/mediatek/mt76/mt7615/mac.c if (idx && (cur_rate->idx != info->status.rates[i].idx || idx 966 drivers/net/wireless/mediatek/mt76/mt7615/mac.c info->status.rates[i].idx = final_rate; idx 990 drivers/net/wireless/mediatek/mt76/mt7615/mac.c info->status.rates[0].idx = -1; idx 72 drivers/net/wireless/mediatek/mt76/mt7615/main.c int idx, ret = 0; idx 76 drivers/net/wireless/mediatek/mt76/mt7615/main.c mvif->idx = ffs(~dev->vif_mask) - 1; idx 77 drivers/net/wireless/mediatek/mt76/mt7615/main.c if (mvif->idx >= MT7615_MAX_INTERFACES) { idx 82 drivers/net/wireless/mediatek/mt76/mt7615/main.c idx = get_omac_idx(vif->type, dev->omac_mask); idx 83 drivers/net/wireless/mediatek/mt76/mt7615/main.c if (idx < 0) { idx 87 drivers/net/wireless/mediatek/mt76/mt7615/main.c mvif->omac_idx = idx; idx 91 drivers/net/wireless/mediatek/mt76/mt7615/main.c mvif->wmm_idx = mvif->idx % MT7615_MAX_WMM_SETS; idx 97 drivers/net/wireless/mediatek/mt76/mt7615/main.c dev->vif_mask |= BIT(mvif->idx); idx 99 drivers/net/wireless/mediatek/mt76/mt7615/main.c idx = MT7615_WTBL_RESERVED - mvif->idx; idx 100 drivers/net/wireless/mediatek/mt76/mt7615/main.c mvif->sta.wcid.idx = idx; idx 103 drivers/net/wireless/mediatek/mt76/mt7615/main.c rcu_assign_pointer(dev->mt76.wcid[idx], &mvif->sta.wcid); idx 119 drivers/net/wireless/mediatek/mt76/mt7615/main.c int idx = mvif->sta.wcid.idx; idx 125 drivers/net/wireless/mediatek/mt76/mt7615/main.c rcu_assign_pointer(dev->mt76.wcid[idx], NULL); idx 129 drivers/net/wireless/mediatek/mt76/mt7615/main.c dev->vif_mask &= ~BIT(mvif->idx); idx 176 drivers/net/wireless/mediatek/mt76/mt7615/main.c int idx = key->keyidx; idx 207 drivers/net/wireless/mediatek/mt76/mt7615/main.c key->hw_key_idx = wcid->idx; idx 208 drivers/net/wireless/mediatek/mt76/mt7615/main.c wcid->hw_key_idx = idx; idx 209 drivers/net/wireless/mediatek/mt76/mt7615/main.c } else if (idx == wcid->hw_key_idx) { idx 345 drivers/net/wireless/mediatek/mt76/mt7615/main.c int idx; idx 347 drivers/net/wireless/mediatek/mt76/mt7615/main.c idx = mt76_wcid_alloc(dev->mt76.wcid_mask, MT7615_WTBL_STA - 1); idx 348 drivers/net/wireless/mediatek/mt76/mt7615/main.c if (idx < 0) idx 353 drivers/net/wireless/mediatek/mt76/mt7615/main.c msta->wcid.idx = idx; idx 390 drivers/net/wireless/mediatek/mt76/mt7615/main.c msta->rates[i].idx = sta_rates->rate[i].idx; idx 394 drivers/net/wireless/mediatek/mt76/mt7615/main.c if (msta->rates[i].idx < 0 || !msta->rates[i].count) idx 749 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c u8 idx; idx 751 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c idx = mvif->omac_idx > EXT_BSSID_START ? HW_BSSID_0 : mvif->omac_idx; idx 754 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c hdr->hw_bss_idx = idx; idx 828 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c tx_wlan_idx = mvif->sta.wcid.idx; idx 845 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c tx_wlan_idx = msta->wcid.idx; idx 861 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c hdr->bss_idx = mvif->idx; idx 906 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = mvif->sta.wcid.idx, idx 936 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = mvif->sta.wcid.idx, idx 958 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = msta->wcid.idx, idx 988 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = msta->wcid.idx, idx 1015 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .bss_idx = mvif->idx, idx 1016 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = mvif->sta.wcid.idx, idx 1053 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .bss_idx = mvif->idx, idx 1054 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = msta->wcid.idx, idx 1118 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = wcid->idx, idx 1361 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c wtbl_hdr->wlan_idx = msta->wcid.idx; idx 1432 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c sta_hdr->bss_idx = mvif->idx; idx 1433 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c sta_hdr->wlan_idx = msta->wcid.idx; idx 1477 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = msta->wcid.idx, idx 1495 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .bss_idx = mvif->idx, idx 1496 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = msta->wcid.idx, idx 1515 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c u8 idx, ba_range[] = { 4, 8, 12, 24, 36, 48, 54, 64 }; idx 1517 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c for (idx = 7; idx > 0; idx--) { idx 1518 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c if (params->buf_size >= ba_range[idx]) idx 1522 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c wtbl_req.ba.ba_winsize_idx = idx; idx 1545 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = msta->wcid.idx, idx 1564 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .bss_idx = mvif->idx, idx 1565 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c .wlan_idx = msta->wcid.idx, idx 71 drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h u8 idx; idx 228 drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c int i, idx = 0; idx 246 drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c idx = (chan->hw_value == chan_map[i].chan); idx 280 drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c *tp = data >> (8 * idx); idx 239 drivers/net/wireless/mediatek/mt76/mt76x02.h mt76x02_rx_get_sta(struct mt76_dev *dev, u8 idx) idx 243 drivers/net/wireless/mediatek/mt76/mt76x02.h if (idx >= ARRAY_SIZE(dev->wcid)) idx 246 drivers/net/wireless/mediatek/mt76/mt76x02.h wcid = rcu_dereference(dev->wcid[idx]); idx 119 drivers/net/wireless/mediatek/mt76/mt76x02_beacon.c dev->mt76.beacon_mask |= BIT(mvif->idx); idx 121 drivers/net/wireless/mediatek/mt76/mt76x02_beacon.c dev->mt76.beacon_mask &= ~BIT(mvif->idx); idx 122 drivers/net/wireless/mediatek/mt76/mt76x02_beacon.c mt76x02_mac_set_beacon(dev, mvif->idx, NULL); idx 181 drivers/net/wireless/mediatek/mt76/mt76x02_beacon.c if (!(dev->mt76.beacon_mask & BIT(mvif->idx))) idx 188 drivers/net/wireless/mediatek/mt76/mt76x02_beacon.c mt76x02_mac_set_beacon(dev, mvif->idx, skb); idx 201 drivers/net/wireless/mediatek/mt76/mt76x02_beacon.c if (!(dev->mt76.beacon_mask & BIT(mvif->idx))) idx 213 drivers/net/wireless/mediatek/mt76/mt76x02_beacon.c data->tail[mvif->idx] = skb; idx 59 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c void mt76x02_mac_wcid_sync_pn(struct mt76x02_dev *dev, u8 idx, idx 68 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c iv = mt76_rr(dev, MT_WCID_IV(idx)); idx 69 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c eiv = mt76_rr(dev, MT_WCID_IV(idx) + 4); idx 84 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c int mt76x02_mac_wcid_set_key(struct mt76x02_dev *dev, u8 idx, idx 96 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_wr_copy(dev, MT_WCID_KEY(idx), key_data, sizeof(key_data)); idx 97 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_rmw_field(dev, MT_WCID_ATTR(idx), MT_WCID_ATTR_PKEY_MODE, cipher); idx 101 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_rmw_field(dev, MT_WCID_ATTR(idx), MT_WCID_ATTR_PAIRWISE, idx 121 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_wr_copy(dev, MT_WCID_IV(idx), iv_data, sizeof(iv_data)); idx 126 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c void mt76x02_mac_wcid_setup(struct mt76x02_dev *dev, u8 idx, idx 135 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_wr(dev, MT_WCID_ATTR(idx), attr); idx 137 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (idx >= 128) idx 143 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_wr_copy(dev, MT_WCID_ADDR(idx), &addr, sizeof(addr)); idx 147 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c void mt76x02_mac_wcid_set_drop(struct mt76x02_dev *dev, u8 idx, bool drop) idx 149 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c u32 val = mt76_rr(dev, MT_WCID_DROP(idx)); idx 150 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c u32 bit = MT_WCID_DROP_MASK(idx); idx 154 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_wr(dev, MT_WCID_DROP(idx), (val & ~bit) | (bit * drop)); idx 165 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c rate_idx = rate->idx; idx 166 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c nss = 1 + (rate->idx >> 4); idx 173 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c rate_idx = rate->idx; idx 174 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c nss = 1 + (rate->idx >> 3); idx 185 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c r = &dev->mt76.hw->wiphy->bands[band]->bitrates[rate->idx]; idx 260 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c u8 idx = FIELD_GET(MT_RXWI_RATE_INDEX, rate); idx 262 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c txrate->idx = 0; idx 269 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c idx += 4; idx 271 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c txrate->idx = idx; idx 274 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (idx >= 8) idx 275 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c idx -= 8; idx 277 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c txrate->idx = idx; idx 284 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c txrate->idx = idx; idx 288 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c txrate->idx = idx; idx 338 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c txwi->wcid = wcid->idx; idx 357 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (wcid && (rate->idx < 0 || !rate->count)) { idx 412 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76x02_tx_rate_fallback(struct ieee80211_tx_rate *rates, int idx, int phy) idx 416 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (!idx) idx 419 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c rates += idx - 1; idx 436 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (rates[0].idx == 8) { idx 437 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c rates[1].idx = 0; idx 442 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c rates[1].idx = max_t(int, rates[0].idx - 1, 0); idx 470 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c } else if (rate[0].idx < 0) { idx 490 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (info->status.rates[i].idx == last_rate.idx) idx 495 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c info->status.rates[i + 1].idx = -1; idx 597 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c u8 idx = FIELD_GET(MT_RXWI_RATE_INDEX, rate); idx 601 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (idx >= 8) idx 602 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c idx = 0; idx 605 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c idx += 4; idx 607 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c status->rate_idx = idx; idx 610 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (idx >= 8) { idx 611 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c idx -= 8; idx 615 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c if (idx >= 4) idx 616 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c idx = 0; idx 618 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c status->rate_idx = idx; idx 625 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c status->rate_idx = idx; idx 631 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c status->rate_idx = FIELD_GET(MT_RATE_INDEX_VHT_IDX, idx); idx 633 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c FIELD_GET(MT_RATE_INDEX_VHT_NSS, idx) + 1); idx 1093 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c int i, idx; idx 1098 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c for (i = 0, idx = 0; i < 16; i++) { idx 1101 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c dev->aggr_stats[idx++] += val & 0xffff; idx 1102 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c dev->aggr_stats[idx++] += val >> 16; idx 1119 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c void mt76x02_mac_set_bssid(struct mt76x02_dev *dev, u8 idx, const u8 *addr) idx 1121 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c idx &= 7; idx 1122 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_wr(dev, MT_MAC_APC_BSSID_L(idx), get_unaligned_le32(addr)); idx 1123 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_rmw_field(dev, MT_MAC_APC_BSSID_H(idx), MT_MAC_APC_BSSID_H_ADDR, idx 28 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h u8 idx; idx 167 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h int mt76x02_mac_wcid_set_key(struct mt76x02_dev *dev, u8 idx, idx 169 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h void mt76x02_mac_wcid_sync_pn(struct mt76x02_dev *dev, u8 idx, idx 171 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h void mt76x02_mac_wcid_setup(struct mt76x02_dev *dev, u8 idx, u8 vif_idx, idx 173 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h void mt76x02_mac_wcid_set_drop(struct mt76x02_dev *dev, u8 idx, bool drop); idx 195 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h void mt76x02_mac_set_bssid(struct mt76x02_dev *dev, u8 idx, const u8 *addr); idx 101 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c int idx, int n_desc) idx 110 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c err = mt76_queue_alloc(dev, hwq, idx, n_desc, 0, MT_TX_RING_BASE); idx 117 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c mt76x02_irq_enable(dev, MT_INT_TX_DONE(idx)); idx 124 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c int idx, int n_desc, int bufsize) idx 128 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c err = mt76_queue_alloc(dev, q, idx, n_desc, bufsize, idx 133 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c mt76x02_irq_enable(dev, MT_INT_RX_DONE(idx)); idx 395 drivers/net/wireless/mediatek/mt76/mt76x02_mmio.c mt76x02_mac_wcid_sync_pn(dev, wcid->idx, key); idx 68 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c u8 nss, idx; idx 71 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c idx = ((nss - 1) << 3) + mcs; idx 72 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c max_txpwr = dev->mt76.rate_power.ht[idx & 0xf]; idx 75 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c max_txpwr = dev->mt76.rate_power.ht[rate->idx & 0xf]; idx 84 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c r = &wiphy->bands[band]->bitrates[rate->idx]; idx 90 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c max_txpwr = dev->mt76.rate_power.ofdm[rate->idx & 0x7]; idx 149 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c if (qid == MT_TXQ_PSD && wcid && wcid->idx < 128) idx 150 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c mt76x02_mac_wcid_set_drop(dev, wcid->idx, false); idx 195 drivers/net/wireless/mediatek/mt76/mt76x02_util.c dev->mt76.global_wcid.idx = 255; idx 253 drivers/net/wireless/mediatek/mt76/mt76x02_util.c int idx = 0; idx 257 drivers/net/wireless/mediatek/mt76/mt76x02_util.c idx = mt76_wcid_alloc(dev->mt76.wcid_mask, ARRAY_SIZE(dev->mt76.wcid)); idx 258 drivers/net/wireless/mediatek/mt76/mt76x02_util.c if (idx < 0) idx 263 drivers/net/wireless/mediatek/mt76/mt76x02_util.c msta->wcid.idx = idx; idx 265 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76x02_mac_wcid_setup(dev, idx, mvif->idx, sta->addr); idx 266 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76x02_mac_wcid_set_drop(dev, idx, false); idx 280 drivers/net/wireless/mediatek/mt76/mt76x02_util.c int idx = wcid->idx; idx 282 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76x02_mac_wcid_set_drop(dev, idx, true); idx 283 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76x02_mac_wcid_setup(dev, idx, 0, NULL); idx 289 drivers/net/wireless/mediatek/mt76/mt76x02_util.c unsigned int idx) idx 296 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mvif->idx = idx; idx 297 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mvif->group_wcid.idx = MT_VIF_WCID(idx); idx 309 drivers/net/wireless/mediatek/mt76/mt76x02_util.c unsigned int idx = 0; idx 318 drivers/net/wireless/mediatek/mt76/mt76x02_util.c idx = 1 + (((dev->mt76.macaddr[0] ^ vif->addr[0]) >> 2) & 7); idx 334 drivers/net/wireless/mediatek/mt76/mt76x02_util.c idx += 8; idx 336 drivers/net/wireless/mediatek/mt76/mt76x02_util.c if (dev->vif_mask & BIT(idx)) idx 339 drivers/net/wireless/mediatek/mt76/mt76x02_util.c dev->vif_mask |= BIT(idx); idx 341 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76x02_vif_init(dev, vif, idx); idx 353 drivers/net/wireless/mediatek/mt76/mt76x02_util.c dev->vif_mask &= ~BIT(mvif->idx); idx 378 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76_set(dev, MT_WCID_ADDR(msta->wcid.idx) + 4, BIT(16 + tid)); idx 382 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76_clear(dev, MT_WCID_ADDR(msta->wcid.idx) + 4, idx 416 drivers/net/wireless/mediatek/mt76/mt76x02_util.c int idx = key->keyidx; idx 455 drivers/net/wireless/mediatek/mt76/mt76x02_util.c key->hw_key_idx = wcid->idx; idx 456 drivers/net/wireless/mediatek/mt76/mt76x02_util.c wcid->hw_key_idx = idx; idx 462 drivers/net/wireless/mediatek/mt76/mt76x02_util.c if (idx == wcid->hw_key_idx) { idx 472 drivers/net/wireless/mediatek/mt76/mt76x02_util.c if (key || wcid->hw_key_idx == idx) { idx 473 drivers/net/wireless/mediatek/mt76/mt76x02_util.c ret = mt76x02_mac_wcid_set_key(dev, wcid->idx, key); idx 478 drivers/net/wireless/mediatek/mt76/mt76x02_util.c return mt76x02_mac_shared_key_setup(dev, mvif->idx, idx, key); idx 481 drivers/net/wireless/mediatek/mt76/mt76x02_util.c return mt76x02_mac_wcid_set_key(dev, msta->wcid.idx, key); idx 586 drivers/net/wireless/mediatek/mt76/mt76x02_util.c rate.idx = rates->rate[0].idx; idx 624 drivers/net/wireless/mediatek/mt76/mt76x02_util.c int idx = msta->wcid.idx; idx 628 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76x02_mac_wcid_set_drop(dev, idx, ps); idx 643 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76x02_mac_set_bssid(dev, mvif->idx, info->bssid); idx 19 drivers/net/wireless/mediatek/mt76/mt76x2/mcu.c u8 idx; idx 29 drivers/net/wireless/mediatek/mt76/mt76x2/mcu.c .idx = channel, idx 141 drivers/net/wireless/mediatek/mt76/tx.c info->status.rates[0].idx = -1; idx 183 drivers/net/wireless/mediatek/mt76/tx.c cb->wcid = wcid->idx; idx 203 drivers/net/wireless/mediatek/mt76/tx.c if (wcid && cb->wcid != wcid->idx) idx 393 drivers/net/wireless/mediatek/mt76/tx.c int idx; idx 419 drivers/net/wireless/mediatek/mt76/tx.c idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, txq->sta); idx 421 drivers/net/wireless/mediatek/mt76/tx.c if (idx < 0) idx 422 drivers/net/wireless/mediatek/mt76/tx.c return idx; idx 453 drivers/net/wireless/mediatek/mt76/tx.c idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, idx 455 drivers/net/wireless/mediatek/mt76/tx.c if (idx < 0) idx 456 drivers/net/wireless/mediatek/mt76/tx.c return idx; idx 462 drivers/net/wireless/mediatek/mt76/tx.c hwq->entry[idx].qid = sq - dev->q_tx; idx 463 drivers/net/wireless/mediatek/mt76/tx.c hwq->entry[idx].schedule = true; idx 776 drivers/net/wireless/mediatek/mt76/usb.c u16 idx = q->tail; idx 787 drivers/net/wireless/mediatek/mt76/usb.c err = mt76u_tx_setup_buffers(dev, tx_info.skb, q->entry[idx].urb); idx 792 drivers/net/wireless/mediatek/mt76/usb.c q->entry[idx].urb, mt76u_complete_tx, idx 793 drivers/net/wireless/mediatek/mt76/usb.c &q->entry[idx]); idx 796 drivers/net/wireless/mediatek/mt76/usb.c q->entry[idx].skb = tx_info.skb; idx 799 drivers/net/wireless/mediatek/mt76/usb.c return idx; idx 47 drivers/net/wireless/mediatek/mt76/util.c int i, idx = 0, cur; idx 50 drivers/net/wireless/mediatek/mt76/util.c idx = ffs(~mask[i]); idx 51 drivers/net/wireless/mediatek/mt76/util.c if (!idx) idx 54 drivers/net/wireless/mediatek/mt76/util.c idx--; idx 55 drivers/net/wireless/mediatek/mt76/util.c cur = i * BITS_PER_LONG + idx; idx 59 drivers/net/wireless/mediatek/mt76/util.c mask[i] |= BIT(idx); idx 20 drivers/net/wireless/mediatek/mt76/util.h mt76_wcid_free(unsigned long *mask, int idx) idx 22 drivers/net/wireless/mediatek/mt76/util.h mask[idx / BITS_PER_LONG] &= ~BIT(idx % BITS_PER_LONG); idx 182 drivers/net/wireless/mediatek/mt7601u/eeprom.c int idx = -1; idx 185 drivers/net/wireless/mediatek/mt7601u/eeprom.c idx = val; idx 187 drivers/net/wireless/mediatek/mt7601u/eeprom.c idx = val - 32 + 8; idx 189 drivers/net/wireless/mediatek/mt7601u/eeprom.c if (idx != -1) idx 192 drivers/net/wireless/mediatek/mt7601u/eeprom.c val, chan_bounds[idx].start, idx 193 drivers/net/wireless/mediatek/mt7601u/eeprom.c chan_bounds[idx].start + chan_bounds[idx].num - 1); idx 195 drivers/net/wireless/mediatek/mt7601u/eeprom.c idx = 5; /* channels 1 - 14 */ idx 197 drivers/net/wireless/mediatek/mt7601u/eeprom.c dev->ee->reg = chan_bounds[idx]; idx 590 drivers/net/wireless/mediatek/mt7601u/init.c dev->mon_wcid->idx = 0xff; idx 30 drivers/net/wireless/mediatek/mt7601u/mac.c u8 idx = FIELD_GET(MT_TXWI_RATE_MCS, rate); idx 32 drivers/net/wireless/mediatek/mt7601u/mac.c txrate->idx = 0; idx 38 drivers/net/wireless/mediatek/mt7601u/mac.c txrate->idx = idx + 4; idx 41 drivers/net/wireless/mediatek/mt7601u/mac.c if (idx >= 8) idx 42 drivers/net/wireless/mediatek/mt7601u/mac.c idx -= 8; idx 44 drivers/net/wireless/mediatek/mt7601u/mac.c txrate->idx = idx; idx 51 drivers/net/wireless/mediatek/mt7601u/mac.c txrate->idx = idx; idx 76 drivers/net/wireless/mediatek/mt7601u/mac.c rate[last_rate + 1].idx = -1; idx 78 drivers/net/wireless/mediatek/mt7601u/mac.c cur_idx = rate[last_rate].idx + st->retry; idx 81 drivers/net/wireless/mediatek/mt7601u/mac.c rate[i].idx = max_t(int, 0, cur_idx - i); idx 113 drivers/net/wireless/mediatek/mt7601u/mac.c rate_idx = rate->idx; idx 114 drivers/net/wireless/mediatek/mt7601u/mac.c nss = 1 + (rate->idx >> 3); idx 125 drivers/net/wireless/mediatek/mt7601u/mac.c r = &dev->hw->wiphy->bands[band]->bitrates[rate->idx]; idx 355 drivers/net/wireless/mediatek/mt7601u/mac.c mt7601u_mac_wcid_setup(struct mt7601u_dev *dev, u8 idx, u8 vif_idx, u8 *mac) idx 363 drivers/net/wireless/mediatek/mt7601u/mac.c mt76_wr(dev, MT_WCID_ATTR(idx), attr); idx 368 drivers/net/wireless/mediatek/mt7601u/mac.c mt7601u_addr_wr(dev, MT_WCID_ADDR(idx), zmac); idx 399 drivers/net/wireless/mediatek/mt7601u/mac.c u8 idx = FIELD_GET(MT_RXWI_RATE_MCS, rate); idx 403 drivers/net/wireless/mediatek/mt7601u/mac.c if (WARN_ON(idx >= 8)) idx 404 drivers/net/wireless/mediatek/mt7601u/mac.c idx = 0; idx 405 drivers/net/wireless/mediatek/mt7601u/mac.c idx += 4; idx 407 drivers/net/wireless/mediatek/mt7601u/mac.c status->rate_idx = idx; idx 410 drivers/net/wireless/mediatek/mt7601u/mac.c if (idx >= 8) { idx 411 drivers/net/wireless/mediatek/mt7601u/mac.c idx -= 8; idx 415 drivers/net/wireless/mediatek/mt7601u/mac.c if (WARN_ON(idx >= 4)) idx 416 drivers/net/wireless/mediatek/mt7601u/mac.c idx = 0; idx 418 drivers/net/wireless/mediatek/mt7601u/mac.c status->rate_idx = idx; idx 425 drivers/net/wireless/mediatek/mt7601u/mac.c status->rate_idx = idx; idx 530 drivers/net/wireless/mediatek/mt7601u/mac.c int mt76_mac_wcid_set_key(struct mt7601u_dev *dev, u8 idx, idx 542 drivers/net/wireless/mediatek/mt7601u/mac.c trace_set_key(dev, idx); idx 544 drivers/net/wireless/mediatek/mt7601u/mac.c mt7601u_wr_copy(dev, MT_WCID_KEY(idx), key_data, sizeof(key_data)); idx 557 drivers/net/wireless/mediatek/mt7601u/mac.c mt7601u_wr_copy(dev, MT_WCID_IV(idx), iv_data, sizeof(iv_data)); idx 559 drivers/net/wireless/mediatek/mt7601u/mac.c val = mt7601u_rr(dev, MT_WCID_ATTR(idx)); idx 566 drivers/net/wireless/mediatek/mt7601u/mac.c mt7601u_wr(dev, MT_WCID_ATTR(idx), val); idx 157 drivers/net/wireless/mediatek/mt7601u/mac.h int mt76_mac_wcid_set_key(struct mt7601u_dev *dev, u8 idx, idx 49 drivers/net/wireless/mediatek/mt7601u/main.c unsigned int idx = 0; idx 50 drivers/net/wireless/mediatek/mt7601u/main.c unsigned int wcid = GROUP_WCID(idx); idx 57 drivers/net/wireless/mediatek/mt7601u/main.c mvif->idx = idx; idx 65 drivers/net/wireless/mediatek/mt7601u/main.c mvif->group_wcid.idx = wcid; idx 76 drivers/net/wireless/mediatek/mt7601u/main.c unsigned int wcid = mvif->group_wcid.idx; idx 189 drivers/net/wireless/mediatek/mt7601u/main.c int i, idx = 0; idx 192 drivers/net/wireless/mediatek/mt7601u/main.c idx = ffs(~dev->wcid_mask[i]); idx 193 drivers/net/wireless/mediatek/mt7601u/main.c if (!idx) idx 196 drivers/net/wireless/mediatek/mt7601u/main.c idx--; idx 197 drivers/net/wireless/mediatek/mt7601u/main.c dev->wcid_mask[i] |= BIT(idx); idx 201 drivers/net/wireless/mediatek/mt7601u/main.c idx = i * BITS_PER_LONG + idx; idx 202 drivers/net/wireless/mediatek/mt7601u/main.c if (idx > 119) idx 205 drivers/net/wireless/mediatek/mt7601u/main.c return idx; idx 216 drivers/net/wireless/mediatek/mt7601u/main.c int idx = 0; idx 220 drivers/net/wireless/mediatek/mt7601u/main.c idx = mt76_wcid_alloc(dev); idx 221 drivers/net/wireless/mediatek/mt7601u/main.c if (idx < 0) { idx 226 drivers/net/wireless/mediatek/mt7601u/main.c msta->wcid.idx = idx; idx 228 drivers/net/wireless/mediatek/mt7601u/main.c mt7601u_mac_wcid_setup(dev, idx, mvif->idx, sta->addr); idx 229 drivers/net/wireless/mediatek/mt7601u/main.c mt76_clear(dev, MT_WCID_DROP(idx), MT_WCID_DROP_MASK(idx)); idx 230 drivers/net/wireless/mediatek/mt7601u/main.c rcu_assign_pointer(dev->wcid[idx], &msta->wcid); idx 245 drivers/net/wireless/mediatek/mt7601u/main.c int idx = msta->wcid.idx; idx 248 drivers/net/wireless/mediatek/mt7601u/main.c rcu_assign_pointer(dev->wcid[idx], NULL); idx 249 drivers/net/wireless/mediatek/mt7601u/main.c mt76_set(dev, MT_WCID_DROP(idx), MT_WCID_DROP_MASK(idx)); idx 250 drivers/net/wireless/mediatek/mt7601u/main.c dev->wcid_mask[idx / BITS_PER_LONG] &= ~BIT(idx % BITS_PER_LONG); idx 251 drivers/net/wireless/mediatek/mt7601u/main.c mt7601u_mac_wcid_setup(dev, idx, 0, NULL); idx 300 drivers/net/wireless/mediatek/mt7601u/main.c int idx = key->keyidx; idx 315 drivers/net/wireless/mediatek/mt7601u/main.c key->hw_key_idx = wcid->idx; idx 316 drivers/net/wireless/mediatek/mt7601u/main.c wcid->hw_key_idx = idx; idx 318 drivers/net/wireless/mediatek/mt7601u/main.c if (idx == wcid->hw_key_idx) idx 325 drivers/net/wireless/mediatek/mt7601u/main.c if (key || wcid->hw_key_idx == idx) { idx 326 drivers/net/wireless/mediatek/mt7601u/main.c ret = mt76_mac_wcid_set_key(dev, wcid->idx, key); idx 331 drivers/net/wireless/mediatek/mt7601u/main.c return mt76_mac_shared_key_setup(dev, mvif->idx, idx, key); idx 334 drivers/net/wireless/mediatek/mt7601u/main.c return mt76_mac_wcid_set_key(dev, msta->wcid.idx, key); idx 357 drivers/net/wireless/mediatek/mt7601u/main.c WARN_ON(msta->wcid.idx > GROUP_WCID(0)); idx 361 drivers/net/wireless/mediatek/mt7601u/main.c mt76_set(dev, MT_WCID_ADDR(msta->wcid.idx) + 4, BIT(16 + tid)); idx 364 drivers/net/wireless/mediatek/mt7601u/main.c mt76_clear(dev, MT_WCID_ADDR(msta->wcid.idx) + 4, idx 400 drivers/net/wireless/mediatek/mt7601u/main.c rate.idx = rates->rate[0].idx; idx 106 drivers/net/wireless/mediatek/mt7601u/mt7601u.h #define GROUP_WCID(idx) (N_WCIDS - 2 - idx) idx 253 drivers/net/wireless/mediatek/mt7601u/mt7601u.h u8 idx; idx 262 drivers/net/wireless/mediatek/mt7601u/mt7601u.h u8 idx; idx 364 drivers/net/wireless/mediatek/mt7601u/mt7601u.h mt7601u_mac_wcid_setup(struct mt7601u_dev *dev, u8 idx, u8 vif_idx, u8 *mac); idx 109 drivers/net/wireless/mediatek/mt7601u/tx.c info->status.rates[0].idx = -1; idx 151 drivers/net/wireless/mediatek/mt7601u/tx.c if (rate->idx < 0 || !rate->count) idx 180 drivers/net/wireless/mediatek/mt7601u/tx.c txwi->wcid = wcid->idx; idx 470 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c int idx, u8 *mac, struct station_info *sinfo) idx 478 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c if (idx != 0 || !vif->wdev.current_bss) idx 484 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c sta_node = qtnf_sta_list_lookup_index(&vif->sta_list, idx); idx 728 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c int idx, struct survey_info *survey) idx 739 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c if (sband && idx >= sband->n_channels) { idx 740 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c idx -= sband->n_channels; idx 747 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c if (!sband || idx >= sband->n_channels) idx 750 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c chan = &sband->channels[idx]; idx 159 drivers/net/wireless/quantenna/qtnfmac/qlink_util.c unsigned int idx = bit / BITS_PER_BYTE; idx 160 drivers/net/wireless/quantenna/qtnfmac/qlink_util.c u8 mask = 1 << (bit - (idx * BITS_PER_BYTE)); idx 162 drivers/net/wireless/quantenna/qtnfmac/qlink_util.c if (idx >= arr_max_len) idx 165 drivers/net/wireless/quantenna/qtnfmac/qlink_util.c return arr[idx] & mask; idx 926 drivers/net/wireless/ralink/rt2x00/rt2800lib.c u8 idx = rt2x00_get_field32(status, TX_STA_FIFO_MCS); idx 937 drivers/net/wireless/ralink/rt2x00/rt2800lib.c idx += 4; idx 940 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (idx >= 8) idx 941 drivers/net/wireless/ralink/rt2x00/rt2800lib.c idx -= 8; idx 951 drivers/net/wireless/ralink/rt2x00/rt2800lib.c skbdesc->tx_rate_idx = idx; idx 3262 drivers/net/wireless/ralink/rt2x00/rt2800lib.c int idx = rf->channel-1; idx 3311 drivers/net/wireless/ralink/rt2x00/rt2800lib.c r55_bt_rev[idx]); idx 3313 drivers/net/wireless/ralink/rt2x00/rt2800lib.c r59_bt_rev[idx]); idx 3319 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2800_rfcsr_write(rt2x00dev, 59, r59_bt[idx]); idx 3331 drivers/net/wireless/ralink/rt2x00/rt2800lib.c r55_nonbt_rev[idx]); idx 3333 drivers/net/wireless/ralink/rt2x00/rt2800lib.c r59_nonbt_rev[idx]); idx 3342 drivers/net/wireless/ralink/rt2x00/rt2800lib.c r59_non_bt[idx]); idx 3349 drivers/net/wireless/ralink/rt2x00/rt2800lib.c r59_non_bt[idx]); idx 10507 drivers/net/wireless/ralink/rt2x00/rt2800lib.c int rt2800_get_survey(struct ieee80211_hw *hw, int idx, idx 10514 drivers/net/wireless/ralink/rt2x00/rt2800lib.c if (idx != 0) idx 253 drivers/net/wireless/ralink/rt2x00/rt2800lib.h int rt2800_get_survey(struct ieee80211_hw *hw, int idx, idx 31 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c int idx, qid; idx 39 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c idx = rt2x00mmio_register_read(rt2x00dev, TX_DTX_IDX(qid)); idx 42 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c idx = rt2x00mmio_register_read(rt2x00dev, TX_DTX_IDX(5)); idx 46 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c idx = entry->entry_idx; idx 50 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c idx = 0; idx 54 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c return idx; idx 332 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c tx_info->status.rates[i].idx = rate_idx - i; idx 347 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c tx_info->status.rates[i].idx = -1; /* terminate */ idx 314 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.mcs = txrate->idx; idx 624 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rate_idx = tx_info->control.rates[0].idx; idx 303 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c int idx, bb; idx 305 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c idx = (agc & 0x60) >> 5; idx 308 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c signal = 4 - bb - rtl8187se_lna_gain[idx]; idx 348 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_tx_desc *entry = &ring->desc[ring->idx]; idx 356 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c ring->idx = (ring->idx + 1) % ring->entries; idx 464 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c unsigned int idx, prio, hw_prio; idx 544 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c idx = (ring->idx + skb_queue_len(&ring->queue)) % ring->entries; idx 545 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c entry = &ring->desc[idx]; idx 1086 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c priv->tx_ring[prio].idx = 0; idx 1103 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c struct rtl8180_tx_desc *entry = &ring->desc[ring->idx]; idx 1109 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c ring->idx = (ring->idx + 1) % ring->entries; idx 91 drivers/net/wireless/realtek/rtl818x/rtl8180/rtl8180.h unsigned int idx; idx 161 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h u8 *addr, u8 idx); idx 169 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h __le16 *addr, u8 idx); idx 177 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h __le32 *addr, u8 idx); idx 185 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h u8 *addr, u8 val, u8 idx); idx 193 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h __le16 *addr, u16 val, u8 idx); idx 202 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h __le32 *addr, u32 val, u8 idx); idx 23 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c u8 *addr, u8 idx) idx 30 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c (unsigned long)addr, idx & 0x03, idx 40 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c __le16 *addr, u8 idx) idx 47 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c (unsigned long)addr, idx & 0x03, idx 57 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c __le32 *addr, u8 idx) idx 64 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c (unsigned long)addr, idx & 0x03, idx 74 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c u8 *addr, u8 val, u8 idx) idx 81 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c (unsigned long)addr, idx & 0x03, idx 88 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c __le16 *addr, u16 val, u8 idx) idx 95 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c (unsigned long)addr, idx & 0x03, idx 102 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c __le32 *addr, u32 val, u8 idx) idx 109 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8225.c (unsigned long)addr, idx & 0x03, idx 4715 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c tx_info->status.rates[0].idx = -1; idx 4784 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c rate = tx_info->control.rates[0].idx + DESC_RATE_MCS0; idx 4857 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c rate = tx_info->control.rates[0].idx + DESC_RATE_MCS0; idx 1197 drivers/net/wireless/realtek/rtlwifi/base.c r->idx; idx 1259 drivers/net/wireless/realtek/rtlwifi/base.c if (info->control.rates[0].idx == 0 || idx 2060 drivers/net/wireless/realtek/rtlwifi/base.c u8 idx, tid; idx 2084 drivers/net/wireless/realtek/rtlwifi/base.c for (idx = 0; idx <= 2; idx++) { idx 2085 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.num_rx_in4period[idx] = idx 2086 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.num_rx_in4period[idx + 1]; idx 2087 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.num_tx_in4period[idx] = idx 2088 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.num_tx_in4period[idx + 1]; idx 2094 drivers/net/wireless/realtek/rtlwifi/base.c for (idx = 0; idx <= 3; idx++) { idx 2096 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.num_rx_in4period[idx]; idx 2098 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.num_tx_in4period[idx]; idx 2124 drivers/net/wireless/realtek/rtlwifi/base.c for (idx = 0; idx <= 2; idx++) idx 2125 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.tidtx_in4period[tid][idx] = idx 2127 drivers/net/wireless/realtek/rtlwifi/base.c [idx + 1]; idx 2131 drivers/net/wireless/realtek/rtlwifi/base.c for (idx = 0; idx <= 3; idx++) idx 2133 drivers/net/wireless/realtek/rtlwifi/base.c rtlpriv->link_info.tidtx_in4period[tid][idx]; idx 2504 drivers/net/wireless/realtek/rtlwifi/base.c info->control.rates[0].idx = 0; idx 541 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->buffer_desc[ring->idx]); idx 543 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); idx 545 drivers/net/wireless/realtek/rtlwifi/pci.c if (!rtlpriv->cfg->ops->is_tx_desc_closed(hw, prio, ring->idx)) idx 547 drivers/net/wireless/realtek/rtlwifi/pci.c ring->idx = (ring->idx + 1) % ring->entries; idx 562 drivers/net/wireless/realtek/rtlwifi/pci.c ring->idx, idx 613 drivers/net/wireless/realtek/rtlwifi/pci.c prio, ring->idx, idx 748 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx]; idx 759 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx]; idx 763 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx]; idx 884 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx); idx 888 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx); idx 889 drivers/net/wireless/realtek/rtlwifi/pci.c if (rtlpci->rx_ring[rxring_idx].idx == idx 896 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx = idx 897 drivers/net/wireless/realtek/rtlwifi/pci.c (rtlpci->rx_ring[rxring_idx].idx + 1) % idx 1091 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->buffer_desc[ring->idx]); idx 1093 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); idx 1250 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->tx_ring[prio].idx = 0; idx 1295 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx = 0; idx 1318 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx = 0; idx 1346 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->buffer_desc[ring->idx]); idx 1348 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); idx 1356 drivers/net/wireless/realtek/rtlwifi/pci.c ring->idx = (ring->idx + 1) % ring->entries; idx 1476 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx = 0; idx 1508 drivers/net/wireless/realtek/rtlwifi/pci.c rtlpci->rx_ring[rxring_idx].idx = 0; idx 1526 drivers/net/wireless/realtek/rtlwifi/pci.c [ring->idx]); idx 1528 drivers/net/wireless/realtek/rtlwifi/pci.c entry = (u8 *)(&ring->desc[ring->idx]); idx 1538 drivers/net/wireless/realtek/rtlwifi/pci.c ring->idx = (ring->idx + 1) % ring->entries; idx 1546 drivers/net/wireless/realtek/rtlwifi/pci.c ring->idx = 0; idx 1604 drivers/net/wireless/realtek/rtlwifi/pci.c u16 idx; idx 1636 drivers/net/wireless/realtek/rtlwifi/pci.c idx = ring->cur_tx_wp; idx 1638 drivers/net/wireless/realtek/rtlwifi/pci.c idx = (ring->idx + skb_queue_len(&ring->queue)) % idx 1641 drivers/net/wireless/realtek/rtlwifi/pci.c idx = 0; idx 1644 drivers/net/wireless/realtek/rtlwifi/pci.c pdesc = &ring->desc[idx]; idx 1646 drivers/net/wireless/realtek/rtlwifi/pci.c ptx_bd_desc = &ring->buffer_desc[idx]; idx 1654 drivers/net/wireless/realtek/rtlwifi/pci.c hw_queue, ring->idx, idx, idx 1691 drivers/net/wireless/realtek/rtlwifi/pci.c hw_queue, ring->idx, idx, idx 148 drivers/net/wireless/realtek/rtlwifi/pci.h unsigned int idx; idx 161 drivers/net/wireless/realtek/rtlwifi/pci.h unsigned int idx; idx 108 drivers/net/wireless/realtek/rtlwifi/rc.c return rate.idx; idx 131 drivers/net/wireless/realtek/rtlwifi/rc.c rate->idx = rix >= 0x00 ? rix : 0x00; idx 182 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c u8 idx; idx 286 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c for (idx = 0; idx < 4; idx++) { idx 287 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 288 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c boxcontent[idx]); idx 301 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c for (idx = 0; idx < 2; idx++) { idx 302 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 303 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c boxextcontent[idx]); idx 306 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c for (idx = 0; idx < 4; idx++) { idx 307 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 308 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/fw.c boxcontent[idx]); idx 75 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c struct rtl_tx_desc *entry = &ring->desc[ring->idx]; idx 84 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c ring->idx = (ring->idx + 1) % ring->entries; idx 355 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c u8 idx; idx 359 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 360 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), idx 361 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c val[idx]); idx 381 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 382 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), idx 383 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c val[idx]); idx 2294 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c u8 idx = 0; idx 2300 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c for (idx = 0; idx < clear_number; idx++) { idx 2301 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 2302 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 2304 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c if (idx < 5) { idx 2305 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 2307 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c rtlpriv->sec.key_len[idx] = 0; idx 814 drivers/net/wireless/realtek/rtlwifi/rtl8188ee/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); idx 199 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c u8 idx; idx 301 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c for (idx = 0; idx < 4; idx++) { idx 302 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c rtl_write_byte(rtlpriv, box_reg + idx, idx 303 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c boxcontent[idx]); idx 311 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c for (idx = 0; idx < 4; idx++) { idx 312 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c rtl_write_byte(rtlpriv, box_reg + idx, idx 313 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c boxcontent[idx]); idx 321 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c for (idx = 0; idx < 4; idx++) { idx 322 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c rtl_write_byte(rtlpriv, box_reg + idx, idx 323 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c boxcontent[idx]); idx 333 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c for (idx = 0; idx < 2; idx++) { idx 334 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 335 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c boxextcontent[idx]); idx 338 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c for (idx = 0; idx < 4; idx++) { idx 339 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c rtl_write_byte(rtlpriv, box_reg + idx, idx 340 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c boxcontent[idx]); idx 350 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c for (idx = 0; idx < 2; idx++) { idx 351 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 352 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c boxextcontent[idx]); idx 355 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c for (idx = 0; idx < 4; idx++) { idx 356 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c rtl_write_byte(rtlpriv, box_reg + idx, idx 357 drivers/net/wireless/realtek/rtlwifi/rtl8192c/fw_common.c boxcontent[idx]); idx 581 drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c u8 idx; idx 594 drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c for (idx = 0; idx < 14; idx++) { idx 596 drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c rtlefuse->txpwrlevel_cck[rf_path][idx] = ccktxpwridx; idx 597 drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c rtlefuse->txpwrlevel_ht40_1s[rf_path][idx] = idx 599 drivers/net/wireless/realtek/rtlwifi/rtl8192c/phy_common.c rtlefuse->txpwrlevel_ht40_2s[rf_path][idx] = idx 134 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c u8 idx; idx 138 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 139 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), idx 140 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c val[idx]); idx 162 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 163 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), idx 164 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c val[idx]); idx 2089 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c u8 idx = 0; idx 2095 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c for (idx = 0; idx < clear_number; idx++) { idx 2096 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 2097 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 2099 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c if (idx < 5) { idx 2100 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 2102 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.c rtlpriv->sec.key_len[idx] = 0; idx 712 drivers/net/wireless/realtek/rtlwifi/rtl8192ce/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); idx 1556 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c u8 idx = 0; idx 1560 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 1561 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), idx 1562 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c val[idx]); idx 1589 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 1590 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), idx 1591 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/hw.c val[idx]); idx 214 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c u8 idx = 0; idx 219 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c for (idx = 0; idx < clear_number; idx++) { idx 220 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 221 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 222 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c if (idx < 5) { idx 223 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c memset(rtlpriv->sec.key_buf[idx], 0, idx 225 drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c rtlpriv->sec.key_len[idx] = 0; idx 286 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c u8 idx; idx 384 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c for (idx = 0; idx < 4; idx++) idx 385 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 386 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c boxcontent[idx]); idx 391 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c for (idx = 0; idx < 4; idx++) idx 392 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 393 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c boxcontent[idx]); idx 398 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c for (idx = 0; idx < 4; idx++) idx 399 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 400 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c boxcontent[idx]); idx 406 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c for (idx = 0; idx < 2; idx++) idx 407 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 408 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c boxextcontent[idx]); idx 409 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c for (idx = 0; idx < 4; idx++) idx 410 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 411 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c boxcontent[idx]); idx 417 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c for (idx = 0; idx < 2; idx++) idx 418 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 419 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c boxextcontent[idx]); idx 420 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c for (idx = 0; idx < 4; idx++) idx 421 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 422 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c boxcontent[idx]); idx 461 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c u8 idx = 0; idx 469 drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c pdesc = &ring->desc[idx]; idx 157 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c u8 idx; idx 161 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 162 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), idx 163 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c val[idx]); idx 187 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 188 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), idx 189 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c val[idx]); idx 2110 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c u8 idx; idx 2114 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c for (idx = 0; idx < clear_number; idx++) { idx 2115 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 2116 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 2118 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c if (idx < 5) { idx 2119 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 2121 drivers/net/wireless/realtek/rtlwifi/rtl8192de/hw.c rtlpriv->sec.key_len[idx] = 0; idx 831 drivers/net/wireless/realtek/rtlwifi/rtl8192de/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); idx 182 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c u8 idx; idx 309 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c for (idx = 0; idx < 4; idx++) { idx 310 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 311 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c boxcontent[idx]); idx 324 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c for (idx = 0; idx < 4; idx++) { idx 325 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 326 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c boxextcontent[idx]); idx 329 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c for (idx = 0; idx < 4; idx++) { idx 330 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 331 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/fw.c boxcontent[idx]); idx 414 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c u8 idx; idx 418 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c for (idx = 0; idx < ETH_ALEN; idx++) idx 419 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), val[idx]); idx 431 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c for (idx = 0; idx < ETH_ALEN; idx++) idx 432 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), val[idx]); idx 2020 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c u8 rf, idx; idx 2028 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c idx = _rtl92ee_get_chnl_group(i + 1); idx 2034 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c pwr2g.index_bw40_base[rf][idx]; idx 2037 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c pwr2g.index_cck_base[rf][idx]; idx 2039 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c pwr2g.index_bw40_base[rf][idx]; idx 2043 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c idx = _rtl92ee_get_chnl_group(channel5g[i]); idx 2045 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c pwr5g.index_bw40_base[rf][idx]; idx 2050 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c idx = _rtl92ee_get_chnl_group(channel5g_80m[i]); idx 2051 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c upper = pwr5g.index_bw40_base[rf][idx]; idx 2052 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c lower = pwr5g.index_bw40_base[rf][idx + 1]; idx 2437 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c u8 idx = 0; idx 2443 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c for (idx = 0; idx < clear_number; idx++) { idx 2444 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 2445 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 2447 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c if (idx < 5) { idx 2448 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 2450 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c rtlpriv->sec.key_len[idx] = 0; idx 2807 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c u8 idx; idx 2914 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c idx = rtl92ee_get_rightchnlplace_for_iqk(rtlphy->current_channel); idx 2919 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c rtlphy->iqk_matrix[idx].value[0][i] = idx 2922 drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c rtlphy->iqk_matrix[idx].iqk_done = true; idx 117 drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c u8 idx = 0; idx 123 drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c idx = (ring->idx + skb_queue_len(&ring->queue)) % ring->entries; idx 124 drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c pdesc = &ring->desc[idx]; idx 2403 drivers/net/wireless/realtek/rtlwifi/rtl8192se/hw.c u8 idx = 0; idx 2409 drivers/net/wireless/realtek/rtlwifi/rtl8192se/hw.c for (idx = 0; idx < clear_number; idx++) { idx 2410 drivers/net/wireless/realtek/rtlwifi/rtl8192se/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 2411 drivers/net/wireless/realtek/rtlwifi/rtl8192se/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 2413 drivers/net/wireless/realtek/rtlwifi/rtl8192se/hw.c if (idx < 5) { idx 2414 drivers/net/wireless/realtek/rtlwifi/rtl8192se/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 2416 drivers/net/wireless/realtek/rtlwifi/rtl8192se/hw.c rtlpriv->sec.key_len[idx] = 0; idx 219 drivers/net/wireless/realtek/rtlwifi/rtl8192se/sw.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); idx 44 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c u8 idx; idx 149 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c for (idx = 0; idx < 4; idx++) { idx 150 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 151 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c boxcontent[idx]); idx 159 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c for (idx = 0; idx < 4; idx++) { idx 160 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 161 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c boxcontent[idx]); idx 169 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c for (idx = 0; idx < 4; idx++) { idx 170 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 171 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c boxcontent[idx]); idx 181 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c for (idx = 0; idx < 2; idx++) { idx 182 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 183 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c boxextcontent[idx]); idx 186 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c for (idx = 0; idx < 4; idx++) { idx 187 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 188 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c boxcontent[idx]); idx 198 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c for (idx = 0; idx < 2; idx++) { idx 199 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 200 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c boxextcontent[idx]); idx 203 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c for (idx = 0; idx < 4; idx++) { idx 204 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 205 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/fw.c boxcontent[idx]); idx 138 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c u8 idx; idx 142 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 143 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), idx 144 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c val[idx]); idx 166 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 167 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), idx 168 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c val[idx]); idx 2169 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c u8 idx = 0; idx 2175 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c for (idx = 0; idx < clear_number; idx++) { idx 2176 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 2177 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 2179 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c if (idx < 5) { idx 2180 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 2182 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c rtlpriv->sec.key_len[idx] = 0; idx 646 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c u8 idx; idx 661 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c for (idx = 0; idx < 14; idx++) { idx 663 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c rtlefuse->txpwrlevel_cck[rf_path][idx] = ccktxpwridx; idx 664 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c rtlefuse->txpwrlevel_ht40_1s[rf_path][idx] = idx 666 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c rtlefuse->txpwrlevel_ht40_2s[rf_path][idx] = idx 673 drivers/net/wireless/realtek/rtlwifi/rtl8723ae/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); idx 650 drivers/net/wireless/realtek/rtlwifi/rtl8723be/dm.c u8 rfpath, u8 idx) idx 707 drivers/net/wireless/realtek/rtlwifi/rtl8723be/dm.c rtlphy->iqk_matrix[idx].value[0][0], idx 708 drivers/net/wireless/realtek/rtlwifi/rtl8723be/dm.c rtlphy->iqk_matrix[idx].value[0][1]); idx 717 drivers/net/wireless/realtek/rtlwifi/rtl8723be/dm.c rtlphy->iqk_matrix[idx].value[0][4], idx 718 drivers/net/wireless/realtek/rtlwifi/rtl8723be/dm.c rtlphy->iqk_matrix[idx].value[0][5]); idx 42 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c u8 idx; idx 148 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c for (idx = 0; idx < 4; idx++) { idx 149 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 150 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c boxcontent[idx]); idx 163 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c for (idx = 0; idx < 4; idx++) { idx 164 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 165 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c boxextcontent[idx]); idx 168 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c for (idx = 0; idx < 4; idx++) { idx 169 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 170 drivers/net/wireless/realtek/rtlwifi/rtl8723be/fw.c boxcontent[idx]); idx 37 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c struct rtl_tx_desc *entry = &ring->desc[ring->idx]; idx 46 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c ring->idx = (ring->idx + 1) % ring->entries; idx 397 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c u8 idx; idx 401 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c for (idx = 0; idx < ETH_ALEN; idx++) idx 402 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), val[idx]); idx 419 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c for (idx = 0; idx < ETH_ALEN; idx++) idx 420 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), val[idx]); idx 2558 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c u8 idx = 0; idx 2564 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c for (idx = 0; idx < clear_number; idx++) { idx 2565 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 2566 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 2568 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c if (idx < 5) { idx 2569 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 2571 drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c rtlpriv->sec.key_len[idx] = 0; idx 2252 drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c u8 i, final_candidate, idx; idx 2373 drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c idx = _get_right_chnl_place_for_iqk(rtlphy->current_channel); idx 2377 drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c rtlphy->iqk_matrix[idx].value[0][i] = idx 2379 drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c rtlphy->iqk_matrix[idx].iqk_done = true; idx 726 drivers/net/wireless/realtek/rtlwifi/rtl8723be/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); idx 250 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c u8 idx = 0; idx 366 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c for (idx = 0; idx < 4; idx++) { idx 367 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 368 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c boxcontent[idx]); idx 381 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c for (idx = 0; idx < 4; idx++) { idx 382 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c rtl_write_byte(rtlpriv, box_extreg + idx, idx 383 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c boxextcontent[idx]); idx 386 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c for (idx = 0; idx < 4; idx++) { idx 387 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c rtl_write_byte(rtlpriv, box_reg + idx, idx 388 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/fw.c boxcontent[idx]); idx 33 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c struct rtl_tx_desc *entry = &ring->desc[ring->idx]; idx 42 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c ring->idx = (ring->idx + 1) % ring->entries; idx 475 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c u8 idx; idx 479 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 480 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c rtl_write_byte(rtlpriv, (REG_MACID + idx), idx 481 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c val[idx]); idx 492 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c for (idx = 0; idx < ETH_ALEN; idx++) { idx 493 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c rtl_write_byte(rtlpriv, (REG_BSSID + idx), idx 494 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c val[idx]); idx 3840 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c u8 idx = 0; idx 3846 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c for (idx = 0; idx < clear_number; idx++) { idx 3847 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c rtl_cam_mark_invalid(hw, cam_offset + idx); idx 3848 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c rtl_cam_empty_entry(hw, cam_offset + idx); idx 3850 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c if (idx < 5) { idx 3851 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c memset(rtlpriv->sec.key_buf[idx], 0, idx 3853 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c rtlpriv->sec.key_len[idx] = 0; idx 966 drivers/net/wireless/realtek/rtlwifi/rtl8821ae/trx.c u8 *entry = (u8 *)(&ring->desc[ring->idx]); idx 468 drivers/net/wireless/realtek/rtw88/debug.c u8 idx = rate - DESC_RATE1M; idx 470 drivers/net/wireless/realtek/rtw88/debug.c seq_printf(m, " CCK_%-5s", cck_rate[idx]); idx 477 drivers/net/wireless/realtek/rtw88/debug.c u8 idx = rate - DESC_RATE6M; idx 479 drivers/net/wireless/realtek/rtw88/debug.c seq_printf(m, " OFDM_%-4s", ofdm_rate[idx]); idx 491 drivers/net/wireless/realtek/rtw88/debug.c u8 idx = rate - DESC_RATEVHT1SS_MCS0; idx 495 drivers/net/wireless/realtek/rtw88/debug.c n_ss = 1 + idx / 10; idx 497 drivers/net/wireless/realtek/rtw88/debug.c mcs_n = idx % 10; idx 93 drivers/net/wireless/realtek/rtw88/fw.c int idx; idx 135 drivers/net/wireless/realtek/rtw88/fw.c for (idx = 0; idx < 4; idx++) idx 136 drivers/net/wireless/realtek/rtw88/fw.c rtw_write8(rtwdev, box_reg + idx, h2c[idx]); idx 137 drivers/net/wireless/realtek/rtw88/fw.c for (idx = 0; idx < 4; idx++) idx 138 drivers/net/wireless/realtek/rtw88/fw.c rtw_write8(rtwdev, box_ex_reg + idx, h2c[idx + 4]); idx 761 drivers/net/wireless/realtek/rtw88/fw.c u16 idx = 0; idx 786 drivers/net/wireless/realtek/rtw88/fw.c buf[idx++] = rtw_read32(rtwdev, i); idx 189 drivers/net/wireless/realtek/rtw88/mac.c u32 idx = 0; idx 207 drivers/net/wireless/realtek/rtw88/mac.c cmd = cmd_seq[idx]; idx 215 drivers/net/wireless/realtek/rtw88/mac.c idx++; idx 86 drivers/net/wireless/realtek/rtw88/pci.c static inline void *rtw_pci_get_tx_desc(struct rtw_pci_tx_ring *tx_ring, u8 idx) idx 88 drivers/net/wireless/realtek/rtw88/pci.c int offset = tx_ring->r.desc_size * idx; idx 206 drivers/net/wireless/realtek/rtw88/pci.c u32 idx, u32 desc_sz) idx 222 drivers/net/wireless/realtek/rtw88/pci.c idx * desc_sz); idx 232 drivers/net/wireless/realtek/rtw88/pci.c u32 idx, u32 desc_sz) idx 241 drivers/net/wireless/realtek/rtw88/pci.c idx * desc_sz); idx 584 drivers/net/wireless/realtek/rtw88/pci.c u32 idx) idx 593 drivers/net/wireless/realtek/rtw88/pci.c idx * desc_sz); idx 511 drivers/net/wireless/rndis_wlan.c int idx, u8 *mac, struct station_info *sinfo); idx 573 drivers/net/wireless/rndis_wlan.c static bool is_wpa_key(struct rndis_wlan_private *priv, u8 idx) idx 575 drivers/net/wireless/rndis_wlan.c int cipher = priv->encr_keys[idx].cipher; idx 1511 drivers/net/wireless/rndis_wlan.c static void clear_key(struct rndis_wlan_private *priv, u8 idx) idx 1513 drivers/net/wireless/rndis_wlan.c memset(&priv->encr_keys[idx], 0, sizeof(priv->encr_keys[idx])); idx 2496 drivers/net/wireless/rndis_wlan.c int idx, u8 *mac, struct station_info *sinfo) idx 2501 drivers/net/wireless/rndis_wlan.c if (idx != 0) idx 890 drivers/net/wireless/rsi/rsi_91x_mac80211.c u8 idx = 0; idx 904 drivers/net/wireless/rsi/rsi_91x_mac80211.c idx = VO_Q; idx 907 drivers/net/wireless/rsi/rsi_91x_mac80211.c idx = VI_Q; idx 910 drivers/net/wireless/rsi/rsi_91x_mac80211.c idx = BE_Q; idx 913 drivers/net/wireless/rsi/rsi_91x_mac80211.c idx = BK_Q; idx 916 drivers/net/wireless/rsi/rsi_91x_mac80211.c idx = BE_Q; idx 920 drivers/net/wireless/rsi/rsi_91x_mac80211.c memcpy(&common->edca_params[idx], idx 925 drivers/net/wireless/rsi/rsi_91x_mac80211.c common->uapsd_bitmap |= idx; idx 927 drivers/net/wireless/rsi/rsi_91x_mac80211.c common->uapsd_bitmap &= (~idx); idx 565 drivers/net/wireless/rsi/rsi_91x_usb.c u8 idx, num_rx_cb; idx 569 drivers/net/wireless/rsi/rsi_91x_usb.c for (idx = 0; idx < num_rx_cb; idx++) { idx 570 drivers/net/wireless/rsi/rsi_91x_usb.c rx_cb = &dev->rx_cb[idx]; idx 574 drivers/net/wireless/rsi/rsi_91x_usb.c rsi_dbg(ERR_ZONE, "Failed alloc rx urb[%d]\n", idx); idx 577 drivers/net/wireless/rsi/rsi_91x_usb.c rx_cb->ep_num = idx + 1; idx 694 drivers/net/wireless/st/cw1200/sta.c int idx = cw1200_alloc_key(priv); idx 695 drivers/net/wireless/st/cw1200/sta.c struct wsm_add_key *wsm_key = &priv->keys[idx]; idx 697 drivers/net/wireless/st/cw1200/sta.c if (idx < 0) { idx 712 drivers/net/wireless/st/cw1200/sta.c cw1200_free_key(priv, idx); idx 811 drivers/net/wireless/st/cw1200/sta.c cw1200_free_key(priv, idx); idx 817 drivers/net/wireless/st/cw1200/sta.c key->hw_key_idx = idx; idx 819 drivers/net/wireless/st/cw1200/sta.c cw1200_free_key(priv, idx); idx 72 drivers/net/wireless/st/cw1200/txrx.c BUG_ON(rates[0].idx < 0); idx 77 drivers/net/wireless/st/cw1200/txrx.c if (rates[i].idx < 0) { idx 81 drivers/net/wireless/st/cw1200/txrx.c if (rates[i].idx > rates[i - 1].idx) { idx 91 drivers/net/wireless/st/cw1200/txrx.c if (rates[j].idx == rates[i].idx) { idx 93 drivers/net/wireless/st/cw1200/txrx.c } else if (rates[j].idx > rates[i].idx) { idx 125 drivers/net/wireless/st/cw1200/txrx.c rates[0].idx > 4 && rates[0].count > 2 && idx 126 drivers/net/wireless/st/cw1200/txrx.c rates[1].idx < 2) { idx 127 drivers/net/wireless/st/cw1200/txrx.c int mid_rate = (rates[0].idx + 4) >> 1; idx 137 drivers/net/wireless/st/cw1200/txrx.c rates[2].idx = 4; idx 142 drivers/net/wireless/st/cw1200/txrx.c rates[1].idx = mid_rate; idx 162 drivers/net/wireless/st/cw1200/txrx.c rates[1].idx = 4; idx 190 drivers/net/wireless/st/cw1200/txrx.c rates[0].idx, rates[0].count, idx 191 drivers/net/wireless/st/cw1200/txrx.c rates[1].idx, rates[1].count, idx 192 drivers/net/wireless/st/cw1200/txrx.c rates[2].idx, rates[2].count, idx 193 drivers/net/wireless/st/cw1200/txrx.c rates[3].idx, rates[3].count); idx 252 drivers/net/wireless/st/cw1200/txrx.c int idx, locked; idx 260 drivers/net/wireless/st/cw1200/txrx.c for (idx = 0; idx < TX_POLICY_CACHE_SIZE; idx++) { idx 261 drivers/net/wireless/st/cw1200/txrx.c entry = &cache->cache[idx]; idx 300 drivers/net/wireless/st/cw1200/txrx.c int idx; idx 311 drivers/net/wireless/st/cw1200/txrx.c idx = tx_policy_find(cache, &wanted); idx 312 drivers/net/wireless/st/cw1200/txrx.c if (idx >= 0) { idx 313 drivers/net/wireless/st/cw1200/txrx.c pr_debug("[TX policy] Used TX policy: %d\n", idx); idx 324 drivers/net/wireless/st/cw1200/txrx.c idx = entry - cache->cache; idx 325 drivers/net/wireless/st/cw1200/txrx.c pr_debug("[TX policy] New TX policy: %d\n", idx); idx 328 drivers/net/wireless/st/cw1200/txrx.c tx_policy_use(cache, &cache->cache[idx]); idx 334 drivers/net/wireless/st/cw1200/txrx.c return idx; idx 337 drivers/net/wireless/st/cw1200/txrx.c static void tx_policy_put(struct cw1200_common *priv, int idx) idx 344 drivers/net/wireless/st/cw1200/txrx.c usage = tx_policy_release(cache, &cache->cache[idx]); idx 428 drivers/net/wireless/st/cw1200/txrx.c if (rate->idx < 0) idx 431 drivers/net/wireless/st/cw1200/txrx.c return &priv->mcs_rates[rate->idx]; idx 433 drivers/net/wireless/st/cw1200/txrx.c bitrates[rate->idx]; idx 947 drivers/net/wireless/st/cw1200/txrx.c tx->status.rates[i].idx = -1; idx 1231 drivers/net/wireless/st/cw1200/txrx.c int idx; idx 1233 drivers/net/wireless/st/cw1200/txrx.c idx = ffs(~priv->key_map) - 1; idx 1234 drivers/net/wireless/st/cw1200/txrx.c if (idx < 0 || idx > WSM_KEY_MAX_INDEX) idx 1237 drivers/net/wireless/st/cw1200/txrx.c priv->key_map |= BIT(idx); idx 1238 drivers/net/wireless/st/cw1200/txrx.c priv->keys[idx].index = idx; idx 1239 drivers/net/wireless/st/cw1200/txrx.c return idx; idx 1242 drivers/net/wireless/st/cw1200/txrx.c void cw1200_free_key(struct cw1200_common *priv, int idx) idx 1244 drivers/net/wireless/st/cw1200/txrx.c BUG_ON(!(priv->key_map & BIT(idx))); idx 1245 drivers/net/wireless/st/cw1200/txrx.c memset(&priv->keys[idx], 0, sizeof(priv->keys[idx])); idx 1246 drivers/net/wireless/st/cw1200/txrx.c priv->key_map &= ~BIT(idx); idx 1257 drivers/net/wireless/st/cw1200/txrx.c int idx, ret = 0; idx 1258 drivers/net/wireless/st/cw1200/txrx.c for (idx = 0; idx <= WSM_KEY_MAX_INDEX; ++idx) idx 1259 drivers/net/wireless/st/cw1200/txrx.c if (priv->key_map & BIT(idx)) { idx 1260 drivers/net/wireless/st/cw1200/txrx.c ret = wsm_add_key(priv, &priv->keys[idx]); idx 87 drivers/net/wireless/st/cw1200/txrx.h void cw1200_free_key(struct cw1200_common *priv, int idx); idx 1625 drivers/net/wireless/st/cw1200/wsm.c int idx; idx 1632 drivers/net/wireless/st/cw1200/wsm.c idx = cw1200_get_prio_queue(priv, idx 1634 drivers/net/wireless/st/cw1200/wsm.c if (idx >= 0) { idx 1649 drivers/net/wireless/st/cw1200/wsm.c idx = cw1200_get_prio_queue(priv, idx 1651 drivers/net/wireless/st/cw1200/wsm.c if (idx < 0) idx 1655 drivers/net/wireless/st/cw1200/wsm.c *queue_p = &priv->tx_queue[idx]; idx 490 drivers/net/wireless/ti/wl1251/acx.c int idx = 0; idx 501 drivers/net/wireless/ti/wl1251/acx.c ie_table->table[idx++] = BEACON_FILTER_IE_ID_CHANNEL_SWITCH_ANN; idx 502 drivers/net/wireless/ti/wl1251/acx.c ie_table->table[idx++] = BEACON_RULE_PASS_ON_APPEARANCE; idx 1328 drivers/net/wireless/ti/wl1251/main.c static int wl1251_op_get_survey(struct ieee80211_hw *hw, int idx, idx 1334 drivers/net/wireless/ti/wl1251/main.c if (idx != 0) idx 293 drivers/net/wireless/ti/wl1251/tx.c u8 idx; idx 301 drivers/net/wireless/ti/wl1251/tx.c idx = info->control.hw_key->hw_key_idx; idx 302 drivers/net/wireless/ti/wl1251/tx.c if (unlikely(wl->default_key != idx)) { idx 303 drivers/net/wireless/ti/wl1251/tx.c ret = wl1251_acx_default_key(wl, idx); idx 25 drivers/net/wireless/ti/wl18xx/tx.c rate->idx = 0; idx 31 drivers/net/wireless/ti/wl18xx/tx.c rate->idx = fw_rate; idx 33 drivers/net/wireless/ti/wl18xx/tx.c rate->idx -= CONF_HW_RATE_INDEX_6MBPS; idx 37 drivers/net/wireless/ti/wl18xx/tx.c rate->idx = fw_rate - CONF_HW_RATE_INDEX_MCS0; idx 41 drivers/net/wireless/ti/wl18xx/tx.c (rate->idx)--; idx 43 drivers/net/wireless/ti/wl18xx/tx.c (rate->idx)--; idx 55 drivers/net/wireless/ti/wl18xx/tx.c rate->idx -= 8; idx 384 drivers/net/wireless/ti/wlcore/acx.c int i, idx = 0; idx 401 drivers/net/wireless/ti/wlcore/acx.c ie_table->table[idx++] = r->ie; idx 402 drivers/net/wireless/ti/wlcore/acx.c ie_table->table[idx++] = r->rule; idx 411 drivers/net/wireless/ti/wlcore/acx.c memcpy(&(ie_table->table[idx]), r->oui, idx 413 drivers/net/wireless/ti/wlcore/acx.c idx += CONF_BCN_IE_OUI_LEN; idx 414 drivers/net/wireless/ti/wlcore/acx.c ie_table->table[idx++] = r->type; idx 415 drivers/net/wireless/ti/wlcore/acx.c memcpy(&(ie_table->table[idx]), r->version, idx 417 drivers/net/wireless/ti/wlcore/acx.c idx += CONF_BCN_IE_VER_LEN; idx 793 drivers/net/wireless/ti/wlcore/acx.c u8 idx) idx 799 drivers/net/wireless/ti/wlcore/acx.c idx, c->enabled_rates); idx 812 drivers/net/wireless/ti/wlcore/acx.c acx->rate_policy_idx = cpu_to_le32(idx); idx 1067 drivers/net/wireless/ti/wlcore/acx.h u8 idx); idx 2143 drivers/net/wireless/ti/wlcore/main.c static int wl12xx_allocate_rate_policy(struct wl1271 *wl, u8 *idx) idx 2151 drivers/net/wireless/ti/wlcore/main.c *idx = policy; idx 2155 drivers/net/wireless/ti/wlcore/main.c static void wl12xx_free_rate_policy(struct wl1271 *wl, u8 *idx) idx 2157 drivers/net/wireless/ti/wlcore/main.c if (WARN_ON(*idx >= WL12XX_MAX_RATE_POLICIES)) idx 2160 drivers/net/wireless/ti/wlcore/main.c __clear_bit(*idx, wl->rate_policies_map); idx 2161 drivers/net/wireless/ti/wlcore/main.c *idx = WL12XX_MAX_RATE_POLICIES; idx 2164 drivers/net/wireless/ti/wlcore/main.c static int wlcore_allocate_klv_template(struct wl1271 *wl, u8 *idx) idx 2172 drivers/net/wireless/ti/wlcore/main.c *idx = policy; idx 2176 drivers/net/wireless/ti/wlcore/main.c static void wlcore_free_klv_template(struct wl1271 *wl, u8 *idx) idx 2178 drivers/net/wireless/ti/wlcore/main.c if (WARN_ON(*idx >= WLCORE_MAX_KLV_TEMPLATES)) idx 2181 drivers/net/wireless/ti/wlcore/main.c __clear_bit(*idx, wl->klv_templates_map); idx 2182 drivers/net/wireless/ti/wlcore/main.c *idx = WLCORE_MAX_KLV_TEMPLATES; idx 5005 drivers/net/wireless/ti/wlcore/main.c static int wl1271_op_get_survey(struct ieee80211_hw *hw, int idx, idx 5010 drivers/net/wireless/ti/wlcore/main.c if (idx != 0) idx 6046 drivers/net/wireless/ti/wlcore/main.c u8 idx; idx 6055 drivers/net/wireless/ti/wlcore/main.c idx = wl->band_rate_to_idx[band][rate]; idx 6056 drivers/net/wireless/ti/wlcore/main.c if (unlikely(idx == CONF_HW_RXTX_RATE_UNSUPPORTED)) { idx 6061 drivers/net/wireless/ti/wlcore/main.c return idx; idx 6092 drivers/net/wireless/ti/wlcore/main.c int idx = WLCORE_NUM_MAC_ADDRESSES - 1; idx 6093 drivers/net/wireless/ti/wlcore/main.c memcpy(&wl->addresses[idx], &wl->addresses[0], idx 6096 drivers/net/wireless/ti/wlcore/main.c wl->addresses[idx].addr[0] |= BIT(1); idx 97 drivers/net/wireless/ti/wlcore/ps.c info->status.rates[0].idx = -1; idx 391 drivers/net/wireless/ti/wlcore/tx.c u8 idx = info->control.hw_key->hw_key_idx; idx 397 drivers/net/wireless/ti/wlcore/tx.c if (WARN_ON(is_wep && wlvif && wlvif->default_key != idx)) { idx 398 drivers/net/wireless/ti/wlcore/tx.c ret = wl1271_set_default_wep_key(wl, wlvif, idx); idx 401 drivers/net/wireless/ti/wlcore/tx.c wlvif->default_key = idx; idx 940 drivers/net/wireless/ti/wlcore/tx.c info->status.rates[0].idx = rate; idx 1034 drivers/net/wireless/ti/wlcore/tx.c info->status.rates[0].idx = -1; idx 1123 drivers/net/wireless/ti/wlcore/tx.c info->status.rates[0].idx = -1; idx 320 drivers/net/wireless/virt_wifi.c int idx, u8 *mac, struct station_info *sinfo) idx 326 drivers/net/wireless/virt_wifi.c if (idx != 0 || !priv->is_connected) idx 455 drivers/net/wireless/zydas/zd1211rw/zd_mac.c first_idx = info->status.rates[0].idx; idx 460 drivers/net/wireless/zydas/zd1211rw/zd_mac.c info->status.rates[0].idx = retries->rate[0]; idx 464 drivers/net/wireless/zydas/zd1211rw/zd_mac.c info->status.rates[i].idx = retries->rate[i]; idx 468 drivers/net/wireless/zydas/zd1211rw/zd_mac.c info->status.rates[i].idx = retries->rate[retry - 1]; idx 472 drivers/net/wireless/zydas/zd1211rw/zd_mac.c info->status.rates[i].idx = -1; /* terminate */ idx 527 drivers/net/wireless/zydas/zd1211rw/zd_mac.c first_idx = info->status.rates[0].idx; idx 246 drivers/net/wireless/zydas/zd1211rw/zd_rf_uw2453.c int idx = channel - 1; idx 250 drivers/net/wireless/zydas/zd1211rw/zd_rf_uw2453.c val = UW2453_REGWRITE(1, uw2453_autocal_synth[idx]); idx 252 drivers/net/wireless/zydas/zd1211rw/zd_rf_uw2453.c val = UW2453_REGWRITE(1, uw2453_std_synth[idx]); idx 259 drivers/net/wireless/zydas/zd1211rw/zd_rf_uw2453.c UW2453_REGWRITE(2, uw2453_synth_divide[idx]), RF_RV_BITS); idx 133 drivers/net/xen-netback/common.h RING_IDX idx[COPY_BATCH_SIZE]; idx 111 drivers/net/xen-netback/netback.c u16 idx) idx 113 drivers/net/xen-netback/netback.c return page_to_pfn(queue->mmap_pages[idx]); idx 117 drivers/net/xen-netback/netback.c u16 idx) idx 119 drivers/net/xen-netback/netback.c return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx)); idx 804 drivers/net/xen-netback/netback.c RING_IDX idx; idx 824 drivers/net/xen-netback/netback.c idx = queue->tx.req_cons; idx 826 drivers/net/xen-netback/netback.c RING_COPY_REQUEST(&queue->tx, idx, &txreq); idx 836 drivers/net/xen-netback/netback.c queue->tx.req_cons = ++idx; idx 844 drivers/net/xen-netback/netback.c idx = queue->tx.req_cons; idx 880 drivers/net/xen-netback/netback.c idx += ret; idx 885 drivers/net/xen-netback/netback.c xenvif_tx_err(queue, &txreq, extra_count, idx); idx 910 drivers/net/xen-netback/netback.c xenvif_tx_err(queue, &txreq, extra_count, idx); idx 930 drivers/net/xen-netback/netback.c xenvif_tx_err(queue, &txreq, extra_count, idx); idx 1018 drivers/net/xen-netback/netback.c queue->tx.req_cons = idx; idx 1517 drivers/net/xen-netback/netback.c RING_IDX idx = vif->ctrl.rsp_prod_pvt; idx 1525 drivers/net/xen-netback/netback.c *RING_GET_RESPONSE(&vif->ctrl, idx) = rsp; idx 1526 drivers/net/xen-netback/netback.c vif->ctrl.rsp_prod_pvt = ++idx; idx 152 drivers/net/xen-netback/rx.c queue->rx_copy.idx[i]); idx 200 drivers/net/xen-netback/rx.c queue->rx_copy.idx[queue->rx_copy.num] = queue->rx.req_cons; idx 200 drivers/net/xen-netfront.c static int xennet_rxidx(RING_IDX idx) idx 202 drivers/net/xen-netfront.c return idx & (NET_RX_RING_SIZE - 1); idx 321 drivers/nfc/microread/microread.c r = nfc_dep_link_is_up(hdev->ndev, target->idx, comm_mode, idx 355 drivers/nfc/nfcsim.c u32 idx; idx 363 drivers/nfc/nfcsim.c idx = dev->nfc_digital_dev->nfc_dev->idx; idx 364 drivers/nfc/nfcsim.c n = snprintf(devname, sizeof(devname), "nfc%d", idx); idx 366 drivers/nfc/nfcsim.c NFCSIM_ERR(dev, "Could not compute dev name for dev %d\n", idx); idx 373 drivers/nfc/nfcsim.c idx); idx 1322 drivers/nfc/pn533/pn533.c dev->nfc_dev->targets[0].idx, idx 1792 drivers/nfc/pn533/pn533.c dev->nfc_dev->targets[0].idx, idx 467 drivers/nfc/pn544/pn544.c r = nfc_dep_link_is_up(hdev->ndev, target->idx, comm_mode, idx 603 drivers/nfc/st21nfca/core.c info->dep_info.idx = target->idx; idx 497 drivers/nfc/st21nfca/dep.c r = nfc_dep_link_is_up(info->hdev->ndev, info->dep_info.idx, idx 671 drivers/nfc/st21nfca/dep.c info->dep_info.idx = 0; idx 119 drivers/nfc/st21nfca/st21nfca.h u32 idx; idx 76 drivers/ntb/hw/amd/ntb_hw_amd.c static int ndev_mw_to_bar(struct amd_ntb_dev *ndev, int idx) idx 78 drivers/ntb/hw/amd/ntb_hw_amd.c if (idx < 0 || idx > ndev->mw_count) idx 81 drivers/ntb/hw/amd/ntb_hw_amd.c return ndev->dev_data->mw_idx << idx; idx 92 drivers/ntb/hw/amd/ntb_hw_amd.c static int amd_ntb_mw_get_align(struct ntb_dev *ntb, int pidx, int idx, idx 103 drivers/ntb/hw/amd/ntb_hw_amd.c bar = ndev_mw_to_bar(ndev, idx); idx 119 drivers/ntb/hw/amd/ntb_hw_amd.c static int amd_ntb_mw_set_trans(struct ntb_dev *ntb, int pidx, int idx, idx 132 drivers/ntb/hw/amd/ntb_hw_amd.c bar = ndev_mw_to_bar(ndev, idx); idx 300 drivers/ntb/hw/amd/ntb_hw_amd.c static int amd_ntb_peer_mw_get_addr(struct ntb_dev *ntb, int idx, idx 306 drivers/ntb/hw/amd/ntb_hw_amd.c bar = ndev_mw_to_bar(ndev, idx); idx 406 drivers/ntb/hw/amd/ntb_hw_amd.c static u32 amd_ntb_spad_read(struct ntb_dev *ntb, int idx) idx 412 drivers/ntb/hw/amd/ntb_hw_amd.c if (idx < 0 || idx >= ndev->spad_count) idx 415 drivers/ntb/hw/amd/ntb_hw_amd.c offset = ndev->self_spad + (idx << 2); idx 420 drivers/ntb/hw/amd/ntb_hw_amd.c int idx, u32 val) idx 426 drivers/ntb/hw/amd/ntb_hw_amd.c if (idx < 0 || idx >= ndev->spad_count) idx 429 drivers/ntb/hw/amd/ntb_hw_amd.c offset = ndev->self_spad + (idx << 2); idx 1091 drivers/ntb/hw/idt/ntb_hw_idt.c mws[*mw_cnt].idx = widx; idx 1263 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->mws[widx].idx * ndev->mws[widx].size_max; idx 1329 drivers/ntb/hw/idt/ntb_hw_idt.c lutoff = SET_FIELD(LUTOFFSET_INDEX, 0, mw_cfg->idx) | idx 1384 drivers/ntb/hw/idt/ntb_hw_idt.c lutoff = SET_FIELD(LUTOFFSET_INDEX, 0, mw_cfg->idx) | idx 2338 drivers/ntb/hw/idt/ntb_hw_idt.c unsigned char idx, pidx, cnt; idx 2365 drivers/ntb/hw/idt/ntb_hw_idt.c for (idx = 0; idx < ndev->peer_cnt; idx++) { idx 2368 drivers/ntb/hw/idt/ntb_hw_idt.c idx, ndev->peers[idx].port, ndev->peers[idx].part); idx 2380 drivers/ntb/hw/idt/ntb_hw_idt.c for (idx = 0; idx < IDT_MTBL_ENTRY_CNT; idx++) { idx 2382 drivers/ntb/hw/idt/ntb_hw_idt.c idt_nt_write(ndev, IDT_NT_NTMTBLADDR, idx); idx 2390 drivers/ntb/hw/idt/ntb_hw_idt.c idx, GET_FIELD(NTMTBLDATA_PART, data), idx 2399 drivers/ntb/hw/idt/ntb_hw_idt.c for (idx = 0; idx < ndev->mw_cnt; idx += cnt) { idx 2400 drivers/ntb/hw/idt/ntb_hw_idt.c data = ndev->mws[idx].type; idx 2406 drivers/ntb/hw/idt/ntb_hw_idt.c "\t%hhu.\t", idx); idx 2409 drivers/ntb/hw/idt/ntb_hw_idt.c "\t%hhu-%hhu.\t", idx, idx + cnt - 1); idx 2412 drivers/ntb/hw/idt/ntb_hw_idt.c idt_get_mw_name(data), ndev->mws[idx].bar); idx 2415 drivers/ntb/hw/idt/ntb_hw_idt.c "Address align 0x%08llx, ", ndev->mws[idx].addr_align); idx 2419 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->mws[idx].size_align, ndev->mws[idx].size_max); idx 2429 drivers/ntb/hw/idt/ntb_hw_idt.c for (idx = 0; idx < ndev->peers[pidx].mw_cnt; idx += cnt) { idx 2430 drivers/ntb/hw/idt/ntb_hw_idt.c data = ndev->peers[pidx].mws[idx].type; idx 2435 drivers/ntb/hw/idt/ntb_hw_idt.c "\t%hhu.\t", idx); idx 2438 drivers/ntb/hw/idt/ntb_hw_idt.c "\t%hhu-%hhu.\t", idx, idx + cnt - 1); idx 2442 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->peers[pidx].mws[idx].bar); idx 2446 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->peers[pidx].mws[idx].addr_align); idx 2450 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->peers[pidx].mws[idx].size_align, idx 2451 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->peers[pidx].mws[idx].size_max); idx 2479 drivers/ntb/hw/idt/ntb_hw_idt.c for (idx = 0; idx < IDT_MSG_CNT; idx++) { idx 2481 drivers/ntb/hw/idt/ntb_hw_idt.c data = idt_ntb_msg_read(&ndev->ntb, &src, idx); idx 2484 drivers/ntb/hw/idt/ntb_hw_idt.c idx, data, src, ndev->peers[src].port); idx 1069 drivers/ntb/hw/idt/ntb_hw_idt.h unsigned char idx; idx 176 drivers/ntb/hw/intel/ntb_hw_gen1.c int ndev_mw_to_bar(struct intel_ntb_dev *ndev, int idx) idx 178 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx < 0 || idx >= ndev->mw_count) idx 180 drivers/ntb/hw/intel/ntb_hw_gen1.c return ndev->reg->mw_bar[idx]; idx 276 drivers/ntb/hw/intel/ntb_hw_gen1.c static inline int ndev_spad_addr(struct intel_ntb_dev *ndev, int idx, idx 283 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx < 0 || idx >= ndev->spad_count) idx 287 drivers/ntb/hw/intel/ntb_hw_gen1.c *spad_addr = reg_addr + reg + (idx << 2); idx 295 drivers/ntb/hw/intel/ntb_hw_gen1.c static inline u32 ndev_spad_read(struct intel_ntb_dev *ndev, int idx, idx 301 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx < 0 || idx >= ndev->spad_count) idx 304 drivers/ntb/hw/intel/ntb_hw_gen1.c return ioread32(mmio + (idx << 2)); idx 307 drivers/ntb/hw/intel/ntb_hw_gen1.c static inline int ndev_spad_write(struct intel_ntb_dev *ndev, int idx, u32 val, idx 313 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx < 0 || idx >= ndev->spad_count) idx 316 drivers/ntb/hw/intel/ntb_hw_gen1.c iowrite32(val, mmio + (idx << 2)); idx 801 drivers/ntb/hw/intel/ntb_hw_gen1.c int intel_ntb_mw_get_align(struct ntb_dev *ntb, int pidx, int idx, idx 813 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx >= ndev->b2b_idx && !ndev->b2b_off) idx 814 drivers/ntb/hw/intel/ntb_hw_gen1.c idx += 1; idx 816 drivers/ntb/hw/intel/ntb_hw_gen1.c bar = ndev_mw_to_bar(ndev, idx); idx 822 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx == ndev->b2b_idx) idx 839 drivers/ntb/hw/intel/ntb_hw_gen1.c static int intel_ntb_mw_set_trans(struct ntb_dev *ntb, int pidx, int idx, idx 852 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx >= ndev->b2b_idx && !ndev->b2b_off) idx 853 drivers/ntb/hw/intel/ntb_hw_gen1.c idx += 1; idx 855 drivers/ntb/hw/intel/ntb_hw_gen1.c bar = ndev_mw_to_bar(ndev, idx); idx 861 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx == ndev->b2b_idx) idx 1023 drivers/ntb/hw/intel/ntb_hw_gen1.c int intel_ntb_peer_mw_get_addr(struct ntb_dev *ntb, int idx, idx 1029 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx >= ndev->b2b_idx && !ndev->b2b_off) idx 1030 drivers/ntb/hw/intel/ntb_hw_gen1.c idx += 1; idx 1032 drivers/ntb/hw/intel/ntb_hw_gen1.c bar = ndev_mw_to_bar(ndev, idx); idx 1038 drivers/ntb/hw/intel/ntb_hw_gen1.c (idx == ndev->b2b_idx ? ndev->b2b_off : 0); idx 1042 drivers/ntb/hw/intel/ntb_hw_gen1.c (idx == ndev->b2b_idx ? ndev->b2b_off : 0); idx 1159 drivers/ntb/hw/intel/ntb_hw_gen1.c u32 intel_ntb_spad_read(struct ntb_dev *ntb, int idx) idx 1163 drivers/ntb/hw/intel/ntb_hw_gen1.c return ndev_spad_read(ndev, idx, idx 1168 drivers/ntb/hw/intel/ntb_hw_gen1.c int intel_ntb_spad_write(struct ntb_dev *ntb, int idx, u32 val) idx 1172 drivers/ntb/hw/intel/ntb_hw_gen1.c return ndev_spad_write(ndev, idx, val, idx 156 drivers/ntb/hw/intel/ntb_hw_gen1.h int ndev_mw_to_bar(struct intel_ntb_dev *ndev, int idx); idx 158 drivers/ntb/hw/intel/ntb_hw_gen1.h int intel_ntb_mw_get_align(struct ntb_dev *ntb, int pidx, int idx, idx 162 drivers/ntb/hw/intel/ntb_hw_gen1.h int intel_ntb_peer_mw_get_addr(struct ntb_dev *ntb, int idx, idx 174 drivers/ntb/hw/intel/ntb_hw_gen1.h u32 intel_ntb_spad_read(struct ntb_dev *ntb, int idx); idx 175 drivers/ntb/hw/intel/ntb_hw_gen1.h int intel_ntb_spad_write(struct ntb_dev *ntb, int idx, u32 val); idx 444 drivers/ntb/hw/intel/ntb_hw_gen3.c static int intel_ntb3_mw_set_trans(struct ntb_dev *ntb, int pidx, int idx, idx 457 drivers/ntb/hw/intel/ntb_hw_gen3.c if (idx >= ndev->b2b_idx && !ndev->b2b_off) idx 458 drivers/ntb/hw/intel/ntb_hw_gen3.c idx += 1; idx 460 drivers/ntb/hw/intel/ntb_hw_gen3.c bar = ndev_mw_to_bar(ndev, idx); idx 466 drivers/ntb/hw/intel/ntb_hw_gen3.c if (idx == ndev->b2b_idx) idx 480 drivers/ntb/hw/intel/ntb_hw_gen3.c xlat_reg = ndev->xlat_reg->bar2_xlat + (idx * 0x10); idx 481 drivers/ntb/hw/intel/ntb_hw_gen3.c limit_reg = ndev->xlat_reg->bar2_limit + (idx * 0x10); idx 512 drivers/ntb/hw/intel/ntb_hw_gen3.c limit_reg = ndev->xlat_reg->bar2_limit + (idx * 0x10) + 0x4000; idx 513 drivers/ntb/hw/intel/ntb_hw_gen3.c base = ioread64(mmio + GEN3_EMBAR1_OFFSET + (8 * idx)); idx 155 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static int switchtec_ntb_send_msg(struct switchtec_ntb *sndev, int idx, idx 158 drivers/ntb/hw/mscc/ntb_hw_switchtec.c if (idx < 0 || idx >= ARRAY_SIZE(sndev->mmio_peer_dbmsg->omsg)) idx 161 drivers/ntb/hw/mscc/ntb_hw_switchtec.c iowrite32(val, &sndev->mmio_peer_dbmsg->omsg[idx].msg); idx 221 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static void switchtec_ntb_mw_clr_direct(struct switchtec_ntb *sndev, int idx) idx 224 drivers/ntb/hw/mscc/ntb_hw_switchtec.c int bar = sndev->peer_direct_mw_to_bar[idx]; idx 235 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static void switchtec_ntb_mw_clr_lut(struct switchtec_ntb *sndev, int idx) idx 239 drivers/ntb/hw/mscc/ntb_hw_switchtec.c iowrite64(0, &ctl->lut_entry[peer_lut_index(sndev, idx)]); idx 242 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static void switchtec_ntb_mw_set_direct(struct switchtec_ntb *sndev, int idx, idx 246 drivers/ntb/hw/mscc/ntb_hw_switchtec.c int bar = sndev->peer_direct_mw_to_bar[idx]; idx 261 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static void switchtec_ntb_mw_set_lut(struct switchtec_ntb *sndev, int idx, idx 267 drivers/ntb/hw/mscc/ntb_hw_switchtec.c &ctl->lut_entry[peer_lut_index(sndev, idx)]); idx 350 drivers/ntb/hw/mscc/ntb_hw_switchtec.c int idx, phys_addr_t *base, idx 353 drivers/ntb/hw/mscc/ntb_hw_switchtec.c int bar = sndev->direct_mw_to_bar[idx]; idx 359 drivers/ntb/hw/mscc/ntb_hw_switchtec.c if (idx == 0) { idx 385 drivers/ntb/hw/mscc/ntb_hw_switchtec.c int idx, phys_addr_t *base, idx 391 drivers/ntb/hw/mscc/ntb_hw_switchtec.c offset = LUT_SIZE * lut_index(sndev, idx); idx 402 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static int switchtec_ntb_peer_mw_get_addr(struct ntb_dev *ntb, int idx, idx 408 drivers/ntb/hw/mscc/ntb_hw_switchtec.c if (idx < sndev->nr_direct_mw) idx 409 drivers/ntb/hw/mscc/ntb_hw_switchtec.c return switchtec_ntb_direct_get_addr(sndev, idx, base, size); idx 410 drivers/ntb/hw/mscc/ntb_hw_switchtec.c else if (idx < switchtec_ntb_peer_mw_count(ntb)) idx 411 drivers/ntb/hw/mscc/ntb_hw_switchtec.c return switchtec_ntb_lut_get_addr(sndev, idx, base, size); idx 729 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static u32 switchtec_ntb_spad_read(struct ntb_dev *ntb, int idx) idx 733 drivers/ntb/hw/mscc/ntb_hw_switchtec.c if (idx < 0 || idx >= ARRAY_SIZE(sndev->self_shared->spad)) idx 739 drivers/ntb/hw/mscc/ntb_hw_switchtec.c return sndev->self_shared->spad[idx]; idx 742 drivers/ntb/hw/mscc/ntb_hw_switchtec.c static int switchtec_ntb_spad_write(struct ntb_dev *ntb, int idx, u32 val) idx 746 drivers/ntb/hw/mscc/ntb_hw_switchtec.c if (idx < 0 || idx >= ARRAY_SIZE(sndev->self_shared->spad)) idx 752 drivers/ntb/hw/mscc/ntb_hw_switchtec.c sndev->self_shared->spad[idx] = val; idx 1317 drivers/ntb/hw/mscc/ntb_hw_switchtec.c int idx = sndev->nr_direct_mw + i; idx 1319 drivers/ntb/hw/mscc/ntb_hw_switchtec.c sndev->self_shared->mw_sizes[idx] = LUT_SIZE; idx 366 drivers/ntb/msi.c int idx; idx 371 drivers/ntb/msi.c idx = desc->addr_offset / sizeof(*ntb->msi->peer_mws[peer]); idx 373 drivers/ntb/msi.c iowrite32(desc->data, &ntb->msi->peer_mws[peer][idx]); idx 184 drivers/ntb/test/ntb_msi_test.c static int ntb_msit_dbgfs_trigger(void *data, u64 idx) idx 188 drivers/ntb/test/ntb_msi_test.c if (idx >= peer->num_irqs) idx 192 drivers/ntb/test/ntb_msi_test.c idx, peer->pidx); idx 195 drivers/ntb/test/ntb_msi_test.c &peer->msi_desc[idx]); idx 217 drivers/nvdimm/btt.c int idx) idx 226 drivers/nvdimm/btt.c snprintf(dirname, 32, "arena%d", idx); idx 1091 drivers/nvdimm/btt.c __acquires(&arena->map_locks[idx].lock) idx 1093 drivers/nvdimm/btt.c u32 idx = (premap * MAP_ENT_SIZE / L1_CACHE_BYTES) % arena->nfree; idx 1095 drivers/nvdimm/btt.c spin_lock(&arena->map_locks[idx].lock); idx 1099 drivers/nvdimm/btt.c __releases(&arena->map_locks[idx].lock) idx 1101 drivers/nvdimm/btt.c u32 idx = (premap * MAP_ENT_SIZE / L1_CACHE_BYTES) % arena->nfree; idx 1103 drivers/nvdimm/btt.c spin_unlock(&arena->map_locks[idx].lock); idx 819 drivers/nvdimm/bus.c const struct nd_cmd_desc *desc, int idx, void *buf) idx 821 drivers/nvdimm/bus.c if (idx >= desc->in_num) idx 824 drivers/nvdimm/bus.c if (desc->in_sizes[idx] < UINT_MAX) idx 825 drivers/nvdimm/bus.c return desc->in_sizes[idx]; idx 827 drivers/nvdimm/bus.c if (nvdimm && cmd == ND_CMD_SET_CONFIG_DATA && idx == 2) { idx 831 drivers/nvdimm/bus.c } else if (nvdimm && cmd == ND_CMD_VENDOR && idx == 2) { idx 846 drivers/nvdimm/bus.c const struct nd_cmd_desc *desc, int idx, const u32 *in_field, idx 849 drivers/nvdimm/bus.c if (idx >= desc->out_num) idx 852 drivers/nvdimm/bus.c if (desc->out_sizes[idx] < UINT_MAX) idx 853 drivers/nvdimm/bus.c return desc->out_sizes[idx]; idx 855 drivers/nvdimm/bus.c if (nvdimm && cmd == ND_CMD_GET_CONFIG_DATA && idx == 1) idx 857 drivers/nvdimm/bus.c else if (nvdimm && cmd == ND_CMD_VENDOR && idx == 2) idx 859 drivers/nvdimm/bus.c else if (!nvdimm && cmd == ND_CMD_ARS_STATUS && idx == 2) { idx 307 drivers/nvdimm/label.c static bool preamble_index(struct nvdimm_drvdata *ndd, int idx, idx 313 drivers/nvdimm/label.c nsindex = to_namespace_index(ndd, idx); idx 767 drivers/nvdimm/region_devs.c #define REGION_MAPPING(idx) \ idx 768 drivers/nvdimm/region_devs.c static ssize_t mapping##idx##_show(struct device *dev, \ idx 771 drivers/nvdimm/region_devs.c return mappingN(dev, buf, idx); \ idx 773 drivers/nvdimm/region_devs.c static DEVICE_ATTR_RO(mapping##idx) idx 1102 drivers/nvdimm/region_devs.c int i, idx; idx 1108 drivers/nvdimm/region_devs.c idx = this_cpu_read(flush_idx); idx 1109 drivers/nvdimm/region_devs.c idx = this_cpu_add_return(flush_idx, hash_32(current->pid + idx, 8)); idx 1121 drivers/nvdimm/region_devs.c writeq(1, ndrd_get_flush_wpq(ndrd, i, idx)); idx 1519 drivers/nvme/host/core.c static void nvme_put_ns_from_disk(struct nvme_ns_head *head, int idx) idx 1522 drivers/nvme/host/core.c srcu_read_unlock(&head->srcu, idx); idx 339 drivers/nvme/host/fc.c int ret, idx; idx 379 drivers/nvme/host/fc.c idx = ida_simple_get(&nvme_fc_local_port_cnt, 0, 0, GFP_KERNEL); idx 380 drivers/nvme/host/fc.c if (idx < 0) { idx 403 drivers/nvme/host/fc.c newrec->localport.port_num = idx; idx 416 drivers/nvme/host/fc.c ida_simple_remove(&nvme_fc_local_port_cnt, idx); idx 653 drivers/nvme/host/fc.c int ret, idx; idx 690 drivers/nvme/host/fc.c idx = ida_simple_get(&lport->endp_cnt, 0, 0, GFP_KERNEL); idx 691 drivers/nvme/host/fc.c if (idx < 0) { idx 712 drivers/nvme/host/fc.c newrec->remoteport.port_num = idx; idx 1868 drivers/nvme/host/fc.c nvme_fc_init_queue(struct nvme_fc_ctrl *ctrl, int idx) idx 1872 drivers/nvme/host/fc.c queue = &ctrl->queues[idx]; idx 1875 drivers/nvme/host/fc.c queue->qnum = idx; idx 1879 drivers/nvme/host/fc.c if (idx > 0) idx 3051 drivers/nvme/host/fc.c int ret, idx; idx 3071 drivers/nvme/host/fc.c idx = ida_simple_get(&nvme_fc_ctrl_cnt, 0, 0, GFP_KERNEL); idx 3072 drivers/nvme/host/fc.c if (idx < 0) { idx 3087 drivers/nvme/host/fc.c ctrl->cnum = idx; idx 947 drivers/nvme/host/pci.c static inline void nvme_handle_cqe(struct nvme_queue *nvmeq, u16 idx) idx 949 drivers/nvme/host/pci.c volatile struct nvme_completion *cqe = &nvmeq->cqes[idx]; idx 438 drivers/nvme/host/rdma.c int comp_vector, idx = nvme_rdma_queue_idx(queue); idx 454 drivers/nvme/host/rdma.c comp_vector = idx == 0 ? idx : idx - 1; idx 495 drivers/nvme/host/rdma.c queue->queue_size, idx); idx 516 drivers/nvme/host/rdma.c int idx, size_t queue_size) idx 522 drivers/nvme/host/rdma.c queue = &ctrl->queues[idx]; idx 526 drivers/nvme/host/rdma.c if (idx > 0) idx 609 drivers/nvme/host/rdma.c static int nvme_rdma_start_queue(struct nvme_rdma_ctrl *ctrl, int idx) idx 611 drivers/nvme/host/rdma.c struct nvme_rdma_queue *queue = &ctrl->queues[idx]; idx 615 drivers/nvme/host/rdma.c if (idx) idx 616 drivers/nvme/host/rdma.c ret = nvmf_connect_io_queue(&ctrl->ctrl, idx, poll); idx 626 drivers/nvme/host/rdma.c "failed to connect queue: %d ret=%d\n", idx, ret); idx 1442 drivers/nvme/host/tcp.c static int nvme_tcp_start_queue(struct nvme_ctrl *nctrl, int idx) idx 1447 drivers/nvme/host/tcp.c if (idx) idx 1448 drivers/nvme/host/tcp.c ret = nvmf_connect_io_queue(nctrl, idx, false); idx 1453 drivers/nvme/host/tcp.c set_bit(NVME_TCP_Q_LIVE, &ctrl->queues[idx].flags); idx 1455 drivers/nvme/host/tcp.c if (test_bit(NVME_TCP_Q_ALLOCATED, &ctrl->queues[idx].flags)) idx 1456 drivers/nvme/host/tcp.c __nvme_tcp_stop_queue(&ctrl->queues[idx]); idx 1458 drivers/nvme/host/tcp.c "failed to connect queue: %d ret=%d\n", idx, ret); idx 796 drivers/nvme/target/fc.c int idx; idx 803 drivers/nvme/target/fc.c idx = ida_simple_get(&tgtport->assoc_cnt, 0, 0, GFP_KERNEL); idx 804 drivers/nvme/target/fc.c if (idx < 0) idx 811 drivers/nvme/target/fc.c assoc->a_id = idx; idx 837 drivers/nvme/target/fc.c ida_simple_remove(&tgtport->assoc_cnt, idx); idx 1020 drivers/nvme/target/fc.c int ret, idx; idx 1038 drivers/nvme/target/fc.c idx = ida_simple_get(&nvmet_fc_tgtport_cnt, 0, 0, GFP_KERNEL); idx 1039 drivers/nvme/target/fc.c if (idx < 0) { idx 1053 drivers/nvme/target/fc.c newrec->fc_target_port.port_num = idx; idx 1083 drivers/nvme/target/fc.c ida_simple_remove(&nvmet_fc_tgtport_cnt, idx); idx 291 drivers/nvme/target/fcloop.c unsigned int idx, void *handle) idx 97 drivers/nvme/target/rdma.c int idx; idx 738 drivers/nvme/target/rdma.c 1 + rsp->n_rdma, queue->idx, idx 985 drivers/nvme/target/rdma.c queue->idx % ndev->device->num_comp_vectors; idx 1066 drivers/nvme/target/rdma.c pr_debug("freeing queue %d\n", queue->idx); idx 1077 drivers/nvme/target/rdma.c ida_simple_remove(&nvmet_rdma_queue_ida, queue->idx); idx 1177 drivers/nvme/target/rdma.c queue->idx = ida_simple_get(&nvmet_rdma_queue_ida, 0, 0, GFP_KERNEL); idx 1178 drivers/nvme/target/rdma.c if (queue->idx < 0) { idx 1218 drivers/nvme/target/rdma.c ida_simple_remove(&nvmet_rdma_queue_ida, queue->idx); idx 1388 drivers/nvme/target/rdma.c pr_err("failed to connect queue %d\n", queue->idx); idx 1645 drivers/nvme/target/rdma.c pr_info("Removing queue %d\n", queue->idx); idx 120 drivers/nvme/target/tcp.c int idx; idx 236 drivers/nvme/target/tcp.c queue->idx); idx 245 drivers/nvme/target/tcp.c queue->idx, le32_to_cpu(recv_digest), idx 263 drivers/nvme/target/tcp.c pr_err("queue %d: data digest flag is cleared\n", queue->idx); idx 773 drivers/nvme/target/tcp.c pr_err("queue %d: bad pfv %d\n", queue->idx, icreq->pfv); idx 778 drivers/nvme/target/tcp.c pr_err("queue %d: unsupported hpda %d\n", queue->idx, idx 835 drivers/nvme/target/tcp.c pr_err("queue %d: failed to map data\n", queue->idx); idx 899 drivers/nvme/target/tcp.c queue->idx, queue->nr_cmds, queue->send_list_len, idx 921 drivers/nvme/target/tcp.c pr_err("queue %d: failed to map data\n", queue->idx); idx 955 drivers/nvme/target/tcp.c size_t idx = type; idx 957 drivers/nvme/target/tcp.c return (idx < ARRAY_SIZE(nvme_tcp_pdu_sizes) && idx 958 drivers/nvme/target/tcp.c nvme_tcp_pdu_sizes[idx]) ? idx 959 drivers/nvme/target/tcp.c nvme_tcp_pdu_sizes[idx] : 0; idx 1090 drivers/nvme/target/tcp.c queue->idx, cmd->req.cmd->common.command_id, idx 1354 drivers/nvme/target/tcp.c ida_simple_remove(&nvmet_tcp_queue_ida, queue->idx); idx 1411 drivers/nvme/target/tcp.c queue->idx, sk->sk_state); idx 1488 drivers/nvme/target/tcp.c queue->idx = ida_simple_get(&nvmet_tcp_queue_ida, 0, 0, GFP_KERNEL); idx 1489 drivers/nvme/target/tcp.c if (queue->idx < 0) { idx 1490 drivers/nvme/target/tcp.c ret = queue->idx; idx 1526 drivers/nvme/target/tcp.c ida_simple_remove(&nvmet_tcp_queue_ida, queue->idx); idx 92 drivers/nvmem/rave-sp-eeprom.c u16 idx, idx 110 drivers/nvmem/rave-sp-eeprom.c cmd[offset++] = idx; idx 118 drivers/nvmem/rave-sp-eeprom.c cmd[offset++] = idx >> 8; idx 313 drivers/of/of_reserved_mem.c struct device_node *np, int idx) idx 323 drivers/of/of_reserved_mem.c target = of_parse_phandle(np, "memory-region", idx); idx 297 drivers/parisc/ccio-dma.c unsigned int idx;\ idx 298 drivers/parisc/ccio-dma.c idx = (unsigned int)((unsigned long)res_ptr - (unsigned long)ioc->res_map); \ idx 299 drivers/parisc/ccio-dma.c ret = iommu_is_span_boundary(idx << 3, pages_needed, 0, boundary_size);\ idx 302 drivers/parisc/ccio-dma.c res_idx = idx;\ idx 670 drivers/parisc/ccio-dma.c unsigned int idx = PDIR_INDEX(iovp); idx 671 drivers/parisc/ccio-dma.c char *pdir_ptr = (char *) &(ioc->pdir_base[idx]); idx 673 drivers/parisc/ccio-dma.c BUG_ON(idx >= (ioc->pdir_size / sizeof(u64))); idx 727 drivers/parisc/ccio-dma.c int idx; idx 754 drivers/parisc/ccio-dma.c idx = ccio_alloc_range(ioc, dev, size); idx 755 drivers/parisc/ccio-dma.c iovp = (dma_addr_t)MKIOVP(idx); idx 757 drivers/parisc/ccio-dma.c pdir_start = &(ioc->pdir_base[idx]); idx 1160 drivers/parisc/ccio-dma.c unsigned int idx; idx 1167 drivers/parisc/ccio-dma.c idx = PDIR_INDEX(iovp) >> 3; idx 1169 drivers/parisc/ccio-dma.c while (idx < ioc->res_size) { idx 1170 drivers/parisc/ccio-dma.c res_ptr[idx] |= 0xff; idx 1171 drivers/parisc/ccio-dma.c idx += PDIR_INDEX(CUJO_20_STEP) >> 3; idx 170 drivers/parisc/iosapic.c #define IOSAPIC_IRDT_ENTRY(idx) (0x10+(idx)*2) idx 171 drivers/parisc/iosapic.c #define IOSAPIC_IRDT_ENTRY_HI(idx) (0x11+(idx)*2) idx 529 drivers/parisc/iosapic.c u8 idx = vi->irqline; idx 531 drivers/parisc/iosapic.c *dp0 = iosapic_read(isp->addr, IOSAPIC_IRDT_ENTRY(idx)); idx 532 drivers/parisc/iosapic.c *dp1 = iosapic_read(isp->addr, IOSAPIC_IRDT_ENTRY_HI(idx)); idx 669 drivers/parisc/lba_pci.c int idx; idx 672 drivers/parisc/lba_pci.c for (idx = PCI_BRIDGE_RESOURCES; idx < PCI_NUM_RESOURCES; idx++) { idx 673 drivers/parisc/lba_pci.c r = &dev->resource[idx]; idx 678 drivers/parisc/lba_pci.c if (!r->start || pci_claim_bridge_resource(dev, idx) < 0) { idx 261 drivers/pci/bus.c bool pci_bus_clip_resource(struct pci_dev *dev, int idx) idx 264 drivers/pci/bus.c struct resource *res = &dev->resource[idx]; idx 1032 drivers/pci/controller/pcie-rcar.c int idx = *index; idx 1061 drivers/pci/controller/pcie-rcar.c PCIEPRAR(idx)); idx 1062 drivers/pci/controller/pcie-rcar.c rcar_pci_write_reg(pcie, lower_32_bits(cpu_addr), PCIELAR(idx)); idx 1064 drivers/pci/controller/pcie-rcar.c PCIELAMR(idx)); idx 1067 drivers/pci/controller/pcie-rcar.c PCIEPRAR(idx + 1)); idx 1069 drivers/pci/controller/pcie-rcar.c PCIELAR(idx + 1)); idx 1070 drivers/pci/controller/pcie-rcar.c rcar_pci_write_reg(pcie, 0, PCIELAMR(idx + 1)); idx 1074 drivers/pci/controller/pcie-rcar.c idx += 2; idx 1076 drivers/pci/controller/pcie-rcar.c if (idx > MAX_NR_INBOUND_MAPS) { idx 1081 drivers/pci/controller/pcie-rcar.c *index = idx; idx 29 drivers/pci/controller/pcie-tango.c unsigned long status, base, virq, idx, pos = 0; idx 37 drivers/pci/controller/pcie-tango.c for_each_set_bit(idx, &status, 32) { idx 38 drivers/pci/controller/pcie-tango.c virq = irq_find_mapping(pcie->dom, base + idx); idx 723 drivers/pci/controller/vmd.c int idx; idx 725 drivers/pci/controller/vmd.c idx = srcu_read_lock(&irqs->srcu); idx 728 drivers/pci/controller/vmd.c srcu_read_unlock(&irqs->srcu, idx); idx 1595 drivers/pci/pci-driver.c int idx = 0; idx 1601 drivers/pci/pci-driver.c envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; idx 1602 drivers/pci/pci-driver.c envp[idx++] = "DEVICE_ONLINE=0"; idx 1605 drivers/pci/pci-driver.c envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; idx 1606 drivers/pci/pci-driver.c envp[idx++] = "DEVICE_ONLINE=1"; idx 1609 drivers/pci/pci-driver.c envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; idx 1610 drivers/pci/pci-driver.c envp[idx++] = "DEVICE_ONLINE=0"; idx 1616 drivers/pci/pci-driver.c if (idx > 0) { idx 1617 drivers/pci/pci-driver.c envp[idx++] = NULL; idx 286 drivers/pci/pci.h bool pci_bus_clip_resource(struct pci_dev *dev, int idx); idx 220 drivers/pci/setup-bus.c int idx; idx 240 drivers/pci/setup-bus.c idx = res - &add_res->dev->resource[0]; idx 246 drivers/pci/setup-bus.c if (pci_assign_resource(add_res->dev, idx)) idx 251 drivers/pci/setup-bus.c if (pci_reassign_resource(add_res->dev, idx, idx 254 drivers/pci/setup-bus.c (unsigned long long) add_size, idx, idx 278 drivers/pci/setup-bus.c int idx; idx 282 drivers/pci/setup-bus.c idx = res - &dev_res->dev->resource[0]; idx 284 drivers/pci/setup-bus.c pci_assign_resource(dev_res->dev, idx)) { idx 291 drivers/pci/setup-bus.c if (!((idx == PCI_ROM_RESOURCE) && idx 1492 drivers/pci/setup-bus.c int idx = 1; idx 1507 drivers/pci/setup-bus.c idx = 0; idx 1509 drivers/pci/setup-bus.c idx = 1; idx 1512 drivers/pci/setup-bus.c idx = 2; idx 1515 drivers/pci/setup-bus.c idx = 2; idx 1517 drivers/pci/setup-bus.c idx = 1; idx 1519 drivers/pci/setup-bus.c r = &b_res[idx]; idx 1529 drivers/pci/setup-bus.c PCI_BRIDGE_RESOURCES + idx, r); idx 1788 drivers/pci/setup-bus.c int idx; idx 1795 drivers/pci/setup-bus.c idx = res - &fail_res->dev->resource[0]; idx 1796 drivers/pci/setup-bus.c if (idx >= PCI_BRIDGE_RESOURCES && idx 1797 drivers/pci/setup-bus.c idx <= PCI_BRIDGE_RESOURCE_END) idx 2046 drivers/pci/setup-bus.c int idx; idx 2053 drivers/pci/setup-bus.c idx = res - &fail_res->dev->resource[0]; idx 2054 drivers/pci/setup-bus.c if (idx >= PCI_BRIDGE_RESOURCES && idx 2055 drivers/pci/setup-bus.c idx <= PCI_BRIDGE_RESOURCE_END) idx 189 drivers/pci/setup-res.c resource_size_t __weak pcibios_retrieve_fw_addr(struct pci_dev *dev, int idx) idx 1043 drivers/pci/switch/switchtec.c int idx; idx 1048 drivers/pci/switch/switchtec.c for (idx = 0; idx < stdev->pff_csr_count; idx++) { idx 1049 drivers/pci/switch/switchtec.c reg = ioread32(&stdev->mmio_pff_csr[idx].link_state_hdr); idx 1050 drivers/pci/switch/switchtec.c dev_dbg(&stdev->dev, "link_state: %d->%08x\n", idx, reg); idx 1053 drivers/pci/switch/switchtec.c if (count != stdev->link_event_count[idx]) { idx 1055 drivers/pci/switch/switchtec.c stdev->link_event_count[idx] = count; idx 1065 drivers/pci/switch/switchtec.c int idx; idx 1067 drivers/pci/switch/switchtec.c for (idx = 0; idx < stdev->pff_csr_count; idx++) { idx 1070 drivers/pci/switch/switchtec.c &stdev->mmio_pff_csr[idx].link_state_hdr); idx 1172 drivers/pci/switch/switchtec.c static int mask_event(struct switchtec_dev *stdev, int eid, int idx) idx 1178 drivers/pci/switch/switchtec.c hdr_reg = event_regs[eid].map_reg(stdev, off, idx); idx 1188 drivers/pci/switch/switchtec.c dev_dbg(&stdev->dev, "%s: %d %d %x\n", __func__, eid, idx, hdr); idx 1197 drivers/pci/switch/switchtec.c int idx; idx 1201 drivers/pci/switch/switchtec.c for (idx = 0; idx < stdev->partition_count; idx++) idx 1202 drivers/pci/switch/switchtec.c count += mask_event(stdev, eid, idx); idx 1204 drivers/pci/switch/switchtec.c for (idx = 0; idx < stdev->pff_csr_count; idx++) { idx 1205 drivers/pci/switch/switchtec.c if (!stdev->pff_local[idx]) idx 1208 drivers/pci/switch/switchtec.c count += mask_event(stdev, eid, idx); idx 39 drivers/perf/arm-cci.c #define CCI_PMU_CNTR_BASE(model, idx) ((idx) * CCI_PMU_CNTR_SIZE(model)) idx 316 drivers/perf/arm-cci.c int idx; idx 326 drivers/perf/arm-cci.c for (idx = CCI400_PMU_CNTR0_IDX; idx <= CCI_PMU_CNTR_LAST(cci_pmu); ++idx) idx 327 drivers/perf/arm-cci.c if (!test_and_set_bit(idx, hw->used_mask)) idx 328 drivers/perf/arm-cci.c return idx; idx 712 drivers/perf/arm-cci.c static int pmu_is_valid_counter(struct cci_pmu *cci_pmu, int idx) idx 714 drivers/perf/arm-cci.c return 0 <= idx && idx <= CCI_PMU_CNTR_LAST(cci_pmu); idx 717 drivers/perf/arm-cci.c static u32 pmu_read_register(struct cci_pmu *cci_pmu, int idx, unsigned int offset) idx 720 drivers/perf/arm-cci.c CCI_PMU_CNTR_BASE(cci_pmu->model, idx) + offset); idx 724 drivers/perf/arm-cci.c int idx, unsigned int offset) idx 727 drivers/perf/arm-cci.c CCI_PMU_CNTR_BASE(cci_pmu->model, idx) + offset); idx 730 drivers/perf/arm-cci.c static void pmu_disable_counter(struct cci_pmu *cci_pmu, int idx) idx 732 drivers/perf/arm-cci.c pmu_write_register(cci_pmu, 0, idx, CCI_PMU_CNTR_CTRL); idx 735 drivers/perf/arm-cci.c static void pmu_enable_counter(struct cci_pmu *cci_pmu, int idx) idx 737 drivers/perf/arm-cci.c pmu_write_register(cci_pmu, 1, idx, CCI_PMU_CNTR_CTRL); idx 741 drivers/perf/arm-cci.c pmu_counter_is_enabled(struct cci_pmu *cci_pmu, int idx) idx 743 drivers/perf/arm-cci.c return (pmu_read_register(cci_pmu, idx, CCI_PMU_CNTR_CTRL) & 0x1) != 0; idx 746 drivers/perf/arm-cci.c static void pmu_set_event(struct cci_pmu *cci_pmu, int idx, unsigned long event) idx 748 drivers/perf/arm-cci.c pmu_write_register(cci_pmu, event, idx, CCI_PMU_EVT_SEL); idx 803 drivers/perf/arm-cci.c int idx; idx 809 drivers/perf/arm-cci.c for(idx = 0; idx <= CCI_PMU_CNTR_LAST(cci_pmu); idx++) idx 810 drivers/perf/arm-cci.c if (!test_and_set_bit(idx, hw->used_mask)) idx 811 drivers/perf/arm-cci.c return idx; idx 879 drivers/perf/arm-cci.c int idx = hw_counter->idx; idx 882 drivers/perf/arm-cci.c if (unlikely(!pmu_is_valid_counter(cci_pmu, idx))) { idx 883 drivers/perf/arm-cci.c dev_err(&cci_pmu->plat_device->dev, "Invalid CCI PMU counter %d\n", idx); idx 886 drivers/perf/arm-cci.c value = pmu_read_register(cci_pmu, idx, CCI_PMU_CNTR); idx 891 drivers/perf/arm-cci.c static void pmu_write_counter(struct cci_pmu *cci_pmu, u32 value, int idx) idx 893 drivers/perf/arm-cci.c pmu_write_register(cci_pmu, value, idx, CCI_PMU_CNTR); idx 1032 drivers/perf/arm-cci.c int idx, handled = IRQ_NONE; idx 1043 drivers/perf/arm-cci.c for (idx = 0; idx <= CCI_PMU_CNTR_LAST(cci_pmu); idx++) { idx 1044 drivers/perf/arm-cci.c struct perf_event *event = events->events[idx]; idx 1050 drivers/perf/arm-cci.c if (!(pmu_read_register(cci_pmu, idx, CCI_PMU_OVRFLW) & idx 1054 drivers/perf/arm-cci.c pmu_write_register(cci_pmu, CCI_PMU_OVRFLW_FLAG, idx, idx 1128 drivers/perf/arm-cci.c static bool pmu_fixed_hw_idx(struct cci_pmu *cci_pmu, int idx) idx 1130 drivers/perf/arm-cci.c return (idx >= 0) && (idx < cci_pmu->model->fixed_hw_cntrs); idx 1138 drivers/perf/arm-cci.c int idx = hwc->idx; idx 1150 drivers/perf/arm-cci.c if (unlikely(!pmu_is_valid_counter(cci_pmu, idx))) { idx 1151 drivers/perf/arm-cci.c dev_err(&cci_pmu->plat_device->dev, "Invalid CCI PMU counter %d\n", idx); idx 1158 drivers/perf/arm-cci.c if (!pmu_fixed_hw_idx(cci_pmu, idx)) idx 1159 drivers/perf/arm-cci.c pmu_set_event(cci_pmu, idx, hwc->config_base); idx 1162 drivers/perf/arm-cci.c pmu_enable_counter(cci_pmu, idx); idx 1171 drivers/perf/arm-cci.c int idx = hwc->idx; idx 1176 drivers/perf/arm-cci.c if (unlikely(!pmu_is_valid_counter(cci_pmu, idx))) { idx 1177 drivers/perf/arm-cci.c dev_err(&cci_pmu->plat_device->dev, "Invalid CCI PMU counter %d\n", idx); idx 1185 drivers/perf/arm-cci.c pmu_disable_counter(cci_pmu, idx); idx 1195 drivers/perf/arm-cci.c int idx; idx 1198 drivers/perf/arm-cci.c idx = pmu_get_event_idx(hw_events, event); idx 1199 drivers/perf/arm-cci.c if (idx < 0) idx 1200 drivers/perf/arm-cci.c return idx; idx 1202 drivers/perf/arm-cci.c event->hw.idx = idx; idx 1203 drivers/perf/arm-cci.c hw_events->events[idx] = event; idx 1220 drivers/perf/arm-cci.c int idx = hwc->idx; idx 1223 drivers/perf/arm-cci.c hw_events->events[idx] = NULL; idx 1224 drivers/perf/arm-cci.c clear_bit(idx, hw_events->used_mask); idx 1299 drivers/perf/arm-cci.c hwc->idx = -1; idx 654 drivers/perf/arm-ccn.c hw->idx = CCN_IDX_PMU_CYCLE_COUNTER; idx 661 drivers/perf/arm-ccn.c hw->idx = arm_ccn_pmu_alloc_bit(ccn->dt.pmu_counters_mask, idx 663 drivers/perf/arm-ccn.c if (hw->idx < 0) { idx 672 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].source = source; idx 684 drivers/perf/arm-ccn.c clear_bit(hw->idx, ccn->dt.pmu_counters_mask); idx 689 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].event = event; idx 699 drivers/perf/arm-ccn.c if (hw->idx == CCN_IDX_PMU_CYCLE_COUNTER) { idx 703 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].source; idx 711 drivers/perf/arm-ccn.c clear_bit(hw->idx, ccn->dt.pmu_counters_mask); idx 714 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].source = NULL; idx 715 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].event = NULL; idx 849 drivers/perf/arm-ccn.c static u64 arm_ccn_pmu_read_counter(struct arm_ccn *ccn, int idx) idx 853 drivers/perf/arm-ccn.c if (idx == CCN_IDX_PMU_CYCLE_COUNTER) { idx 867 drivers/perf/arm-ccn.c res = readl(ccn->dt.base + CCN_DT_PMEVCNT(idx)); idx 881 drivers/perf/arm-ccn.c new_count = arm_ccn_pmu_read_counter(ccn, hw->idx); idx 884 drivers/perf/arm-ccn.c mask = (1LLU << (hw->idx == CCN_IDX_PMU_CYCLE_COUNTER ? 40 : 32)) - 1; idx 897 drivers/perf/arm-ccn.c if (hw->idx == CCN_IDX_PMU_CYCLE_COUNTER) idx 915 drivers/perf/arm-ccn.c CCN_XP_DT_CONFIG__DT_CFG__SHIFT(hw->idx)); idx 916 drivers/perf/arm-ccn.c val |= dt_cfg << CCN_XP_DT_CONFIG__DT_CFG__SHIFT(hw->idx); idx 928 drivers/perf/arm-ccn.c arm_ccn_pmu_read_counter(ccn, hw->idx)); idx 953 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].source; idx 1001 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].source; idx 1022 drivers/perf/arm-ccn.c ccn->dt.pmu_counters[hw->idx].source; idx 1061 drivers/perf/arm-ccn.c if (hw->idx == CCN_IDX_PMU_CYCLE_COUNTER) idx 1072 drivers/perf/arm-ccn.c offset = (hw->idx / 4) * 4; idx 1075 drivers/perf/arm-ccn.c CCN_DT_ACTIVE_DSM__DSM_ID__SHIFT(hw->idx % 4)); idx 1076 drivers/perf/arm-ccn.c val |= xp << CCN_DT_ACTIVE_DSM__DSM_ID__SHIFT(hw->idx % 4); idx 1165 drivers/perf/arm-ccn.c int idx; idx 1174 drivers/perf/arm-ccn.c for (idx = 0; idx < CCN_NUM_PMU_EVENT_COUNTERS + 1; idx++) { idx 1175 drivers/perf/arm-ccn.c struct perf_event *event = dt->pmu_counters[idx].event; idx 1176 drivers/perf/arm-ccn.c int overflowed = pmovsr & BIT(idx); idx 1179 drivers/perf/arm-ccn.c idx != CCN_IDX_PMU_CYCLE_COUNTER); idx 69 drivers/perf/arm_dsu_pmu.c #define DSU_PMU_COUNTER_WIDTH(idx) \ idx 70 drivers/perf/arm_dsu_pmu.c (((idx) == DSU_PMU_IDX_CYCLE_COUNTER) ? 64 : 32) idx 72 drivers/perf/arm_dsu_pmu.c #define DSU_PMU_COUNTER_MASK(idx) \ idx 73 drivers/perf/arm_dsu_pmu.c GENMASK_ULL((DSU_PMU_COUNTER_WIDTH((idx)) - 1), 0) idx 242 drivers/perf/arm_dsu_pmu.c static inline bool dsu_pmu_counter_valid(struct dsu_pmu *dsu_pmu, u32 idx) idx 244 drivers/perf/arm_dsu_pmu.c return (idx < dsu_pmu->num_counters) || idx 245 drivers/perf/arm_dsu_pmu.c (idx == DSU_PMU_IDX_CYCLE_COUNTER); idx 253 drivers/perf/arm_dsu_pmu.c int idx = event->hw.idx; idx 259 drivers/perf/arm_dsu_pmu.c if (!dsu_pmu_counter_valid(dsu_pmu, idx)) { idx 261 drivers/perf/arm_dsu_pmu.c "Trying reading invalid counter %d\n", idx); idx 266 drivers/perf/arm_dsu_pmu.c if (idx == DSU_PMU_IDX_CYCLE_COUNTER) idx 269 drivers/perf/arm_dsu_pmu.c val = __dsu_pmu_read_counter(idx); idx 279 drivers/perf/arm_dsu_pmu.c int idx = event->hw.idx; idx 285 drivers/perf/arm_dsu_pmu.c if (!dsu_pmu_counter_valid(dsu_pmu, idx)) { idx 287 drivers/perf/arm_dsu_pmu.c "writing to invalid counter %d\n", idx); idx 292 drivers/perf/arm_dsu_pmu.c if (idx == DSU_PMU_IDX_CYCLE_COUNTER) idx 295 drivers/perf/arm_dsu_pmu.c __dsu_pmu_write_counter(idx, val); idx 302 drivers/perf/arm_dsu_pmu.c int idx; idx 313 drivers/perf/arm_dsu_pmu.c idx = find_first_zero_bit(used_mask, dsu_pmu->num_counters); idx 314 drivers/perf/arm_dsu_pmu.c if (idx >= dsu_pmu->num_counters) idx 316 drivers/perf/arm_dsu_pmu.c set_bit(idx, hw_events->used_mask); idx 317 drivers/perf/arm_dsu_pmu.c return idx; idx 320 drivers/perf/arm_dsu_pmu.c static void dsu_pmu_enable_counter(struct dsu_pmu *dsu_pmu, int idx) idx 322 drivers/perf/arm_dsu_pmu.c __dsu_pmu_counter_interrupt_enable(idx); idx 323 drivers/perf/arm_dsu_pmu.c __dsu_pmu_enable_counter(idx); idx 326 drivers/perf/arm_dsu_pmu.c static void dsu_pmu_disable_counter(struct dsu_pmu *dsu_pmu, int idx) idx 328 drivers/perf/arm_dsu_pmu.c __dsu_pmu_disable_counter(idx); idx 329 drivers/perf/arm_dsu_pmu.c __dsu_pmu_counter_interrupt_disable(idx); idx 335 drivers/perf/arm_dsu_pmu.c int idx = event->hw.idx; idx 338 drivers/perf/arm_dsu_pmu.c if (!dsu_pmu_counter_valid(dsu_pmu, idx)) { idx 340 drivers/perf/arm_dsu_pmu.c "Trying to set invalid counter %d\n", idx); idx 345 drivers/perf/arm_dsu_pmu.c __dsu_pmu_set_event(idx, event->hw.config_base); idx 360 drivers/perf/arm_dsu_pmu.c delta = (new_count - prev_count) & DSU_PMU_COUNTER_MASK(hwc->idx); idx 383 drivers/perf/arm_dsu_pmu.c int idx = event->hw.idx; idx 384 drivers/perf/arm_dsu_pmu.c u64 val = DSU_PMU_COUNTER_MASK(idx) >> 1; idx 423 drivers/perf/arm_dsu_pmu.c if (event->hw.idx != DSU_PMU_IDX_CYCLE_COUNTER) idx 426 drivers/perf/arm_dsu_pmu.c dsu_pmu_enable_counter(dsu_pmu, event->hw.idx); idx 435 drivers/perf/arm_dsu_pmu.c dsu_pmu_disable_counter(dsu_pmu, event->hw.idx); idx 445 drivers/perf/arm_dsu_pmu.c int idx; idx 451 drivers/perf/arm_dsu_pmu.c idx = dsu_pmu_get_event_idx(hw_events, event); idx 452 drivers/perf/arm_dsu_pmu.c if (idx < 0) idx 453 drivers/perf/arm_dsu_pmu.c return idx; idx 455 drivers/perf/arm_dsu_pmu.c hwc->idx = idx; idx 456 drivers/perf/arm_dsu_pmu.c hw_events->events[idx] = event; idx 471 drivers/perf/arm_dsu_pmu.c int idx = hwc->idx; idx 474 drivers/perf/arm_dsu_pmu.c hw_events->events[idx] = NULL; idx 475 drivers/perf/arm_dsu_pmu.c clear_bit(idx, hw_events->used_mask); idx 238 drivers/perf/arm_pmu.c int idx = hwc->idx; idx 241 drivers/perf/arm_pmu.c hw_events->events[idx] = NULL; idx 245 drivers/perf/arm_pmu.c hwc->idx = -1; idx 254 drivers/perf/arm_pmu.c int idx; idx 261 drivers/perf/arm_pmu.c idx = armpmu->get_event_idx(hw_events, event); idx 262 drivers/perf/arm_pmu.c if (idx < 0) idx 263 drivers/perf/arm_pmu.c return idx; idx 269 drivers/perf/arm_pmu.c event->hw.idx = idx; idx 271 drivers/perf/arm_pmu.c hw_events->events[idx] = event; idx 382 drivers/perf/arm_pmu.c hwc->idx = -1; idx 664 drivers/perf/arm_pmu.c int idx; idx 666 drivers/perf/arm_pmu.c for (idx = 0; idx < armpmu->num_events; idx++) { idx 667 drivers/perf/arm_pmu.c event = hw_events->events[idx]; idx 150 drivers/perf/arm_smmuv3_pmu.c u32 idx, u64 value) idx 153 drivers/perf/arm_smmuv3_pmu.c writeq(value, smmu_pmu->reloc_base + SMMU_PMCG_EVCNTR(idx, 8)); idx 155 drivers/perf/arm_smmuv3_pmu.c writel(value, smmu_pmu->reloc_base + SMMU_PMCG_EVCNTR(idx, 4)); idx 158 drivers/perf/arm_smmuv3_pmu.c static inline u64 smmu_pmu_counter_get_value(struct smmu_pmu *smmu_pmu, u32 idx) idx 163 drivers/perf/arm_smmuv3_pmu.c value = readq(smmu_pmu->reloc_base + SMMU_PMCG_EVCNTR(idx, 8)); idx 165 drivers/perf/arm_smmuv3_pmu.c value = readl(smmu_pmu->reloc_base + SMMU_PMCG_EVCNTR(idx, 4)); idx 170 drivers/perf/arm_smmuv3_pmu.c static inline void smmu_pmu_counter_enable(struct smmu_pmu *smmu_pmu, u32 idx) idx 172 drivers/perf/arm_smmuv3_pmu.c writeq(BIT(idx), smmu_pmu->reg_base + SMMU_PMCG_CNTENSET0); idx 175 drivers/perf/arm_smmuv3_pmu.c static inline void smmu_pmu_counter_disable(struct smmu_pmu *smmu_pmu, u32 idx) idx 177 drivers/perf/arm_smmuv3_pmu.c writeq(BIT(idx), smmu_pmu->reg_base + SMMU_PMCG_CNTENCLR0); idx 180 drivers/perf/arm_smmuv3_pmu.c static inline void smmu_pmu_interrupt_enable(struct smmu_pmu *smmu_pmu, u32 idx) idx 182 drivers/perf/arm_smmuv3_pmu.c writeq(BIT(idx), smmu_pmu->reg_base + SMMU_PMCG_INTENSET0); idx 186 drivers/perf/arm_smmuv3_pmu.c u32 idx) idx 188 drivers/perf/arm_smmuv3_pmu.c writeq(BIT(idx), smmu_pmu->reg_base + SMMU_PMCG_INTENCLR0); idx 191 drivers/perf/arm_smmuv3_pmu.c static inline void smmu_pmu_set_evtyper(struct smmu_pmu *smmu_pmu, u32 idx, idx 194 drivers/perf/arm_smmuv3_pmu.c writel(val, smmu_pmu->reg_base + SMMU_PMCG_EVTYPER(idx)); idx 197 drivers/perf/arm_smmuv3_pmu.c static inline void smmu_pmu_set_smr(struct smmu_pmu *smmu_pmu, u32 idx, u32 val) idx 199 drivers/perf/arm_smmuv3_pmu.c writel(val, smmu_pmu->reg_base + SMMU_PMCG_SMR(idx)); idx 207 drivers/perf/arm_smmuv3_pmu.c u32 idx = hwc->idx; idx 211 drivers/perf/arm_smmuv3_pmu.c now = smmu_pmu_counter_get_value(smmu_pmu, idx); idx 224 drivers/perf/arm_smmuv3_pmu.c u32 idx = hwc->idx; idx 235 drivers/perf/arm_smmuv3_pmu.c new = smmu_pmu_counter_get_value(smmu_pmu, idx); idx 244 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_counter_set_value(smmu_pmu, idx, new); idx 251 drivers/perf/arm_smmuv3_pmu.c int idx, u32 span, u32 sid) idx 257 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_set_evtyper(smmu_pmu, idx, evtyper); idx 258 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_set_smr(smmu_pmu, idx, sid); idx 275 drivers/perf/arm_smmuv3_pmu.c struct perf_event *event, int idx) idx 288 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_set_event_filter(event, idx, span, sid); idx 293 drivers/perf/arm_smmuv3_pmu.c idx = find_first_bit(smmu_pmu->used_counters, num_ctrs); idx 294 drivers/perf/arm_smmuv3_pmu.c if (idx == num_ctrs || idx 295 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_check_global_filter(smmu_pmu->events[idx], event)) { idx 306 drivers/perf/arm_smmuv3_pmu.c int idx, err; idx 309 drivers/perf/arm_smmuv3_pmu.c idx = find_first_zero_bit(smmu_pmu->used_counters, num_ctrs); idx 310 drivers/perf/arm_smmuv3_pmu.c if (idx == num_ctrs) idx 314 drivers/perf/arm_smmuv3_pmu.c err = smmu_pmu_apply_event_filter(smmu_pmu, event, idx); idx 318 drivers/perf/arm_smmuv3_pmu.c set_bit(idx, smmu_pmu->used_counters); idx 320 drivers/perf/arm_smmuv3_pmu.c return idx; idx 391 drivers/perf/arm_smmuv3_pmu.c hwc->idx = -1; idx 406 drivers/perf/arm_smmuv3_pmu.c int idx = hwc->idx; idx 412 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_counter_enable(smmu_pmu, idx); idx 419 drivers/perf/arm_smmuv3_pmu.c int idx = hwc->idx; idx 424 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_counter_disable(smmu_pmu, idx); idx 433 drivers/perf/arm_smmuv3_pmu.c int idx; idx 436 drivers/perf/arm_smmuv3_pmu.c idx = smmu_pmu_get_event_idx(smmu_pmu, event); idx 437 drivers/perf/arm_smmuv3_pmu.c if (idx < 0) idx 438 drivers/perf/arm_smmuv3_pmu.c return idx; idx 440 drivers/perf/arm_smmuv3_pmu.c hwc->idx = idx; idx 442 drivers/perf/arm_smmuv3_pmu.c smmu_pmu->events[idx] = event; idx 445 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_interrupt_enable(smmu_pmu, idx); idx 460 drivers/perf/arm_smmuv3_pmu.c int idx = hwc->idx; idx 463 drivers/perf/arm_smmuv3_pmu.c smmu_pmu_interrupt_disable(smmu_pmu, idx); idx 464 drivers/perf/arm_smmuv3_pmu.c smmu_pmu->events[idx] = NULL; idx 465 drivers/perf/arm_smmuv3_pmu.c clear_bit(idx, smmu_pmu->used_counters); idx 609 drivers/perf/arm_smmuv3_pmu.c unsigned int idx; idx 617 drivers/perf/arm_smmuv3_pmu.c for_each_set_bit(idx, (unsigned long *)&ovsr, smmu_pmu->num_counters) { idx 618 drivers/perf/arm_smmuv3_pmu.c struct perf_event *event = smmu_pmu->events[idx]; idx 78 drivers/perf/arm_spe_pmu.c #define PERF_IDX2OFF(idx, buf) ((idx) % ((buf)->nr_pages << PAGE_SHIFT)) idx 881 drivers/perf/arm_spe_pmu.c int idx; idx 910 drivers/perf/arm_spe_pmu.c idx = atomic_inc_return(&pmu_idx); idx 911 drivers/perf/arm_spe_pmu.c name = devm_kasprintf(dev, GFP_KERNEL, "%s_%d", PMUNAME, idx); idx 913 drivers/perf/arm_spe_pmu.c dev_err(dev, "failed to allocate name for pmu %d\n", idx); idx 298 drivers/perf/fsl_imx8_ddr_perf.c hwc->idx = -1; idx 309 drivers/perf/fsl_imx8_ddr_perf.c int counter = hwc->idx; idx 350 drivers/perf/fsl_imx8_ddr_perf.c int counter = hwc->idx; idx 391 drivers/perf/fsl_imx8_ddr_perf.c hwc->idx = counter; idx 405 drivers/perf/fsl_imx8_ddr_perf.c int counter = hwc->idx; idx 417 drivers/perf/fsl_imx8_ddr_perf.c int counter = hwc->idx; idx 423 drivers/perf/fsl_imx8_ddr_perf.c hwc->idx = -1; idx 504 drivers/perf/fsl_imx8_ddr_perf.c if (event->hw.idx == EVENT_CYCLES_COUNTER) idx 68 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c u32 idx = GET_DDRC_EVENTID(hwc); idx 70 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c if (!hisi_uncore_pmu_counter_valid(ddrc_pmu, idx)) { idx 71 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c dev_err(ddrc_pmu->dev, "Unsupported event index:%d!\n", idx); idx 75 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c return readl(ddrc_pmu->base + hisi_ddrc_pmu_get_counter_offset(idx)); idx 81 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c u32 idx = GET_DDRC_EVENTID(hwc); idx 83 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c if (!hisi_uncore_pmu_counter_valid(ddrc_pmu, idx)) { idx 84 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c dev_err(ddrc_pmu->dev, "Unsupported event index:%d!\n", idx); idx 89 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c ddrc_pmu->base + hisi_ddrc_pmu_get_counter_offset(idx)); idx 96 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c static void hisi_ddrc_pmu_write_evtype(struct hisi_pmu *hha_pmu, int idx, idx 149 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c int idx = GET_DDRC_EVENTID(hwc); idx 151 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c if (test_bit(idx, used_mask)) idx 154 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c set_bit(idx, used_mask); idx 156 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c return idx; idx 186 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c int idx; idx 197 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c for_each_set_bit(idx, &overflown, DDRC_NR_COUNTERS) { idx 199 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c writel((1 << idx), ddrc_pmu->base + DDRC_INT_CLEAR); idx 202 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c event = ddrc_pmu->pmu_events.hw_events[idx]; idx 53 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c u32 idx = hwc->idx; idx 55 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c if (!hisi_uncore_pmu_counter_valid(hha_pmu, idx)) { idx 56 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c dev_err(hha_pmu->dev, "Unsupported event index:%d!\n", idx); idx 61 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c return readq(hha_pmu->base + hisi_hha_pmu_get_counter_offset(idx)); idx 67 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c u32 idx = hwc->idx; idx 69 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c if (!hisi_uncore_pmu_counter_valid(hha_pmu, idx)) { idx 70 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c dev_err(hha_pmu->dev, "Unsupported event index:%d!\n", idx); idx 75 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c writeq(val, hha_pmu->base + hisi_hha_pmu_get_counter_offset(idx)); idx 78 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c static void hisi_hha_pmu_write_evtype(struct hisi_pmu *hha_pmu, int idx, idx 90 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c reg = HHA_EVENT_TYPE0 + 4 * (idx / 4); idx 91 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c reg_idx = idx % 4; idx 134 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c val |= (1 << hwc->idx); idx 145 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c val &= ~(1 << hwc->idx); idx 156 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c val &= ~(1 << hwc->idx); idx 167 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c val |= (1 << hwc->idx); idx 176 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c int idx; idx 187 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c for_each_set_bit(idx, &overflown, HHA_NR_COUNTERS) { idx 189 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c writel((1 << idx), hha_pmu->base + HHA_INT_CLEAR); idx 192 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c event = hha_pmu->pmu_events.hw_events[idx]; idx 52 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c u32 idx = hwc->idx; idx 54 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c if (!hisi_uncore_pmu_counter_valid(l3c_pmu, idx)) { idx 55 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c dev_err(l3c_pmu->dev, "Unsupported event index:%d!\n", idx); idx 60 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c return readq(l3c_pmu->base + hisi_l3c_pmu_get_counter_offset(idx)); idx 66 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c u32 idx = hwc->idx; idx 68 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c if (!hisi_uncore_pmu_counter_valid(l3c_pmu, idx)) { idx 69 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c dev_err(l3c_pmu->dev, "Unsupported event index:%d!\n", idx); idx 74 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c writeq(val, l3c_pmu->base + hisi_l3c_pmu_get_counter_offset(idx)); idx 77 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c static void hisi_l3c_pmu_write_evtype(struct hisi_pmu *l3c_pmu, int idx, idx 89 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c reg = L3C_EVENT_TYPE0 + (idx / 4) * 4; idx 90 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c reg_idx = idx % 4; idx 133 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c val |= (1 << hwc->idx); idx 144 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c val &= ~(1 << hwc->idx); idx 155 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c val &= ~(1 << hwc->idx); idx 166 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c val |= (1 << hwc->idx); idx 175 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c int idx; idx 186 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c for_each_set_bit(idx, &overflown, L3C_NR_COUNTERS) { idx 188 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c writel((1 << idx), l3c_pmu->base + L3C_INT_CLEAR); idx 191 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c event = l3c_pmu->pmu_events.hw_events[idx]; idx 95 drivers/perf/hisilicon/hisi_uncore_pmu.c int hisi_uncore_pmu_counter_valid(struct hisi_pmu *hisi_pmu, int idx) idx 97 drivers/perf/hisilicon/hisi_uncore_pmu.c return idx >= 0 && idx < hisi_pmu->num_counters; idx 105 drivers/perf/hisilicon/hisi_uncore_pmu.c int idx; idx 107 drivers/perf/hisilicon/hisi_uncore_pmu.c idx = find_first_zero_bit(used_mask, num_counters); idx 108 drivers/perf/hisilicon/hisi_uncore_pmu.c if (idx == num_counters) idx 111 drivers/perf/hisilicon/hisi_uncore_pmu.c set_bit(idx, used_mask); idx 113 drivers/perf/hisilicon/hisi_uncore_pmu.c return idx; idx 116 drivers/perf/hisilicon/hisi_uncore_pmu.c static void hisi_uncore_pmu_clear_event_idx(struct hisi_pmu *hisi_pmu, int idx) idx 118 drivers/perf/hisilicon/hisi_uncore_pmu.c if (!hisi_uncore_pmu_counter_valid(hisi_pmu, idx)) { idx 119 drivers/perf/hisilicon/hisi_uncore_pmu.c dev_err(hisi_pmu->dev, "Unsupported event index:%d!\n", idx); idx 123 drivers/perf/hisilicon/hisi_uncore_pmu.c clear_bit(idx, hisi_pmu->pmu_events.used_mask); idx 167 drivers/perf/hisilicon/hisi_uncore_pmu.c hwc->idx = -1; idx 185 drivers/perf/hisilicon/hisi_uncore_pmu.c hisi_pmu->ops->write_evtype(hisi_pmu, hwc->idx, idx 285 drivers/perf/hisilicon/hisi_uncore_pmu.c int idx; idx 290 drivers/perf/hisilicon/hisi_uncore_pmu.c idx = hisi_pmu->ops->get_event_idx(event); idx 291 drivers/perf/hisilicon/hisi_uncore_pmu.c if (idx < 0) idx 292 drivers/perf/hisilicon/hisi_uncore_pmu.c return idx; idx 294 drivers/perf/hisilicon/hisi_uncore_pmu.c event->hw.idx = idx; idx 295 drivers/perf/hisilicon/hisi_uncore_pmu.c hisi_pmu->pmu_events.hw_events[idx] = event; idx 309 drivers/perf/hisilicon/hisi_uncore_pmu.c hisi_uncore_pmu_clear_event_idx(hisi_pmu, hwc->idx); idx 311 drivers/perf/hisilicon/hisi_uncore_pmu.c hisi_pmu->pmu_events.hw_events[hwc->idx] = NULL; idx 79 drivers/perf/hisilicon/hisi_uncore_pmu.h int hisi_uncore_pmu_counter_valid(struct hisi_pmu *hisi_pmu, int idx); idx 197 drivers/perf/qcom_l2_pmu.c static inline u32 idx_to_reg_bit(u32 idx) idx 199 drivers/perf/qcom_l2_pmu.c if (idx == l2_cycle_ctr_idx) idx 202 drivers/perf/qcom_l2_pmu.c return BIT(idx); idx 230 drivers/perf/qcom_l2_pmu.c static inline void cluster_pmu_counter_set_value(u32 idx, u64 value) idx 232 drivers/perf/qcom_l2_pmu.c if (idx == l2_cycle_ctr_idx) idx 235 drivers/perf/qcom_l2_pmu.c set_l2_indirect_reg(reg_idx(IA_L2PMXEVCNTR, idx), value); idx 238 drivers/perf/qcom_l2_pmu.c static inline u64 cluster_pmu_counter_get_value(u32 idx) idx 242 drivers/perf/qcom_l2_pmu.c if (idx == l2_cycle_ctr_idx) idx 245 drivers/perf/qcom_l2_pmu.c value = get_l2_indirect_reg(reg_idx(IA_L2PMXEVCNTR, idx)); idx 250 drivers/perf/qcom_l2_pmu.c static inline void cluster_pmu_counter_enable(u32 idx) idx 252 drivers/perf/qcom_l2_pmu.c set_l2_indirect_reg(L2PMCNTENSET, idx_to_reg_bit(idx)); idx 255 drivers/perf/qcom_l2_pmu.c static inline void cluster_pmu_counter_disable(u32 idx) idx 257 drivers/perf/qcom_l2_pmu.c set_l2_indirect_reg(L2PMCNTENCLR, idx_to_reg_bit(idx)); idx 260 drivers/perf/qcom_l2_pmu.c static inline void cluster_pmu_counter_enable_interrupt(u32 idx) idx 262 drivers/perf/qcom_l2_pmu.c set_l2_indirect_reg(L2PMINTENSET, idx_to_reg_bit(idx)); idx 265 drivers/perf/qcom_l2_pmu.c static inline void cluster_pmu_counter_disable_interrupt(u32 idx) idx 267 drivers/perf/qcom_l2_pmu.c set_l2_indirect_reg(L2PMINTENCLR, idx_to_reg_bit(idx)); idx 334 drivers/perf/qcom_l2_pmu.c static inline bool cluster_pmu_counter_has_overflowed(u32 ovsr, u32 idx) idx 336 drivers/perf/qcom_l2_pmu.c return !!(ovsr & idx_to_reg_bit(idx)); idx 343 drivers/perf/qcom_l2_pmu.c u32 idx = hwc->idx; idx 347 drivers/perf/qcom_l2_pmu.c now = cluster_pmu_counter_get_value(idx); idx 355 drivers/perf/qcom_l2_pmu.c if (idx != l2_cycle_ctr_idx) idx 364 drivers/perf/qcom_l2_pmu.c u32 idx = hwc->idx; idx 372 drivers/perf/qcom_l2_pmu.c if (idx == l2_cycle_ctr_idx) idx 378 drivers/perf/qcom_l2_pmu.c cluster_pmu_counter_set_value(idx, new); idx 385 drivers/perf/qcom_l2_pmu.c int idx; idx 396 drivers/perf/qcom_l2_pmu.c idx = find_first_zero_bit(cluster->used_counters, num_ctrs); idx 397 drivers/perf/qcom_l2_pmu.c if (idx == num_ctrs) idx 410 drivers/perf/qcom_l2_pmu.c set_bit(idx, cluster->used_counters); idx 413 drivers/perf/qcom_l2_pmu.c return idx; idx 420 drivers/perf/qcom_l2_pmu.c int idx = hwc->idx; idx 422 drivers/perf/qcom_l2_pmu.c clear_bit(idx, cluster->used_counters); idx 432 drivers/perf/qcom_l2_pmu.c int idx; idx 438 drivers/perf/qcom_l2_pmu.c for_each_set_bit(idx, cluster->used_counters, num_counters) { idx 439 drivers/perf/qcom_l2_pmu.c struct perf_event *event = cluster->events[idx]; idx 445 drivers/perf/qcom_l2_pmu.c if (!cluster_pmu_counter_has_overflowed(ovsr, idx)) idx 571 drivers/perf/qcom_l2_pmu.c hwc->idx = -1; idx 587 drivers/perf/qcom_l2_pmu.c int idx = hwc->idx; idx 604 drivers/perf/qcom_l2_pmu.c cluster_pmu_set_evcntcr(idx, 0); idx 605 drivers/perf/qcom_l2_pmu.c cluster_pmu_set_evtyper(idx, event_group); idx 607 drivers/perf/qcom_l2_pmu.c cluster_pmu_set_evfilter_sys_mode(idx); idx 610 drivers/perf/qcom_l2_pmu.c cluster_pmu_counter_enable_interrupt(idx); idx 611 drivers/perf/qcom_l2_pmu.c cluster_pmu_counter_enable(idx); idx 617 drivers/perf/qcom_l2_pmu.c int idx = hwc->idx; idx 622 drivers/perf/qcom_l2_pmu.c cluster_pmu_counter_disable_interrupt(idx); idx 623 drivers/perf/qcom_l2_pmu.c cluster_pmu_counter_disable(idx); idx 633 drivers/perf/qcom_l2_pmu.c int idx; idx 639 drivers/perf/qcom_l2_pmu.c idx = l2_cache_get_event_idx(cluster, event); idx 640 drivers/perf/qcom_l2_pmu.c if (idx < 0) idx 641 drivers/perf/qcom_l2_pmu.c return idx; idx 643 drivers/perf/qcom_l2_pmu.c hwc->idx = idx; idx 645 drivers/perf/qcom_l2_pmu.c cluster->events[idx] = event; idx 661 drivers/perf/qcom_l2_pmu.c int idx = hwc->idx; idx 666 drivers/perf/qcom_l2_pmu.c cluster->events[idx] = NULL; idx 196 drivers/perf/qcom_l3_pmu.c int idx = event->hw.idx; idx 202 drivers/perf/qcom_l3_pmu.c gang |= GANG_EN(idx + 1); idx 207 drivers/perf/qcom_l3_pmu.c writel_relaxed(0, l3pmu->regs + L3_HML3_PM_EVCNTR(idx + 1)); idx 208 drivers/perf/qcom_l3_pmu.c writel_relaxed(0, l3pmu->regs + L3_HML3_PM_EVCNTR(idx)); idx 214 drivers/perf/qcom_l3_pmu.c writel_relaxed(EVSEL(0), l3pmu->regs + L3_HML3_PM_EVTYPE(idx + 1)); idx 215 drivers/perf/qcom_l3_pmu.c writel_relaxed(EVSEL(evsel), l3pmu->regs + L3_HML3_PM_EVTYPE(idx)); idx 218 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNT_RESET, l3pmu->regs + L3_HML3_PM_CNTCTL(idx + 1)); idx 219 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNTENSET(idx + 1), l3pmu->regs + L3_M_BC_CNTENSET); idx 220 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNT_RESET, l3pmu->regs + L3_HML3_PM_CNTCTL(idx)); idx 221 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNTENSET(idx), l3pmu->regs + L3_M_BC_CNTENSET); idx 228 drivers/perf/qcom_l3_pmu.c int idx = event->hw.idx; idx 232 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNTENCLR(idx), l3pmu->regs + L3_M_BC_CNTENCLR); idx 233 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNTENCLR(idx + 1), l3pmu->regs + L3_M_BC_CNTENCLR); idx 236 drivers/perf/qcom_l3_pmu.c writel_relaxed(gang & ~GANG_EN(idx + 1), l3pmu->regs + L3_M_BC_GANG); idx 242 drivers/perf/qcom_l3_pmu.c int idx = event->hw.idx; idx 249 drivers/perf/qcom_l3_pmu.c hi = readl_relaxed(l3pmu->regs + L3_HML3_PM_EVCNTR(idx + 1)); idx 250 drivers/perf/qcom_l3_pmu.c lo = readl_relaxed(l3pmu->regs + L3_HML3_PM_EVCNTR(idx)); idx 251 drivers/perf/qcom_l3_pmu.c } while (hi != readl_relaxed(l3pmu->regs + L3_HML3_PM_EVCNTR(idx + 1))); idx 277 drivers/perf/qcom_l3_pmu.c int idx = event->hw.idx; idx 282 drivers/perf/qcom_l3_pmu.c writel_relaxed(irqctl | PMIRQONMSBEN(idx), l3pmu->regs + L3_M_BC_IRQCTL); idx 286 drivers/perf/qcom_l3_pmu.c writel_relaxed(0, l3pmu->regs + L3_HML3_PM_EVCNTR(idx)); idx 289 drivers/perf/qcom_l3_pmu.c writel_relaxed(EVSEL(evsel), l3pmu->regs + L3_HML3_PM_EVTYPE(idx)); idx 292 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMINTENSET(idx), l3pmu->regs + L3_M_BC_INTENSET); idx 295 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNT_RESET, l3pmu->regs + L3_HML3_PM_CNTCTL(idx)); idx 296 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNTENSET(idx), l3pmu->regs + L3_M_BC_CNTENSET); idx 303 drivers/perf/qcom_l3_pmu.c int idx = event->hw.idx; idx 307 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMCNTENCLR(idx), l3pmu->regs + L3_M_BC_CNTENCLR); idx 310 drivers/perf/qcom_l3_pmu.c writel_relaxed(PMINTENCLR(idx), l3pmu->regs + L3_M_BC_INTENCLR); idx 313 drivers/perf/qcom_l3_pmu.c writel_relaxed(irqctl & ~PMIRQONMSBEN(idx), l3pmu->regs + L3_M_BC_IRQCTL); idx 319 drivers/perf/qcom_l3_pmu.c int idx = event->hw.idx; idx 324 drivers/perf/qcom_l3_pmu.c new = readl_relaxed(l3pmu->regs + L3_HML3_PM_EVCNTR(idx)); idx 393 drivers/perf/qcom_l3_pmu.c int idx; idx 401 drivers/perf/qcom_l3_pmu.c for_each_set_bit(idx, &status, L3_NUM_COUNTERS) { idx 405 drivers/perf/qcom_l3_pmu.c event = l3pmu->events[idx]; idx 506 drivers/perf/qcom_l3_pmu.c hwc->idx = -1; idx 552 drivers/perf/qcom_l3_pmu.c int idx; idx 557 drivers/perf/qcom_l3_pmu.c idx = bitmap_find_free_region(l3pmu->used_mask, L3_NUM_COUNTERS, order); idx 558 drivers/perf/qcom_l3_pmu.c if (idx < 0) idx 562 drivers/perf/qcom_l3_pmu.c hwc->idx = idx; idx 564 drivers/perf/qcom_l3_pmu.c l3pmu->events[idx] = event; idx 583 drivers/perf/qcom_l3_pmu.c l3pmu->events[hwc->idx] = NULL; idx 584 drivers/perf/qcom_l3_pmu.c bitmap_release_region(l3pmu->used_mask, hwc->idx, order); idx 25 drivers/perf/thunderx2_pmu.c #define GET_COUNTERID(ev) ((ev->hw.idx) & 0x3) idx 29 drivers/perf/thunderx2_pmu.c #define DMC_EVENT_CFG(idx, val) ((val) << (((idx) * 8) + 1)) idx 289 drivers/perf/thunderx2_pmu.c int idx = GET_COUNTERID(event); idx 296 drivers/perf/thunderx2_pmu.c val &= ~DMC_EVENT_CFG(idx, 0x1f); idx 297 drivers/perf/thunderx2_pmu.c val |= DMC_EVENT_CFG(idx, event_id); idx 307 drivers/perf/thunderx2_pmu.c int idx = GET_COUNTERID(event); idx 311 drivers/perf/thunderx2_pmu.c val &= ~DMC_EVENT_CFG(idx, 0x1f); idx 494 drivers/perf/thunderx2_pmu.c hwc->idx = alloc_counter(tx2_pmu); idx 495 drivers/perf/thunderx2_pmu.c if (hwc->idx < 0) idx 498 drivers/perf/thunderx2_pmu.c tx2_pmu->events[hwc->idx] = event; idx 520 drivers/perf/thunderx2_pmu.c tx2_pmu->events[hwc->idx] = NULL; idx 521 drivers/perf/thunderx2_pmu.c hwc->idx = -1; idx 532 drivers/perf/thunderx2_pmu.c int max_counters, idx; idx 540 drivers/perf/thunderx2_pmu.c for_each_set_bit(idx, tx2_pmu->active_counters, max_counters) { idx 541 drivers/perf/thunderx2_pmu.c struct perf_event *event = tx2_pmu->events[idx]; idx 81 drivers/perf/xgene_pmu.c #define GET_CNTR(ev) (ev->hw.idx) idx 106 drivers/perf/xgene_pmu.c u64 (*read_counter)(struct xgene_pmu_dev *pmu, int idx); idx 107 drivers/perf/xgene_pmu.c void (*write_counter)(struct xgene_pmu_dev *pmu, int idx, u64 val); idx 108 drivers/perf/xgene_pmu.c void (*write_evttype)(struct xgene_pmu_dev *pmu_dev, int idx, u32 val); idx 111 drivers/perf/xgene_pmu.c void (*enable_counter)(struct xgene_pmu_dev *pmu_dev, int idx); idx 112 drivers/perf/xgene_pmu.c void (*disable_counter)(struct xgene_pmu_dev *pmu_dev, int idx); idx 113 drivers/perf/xgene_pmu.c void (*enable_counter_int)(struct xgene_pmu_dev *pmu_dev, int idx); idx 114 drivers/perf/xgene_pmu.c void (*disable_counter_int)(struct xgene_pmu_dev *pmu_dev, int idx); idx 735 drivers/perf/xgene_pmu.c int idx) idx 737 drivers/perf/xgene_pmu.c return readl(pmu_dev->inf->csr + PMU_PMEVCNTR0 + (4 * idx)); idx 741 drivers/perf/xgene_pmu.c int idx) idx 752 drivers/perf/xgene_pmu.c hi = xgene_pmu_read_counter32(pmu_dev, 2 * idx + 1); idx 753 drivers/perf/xgene_pmu.c lo = xgene_pmu_read_counter32(pmu_dev, 2 * idx); idx 754 drivers/perf/xgene_pmu.c } while (hi != xgene_pmu_read_counter32(pmu_dev, 2 * idx + 1)); idx 760 drivers/perf/xgene_pmu.c xgene_pmu_write_counter32(struct xgene_pmu_dev *pmu_dev, int idx, u64 val) idx 762 drivers/perf/xgene_pmu.c writel(val, pmu_dev->inf->csr + PMU_PMEVCNTR0 + (4 * idx)); idx 766 drivers/perf/xgene_pmu.c xgene_pmu_write_counter64(struct xgene_pmu_dev *pmu_dev, int idx, u64 val) idx 774 drivers/perf/xgene_pmu.c xgene_pmu_write_counter32(pmu_dev, 2 * idx, cnt_lo); idx 775 drivers/perf/xgene_pmu.c xgene_pmu_write_counter32(pmu_dev, 2 * idx + 1, cnt_hi); idx 779 drivers/perf/xgene_pmu.c xgene_pmu_write_evttype(struct xgene_pmu_dev *pmu_dev, int idx, u32 val) idx 781 drivers/perf/xgene_pmu.c writel(val, pmu_dev->inf->csr + PMU_PMEVTYPER0 + (4 * idx)); idx 803 drivers/perf/xgene_pmu.c xgene_pmu_enable_counter(struct xgene_pmu_dev *pmu_dev, int idx) idx 808 drivers/perf/xgene_pmu.c val |= 1 << idx; idx 813 drivers/perf/xgene_pmu.c xgene_pmu_disable_counter(struct xgene_pmu_dev *pmu_dev, int idx) idx 818 drivers/perf/xgene_pmu.c val |= 1 << idx; idx 823 drivers/perf/xgene_pmu.c xgene_pmu_enable_counter_int(struct xgene_pmu_dev *pmu_dev, int idx) idx 828 drivers/perf/xgene_pmu.c val |= 1 << idx; idx 833 drivers/perf/xgene_pmu.c xgene_pmu_disable_counter_int(struct xgene_pmu_dev *pmu_dev, int idx) idx 838 drivers/perf/xgene_pmu.c val |= 1 << idx; idx 987 drivers/perf/xgene_pmu.c xgene_pmu->ops->write_counter(pmu_dev, hw->idx, val); idx 1066 drivers/perf/xgene_pmu.c hw->idx = get_next_avail_cntr(pmu_dev); idx 1067 drivers/perf/xgene_pmu.c if (hw->idx < 0) idx 1071 drivers/perf/xgene_pmu.c pmu_dev->pmu_counter_event[hw->idx] = event; idx 1090 drivers/perf/xgene_pmu.c pmu_dev->pmu_counter_event[hw->idx] = NULL; idx 1197 drivers/perf/xgene_pmu.c int idx; idx 1217 drivers/perf/xgene_pmu.c for (idx = 0; idx < PMU_MAX_COUNTERS; idx++) { idx 1218 drivers/perf/xgene_pmu.c struct perf_event *event = pmu_dev->pmu_counter_event[idx]; idx 1219 drivers/perf/xgene_pmu.c int overflowed = pmovsr & BIT(idx); idx 290 drivers/phy/broadcom/phy-bcm-sr-usb.c int idx; idx 299 drivers/phy/broadcom/phy-bcm-sr-usb.c for (idx = 0; idx < NUM_BCM_SR_USB_COMBO_PHYS; idx++) { idx 300 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].regs = regs; idx 301 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].version = version; idx 302 drivers/phy/broadcom/phy-bcm-sr-usb.c if (idx == 0) { idx 303 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].offset = bcm_usb_combo_phy_hs; idx 304 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].type = USB_HS_PHY; idx 306 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].offset = bcm_usb_combo_phy_ss; idx 307 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].type = USB_SS_PHY; idx 309 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_cfg[idx].phy = devm_phy_create(dev, node, idx 311 drivers/phy/broadcom/phy-bcm-sr-usb.c if (IS_ERR(phy_cfg[idx].phy)) idx 312 drivers/phy/broadcom/phy-bcm-sr-usb.c return PTR_ERR(phy_cfg[idx].phy); idx 314 drivers/phy/broadcom/phy-bcm-sr-usb.c phy_set_drvdata(phy_cfg[idx].phy, &phy_cfg[idx]); idx 52 drivers/phy/marvell/phy-mvebu-a3700-comphy.c #define COMPHY_FW_NET(mode, idx, speed) (COMPHY_FW_MODE(mode) | \ idx 53 drivers/phy/marvell/phy-mvebu-a3700-comphy.c ((idx) << 8) | \ idx 55 drivers/phy/marvell/phy-mvebu-a3700-comphy.c #define COMPHY_FW_PCIE(mode, idx, speed, width) (COMPHY_FW_NET(mode, idx, speed) | \ idx 28 drivers/phy/mscc/phy-ocelot-serdes.c u8 idx; idx 337 drivers/phy/mscc/phy-ocelot-serdes.c u8 idx; idx 346 drivers/phy/mscc/phy-ocelot-serdes.c .idx = _idx, \ idx 409 drivers/phy/mscc/phy-ocelot-serdes.c if (macro->idx != ocelot_serdes_muxes[i].idx || idx 424 drivers/phy/mscc/phy-ocelot-serdes.c if (macro->idx <= SERDES1G_MAX) idx 425 drivers/phy/mscc/phy-ocelot-serdes.c return serdes_init_s1g(macro->ctrl->regs, macro->idx); idx 426 drivers/phy/mscc/phy-ocelot-serdes.c else if (macro->idx <= SERDES6G_MAX) idx 428 drivers/phy/mscc/phy-ocelot-serdes.c macro->idx - (SERDES1G_MAX + 1), idx 447 drivers/phy/mscc/phy-ocelot-serdes.c unsigned int port, idx, i; idx 453 drivers/phy/mscc/phy-ocelot-serdes.c idx = args->args[1]; idx 458 drivers/phy/mscc/phy-ocelot-serdes.c if (idx != macro->idx) idx 462 drivers/phy/mscc/phy-ocelot-serdes.c if (idx != SERDES6G(0) && macro->port >= 0) idx 472 drivers/phy/mscc/phy-ocelot-serdes.c static int serdes_phy_create(struct serdes_ctrl *ctrl, u8 idx, struct phy **phy) idx 484 drivers/phy/mscc/phy-ocelot-serdes.c macro->idx = idx; idx 487 drivers/pinctrl/aspeed/pinmux-aspeed.h #define SIG_DESC_IP_BIT(ip, reg, idx, val) \ idx 488 drivers/pinctrl/aspeed/pinmux-aspeed.h { ip, reg, BIT_MASK(idx), val, (((val) + 1) & 1) } idx 498 drivers/pinctrl/aspeed/pinmux-aspeed.h #define SIG_DESC_BIT(reg, idx, val) \ idx 499 drivers/pinctrl/aspeed/pinmux-aspeed.h SIG_DESC_IP_BIT(ASPEED_IP_SCU, reg, idx, val) idx 501 drivers/pinctrl/aspeed/pinmux-aspeed.h #define SIG_DESC_IP_SET(ip, reg, idx) SIG_DESC_IP_BIT(ip, reg, idx, 1) idx 510 drivers/pinctrl/aspeed/pinmux-aspeed.h #define SIG_DESC_SET(reg, idx) SIG_DESC_IP_BIT(ASPEED_IP_SCU, reg, idx, 1) idx 511 drivers/pinctrl/aspeed/pinmux-aspeed.h #define SIG_DESC_CLEAR(reg, idx) { ASPEED_IP_SCU, reg, BIT_MASK(idx), 0, 0 } idx 348 drivers/pinctrl/freescale/pinctrl-mxs.c struct device_node *np, int idx, idx 352 drivers/pinctrl/freescale/pinctrl-mxs.c struct mxs_group *g = &d->soc->groups[idx]; idx 1005 drivers/pinctrl/pinctrl-bm1880.c #define BM1880_PINCONF(pin, idx) ((!((pin + 1) & 1) << 4) + idx) idx 1606 drivers/pinctrl/pinctrl-ingenic.c unsigned int idx = pin % PINS_PER_GPIO_CHIP; idx 1610 drivers/pinctrl/pinctrl-ingenic.c (set ? REG_SET(reg) : REG_CLEAR(reg)), BIT(idx)); idx 1616 drivers/pinctrl/pinctrl-ingenic.c unsigned int idx = pin % PINS_PER_GPIO_CHIP; idx 1619 drivers/pinctrl/pinctrl-ingenic.c (set ? REG_SET(reg) : REG_CLEAR(reg)), BIT(idx)); idx 1631 drivers/pinctrl/pinctrl-ingenic.c unsigned int idx = pin % PINS_PER_GPIO_CHIP; idx 1637 drivers/pinctrl/pinctrl-ingenic.c return val & BIT(idx); idx 1666 drivers/pinctrl/pinctrl-ingenic.c unsigned int idx = pin % PINS_PER_GPIO_CHIP; idx 1670 drivers/pinctrl/pinctrl-ingenic.c 'A' + offt, idx, func); idx 1725 drivers/pinctrl/pinctrl-ingenic.c unsigned int idx = pin % PINS_PER_GPIO_CHIP; idx 1729 drivers/pinctrl/pinctrl-ingenic.c 'A' + offt, idx, input ? "in" : "out"); idx 1762 drivers/pinctrl/pinctrl-ingenic.c unsigned int idx = pin % PINS_PER_GPIO_CHIP; idx 1778 drivers/pinctrl/pinctrl-ingenic.c if (!pull || !(jzpc->info->pull_ups[offt] & BIT(idx))) idx 1783 drivers/pinctrl/pinctrl-ingenic.c if (!pull || !(jzpc->info->pull_downs[offt] & BIT(idx))) idx 1808 drivers/pinctrl/pinctrl-ingenic.c unsigned int idx = pin % PINS_PER_GPIO_CHIP; idx 1827 drivers/pinctrl/pinctrl-ingenic.c 'A' + offt, idx); idx 1832 drivers/pinctrl/pinctrl-ingenic.c if (!(jzpc->info->pull_ups[offt] & BIT(idx))) idx 1835 drivers/pinctrl/pinctrl-ingenic.c 'A' + offt, idx); idx 1840 drivers/pinctrl/pinctrl-ingenic.c if (!(jzpc->info->pull_downs[offt] & BIT(idx))) idx 1843 drivers/pinctrl/pinctrl-ingenic.c 'A' + offt, idx); idx 1164 drivers/pinctrl/pinctrl-st.c struct st_pctl_group *grp, struct st_pinctrl *info, int idx) idx 1004 drivers/pinctrl/pinctrl-sx150x.c unsigned int idx, val; idx 1035 drivers/pinctrl/pinctrl-sx150x.c for (n = width, val = 0, idx = reg; n > 0; n -= 8, idx++) { idx 1038 drivers/pinctrl/pinctrl-sx150x.c ret = i2c_smbus_read_byte_data(i2c, idx); idx 480 drivers/pinctrl/samsung/pinctrl-exynos.c int idx, irq; idx 529 drivers/pinctrl/samsung/pinctrl-exynos.c for (idx = 0; idx < bank->nr_pins; ++idx) { idx 530 drivers/pinctrl/samsung/pinctrl-exynos.c irq = irq_of_parse_and_map(bank->of_node, idx); idx 533 drivers/pinctrl/samsung/pinctrl-exynos.c bank->name, idx); idx 536 drivers/pinctrl/samsung/pinctrl-exynos.c weint_data[idx].irq = idx; idx 537 drivers/pinctrl/samsung/pinctrl-exynos.c weint_data[idx].bank = bank; idx 540 drivers/pinctrl/samsung/pinctrl-exynos.c &weint_data[idx]); idx 565 drivers/pinctrl/samsung/pinctrl-exynos.c idx = 0; idx 570 drivers/pinctrl/samsung/pinctrl-exynos.c muxed_data->banks[idx++] = bank; idx 625 drivers/pinctrl/sh-pfc/core.c static void sh_pfc_nop_reg(struct sh_pfc *pfc, u32 reg, unsigned int idx) idx 629 drivers/pinctrl/sh-pfc/core.c static void sh_pfc_save_reg(struct sh_pfc *pfc, u32 reg, unsigned int idx) idx 631 drivers/pinctrl/sh-pfc/core.c pfc->saved_regs[idx] = sh_pfc_read(pfc, reg); idx 634 drivers/pinctrl/sh-pfc/core.c static void sh_pfc_restore_reg(struct sh_pfc *pfc, u32 reg, unsigned int idx) idx 636 drivers/pinctrl/sh-pfc/core.c sh_pfc_write(pfc, reg, pfc->saved_regs[idx]); idx 640 drivers/pinctrl/sh-pfc/core.c void (*do_reg)(struct sh_pfc *pfc, u32 reg, unsigned int idx)) idx 48 drivers/pinctrl/sh-pfc/gpio.c int idx = sh_pfc_get_pin_index(chip->pfc, offset); idx 49 drivers/pinctrl/sh-pfc/gpio.c struct sh_pfc_gpio_pin *gpio_pin = &chip->pins[idx]; idx 73 drivers/pinctrl/sh-pfc/gpio.c static void gpio_setup_data_reg(struct sh_pfc_chip *chip, unsigned idx) idx 76 drivers/pinctrl/sh-pfc/gpio.c struct sh_pfc_gpio_pin *gpio_pin = &chip->pins[idx]; idx 77 drivers/pinctrl/sh-pfc/gpio.c const struct sh_pfc_pin *pin = &pfc->info->pins[idx]; idx 134 drivers/pinctrl/sh-pfc/gpio.c int idx = sh_pfc_get_pin_index(pfc, offset); idx 136 drivers/pinctrl/sh-pfc/gpio.c if (idx < 0 || pfc->info->pins[idx].enum_id == 0) idx 28 drivers/pinctrl/sh-pfc/pfc-r8a7740.c #define IRQC_PINS_MUX(irq, idx, pin) \ idx 29 drivers/pinctrl/sh-pfc/pfc-r8a7740.c static const unsigned int intc_irq##irq##_##idx##_pins[] = { \ idx 32 drivers/pinctrl/sh-pfc/pfc-r8a7740.c static const unsigned int intc_irq##irq##_##idx##_mux[] = { \ idx 111 drivers/pinctrl/sh-pfc/pinctrl.c unsigned int idx = *index; idx 203 drivers/pinctrl/sh-pfc/pinctrl.c maps[idx].type = PIN_MAP_TYPE_MUX_GROUP; idx 204 drivers/pinctrl/sh-pfc/pinctrl.c maps[idx].data.mux.group = group; idx 205 drivers/pinctrl/sh-pfc/pinctrl.c maps[idx].data.mux.function = function; idx 206 drivers/pinctrl/sh-pfc/pinctrl.c idx++; idx 210 drivers/pinctrl/sh-pfc/pinctrl.c ret = sh_pfc_map_add_config(&maps[idx], group, idx 216 drivers/pinctrl/sh-pfc/pinctrl.c idx++; idx 226 drivers/pinctrl/sh-pfc/pinctrl.c ret = sh_pfc_map_add_config(&maps[idx], pin, idx 232 drivers/pinctrl/sh-pfc/pinctrl.c idx++; idx 236 drivers/pinctrl/sh-pfc/pinctrl.c *index = idx; idx 357 drivers/pinctrl/sh-pfc/pinctrl.c int idx = sh_pfc_get_pin_index(pfc, grp->pins[i]); idx 358 drivers/pinctrl/sh-pfc/pinctrl.c struct sh_pfc_pin_config *cfg = &pmx->configs[idx]; idx 376 drivers/pinctrl/sh-pfc/pinctrl.c int idx = sh_pfc_get_pin_index(pfc, grp->pins[i]); idx 377 drivers/pinctrl/sh-pfc/pinctrl.c struct sh_pfc_pin_config *cfg = &pmx->configs[idx]; idx 394 drivers/pinctrl/sh-pfc/pinctrl.c int idx = sh_pfc_get_pin_index(pfc, offset); idx 395 drivers/pinctrl/sh-pfc/pinctrl.c struct sh_pfc_pin_config *cfg = &pmx->configs[idx]; idx 405 drivers/pinctrl/sh-pfc/pinctrl.c const struct sh_pfc_pin *pin = &pfc->info->pins[idx]; idx 428 drivers/pinctrl/sh-pfc/pinctrl.c int idx = sh_pfc_get_pin_index(pfc, offset); idx 429 drivers/pinctrl/sh-pfc/pinctrl.c struct sh_pfc_pin_config *cfg = &pmx->configs[idx]; idx 447 drivers/pinctrl/sh-pfc/pinctrl.c int idx = sh_pfc_get_pin_index(pfc, offset); idx 448 drivers/pinctrl/sh-pfc/pinctrl.c const struct sh_pfc_pin *pin = &pfc->info->pins[idx]; idx 572 drivers/pinctrl/sh-pfc/pinctrl.c int idx = sh_pfc_get_pin_index(pfc, _pin); idx 573 drivers/pinctrl/sh-pfc/pinctrl.c const struct sh_pfc_pin *pin = &pfc->info->pins[idx]; idx 4778 drivers/pinctrl/sirf/pinctrl-atlas7.c u8 idx; idx 5098 drivers/pinctrl/sirf/pinctrl-atlas7.c int idx, ret; idx 5121 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < grp_mux->pad_mux_count; idx++) { idx 5122 drivers/pinctrl/sirf/pinctrl-atlas7.c mux = &grp_mux->pad_mux_list[idx]; idx 5142 drivers/pinctrl/sirf/pinctrl-atlas7.c int idx; idx 5144 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < ARRAY_SIZE(atlas7_ma2ds_map); idx++) { idx 5145 drivers/pinctrl/sirf/pinctrl-atlas7.c if (atlas7_ma2ds_map[idx].ma != ma) idx 5149 drivers/pinctrl/sirf/pinctrl-atlas7.c return atlas7_ma2ds_map[idx].ds_4we; idx 5151 drivers/pinctrl/sirf/pinctrl-atlas7.c return atlas7_ma2ds_map[idx].ds_16st; idx 5153 drivers/pinctrl/sirf/pinctrl-atlas7.c return atlas7_ma2ds_map[idx].ds_0204m31; idx 5155 drivers/pinctrl/sirf/pinctrl-atlas7.c return atlas7_ma2ds_map[idx].ds_0610m31; idx 5241 drivers/pinctrl/sirf/pinctrl-atlas7.c u32 idx; idx 5245 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < range->npins; idx++) { idx 5246 drivers/pinctrl/sirf/pinctrl-atlas7.c if (pin == range->pins[idx]) idx 5250 drivers/pinctrl/sirf/pinctrl-atlas7.c if (idx >= range->npins) { idx 5325 drivers/pinctrl/sirf/pinctrl-atlas7.c int idx, err; idx 5327 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < num_configs; idx++) { idx 5328 drivers/pinctrl/sirf/pinctrl-atlas7.c param = pinconf_to_config_param(configs[idx]); idx 5329 drivers/pinctrl/sirf/pinctrl-atlas7.c arg = pinconf_to_config_argument(configs[idx]); idx 5404 drivers/pinctrl/sirf/pinctrl-atlas7.c int ret, idx; idx 5442 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < banks; idx++) { idx 5443 drivers/pinctrl/sirf/pinctrl-atlas7.c pmx->regs[idx] = of_iomap(np, idx); idx 5444 drivers/pinctrl/sirf/pinctrl-atlas7.c if (!pmx->regs[idx]) { idx 5446 drivers/pinctrl/sirf/pinctrl-atlas7.c "can't map ioc bank#%d registers\n", idx); idx 5467 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < banks; idx++) { idx 5468 drivers/pinctrl/sirf/pinctrl-atlas7.c if (!pmx->regs[idx]) idx 5470 drivers/pinctrl/sirf/pinctrl-atlas7.c iounmap(pmx->regs[idx]); idx 5484 drivers/pinctrl/sirf/pinctrl-atlas7.c int idx; idx 5488 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < pmx->pctl_desc.npins; idx++) { idx 5490 drivers/pinctrl/sirf/pinctrl-atlas7.c conf = &pmx->pctl_data->confs[idx]; idx 5491 drivers/pinctrl/sirf/pinctrl-atlas7.c bank = atlas7_pin_to_bank(idx); idx 5492 drivers/pinctrl/sirf/pinctrl-atlas7.c status = &pmx->sleep_data[idx]; idx 5527 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < NUM_OF_IN_DISABLE_REG; idx++) { idx 5528 drivers/pinctrl/sirf/pinctrl-atlas7.c pmx->status_ds[idx] = readl(pmx->regs[BANK_DS] + idx 5529 drivers/pinctrl/sirf/pinctrl-atlas7.c IN_DISABLE_0_REG_SET + 0x8 * idx); idx 5530 drivers/pinctrl/sirf/pinctrl-atlas7.c pmx->status_dsv[idx] = readl(pmx->regs[BANK_DS] + idx 5531 drivers/pinctrl/sirf/pinctrl-atlas7.c IN_DISABLE_VAL_0_REG_SET + 0x8 * idx); idx 5541 drivers/pinctrl/sirf/pinctrl-atlas7.c int idx; idx 5543 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < pmx->pctl_desc.npins; idx++) { idx 5545 drivers/pinctrl/sirf/pinctrl-atlas7.c status = &pmx->sleep_data[idx]; idx 5548 drivers/pinctrl/sirf/pinctrl-atlas7.c __atlas7_pmx_pin_enable(pmx, idx, (u32)status->func & 0xff); idx 5554 drivers/pinctrl/sirf/pinctrl-atlas7.c __altas7_pinctrl_set_drive_strength_sel(pmx->pctl, idx, idx 5559 drivers/pinctrl/sirf/pinctrl-atlas7.c altas7_pinctrl_set_pull_sel(pmx->pctl, idx, idx 5567 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < NUM_OF_IN_DISABLE_REG; idx++) { idx 5569 drivers/pinctrl/sirf/pinctrl-atlas7.c IN_DISABLE_0_REG_CLR + 0x8 * idx); idx 5570 drivers/pinctrl/sirf/pinctrl-atlas7.c writel(pmx->status_ds[idx], pmx->regs[BANK_DS] + idx 5571 drivers/pinctrl/sirf/pinctrl-atlas7.c IN_DISABLE_0_REG_SET + 0x8 * idx); idx 5573 drivers/pinctrl/sirf/pinctrl-atlas7.c IN_DISABLE_VAL_0_REG_CLR + 0x8 * idx); idx 5574 drivers/pinctrl/sirf/pinctrl-atlas7.c writel(pmx->status_dsv[idx], pmx->regs[BANK_DS] + idx 5575 drivers/pinctrl/sirf/pinctrl-atlas7.c IN_DISABLE_VAL_0_REG_SET + 0x8 * idx); idx 5656 drivers/pinctrl/sirf/pinctrl-atlas7.c static void __atlas7_gpio_irq_mask(struct atlas7_gpio_chip *a7gc, int idx) idx 5662 drivers/pinctrl/sirf/pinctrl-atlas7.c bank = atlas7_gpio_to_bank(a7gc, idx); idx 5663 drivers/pinctrl/sirf/pinctrl-atlas7.c pin_in_bank = idx - bank->gpio_offset; idx 5784 drivers/pinctrl/sirf/pinctrl-atlas7.c int pin_in_bank = 0, idx; idx 5788 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < a7gc->nbank; idx++) { idx 5789 drivers/pinctrl/sirf/pinctrl-atlas7.c bank = &a7gc->banks[idx]; idx 5793 drivers/pinctrl/sirf/pinctrl-atlas7.c BUG_ON(idx == a7gc->nbank); idx 5998 drivers/pinctrl/sirf/pinctrl-atlas7.c int ret, idx; idx 6069 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < nbank; idx++) { idx 6072 drivers/pinctrl/sirf/pinctrl-atlas7.c bank = &a7gc->banks[idx]; idx 6074 drivers/pinctrl/sirf/pinctrl-atlas7.c bank->base = ATLAS7_GPIO_BASE(a7gc, idx); idx 6075 drivers/pinctrl/sirf/pinctrl-atlas7.c bank->gpio_offset = idx * NGPIO_OF_BANK; idx 6078 drivers/pinctrl/sirf/pinctrl-atlas7.c ret = of_irq_get(np, idx); idx 6105 drivers/pinctrl/sirf/pinctrl-atlas7.c u32 idx, pin; idx 6107 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < a7gc->nbank; idx++) { idx 6108 drivers/pinctrl/sirf/pinctrl-atlas7.c bank = &a7gc->banks[idx]; idx 6123 drivers/pinctrl/sirf/pinctrl-atlas7.c u32 idx, pin; idx 6125 drivers/pinctrl/sirf/pinctrl-atlas7.c for (idx = 0; idx < a7gc->nbank; idx++) { idx 6126 drivers/pinctrl/sirf/pinctrl-atlas7.c bank = &a7gc->banks[idx]; idx 425 drivers/pinctrl/sirf/pinctrl-sirf.c int idx = sirfsoc_gpio_to_bankoff(d->hwirq); idx 429 drivers/pinctrl/sirf/pinctrl-sirf.c offset = SIRFSOC_GPIO_CTRL(bank->id, idx); idx 442 drivers/pinctrl/sirf/pinctrl-sirf.c int idx) idx 447 drivers/pinctrl/sirf/pinctrl-sirf.c offset = SIRFSOC_GPIO_CTRL(bank->id, idx); idx 473 drivers/pinctrl/sirf/pinctrl-sirf.c int idx = sirfsoc_gpio_to_bankoff(d->hwirq); idx 477 drivers/pinctrl/sirf/pinctrl-sirf.c offset = SIRFSOC_GPIO_CTRL(bank->id, idx); idx 494 drivers/pinctrl/sirf/pinctrl-sirf.c int idx = sirfsoc_gpio_to_bankoff(d->hwirq); idx 498 drivers/pinctrl/sirf/pinctrl-sirf.c offset = SIRFSOC_GPIO_CTRL(bank->id, idx); idx 557 drivers/pinctrl/sirf/pinctrl-sirf.c int idx = 0; idx 580 drivers/pinctrl/sirf/pinctrl-sirf.c ctrl = readl(sgpio->chip.regs + SIRFSOC_GPIO_CTRL(bank->id, idx)); idx 588 drivers/pinctrl/sirf/pinctrl-sirf.c __func__, bank->id, idx); idx 589 drivers/pinctrl/sirf/pinctrl-sirf.c generic_handle_irq(irq_find_mapping(gc->irq.domain, idx + idx 593 drivers/pinctrl/sirf/pinctrl-sirf.c idx++; idx 653 drivers/pinctrl/sirf/pinctrl-sirf.c int idx = sirfsoc_gpio_to_bankoff(gpio); idx 657 drivers/pinctrl/sirf/pinctrl-sirf.c offset = SIRFSOC_GPIO_CTRL(bank->id, idx); idx 696 drivers/pinctrl/sirf/pinctrl-sirf.c int idx = sirfsoc_gpio_to_bankoff(gpio); idx 700 drivers/pinctrl/sirf/pinctrl-sirf.c offset = SIRFSOC_GPIO_CTRL(bank->id, idx); idx 267 drivers/pinctrl/sunxi/pinctrl-sunxi.c unsigned int configlen = 0, idx = 0; idx 292 drivers/pinctrl/sunxi/pinctrl-sunxi.c pinconfig[idx++] = pinconf_to_config_packed(PIN_CONFIG_DRIVE_STRENGTH, idx 307 drivers/pinctrl/sunxi/pinctrl-sunxi.c pinconfig[idx++] = pinconf_to_config_packed(pull, arg); idx 1025 drivers/pinctrl/sunxi/pinctrl-sunxi.c u8 idx = sunxi_irq_ctrl_offset(d->hwirq); idx 1033 drivers/pinctrl/sunxi/pinctrl-sunxi.c writel(val & ~(1 << idx), pctl->membase + reg); idx 1042 drivers/pinctrl/sunxi/pinctrl-sunxi.c u8 idx = sunxi_irq_ctrl_offset(d->hwirq); idx 1050 drivers/pinctrl/sunxi/pinctrl-sunxi.c writel(val | (1 << idx), pctl->membase + reg); idx 27 drivers/platform/chrome/cros_ec_debugfs.c #define CIRC_ADD(idx, size, value) (((idx) + (value)) & ((size) - 1)) idx 73 drivers/platform/chrome/cros_ec_debugfs.c int idx; idx 103 drivers/platform/chrome/cros_ec_debugfs.c idx = 0; idx 104 drivers/platform/chrome/cros_ec_debugfs.c while (idx < ret && ec_buffer[idx] != '\0' && buf_space > 0) { idx 105 drivers/platform/chrome/cros_ec_debugfs.c cb->buf[cb->head] = ec_buffer[idx]; idx 107 drivers/platform/chrome/cros_ec_debugfs.c idx++; idx 292 drivers/platform/mellanox/mlxbf-tmfifo.c unsigned int idx, head; idx 294 drivers/platform/mellanox/mlxbf-tmfifo.c if (vring->next_avail == virtio16_to_cpu(vdev, vr->avail->idx)) idx 297 drivers/platform/mellanox/mlxbf-tmfifo.c idx = vring->next_avail % vr->num; idx 298 drivers/platform/mellanox/mlxbf-tmfifo.c head = virtio16_to_cpu(vdev, vr->avail->ring[idx]); idx 313 drivers/platform/mellanox/mlxbf-tmfifo.c u16 idx, vr_idx; idx 315 drivers/platform/mellanox/mlxbf-tmfifo.c vr_idx = virtio16_to_cpu(vdev, vr->used->idx); idx 316 drivers/platform/mellanox/mlxbf-tmfifo.c idx = vr_idx % vr->num; idx 317 drivers/platform/mellanox/mlxbf-tmfifo.c vr->used->ring[idx].id = cpu_to_virtio32(vdev, desc - vr->desc); idx 318 drivers/platform/mellanox/mlxbf-tmfifo.c vr->used->ring[idx].len = cpu_to_virtio32(vdev, len); idx 326 drivers/platform/mellanox/mlxbf-tmfifo.c vr->used->idx = cpu_to_virtio16(vdev, vr_idx + 1); idx 335 drivers/platform/mellanox/mlxbf-tmfifo.c u32 len = 0, idx; idx 341 drivers/platform/mellanox/mlxbf-tmfifo.c idx = virtio16_to_cpu(vdev, desc->next); idx 342 drivers/platform/mellanox/mlxbf-tmfifo.c desc = &vr->desc[idx]; idx 417 drivers/platform/mellanox/mlxbf-tmfifo.c u32 len, idx, seg; idx 438 drivers/platform/mellanox/mlxbf-tmfifo.c idx = virtio16_to_cpu(vdev, desc->next); idx 439 drivers/platform/mellanox/mlxbf-tmfifo.c desc = &vr->desc[idx]; idx 680 drivers/platform/mellanox/mlxbf-tmfifo.c u32 len, idx; idx 723 drivers/platform/mellanox/mlxbf-tmfifo.c idx = virtio16_to_cpu(vdev, desc->next); idx 724 drivers/platform/mellanox/mlxbf-tmfifo.c desc = &vr->desc[idx]; idx 1578 drivers/platform/x86/asus-laptop.c int idx) idx 1537 drivers/platform/x86/asus-wmi.c struct attribute *attr, int idx) idx 2159 drivers/platform/x86/asus-wmi.c struct attribute *attr, int idx) idx 101 drivers/platform/x86/dell_rbu.c unsigned int idx = 0; idx 171 drivers/platform/x86/dell_rbu.c invalid_addr_packet_array[idx++] = packet_data_temp_buf; idx 202 drivers/platform/x86/dell_rbu.c for (;idx>0;idx--) { idx 205 drivers/platform/x86/dell_rbu.c invalid_addr_packet_array[idx-1])); idx 206 drivers/platform/x86/dell_rbu.c free_pages((unsigned long)invalid_addr_packet_array[idx-1], idx 520 drivers/platform/x86/ideapad-laptop.c int idx) idx 298 drivers/platform/x86/intel_menlow.c char *buf, int idx) idx 304 drivers/platform/x86/intel_menlow.c result = sensor_get_auxtrip(attr->handle, idx, &value); idx 322 drivers/platform/x86/intel_menlow.c const char *buf, size_t count, int idx) idx 335 drivers/platform/x86/intel_menlow.c result = sensor_set_auxtrip(attr->handle, idx, idx 69 drivers/platform/x86/intel_telemetry_debugfs.c for (idx = 0; idx < (EVTNUM); idx++) \ idx 70 drivers/platform/x86/intel_telemetry_debugfs.c (BUF)[idx] = ((EVTLOG) >> (EVTDAT)[idx].bit_pos) & \ idx 342 drivers/platform/x86/intel_telemetry_debugfs.c int index, idx, ret, err = 0; idx 509 drivers/platform/x86/intel_telemetry_debugfs.c int index, idx, ret; idx 623 drivers/platform/x86/intel_telemetry_debugfs.c for (idx = 0; idx < conf->pss_idle_evts - 1; idx++) { idx 624 drivers/platform/x86/intel_telemetry_debugfs.c pss_idle[idx] = (evtlog->telem_evtlog >> idx 625 drivers/platform/x86/intel_telemetry_debugfs.c conf->pss_idle_data[idx].bit_pos) & idx 629 drivers/platform/x86/intel_telemetry_debugfs.c pss_idle[idx] = (evtlog->telem_evtlog >> idx 630 drivers/platform/x86/intel_telemetry_debugfs.c conf->pss_idle_data[idx].bit_pos) & idx 285 drivers/platform/x86/intel_telemetry_pltdrv.c int ret, index, idx; idx 333 drivers/platform/x86/intel_telemetry_pltdrv.c for (idx = 0; idx < num_ioss_evts; idx++) { idx 335 drivers/platform/x86/intel_telemetry_pltdrv.c telm_conf->ioss_config.telem_evts[idx].evt_id, idx 336 drivers/platform/x86/intel_telemetry_pltdrv.c idx)) { idx 338 drivers/platform/x86/intel_telemetry_pltdrv.c telm_conf->ioss_config.telem_evts[idx].evt_id); idx 380 drivers/platform/x86/intel_telemetry_pltdrv.c for (index = telm_conf->ioss_config.ssram_evts_used, idx = 0; idx 381 drivers/platform/x86/intel_telemetry_pltdrv.c idx < num_ioss_evts; index++, idx++) { idx 383 drivers/platform/x86/intel_telemetry_pltdrv.c ioss_evtmap[idx]; idx 389 drivers/platform/x86/intel_telemetry_pltdrv.c ioss_evtmap[idx]); idx 421 drivers/platform/x86/intel_telemetry_pltdrv.c int ret, index, idx; idx 461 drivers/platform/x86/intel_telemetry_pltdrv.c for (idx = 0; idx < num_pss_evts; idx++) { idx 463 drivers/platform/x86/intel_telemetry_pltdrv.c telm_conf->pss_config.telem_evts[idx].evt_id, idx 464 drivers/platform/x86/intel_telemetry_pltdrv.c idx)) { idx 466 drivers/platform/x86/intel_telemetry_pltdrv.c telm_conf->pss_config.telem_evts[idx].evt_id); idx 506 drivers/platform/x86/intel_telemetry_pltdrv.c for (index = telm_conf->pss_config.ssram_evts_used, idx = 0; idx 507 drivers/platform/x86/intel_telemetry_pltdrv.c idx < num_pss_evts; index++, idx++) { idx 510 drivers/platform/x86/intel_telemetry_pltdrv.c pss_evtmap[idx]; idx 516 drivers/platform/x86/intel_telemetry_pltdrv.c pss_evtmap[idx]); idx 1212 drivers/platform/x86/samsung-laptop.c struct attribute *attr, int idx) idx 2186 drivers/platform/x86/sony-laptop.c size_t idx = 0; idx 2190 drivers/platform/x86/sony-laptop.c idx += snprintf(buffer + idx, PAGE_SIZE - idx, "%s ", idx 2193 drivers/platform/x86/sony-laptop.c idx += snprintf(buffer + idx, PAGE_SIZE - idx, "\n"); idx 2195 drivers/platform/x86/sony-laptop.c return idx; idx 6287 drivers/platform/x86/thinkpad_acpi.c static int thermal_get_sensor(int idx, s32 *value) idx 6298 drivers/platform/x86/thinkpad_acpi.c if (idx >= 8 && idx <= 15) { idx 6300 drivers/platform/x86/thinkpad_acpi.c idx -= 8; idx 6305 drivers/platform/x86/thinkpad_acpi.c if (idx <= 7) { idx 6306 drivers/platform/x86/thinkpad_acpi.c if (!acpi_ec_read(t + idx, &tmp)) idx 6314 drivers/platform/x86/thinkpad_acpi.c if (idx <= 7) { idx 6315 drivers/platform/x86/thinkpad_acpi.c snprintf(tmpi, sizeof(tmpi), "TMP%c", '0' + idx); idx 6326 drivers/platform/x86/thinkpad_acpi.c if (idx <= 7) { idx 6327 drivers/platform/x86/thinkpad_acpi.c snprintf(tmpi, sizeof(tmpi), "TMP%c", '0' + idx); idx 6397 drivers/platform/x86/thinkpad_acpi.c int idx = sensor_attr->index; idx 6401 drivers/platform/x86/thinkpad_acpi.c res = thermal_get_sensor(idx, &value); idx 2365 drivers/platform/x86/toshiba_acpi.c struct attribute *attr, int idx) idx 115 drivers/pnp/isapnp/core.c unsigned char isapnp_read_byte(unsigned char idx) idx 117 drivers/pnp/isapnp/core.c write_address(idx); idx 121 drivers/pnp/isapnp/core.c static unsigned short isapnp_read_word(unsigned char idx) idx 125 drivers/pnp/isapnp/core.c val = isapnp_read_byte(idx); idx 126 drivers/pnp/isapnp/core.c val = (val << 8) + isapnp_read_byte(idx + 1); idx 130 drivers/pnp/isapnp/core.c void isapnp_write_byte(unsigned char idx, unsigned char val) idx 132 drivers/pnp/isapnp/core.c write_address(idx); idx 136 drivers/pnp/isapnp/core.c static void isapnp_write_word(unsigned char idx, unsigned short val) idx 138 drivers/pnp/isapnp/core.c isapnp_write_byte(idx, val >> 8); idx 139 drivers/pnp/isapnp/core.c isapnp_write_byte(idx + 1, val); idx 38 drivers/pnp/manager.c static int pnp_assign_port(struct pnp_dev *dev, struct pnp_port *rule, int idx) idx 42 drivers/pnp/manager.c res = pnp_find_resource(dev, rule->flags, IORESOURCE_IO, idx); idx 45 drivers/pnp/manager.c "flags %#lx\n", idx, (unsigned long long) res->start, idx 57 drivers/pnp/manager.c pnp_dbg(&dev->dev, " io %d disabled\n", idx); idx 69 drivers/pnp/manager.c "(min %#llx max %#llx)\n", idx, idx 81 drivers/pnp/manager.c static int pnp_assign_mem(struct pnp_dev *dev, struct pnp_mem *rule, int idx) idx 85 drivers/pnp/manager.c res = pnp_find_resource(dev, rule->flags, IORESOURCE_MEM, idx); idx 88 drivers/pnp/manager.c "flags %#lx\n", idx, (unsigned long long) res->start, idx 108 drivers/pnp/manager.c pnp_dbg(&dev->dev, " mem %d disabled\n", idx); idx 120 drivers/pnp/manager.c "(min %#llx max %#llx)\n", idx, idx 132 drivers/pnp/manager.c static int pnp_assign_irq(struct pnp_dev *dev, struct pnp_irq *rule, int idx) idx 142 drivers/pnp/manager.c res = pnp_find_resource(dev, rule->flags, IORESOURCE_IRQ, idx); idx 145 drivers/pnp/manager.c idx, (int) res->start, res->flags); idx 156 drivers/pnp/manager.c pnp_dbg(&dev->dev, " irq %d disabled\n", idx); idx 178 drivers/pnp/manager.c pnp_dbg(&dev->dev, " irq %d disabled (optional)\n", idx); idx 182 drivers/pnp/manager.c pnp_dbg(&dev->dev, " couldn't assign irq %d\n", idx); idx 191 drivers/pnp/manager.c static int pnp_assign_dma(struct pnp_dev *dev, struct pnp_dma *rule, int idx) idx 201 drivers/pnp/manager.c res = pnp_find_resource(dev, rule->flags, IORESOURCE_DMA, idx); idx 204 drivers/pnp/manager.c idx, (int) res->start, res->flags); idx 215 drivers/pnp/manager.c pnp_dbg(&dev->dev, " dma %d disabled\n", idx); idx 227 drivers/pnp/manager.c pnp_dbg(&dev->dev, " couldn't assign dma %d\n", idx); idx 69 drivers/power/avs/rockchip-io-domain.c int idx; idx 88 drivers/power/avs/rockchip-io-domain.c val <<= supply->idx; idx 91 drivers/power/avs/rockchip-io-domain.c val |= (BIT(supply->idx) << 16); idx 562 drivers/power/avs/rockchip-io-domain.c supply->idx = i; idx 211 drivers/power/reset/at91-reset.c int ret, idx = 0; idx 222 drivers/power/reset/at91-reset.c at91_ramc_base[idx] = of_iomap(np, 0); idx 223 drivers/power/reset/at91-reset.c if (!at91_ramc_base[idx]) { idx 228 drivers/power/reset/at91-reset.c idx++; idx 315 drivers/power/supply/bq24190_charger.c u8 idx; idx 317 drivers/power/supply/bq24190_charger.c idx = bq24190_find_idx(tbl, tbl_size, val); idx 319 drivers/power/supply/bq24190_charger.c return bq24190_write_mask(bdi, reg, mask, shift, idx); idx 980 drivers/power/supply/bq24190_charger.c int idx = ARRAY_SIZE(bq24190_ccc_ichg_values) - 1; idx 982 drivers/power/supply/bq24190_charger.c val->intval = bq24190_ccc_ichg_values[idx]; idx 1027 drivers/power/supply/bq24190_charger.c int idx = ARRAY_SIZE(bq24190_cvc_vreg_values) - 1; idx 1029 drivers/power/supply/bq24190_charger.c val->intval = bq24190_cvc_vreg_values[idx]; idx 231 drivers/power/supply/bq24257_charger.c u8 idx; idx 233 drivers/power/supply/bq24257_charger.c for (idx = 1; idx < map_size; idx++) idx 234 drivers/power/supply/bq24257_charger.c if (value < map[idx]) idx 237 drivers/power/supply/bq24257_charger.c return idx - 1; idx 310 drivers/power/supply/bq25890_charger.c u8 idx; idx 316 drivers/power/supply/bq25890_charger.c for (idx = 1; idx < tbl_size && tbl[idx] <= value; idx++) idx 324 drivers/power/supply/bq25890_charger.c for (idx = 1; idx 325 drivers/power/supply/bq25890_charger.c idx < rtbl_size && (idx * rtbl->step + rtbl->min <= value); idx 326 drivers/power/supply/bq25890_charger.c idx++) idx 330 drivers/power/supply/bq25890_charger.c return idx - 1; idx 333 drivers/power/supply/bq25890_charger.c static u32 bq25890_find_val(u8 idx, enum bq25890_table_ids id) idx 339 drivers/power/supply/bq25890_charger.c return bq25890_tables[id].lt.tbl[idx]; idx 344 drivers/power/supply/bq25890_charger.c return (rtbl->min + idx * rtbl->step); idx 292 drivers/power/supply/rt9455_charger.c unsigned int idx = rt9455_find_idx(tbl, tbl_size, val); idx 294 drivers/power/supply/rt9455_charger.c return regmap_field_write(info->regmap_fields[field], idx); idx 545 drivers/power/supply/rt9455_charger.c int idx = ARRAY_SIZE(rt9455_ichrg_values) - 1; idx 547 drivers/power/supply/rt9455_charger.c val->intval = rt9455_ichrg_values[idx]; idx 575 drivers/power/supply/rt9455_charger.c int idx = ARRAY_SIZE(rt9455_vmreg_values) - 1; idx 577 drivers/power/supply/rt9455_charger.c val->intval = rt9455_vmreg_values[idx]; idx 655 drivers/power/supply/rt9455_charger.c int idx, ret; idx 734 drivers/power/supply/rt9455_charger.c idx = ARRAY_SIZE(rt9455_vmreg_values) - 1; idx 738 drivers/power/supply/rt9455_charger.c rt9455_vmreg_values[idx]); idx 248 drivers/power/supply/ucs1002_power.c int ret, idx; idx 250 drivers/power/supply/ucs1002_power.c for (idx = 0; idx < ARRAY_SIZE(ucs1002_current_limit_uA); idx++) { idx 251 drivers/power/supply/ucs1002_power.c if (val == ucs1002_current_limit_uA[idx]) idx 255 drivers/power/supply/ucs1002_power.c if (idx == ARRAY_SIZE(ucs1002_current_limit_uA)) idx 258 drivers/power/supply/ucs1002_power.c ret = regmap_write(info->regmap, UCS1002_REG_ILIMIT, idx); idx 270 drivers/power/supply/ucs1002_power.c if (reg != idx) idx 1964 drivers/rapidio/devices/tsi721.c int idx = TSI721_VECT_OMB0_DONE + mbox; idx 1967 drivers/rapidio/devices/tsi721.c rc = request_irq(priv->msix[idx].vector, tsi721_omsg_msix, 0, idx 1968 drivers/rapidio/devices/tsi721.c priv->msix[idx].irq_name, (void *)priv); idx 1977 drivers/rapidio/devices/tsi721.c idx = TSI721_VECT_OMB0_INT + mbox; idx 1978 drivers/rapidio/devices/tsi721.c rc = request_irq(priv->msix[idx].vector, tsi721_omsg_msix, 0, idx 1979 drivers/rapidio/devices/tsi721.c priv->msix[idx].irq_name, (void *)priv); idx 1984 drivers/rapidio/devices/tsi721.c idx = TSI721_VECT_OMB0_DONE + mbox; idx 1985 drivers/rapidio/devices/tsi721.c free_irq(priv->msix[idx].vector, (void *)priv); idx 2279 drivers/rapidio/devices/tsi721.c int idx = TSI721_VECT_IMB0_RCV + mbox; idx 2282 drivers/rapidio/devices/tsi721.c rc = request_irq(priv->msix[idx].vector, tsi721_imsg_msix, 0, idx 2283 drivers/rapidio/devices/tsi721.c priv->msix[idx].irq_name, (void *)priv); idx 2292 drivers/rapidio/devices/tsi721.c idx = TSI721_VECT_IMB0_INT + mbox; idx 2293 drivers/rapidio/devices/tsi721.c rc = request_irq(priv->msix[idx].vector, tsi721_imsg_msix, 0, idx 2294 drivers/rapidio/devices/tsi721.c priv->msix[idx].irq_name, (void *)priv); idx 148 drivers/rapidio/devices/tsi721_dma.c int rc, idx; idx 150 drivers/rapidio/devices/tsi721_dma.c idx = TSI721_VECT_DMA0_DONE + bdma_chan->id; idx 152 drivers/rapidio/devices/tsi721_dma.c rc = request_irq(priv->msix[idx].vector, tsi721_bdma_msix, 0, idx 153 drivers/rapidio/devices/tsi721_dma.c priv->msix[idx].irq_name, (void *)bdma_chan); idx 162 drivers/rapidio/devices/tsi721_dma.c idx = TSI721_VECT_DMA0_INT + bdma_chan->id; idx 164 drivers/rapidio/devices/tsi721_dma.c rc = request_irq(priv->msix[idx].vector, tsi721_bdma_msix, 0, idx 165 drivers/rapidio/devices/tsi721_dma.c priv->msix[idx].irq_name, (void *)bdma_chan); idx 421 drivers/rapidio/devices/tsi721_dma.c u32 idx, rd_idx; idx 443 drivers/rapidio/devices/tsi721_dma.c idx = bdma_chan->wr_count_next % (bdma_chan->bd_num + 1); idx 444 drivers/rapidio/devices/tsi721_dma.c if (idx == bdma_chan->bd_num) { idx 446 drivers/rapidio/devices/tsi721_dma.c idx = 0; idx 451 drivers/rapidio/devices/tsi721_dma.c bdma_chan->id, rd_idx, idx); idx 484 drivers/rapidio/devices/tsi721_dma.c if (i && idx == rd_idx) { idx 493 drivers/rapidio/devices/tsi721_dma.c bd_ptr = &((struct tsi721_dma_desc *)bdma_chan->bd_base)[idx]; idx 507 drivers/rapidio/devices/tsi721_dma.c if (++idx == bdma_chan->bd_num) { idx 509 drivers/rapidio/devices/tsi721_dma.c idx = 0; idx 234 drivers/ras/cec.c static void del_elem(struct ce_array *ca, int idx) idx 237 drivers/ras/cec.c if (ca->n - (idx + 1)) idx 238 drivers/ras/cec.c memmove((void *)&ca->array[idx], idx 239 drivers/ras/cec.c (void *)&ca->array[idx + 1], idx 240 drivers/ras/cec.c (ca->n - (idx + 1)) * sizeof(u64)); idx 4664 drivers/regulator/core.c struct attribute *attr, int idx) idx 40 drivers/regulator/isl6271a-regulator.c int idx; idx 44 drivers/regulator/isl6271a-regulator.c idx = i2c_smbus_read_byte(pmic->client); idx 45 drivers/regulator/isl6271a-regulator.c if (idx < 0) idx 49 drivers/regulator/isl6271a-regulator.c return idx; idx 207 drivers/regulator/lp8788-buck.c u8 val, idx, addr; idx 217 drivers/regulator/lp8788-buck.c idx = gpio_get_value(b1_dvs->gpio) ? 1 : 0; idx 220 drivers/regulator/lp8788-buck.c idx = (val & LP8788_BUCK1_DVS_M) >> LP8788_BUCK1_DVS_S; idx 222 drivers/regulator/lp8788-buck.c addr = LP8788_BUCK1_VOUT0 + idx; idx 234 drivers/regulator/lp8788-buck.c idx = 0; idx 236 drivers/regulator/lp8788-buck.c idx = 2; idx 238 drivers/regulator/lp8788-buck.c idx = 1; idx 240 drivers/regulator/lp8788-buck.c idx = 3; idx 243 drivers/regulator/lp8788-buck.c idx = (val & LP8788_BUCK2_DVS_M) >> LP8788_BUCK2_DVS_S; idx 245 drivers/regulator/lp8788-buck.c addr = LP8788_BUCK2_VOUT0 + idx; idx 256 drivers/regulator/mt6358-regulator.c int idx, ret; idx 262 drivers/regulator/mt6358-regulator.c idx = pvol[selector]; idx 265 drivers/regulator/mt6358-regulator.c idx << info->vsel_shift); idx 272 drivers/regulator/mt6358-regulator.c int idx, ret; idx 287 drivers/regulator/mt6358-regulator.c for (idx = 0; idx < info->desc.n_voltages; idx++) { idx 288 drivers/regulator/mt6358-regulator.c if (pvol[idx] == selector) idx 289 drivers/regulator/mt6358-regulator.c return idx; idx 1473 drivers/regulator/palmas-regulator.c int idx, ret; idx 1489 drivers/regulator/palmas-regulator.c for (idx = 0; idx < ddata->max_reg; idx++) { idx 1494 drivers/regulator/palmas-regulator.c match = &ddata->palmas_matches[idx]; idx 1504 drivers/regulator/palmas-regulator.c pdata->reg_data[idx] = match->init_data; idx 1505 drivers/regulator/palmas-regulator.c pdata->reg_init[idx] = rinit; idx 1545 drivers/regulator/palmas-regulator.c if (idx == PALMAS_REG_LDO8) idx 166 drivers/regulator/pbias-regulator.c int count, idx, data_idx = 0; idx 203 drivers/regulator/pbias-regulator.c for (idx = 0; idx < PBIAS_NUM_REGS && data_idx < count; idx++) { idx 204 drivers/regulator/pbias-regulator.c if (!pbias_matches[idx].init_data || idx 205 drivers/regulator/pbias-regulator.c !pbias_matches[idx].of_node) idx 208 drivers/regulator/pbias-regulator.c info = pbias_matches[idx].driver_data; idx 228 drivers/regulator/pbias-regulator.c cfg.init_data = pbias_matches[idx].init_data; idx 230 drivers/regulator/pbias-regulator.c cfg.of_node = pbias_matches[idx].of_node; idx 316 drivers/regulator/tps65090-regulator.c int idx = 0, ret; idx 346 drivers/regulator/tps65090-regulator.c for (idx = 0; idx < ARRAY_SIZE(tps65090_matches); idx++) { idx 350 drivers/regulator/tps65090-regulator.c rpdata = ®_pdata[idx]; idx 351 drivers/regulator/tps65090-regulator.c ri_data = tps65090_matches[idx].init_data; idx 352 drivers/regulator/tps65090-regulator.c if (!ri_data || !tps65090_matches[idx].of_node) idx 357 drivers/regulator/tps65090-regulator.c tps65090_matches[idx].of_node, idx 370 drivers/regulator/tps65090-regulator.c tps65090_matches[idx].of_node, idx 382 drivers/regulator/tps65090-regulator.c if (of_property_read_u32(tps65090_matches[idx].of_node, idx 387 drivers/regulator/tps65090-regulator.c tps65090_pdata->reg_pdata[idx] = rpdata; idx 1006 drivers/regulator/tps65910-regulator.c int idx = 0, ret, count; idx 1045 drivers/regulator/tps65910-regulator.c for (idx = 0; idx < count; idx++) { idx 1046 drivers/regulator/tps65910-regulator.c if (!matches[idx].of_node) idx 1049 drivers/regulator/tps65910-regulator.c pmic_plat_data->tps65910_pmic_init_data[idx] = idx 1050 drivers/regulator/tps65910-regulator.c matches[idx].init_data; idx 1052 drivers/regulator/tps65910-regulator.c ret = of_property_read_u32(matches[idx].of_node, idx 1055 drivers/regulator/tps65910-regulator.c pmic_plat_data->regulator_ext_sleep_control[idx] = prop; idx 403 drivers/remoteproc/remoteproc_core.c int idx = rvring - rvring->rvdev->vring; idx 410 drivers/remoteproc/remoteproc_core.c rsc->vring[idx].da = 0; idx 411 drivers/remoteproc/remoteproc_core.c rsc->vring[idx].notifyid = -1; idx 411 drivers/remoteproc/stm32_rproc.c int err, dummy_data, idx; idx 415 drivers/remoteproc/stm32_rproc.c idx = stm32_rproc_mbox_idx(rproc, STM32_MBX_SHUTDOWN); idx 416 drivers/remoteproc/stm32_rproc.c if (idx >= 0 && ddata->mb[idx].chan) { idx 418 drivers/remoteproc/stm32_rproc.c err = mbox_send_message(ddata->mb[idx].chan, idx 44 drivers/reset/hisilicon/hi6220_reset.c unsigned long idx) idx 48 drivers/reset/hisilicon/hi6220_reset.c u32 bank = idx >> 8; idx 49 drivers/reset/hisilicon/hi6220_reset.c u32 offset = idx & 0xff; idx 56 drivers/reset/hisilicon/hi6220_reset.c unsigned long idx) idx 60 drivers/reset/hisilicon/hi6220_reset.c u32 bank = idx >> 8; idx 61 drivers/reset/hisilicon/hi6220_reset.c u32 offset = idx & 0xff; idx 73 drivers/reset/hisilicon/hi6220_reset.c unsigned long idx) idx 78 drivers/reset/hisilicon/hi6220_reset.c return regmap_write(regmap, SC_MEDIA_RSTEN, BIT(idx)); idx 82 drivers/reset/hisilicon/hi6220_reset.c unsigned long idx) idx 87 drivers/reset/hisilicon/hi6220_reset.c return regmap_write(regmap, SC_MEDIA_RSTDIS, BIT(idx)); idx 23 drivers/reset/hisilicon/reset-hi3660.c unsigned long idx, bool assert) idx 26 drivers/reset/hisilicon/reset-hi3660.c unsigned int offset = idx >> 8; idx 27 drivers/reset/hisilicon/reset-hi3660.c unsigned int mask = BIT(idx & 0x1f); idx 36 drivers/reset/hisilicon/reset-hi3660.c unsigned long idx) idx 38 drivers/reset/hisilicon/reset-hi3660.c return hi3660_reset_program_hw(rcdev, idx, true); idx 42 drivers/reset/hisilicon/reset-hi3660.c unsigned long idx) idx 44 drivers/reset/hisilicon/reset-hi3660.c return hi3660_reset_program_hw(rcdev, idx, false); idx 48 drivers/reset/hisilicon/reset-hi3660.c unsigned long idx) idx 52 drivers/reset/hisilicon/reset-hi3660.c err = hi3660_reset_assert(rcdev, idx); idx 56 drivers/reset/hisilicon/reset-hi3660.c return hi3660_reset_deassert(rcdev, idx); idx 51 drivers/reset/reset-qcom-aoss.c unsigned long idx) idx 54 drivers/reset/reset-qcom-aoss.c const struct qcom_aoss_reset_map *map = &data->desc->resets[idx]; idx 63 drivers/reset/reset-qcom-aoss.c unsigned long idx) idx 66 drivers/reset/reset-qcom-aoss.c const struct qcom_aoss_reset_map *map = &data->desc->resets[idx]; idx 75 drivers/reset/reset-qcom-aoss.c unsigned long idx) idx 77 drivers/reset/reset-qcom-aoss.c qcom_aoss_control_assert(rcdev, idx); idx 79 drivers/reset/reset-qcom-aoss.c return qcom_aoss_control_deassert(rcdev, idx); idx 54 drivers/reset/reset-qcom-pdc.c unsigned long idx) idx 59 drivers/reset/reset-qcom-pdc.c BIT(sdm845_pdc_resets[idx].bit), idx 60 drivers/reset/reset-qcom-pdc.c BIT(sdm845_pdc_resets[idx].bit)); idx 64 drivers/reset/reset-qcom-pdc.c unsigned long idx) idx 69 drivers/reset/reset-qcom-pdc.c BIT(sdm845_pdc_resets[idx].bit), 0); idx 49 drivers/reset/sti/reset-syscfg.c unsigned long idx, int assert) idx 56 drivers/reset/sti/reset-syscfg.c if (idx >= rcdev->nr_resets) idx 59 drivers/reset/sti/reset-syscfg.c ch = &rst->channels[idx]; idx 88 drivers/reset/sti/reset-syscfg.c unsigned long idx) idx 90 drivers/reset/sti/reset-syscfg.c return syscfg_reset_program_hw(rcdev, idx, true); idx 94 drivers/reset/sti/reset-syscfg.c unsigned long idx) idx 96 drivers/reset/sti/reset-syscfg.c return syscfg_reset_program_hw(rcdev, idx, false); idx 100 drivers/reset/sti/reset-syscfg.c unsigned long idx) idx 104 drivers/reset/sti/reset-syscfg.c err = syscfg_reset_assert(rcdev, idx); idx 108 drivers/reset/sti/reset-syscfg.c return syscfg_reset_deassert(rcdev, idx); idx 112 drivers/reset/sti/reset-syscfg.c unsigned long idx) idx 119 drivers/reset/sti/reset-syscfg.c if (idx >= rcdev->nr_resets) idx 122 drivers/reset/sti/reset-syscfg.c ch = &rst->channels[idx]; idx 60 drivers/rtc/rtc-da9052.c int idx = 1; idx 71 drivers/rtc/rtc-da9052.c DA9052_ALARM_MI_REG, 5, &v[idx][0]); idx 89 drivers/rtc/rtc-da9052.c idx = (1-idx); idx 160 drivers/rtc/rtc-da9052.c int idx = 1; idx 171 drivers/rtc/rtc-da9052.c DA9052_COUNT_S_REG, 6, &v[idx][0]); idx 188 drivers/rtc/rtc-da9052.c idx = (1-idx); idx 3232 drivers/s390/block/dasd.c unsigned int idx) idx 3245 drivers/s390/block/dasd.c static void dasd_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int idx) idx 332 drivers/s390/block/scm_blk.c unsigned int idx) idx 345 drivers/s390/block/scm_blk.c static void scm_blk_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int idx) idx 162 drivers/s390/cio/chp.c int idx; idx 166 drivers/s390/cio/chp.c idx = chpid.id; idx 169 drivers/s390/cio/chp.c idx = chpid.id - 128; idx 171 drivers/s390/cio/chp.c entry = area + (idx * sizeof(struct cmg_entry)); idx 1115 drivers/s390/cio/chsc.c int __init chsc_get_cssid(int idx) idx 1151 drivers/s390/cio/chsc.c if ((addr_t) &sdcal_area->list[idx] < idx 1153 drivers/s390/cio/chsc.c ret = sdcal_area->list[idx].cssid; idx 254 drivers/s390/cio/chsc.h int __init chsc_get_cssid(int idx); idx 977 drivers/s390/cio/cmf.c static ssize_t cmb_show_attr(struct device *dev, char *buf, enum cmb_index idx) idx 980 drivers/s390/cio/cmf.c (unsigned long long) cmf_read(to_ccwdev(dev), idx)); idx 347 drivers/s390/cio/vfio_ccw_cp.c static void ccwchain_cda_free(struct ccwchain *chain, int idx) idx 349 drivers/s390/cio/vfio_ccw_cp.c struct ccw1 *ccw = chain->ch_ccw + idx; idx 487 drivers/s390/cio/vfio_ccw_cp.c int idx, idx 490 drivers/s390/cio/vfio_ccw_cp.c struct ccw1 *ccw = chain->ch_ccw + idx; idx 507 drivers/s390/cio/vfio_ccw_cp.c int idx, idx 519 drivers/s390/cio/vfio_ccw_cp.c ccw = chain->ch_ccw + idx; idx 550 drivers/s390/cio/vfio_ccw_cp.c pa = chain->ch_pa + idx; idx 607 drivers/s390/cio/vfio_ccw_cp.c int idx, idx 610 drivers/s390/cio/vfio_ccw_cp.c struct ccw1 *ccw = chain->ch_ccw + idx; idx 613 drivers/s390/cio/vfio_ccw_cp.c return ccwchain_fetch_tic(chain, idx, cp); idx 615 drivers/s390/cio/vfio_ccw_cp.c return ccwchain_fetch_direct(chain, idx, cp); idx 732 drivers/s390/cio/vfio_ccw_cp.c int len, idx, ret; idx 740 drivers/s390/cio/vfio_ccw_cp.c for (idx = 0; idx < len; idx++) { idx 741 drivers/s390/cio/vfio_ccw_cp.c ret = ccwchain_fetch_one(chain, idx, cp); idx 750 drivers/s390/cio/vfio_ccw_cp.c chain->ch_len = idx; idx 90 drivers/s390/net/fsm.c int idx = 0; idx 94 drivers/s390/net/fsm.c idx = fi->history_index; idx 98 drivers/s390/net/fsm.c int e = fi->history[idx].event; idx 99 drivers/s390/net/fsm.c int s = fi->history[idx++].state; idx 100 drivers/s390/net/fsm.c idx %= FSM_HISTORY_SIZE; idx 173 drivers/s390/net/ism.h u64 idx; idx 208 drivers/s390/net/ism.h #define ISM_CREATE_REQ(dmb, idx, sf, offset) \ idx 209 drivers/s390/net/ism.h ((dmb) | (idx) << 24 | (sf) << 23 | (offset)) idx 365 drivers/s390/net/ism_drv.c static int ism_move(struct smcd_dev *smcd, u64 dmb_tok, unsigned int idx, idx 375 drivers/s390/net/ism_drv.c dmb_req = ISM_CREATE_REQ(dmb_tok, idx, size == bytes ? sf : 0, idx 394 drivers/s390/net/ism_drv.c while ((ism->ieq_idx + 1) != READ_ONCE(ism->ieq->header.idx)) { idx 2449 drivers/s390/scsi/zfcp_fsf.c int idx; idx 2451 drivers/s390/scsi/zfcp_fsf.c for (idx = 0; idx < QDIO_MAX_ELEMENTS_PER_BUFFER; idx++) { idx 2453 drivers/s390/scsi/zfcp_fsf.c sbale = &sbal->element[idx]; idx 63 drivers/s390/scsi/zfcp_qdio.c int queue_no, int idx, int count, idx 74 drivers/s390/scsi/zfcp_qdio.c zfcp_qdio_zero_sbals(qdio->req_q, idx, count); idx 84 drivers/s390/scsi/zfcp_qdio.c int queue_no, int idx, int count, idx 100 drivers/s390/scsi/zfcp_qdio.c sbale = qdio->res_q[idx]->element; idx 107 drivers/s390/scsi/zfcp_qdio.c sbal_idx = (idx + sbal_no) % idx 122 drivers/s390/scsi/zfcp_qdio.c sbal_idx = (idx + sbal_no) % QDIO_MAX_BUFFERS_PER_Q; idx 130 drivers/s390/scsi/zfcp_qdio.c if (do_QDIO(cdev, QDIO_FLAG_SYNC_INPUT, 0, idx, count)) idx 345 drivers/s390/scsi/zfcp_qdio.c int idx, count; idx 362 drivers/s390/scsi/zfcp_qdio.c idx = (qdio->req_q_idx + count) % QDIO_MAX_BUFFERS_PER_Q; idx 364 drivers/s390/scsi/zfcp_qdio.c zfcp_qdio_zero_sbals(qdio->req_q, idx, count); idx 438 drivers/sbus/char/oradax.c static int dax_lock_pages(struct dax_ctx *ctx, int idx, idx 455 drivers/sbus/char/oradax.c &ctx->pages[i + idx][OUT]) != 0) { idx 465 drivers/sbus/char/oradax.c &ctx->pages[i + idx][PRI]) != 0) { idx 475 drivers/sbus/char/oradax.c &ctx->pages[i + idx][SEC]) != 0) { idx 485 drivers/sbus/char/oradax.c &ctx->pages[i + idx][TBL]) != 0) { idx 499 drivers/sbus/char/oradax.c dax_unlock_pages(ctx, idx, nelem); idx 503 drivers/sbus/char/oradax.c static void dax_ccb_wait(struct dax_ctx *ctx, int idx) idx 508 drivers/sbus/char/oradax.c dax_dbg("idx=%d", idx); idx 511 drivers/sbus/char/oradax.c if (ctx->ca_buf[idx].status == CCA_STAT_NOT_COMPLETED) idx 517 drivers/sbus/char/oradax.c (void *)ctx, idx, DAX_CCB_USEC, DAX_CCB_RETRIES); idx 519 drivers/sbus/char/oradax.c ret = dax_ccb_kill(ctx->ca_buf_ra + idx * sizeof(struct dax_cca), idx 521 drivers/sbus/char/oradax.c dax_dbg("Kill CCB[%d] %s", idx, ret ? "failed" : "succeeded"); idx 570 drivers/sbus/char/oradax.c int i, idx, ret; idx 608 drivers/sbus/char/oradax.c idx = hdr.ca_offset / sizeof(struct dax_cca); idx 609 drivers/sbus/char/oradax.c ctx->ca_buf[idx].status = CCA_STAT_KILLED; idx 610 drivers/sbus/char/oradax.c ctx->ca_buf[idx].err = CCA_ERR_KILLED; idx 779 drivers/sbus/char/oradax.c static int dax_preprocess_usr_ccbs(struct dax_ctx *ctx, int idx, int nelem) idx 837 drivers/sbus/char/oradax.c ca_offset = (idx + i) * sizeof(struct dax_cca); idx 839 drivers/sbus/char/oradax.c memset(&ctx->ca_buf[idx + i], 0, sizeof(struct dax_cca)); idx 856 drivers/sbus/char/oradax.c int i, idx, nccbs, naccepted; idx 859 drivers/sbus/char/oradax.c idx = *ppos; idx 867 drivers/sbus/char/oradax.c dax_dbg("args: ccb_buf_len=%ld, idx=%d", count, idx); idx 870 drivers/sbus/char/oradax.c if (idx < 0 || idx > (DAX_CA_ELEMS - nccbs)) { idx 886 drivers/sbus/char/oradax.c for (i = idx; i < idx + nccbs; i++) { idx 893 drivers/sbus/char/oradax.c dax_unlock_pages(ctx, idx, nccbs); idx 895 drivers/sbus/char/oradax.c ctx->result.exec.status = dax_preprocess_usr_ccbs(ctx, idx, nccbs); idx 899 drivers/sbus/char/oradax.c ctx->result.exec.status = dax_lock_pages(ctx, idx, nccbs, idx 976 drivers/sbus/char/oradax.c dax_unlock_pages(ctx, idx + naccepted, nccbs - naccepted); idx 979 drivers/sbus/char/oradax.c for (i = idx + naccepted; i < idx + nccbs; i++) idx 668 drivers/scsi/a100u2w.c unsigned long idx; idx 677 drivers/scsi/a100u2w.c idx = index + 32 * i; idx 681 drivers/scsi/a100u2w.c return host->scb_virt + idx; idx 360 drivers/scsi/aacraid/commsup.c unsigned long idx; idx 371 drivers/scsi/aacraid/commsup.c idx = *index = le32_to_cpu(*(q->headers.producer)); idx 373 drivers/scsi/aacraid/commsup.c if (idx != le32_to_cpu(*(q->headers.consumer))) { idx 374 drivers/scsi/aacraid/commsup.c if (--idx == 0) { idx 376 drivers/scsi/aacraid/commsup.c idx = ADAP_NORM_CMD_ENTRIES; idx 378 drivers/scsi/aacraid/commsup.c idx = ADAP_NORM_RESP_ENTRIES; idx 380 drivers/scsi/aacraid/commsup.c if (idx != le32_to_cpu(*(q->headers.consumer))) idx 2605 drivers/scsi/be2iscsi/be_main.c unsigned int num_cxn_wrb = 0, j, idx = 0, index; idx 2647 drivers/scsi/be2iscsi/be_main.c mem_descr_wrbh->mem_array[idx].virtual_address; idx 2648 drivers/scsi/be2iscsi/be_main.c num_cxn_wrbh = ((mem_descr_wrbh->mem_array[idx].size) / idx 2651 drivers/scsi/be2iscsi/be_main.c idx++; idx 2670 drivers/scsi/be2iscsi/be_main.c idx = 0; idx 2674 drivers/scsi/be2iscsi/be_main.c pwrb = mem_descr_wrb->mem_array[idx].virtual_address; idx 2675 drivers/scsi/be2iscsi/be_main.c num_cxn_wrb = (mem_descr_wrb->mem_array[idx].size) / idx 2678 drivers/scsi/be2iscsi/be_main.c idx++; idx 2707 drivers/scsi/be2iscsi/be_main.c unsigned int index, idx, num_per_mem, num_async_data; idx 2855 drivers/scsi/be2iscsi/be_main.c idx = 0; idx 2859 drivers/scsi/be2iscsi/be_main.c mem_descr->mem_array[idx].virtual_address; idx 2861 drivers/scsi/be2iscsi/be_main.c mem_descr->mem_array[idx]. idx 2864 drivers/scsi/be2iscsi/be_main.c num_async_data = ((mem_descr->mem_array[idx].size) / idx 2897 drivers/scsi/be2iscsi/be_main.c idx++; idx 2899 drivers/scsi/be2iscsi/be_main.c mem_descr->mem_array[idx]. idx 2903 drivers/scsi/be2iscsi/be_main.c mem_descr->mem_array[idx]. idx 2906 drivers/scsi/be2iscsi/be_main.c ((mem_descr->mem_array[idx]. idx 3131 drivers/scsi/be2iscsi/be_main.c unsigned int idx; idx 3138 drivers/scsi/be2iscsi/be_main.c idx = 0; idx 3145 drivers/scsi/be2iscsi/be_main.c dq_vaddress = mem_descr->mem_array[idx].virtual_address; idx 3156 drivers/scsi/be2iscsi/be_main.c mem->dma = (unsigned long)mem_descr->mem_array[idx]. idx 3183 drivers/scsi/be2iscsi/be_main.c unsigned int idx; idx 3190 drivers/scsi/be2iscsi/be_main.c idx = 0; idx 3197 drivers/scsi/be2iscsi/be_main.c dq_vaddress = mem_descr->mem_array[idx].virtual_address; idx 3209 drivers/scsi/be2iscsi/be_main.c mem->dma = (unsigned long)mem_descr->mem_array[idx]. idx 3339 drivers/scsi/be2iscsi/be_main.c unsigned int idx, num, i, ulp_num; idx 3349 drivers/scsi/be2iscsi/be_main.c idx = 0; idx 3360 drivers/scsi/be2iscsi/be_main.c wrb_vaddr = mem_descr->mem_array[idx].virtual_address; idx 3361 drivers/scsi/be2iscsi/be_main.c pa_addr_lo = mem_descr->mem_array[idx].bus_address.u.a64.address; idx 3362 drivers/scsi/be2iscsi/be_main.c num_wrb_rings = mem_descr->mem_array[idx].size / idx 3375 drivers/scsi/be2iscsi/be_main.c idx++; idx 3376 drivers/scsi/be2iscsi/be_main.c wrb_vaddr = mem_descr->mem_array[idx].virtual_address; idx 3377 drivers/scsi/be2iscsi/be_main.c pa_addr_lo = mem_descr->mem_array[idx].\ idx 3379 drivers/scsi/be2iscsi/be_main.c num_wrb_rings = mem_descr->mem_array[idx].size / idx 3885 drivers/scsi/be2iscsi/be_main.c unsigned int arr_index, i, idx; idx 3920 drivers/scsi/be2iscsi/be_main.c idx = 0; idx 3921 drivers/scsi/be2iscsi/be_main.c while (idx < mem_descr_sglh->num_elements) { idx 3922 drivers/scsi/be2iscsi/be_main.c psgl_handle = mem_descr_sglh->mem_array[idx].virtual_address; idx 3924 drivers/scsi/be2iscsi/be_main.c for (i = 0; i < (mem_descr_sglh->mem_array[idx].size / idx 3939 drivers/scsi/be2iscsi/be_main.c idx++; idx 3960 drivers/scsi/be2iscsi/be_main.c idx = 0; idx 3961 drivers/scsi/be2iscsi/be_main.c while (idx < mem_descr_sg->num_elements) { idx 3962 drivers/scsi/be2iscsi/be_main.c pfrag = mem_descr_sg->mem_array[idx].virtual_address; idx 3965 drivers/scsi/be2iscsi/be_main.c i < (mem_descr_sg->mem_array[idx].size) / idx 3979 drivers/scsi/be2iscsi/be_main.c idx++; idx 444 drivers/scsi/bfa/bfa_fcpim.c u32 val, idx; idx 447 drivers/scsi/bfa/bfa_fcpim.c idx = bfa_ioim_get_index(scsi_bufflen((struct scsi_cmnd *)ioim->dio)); idx 448 drivers/scsi/bfa/bfa_fcpim.c bfa_itnim_ioprofile_update(ioim->itnim, idx); idx 450 drivers/scsi/bfa/bfa_fcpim.c io_lat->count[idx]++; idx 451 drivers/scsi/bfa/bfa_fcpim.c io_lat->min[idx] = (io_lat->min[idx] < val) ? io_lat->min[idx] : val; idx 452 drivers/scsi/bfa/bfa_fcpim.c io_lat->max[idx] = (io_lat->max[idx] > val) ? io_lat->max[idx] : val; idx 453 drivers/scsi/bfa/bfa_fcpim.c io_lat->avg[idx] += val; idx 3636 drivers/scsi/bfa/bfa_fcpim.c u16 nsegs, idx, per_seg_ios, num_io_req; idx 3673 drivers/scsi/bfa/bfa_fcpim.c bfa_mem_dma_seg_iter(fcp, seg_ptr, nsegs, idx) { idx 3693 drivers/scsi/bfa/bfa_fcpim.c u16 idx, nsegs, num_io_req; idx 3708 drivers/scsi/bfa/bfa_fcpim.c bfa_mem_dma_seg_iter(fcp, seg_ptr, nsegs, idx) { idx 3713 drivers/scsi/bfa/bfa_fcpim.c fcp->snsbase[idx].pa = bfa_mem_dma_phys(seg_ptr); idx 3714 drivers/scsi/bfa/bfa_fcpim.c fcp->snsbase[idx].kva = bfa_mem_dma_virt(seg_ptr); idx 3715 drivers/scsi/bfa/bfa_fcpim.c bfa_iocfc_set_snsbase(bfa, idx, fcp->snsbase[idx].pa); idx 458 drivers/scsi/bfa/bfa_svc.c u16 nsegs, idx, per_seg_fcxp; idx 474 drivers/scsi/bfa/bfa_svc.c bfa_mem_dma_seg_iter(fcxp_mod, seg_ptr, nsegs, idx) { idx 5150 drivers/scsi/bfa/bfa_svc.c u16 nsegs, idx, per_seg_sgpg, num_sgpg; idx 5163 drivers/scsi/bfa/bfa_svc.c bfa_mem_dma_seg_iter(sgpg_mod, seg_ptr, nsegs, idx) { idx 5188 drivers/scsi/bfa/bfa_svc.c u16 i, idx, nsegs, per_seg_sgpg, num_sgpg; idx 5208 drivers/scsi/bfa/bfa_svc.c bfa_mem_dma_seg_iter(mod, seg_ptr, nsegs, idx) { idx 5432 drivers/scsi/bfa/bfa_svc.c u16 nsegs, idx, per_seg_uf = 0; idx 5437 drivers/scsi/bfa/bfa_svc.c bfa_mem_dma_seg_iter(ufm, seg_ptr, nsegs, idx) { idx 279 drivers/scsi/bnx2i/bnx2i_iscsi.c int idx; idx 284 drivers/scsi/bnx2i/bnx2i_iscsi.c idx = hba->cid_que.cid_q_cons_idx; idx 290 drivers/scsi/bnx2i/bnx2i_iscsi.c return hba->cid_que.cid_que[idx]; idx 301 drivers/scsi/bnx2i/bnx2i_iscsi.c int idx; idx 308 drivers/scsi/bnx2i/bnx2i_iscsi.c idx = hba->cid_que.cid_q_prod_idx; idx 309 drivers/scsi/bnx2i/bnx2i_iscsi.c hba->cid_que.cid_que[idx] = iscsi_cid; idx 3471 drivers/scsi/csiostor/csio_hw.c static void csio_mem_intr_handler(struct csio_hw *hw, int idx) idx 3477 drivers/scsi/csiostor/csio_hw.c if (idx <= MEM_EDC1) { idx 3478 drivers/scsi/csiostor/csio_hw.c addr = EDC_REG(EDC_INT_CAUSE_A, idx); idx 3479 drivers/scsi/csiostor/csio_hw.c cnt_addr = EDC_REG(EDC_ECC_STATUS_A, idx); idx 3487 drivers/scsi/csiostor/csio_hw.c csio_fatal(hw, "%s FIFO parity error\n", name[idx]); idx 3493 drivers/scsi/csiostor/csio_hw.c cnt, name[idx], cnt > 1 ? "s" : ""); idx 3496 drivers/scsi/csiostor/csio_hw.c csio_fatal(hw, "%s uncorrectable ECC data error\n", name[idx]); idx 82 drivers/scsi/csiostor/csio_hw_chip.h #define CSIO_DEVICE(devid, idx) \ idx 83 drivers/scsi/csiostor/csio_hw_chip.h { PCI_VENDOR_ID_CHELSIO, (devid), PCI_ANY_ID, PCI_ANY_ID, 0, 0, (idx) } idx 146 drivers/scsi/csiostor/csio_hw_t5.c csio_t5_mc_read(struct csio_hw *hw, int idx, uint32_t addr, __be32 *data, idx 153 drivers/scsi/csiostor/csio_hw_t5.c mc_bist_cmd_reg = MC_REG(MC_P_BIST_CMD_A, idx); idx 154 drivers/scsi/csiostor/csio_hw_t5.c mc_bist_cmd_addr_reg = MC_REG(MC_P_BIST_CMD_ADDR_A, idx); idx 155 drivers/scsi/csiostor/csio_hw_t5.c mc_bist_cmd_len_reg = MC_REG(MC_P_BIST_CMD_LEN_A, idx); idx 156 drivers/scsi/csiostor/csio_hw_t5.c mc_bist_status_rdata_reg = MC_REG(MC_P_BIST_STATUS_RDATA_A, idx); idx 157 drivers/scsi/csiostor/csio_hw_t5.c mc_bist_data_pattern_reg = MC_REG(MC_P_BIST_DATA_PATTERN_A, idx); idx 194 drivers/scsi/csiostor/csio_hw_t5.c csio_t5_edc_read(struct csio_hw *hw, int idx, uint32_t addr, __be32 *data, idx 205 drivers/scsi/csiostor/csio_hw_t5.c #define EDC_REG_T5(reg, idx) (reg + EDC_STRIDE_T5 * idx) idx 207 drivers/scsi/csiostor/csio_hw_t5.c edc_bist_cmd_reg = EDC_REG_T5(EDC_H_BIST_CMD_A, idx); idx 208 drivers/scsi/csiostor/csio_hw_t5.c edc_bist_cmd_addr_reg = EDC_REG_T5(EDC_H_BIST_CMD_ADDR_A, idx); idx 209 drivers/scsi/csiostor/csio_hw_t5.c edc_bist_cmd_len_reg = EDC_REG_T5(EDC_H_BIST_CMD_LEN_A, idx); idx 210 drivers/scsi/csiostor/csio_hw_t5.c edc_bist_cmd_data_pattern = EDC_REG_T5(EDC_H_BIST_DATA_PATTERN_A, idx); idx 211 drivers/scsi/csiostor/csio_hw_t5.c edc_bist_status_rdata_reg = EDC_REG_T5(EDC_H_BIST_STATUS_RDATA_A, idx); idx 227 drivers/scsi/csiostor/csio_hw_t5.c #define EDC_DATA(i) (EDC_BIST_STATUS_REG(EDC_BIST_STATUS_RDATA_A, i) + idx) idx 114 drivers/scsi/csiostor/csio_init.c unsigned int idx, unsigned int size_mb) idx 117 drivers/scsi/csiostor/csio_init.c (void *)hw + idx, &csio_mem_debugfs_fops, idx 339 drivers/scsi/csiostor/csio_init.c int i, j, idx, k = 0; idx 418 drivers/scsi/csiostor/csio_init.c idx = csio_wr_alloc_q(hw, csio_scsi_eqsize, 0, idx 421 drivers/scsi/csiostor/csio_init.c if (idx == -1) { idx 423 drivers/scsi/csiostor/csio_init.c idx); idx 427 drivers/scsi/csiostor/csio_init.c sqset->eq_idx = idx; idx 429 drivers/scsi/csiostor/csio_init.c idx = csio_wr_alloc_q(hw, CSIO_SCSI_IQSIZE, idx 433 drivers/scsi/csiostor/csio_init.c if (idx == -1) { idx 435 drivers/scsi/csiostor/csio_init.c idx); idx 438 drivers/scsi/csiostor/csio_init.c sqset->iq_idx = idx; idx 1393 drivers/scsi/csiostor/csio_lnode.c int idx; idx 1402 drivers/scsi/csiostor/csio_lnode.c for (idx = 1; idx <= 3; idx++) { idx 1403 drivers/scsi/csiostor/csio_lnode.c portparams.idx = (idx-1)*6 + 1; idx 1405 drivers/scsi/csiostor/csio_lnode.c if (idx == 3) idx 1045 drivers/scsi/csiostor/csio_mb.c cmdp->u.ctl.port_valid_ix = FW_FCOE_STATS_CMD_IX(portparams->idx) | idx 1067 drivers/scsi/csiostor/csio_mb.c dst = (uint8_t *)(&stats) + ((portparams->idx - 1) * 8); idx 1070 drivers/scsi/csiostor/csio_mb.c if (portparams->idx == 1) { idx 1079 drivers/scsi/csiostor/csio_mb.c if (portparams->idx == 7) { idx 1092 drivers/scsi/csiostor/csio_mb.c if (portparams->idx == 13) { idx 50 drivers/scsi/csiostor/csio_mb.h uint8_t idx; idx 174 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c V_L2T_IDX(e->idx) | V_TX_CHANNEL(e->smt_idx)); idx 183 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c csk->mss_idx, e->idx, e->smt_idx); idx 1090 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c unsigned int idx = ttinfo->idx; idx 1096 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c unsigned int pm_addr = (idx << PPOD_SIZE_SHIFT) + ppm->llimit; idx 1099 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c for (i = 0; i < npods; i++, idx++, pm_addr += IPPOD_SIZE) { idx 1120 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c unsigned int idx = ttinfo->idx; idx 1121 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c unsigned int pm_addr = (idx << PPOD_SIZE_SHIFT) + ppm->llimit; idx 1127 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c cdev, idx, npods); idx 1129 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c for (i = 0; i < npods; i++, idx++, pm_addr += IPPOD_SIZE) { idx 1135 drivers/scsi/cxgbi/cxgb3i/cxgb3i.c cdev, idx, i, npods); idx 216 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c L2T_IDX_V(((struct l2t_entry *)csk->l2t)->idx) | idx 335 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c L2T_IDX_V(((struct l2t_entry *)csk->l2t)->idx) | idx 1914 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c unsigned int idx, unsigned int npods, idx 1917 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c unsigned int pm_addr = (idx << PPOD_SIZE_SHIFT) + ppm->llimit; idx 1925 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c __func__, ppm->ndev->name, idx, npods); idx 1937 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c unsigned int idx, unsigned int npods, idx 1942 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c struct sk_buff *skb = ddp_ppod_init_idata(cdev, ppm, idx, npods, idx 1973 drivers/scsi/cxgbi/cxgb4i/cxgb4i.c unsigned int pidx = ttinfo->idx; idx 67 drivers/scsi/cxgbi/libcxgbi.c static inline void cxgbi_decode_sw_tag(u32 sw_tag, int *idx, int *age) idx 71 drivers/scsi/cxgbi/libcxgbi.c if (idx) idx 72 drivers/scsi/cxgbi/libcxgbi.c *idx = (sw_tag >> 16) & 0x7FFF; idx 442 drivers/scsi/cxgbi/libcxgbi.c int idx; idx 470 drivers/scsi/cxgbi/libcxgbi.c start = idx = pmap->next; idx 472 drivers/scsi/cxgbi/libcxgbi.c if (++idx >= pmap->max_connect) idx 473 drivers/scsi/cxgbi/libcxgbi.c idx = 0; idx 474 drivers/scsi/cxgbi/libcxgbi.c if (!pmap->port_csk[idx]) { idx 476 drivers/scsi/cxgbi/libcxgbi.c *port = htons(pmap->sport_base + idx); idx 477 drivers/scsi/cxgbi/libcxgbi.c pmap->next = idx; idx 478 drivers/scsi/cxgbi/libcxgbi.c pmap->port_csk[idx] = csk; idx 485 drivers/scsi/cxgbi/libcxgbi.c pmap->sport_base + idx, pmap->next); idx 488 drivers/scsi/cxgbi/libcxgbi.c } while (idx != start); idx 510 drivers/scsi/cxgbi/libcxgbi.c int idx = ntohs(*port) - pmap->sport_base; idx 513 drivers/scsi/cxgbi/libcxgbi.c if (idx < 0 || idx >= pmap->max_connect) { idx 522 drivers/scsi/cxgbi/libcxgbi.c pmap->port_csk[idx] = NULL; idx 529 drivers/scsi/cxgbi/libcxgbi.c pmap->sport_base + idx); idx 1098 drivers/scsi/cxgbi/libcxgbi.c unsigned int idx; idx 1107 drivers/scsi/cxgbi/libcxgbi.c idx = cxgbi_sock_find_best_mtu(csk, csk->advmss + 40); idx 1109 drivers/scsi/cxgbi/libcxgbi.c return idx; idx 1376 drivers/scsi/cxgbi/libcxgbi.c err = cxgbi_ppm_ppods_reserve(ppm, ttinfo->nr_pages, 0, &ttinfo->idx, idx 1418 drivers/scsi/cxgbi/libcxgbi.c cxgbi_ppm_ppod_release(ppm, ttinfo->idx); idx 1446 drivers/scsi/cxgbi/libcxgbi.c cxgbi_ppm_ppod_release(ppm, ttinfo->idx); idx 1452 drivers/scsi/cxgbi/libcxgbi.c static inline u32 cxgbi_build_sw_tag(u32 idx, u32 age) idx 1455 drivers/scsi/cxgbi/libcxgbi.c return (idx << 16) | age; idx 1501 drivers/scsi/cxgbi/libcxgbi.c void cxgbi_parse_pdu_itt(struct iscsi_conn *conn, itt_t itt, int *idx, int *age) idx 1519 drivers/scsi/cxgbi/libcxgbi.c cxgbi_decode_sw_tag(sw_bits, idx, age); idx 1522 drivers/scsi/cxgbi/libcxgbi.c cdev, tag, itt, sw_bits, idx ? *idx : 0xFFFFF, idx 3508 drivers/scsi/cxlflash/main.c int idx = 0; idx 3526 drivers/scsi/cxlflash/main.c __func__, cmd, idx, sizeof(ioctl_tbl)); idx 3532 drivers/scsi/cxlflash/main.c idx = _IOC_NR(HT_CXLFLASH_LUN_PROVISION) - _IOC_NR(cmd); idx 3533 drivers/scsi/cxlflash/main.c size = ioctl_tbl[idx].size; idx 3534 drivers/scsi/cxlflash/main.c do_ioctl = ioctl_tbl[idx].ioctl; idx 2106 drivers/scsi/cxlflash/superpipe.c int idx; idx 2164 drivers/scsi/cxlflash/superpipe.c idx = _IOC_NR(cmd) - _IOC_NR(DK_CXLFLASH_ATTACH); idx 2165 drivers/scsi/cxlflash/superpipe.c size = ioctl_tbl[idx].size; idx 2166 drivers/scsi/cxlflash/superpipe.c do_ioctl = ioctl_tbl[idx].ioctl; idx 239 drivers/scsi/cxlflash/vlun.c int idx = 0, bit_pos = 0; idx 241 drivers/scsi/cxlflash/vlun.c idx = aun / BITS_PER_LONG; idx 244 drivers/scsi/cxlflash/vlun.c if (test_bit(bit_pos, (ulong *)&bali->lun_alloc_map[idx])) idx 259 drivers/scsi/cxlflash/vlun.c int idx = 0, bit_pos = 0; idx 282 drivers/scsi/cxlflash/vlun.c idx = to_free / BITS_PER_LONG; idx 285 drivers/scsi/cxlflash/vlun.c set_bit(bit_pos, (ulong *)&bali->lun_alloc_map[idx]); idx 288 drivers/scsi/cxlflash/vlun.c if (idx < bali->free_low_idx) idx 289 drivers/scsi/cxlflash/vlun.c bali->free_low_idx = idx; idx 290 drivers/scsi/cxlflash/vlun.c else if (idx > bali->free_high_idx) idx 291 drivers/scsi/cxlflash/vlun.c bali->free_high_idx = idx; idx 294 drivers/scsi/cxlflash/vlun.c "lun_id=%016llx free_aun_cnt=%llx\n", __func__, bit_pos, idx, idx 648 drivers/scsi/dc395x.c u8 idx = 0; idx 649 drivers/scsi/dc395x.c while (idx < 7 && eeprom_index_to_delay_map[idx] < delay) idx 650 drivers/scsi/dc395x.c idx++; idx 651 drivers/scsi/dc395x.c return idx; idx 1834 drivers/scsi/dc395x.c unsigned idx = srb->sg_index; idx 1835 drivers/scsi/dc395x.c struct SGentry *psge = srb->segment_x + idx; idx 1836 drivers/scsi/dc395x.c for (; idx < srb->sg_count; psge++, idx++) idx 1852 drivers/scsi/dc395x.c u8 idx; idx 1866 drivers/scsi/dc395x.c for (idx = srb->sg_index; idx < srb->sg_count; idx++) { idx 1877 drivers/scsi/dc395x.c srb->sg_index = idx; idx 3189 drivers/scsi/dpt_i2o.c int idx = 16; idx 3212 drivers/scsi/dpt_i2o.c idx = 0; break; idx 3214 drivers/scsi/dpt_i2o.c idx = 1; break; idx 3216 drivers/scsi/dpt_i2o.c idx = 2; break; idx 3218 drivers/scsi/dpt_i2o.c idx = 3; break; idx 3220 drivers/scsi/dpt_i2o.c idx = 4; break; idx 3222 drivers/scsi/dpt_i2o.c idx = 5; break; idx 3224 drivers/scsi/dpt_i2o.c idx = 6; break; idx 3226 drivers/scsi/dpt_i2o.c idx = 7; break; idx 3228 drivers/scsi/dpt_i2o.c idx = 8; break; idx 3230 drivers/scsi/dpt_i2o.c idx = 9; break; idx 3232 drivers/scsi/dpt_i2o.c idx = 10; break; idx 3234 drivers/scsi/dpt_i2o.c idx = 11; break; idx 3236 drivers/scsi/dpt_i2o.c idx = 12; break; idx 3238 drivers/scsi/dpt_i2o.c idx = 13; break; idx 3240 drivers/scsi/dpt_i2o.c idx = 14; break; idx 3242 drivers/scsi/dpt_i2o.c idx = 15; break; idx 3244 drivers/scsi/dpt_i2o.c return i2o_class_name[idx]; idx 133 drivers/scsi/esp_scsi.c int idx = esp->esp_event_cur; idx 135 drivers/scsi/esp_scsi.c p = &esp->esp_event_log[idx]; idx 140 drivers/scsi/esp_scsi.c esp->esp_event_cur = (idx + 1) & (ESP_EVENT_LOG_SZ - 1); idx 168 drivers/scsi/esp_scsi.c int idx = esp->esp_event_cur; idx 170 drivers/scsi/esp_scsi.c p = &esp->esp_event_log[idx]; idx 175 drivers/scsi/esp_scsi.c esp->esp_event_cur = (idx + 1) & (ESP_EVENT_LOG_SZ - 1); idx 182 drivers/scsi/esp_scsi.c int idx = esp->esp_event_cur; idx 183 drivers/scsi/esp_scsi.c int stop = idx; idx 187 drivers/scsi/esp_scsi.c struct esp_event_ent *p = &esp->esp_event_log[idx]; idx 192 drivers/scsi/esp_scsi.c idx, idx 197 drivers/scsi/esp_scsi.c idx = (idx + 1) & (ESP_EVENT_LOG_SZ - 1); idx 198 drivers/scsi/esp_scsi.c } while (idx != stop); idx 221 drivers/scsi/esp_scsi.c int idx = 0; idx 224 drivers/scsi/esp_scsi.c esp->fifo[idx++] = esp_read8(ESP_FDATA); idx 225 drivers/scsi/esp_scsi.c esp->fifo[idx++] = esp_read8(ESP_FDATA); idx 229 drivers/scsi/esp_scsi.c esp->fifo[idx++] = esp_read8(ESP_FDATA); idx 232 drivers/scsi/esp_scsi.c esp->fifo_cnt = idx; idx 169 drivers/scsi/g_NCR5380.c static void magic_configure(int idx, u8 irq, u8 magic[]) idx 182 drivers/scsi/g_NCR5380.c if (idx >= 0 && idx <= 7) idx 183 drivers/scsi/g_NCR5380.c cfg = 0x80 | idx | (irq << 4); idx 131 drivers/scsi/gdth.c u16 idx, gdth_evt_data *evt); idx 2359 drivers/scsi/gdth.c u16 idx, gdth_evt_data *evt) idx 2364 drivers/scsi/gdth.c TRACE2(("gdth_store_event() source %d idx %d\n", source, idx)); idx 2369 drivers/scsi/gdth.c ebuffer[elastidx].event_idx == idx && idx 2392 drivers/scsi/gdth.c e->event_idx = idx; idx 236 drivers/scsi/hisi_sas/hisi_sas.h u16 idx; idx 256 drivers/scsi/hisi_sas/hisi_sas_main.c hisi_sas_slot_index_free(hisi_hba, slot->idx); idx 1641 drivers/scsi/hisi_sas/hisi_sas_main.c u16 tag = slot->idx; idx 1685 drivers/scsi/hisi_sas/hisi_sas_main.c u32 tag = slot->idx; idx 1908 drivers/scsi/hisi_sas/hisi_sas_main.c u32 tag = slot->idx; idx 2395 drivers/scsi/hisi_sas/hisi_sas_main.c slot->idx = slot_index; idx 943 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx << CMD_HDR_IPTT_OFF); idx 1007 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx << CMD_HDR_IPTT_OFF); idx 1509 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c int idx; idx 1514 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c idx = (cmplt_hdr_data & CMPLT_HDR_IPTT_MSK) >> idx 1516 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c slot = &hisi_hba->slot_info[idx]; idx 1640 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c int i, j, irq, rc, idx; idx 1645 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c idx = i * HISI_SAS_PHY_INT_NR; idx 1646 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c for (j = 0; j < HISI_SAS_PHY_INT_NR; j++, idx++) { idx 1647 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c irq = platform_get_irq(pdev, idx); idx 1650 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c idx); idx 1664 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c idx = hisi_hba->n_phy * HISI_SAS_PHY_INT_NR; idx 1665 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c for (i = 0; i < hisi_hba->queue_count; i++, idx++) { idx 1666 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c irq = platform_get_irq(pdev, idx); idx 1669 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c idx); idx 1682 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c idx = (hisi_hba->n_phy * HISI_SAS_PHY_INT_NR) + hisi_hba->queue_count; idx 1683 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c for (i = 0; i < HISI_SAS_FATAL_INT_NR; i++, idx++) { idx 1684 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c irq = platform_get_irq(pdev, idx); idx 1687 drivers/scsi/hisi_sas/hisi_sas_v1_hw.c idx); idx 821 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c static bool sata_index_alloc_v2_hw(struct hisi_hba *hisi_hba, int *idx) idx 834 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c *idx = index; idx 1724 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx << CMD_HDR_IPTT_OFF); idx 1781 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx); idx 2397 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c slot->idx, task, sas_dev->device_id, idx 2550 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx); idx 2629 drivers/scsi/hisi_sas/hisi_sas_v2_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx); idx 833 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c (slot->idx << CFG_SET_ABORTED_IPTT_OFF); idx 1218 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx); idx 1322 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx << CMD_HDR_IPTT_OFF); idx 1390 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx); idx 1431 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c hdr->transfer_tags = cpu_to_le32(slot->idx); idx 2215 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c slot->idx, task, sas_dev->device_id, idx 5838 drivers/scsi/hpsa.c int idx = scmd->request->tag; idx 5840 drivers/scsi/hpsa.c if (idx < 0) idx 5841 drivers/scsi/hpsa.c return idx; idx 5844 drivers/scsi/hpsa.c return idx += HPSA_NRESERVED_CMDS; idx 6075 drivers/scsi/hpsa.c int idx = hpsa_get_cmd_index(scmd); idx 6076 drivers/scsi/hpsa.c struct CommandList *c = h->cmd_pool + idx; idx 6078 drivers/scsi/hpsa.c if (idx < HPSA_NRESERVED_CMDS || idx >= h->nr_cmds) { idx 6080 drivers/scsi/hpsa.c idx, HPSA_NRESERVED_CMDS, h->nr_cmds - 1); idx 6094 drivers/scsi/hpsa.c if (idx != h->last_collision_tag) { /* Print once per tag */ idx 6096 drivers/scsi/hpsa.c "%s: tag collision (tag=%d)\n", __func__, idx); idx 6099 drivers/scsi/hpsa.c h->last_collision_tag = idx; idx 6106 drivers/scsi/hpsa.c hpsa_cmd_partial_init(h, idx, c); idx 852 drivers/scsi/hptiop.c int idx, nseg; idx 864 drivers/scsi/hptiop.c scsi_for_each_sg(scp, sg, HPT_SCP(scp)->sgcnt, idx) { idx 865 drivers/scsi/hptiop.c psg[idx].pci_address = cpu_to_le64(sg_dma_address(sg)) | idx 867 drivers/scsi/hptiop.c psg[idx].size = cpu_to_le32(sg_dma_len(sg)); idx 868 drivers/scsi/hptiop.c psg[idx].eot = (idx == HPT_SCP(scp)->sgcnt - 1) ? idx 2123 drivers/scsi/initio.c u8 len, idx; idx 2136 drivers/scsi/initio.c for (idx = 1; len != 0; len--) { idx 2144 drivers/scsi/initio.c host->msg[idx++] = inb(host->addr + TUL_SFifo); idx 75 drivers/scsi/isci/request.c int idx) idx 77 drivers/scsi/isci/request.c if (idx == 0) idx 79 drivers/scsi/isci/request.c else if (idx == 1) idx 81 drivers/scsi/isci/request.c else if (idx < 0) idx 84 drivers/scsi/isci/request.c return &ireq->sg_table[idx - 2]; idx 88 drivers/scsi/isci/request.c struct isci_request *ireq, u32 idx) idx 92 drivers/scsi/isci/request.c if (idx == 0) { idx 96 drivers/scsi/isci/request.c } else if (idx == 1) { idx 102 drivers/scsi/isci/request.c return sci_io_request_get_dma_addr(ireq, &ireq->sg_table[idx - 2]); idx 174 drivers/scsi/lpfc/lpfc.h uint16_t idx; idx 80 drivers/scsi/lpfc/lpfc_crtn.h struct lpfc_queue *drq, int count, int idx); idx 570 drivers/scsi/lpfc/lpfc_crtn.h void lpfc_nvmet_unsol_fcp_event(struct lpfc_hba *phba, uint32_t idx, idx 6241 drivers/scsi/lpfc/lpfc_debugfs.c int idx; idx 6250 drivers/scsi/lpfc/lpfc_debugfs.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) idx 6251 drivers/scsi/lpfc/lpfc_debugfs.c lpfc_debug_dump_wq(phba, DUMP_IO, idx); idx 6262 drivers/scsi/lpfc/lpfc_debugfs.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) idx 6263 drivers/scsi/lpfc/lpfc_debugfs.c lpfc_debug_dump_cq(phba, DUMP_IO, idx); idx 6268 drivers/scsi/lpfc/lpfc_debugfs.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) idx 6269 drivers/scsi/lpfc/lpfc_debugfs.c lpfc_debug_dump_hba_eq(phba, idx); idx 333 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_qe(struct lpfc_queue *q, uint32_t idx) idx 342 drivers/scsi/lpfc/lpfc_debugfs.h if (idx >= q->entry_count) idx 347 drivers/scsi/lpfc/lpfc_debugfs.h pword = lpfc_sli4_qe(q, idx); idx 350 drivers/scsi/lpfc/lpfc_debugfs.h len += scnprintf(line_buf+len, LPFC_LBUF_SZ-len, "QE[%04d]: ", idx); idx 383 drivers/scsi/lpfc/lpfc_debugfs.h int idx, entry_count; idx 398 drivers/scsi/lpfc/lpfc_debugfs.h for (idx = 0; idx < entry_count; idx++) idx 399 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_qe(q, idx); idx 1049 drivers/scsi/lpfc/lpfc_init.c int cnt, idx; idx 1081 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 1082 drivers/scsi/lpfc/lpfc_init.c qp = &phba->sli4_hba.hdwq[idx]; idx 3618 drivers/scsi/lpfc/lpfc_init.c int idx; idx 3620 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 3621 drivers/scsi/lpfc/lpfc_init.c qp = &phba->sli4_hba.hdwq[idx]; idx 3905 drivers/scsi/lpfc/lpfc_init.c int idx, cnt, xri, inserted; idx 3908 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 3909 drivers/scsi/lpfc/lpfc_init.c qp = &phba->sli4_hba.hdwq[idx]; idx 3931 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < cnt; idx++) { idx 3935 drivers/scsi/lpfc/lpfc_init.c if (idx == 0) { idx 3965 drivers/scsi/lpfc/lpfc_init.c int idx, cnt; idx 3970 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 3976 drivers/scsi/lpfc/lpfc_init.c qp = &phba->sli4_hba.hdwq[idx]; idx 3977 drivers/scsi/lpfc/lpfc_init.c lpfc_cmd->hdwq_no = idx; idx 8652 drivers/scsi/lpfc/lpfc_init.c lpfc_alloc_io_wq_cq(struct lpfc_hba *phba, int idx) idx 8658 drivers/scsi/lpfc/lpfc_init.c cpu = lpfc_find_cpu_handle(phba, idx, LPFC_FIND_BY_HDWQ); idx 8672 drivers/scsi/lpfc/lpfc_init.c "0499 Failed allocate fast-path IO CQ (%d)\n", idx); idx 8676 drivers/scsi/lpfc/lpfc_init.c qdesc->hdwq = idx; idx 8678 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hdwq[idx].io_cq = qdesc; idx 8696 drivers/scsi/lpfc/lpfc_init.c idx); idx 8699 drivers/scsi/lpfc/lpfc_init.c qdesc->hdwq = idx; idx 8701 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hdwq[idx].io_wq = qdesc; idx 8724 drivers/scsi/lpfc/lpfc_init.c int idx, cpu, eqcpu; idx 8756 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 8757 drivers/scsi/lpfc/lpfc_init.c qp = &phba->sli4_hba.hdwq[idx]; idx 8869 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 8870 drivers/scsi/lpfc/lpfc_init.c if (lpfc_alloc_io_wq_cq(phba, idx)) idx 8875 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_nvmet_mrq; idx++) { idx 8876 drivers/scsi/lpfc/lpfc_init.c cpu = lpfc_find_cpu_handle(phba, idx, idx 8886 drivers/scsi/lpfc/lpfc_init.c "CQ Set (%d)\n", idx); idx 8890 drivers/scsi/lpfc/lpfc_init.c qdesc->hdwq = idx; idx 8892 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.nvmet_cqset[idx] = qdesc; idx 9017 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_nvmet_mrq; idx++) { idx 9018 drivers/scsi/lpfc/lpfc_init.c cpu = lpfc_find_cpu_handle(phba, idx, idx 9032 drivers/scsi/lpfc/lpfc_init.c qdesc->hdwq = idx; idx 9033 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.nvmet_mrq_hdr[idx] = qdesc; idx 9061 drivers/scsi/lpfc/lpfc_init.c qdesc->hdwq = idx; idx 9062 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.nvmet_mrq_data[idx] = qdesc; idx 9068 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 9069 drivers/scsi/lpfc/lpfc_init.c memset(&phba->sli4_hba.hdwq[idx].nvme_cstat, 0, idx 9070 drivers/scsi/lpfc/lpfc_init.c sizeof(phba->sli4_hba.hdwq[idx].nvme_cstat)); idx 9076 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 9077 drivers/scsi/lpfc/lpfc_init.c memset(&phba->sli4_hba.hdwq[idx].scsi_cstat, 0, idx 9078 drivers/scsi/lpfc/lpfc_init.c sizeof(phba->sli4_hba.hdwq[idx].scsi_cstat)); idx 9101 drivers/scsi/lpfc/lpfc_init.c int idx; idx 9106 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < max; idx++) idx 9107 drivers/scsi/lpfc/lpfc_init.c __lpfc_sli4_release_queue(&(*qs)[idx]); idx 9118 drivers/scsi/lpfc/lpfc_init.c uint32_t idx; idx 9123 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 9125 drivers/scsi/lpfc/lpfc_init.c lpfc_sli4_queue_free(hdwq[idx].io_cq); idx 9126 drivers/scsi/lpfc/lpfc_init.c lpfc_sli4_queue_free(hdwq[idx].io_wq); idx 9127 drivers/scsi/lpfc/lpfc_init.c hdwq[idx].io_cq = NULL; idx 9128 drivers/scsi/lpfc/lpfc_init.c hdwq[idx].io_wq = NULL; idx 9130 drivers/scsi/lpfc/lpfc_init.c lpfc_free_sgl_per_hdwq(phba, &hdwq[idx]); idx 9131 drivers/scsi/lpfc/lpfc_init.c lpfc_free_cmd_rsp_buf_per_hdwq(phba, &hdwq[idx]); idx 9134 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_irq_chann; idx++) { idx 9136 drivers/scsi/lpfc/lpfc_init.c eq = phba->sli4_hba.hba_eq_hdl[idx].eq; idx 9138 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hba_eq_hdl[idx].eq = NULL; idx 10618 drivers/scsi/lpfc/lpfc_init.c int idx; idx 10620 drivers/scsi/lpfc/lpfc_init.c for_each_present_cpu(idx) { idx 10621 drivers/scsi/lpfc/lpfc_init.c cpup = &phba->sli4_hba.cpu_map[idx]; idx 10625 drivers/scsi/lpfc/lpfc_init.c (cpu != idx)) idx 10645 drivers/scsi/lpfc/lpfc_init.c int i, cpu, idx, next_idx, new_cpu, start_cpu, first_cpu; idx 10719 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_irq_chann; idx++) { idx 10721 drivers/scsi/lpfc/lpfc_init.c maskp = pci_irq_get_affinity(phba->pcidev, idx); idx 10727 drivers/scsi/lpfc/lpfc_init.c idx, phba->cfg_irq_chann); idx 10728 drivers/scsi/lpfc/lpfc_init.c if (!idx) { idx 10731 drivers/scsi/lpfc/lpfc_init.c cpup->eq = idx; idx 10732 drivers/scsi/lpfc/lpfc_init.c cpup->irq = pci_irq_vector(phba->pcidev, idx); idx 10743 drivers/scsi/lpfc/lpfc_init.c cpup->eq = idx; idx 10744 drivers/scsi/lpfc/lpfc_init.c cpup->irq = pci_irq_vector(phba->pcidev, idx); idx 10873 drivers/scsi/lpfc/lpfc_init.c idx = 0; idx 10882 drivers/scsi/lpfc/lpfc_init.c cpup->hdwq = idx; idx 10883 drivers/scsi/lpfc/lpfc_init.c idx++; idx 10898 drivers/scsi/lpfc/lpfc_init.c next_idx = idx; idx 10900 drivers/scsi/lpfc/lpfc_init.c idx = 0; idx 10952 drivers/scsi/lpfc/lpfc_init.c cpup->hdwq = idx % phba->cfg_hdw_queue; idx 10953 drivers/scsi/lpfc/lpfc_init.c idx++; idx 10991 drivers/scsi/lpfc/lpfc_init.c u16 idx; idx 10993 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_irq_chann; idx++) { idx 10994 drivers/scsi/lpfc/lpfc_init.c maskp = pci_irq_get_affinity(phba->pcidev, idx); idx 11021 drivers/scsi/lpfc/lpfc_init.c if (!(map->irq == pci_irq_vector(phba->pcidev, idx))) idx 11176 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hba_eq_hdl[index].idx = index; idx 11256 drivers/scsi/lpfc/lpfc_init.c phba->sli4_hba.hba_eq_hdl[index].idx = index; idx 11283 drivers/scsi/lpfc/lpfc_init.c int retval, idx; idx 11320 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_irq_chann; idx++) { idx 11321 drivers/scsi/lpfc/lpfc_init.c eqhdl = &phba->sli4_hba.hba_eq_hdl[idx]; idx 11322 drivers/scsi/lpfc/lpfc_init.c eqhdl->idx = idx; idx 11415 drivers/scsi/lpfc/lpfc_init.c int idx, ccnt; idx 11432 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 11433 drivers/scsi/lpfc/lpfc_init.c qp = &phba->sli4_hba.hdwq[idx]; idx 11471 drivers/scsi/lpfc/lpfc_init.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 11472 drivers/scsi/lpfc/lpfc_init.c qp = &phba->sli4_hba.hdwq[idx]; idx 61 drivers/scsi/lpfc/lpfc_nvme.c int idx, int expedite); idx 972 drivers/scsi/lpfc/lpfc_nvme.c uint32_t code, status, idx; idx 1002 drivers/scsi/lpfc/lpfc_nvme.c idx = lpfc_ncmd->cur_iocbq.hba_wqidx; idx 1003 drivers/scsi/lpfc/lpfc_nvme.c phba->sli4_hba.hdwq[idx].nvme_cstat.io_cmpls++; idx 1149 drivers/scsi/lpfc/lpfc_nvme.c idx = lpfc_ncmd->cur_iocbq.hba_wqidx; idx 1158 drivers/scsi/lpfc/lpfc_nvme.c phba->sli4_hba.hdwq[idx].cpucheck_cmpl_io[cpu]++; idx 1501 drivers/scsi/lpfc/lpfc_nvme.c int idx, cpu; idx 1627 drivers/scsi/lpfc/lpfc_nvme.c idx = lpfc_queue_info->index; idx 1630 drivers/scsi/lpfc/lpfc_nvme.c idx = phba->sli4_hba.cpu_map[cpu].hdwq; idx 1633 drivers/scsi/lpfc/lpfc_nvme.c lpfc_ncmd = lpfc_get_nvme_buf(phba, ndlp, idx, expedite); idx 1671 drivers/scsi/lpfc/lpfc_nvme.c lpfc_ncmd->cur_iocbq.hba_wqidx = idx; idx 1672 drivers/scsi/lpfc/lpfc_nvme.c cstat = &phba->sli4_hba.hdwq[idx].nvme_cstat; idx 1712 drivers/scsi/lpfc/lpfc_nvme.c if (idx != cpu) idx 1719 drivers/scsi/lpfc/lpfc_nvme.c phba->sli4_hba.hdwq[idx].cpucheck_xmt_io[cpu]++; idx 2018 drivers/scsi/lpfc/lpfc_nvme.c int idx, int expedite) idx 2026 drivers/scsi/lpfc/lpfc_nvme.c lpfc_ncmd = lpfc_get_io_buf(phba, NULL, idx, expedite); idx 2059 drivers/scsi/lpfc/lpfc_nvme.c qp = &phba->sli4_hba.hdwq[idx]; idx 495 drivers/scsi/lpfc/lpfc_nvmet.c infop = lpfc_get_ctx_list(phba, cpu, ctxp->idx); idx 1297 drivers/scsi/lpfc/lpfc_nvmet.c int i, j, idx, cpu; idx 1360 drivers/scsi/lpfc/lpfc_nvmet.c idx = 0; idx 1416 drivers/scsi/lpfc/lpfc_nvmet.c infop = lpfc_get_ctx_list(phba, cpu, idx); idx 1423 drivers/scsi/lpfc/lpfc_nvmet.c idx++; idx 1424 drivers/scsi/lpfc/lpfc_nvmet.c if (idx >= phba->cfg_nvmet_mrq) { idx 1425 drivers/scsi/lpfc/lpfc_nvmet.c idx = 0; idx 2122 drivers/scsi/lpfc/lpfc_nvmet.c qno = nvmebuf->idx; idx 2230 drivers/scsi/lpfc/lpfc_nvmet.c uint32_t idx, idx 2264 drivers/scsi/lpfc/lpfc_nvmet.c current_infop = lpfc_get_ctx_list(phba, current_cpu, idx); idx 2282 drivers/scsi/lpfc/lpfc_nvmet.c if (idx != current_cpu) idx 2286 drivers/scsi/lpfc/lpfc_nvmet.c current_cpu, idx); idx 2287 drivers/scsi/lpfc/lpfc_nvmet.c phba->sli4_hba.hdwq[idx].cpucheck_rcv_io[current_cpu]++; idx 2308 drivers/scsi/lpfc/lpfc_nvmet.c qno = nvmebuf->idx; idx 2335 drivers/scsi/lpfc/lpfc_nvmet.c ctxp->idx = idx; idx 2427 drivers/scsi/lpfc/lpfc_nvmet.c uint32_t idx, idx 2436 drivers/scsi/lpfc/lpfc_nvmet.c lpfc_nvmet_unsol_fcp_buffer(phba, idx, nvmebuf, isr_timestamp, cqflag); idx 124 drivers/scsi/lpfc/lpfc_nvmet.h uint16_t idx; idx 472 drivers/scsi/lpfc/lpfc_scsi.c int idx; idx 478 drivers/scsi/lpfc/lpfc_scsi.c for (idx = 0; idx < phba->cfg_hdw_queue; idx++) { idx 479 drivers/scsi/lpfc/lpfc_scsi.c qp = &phba->sli4_hba.hdwq[idx]; idx 506 drivers/scsi/lpfc/lpfc_scsi.c struct sli4_wcqe_xri_aborted *axri, int idx) idx 522 drivers/scsi/lpfc/lpfc_scsi.c qp = &phba->sli4_hba.hdwq[idx]; idx 639 drivers/scsi/lpfc/lpfc_scsi.c uint32_t cpu, idx; idx 646 drivers/scsi/lpfc/lpfc_scsi.c idx = blk_mq_unique_tag_to_hwq(tag); idx 648 drivers/scsi/lpfc/lpfc_scsi.c idx = phba->sli4_hba.cpu_map[cpu].hdwq; idx 651 drivers/scsi/lpfc/lpfc_scsi.c lpfc_cmd = lpfc_get_io_buf(phba, ndlp, idx, idx 654 drivers/scsi/lpfc/lpfc_scsi.c qp = &phba->sli4_hba.hdwq[idx]; idx 3806 drivers/scsi/lpfc/lpfc_scsi.c int idx; idx 3824 drivers/scsi/lpfc/lpfc_scsi.c idx = lpfc_cmd->cur_iocbq.hba_wqidx; idx 3826 drivers/scsi/lpfc/lpfc_scsi.c phba->sli4_hba.hdwq[idx].scsi_cstat.io_cmpls++; idx 3832 drivers/scsi/lpfc/lpfc_scsi.c phba->sli4_hba.hdwq[idx].cpucheck_cmpl_io[cpu]++; idx 4088 drivers/scsi/lpfc/lpfc_scsi.c int idx; idx 4114 drivers/scsi/lpfc/lpfc_scsi.c idx = lpfc_cmd->hdwq_no; idx 4116 drivers/scsi/lpfc/lpfc_scsi.c hdwq = &phba->sli4_hba.hdwq[idx]; idx 4505 drivers/scsi/lpfc/lpfc_scsi.c int err, idx; idx 4673 drivers/scsi/lpfc/lpfc_scsi.c idx = lpfc_cmd->hdwq_no; idx 4678 drivers/scsi/lpfc/lpfc_scsi.c phba->sli4_hba.hdwq[idx].scsi_cstat.output_requests--; idx 4681 drivers/scsi/lpfc/lpfc_scsi.c phba->sli4_hba.hdwq[idx].scsi_cstat.input_requests--; idx 4684 drivers/scsi/lpfc/lpfc_scsi.c phba->sli4_hba.hdwq[idx].scsi_cstat.control_requests--; idx 143 drivers/scsi/lpfc/lpfc_sli.c uint32_t idx; idx 154 drivers/scsi/lpfc/lpfc_sli.c idx = ((q->host_index + 1) % q->entry_count); idx 155 drivers/scsi/lpfc/lpfc_sli.c if (idx == q->hba_index) { idx 187 drivers/scsi/lpfc/lpfc_sli.c q->host_index = idx; idx 7095 drivers/scsi/lpfc/lpfc_sli.c struct lpfc_queue *drq, int count, int idx) idx 7116 drivers/scsi/lpfc/lpfc_sli.c rqb_buffer->idx = idx; idx 7818 drivers/scsi/lpfc/lpfc_sli.c uint32_t idx; idx 7830 drivers/scsi/lpfc/lpfc_sli.c idx = mcq->hba_index; idx 7833 drivers/scsi/lpfc/lpfc_sli.c (struct lpfc_cqe *)lpfc_sli4_qe(mcq, idx)) == qe_valid) { idx 7834 drivers/scsi/lpfc/lpfc_sli.c mcqe = (struct lpfc_mcqe *)(lpfc_sli4_qe(mcq, idx)); idx 7840 drivers/scsi/lpfc/lpfc_sli.c idx = (idx + 1) % mcq->entry_count; idx 7841 drivers/scsi/lpfc/lpfc_sli.c if (mcq->hba_index == idx) idx 7845 drivers/scsi/lpfc/lpfc_sli.c if (phba->sli4_hba.pc_sli4_params.cqav && !idx) idx 13882 drivers/scsi/lpfc/lpfc_sli.c uint32_t fctl, idx; idx 13888 drivers/scsi/lpfc/lpfc_sli.c idx = cq->queue_id - phba->sli4_hba.nvmet_cqset[0]->queue_id; idx 13889 drivers/scsi/lpfc/lpfc_sli.c hrq = phba->sli4_hba.nvmet_mrq_hdr[idx]; idx 13890 drivers/scsi/lpfc/lpfc_sli.c drq = phba->sli4_hba.nvmet_mrq_data[idx]; idx 13938 drivers/scsi/lpfc/lpfc_sli.c phba, idx, dma_buf, cq->isr_timestamp, idx 14226 drivers/scsi/lpfc/lpfc_sli.c hba_eqidx = hba_eq_hdl->idx; idx 15036 drivers/scsi/lpfc/lpfc_sli.c int cnt, idx, numcq, page_idx = 0; idx 15068 drivers/scsi/lpfc/lpfc_sli.c for (idx = 0; idx < numcq; idx++) { idx 15069 drivers/scsi/lpfc/lpfc_sli.c cq = cqp[idx]; idx 15070 drivers/scsi/lpfc/lpfc_sli.c eq = hdwq[idx].hba_eq; idx 15078 drivers/scsi/lpfc/lpfc_sli.c switch (idx) { idx 15208 drivers/scsi/lpfc/lpfc_sli.c cq->chann = idx; idx 15242 drivers/scsi/lpfc/lpfc_sli.c for (idx = 0; idx < numcq; idx++) { idx 15243 drivers/scsi/lpfc/lpfc_sli.c cq = cqp[idx]; idx 15244 drivers/scsi/lpfc/lpfc_sli.c cq->queue_id = rc + idx; idx 16055 drivers/scsi/lpfc/lpfc_sli.c int cnt, idx, numrq, page_idx = 0; idx 16095 drivers/scsi/lpfc/lpfc_sli.c for (idx = 0; idx < numrq; idx++) { idx 16096 drivers/scsi/lpfc/lpfc_sli.c hrq = hrqp[idx]; idx 16097 drivers/scsi/lpfc/lpfc_sli.c drq = drqp[idx]; idx 16098 drivers/scsi/lpfc/lpfc_sli.c cq = cqp[idx]; idx 16111 drivers/scsi/lpfc/lpfc_sli.c if (idx == 0) { idx 16203 drivers/scsi/lpfc/lpfc_sli.c for (idx = 0; idx < numrq; idx++) { idx 16204 drivers/scsi/lpfc/lpfc_sli.c hrq = hrqp[idx]; idx 16205 drivers/scsi/lpfc/lpfc_sli.c hrq->queue_id = rc + (2 * idx); idx 16206 drivers/scsi/lpfc/lpfc_sli.c drq = drqp[idx]; idx 16207 drivers/scsi/lpfc/lpfc_sli.c drq->queue_id = rc + (2 * idx) + 1; idx 20478 drivers/scsi/lpfc/lpfc_sli.c lpfc_io_buf(struct lpfc_hba *phba, struct lpfc_nodelist *ndlp, int idx) idx 20483 drivers/scsi/lpfc/lpfc_sli.c qp = &phba->sli4_hba.hdwq[idx]; idx 20496 drivers/scsi/lpfc/lpfc_sli.c lpfc_cmd->hdwq_no = idx; idx 468 drivers/scsi/lpfc/lpfc_sli4.h uint32_t idx; idx 1076 drivers/scsi/lpfc/lpfc_sli4.h struct sli4_wcqe_xri_aborted *axri, int idx); idx 1122 drivers/scsi/lpfc/lpfc_sli4.h static inline void *lpfc_sli4_qe(struct lpfc_queue *q, uint16_t idx) idx 1124 drivers/scsi/lpfc/lpfc_sli4.h return q->q_pgs[idx / q->entry_cnt_per_pg] + idx 1125 drivers/scsi/lpfc/lpfc_sli4.h (q->entry_size * (idx % q->entry_cnt_per_pg)); idx 1115 drivers/scsi/megaraid.c mbox->m_out.cmdid = scb->idx; /* Set cmdid */ idx 1469 drivers/scsi/megaraid.c scb->idx); idx 1488 drivers/scsi/megaraid.c scb->idx); idx 1723 drivers/scsi/megaraid.c int idx; idx 1748 drivers/scsi/megaraid.c scsi_for_each_sg(cmd, sg, sgcnt, idx) { idx 1750 drivers/scsi/megaraid.c scb->sgl64[idx].address = sg_dma_address(sg); idx 1751 drivers/scsi/megaraid.c *len += scb->sgl64[idx].length = sg_dma_len(sg); idx 1753 drivers/scsi/megaraid.c scb->sgl[idx].address = sg_dma_address(sg); idx 1754 drivers/scsi/megaraid.c *len += scb->sgl[idx].length = sg_dma_len(sg); idx 1964 drivers/scsi/megaraid.c scb->idx); idx 1977 drivers/scsi/megaraid.c scb->idx); idx 2892 drivers/scsi/megaraid.c scb->idx = i; idx 4095 drivers/scsi/megaraid.c scb->idx = CMDID_INT_CMDS; idx 187 drivers/scsi/megaraid.h int idx; idx 150 drivers/scsi/qedf/qedf.h int idx; idx 431 drivers/scsi/qedf/qedf.h u16 idx; idx 239 drivers/scsi/qedf/qedf_debugfs.c int i, idx = 0; idx 253 drivers/scsi/qedf/qedf_debugfs.c idx = qedf->io_trace_idx; idx 255 drivers/scsi/qedf/qedf_debugfs.c io_log = &qedf->io_trace_buf[idx]; idx 273 drivers/scsi/qedf/qedf_debugfs.c idx++; idx 274 drivers/scsi/qedf/qedf_debugfs.c if (idx == QEDF_IO_TRACE_SIZE) idx 275 drivers/scsi/qedf/qedf_debugfs.c idx = 0; idx 344 drivers/scsi/qedf/qedf_io.c io_req = &cmd_mgr->cmds[cmd_mgr->idx]; idx 345 drivers/scsi/qedf/qedf_io.c cmd_mgr->idx++; idx 346 drivers/scsi/qedf/qedf_io.c if (cmd_mgr->idx == FCOE_PARAMS_NUM_TASKS) idx 347 drivers/scsi/qedf/qedf_io.c cmd_mgr->idx = 0; idx 179 drivers/scsi/qedi/qedi_debugfs.c int id, idx = 0; idx 186 drivers/scsi/qedi/qedi_debugfs.c idx = qedi->io_trace_idx; idx 188 drivers/scsi/qedi/qedi_debugfs.c io_log = &qedi->io_trace_buf[idx]; idx 205 drivers/scsi/qedi/qedi_debugfs.c idx++; idx 206 drivers/scsi/qedi/qedi_debugfs.c if (idx == QEDI_IO_TRACE_SIZE) idx 207 drivers/scsi/qedi/qedi_debugfs.c idx = 0; idx 316 drivers/scsi/qedi/qedi_fw.c u16 idx = 0; idx 324 drivers/scsi/qedi/qedi_fw.c idx = cqe->rqe_opaque; idx 325 drivers/scsi/qedi/qedi_fw.c if (idx > (QEDI_BDQ_NUM - 1)) { idx 328 drivers/scsi/qedi/qedi_fw.c idx); idx 333 drivers/scsi/qedi/qedi_fw.c "rqe_opaque [0x%p], idx [%d]\n", cqe->rqe_opaque, idx); idx 341 drivers/scsi/qedi/qedi_fw.c memcpy(ptr, (void *)qedi->bdq[idx].buf_addr, len); idx 356 drivers/scsi/qedi/qedi_fw.c u16 idx = 0; idx 360 drivers/scsi/qedi/qedi_fw.c idx = cqe->rqe_opaque; idx 361 drivers/scsi/qedi/qedi_fw.c if (idx > (QEDI_BDQ_NUM - 1)) { idx 364 drivers/scsi/qedi/qedi_fw.c idx); idx 370 drivers/scsi/qedi/qedi_fw.c pbl->address.hi = cpu_to_le32(QEDI_U64_HI(qedi->bdq[idx].buf_dma)); idx 371 drivers/scsi/qedi/qedi_fw.c pbl->address.lo = cpu_to_le32(QEDI_U64_LO(qedi->bdq[idx].buf_dma)); idx 374 drivers/scsi/qedi/qedi_fw.c pbl, pbl->address.hi, pbl->address.lo, idx); idx 378 drivers/scsi/qedi/qedi_fw.c pbl->opaque.iscsi_opaque.opaque = cpu_to_le32(idx); idx 43 drivers/scsi/qedi/qedi_gbl.h void qedi_clear_task_idx(struct qedi_ctx *qedi, int idx); idx 1316 drivers/scsi/qedi/qedi_main.c u16 idx; idx 1320 drivers/scsi/qedi/qedi_main.c idx = i * qedi->dev_info.common.num_hwfns + idx 1324 drivers/scsi/qedi/qedi_main.c "Freeing IRQ #%d vector_idx=%d.\n", i, idx); idx 1326 drivers/scsi/qedi/qedi_main.c synchronize_irq(qedi->int_info.msix[idx].vector); idx 1327 drivers/scsi/qedi/qedi_main.c irq_set_affinity_hint(qedi->int_info.msix[idx].vector, idx 1329 drivers/scsi/qedi/qedi_main.c free_irq(qedi->int_info.msix[idx].vector, idx 1344 drivers/scsi/qedi/qedi_main.c u16 idx; idx 1348 drivers/scsi/qedi/qedi_main.c idx = i * qedi->dev_info.common.num_hwfns + idx 1356 drivers/scsi/qedi/qedi_main.c rc = request_irq(qedi->int_info.msix[idx].vector, idx 1365 drivers/scsi/qedi/qedi_main.c rc = irq_set_affinity_hint(qedi->int_info.msix[idx].vector, idx 1776 drivers/scsi/qedi/qedi_main.c void qedi_clear_task_idx(struct qedi_ctx *qedi, int idx) idx 1778 drivers/scsi/qedi/qedi_main.c if (!test_and_clear_bit(idx, qedi->task_idx_map)) idx 1780 drivers/scsi/qedi/qedi_main.c "FW task context, already cleared, tid=0x%x\n", idx); idx 2112 drivers/scsi/qedi/qedi_main.c char *buf, enum qedi_nvm_tgts idx) idx 2126 drivers/scsi/qedi/qedi_main.c GET_FIELD2(block->id, NVM_ISCSI_CFG_BLK_MAPPED_PF_ID), idx); idx 2128 drivers/scsi/qedi/qedi_main.c ctrl_flags = block->target[idx].ctrl_flags & idx 2147 drivers/scsi/qedi/qedi_main.c mchap_name = mchap_en ? block->target[idx].chap_name.byte : NULL; idx 2148 drivers/scsi/qedi/qedi_main.c mchap_secret = mchap_en ? block->target[idx].chap_password.byte : NULL; idx 2153 drivers/scsi/qedi/qedi_main.c block->target[idx].target_name.byte); idx 2158 drivers/scsi/qedi/qedi_main.c block->target[idx].ipv6_addr.byte); idx 2161 drivers/scsi/qedi/qedi_main.c block->target[idx].ipv4_addr.byte); idx 2165 drivers/scsi/qedi/qedi_main.c GET_FIELD2(block->target[idx].generic_cont0, idx 2170 drivers/scsi/qedi/qedi_main.c block->target[idx].lun.value[1], idx 2171 drivers/scsi/qedi/qedi_main.c block->target[idx].lun.value[0]); idx 373 drivers/scsi/qla2xxx/qla_dbg.c uint32_t cnt, stat, timer, words, idx; idx 441 drivers/scsi/qla2xxx/qla_dbg.c for (idx = 0; idx < words; idx++) idx 442 drivers/scsi/qla2xxx/qla_dbg.c ram[cnt + idx] = swab16(dump[idx]); idx 4551 drivers/scsi/qla2xxx/qla_def.h uint8_t idx; idx 1585 drivers/scsi/qla2xxx/qla_nx.c int idx = cpu_to_le32(*((int *)&unirom[ha->file_prd_off] + idx_offset)); idx 1594 drivers/scsi/qla2xxx/qla_nx.c (cpu_to_le32(tab_desc->entry_size) * idx); idx 87 drivers/scsi/qla4xxx/ql4_glbl.h uint16_t idx, int bidi); idx 170 drivers/scsi/qla4xxx/ql4_glbl.h char *password, uint16_t idx); idx 916 drivers/scsi/qla4xxx/ql4_init.c uint32_t idx = 0, next_idx = 0; idx 922 drivers/scsi/qla4xxx/ql4_init.c for (idx = 0; idx < max_ddbs; idx = next_idx) { idx 923 drivers/scsi/qla4xxx/ql4_init.c ret = qla4xxx_get_fwddb_entry(ha, idx, NULL, 0, NULL, idx 933 drivers/scsi/qla4xxx/ql4_init.c "Freeing DDB index = 0x%x\n", idx)); idx 934 drivers/scsi/qla4xxx/ql4_init.c ret = qla4xxx_clear_ddb_entry(ha, idx); idx 938 drivers/scsi/qla4xxx/ql4_init.c "0x%x\n", idx); idx 1576 drivers/scsi/qla4xxx/ql4_mbx.c uint16_t idx) idx 1591 drivers/scsi/qla4xxx/ql4_mbx.c offset = FLASH_CHAP_OFFSET | (idx * chap_size); idx 1599 drivers/scsi/qla4xxx/ql4_mbx.c offset += (idx * chap_size); idx 1638 drivers/scsi/qla4xxx/ql4_mbx.c uint16_t idx, int bidi) idx 1674 drivers/scsi/qla4xxx/ql4_mbx.c offset += (idx * sizeof(struct ql4_chap_table)); idx 1681 drivers/scsi/qla4xxx/ql4_mbx.c memcpy((struct ql4_chap_table *)ha->chap_list + idx, idx 1998 drivers/scsi/qla4xxx/ql4_mbx.c uint16_t idx, *ptid; idx 2079 drivers/scsi/qla4xxx/ql4_mbx.c LOCAL_CHAP, &idx); idx 2083 drivers/scsi/qla4xxx/ql4_mbx.c fw_ddb_entry->chap_tbl_idx = cpu_to_le16(idx); idx 2094 drivers/scsi/qla4xxx/ql4_mbx.c BIDI_CHAP, &idx); idx 2825 drivers/scsi/qla4xxx/ql4_os.c uint16_t idx; idx 2832 drivers/scsi/qla4xxx/ql4_os.c &idx); idx 2836 drivers/scsi/qla4xxx/ql4_os.c len = sprintf(buf, "%hu\n", idx); idx 2841 drivers/scsi/qla4xxx/ql4_os.c idx = ddb_entry->chap_tbl_idx; idx 2849 drivers/scsi/qla4xxx/ql4_os.c LOCAL_CHAP, &idx); idx 2854 drivers/scsi/qla4xxx/ql4_os.c len = sprintf(buf, "%hu\n", idx); idx 2865 drivers/scsi/qla4xxx/ql4_os.c idx = ddb_entry->chap_tbl_idx; idx 2868 drivers/scsi/qla4xxx/ql4_os.c idx); idx 2983 drivers/scsi/qla4xxx/ql4_os.c int idx = 0, max_ddbs, rval; idx 3000 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < max_ddbs; idx++) { idx 3001 drivers/scsi/qla4xxx/ql4_os.c ddb_entry = qla4xxx_lookup_ddb_by_fw_index(ha, idx); idx 3043 drivers/scsi/qla4xxx/ql4_os.c if (idx == max_ddbs) idx 6054 drivers/scsi/qla4xxx/ql4_os.c uint16_t idx; idx 6094 drivers/scsi/qla4xxx/ql4_os.c idx = __le16_to_cpu(fw_ddb_entry->chap_tbl_idx); idx 6103 drivers/scsi/qla4xxx/ql4_os.c idx); idx 6342 drivers/scsi/qla4xxx/ql4_os.c int idx; idx 6363 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < MAX_DDB_ENTRIES; idx++) { idx 6364 drivers/scsi/qla4xxx/ql4_os.c ddb_entry = qla4xxx_lookup_ddb_by_fw_index(ha, idx); idx 6372 drivers/scsi/qla4xxx/ql4_os.c *index = idx; idx 6587 drivers/scsi/qla4xxx/ql4_os.c static int qla4xxx_verify_boot_idx(struct scsi_qla_host *ha, uint16_t idx) idx 6591 drivers/scsi/qla4xxx/ql4_os.c if (idx == ha->pri_ddb_idx || idx == ha->sec_ddb_idx) idx 6598 drivers/scsi/qla4xxx/ql4_os.c uint16_t idx) idx 6621 drivers/scsi/qla4xxx/ql4_os.c (idx == ha->pri_ddb_idx || idx == ha->sec_ddb_idx)) idx 6627 drivers/scsi/qla4xxx/ql4_os.c uint32_t idx = 0; idx 6636 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < IP_ADDR_COUNT; idx++) { idx 6637 drivers/scsi/qla4xxx/ql4_os.c if (ip_idx[idx] == -1) idx 6640 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_get_ip_state(ha, 0, ip_idx[idx], sts); idx 6643 drivers/scsi/qla4xxx/ql4_os.c ip_idx[idx] = -1; idx 6651 drivers/scsi/qla4xxx/ql4_os.c ip_idx[idx], ip_state)); idx 6657 drivers/scsi/qla4xxx/ql4_os.c ip_idx[idx] = -1; idx 6700 drivers/scsi/qla4xxx/ql4_os.c uint32_t idx = 0; idx 6725 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < max_ddbs; idx++) { idx 6727 drivers/scsi/qla4xxx/ql4_os.c flash_ddb_entry_dma, idx); idx 6733 drivers/scsi/qla4xxx/ql4_os.c *flash_index = idx; idx 6739 drivers/scsi/qla4xxx/ql4_os.c if (idx == max_ddbs) idx 6760 drivers/scsi/qla4xxx/ql4_os.c uint32_t idx = 0, next_idx = 0; idx 6776 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < max_ddbs; idx = next_idx) { idx 6777 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_get_fwddb_entry(ha, idx, fw_ddb_entry, fw_ddb_dma, idx 6795 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_find_flash_st_idx(ha, fw_ddb_entry, idx, idx 6799 drivers/scsi/qla4xxx/ql4_os.c "No flash entry for ST at idx [%d]\n", idx); idx 6800 drivers/scsi/qla4xxx/ql4_os.c st_ddb_idx->flash_ddb_idx = idx; idx 6804 drivers/scsi/qla4xxx/ql4_os.c idx, flash_index); idx 6808 drivers/scsi/qla4xxx/ql4_os.c st_ddb_idx->fw_ddb_idx = idx; idx 6876 drivers/scsi/qla4xxx/ql4_os.c int is_reset, uint16_t idx) idx 6914 drivers/scsi/qla4xxx/ql4_os.c qla4xxx_setup_flash_ddb_entry(ha, ddb_entry, idx); idx 6987 drivers/scsi/qla4xxx/ql4_os.c uint32_t idx = 0, next_idx = 0; idx 7004 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < max_ddbs; idx = next_idx) { idx 7005 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_get_fwddb_entry(ha, idx, fw_ddb_entry, fw_ddb_dma, idx 7011 drivers/scsi/qla4xxx/ql4_os.c if (qla4xxx_verify_boot_idx(ha, idx) != QLA_SUCCESS) idx 7028 drivers/scsi/qla4xxx/ql4_os.c "Adding DDB to session = 0x%x\n", idx)); idx 7035 drivers/scsi/qla4xxx/ql4_os.c nt_ddb_idx->fw_ddb_idx = idx; idx 7072 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_sess_conn_setup(ha, fw_ddb_entry, is_reset, idx); idx 7095 drivers/scsi/qla4xxx/ql4_os.c uint32_t idx = 0, next_idx = 0; idx 7110 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < max_ddbs; idx = next_idx) { idx 7111 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_get_fwddb_entry(ha, idx, fw_ddb_entry, fw_ddb_dma, idx 7125 drivers/scsi/qla4xxx/ql4_os.c "Adding DDB to session = 0x%x\n", idx)); idx 7131 drivers/scsi/qla4xxx/ql4_os.c nt_ddb_idx->fw_ddb_idx = idx; idx 7146 drivers/scsi/qla4xxx/ql4_os.c idx); idx 7199 drivers/scsi/qla4xxx/ql4_os.c uint16_t *idx, int user) idx 7205 drivers/scsi/qla4xxx/ql4_os.c fnode_sess = iscsi_create_flashnode_sess(ha->host, *idx, idx 7210 drivers/scsi/qla4xxx/ql4_os.c __func__, *idx, ha->host_no); idx 7219 drivers/scsi/qla4xxx/ql4_os.c __func__, *idx, ha->host_no); idx 7228 drivers/scsi/qla4xxx/ql4_os.c if (*idx == ha->pri_ddb_idx || *idx == ha->sec_ddb_idx) idx 7270 drivers/scsi/qla4xxx/ql4_os.c uint16_t idx = 0; idx 7305 drivers/scsi/qla4xxx/ql4_os.c for (idx = 2; idx < max_ddbs; idx++) { idx 7307 drivers/scsi/qla4xxx/ql4_os.c fw_ddb_entry_dma, idx)) idx 7311 drivers/scsi/qla4xxx/ql4_os.c if (idx == max_ddbs) idx 7321 drivers/scsi/qla4xxx/ql4_os.c rval = qla4xxx_sysfs_ddb_tgt_create(ha, fw_ddb_entry, &idx, 1); idx 7328 drivers/scsi/qla4xxx/ql4_os.c return idx; idx 7399 drivers/scsi/qla4xxx/ql4_os.c uint16_t idx) idx 7420 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_set_ddb_entry(ha, idx, ddb_entry_dma, &mbx_sts); idx 7424 drivers/scsi/qla4xxx/ql4_os.c __func__, idx)); idx 7428 drivers/scsi/qla4xxx/ql4_os.c qla4xxx_conn_open(ha, idx); idx 7440 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_get_fwddb_entry(ha, idx, NULL, 0, NULL, idx 7510 drivers/scsi/qla4xxx/ql4_os.c uint16_t idx) idx 7517 drivers/scsi/qla4xxx/ql4_os.c idx); idx 7728 drivers/scsi/qla4xxx/ql4_os.c int idx, index; idx 7784 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < MAX_DDB_ENTRIES; idx++) { idx 7785 drivers/scsi/qla4xxx/ql4_os.c ddb_entry = qla4xxx_lookup_ddb_by_fw_index(ha, idx); idx 7812 drivers/scsi/qla4xxx/ql4_os.c if (idx == MAX_DDB_ENTRIES) idx 8416 drivers/scsi/qla4xxx/ql4_os.c uint16_t idx = 0; idx 8432 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < max_ddbs; idx++) { idx 8434 drivers/scsi/qla4xxx/ql4_os.c idx)) idx 8437 drivers/scsi/qla4xxx/ql4_os.c ret = qla4xxx_sysfs_ddb_tgt_create(ha, fw_ddb_entry, &idx, 0); idx 8535 drivers/scsi/qla4xxx/ql4_os.c int max_ddbs, idx, ret; idx 8550 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < max_ddbs; idx++) { idx 8551 drivers/scsi/qla4xxx/ql4_os.c ddb_entry = qla4xxx_lookup_ddb_by_fw_index(ha, idx); idx 8956 drivers/scsi/qla4xxx/ql4_os.c int idx; idx 8958 drivers/scsi/qla4xxx/ql4_os.c for (idx = 0; idx < MAX_DDB_ENTRIES; idx++) { idx 8960 drivers/scsi/qla4xxx/ql4_os.c ddb_entry = qla4xxx_lookup_ddb_by_fw_index(ha, idx); idx 2365 drivers/scsi/scsi_lib.c int idx = 0; idx 2370 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_MEDIA_CHANGE=1"; idx 2374 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_UA=INQUIRY_DATA_HAS_CHANGED"; idx 2377 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_UA=CAPACITY_DATA_HAS_CHANGED"; idx 2380 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_UA=THIN_PROVISIONING_SOFT_THRESHOLD_REACHED"; idx 2383 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_UA=MODE_PARAMETERS_CHANGED"; idx 2386 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_UA=REPORTED_LUNS_DATA_HAS_CHANGED"; idx 2389 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_UA=ASYMMETRIC_ACCESS_STATE_CHANGED"; idx 2392 drivers/scsi/scsi_lib.c envp[idx++] = "SDEV_UA=POWER_ON_RESET_OCCURRED"; idx 2399 drivers/scsi/scsi_lib.c envp[idx++] = NULL; idx 1007 drivers/scsi/scsi_scan.c unsigned term = 0, idx; idx 1009 drivers/scsi/scsi_scan.c for (idx = 0; idx + first < end && idx + first < inq[4] + 5; idx++) { idx 1010 drivers/scsi/scsi_scan.c if (inq[idx+first] > ' ') { idx 1011 drivers/scsi/scsi_scan.c buf[idx] = inq[idx+first]; idx 1012 drivers/scsi/scsi_scan.c term = idx+1; idx 1014 drivers/scsi/scsi_scan.c buf[idx] = ' '; idx 1372 drivers/scsi/scsi_transport_iscsi.c iscsi_get_flashnode_by_index(struct Scsi_Host *shost, uint32_t idx) idx 1377 drivers/scsi/scsi_transport_iscsi.c dev = device_find_child(&shost->shost_gendev, &idx, idx 3164 drivers/scsi/scsi_transport_iscsi.c uint32_t idx; idx 3180 drivers/scsi/scsi_transport_iscsi.c idx = ev->u.set_flashnode.flashnode_idx; idx 3181 drivers/scsi/scsi_transport_iscsi.c fnode_sess = iscsi_get_flashnode_by_index(shost, idx); idx 3184 drivers/scsi/scsi_transport_iscsi.c __func__, idx, ev->u.set_flashnode.host_no); idx 3249 drivers/scsi/scsi_transport_iscsi.c uint32_t idx; idx 3265 drivers/scsi/scsi_transport_iscsi.c idx = ev->u.del_flashnode.flashnode_idx; idx 3266 drivers/scsi/scsi_transport_iscsi.c fnode_sess = iscsi_get_flashnode_by_index(shost, idx); idx 3269 drivers/scsi/scsi_transport_iscsi.c __func__, idx, ev->u.del_flashnode.host_no); idx 3291 drivers/scsi/scsi_transport_iscsi.c uint32_t idx; idx 3307 drivers/scsi/scsi_transport_iscsi.c idx = ev->u.login_flashnode.flashnode_idx; idx 3308 drivers/scsi/scsi_transport_iscsi.c fnode_sess = iscsi_get_flashnode_by_index(shost, idx); idx 3311 drivers/scsi/scsi_transport_iscsi.c __func__, idx, ev->u.login_flashnode.host_no); idx 3343 drivers/scsi/scsi_transport_iscsi.c uint32_t idx; idx 3359 drivers/scsi/scsi_transport_iscsi.c idx = ev->u.logout_flashnode.flashnode_idx; idx 3360 drivers/scsi/scsi_transport_iscsi.c fnode_sess = iscsi_get_flashnode_by_index(shost, idx); idx 3363 drivers/scsi/scsi_transport_iscsi.c __func__, idx, ev->u.logout_flashnode.host_no); idx 200 drivers/sh/clk/core.c int idx; idx 202 drivers/sh/clk/core.c cpufreq_for_each_valid_entry_idx(pos, freq_table, idx) idx 204 drivers/sh/clk/core.c return idx; idx 133 drivers/sh/clk/cpg.c unsigned int idx; idx 138 drivers/sh/clk/cpg.c idx = (sh_clk_read(clk) >> clk->enable_bit) & clk->div_mask; idx 140 drivers/sh/clk/cpg.c return clk->freq_table[idx].frequency; idx 147 drivers/sh/clk/cpg.c int idx; idx 149 drivers/sh/clk/cpg.c idx = clk_rate_table_find(clk, clk->freq_table, rate); idx 150 drivers/sh/clk/cpg.c if (idx < 0) idx 151 drivers/sh/clk/cpg.c return idx; idx 155 drivers/sh/clk/cpg.c value |= (idx << clk->enable_bit); idx 448 drivers/sh/clk/cpg.c int idx; idx 450 drivers/sh/clk/cpg.c idx = (clk->parent->rate / rate) & 0xffff; idx 451 drivers/sh/clk/cpg.c if (idx < 2) idx 454 drivers/sh/clk/cpg.c __raw_writel(idx << 16, clk->mapping->base); idx 134 drivers/slimbus/qcom-ctrl.c int idx; idx 142 drivers/slimbus/qcom-ctrl.c idx = ctrl->rx.tail; idx 146 drivers/slimbus/qcom-ctrl.c return ctrl->rx.base + (idx * ctrl->rx.sl_sz); idx 153 drivers/slimbus/qcom-ctrl.c int idx; idx 156 drivers/slimbus/qcom-ctrl.c idx = ctrl->tx.head; idx 160 drivers/slimbus/qcom-ctrl.c comp = ctrl->wr_comp[idx]; idx 161 drivers/slimbus/qcom-ctrl.c ctrl->wr_comp[idx] = NULL; idx 307 drivers/slimbus/qcom-ctrl.c int idx; idx 315 drivers/slimbus/qcom-ctrl.c idx = ctrl->tx.tail; idx 316 drivers/slimbus/qcom-ctrl.c ctrl->wr_comp[idx] = done; idx 321 drivers/slimbus/qcom-ctrl.c return ctrl->tx.base + (idx * ctrl->tx.sl_sz); idx 80 drivers/soc/bcm/brcmstb/pm/pm-mips.c #define AON_SAVE_SRAM(base, idx, val) \ idx 81 drivers/soc/bcm/brcmstb/pm/pm-mips.c __raw_writel(val, base + (idx << 2)) idx 38 drivers/soc/fsl/dpio/dpio-service.c unsigned int idx; /* position of the next-to-be-returned entry */ idx 569 drivers/soc/fsl/dpio/dpio-service.c ret->idx = 0; idx 610 drivers/soc/fsl/dpio/dpio-service.c struct dpaa2_dq *ret = &s->vaddr[s->idx]; idx 618 drivers/soc/fsl/dpio/dpio-service.c s->idx++; idx 622 drivers/soc/fsl/dpio/dpio-service.c s->idx = 0; idx 631 drivers/soc/fsl/dpio/dpio-service.c prefetch(&s->vaddr[s->idx]); idx 445 drivers/soc/fsl/dpio/qbman-portal.c u8 idx) idx 447 drivers/soc/fsl/dpio/qbman-portal.c if (idx < 16) idx 448 drivers/soc/fsl/dpio/qbman-portal.c qbman_write_register(p, QBMAN_CINH_SWP_EQCR_AM_RT + idx * 4, idx 452 drivers/soc/fsl/dpio/qbman-portal.c (idx - 16) * 4, idx 37 drivers/soc/fsl/qbman/dpaa_sys.c int qbman_init_private_mem(struct device *dev, int idx, dma_addr_t *addr, idx 46 drivers/soc/fsl/qbman/dpaa_sys.c mem_node = of_parse_phandle(dev->of_node, "memory-region", idx); idx 48 drivers/soc/fsl/qbman/dpaa_sys.c dev_err(dev, "No memory-region found for index %d\n", idx); idx 104 drivers/soc/fsl/qbman/dpaa_sys.h int qbman_init_private_mem(struct device *dev, int idx, dma_addr_t *addr, idx 128 drivers/soc/fsl/qbman/qman.c #define qm_cl(base, idx) ((void *)base + ((idx) << 6)) idx 703 drivers/soc/fsl/qbman/qman.c int idx = dqrr_ptr2idx(dq); idx 706 drivers/soc/fsl/qbman/qman.c DPAA_ASSERT((dqrr->ring + idx) == dq); idx 707 drivers/soc/fsl/qbman/qman.c DPAA_ASSERT(idx < QM_DQRR_SIZE); idx 710 drivers/soc/fsl/qbman/qman.c idx); /* DQRR_DCAP::DCAP_CI */ idx 1124 drivers/soc/fsl/qbman/qman.c static struct qman_fq *idx_to_fq(u32 idx) idx 1129 drivers/soc/fsl/qbman/qman.c if (WARN_ON(idx >= num_fqids * 2)) idx 1132 drivers/soc/fsl/qbman/qman.c fq = fq_table[idx]; idx 1133 drivers/soc/fsl/qbman/qman.c DPAA_ASSERT(!fq || idx == fq->idx); idx 1159 drivers/soc/fsl/qbman/qman.c return fq->idx; idx 1812 drivers/soc/fsl/qbman/qman.c fq->idx = fqid * 2; idx 1814 drivers/soc/fsl/qbman/qman.c fq->idx++; idx 1816 drivers/soc/fsl/qbman/qman.c WARN_ON(fq_table[fq->idx]); idx 1817 drivers/soc/fsl/qbman/qman.c fq_table[fq->idx] = fq; idx 1835 drivers/soc/fsl/qbman/qman.c DPAA_ASSERT(fq_table[fq->idx]); idx 1836 drivers/soc/fsl/qbman/qman.c fq_table[fq->idx] = NULL; idx 648 drivers/soc/fsl/qbman/qman_ccsr.c int idx = channel - QM_CHANNEL_SWPORTAL0; idx 651 drivers/soc/fsl/qbman/qman_ccsr.c before = qm_ccsr_in(REG_REV3_QCSP_LIO_CFG(idx)); idx 653 drivers/soc/fsl/qbman/qman_ccsr.c before = qm_ccsr_in(REG_QCSP_LIO_CFG(idx)); idx 661 drivers/soc/fsl/qbman/qman_ccsr.c qm_ccsr_out(REG_REV3_QCSP_LIO_CFG(idx), after); idx 663 drivers/soc/fsl/qbman/qman_ccsr.c qm_ccsr_out(REG_QCSP_LIO_CFG(idx), after); idx 669 drivers/soc/fsl/qbman/qman_ccsr.c int idx = channel - QM_CHANNEL_SWPORTAL0; idx 673 drivers/soc/fsl/qbman/qman_ccsr.c before = qm_ccsr_in(REG_REV3_QCSP_IO_CFG(idx)); idx 677 drivers/soc/fsl/qbman/qman_ccsr.c qm_ccsr_out(REG_REV3_QCSP_IO_CFG(idx), after); idx 679 drivers/soc/fsl/qbman/qman_ccsr.c before = qm_ccsr_in(REG_QCSP_IO_CFG(idx)); idx 681 drivers/soc/fsl/qbman/qman_ccsr.c qm_ccsr_out(REG_QCSP_IO_CFG(idx), after); idx 386 drivers/soc/qcom/qcom-geni-se.c int idx = idx_start; idx 397 drivers/soc/qcom/qcom-geni-se.c cfg[i] = idx << PACKING_START_SHIFT; idx 402 drivers/soc/qcom/qcom-geni-se.c idx = ((i + 1) * BITS_PER_BYTE) + idx_start; idx 405 drivers/soc/qcom/qcom-geni-se.c idx = idx + idx_delta; idx 205 drivers/soc/qcom/socinfo.c int idx; idx 207 drivers/soc/qcom/socinfo.c for (idx = 0; idx < ARRAY_SIZE(soc_id); idx++) { idx 208 drivers/soc/qcom/socinfo.c if (soc_id[idx].id == id) idx 209 drivers/soc/qcom/socinfo.c return soc_id[idx].name; idx 434 drivers/soc/renesas/rcar-sysc.c static int rcar_sysc_power_cpu(unsigned int idx, bool on) idx 449 drivers/soc/renesas/rcar-sysc.c if (!(pd->flags & PD_CPU) || pd->ch.chan_bit != idx) idx 288 drivers/soc/renesas/rmobile-sysc.c u32 idx = ~0; idx 290 drivers/soc/renesas/rmobile-sysc.c if (of_property_read_u32(np, "reg", &idx)) { idx 302 drivers/soc/renesas/rmobile-sysc.c pd->bit_shift = idx; idx 572 drivers/soc/rockchip/pm_domains.c u32 idx; idx 574 drivers/soc/rockchip/pm_domains.c error = of_property_read_u32(parent, "reg", &idx); idx 581 drivers/soc/rockchip/pm_domains.c parent_domain = pmu->genpd_data.domains[idx]; idx 590 drivers/soc/rockchip/pm_domains.c error = of_property_read_u32(np, "reg", &idx); idx 597 drivers/soc/rockchip/pm_domains.c child_domain = pmu->genpd_data.domains[idx]; idx 86 drivers/soc/ti/knav_qmss_acc.c u32 *list, *list_cpu, val, idx, notifies; idx 170 drivers/soc/ti/knav_qmss_acc.c idx = atomic_inc_return(&kq->desc_tail) & ACC_DESCS_MASK; idx 171 drivers/soc/ti/knav_qmss_acc.c kq->descs[idx] = val; idx 174 drivers/soc/ti/knav_qmss_acc.c val, idx, queue + range_base); idx 52 drivers/soc/ti/knav_qmss_queue.c #define knav_queue_idx_to_inst(kdev, idx) \ idx 53 drivers/soc/ti/knav_qmss_queue.c (kdev->instances + (idx << kdev->inst_shift)) idx 58 drivers/soc/ti/knav_qmss_queue.c #define for_each_instance(idx, inst, kdev) \ idx 59 drivers/soc/ti/knav_qmss_queue.c for (idx = 0, inst = kdev->instances; \ idx 60 drivers/soc/ti/knav_qmss_queue.c idx < (kdev)->num_queues_in_use; \ idx 61 drivers/soc/ti/knav_qmss_queue.c idx++, inst = knav_queue_idx_to_inst(kdev, idx)) idx 193 drivers/soc/ti/knav_qmss_queue.c int idx; idx 195 drivers/soc/ti/knav_qmss_queue.c for_each_instance(idx, inst, kdev) { idx 292 drivers/soc/ti/knav_qmss_queue.c int idx; idx 296 drivers/soc/ti/knav_qmss_queue.c for_each_instance(idx, inst, kdev) { idx 465 drivers/soc/ti/knav_qmss_queue.c int idx; idx 471 drivers/soc/ti/knav_qmss_queue.c for_each_instance(idx, inst, kdev) idx 666 drivers/soc/ti/knav_qmss_queue.c u32 val, idx; idx 674 drivers/soc/ti/knav_qmss_queue.c idx = atomic_inc_return(&inst->desc_head); idx 675 drivers/soc/ti/knav_qmss_queue.c idx &= ACC_DESCS_MASK; idx 676 drivers/soc/ti/knav_qmss_queue.c val = inst->descs[idx]; idx 1715 drivers/soc/ti/knav_qmss_queue.c int idx = 0, ret = 0; idx 1732 drivers/soc/ti/knav_qmss_queue.c base_idx = idx; idx 1734 drivers/soc/ti/knav_qmss_queue.c id < range->queue_base + range->num_queues; id++, idx++) { idx 1736 drivers/soc/ti/knav_qmss_queue.c knav_queue_idx_to_inst(kdev, idx), id); idx 29 drivers/soc/ti/ti_sci_pm_domains.c int idx; idx 59 drivers/soc/ti/ti_sci_pm_domains.c return sci_dev_data->idx; idx 92 drivers/soc/ti/ti_sci_pm_domains.c int idx = ti_sci_dev_id(dev); idx 95 drivers/soc/ti/ti_sci_pm_domains.c return ti_sci->ops.dev_ops.get_device_exclusive(ti_sci, idx); idx 97 drivers/soc/ti/ti_sci_pm_domains.c return ti_sci->ops.dev_ops.get_device(ti_sci, idx); idx 107 drivers/soc/ti/ti_sci_pm_domains.c int idx = ti_sci_dev_id(dev); idx 109 drivers/soc/ti/ti_sci_pm_domains.c return ti_sci->ops.dev_ops.put_device(ti_sci, idx); idx 121 drivers/soc/ti/ti_sci_pm_domains.c int idx, ret = 0; idx 131 drivers/soc/ti/ti_sci_pm_domains.c idx = pd_args.args[0]; idx 137 drivers/soc/ti/ti_sci_pm_domains.c ret = ti_sci->ops.dev_ops.is_valid(ti_sci, idx); idx 145 drivers/soc/ti/ti_sci_pm_domains.c sci_dev_data->idx = idx; idx 227 drivers/soc/xilinx/zynqmp_pm_domains.c unsigned int i, idx = genpdspec->args[0]; idx 237 drivers/soc/xilinx/zynqmp_pm_domains.c if (pd[i].node_id == idx) idx 247 drivers/soc/xilinx/zynqmp_pm_domains.c pd[i].node_id = idx; idx 125 drivers/spi/spi-fsl-qspi.c #define QUADSPI_LUT_REG(idx) \ idx 126 drivers/spi/spi-fsl-qspi.c (QUADSPI_LUT_BASE + QUADSPI_LUT_OFFSET + (idx) * 4) idx 164 drivers/spi/spi-fsl-qspi.c #define LUT_DEF(idx, ins, pad, opr) \ idx 165 drivers/spi/spi-fsl-qspi.c ((((ins) << 10) | ((pad) << 8) | (opr)) << (((idx) % 2) * 16)) idx 204 drivers/spi/spi-geni-qcom.c u32 demux_sel, clk_sel, m_clk_cfg, idx, div; idx 231 drivers/spi/spi-geni-qcom.c ret = get_spi_clk_cfg(mas->cur_speed_hz, mas, &idx, &div); idx 238 drivers/spi/spi-geni-qcom.c clk_sel = idx & CLK_SEL_MSK; idx 321 drivers/spi/spi-geni-qcom.c unsigned int idx, div; idx 323 drivers/spi/spi-geni-qcom.c ret = get_spi_clk_cfg(xfer->speed_hz, mas, &idx, &div); idx 336 drivers/spi/spi-geni-qcom.c clk_sel = idx & CLK_SEL_MSK; idx 427 drivers/spi/spi-npcm-fiu.c u32 idx = 0; idx 445 drivers/spi/spi-npcm-fiu.c for (idx = 0; idx < num_data_chunks; ++idx) { idx 246 drivers/spi/spi-nxp-fspi.c #define FSPI_LUT_REG(idx) \ idx 247 drivers/spi/spi-nxp-fspi.c (FSPI_LUT_BASE + FSPI_LUT_OFFSET + (idx) * 4) idx 304 drivers/spi/spi-nxp-fspi.c #define LUT_DEF(idx, ins, pad, opr) \ idx 306 drivers/spi/spi-nxp-fspi.c (opr)) << (((idx) % 2) * OPRND_SHIFT)) idx 108 drivers/spi/spi-omap-uwire.c static inline void uwire_write_reg(int idx, u16 val) idx 110 drivers/spi/spi-omap-uwire.c __raw_writew(val, uwire_base + (idx << uwire_idx_shift)); idx 113 drivers/spi/spi-omap-uwire.c static inline u16 uwire_read_reg(int idx) idx 115 drivers/spi/spi-omap-uwire.c return __raw_readw(uwire_base + (idx << uwire_idx_shift)); idx 146 drivers/spi/spi-omap2-mcspi.c int idx, u32 val) idx 150 drivers/spi/spi-omap2-mcspi.c writel_relaxed(val, mcspi->base + idx); idx 153 drivers/spi/spi-omap2-mcspi.c static inline u32 mcspi_read_reg(struct spi_master *master, int idx) idx 157 drivers/spi/spi-omap2-mcspi.c return readl_relaxed(mcspi->base + idx); idx 161 drivers/spi/spi-omap2-mcspi.c int idx, u32 val) idx 165 drivers/spi/spi-omap2-mcspi.c writel_relaxed(val, cs->base + idx); idx 168 drivers/spi/spi-omap2-mcspi.c static inline u32 mcspi_read_cs_reg(const struct spi_device *spi, int idx) idx 172 drivers/spi/spi-omap2-mcspi.c return readl_relaxed(cs->base + idx); idx 218 drivers/spi/spi-topcliff-pch.c static inline void pch_spi_writereg(struct spi_master *master, int idx, u32 val) idx 221 drivers/spi/spi-topcliff-pch.c iowrite32(val, (data->io_remap_addr + idx)); idx 229 drivers/spi/spi-topcliff-pch.c static inline u32 pch_spi_readreg(struct spi_master *master, int idx) idx 232 drivers/spi/spi-topcliff-pch.c return ioread32(data->io_remap_addr + idx); idx 235 drivers/spi/spi-topcliff-pch.c static inline void pch_spi_setclr_reg(struct spi_master *master, int idx, idx 238 drivers/spi/spi-topcliff-pch.c u32 tmp = pch_spi_readreg(master, idx); idx 240 drivers/spi/spi-topcliff-pch.c pch_spi_writereg(master, idx, tmp); idx 146 drivers/staging/comedi/drivers/gsc_hpdi.c unsigned int idx; idx 154 drivers/staging/comedi/drivers/gsc_hpdi.c idx = devpriv->dma_desc_index; idx 155 drivers/staging/comedi/drivers/gsc_hpdi.c start = le32_to_cpu(devpriv->dma_desc[idx].pci_start_addr); idx 166 drivers/staging/comedi/drivers/gsc_hpdi.c comedi_buf_write_samples(s, devpriv->desc_dio_buffer[idx], idx 168 drivers/staging/comedi/drivers/gsc_hpdi.c idx++; idx 169 drivers/staging/comedi/drivers/gsc_hpdi.c idx %= devpriv->num_dma_descriptors; idx 170 drivers/staging/comedi/drivers/gsc_hpdi.c start = le32_to_cpu(devpriv->dma_desc[idx].pci_start_addr); idx 172 drivers/staging/comedi/drivers/gsc_hpdi.c devpriv->dma_desc_index = idx; idx 419 drivers/staging/comedi/drivers/gsc_hpdi.c unsigned int idx = 0; idx 428 drivers/staging/comedi/drivers/gsc_hpdi.c for (i = 0; i < NUM_DMA_DESCRIPTORS && idx < NUM_DMA_BUFFERS; i++) { idx 430 drivers/staging/comedi/drivers/gsc_hpdi.c cpu_to_le32(devpriv->dio_buffer_phys_addr[idx] + offset); idx 436 drivers/staging/comedi/drivers/gsc_hpdi.c devpriv->desc_dio_buffer[i] = devpriv->dio_buffer[idx] + idx 442 drivers/staging/comedi/drivers/gsc_hpdi.c idx++; idx 151 drivers/staging/fbtft/fb_hx8340bn.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 95 drivers/staging/fbtft/fb_hx8347d.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 195 drivers/staging/fbtft/fb_ili9163.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 214 drivers/staging/fbtft/fb_ili9320.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 208 drivers/staging/fbtft/fb_ili9325.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 114 drivers/staging/fbtft/fb_ili9341.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 123 drivers/staging/fbtft/fb_s6d1121.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 129 drivers/staging/fbtft/fb_ssd1289.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 136 drivers/staging/fbtft/fb_st7735r.c #define CURVE(num, idx) curves[(num) * par->gamma.num_values + (idx)] idx 128 drivers/staging/fieldbus/anybuss/arcx-anybus.c int idx) idx 134 drivers/staging/fieldbus/anybuss/arcx-anybus.c res = platform_get_resource(pdev, IORESOURCE_MEM, idx + 1); idx 142 drivers/staging/fieldbus/anybuss/arcx-anybus.c create_anybus_host(struct platform_device *pdev, int idx) idx 146 drivers/staging/fieldbus/anybuss/arcx-anybus.c switch (idx) { idx 156 drivers/staging/fieldbus/anybuss/arcx-anybus.c ops.host_idx = idx; idx 157 drivers/staging/fieldbus/anybuss/arcx-anybus.c ops.regmap = create_parallel_regmap(pdev, idx); idx 160 drivers/staging/fieldbus/anybuss/arcx-anybus.c ops.irq = platform_get_irq(pdev, idx); idx 65 drivers/staging/fsl-dpaa2/ethsw/ethsw-ethtool.c port_priv->idx, idx 101 drivers/staging/fsl-dpaa2/ethsw/ethsw-ethtool.c port_priv->idx); idx 120 drivers/staging/fsl-dpaa2/ethsw/ethsw-ethtool.c port_priv->idx, idx 126 drivers/staging/fsl-dpaa2/ethsw/ethsw-ethtool.c port_priv->idx); idx 169 drivers/staging/fsl-dpaa2/ethsw/ethsw-ethtool.c port_priv->idx, idx 57 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, &tci_cfg); idx 70 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx); idx 78 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, &tci_cfg); idx 93 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx); idx 117 drivers/staging/fsl-dpaa2/ethsw/ethsw.c vcfg.if_id[0] = port_priv->idx; idx 174 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, enable); idx 198 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, &stp_cfg); idx 239 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_egress = port_priv->idx; idx 258 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_egress = port_priv->idx; idx 281 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_id[0] = port_priv->idx; idx 302 drivers/staging/fsl-dpaa2/ethsw/ethsw.c entry.if_id[0] = port_priv->idx; idx 348 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 355 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 362 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 369 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 376 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 384 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 393 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 432 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, idx 452 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx, &state); idx 480 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx); idx 499 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx); idx 510 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx); idx 545 drivers/staging/fsl-dpaa2/ethsw/ethsw.c err = snprintf(name, len, "p%d", port_priv->idx); idx 556 drivers/staging/fsl-dpaa2/ethsw/ethsw.c int idx; idx 568 drivers/staging/fsl-dpaa2/ethsw/ethsw.c if (dump->idx < dump->cb->args[2]) idx 591 drivers/staging/fsl-dpaa2/ethsw/ethsw.c dump->idx++; idx 602 drivers/staging/fsl-dpaa2/ethsw/ethsw.c int idx = port_priv->idx; idx 606 drivers/staging/fsl-dpaa2/ethsw/ethsw.c valid = entry->if_info == port_priv->idx; idx 608 drivers/staging/fsl-dpaa2/ethsw/ethsw.c valid = entry->if_mask[idx / 8] & BIT(idx % 8); idx 615 drivers/staging/fsl-dpaa2/ethsw/ethsw.c struct net_device *filter_dev, int *idx) idx 626 drivers/staging/fsl-dpaa2/ethsw/ethsw.c .idx = *idx, idx 669 drivers/staging/fsl-dpaa2/ethsw/ethsw.c *idx = dump.idx; idx 998 drivers/staging/fsl-dpaa2/ethsw/ethsw.c vcfg.if_id[0] = port_priv->idx; idx 1470 drivers/staging/fsl-dpaa2/ethsw/ethsw.c vcfg.if_id[0] = port_priv->idx; idx 1577 drivers/staging/fsl-dpaa2/ethsw/ethsw.c port_priv->idx = port_idx; idx 47 drivers/staging/fsl-dpaa2/ethsw/ethsw.h u16 idx; idx 407 drivers/staging/gdm724x/gdm_lte.c int idx; idx 461 drivers/staging/gdm724x/gdm_lte.c ret = sscanf(dev->name, "lte%d", &idx); idx 470 drivers/staging/gdm724x/gdm_lte.c tx_complete, nic, idx, idx 507 drivers/staging/gdm724x/gdm_lte.c int idx; idx 510 drivers/staging/gdm724x/gdm_lte.c ret = sscanf(dev->name, "lte%d", &idx); idx 516 drivers/staging/gdm724x/gdm_lte.c return netlink_send(lte_event.sock, idx, 0, buf, length); idx 486 drivers/staging/iio/adc/ad7192.c int idx = 0; idx 497 drivers/staging/iio/adc/ad7192.c idx = i; idx 501 drivers/staging/iio/adc/ad7192.c switch (idx) { idx 223 drivers/staging/iio/cdc/ad7746.c int ret, delay, idx; idx 229 drivers/staging/iio/cdc/ad7746.c idx = (chip->config & AD7746_CONF_CAPFS_MASK) >> idx 231 drivers/staging/iio/cdc/ad7746.c delay = ad7746_cap_filter_rate_table[idx][1]; idx 252 drivers/staging/iio/cdc/ad7746.c idx = (chip->config & AD7746_CONF_VTFS_MASK) >> idx 254 drivers/staging/iio/cdc/ad7746.c delay = ad7746_cap_filter_rate_table[idx][1]; idx 534 drivers/staging/iio/cdc/ad7746.c int ret, delay, idx; idx 643 drivers/staging/iio/cdc/ad7746.c idx = (chip->config & AD7746_CONF_CAPFS_MASK) >> idx 645 drivers/staging/iio/cdc/ad7746.c *val = ad7746_cap_filter_rate_table[idx][0]; idx 649 drivers/staging/iio/cdc/ad7746.c idx = (chip->config & AD7746_CONF_VTFS_MASK) >> idx 651 drivers/staging/iio/cdc/ad7746.c *val = ad7746_vt_filter_rate_table[idx][0]; idx 162 drivers/staging/kpc2000/kpc2000_spi.c kp_spi_read_reg(struct kp_spi_controller_state *cs, int idx) idx 167 drivers/staging/kpc2000/kpc2000_spi.c addr += idx; idx 168 drivers/staging/kpc2000/kpc2000_spi.c if ((idx == KP_SPI_REG_CONFIG) && (cs->conf_cache >= 0)) idx 176 drivers/staging/kpc2000/kpc2000_spi.c kp_spi_write_reg(struct kp_spi_controller_state *cs, int idx, u64 val) idx 180 drivers/staging/kpc2000/kpc2000_spi.c addr += idx; idx 182 drivers/staging/kpc2000/kpc2000_spi.c if (idx == KP_SPI_REG_CONFIG) idx 187 drivers/staging/kpc2000/kpc2000_spi.c kp_spi_wait_for_reg_bit(struct kp_spi_controller_state *cs, int idx, idx 193 drivers/staging/kpc2000/kpc2000_spi.c while (!(kp_spi_read_reg(cs, idx) & bit)) { idx 195 drivers/staging/kpc2000/kpc2000_spi.c if (!(kp_spi_read_reg(cs, idx) & bit)) idx 94 drivers/staging/media/tegra-vde/vde.c unsigned int idx; idx 107 drivers/staging/media/tegra-vde/vde.c for (idx = 0, frame_idx = 1; idx < refs_nb; idx++, frame_idx++) { idx 113 drivers/staging/media/tegra-vde/vde.c frame_idx_enb_mask |= frame_idx << (6 * (idx % 4)); idx 115 drivers/staging/media/tegra-vde/vde.c if (idx % 4 == 3 || idx == refs_nb - 1) { idx 117 drivers/staging/media/tegra-vde/vde.c value |= (idx >> 2) << 24; idx 205 drivers/staging/media/tegra-vde/vde.c unsigned int idx; idx 207 drivers/staging/media/tegra-vde/vde.c for (idx = 0; idx < frames_nb; idx++) idx 208 drivers/staging/media/tegra-vde/vde.c tegra_vde_setup_frameid(vde, &frames[idx], idx, idx 211 drivers/staging/media/tegra-vde/vde.c for (; idx < 17; idx++) idx 212 drivers/staging/media/tegra-vde/vde.c tegra_vde_setup_frameid(vde, NULL, idx, 0, 0); idx 324 drivers/staging/most/dim2/dim2.c int idx = 0; idx 329 drivers/staging/most/dim2/dim2.c buffer[idx++] = &dev->hch[ch_idx].ch; idx 331 drivers/staging/most/dim2/dim2.c buffer[idx++] = NULL; idx 228 drivers/staging/most/dim2/hal.c u8 const idx = (ch_addr % 8) / 2; idx 233 drivers/staging/most/dim2/hal.c mask[idx] = (u32)0xFFFF << shift; idx 234 drivers/staging/most/dim2/hal.c value[idx] = cat << shift; idx 241 drivers/staging/most/dim2/hal.c u8 const idx = (ch_addr % 8) / 2; idx 246 drivers/staging/most/dim2/hal.c mask[idx] = (u32)0xFFFF << shift; idx 297 drivers/staging/most/dim2/hal.c static void dim2_start_ctrl_async(u8 ch_addr, u8 idx, u32 buf_addr, idx 300 drivers/staging/most/dim2/hal.c u8 const shift = idx * 16; idx 314 drivers/staging/most/dim2/hal.c mask[idx + 2] = 0xFFFFFFFF; idx 315 drivers/staging/most/dim2/hal.c adt[idx + 2] = buf_addr; idx 320 drivers/staging/most/dim2/hal.c static void dim2_start_isoc_sync(u8 ch_addr, u8 idx, u32 buf_addr, idx 323 drivers/staging/most/dim2/hal.c u8 const shift = idx * 16; idx 335 drivers/staging/most/dim2/hal.c mask[idx + 2] = 0xFFFFFFFF; idx 336 drivers/staging/most/dim2/hal.c adt[idx + 2] = buf_addr; idx 574 drivers/staging/most/dim2/hal.c static inline bool service_channel(u8 ch_addr, u8 idx) idx 576 drivers/staging/most/dim2/hal.c u8 const shift = idx * 16; idx 87 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c u8 idx = 0; idx 89 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c for (idx = 0; idx < 5; idx++) idx 90 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c if (DynamicTxRPTTiming[idx] == pRaInfo->RptTime) idx 94 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c idx = 0; /* 200ms */ idx 96 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c idx += 1; idx 97 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c if (idx > 5) idx 98 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c idx = 5; idx 100 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c if (idx != 0) idx 101 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c idx -= 1; idx 103 drivers/staging/rtl8188eu/hal/hal8188e_rate_adaptive.c pRaInfo->RptTime = DynamicTxRPTTiming[idx]; idx 1188 drivers/staging/rtl8188eu/hal/usb_halinit.c u8 idx = 0; idx 1193 drivers/staging/rtl8188eu/hal/usb_halinit.c for (idx = 0; idx < 6; idx++) idx 1194 drivers/staging/rtl8188eu/hal/usb_halinit.c usb_write8(Adapter, (reg_macid + idx), val[idx]); idx 1199 drivers/staging/rtl8188eu/hal/usb_halinit.c u8 idx = 0; idx 1204 drivers/staging/rtl8188eu/hal/usb_halinit.c for (idx = 0; idx < 6; idx++) idx 1205 drivers/staging/rtl8188eu/hal/usb_halinit.c usb_write8(Adapter, (reg_bssid + idx), val[idx]); idx 217 drivers/staging/rtl8188eu/include/ieee80211.h u8 idx; idx 367 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c if (param->u.crypt.idx >= WEP_KEYS) { idx 384 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c wep_key_idx = param->u.crypt.idx; idx 463 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(padapter->securitypriv.dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, min_t(u16, param->u.crypt.key_len, 16 )); idx 464 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(padapter->securitypriv.dot118021XGrptxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[16]), 8); idx 465 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(padapter->securitypriv.dot118021XGrprxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[24]), 8); idx 469 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c padapter->securitypriv.dot118021XGrpKeyid = param->u.crypt.idx; idx 471 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c rtw_set_key(padapter, &padapter->securitypriv, param->u.crypt.idx, 1); idx 1909 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c param->u.crypt.idx = (pencoding->flags&0x00FF) - 1; idx 2211 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c if (param->u.crypt.idx >= WEP_KEYS) { idx 2226 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c DBG_88E("clear default encryption keys, keyid =%d\n", param->u.crypt.idx); idx 2231 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c wep_key_idx = param->u.crypt.idx; idx 2298 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, idx 2307 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, idx 2310 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrptxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[16]), 8); idx 2311 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrprxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[24]), 8); idx 2317 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, idx 2323 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c psecuritypriv->dot118021XGrpKeyid = param->u.crypt.idx; idx 2326 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c set_group_key(padapter, param->u.crypt.key, psecuritypriv->dot118021XGrpPrivacy, param->u.crypt.idx); idx 2372 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, idx 2380 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, idx 2384 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrptxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[16]), 8); idx 2385 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrprxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[24]), 8); idx 2391 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, idx 2397 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c psecuritypriv->dot118021XGrpKeyid = param->u.crypt.idx; idx 2403 drivers/staging/rtl8188eu/os_dep/ioctl_linux.c set_group_key(padapter, param->u.crypt.key, psecuritypriv->dot118021XGrpPrivacy, param->u.crypt.idx); idx 1583 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct tx_desc *entry = &ring->desc[ring->idx]; idx 1589 drivers/staging/rtl8192e/rtl8192e/rtl_core.c ring->idx = (ring->idx + 1) % ring->entries; idx 1670 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct tx_desc *entry = &ring->desc[ring->idx]; idx 1676 drivers/staging/rtl8192e/rtl8192e/rtl_core.c ring->idx = (ring->idx + 1) % ring->entries; idx 1694 drivers/staging/rtl8192e/rtl8192e/rtl_core.c unsigned int idx; idx 1701 drivers/staging/rtl8192e/rtl8192e/rtl_core.c idx = (ring->idx + skb_queue_len(&ring->queue)) % ring->entries; idx 1702 drivers/staging/rtl8192e/rtl8192e/rtl_core.c entry = (struct tx_desc_cmd *)&ring->desc[idx]; idx 1723 drivers/staging/rtl8192e/rtl8192e/rtl_core.c int idx; idx 1751 drivers/staging/rtl8192e/rtl8192e/rtl_core.c idx = (ring->idx + skb_queue_len(&ring->queue)) % ring->entries; idx 1753 drivers/staging/rtl8192e/rtl8192e/rtl_core.c idx = 0; idx 1755 drivers/staging/rtl8192e/rtl8192e/rtl_core.c pdesc = &ring->desc[idx]; idx 1759 drivers/staging/rtl8192e/rtl8192e/rtl_core.c tcb_desc->queue_index, ring->idx, idx, skb->len, idx 1843 drivers/staging/rtl8192e/rtl8192e/rtl_core.c priv->tx_ring[prio].idx = 0; idx 1905 drivers/staging/rtl8192e/rtl8192e/rtl_core.c struct tx_desc *entry = &ring->desc[ring->idx]; idx 1913 drivers/staging/rtl8192e/rtl8192e/rtl_core.c ring->idx = (ring->idx + 1) % ring->entries; idx 1915 drivers/staging/rtl8192e/rtl8192e/rtl_core.c ring->idx = 0; idx 264 drivers/staging/rtl8192e/rtl8192e/rtl_core.h unsigned int idx; idx 171 drivers/staging/rtl8192e/rtl8192e/rtl_dm.h #define dm_tx_bb_gain_idx_to_amplify(idx) (-idx + 12) idx 889 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c u8 idx = 0, alg = 0, group = 0; idx 902 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c idx = encoding->flags & IW_ENCODE_INDEX; idx 903 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c if (idx) idx 904 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c idx--; idx 919 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c rtl92e_set_key(dev, idx, idx, alg, zero, 0, key); idx 920 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c rtl92e_set_swcam(dev, idx, idx, alg, zero, 0, key, 0); idx 923 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c rtl92e_set_key(dev, idx, idx, alg, broadcast_addr, 0, idx 925 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c rtl92e_set_swcam(dev, idx, idx, alg, broadcast_addr, 0, idx 931 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c rtl92e_set_key(dev, 4, idx, alg, idx 933 drivers/staging/rtl8192e/rtl8192e/rtl_wx.c rtl92e_set_swcam(dev, 4, idx, alg, idx 1048 drivers/staging/rtl8192e/rtllib_rx.c int idx = 0; idx 1052 drivers/staging/rtl8192e/rtllib_rx.c idx = skb->data[hdrlen + 3] >> 6; idx 1054 drivers/staging/rtl8192e/rtllib_rx.c *crypt = ieee->crypt_info.crypt[idx]; idx 517 drivers/staging/rtl8192e/rtllib_wx.c int i, idx; idx 526 drivers/staging/rtl8192e/rtllib_wx.c idx = encoding->flags & IW_ENCODE_INDEX; idx 527 drivers/staging/rtl8192e/rtllib_wx.c if (idx) { idx 528 drivers/staging/rtl8192e/rtllib_wx.c if (idx < 1 || idx > NUM_WEP_KEYS) idx 530 drivers/staging/rtl8192e/rtllib_wx.c idx--; idx 532 drivers/staging/rtl8192e/rtllib_wx.c idx = ieee->crypt_info.tx_keyidx; idx 535 drivers/staging/rtl8192e/rtllib_wx.c crypt = &ieee->crypt_info.crypt[idx]; idx 539 drivers/staging/rtl8192e/rtllib_wx.c if (idx != 0 && ext->alg != IW_ENCODE_ALG_WEP) idx 542 drivers/staging/rtl8192e/rtllib_wx.c crypt = &ieee->crypt_info.crypt[idx]; idx 613 drivers/staging/rtl8192e/rtllib_wx.c new_crypt->priv = new_crypt->ops->init(idx); idx 632 drivers/staging/rtl8192e/rtllib_wx.c ieee->crypt_info.tx_keyidx = idx; idx 633 drivers/staging/rtl8192e/rtllib_wx.c sec.active_key = idx; idx 637 drivers/staging/rtl8192e/rtllib_wx.c sec.key_sizes[idx] = ext->key_len; idx 638 drivers/staging/rtl8192e/rtllib_wx.c sec.flags |= (1 << idx); idx 316 drivers/staging/rtl8192u/ieee80211/ieee80211.h u8 idx; idx 953 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c int idx = 0; idx 955 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c idx = skb->data[hdrlen + 3] >> 6; idx 956 drivers/staging/rtl8192u/ieee80211/ieee80211_rx.c crypt = ieee->crypt[idx]; idx 2812 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c if (param->u.crypt.idx >= WEP_KEYS) idx 2814 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c crypt = &ieee->crypt[param->u.crypt.idx]; idx 2870 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c new_crypt->ops->init(param->u.crypt.idx); idx 2894 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c ieee->tx_keyidx = param->u.crypt.idx; idx 2895 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c sec.active_key = param->u.crypt.idx; idx 2900 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c memcpy(sec.keys[param->u.crypt.idx], idx 2903 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c sec.key_sizes[param->u.crypt.idx] = param->u.crypt.key_len; idx 2904 drivers/staging/rtl8192u/ieee80211/ieee80211_softmac.c sec.flags |= (1 << param->u.crypt.idx); idx 497 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c int i, idx; idx 506 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c idx = encoding->flags & IW_ENCODE_INDEX; idx 507 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c if (idx) { idx 508 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c if (idx < 1 || idx > WEP_KEYS) idx 510 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c idx--; idx 512 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c idx = ieee->tx_keyidx; idx 516 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c crypt = &ieee->crypt[idx]; idx 521 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c if (idx != 0 && ext->alg != IW_ENCODE_ALG_WEP) idx 525 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c crypt = &ieee->crypt[idx]; idx 596 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c new_crypt->priv = new_crypt->ops->init(idx); idx 615 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c ieee->tx_keyidx = idx; idx 616 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c sec.active_key = idx; idx 622 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c sec.key_sizes[idx] = ext->key_len; idx 623 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c sec.flags |= BIT(idx); idx 662 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c int idx, max_key_len; idx 668 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c idx = encoding->flags & IW_ENCODE_INDEX; idx 669 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c if (idx) { idx 670 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c if (idx < 1 || idx > WEP_KEYS) idx 672 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c idx--; idx 674 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c idx = ieee->tx_keyidx; idx 678 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c if (idx != 0 || ieee->iw_mode != IW_MODE_INFRA) idx 681 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c crypt = ieee->crypt[idx]; idx 682 drivers/staging/rtl8192u/ieee80211/ieee80211_wx.c encoding->flags = idx + 1; idx 3574 drivers/staging/rtl8192u/r8192U_core.c setKey(dev, 4, ipw->u.crypt.idx, idx 3579 drivers/staging/rtl8192u/r8192U_core.c setKey(dev, ipw->u.crypt.idx, idx 3580 drivers/staging/rtl8192u/r8192U_core.c ipw->u.crypt.idx, idx 3601 drivers/staging/rtl8192u/r8192U_core.c setKey(dev, ipw->u.crypt.idx, idx 3603 drivers/staging/rtl8192u/r8192U_core.c ipw->u.crypt.idx, idx 739 drivers/staging/rtl8192u/r8192U_wx.c u8 idx = 0, alg = 0, group = 0; idx 747 drivers/staging/rtl8192u/r8192U_wx.c idx = encoding->flags & IW_ENCODE_INDEX; idx 748 drivers/staging/rtl8192u/r8192U_wx.c if (idx) idx 749 drivers/staging/rtl8192u/r8192U_wx.c idx--; idx 763 drivers/staging/rtl8192u/r8192U_wx.c idx, /* EntryNao */ idx 764 drivers/staging/rtl8192u/r8192U_wx.c idx, /* KeyIndex */ idx 772 drivers/staging/rtl8192u/r8192U_wx.c idx, /* EntryNo */ idx 773 drivers/staging/rtl8192u/r8192U_wx.c idx, /* KeyIndex */ idx 781 drivers/staging/rtl8192u/r8192U_wx.c idx, /* KeyIndex */ idx 97 drivers/staging/rtl8712/ieee80211.h u8 idx; idx 457 drivers/staging/rtl8712/rtl8712_efuse.c int i, idx; idx 469 drivers/staging/rtl8712/rtl8712_efuse.c idx = 0; /* data index */ idx 473 drivers/staging/rtl8712/rtl8712_efuse.c data[idx++] = pktdata[i]; idx 474 drivers/staging/rtl8712/rtl8712_efuse.c if (idx == cnts) idx 490 drivers/staging/rtl8712/rtl8712_efuse.c int i, j, idx; idx 512 drivers/staging/rtl8712/rtl8712_efuse.c idx = 0; /* data index */ idx 516 drivers/staging/rtl8712/rtl8712_efuse.c if (data[idx] != pktdata[i]) { idx 519 drivers/staging/rtl8712/rtl8712_efuse.c newdata[j++] = data[idx]; idx 522 drivers/staging/rtl8712/rtl8712_efuse.c idx++; idx 526 drivers/staging/rtl8712/rtl8712_efuse.c if ((cnts - idx) == 1) { idx 527 drivers/staging/rtl8712/rtl8712_efuse.c if (data[idx] != pktdata[i]) { idx 529 drivers/staging/rtl8712/rtl8712_efuse.c newdata[j++] = data[idx]; idx 532 drivers/staging/rtl8712/rtl8712_efuse.c idx++; idx 536 drivers/staging/rtl8712/rtl8712_efuse.c if ((data[idx] != pktdata[i]) || (data[idx + 1] != idx 539 drivers/staging/rtl8712/rtl8712_efuse.c newdata[j++] = data[idx]; idx 540 drivers/staging/rtl8712/rtl8712_efuse.c newdata[j++] = data[idx + 1]; idx 542 drivers/staging/rtl8712/rtl8712_efuse.c idx += 2; idx 544 drivers/staging/rtl8712/rtl8712_efuse.c if (idx == cnts) idx 552 drivers/staging/rtl8712/rtl8712_efuse.c if (idx == cnts) idx 104 drivers/staging/rtl8712/rtl871x_ioctl_linux.c if (param->u.crypt.idx > 0 && idx 105 drivers/staging/rtl8712/rtl871x_ioctl_linux.c param->u.crypt.idx < 3) { idx 107 drivers/staging/rtl8712/rtl871x_ioctl_linux.c memcpy(gk[param->u.crypt.idx - 1].skey, idx 111 drivers/staging/rtl8712/rtl871x_ioctl_linux.c memcpy(gtk[param->u.crypt.idx - 1].skey, idx 113 drivers/staging/rtl8712/rtl871x_ioctl_linux.c memcpy(grk[param->u.crypt.idx - 1].skey, idx 117 drivers/staging/rtl8712/rtl871x_ioctl_linux.c param->u.crypt.idx); idx 390 drivers/staging/rtl8712/rtl871x_ioctl_linux.c if (param->u.crypt.idx >= WEP_KEYS) { idx 392 drivers/staging/rtl8712/rtl871x_ioctl_linux.c param->u.crypt.idx = 0; idx 400 drivers/staging/rtl8712/rtl871x_ioctl_linux.c wep_key_idx = param->u.crypt.idx; idx 1796 drivers/staging/rtl8712/rtl871x_ioctl_linux.c param->u.crypt.idx = (pencoding->flags & 0x00FF) - 1; idx 145 drivers/staging/rtl8712/rtl871x_recv.c u8 *mickey, idx, *iv; idx 157 drivers/staging/rtl8712/rtl871x_recv.c idx = iv[3]; idx 158 drivers/staging/rtl8712/rtl871x_recv.c mickey = &psecuritypriv->XGrprxmickey[(((idx >> idx 648 drivers/staging/rtl8712/rtl871x_security.c u8 *pframe, *payload, *iv, *prwskey, idx = 0; idx 669 drivers/staging/rtl8712/rtl871x_security.c idx = iv[3]; idx 671 drivers/staging/rtl8712/rtl871x_security.c ((idx >> 6) & 0x3) - 1].skey[0]; idx 1348 drivers/staging/rtl8712/rtl871x_security.c u8 *pframe, *prwskey, *iv, idx; idx 1363 drivers/staging/rtl8712/rtl871x_security.c idx = iv[3]; idx 1365 drivers/staging/rtl8712/rtl871x_security.c ((idx >> 6) & 0x3) - 1].skey[0]; idx 194 drivers/staging/rtl8723bs/core/rtw_eeprom.c u16 reg, stmp, i = 0, idx = 0; idx 202 drivers/staging/rtl8723bs/core/rtw_eeprom.c rbuf[idx++] = (u8) ((stmp>>8)&0xff); /*return hogh-part of the short*/ idx 211 drivers/staging/rtl8723bs/core/rtw_eeprom.c rbuf[idx++] = (u8) (stmp&0xff); idx 212 drivers/staging/rtl8723bs/core/rtw_eeprom.c rbuf[idx++] = (u8) ((stmp>>8)&0xff); idx 218 drivers/staging/rtl8723bs/core/rtw_eeprom.c rbuf[idx] = (u8)(stmp & 0xff); idx 2834 drivers/staging/rtl8723bs/hal/hal_com_phycfg.c sscanf(token, "%d", &idx);\ idx 2835 drivers/staging/rtl8723bs/hal/hal_com_phycfg.c _array[_iteratedIdx++] = (u8)idx;\ idx 2844 drivers/staging/rtl8723bs/hal/hal_com_phycfg.c u32 idx = 0; idx 3414 drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c u8 idx = 0; idx 3419 drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c for (idx = 0 ; idx < 6; idx++) idx 3420 drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c rtw_write8(GET_PRIMARY_ADAPTER(padapter), (reg_macid+idx), val[idx]); idx 3425 drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c u8 idx = 0; idx 3430 drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c for (idx = 0 ; idx < 6; idx++) idx 3431 drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c rtw_write8(padapter, (reg_bssid+idx), val[idx]); idx 205 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c u8 idx, hwentry; idx 224 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c for (idx = 0; idx < 4; idx++) idx 225 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c inx[idx] = pxmitpriv->wmm_para_seq[idx]; idx 234 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c for (idx = 0; idx < hwentry; idx++) { idx 235 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c phwxmit = hwxmits + inx[idx]; idx 247 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c max_xmit_len = rtw_hal_get_sdio_tx_max_length(padapter, inx[idx]); idx 263 drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c idx, idx 226 drivers/staging/rtl8723bs/include/ieee80211.h u8 idx; idx 610 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c if (param->u.crypt.idx >= WEP_KEYS) idx 631 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c DBG_8192C("clear default encryption keys, keyid =%d\n", param->u.crypt.idx); idx 641 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c wep_key_idx = param->u.crypt.idx; idx 694 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 709 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 713 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrptxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[16]), 8); idx 714 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrprxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[24]), 8); idx 725 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 734 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c psecuritypriv->dot118021XGrpKeyid = param->u.crypt.idx; idx 740 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c rtw_ap_set_group_key(padapter, param->u.crypt.key, psecuritypriv->dot118021XGrpPrivacy, param->u.crypt.idx); idx 812 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 824 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 828 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrptxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[16]), 8); idx 829 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrprxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[24]), 8); idx 838 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 845 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c psecuritypriv->dot118021XGrpKeyid = param->u.crypt.idx; idx 851 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c rtw_ap_set_group_key(padapter, param->u.crypt.key, psecuritypriv->dot118021XGrpPrivacy, param->u.crypt.idx); idx 895 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c if (param->u.crypt.idx >= WEP_KEYS idx 896 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c || param->u.crypt.idx >= BIP_MAX_KEYID idx 914 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c wep_key_idx = param->u.crypt.idx; idx 1004 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(padapter->securitypriv.dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 1005 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(padapter->securitypriv.dot118021XGrptxmickey[param->u.crypt.idx].skey,&(param->u.crypt.key[16]), 8); idx 1006 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(padapter->securitypriv.dot118021XGrprxmickey[param->u.crypt.idx].skey,&(param->u.crypt.key[24]), 8); idx 1011 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c padapter->securitypriv.dot118021XGrpKeyid = param->u.crypt.idx; idx 1012 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c rtw_set_key(padapter,&padapter->securitypriv, param->u.crypt.idx, 1, true); idx 1018 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c memcpy(padapter->securitypriv.dot11wBIPKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 1023 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c padapter->securitypriv.dot11wBIPKeyid = param->u.crypt.idx; idx 1122 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c param->u.crypt.idx = key_index; idx 2933 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c static struct sta_info *rtw_sta_info_get_by_idx(const int idx, struct sta_priv *pstapriv) idx 2946 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c if (idx == i) psta = LIST_CONTAINOR(plist, struct sta_info, asoc_list); idx 2954 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c int idx, u8 *mac, struct station_info *sinfo) idx 2964 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c psta = rtw_sta_info_get_by_idx(idx, pstapriv); idx 442 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c if (param->u.crypt.idx >= WEP_KEYS || idx 443 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c param->u.crypt.idx >= BIP_MAX_KEYID) { idx 462 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c wep_key_idx = param->u.crypt.idx; idx 559 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(padapter->securitypriv.dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 562 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(padapter->securitypriv.dot118021XGrptxmickey[param->u.crypt.idx].skey,&(param->u.crypt.key[16]), 8); idx 563 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(padapter->securitypriv.dot118021XGrprxmickey[param->u.crypt.idx].skey,&(param->u.crypt.key[24]), 8); idx 569 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c padapter->securitypriv.dot118021XGrpKeyid = param->u.crypt.idx; idx 571 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c rtw_set_key(padapter,&padapter->securitypriv, param->u.crypt.idx, 1, true); idx 575 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(padapter->securitypriv.dot11wBIPKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 580 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c padapter->securitypriv.dot11wBIPKeyid = param->u.crypt.idx; idx 2184 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c param->u.crypt.idx = (pencoding->flags&0x00FF) -1 ; idx 3464 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c if (param->u.crypt.idx >= WEP_KEYS) { idx 3485 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c DBG_871X("clear default encryption keys, keyid =%d\n", param->u.crypt.idx); idx 3494 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c wep_key_idx = param->u.crypt.idx; idx 3567 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 3578 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 3582 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrptxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[16]), 8); idx 3583 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrprxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[24]), 8); idx 3593 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 3600 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c psecuritypriv->dot118021XGrpKeyid = param->u.crypt.idx; idx 3606 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c rtw_ap_set_group_key(padapter, param->u.crypt.key, psecuritypriv->dot118021XGrpPrivacy, param->u.crypt.idx); idx 3659 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 3667 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 3671 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrptxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[16]), 8); idx 3672 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrprxmickey[param->u.crypt.idx].skey, &(param->u.crypt.key[24]), 8); idx 3679 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c memcpy(psecuritypriv->dot118021XGrpKey[param->u.crypt.idx].skey, param->u.crypt.key, (param->u.crypt.key_len>16 ?16:param->u.crypt.key_len)); idx 3684 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c psecuritypriv->dot118021XGrpKeyid = param->u.crypt.idx; idx 3690 drivers/staging/rtl8723bs/os_dep/ioctl_linux.c rtw_ap_set_group_key(padapter, param->u.crypt.key, psecuritypriv->dot118021XGrpPrivacy, param->u.crypt.idx); idx 2280 drivers/staging/rts5208/ms.c u16 start, end, phy_blk, log_blk, tmp_blk, idx; idx 2392 drivers/staging/rts5208/ms.c idx = log_blk - ms_start_idx[seg_no]; idx 2394 drivers/staging/rts5208/ms.c if (segment->l2p_table[idx] == 0xFFFF) { idx 2395 drivers/staging/rts5208/ms.c segment->l2p_table[idx] = phy_blk; idx 2400 drivers/staging/rts5208/ms.c tmp_blk = segment->l2p_table[idx]; idx 2431 drivers/staging/rts5208/ms.c idx = log_blk - ms_start_idx[seg_no]; idx 2432 drivers/staging/rts5208/ms.c if (segment->l2p_table[idx] == 0xFFFF) { idx 2442 drivers/staging/rts5208/ms.c segment->l2p_table[idx] = phy_blk; idx 325 drivers/staging/rts5208/rtsx_chip.h #define SET_BIT(data, idx) ((data) |= 1 << (idx)) idx 326 drivers/staging/rts5208/rtsx_chip.h #define CLR_BIT(data, idx) ((data) &= ~(1 << (idx))) idx 327 drivers/staging/rts5208/rtsx_chip.h #define CHK_BIT(data, idx) ((data) & (1 << (idx))) idx 1723 drivers/staging/rts5208/rtsx_scsi.c u8 cmd_type, mask, value, idx; idx 1757 drivers/staging/rts5208/rtsx_scsi.c idx = srb->cmnd[4]; idx 1758 drivers/staging/rts5208/rtsx_scsi.c value = *(rtsx_get_cmd_data(chip) + idx); idx 1647 drivers/staging/rts5208/sd.c int idx = cont_path_cnt - 1; idx 1649 drivers/staging/rts5208/sd.c path[idx].len = path[idx].end - idx 1650 drivers/staging/rts5208/sd.c path[idx].start + 1; idx 1651 drivers/staging/rts5208/sd.c path[idx].mid = path[idx].start + idx 1652 drivers/staging/rts5208/sd.c path[idx].len / 2; idx 1661 drivers/staging/rts5208/sd.c int idx = cont_path_cnt - 1; idx 1663 drivers/staging/rts5208/sd.c path[idx].len = path[idx].end - path[idx].start + 1; idx 1664 drivers/staging/rts5208/sd.c path[idx].mid = path[idx].start + path[idx].len / 2; idx 603 drivers/staging/speakup/kobjects.c int *synth_default_value, int idx) idx 610 drivers/staging/speakup/kobjects.c spk_set_num_var(synth_default_value[idx], idx 91 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c int idx; idx 95 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c idx = substream->number; idx 100 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c } else if (!spdif && (chip->opened & (1 << idx))) { idx 104 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c if (idx >= MAX_SUBSTREAMS) { idx 107 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c idx, MAX_SUBSTREAMS); idx 121 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c alsa_stream->idx = idx; idx 148 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c chip->alsa_stream[idx] = alsa_stream; idx 150 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c chip->opened |= (1 << idx); idx 183 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c alsa_stream->chip->alsa_stream[alsa_stream->idx] = NULL; idx 342 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c int idx, enum snd_bcm2835_route route, idx 348 drivers/staging/vc04_services/bcm2835-audio/bcm2835-pcm.c err = snd_pcm_new(chip->card, name, idx, numchannels, 0, &pcm); idx 83 drivers/staging/vc04_services/bcm2835-audio/bcm2835.h int idx; idx 88 drivers/staging/vc04_services/bcm2835-audio/bcm2835.h int idx, enum snd_bcm2835_route route, idx 1618 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c int idx; /* port index */ idx 1648 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c for (idx = 0; idx < component->inputs; idx++) { idx 1649 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->input[idx].type = MMAL_PORT_TYPE_INPUT; idx 1650 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->input[idx].index = idx; idx 1651 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->input[idx].component = component; idx 1652 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c spin_lock_init(&component->input[idx].slock); idx 1653 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c INIT_LIST_HEAD(&component->input[idx].buffers); idx 1654 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c ret = port_info_get(instance, &component->input[idx]); idx 1659 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c for (idx = 0; idx < component->outputs; idx++) { idx 1660 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->output[idx].type = MMAL_PORT_TYPE_OUTPUT; idx 1661 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->output[idx].index = idx; idx 1662 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->output[idx].component = component; idx 1663 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c spin_lock_init(&component->output[idx].slock); idx 1664 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c INIT_LIST_HEAD(&component->output[idx].buffers); idx 1665 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c ret = port_info_get(instance, &component->output[idx]); idx 1670 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c for (idx = 0; idx < component->clocks; idx++) { idx 1671 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->clock[idx].type = MMAL_PORT_TYPE_CLOCK; idx 1672 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->clock[idx].index = idx; idx 1673 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c component->clock[idx].component = component; idx 1674 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c spin_lock_init(&component->clock[idx].slock); idx 1675 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c INIT_LIST_HEAD(&component->clock[idx].buffers); idx 1676 drivers/staging/vc04_services/bcm2835-camera/mmal-vchiq.c ret = port_info_get(instance, &component->clock[idx]); idx 234 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c int idx = *pidx; idx 237 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c while (idx < state->unused_service) { idx 238 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c struct vchiq_service *srv = state->services[idx++]; idx 250 drivers/staging/vc04_services/interface/vchiq_arm/vchiq_core.c *pidx = idx; idx 133 drivers/staging/vt6655/device_main.c static int device_rx_srv(struct vnt_private *priv, unsigned int idx); idx 134 drivers/staging/vt6655/device_main.c static int device_tx_srv(struct vnt_private *priv, unsigned int idx); idx 749 drivers/staging/vt6655/device_main.c static int device_rx_srv(struct vnt_private *priv, unsigned int idx) idx 754 drivers/staging/vt6655/device_main.c for (rd = priv->pCurrRD[idx]; idx 773 drivers/staging/vt6655/device_main.c priv->pCurrRD[idx] = rd; idx 841 drivers/staging/vt6655/device_main.c s8 idx; idx 854 drivers/staging/vt6655/device_main.c idx = info->control.rates[0].idx; idx 872 drivers/staging/vt6655/device_main.c idx = tx_rate - RATE_6M; idx 874 drivers/staging/vt6655/device_main.c idx = tx_rate; idx 882 drivers/staging/vt6655/device_main.c info->status.rates[0].idx = idx; idx 893 drivers/staging/vt6655/device_main.c static int device_tx_srv(struct vnt_private *priv, unsigned int idx) idx 900 drivers/staging/vt6655/device_main.c for (desc = priv->apTailTD[idx]; priv->iTDUsed[idx] > 0; desc = desc->next) { idx 915 drivers/staging/vt6655/device_main.c (int)idx, byTsr1, idx 920 drivers/staging/vt6655/device_main.c (int)idx, byTsr1, byTsr0); idx 927 drivers/staging/vt6655/device_main.c (int)idx, byTsr1, byTsr0); idx 934 drivers/staging/vt6655/device_main.c priv->iTDUsed[idx]--; idx 938 drivers/staging/vt6655/device_main.c priv->apTailTD[idx] = desc; idx 64 drivers/staging/vt6656/int.c s8 idx; idx 75 drivers/staging/vt6656/int.c idx = info->control.rates[0].idx; idx 93 drivers/staging/vt6656/int.c idx = tx_rate - RATE_6M; idx 95 drivers/staging/vt6656/int.c idx = tx_rate; idx 103 drivers/staging/vt6656/int.c info->status.rates[0].idx = idx; idx 793 drivers/staging/vt6656/rf.c u8 idx = ((rssi & 0xc0) >> 6) & 0x03; idx 805 drivers/staging/vt6656/rf.c a = airoharf[idx]; idx 126 drivers/staging/wilc1000/wilc_hif.c u8 idx; idx 175 drivers/staging/wilc1000/wilc_hif.c return vif->idx + 1; idx 183 drivers/staging/wilc1000/wilc_hif.c static struct wilc_vif *wilc_get_vif_from_idx(struct wilc *wilc, int idx) idx 185 drivers/staging/wilc1000/wilc_hif.c int index = idx - 1; idx 509 drivers/staging/wilc1000/wilc_hif.c param->idx = noa_attr.index; idx 630 drivers/staging/wilc1000/wilc_netdev.c vif->idx); idx 925 drivers/staging/wilc1000/wilc_netdev.c vif->idx = wl->vif_num; idx 464 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c static int wilc_wfi_cfg_allocate_wpa_entry(struct wilc_priv *priv, u8 idx) idx 466 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c if (!priv->wilc_gtk[idx]) { idx 467 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c priv->wilc_gtk[idx] = kzalloc(sizeof(*priv->wilc_gtk[idx]), idx 469 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c if (!priv->wilc_gtk[idx]) idx 473 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c if (!priv->wilc_ptk[idx]) { idx 474 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c priv->wilc_ptk[idx] = kzalloc(sizeof(*priv->wilc_ptk[idx]), idx 476 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c if (!priv->wilc_ptk[idx]) idx 1369 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c int idx, u8 *mac, struct station_info *sinfo) idx 1374 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c if (idx != 0) idx 1425 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c WILC_STATION_MODE, vif->idx); idx 1440 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c WILC_STATION_MODE, vif->idx); idx 1450 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c WILC_AP_MODE, vif->idx); idx 1460 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c WILC_AP_MODE, vif->idx); idx 1669 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c for (i = vif->idx; i < wl->vif_num; i++) { idx 1674 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c vif->idx = i; idx 1677 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c vif->iftype, vif->idx); idx 194 drivers/staging/wilc1000/wilc_wfi_netdevice.h u8 idx; idx 273 drivers/staging/wusbcore/devconnect.c unsigned idx; idx 278 drivers/staging/wusbcore/devconnect.c for (idx = 0; idx < wusbhc->ports_max; idx++) { idx 279 drivers/staging/wusbcore/devconnect.c port = wusb_port_by_idx(wusbhc, idx); idx 285 drivers/staging/wusbcore/devconnect.c for (idx = 0; idx < wusbhc->ports_max; idx++) { idx 286 drivers/staging/wusbcore/devconnect.c port = wusb_port_by_idx(wusbhc, idx); idx 291 drivers/staging/wusbcore/devconnect.c if (idx >= wusbhc->ports_max) { idx 302 drivers/staging/wusbcore/devconnect.c wusbhc->set_ptk(wusbhc, idx, 0, NULL, 0); idx 306 drivers/staging/wusbcore/devconnect.c wusb_dev = wusbhc_cack_add(wusbhc, dnc, pr_cdid, idx); idx 110 drivers/staging/wusbcore/host/whci/whci-hc.h __u8 idx; /*< index into page list */ idx 14 drivers/staging/wusbcore/host/whci/wusb.c static int whc_update_di(struct whc *whc, int idx) idx 16 drivers/staging/wusbcore/host/whci/wusb.c int offset = idx / 32; idx 17 drivers/staging/wusbcore/host/whci/wusb.c u32 bit = 1 << (idx % 32); idx 95 drivers/staging/wusbcore/host/whci/wusb.c int idx = wusb_dev->port_idx; idx 96 drivers/staging/wusbcore/host/whci/wusb.c struct di_buf_entry *di = &whc->di_buf[idx]; idx 105 drivers/staging/wusbcore/host/whci/wusb.c ret = whc_update_di(whc, idx); idx 1168 drivers/target/iscsi/cxgbit/cxgbit_cm.c L2T_IDX_V(csk->l2t->idx) | idx 65 drivers/target/iscsi/cxgbit/cxgbit_ddp.c unsigned int idx, unsigned int npods, unsigned int tid) idx 69 drivers/target/iscsi/cxgbit/cxgbit_ddp.c unsigned int pm_addr = (idx << PPOD_SIZE_SHIFT) + ppm->llimit; idx 99 drivers/target/iscsi/cxgbit/cxgbit_ddp.c struct cxgbi_task_tag_info *ttinfo, unsigned int idx, idx 110 drivers/target/iscsi/cxgbit/cxgbit_ddp.c skb = cxgbit_ppod_init_idata(cdev, ppm, idx, npods, csk->tid); idx 130 drivers/target/iscsi/cxgbit/cxgbit_ddp.c unsigned int pidx = ttinfo->idx; idx 197 drivers/target/iscsi/cxgbit/cxgbit_ddp.c ret = cxgbi_ppm_ppods_reserve(ppm, ttinfo->nr_pages, 0, &ttinfo->idx, idx 225 drivers/target/iscsi/cxgbit/cxgbit_ddp.c cxgbi_ppm_ppod_release(ppm, ttinfo->idx); idx 283 drivers/target/iscsi/cxgbit/cxgbit_ddp.c cxgbi_ppm_ppod_release(ppm, ttinfo->idx); idx 1367 drivers/target/iscsi/cxgbit/cxgbit_target.c cxgbit_process_iscsi_pdu(struct cxgbit_sock *csk, struct sk_buff *skb, int idx) idx 1369 drivers/target/iscsi/cxgbit/cxgbit_target.c struct cxgbit_lro_pdu_cb *pdu_cb = cxgbit_skb_lro_pdu_cb(skb, idx); idx 1168 drivers/target/sbp/sbp_target.c int data_size, direction, idx; idx 1183 drivers/target/sbp/sbp_target.c for (idx = 0; idx < data_size; idx++) { idx 1185 drivers/target/sbp/sbp_target.c req->pg_tbl[idx].segment_length); idx 1814 drivers/target/sbp/sbp_target.c int num_luns, num_entries, idx = 0, mgt_agt_addr, ret; idx 1850 drivers/target/sbp/sbp_target.c data[idx++] = num_entries << 16; idx 1854 drivers/target/sbp/sbp_target.c data[idx++] = (CSR_DIRECTORY_ID << 24) | tport->directory_id; idx 1857 drivers/target/sbp/sbp_target.c memcpy(&data[idx], sbp_unit_directory_template, idx 1859 drivers/target/sbp/sbp_target.c idx += ARRAY_SIZE(sbp_unit_directory_template); idx 1863 drivers/target/sbp/sbp_target.c data[idx++] = 0x54000000 | (mgt_agt_addr & 0x00ffffff); idx 1866 drivers/target/sbp/sbp_target.c data[idx++] = 0x3a000000 | idx 1871 drivers/target/sbp/sbp_target.c data[idx++] = 0x3d000000 | (tport->max_reconnect_timeout & 0xffff); idx 1874 drivers/target/sbp/sbp_target.c data[idx++] = 0x8d000000 | (num_luns + 1); idx 1888 drivers/target/sbp/sbp_target.c data[idx++] = 0x14000000 | idx 1895 drivers/target/sbp/sbp_target.c data[idx++] = 2 << 16; idx 1896 drivers/target/sbp/sbp_target.c data[idx++] = tport->guid >> 32; idx 1897 drivers/target/sbp/sbp_target.c data[idx++] = tport->guid; idx 1899 drivers/target/sbp/sbp_target.c tport->unit_directory.length = idx; idx 99 drivers/tee/optee/device.c u32 shm_size = 0, idx, num_devices = 0; idx 146 drivers/tee/optee/device.c for (idx = 0; idx < num_devices; idx++) { idx 147 drivers/tee/optee/device.c rc = optee_register_device(&device_uuid[idx], idx); idx 92 drivers/thermal/db8500_thermal.c unsigned int idx, idx 99 drivers/thermal/db8500_thermal.c th->cur_index = idx; idx 114 drivers/thermal/db8500_thermal.c unsigned int idx = th->cur_index; idx 117 drivers/thermal/db8500_thermal.c if (idx == 0) idx 121 drivers/thermal/db8500_thermal.c if (idx == 1) { idx 125 drivers/thermal/db8500_thermal.c next_high = db8500_thermal_points[idx - 1]; idx 126 drivers/thermal/db8500_thermal.c next_low = db8500_thermal_points[idx - 2]; idx 128 drivers/thermal/db8500_thermal.c idx -= 1; idx 130 drivers/thermal/db8500_thermal.c db8500_thermal_update_config(th, idx, THERMAL_TREND_DROPPING, idx 143 drivers/thermal/db8500_thermal.c unsigned int idx = th->cur_index; idx 147 drivers/thermal/db8500_thermal.c if (idx < num_points - 1) { idx 148 drivers/thermal/db8500_thermal.c next_high = db8500_thermal_points[idx+1]; idx 149 drivers/thermal/db8500_thermal.c next_low = db8500_thermal_points[idx]; idx 150 drivers/thermal/db8500_thermal.c idx += 1; idx 152 drivers/thermal/db8500_thermal.c db8500_thermal_update_config(th, idx, THERMAL_TREND_RAISING, idx 157 drivers/thermal/db8500_thermal.c } else if (idx == num_points - 1) idx 159 drivers/thermal/db8500_thermal.c th->interpolated_temp = db8500_thermal_points[idx] + 1; idx 45 drivers/thermal/tango_thermal.c int idx = priv->thresh_idx; idx 47 drivers/thermal/tango_thermal.c if (temp_above_thresh(priv->base, idx)) { idx 49 drivers/thermal/tango_thermal.c while (idx < IDX_MAX && temp_above_thresh(priv->base, ++idx)) idx 51 drivers/thermal/tango_thermal.c idx = idx - 1; /* always return lower bound */ idx 54 drivers/thermal/tango_thermal.c while (idx > IDX_MIN && !temp_above_thresh(priv->base, --idx)) idx 58 drivers/thermal/tango_thermal.c *res = (idx * 9 / 2 - 38) * 1000; /* millidegrees Celsius */ idx 59 drivers/thermal/tango_thermal.c priv->thresh_idx = idx; idx 23 drivers/thermal/tegra/tegra-bpmp-thermal.c unsigned int idx; idx 43 drivers/thermal/tegra/tegra-bpmp-thermal.c req.get_temp.zone = zone->idx; idx 69 drivers/thermal/tegra/tegra-bpmp-thermal.c req.set_trip.zone = zone->idx; idx 109 drivers/thermal/tegra/tegra-bpmp-thermal.c if (tegra->zones[i]->idx != req->host_trip_reached.zone) idx 189 drivers/thermal/tegra/tegra-bpmp-thermal.c zone->idx = i; idx 338 drivers/thunderbolt/nhi.c int idx = ring_interrupt_index(ring); idx 339 drivers/thunderbolt/nhi.c int reg = REG_RING_INTERRUPT_BASE + idx / 32 * 4; idx 340 drivers/thunderbolt/nhi.c int bit = idx % 32; idx 1620 drivers/tty/isicom.c int retval, idx, channel; idx 1623 drivers/tty/isicom.c for (idx = 0; idx < BOARD_COUNT; idx++) { idx 1624 drivers/tty/isicom.c port = &isi_ports[idx * 16]; idx 1625 drivers/tty/isicom.c isi_card[idx].ports = port; idx 1626 drivers/tty/isicom.c spin_lock_init(&isi_card[idx].card_lock); idx 1629 drivers/tty/isicom.c port->card = &isi_card[idx]; idx 1634 drivers/tty/isicom.c isi_card[idx].base = 0; idx 1635 drivers/tty/isicom.c isi_card[idx].irq = 0; idx 263 drivers/tty/mxser.c unsigned int idx; idx 2434 drivers/tty/mxser.c tty_unregister_device(mxvar_sdriver, brd->idx + i); idx 2576 drivers/tty/mxser.c brd->idx = i * MXSER_PORTS_PER_BOARD; idx 2648 drivers/tty/mxser.c mxvar_sdriver, brd->idx + i, &pdev->dev); idx 2653 drivers/tty/mxser.c brd->idx + i - 1); idx 2754 drivers/tty/mxser.c brd->idx = m * MXSER_PORTS_PER_BOARD; idx 2757 drivers/tty/mxser.c mxvar_sdriver, brd->idx + i, NULL); idx 2761 drivers/tty/mxser.c brd->idx + i - 1); idx 384 drivers/tty/pty.c int idx = tty->index; idx 399 drivers/tty/pty.c o_tty = alloc_tty_struct(driver->other, idx); idx 412 drivers/tty/pty.c driver->other->ttys[idx] = o_tty; idx 413 drivers/tty/pty.c driver->ttys[idx] = tty; idx 709 drivers/tty/pty.c struct file *file, int idx) idx 725 drivers/tty/pty.c struct file *file, int idx) idx 266 drivers/tty/serdev/serdev-ttyport.c struct tty_driver *drv, int idx) idx 281 drivers/tty/serdev/serdev-ttyport.c serport->tty_idx = idx; idx 293 drivers/tty/serdev/serdev-ttyport.c dev_info(&ctrl->dev, "tty port %s%d registered\n", drv->name, idx); idx 629 drivers/tty/serial/8250/8250_core.c static int univ8250_console_match(struct console *co, char *name, int idx, idx 198 drivers/tty/serial/8250/8250_exar.c int idx, unsigned int offset, idx 235 drivers/tty/serial/8250/8250_exar.c struct uart_8250_port *port, int idx) idx 237 drivers/tty/serial/8250/8250_exar.c unsigned int offset = idx * 0x200; idx 244 drivers/tty/serial/8250/8250_exar.c err = default_setup(priv, pcidev, idx, offset, port); idx 258 drivers/tty/serial/8250/8250_exar.c if (idx == 0) { idx 283 drivers/tty/serial/8250/8250_exar.c struct uart_8250_port *port, int idx) idx 285 drivers/tty/serial/8250/8250_exar.c unsigned int offset = idx * 0x200; idx 289 drivers/tty/serial/8250/8250_exar.c return default_setup(priv, pcidev, idx, offset, port); idx 294 drivers/tty/serial/8250/8250_exar.c struct uart_8250_port *port, int idx) idx 296 drivers/tty/serial/8250/8250_exar.c unsigned int offset = idx * 0x200; idx 300 drivers/tty/serial/8250/8250_exar.c return default_setup(priv, pcidev, idx, offset, port); idx 469 drivers/tty/serial/8250/8250_exar.c struct uart_8250_port *port, int idx) idx 473 drivers/tty/serial/8250/8250_exar.c unsigned int offset = idx * 0x400; idx 491 drivers/tty/serial/8250/8250_exar.c if (idx >= 8) idx 494 drivers/tty/serial/8250/8250_exar.c ret = default_setup(priv, pcidev, idx, offset, port); idx 505 drivers/tty/serial/8250/8250_exar.c if (idx == 0) { idx 1485 drivers/tty/serial/8250/8250_omap.c u8 idx; idx 1498 drivers/tty/serial/8250/8250_omap.c idx = *omap_str - '0'; idx 1510 drivers/tty/serial/8250/8250_omap.c add_preferred_console("ttyS", idx, options); idx 1512 drivers/tty/serial/8250/8250_omap.c idx, idx); idx 48 drivers/tty/serial/8250/8250_pci.c int idx; idx 119 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 124 drivers/tty/serial/8250/8250_pci.c if (idx < 2) { idx 125 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset; idx 126 drivers/tty/serial/8250/8250_pci.c } else if ((idx >= 2) && (idx < 4)) { idx 128 drivers/tty/serial/8250/8250_pci.c offset += ((idx - 2) * board->uart_offset); idx 129 drivers/tty/serial/8250/8250_pci.c } else if ((idx >= 4) && (idx < 6)) { idx 131 drivers/tty/serial/8250/8250_pci.c offset += ((idx - 4) * board->uart_offset); idx 132 drivers/tty/serial/8250/8250_pci.c } else if (idx >= 6) { idx 134 drivers/tty/serial/8250/8250_pci.c offset += ((idx - 6) * board->uart_offset); idx 146 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 151 drivers/tty/serial/8250/8250_pci.c if (idx < 4) idx 152 drivers/tty/serial/8250/8250_pci.c bar += idx; idx 155 drivers/tty/serial/8250/8250_pci.c offset += (idx - 4) * board->uart_offset; idx 201 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 208 drivers/tty/serial/8250/8250_pci.c if (idx == 3) idx 209 drivers/tty/serial/8250/8250_pci.c idx++; idx 212 drivers/tty/serial/8250/8250_pci.c if (idx > 0) idx 213 drivers/tty/serial/8250/8250_pci.c idx++; idx 214 drivers/tty/serial/8250/8250_pci.c if (idx > 2) idx 215 drivers/tty/serial/8250/8250_pci.c idx++; idx 218 drivers/tty/serial/8250/8250_pci.c if (idx > 2) idx 221 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset; idx 370 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 376 drivers/tty/serial/8250/8250_pci.c if (idx < 4) { idx 378 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset; idx 379 drivers/tty/serial/8250/8250_pci.c } else if (idx < 8) { idx 381 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset + 0xC00; idx 525 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 527 drivers/tty/serial/8250/8250_pci.c unsigned int bar = FL_GET_BASE(board->flags) + idx, offset = 0; idx 529 drivers/tty/serial/8250/8250_pci.c if (idx > 3) { idx 531 drivers/tty/serial/8250/8250_pci.c offset = (idx - 4) * 8; idx 619 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 623 drivers/tty/serial/8250/8250_pci.c switch (idx) { idx 641 drivers/tty/serial/8250/8250_pci.c bar = idx - 2; idx 653 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 657 drivers/tty/serial/8250/8250_pci.c switch (idx) { idx 666 drivers/tty/serial/8250/8250_pci.c offset = (idx - 2) * board->uart_offset; idx 754 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 760 drivers/tty/serial/8250/8250_pci.c if (idx >= board->num_ports) idx 764 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset; idx 781 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 790 drivers/tty/serial/8250/8250_pci.c bar = 3 * idx; idx 794 drivers/tty/serial/8250/8250_pci.c return pci_default_setup(priv, board, port, idx); idx 1069 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1072 drivers/tty/serial/8250/8250_pci.c return pci_default_setup(priv, board, port, idx); idx 1308 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1317 drivers/tty/serial/8250/8250_pci.c return pci_default_setup(priv, board, port, idx); idx 1326 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1332 drivers/tty/serial/8250/8250_pci.c bar += idx; idx 1334 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset; idx 1339 drivers/tty/serial/8250/8250_pci.c if (board->flags & FL_REGION_SZ_CAP && idx >= maxnr) idx 1376 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1382 drivers/tty/serial/8250/8250_pci.c bar += idx; idx 1384 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset; idx 1390 drivers/tty/serial/8250/8250_pci.c if (board->flags & FL_REGION_SZ_CAP && idx >= maxnr) idx 1400 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1406 drivers/tty/serial/8250/8250_pci.c bar += idx; idx 1408 drivers/tty/serial/8250/8250_pci.c offset += idx * board->uart_offset; idx 1410 drivers/tty/serial/8250/8250_pci.c if (idx==3) idx 1416 drivers/tty/serial/8250/8250_pci.c if (board->flags & FL_REGION_SZ_CAP && idx >= maxnr) idx 1427 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1431 drivers/tty/serial/8250/8250_pci.c ret = setup_port(priv, port, idx, 0, board->reg_shift); idx 1443 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1445 drivers/tty/serial/8250/8250_pci.c return setup_port(priv, port, 2, idx * 8, 0); idx 1451 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1453 drivers/tty/serial/8250/8250_pci.c int ret = pci_default_setup(priv, board, port, idx); idx 1514 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1521 drivers/tty/serial/8250/8250_pci.c config_base = 0x40 + 0x08 * idx; idx 1526 drivers/tty/serial/8250/8250_pci.c dev_dbg(&pdev->dev, "%s: idx=%d iobase=0x%x", __func__, idx, iobase); idx 1537 drivers/tty/serial/8250/8250_pci.c *data = idx; idx 1627 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1636 drivers/tty/serial/8250/8250_pci.c data->idx = idx; idx 1642 drivers/tty/serial/8250/8250_pci.c port->port.mapbase = pci_resource_start(pdev, 0) + 8 * idx; idx 1687 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1695 drivers/tty/serial/8250/8250_pci.c return pci_default_setup(priv, board, port, idx); idx 1734 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1739 drivers/tty/serial/8250/8250_pci.c return skip_tx_en_setup(priv, board, port, idx); idx 1754 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1758 drivers/tty/serial/8250/8250_pci.c return pci_default_setup(priv, board, port, idx); idx 1764 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1768 drivers/tty/serial/8250/8250_pci.c return pci_default_setup(priv, board, port, idx); idx 1774 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1778 drivers/tty/serial/8250/8250_pci.c return pci_default_setup(priv, board, port, idx); idx 1784 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1792 drivers/tty/serial/8250/8250_pci.c if (idx < 4) { idx 1794 drivers/tty/serial/8250/8250_pci.c offset = idx * board->uart_offset; idx 1797 drivers/tty/serial/8250/8250_pci.c idx -= 4; idx 1798 drivers/tty/serial/8250/8250_pci.c idx = div_s64_rem(idx, 4, &offset); idx 1799 drivers/tty/serial/8250/8250_pci.c offset = idx * 64 + offset * board->uart_offset; idx 1808 drivers/tty/serial/8250/8250_pci.c struct uart_8250_port *port, int idx) idx 1813 drivers/tty/serial/8250/8250_pci.c if (board->num_ports == 4 && idx == 3) idx 1816 drivers/tty/serial/8250/8250_pci.c offset = idx * board->uart_offset; idx 2358 drivers/tty/serial/amba-pl011.c static int __init pl011_console_match(struct console *co, char *name, int idx, idx 68 drivers/tty/serial/kgdb_nmi.c static struct tty_driver *kgdb_nmi_console_device(struct console *co, int *idx) idx 70 drivers/tty/serial/kgdb_nmi.c *idx = co->index; idx 174 drivers/tty/serial/kgdboc.c int idx; idx 175 drivers/tty/serial/kgdboc.c if (cons->device && cons->device(cons, &idx) == p && idx 176 drivers/tty/serial/kgdboc.c idx == tty_line) { idx 1743 drivers/tty/serial/mpc52xx_uart.c int idx = -1; idx 1750 drivers/tty/serial/mpc52xx_uart.c for (idx = 0; idx < MPC52xx_PSC_MAXNUM; idx++) idx 1751 drivers/tty/serial/mpc52xx_uart.c if (mpc52xx_uart_nodes[idx] == op->dev.of_node) idx 1753 drivers/tty/serial/mpc52xx_uart.c if (idx >= MPC52xx_PSC_MAXNUM) idx 1756 drivers/tty/serial/mpc52xx_uart.c mpc52xx_uart_nodes[idx], idx); idx 1768 drivers/tty/serial/mpc52xx_uart.c port = &mpc52xx_uart_ports[idx]; idx 1776 drivers/tty/serial/mpc52xx_uart.c port->line = idx; idx 426 drivers/tty/serial/pic32_uart.c sport->idx); idx 444 drivers/tty/serial/pic32_uart.c sport->idx); idx 462 drivers/tty/serial/pic32_uart.c sport->idx); idx 822 drivers/tty/serial/pic32_uart.c sport->idx = uart_idx; idx 904 drivers/tty/serial/pic32_uart.c pic32_sports[sport->idx] = NULL; idx 48 drivers/tty/serial/pic32_uart.h int idx; idx 180 drivers/tty/serial/rp2.c int idx; idx 530 drivers/tty/serial/rp2.c rp2_mask_ch_irq(up, up->idx, 1); idx 543 drivers/tty/serial/rp2.c rp2_mask_ch_irq(up, up->idx, 0); idx 690 drivers/tty/serial/rp2.c rp->idx = j; idx 670 drivers/tty/serial/sa1100.c void __init sa1100_register_uart(int idx, int port) idx 672 drivers/tty/serial/sa1100.c if (idx >= NR_PORTS) { idx 673 drivers/tty/serial/sa1100.c printk(KERN_ERR "%s: bad index number %d\n", __func__, idx); idx 679 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.membase = (void __iomem *)&Ser1UTCR0; idx 680 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.mapbase = _Ser1UTCR0; idx 681 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.irq = IRQ_Ser1UART; idx 682 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.flags = UPF_BOOT_AUTOCONF; idx 686 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.membase = (void __iomem *)&Ser2UTCR0; idx 687 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.mapbase = _Ser2UTCR0; idx 688 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.irq = IRQ_Ser2ICP; idx 689 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.flags = UPF_BOOT_AUTOCONF; idx 693 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.membase = (void __iomem *)&Ser3UTCR0; idx 694 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.mapbase = _Ser3UTCR0; idx 695 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.irq = IRQ_Ser3UART; idx 696 drivers/tty/serial/sa1100.c sa1100_ports[idx].port.flags = UPF_BOOT_AUTOCONF; idx 121 drivers/tty/serial/samsung.h static inline void s3c24xx_set_bit(struct uart_port *port, int idx, idx 129 drivers/tty/serial/samsung.h val |= (1 << idx); idx 134 drivers/tty/serial/samsung.h static inline void s3c24xx_clear_bit(struct uart_port *port, int idx, idx 142 drivers/tty/serial/samsung.h val &= ~(1 << idx); idx 1941 drivers/tty/serial/serial_core.c int idx = co->index; idx 1943 drivers/tty/serial/serial_core.c if (idx < 0 || idx >= nr || (ports[idx].iobase == 0 && idx 1944 drivers/tty/serial/serial_core.c ports[idx].membase == NULL)) idx 1945 drivers/tty/serial/serial_core.c for (idx = 0; idx < nr; idx++) idx 1946 drivers/tty/serial/serial_core.c if (ports[idx].iobase != 0 || idx 1947 drivers/tty/serial/serial_core.c ports[idx].membase != NULL) idx 1950 drivers/tty/serial/serial_core.c co->index = idx; idx 1952 drivers/tty/serial/serial_core.c return ports + idx; idx 40 drivers/tty/serial/serial_mctrl_gpio.c static bool mctrl_gpio_flags_is_dir_out(unsigned int idx) idx 42 drivers/tty/serial/serial_mctrl_gpio.c return mctrl_gpios_desc[idx].flags & GPIOD_FLAGS_BIT_DIR_OUT; idx 117 drivers/tty/serial/serial_mctrl_gpio.c struct mctrl_gpios *mctrl_gpio_init_noauto(struct device *dev, unsigned int idx) idx 144 drivers/tty/serial/serial_mctrl_gpio.c idx, idx 192 drivers/tty/serial/serial_mctrl_gpio.c struct mctrl_gpios *mctrl_gpio_init(struct uart_port *port, unsigned int idx) idx 197 drivers/tty/serial/serial_mctrl_gpio.c gpios = mctrl_gpio_init_noauto(port->dev, idx); idx 213 drivers/tty/serial/serial_mctrl_gpio.c mctrl_gpios_desc[i].name, idx, ret); idx 229 drivers/tty/serial/serial_mctrl_gpio.c mctrl_gpios_desc[i].name, idx, ret); idx 66 drivers/tty/serial/serial_mctrl_gpio.h struct mctrl_gpios *mctrl_gpio_init(struct uart_port *port, unsigned int idx); idx 75 drivers/tty/serial/serial_mctrl_gpio.h unsigned int idx); idx 121 drivers/tty/serial/serial_mctrl_gpio.h struct mctrl_gpios *mctrl_gpio_init(struct uart_port *port, unsigned int idx) idx 127 drivers/tty/serial/serial_mctrl_gpio.h struct mctrl_gpios *mctrl_gpio_init_noauto(struct device *dev, unsigned int idx) idx 159 drivers/tty/tty_io.c static void release_tty(struct tty_struct *tty, int idx); idx 1155 drivers/tty/tty_io.c struct file *file, int idx) idx 1163 drivers/tty/tty_io.c tty = driver->ops->lookup(driver, file, idx); idx 1165 drivers/tty/tty_io.c tty = driver->ttys[idx]; idx 1183 drivers/tty/tty_io.c int idx = tty->index; idx 1189 drivers/tty/tty_io.c tp = tty->driver->termios[idx]; idx 1317 drivers/tty/tty_io.c struct tty_struct *tty_init_dev(struct tty_driver *driver, int idx) idx 1333 drivers/tty/tty_io.c tty = alloc_tty_struct(driver, idx); idx 1345 drivers/tty/tty_io.c tty->port = driver->ports[idx]; idx 1379 drivers/tty/tty_io.c retval, idx); idx 1382 drivers/tty/tty_io.c release_tty(tty, idx); idx 1395 drivers/tty/tty_io.c int idx = tty->index; idx 1402 drivers/tty/tty_io.c tp = tty->driver->termios[idx]; idx 1407 drivers/tty/tty_io.c tty->driver->termios[idx] = tp; idx 1504 drivers/tty/tty_io.c static void release_tty(struct tty_struct *tty, int idx) idx 1507 drivers/tty/tty_io.c WARN_ON(tty->index != idx); idx 1533 drivers/tty/tty_io.c static int tty_release_checks(struct tty_struct *tty, int idx) idx 1536 drivers/tty/tty_io.c if (idx < 0 || idx >= tty->driver->num) { idx 1537 drivers/tty/tty_io.c tty_debug(tty, "bad idx %d\n", idx); idx 1545 drivers/tty/tty_io.c if (tty != tty->driver->ttys[idx]) { idx 1547 drivers/tty/tty_io.c idx, tty->driver->ttys[idx]); idx 1553 drivers/tty/tty_io.c if (o_tty != tty->driver->other->ttys[idx]) { idx 1555 drivers/tty/tty_io.c idx, tty->driver->other->ttys[idx]); idx 1607 drivers/tty/tty_io.c void tty_release_struct(struct tty_struct *tty, int idx) idx 1625 drivers/tty/tty_io.c release_tty(tty, idx); idx 1654 drivers/tty/tty_io.c int idx; idx 1666 drivers/tty/tty_io.c idx = tty->index; idx 1671 drivers/tty/tty_io.c if (tty_release_checks(tty, idx)) { idx 1785 drivers/tty/tty_io.c tty_release_struct(tty, idx); idx 2973 drivers/tty/tty_io.c struct tty_struct *alloc_tty_struct(struct tty_driver *driver, int idx) idx 3006 drivers/tty/tty_io.c tty->index = idx; idx 3007 drivers/tty/tty_io.c tty_line_name(driver, idx, tty->name); idx 87 drivers/usb/atm/ueagle-atm.c u16 idx; idx 1150 drivers/usb/atm/ueagle-atm.c sc->cmv_dsc.e1.idx++; idx 1157 drivers/usb/atm/ueagle-atm.c cmv.wIndex = cpu_to_le16(sc->cmv_dsc.e1.idx); idx 1982 drivers/usb/atm/ueagle-atm.c cmv->wIndex = cpu_to_le16(dsc->idx); idx 1998 drivers/usb/atm/ueagle-atm.c if (le16_to_cpu(cmv->wIndex) != dsc->idx || idx 129 drivers/usb/dwc2/debugfs.c int idx; idx 153 drivers/usb/dwc2/debugfs.c for (idx = 0; idx < hsotg->num_of_eps; idx++) { idx 156 drivers/usb/dwc2/debugfs.c in = dwc2_readl(hsotg, DIEPCTL(idx)); idx 157 drivers/usb/dwc2/debugfs.c out = dwc2_readl(hsotg, DOEPCTL(idx)); idx 160 drivers/usb/dwc2/debugfs.c idx, in, out); idx 162 drivers/usb/dwc2/debugfs.c in = dwc2_readl(hsotg, DIEPTSIZ(idx)); idx 163 drivers/usb/dwc2/debugfs.c out = dwc2_readl(hsotg, DOEPTSIZ(idx)); idx 187 drivers/usb/dwc2/debugfs.c int idx; idx 199 drivers/usb/dwc2/debugfs.c for (idx = 1; idx < hsotg->num_of_eps; idx++) { idx 200 drivers/usb/dwc2/debugfs.c val = dwc2_readl(hsotg, DPTXFSIZN(idx)); idx 202 drivers/usb/dwc2/debugfs.c seq_printf(seq, "\tDPTXFIFO%2d: Size %d, Start 0x%08x\n", idx, idx 1536 drivers/usb/dwc2/gadget.c int idx = windex & 0x7F; idx 1541 drivers/usb/dwc2/gadget.c if (idx > hsotg->num_of_eps) idx 1544 drivers/usb/dwc2/gadget.c ep = index_to_ep(hsotg, idx, dir); idx 1546 drivers/usb/dwc2/gadget.c if (idx && ep->dir_in != dir) idx 2605 drivers/usb/dwc2/gadget.c static void dwc2_hsotg_txfifo_flush(struct dwc2_hsotg *hsotg, unsigned int idx) idx 2607 drivers/usb/dwc2/gadget.c dwc2_writel(hsotg, GRSTCTL_TXFNUM(idx) | GRSTCTL_TXFFLSH, idx 2757 drivers/usb/dwc2/gadget.c unsigned int idx, int dir_in) idx 2760 drivers/usb/dwc2/gadget.c u32 epint_reg = dir_in ? DIEPINT(idx) : DOEPINT(idx); idx 2767 drivers/usb/dwc2/gadget.c mask |= ((diepempmsk >> idx) & 0x1) ? DIEPMSK_TXFIFOEMPTY : 0; idx 2792 drivers/usb/dwc2/gadget.c unsigned char idx = hs_ep->index; idx 2794 drivers/usb/dwc2/gadget.c u32 epctl_reg = dir_in ? DIEPCTL(idx) : DOEPCTL(idx); idx 2972 drivers/usb/dwc2/gadget.c static void dwc2_hsotg_epint(struct dwc2_hsotg *hsotg, unsigned int idx, idx 2975 drivers/usb/dwc2/gadget.c struct dwc2_hsotg_ep *hs_ep = index_to_ep(hsotg, idx, dir_in); idx 2976 drivers/usb/dwc2/gadget.c u32 epint_reg = dir_in ? DIEPINT(idx) : DOEPINT(idx); idx 2977 drivers/usb/dwc2/gadget.c u32 epctl_reg = dir_in ? DIEPCTL(idx) : DOEPCTL(idx); idx 2978 drivers/usb/dwc2/gadget.c u32 epsiz_reg = dir_in ? DIEPTSIZ(idx) : DOEPTSIZ(idx); idx 2982 drivers/usb/dwc2/gadget.c ints = dwc2_gadget_read_ep_interrupts(hsotg, idx, dir_in); idx 2990 drivers/usb/dwc2/gadget.c __func__, idx, dir_in ? "in" : "out"); idx 2995 drivers/usb/dwc2/gadget.c __func__, idx, dir_in ? "in" : "out", ints); idx 2998 drivers/usb/dwc2/gadget.c if (idx == 0 && (ints & (DXEPINT_SETUP | DXEPINT_SETUP_RCVD))) idx 3007 drivers/usb/dwc2/gadget.c if (using_desc_dma(hsotg) && idx == 0 && !hs_ep->dir_in && idx 3035 drivers/usb/dwc2/gadget.c if (idx == 0 && !hs_ep->req) idx 3045 drivers/usb/dwc2/gadget.c dwc2_hsotg_handle_outdone(hsotg, idx); idx 3064 drivers/usb/dwc2/gadget.c if (using_dma(hsotg) && idx == 0) { idx 3116 drivers/usb/dwc2/gadget.c __func__, idx); idx 3122 drivers/usb/dwc2/gadget.c __func__, idx); idx 3129 drivers/usb/dwc2/gadget.c __func__, idx); idx 3564 drivers/usb/dwc2/gadget.c u32 idx; idx 3570 drivers/usb/dwc2/gadget.c for (idx = 1; idx < hsotg->num_of_eps; idx++) { idx 3571 drivers/usb/dwc2/gadget.c hs_ep = hsotg->eps_in[idx]; idx 3573 drivers/usb/dwc2/gadget.c if ((BIT(idx) & ~daintmsk) || !hs_ep->isochronous) idx 3576 drivers/usb/dwc2/gadget.c epctrl = dwc2_readl(hsotg, DIEPCTL(idx)); idx 3581 drivers/usb/dwc2/gadget.c dwc2_writel(hsotg, epctrl, DIEPCTL(idx)); idx 3609 drivers/usb/dwc2/gadget.c int idx; idx 3616 drivers/usb/dwc2/gadget.c for (idx = 1; idx < hsotg->num_of_eps; idx++) { idx 3617 drivers/usb/dwc2/gadget.c hs_ep = hsotg->eps_out[idx]; idx 3619 drivers/usb/dwc2/gadget.c if ((BIT(idx) & ~daintmsk) || !hs_ep->isochronous) idx 3622 drivers/usb/dwc2/gadget.c epctrl = dwc2_readl(hsotg, DOEPCTL(idx)); idx 3776 drivers/usb/dwc2/gadget.c u8 idx; idx 3790 drivers/usb/dwc2/gadget.c for (idx = 1; idx < hsotg->num_of_eps; idx++) { idx 3791 drivers/usb/dwc2/gadget.c hs_ep = hsotg->eps_out[idx]; idx 3793 drivers/usb/dwc2/gadget.c if ((BIT(idx) & ~daintmsk) || !hs_ep->isochronous) idx 3796 drivers/usb/dwc2/gadget.c epctrl = dwc2_readl(hsotg, DOEPCTL(idx)); idx 3801 drivers/usb/dwc2/gadget.c dwc2_writel(hsotg, epctrl, DOEPCTL(idx)); idx 4759 drivers/usb/dwc2/gadget.c int idx; idx 4773 drivers/usb/dwc2/gadget.c for (idx = 1; idx < hsotg->num_of_eps; idx++) { idx 4774 drivers/usb/dwc2/gadget.c val = dwc2_readl(hsotg, DPTXFSIZN(idx)); idx 4775 drivers/usb/dwc2/gadget.c dev_info(dev, "DPTx[%d] FSize=%d, StAddr=0x%08x\n", idx, idx 4780 drivers/usb/dwc2/gadget.c for (idx = 0; idx < hsotg->num_of_eps; idx++) { idx 4782 drivers/usb/dwc2/gadget.c "ep%d-in: EPCTL=0x%08x, SIZ=0x%08x, DMA=0x%08x\n", idx, idx 4783 drivers/usb/dwc2/gadget.c dwc2_readl(hsotg, DIEPCTL(idx)), idx 4784 drivers/usb/dwc2/gadget.c dwc2_readl(hsotg, DIEPTSIZ(idx)), idx 4785 drivers/usb/dwc2/gadget.c dwc2_readl(hsotg, DIEPDMA(idx))); idx 4787 drivers/usb/dwc2/gadget.c val = dwc2_readl(hsotg, DOEPCTL(idx)); idx 4790 drivers/usb/dwc2/gadget.c idx, dwc2_readl(hsotg, DOEPCTL(idx)), idx 4791 drivers/usb/dwc2/gadget.c dwc2_readl(hsotg, DOEPTSIZ(idx)), idx 4792 drivers/usb/dwc2/gadget.c dwc2_readl(hsotg, DOEPDMA(idx))); idx 61 drivers/usb/dwc2/hcd_ddma.c static u16 dwc2_desclist_idx_inc(u16 idx, u16 inc, u8 speed) idx 63 drivers/usb/dwc2/hcd_ddma.c return (idx + inc) & idx 68 drivers/usb/dwc2/hcd_ddma.c static u16 dwc2_desclist_idx_dec(u16 idx, u16 inc, u8 speed) idx 70 drivers/usb/dwc2/hcd_ddma.c return (idx - inc) & idx 544 drivers/usb/dwc2/hcd_ddma.c u16 idx) idx 546 drivers/usb/dwc2/hcd_ddma.c struct dwc2_dma_desc *dma_desc = &qh->desc_list[idx]; idx 553 drivers/usb/dwc2/hcd_ddma.c qh->n_bytes[idx] = max_xfer_size; idx 555 drivers/usb/dwc2/hcd_ddma.c qh->n_bytes[idx] = frame_desc->length; idx 558 drivers/usb/dwc2/hcd_ddma.c dma_desc->status = qh->n_bytes[idx] << HOST_DMA_ISOC_NBYTES_SHIFT & idx 575 drivers/usb/dwc2/hcd_ddma.c (idx * sizeof(struct dwc2_dma_desc)), idx 585 drivers/usb/dwc2/hcd_ddma.c u16 idx, inc, n_desc = 0, ntd_max = 0; idx 589 drivers/usb/dwc2/hcd_ddma.c idx = qh->td_last; idx 608 drivers/usb/dwc2/hcd_ddma.c idx = qh->td_last; idx 628 drivers/usb/dwc2/hcd_ddma.c qtd->isoc_td_first = idx; idx 632 drivers/usb/dwc2/hcd_ddma.c max_xfer_size, idx); idx 633 drivers/usb/dwc2/hcd_ddma.c idx = dwc2_desclist_idx_inc(idx, inc, qh->dev_speed); idx 636 drivers/usb/dwc2/hcd_ddma.c qtd->isoc_td_last = idx; idx 640 drivers/usb/dwc2/hcd_ddma.c qh->td_last = idx; idx 645 drivers/usb/dwc2/hcd_ddma.c idx = dwc2_desclist_idx_dec(qh->td_last, inc, qh->dev_speed); idx 646 drivers/usb/dwc2/hcd_ddma.c qh->desc_list[idx].status |= HOST_DMA_IOC; idx 648 drivers/usb/dwc2/hcd_ddma.c qh->desc_list_dma + (idx * idx 670 drivers/usb/dwc2/hcd_ddma.c idx = dwc2_desclist_idx_dec(idx, inc * ((qh->ntd + 1) / 2), idx 678 drivers/usb/dwc2/hcd_ddma.c idx = dwc2_desclist_idx_dec(qh->td_last, inc, qh->dev_speed); idx 680 drivers/usb/dwc2/hcd_ddma.c qh->desc_list[idx].status |= HOST_DMA_IOC; idx 683 drivers/usb/dwc2/hcd_ddma.c (idx * sizeof(struct dwc2_dma_desc)), idx 895 drivers/usb/dwc2/hcd_ddma.c struct dwc2_qh *qh, u16 idx) idx 905 drivers/usb/dwc2/hcd_ddma.c dma_sync_single_for_cpu(hsotg->dev, qh->desc_list_dma + (idx * idx 910 drivers/usb/dwc2/hcd_ddma.c dma_desc = &qh->desc_list[idx]; idx 925 drivers/usb/dwc2/hcd_ddma.c frame_desc->actual_length = qh->n_bytes[idx] - remain; idx 929 drivers/usb/dwc2/hcd_ddma.c frame_desc->actual_length = qh->n_bytes[idx] - remain; idx 968 drivers/usb/dwc2/hcd_ddma.c u16 idx; idx 972 drivers/usb/dwc2/hcd_ddma.c idx = qh->td_first; idx 997 drivers/usb/dwc2/hcd_ddma.c for (idx = 0; idx < qtd->urb->packet_count; idx 998 drivers/usb/dwc2/hcd_ddma.c idx++) { idx 999 drivers/usb/dwc2/hcd_ddma.c frame_desc = &qtd->urb->iso_descs[idx]; idx 1021 drivers/usb/dwc2/hcd_ddma.c if (idx != qtd->isoc_td_first) { idx 1024 drivers/usb/dwc2/hcd_ddma.c idx, qtd->isoc_td_first); idx 1025 drivers/usb/dwc2/hcd_ddma.c idx = qtd->isoc_td_first; idx 1033 drivers/usb/dwc2/hcd_ddma.c idx); idx 1036 drivers/usb/dwc2/hcd_ddma.c idx = dwc2_desclist_idx_inc(idx, qh->host_interval, idx 1049 drivers/usb/dwc2/hcd_ddma.c qh->td_first = idx; idx 1060 drivers/usb/dwc2/hcd_ddma.c } while (idx != qh->td_first); idx 1064 drivers/usb/dwc2/hcd_ddma.c qh->td_first = idx; idx 2817 drivers/usb/gadget/function/f_fs.c int idx; idx 2843 drivers/usb/gadget/function/f_fs.c idx = ffs_ep_addr2idx(func->ffs, ds->bEndpointAddress) - 1; idx 2844 drivers/usb/gadget/function/f_fs.c if (idx < 0) idx 2845 drivers/usb/gadget/function/f_fs.c return idx; idx 2847 drivers/usb/gadget/function/f_fs.c ffs_ep = func->eps + idx; idx 2882 drivers/usb/gadget/function/f_fs.c ep->driver_data = func->eps + idx; idx 2891 drivers/usb/gadget/function/f_fs.c USB_ENDPOINT_NUMBER_MASK] = idx + 1; idx 2914 drivers/usb/gadget/function/f_fs.c unsigned idx; idx 2924 drivers/usb/gadget/function/f_fs.c idx = *valuep; idx 2925 drivers/usb/gadget/function/f_fs.c if (func->interfaces_nums[idx] < 0) { idx 2929 drivers/usb/gadget/function/f_fs.c func->interfaces_nums[idx] = id; idx 2931 drivers/usb/gadget/function/f_fs.c newValue = func->interfaces_nums[idx]; idx 2947 drivers/usb/gadget/function/f_fs.c idx = (*valuep & USB_ENDPOINT_NUMBER_MASK) - 1; idx 2948 drivers/usb/gadget/function/f_fs.c if (unlikely(!func->eps[idx].ep)) idx 2953 drivers/usb/gadget/function/f_fs.c descs = func->eps[idx].descs; idx 2716 drivers/usb/gadget/function/f_mass_storage.c struct attribute *attr, int idx) idx 536 drivers/usb/gadget/udc/aspeed-vhub/dev.c int ast_vhub_init_dev(struct ast_vhub *vhub, unsigned int idx) idx 538 drivers/usb/gadget/udc/aspeed-vhub/dev.c struct ast_vhub_dev *d = &vhub->ports[idx].dev; idx 543 drivers/usb/gadget/udc/aspeed-vhub/dev.c d->index = idx; idx 544 drivers/usb/gadget/udc/aspeed-vhub/dev.c d->name = devm_kasprintf(parent, GFP_KERNEL, "port%d", idx+1); idx 545 drivers/usb/gadget/udc/aspeed-vhub/dev.c d->regs = vhub->regs + 0x100 + 0x10 * idx; idx 560 drivers/usb/gadget/udc/aspeed-vhub/dev.c dev_set_name(d->port_dev, "%s:p%d", dev_name(parent), idx + 1); idx 534 drivers/usb/gadget/udc/aspeed-vhub/vhub.h int ast_vhub_init_dev(struct ast_vhub *vhub, unsigned int idx); idx 425 drivers/usb/gadget/udc/bcm63xx_udc.c static void bcm63xx_ep_dma_select(struct bcm63xx_udc *udc, int idx) idx 430 drivers/usb/gadget/udc/bcm63xx_udc.c val |= idx << USBD_CONTROL_INIT_SEL_SHIFT; idx 562 drivers/usb/gadget/udc/bcm63xx_udc.c int idx = cfg->ep_num; idx 566 drivers/usb/gadget/udc/bcm63xx_udc.c if (idx < 0) idx 568 drivers/usb/gadget/udc/bcm63xx_udc.c usb_ep_set_maxpacket_limit(&udc->bep[idx].ep, max_pkt); idx 570 drivers/usb/gadget/udc/bcm63xx_udc.c val = (idx << USBD_CSR_EP_LOG_SHIFT) | idx 577 drivers/usb/gadget/udc/bcm63xx_udc.c usbd_writel(udc, val, USBD_CSR_EP_REG(idx)); idx 174 drivers/usb/gadget/udc/pxa27x_udc.h #define ofs_UDCCR(ep) (UDCCRn(ep->idx)) idx 175 drivers/usb/gadget/udc/pxa27x_udc.h #define ofs_UDCCSR(ep) (UDCCSRn(ep->idx)) idx 176 drivers/usb/gadget/udc/pxa27x_udc.h #define ofs_UDCBCR(ep) (UDCBCRn(ep->idx)) idx 177 drivers/usb/gadget/udc/pxa27x_udc.h #define ofs_UDCDR(ep) (UDCDRn(ep->idx)) idx 198 drivers/usb/gadget/udc/pxa27x_udc.h #define EPIDX(ep) (ep->idx) idx 202 drivers/usb/gadget/udc/pxa27x_udc.h #define is_ep0(ep) (!ep->idx) idx 267 drivers/usb/gadget/udc/pxa27x_udc.h .idx = _idx, .enabled = 0, \ idx 353 drivers/usb/gadget/udc/pxa27x_udc.h unsigned idx:5; idx 307 drivers/usb/gadget/udc/s3c2410_udc.c u32 idx; idx 311 drivers/usb/gadget/udc/s3c2410_udc.c idx = ep->bEndpointAddress & 0x7F; idx 312 drivers/usb/gadget/udc/s3c2410_udc.c switch (idx) { idx 314 drivers/usb/gadget/udc/s3c2410_udc.c idx = 0; idx 344 drivers/usb/gadget/udc/s3c2410_udc.c if (idx == 0) idx 347 drivers/usb/gadget/udc/s3c2410_udc.c idx, count, req->req.actual, req->req.length, idx 354 drivers/usb/gadget/udc/s3c2410_udc.c if (idx == 0) { idx 361 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 363 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 371 drivers/usb/gadget/udc/s3c2410_udc.c if (idx == 0) { idx 377 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 379 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 412 drivers/usb/gadget/udc/s3c2410_udc.c u32 idx; idx 415 drivers/usb/gadget/udc/s3c2410_udc.c idx = ep->bEndpointAddress & 0x7F; idx 417 drivers/usb/gadget/udc/s3c2410_udc.c switch (idx) { idx 419 drivers/usb/gadget/udc/s3c2410_udc.c idx = 0; idx 448 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 463 drivers/usb/gadget/udc/s3c2410_udc.c if (idx != 0 && fifo_count < ep->ep.maxpacket) { idx 472 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 476 drivers/usb/gadget/udc/s3c2410_udc.c if (idx == 0) idx 481 drivers/usb/gadget/udc/s3c2410_udc.c if (idx == 0) { idx 485 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 487 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 494 drivers/usb/gadget/udc/s3c2410_udc.c if (idx == 0) { idx 497 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 499 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 799 drivers/usb/gadget/udc/s3c2410_udc.c u32 idx; idx 807 drivers/usb/gadget/udc/s3c2410_udc.c idx = ep->bEndpointAddress & 0x7F; idx 810 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 813 drivers/usb/gadget/udc/s3c2410_udc.c idx, ep_csr1, req ? 1 : 0); idx 817 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 826 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 828 drivers/usb/gadget/udc/s3c2410_udc.c dprintk(DEBUG_VERBOSE, "ep%01d rd csr:%02x\n", idx, ep_csr1); idx 831 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 855 drivers/usb/gadget/udc/s3c2410_udc.c u32 idx, idx2; idx 870 drivers/usb/gadget/udc/s3c2410_udc.c idx = udc_read(S3C2410_UDC_INDEX_REG); idx 911 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 987 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 1323 drivers/usb/gadget/udc/s3c2410_udc.c u32 idx; idx 1332 drivers/usb/gadget/udc/s3c2410_udc.c idx = ep->bEndpointAddress & 0x7F; idx 1334 drivers/usb/gadget/udc/s3c2410_udc.c if (idx == 0) { idx 1338 drivers/usb/gadget/udc/s3c2410_udc.c udc_write(idx, S3C2410_UDC_INDEX_REG); idx 179 drivers/usb/host/fsl-mph-dr-of.c static unsigned int idx; idx 250 drivers/usb/host/fsl-mph-dr-of.c dev_data->drivers[i], idx); idx 256 drivers/usb/host/fsl-mph-dr-of.c idx++; idx 1676 drivers/usb/host/max3421-hcd.c u8 mask, idx; idx 1683 drivers/usb/host/max3421-hcd.c idx = pin_number / 4; idx 1686 drivers/usb/host/max3421-hcd.c max3421_hcd->iopins[idx] |= mask; idx 1688 drivers/usb/host/max3421-hcd.c max3421_hcd->iopins[idx] &= ~mask; idx 35 drivers/usb/host/ohci-s3c2410.c #define valid_port(idx) ((idx) == 1 || (idx) == 2) idx 741 drivers/usb/host/pci-quirks.c static int mmio_resource_enabled(struct pci_dev *pdev, int idx) idx 743 drivers/usb/host/pci-quirks.c return pci_resource_start(pdev, idx) && mmio_enabled(pdev); idx 283 drivers/usb/host/r8a66597-hcd.c int idx; idx 285 drivers/usb/host/r8a66597-hcd.c idx = address / 32; idx 286 drivers/usb/host/r8a66597-hcd.c r8a66597->child_connect_map[idx] |= 1 << (address % 32); idx 291 drivers/usb/host/r8a66597-hcd.c int idx; idx 293 drivers/usb/host/r8a66597-hcd.c idx = address / 32; idx 294 drivers/usb/host/r8a66597-hcd.c r8a66597->child_connect_map[idx] &= ~(1 << (address % 32)); idx 2141 drivers/usb/host/xhci-ring.c int idx; idx 2151 drivers/usb/host/xhci-ring.c idx = urb_priv->num_tds_done; idx 2152 drivers/usb/host/xhci-ring.c frame = &td->urb->iso_frame_desc[idx]; idx 2226 drivers/usb/host/xhci-ring.c int idx; idx 2230 drivers/usb/host/xhci-ring.c idx = urb_priv->num_tds_done; idx 2231 drivers/usb/host/xhci-ring.c frame = &td->urb->iso_frame_desc[idx]; idx 1233 drivers/usb/misc/sisusbvga/sisusb.c int sisusb_setidxregandor(struct sisusb_usb_data *sisusb, u32 port, u8 idx, idx 1239 drivers/usb/misc/sisusbvga/sisusb.c ret = sisusb_write_memio_byte(sisusb, SISUSB_TYPE_IO, port, idx); idx 1248 drivers/usb/misc/sisusbvga/sisusb.c u32 port, u8 idx, u8 data, u8 mask) idx 1253 drivers/usb/misc/sisusbvga/sisusb.c ret = sisusb_write_memio_byte(sisusb, SISUSB_TYPE_IO, port, idx); idx 1268 drivers/usb/misc/sisusbvga/sisusb.c u8 idx, u8 myand) idx 1270 drivers/usb/misc/sisusbvga/sisusb.c return sisusb_setidxregandor(sisusb, port, idx, myand, 0x00); idx 1639 drivers/usb/misc/sisusbvga/sisusb.c int *iret, int rankno, int idx, int bw, const u8 rtype[][5]) idx 1647 drivers/usb/misc/sisusbvga/sisusb.c inc = 1 << (rtype[idx][2] + rtype[idx][1] + rtype[idx][0] + idx 1654 drivers/usb/misc/sisusbvga/sisusb.c inc = 1 << (rtype[idx][2] + bw / 64 + 2); idx 822 drivers/usb/misc/sisusbvga/sisusb_init.h u8 idx, u8 myand, u8 myor); idx 826 drivers/usb/misc/sisusbvga/sisusb_init.h u8 idx, u8 myand); idx 64 drivers/usb/musb/musbhsdma.c u8 idx; idx 117 drivers/usb/musb/musbhsdma.c musb_channel->idx = bit; idx 143 drivers/usb/musb/musbhsdma.c ~(1 << musb_channel->idx); idx 156 drivers/usb/musb/musbhsdma.c u8 bchannel = musb_channel->idx; idx 231 drivers/usb/musb/musbhsdma.c u8 bchannel = musb_channel->idx; idx 402 drivers/usb/serial/f81232.c int idx; idx 404 drivers/usb/serial/f81232.c for (idx = 0; idx < ARRAY_SIZE(baudrate_table); ++idx) { idx 405 drivers/usb/serial/f81232.c if (baudrate <= baudrate_table[idx] && idx 406 drivers/usb/serial/f81232.c baudrate_table[idx] % baudrate == 0) idx 407 drivers/usb/serial/f81232.c return idx; idx 422 drivers/usb/serial/f81232.c int idx; idx 426 drivers/usb/serial/f81232.c idx = f81232_find_clk(baud_list[i]); idx 427 drivers/usb/serial/f81232.c if (idx >= 0) { idx 434 drivers/usb/serial/f81232.c if (idx < 0) idx 437 drivers/usb/serial/f81232.c priv->baud_base = baudrate_table[idx]; idx 441 drivers/usb/serial/f81232.c F81232_CLK_MASK, clock_table[idx]); idx 550 drivers/usb/serial/f81534.c int idx; idx 552 drivers/usb/serial/f81534.c for (idx = 0; idx < ARRAY_SIZE(baudrate_table); ++idx) { idx 553 drivers/usb/serial/f81534.c if (baudrate <= baudrate_table[idx] && idx 554 drivers/usb/serial/f81534.c baudrate_table[idx] % baudrate == 0) idx 555 drivers/usb/serial/f81534.c return idx; idx 568 drivers/usb/serial/f81534.c int idx; idx 573 drivers/usb/serial/f81534.c idx = f81534_find_clk(baud_list[i]); idx 574 drivers/usb/serial/f81534.c if (idx >= 0) { idx 581 drivers/usb/serial/f81534.c if (idx < 0) idx 584 drivers/usb/serial/f81534.c port_priv->baud_base = baudrate_table[idx]; idx 586 drivers/usb/serial/f81534.c port_priv->shadow_clk |= clock_table[idx]; idx 1337 drivers/usb/serial/f81534.c u8 idx; idx 1343 drivers/usb/serial/f81534.c idx = F81534_CONF_INIT_GPIO_OFFSET + port_priv->phy_num; idx 1344 drivers/usb/serial/f81534.c value = serial_priv->conf_data[idx]; idx 1350 drivers/usb/serial/f81534.c idx = F81534_CONF_WORK_GPIO_OFFSET + port_priv->phy_num; idx 1351 drivers/usb/serial/f81534.c value = serial_priv->conf_data[idx]; idx 178 drivers/usb/serial/usb-serial.c int idx = tty->index; idx 184 drivers/usb/serial/usb-serial.c port = usb_serial_port_get_by_minor(idx); idx 196 drivers/usb/serial/usb-serial.c init_termios = (driver->termios[idx] == NULL); idx 120 drivers/usb/storage/realtek_cr.c #define SET_BIT(data, idx) ((data) |= 1 << (idx)) idx 121 drivers/usb/storage/realtek_cr.c #define CLR_BIT(data, idx) ((data) &= ~(1 << (idx))) idx 122 drivers/usb/storage/realtek_cr.c #define CHK_BIT(data, idx) ((data) & (1 << (idx))) idx 309 drivers/usb/storage/uas.c unsigned int idx; idx 324 drivers/usb/storage/uas.c idx = be16_to_cpup(&iu->tag) - 1; idx 325 drivers/usb/storage/uas.c if (idx >= MAX_CMNDS || !devinfo->cmnd[idx]) { idx 327 drivers/usb/storage/uas.c "stat urb: no pending cmd for uas-tag %d\n", idx + 1); idx 331 drivers/usb/storage/uas.c cmnd = devinfo->cmnd[idx]; idx 643 drivers/usb/storage/uas.c int idx, err; idx 670 drivers/usb/storage/uas.c for (idx = 0; idx < devinfo->qdepth; idx++) { idx 671 drivers/usb/storage/uas.c if (!devinfo->cmnd[idx]) idx 674 drivers/usb/storage/uas.c if (idx == devinfo->qdepth) { idx 682 drivers/usb/storage/uas.c cmdinfo->uas_tag = idx + 1; /* uas-tag == usb-stream-id, so 1 based */ idx 711 drivers/usb/storage/uas.c devinfo->cmnd[idx] = cmnd; idx 51 drivers/usb/usbip/stub_main.c int idx = -1; idx 57 drivers/usb/usbip/stub_main.c idx = i; idx 63 drivers/usb/usbip/stub_main.c return idx; idx 69 drivers/usb/usbip/stub_main.c int idx; idx 73 drivers/usb/usbip/stub_main.c idx = get_busid_idx(busid); idx 74 drivers/usb/usbip/stub_main.c if (idx >= 0) { idx 75 drivers/usb/usbip/stub_main.c bid = &(busid_table[idx]); idx 124 drivers/usb/usbip/stub_main.c int idx; idx 128 drivers/usb/usbip/stub_main.c idx = get_busid_idx(busid); idx 129 drivers/usb/usbip/stub_main.c if (idx < 0) idx 135 drivers/usb/usbip/stub_main.c spin_lock(&busid_table[idx].busid_lock); idx 137 drivers/usb/usbip/stub_main.c if (busid_table[idx].status == STUB_BUSID_OTHER) idx 138 drivers/usb/usbip/stub_main.c memset(busid_table[idx].name, 0, BUSID_SIZE); idx 140 drivers/usb/usbip/stub_main.c if ((busid_table[idx].status != STUB_BUSID_OTHER) && idx 141 drivers/usb/usbip/stub_main.c (busid_table[idx].status != STUB_BUSID_ADDED)) idx 142 drivers/usb/usbip/stub_main.c busid_table[idx].status = STUB_BUSID_REMOV; idx 144 drivers/usb/usbip/stub_main.c spin_unlock(&busid_table[idx].busid_lock); idx 244 drivers/vhost/scsi.c int idx, i; idx 252 drivers/vhost/scsi.c idx = vs->vqs[i].inflight_idx; idx 254 drivers/vhost/scsi.c old_inflight[i] = &vs->vqs[i].inflights[idx]; idx 257 drivers/vhost/scsi.c vs->vqs[i].inflight_idx = idx ^ 1; idx 258 drivers/vhost/scsi.c new_inflight = &vs->vqs[i].inflights[idx ^ 1]; idx 94 drivers/vhost/vhost.c static long vhost_get_vring_endian(struct vhost_virtqueue *vq, u32 idx, idx 98 drivers/vhost/vhost.c .index = idx, idx 127 drivers/vhost/vhost.c static long vhost_get_vring_endian(struct vhost_virtqueue *vq, u32 idx, idx 921 drivers/vhost/vhost.c struct vring_used_elem *head, int idx, idx 924 drivers/vhost/vhost.c return vhost_copy_to_user(vq, vq->used->ring + idx, head, idx 939 drivers/vhost/vhost.c &vq->used->idx); idx 981 drivers/vhost/vhost.c __virtio16 *idx) idx 983 drivers/vhost/vhost.c return vhost_get_avail(vq, *idx, &vq->avail->idx); idx 987 drivers/vhost/vhost.c __virtio16 *head, int idx) idx 990 drivers/vhost/vhost.c &vq->avail->ring[idx & (vq->num - 1)]); idx 1006 drivers/vhost/vhost.c __virtio16 *idx) idx 1008 drivers/vhost/vhost.c return vhost_get_used(vq, *idx, &vq->used->idx); idx 1012 drivers/vhost/vhost.c struct vring_desc *desc, int idx) idx 1014 drivers/vhost/vhost.c return vhost_copy_from_user(vq, desc, vq->desc + idx, sizeof(*desc)); idx 1594 drivers/vhost/vhost.c u32 idx; idx 1597 drivers/vhost/vhost.c r = get_user(idx, idxp); idx 1600 drivers/vhost/vhost.c if (idx >= d->nvqs) idx 1603 drivers/vhost/vhost.c idx = array_index_nospec(idx, d->nvqs); idx 1604 drivers/vhost/vhost.c vq = d->vqs[idx]; idx 1634 drivers/vhost/vhost.c s.index = idx; idx 1683 drivers/vhost/vhost.c r = vhost_get_vring_endian(vq, idx, argp); idx 1693 drivers/vhost/vhost.c s.index = idx; idx 2021 drivers/vhost/vhost.c !access_ok(&vq->used->idx, sizeof vq->used->idx)) { idx 2028 drivers/vhost/vhost.c &vq->used->idx); idx 2226 drivers/vhost/vhost.c &vq->avail->idx); idx 2429 drivers/vhost/vhost.c log_used(vq, offsetof(struct vring_used, idx), idx 2430 drivers/vhost/vhost.c sizeof vq->used->idx); idx 2552 drivers/vhost/vhost.c &vq->avail->idx, r); idx 41 drivers/vhost/vringh.c err = getu16(vrh, &avail_idx, &vrh->vring.avail->idx); idx 44 drivers/vhost/vringh.c &vrh->vring.avail->idx); idx 439 drivers/vhost/vringh.c err = putu16(vrh, &vrh->vring.used->idx, used_idx + num_used); idx 442 drivers/vhost/vringh.c &vrh->vring.used->idx); idx 526 drivers/vhost/vringh.c if (getu16(vrh, &avail, &vrh->vring.avail->idx) != 0) { idx 528 drivers/vhost/vringh.c &vrh->vring.avail->idx); idx 232 drivers/video/backlight/pm8941-wled.c static u32 pm8941_wled_num_strings_values_fn(u32 idx) idx 234 drivers/video/backlight/pm8941-wled.c return idx + 1; idx 242 drivers/video/backlight/pm8941-wled.c static u32 pm8941_wled_switch_freq_values_fn(u32 idx) idx 244 drivers/video/backlight/pm8941-wled.c return 19200 / (2 * (1 + idx)); idx 256 drivers/video/backlight/pm8941-wled.c static u32 pm8941_wled_values(const struct pm8941_wled_var_cfg *cfg, u32 idx) idx 258 drivers/video/backlight/pm8941-wled.c if (idx >= cfg->size) idx 261 drivers/video/backlight/pm8941-wled.c return cfg->fn(idx); idx 263 drivers/video/backlight/pm8941-wled.c return cfg->values[idx]; idx 264 drivers/video/backlight/pm8941-wled.c return idx; idx 1096 drivers/video/fbdev/au1200fb.c int idx; idx 1097 drivers/video/fbdev/au1200fb.c idx = (win->w[0].mode_winctrl1 & LCD_WINCTRL1_FRM) >> 25; idx 1098 drivers/video/fbdev/au1200fb.c var->red = rgb_bitfields[idx][0]; idx 1099 drivers/video/fbdev/au1200fb.c var->green = rgb_bitfields[idx][1]; idx 1100 drivers/video/fbdev/au1200fb.c var->blue = rgb_bitfields[idx][2]; idx 1101 drivers/video/fbdev/au1200fb.c var->transp = rgb_bitfields[idx][3]; idx 1109 drivers/video/fbdev/au1200fb.c int idx; idx 1110 drivers/video/fbdev/au1200fb.c idx = (win->w[0].mode_winctrl1 & LCD_WINCTRL1_FRM) >> 25; idx 1111 drivers/video/fbdev/au1200fb.c var->red = rgb_bitfields[idx][0]; idx 1112 drivers/video/fbdev/au1200fb.c var->green = rgb_bitfields[idx][1]; idx 1113 drivers/video/fbdev/au1200fb.c var->blue = rgb_bitfields[idx][2]; idx 1114 drivers/video/fbdev/au1200fb.c var->transp = rgb_bitfields[idx][3]; idx 83 drivers/video/fbdev/core/bitblit.c u32 idx = vc->vc_font.width >> 3; idx 95 drivers/video/fbdev/core/bitblit.c if (likely(idx == 1)) idx 96 drivers/video/fbdev/core/bitblit.c __fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 99 drivers/video/fbdev/core/bitblit.c fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 118 drivers/video/fbdev/core/bitblit.c u32 idx = vc->vc_font.width >> 3; idx 130 drivers/video/fbdev/core/bitblit.c fb_pad_unaligned_buffer(dst, d_pitch, src, idx, idx 546 drivers/video/fbdev/core/fbcon.c static int search_fb_in_map(int idx) idx 551 drivers/video/fbdev/core/fbcon.c if (con2fb_map[i] == idx) idx 1243 drivers/video/fbdev/core/fbcon.c int idx; idx 1246 drivers/video/fbdev/core/fbcon.c idx = con2fb_map[vc->vc_num]; idx 1248 drivers/video/fbdev/core/fbcon.c if (idx == -1) idx 1251 drivers/video/fbdev/core/fbcon.c info = registered_fb[idx]; idx 3090 drivers/video/fbdev/core/fbcon.c int idx = info->node; idx 3098 drivers/video/fbdev/core/fbcon.c if (con2fb_map[i] != idx && idx 3107 drivers/video/fbdev/core/fbcon.c if (con2fb_map[i] == idx) idx 3111 drivers/video/fbdev/core/fbcon.c struct fb_info *info = registered_fb[idx]; idx 3120 drivers/video/fbdev/core/fbcon.c if (con2fb_map[i] == idx) { idx 3122 drivers/video/fbdev/core/fbcon.c if (!search_fb_in_map(idx)) { idx 3125 drivers/video/fbdev/core/fbcon.c idx, 0); idx 3127 drivers/video/fbdev/core/fbcon.c con2fb_map[i] = idx; idx 3140 drivers/video/fbdev/core/fbcon.c int i, idx; idx 3147 drivers/video/fbdev/core/fbcon.c idx = info->node; idx 3149 drivers/video/fbdev/core/fbcon.c if (con2fb_map[i] == idx) idx 3153 drivers/video/fbdev/core/fbcon.c if (idx == info_idx) { idx 3169 drivers/video/fbdev/core/fbcon.c if (primary_device == idx) idx 3178 drivers/video/fbdev/core/fbcon.c int i, idx = info->node; idx 3183 drivers/video/fbdev/core/fbcon.c con2fb_map_boot[i] = idx; idx 3190 drivers/video/fbdev/core/fbcon.c set_con2fb_map(i, idx, 0); idx 3194 drivers/video/fbdev/core/fbcon.c "fb%i, to tty %i-%i\n", idx, idx 3196 drivers/video/fbdev/core/fbcon.c info_idx = idx; idx 3234 drivers/video/fbdev/core/fbcon.c int ret = 0, i, idx; idx 3238 drivers/video/fbdev/core/fbcon.c idx = info->node; idx 3248 drivers/video/fbdev/core/fbcon.c if (con2fb_map_boot[i] == idx) { idx 3249 drivers/video/fbdev/core/fbcon.c info_idx = idx; idx 3258 drivers/video/fbdev/core/fbcon.c if (con2fb_map_boot[i] == idx) idx 3259 drivers/video/fbdev/core/fbcon.c set_con2fb_map(i, idx, 0); idx 3422 drivers/video/fbdev/core/fbcon.c int rotate, idx; idx 3426 drivers/video/fbdev/core/fbcon.c idx = con2fb_map[fg_console]; idx 3428 drivers/video/fbdev/core/fbcon.c if (idx == -1 || registered_fb[idx] == NULL) idx 3431 drivers/video/fbdev/core/fbcon.c info = registered_fb[idx]; idx 3444 drivers/video/fbdev/core/fbcon.c int rotate, idx; idx 3448 drivers/video/fbdev/core/fbcon.c idx = con2fb_map[fg_console]; idx 3450 drivers/video/fbdev/core/fbcon.c if (idx == -1 || registered_fb[idx] == NULL) idx 3453 drivers/video/fbdev/core/fbcon.c info = registered_fb[idx]; idx 3465 drivers/video/fbdev/core/fbcon.c int rotate = 0, idx; idx 3468 drivers/video/fbdev/core/fbcon.c idx = con2fb_map[fg_console]; idx 3470 drivers/video/fbdev/core/fbcon.c if (idx == -1 || registered_fb[idx] == NULL) idx 3473 drivers/video/fbdev/core/fbcon.c info = registered_fb[idx]; idx 3485 drivers/video/fbdev/core/fbcon.c int idx, blink = -1; idx 3488 drivers/video/fbdev/core/fbcon.c idx = con2fb_map[fg_console]; idx 3490 drivers/video/fbdev/core/fbcon.c if (idx == -1 || registered_fb[idx] == NULL) idx 3493 drivers/video/fbdev/core/fbcon.c info = registered_fb[idx]; idx 3510 drivers/video/fbdev/core/fbcon.c int blink, idx; idx 3514 drivers/video/fbdev/core/fbcon.c idx = con2fb_map[fg_console]; idx 3516 drivers/video/fbdev/core/fbcon.c if (idx == -1 || registered_fb[idx] == NULL) idx 3519 drivers/video/fbdev/core/fbcon.c info = registered_fb[idx]; idx 105 drivers/video/fbdev/core/fbcon_ccw.c u32 idx = (vc->vc_font.height + 7) >> 3; idx 116 drivers/video/fbdev/core/fbcon_ccw.c if (likely(idx == 1)) idx 117 drivers/video/fbdev/core/fbcon_ccw.c __fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 120 drivers/video/fbdev/core/fbcon_ccw.c fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 90 drivers/video/fbdev/core/fbcon_cw.c u32 idx = (vc->vc_font.height + 7) >> 3; idx 101 drivers/video/fbdev/core/fbcon_cw.c if (likely(idx == 1)) idx 102 drivers/video/fbdev/core/fbcon_cw.c __fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 105 drivers/video/fbdev/core/fbcon_cw.c fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 92 drivers/video/fbdev/core/fbcon_ud.c u32 idx = vc->vc_font.width >> 3; idx 103 drivers/video/fbdev/core/fbcon_ud.c if (likely(idx == 1)) idx 104 drivers/video/fbdev/core/fbcon_ud.c __fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 107 drivers/video/fbdev/core/fbcon_ud.c fb_pad_aligned_buffer(dst, d_pitch, src, idx, idx 127 drivers/video/fbdev/core/fbcon_ud.c u32 idx = vc->vc_font.width >> 3; idx 138 drivers/video/fbdev/core/fbcon_ud.c fb_pad_unaligned_buffer(dst, d_pitch, src, idx, idx 59 drivers/video/fbdev/core/fbmem.c static struct fb_info *get_fb_info(unsigned int idx) idx 63 drivers/video/fbdev/core/fbmem.c if (idx >= FB_MAX) idx 67 drivers/video/fbdev/core/fbmem.c fb_info = registered_fb[idx]; idx 119 drivers/video/fbdev/core/fbmem.c void fb_pad_unaligned_buffer(u8 *dst, u32 d_pitch, u8 *src, u32 idx, u32 height, idx 126 drivers/video/fbdev/core/fbmem.c for (j = 0; j < idx; j++) { idx 135 drivers/video/fbdev/core/fbmem.c tmp = dst[idx]; idx 138 drivers/video/fbdev/core/fbmem.c dst[idx] = tmp; idx 141 drivers/video/fbdev/core/fbmem.c dst[idx+1] = tmp; idx 1774 drivers/video/fbdev/core/fbmem.c int err, idx, bar; idx 1777 drivers/video/fbdev/core/fbmem.c for (idx = 0, bar = 0; bar < PCI_ROM_RESOURCE; bar++) { idx 1780 drivers/video/fbdev/core/fbmem.c idx++; idx 1783 drivers/video/fbdev/core/fbmem.c ap = alloc_apertures(idx); idx 1787 drivers/video/fbdev/core/fbmem.c for (idx = 0, bar = 0; bar < PCI_ROM_RESOURCE; bar++) { idx 1790 drivers/video/fbdev/core/fbmem.c ap->ranges[idx].base = pci_resource_start(pdev, bar); idx 1791 drivers/video/fbdev/core/fbmem.c ap->ranges[idx].size = pci_resource_len(pdev, bar); idx 1795 drivers/video/fbdev/core/fbmem.c idx++; idx 1123 drivers/video/fbdev/cyber2000fb.c int cyber2000fb_attach(struct cyberpro_info *info, int idx) idx 1149 drivers/video/fbdev/cyber2000fb.c void cyber2000fb_detach(int idx) idx 491 drivers/video/fbdev/cyber2000fb.h int cyber2000fb_attach(struct cyberpro_info *info, int idx); idx 492 drivers/video/fbdev/cyber2000fb.h void cyber2000fb_detach(int idx); idx 620 drivers/video/fbdev/efifb.c static void record_efifb_bar_resource(struct pci_dev *dev, int idx, u64 offset) idx 631 drivers/video/fbdev/efifb.c idx); idx 635 drivers/video/fbdev/efifb.c bar_resource = &dev->resource[idx]; idx 638 drivers/video/fbdev/efifb.c dev_info(&dev->dev, "BAR %d: assigned to efifb\n", idx); idx 1271 drivers/video/fbdev/imsttfb.c __u8 idx[2]; idx 1299 drivers/video/fbdev/imsttfb.c if (copy_from_user(idx, argp, 2)) idx 1302 drivers/video/fbdev/imsttfb.c par->cmap_regs[PIDXLO] = idx[0]; eieio(); idx 1303 drivers/video/fbdev/imsttfb.c par->cmap_regs[PIDXDATA] = idx[1]; eieio(); idx 1306 drivers/video/fbdev/imsttfb.c if (copy_from_user(idx, argp, 1)) idx 1309 drivers/video/fbdev/imsttfb.c par->cmap_regs[PIDXLO] = idx[0]; eieio(); idx 1310 drivers/video/fbdev/imsttfb.c idx[1] = par->cmap_regs[PIDXDATA]; idx 1311 drivers/video/fbdev/imsttfb.c if (copy_to_user((void __user *)(arg + 1), &idx[1], 1)) idx 259 drivers/video/fbdev/matrox/g450_pll.c unsigned int idx; idx 262 drivers/video/fbdev/matrox/g450_pll.c for (idx = 0; idx < mnpcount; idx++) { idx 269 drivers/video/fbdev/matrox/g450_pll.c mnp = mnparray[idx]; idx 439 drivers/video/fbdev/matrox/g450_pll.c unsigned int idx; idx 464 drivers/video/fbdev/matrox/g450_pll.c for (idx = mnpcount; idx > 0; idx--) { idx 469 drivers/video/fbdev/matrox/g450_pll.c if (delta <= deltaarray[idx-1]) { idx 474 drivers/video/fbdev/matrox/g450_pll.c if (delta == deltaarray[idx-1] idx 475 drivers/video/fbdev/matrox/g450_pll.c && vco != g450_mnp2vco(minfo, mnparray[idx-1]) idx 479 drivers/video/fbdev/matrox/g450_pll.c mnparray[idx] = mnparray[idx-1]; idx 480 drivers/video/fbdev/matrox/g450_pll.c deltaarray[idx] = deltaarray[idx-1]; idx 485 drivers/video/fbdev/matrox/g450_pll.c mnparray[idx] = mnp; idx 486 drivers/video/fbdev/matrox/g450_pll.c deltaarray[idx] = delta; idx 671 drivers/video/fbdev/matrox/matroxfb_base.h #define mga_readr(port,idx) (mga_outb((port),(idx)), mga_inb((port)+1)) idx 84 drivers/video/fbdev/matrox/matroxfb_g450.c static inline int *get_ctrl_ptr(struct matrox_fb_info *minfo, unsigned int idx) idx 86 drivers/video/fbdev/matrox/matroxfb_g450.c return (int*)((char*)minfo + g450_controls[idx].control); idx 135 drivers/video/fbdev/matrox/matroxfb_maven.c static int* get_ctrl_ptr(struct maven_data* md, int idx) { idx 136 drivers/video/fbdev/matrox/matroxfb_maven.c return (int*)((char*)(md->primary_head) + maven_controls[idx].control); idx 189 drivers/video/fbdev/omap/omapfb.h int idx; idx 192 drivers/video/fbdev/omap/omapfb_main.c r = fbdev->ctrl->setup_plane(plane->idx, plane->info.channel_out, idx 206 drivers/video/fbdev/omap/omapfb_main.c r = fbdev->ctrl->set_scale(plane->idx, idx 378 drivers/video/fbdev/omap/omapfb_main.c rg = &plane->fbdev->mem_desc.region[plane->idx]; idx 525 drivers/video/fbdev/omap/omapfb_main.c max_frame_size = fbdev->mem_desc.region[plane->idx].size; idx 777 drivers/video/fbdev/omap/omapfb_main.c if (pi->enabled && !fbdev->mem_desc.region[plane->idx].size) { idx 794 drivers/video/fbdev/omap/omapfb_main.c r = fbdev->ctrl->enable_plane(plane->idx, pi->enabled); idx 816 drivers/video/fbdev/omap/omapfb_main.c struct omapfb_mem_region *rg = &fbdev->mem_desc.region[plane->idx]; idx 856 drivers/video/fbdev/omap/omapfb_main.c r = fbdev->ctrl->setup_mem(plane->idx, size, mi->type, &paddr); idx 894 drivers/video/fbdev/omap/omapfb_main.c rg = &fbdev->mem_desc.region[plane->idx]; idx 1158 drivers/video/fbdev/omap/omapfb_main.c omapfb_get_caps(fbdev, plane->idx, &p.caps); idx 1510 drivers/video/fbdev/omap/omapfb_main.c plane->idx = i; idx 48 drivers/video/fbdev/omap2/omapfb/dss/dispc.c #define REG_GET(idx, start, end) \ idx 49 drivers/video/fbdev/omap2/omapfb/dss/dispc.c FLD_GET(dispc_read_reg(idx), start, end) idx 51 drivers/video/fbdev/omap2/omapfb/dss/dispc.c #define REG_FLD_MOD(idx, val, start, end) \ idx 52 drivers/video/fbdev/omap2/omapfb/dss/dispc.c dispc_write_reg(idx, FLD_MOD(dispc_read_reg(idx), val, start, end)) idx 251 drivers/video/fbdev/omap2/omapfb/dss/dispc.c static inline void dispc_write_reg(const u16 idx, u32 val) idx 253 drivers/video/fbdev/omap2/omapfb/dss/dispc.c __raw_writel(val, dispc.base + idx); idx 256 drivers/video/fbdev/omap2/omapfb/dss/dispc.c static inline u32 dispc_read_reg(const u16 idx) idx 258 drivers/video/fbdev/omap2/omapfb/dss/dispc.c return __raw_readl(dispc.base + idx); idx 1494 drivers/video/fbdev/omap2/omapfb/dss/dispc.c int idx; idx 1529 drivers/video/fbdev/omap2/omapfb/dss/dispc.c idx = 0; idx 1532 drivers/video/fbdev/omap2/omapfb/dss/dispc.c idx = 1; idx 1535 drivers/video/fbdev/omap2/omapfb/dss/dispc.c idx = 2; idx 1538 drivers/video/fbdev/omap2/omapfb/dss/dispc.c idx = 3; idx 1561 drivers/video/fbdev/omap2/omapfb/dss/dispc.c accu_val = &accu_table[idx]; idx 42 drivers/video/fbdev/omap2/omapfb/dss/dsi.c struct dsi_reg { u16 module; u16 idx; }; idx 44 drivers/video/fbdev/omap2/omapfb/dss/dsi.c #define DSI_REG(mod, idx) ((const struct dsi_reg) { mod, idx }) idx 110 drivers/video/fbdev/omap2/omapfb/dss/dsi.c #define REG_GET(dsidev, idx, start, end) \ idx 111 drivers/video/fbdev/omap2/omapfb/dss/dsi.c FLD_GET(dsi_read_reg(dsidev, idx), start, end) idx 113 drivers/video/fbdev/omap2/omapfb/dss/dsi.c #define REG_FLD_MOD(dsidev, idx, val, start, end) \ idx 114 drivers/video/fbdev/omap2/omapfb/dss/dsi.c dsi_write_reg(dsidev, idx, FLD_MOD(dsi_read_reg(dsidev, idx), val, start, end)) idx 432 drivers/video/fbdev/omap2/omapfb/dss/dsi.c const struct dsi_reg idx, u32 val) idx 437 drivers/video/fbdev/omap2/omapfb/dss/dsi.c switch(idx.module) { idx 444 drivers/video/fbdev/omap2/omapfb/dss/dsi.c __raw_writel(val, base + idx.idx); idx 448 drivers/video/fbdev/omap2/omapfb/dss/dsi.c const struct dsi_reg idx) idx 453 drivers/video/fbdev/omap2/omapfb/dss/dsi.c switch(idx.module) { idx 460 drivers/video/fbdev/omap2/omapfb/dss/dsi.c return __raw_readl(base + idx.idx); idx 492 drivers/video/fbdev/omap2/omapfb/dss/dsi.c const struct dsi_reg idx, int bitnum, int value) idx 501 drivers/video/fbdev/omap2/omapfb/dss/dsi.c if (REG_GET(dsidev, idx, bitnum, bitnum) == value) idx 508 drivers/video/fbdev/omap2/omapfb/dss/dsi.c if (REG_GET(dsidev, idx, bitnum, bitnum) == value) idx 42 drivers/video/fbdev/omap2/omapfb/dss/dss.c u16 idx; idx 45 drivers/video/fbdev/omap2/omapfb/dss/dss.c #define DSS_REG(idx) ((const struct dss_reg) { idx }) idx 55 drivers/video/fbdev/omap2/omapfb/dss/dss.c #define REG_GET(idx, start, end) \ idx 56 drivers/video/fbdev/omap2/omapfb/dss/dss.c FLD_GET(dss_read_reg(idx), start, end) idx 58 drivers/video/fbdev/omap2/omapfb/dss/dss.c #define REG_FLD_MOD(idx, val, start, end) \ idx 59 drivers/video/fbdev/omap2/omapfb/dss/dss.c dss_write_reg(idx, FLD_MOD(dss_read_reg(idx), val, start, end)) idx 113 drivers/video/fbdev/omap2/omapfb/dss/dss.c static inline void dss_write_reg(const struct dss_reg idx, u32 val) idx 115 drivers/video/fbdev/omap2/omapfb/dss/dss.c __raw_writel(val, dss.base + idx.idx); idx 118 drivers/video/fbdev/omap2/omapfb/dss/dss.c static inline u32 dss_read_reg(const struct dss_reg idx) idx 120 drivers/video/fbdev/omap2/omapfb/dss/dss.c return __raw_readl(dss.base + idx.idx); idx 124 drivers/video/fbdev/omap2/omapfb/dss/dss.c dss.ctx[(DSS_##reg).idx / sizeof(u32)] = dss_read_reg(DSS_##reg) idx 126 drivers/video/fbdev/omap2/omapfb/dss/dss.c dss_write_reg(DSS_##reg, dss.ctx[(DSS_##reg).idx / sizeof(u32)]) idx 247 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h static inline void hdmi_write_reg(void __iomem *base_addr, const u32 idx, idx 250 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h __raw_writel(val, base_addr + idx); idx 253 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h static inline u32 hdmi_read_reg(void __iomem *base_addr, const u32 idx) idx 255 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h return __raw_readl(base_addr + idx); idx 258 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h #define REG_FLD_MOD(base, idx, val, start, end) \ idx 259 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h hdmi_write_reg(base, idx, FLD_MOD(hdmi_read_reg(base, idx),\ idx 261 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h #define REG_GET(base, idx, start, end) \ idx 262 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h FLD_GET(hdmi_read_reg(base, idx), start, end) idx 265 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h const u32 idx, int b2, int b1, u32 val) idx 268 drivers/video/fbdev/omap2/omapfb/dss/hdmi.h while (val != (v = REG_GET(base_addr, idx, b2, b1))) { idx 301 drivers/video/fbdev/omap2/omapfb/dss/venc.c static inline void venc_write_reg(int idx, u32 val) idx 303 drivers/video/fbdev/omap2/omapfb/dss/venc.c __raw_writel(val, venc.base + idx); idx 306 drivers/video/fbdev/omap2/omapfb/dss/venc.c static inline u32 venc_read_reg(int idx) idx 308 drivers/video/fbdev/omap2/omapfb/dss/venc.c u32 l = __raw_readl(venc.base + idx); idx 154 drivers/video/fbdev/pm2fb.c static inline u32 pm2_RDAC_RD(struct pm2fb_par *p, s32 idx) idx 156 drivers/video/fbdev/pm2fb.c pm2_WR(p, PM2R_RD_PALETTE_WRITE_ADDRESS, idx); idx 161 drivers/video/fbdev/pm2fb.c static inline u32 pm2v_RDAC_RD(struct pm2fb_par *p, s32 idx) idx 163 drivers/video/fbdev/pm2fb.c pm2_WR(p, PM2VR_RD_INDEX_LOW, idx & 0xff); idx 168 drivers/video/fbdev/pm2fb.c static inline void pm2_RDAC_WR(struct pm2fb_par *p, s32 idx, u32 v) idx 170 drivers/video/fbdev/pm2fb.c pm2_WR(p, PM2R_RD_PALETTE_WRITE_ADDRESS, idx); idx 176 drivers/video/fbdev/pm2fb.c static inline void pm2v_RDAC_WR(struct pm2fb_par *p, s32 idx, u32 v) idx 178 drivers/video/fbdev/pm2fb.c pm2_WR(p, PM2VR_RD_INDEX_LOW, idx & 0xff); idx 465 drivers/video/fbdev/sis/init301.c unsigned short romindex = 0, reg = 0, idx = 0; idx 480 drivers/video/fbdev/sis/init301.c idx = (SiS_GetReg(SiS_Pr->SiS_P3d4,reg) & 0x1f) * 26; idx 482 drivers/video/fbdev/sis/init301.c if(idx < (8*26)) { idx 483 drivers/video/fbdev/sis/init301.c myptr = (unsigned char *)&SiS_LCDStruct661[idx]; idx 487 drivers/video/fbdev/sis/init301.c romindex += idx; idx 2153 drivers/video/fbdev/sis/init301.c int idx; idx 2156 drivers/video/fbdev/sis/init301.c if(SiS_Pr->ChipType < SIS_315H) idx = VCLK_CUSTOM_300; idx 2157 drivers/video/fbdev/sis/init301.c else idx = VCLK_CUSTOM_315; idx 2158 drivers/video/fbdev/sis/init301.c SiS_Pr->SiS_VCLKData[idx].CLOCK = idx 2159 drivers/video/fbdev/sis/init301.c SiS_Pr->SiS_VBVCLKData[idx].CLOCK = SiS_Pr->CP_PrefClock; idx 2160 drivers/video/fbdev/sis/init301.c SiS_Pr->SiS_VCLKData[idx].SR2B = idx 2161 drivers/video/fbdev/sis/init301.c SiS_Pr->SiS_VBVCLKData[idx].Part4_A = SiS_Pr->CP_PrefSR2B; idx 2162 drivers/video/fbdev/sis/init301.c SiS_Pr->SiS_VCLKData[idx].SR2C = idx 2163 drivers/video/fbdev/sis/init301.c SiS_Pr->SiS_VBVCLKData[idx].Part4_B = SiS_Pr->CP_PrefSR2C; idx 724 drivers/video/fbdev/sis/sis_main.c while((sisfb_vrate[i].idx != 0) && (sisfb_vrate[i].xres <= xres)) { idx 727 drivers/video/fbdev/sis/sis_main.c ivideo->rate_idx = sisfb_vrate[i].idx; idx 733 drivers/video/fbdev/sis/sis_main.c ivideo->rate_idx = sisfb_vrate[i].idx; idx 735 drivers/video/fbdev/sis/sis_main.c } else if((sisfb_vrate[i].idx != 1) && idx 739 drivers/video/fbdev/sis/sis_main.c ivideo->rate_idx = sisfb_vrate[i-1].idx; idx 746 drivers/video/fbdev/sis/sis_main.c ivideo->rate_idx = sisfb_vrate[i].idx; idx 852 drivers/video/fbdev/sis/sis_main.c u8 idx, reg1, reg2, reg3, reg4; idx 865 drivers/video/fbdev/sis/sis_main.c case SIS_300_VGA: idx = 0x25; break; idx 867 drivers/video/fbdev/sis/sis_main.c case SIS_315_VGA: idx = 0x30; break; idx 869 drivers/video/fbdev/sis/sis_main.c reg1 = SiS_GetReg(SISPART1, (idx+0)); /* 30 */ idx 870 drivers/video/fbdev/sis/sis_main.c reg2 = SiS_GetReg(SISPART1, (idx+1)); /* 31 */ idx 871 drivers/video/fbdev/sis/sis_main.c reg3 = SiS_GetReg(SISPART1, (idx+2)); /* 32 */ idx 872 drivers/video/fbdev/sis/sis_main.c reg4 = SiS_GetReg(SISPART1, (idx+3)); /* 33 */ idx 397 drivers/video/fbdev/sis/sis_main.h u16 idx; idx 169 drivers/video/fbdev/tdfxfb.c static inline void gra_outb(struct tdfx_par *par, u32 idx, u8 val) idx 171 drivers/video/fbdev/tdfxfb.c vga_outb(par, GRA_I, idx); idx 177 drivers/video/fbdev/tdfxfb.c static inline void seq_outb(struct tdfx_par *par, u32 idx, u8 val) idx 179 drivers/video/fbdev/tdfxfb.c vga_outb(par, SEQ_I, idx); idx 185 drivers/video/fbdev/tdfxfb.c static inline u8 seq_inb(struct tdfx_par *par, u32 idx) idx 187 drivers/video/fbdev/tdfxfb.c vga_outb(par, SEQ_I, idx); idx 192 drivers/video/fbdev/tdfxfb.c static inline void crt_outb(struct tdfx_par *par, u32 idx, u8 val) idx 194 drivers/video/fbdev/tdfxfb.c vga_outb(par, CRT_I, idx); idx 200 drivers/video/fbdev/tdfxfb.c static inline u8 crt_inb(struct tdfx_par *par, u32 idx) idx 202 drivers/video/fbdev/tdfxfb.c vga_outb(par, CRT_I, idx); idx 207 drivers/video/fbdev/tdfxfb.c static inline void att_outb(struct tdfx_par *par, u32 idx, u8 val) idx 212 drivers/video/fbdev/tdfxfb.c vga_outb(par, ATT_IW, idx); idx 34 drivers/video/fbdev/uvesafb.c .idx = CN_IDX_V86D, idx 271 drivers/virtio/virtio_balloon.c static inline void update_stat(struct virtio_balloon *vb, int idx, idx 274 drivers/virtio/virtio_balloon.c BUG_ON(idx >= VIRTIO_BALLOON_S_NR); idx 275 drivers/virtio/virtio_balloon.c vb->stats[idx].tag = cpu_to_virtio16(vb->vdev, tag); idx 276 drivers/virtio/virtio_balloon.c vb->stats[idx].val = cpu_to_virtio64(vb->vdev, val); idx 285 drivers/virtio/virtio_balloon.c unsigned int idx = 0; idx 296 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_SWAP_IN, idx 298 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_SWAP_OUT, idx 300 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_MAJFLT, events[PGMAJFAULT]); idx 301 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_MINFLT, events[PGFAULT]); idx 303 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_HTLB_PGALLOC, idx 305 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_HTLB_PGFAIL, idx 309 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_MEMFREE, idx 311 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_MEMTOT, idx 313 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_AVAIL, idx 315 drivers/virtio/virtio_balloon.c update_stat(vb, idx++, VIRTIO_BALLOON_S_CACHES, idx 318 drivers/virtio/virtio_balloon.c return idx; idx 99 drivers/virtio/virtio_pci_common.h unsigned idx, idx 553 drivers/virtio/virtio_ring.c vq->split.vring.avail->idx = cpu_to_virtio16(_vq->vdev, idx 675 drivers/virtio/virtio_ring.c vq->split.vring.used->idx); idx 779 drivers/virtio/virtio_ring.c vq->split.vring.used->idx); idx 808 drivers/virtio/virtio_ring.c if (unlikely((u16)(virtio16_to_cpu(_vq->vdev, vq->split.vring.used->idx) idx 833 drivers/virtio/virtio_ring.c vq->split.vring.avail->idx = cpu_to_virtio16(_vq->vdev, idx 1328 drivers/virtio/virtio_ring.c u16 idx, bool used_wrap_counter) idx 1333 drivers/virtio/virtio_ring.c flags = le16_to_cpu(vq->packed.vring.desc[idx].flags); idx 451 drivers/w1/masters/ds2482.c int idx; idx 495 drivers/w1/masters/ds2482.c for (idx = 0; idx < data->w1_count; idx++) { idx 496 drivers/w1/masters/ds2482.c data->w1_ch[idx].pdev = data; idx 497 drivers/w1/masters/ds2482.c data->w1_ch[idx].channel = idx; idx 500 drivers/w1/masters/ds2482.c data->w1_ch[idx].w1_bm.data = &data->w1_ch[idx]; idx 501 drivers/w1/masters/ds2482.c data->w1_ch[idx].w1_bm.read_byte = ds2482_w1_read_byte; idx 502 drivers/w1/masters/ds2482.c data->w1_ch[idx].w1_bm.write_byte = ds2482_w1_write_byte; idx 503 drivers/w1/masters/ds2482.c data->w1_ch[idx].w1_bm.touch_bit = ds2482_w1_touch_bit; idx 504 drivers/w1/masters/ds2482.c data->w1_ch[idx].w1_bm.triplet = ds2482_w1_triplet; idx 505 drivers/w1/masters/ds2482.c data->w1_ch[idx].w1_bm.reset_bus = ds2482_w1_reset_bus; idx 506 drivers/w1/masters/ds2482.c data->w1_ch[idx].w1_bm.set_pullup = ds2482_w1_set_pullup; idx 508 drivers/w1/masters/ds2482.c err = w1_add_master_device(&data->w1_ch[idx].w1_bm); idx 510 drivers/w1/masters/ds2482.c data->w1_ch[idx].pdev = NULL; idx 518 drivers/w1/masters/ds2482.c for (idx = 0; idx < data->w1_count; idx++) { idx 519 drivers/w1/masters/ds2482.c if (data->w1_ch[idx].pdev != NULL) idx 520 drivers/w1/masters/ds2482.c w1_remove_master_device(&data->w1_ch[idx].w1_bm); idx 531 drivers/w1/masters/ds2482.c int idx; idx 534 drivers/w1/masters/ds2482.c for (idx = 0; idx < data->w1_count; idx++) { idx 535 drivers/w1/masters/ds2482.c if (data->w1_ch[idx].pdev != NULL) idx 536 drivers/w1/masters/ds2482.c w1_remove_master_device(&data->w1_ch[idx].w1_bm); idx 208 drivers/w1/slaves/w1_ds2433.c int addr, len, idx; idx 222 drivers/w1/slaves/w1_ds2433.c for (idx = 0; idx < count; idx += W1_PAGE_SIZE) { idx 223 drivers/w1/slaves/w1_ds2433.c if (crc16(CRC16_INIT, &buf[idx], W1_PAGE_SIZE) != CRC16_VALID) { idx 233 drivers/w1/slaves/w1_ds2433.c idx = 0; idx 234 drivers/w1/slaves/w1_ds2433.c while (idx < count) { idx 235 drivers/w1/slaves/w1_ds2433.c addr = off + idx; idx 237 drivers/w1/slaves/w1_ds2433.c if (len > (count - idx)) idx 238 drivers/w1/slaves/w1_ds2433.c len = count - idx; idx 240 drivers/w1/slaves/w1_ds2433.c if (w1_f23_write(sl, addr, len, &buf[idx]) < 0) { idx 244 drivers/w1/slaves/w1_ds2433.c idx += len; idx 228 drivers/w1/slaves/w1_ds28e04.c int addr, len, idx; idx 243 drivers/w1/slaves/w1_ds28e04.c for (idx = 0; idx < count; idx += W1_PAGE_SIZE) { idx 244 drivers/w1/slaves/w1_ds28e04.c if (crc16(CRC16_INIT, &buf[idx], W1_PAGE_SIZE) idx 256 drivers/w1/slaves/w1_ds28e04.c idx = 0; idx 257 drivers/w1/slaves/w1_ds28e04.c while (idx < count) { idx 258 drivers/w1/slaves/w1_ds28e04.c addr = off + idx; idx 260 drivers/w1/slaves/w1_ds28e04.c if (len > (count - idx)) idx 261 drivers/w1/slaves/w1_ds28e04.c len = count - idx; idx 263 drivers/w1/slaves/w1_ds28e04.c if (w1_f1C_write(sl, addr, len, &buf[idx]) < 0) { idx 267 drivers/w1/slaves/w1_ds28e04.c idx += len; idx 222 drivers/w1/w1_netlink.c packet.cn.id.idx = CN_W1_IDX; idx 408 drivers/w1/w1_netlink.c cn->id.idx = CN_W1_IDX; idx 665 drivers/w1/w1_netlink.c __func__, cn->id.idx, cn->id.val, idx 715 drivers/w1/w1_netlink.c struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL}; idx 722 drivers/w1/w1_netlink.c struct cb_id w1_id = {.idx = CN_W1_IDX, .val = CN_W1_VAL}; idx 147 drivers/xen/events/events_2l.c unsigned int idx) idx 149 drivers/xen/events/events_2l.c return sh->evtchn_pending[idx] & idx 150 drivers/xen/events/events_2l.c per_cpu(cpu_evtchn_mask, cpu)[idx] & idx 151 drivers/xen/events/events_2l.c ~sh->evtchn_mask[idx]; idx 96 drivers/xen/evtchn.c unsigned int idx) idx 98 drivers/xen/evtchn.c return idx & (u->ring_size - 1); idx 102 drivers/xen/evtchn.c unsigned int idx) idx 104 drivers/xen/evtchn.c return u->ring + evtchn_ring_offset(u, idx); idx 1251 drivers/xen/grant-table.c xatp.idx = i; idx 225 drivers/xen/manage.c int idx; idx 242 drivers/xen/manage.c for (idx = 0; idx < ARRAY_SIZE(shutdown_handlers); idx++) { idx 243 drivers/xen/manage.c if (strcmp(str, shutdown_handlers[idx].command) == 0) idx 248 drivers/xen/manage.c if (idx < ARRAY_SIZE(shutdown_handlers)) idx 257 drivers/xen/manage.c if (idx < ARRAY_SIZE(shutdown_handlers)) { idx 258 drivers/xen/manage.c shutdown_handlers[idx].cb(); idx 331 drivers/xen/manage.c int idx; idx 350 drivers/xen/manage.c for (idx = 0; idx < ARRAY_SIZE(shutdown_handlers); idx++) { idx 351 drivers/xen/manage.c if (!shutdown_handlers[idx].flag) idx 354 drivers/xen/manage.c shutdown_handlers[idx].command); idx 141 drivers/xen/pcpu.c struct attribute *attr, int idx) idx 782 drivers/xen/privcmd.c xdata.frame = kdata.idx; idx 529 drivers/xen/xenbus/xenbus_client.c unsigned int idx; idx 544 drivers/xen/xenbus/xenbus_client.c info->phys_addrs[info->idx] = vaddr; idx 545 drivers/xen/xenbus/xenbus_client.c info->addrs[info->idx] = vaddr; idx 547 drivers/xen/xenbus/xenbus_client.c info->idx++; idx 560 drivers/xen/xenbus/xenbus_client.c .idx = 0, idx 805 drivers/xen/xenbus/xenbus_client.c unsigned int idx; idx 816 drivers/xen/xenbus/xenbus_client.c info->addrs[info->idx] = (unsigned long)gfn_to_virt(gfn); idx 818 drivers/xen/xenbus/xenbus_client.c info->idx++; idx 827 drivers/xen/xenbus/xenbus_client.c .idx = 0, idx 342 drivers/xen/xenbus/xenbus_comms.c int idx; idx 357 drivers/xen/xenbus/xenbus_comms.c state.idx = -1; idx 364 drivers/xen/xenbus/xenbus_comms.c while (state.idx < state.req->num_vecs) { idx 365 drivers/xen/xenbus/xenbus_comms.c if (state.idx < 0) { idx 369 drivers/xen/xenbus/xenbus_comms.c base = state.req->vec[state.idx].iov_base; idx 370 drivers/xen/xenbus/xenbus_comms.c len = state.req->vec[state.idx].iov_len; idx 379 drivers/xen/xenbus/xenbus_comms.c state.idx++; idx 194 drivers/xen/xlate_mmu.c unsigned int idx; idx 201 drivers/xen/xlate_mmu.c info->pfns[info->idx++] = gfn; idx 245 drivers/xen/xlate_mmu.c data.idx = 0; idx 48 fs/adfs/dir_f.c #define dir_u8(idx) \ idx 49 fs/adfs/dir_f.c ({ int _buf = idx >> blocksize_bits; \ idx 50 fs/adfs/dir_f.c int _off = idx - (_buf << blocksize_bits);\ idx 54 fs/adfs/dir_f.c #define dir_u32(idx) \ idx 55 fs/adfs/dir_f.c ({ int _buf = idx >> blocksize_bits; \ idx 56 fs/adfs/dir_f.c int _off = idx - (_buf << blocksize_bits);\ idx 176 fs/affs/file.c u32 tmp, idx; idx 243 fs/affs/file.c idx = ac_idx; idx 245 fs/affs/file.c idx = (idx - 1) & AFFS_AC_MASK; idx 246 fs/affs/file.c if (AFFS_I(inode)->i_ac[idx].ext == tmp) { idx 247 fs/affs/file.c ext_key = AFFS_I(inode)->i_ac[idx].key; idx 382 fs/aio.c pgoff_t idx; idx 412 fs/aio.c idx = old->index; idx 413 fs/aio.c if (idx < (pgoff_t)ctx->nr_pages) { idx 415 fs/aio.c if (ctx->ring_pages[idx] != old) idx 439 fs/aio.c BUG_ON(ctx->ring_pages[idx] != old); idx 440 fs/aio.c ctx->ring_pages[idx] = new; idx 582 fs/autofs/dev-ioctl.c unsigned int idx = cmd_idx(cmd); idx 584 fs/autofs/dev-ioctl.c return (idx >= ARRAY_SIZE(_ioctls)) ? NULL : _ioctls[idx]; idx 345 fs/btrfs/relocation.c int idx = *index; idx 350 fs/btrfs/relocation.c edges[idx++] = edge; idx 354 fs/btrfs/relocation.c *index = idx; idx 366 fs/btrfs/relocation.c int idx = *index; idx 368 fs/btrfs/relocation.c while (idx > 0) { idx 369 fs/btrfs/relocation.c edge = edges[idx - 1]; idx 372 fs/btrfs/relocation.c idx--; idx 377 fs/btrfs/relocation.c edges[idx - 1] = edge; idx 378 fs/btrfs/relocation.c *index = idx; idx 1546 fs/btrfs/send.c u64 idx = 0; idx 1554 fs/btrfs/send.c ino, gen, idx); idx 1567 fs/btrfs/send.c idx++; idx 1587 fs/btrfs/send.c idx++; idx 124 fs/ceph/dir.c __dcache_find_get_entry(struct dentry *parent, u64 idx, idx 130 fs/ceph/dir.c loff_t ptr_pos = idx * sizeof(struct dentry *); idx 149 fs/ceph/dir.c cache_ctl->index = idx & idx_mask; idx 186 fs/ceph/dir.c u64 idx = 0; idx 196 fs/ceph/dir.c dentry = __dcache_find_get_entry(parent, idx + step, idx 200 fs/ceph/dir.c idx = 0; idx 210 fs/ceph/dir.c idx += step + 1; idx 219 fs/ceph/dir.c dout("__dcache_readdir %p cache idx %llu\n", dir, idx); idx 225 fs/ceph/dir.c dentry = __dcache_find_get_entry(parent, idx++, &cache_ctl); idx 94 fs/ceph/file.c int idx = 0; idx 104 fs/ceph/file.c for ( ; bytes; idx++, bvec_idx++) { idx 106 fs/ceph/file.c .bv_page = pages[idx], idx 687 fs/ceph/file.c int idx = 0; idx 693 fs/ceph/file.c copied = copy_page_to_iter(pages[idx++], idx 1524 fs/ceph/inode.c unsigned idx = ctl->index % nsize; idx 1529 fs/ceph/inode.c if (idx == 0) idx 1535 fs/ceph/inode.c return idx == 0 ? -ENOMEM : 0; idx 1541 fs/ceph/inode.c if (idx == 0) idx 1548 fs/ceph/inode.c ctl->dentries[idx] = dn; idx 384 fs/char_dev.c int idx; idx 386 fs/char_dev.c kobj = kobj_lookup(cdev_map, inode->i_rdev, &idx); idx 776 fs/cifs/netmisc.c int idx = 0; idx 778 fs/cifs/netmisc.c while (nt_errs[idx].nt_errstr != NULL) { idx 779 fs/cifs/netmisc.c if (((nt_errs[idx].nt_errcode) & 0xFFFFFF) == idx 782 fs/cifs/netmisc.c status_code, nt_errs[idx].nt_errstr); idx 784 fs/cifs/netmisc.c idx++; idx 2440 fs/cifs/smb2maperror.c int idx = 0; idx 2442 fs/cifs/smb2maperror.c while (smb2_error_map_table[idx].status_string != NULL) { idx 2443 fs/cifs/smb2maperror.c if ((smb2_error_map_table[idx].smb2_status) == status) { idx 2445 fs/cifs/smb2maperror.c smb2_error_map_table[idx].status_string); idx 2447 fs/cifs/smb2maperror.c idx++; idx 3597 fs/cifs/smb2ops.c unsigned int idx = 0; idx 3616 fs/cifs/smb2ops.c smb2_sg_set_buf(&sg[idx++], idx 3625 fs/cifs/smb2ops.c sg_set_page(&sg[idx++], rqst[i].rq_pages[j], len, offset); idx 3628 fs/cifs/smb2ops.c smb2_sg_set_buf(&sg[idx], sign, SMB2_SIGNATURE_SIZE); idx 629 fs/cifs/winucase.c unsigned char idx; idx 634 fs/cifs/winucase.c idx = (in & 0xff00) >> 8; idx 637 fs/cifs/winucase.c tbl = toplevel[idx]; idx 642 fs/cifs/winucase.c idx = in & 0xff; idx 645 fs/cifs/winucase.c out = tbl[idx]; idx 112 fs/coda/inode.c int idx; idx 133 fs/coda/inode.c idx = iminor(inode); idx 136 fs/coda/inode.c if (idx < 0 || idx >= MAX_CODADEVS) { idx 141 fs/coda/inode.c return idx; idx 153 fs/coda/inode.c int idx; idx 158 fs/coda/inode.c idx = get_device_index((struct coda_mount_data *) data); idx 161 fs/coda/inode.c if(idx == -1) idx 162 fs/coda/inode.c idx = 0; idx 164 fs/coda/inode.c pr_info("%s: device index: %i\n", __func__, idx); idx 166 fs/coda/inode.c vc = &coda_comms[idx]; idx 275 fs/coda/psdev.c int idx, err; idx 283 fs/coda/psdev.c idx = iminor(inode); idx 284 fs/coda/psdev.c if (idx < 0 || idx >= MAX_CODADEVS) idx 288 fs/coda/psdev.c vcp = &coda_comms[idx]; idx 547 fs/devpts/inode.c void devpts_kill_index(struct pts_fs_info *fsi, int idx) idx 549 fs/devpts/inode.c ida_free(&fsi->allocated_ptys, idx); idx 47 fs/erofs/xattr.h static inline const struct xattr_handler *erofs_xattr_handler(unsigned int idx) idx 62 fs/erofs/xattr.h return idx && idx < ARRAY_SIZE(xattr_handler_map) ? idx 63 fs/erofs/xattr.h xattr_handler_map[idx] : NULL; idx 689 fs/ext4/extents.c struct ext4_extent_idx *idx; idx 690 fs/ext4/extents.c idx = path[level].p_idx; idx 691 fs/ext4/extents.c while (idx <= EXT_MAX_INDEX(path[level].p_hdr)) { idx 693 fs/ext4/extents.c le32_to_cpu(idx->ei_block), idx 694 fs/ext4/extents.c ext4_idx_pblock(idx), idx 696 fs/ext4/extents.c idx++; idx 3140 fs/ext4/inode.c unsigned int idx; idx 3144 fs/ext4/inode.c idx = offset >> inode->i_blkbits; idx 3146 fs/ext4/inode.c for (i = 0; i < idx; i++) idx 2395 fs/ext4/mballoc.c int idx = group >> EXT4_DESC_PER_BLOCK_BITS(sb); idx 2415 fs/ext4/mballoc.c rcu_dereference(sbi->s_group_info)[idx] = meta_group_info; idx 2419 fs/ext4/mballoc.c meta_group_info = sbi_array_rcu_deref(sbi, s_group_info, idx); idx 2471 fs/ext4/mballoc.c kfree(group_info[idx]); idx 2472 fs/ext4/mballoc.c group_info[idx] = NULL; idx 186 fs/ext4/readpage.c static inline bool ext4_need_verity(const struct inode *inode, pgoff_t idx) idx 189 fs/ext4/readpage.c idx < DIV_ROUND_UP(inode->i_size, PAGE_SIZE); idx 2955 fs/ext4/xattr.c int idx; idx 2960 fs/ext4/xattr.c for (idx = 0; idx < ea_inode_array->count; ++idx) idx 2961 fs/ext4/xattr.c iput(ea_inode_array->inodes[idx]); idx 670 fs/f2fs/data.c static inline bool f2fs_need_verity(const struct inode *inode, pgoff_t idx) idx 673 fs/f2fs/data.c idx < DIV_ROUND_UP(inode->i_size, PAGE_SIZE); idx 74 fs/f2fs/dir.c int dir_level, unsigned int idx) idx 81 fs/f2fs/dir.c bidx += idx * bucket_blocks(level); idx 1443 fs/f2fs/file.c pgoff_t nr, pg_start, pg_end, delta, idx; idx 1479 fs/f2fs/file.c idx = DIV_ROUND_UP(i_size_read(inode), PAGE_SIZE); idx 1486 fs/f2fs/file.c while (!ret && idx > pg_start) { idx 1487 fs/f2fs/file.c nr = idx - pg_start; idx 1490 fs/f2fs/file.c idx -= nr; idx 1495 fs/f2fs/file.c ret = __exchange_data_block(inode, inode, idx, idx 1496 fs/f2fs/file.c idx + delta, nr, false); idx 2468 fs/f2fs/file.c pgoff_t idx; idx 2484 fs/f2fs/file.c idx = map.m_lblk; idx 2485 fs/f2fs/file.c while (idx < map.m_lblk + map.m_len && cnt < blk_per_seg) { idx 2488 fs/f2fs/file.c page = f2fs_get_lock_data_page(inode, idx, true); idx 2497 fs/f2fs/file.c idx++; idx 2502 fs/f2fs/file.c map.m_lblk = idx; idx 979 fs/f2fs/node.c int idx = depth - 2; idx 986 fs/f2fs/node.c for (i = 0; i < idx + 1; i++) { idx 991 fs/f2fs/node.c idx = i - 1; idx 997 fs/f2fs/node.c f2fs_ra_node_pages(pages[idx], offset[idx + 1], NIDS_PER_BLOCK); idx 1000 fs/f2fs/node.c for (i = offset[idx + 1]; i < NIDS_PER_BLOCK; i++) { idx 1001 fs/f2fs/node.c child_nid = get_nid(pages[idx], i, false); idx 1008 fs/f2fs/node.c if (set_nid(pages[idx], i, 0, false)) idx 1012 fs/f2fs/node.c if (offset[idx + 1] == 0) { idx 1013 fs/f2fs/node.c dn->node_page = pages[idx]; idx 1014 fs/f2fs/node.c dn->nid = nid[idx]; idx 1019 fs/f2fs/node.c f2fs_put_page(pages[idx], 1); idx 1021 fs/f2fs/node.c offset[idx]++; idx 1022 fs/f2fs/node.c offset[idx + 1] = 0; idx 1023 fs/f2fs/node.c idx--; idx 1025 fs/f2fs/node.c for (i = idx; i >= 0; i--) idx 2279 fs/f2fs/node.c unsigned int i, idx; idx 2289 fs/f2fs/node.c for (idx = 0; idx < NAT_ENTRY_PER_BLOCK; idx++) { idx 2290 fs/f2fs/node.c idx = find_next_bit_le(nm_i->free_nid_bitmap[i], idx 2291 fs/f2fs/node.c NAT_ENTRY_PER_BLOCK, idx); idx 2292 fs/f2fs/node.c if (idx >= NAT_ENTRY_PER_BLOCK) idx 2295 fs/f2fs/node.c nid = i * NAT_ENTRY_PER_BLOCK + idx; idx 2667 fs/f2fs/node.c int i, idx, last_offset, nrpages; idx 2680 fs/f2fs/node.c for (idx = addr; idx < addr + nrpages; idx++) { idx 2681 fs/f2fs/node.c struct page *page = f2fs_get_tmp_page(sbi, idx); idx 2903 fs/f2fs/node.c unsigned idx; idx 2905 fs/f2fs/node.c for (idx = 0; idx < found; idx++) idx 2906 fs/f2fs/node.c __adjust_nat_entry_set(setvec[idx], &sets, idx 3148 fs/f2fs/node.c unsigned idx; idx 3151 fs/f2fs/node.c for (idx = 0; idx < found; idx++) { idx 3153 fs/f2fs/node.c list_del(&natvec[idx]->list); idx 3156 fs/f2fs/node.c __del_from_nat_cache(nm_i, natvec[idx]); idx 3165 fs/f2fs/node.c unsigned idx; idx 3168 fs/f2fs/node.c for (idx = 0; idx < found; idx++) { idx 3170 fs/f2fs/node.c f2fs_bug_on(sbi, !list_empty(&setvec[idx]->entry_list)); idx 3171 fs/f2fs/node.c radix_tree_delete(&nm_i->nat_set_root, setvec[idx]->set); idx 3172 fs/f2fs/node.c kmem_cache_free(nat_entry_set_slab, setvec[idx]); idx 36 fs/f2fs/segment.c int shift = 24, idx = 0; idx 42 fs/f2fs/segment.c tmp |= (unsigned long)str[idx++] << shift; idx 158 fs/f2fs/trace.c unsigned idx; idx 161 fs/f2fs/trace.c for (idx = 0; idx < found; idx++) idx 162 fs/f2fs/trace.c radix_tree_delete(&pids, pid[idx]); idx 1941 fs/fuse/dev.c unsigned idx; idx 1963 fs/fuse/dev.c for (idx = 0; idx < pipe->nrbufs && rem < len; idx++) idx 1964 fs/fuse/dev.c rem += pipe->bufs[(pipe->curbuf + idx) & (pipe->buffers - 1)].len; idx 2012 fs/fuse/dev.c for (idx = 0; idx < nbuf; idx++) idx 2013 fs/fuse/dev.c pipe_buf_release(pipe, &bufs[idx]); idx 166 fs/hfs/bfind.c int idx, res = 0; idx 177 fs/hfs/bfind.c idx = bnode->prev; idx 178 fs/hfs/bfind.c if (!idx) { idx 183 fs/hfs/bfind.c bnode = hfs_bnode_find(tree, idx); idx 195 fs/hfs/bfind.c idx = bnode->next; idx 196 fs/hfs/bfind.c if (!idx) { idx 201 fs/hfs/bfind.c bnode = hfs_bnode_find(tree, idx); idx 188 fs/hfs/btree.c static struct hfs_bnode *hfs_bmap_new_bmap(struct hfs_bnode *prev, u32 idx) idx 195 fs/hfs/btree.c node = hfs_bnode_create(tree, idx); idx 202 fs/hfs/btree.c prev->next = idx; idx 203 fs/hfs/btree.c cnid = cpu_to_be32(idx); idx 251 fs/hfs/btree.c u32 nidx, idx; idx 273 fs/hfs/btree.c idx = 0; idx 281 fs/hfs/btree.c idx += i; idx 288 fs/hfs/btree.c return hfs_bnode_create(tree, idx); idx 297 fs/hfs/btree.c idx += 8; idx 304 fs/hfs/btree.c next_node = hfs_bmap_new_bmap(node, idx); idx 235 fs/hfsplus/bfind.c int idx, res = 0; idx 246 fs/hfsplus/bfind.c idx = bnode->prev; idx 247 fs/hfsplus/bfind.c if (!idx) { idx 252 fs/hfsplus/bfind.c bnode = hfs_bnode_find(tree, idx); idx 264 fs/hfsplus/bfind.c idx = bnode->next; idx 265 fs/hfsplus/bfind.c if (!idx) { idx 270 fs/hfsplus/bfind.c bnode = hfs_bnode_find(tree, idx); idx 312 fs/hfsplus/btree.c static struct hfs_bnode *hfs_bmap_new_bmap(struct hfs_bnode *prev, u32 idx) idx 319 fs/hfsplus/btree.c node = hfs_bnode_create(tree, idx); idx 324 fs/hfsplus/btree.c prev->next = idx; idx 325 fs/hfsplus/btree.c cnid = cpu_to_be32(idx); idx 377 fs/hfsplus/btree.c u32 nidx, idx; idx 399 fs/hfsplus/btree.c idx = 0; idx 407 fs/hfsplus/btree.c idx += i; idx 415 fs/hfsplus/btree.c idx); idx 424 fs/hfsplus/btree.c idx += 8; idx 431 fs/hfsplus/btree.c next_node = hfs_bmap_new_bmap(node, idx); idx 489 fs/hpfs/alloc.c static unsigned find_run(__le32 *bmp, unsigned *idx) idx 492 fs/hpfs/alloc.c while (tstbits(bmp, *idx, 1)) { idx 493 fs/hpfs/alloc.c (*idx)++; idx 494 fs/hpfs/alloc.c if (unlikely(*idx >= 0x4000)) idx 498 fs/hpfs/alloc.c while (!tstbits(bmp, *idx + len, 1)) idx 529 fs/hpfs/alloc.c unsigned idx, len, start_bmp, end_bmp; idx 550 fs/hpfs/alloc.c idx = 0; idx 551 fs/hpfs/alloc.c while ((len = find_run(bmp, &idx)) && !err) { idx 552 fs/hpfs/alloc.c err = do_trim(s, sbi->sb_dirband_start + idx * 4, len * 4, start, end, minlen, result); idx 553 fs/hpfs/alloc.c idx += len; idx 571 fs/hpfs/alloc.c idx = 0; idx 572 fs/hpfs/alloc.c while ((len = find_run(bmp, &idx)) && !err) { idx 573 fs/hpfs/alloc.c err = do_trim(s, (start_bmp << 14) + idx, len, start, end, minlen, result); idx 574 fs/hpfs/alloc.c idx += len; idx 890 fs/nfs/blocklayout/blocklayout.c static u64 pnfs_num_cont_bytes(struct inode *inode, pgoff_t idx) idx 899 fs/nfs/blocklayout/blocklayout.c end = page_cache_next_miss(mapping, idx + 1, ULONG_MAX); idx 904 fs/nfs/blocklayout/blocklayout.c return i_size_read(inode) - (idx << PAGE_SHIFT); idx 906 fs/nfs/blocklayout/blocklayout.c return (end - idx) << PAGE_SHIFT; idx 231 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask); idx 236 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) idx 238 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *v = &volumes[idx]; idx 353 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) idx 355 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *v = &volumes[idx]; idx 402 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) idx 404 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *v = &volumes[idx]; idx 418 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) idx 420 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *v = &volumes[idx]; idx 447 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) idx 449 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *v = &volumes[idx]; idx 476 fs/nfs/blocklayout/dev.c struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) idx 478 fs/nfs/blocklayout/dev.c switch (volumes[idx].type) { idx 480 fs/nfs/blocklayout/dev.c return bl_parse_simple(server, d, volumes, idx, gfp_mask); idx 482 fs/nfs/blocklayout/dev.c return bl_parse_slice(server, d, volumes, idx, gfp_mask); idx 484 fs/nfs/blocklayout/dev.c return bl_parse_concat(server, d, volumes, idx, gfp_mask); idx 486 fs/nfs/blocklayout/dev.c return bl_parse_stripe(server, d, volumes, idx, gfp_mask); idx 488 fs/nfs/blocklayout/dev.c return bl_parse_scsi(server, d, volumes, idx, gfp_mask); idx 490 fs/nfs/blocklayout/dev.c dprintk("unsupported volume type: %d\n", volumes[idx].type); idx 455 fs/nfs/filelayout/filelayout.c u32 j, idx; idx 464 fs/nfs/filelayout/filelayout.c idx = nfs4_fl_calc_ds_index(lseg, j); idx 465 fs/nfs/filelayout/filelayout.c ds = nfs4_fl_prepare_ds(lseg, idx); idx 479 fs/nfs/filelayout/filelayout.c hdr->ds_commit_idx = idx; idx 502 fs/nfs/filelayout/filelayout.c u32 j, idx; idx 507 fs/nfs/filelayout/filelayout.c idx = nfs4_fl_calc_ds_index(lseg, j); idx 508 fs/nfs/filelayout/filelayout.c ds = nfs4_fl_prepare_ds(lseg, idx); idx 523 fs/nfs/filelayout/filelayout.c hdr->ds_commit_idx = idx; idx 1052 fs/nfs/filelayout/filelayout.c u32 idx; idx 1055 fs/nfs/filelayout/filelayout.c idx = calc_ds_index_from_commit(lseg, data->ds_commit_index); idx 1056 fs/nfs/filelayout/filelayout.c ds = nfs4_fl_prepare_ds(lseg, idx); idx 796 fs/nfs/flexfilelayout/flexfilelayout.c ff_layout_mark_ds_unreachable(struct pnfs_layout_segment *lseg, int idx) idx 798 fs/nfs/flexfilelayout/flexfilelayout.c struct nfs4_deviceid_node *devid = FF_LAYOUT_DEVID_NODE(lseg, idx); idx 805 fs/nfs/flexfilelayout/flexfilelayout.c ff_layout_mark_ds_reachable(struct pnfs_layout_segment *lseg, int idx) idx 807 fs/nfs/flexfilelayout/flexfilelayout.c struct nfs4_deviceid_node *devid = FF_LAYOUT_DEVID_NODE(lseg, idx); idx 822 fs/nfs/flexfilelayout/flexfilelayout.c int idx; idx 825 fs/nfs/flexfilelayout/flexfilelayout.c for (idx = start_idx; idx < fls->mirror_array_cnt; idx++) { idx 826 fs/nfs/flexfilelayout/flexfilelayout.c if (idx+1 == fls->mirror_array_cnt) idx 829 fs/nfs/flexfilelayout/flexfilelayout.c mirror = FF_LAYOUT_COMP(lseg, idx); idx 838 fs/nfs/flexfilelayout/flexfilelayout.c *best_idx = idx; idx 1132 fs/nfs/flexfilelayout/flexfilelayout.c int idx) idx 1136 fs/nfs/flexfilelayout/flexfilelayout.c struct nfs4_deviceid_node *devid = FF_LAYOUT_DEVID_NODE(lseg, idx); idx 1206 fs/nfs/flexfilelayout/flexfilelayout.c int idx) idx 1208 fs/nfs/flexfilelayout/flexfilelayout.c struct nfs4_deviceid_node *devid = FF_LAYOUT_DEVID_NODE(lseg, idx); idx 1241 fs/nfs/flexfilelayout/flexfilelayout.c int idx) idx 1246 fs/nfs/flexfilelayout/flexfilelayout.c ff_layout_mark_ds_reachable(lseg, idx); idx 1256 fs/nfs/flexfilelayout/flexfilelayout.c return ff_layout_async_handle_error_v3(task, lseg, idx); idx 1259 fs/nfs/flexfilelayout/flexfilelayout.c lseg, idx); idx 1268 fs/nfs/flexfilelayout/flexfilelayout.c int idx, u64 offset, u64 length, idx 1307 fs/nfs/flexfilelayout/flexfilelayout.c mirror = FF_LAYOUT_COMP(lseg, idx); idx 1312 fs/nfs/flexfilelayout/flexfilelayout.c ff_layout_mark_ds_unreachable(lseg, idx); idx 1786 fs/nfs/flexfilelayout/flexfilelayout.c u32 idx = hdr->pgio_mirror_idx; idx 1794 fs/nfs/flexfilelayout/flexfilelayout.c mirror = FF_LAYOUT_COMP(lseg, idx); idx 1858 fs/nfs/flexfilelayout/flexfilelayout.c int idx = hdr->pgio_mirror_idx; idx 1860 fs/nfs/flexfilelayout/flexfilelayout.c mirror = FF_LAYOUT_COMP(lseg, idx); idx 1884 fs/nfs/flexfilelayout/flexfilelayout.c hdr->ds_commit_idx = idx; idx 1937 fs/nfs/flexfilelayout/flexfilelayout.c u32 idx; idx 1945 fs/nfs/flexfilelayout/flexfilelayout.c idx = calc_ds_index_from_commit(lseg, data->ds_commit_index); idx 1946 fs/nfs/flexfilelayout/flexfilelayout.c mirror = FF_LAYOUT_COMP(lseg, idx); idx 142 fs/nfs/flexfilelayout/flexfilelayout.h FF_LAYOUT_COMP(struct pnfs_layout_segment *lseg, u32 idx) idx 146 fs/nfs/flexfilelayout/flexfilelayout.h if (idx < fls->mirror_array_cnt) idx 147 fs/nfs/flexfilelayout/flexfilelayout.h return fls->mirror_array[idx]; idx 152 fs/nfs/flexfilelayout/flexfilelayout.h FF_LAYOUT_DEVID_NODE(struct pnfs_layout_segment *lseg, u32 idx) idx 154 fs/nfs/flexfilelayout/flexfilelayout.h struct nfs4_ff_layout_mirror *mirror = FF_LAYOUT_COMP(lseg, idx); idx 554 fs/nfs/flexfilelayout/flexfilelayoutdev.c u32 idx; idx 556 fs/nfs/flexfilelayout/flexfilelayoutdev.c for (idx = 0; idx < FF_LAYOUT_MIRROR_COUNT(lseg); idx++) { idx 557 fs/nfs/flexfilelayout/flexfilelayoutdev.c mirror = FF_LAYOUT_COMP(lseg, idx); idx 576 fs/nfs/flexfilelayout/flexfilelayoutdev.c u32 idx; idx 578 fs/nfs/flexfilelayout/flexfilelayoutdev.c for (idx = 0; idx < FF_LAYOUT_MIRROR_COUNT(lseg); idx++) { idx 579 fs/nfs/flexfilelayout/flexfilelayoutdev.c mirror = FF_LAYOUT_COMP(lseg, idx); idx 159 fs/nfs/pnfs_nfs.c static void pnfs_generic_retry_commit(struct nfs_commit_info *cinfo, int idx) idx 169 fs/nfs/pnfs_nfs.c for (i = idx; i < fl_cinfo->nbuckets; i++) { idx 50 fs/nfsd/nfs4layouts.c static inline u32 devid_hashfn(u64 idx) idx 52 fs/nfsd/nfs4layouts.c return jhash_2words(idx, idx >> 32, 0) & DEVID_HASH_MASK; idx 87 fs/nfsd/nfs4layouts.c map->idx = nfsd_devid_seq++; idx 88 fs/nfsd/nfs4layouts.c list_add_tail_rcu(&map->hash, &nfsd_devid_hash[devid_hashfn(map->idx)]); idx 98 fs/nfsd/nfs4layouts.c nfsd4_find_devid_map(int idx) idx 103 fs/nfsd/nfs4layouts.c list_for_each_entry_rcu(map, &nfsd_devid_hash[devid_hashfn(idx)], hash) idx 104 fs/nfsd/nfs4layouts.c if (map->idx == idx) idx 121 fs/nfsd/nfs4layouts.c id->fsid_idx = fhp->fh_export->ex_devid_map->idx; idx 1766 fs/nfsd/nfs4state.c int idx; idx 1779 fs/nfsd/nfs4state.c idx = hash_sessionid(&new->se_sessionid); idx 1780 fs/nfsd/nfs4state.c list_add(&new->se_hash, &nn->sessionid_hashtbl[idx]); idx 1804 fs/nfsd/nfs4state.c int idx; idx 1810 fs/nfsd/nfs4state.c idx = hash_sessionid(sessionid); idx 1812 fs/nfsd/nfs4state.c list_for_each_entry(elem, &nn->sessionid_hashtbl[idx], se_hash) { idx 294 fs/nfsd/nfscache.c int idx; idx 308 fs/nfsd/nfscache.c idx = buf->page_base / PAGE_SIZE; idx 311 fs/nfsd/nfscache.c p = page_address(buf->pages[idx]) + base; idx 316 fs/nfsd/nfscache.c ++idx; idx 16 fs/nfsd/pnfs.h u64 idx; idx 67 fs/nfsd/pnfs.h struct nfsd4_deviceid_map *nfsd4_find_devid_map(int idx); idx 520 fs/notify/mark.c int idx; idx 522 fs/notify/mark.c idx = srcu_read_lock(&fsnotify_mark_srcu); idx 529 fs/notify/mark.c srcu_read_unlock(&fsnotify_mark_srcu, idx); idx 533 fs/notify/mark.c srcu_read_unlock(&fsnotify_mark_srcu, idx); idx 2485 fs/ntfs/attrib.c pgoff_t idx, end; idx 2502 fs/ntfs/attrib.c idx = ofs >> PAGE_SHIFT; idx 2515 fs/ntfs/attrib.c page = read_mapping_page(mapping, idx, NULL); idx 2518 fs/ntfs/attrib.c "page (error, index 0x%lx).", idx); idx 2526 fs/ntfs/attrib.c if (idx == end) idx 2536 fs/ntfs/attrib.c if (idx == end) idx 2538 fs/ntfs/attrib.c idx++; idx 2541 fs/ntfs/attrib.c for (; idx < end; idx++) { idx 2543 fs/ntfs/attrib.c page = grab_cache_page(mapping, idx); idx 2546 fs/ntfs/attrib.c "page (index 0x%lx).", idx); idx 2580 fs/ntfs/attrib.c page = read_mapping_page(mapping, idx, NULL); idx 2583 fs/ntfs/attrib.c "(error, index 0x%lx).", idx); idx 1780 fs/ntfs/file.c pgoff_t idx, start_idx; idx 1784 fs/ntfs/file.c start_idx = idx = pos >> PAGE_SHIFT; idx 264 fs/ntfs/logfile.c u16 nr_clients, idx; idx 280 fs/ntfs/logfile.c idx = le16_to_cpu(ra->client_free_list); idx 283 fs/ntfs/logfile.c for (idx_is_first = true; idx != LOGFILE_NO_CLIENT_CPU; nr_clients--, idx 284 fs/ntfs/logfile.c idx = le16_to_cpu(cr->next_client)) { idx 285 fs/ntfs/logfile.c if (!nr_clients || idx >= le16_to_cpu(ra->log_clients)) idx 288 fs/ntfs/logfile.c cr = ca + idx; idx 299 fs/ntfs/logfile.c idx = le16_to_cpu(ra->client_in_use_list); idx 375 fs/ntfs/logfile.c pgoff_t idx; idx 384 fs/ntfs/logfile.c idx = (pos + size) >> PAGE_SHIFT; idx 387 fs/ntfs/logfile.c page = ntfs_map_page(vi->i_mapping, idx); idx 390 fs/ntfs/logfile.c "page (index %lu).", idx); idx 401 fs/ntfs/logfile.c idx++; idx 529 fs/ntfs/logfile.c pgoff_t idx = pos >> PAGE_SHIFT; idx 530 fs/ntfs/logfile.c if (!page || page->index != idx) { idx 533 fs/ntfs/logfile.c page = ntfs_map_page(mapping, idx); idx 536 fs/ntfs/logfile.c "page (index %lu).", idx); idx 724 fs/ocfs2/alloc.c int idx) idx 731 fs/ocfs2/alloc.c if (idx) idx 734 fs/ocfs2/alloc.c return access(handle, ci, path->p_node[idx].bh, idx 2043 fs/ocfs2/alloc.c int i, idx; idx 2072 fs/ocfs2/alloc.c idx = le16_to_cpu(left_el->l_next_free_rec) - 1; idx 2073 fs/ocfs2/alloc.c left_rec = &el->l_recs[idx]; idx 2528 fs/ocfs2/alloc.c int i, idx, ret; idx 2546 fs/ocfs2/alloc.c idx = le16_to_cpu(el->l_next_free_rec) - 1; idx 2547 fs/ocfs2/alloc.c rec = &el->l_recs[idx]; idx 2552 fs/ocfs2/alloc.c idx = le16_to_cpu(el->l_next_free_rec) - 1; idx 2553 fs/ocfs2/alloc.c rec = &el->l_recs[idx]; idx 301 fs/ocfs2/alloc.h int idx); idx 742 fs/ocfs2/cluster/heartbeat.c int idx) idx 748 fs/ocfs2/cluster/heartbeat.c (f->hc_func)(node, idx, f->hc_data); idx 380 fs/ocfs2/dlm/dlmcommon.h static inline char *dlm_list_in_text(enum dlm_lockres_list idx) idx 382 fs/ocfs2/dlm/dlmcommon.h if (idx == DLM_GRANTED_LIST) idx 384 fs/ocfs2/dlm/dlmcommon.h else if (idx == DLM_CONVERTING_LIST) idx 386 fs/ocfs2/dlm/dlmcommon.h else if (idx == DLM_BLOCKED_LIST) idx 393 fs/ocfs2/dlm/dlmcommon.h dlm_list_idx_to_ptr(struct dlm_lock_resource *res, enum dlm_lockres_list idx) idx 396 fs/ocfs2/dlm/dlmcommon.h if (idx == DLM_GRANTED_LIST) idx 398 fs/ocfs2/dlm/dlmcommon.h else if (idx == DLM_CONVERTING_LIST) idx 400 fs/ocfs2/dlm/dlmcommon.h else if (idx == DLM_BLOCKED_LIST) idx 952 fs/ocfs2/dlm/dlmcommon.h void dlm_hb_node_down_cb(struct o2nm_node *node, int idx, void *data); idx 953 fs/ocfs2/dlm/dlmcommon.h void dlm_hb_node_up_cb(struct o2nm_node *node, int idx, void *data); idx 1032 fs/ocfs2/dlm/dlmcommon.h void dlm_hb_event_notify_attached(struct dlm_ctxt *dlm, int idx, int node_up); idx 43 fs/ocfs2/dlm/dlmmaster.c int idx); idx 47 fs/ocfs2/dlm/dlmmaster.c int idx); idx 346 fs/ocfs2/dlm/dlmmaster.c void dlm_hb_event_notify_attached(struct dlm_ctxt *dlm, int idx, int node_up) idx 354 fs/ocfs2/dlm/dlmmaster.c dlm_mle_node_up(dlm, mle, NULL, idx); idx 356 fs/ocfs2/dlm/dlmmaster.c dlm_mle_node_down(dlm, mle, NULL, idx); idx 362 fs/ocfs2/dlm/dlmmaster.c struct o2nm_node *node, int idx) idx 366 fs/ocfs2/dlm/dlmmaster.c if (!test_bit(idx, mle->node_map)) idx 367 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u already removed from nodemap!\n", idx); idx 369 fs/ocfs2/dlm/dlmmaster.c clear_bit(idx, mle->node_map); idx 376 fs/ocfs2/dlm/dlmmaster.c struct o2nm_node *node, int idx) idx 380 fs/ocfs2/dlm/dlmmaster.c if (test_bit(idx, mle->node_map)) idx 381 fs/ocfs2/dlm/dlmmaster.c mlog(0, "node %u already in node map!\n", idx); idx 383 fs/ocfs2/dlm/dlmmaster.c set_bit(idx, mle->node_map); idx 2487 fs/ocfs2/dlm/dlmmaster.c enum dlm_lockres_list idx; idx 2507 fs/ocfs2/dlm/dlmmaster.c for (idx = DLM_GRANTED_LIST; idx <= DLM_BLOCKED_LIST; idx++) { idx 2508 fs/ocfs2/dlm/dlmmaster.c queue = dlm_list_idx_to_ptr(res, idx); idx 2520 fs/ocfs2/dlm/dlmmaster.c dlm_list_in_text(idx)); idx 2984 fs/ocfs2/dlm/dlmmaster.c enum dlm_lockres_list idx; idx 2994 fs/ocfs2/dlm/dlmmaster.c for (idx = DLM_GRANTED_LIST; idx <= DLM_BLOCKED_LIST; idx++) { idx 2995 fs/ocfs2/dlm/dlmmaster.c queue = dlm_list_idx_to_ptr(res, idx); idx 2429 fs/ocfs2/dlm/dlmrecovery.c static void __dlm_hb_node_down(struct dlm_ctxt *dlm, int idx) idx 2433 fs/ocfs2/dlm/dlmrecovery.c if (dlm->reco.new_master == idx) { idx 2435 fs/ocfs2/dlm/dlmrecovery.c dlm->name, idx); idx 2441 fs/ocfs2/dlm/dlmrecovery.c "finalize1 state, clearing\n", dlm->name, idx); idx 2448 fs/ocfs2/dlm/dlmrecovery.c if (dlm->joining_node == idx) { idx 2449 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "Clearing join state for node %u\n", idx); idx 2454 fs/ocfs2/dlm/dlmrecovery.c if (!test_bit(idx, dlm->live_nodes_map)) { idx 2457 fs/ocfs2/dlm/dlmrecovery.c dlm->name, idx); idx 2462 fs/ocfs2/dlm/dlmrecovery.c if (!test_bit(idx, dlm->domain_map)) { idx 2465 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u already removed from domain!\n", idx); idx 2469 fs/ocfs2/dlm/dlmrecovery.c clear_bit(idx, dlm->live_nodes_map); idx 2472 fs/ocfs2/dlm/dlmrecovery.c if (!test_bit(idx, dlm->recovery_map)) idx 2473 fs/ocfs2/dlm/dlmrecovery.c dlm_do_local_recovery_cleanup(dlm, idx); idx 2476 fs/ocfs2/dlm/dlmrecovery.c dlm_hb_event_notify_attached(dlm, idx, 0); idx 2478 fs/ocfs2/dlm/dlmrecovery.c mlog(0, "node %u being removed from domain map!\n", idx); idx 2479 fs/ocfs2/dlm/dlmrecovery.c clear_bit(idx, dlm->domain_map); idx 2480 fs/ocfs2/dlm/dlmrecovery.c clear_bit(idx, dlm->exit_domain_map); idx 2485 fs/ocfs2/dlm/dlmrecovery.c set_bit(idx, dlm->recovery_map); idx 2488 fs/ocfs2/dlm/dlmrecovery.c void dlm_hb_node_down_cb(struct o2nm_node *node, int idx, void *data) idx 2499 fs/ocfs2/dlm/dlmrecovery.c if (test_bit(idx, dlm->domain_map)) idx 2500 fs/ocfs2/dlm/dlmrecovery.c dlm_fire_domain_eviction_callbacks(dlm, idx); idx 2503 fs/ocfs2/dlm/dlmrecovery.c __dlm_hb_node_down(dlm, idx); idx 2509 fs/ocfs2/dlm/dlmrecovery.c void dlm_hb_node_up_cb(struct o2nm_node *node, int idx, void *data) idx 2517 fs/ocfs2/dlm/dlmrecovery.c set_bit(idx, dlm->live_nodes_map); idx 764 fs/ocfs2/ioctl.c static int ocfs2_get_request_ptr(struct ocfs2_info *info, int idx, idx 783 fs/ocfs2/ioctl.c if (o2info_from_user(*req_addr, bp + idx)) idx 94 fs/overlayfs/export.c return oe->lowerstack[0].layer->idx; idx 115 fs/overlayfs/export.c origin_layer = OVL_E(dentry)->lowerstack[0].layer->idx; idx 346 fs/overlayfs/export.c static struct dentry *ovl_dentry_real_at(struct dentry *dentry, int idx) idx 351 fs/overlayfs/export.c if (!idx) idx 355 fs/overlayfs/export.c if (oe->lowerstack[i].layer->idx == idx) idx 387 fs/overlayfs/export.c if (ovl_dentry_real_at(connected, layer->idx) != parent) idx 405 fs/overlayfs/export.c } else if (ovl_dentry_real_at(this, layer->idx) != real) { idx 419 fs/overlayfs/export.c real, layer->idx, connected, err); idx 445 fs/overlayfs/export.c inode = ovl_lookup_inode(sb, real, !layer->idx); idx 457 fs/overlayfs/export.c if (!this && layer->idx && ofs->indexdir && !WARN_ON(!d_is_dir(real))) { idx 485 fs/overlayfs/export.c if (WARN_ON(ovl_dentry_real_at(this, layer->idx) != real)) { idx 565 fs/overlayfs/export.c layer->idx); idx 644 fs/overlayfs/export.c real, layer->idx, connected, err); idx 774 fs/overlayfs/namei.c int ovl_path_next(int idx, struct dentry *dentry, struct path *path) idx 778 fs/overlayfs/namei.c BUG_ON(idx < 0); idx 779 fs/overlayfs/namei.c if (idx == 0) { idx 783 fs/overlayfs/namei.c idx++; idx 785 fs/overlayfs/namei.c BUG_ON(idx > oe->numlower); idx 786 fs/overlayfs/namei.c path->dentry = oe->lowerstack[idx - 1].dentry; idx 787 fs/overlayfs/namei.c path->mnt = oe->lowerstack[idx - 1].layer->mnt; idx 789 fs/overlayfs/namei.c return (idx < oe->numlower) ? idx + 1 : -1; idx 906 fs/overlayfs/namei.c d.last = lower.layer->idx == roe->numlower; idx 987 fs/overlayfs/namei.c i = lower.layer->idx - 1; idx 318 fs/overlayfs/overlayfs.h int ovl_path_next(int idx, struct dentry *dentry, struct path *path); idx 35 fs/overlayfs/ovl_entry.h int idx; idx 367 fs/overlayfs/readdir.c int idx, next; idx 369 fs/overlayfs/readdir.c for (idx = 0; idx != -1; idx = next) { idx 370 fs/overlayfs/readdir.c next = ovl_path_next(idx, dentry, &realpath); idx 1373 fs/overlayfs/super.c ofs->lower_layers[ofs->numlower].idx = i + 1; idx 301 fs/splice.c int idx, ret; idx 304 fs/splice.c idx = to.idx; idx 312 fs/splice.c to.idx = idx; idx 403 fs/splice.c pipe->bufs[to.idx].offset = offset; idx 404 fs/splice.c pipe->bufs[to.idx].len -= offset; idx 703 fs/splice.c int n, idx; idx 722 fs/splice.c for (n = 0, idx = pipe->curbuf; left && n < pipe->nrbufs; n++, idx++) { idx 723 fs/splice.c struct pipe_buffer *buf = pipe->bufs + idx; idx 729 fs/splice.c if (idx == pipe->buffers - 1) idx 730 fs/splice.c idx = -1; idx 501 fs/ubifs/commit.c struct ubifs_idx_node idx __aligned(8); idx 516 fs/ubifs/commit.c struct ubifs_idx_node *idx; idx 525 fs/ubifs/commit.c idx = kmalloc(c->max_idx_node_sz, GFP_NOFS); idx 526 fs/ubifs/commit.c if (!idx) idx 529 fs/ubifs/commit.c err = ubifs_read_node(c, idx, UBIFS_IDX_NODE, len, lnum, offs); idx 533 fs/ubifs/commit.c d->old_zroot_level = le16_to_cpu(idx->level); idx 534 fs/ubifs/commit.c d->old_zroot_sqnum = le64_to_cpu(idx->ch.sqnum); idx 536 fs/ubifs/commit.c kfree(idx); idx 560 fs/ubifs/commit.c struct ubifs_idx_node *idx; idx 596 fs/ubifs/commit.c idx = &i->idx; idx 597 fs/ubifs/commit.c err = ubifs_read_node(c, idx, UBIFS_IDX_NODE, len, lnum, offs); idx 601 fs/ubifs/commit.c child_cnt = le16_to_cpu(idx->child_cnt); idx 609 fs/ubifs/commit.c if (le16_to_cpu(idx->level) != d->old_zroot_level) { idx 613 fs/ubifs/commit.c if (le64_to_cpu(idx->ch.sqnum) != d->old_zroot_sqnum) { idx 618 fs/ubifs/commit.c last_level = le16_to_cpu(idx->level) + 1; idx 619 fs/ubifs/commit.c last_sqnum = le64_to_cpu(idx->ch.sqnum) + 1; idx 620 fs/ubifs/commit.c key_read(c, ubifs_idx_key(c, idx), &lower_key); idx 624 fs/ubifs/commit.c if (le16_to_cpu(idx->level) != last_level - 1) { idx 632 fs/ubifs/commit.c if (le64_to_cpu(idx->ch.sqnum) >= last_sqnum) { idx 637 fs/ubifs/commit.c key_read(c, ubifs_idx_key(c, idx), &l_key); idx 638 fs/ubifs/commit.c br = ubifs_idx_branch(c, idx, child_cnt - 1); idx 654 fs/ubifs/commit.c if (le16_to_cpu(idx->level) == 0) { idx 666 fs/ubifs/commit.c idx = &i->idx; idx 668 fs/ubifs/commit.c if (iip + 1 < le16_to_cpu(idx->child_cnt)) { idx 682 fs/ubifs/commit.c last_level = le16_to_cpu(idx->level); idx 683 fs/ubifs/commit.c last_sqnum = le64_to_cpu(idx->ch.sqnum); idx 684 fs/ubifs/commit.c br = ubifs_idx_branch(c, idx, iip); idx 689 fs/ubifs/commit.c if (iip + 1 < le16_to_cpu(idx->child_cnt)) { idx 690 fs/ubifs/commit.c br = ubifs_idx_branch(c, idx, iip + 1); idx 704 fs/ubifs/commit.c ubifs_dump_node(c, idx); idx 710 fs/ubifs/commit.c ubifs_dump_node(c, &i->idx); idx 497 fs/ubifs/debug.c const struct ubifs_idx_node *idx = node; idx 499 fs/ubifs/debug.c n = le16_to_cpu(idx->child_cnt); idx 501 fs/ubifs/debug.c pr_err("\tlevel %d\n", (int)le16_to_cpu(idx->level)); idx 507 fs/ubifs/debug.c br = ubifs_idx_branch(c, idx, i); idx 1601 fs/ubifs/debug.c int idx; idx 1615 fs/ubifs/debug.c for (idx = 0; idx < znode->child_cnt; idx++) { idx 1616 fs/ubifs/debug.c zbr = &znode->zbranch[idx]; idx 1629 fs/ubifs/debug.c idx = znode->iip + 1; idx 1631 fs/ubifs/debug.c if (idx < znode->child_cnt) { idx 1633 fs/ubifs/debug.c zbr = &znode->zbranch[idx]; idx 1636 fs/ubifs/debug.c child = ubifs_load_znode(c, zbr, znode, idx); idx 554 fs/ubifs/gc.c struct ubifs_idx_node *idx = snod->node; idx 555 fs/ubifs/gc.c int level = le16_to_cpu(idx->level); idx 558 fs/ubifs/gc.c key_read(c, ubifs_idx_key(c, idx), &snod->key); idx 1128 fs/ubifs/lprops.c struct ubifs_idx_node *idx = snod->node; idx 1130 fs/ubifs/lprops.c key_read(c, ubifs_idx_key(c, idx), &snod->key); idx 1131 fs/ubifs/lprops.c level = le16_to_cpu(idx->level); idx 200 fs/ubifs/misc.h const struct ubifs_idx_node *idx, idx 203 fs/ubifs/misc.h return (struct ubifs_branch *)((void *)idx->branches + idx 213 fs/ubifs/misc.h const struct ubifs_idx_node *idx) idx 215 fs/ubifs/misc.h return (void *)((struct ubifs_branch *)idx->branches)->key; idx 76 fs/ubifs/sb.c struct ubifs_idx_node *idx; idx 164 fs/ubifs/sb.c idx = kzalloc(ALIGN(idx_node_size, c->min_io_size), GFP_KERNEL); idx 168 fs/ubifs/sb.c if (!sup || !mst || !idx || !ino || !cs) { idx 276 fs/ubifs/sb.c idx->ch.node_type = UBIFS_IDX_NODE; idx 277 fs/ubifs/sb.c idx->child_cnt = cpu_to_le16(1); idx 279 fs/ubifs/sb.c br = ubifs_idx_branch(c, idx, 0); idx 334 fs/ubifs/sb.c err = ubifs_write_node(c, idx, idx_node_size, main_first + DEFAULT_IDX_LEB, 0); idx 338 fs/ubifs/sb.c ubifs_node_calc_hash(c, idx, hash); idx 361 fs/ubifs/sb.c kfree(idx); idx 25 fs/ubifs/tnc_commit.c static int make_idx_node(struct ubifs_info *c, struct ubifs_idx_node *idx, idx 33 fs/ubifs/tnc_commit.c idx->ch.node_type = UBIFS_IDX_NODE; idx 34 fs/ubifs/tnc_commit.c idx->child_cnt = cpu_to_le16(znode->child_cnt); idx 35 fs/ubifs/tnc_commit.c idx->level = cpu_to_le16(znode->level); idx 37 fs/ubifs/tnc_commit.c struct ubifs_branch *br = ubifs_idx_branch(c, idx, i); idx 54 fs/ubifs/tnc_commit.c ubifs_prepare_node(c, idx, len, 0); idx 55 fs/ubifs/tnc_commit.c ubifs_node_calc_hash(c, idx, hash); idx 252 fs/ubifs/tnc_commit.c struct ubifs_idx_node *idx; idx 256 fs/ubifs/tnc_commit.c idx = snod->node; idx 257 fs/ubifs/tnc_commit.c key_read(c, ubifs_idx_key(c, idx), &snod->key); idx 258 fs/ubifs/tnc_commit.c level = le16_to_cpu(idx->level); idx 826 fs/ubifs/tnc_commit.c struct ubifs_idx_node *idx; idx 863 fs/ubifs/tnc_commit.c idx = c->cbuf + used; idx 866 fs/ubifs/tnc_commit.c idx->ch.node_type = UBIFS_IDX_NODE; idx 867 fs/ubifs/tnc_commit.c idx->child_cnt = cpu_to_le16(znode->child_cnt); idx 868 fs/ubifs/tnc_commit.c idx->level = cpu_to_le16(znode->level); idx 870 fs/ubifs/tnc_commit.c struct ubifs_branch *br = ubifs_idx_branch(c, idx, i); idx 888 fs/ubifs/tnc_commit.c ubifs_prepare_node(c, idx, len, 0); idx 889 fs/ubifs/tnc_commit.c ubifs_node_calc_hash(c, idx, hash); idx 272 fs/ubifs/tnc_misc.c struct ubifs_idx_node *idx; idx 274 fs/ubifs/tnc_misc.c idx = kmalloc(c->max_idx_node_sz, GFP_NOFS); idx 275 fs/ubifs/tnc_misc.c if (!idx) idx 278 fs/ubifs/tnc_misc.c err = ubifs_read_node(c, idx, UBIFS_IDX_NODE, len, lnum, offs); idx 280 fs/ubifs/tnc_misc.c kfree(idx); idx 284 fs/ubifs/tnc_misc.c err = ubifs_node_check_hash(c, idx, zzbr->hash); idx 286 fs/ubifs/tnc_misc.c ubifs_bad_hash(c, idx, zzbr->hash, lnum, offs); idx 287 fs/ubifs/tnc_misc.c kfree(idx); idx 291 fs/ubifs/tnc_misc.c znode->child_cnt = le16_to_cpu(idx->child_cnt); idx 292 fs/ubifs/tnc_misc.c znode->level = le16_to_cpu(idx->level); idx 307 fs/ubifs/tnc_misc.c struct ubifs_branch *br = ubifs_idx_branch(c, idx, i); idx 388 fs/ubifs/tnc_misc.c kfree(idx); idx 393 fs/ubifs/tnc_misc.c ubifs_dump_node(c, idx); idx 394 fs/ubifs/tnc_misc.c kfree(idx); idx 158 fs/udf/unicode.c int idx, len; idx 203 fs/udf/unicode.c for (idx = ocu_len - u_ch, ext_i_len = 0; idx 204 fs/udf/unicode.c (idx >= 0) && (ext_i_len < EXT_SIZE); idx 205 fs/udf/unicode.c idx -= u_ch, ext_i_len++) { idx 206 fs/udf/unicode.c c = ocu[idx]; idx 208 fs/udf/unicode.c c = (c << 8) | ocu[idx + 1]; idx 212 fs/udf/unicode.c i_ext = idx; idx 220 fs/udf/unicode.c idx = i_ext + u_ch; idx 222 fs/udf/unicode.c ocu, ocu_len, &idx, idx 231 fs/udf/unicode.c idx = 0; idx 233 fs/udf/unicode.c if (translate && (idx == i_ext)) { idx 240 fs/udf/unicode.c ocu, ocu_len, &idx, idx 794 fs/xfs/libxfs/xfs_da_format.h xfs_attr3_leaf_name(xfs_attr_leafblock_t *leafp, int idx) idx 798 fs/xfs/libxfs/xfs_da_format.h return &((char *)leafp)[be16_to_cpu(entries[idx].nameidx)]; idx 802 fs/xfs/libxfs/xfs_da_format.h xfs_attr3_leaf_name_remote(xfs_attr_leafblock_t *leafp, int idx) idx 804 fs/xfs/libxfs/xfs_da_format.h return (xfs_attr_leaf_name_remote_t *)xfs_attr3_leaf_name(leafp, idx); idx 808 fs/xfs/libxfs/xfs_da_format.h xfs_attr3_leaf_name_local(xfs_attr_leafblock_t *leafp, int idx) idx 810 fs/xfs/libxfs/xfs_da_format.h return (xfs_attr_leaf_name_local_t *)xfs_attr3_leaf_name(leafp, idx); idx 249 fs/xfs/scrub/attr.c int idx, idx 278 fs/xfs/scrub/attr.c lentry = xfs_attr3_leaf_name_local(leaf, idx); idx 285 fs/xfs/scrub/attr.c rentry = xfs_attr3_leaf_name_remote(leaf, idx); idx 2107 fs/xfs/xfs_buf.c int idx; idx 2109 fs/xfs/xfs_buf.c idx = xfs_sb_version_hascrc(&mp->m_sb); idx 2110 fs/xfs/xfs_buf.c if (WARN_ON(!bp->b_ops || !bp->b_ops->magic[idx])) idx 2112 fs/xfs/xfs_buf.c return dmagic == bp->b_ops->magic[idx]; idx 2125 fs/xfs/xfs_buf.c int idx; idx 2127 fs/xfs/xfs_buf.c idx = xfs_sb_version_hascrc(&mp->m_sb); idx 2128 fs/xfs/xfs_buf.c if (WARN_ON(!bp->b_ops || !bp->b_ops->magic16[idx])) idx 2130 fs/xfs/xfs_buf.c return dmagic == bp->b_ops->magic16[idx]; idx 1837 fs/xfs/xfs_ioctl.c unsigned int idx; idx 1851 fs/xfs/xfs_ioctl.c if (copy_to_user(&info->data->fmh_recs[info->idx++], &fm, idx 1896 fs/xfs/xfs_ioctl.c if (!aborted && info.idx) { idx 1898 fs/xfs/xfs_ioctl.c if (copy_to_user(&info.data->fmh_recs[info.idx - 1].fmr_flags, idx 125 fs/xfs/xfs_iwalk.c int idx; /* index into inode chunk */ idx 128 fs/xfs/xfs_iwalk.c idx = agino - irec->ir_startino; idx 135 fs/xfs/xfs_iwalk.c for (i = 0; i < idx; i++) { idx 140 fs/xfs/xfs_iwalk.c irec->ir_free |= xfs_inobt_maskn(0, idx); idx 3814 fs/xfs/xfs_log.c int idx; idx 3852 fs/xfs/xfs_log.c idx = BTOBBT((char *)&ophead->oh_clientid - iclog->ic_datap); idx 3853 fs/xfs/xfs_log.c if (idx >= (XLOG_HEADER_CYCLE_SIZE / BBSIZE)) { idx 3854 fs/xfs/xfs_log.c j = idx / (XLOG_HEADER_CYCLE_SIZE / BBSIZE); idx 3855 fs/xfs/xfs_log.c k = idx % (XLOG_HEADER_CYCLE_SIZE / BBSIZE); idx 3860 fs/xfs/xfs_log.c iclog->ic_header.h_cycle_data[idx]); idx 3875 fs/xfs/xfs_log.c idx = BTOBBT((uintptr_t)&ophead->oh_len - idx 3877 fs/xfs/xfs_log.c if (idx >= (XLOG_HEADER_CYCLE_SIZE / BBSIZE)) { idx 3878 fs/xfs/xfs_log.c j = idx / (XLOG_HEADER_CYCLE_SIZE / BBSIZE); idx 3879 fs/xfs/xfs_log.c k = idx % (XLOG_HEADER_CYCLE_SIZE / BBSIZE); idx 3882 fs/xfs/xfs_log.c op_len = be32_to_cpu(iclog->ic_header.h_cycle_data[idx]); idx 10 fs/xfs/xfs_stats.c static int counter_val(struct xfsstats __percpu *stats, int idx) idx 15 fs/xfs/xfs_stats.c val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx)); idx 1840 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_da_args *args, int idx), idx 1841 fs/xfs/xfs_trace.h TP_ARGS(args, idx), idx 1846 fs/xfs/xfs_trace.h __field(int, idx) idx 1852 fs/xfs/xfs_trace.h __entry->idx = idx; idx 1858 fs/xfs/xfs_trace.h __entry->idx) idx 1863 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_da_args *args, int idx), \ idx 1864 fs/xfs/xfs_trace.h TP_ARGS(args, idx)) idx 30 include/asm-generic/fixmap.h static __always_inline unsigned long fix_to_virt(const unsigned int idx) idx 32 include/asm-generic/fixmap.h BUILD_BUG_ON(idx >= __end_of_fixed_addresses); idx 33 include/asm-generic/fixmap.h return __fix_to_virt(idx); idx 64 include/asm-generic/fixmap.h #define set_fixmap(idx, phys) \ idx 65 include/asm-generic/fixmap.h __set_fixmap(idx, phys, FIXMAP_PAGE_NORMAL) idx 69 include/asm-generic/fixmap.h #define clear_fixmap(idx) \ idx 70 include/asm-generic/fixmap.h __set_fixmap(idx, 0, FIXMAP_PAGE_CLEAR) idx 74 include/asm-generic/fixmap.h #define __set_fixmap_offset(idx, phys, flags) \ idx 77 include/asm-generic/fixmap.h __set_fixmap(idx, phys, flags); \ idx 78 include/asm-generic/fixmap.h ________addr = fix_to_virt(idx) + ((phys) & (PAGE_SIZE - 1)); \ idx 82 include/asm-generic/fixmap.h #define set_fixmap_offset(idx, phys) \ idx 83 include/asm-generic/fixmap.h __set_fixmap_offset(idx, phys, FIXMAP_PAGE_NORMAL) idx 88 include/asm-generic/fixmap.h #define set_fixmap_nocache(idx, phys) \ idx 89 include/asm-generic/fixmap.h __set_fixmap(idx, phys, FIXMAP_PAGE_NOCACHE) idx 91 include/asm-generic/fixmap.h #define set_fixmap_offset_nocache(idx, phys) \ idx 92 include/asm-generic/fixmap.h __set_fixmap_offset(idx, phys, FIXMAP_PAGE_NOCACHE) idx 97 include/asm-generic/fixmap.h #define set_fixmap_io(idx, phys) \ idx 98 include/asm-generic/fixmap.h __set_fixmap(idx, phys, FIXMAP_PAGE_IO) idx 100 include/asm-generic/fixmap.h #define set_fixmap_offset_io(idx, phys) \ idx 101 include/asm-generic/fixmap.h __set_fixmap_offset(idx, phys, FIXMAP_PAGE_IO) idx 1177 include/drm/drm_crtc.h struct drm_crtc *drm_crtc_from_index(struct drm_device *dev, int idx); idx 95 include/drm/drm_displayid.h #define for_each_displayid_db(displayid, block, idx, length) \ idx 96 include/drm/drm_displayid.h for ((block) = (struct displayid_block *)&(displayid)[idx]; \ idx 97 include/drm/drm_displayid.h (idx) + sizeof(struct displayid_block) <= (length) && \ idx 98 include/drm/drm_displayid.h (idx) + sizeof(struct displayid_block) + (block)->num_bytes <= (length) && \ idx 100 include/drm/drm_displayid.h (idx) += (block)->num_bytes + sizeof(struct displayid_block), \ idx 101 include/drm/drm_displayid.h (block) = (struct displayid_block *)&(displayid)[idx]) idx 799 include/drm/drm_drv.h bool drm_dev_enter(struct drm_device *dev, int *idx); idx 800 include/drm/drm_drv.h void drm_dev_exit(int idx); idx 818 include/drm/drm_drv.h int idx; idx 820 include/drm/drm_drv.h if (drm_dev_enter(dev, &idx)) { idx 821 include/drm/drm_drv.h drm_dev_exit(idx); idx 54 include/drm/drm_legacy.h int idx; /**< Index into master buflist */ idx 751 include/drm/drm_plane.h struct drm_plane * drm_plane_from_index(struct drm_device *dev, int idx); idx 19 include/kvm/arm_pmu.h u8 idx; /* index into the pmu->pmc array */ idx 127 include/linux/bio.h if (iter->idx >= bio->bi_vcnt) idx 130 include/linux/bio.h bvec_advance(&bio->bi_io_vec[iter->idx], iter); idx 255 include/linux/bio.h int idx; idx 265 include/linux/bio.h idx = iter.bi_idx - 1; idx 267 include/linux/bio.h idx = iter.bi_idx; idx 269 include/linux/bio.h *bv = bio->bi_io_vec[idx]; idx 482 include/linux/bio.h extern unsigned int bvec_nr_vecs(unsigned short idx); idx 206 include/linux/blk-cgroup.h unsigned int idx) idx 208 include/linux/blk-cgroup.h return atomic64_read(&rwstat->aux_cnt[idx]) + idx 209 include/linux/blk-cgroup.h percpu_counter_sum_positive(&rwstat->cpu_cnt[idx]); idx 194 include/linux/bpf_verifier.h u32 idx; idx 37 include/linux/bvec.h int idx; idx 131 include/linux/bvec.h iter_all->idx = 0; idx 153 include/linux/bvec.h iter_all->idx++; idx 205 include/linux/can/dev.h unsigned int idx); idx 206 include/linux/can/dev.h struct sk_buff *__can_get_echo_skb(struct net_device *dev, unsigned int idx, idx 208 include/linux/can/dev.h unsigned int can_get_echo_skb(struct net_device *dev, unsigned int idx); idx 209 include/linux/can/dev.h void can_free_echo_skb(struct net_device *dev, unsigned int idx); idx 112 include/linux/can/dev/peak_canfd.h __le16 idx; idx 123 include/linux/can/dev/peak_canfd.h u8 idx; idx 44 include/linux/can/rx-offload.h unsigned int idx, u32 timestamp); idx 152 include/linux/console.h int (*match)(struct console *, char *name, int idx, char *options); idx 174 include/linux/console.h extern int add_preferred_console(char *name, int idx, char *options); idx 666 include/linux/cpufreq.h #define cpufreq_for_each_entry_idx(pos, table, idx) \ idx 667 include/linux/cpufreq.h for (pos = table, idx = 0; pos->frequency != CPUFREQ_TABLE_END; \ idx 668 include/linux/cpufreq.h pos++, idx++) idx 691 include/linux/cpufreq.h #define cpufreq_for_each_valid_entry_idx(pos, table, idx) \ idx 692 include/linux/cpufreq.h cpufreq_for_each_entry_idx(pos, table, idx) \ idx 726 include/linux/cpufreq.h int idx, best = -1; idx 728 include/linux/cpufreq.h cpufreq_for_each_valid_entry_idx(pos, table, idx) { idx 732 include/linux/cpufreq.h return idx; idx 734 include/linux/cpufreq.h best = idx; idx 747 include/linux/cpufreq.h int idx, best = -1; idx 749 include/linux/cpufreq.h cpufreq_for_each_valid_entry_idx(pos, table, idx) { idx 753 include/linux/cpufreq.h return idx; idx 756 include/linux/cpufreq.h best = idx; idx 762 include/linux/cpufreq.h return idx; idx 789 include/linux/cpufreq.h int idx, best = -1; idx 791 include/linux/cpufreq.h cpufreq_for_each_valid_entry_idx(pos, table, idx) { idx 795 include/linux/cpufreq.h return idx; idx 798 include/linux/cpufreq.h best = idx; idx 804 include/linux/cpufreq.h return idx; idx 819 include/linux/cpufreq.h int idx, best = -1; idx 821 include/linux/cpufreq.h cpufreq_for_each_valid_entry_idx(pos, table, idx) { idx 825 include/linux/cpufreq.h return idx; idx 827 include/linux/cpufreq.h best = idx; idx 852 include/linux/cpufreq.h int idx, best = -1; idx 854 include/linux/cpufreq.h cpufreq_for_each_valid_entry_idx(pos, table, idx) { idx 858 include/linux/cpufreq.h return idx; idx 861 include/linux/cpufreq.h best = idx; idx 867 include/linux/cpufreq.h return idx; idx 871 include/linux/cpufreq.h return idx; idx 886 include/linux/cpufreq.h int idx, best = -1; idx 888 include/linux/cpufreq.h cpufreq_for_each_valid_entry_idx(pos, table, idx) { idx 892 include/linux/cpufreq.h return idx; idx 895 include/linux/cpufreq.h best = idx; idx 901 include/linux/cpufreq.h return idx; idx 905 include/linux/cpufreq.h return idx; idx 270 include/linux/cpuidle.h idx, \ idx 276 include/linux/cpuidle.h if (!idx) { \ idx 278 include/linux/cpuidle.h return idx; \ idx 289 include/linux/cpuidle.h __ret ? -1 : idx; \ idx 292 include/linux/cpuidle.h #define CPU_PM_CPU_IDLE_ENTER(low_level_idle_enter, idx) \ idx 293 include/linux/cpuidle.h __CPU_PM_CPU_IDLE_ENTER(low_level_idle_enter, idx, idx, 0) idx 295 include/linux/cpuidle.h #define CPU_PM_CPU_IDLE_ENTER_RETENTION(low_level_idle_enter, idx) \ idx 296 include/linux/cpuidle.h __CPU_PM_CPU_IDLE_ENTER(low_level_idle_enter, idx, idx, 1) idx 298 include/linux/cpuidle.h #define CPU_PM_CPU_IDLE_ENTER_PARAM(low_level_idle_enter, idx, state) \ idx 299 include/linux/cpuidle.h __CPU_PM_CPU_IDLE_ENTER(low_level_idle_enter, idx, state, 0) idx 301 include/linux/cpuidle.h #define CPU_PM_CPU_IDLE_ENTER_RETENTION_PARAM(low_level_idle_enter, idx, state) \ idx 302 include/linux/cpuidle.h __CPU_PM_CPU_IDLE_ENTER(low_level_idle_enter, idx, state, 1) idx 543 include/linux/dma-fence.h uint32_t *idx); idx 617 include/linux/fb.h extern void fb_pad_unaligned_buffer(u8 *dst, u32 d_pitch, u8 *src, u32 idx, idx 22 include/linux/firmware/imx/dsp.h int idx; idx 17 include/linux/font.h int idx; idx 785 include/linux/ftrace.h ftrace_graph_get_ret_stack(struct task_struct *task, int idx); idx 787 include/linux/ftrace.h unsigned long ftrace_graph_ret_addr(struct task_struct *task, int *idx, idx 836 include/linux/ftrace.h ftrace_graph_ret_addr(struct task_struct *task, int *idx, unsigned long ret, idx 99 include/linux/generic-radix-tree.h static inline size_t __idx_to_offset(size_t idx, size_t obj_size) idx 109 include/linux/generic-radix-tree.h return (idx / objs_per_page) * PAGE_SIZE + idx 110 include/linux/generic-radix-tree.h (idx % objs_per_page) * obj_size; idx 112 include/linux/generic-radix-tree.h return idx * obj_size; idx 278 include/linux/genhd.h int idx; idx 69 include/linux/gpio/consumer.h unsigned int idx, idx 92 include/linux/gpio/consumer.h unsigned int idx, idx 202 include/linux/gpio/consumer.h unsigned int idx, idx 272 include/linux/gpio/consumer.h unsigned int idx, idx 36 include/linux/gpio/machine.h unsigned int idx; idx 79 include/linux/gpio/machine.h .idx = _idx, \ idx 118 include/linux/highmem.h int idx = __this_cpu_inc_return(__kmap_atomic_idx) - 1; idx 122 include/linux/highmem.h BUG_ON(idx >= KM_TYPE_NR); idx 124 include/linux/highmem.h return idx; idx 135 include/linux/highmem.h int idx = __this_cpu_dec_return(__kmap_atomic_idx); idx 137 include/linux/highmem.h BUG_ON(idx < 0); idx 66 include/linux/hp_sdc.h int idx; /* Index within the act */ idx 109 include/linux/hugetlb.h pgoff_t idx, unsigned long address); idx 370 include/linux/hugetlb.h pgoff_t idx); idx 54 include/linux/hugetlb_cgroup.h extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, idx 56 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, idx 59 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, idx 61 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, idx 85 include/linux/hugetlb_cgroup.h hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, idx 92 include/linux/hugetlb_cgroup.h hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, idx 99 include/linux/hugetlb_cgroup.h hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) idx 104 include/linux/hugetlb_cgroup.h hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, idx 319 include/linux/intel-iommu.h #define QI_IEC_IIDEX(idx) (((u64)(idx & 0xffff) << 32)) idx 379 include/linux/intel-iommu.h #define QI_PGRP_IDX(idx) (((u64)(idx)) << 3) idx 66 include/linux/isapnp.h unsigned char isapnp_read_byte(unsigned char idx); idx 67 include/linux/isapnp.h void isapnp_write_byte(unsigned char idx, unsigned char val); idx 92 include/linux/isapnp.h static inline unsigned char isapnp_read_byte(unsigned char idx) { return 0xff; } idx 93 include/linux/isapnp.h static inline void isapnp_write_byte(unsigned char idx, unsigned char val) { ; } idx 532 include/linux/kvm_host.h static inline struct kvm_io_bus *kvm_get_bus(struct kvm *kvm, enum kvm_bus idx) idx 534 include/linux/kvm_host.h return srcu_dereference_check(kvm->buses[idx], &kvm->srcu, idx 549 include/linux/kvm_host.h #define kvm_for_each_vcpu(idx, vcpup, kvm) \ idx 550 include/linux/kvm_host.h for (idx = 0; \ idx 551 include/linux/kvm_host.h idx < atomic_read(&kvm->online_vcpus) && \ idx 552 include/linux/kvm_host.h (vcpup = kvm_get_vcpu(kvm, idx)) != NULL; \ idx 553 include/linux/kvm_host.h idx++) idx 575 include/linux/kvm_host.h int idx; idx 577 include/linux/kvm_host.h kvm_for_each_vcpu(idx, tmp, vcpu->kvm) idx 579 include/linux/kvm_host.h return idx; idx 245 include/linux/libnvdimm.h const struct nd_cmd_desc *desc, int idx, void *buf); idx 247 include/linux/libnvdimm.h const struct nd_cmd_desc *desc, int idx, const u32 *in_field, idx 134 include/linux/memblock.h void __next_mem_range(u64 *idx, int nid, enum memblock_flags flags, idx 139 include/linux/memblock.h void __next_mem_range_rev(u64 *idx, int nid, enum memblock_flags flags, idx 144 include/linux/memblock.h void __next_reserved_mem_region(u64 *idx, phys_addr_t *out_start, idx 222 include/linux/memblock.h void __next_mem_pfn_range(int *idx, int nid, unsigned long *out_start_pfn, idx 241 include/linux/memblock.h void __next_mem_pfn_range_in_zone(u64 *idx, struct zone *zone, idx 594 include/linux/memcontrol.h static inline unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx) idx 596 include/linux/memcontrol.h long x = atomic_long_read(&memcg->vmstats[idx]); idx 609 include/linux/memcontrol.h int idx) idx 615 include/linux/memcontrol.h x += per_cpu(memcg->vmstats_local->stat[idx], cpu); idx 623 include/linux/memcontrol.h void __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val); idx 627 include/linux/memcontrol.h int idx, int val) idx 632 include/linux/memcontrol.h __mod_memcg_state(memcg, idx, val); idx 654 include/linux/memcontrol.h int idx, int val) idx 657 include/linux/memcontrol.h __mod_memcg_state(page->mem_cgroup, idx, val); idx 661 include/linux/memcontrol.h int idx, int val) idx 664 include/linux/memcontrol.h mod_memcg_state(page->mem_cgroup, idx, val); idx 668 include/linux/memcontrol.h enum node_stat_item idx) idx 674 include/linux/memcontrol.h return node_page_state(lruvec_pgdat(lruvec), idx); idx 677 include/linux/memcontrol.h x = atomic_long_read(&pn->lruvec_stat[idx]); idx 686 include/linux/memcontrol.h enum node_stat_item idx) idx 693 include/linux/memcontrol.h return node_page_state(lruvec_pgdat(lruvec), idx); idx 697 include/linux/memcontrol.h x += per_cpu(pn->lruvec_stat_local->count[idx], cpu); idx 705 include/linux/memcontrol.h void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, idx 707 include/linux/memcontrol.h void __mod_lruvec_slab_state(void *p, enum node_stat_item idx, int val); idx 708 include/linux/memcontrol.h void mod_memcg_obj_state(void *p, int idx, int val); idx 711 include/linux/memcontrol.h enum node_stat_item idx, int val) idx 716 include/linux/memcontrol.h __mod_lruvec_state(lruvec, idx, val); idx 721 include/linux/memcontrol.h enum node_stat_item idx, int val) idx 728 include/linux/memcontrol.h __mod_node_page_state(pgdat, idx, val); idx 733 include/linux/memcontrol.h __mod_lruvec_state(lruvec, idx, val); idx 737 include/linux/memcontrol.h enum node_stat_item idx, int val) idx 742 include/linux/memcontrol.h __mod_lruvec_page_state(page, idx, val); idx 750 include/linux/memcontrol.h void __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, idx 754 include/linux/memcontrol.h enum vm_event_item idx, idx 760 include/linux/memcontrol.h __count_memcg_events(memcg, idx, count); idx 765 include/linux/memcontrol.h enum vm_event_item idx) idx 768 include/linux/memcontrol.h count_memcg_events(page->mem_cgroup, idx, 1); idx 772 include/linux/memcontrol.h enum vm_event_item idx) idx 782 include/linux/memcontrol.h count_memcg_events(memcg, idx, 1); idx 1055 include/linux/memcontrol.h static inline unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx) idx 1061 include/linux/memcontrol.h int idx) idx 1067 include/linux/memcontrol.h int idx, idx 1073 include/linux/memcontrol.h int idx, idx 1079 include/linux/memcontrol.h int idx, idx 1085 include/linux/memcontrol.h int idx, idx 1091 include/linux/memcontrol.h enum node_stat_item idx) idx 1093 include/linux/memcontrol.h return node_page_state(lruvec_pgdat(lruvec), idx); idx 1097 include/linux/memcontrol.h enum node_stat_item idx) idx 1099 include/linux/memcontrol.h return node_page_state(lruvec_pgdat(lruvec), idx); idx 1103 include/linux/memcontrol.h enum node_stat_item idx, int val) idx 1105 include/linux/memcontrol.h __mod_node_page_state(lruvec_pgdat(lruvec), idx, val); idx 1109 include/linux/memcontrol.h enum node_stat_item idx, int val) idx 1111 include/linux/memcontrol.h mod_node_page_state(lruvec_pgdat(lruvec), idx, val); idx 1115 include/linux/memcontrol.h enum node_stat_item idx, int val) idx 1117 include/linux/memcontrol.h __mod_node_page_state(page_pgdat(page), idx, val); idx 1121 include/linux/memcontrol.h enum node_stat_item idx, int val) idx 1123 include/linux/memcontrol.h mod_node_page_state(page_pgdat(page), idx, val); idx 1126 include/linux/memcontrol.h static inline void __mod_lruvec_slab_state(void *p, enum node_stat_item idx, idx 1131 include/linux/memcontrol.h __mod_node_page_state(page_pgdat(page), idx, val); idx 1134 include/linux/memcontrol.h static inline void mod_memcg_obj_state(void *p, int idx, int val) idx 1151 include/linux/memcontrol.h enum vm_event_item idx, idx 1157 include/linux/memcontrol.h enum vm_event_item idx, idx 1163 include/linux/memcontrol.h int idx) idx 1168 include/linux/memcontrol.h void count_memcg_event_mm(struct mm_struct *mm, enum vm_event_item idx) idx 1175 include/linux/memcontrol.h int idx) idx 1177 include/linux/memcontrol.h __mod_memcg_state(memcg, idx, 1); idx 1182 include/linux/memcontrol.h int idx) idx 1184 include/linux/memcontrol.h __mod_memcg_state(memcg, idx, -1); idx 1189 include/linux/memcontrol.h int idx) idx 1191 include/linux/memcontrol.h __mod_memcg_page_state(page, idx, 1); idx 1196 include/linux/memcontrol.h int idx) idx 1198 include/linux/memcontrol.h __mod_memcg_page_state(page, idx, -1); idx 1202 include/linux/memcontrol.h enum node_stat_item idx) idx 1204 include/linux/memcontrol.h __mod_lruvec_state(lruvec, idx, 1); idx 1208 include/linux/memcontrol.h enum node_stat_item idx) idx 1210 include/linux/memcontrol.h __mod_lruvec_state(lruvec, idx, -1); idx 1214 include/linux/memcontrol.h enum node_stat_item idx) idx 1216 include/linux/memcontrol.h __mod_lruvec_page_state(page, idx, 1); idx 1220 include/linux/memcontrol.h enum node_stat_item idx) idx 1222 include/linux/memcontrol.h __mod_lruvec_page_state(page, idx, -1); idx 1225 include/linux/memcontrol.h static inline void __inc_lruvec_slab_state(void *p, enum node_stat_item idx) idx 1227 include/linux/memcontrol.h __mod_lruvec_slab_state(p, idx, 1); idx 1230 include/linux/memcontrol.h static inline void __dec_lruvec_slab_state(void *p, enum node_stat_item idx) idx 1232 include/linux/memcontrol.h __mod_lruvec_slab_state(p, idx, -1); idx 1237 include/linux/memcontrol.h int idx) idx 1239 include/linux/memcontrol.h mod_memcg_state(memcg, idx, 1); idx 1244 include/linux/memcontrol.h int idx) idx 1246 include/linux/memcontrol.h mod_memcg_state(memcg, idx, -1); idx 1251 include/linux/memcontrol.h int idx) idx 1253 include/linux/memcontrol.h mod_memcg_page_state(page, idx, 1); idx 1258 include/linux/memcontrol.h int idx) idx 1260 include/linux/memcontrol.h mod_memcg_page_state(page, idx, -1); idx 1264 include/linux/memcontrol.h enum node_stat_item idx) idx 1266 include/linux/memcontrol.h mod_lruvec_state(lruvec, idx, 1); idx 1270 include/linux/memcontrol.h enum node_stat_item idx) idx 1272 include/linux/memcontrol.h mod_lruvec_state(lruvec, idx, -1); idx 1276 include/linux/memcontrol.h enum node_stat_item idx) idx 1278 include/linux/memcontrol.h mod_lruvec_page_state(page, idx, 1); idx 1282 include/linux/memcontrol.h enum node_stat_item idx) idx 1284 include/linux/memcontrol.h mod_lruvec_page_state(page, idx, -1); idx 137 include/linux/mempolicy.h unsigned long idx); idx 241 include/linux/mempolicy.h mpol_shared_policy_lookup(struct shared_policy *sp, unsigned long idx) idx 462 include/linux/mfd/rk808.h #define RK817_LDO_ON_VSEL_REG(idx) (0xcc + (idx) * 2) idx 1410 include/linux/mlx4/device.h int mlx4_find_cached_mac(struct mlx4_dev *dev, u8 port, u64 mac, int *idx); idx 1411 include/linux/mlx4/device.h int mlx4_find_cached_vlan(struct mlx4_dev *dev, u8 port, u16 vid, int *idx); idx 1442 include/linux/mlx4/device.h int mlx4_counter_alloc(struct mlx4_dev *dev, u32 *idx, u8 usage); idx 1443 include/linux/mlx4/device.h void mlx4_counter_free(struct mlx4_dev *dev, u32 idx); idx 82 include/linux/mlx5/device.h #define MLX5_ARRAY_SET(typ, p, fld, idx, v) do { \ idx 84 include/linux/mlx5/device.h MLX5_SET(typ, p, fld[idx], v); \ idx 115 include/linux/mlx5/device.h #define MLX5_ARRAY_SET64(typ, p, fld, idx, v) do { \ idx 117 include/linux/mlx5/device.h __MLX5_SET64(typ, p, fld[idx], v); \ idx 758 include/linux/mlx5/driver.h int idx; idx 19 include/linux/mmc/slot-gpio.h unsigned int idx, bool override_active_level, idx 22 include/linux/mmc/slot-gpio.h unsigned int idx, idx 915 include/linux/mmzone.h static inline int is_highmem_idx(enum zone_type idx) idx 918 include/linux/mmzone.h return (idx == ZONE_HIGHMEM || idx 919 include/linux/mmzone.h (idx == ZONE_MOVABLE && zone_movable_is_highmem())); idx 1333 include/linux/mmzone.h int idx = subsection_map_index(pfn); idx 1335 include/linux/mmzone.h return test_bit(idx, ms->usage->subsection_map); idx 52 include/linux/moxtet.h u8 idx; idx 94 include/linux/moxtet.h unsigned int idx; idx 266 include/linux/mtd/spi-nor.h u8 idx; idx 1391 include/linux/netdevice.h int *idx); idx 94 include/linux/netfilter/nf_conntrack_h323.h int idx, __be16 port, idx 35 include/linux/of_reserved_mem.h struct device_node *np, int idx); idx 44 include/linux/of_reserved_mem.h struct device_node *np, int idx) idx 73 include/linux/omap-gpmc.h extern void gpmc_cs_write_reg(int cs, int idx, u32 val); idx 1268 include/linux/pci.h resource_size_t pcibios_retrieve_fw_addr(struct pci_dev *dev, int idx); idx 115 include/linux/perf_event.h int idx; /* index in shared_regs->regs[] */ idx 130 include/linux/perf_event.h int idx; idx 19 include/linux/perf_regs.h u64 perf_reg_value(struct pt_regs *regs, int idx); idx 29 include/linux/perf_regs.h static inline u64 perf_reg_value(struct pt_regs *regs, int idx) idx 90 include/linux/platform_data/dma-iop32x.h u16 idx; idx 108 include/linux/platform_data/dma-iop32x.h #define iop_hw_desc_slot_idx(hw_desc, idx) \ idx 109 include/linux/platform_data/dma-iop32x.h ( (void *) (((unsigned long) hw_desc) + ((idx) << 5)) ) idx 11 include/linux/platform_data/gpio-dwapb.h unsigned int idx; idx 28 include/linux/platform_data/sa11x0-serial.h void sa1100_register_uart(int idx, int port); idx 33 include/linux/platform_data/sa11x0-serial.h static inline void sa1100_register_uart(int idx, int port) idx 300 include/linux/qed/qed_chain.h #define is_unusable_idx(p, idx) \ idx 301 include/linux/qed/qed_chain.h (((p)->u.chain16.idx & (p)->elem_per_page_mask) == (p)->usable_per_page) idx 303 include/linux/qed/qed_chain.h #define is_unusable_idx_u32(p, idx) \ idx 304 include/linux/qed/qed_chain.h (((p)->u.chain32.idx & (p)->elem_per_page_mask) == (p)->usable_per_page) idx 305 include/linux/qed/qed_chain.h #define is_unusable_next_idx(p, idx) \ idx 306 include/linux/qed/qed_chain.h ((((p)->u.chain16.idx + 1) & (p)->elem_per_page_mask) == \ idx 309 include/linux/qed/qed_chain.h #define is_unusable_next_idx_u32(p, idx) \ idx 310 include/linux/qed/qed_chain.h ((((p)->u.chain32.idx + 1) & (p)->elem_per_page_mask) == \ idx 313 include/linux/qed/qed_chain.h #define test_and_skip(p, idx) \ idx 316 include/linux/qed/qed_chain.h if (is_unusable_idx(p, idx)) \ idx 317 include/linux/qed/qed_chain.h (p)->u.chain16.idx += (p)->elem_unusable; \ idx 319 include/linux/qed/qed_chain.h if (is_unusable_idx_u32(p, idx)) \ idx 320 include/linux/qed/qed_chain.h (p)->u.chain32.idx += (p)->elem_unusable; \ idx 70 include/linux/rbtree_latch.h __lt_from_rb(struct rb_node *node, int idx) idx 72 include/linux/rbtree_latch.h return container_of(node, struct latch_tree_node, node[idx]); idx 76 include/linux/rbtree_latch.h __lt_insert(struct latch_tree_node *ltn, struct latch_tree_root *ltr, int idx, idx 79 include/linux/rbtree_latch.h struct rb_root *root = <r->tree[idx]; idx 81 include/linux/rbtree_latch.h struct rb_node *node = <n->node[idx]; idx 87 include/linux/rbtree_latch.h ltp = __lt_from_rb(parent, idx); idx 100 include/linux/rbtree_latch.h __lt_erase(struct latch_tree_node *ltn, struct latch_tree_root *ltr, int idx) idx 102 include/linux/rbtree_latch.h rb_erase(<n->node[idx], <r->tree[idx]); idx 106 include/linux/rbtree_latch.h __lt_find(void *key, struct latch_tree_root *ltr, int idx, idx 109 include/linux/rbtree_latch.h struct rb_node *node = rcu_dereference_raw(ltr->tree[idx].rb_node); idx 114 include/linux/rbtree_latch.h ltn = __lt_from_rb(node, idx); idx 117 include/linux/rtnetlink.h int *idx); idx 315 include/linux/serdev.h struct tty_driver *drv, int idx); idx 320 include/linux/serdev.h struct tty_driver *drv, int idx) idx 61 include/linux/srcu.h void __srcu_read_unlock(struct srcu_struct *ssp, int idx) __releases(ssp); idx 176 include/linux/srcu.h static inline void srcu_read_unlock(struct srcu_struct *ssp, int idx) idx 179 include/linux/srcu.h WARN_ON_ONCE(idx & ~0x1); idx 181 include/linux/srcu.h __srcu_read_unlock(ssp, idx); idx 186 include/linux/srcu.h srcu_read_unlock_notrace(struct srcu_struct *ssp, int idx) __releases(ssp) idx 188 include/linux/srcu.h __srcu_read_unlock(ssp, idx); idx 60 include/linux/srcutiny.h int idx; idx 62 include/linux/srcutiny.h idx = READ_ONCE(ssp->srcu_idx); idx 63 include/linux/srcutiny.h WRITE_ONCE(ssp->srcu_lock_nesting[idx], ssp->srcu_lock_nesting[idx] + 1); idx 64 include/linux/srcutiny.h return idx; idx 81 include/linux/srcutiny.h int idx; idx 83 include/linux/srcutiny.h idx = READ_ONCE(ssp->srcu_idx) & 0x1; idx 85 include/linux/srcutiny.h tt, tf, idx, idx 86 include/linux/srcutiny.h READ_ONCE(ssp->srcu_lock_nesting[!idx]), idx 87 include/linux/srcutiny.h READ_ONCE(ssp->srcu_lock_nesting[idx])); idx 102 include/linux/trace_events.h long idx; idx 560 include/linux/tty.h extern struct tty_struct *alloc_tty_struct(struct tty_driver *driver, int idx); idx 564 include/linux/tty.h extern struct tty_struct *tty_init_dev(struct tty_driver *driver, int idx); idx 565 include/linux/tty.h extern void tty_release_struct(struct tty_struct *tty, int idx); idx 256 include/linux/tty_driver.h struct file *filp, int idx); idx 48 include/linux/uio.h int idx; idx 358 include/linux/usb/pd.h #define RDO_OBJ(idx) (((idx) & RDO_OBJ_POS_MASK) << RDO_OBJ_POS_SHIFT) idx 363 include/linux/usb/pd.h #define RDO_FIXED(idx, op_ma, max_ma, flags) \ idx 364 include/linux/usb/pd.h (RDO_OBJ(idx) | (flags) | \ idx 373 include/linux/usb/pd.h #define RDO_BATT(idx, op_mw, max_mw, flags) \ idx 374 include/linux/usb/pd.h (RDO_OBJ(idx) | (flags) | \ idx 391 include/linux/usb/pd.h #define RDO_PROG(idx, out_mv, op_ma, flags) \ idx 392 include/linux/usb/pd.h (RDO_OBJ(idx) | (flags) | \ idx 240 include/media/dvb_ringbuffer.h size_t idx, idx 257 include/media/dvb_ringbuffer.h extern ssize_t dvb_ringbuffer_pkt_read(struct dvb_ringbuffer *rbuf, size_t idx, idx 266 include/media/dvb_ringbuffer.h extern void dvb_ringbuffer_pkt_dispose(struct dvb_ringbuffer *rbuf, size_t idx); idx 278 include/media/dvb_ringbuffer.h size_t idx, size_t *pktlen); idx 105 include/media/v4l2-ctrls.h bool (*equal)(const struct v4l2_ctrl *ctrl, u32 idx, idx 108 include/media/v4l2-ctrls.h void (*init)(const struct v4l2_ctrl *ctrl, u32 idx, idx 111 include/media/v4l2-ctrls.h int (*validate)(const struct v4l2_ctrl *ctrl, u32 idx, idx 613 include/media/v4l2-mem2mem.h v4l2_m2m_buf_remove_by_idx(struct v4l2_m2m_queue_ctx *q_ctx, unsigned int idx); idx 616 include/media/v4l2-mem2mem.h v4l2_m2m_src_buf_remove_by_idx(struct v4l2_m2m_ctx *m2m_ctx, unsigned int idx) idx 618 include/media/v4l2-mem2mem.h return v4l2_m2m_buf_remove_by_idx(&m2m_ctx->out_q_ctx, idx); idx 622 include/media/v4l2-mem2mem.h v4l2_m2m_dst_buf_remove_by_idx(struct v4l2_m2m_ctx *m2m_ctx, unsigned int idx) idx 624 include/media/v4l2-mem2mem.h return v4l2_m2m_buf_remove_by_idx(&m2m_ctx->cap_q_ctx, idx); idx 21 include/misc/ocxl.h u8 idx; idx 3683 include/net/cfg80211.h int idx, u8 *mac, struct station_info *sinfo); idx 3694 include/net/cfg80211.h int idx, u8 *dst, u8 *next_hop, idx 3699 include/net/cfg80211.h int idx, u8 *dst, u8 *mpp, idx 3784 include/net/cfg80211.h int idx, struct survey_info *info); idx 168 include/net/erspan.h __be32 *idx; idx 198 include/net/erspan.h idx = (__be32 *)(ershdr + 1); idx 199 include/net/erspan.h *idx = htonl(index & INDEX_MASK); idx 117 include/net/fq_impl.h struct fq_tin *tin, u32 idx, idx 125 include/net/fq_impl.h flow = &fq->flows[idx]; idx 127 include/net/fq_impl.h flow = get_default_func(fq, tin, idx, skb); idx 157 include/net/fq_impl.h struct fq_tin *tin, u32 idx, idx 167 include/net/fq_impl.h flow = fq_flow_classify(fq, tin, idx, skb, get_default_func); idx 193 include/net/ip_vs.h int *idx) idx 198 include/net/ip_vs.h len = snprintf(&buf[*idx], buf_len - *idx, "[%pI6c]", idx 202 include/net/ip_vs.h len = snprintf(&buf[*idx], buf_len - *idx, "%pI4", idx 205 include/net/ip_vs.h *idx += len; idx 206 include/net/ip_vs.h BUG_ON(*idx > buf_len + 1); idx 207 include/net/ip_vs.h return &buf[*idx - len]; idx 931 include/net/mac80211.h s8 idx; idx 943 include/net/mac80211.h rate->idx = ((nss - 1) << 4) | mcs; idx 949 include/net/mac80211.h return rate->idx & 0xF; idx 955 include/net/mac80211.h return (rate->idx >> 4) + 1; idx 1900 include/net/mac80211.h s8 idx; idx 2601 include/net/mac80211.h if (WARN_ON_ONCE(c->control.rates[0].idx < 0)) idx 2603 include/net/mac80211.h return &hw->wiphy->bands[c->band]->bitrates[c->control.rates[0].idx]; idx 2617 include/net/mac80211.h const struct ieee80211_tx_info *c, int idx) idx 2619 include/net/mac80211.h if (c->control.rates[idx + 1].idx < 0) idx 2621 include/net/mac80211.h return &hw->wiphy->bands[c->band]->bitrates[c->control.rates[idx + 1].idx]; idx 3807 include/net/mac80211.h struct ieee80211_vif *vif, int idx); idx 3905 include/net/mac80211.h int (*get_survey)(struct ieee80211_hw *hw, int idx, idx 25 include/net/netprio_cgroup.h u32 idx; idx 29 include/net/netprio_cgroup.h idx = css->cgroup->id; idx 31 include/net/netprio_cgroup.h return idx; idx 261 include/net/nfc/hci.h int nfc_hci_get_param(struct nfc_hci_dev *hdev, u8 gate, u8 idx, idx 263 include/net/nfc/hci.h int nfc_hci_set_param(struct nfc_hci_dev *hdev, u8 gate, u8 idx, idx 309 include/net/nfc/nci_core.h int nci_hci_set_param(struct nci_dev *ndev, u8 gate, u8 idx, idx 311 include/net/nfc/nci_core.h int nci_hci_get_param(struct nci_dev *ndev, u8 gate, u8 idx, idx 91 include/net/nfc/nfc.h u32 idx; idx 122 include/net/nfc/nfc.h u32 idx; idx 164 include/net/nfc/nfc.h int idx; idx 61 include/net/smc.h int (*move_data)(struct smcd_dev *dev, u64 dmb_tok, unsigned int idx, idx 1683 include/net/xfrm.h struct xfrm_algo_desc *xfrm_aalg_get_byidx(unsigned int idx); idx 1684 include/net/xfrm.h struct xfrm_algo_desc *xfrm_ealg_get_byidx(unsigned int idx); idx 661 include/rdma/uverbs_ioctl.h unsigned int idx) idx 663 include/rdma/uverbs_ioctl.h return test_bit(uapi_bkey_attr(uapi_key_attr(idx)), idx 687 include/rdma/uverbs_ioctl.h u16 idx) idx 689 include/rdma/uverbs_ioctl.h if (!uverbs_attr_is_valid(attrs_bundle, idx)) idx 692 include/rdma/uverbs_ioctl.h return &attrs_bundle->attrs[uapi_bkey_attr(uapi_key_attr(idx))]; idx 696 include/rdma/uverbs_ioctl.h u16 idx) idx 698 include/rdma/uverbs_ioctl.h const struct uverbs_attr *attr = uverbs_attr_get(attrs_bundle, idx); idx 707 include/rdma/uverbs_ioctl.h u16 idx) idx 711 include/rdma/uverbs_ioctl.h attr = uverbs_attr_get(attrs_bundle, idx); idx 719 include/rdma/uverbs_ioctl.h u16 idx) idx 721 include/rdma/uverbs_ioctl.h const struct uverbs_attr *attr = uverbs_attr_get(attrs_bundle, idx); idx 730 include/rdma/uverbs_ioctl.h uverbs_attr_get_len(const struct uverbs_attr_bundle *attrs_bundle, u16 idx) idx 732 include/rdma/uverbs_ioctl.h const struct uverbs_attr *attr = uverbs_attr_get(attrs_bundle, idx); idx 748 include/rdma/uverbs_ioctl.h uverbs_attr_ptr_get_array_size(struct uverbs_attr_bundle *attrs, u16 idx, idx 751 include/rdma/uverbs_ioctl.h int size = uverbs_attr_get_len(attrs, idx); idx 793 include/rdma/uverbs_ioctl.h const struct uverbs_attr_bundle *attrs_bundle, u16 idx) idx 795 include/rdma/uverbs_ioctl.h const struct uverbs_attr *attr = uverbs_attr_get(attrs_bundle, idx); idx 806 include/rdma/uverbs_ioctl.h size_t idx, idx 809 include/rdma/uverbs_ioctl.h const struct uverbs_attr *attr = uverbs_attr_get(attrs_bundle, idx); idx 833 include/rdma/uverbs_ioctl.h size_t idx, idx 836 include/rdma/uverbs_ioctl.h const struct uverbs_attr *attr = uverbs_attr_get(attrs_bundle, idx); idx 856 include/rdma/uverbs_ioctl.h #define uverbs_copy_from(to, attrs_bundle, idx) \ idx 857 include/rdma/uverbs_ioctl.h _uverbs_copy_from(to, attrs_bundle, idx, sizeof(*to)) idx 859 include/rdma/uverbs_ioctl.h #define uverbs_copy_from_or_zero(to, attrs_bundle, idx) \ idx 860 include/rdma/uverbs_ioctl.h _uverbs_copy_from_or_zero(to, attrs_bundle, idx, sizeof(*to)) idx 870 include/rdma/uverbs_ioctl.h size_t idx, u64 allowed_bits); idx 872 include/rdma/uverbs_ioctl.h size_t idx, u64 allowed_bits); idx 873 include/rdma/uverbs_ioctl.h int uverbs_copy_to(const struct uverbs_attr_bundle *attrs_bundle, size_t idx, idx 890 include/rdma/uverbs_ioctl.h size_t idx, s64 lower_bound, u64 upper_bound, idx 893 include/rdma/uverbs_ioctl.h size_t idx, const void *from, size_t size); idx 897 include/rdma/uverbs_ioctl.h size_t idx, u64 allowed_bits) idx 903 include/rdma/uverbs_ioctl.h size_t idx, u64 allowed_bits) idx 908 include/rdma/uverbs_ioctl.h size_t idx, const void *from, size_t size) idx 924 include/rdma/uverbs_ioctl.h size_t idx, s64 lower_bound, u64 upper_bound, idx 931 include/rdma/uverbs_ioctl.h size_t idx, const void *from, size_t size) idx 131 include/soc/at91/atmel_tcb.h #define ATMEL_TC_CHAN(idx) ((idx)*0x40) idx 132 include/soc/at91/atmel_tcb.h #define ATMEL_TC_REG(idx, reg) (ATMEL_TC_CHAN(idx) + ATMEL_TC_ ## reg) idx 222 include/sound/core.h int snd_card_new(struct device *parent, int idx, const char *xid, idx 403 include/sound/hda_codec.h void snd_hda_spdif_ctls_unassign(struct hda_codec *codec, int idx); idx 404 include/sound/hda_codec.h void snd_hda_spdif_ctls_assign(struct hda_codec *codec, int idx, hda_nid_t nid); idx 15 include/sound/hda_component.h void snd_hdac_display_power(struct hdac_bus *bus, unsigned int idx, idx 34 include/sound/hda_component.h unsigned int idx, bool enable) idx 50 include/sound/hda_regmap.h #define snd_hdac_regmap_encode_amp(nid, ch, dir, idx) \ idx 54 include/sound/hda_regmap.h (idx)) idx 64 include/sound/hda_regmap.h #define snd_hdac_regmap_encode_amp_stereo(nid, dir, idx) \ idx 68 include/sound/hda_regmap.h (idx)) idx 137 include/sound/hda_regmap.h int ch, int dir, int idx) idx 139 include/sound/hda_regmap.h unsigned int cmd = snd_hdac_regmap_encode_amp(nid, ch, dir, idx); idx 161 include/sound/hda_regmap.h int ch, int dir, int idx, int mask, int val) idx 163 include/sound/hda_regmap.h unsigned int cmd = snd_hdac_regmap_encode_amp(nid, ch, dir, idx); idx 182 include/sound/hda_regmap.h int dir, int idx) idx 184 include/sound/hda_regmap.h unsigned int cmd = snd_hdac_regmap_encode_amp_stereo(nid, dir, idx); idx 206 include/sound/hda_regmap.h int dir, int idx, int mask, int val) idx 208 include/sound/hda_regmap.h unsigned int cmd = snd_hdac_regmap_encode_amp_stereo(nid, dir, idx); idx 360 include/sound/hdaudio.h int idx; idx 543 include/sound/hdaudio.h int idx, int direction, int tag); idx 647 include/sound/hdaudio.h static inline void *snd_array_elem(struct snd_array *array, unsigned int idx) idx 649 include/sound/hdaudio.h return array->list + idx * array->elem_size; idx 658 include/sound/hdaudio.h #define snd_array_for_each(array, idx, ptr) \ idx 659 include/sound/hdaudio.h for ((idx) = 0, (ptr) = (array)->list; (idx) < (array)->used; \ idx 660 include/sound/hdaudio.h (ptr) = snd_array_elem(array, ++(idx))) idx 81 include/sound/hdaudio_ext.h struct hdac_ext_stream *stream, int idx, idx 1383 include/sound/pcm.h snd_pcm_chmap_substream(struct snd_pcm_chmap *info, unsigned int idx) idx 1387 include/sound/pcm.h if (s->number == idx) idx 677 include/sound/soc.h unsigned int idx; idx 90 include/sound/vx_core.h int (*load_dsp)(struct vx_core *chip, int idx, const struct firmware *fw); idx 207 include/trace/events/compaction.h __field(enum zone_type, idx) idx 214 include/trace/events/compaction.h __entry->idx = zone_idx(zone); idx 221 include/trace/events/compaction.h __print_symbolic(__entry->idx, ZONE_TYPE), idx 252 include/trace/events/compaction.h __field(enum zone_type, idx) idx 261 include/trace/events/compaction.h __entry->idx = zone_idx(zone); idx 270 include/trace/events/compaction.h __print_symbolic(__entry->idx, ZONE_TYPE), idx 14 include/trace/events/cpuhp.h int idx, idx 17 include/trace/events/cpuhp.h TP_ARGS(cpu, target, idx, fun), idx 22 include/trace/events/cpuhp.h __field( int, idx ) idx 29 include/trace/events/cpuhp.h __entry->idx = idx; idx 34 include/trace/events/cpuhp.h __entry->cpu, __entry->target, __entry->idx, __entry->fun) idx 41 include/trace/events/cpuhp.h int idx, idx 45 include/trace/events/cpuhp.h TP_ARGS(cpu, target, idx, fun, node), idx 50 include/trace/events/cpuhp.h __field( int, idx ) idx 57 include/trace/events/cpuhp.h __entry->idx = idx; idx 62 include/trace/events/cpuhp.h __entry->cpu, __entry->target, __entry->idx, __entry->fun) idx 69 include/trace/events/cpuhp.h int idx, idx 72 include/trace/events/cpuhp.h TP_ARGS(cpu, state, idx, ret), idx 77 include/trace/events/cpuhp.h __field( int, idx ) idx 84 include/trace/events/cpuhp.h __entry->idx = idx; idx 89 include/trace/events/cpuhp.h __entry->cpu, __entry->state, __entry->idx, __entry->ret) idx 23 include/trace/events/fsi.h __entry->master_idx = master->idx; idx 51 include/trace/events/fsi.h __entry->master_idx = master->idx; idx 85 include/trace/events/fsi.h __entry->master_idx = master->idx; idx 116 include/trace/events/fsi.h __entry->master_idx = master->idx; idx 18 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 36 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 58 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 78 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 94 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 110 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 126 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 141 include/trace/events/fsi_master_ast_cf.h __entry->master_idx = master->master.idx; idx 20 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 41 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 61 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 76 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 90 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 104 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 119 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 134 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 149 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 163 include/trace/events/fsi_master_gpio.h __entry->master_idx = master->master.idx; idx 182 include/trace/events/xdp.h #define _trace_xdp_redirect_map(dev, xdp, fwd, map, idx) \ idx 184 include/trace/events/xdp.h 0, map, idx) idx 186 include/trace/events/xdp.h #define _trace_xdp_redirect_map_err(dev, xdp, fwd, map, idx, err) \ idx 188 include/trace/events/xdp.h err, map, idx) idx 238 include/uapi/drm/drm.h int idx; /**< Which client desired? */ idx 376 include/uapi/drm/drm.h int idx; /**< Index into the master buffer list */ idx 247 include/uapi/drm/i810_drm.h int idx; /* buffer index */ idx 253 include/uapi/drm/i810_drm.h int idx; /* buffer index */ idx 281 include/uapi/drm/i810_drm.h int idx; /* buffer index */ idx 376 include/uapi/drm/mga_drm.h int idx; /* buffer to queue */ idx 382 include/uapi/drm/mga_drm.h int idx; /* buffer to queue */ idx 389 include/uapi/drm/mga_drm.h int idx; idx 268 include/uapi/drm/r128_drm.h int idx; /* Index of vertex buffer */ idx 275 include/uapi/drm/r128_drm.h int idx; idx 282 include/uapi/drm/r128_drm.h int idx; idx 309 include/uapi/drm/r128_drm.h int idx; idx 627 include/uapi/drm/radeon_drm.h int idx; /* Index of vertex buffer */ idx 634 include/uapi/drm/radeon_drm.h int idx; idx 645 include/uapi/drm/radeon_drm.h int idx; /* Index of vertex buffer */ idx 690 include/uapi/drm/radeon_drm.h int idx; idx 203 include/uapi/drm/savage_drm.h } idx; /* SAVAGE_CMD_DMA_IDX, SAVAGE_CMD_VB_IDX */ idx 66 include/uapi/linux/connector.h __u32 idx; idx 7 include/uapi/linux/netfilter/xt_TCPOPTSTRIP.h #define tcpoptstrip_set_bit(bmap, idx) \ idx 8 include/uapi/linux/netfilter/xt_TCPOPTSTRIP.h (bmap[(idx) >> 5] |= 1U << (idx & 31)) idx 9 include/uapi/linux/netfilter/xt_TCPOPTSTRIP.h #define tcpoptstrip_test_bit(bmap, idx) \ idx 10 include/uapi/linux/netfilter/xt_TCPOPTSTRIP.h (((1U << (idx & 31)) & bmap[(idx) >> 5]) != 0) idx 84 include/uapi/linux/rfkill.h __u32 idx; idx 103 include/uapi/linux/virtio_ring.h __virtio16 idx; idx 117 include/uapi/linux/virtio_ring.h __virtio16 idx; idx 96 include/uapi/xen/privcmd.h __u32 idx; idx 39 include/xen/hvm.h static inline int hvm_get_parameter(int idx, uint64_t *value) idx 45 include/xen/hvm.h xhv.index = idx; idx 49 include/xen/hvm.h param_name(idx), idx, r); idx 15 include/xen/interface/io/console.h #define MASK_XENCONS_IDX(idx, ring) ((idx) & (sizeof(ring)-1)) idx 859 include/xen/interface/io/displif.h #define XENDISPL_IN_RING_REF(page, idx) \ idx 860 include/xen/interface/io/displif.h (XENDISPL_IN_RING((page))[(idx) % XENDISPL_IN_RING_LEN]) idx 97 include/xen/interface/io/fbif.h #define XENFB_IN_RING_REF(page, idx) \ idx 98 include/xen/interface/io/fbif.h (XENFB_IN_RING((page))[(idx) % XENFB_IN_RING_LEN]) idx 105 include/xen/interface/io/fbif.h #define XENFB_OUT_RING_REF(page, idx) \ idx 106 include/xen/interface/io/fbif.h (XENFB_OUT_RING((page))[(idx) % XENFB_OUT_RING_LEN]) idx 544 include/xen/interface/io/kbdif.h #define XENKBD_IN_RING_REF(page, idx) \ idx 545 include/xen/interface/io/kbdif.h (XENKBD_IN_RING((page))[(idx) % XENKBD_IN_RING_LEN]) idx 552 include/xen/interface/io/kbdif.h #define XENKBD_OUT_RING_REF(page, idx) \ idx 553 include/xen/interface/io/kbdif.h (XENKBD_OUT_RING((page))[(idx) % XENKBD_OUT_RING_LEN]) idx 342 include/xen/interface/io/ring.h static inline RING_IDX name##_mask(RING_IDX idx, RING_IDX ring_size) \ idx 344 include/xen/interface/io/ring.h return idx & (ring_size - 1); \ idx 348 include/xen/interface/io/ring.h RING_IDX idx, \ idx 351 include/xen/interface/io/ring.h return buf + name##_mask(idx, ring_size); \ idx 1078 include/xen/interface/io/sndif.h #define XENSND_IN_RING_REF(page, idx) \ idx 1079 include/xen/interface/io/sndif.h (XENSND_IN_RING((page))[(idx) % XENSND_IN_RING_LEN]) idx 84 include/xen/interface/io/xs_wire.h #define MASK_XENSTORE_IDX(idx) ((idx) & (XENSTORE_RING_SIZE-1)) idx 183 include/xen/interface/memory.h xen_ulong_t idx; idx 371 ipc/sem.c int idx; idx 389 ipc/sem.c idx = array_index_nospec(sops->sem_num, sma->sem_nsems); idx 390 ipc/sem.c sem = &sma->sems[idx]; idx 639 ipc/sem.c int idx = array_index_nospec(sop->sem_num, sma->sem_nsems); idx 640 ipc/sem.c curr = &sma->sems[idx]; idx 720 ipc/sem.c int idx = array_index_nospec(sop->sem_num, sma->sem_nsems); idx 722 ipc/sem.c curr = &sma->sems[idx]; idx 2123 ipc/sem.c int idx = array_index_nospec(sops->sem_num, sma->sem_nsems); idx 2124 ipc/sem.c curr = &sma->sems[idx]; idx 203 ipc/util.c int idx, next_id = -1; idx 229 ipc/util.c idx = idr_alloc_cyclic(&ids->ipcs_idr, NULL, 0, max_idx, idx 232 ipc/util.c if (idx >= 0) { idx 238 ipc/util.c if (idx <= ids->last_idx) { idx 243 ipc/util.c ids->last_idx = idx; idx 250 ipc/util.c idr_replace(&ids->ipcs_idr, new, idx); idx 254 ipc/util.c idx = idr_alloc(&ids->ipcs_idr, new, ipcid_to_idx(next_id), idx 257 ipc/util.c if (idx >= 0) idx 258 ipc/util.c new->id = (new->seq << ipcmni_seq_shift()) + idx; idx 259 ipc/util.c return idx; idx 281 ipc/util.c int idx, err; idx 304 ipc/util.c idx = ipc_idr_alloc(ids, new); idx 307 ipc/util.c if (idx >= 0 && new->key != IPC_PRIVATE) { idx 311 ipc/util.c idr_remove(&ids->ipcs_idr, idx); idx 312 ipc/util.c idx = err; idx 315 ipc/util.c if (idx < 0) { idx 319 ipc/util.c return idx; idx 323 ipc/util.c if (idx > ids->max_idx) idx 324 ipc/util.c ids->max_idx = idx; idx 325 ipc/util.c return idx; idx 463 ipc/util.c int idx = ipcid_to_idx(ipcp->id); idx 465 ipc/util.c idr_remove(&ids->ipcs_idr, idx); idx 470 ipc/util.c if (unlikely(idx == ids->max_idx)) { idx 472 ipc/util.c idx--; idx 473 ipc/util.c if (idx == -1) idx 475 ipc/util.c } while (!idr_find(&ids->ipcs_idr, idx)); idx 476 ipc/util.c ids->max_idx = idx; idx 596 ipc/util.c int idx = ipcid_to_idx(id); idx 598 ipc/util.c out = idr_find(&ids->ipcs_idr, idx); idx 72 kernel/bpf/devmap.c unsigned int idx; /* keep track of map index for tracepoint */ idx 105 kernel/bpf/devmap.c int idx) idx 107 kernel/bpf/devmap.c return &dtab->dev_index_head[idx & (dtab->n_buckets - 1)]; idx 298 kernel/bpf/devmap.c if (dev->idx == key) idx 308 kernel/bpf/devmap.c u32 idx, *next = next_key; idx 316 kernel/bpf/devmap.c idx = *(u32 *)key; idx 318 kernel/bpf/devmap.c dev = __dev_map_hash_lookup_elem(map, idx); idx 326 kernel/bpf/devmap.c *next = next_dev->idx; idx 330 kernel/bpf/devmap.c i = idx & (dtab->n_buckets - 1); idx 341 kernel/bpf/devmap.c *next = next_dev->idx; idx 376 kernel/bpf/devmap.c trace_xdp_devmap_xmit(&obj->dtab->map, obj->idx, idx 587 kernel/bpf/devmap.c unsigned int idx) idx 617 kernel/bpf/devmap.c dev->idx = idx; idx 670 kernel/bpf/devmap.c u32 idx = *(u32 *)key; idx 679 kernel/bpf/devmap.c old_dev = __dev_map_hash_lookup_elem(map, idx); idx 683 kernel/bpf/devmap.c dev = __dev_map_alloc_node(net, dtab, ifindex, idx); idx 701 kernel/bpf/devmap.c dev_map_index_hash(dtab, idx)); idx 1414 kernel/bpf/verifier.c p[cnt - 1].idx = env->insn_idx; idx 1429 kernel/bpf/verifier.c if (cnt && st->jmp_history[cnt - 1].idx == i) { idx 1442 kernel/bpf/verifier.c static int backtrack_insn(struct bpf_verifier_env *env, int idx, idx 1449 kernel/bpf/verifier.c struct bpf_insn *insn = env->prog->insnsi + idx; idx 1461 kernel/bpf/verifier.c verbose(env, "%d: ", idx); idx 2608 kernel/bpf/verifier.c int depth = 0, frame = 0, idx = 0, i = 0, subprog_end; idx 2618 kernel/bpf/verifier.c depth += round_up(max_t(u32, subprog[idx].stack_depth, 1), 32); idx 2625 kernel/bpf/verifier.c subprog_end = subprog[idx + 1].start; idx 2633 kernel/bpf/verifier.c ret_prog[frame] = idx; idx 2637 kernel/bpf/verifier.c idx = find_subprog(env, i); idx 2638 kernel/bpf/verifier.c if (idx < 0) { idx 2656 kernel/bpf/verifier.c depth -= round_up(max_t(u32, subprog[idx].stack_depth, 1), 32); idx 2659 kernel/bpf/verifier.c idx = ret_prog[frame]; idx 2665 kernel/bpf/verifier.c const struct bpf_insn *insn, int idx) idx 2667 kernel/bpf/verifier.c int start = idx + insn->imm + 1, subprog; idx 6288 kernel/bpf/verifier.c int idx) idx 6293 kernel/bpf/verifier.c return &env->explored_states[(idx ^ state->callsite) % state_htab_size(env)]; idx 6296 kernel/bpf/verifier.c static void init_explored_state(struct bpf_verifier_env *env, int idx) idx 6298 kernel/bpf/verifier.c env->insn_aux_data[idx].prune_point = true; idx 139 kernel/cgroup/debug.c int idx = 0; idx 143 kernel/cgroup/debug.c seq_puts(seq, idx ? "," : "<="); idx 145 kernel/cgroup/debug.c idx++; idx 345 kernel/debug/gdbstub.c int idx = 0; idx 349 kernel/debug/gdbstub.c dbg_get_reg(i, ptr + idx, regs); idx 350 kernel/debug/gdbstub.c idx += dbg_reg_def[i].size; idx 357 kernel/debug/gdbstub.c int idx = 0; idx 361 kernel/debug/gdbstub.c dbg_set_reg(i, ptr + idx, regs); idx 362 kernel/debug/gdbstub.c idx += dbg_reg_def[i].size; idx 247 kernel/dma/debug.c __acquires(&dma_entry_hash[idx].lock) idx 249 kernel/dma/debug.c int idx = hash_fn(entry); idx 252 kernel/dma/debug.c spin_lock_irqsave(&dma_entry_hash[idx].lock, __flags); idx 254 kernel/dma/debug.c return &dma_entry_hash[idx]; idx 404 kernel/dma/debug.c int idx; idx 406 kernel/dma/debug.c for (idx = 0; idx < HASH_SIZE; idx++) { idx 407 kernel/dma/debug.c struct hash_bucket *bucket = &dma_entry_hash[idx]; idx 417 kernel/dma/debug.c type2name[entry->type], idx, idx 824 kernel/dma/debug.c int idx; idx 826 kernel/dma/debug.c for (idx = 0; idx < HASH_SIZE; idx++) { idx 827 kernel/dma/debug.c struct hash_bucket *bucket = &dma_entry_hash[idx]; idx 837 kernel/dma/debug.c type2name[entry->type], idx, idx 10259 kernel/events/core.c int idx; idx 10262 kernel/events/core.c idx = srcu_read_lock(&pmus_srcu); idx 10294 kernel/events/core.c srcu_read_unlock(&pmus_srcu, idx); idx 593 kernel/events/ring_buffer.c static void rb_free_aux_page(struct ring_buffer *rb, int idx) idx 595 kernel/events/ring_buffer.c struct page *page = virt_to_page(rb->aux_pages[idx]); idx 343 kernel/irq/generic-chip.c int idx; idx 347 kernel/irq/generic-chip.c idx = hw_irq / dgc->irqs_per_chip; idx 348 kernel/irq/generic-chip.c if (idx >= dgc->num_chips) idx 350 kernel/irq/generic-chip.c return dgc->gc[idx]; idx 386 kernel/irq/generic-chip.c int idx; idx 392 kernel/irq/generic-chip.c idx = hw_irq % dgc->irqs_per_chip; idx 394 kernel/irq/generic-chip.c if (test_bit(idx, &gc->unused)) idx 397 kernel/irq/generic-chip.c if (test_bit(idx, &gc->installed)) idx 411 kernel/irq/generic-chip.c set_bit(idx, &gc->installed); idx 420 kernel/irq/generic-chip.c data->mask = 1 << idx; idx 345 kernel/irq/timings.c int idx = period; idx 354 kernel/irq/timings.c while (!memcmp(buffer, &buffer[idx], size * sizeof(int))) { idx 359 kernel/irq/timings.c idx += size; idx 366 kernel/irq/timings.c if (idx == len) idx 374 kernel/irq/timings.c if (len - idx < period) idx 375 kernel/irq/timings.c size = len - idx; idx 145 kernel/kallsyms.c static unsigned long kallsyms_sym_address(int idx) idx 148 kernel/kallsyms.c return kallsyms_addresses[idx]; idx 152 kernel/kallsyms.c return kallsyms_relative_base + (u32)kallsyms_offsets[idx]; idx 155 kernel/kallsyms.c if (kallsyms_offsets[idx] >= 0) idx 156 kernel/kallsyms.c return kallsyms_offsets[idx]; idx 159 kernel/kallsyms.c return kallsyms_relative_base - 1 - kallsyms_offsets[idx]; idx 62 kernel/kcmp.c get_file_raw_ptr(struct task_struct *task, unsigned int idx) idx 70 kernel/kcmp.c file = fcheck_files(task->files, idx); idx 204 kernel/kexec_elf.c int idx) idx 207 kernel/kexec_elf.c struct elf_phdr *phdr = (struct elf_phdr *) &elf_info->proghdrs[idx]; idx 212 kernel/kexec_elf.c pbuf = buf + elf_info->ehdr->e_phoff + (idx * sizeof(*buf_phdr)); idx 192 kernel/kprobes.c static int collect_one_slot(struct kprobe_insn_page *kip, int idx) idx 194 kernel/kprobes.c kip->slot_used[idx] = SLOT_CLEAN; idx 239 kernel/kprobes.c long idx; idx 244 kernel/kprobes.c idx = ((long)slot - (long)kip->insns) / idx 246 kernel/kprobes.c if (idx >= 0 && idx < slots_per_page(c)) idx 257 kernel/kprobes.c WARN_ON(kip->slot_used[idx] != SLOT_USED); idx 259 kernel/kprobes.c kip->slot_used[idx] = SLOT_DIRTY; idx 264 kernel/kprobes.c collect_one_slot(kip, idx); idx 364 kernel/locking/lockdep.c static inline u64 iterate_chain_key(u64 key, u32 idx) idx 368 kernel/locking/lockdep.c __jhash_mix(idx, k0, k1); /* Macro that modifies arguments! */ idx 1283 kernel/locking/lockdep.c int idx = find_first_zero_bit(list_entries_in_use, idx 1286 kernel/locking/lockdep.c if (idx >= ARRAY_SIZE(list_entries)) { idx 1295 kernel/locking/lockdep.c __set_bit(idx, list_entries_in_use); idx 1296 kernel/locking/lockdep.c return list_entries + idx; idx 2780 kernel/locking/lockdep.c int idx = find_first_zero_bit(lock_chains_in_use, idx 2783 kernel/locking/lockdep.c if (unlikely(idx >= ARRAY_SIZE(lock_chains))) idx 2785 kernel/locking/lockdep.c __set_bit(idx, lock_chains_in_use); idx 2786 kernel/locking/lockdep.c return lock_chains + idx; idx 4053 kernel/locking/lockdep.c unsigned int depth, int *idx) idx 4082 kernel/locking/lockdep.c *idx = i; idx 4087 kernel/locking/lockdep.c int idx, unsigned int *merged) idx 4090 kernel/locking/lockdep.c int first_idx = idx; idx 4095 kernel/locking/lockdep.c for (hlock = curr->held_locks + idx; idx < depth; idx++, hlock++) { idx 4108 kernel/locking/lockdep.c *merged += (idx == first_idx); idx 221 kernel/locking/lockdep_internals.h int idx; idx 223 kernel/locking/lockdep_internals.h idx = class - lock_classes; idx 224 kernel/locking/lockdep_internals.h __debug_atomic_inc(lock_class_ops[idx]); idx 229 kernel/locking/lockdep_internals.h int idx, cpu; idx 232 kernel/locking/lockdep_internals.h idx = class - lock_classes; idx 234 kernel/locking/lockdep_internals.h ops += per_cpu(lockdep_stats.lock_class_ops[idx], cpu); idx 114 kernel/locking/qspinlock.c static inline __pure u32 encode_tail(int cpu, int idx) idx 119 kernel/locking/qspinlock.c tail |= idx << _Q_TAIL_IDX_OFFSET; /* assume < 4 */ idx 127 kernel/locking/qspinlock.c int idx = (tail & _Q_TAIL_IDX_MASK) >> _Q_TAIL_IDX_OFFSET; idx 129 kernel/locking/qspinlock.c return per_cpu_ptr(&qnodes[idx].mcs, cpu); idx 133 kernel/locking/qspinlock.c struct mcs_spinlock *grab_mcs_node(struct mcs_spinlock *base, int idx) idx 135 kernel/locking/qspinlock.c return &((struct qnode *)base + idx)->mcs; idx 318 kernel/locking/qspinlock.c int idx; idx 400 kernel/locking/qspinlock.c idx = node->count++; idx 401 kernel/locking/qspinlock.c tail = encode_tail(smp_processor_id(), idx); idx 412 kernel/locking/qspinlock.c if (unlikely(idx >= MAX_NODES)) { idx 419 kernel/locking/qspinlock.c node = grab_mcs_node(node, idx); idx 424 kernel/locking/qspinlock.c lockevent_cond_inc(lock_use_node2 + idx - 1, idx); idx 398 kernel/module.c #define symversion(base, idx) NULL idx 400 kernel/module.c #define symversion(base, idx) ((base != NULL) ? ((base) + (idx)) : NULL) idx 497 kernel/notifier.c int idx; idx 499 kernel/notifier.c idx = srcu_read_lock(&nh->srcu); idx 501 kernel/notifier.c srcu_read_unlock(&nh->srcu, idx); idx 500 kernel/printk/printk.c static struct printk_log *log_from_idx(u32 idx) idx 502 kernel/printk/printk.c struct printk_log *msg = (struct printk_log *)(log_buf + idx); idx 514 kernel/printk/printk.c static u32 log_next(u32 idx) idx 516 kernel/printk/printk.c struct printk_log *msg = (struct printk_log *)(log_buf + idx); idx 528 kernel/printk/printk.c return idx + msg->len; idx 798 kernel/printk/printk.c u32 idx; idx 913 kernel/printk/printk.c user->idx = log_first_idx; idx 920 kernel/printk/printk.c msg = log_from_idx(user->idx); idx 927 kernel/printk/printk.c user->idx = log_next(user->idx); idx 960 kernel/printk/printk.c user->idx = log_first_idx; idx 969 kernel/printk/printk.c user->idx = clear_idx; idx 974 kernel/printk/printk.c user->idx = log_next_idx; idx 1033 kernel/printk/printk.c user->idx = log_first_idx; idx 1453 kernel/printk/printk.c u32 idx; idx 1467 kernel/printk/printk.c idx = clear_idx; idx 1469 kernel/printk/printk.c struct printk_log *msg = log_from_idx(idx); idx 1472 kernel/printk/printk.c idx = log_next(idx); idx 1478 kernel/printk/printk.c idx = clear_idx; idx 1480 kernel/printk/printk.c struct printk_log *msg = log_from_idx(idx); idx 1483 kernel/printk/printk.c idx = log_next(idx); idx 1492 kernel/printk/printk.c struct printk_log *msg = log_from_idx(idx); idx 1496 kernel/printk/printk.c idx = log_next(idx); idx 1509 kernel/printk/printk.c idx = log_first_idx; idx 1618 kernel/printk/printk.c u32 idx = syslog_idx; idx 1622 kernel/printk/printk.c struct printk_log *msg = log_from_idx(idx); idx 1627 kernel/printk/printk.c idx = log_next(idx); idx 2107 kernel/printk/printk.c static struct printk_log *log_from_idx(u32 idx) { return NULL; } idx 2108 kernel/printk/printk.c static u32 log_next(u32 idx) { return 0; } idx 2145 kernel/printk/printk.c static int __add_preferred_console(char *name, int idx, char *options, idx 2158 kernel/printk/printk.c if (strcmp(c->name, name) == 0 && c->index == idx) { idx 2172 kernel/printk/printk.c c->index = idx; idx 2194 kernel/printk/printk.c int idx; idx 2221 kernel/printk/printk.c idx = simple_strtoul(s, NULL, 10); idx 2224 kernel/printk/printk.c __add_preferred_console(buf, idx, options, brl_options); idx 2243 kernel/printk/printk.c int add_preferred_console(char *name, int idx, char *options) idx 2245 kernel/printk/printk.c return __add_preferred_console(name, idx, options, NULL); idx 3284 kernel/printk/printk.c u32 idx; idx 3309 kernel/printk/printk.c idx = dumper->cur_idx; idx 3311 kernel/printk/printk.c struct printk_log *msg = log_from_idx(idx); idx 3314 kernel/printk/printk.c idx = log_next(idx); idx 3320 kernel/printk/printk.c idx = dumper->cur_idx; idx 3322 kernel/printk/printk.c struct printk_log *msg = log_from_idx(idx); idx 3325 kernel/printk/printk.c idx = log_next(idx); idx 3331 kernel/printk/printk.c next_idx = idx; idx 3335 kernel/printk/printk.c struct printk_log *msg = log_from_idx(idx); idx 3338 kernel/printk/printk.c idx = log_next(idx); idx 133 kernel/rcu/rcuperf.c void (*readunlock)(int idx); idx 156 kernel/rcu/rcuperf.c static void rcu_perf_read_unlock(int idx) __releases(RCU) idx 197 kernel/rcu/rcuperf.c static void srcu_perf_read_unlock(int idx) __releases(srcu_ctlp) idx 199 kernel/rcu/rcuperf.c srcu_read_unlock(srcu_ctlp, idx); idx 280 kernel/rcu/rcuperf.c static void tasks_perf_read_unlock(int idx) idx 325 kernel/rcu/rcuperf.c int idx; idx 335 kernel/rcu/rcuperf.c idx = cur_ops->readlock(); idx 336 kernel/rcu/rcuperf.c cur_ops->readunlock(idx); idx 296 kernel/rcu/rcutorture.c void (*readunlock)(int idx); idx 364 kernel/rcu/rcutorture.c static void rcu_torture_read_unlock(int idx) __releases(RCU) idx 538 kernel/rcu/rcutorture.c static void srcu_torture_read_unlock(int idx) __releases(srcu_ctlp) idx 540 kernel/rcu/rcutorture.c srcu_read_unlock(srcu_ctlp, idx); idx 657 kernel/rcu/rcutorture.c static void tasks_torture_read_unlock(int idx) idx 706 kernel/rcu/rcutorture.c static void rcu_torture_read_unlock_trivial(int idx) __releases(RCU) idx 1773 kernel/rcu/rcutorture.c int idx; idx 1802 kernel/rcu/rcutorture.c idx = cur_ops->readlock(); idx 1804 kernel/rcu/rcutorture.c cur_ops->readunlock(idx); idx 94 kernel/rcu/srcutiny.c void __srcu_read_unlock(struct srcu_struct *ssp, int idx) idx 96 kernel/rcu/srcutiny.c int newval = ssp->srcu_lock_nesting[idx] - 1; idx 98 kernel/rcu/srcutiny.c WRITE_ONCE(ssp->srcu_lock_nesting[idx], newval); idx 111 kernel/rcu/srcutiny.c int idx; idx 127 kernel/rcu/srcutiny.c idx = ssp->srcu_idx; idx 130 kernel/rcu/srcutiny.c swait_event_exclusive(ssp->srcu_wq, !READ_ONCE(ssp->srcu_lock_nesting[idx])); idx 247 kernel/rcu/srcutree.c static unsigned long srcu_readers_lock_idx(struct srcu_struct *ssp, int idx) idx 255 kernel/rcu/srcutree.c sum += READ_ONCE(cpuc->srcu_lock_count[idx]); idx 264 kernel/rcu/srcutree.c static unsigned long srcu_readers_unlock_idx(struct srcu_struct *ssp, int idx) idx 272 kernel/rcu/srcutree.c sum += READ_ONCE(cpuc->srcu_unlock_count[idx]); idx 281 kernel/rcu/srcutree.c static bool srcu_readers_active_idx_check(struct srcu_struct *ssp, int idx) idx 285 kernel/rcu/srcutree.c unlocks = srcu_readers_unlock_idx(ssp, idx); idx 321 kernel/rcu/srcutree.c return srcu_readers_lock_idx(ssp, idx) == unlocks; idx 405 kernel/rcu/srcutree.c int idx; idx 407 kernel/rcu/srcutree.c idx = READ_ONCE(ssp->srcu_idx) & 0x1; idx 408 kernel/rcu/srcutree.c this_cpu_inc(ssp->sda->srcu_lock_count[idx]); idx 410 kernel/rcu/srcutree.c return idx; idx 419 kernel/rcu/srcutree.c void __srcu_read_unlock(struct srcu_struct *ssp, int idx) idx 422 kernel/rcu/srcutree.c this_cpu_inc(ssp->sda->srcu_unlock_count[idx]); idx 520 kernel/rcu/srcutree.c int idx; idx 530 kernel/rcu/srcutree.c idx = rcu_seq_state(ssp->srcu_gp_seq); idx 531 kernel/rcu/srcutree.c WARN_ON_ONCE(idx != SRCU_STATE_SCAN2); idx 543 kernel/rcu/srcutree.c idx = rcu_seq_ctr(gpseq) % ARRAY_SIZE(snp->srcu_have_cbs); idx 549 kernel/rcu/srcutree.c cbs = snp->srcu_have_cbs[idx] == gpseq; idx 550 kernel/rcu/srcutree.c snp->srcu_have_cbs[idx] = gpseq; idx 551 kernel/rcu/srcutree.c rcu_seq_set_state(&snp->srcu_have_cbs[idx], 1); idx 554 kernel/rcu/srcutree.c mask = snp->srcu_data_have_cbs[idx]; idx 555 kernel/rcu/srcutree.c snp->srcu_data_have_cbs[idx] = 0; idx 635 kernel/rcu/srcutree.c int idx = rcu_seq_ctr(s) % ARRAY_SIZE(sdp->mynode->srcu_have_cbs); idx 644 kernel/rcu/srcutree.c if (ULONG_CMP_GE(snp->srcu_have_cbs[idx], s)) { idx 645 kernel/rcu/srcutree.c snp_seq = snp->srcu_have_cbs[idx]; idx 647 kernel/rcu/srcutree.c snp->srcu_data_have_cbs[idx] |= sdp->grpmask; idx 659 kernel/rcu/srcutree.c snp->srcu_have_cbs[idx] = s; idx 661 kernel/rcu/srcutree.c snp->srcu_data_have_cbs[idx] |= sdp->grpmask; idx 698 kernel/rcu/srcutree.c static bool try_check_zero(struct srcu_struct *ssp, int idx, int trycount) idx 701 kernel/rcu/srcutree.c if (srcu_readers_active_idx_check(ssp, idx)) idx 839 kernel/rcu/srcutree.c int idx; idx 853 kernel/rcu/srcutree.c idx = srcu_read_lock(ssp); idx 875 kernel/rcu/srcutree.c srcu_read_unlock(ssp, idx); idx 1093 kernel/rcu/srcutree.c int idx; idx 1107 kernel/rcu/srcutree.c idx = rcu_seq_state(smp_load_acquire(&ssp->srcu_gp_seq)); /* ^^^ */ idx 1108 kernel/rcu/srcutree.c if (idx == SRCU_STATE_IDLE) { idx 1116 kernel/rcu/srcutree.c idx = rcu_seq_state(READ_ONCE(ssp->srcu_gp_seq)); idx 1117 kernel/rcu/srcutree.c if (idx == SRCU_STATE_IDLE) idx 1120 kernel/rcu/srcutree.c if (idx != SRCU_STATE_IDLE) { idx 1127 kernel/rcu/srcutree.c idx = 1 ^ (ssp->srcu_idx & 1); idx 1128 kernel/rcu/srcutree.c if (!try_check_zero(ssp, idx, 1)) { idx 1142 kernel/rcu/srcutree.c idx = 1 ^ (ssp->srcu_idx & 1); idx 1143 kernel/rcu/srcutree.c if (!try_check_zero(ssp, idx, 2)) { idx 1256 kernel/rcu/srcutree.c int idx; idx 1259 kernel/rcu/srcutree.c idx = ssp->srcu_idx & 0x1; idx 1261 kernel/rcu/srcutree.c tt, tf, rcu_seq_current(&ssp->srcu_gp_seq), idx); idx 1269 kernel/rcu/srcutree.c u0 = sdp->srcu_unlock_count[!idx]; idx 1270 kernel/rcu/srcutree.c u1 = sdp->srcu_unlock_count[idx]; idx 1278 kernel/rcu/srcutree.c l0 = sdp->srcu_lock_count[!idx]; idx 1279 kernel/rcu/srcutree.c l1 = sdp->srcu_lock_count[idx]; idx 213 kernel/sched/autogroup.c int err, idx; idx 232 kernel/sched/autogroup.c idx = array_index_nospec(nice + 20, 40); idx 233 kernel/sched/autogroup.c shares = scale_load(sched_prio_to_weight[idx]); idx 7788 kernel/sched/core.c int idx; idx 7793 kernel/sched/core.c idx = NICE_TO_PRIO(nice) - MAX_RT_PRIO; idx 7794 kernel/sched/core.c idx = array_index_nospec(idx, 40); idx 7795 kernel/sched/core.c weight = sched_prio_to_weight[idx]; idx 26 kernel/sched/cpudeadline.c static void cpudl_heapify_down(struct cpudl *cp, int idx) idx 30 kernel/sched/cpudeadline.c int orig_cpu = cp->elements[idx].cpu; idx 31 kernel/sched/cpudeadline.c u64 orig_dl = cp->elements[idx].dl; idx 33 kernel/sched/cpudeadline.c if (left_child(idx) >= cp->size) idx 40 kernel/sched/cpudeadline.c l = left_child(idx); idx 41 kernel/sched/cpudeadline.c r = right_child(idx); idx 42 kernel/sched/cpudeadline.c largest = idx; idx 54 kernel/sched/cpudeadline.c if (largest == idx) idx 58 kernel/sched/cpudeadline.c cp->elements[idx].cpu = cp->elements[largest].cpu; idx 59 kernel/sched/cpudeadline.c cp->elements[idx].dl = cp->elements[largest].dl; idx 60 kernel/sched/cpudeadline.c cp->elements[cp->elements[idx].cpu].idx = idx; idx 61 kernel/sched/cpudeadline.c idx = largest; idx 64 kernel/sched/cpudeadline.c cp->elements[idx].cpu = orig_cpu; idx 65 kernel/sched/cpudeadline.c cp->elements[idx].dl = orig_dl; idx 66 kernel/sched/cpudeadline.c cp->elements[cp->elements[idx].cpu].idx = idx; idx 69 kernel/sched/cpudeadline.c static void cpudl_heapify_up(struct cpudl *cp, int idx) idx 73 kernel/sched/cpudeadline.c int orig_cpu = cp->elements[idx].cpu; idx 74 kernel/sched/cpudeadline.c u64 orig_dl = cp->elements[idx].dl; idx 76 kernel/sched/cpudeadline.c if (idx == 0) idx 80 kernel/sched/cpudeadline.c p = parent(idx); idx 84 kernel/sched/cpudeadline.c cp->elements[idx].cpu = cp->elements[p].cpu; idx 85 kernel/sched/cpudeadline.c cp->elements[idx].dl = cp->elements[p].dl; idx 86 kernel/sched/cpudeadline.c cp->elements[cp->elements[idx].cpu].idx = idx; idx 87 kernel/sched/cpudeadline.c idx = p; idx 88 kernel/sched/cpudeadline.c } while (idx != 0); idx 90 kernel/sched/cpudeadline.c cp->elements[idx].cpu = orig_cpu; idx 91 kernel/sched/cpudeadline.c cp->elements[idx].dl = orig_dl; idx 92 kernel/sched/cpudeadline.c cp->elements[cp->elements[idx].cpu].idx = idx; idx 95 kernel/sched/cpudeadline.c static void cpudl_heapify(struct cpudl *cp, int idx) idx 97 kernel/sched/cpudeadline.c if (idx > 0 && dl_time_before(cp->elements[parent(idx)].dl, idx 98 kernel/sched/cpudeadline.c cp->elements[idx].dl)) idx 99 kernel/sched/cpudeadline.c cpudl_heapify_up(cp, idx); idx 101 kernel/sched/cpudeadline.c cpudl_heapify_down(cp, idx); idx 159 kernel/sched/cpudeadline.c old_idx = cp->elements[cpu].idx; idx 171 kernel/sched/cpudeadline.c cp->elements[new_cpu].idx = old_idx; idx 172 kernel/sched/cpudeadline.c cp->elements[cpu].idx = IDX_INVALID; idx 199 kernel/sched/cpudeadline.c old_idx = cp->elements[cpu].idx; idx 205 kernel/sched/cpudeadline.c cp->elements[cpu].idx = new_idx; idx 259 kernel/sched/cpudeadline.c cp->elements[i].idx = IDX_INVALID; idx 8 kernel/sched/cpudeadline.h int idx; idx 62 kernel/sched/cpupri.c int idx = 0; idx 67 kernel/sched/cpupri.c for (idx = 0; idx < task_pri; idx++) { idx 68 kernel/sched/cpupri.c struct cpupri_vec *vec = &cp->pri_to_cpu[idx]; idx 35 kernel/sched/cputime.c enum cpu_usage_stat idx) idx 40 kernel/sched/cputime.c cpustat[idx] += delta; idx 211 kernel/sched/loadavg.c int idx = calc_load_idx; idx 224 kernel/sched/loadavg.c idx++; idx 226 kernel/sched/loadavg.c return idx & 1; idx 240 kernel/sched/loadavg.c int idx = calc_load_write_idx(); idx 242 kernel/sched/loadavg.c atomic_long_add(delta, &calc_load_nohz[idx]); idx 286 kernel/sched/loadavg.c int idx = calc_load_read_idx(); idx 289 kernel/sched/loadavg.c if (atomic_long_read(&calc_load_nohz[idx])) idx 290 kernel/sched/loadavg.c delta = atomic_long_xchg(&calc_load_nohz[idx], 0); idx 1545 kernel/sched/rt.c int idx; idx 1547 kernel/sched/rt.c idx = sched_find_first_bit(array->bitmap); idx 1548 kernel/sched/rt.c BUG_ON(idx >= MAX_RT_PRIO); idx 1550 kernel/sched/rt.c queue = array->queue + idx; idx 489 kernel/time/hrtimer.c unsigned int idx; idx 494 kernel/time/hrtimer.c idx = __ffs(*active); idx 495 kernel/time/hrtimer.c *active &= ~(1U << idx); idx 497 kernel/time/hrtimer.c return &cpu_base->clock_base[idx]; idx 1359 kernel/time/posix-timers.c clockid_t idx = id; idx 1369 kernel/time/posix-timers.c return posix_clocks[array_index_nospec(idx, ARRAY_SIZE(posix_clocks))]; idx 481 kernel/time/timer.c static inline void timer_set_idx(struct timer_list *timer, unsigned int idx) idx 484 kernel/time/timer.c idx << TIMER_ARRAYSHIFT; idx 500 kernel/time/timer.c unsigned int idx; idx 503 kernel/time/timer.c idx = calc_index(expires, 0); idx 505 kernel/time/timer.c idx = calc_index(expires, 1); idx 507 kernel/time/timer.c idx = calc_index(expires, 2); idx 509 kernel/time/timer.c idx = calc_index(expires, 3); idx 511 kernel/time/timer.c idx = calc_index(expires, 4); idx 513 kernel/time/timer.c idx = calc_index(expires, 5); idx 515 kernel/time/timer.c idx = calc_index(expires, 6); idx 517 kernel/time/timer.c idx = calc_index(expires, 7); idx 519 kernel/time/timer.c idx = clk & LVL_MASK; idx 528 kernel/time/timer.c idx = calc_index(expires, LVL_DEPTH - 1); idx 530 kernel/time/timer.c return idx; idx 538 kernel/time/timer.c unsigned int idx) idx 540 kernel/time/timer.c hlist_add_head(&timer->entry, base->vectors + idx); idx 541 kernel/time/timer.c __set_bit(idx, base->pending_map); idx 542 kernel/time/timer.c timer_set_idx(timer, idx); idx 550 kernel/time/timer.c unsigned int idx; idx 552 kernel/time/timer.c idx = calc_wheel_index(timer->expires, base->clk); idx 553 kernel/time/timer.c enqueue_timer(base, timer, idx); idx 824 kernel/time/timer.c unsigned idx = timer_get_idx(timer); idx 829 kernel/time/timer.c if (hlist_is_singular_node(&timer->entry, base->vectors + idx)) idx 830 kernel/time/timer.c __clear_bit(idx, base->pending_map); idx 952 kernel/time/timer.c unsigned int idx = UINT_MAX; idx 992 kernel/time/timer.c idx = calc_wheel_index(expires, clk); idx 999 kernel/time/timer.c if (idx == timer_get_idx(timer)) { idx 1049 kernel/time/timer.c if (idx != UINT_MAX && clk == base->clk) { idx 1050 kernel/time/timer.c enqueue_timer(base, timer, idx); idx 1463 kernel/time/timer.c unsigned int idx; idx 1466 kernel/time/timer.c idx = (clk & LVL_MASK) + i * LVL_SIZE; idx 1468 kernel/time/timer.c if (__test_and_clear_bit(idx, base->pending_map)) { idx 1469 kernel/time/timer.c vec = base->vectors + idx; idx 57 kernel/time/timer_list.c int idx, u64 now) idx 59 kernel/time/timer_list.c SEQ_printf(m, " #%d: ", idx); idx 563 kernel/trace/bpf_trace.c BPF_CALL_2(bpf_current_task_under_cgroup, struct bpf_map *, map, u32, idx) idx 568 kernel/trace/bpf_trace.c if (unlikely(idx >= array->map.max_entries)) idx 571 kernel/trace/bpf_trace.c cgrp = READ_ONCE(array->ptrs[idx]); idx 247 kernel/trace/fgraph.c ftrace_graph_get_ret_stack(struct task_struct *task, int idx) idx 249 kernel/trace/fgraph.c idx = task->curr_ret_stack - idx; idx 251 kernel/trace/fgraph.c if (idx >= 0 && idx <= task->curr_ret_stack) idx 252 kernel/trace/fgraph.c return &task->ret_stack[idx]; idx 273 kernel/trace/fgraph.c unsigned long ftrace_graph_ret_addr(struct task_struct *task, int *idx, idx 292 kernel/trace/fgraph.c unsigned long ftrace_graph_ret_addr(struct task_struct *task, int *idx, idx 302 kernel/trace/fgraph.c if (!task->ret_stack || task_idx < *idx) idx 305 kernel/trace/fgraph.c task_idx -= *idx; idx 306 kernel/trace/fgraph.c (*idx)++; idx 430 kernel/trace/ftrace.c function_stat_next(void *v, int idx) idx 438 kernel/trace/ftrace.c if (idx != 0) idx 3064 kernel/trace/ftrace.c int idx; idx 3284 kernel/trace/ftrace.c if (iter->idx >= iter->pg->index) { idx 3287 kernel/trace/ftrace.c iter->idx = 0; idx 3291 kernel/trace/ftrace.c rec = &iter->pg->records[iter->idx++]; idx 3391 kernel/trace/ftrace.c iter->idx = 0; idx 4703 kernel/trace/ftrace.c parser->idx, enable); idx 5036 kernel/trace/ftrace.c ftrace_match_records(iter->hash, parser->buffer, parser->idx); idx 5118 kernel/trace/ftrace.c int idx; /* for hash table iteration */ idx 5131 kernel/trace/ftrace.c int i, idx = fgd->idx; idx 5142 kernel/trace/ftrace.c idx++; idx 5145 kernel/trace/ftrace.c for (i = idx; i < 1 << fgd->hash->size_bits; i++) { idx 5149 kernel/trace/ftrace.c fgd->idx = i; idx 5180 kernel/trace/ftrace.c fgd->idx = 0; idx 1436 kernel/trace/trace.c parser->idx = 0; idx 1448 kernel/trace/trace.c if (parser->idx < parser->size - 1) idx 1449 kernel/trace/trace.c parser->buffer[parser->idx++] = ch; idx 1463 kernel/trace/trace.c parser->buffer[parser->idx] = 0; idx 1465 kernel/trace/trace.c } else if (parser->idx < parser->size - 1) { idx 1467 kernel/trace/trace.c parser->buffer[parser->idx++] = ch; idx 1469 kernel/trace/trace.c parser->buffer[parser->idx] = 0; idx 1953 kernel/trace/trace.c static inline char *get_saved_cmdlines(int idx) idx 1955 kernel/trace/trace.c return &savedcmd->saved_cmdlines[idx * TASK_COMM_LEN]; idx 1958 kernel/trace/trace.c static inline void set_cmdline(int idx, const char *cmdline) idx 1960 kernel/trace/trace.c strncpy(get_saved_cmdlines(idx), cmdline, TASK_COMM_LEN); idx 2142 kernel/trace/trace.c unsigned pid, idx; idx 2160 kernel/trace/trace.c idx = savedcmd->map_pid_to_cmdline[tsk->pid]; idx 2161 kernel/trace/trace.c if (idx == NO_CMDLINE_MAP) { idx 2162 kernel/trace/trace.c idx = (savedcmd->cmdline_idx + 1) % savedcmd->cmdline_num; idx 2170 kernel/trace/trace.c pid = savedcmd->map_cmdline_to_pid[idx]; idx 2174 kernel/trace/trace.c savedcmd->map_cmdline_to_pid[idx] = tsk->pid; idx 2175 kernel/trace/trace.c savedcmd->map_pid_to_cmdline[tsk->pid] = idx; idx 2177 kernel/trace/trace.c savedcmd->cmdline_idx = idx; idx 2180 kernel/trace/trace.c set_cmdline(idx, tsk->comm); idx 3270 kernel/trace/trace.c iter->idx++; idx 3392 kernel/trace/trace.c if (iter->idx > i) idx 3395 kernel/trace/trace.c if (iter->idx < 0) idx 3400 kernel/trace/trace.c while (ent && iter->idx < i) idx 3473 kernel/trace/trace.c iter->idx = -1; idx 3718 kernel/trace/trace.c if (iter->idx > 1) idx 6104 kernel/trace/trace.c unsigned int idx) idx 6106 kernel/trace/trace.c __free_page(spd->pages[idx]); idx 1184 kernel/trace/trace.h unsigned idx; idx 1190 kernel/trace/trace.h return (parser->idx != 0); idx 1201 kernel/trace/trace.h parser->idx = 0; idx 323 kernel/trace/trace_branch.c annotated_branch_stat_next(void *v, int idx) idx 413 kernel/trace/trace_branch.c all_branch_stat_next(void *v, int idx) idx 112 kernel/trace/trace_events_hist.c unsigned int idx; idx 1575 kernel/trace/trace_events_hist.c if (hist_field && hist_field->var.idx == var_idx && idx 1664 kernel/trace/trace_events_hist.c if (find_any_var_ref(hist_data, field->var.idx)) { idx 1956 kernel/trace/trace_events_hist.c var_idx = hist_field->var.idx; idx 2635 kernel/trace/trace_events_hist.c ref_field->var.idx = var_field->var.idx; idx 2693 kernel/trace/trace_events_hist.c if (ref_field->var.idx == var_field->var.idx && idx 2727 kernel/trace/trace_events_hist.c if (ref_field->var.idx == var_field->var.idx && idx 3478 kernel/trace/trace_events_hist.c var_idx = var->var.idx; idx 3514 kernel/trace/trace_events_hist.c int idx; idx 3527 kernel/trace/trace_events_hist.c idx = tracing_map_add_var(hist_data->map); idx 3528 kernel/trace/trace_events_hist.c if (idx < 0) { idx 3536 kernel/trace/trace_events_hist.c var->var.idx = idx; idx 3663 kernel/trace/trace_events_hist.c unsigned int track_var_idx = data->track_data.track_var->var.idx; idx 3675 kernel/trace/trace_events_hist.c unsigned int track_var_idx = data->track_data.track_var->var.idx; idx 3835 kernel/trace/trace_events_hist.c save_var_idx = save_var->var.idx; idx 4880 kernel/trace/trace_events_hist.c unsigned int idx; idx 4886 kernel/trace/trace_events_hist.c idx = k++; idx 4891 kernel/trace/trace_events_hist.c sort_key->field_idx = idx; idx 5181 kernel/trace/trace_events_hist.c int i, idx = 0; idx 5200 kernel/trace/trace_events_hist.c idx = tracing_map_add_key_field(map, idx 5204 kernel/trace/trace_events_hist.c idx = tracing_map_add_sum_field(map); idx 5206 kernel/trace/trace_events_hist.c if (idx < 0) idx 5207 kernel/trace/trace_events_hist.c return idx; idx 5210 kernel/trace/trace_events_hist.c idx = tracing_map_add_var(map); idx 5211 kernel/trace/trace_events_hist.c if (idx < 0) idx 5212 kernel/trace/trace_events_hist.c return idx; idx 5213 kernel/trace/trace_events_hist.c hist_field->var.idx = idx; idx 5293 kernel/trace/trace_events_hist.c var_idx = hist_field->var.idx; idx 5304 kernel/trace/trace_events_hist.c var_idx = hist_field->var.idx; idx 5723 kernel/trace/trace_events_hist.c unsigned int idx, first_key_idx; idx 5729 kernel/trace/trace_events_hist.c idx = sort_key->field_idx; idx 5731 kernel/trace/trace_events_hist.c if (WARN_ON(idx >= HIST_FIELDS_MAX)) idx 5737 kernel/trace/trace_events_hist.c if (idx == HITCOUNT_IDX) idx 5740 kernel/trace/trace_events_hist.c if (idx >= first_key_idx) idx 5741 kernel/trace/trace_events_hist.c idx += hist_data->n_vars; idx 5742 kernel/trace/trace_events_hist.c hist_field_print(m, hist_data->fields[idx]); idx 629 kernel/trace/trace_output.c entry->preempt_count, iter->idx); idx 17 kernel/trace/trace_stat.h void *(*stat_next)(void *prev, int idx); idx 274 kernel/trace/tracing_map.c int idx = tracing_map_add_field(map, cmp_fn); idx 276 kernel/trace/tracing_map.c if (idx < 0) idx 277 kernel/trace/tracing_map.c return idx; idx 279 kernel/trace/tracing_map.c map->fields[idx].offset = offset; idx 281 kernel/trace/tracing_map.c map->key_idx[map->n_keys++] = idx; idx 283 kernel/trace/tracing_map.c return idx; idx 452 kernel/trace/tracing_map.c int idx; idx 454 kernel/trace/tracing_map.c idx = atomic_inc_return(&map->next_elt); idx 455 kernel/trace/tracing_map.c if (idx < map->max_elts) { idx 456 kernel/trace/tracing_map.c elt = *(TRACING_MAP_ELT(map->elts, idx)); idx 515 kernel/trace/tracing_map.c u32 idx, key_hash, test_key; idx 523 kernel/trace/tracing_map.c idx = key_hash >> (32 - (map->map_bits + 1)); idx 526 kernel/trace/tracing_map.c idx &= (map->map_size - 1); idx 527 kernel/trace/tracing_map.c entry = TRACING_MAP_ENTRY(map->map, idx); idx 588 kernel/trace/tracing_map.c idx++; idx 173 kernel/trace/tracing_map.h #define TRACING_MAP_ARRAY_ELT(array, idx) \ idx 174 kernel/trace/tracing_map.h (array->pages[idx >> array->entry_shift] + \ idx 175 kernel/trace/tracing_map.h ((idx & array->entry_mask) << array->entry_size_shift)) idx 177 kernel/trace/tracing_map.h #define TRACING_MAP_ENTRY(array, idx) \ idx 178 kernel/trace/tracing_map.h ((struct tracing_map_entry *)TRACING_MAP_ARRAY_ELT(array, idx)) idx 180 kernel/trace/tracing_map.h #define TRACING_MAP_ELT(array, idx) \ idx 181 kernel/trace/tracing_map.h ((struct tracing_map_elt **)TRACING_MAP_ARRAY_ELT(array, idx)) idx 274 kernel/user_namespace.c unsigned idx; idx 280 kernel/user_namespace.c for (idx = 0; idx < extents; idx++) { idx 281 kernel/user_namespace.c first = map->extent[idx].first; idx 282 kernel/user_namespace.c last = first + map->extent[idx].count - 1; idx 285 kernel/user_namespace.c return &map->extent[idx]; idx 323 kernel/user_namespace.c unsigned idx; idx 327 kernel/user_namespace.c for (idx = 0; idx < extents; idx++) { idx 328 kernel/user_namespace.c first = map->extent[idx].lower_first; idx 329 kernel/user_namespace.c last = first + map->extent[idx].count - 1; idx 331 kernel/user_namespace.c return &map->extent[idx]; idx 713 kernel/user_namespace.c unsigned idx; idx 720 kernel/user_namespace.c for (idx = 0; idx < new_map->nr_extents; idx++) { idx 726 kernel/user_namespace.c prev = &new_map->extent[idx]; idx 728 kernel/user_namespace.c prev = &new_map->forward[idx]; idx 853 kernel/user_namespace.c unsigned idx; idx 975 kernel/user_namespace.c for (idx = 0; idx < new_map.nr_extents; idx++) { idx 980 kernel/user_namespace.c e = &new_map.extent[idx]; idx 982 kernel/user_namespace.c e = &new_map.forward[idx]; idx 62 lib/dynamic_debug.c unsigned int idx; idx 731 lib/dynamic_debug.c iter->idx = 0; idx 736 lib/dynamic_debug.c iter->idx = 0; idx 737 lib/dynamic_debug.c return &iter->table->ddebugs[iter->idx]; idx 750 lib/dynamic_debug.c if (++iter->idx == iter->table->num_ddebugs) { idx 752 lib/dynamic_debug.c iter->idx = 0; idx 760 lib/dynamic_debug.c return &iter->table->ddebugs[iter->idx]; idx 100 lib/find_bit.c unsigned long idx; idx 102 lib/find_bit.c for (idx = 0; idx * BITS_PER_LONG < size; idx++) { idx 103 lib/find_bit.c if (addr[idx]) idx 104 lib/find_bit.c return min(idx * BITS_PER_LONG + __ffs(addr[idx]), size); idx 118 lib/find_bit.c unsigned long idx; idx 120 lib/find_bit.c for (idx = 0; idx * BITS_PER_LONG < size; idx++) { idx 121 lib/find_bit.c if (addr[idx] != ~0UL) idx 122 lib/find_bit.c return min(idx * BITS_PER_LONG + ffz(addr[idx]), size); idx 135 lib/find_bit.c unsigned long idx = (size-1) / BITS_PER_LONG; idx 138 lib/find_bit.c val &= addr[idx]; idx 140 lib/find_bit.c return idx * BITS_PER_LONG + __fls(val); idx 143 lib/find_bit.c } while (idx--); idx 5137 lib/fonts/font_10x18.c .idx = FONT10x18_IDX, idx 3081 lib/fonts/font_6x10.c .idx = FONT6x10_IDX, idx 3346 lib/fonts/font_6x11.c .idx = VGA6x11_IDX, idx 4113 lib/fonts/font_7x14.c .idx = FONT7x14_IDX, idx 4627 lib/fonts/font_8x16.c .idx = VGA8x16_IDX, idx 2578 lib/fonts/font_8x8.c .idx = VGA8x8_IDX, idx 266 lib/fonts/font_acorn_8x8.c .idx = ACORN8x8_IDX, idx 2151 lib/fonts/font_mini_4x6.c .idx = MINI4x6_IDX, idx 2582 lib/fonts/font_pearl_8x8.c .idx = PEARL8x8_IDX, idx 6156 lib/fonts/font_sun12x22.c .idx = SUN12x22_IDX, idx 266 lib/fonts/font_sun8x16.c .idx = SUN8x16_IDX, idx 2062 lib/fonts/font_ter16x32.c .idx = TER16x32_IDX, idx 119 lib/fonts/fonts.c if (MACH_IS_AMIGA && f->idx == PEARL8x8_IDX) idx 123 lib/fonts/fonts.c if (MACH_IS_MAC && xres < 640 && f->idx == VGA6x11_IDX) idx 328 lib/iov_iter.c int idx = i->idx; idx 334 lib/iov_iter.c if (unlikely(idx != ((next - 1) & (pipe->buffers - 1)))) idx 337 lib/iov_iter.c p = &pipe->bufs[idx]; idx 341 lib/iov_iter.c if (idx != (next & (pipe->buffers - 1))) idx 346 lib/iov_iter.c printk(KERN_ERR "idx = %d, offset = %zd\n", i->idx, i->iov_offset); idx 349 lib/iov_iter.c for (idx = 0; idx < pipe->buffers; idx++) idx 351 lib/iov_iter.c pipe->bufs[idx].ops, idx 352 lib/iov_iter.c pipe->bufs[idx].page, idx 353 lib/iov_iter.c pipe->bufs[idx].offset, idx 354 lib/iov_iter.c pipe->bufs[idx].len); idx 362 lib/iov_iter.c static inline int next_idx(int idx, struct pipe_inode_info *pipe) idx 364 lib/iov_iter.c return (idx + 1) & (pipe->buffers - 1); idx 373 lib/iov_iter.c int idx; idx 385 lib/iov_iter.c idx = i->idx; idx 386 lib/iov_iter.c buf = &pipe->bufs[idx]; idx 394 lib/iov_iter.c idx = next_idx(idx, pipe); idx 395 lib/iov_iter.c buf = &pipe->bufs[idx]; idx 397 lib/iov_iter.c if (idx == pipe->curbuf && pipe->nrbufs) idx 405 lib/iov_iter.c i->idx = idx; idx 486 lib/iov_iter.c int idx = i->idx; idx 487 lib/iov_iter.c if (off && (!allocated(&i->pipe->bufs[idx]) || off == PAGE_SIZE)) { idx 488 lib/iov_iter.c idx = next_idx(idx, i->pipe); idx 491 lib/iov_iter.c *idxp = idx; idx 500 lib/iov_iter.c int idx; idx 509 lib/iov_iter.c data_start(i, &idx, &off); idx 510 lib/iov_iter.c *idxp = idx; idx 515 lib/iov_iter.c pipe->bufs[idx].len += size; idx 518 lib/iov_iter.c pipe->bufs[idx].len = PAGE_SIZE; idx 519 lib/iov_iter.c idx = next_idx(idx, pipe); idx 521 lib/iov_iter.c while (idx != pipe->curbuf || !pipe->nrbufs) { idx 526 lib/iov_iter.c pipe->bufs[idx].ops = &default_pipe_buf_ops; idx 527 lib/iov_iter.c pipe->bufs[idx].page = page; idx 528 lib/iov_iter.c pipe->bufs[idx].offset = 0; idx 530 lib/iov_iter.c pipe->bufs[idx].len = left; idx 533 lib/iov_iter.c pipe->bufs[idx].len = PAGE_SIZE; idx 535 lib/iov_iter.c idx = next_idx(idx, pipe); idx 545 lib/iov_iter.c int idx; idx 550 lib/iov_iter.c bytes = n = push_pipe(i, bytes, &idx, &off); idx 553 lib/iov_iter.c for ( ; n; idx = next_idx(idx, pipe), off = 0) { idx 555 lib/iov_iter.c memcpy_to_page(pipe->bufs[idx].page, off, addr, chunk); idx 556 lib/iov_iter.c i->idx = idx; idx 579 lib/iov_iter.c int idx; idx 584 lib/iov_iter.c bytes = n = push_pipe(i, bytes, &idx, &r); idx 587 lib/iov_iter.c for ( ; n; idx = next_idx(idx, pipe), r = 0) { idx 589 lib/iov_iter.c char *p = kmap_atomic(pipe->bufs[idx].page); idx 592 lib/iov_iter.c i->idx = idx; idx 649 lib/iov_iter.c int idx; idx 654 lib/iov_iter.c bytes = n = push_pipe(i, bytes, &idx, &off); idx 657 lib/iov_iter.c for ( ; n; idx = next_idx(idx, pipe), off = 0) { idx 661 lib/iov_iter.c rem = memcpy_mcsafe_to_page(pipe->bufs[idx].page, off, addr, idx 663 lib/iov_iter.c i->idx = idx; idx 929 lib/iov_iter.c int idx; idx 934 lib/iov_iter.c bytes = n = push_pipe(i, bytes, &idx, &off); idx 938 lib/iov_iter.c for ( ; n; idx = next_idx(idx, pipe), off = 0) { idx 940 lib/iov_iter.c memzero_page(pipe->bufs[idx].page, off, chunk); idx 941 lib/iov_iter.c i->idx = idx; idx 992 lib/iov_iter.c int idx = i->idx; idx 993 lib/iov_iter.c int nrbufs = (idx - pipe->curbuf) & (pipe->buffers - 1); idx 995 lib/iov_iter.c pipe->bufs[idx].len = off - pipe->bufs[idx].offset; idx 996 lib/iov_iter.c idx = next_idx(idx, pipe); idx 1000 lib/iov_iter.c pipe_buf_release(pipe, &pipe->bufs[idx]); idx 1001 lib/iov_iter.c idx = next_idx(idx, pipe); idx 1015 lib/iov_iter.c int idx = i->idx; idx 1017 lib/iov_iter.c left += off - pipe->bufs[idx].offset; idx 1019 lib/iov_iter.c buf = &pipe->bufs[idx]; idx 1023 lib/iov_iter.c idx = next_idx(idx, pipe); idx 1025 lib/iov_iter.c i->idx = idx; idx 1056 lib/iov_iter.c int idx = i->idx; idx 1059 lib/iov_iter.c size_t n = off - pipe->bufs[idx].offset; idx 1065 lib/iov_iter.c if (!unroll && idx == i->start_idx) { idx 1069 lib/iov_iter.c if (!idx--) idx 1070 lib/iov_iter.c idx = pipe->buffers - 1; idx 1071 lib/iov_iter.c off = pipe->bufs[idx].offset + pipe->bufs[idx].len; idx 1074 lib/iov_iter.c i->idx = idx; idx 1165 lib/iov_iter.c i->idx = (pipe->curbuf + pipe->nrbufs) & (pipe->buffers - 1); idx 1168 lib/iov_iter.c i->start_idx = i->idx; idx 1196 lib/iov_iter.c if (size && i->iov_offset && allocated(&i->pipe->bufs[i->idx])) idx 1234 lib/iov_iter.c int idx, idx 1238 lib/iov_iter.c ssize_t n = push_pipe(i, maxsize, &idx, start); idx 1245 lib/iov_iter.c get_page(*pages++ = pipe->bufs[idx].page); idx 1246 lib/iov_iter.c idx = next_idx(idx, pipe); idx 1259 lib/iov_iter.c int idx; idx 1267 lib/iov_iter.c data_start(i, &idx, start); idx 1269 lib/iov_iter.c npages = ((i->pipe->curbuf - idx - 1) & (i->pipe->buffers - 1)) + 1; idx 1272 lib/iov_iter.c return __pipe_get_pages(i, min(maxsize, capacity), pages, idx, start); idx 1327 lib/iov_iter.c int idx; idx 1336 lib/iov_iter.c data_start(i, &idx, start); idx 1338 lib/iov_iter.c npages = ((i->pipe->curbuf - idx - 1) & (i->pipe->buffers - 1)) + 1; idx 1347 lib/iov_iter.c n = __pipe_get_pages(i, maxsize, p, idx, start); idx 1564 lib/iov_iter.c int idx; idx 1569 lib/iov_iter.c data_start(i, &idx, &off); idx 1571 lib/iov_iter.c npages = ((pipe->curbuf - idx - 1) & (pipe->buffers - 1)) + 1; idx 81 lib/lz4/lz4hc_compress.c U32 idx = hc4->nextToUpdate; idx 83 lib/lz4/lz4hc_compress.c while (idx < target) { idx 84 lib/lz4/lz4hc_compress.c U32 const h = LZ4HC_hashPtr(base + idx); idx 85 lib/lz4/lz4hc_compress.c size_t delta = idx - hashTable[h]; idx 90 lib/lz4/lz4hc_compress.c DELTANEXTU16(idx) = (U16)delta; idx 92 lib/lz4/lz4hc_compress.c hashTable[h] = idx; idx 93 lib/lz4/lz4hc_compress.c idx++; idx 164 lib/radix-tree.c unsigned idx; idx 165 lib/radix-tree.c for (idx = 0; idx < RADIX_TREE_TAG_LONGS; idx++) { idx 166 lib/radix-tree.c if (node->tags[tag][idx]) idx 6753 lib/test_bpf.c int idx = find_test_index(test_name); idx 6755 lib/test_bpf.c if (idx < 0) { idx 6760 lib/test_bpf.c test_range[0] = idx; idx 6761 lib/test_bpf.c test_range[1] = idx; idx 36 lib/test_firmware.c u8 idx; idx 636 lib/test_firmware.c req->idx, req->rc); idx 642 lib/test_firmware.c req->idx, req->fw->size); idx 681 lib/test_firmware.c req->idx = i; idx 686 lib/test_firmware.c "%s-%u", KBUILD_MODNAME, req->idx); idx 688 lib/test_firmware.c pr_err("Setting up thread %u failed\n", req->idx); idx 736 lib/test_firmware.c if (!req->idx) idx 783 lib/test_firmware.c req->idx = i; idx 849 lib/test_firmware.c u8 idx; idx 854 lib/test_firmware.c idx = test_fw_config->read_fw_idx; idx 855 lib/test_firmware.c if (idx >= test_fw_config->num_requests) { idx 865 lib/test_firmware.c req = &test_fw_config->reqs[idx]; idx 867 lib/test_firmware.c pr_err("#%u: failed to async load firmware\n", idx); idx 872 lib/test_firmware.c pr_info("#%u: loaded %zu\n", idx, req->fw->size); idx 180 lib/test_kmod.c unsigned int idx) idx 190 lib/test_kmod.c dev_info(test_dev->dev, "Last thread to run: %u\n", idx); idx 295 lib/test_kmod.c unsigned int idx; idx 303 lib/test_kmod.c for (idx=0; idx < config->num_threads; idx++) { idx 304 lib/test_kmod.c info = &test_dev->info[idx]; idx 319 lib/test_kmod.c static int try_one_request(struct kmod_test_device *test_dev, unsigned int idx) idx 321 lib/test_kmod.c struct kmod_test_device_info *info = &test_dev->info[idx]; idx 326 lib/test_kmod.c info->thread_idx = idx; idx 329 lib/test_kmod.c KBUILD_MODNAME, idx); idx 333 lib/test_kmod.c dev_err(test_dev->dev, "Setting up thread %u failed\n", idx); idx 337 lib/test_kmod.c dev_dbg(test_dev->dev, "Kicked off thread %u\n", idx); idx 391 lib/test_kmod.c unsigned int idx; idx 395 lib/test_kmod.c for (idx=0; idx < config->num_threads; idx++) { idx 401 lib/test_kmod.c ret = try_one_request(test_dev, idx); idx 1080 lib/test_kmod.c static struct kmod_test_device *alloc_test_dev_kmod(int idx) idx 1102 lib/test_kmod.c test_dev->dev_idx = idx; idx 1106 lib/test_kmod.c misc_dev->name = kasprintf(GFP_KERNEL, "test_kmod%d", idx); idx 1238 lib/test_xarray.c static noinline void check_move_small(struct xarray *xa, unsigned long idx) idx 1244 lib/test_xarray.c xa_store_index(xa, idx, GFP_KERNEL); idx 1247 lib/test_xarray.c for (i = 0; i < idx * 4; i++) { idx 1249 lib/test_xarray.c if (i <= idx) idx 1252 lib/test_xarray.c if (i == 0 || i == idx) idx 1263 lib/test_xarray.c if (i <= idx) idx 1266 lib/test_xarray.c if (i == 0 || i == idx) idx 1282 lib/test_xarray.c xa_erase_index(xa, idx); idx 346 lib/vsprintf.c int idx, len; idx 361 lib/vsprintf.c for (idx = 0; idx < width; idx++) idx 362 lib/vsprintf.c buf[idx] = ' '; idx 367 lib/vsprintf.c for (idx = 0; idx < len; ++idx) idx 368 lib/vsprintf.c buf[idx + width] = tmp[len - idx - 1]; idx 1728 lib/zstd/compress.c U32 idx = zc->nextToUpdate; idx 1730 lib/zstd/compress.c while (idx < target) idx 1731 lib/zstd/compress.c idx += ZSTD_insertBt1(zc, base + idx, mls, iend, nbCompares, 0); idx 1759 lib/zstd/compress.c U32 idx = zc->nextToUpdate; idx 1761 lib/zstd/compress.c while (idx < target) idx 1762 lib/zstd/compress.c idx += ZSTD_insertBt1(zc, base + idx, mls, iend, nbCompares, 1); idx 1804 lib/zstd/compress.c U32 idx = zc->nextToUpdate; idx 1806 lib/zstd/compress.c while (idx < target) { /* catch up */ idx 1807 lib/zstd/compress.c size_t const h = ZSTD_hashPtr(base + idx, hashLog, mls); idx 1808 lib/zstd/compress.c NEXT_IN_CHAIN(idx, chainMask) = hashTable[h]; idx 1809 lib/zstd/compress.c hashTable[h] = idx; idx 1810 lib/zstd/compress.c idx++; idx 220 lib/zstd/zstd_opt.h U32 idx = zc->nextToUpdate3; idx 224 lib/zstd/zstd_opt.h while (idx < target) { idx 225 lib/zstd/zstd_opt.h hashTable3[ZSTD_hash3Ptr(base + idx, hashLog3)] = idx; idx 226 lib/zstd/zstd_opt.h idx++; idx 57 mm/early_ioremap.c static inline void __init __late_set_fixmap(enum fixed_addresses idx, idx 65 mm/early_ioremap.c static inline void __init __late_clear_fixmap(enum fixed_addresses idx) idx 111 mm/early_ioremap.c enum fixed_addresses idx; idx 151 mm/early_ioremap.c idx = FIX_BTMAP_BEGIN - NR_FIX_BTMAPS*slot; idx 154 mm/early_ioremap.c __late_set_fixmap(idx, phys_addr, prot); idx 156 mm/early_ioremap.c __early_set_fixmap(idx, phys_addr, prot); idx 158 mm/early_ioremap.c --idx; idx 173 mm/early_ioremap.c enum fixed_addresses idx; idx 203 mm/early_ioremap.c idx = FIX_BTMAP_BEGIN - NR_FIX_BTMAPS*slot; idx 206 mm/early_ioremap.c __late_clear_fixmap(idx); idx 208 mm/early_ioremap.c __early_set_fixmap(idx, 0, FIXMAP_PAGE_CLEAR); idx 209 mm/early_ioremap.c --idx; idx 1990 mm/hugetlb.c pgoff_t idx; idx 1997 mm/hugetlb.c idx = vma_hugecache_offset(h, vma, addr); idx 2000 mm/hugetlb.c ret = region_chg(resv, idx, idx + 1); idx 2003 mm/hugetlb.c ret = region_add(resv, idx, idx + 1); idx 2006 mm/hugetlb.c region_abort(resv, idx, idx + 1); idx 2011 mm/hugetlb.c ret = region_add(resv, idx, idx + 1); idx 2013 mm/hugetlb.c region_abort(resv, idx, idx + 1); idx 2014 mm/hugetlb.c ret = region_del(resv, idx, idx + 1); idx 2122 mm/hugetlb.c int ret, idx; idx 2125 mm/hugetlb.c idx = hstate_index(h); idx 2161 mm/hugetlb.c ret = hugetlb_cgroup_charge_cgroup(idx, pages_per_huge_page(h), &h_cg); idx 2185 mm/hugetlb.c hugetlb_cgroup_commit_charge(idx, pages_per_huge_page(h), h_cg, page); idx 2209 mm/hugetlb.c hugetlb_cgroup_uncharge_cgroup(idx, pages_per_huge_page(h), h_cg); idx 2865 mm/hugetlb.c int idx = hstate_index(h); idx 2866 mm/hugetlb.c if (nhs->hstate_kobjs[idx]) { idx 2867 mm/hugetlb.c kobject_put(nhs->hstate_kobjs[idx]); idx 2868 mm/hugetlb.c nhs->hstate_kobjs[idx] = NULL; idx 3856 mm/hugetlb.c pgoff_t idx; idx 3859 mm/hugetlb.c idx = vma_hugecache_offset(h, vma, address); idx 3861 mm/hugetlb.c return find_lock_page(mapping, idx); idx 3872 mm/hugetlb.c pgoff_t idx; idx 3876 mm/hugetlb.c idx = vma_hugecache_offset(h, vma, address); idx 3878 mm/hugetlb.c page = find_get_page(mapping, idx); idx 3885 mm/hugetlb.c pgoff_t idx) idx 3889 mm/hugetlb.c int err = add_to_page_cache(page, mapping, idx, GFP_KERNEL); idx 3909 mm/hugetlb.c struct address_space *mapping, pgoff_t idx, idx 3938 mm/hugetlb.c page = find_lock_page(mapping, idx); idx 3941 mm/hugetlb.c if (idx >= size) idx 3967 mm/hugetlb.c hash = hugetlb_fault_mutex_hash(h, mapping, idx, haddr); idx 4003 mm/hugetlb.c int err = huge_add_to_page_cache(page, mapping, idx); idx 4048 mm/hugetlb.c if (idx >= size) idx 4095 mm/hugetlb.c pgoff_t idx, unsigned long address) idx 4101 mm/hugetlb.c key[1] = idx; idx 4113 mm/hugetlb.c pgoff_t idx, unsigned long address) idx 4126 mm/hugetlb.c pgoff_t idx; idx 4150 mm/hugetlb.c idx = vma_hugecache_offset(h, vma, haddr); idx 4157 mm/hugetlb.c hash = hugetlb_fault_mutex_hash(h, mapping, idx, haddr); idx 4162 mm/hugetlb.c ret = hugetlb_no_page(mm, vma, mapping, idx, address, ptep, flags); idx 4268 mm/hugetlb.c pgoff_t idx; idx 4307 mm/hugetlb.c idx = vma_hugecache_offset(h, dst_vma, dst_addr); idx 4315 mm/hugetlb.c if (idx >= size) idx 4324 mm/hugetlb.c ret = huge_add_to_page_cache(page, mapping, idx); idx 4343 mm/hugetlb.c if (idx >= size) idx 4805 mm/hugetlb.c unsigned long addr, pgoff_t idx) idx 4807 mm/hugetlb.c unsigned long saddr = ((idx - svma->vm_pgoff) << PAGE_SHIFT) + idx 4883 mm/hugetlb.c pgoff_t idx = ((addr - vma->vm_start) >> PAGE_SHIFT) + idx 4895 mm/hugetlb.c vma_interval_tree_foreach(svma, &mapping->i_mmap, idx, idx) { idx 4899 mm/hugetlb.c saddr = page_table_shareable(svma, vma, addr, idx); idx 61 mm/hugetlb_cgroup.c int idx; idx 63 mm/hugetlb_cgroup.c for (idx = 0; idx < hugetlb_max_hstate; idx++) { idx 64 mm/hugetlb_cgroup.c if (page_counter_read(&h_cg->hugepage[idx])) idx 73 mm/hugetlb_cgroup.c int idx; idx 75 mm/hugetlb_cgroup.c for (idx = 0; idx < HUGE_MAX_HSTATE; idx++) { idx 76 mm/hugetlb_cgroup.c struct page_counter *counter = &h_cgroup->hugepage[idx]; idx 82 mm/hugetlb_cgroup.c parent = &parent_h_cgroup->hugepage[idx]; idx 86 mm/hugetlb_cgroup.c 1 << huge_page_order(&hstates[idx])); idx 125 mm/hugetlb_cgroup.c static void hugetlb_cgroup_move_parent(int idx, struct hugetlb_cgroup *h_cg, idx 146 mm/hugetlb_cgroup.c page_counter_charge(&parent->hugepage[idx], nr_pages); idx 148 mm/hugetlb_cgroup.c counter = &h_cg->hugepage[idx]; idx 166 mm/hugetlb_cgroup.c int idx = 0; idx 172 mm/hugetlb_cgroup.c hugetlb_cgroup_move_parent(idx, h_cg, page); idx 175 mm/hugetlb_cgroup.c idx++; idx 181 mm/hugetlb_cgroup.c int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, idx 194 mm/hugetlb_cgroup.c if (huge_page_order(&hstates[idx]) < HUGETLB_CGROUP_MIN_ORDER) idx 205 mm/hugetlb_cgroup.c if (!page_counter_try_charge(&h_cg->hugepage[idx], nr_pages, &counter)) idx 214 mm/hugetlb_cgroup.c void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, idx 228 mm/hugetlb_cgroup.c void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, idx 240 mm/hugetlb_cgroup.c page_counter_uncharge(&h_cg->hugepage[idx], nr_pages); idx 244 mm/hugetlb_cgroup.c void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, idx 250 mm/hugetlb_cgroup.c if (huge_page_order(&hstates[idx]) < HUGETLB_CGROUP_MIN_ORDER) idx 253 mm/hugetlb_cgroup.c page_counter_uncharge(&h_cg->hugepage[idx], nr_pages); idx 291 mm/hugetlb_cgroup.c int ret, idx; idx 303 mm/hugetlb_cgroup.c idx = MEMFILE_IDX(of_cft(of)->private); idx 304 mm/hugetlb_cgroup.c nr_pages = round_down(nr_pages, 1 << huge_page_order(&hstates[idx])); idx 309 mm/hugetlb_cgroup.c ret = page_counter_set_max(&h_cg->hugepage[idx], nr_pages); idx 353 mm/hugetlb_cgroup.c static void __init __hugetlb_cgroup_file_init(int idx) idx 357 mm/hugetlb_cgroup.c struct hstate *h = &hstates[idx]; idx 365 mm/hugetlb_cgroup.c cft->private = MEMFILE_PRIVATE(idx, RES_LIMIT); idx 372 mm/hugetlb_cgroup.c cft->private = MEMFILE_PRIVATE(idx, RES_USAGE); idx 378 mm/hugetlb_cgroup.c cft->private = MEMFILE_PRIVATE(idx, RES_MAX_USAGE); idx 385 mm/hugetlb_cgroup.c cft->private = MEMFILE_PRIVATE(idx, RES_FAILCNT); idx 322 mm/kasan/common.c u8 idx = 0; idx 327 mm/kasan/common.c idx = alloc_meta->free_track_idx; idx 328 mm/kasan/common.c alloc_meta->free_pointer_tag[idx] = tag; idx 329 mm/kasan/common.c alloc_meta->free_track_idx = (idx + 1) % KASAN_NR_FREE_STACKS; idx 332 mm/kasan/common.c set_track(&alloc_meta->free_track[idx], GFP_NOWAIT); idx 46 mm/list_lru.c list_lru_from_memcg_idx(struct list_lru_node *nlru, int idx) idx 55 mm/list_lru.c if (memcg_lrus && idx >= 0) idx 56 mm/list_lru.c return memcg_lrus->lru[idx]; idx 110 mm/list_lru.c list_lru_from_memcg_idx(struct list_lru_node *nlru, int idx) idx 549 mm/memblock.c int idx, phys_addr_t base, idx 554 mm/memblock.c struct memblock_region *rgn = &type->regions[idx]; idx 557 mm/memblock.c memmove(rgn + 1, rgn, (type->cnt - idx) * sizeof(*rgn)); idx 589 mm/memblock.c int idx, nr_new; idx 614 mm/memblock.c for_each_memblock_type(idx, type, rgn) { idx 633 mm/memblock.c memblock_insert_region(type, idx++, base, idx 645 mm/memblock.c memblock_insert_region(type, idx, base, end - base, idx 728 mm/memblock.c int idx; idx 741 mm/memblock.c for_each_memblock_type(idx, type, rgn) { idx 758 mm/memblock.c memblock_insert_region(type, idx, rbase, base - rbase, idx 769 mm/memblock.c memblock_insert_region(type, idx--, rbase, end - rbase, idx 775 mm/memblock.c *start_rgn = idx; idx 776 mm/memblock.c *end_rgn = idx + 1; idx 941 mm/memblock.c void __init_memblock __next_reserved_mem_region(u64 *idx, idx 947 mm/memblock.c if (*idx < type->cnt) { idx 948 mm/memblock.c struct memblock_region *r = &type->regions[*idx]; idx 957 mm/memblock.c *idx += 1; idx 962 mm/memblock.c *idx = ULLONG_MAX; idx 1014 mm/memblock.c void __init_memblock __next_mem_range(u64 *idx, int nid, idx 1021 mm/memblock.c int idx_a = *idx & 0xffffffff; idx 1022 mm/memblock.c int idx_b = *idx >> 32; idx 1046 mm/memblock.c *idx = (u32)idx_a | (u64)idx_b << 32; idx 1084 mm/memblock.c *idx = (u32)idx_a | (u64)idx_b << 32; idx 1091 mm/memblock.c *idx = ULLONG_MAX; idx 1111 mm/memblock.c void __init_memblock __next_mem_range_rev(u64 *idx, int nid, idx 1118 mm/memblock.c int idx_a = *idx & 0xffffffff; idx 1119 mm/memblock.c int idx_b = *idx >> 32; idx 1124 mm/memblock.c if (*idx == (u64)ULLONG_MAX) { idx 1150 mm/memblock.c *idx = (u32)idx_a | (u64)idx_b << 32; idx 1183 mm/memblock.c *idx = (u32)idx_a | (u64)idx_b << 32; idx 1189 mm/memblock.c *idx = ULLONG_MAX; idx 1196 mm/memblock.c void __init_memblock __next_mem_pfn_range(int *idx, int nid, idx 1203 mm/memblock.c while (++*idx < type->cnt) { idx 1204 mm/memblock.c r = &type->regions[*idx]; idx 1211 mm/memblock.c if (*idx >= type->cnt) { idx 1212 mm/memblock.c *idx = -1; idx 1271 mm/memblock.c __next_mem_pfn_range_in_zone(u64 *idx, struct zone *zone, idx 1278 mm/memblock.c __next_mem_range(idx, zone_nid, MEMBLOCK_NONE, idx 1282 mm/memblock.c while (*idx != U64_MAX) { idx 1293 mm/memblock.c *idx = U64_MAX; idx 1305 mm/memblock.c __next_mem_range(idx, zone_nid, MEMBLOCK_NONE, idx 1622 mm/memblock.c int idx = memblock.memory.cnt - 1; idx 1624 mm/memblock.c return (memblock.memory.regions[idx].base + memblock.memory.regions[idx].size); idx 1778 mm/memblock.c int idx = memblock_search(&memblock.memory, base); idx 1781 mm/memblock.c if (idx == -1) idx 1783 mm/memblock.c return (memblock.memory.regions[idx].base + idx 1784 mm/memblock.c memblock.memory.regions[idx].size) >= end; idx 1843 mm/memblock.c int idx; idx 1848 mm/memblock.c for_each_memblock_type(idx, type, rgn) { idx 1861 mm/memblock.c type->name, idx, &base, &end, &size, nid_buf, flags); idx 691 mm/memcontrol.c void __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val) idx 698 mm/memcontrol.c x = val + __this_cpu_read(memcg->vmstats_percpu->stat[idx]); idx 706 mm/memcontrol.c __this_cpu_add(memcg->vmstats_local->stat[idx], x); idx 708 mm/memcontrol.c atomic_long_add(x, &mi->vmstats[idx]); idx 711 mm/memcontrol.c __this_cpu_write(memcg->vmstats_percpu->stat[idx], x); idx 735 mm/memcontrol.c void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, idx 744 mm/memcontrol.c __mod_node_page_state(pgdat, idx, val); idx 753 mm/memcontrol.c __mod_memcg_state(memcg, idx, val); idx 756 mm/memcontrol.c __this_cpu_add(pn->lruvec_stat_local->count[idx], val); idx 758 mm/memcontrol.c x = val + __this_cpu_read(pn->lruvec_stat_cpu->count[idx]); idx 763 mm/memcontrol.c atomic_long_add(x, &pi->lruvec_stat[idx]); idx 766 mm/memcontrol.c __this_cpu_write(pn->lruvec_stat_cpu->count[idx], x); idx 769 mm/memcontrol.c void __mod_lruvec_slab_state(void *p, enum node_stat_item idx, int val) idx 781 mm/memcontrol.c __mod_node_page_state(pgdat, idx, val); idx 784 mm/memcontrol.c __mod_lruvec_state(lruvec, idx, val); idx 789 mm/memcontrol.c void mod_memcg_obj_state(void *p, int idx, int val) idx 796 mm/memcontrol.c mod_memcg_state(memcg, idx, val); idx 806 mm/memcontrol.c void __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, idx 814 mm/memcontrol.c x = count + __this_cpu_read(memcg->vmstats_percpu->events[idx]); idx 822 mm/memcontrol.c __this_cpu_add(memcg->vmstats_local->events[idx], x); idx 824 mm/memcontrol.c atomic_long_add(x, &mi->vmevents[idx]); idx 827 mm/memcontrol.c __this_cpu_write(memcg->vmstats_percpu->events[idx], x); idx 4408 mm/memcontrol.c static unsigned long memcg_exact_page_state(struct mem_cgroup *memcg, int idx) idx 4410 mm/memcontrol.c long x = atomic_long_read(&memcg->vmstats[idx]); idx 4414 mm/memcontrol.c x += per_cpu_ptr(memcg->vmstats_percpu, cpu)->stat[idx]; idx 4480 mm/memory.c void (*process_subpage)(unsigned long addr, int idx, void *arg), idx 4539 mm/memory.c static void clear_subpage(unsigned long addr, int idx, void *arg) idx 4543 mm/memory.c clear_user_highpage(page + idx, addr); idx 4585 mm/memory.c static void copy_subpage(unsigned long addr, int idx, void *arg) idx 4589 mm/memory.c copy_user_highpage(copy_arg->dst + idx, copy_arg->src + idx, idx 2345 mm/mempolicy.c mpol_shared_policy_lookup(struct shared_policy *sp, unsigned long idx) idx 2353 mm/mempolicy.c sn = sp_lookup(sp, idx, idx+1); idx 103 mm/memtest.c unsigned int idx = 0; idx 110 mm/memtest.c idx = i % ARRAY_SIZE(patterns); idx 111 mm/memtest.c do_one_pass(patterns[idx], start, end); idx 6715 mm/page_alloc.c static void __meminit zone_init_internals(struct zone *zone, enum zone_type idx, int nid, idx 6720 mm/page_alloc.c zone->name = zone_names[idx]; idx 7712 mm/page_alloc.c enum zone_type j, idx; idx 7721 mm/page_alloc.c idx = j; idx 7722 mm/page_alloc.c while (idx) { idx 7725 mm/page_alloc.c idx--; idx 7726 mm/page_alloc.c lower_zone = pgdat->node_zones + idx; idx 7728 mm/page_alloc.c if (sysctl_lowmem_reserve_ratio[idx] < 1) { idx 7729 mm/page_alloc.c sysctl_lowmem_reserve_ratio[idx] = 0; idx 7733 mm/page_alloc.c managed_pages / sysctl_lowmem_reserve_ratio[idx]; idx 375 mm/slab.c unsigned int idx) idx 377 mm/slab.c return page->s_mem + cache->size * idx; idx 2320 mm/slab.c static inline freelist_idx_t get_free_obj(struct page *page, unsigned int idx) idx 2322 mm/slab.c return ((freelist_idx_t *)page->freelist)[idx]; idx 2326 mm/slab.c unsigned int idx, freelist_idx_t val) idx 2328 mm/slab.c ((freelist_idx_t *)(page->freelist))[idx] = val; idx 648 mm/slab_common.c int idx; idx 662 mm/slab_common.c idx = memcg_cache_id(memcg); idx 671 mm/slab_common.c if (arr->entries[idx]) idx 701 mm/slab_common.c arr->entries[idx] = s; idx 795 mm/slab_common.c int idx; idx 800 mm/slab_common.c idx = memcg_cache_id(memcg); idx 809 mm/slab_common.c c = arr->entries[idx]; idx 814 mm/slab_common.c arr->entries[idx] = NULL; idx 1229 mm/slab_common.c int idx = 0; idx 1233 mm/slab_common.c idx++; idx 1236 mm/slab_common.c return kasprintf(GFP_NOWAIT, "%s-%u%c", prefix, size, units[idx]); idx 1240 mm/slab_common.c new_kmalloc_cache(int idx, int type, slab_flags_t flags) idx 1247 mm/slab_common.c kmalloc_info[idx].size); idx 1250 mm/slab_common.c name = kmalloc_info[idx].name; idx 1253 mm/slab_common.c kmalloc_caches[type][idx] = create_kmalloc_cache(name, idx 1254 mm/slab_common.c kmalloc_info[idx].size, flags, 0, idx 1255 mm/slab_common.c kmalloc_info[idx].size); idx 1554 mm/slub.c unsigned int idx; idx 1561 mm/slub.c idx = s->random_seq[*pos]; idx 1565 mm/slub.c } while (unlikely(idx >= page_limit)); idx 1567 mm/slub.c return (char *)start + idx; idx 1576 mm/slub.c unsigned long idx, pos, page_limit, freelist_count; idx 1593 mm/slub.c for (idx = 1; idx < page->objects; idx++) { idx 1622 mm/slub.c int idx; idx 1673 mm/slub.c for (idx = 0, p = start; idx < page->objects - 1; idx++) { idx 225 mm/sparse.c int idx = subsection_map_index(pfn); idx 228 mm/sparse.c bitmap_set(map, idx, end - idx + 1); idx 43 mm/swap_cgroup.c unsigned long idx, max; idx 47 mm/swap_cgroup.c for (idx = 0; idx < ctrl->length; idx++) { idx 51 mm/swap_cgroup.c ctrl->map[idx] = page; idx 53 mm/swap_cgroup.c if (!(idx % SWAP_CLUSTER_MAX)) idx 58 mm/swap_cgroup.c max = idx; idx 59 mm/swap_cgroup.c for (idx = 0; idx < max; idx++) idx 60 mm/swap_cgroup.c __free_page(ctrl->map[idx]); idx 117 mm/swap_state.c pgoff_t idx = swp_offset(entry); idx 118 mm/swap_state.c XA_STATE_ORDER(xas, &address_space->i_pages, idx, compound_order(page)); idx 134 mm/swap_state.c VM_BUG_ON_PAGE(xas.xa_index != idx + i, page); idx 162 mm/swap_state.c pgoff_t idx = swp_offset(entry); idx 163 mm/swap_state.c XA_STATE(xas, &address_space->i_pages, idx); idx 402 mm/swapfile.c unsigned int idx) idx 405 mm/swapfile.c cluster_set_next_flag(&list->head, idx, 0); idx 406 mm/swapfile.c cluster_set_next_flag(&list->tail, idx, 0); idx 417 mm/swapfile.c cluster_set_next(ci_tail, idx); idx 419 mm/swapfile.c cluster_set_next_flag(&list->tail, idx, 0); idx 426 mm/swapfile.c unsigned int idx; idx 428 mm/swapfile.c idx = cluster_next(&list->head); idx 429 mm/swapfile.c if (cluster_next(&list->tail) == idx) { idx 434 mm/swapfile.c cluster_next(&ci[idx]), 0); idx 436 mm/swapfile.c return idx; idx 441 mm/swapfile.c unsigned int idx) idx 449 mm/swapfile.c memset(si->swap_map + idx * SWAPFILE_CLUSTER, idx 452 mm/swapfile.c cluster_list_add_tail(&si->discard_clusters, si->cluster_info, idx); idx 457 mm/swapfile.c static void __free_cluster(struct swap_info_struct *si, unsigned long idx) idx 461 mm/swapfile.c cluster_set_flag(ci + idx, CLUSTER_FLAG_FREE); idx 462 mm/swapfile.c cluster_list_add_tail(&si->free_clusters, ci, idx); idx 472 mm/swapfile.c unsigned int idx; idx 477 mm/swapfile.c idx = cluster_list_del_first(&si->discard_clusters, info); idx 480 mm/swapfile.c discard_swap_cluster(si, idx * SWAPFILE_CLUSTER, idx 484 mm/swapfile.c ci = lock_cluster(si, idx * SWAPFILE_CLUSTER); idx 485 mm/swapfile.c __free_cluster(si, idx); idx 486 mm/swapfile.c memset(si->swap_map + idx * SWAPFILE_CLUSTER, idx 503 mm/swapfile.c static void alloc_cluster(struct swap_info_struct *si, unsigned long idx) idx 507 mm/swapfile.c VM_BUG_ON(cluster_list_first(&si->free_clusters) != idx); idx 509 mm/swapfile.c cluster_set_count_flag(ci + idx, 0, 0); idx 512 mm/swapfile.c static void free_cluster(struct swap_info_struct *si, unsigned long idx) idx 514 mm/swapfile.c struct swap_cluster_info *ci = si->cluster_info + idx; idx 524 mm/swapfile.c swap_cluster_schedule_discard(si, idx); idx 528 mm/swapfile.c __free_cluster(si, idx); idx 538 mm/swapfile.c unsigned long idx = page_nr / SWAPFILE_CLUSTER; idx 542 mm/swapfile.c if (cluster_is_free(&cluster_info[idx])) idx 543 mm/swapfile.c alloc_cluster(p, idx); idx 545 mm/swapfile.c VM_BUG_ON(cluster_count(&cluster_info[idx]) >= SWAPFILE_CLUSTER); idx 546 mm/swapfile.c cluster_set_count(&cluster_info[idx], idx 547 mm/swapfile.c cluster_count(&cluster_info[idx]) + 1); idx 558 mm/swapfile.c unsigned long idx = page_nr / SWAPFILE_CLUSTER; idx 563 mm/swapfile.c VM_BUG_ON(cluster_count(&cluster_info[idx]) == 0); idx 564 mm/swapfile.c cluster_set_count(&cluster_info[idx], idx 565 mm/swapfile.c cluster_count(&cluster_info[idx]) - 1); idx 567 mm/swapfile.c if (cluster_count(&cluster_info[idx]) == 0) idx 568 mm/swapfile.c free_cluster(p, idx); idx 931 mm/swapfile.c unsigned long idx; idx 948 mm/swapfile.c idx = cluster_list_first(&si->free_clusters); idx 949 mm/swapfile.c offset = idx * SWAPFILE_CLUSTER; idx 951 mm/swapfile.c alloc_cluster(si, idx); idx 964 mm/swapfile.c static void swap_free_cluster(struct swap_info_struct *si, unsigned long idx) idx 966 mm/swapfile.c unsigned long offset = idx * SWAPFILE_CLUSTER; idx 972 mm/swapfile.c free_cluster(si, idx); idx 1328 mm/swapfile.c unsigned long idx = offset / SWAPFILE_CLUSTER; idx 1355 mm/swapfile.c swap_free_cluster(si, idx); idx 3015 mm/swapfile.c unsigned long i, idx; idx 3071 mm/swapfile.c idx = i * SWAP_CLUSTER_COLS + j; idx 3072 mm/swapfile.c if (idx >= nr_clusters) idx 3074 mm/swapfile.c if (cluster_count(&cluster_info[idx])) idx 3076 mm/swapfile.c cluster_set_flag(&cluster_info[idx], CLUSTER_FLAG_FREE); idx 3078 mm/swapfile.c idx); idx 189 mm/userfaultfd.c pgoff_t idx; idx 270 mm/userfaultfd.c idx = linear_page_index(dst_vma, dst_addr); idx 272 mm/userfaultfd.c hash = hugetlb_fault_mutex_hash(h, mapping, idx, dst_addr); idx 64 mm/vmacache.c int idx = VMACACHE_HASH(addr); idx 73 mm/vmacache.c struct vm_area_struct *vma = current->vmacache.vmas[idx]; idx 85 mm/vmacache.c if (++idx == VMACACHE_SIZE) idx 86 mm/vmacache.c idx = 0; idx 97 mm/vmacache.c int idx = VMACACHE_HASH(start); idx 106 mm/vmacache.c struct vm_area_struct *vma = current->vmacache.vmas[idx]; idx 112 mm/vmacache.c if (++idx == VMACACHE_SIZE) idx 113 mm/vmacache.c idx = 0; idx 377 mm/z3fold.c int idx = 0; idx 387 mm/z3fold.c idx = __idx(zhdr, bud); idx 388 mm/z3fold.c h += idx; idx 392 mm/z3fold.c slots->slot[idx] = h; idx 393 mm/z3fold.c return (unsigned long)&slots->slot[idx]; idx 541 mm/zsmalloc.c int idx = 0; idx 544 mm/zsmalloc.c idx = DIV_ROUND_UP(size - ZS_MIN_ALLOC_SIZE, idx 547 mm/zsmalloc.c return min_t(int, ZS_SIZE_CLASSES - 1, idx); idx 1897 mm/zsmalloc.c int idx = 0; idx 1902 mm/zsmalloc.c pages[idx] = newpage; idx 1904 mm/zsmalloc.c pages[idx] = page; idx 1905 mm/zsmalloc.c idx++; idx 2052 net/batman-adv/bat_iv_ogm.c int idx = 0; idx 2056 net/batman-adv/bat_iv_ogm.c if (idx++ < *idx_s) idx 2063 net/batman-adv/bat_iv_ogm.c *idx_s = idx - 1; idx 2089 net/batman-adv/bat_iv_ogm.c int idx = cb->args[1]; idx 2099 net/batman-adv/bat_iv_ogm.c &idx, &sub)) idx 2106 net/batman-adv/bat_iv_ogm.c cb->args[1] = idx; idx 2268 net/batman-adv/bat_iv_ogm.c int idx = 0; idx 2272 net/batman-adv/bat_iv_ogm.c if (idx++ < *idx_s) idx 2277 net/batman-adv/bat_iv_ogm.c *idx_s = idx - 1; idx 2301 net/batman-adv/bat_iv_ogm.c int idx = cb->args[1]; idx 2311 net/batman-adv/bat_iv_ogm.c &idx) == 0) idx 2326 net/batman-adv/bat_iv_ogm.c hard_iface, &idx)) { idx 2335 net/batman-adv/bat_iv_ogm.c cb->args[1] = idx; idx 2713 net/batman-adv/bat_iv_ogm.c int idx = 0; idx 2719 net/batman-adv/bat_iv_ogm.c if (idx++ < idx_skip) idx 2724 net/batman-adv/bat_iv_ogm.c idx_skip = idx - 1; idx 2729 net/batman-adv/bat_iv_ogm.c idx_skip = idx; idx 272 net/batman-adv/bat_v.c int idx = 0; idx 276 net/batman-adv/bat_v.c if (idx++ < *idx_s) idx 280 net/batman-adv/bat_v.c *idx_s = idx - 1; idx 305 net/batman-adv/bat_v.c int idx = cb->args[1]; idx 314 net/batman-adv/bat_v.c &idx) == 0) idx 328 net/batman-adv/bat_v.c &idx)) { idx 337 net/batman-adv/bat_v.c cb->args[1] = idx; idx 548 net/batman-adv/bat_v.c int idx = 0; idx 552 net/batman-adv/bat_v.c if (idx++ < *idx_s) idx 558 net/batman-adv/bat_v.c *idx_s = idx - 1; idx 584 net/batman-adv/bat_v.c int idx = cb->args[1]; idx 593 net/batman-adv/bat_v.c bat_priv, if_outgoing, head, &idx, idx 601 net/batman-adv/bat_v.c cb->args[1] = idx; idx 1013 net/batman-adv/bat_v.c int idx = 0; idx 1019 net/batman-adv/bat_v.c if (idx++ < idx_skip) idx 1024 net/batman-adv/bat_v.c idx_skip = idx - 1; idx 1029 net/batman-adv/bat_v.c idx_skip = idx; idx 2173 net/batman-adv/bridge_loop_avoidance.c int idx = 0; idx 2180 net/batman-adv/bridge_loop_avoidance.c if (idx++ < *idx_skip) idx 2186 net/batman-adv/bridge_loop_avoidance.c *idx_skip = idx - 1; idx 2213 net/batman-adv/bridge_loop_avoidance.c int idx = cb->args[1]; idx 2239 net/batman-adv/bridge_loop_avoidance.c hash, bucket, &idx)) idx 2245 net/batman-adv/bridge_loop_avoidance.c cb->args[1] = idx; idx 2411 net/batman-adv/bridge_loop_avoidance.c int idx = 0; idx 2418 net/batman-adv/bridge_loop_avoidance.c if (idx++ < *idx_skip) idx 2424 net/batman-adv/bridge_loop_avoidance.c *idx_skip = idx - 1; idx 2451 net/batman-adv/bridge_loop_avoidance.c int idx = cb->args[1]; idx 2477 net/batman-adv/bridge_loop_avoidance.c hash, bucket, &idx)) idx 2483 net/batman-adv/bridge_loop_avoidance.c cb->args[1] = idx; idx 958 net/batman-adv/distributed-arp-table.c int idx = 0; idx 964 net/batman-adv/distributed-arp-table.c if (idx < *idx_skip) idx 969 net/batman-adv/distributed-arp-table.c *idx_skip = idx; idx 975 net/batman-adv/distributed-arp-table.c idx++; idx 998 net/batman-adv/distributed-arp-table.c int idx = cb->args[1]; idx 1024 net/batman-adv/distributed-arp-table.c &idx)) idx 1028 net/batman-adv/distributed-arp-table.c idx = 0; idx 1032 net/batman-adv/distributed-arp-table.c cb->args[1] = idx; idx 41 net/batman-adv/log.c size_t idx) idx 43 net/batman-adv/log.c return &debug_log->log_buff[idx & BATADV_LOG_BUFF_MASK]; idx 454 net/batman-adv/main.c u8 idx; idx 504 net/batman-adv/main.c idx = batadv_ogm_packet->packet_type; idx 505 net/batman-adv/main.c (*batadv_rx_handler[idx])(skb, hard_iface); idx 359 net/batman-adv/main.h static inline void batadv_add_counter(struct batadv_priv *bat_priv, size_t idx, idx 362 net/batman-adv/main.h this_cpu_add(bat_priv->bat_counters[idx], count); idx 2256 net/batman-adv/multicast.c long idx = 0; idx 2266 net/batman-adv/multicast.c if (idx < *idx_skip) idx 2271 net/batman-adv/multicast.c *idx_skip = idx; idx 2277 net/batman-adv/multicast.c idx++; idx 2298 net/batman-adv/multicast.c struct batadv_priv *bat_priv, long *bucket, long *idx) idx 2302 net/batman-adv/multicast.c long idx_tmp = *idx; idx 2314 net/batman-adv/multicast.c *idx = idx_tmp; idx 2381 net/batman-adv/multicast.c long *idx = &cb->args[1]; idx 2389 net/batman-adv/multicast.c ret = __batadv_mcast_flags_dump(msg, portid, cb, bat_priv, bucket, idx); idx 1272 net/batman-adv/network-coding.c int idx; idx 1280 net/batman-adv/network-coding.c idx = batadv_nc_hash_choose(&nc_path_key, hash->size); idx 1284 net/batman-adv/network-coding.c hlist_for_each_entry_rcu(nc_path, &hash->table[idx], hash_entry) { idx 102 net/batman-adv/soft-interface.c static u64 batadv_sum_counter(struct batadv_priv *bat_priv, size_t idx) idx 109 net/batman-adv/soft-interface.c sum += counters[idx]; idx 1222 net/batman-adv/translation-table.c int idx = 0; idx 1228 net/batman-adv/translation-table.c if (idx++ < *idx_s) idx 1234 net/batman-adv/translation-table.c *idx_s = idx - 1; idx 1261 net/batman-adv/translation-table.c int idx = cb->args[1]; idx 1286 net/batman-adv/translation-table.c hash, bucket, &idx)) idx 1301 net/batman-adv/translation-table.c cb->args[1] = idx; idx 2139 net/batman-adv/translation-table.c int idx = 0; idx 2143 net/batman-adv/translation-table.c if (idx++ < *idx_s) idx 2149 net/batman-adv/translation-table.c *idx_s = idx - 1; idx 2178 net/batman-adv/translation-table.c int idx = cb->args[1]; idx 2207 net/batman-adv/translation-table.c head, &idx, &sub)) idx 2222 net/batman-adv/translation-table.c cb->args[1] = idx; idx 723 net/bridge/br_fdb.c int *idx) idx 733 net/bridge/br_fdb.c err = ndo_dflt_fdb_dump(skb, cb, dev, NULL, idx); idx 740 net/bridge/br_fdb.c if (*idx < cb->args[2]) idx 764 net/bridge/br_fdb.c *idx += 1; idx 130 net/bridge/br_mdb.c int idx = 0, s_idx = cb->args[1], err = 0; idx 146 net/bridge/br_mdb.c if (idx < s_idx) idx 176 net/bridge/br_mdb.c idx++; idx 180 net/bridge/br_mdb.c cb->args[1] = idx; idx 213 net/bridge/br_mdb.c int idx = 0, s_idx; idx 232 net/bridge/br_mdb.c if (idx < s_idx) idx 252 net/bridge/br_mdb.c idx++; idx 260 net/bridge/br_mdb.c cb->args[0] = idx; idx 577 net/bridge/br_private.h struct net_device *dev, struct net_device *fdev, int *idx); idx 108 net/caif/cfmuxl.c int idx = phyid % DN_CACHE_SIZE; idx 111 net/caif/cfmuxl.c RCU_INIT_POINTER(muxl->dn_cache[idx], NULL); idx 126 net/caif/cfmuxl.c int idx = id % UP_CACHE_SIZE; idx 127 net/caif/cfmuxl.c up = rcu_dereference(muxl->up_cache[idx]); idx 131 net/caif/cfmuxl.c rcu_assign_pointer(muxl->up_cache[idx], up); idx 140 net/caif/cfmuxl.c int idx = dev_info->id % DN_CACHE_SIZE; idx 141 net/caif/cfmuxl.c dn = rcu_dereference(muxl->dn_cache[idx]); idx 145 net/caif/cfmuxl.c rcu_assign_pointer(muxl->dn_cache[idx], dn); idx 155 net/caif/cfmuxl.c int idx = id % UP_CACHE_SIZE; idx 167 net/caif/cfmuxl.c RCU_INIT_POINTER(muxl->up_cache[idx], NULL); idx 251 net/can/gw.c static inline int calc_idx(int idx, int rx_len) idx 253 net/can/gw.c if (idx < 0) idx 254 net/can/gw.c return rx_len + idx; idx 256 net/can/gw.c return idx; idx 691 net/can/gw.c int idx = 0; idx 696 net/can/gw.c if (idx < s_idx) idx 704 net/can/gw.c idx++; idx 708 net/can/gw.c cb->args[0] = idx; idx 302 net/can/proc.c static inline void can_rcvlist_proc_show_one(struct seq_file *m, int idx, idx 306 net/can/proc.c if (!hlist_empty(&dev_rcv_lists->rx[idx])) { idx 308 net/can/proc.c can_print_rcvlist(m, &dev_rcv_lists->rx[idx], dev); idx 317 net/can/proc.c int idx = (int)(long)PDE_DATA(m->file->f_inode); idx 322 net/can/proc.c seq_printf(m, "\nreceive list '%s':\n", rx_list_name[idx]); idx 328 net/can/proc.c can_rcvlist_proc_show_one(m, idx, NULL, dev_rcv_lists); idx 333 net/can/proc.c can_rcvlist_proc_show_one(m, idx, dev, dev->ml_priv); idx 655 net/core/devlink.c int idx = 0; idx 662 net/core/devlink.c if (idx < start) { idx 663 net/core/devlink.c idx++; idx 671 net/core/devlink.c idx++; idx 676 net/core/devlink.c cb->args[0] = idx; idx 709 net/core/devlink.c int idx = 0; idx 718 net/core/devlink.c if (idx < start) { idx 719 net/core/devlink.c idx++; idx 731 net/core/devlink.c idx++; idx 738 net/core/devlink.c cb->args[0] = idx; idx 897 net/core/devlink.c int idx = 0; idx 906 net/core/devlink.c if (idx < start) { idx 907 net/core/devlink.c idx++; idx 919 net/core/devlink.c idx++; idx 926 net/core/devlink.c cb->args[0] = idx; idx 1037 net/core/devlink.c int idx = 0; idx 1047 net/core/devlink.c err = __sb_pool_get_dumpit(msg, start, &idx, devlink, idx 1061 net/core/devlink.c cb->args[0] = idx; idx 1235 net/core/devlink.c int idx = 0; idx 1245 net/core/devlink.c err = __sb_port_pool_get_dumpit(msg, start, &idx, idx 1259 net/core/devlink.c cb->args[0] = idx; idx 1462 net/core/devlink.c int idx = 0; idx 1473 net/core/devlink.c err = __sb_tc_pool_bind_get_dumpit(msg, start, &idx, idx 1488 net/core/devlink.c cb->args[0] = idx; idx 3157 net/core/devlink.c int idx = 0; idx 3166 net/core/devlink.c if (idx < start) { idx 3167 net/core/devlink.c idx++; idx 3179 net/core/devlink.c idx++; idx 3186 net/core/devlink.c cb->args[0] = idx; idx 3420 net/core/devlink.c int idx = 0; idx 3431 net/core/devlink.c if (idx < start) { idx 3432 net/core/devlink.c idx++; idx 3446 net/core/devlink.c idx++; idx 3454 net/core/devlink.c cb->args[0] = idx; idx 3686 net/core/devlink.c int idx = 0; idx 3696 net/core/devlink.c if (idx < start) { idx 3697 net/core/devlink.c idx++; idx 3709 net/core/devlink.c idx++; idx 3715 net/core/devlink.c cb->args[0] = idx; idx 4086 net/core/devlink.c int idx = 0; idx 4093 net/core/devlink.c if (idx < start) { idx 4094 net/core/devlink.c idx++; idx 4099 net/core/devlink.c idx++; idx 4111 net/core/devlink.c idx++; idx 4115 net/core/devlink.c cb->args[0] = idx; idx 5031 net/core/devlink.c int idx = 0; idx 5041 net/core/devlink.c if (idx < start) { idx 5042 net/core/devlink.c idx++; idx 5055 net/core/devlink.c idx++; idx 5062 net/core/devlink.c cb->args[0] = idx; idx 5476 net/core/devlink.c int idx = 0; idx 5485 net/core/devlink.c if (idx < start) { idx 5486 net/core/devlink.c idx++; idx 5498 net/core/devlink.c idx++; idx 5505 net/core/devlink.c cb->args[0] = idx; idx 5682 net/core/devlink.c int idx = 0; idx 5692 net/core/devlink.c if (idx < start) { idx 5693 net/core/devlink.c idx++; idx 5705 net/core/devlink.c idx++; idx 5712 net/core/devlink.c cb->args[0] = idx; idx 392 net/core/ethtool.c unsigned int idx = 0; idx 394 net/core/ethtool.c for (; idx < size; idx++) { idx 395 net/core/ethtool.c dst->link_modes.supported[idx] &= idx 396 net/core/ethtool.c src->link_modes.supported[idx]; idx 397 net/core/ethtool.c dst->link_modes.advertising[idx] &= idx 398 net/core/ethtool.c src->link_modes.advertising[idx]; idx 820 net/core/ethtool.c int i, idx = 0, n_bits = 0, ret, rc; idx 852 net/core/ethtool.c info_buf[idx++] = rc; idx 861 net/core/ethtool.c if (copy_to_user(useraddr, info_buf, idx * sizeof(u32))) idx 1041 net/core/fib_rules.c int idx = 0; idx 1047 net/core/fib_rules.c if (idx < cb->args[1]) idx 1056 net/core/fib_rules.c idx++; idx 1059 net/core/fib_rules.c cb->args[1] = idx; idx 1096 net/core/fib_rules.c int idx = 0, family; idx 1119 net/core/fib_rules.c if (idx < cb->args[0] || !try_module_get(ops->owner)) idx 1127 net/core/fib_rules.c idx++; idx 1130 net/core/fib_rules.c cb->args[0] = idx; idx 4056 net/core/filter.c u32, idx) idx 4065 net/core/filter.c if (unlikely(idx >= array->map.max_entries)) idx 4068 net/core/filter.c cgrp = READ_ONCE(array->ptrs[idx]); idx 2552 net/core/neighbour.c int idx, s_idx = idx = cb->args[2]; idx 2565 net/core/neighbour.c for (n = rcu_dereference_bh(nht->hash_buckets[h]), idx = 0; idx 2568 net/core/neighbour.c if (idx < s_idx || !net_eq(dev_net(n->dev), net)) idx 2581 net/core/neighbour.c idx++; idx 2588 net/core/neighbour.c cb->args[2] = idx; idx 2599 net/core/neighbour.c int idx, s_idx = idx = cb->args[4]; idx 2610 net/core/neighbour.c for (n = tbl->phash_buckets[h], idx = 0; n; n = n->next) { idx 2611 net/core/neighbour.c if (idx < s_idx || pneigh_net(n) != net) idx 2624 net/core/neighbour.c idx++; idx 2632 net/core/neighbour.c cb->args[4] = idx; idx 949 net/core/net_namespace.c int idx; idx 958 net/core/net_namespace.c if (net_cb->idx < net_cb->s_idx) idx 969 net/core/net_namespace.c net_cb->idx++; idx 1026 net/core/net_namespace.c .idx = 0, idx 1051 net/core/net_namespace.c cb->args[0] = net_cb.idx; idx 204 net/core/netpoll.c int idx; idx 206 net/core/netpoll.c idx = srcu_read_lock(&netpoll_srcu); idx 210 net/core/netpoll.c srcu_read_unlock(&netpoll_srcu, idx); idx 1934 net/core/rtnetlink.c int idx = 0, s_idx; idx 1992 net/core/rtnetlink.c idx = 0; idx 1997 net/core/rtnetlink.c if (idx < s_idx) idx 2013 net/core/rtnetlink.c idx++; idx 2019 net/core/rtnetlink.c cb->args[1] = idx; idx 3444 net/core/rtnetlink.c int idx; idx 3452 net/core/rtnetlink.c for (idx = 1; idx <= RTNL_FAMILY_MAX; idx++) { idx 3457 net/core/rtnetlink.c if (idx < s_idx || idx == PF_PACKET) idx 3463 net/core/rtnetlink.c tab = rcu_dereference_rtnl(rtnl_msg_handlers[idx]); idx 3475 net/core/rtnetlink.c if (idx > s_idx) { idx 3484 net/core/rtnetlink.c cb->family = idx; idx 3875 net/core/rtnetlink.c int *idx, idx 3886 net/core/rtnetlink.c if (*idx < cb->args[2]) idx 3896 net/core/rtnetlink.c *idx += 1; idx 3916 net/core/rtnetlink.c int *idx) idx 3924 net/core/rtnetlink.c err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->uc); idx 3927 net/core/rtnetlink.c err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->mc); idx 4034 net/core/rtnetlink.c int idx = 0, s_idx; idx 4059 net/core/rtnetlink.c idx = 0; idx 4082 net/core/rtnetlink.c if (idx < s_idx) idx 4111 net/core/rtnetlink.c idx++; idx 4117 net/core/rtnetlink.c cb->args[1] = idx; idx 4448 net/core/rtnetlink.c int idx = 0; idx 4465 net/core/rtnetlink.c if (idx >= cb->args[0]) { idx 4476 net/core/rtnetlink.c idx++; idx 4480 net/core/rtnetlink.c if (idx >= cb->args[0]) { idx 4492 net/core/rtnetlink.c idx++; idx 4498 net/core/rtnetlink.c cb->args[0] = idx; idx 5099 net/core/rtnetlink.c int idx = 0; idx 5120 net/core/rtnetlink.c idx = 0; idx 5123 net/core/rtnetlink.c if (idx < s_idx) idx 5141 net/core/rtnetlink.c idx++; idx 5147 net/core/rtnetlink.c cb->args[1] = idx; idx 3240 net/core/sock.c int cpu, idx = prot->inuse_idx; idx 3244 net/core/sock.c res += per_cpu_ptr(net->core.prot_inuse, cpu)->val[idx]; idx 339 net/core/sock_map.c static int sock_map_update_common(struct bpf_map *map, u32 idx, idx 352 net/core/sock_map.c if (unlikely(idx >= map->max_entries)) idx 369 net/core/sock_map.c osk = stab->sks[idx]; idx 378 net/core/sock_map.c sock_map_add_link(psock, link, map, &stab->sks[idx]); idx 379 net/core/sock_map.c stab->sks[idx] = sk; idx 381 net/core/sock_map.c sock_map_unref(osk, &stab->sks[idx]); idx 409 net/core/sock_map.c u32 idx = *(u32 *)key; idx 431 net/core/sock_map.c ret = sock_map_update_common(map, idx, sk, flags); idx 189 net/dccp/feat.c int idx = dccp_feat_index(feat_num); idx 191 net/dccp/feat.c if (idx < 0) idx 193 net/dccp/feat.c return dccp_feat_table[idx].reconciliation; idx 198 net/dccp/feat.c int idx = dccp_feat_index(feat_num); idx 203 net/dccp/feat.c DCCP_BUG_ON(idx < 0); idx 205 net/dccp/feat.c return idx < 0 ? 0 : dccp_feat_table[idx].default_value; idx 303 net/dccp/feat.c static int __dccp_feat_activate(struct sock *sk, const int idx, idx 309 net/dccp/feat.c if (idx < 0 || idx >= DCCP_FEAT_SUPPORTED_MAX) idx 311 net/dccp/feat.c if (dccp_feat_table[idx].activation_hdlr == NULL) idx 315 net/dccp/feat.c val = dccp_feat_table[idx].default_value; idx 316 net/dccp/feat.c } else if (dccp_feat_table[idx].reconciliation == FEAT_SP) { idx 323 net/dccp/feat.c DCCP_CRIT("Feature #%d undefined: using default", idx); idx 324 net/dccp/feat.c val = dccp_feat_table[idx].default_value; idx 333 net/dccp/feat.c rx = (is_local == (dccp_feat_table[idx].rxtx == FEAT_AT_RX)); idx 336 net/dccp/feat.c dccp_feat_fname(dccp_feat_table[idx].feat_num), idx 339 net/dccp/feat.c return dccp_feat_table[idx].activation_hdlr(sk, val, rx); idx 1501 net/dccp/feat.c int idx; idx 1515 net/dccp/feat.c idx = dccp_feat_index(cur->feat_num); idx 1516 net/dccp/feat.c if (idx < 0) { idx 1527 net/dccp/feat.c fvals[idx][cur->is_local] = &cur->val; idx 1536 net/dccp/feat.c for (idx = DCCP_FEAT_SUPPORTED_MAX; --idx >= 0;) idx 1537 net/dccp/feat.c if (__dccp_feat_activate(sk, idx, 0, fvals[idx][0]) || idx 1538 net/dccp/feat.c __dccp_feat_activate(sk, idx, 1, fvals[idx][1])) { idx 1539 net/dccp/feat.c DCCP_CRIT("Could not activate %d", idx); idx 744 net/decnet/dn_dev.c int idx, dn_idx = 0, skip_ndevs, skip_naddr; idx 755 net/decnet/dn_dev.c idx = 0; idx 758 net/decnet/dn_dev.c if (idx < skip_ndevs) idx 760 net/decnet/dn_dev.c else if (idx > skip_ndevs) { idx 780 net/decnet/dn_dev.c idx++; idx 784 net/decnet/dn_dev.c cb->args[0] = idx; idx 1725 net/decnet/dn_route.c int idx, s_idx; idx 1739 net/decnet/dn_route.c s_idx = idx = cb->args[1]; idx 1746 net/decnet/dn_route.c for(rt = rcu_dereference_bh(dn_rt_hash_table[h].chain), idx = 0; idx 1748 net/decnet/dn_route.c rt = rcu_dereference_bh(rt->dn_next), idx++) { idx 1749 net/decnet/dn_route.c if (idx < s_idx) idx 1766 net/decnet/dn_route.c cb->args[1] = idx; idx 187 net/dsa/slave.c int idx; idx 200 net/dsa/slave.c if (dump->idx < dump->cb->args[2]) idx 226 net/dsa/slave.c dump->idx++; idx 237 net/dsa/slave.c int *idx) idx 244 net/dsa/slave.c .idx = *idx, idx 249 net/dsa/slave.c *idx = dump.idx; idx 442 net/ieee802154/nl-mac.c int idx; idx 447 net/ieee802154/nl-mac.c idx = 0; idx 449 net/ieee802154/nl-mac.c if (idx < s_idx || dev->type != ARPHRD_IEEE802154) idx 457 net/ieee802154/nl-mac.c idx++; idx 459 net/ieee802154/nl-mac.c cb->args[0] = idx; idx 775 net/ieee802154/nl-mac.c int idx = 0; idx 780 net/ieee802154/nl-mac.c if (idx < first_dev || dev->type != ARPHRD_IEEE802154) idx 803 net/ieee802154/nl-mac.c idx++; idx 805 net/ieee802154/nl-mac.c cb->args[0] = idx; idx 951 net/ieee802154/nl-mac.c int rc = 0, idx = 0; idx 954 net/ieee802154/nl-mac.c if (idx++ < data->s_idx) idx 1082 net/ieee802154/nl-mac.c int rc = 0, idx = 0; idx 1085 net/ieee802154/nl-mac.c if (idx++ < data->s_idx) idx 1186 net/ieee802154/nl-mac.c int rc = 0, idx = 0, idx2; idx 1189 net/ieee802154/nl-mac.c if (idx++ < data->s_idx) idx 1318 net/ieee802154/nl-mac.c int rc = 0, idx = 0; idx 1321 net/ieee802154/nl-mac.c if (idx++ < data->s_idx) idx 116 net/ieee802154/nl-phy.c int idx, s_idx; idx 126 net/ieee802154/nl-phy.c if (data->idx++ < data->s_idx) idx 136 net/ieee802154/nl-phy.c data->idx--; idx 149 net/ieee802154/nl-phy.c .idx = 0, idx 156 net/ieee802154/nl-phy.c cb->args[0] = data.idx; idx 597 net/ieee802154/nl802154.c int idx = 0, ret; idx 621 net/ieee802154/nl802154.c if (++idx <= state->start) idx 639 net/ieee802154/nl802154.c idx--; idx 646 net/ieee802154/nl802154.c state->start = idx; idx 1814 net/ipv4/devinet.c int idx, s_idx; idx 1822 net/ipv4/devinet.c s_idx = idx = cb->args[1]; idx 1849 net/ipv4/devinet.c idx = 0; idx 1855 net/ipv4/devinet.c if (idx < s_idx) idx 1857 net/ipv4/devinet.c if (h > s_h || idx > s_idx) idx 1870 net/ipv4/devinet.c idx++; idx 1877 net/ipv4/devinet.c cb->args[1] = idx; idx 2225 net/ipv4/devinet.c int idx, s_idx; idx 2246 net/ipv4/devinet.c s_idx = idx = cb->args[1]; idx 2249 net/ipv4/devinet.c idx = 0; idx 2255 net/ipv4/devinet.c if (idx < s_idx) idx 2273 net/ipv4/devinet.c idx++; idx 2301 net/ipv4/devinet.c cb->args[1] = idx; idx 2489 net/ipv4/fib_trie.c loff_t idx = 0; idx 2502 net/ipv4/fib_trie.c if (pos == idx++) { idx 896 net/ipv4/fou.c int idx = 0, ret; idx 900 net/ipv4/fou.c if (idx++ < cb->args[0]) idx 910 net/ipv4/fou.c cb->args[0] = idx; idx 965 net/ipv4/inet_diag.c int idx, accum, res; idx 1014 net/ipv4/inet_diag.c for (idx = 0; idx < accum; idx++) { idx 1016 net/ipv4/inet_diag.c res = sk_diag_fill(sk_arr[idx], skb, r, idx 1022 net/ipv4/inet_diag.c num = num_arr[idx]; idx 1024 net/ipv4/inet_diag.c sock_gen_put(sk_arr[idx]); idx 347 net/ipv4/ip_input.c u32 idx = skb_dst(skb)->tclassid; idx 348 net/ipv4/ip_input.c st[idx&0xFF].o_packets++; idx 349 net/ipv4/ip_input.c st[idx&0xFF].o_bytes += skb->len; idx 350 net/ipv4/ip_input.c st[(idx>>16)&0xFF].i_packets++; idx 351 net/ipv4/ip_input.c st[(idx>>16)&0xFF].i_bytes += skb->len; idx 682 net/ipv4/netfilter/ipt_CLUSTERIP.c struct clusterip_seq_position *idx; idx 690 net/ipv4/netfilter/ipt_CLUSTERIP.c idx = kmalloc(sizeof(struct clusterip_seq_position), GFP_KERNEL); idx 691 net/ipv4/netfilter/ipt_CLUSTERIP.c if (!idx) idx 694 net/ipv4/netfilter/ipt_CLUSTERIP.c idx->pos = *pos; idx 695 net/ipv4/netfilter/ipt_CLUSTERIP.c idx->weight = weight; idx 696 net/ipv4/netfilter/ipt_CLUSTERIP.c idx->bit = ffs(local_nodes); idx 697 net/ipv4/netfilter/ipt_CLUSTERIP.c idx->val = local_nodes; idx 698 net/ipv4/netfilter/ipt_CLUSTERIP.c clear_bit(idx->bit - 1, &idx->val); idx 700 net/ipv4/netfilter/ipt_CLUSTERIP.c return idx; idx 705 net/ipv4/netfilter/ipt_CLUSTERIP.c struct clusterip_seq_position *idx = v; idx 707 net/ipv4/netfilter/ipt_CLUSTERIP.c *pos = ++idx->pos; idx 708 net/ipv4/netfilter/ipt_CLUSTERIP.c if (*pos >= idx->weight) { idx 712 net/ipv4/netfilter/ipt_CLUSTERIP.c idx->bit = ffs(idx->val); idx 713 net/ipv4/netfilter/ipt_CLUSTERIP.c clear_bit(idx->bit - 1, &idx->val); idx 714 net/ipv4/netfilter/ipt_CLUSTERIP.c return idx; idx 725 net/ipv4/netfilter/ipt_CLUSTERIP.c struct clusterip_seq_position *idx = v; idx 727 net/ipv4/netfilter/ipt_CLUSTERIP.c if (idx->pos != 0) idx 730 net/ipv4/netfilter/ipt_CLUSTERIP.c seq_printf(s, "%u", idx->bit); idx 732 net/ipv4/netfilter/ipt_CLUSTERIP.c if (idx->pos == idx->weight - 1) idx 425 net/ipv4/netfilter/nf_nat_h323.c TransportAddress *taddr, int idx, idx 462 net/ipv4/netfilter/nf_nat_h323.c if (set_h225_addr(skb, protoff, data, 0, &taddr[idx], idx 474 net/ipv4/netfilter/nf_nat_h323.c if (idx > 0 && idx 1667 net/ipv4/nexthop.c u32 idx; idx 1680 net/ipv4/nexthop.c idx = nla_get_u32(tb[i]); idx 1681 net/ipv4/nexthop.c if (idx > INT_MAX) { idx 1685 net/ipv4/nexthop.c *dev_idx = idx; idx 1688 net/ipv4/nexthop.c idx = nla_get_u32(tb[i]); idx 1689 net/ipv4/nexthop.c if (idx > INT_MAX) { idx 1693 net/ipv4/nexthop.c *master_idx = idx; idx 1722 net/ipv4/nexthop.c int idx = 0, s_idx; idx 1734 net/ipv4/nexthop.c if (idx < s_idx) idx 1752 net/ipv4/nexthop.c idx++; idx 1758 net/ipv4/nexthop.c cb->args[0] = idx; idx 60 net/ipv4/tcp_metrics.c enum tcp_metric_index idx) idx 62 net/ipv4/tcp_metrics.c return tm->tcpm_lock & (1 << idx); idx 66 net/ipv4/tcp_metrics.c enum tcp_metric_index idx) idx 68 net/ipv4/tcp_metrics.c return tm->tcpm_vals[idx]; idx 72 net/ipv4/tcp_metrics.c enum tcp_metric_index idx, idx 75 net/ipv4/tcp_metrics.c tm->tcpm_vals[idx] = val; idx 706 net/ipv6/addrconf.c int idx, s_idx; idx 727 net/ipv6/addrconf.c s_idx = idx = cb->args[1]; idx 730 net/ipv6/addrconf.c idx = 0; idx 736 net/ipv6/addrconf.c if (idx < s_idx) idx 754 net/ipv6/addrconf.c idx++; idx 782 net/ipv6/addrconf.c cb->args[1] = idx; idx 5165 net/ipv6/addrconf.c int idx, s_idx, s_ip_idx; idx 5173 net/ipv6/addrconf.c s_idx = idx = cb->args[1]; idx 5203 net/ipv6/addrconf.c idx = 0; idx 5206 net/ipv6/addrconf.c if (idx < s_idx) idx 5208 net/ipv6/addrconf.c if (h > s_h || idx > s_idx) idx 5218 net/ipv6/addrconf.c idx++; idx 5224 net/ipv6/addrconf.c cb->args[1] = idx; idx 5850 net/ipv6/addrconf.c int idx = 0, s_idx; idx 5870 net/ipv6/addrconf.c idx = 0; idx 5873 net/ipv6/addrconf.c if (idx < s_idx) idx 5884 net/ipv6/addrconf.c idx++; idx 5889 net/ipv6/addrconf.c cb->args[1] = idx; idx 491 net/ipv6/addrlabel.c int idx = 0, s_idx = cb->args[0]; idx 502 net/ipv6/addrlabel.c if (idx >= s_idx) { idx 512 net/ipv6/addrlabel.c idx++; idx 515 net/ipv6/addrlabel.c cb->args[0] = idx; idx 43 net/ipv6/netfilter/ip6t_NPT.c unsigned int i, idx; idx 54 net/ipv6/netfilter/ip6t_NPT.c idx = i / 32; idx 55 net/ipv6/netfilter/ip6t_NPT.c addr->s6_addr32[idx] &= mask; idx 56 net/ipv6/netfilter/ip6t_NPT.c addr->s6_addr32[idx] |= ~mask & npt->dst_pfx.in6.s6_addr32[idx]; idx 60 net/ipv6/netfilter/ip6t_NPT.c idx = 3; idx 62 net/ipv6/netfilter/ip6t_NPT.c for (idx = 4; idx < ARRAY_SIZE(addr->s6_addr16); idx++) { idx 63 net/ipv6/netfilter/ip6t_NPT.c if ((__force __sum16)addr->s6_addr16[idx] != idx 67 net/ipv6/netfilter/ip6t_NPT.c if (idx == ARRAY_SIZE(addr->s6_addr16)) idx 71 net/ipv6/netfilter/ip6t_NPT.c sum = ~csum_fold(csum_add(csum_unfold((__force __sum16)addr->s6_addr16[idx]), idx 75 net/ipv6/netfilter/ip6t_NPT.c *(__force __sum16 *)&addr->s6_addr16[idx] = sum; idx 80 net/kcm/kcmproc.c int idx; idx 186 net/kcm/kcmproc.c kcm_format_mux(struct kcm_mux *mux, loff_t idx, struct seq_file *seq) idx 226 net/kcm/kcmproc.c mux_state->idx = 0; idx 229 net/kcm/kcmproc.c kcm_format_mux(v, mux_state->idx, seq); idx 230 net/kcm/kcmproc.c mux_state->idx++; idx 690 net/mac80211/cfg.c rinfo->mcs = rate->idx; idx 702 net/mac80211/cfg.c brate = sband->bitrates[rate->idx].bitrate; idx 719 net/mac80211/cfg.c int idx, u8 *mac, struct station_info *sinfo) idx 728 net/mac80211/cfg.c sta = sta_info_get_by_idx(sdata, idx); idx 741 net/mac80211/cfg.c int idx, struct survey_info *survey) idx 745 net/mac80211/cfg.c return drv_get_survey(local, idx, survey); idx 1858 net/mac80211/cfg.c int idx, u8 *dst, u8 *next_hop, idx 1867 net/mac80211/cfg.c mpath = mesh_path_lookup_by_idx(sdata, idx); idx 1909 net/mac80211/cfg.c int idx, u8 *dst, u8 *mpp, idx 1918 net/mac80211/cfg.c mpath = mpp_path_lookup_by_idx(sdata, idx); idx 597 net/mac80211/debugfs_sta.c u8 idx = ((cap[i] & msk) >> (ffs(msk) - 1)) + off; \ idx 598 net/mac80211/debugfs_sta.c PRINT(fmt, (s << idx) + (m * idx)); \ idx 603 net/mac80211/driver-ops.h static inline int drv_get_survey(struct ieee80211_local *local, int idx, idx 608 net/mac80211/driver-ops.h trace_drv_get_survey(local, idx, survey); idx 611 net/mac80211/driver-ops.h ret = local->ops->get_survey(&local->hw, idx, survey); idx 348 net/mac80211/key.c int idx, bool uni, bool multi) idx 354 net/mac80211/key.c if (idx >= 0 && idx < NUM_DEFAULT_KEYS) idx 355 net/mac80211/key.c key = key_mtx_dereference(sdata->local, sdata->keys[idx]); idx 361 net/mac80211/key.c drv_set_default_unicast_key(sdata->local, sdata, idx); idx 370 net/mac80211/key.c void ieee80211_set_default_key(struct ieee80211_sub_if_data *sdata, int idx, idx 374 net/mac80211/key.c __ieee80211_set_default_key(sdata, idx, uni, multi); idx 379 net/mac80211/key.c __ieee80211_set_default_mgmt_key(struct ieee80211_sub_if_data *sdata, int idx) idx 385 net/mac80211/key.c if (idx >= NUM_DEFAULT_KEYS && idx 386 net/mac80211/key.c idx < NUM_DEFAULT_KEYS + NUM_DEFAULT_MGMT_KEYS) idx 387 net/mac80211/key.c key = key_mtx_dereference(sdata->local, sdata->keys[idx]); idx 395 net/mac80211/key.c int idx) idx 398 net/mac80211/key.c __ieee80211_set_default_mgmt_key(sdata, idx); idx 408 net/mac80211/key.c int idx; idx 429 net/mac80211/key.c idx = old->conf.keyidx; idx 439 net/mac80211/key.c idx = new->conf.keyidx; idx 449 net/mac80211/key.c rcu_assign_pointer(sta->ptk[idx], new); idx 454 net/mac80211/key.c rcu_assign_pointer(sta->gtk[idx], new); idx 480 net/mac80211/key.c rcu_assign_pointer(sdata->keys[idx], new); idx 499 net/mac80211/key.c ieee80211_key_alloc(u32 cipher, int idx, size_t key_len, idx 507 net/mac80211/key.c if (WARN_ON(idx < 0 || idx >= NUM_DEFAULT_KEYS + NUM_DEFAULT_MGMT_KEYS)) idx 522 net/mac80211/key.c key->conf.keyidx = idx; idx 768 net/mac80211/key.c int idx = key->conf.keyidx; idx 783 net/mac80211/key.c old_key = key_mtx_dereference(sdata->local, sta->ptk[idx]); idx 784 net/mac80211/key.c alt_key = key_mtx_dereference(sdata->local, sta->ptk[idx ^ 1]); idx 793 net/mac80211/key.c old_key = key_mtx_dereference(sdata->local, sta->gtk[idx]); idx 795 net/mac80211/key.c old_key = key_mtx_dereference(sdata->local, sdata->keys[idx]); idx 138 net/mac80211/key.h ieee80211_key_alloc(u32 cipher, int idx, size_t key_len, idx 152 net/mac80211/key.h void ieee80211_set_default_key(struct ieee80211_sub_if_data *sdata, int idx, idx 155 net/mac80211/key.h int idx); idx 216 net/mac80211/mesh.c u8 idx; idx 225 net/mac80211/mesh.c idx = le32_to_cpu(mesh_hdr->seqnum) & rmc->idx_mask; idx 226 net/mac80211/mesh.c hlist_for_each_entry_safe(p, n, &rmc->bucket[idx], list) { idx 244 net/mac80211/mesh.c hlist_add_head(&p->list, &rmc->bucket[idx]); idx 269 net/mac80211/mesh.h mesh_path_lookup_by_idx(struct ieee80211_sub_if_data *sdata, int idx); idx 271 net/mac80211/mesh.h mpp_path_lookup_by_idx(struct ieee80211_sub_if_data *sdata, int idx); idx 249 net/mac80211/mesh_pathtbl.c __mesh_path_lookup_by_idx(struct mesh_table *tbl, int idx) idx 255 net/mac80211/mesh_pathtbl.c if (i++ == idx) idx 280 net/mac80211/mesh_pathtbl.c mesh_path_lookup_by_idx(struct ieee80211_sub_if_data *sdata, int idx) idx 282 net/mac80211/mesh_pathtbl.c return __mesh_path_lookup_by_idx(sdata->u.mesh.mesh_paths, idx); idx 295 net/mac80211/mesh_pathtbl.c mpp_path_lookup_by_idx(struct ieee80211_sub_if_data *sdata, int idx) idx 297 net/mac80211/mesh_pathtbl.c return __mesh_path_lookup_by_idx(sdata->u.mesh.mpp_paths, idx); idx 4878 net/mac80211/mlme.c const struct ieee80211_bssid_index *idx; idx 4894 net/mac80211/mlme.c idx = (void *)(idx_ie + 2); idx 4897 net/mac80211/mlme.c *dtim_count = idx->dtim_count; idx 4900 net/mac80211/mlme.c *dtim_period = idx->dtim_period; idx 299 net/mac80211/rate.c static void rc_send_low_basicrate(s8 *idx, u32 basic_rates, idx 306 net/mac80211/rate.c if (*idx < 0) idx 308 net/mac80211/rate.c if (basic_rates & (1 << *idx)) idx 311 net/mac80211/rate.c for (i = *idx + 1; i <= sband->n_bitrates; i++) { idx 313 net/mac80211/rate.c *idx = i; idx 335 net/mac80211/rate.c info->control.rates[0].idx = 0; idx 346 net/mac80211/rate.c info->control.rates[0].idx = i; idx 380 net/mac80211/rate.c info->control.rates[0].idx = mcast_rate - 1; idx 391 net/mac80211/rate.c rc_send_low_basicrate(&info->control.rates[0].idx, idx 607 net/mac80211/rate.c rate = &sband->bitrates[rates[0].idx]; idx 632 net/mac80211/rate.c rates[i].idx = -1; idx 635 net/mac80211/rate.c if (rates[i].idx < 0) { idx 645 net/mac80211/rate.c WARN_ON(rates[i].idx > 76); idx 666 net/mac80211/rate.c if (WARN_ON_ONCE(rates[i].idx >= sband->n_bitrates)) { idx 667 net/mac80211/rate.c rates[i].idx = -1; idx 671 net/mac80211/rate.c rate = &sband->bitrates[rates[i].idx]; idx 702 net/mac80211/rate.c info->control.rates[i].idx >= 0 && idx 707 net/mac80211/rate.c rates[i].idx = ratetbl->rate[i].idx; idx 716 net/mac80211/rate.c rates[i].idx = -1; idx 720 net/mac80211/rate.c if (rates[i].idx < 0 || !rates[i].count) idx 792 net/mac80211/rate.c if (rates->rate[i].idx < 0) idx 795 net/mac80211/rate.c rate_idx_match_mask(&rates->rate[i].idx, &rates->rate[i].flags, idx 830 net/mac80211/rate.c if (rates[i].idx < 0) idx 834 net/mac80211/rate.c rate_idx_match_mask(&rates[i].idx, &rate_flags, sband, idx 862 net/mac80211/rate.c if (dest[0].idx < 0) idx 882 net/mac80211/rate.c info->control.rates[i].idx = -1; idx 114 net/mac80211/rc80211_minstrel.c int offset, int idx) idx 116 net/mac80211/rc80211_minstrel.c struct minstrel_rate *r = &mi->r[idx]; idx 118 net/mac80211/rc80211_minstrel.c ratetbl->rate[offset].idx = r->rix; idx 148 net/mac80211/rc80211_minstrel.c ratetbl->rate[i].idx = mi->lowest_rix; idx 273 net/mac80211/rc80211_minstrel.c if (ar[i].idx < 0) idx 276 net/mac80211/rc80211_minstrel.c ndx = rix_to_ndx(mi, ar[i].idx); idx 282 net/mac80211/rc80211_minstrel.c if ((i != IEEE80211_TX_MAX_RATES - 1) && (ar[i + 1].idx < 0)) idx 430 net/mac80211/rc80211_minstrel.c rate->idx = mi->r[ndx].rix; idx 569 net/mac80211/rc80211_minstrel.c int idx = mi->max_tp_rate[0]; idx 575 net/mac80211/rc80211_minstrel.c tmp_mrs = &mi->r[idx].stats; idx 576 net/mac80211/rc80211_minstrel.c tmp_cur_tp = minstrel_get_tp_avg(&mi->r[idx], tmp_mrs->prob_ewma) * 10; idx 287 net/mac80211/rc80211_minstrel_ht.c return GROUP_IDX((rate->idx / 8) + 1, idx 305 net/mac80211/rc80211_minstrel_ht.c int group, idx; idx 309 net/mac80211/rc80211_minstrel_ht.c idx = rate->idx % 8; idx 312 net/mac80211/rc80211_minstrel_ht.c idx = ieee80211_rate_get_vht_mcs(rate); idx 316 net/mac80211/rc80211_minstrel_ht.c for (idx = 0; idx < ARRAY_SIZE(mp->cck_rates); idx++) idx 317 net/mac80211/rc80211_minstrel_ht.c if (rate->idx == mp->cck_rates[idx]) idx 321 net/mac80211/rc80211_minstrel_ht.c if ((mi->supported[group] & BIT(idx + 4)) && idx 323 net/mac80211/rc80211_minstrel_ht.c idx += 4; idx 325 net/mac80211/rc80211_minstrel_ht.c return &mi->groups[group].rates[idx]; idx 583 net/mac80211/rc80211_minstrel_ht.c int idx; idx 591 net/mac80211/rc80211_minstrel_ht.c idx = g * MCS_GROUP_RATES + i; idx 592 net/mac80211/rc80211_minstrel_ht.c if (idx == mi->max_tp_rate[0]) idx 595 net/mac80211/rc80211_minstrel_ht.c rates[(*n_rates)++] = idx; idx 797 net/mac80211/rc80211_minstrel_ht.c if (rate->idx < 0) idx 807 net/mac80211/rc80211_minstrel_ht.c return rate->idx == mp->cck_rates[0] || idx 808 net/mac80211/rc80211_minstrel_ht.c rate->idx == mp->cck_rates[1] || idx 809 net/mac80211/rc80211_minstrel_ht.c rate->idx == mp->cck_rates[2] || idx 810 net/mac80211/rc80211_minstrel_ht.c rate->idx == mp->cck_rates[3]; idx 836 net/mac80211/rc80211_minstrel_ht.c minstrel_downgrade_rate(struct minstrel_ht_sta *mi, u16 *idx, bool primary) idx 840 net/mac80211/rc80211_minstrel_ht.c orig_group = group = *idx / MCS_GROUP_RATES; idx 852 net/mac80211/rc80211_minstrel_ht.c *idx = mi->groups[group].max_group_tp_rate[0]; idx 854 net/mac80211/rc80211_minstrel_ht.c *idx = mi->groups[group].max_group_tp_rate[1]; idx 1061 net/mac80211/rc80211_minstrel_ht.c u8 idx; idx 1079 net/mac80211/rc80211_minstrel_ht.c idx = mp->cck_rates[index % ARRAY_SIZE(mp->cck_rates)]; idx 1081 net/mac80211/rc80211_minstrel_ht.c idx = ((group->streams - 1) << 4) | idx 1084 net/mac80211/rc80211_minstrel_ht.c idx = index % MCS_GROUP_RATES + (group->streams - 1) * 8; idx 1097 net/mac80211/rc80211_minstrel_ht.c ratetbl->rate[offset].idx = idx; idx 1185 net/mac80211/rc80211_minstrel_ht.c rates->rate[i].idx = -1; idx 1342 net/mac80211/rc80211_minstrel_ht.c int idx = sample_idx % ARRAY_SIZE(mp->cck_rates); idx 1343 net/mac80211/rc80211_minstrel_ht.c rate->idx = mp->cck_rates[idx]; idx 1348 net/mac80211/rc80211_minstrel_ht.c rate->idx = sample_idx + (sample_group->streams - 1) * 8; idx 56 net/mac80211/rc80211_minstrel_ht_debugfs.c int idx = i * MCS_GROUP_RATES + j; idx 76 net/mac80211/rc80211_minstrel_ht_debugfs.c *(p++) = (idx == mi->max_tp_rate[0]) ? 'A' : ' '; idx 77 net/mac80211/rc80211_minstrel_ht_debugfs.c *(p++) = (idx == mi->max_tp_rate[1]) ? 'B' : ' '; idx 78 net/mac80211/rc80211_minstrel_ht_debugfs.c *(p++) = (idx == mi->max_tp_rate[2]) ? 'C' : ' '; idx 79 net/mac80211/rc80211_minstrel_ht_debugfs.c *(p++) = (idx == mi->max_tp_rate[3]) ? 'D' : ' '; idx 80 net/mac80211/rc80211_minstrel_ht_debugfs.c *(p++) = (idx == mi->max_prob_rate) ? 'P' : ' '; idx 92 net/mac80211/rc80211_minstrel_ht_debugfs.c p += sprintf(p, " %3u ", idx); idx 203 net/mac80211/rc80211_minstrel_ht_debugfs.c int idx = i * MCS_GROUP_RATES + j; idx 223 net/mac80211/rc80211_minstrel_ht_debugfs.c p += sprintf(p, "%s" ,((idx == mi->max_tp_rate[0]) ? "A" : "")); idx 224 net/mac80211/rc80211_minstrel_ht_debugfs.c p += sprintf(p, "%s" ,((idx == mi->max_tp_rate[1]) ? "B" : "")); idx 225 net/mac80211/rc80211_minstrel_ht_debugfs.c p += sprintf(p, "%s" ,((idx == mi->max_tp_rate[2]) ? "C" : "")); idx 226 net/mac80211/rc80211_minstrel_ht_debugfs.c p += sprintf(p, "%s" ,((idx == mi->max_tp_rate[3]) ? "D" : "")); idx 227 net/mac80211/rc80211_minstrel_ht_debugfs.c p += sprintf(p, "%s" ,((idx == mi->max_prob_rate) ? "P" : "")); idx 238 net/mac80211/rc80211_minstrel_ht_debugfs.c p += sprintf(p, "%u,", idx); idx 2117 net/mac80211/rx.c int i, idx; idx 2119 net/mac80211/rx.c idx = sdata->fragment_next; idx 2124 net/mac80211/rx.c idx--; idx 2125 net/mac80211/rx.c if (idx < 0) idx 2126 net/mac80211/rx.c idx = IEEE80211_FRAGMENT_MAX - 1; idx 2128 net/mac80211/rx.c entry = &sdata->fragments[idx]; idx 214 net/mac80211/sta_info.c int idx) idx 224 net/mac80211/sta_info.c if (i < idx) { idx 737 net/mac80211/sta_info.h int idx); idx 270 net/mac80211/status.c else if (info->status.rates[0].idx >= 0 && idx 290 net/mac80211/status.c } else if (info->status.rates[0].idx >= 0) { idx 338 net/mac80211/status.c } else if (info->status.rates[0].idx >= 0 && idx 342 net/mac80211/status.c sband->bitrates[info->status.rates[0].idx].bitrate; idx 503 net/mac80211/status.c if ((status && status->rate) || info->status.rates[0].idx < 0) idx 519 net/mac80211/status.c pos[2] = info->status.rates[0].idx; idx 781 net/mac80211/status.c info->status.rates[i].idx = -1; idx 784 net/mac80211/status.c } else if (info->status.rates[i].idx < 0) { idx 788 net/mac80211/status.c info->status.rates[i].idx = -1; idx 1092 net/mac80211/trace.h __field(int, idx) idx 1097 net/mac80211/trace.h __entry->idx = _idx; idx 1102 net/mac80211/trace.h LOCAL_PR_ARG, __entry->idx idx 77 net/mac80211/tx.c if (WARN_ON_ONCE(tx->rate.idx < 0)) idx 81 net/mac80211/tx.c txrate = &sband->bitrates[tx->rate.idx]; idx 691 net/mac80211/tx.c txrc.reported_rate.idx = -1; idx 753 net/mac80211/tx.c if (unlikely(info->control.rates[0].idx < 0)) { idx 756 net/mac80211/tx.c .idx = ratetbl->rate[0].idx, idx 761 net/mac80211/tx.c if (ratetbl->rate[0].idx < 0) idx 772 net/mac80211/tx.c if (txrc.reported_rate.idx < 0) { idx 984 net/mac80211/tx.c info->control.rates[1].idx = -1; idx 985 net/mac80211/tx.c info->control.rates[2].idx = -1; idx 986 net/mac80211/tx.c info->control.rates[3].idx = -1; idx 1390 net/mac80211/tx.c int idx, idx 2169 net/mac80211/tx.c info->control.rates[i].idx = -1; idx 2175 net/mac80211/tx.c info->control.rates[0].idx = rate; idx 2184 net/mac80211/tx.c info->control.rates[0].idx = i; idx 2189 net/mac80211/tx.c if (info->control.rates[0].idx < 0) idx 4628 net/mac80211/tx.c txrc.reported_rate.idx = -1; idx 1307 net/mpls/af_mpls.c int idx, s_idx; idx 1326 net/mpls/af_mpls.c s_idx = idx = cb->args[1]; idx 1329 net/mpls/af_mpls.c idx = 0; idx 1334 net/mpls/af_mpls.c if (idx < s_idx) idx 1350 net/mpls/af_mpls.c idx++; idx 1356 net/mpls/af_mpls.c cb->args[1] = idx; idx 1014 net/netfilter/ipvs/ip_vs_conn.c int idx; idx 1018 net/netfilter/ipvs/ip_vs_conn.c for (idx = 0; idx < ip_vs_conn_tab_size; idx++) { idx 1019 net/netfilter/ipvs/ip_vs_conn.c hlist_for_each_entry_rcu(cp, &ip_vs_conn_tab[idx], c_list) { idx 1024 net/netfilter/ipvs/ip_vs_conn.c iter->l = &ip_vs_conn_tab[idx]; idx 1050 net/netfilter/ipvs/ip_vs_conn.c int idx; idx 1061 net/netfilter/ipvs/ip_vs_conn.c idx = l - ip_vs_conn_tab; idx 1062 net/netfilter/ipvs/ip_vs_conn.c while (++idx < ip_vs_conn_tab_size) { idx 1063 net/netfilter/ipvs/ip_vs_conn.c hlist_for_each_entry_rcu(cp, &ip_vs_conn_tab[idx], c_list) { idx 1064 net/netfilter/ipvs/ip_vs_conn.c iter->l = &ip_vs_conn_tab[idx]; idx 1263 net/netfilter/ipvs/ip_vs_conn.c int idx; idx 1270 net/netfilter/ipvs/ip_vs_conn.c for (idx = 0; idx < (ip_vs_conn_tab_size>>5); idx++) { idx 1334 net/netfilter/ipvs/ip_vs_conn.c int idx; idx 1339 net/netfilter/ipvs/ip_vs_conn.c for (idx = 0; idx < ip_vs_conn_tab_size; idx++) { idx 1341 net/netfilter/ipvs/ip_vs_conn.c hlist_for_each_entry_rcu(cp, &ip_vs_conn_tab[idx], c_list) { idx 1394 net/netfilter/ipvs/ip_vs_conn.c int idx; idx 1424 net/netfilter/ipvs/ip_vs_conn.c for (idx = 0; idx < ip_vs_conn_tab_size; idx++) idx 1425 net/netfilter/ipvs/ip_vs_conn.c INIT_HLIST_HEAD(&ip_vs_conn_tab[idx]); idx 1427 net/netfilter/ipvs/ip_vs_conn.c for (idx = 0; idx < CT_LOCKARRAY_SIZE; idx++) { idx 1428 net/netfilter/ipvs/ip_vs_conn.c spin_lock_init(&__ip_vs_conntbl_lock_array[idx].l); idx 1579 net/netfilter/ipvs/ip_vs_ctl.c int idx; idx 1586 net/netfilter/ipvs/ip_vs_ctl.c for(idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 1587 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry_safe(svc, n, &ip_vs_svc_table[idx], idx 1597 net/netfilter/ipvs/ip_vs_ctl.c for(idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 1598 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry_safe(svc, n, &ip_vs_svc_fwm_table[idx], idx 1652 net/netfilter/ipvs/ip_vs_ctl.c unsigned int idx; idx 1659 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 1660 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry(svc, &ip_vs_svc_table[idx], s_list) { idx 1669 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry(svc, &ip_vs_svc_fwm_table[idx], f_list) { idx 1706 net/netfilter/ipvs/ip_vs_ctl.c int idx; idx 1709 net/netfilter/ipvs/ip_vs_ctl.c for(idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 1710 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry(svc, &ip_vs_svc_table[idx], s_list) { idx 1716 net/netfilter/ipvs/ip_vs_ctl.c for(idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 1717 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry(svc, &ip_vs_svc_fwm_table[idx], f_list) { idx 2024 net/netfilter/ipvs/ip_vs_ctl.c int idx; idx 2028 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 2029 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry_rcu(svc, &ip_vs_svc_table[idx], s_list) { idx 2032 net/netfilter/ipvs/ip_vs_ctl.c iter->bucket = idx; idx 2039 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 2040 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry_rcu(svc, &ip_vs_svc_fwm_table[idx], idx 2044 net/netfilter/ipvs/ip_vs_ctl.c iter->bucket = idx; idx 2582 net/netfilter/ipvs/ip_vs_ctl.c int idx, count=0; idx 2587 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 2588 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry(svc, &ip_vs_svc_table[idx], s_list) { idx 2606 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 2607 net/netfilter/ipvs/ip_vs_ctl.c hlist_for_each_entry(svc, &ip_vs_svc_fwm_table[idx], f_list) { idx 3106 net/netfilter/ipvs/ip_vs_ctl.c int idx = 0, i; idx 3115 net/netfilter/ipvs/ip_vs_ctl.c if (++idx <= start || (svc->ipvs != ipvs)) idx 3118 net/netfilter/ipvs/ip_vs_ctl.c idx--; idx 3126 net/netfilter/ipvs/ip_vs_ctl.c if (++idx <= start || (svc->ipvs != ipvs)) idx 3129 net/netfilter/ipvs/ip_vs_ctl.c idx--; idx 3137 net/netfilter/ipvs/ip_vs_ctl.c cb->args[0] = idx; idx 3317 net/netfilter/ipvs/ip_vs_ctl.c int idx = 0; idx 3338 net/netfilter/ipvs/ip_vs_ctl.c if (++idx <= start) idx 3341 net/netfilter/ipvs/ip_vs_ctl.c idx--; idx 3347 net/netfilter/ipvs/ip_vs_ctl.c cb->args[0] = idx; idx 3983 net/netfilter/ipvs/ip_vs_ctl.c int idx; idx 4002 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < ARRAY_SIZE(vs_vars); idx++) { idx 4003 net/netfilter/ipvs/ip_vs_ctl.c if (tbl[idx].proc_handler == proc_do_defense_mode) idx 4004 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx].extra2 = ipvs; idx 4006 net/netfilter/ipvs/ip_vs_ctl.c idx = 0; idx 4008 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_amemthresh; idx 4010 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_am_droprate; idx 4011 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_drop_entry; idx 4012 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_drop_packet; idx 4014 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_conntrack; idx 4016 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_secure_tcp; idx 4018 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_snat_reroute; idx 4020 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sync_ver; idx 4022 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sync_ports; idx 4023 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sync_persist_mode; idx 4025 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sync_qlen_max; idx 4027 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sync_sock_size; idx 4028 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_cache_bypass; idx 4029 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_expire_nodest_conn; idx 4030 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sloppy_tcp; idx 4031 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sloppy_sctp; idx 4032 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_expire_quiescent_template; idx 4035 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx].data = &ipvs->sysctl_sync_threshold; idx 4036 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].maxlen = sizeof(ipvs->sysctl_sync_threshold); idx 4038 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sync_refresh_period; idx 4040 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_sync_retries; idx 4041 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_nat_icmp_send; idx 4043 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_pmtu_disc; idx 4044 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_backup_only; idx 4046 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_conn_reuse_mode; idx 4047 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_schedule_icmp; idx 4048 net/netfilter/ipvs/ip_vs_ctl.c tbl[idx++].data = &ipvs->sysctl_ignore_tunneled; idx 4094 net/netfilter/ipvs/ip_vs_ctl.c int i, idx; idx 4097 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < IP_VS_RTAB_SIZE; idx++) idx 4098 net/netfilter/ipvs/ip_vs_ctl.c INIT_HLIST_HEAD(&ipvs->rs_table[idx]); idx 4178 net/netfilter/ipvs/ip_vs_ctl.c int idx; idx 4184 net/netfilter/ipvs/ip_vs_ctl.c for (idx = 0; idx < IP_VS_SVC_TAB_SIZE; idx++) { idx 4185 net/netfilter/ipvs/ip_vs_ctl.c INIT_HLIST_HEAD(&ip_vs_svc_table[idx]); idx 4186 net/netfilter/ipvs/ip_vs_ctl.c INIT_HLIST_HEAD(&ip_vs_svc_fwm_table[idx]); idx 15 net/netfilter/ipvs/ip_vs_pe_sip.c int *idx) idx 18 net/netfilter/ipvs/ip_vs_pe_sip.c size_t len = min3(max_len, callid_len, buf_len - *idx - 1); idx 19 net/netfilter/ipvs/ip_vs_pe_sip.c memcpy(buf + *idx, callid, len); idx 20 net/netfilter/ipvs/ip_vs_pe_sip.c buf[*idx+len] = '\0'; idx 21 net/netfilter/ipvs/ip_vs_pe_sip.c *idx += len + 1; idx 22 net/netfilter/ipvs/ip_vs_pe_sip.c return buf + *idx - len; idx 107 net/netfilter/nf_conntrack_h323_main.c unsigned char **data, TransportAddress *taddr, int idx, idx 659 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; idx 670 net/netfilter/nf_tables_api.c if (idx < s_idx) idx 672 net/netfilter/nf_tables_api.c if (idx > s_idx) idx 686 net/netfilter/nf_tables_api.c idx++; idx 690 net/netfilter/nf_tables_api.c cb->args[0] = idx; idx 1364 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; idx 1376 net/netfilter/nf_tables_api.c if (idx < s_idx) idx 1378 net/netfilter/nf_tables_api.c if (idx > s_idx) idx 1394 net/netfilter/nf_tables_api.c idx++; idx 1399 net/netfilter/nf_tables_api.c cb->args[0] = idx; idx 2437 net/netfilter/nf_tables_api.c unsigned int *idx, idx 2450 net/netfilter/nf_tables_api.c if (*idx < s_idx) idx 2452 net/netfilter/nf_tables_api.c if (*idx > s_idx) { idx 2468 net/netfilter/nf_tables_api.c (*idx)++; idx 2480 net/netfilter/nf_tables_api.c unsigned int idx = 0; idx 2505 net/netfilter/nf_tables_api.c __nf_tables_dump_rules(skb, &idx, idx 2513 net/netfilter/nf_tables_api.c if (__nf_tables_dump_rules(skb, &idx, cb, table, chain)) idx 2523 net/netfilter/nf_tables_api.c cb->args[0] = idx; idx 3407 net/netfilter/nf_tables_api.c unsigned int idx, s_idx = cb->args[0]; idx 3432 net/netfilter/nf_tables_api.c idx = 0; idx 3434 net/netfilter/nf_tables_api.c if (idx < s_idx) idx 3446 net/netfilter/nf_tables_api.c cb->args[0] = idx; idx 3452 net/netfilter/nf_tables_api.c idx++; idx 5354 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; idx 5374 net/netfilter/nf_tables_api.c if (idx < s_idx) idx 5376 net/netfilter/nf_tables_api.c if (idx > s_idx) idx 5397 net/netfilter/nf_tables_api.c idx++; idx 5403 net/netfilter/nf_tables_api.c cb->args[0] = idx; idx 6048 net/netfilter/nf_tables_api.c unsigned int idx = 0, s_idx = cb->args[0]; idx 6064 net/netfilter/nf_tables_api.c if (idx < s_idx) idx 6066 net/netfilter/nf_tables_api.c if (idx > s_idx) idx 6082 net/netfilter/nf_tables_api.c idx++; idx 6088 net/netfilter/nf_tables_api.c cb->args[0] = idx; idx 53 net/netfilter/nft_set_bitmap.c u32 *idx, u32 *off) idx 63 net/netfilter/nft_set_bitmap.c *idx = k / BITS_PER_BYTE; idx 71 net/netfilter/nft_set_bitmap.c nft_bitmap_active(const u8 *bitmap, u32 idx, u32 off, u8 genmask) idx 73 net/netfilter/nft_set_bitmap.c return (bitmap[idx] & (0x3 << off)) & (genmask << off); idx 81 net/netfilter/nft_set_bitmap.c u32 idx, off; idx 83 net/netfilter/nft_set_bitmap.c nft_bitmap_location(set, key, &idx, &off); idx 85 net/netfilter/nft_set_bitmap.c return nft_bitmap_active(priv->bitmap, idx, off, genmask); idx 130 net/netfilter/nft_set_bitmap.c u32 idx, off; idx 138 net/netfilter/nft_set_bitmap.c nft_bitmap_location(set, nft_set_ext_key(&new->ext), &idx, &off); idx 140 net/netfilter/nft_set_bitmap.c priv->bitmap[idx] |= (genmask << off); idx 153 net/netfilter/nft_set_bitmap.c u32 idx, off; idx 155 net/netfilter/nft_set_bitmap.c nft_bitmap_location(set, nft_set_ext_key(&be->ext), &idx, &off); idx 157 net/netfilter/nft_set_bitmap.c priv->bitmap[idx] &= ~(genmask << off); idx 168 net/netfilter/nft_set_bitmap.c u32 idx, off; idx 170 net/netfilter/nft_set_bitmap.c nft_bitmap_location(set, nft_set_ext_key(&be->ext), &idx, &off); idx 172 net/netfilter/nft_set_bitmap.c priv->bitmap[idx] |= (genmask << off); idx 182 net/netfilter/nft_set_bitmap.c u32 idx, off; idx 184 net/netfilter/nft_set_bitmap.c nft_bitmap_location(set, nft_set_ext_key(&be->ext), &idx, &off); idx 186 net/netfilter/nft_set_bitmap.c priv->bitmap[idx] &= ~(genmask << off); idx 199 net/netfilter/nft_set_bitmap.c u32 idx, off; idx 201 net/netfilter/nft_set_bitmap.c nft_bitmap_location(set, elem->key.val.data, &idx, &off); idx 208 net/netfilter/nft_set_bitmap.c priv->bitmap[idx] &= ~(genmask << off); idx 611 net/netlabel/netlabel_kapi.c u32 idx; idx 620 net/netlabel/netlabel_kapi.c idx = offset / NETLBL_CATMAP_MAPSIZE; idx 623 net/netlabel/netlabel_kapi.c idx = 0; idx 626 net/netlabel/netlabel_kapi.c bitmap = iter->bitmap[idx] >> bit; idx 635 net/netlabel/netlabel_kapi.c (NETLBL_CATMAP_MAPSIZE * idx) + bit; idx 637 net/netlabel/netlabel_kapi.c if (++idx >= NETLBL_CATMAP_MAPCNT) { idx 640 net/netlabel/netlabel_kapi.c idx = 0; idx 644 net/netlabel/netlabel_kapi.c bitmap = iter->bitmap[idx]; idx 667 net/netlabel/netlabel_kapi.c u32 idx; idx 677 net/netlabel/netlabel_kapi.c idx = offset / NETLBL_CATMAP_MAPSIZE; idx 680 net/netlabel/netlabel_kapi.c idx = 0; idx 686 net/netlabel/netlabel_kapi.c bitmap = iter->bitmap[idx]; idx 692 net/netlabel/netlabel_kapi.c if (prev && idx == 0 && bit == 0) idx 696 net/netlabel/netlabel_kapi.c (NETLBL_CATMAP_MAPSIZE * idx) + bit - 1; idx 697 net/netlabel/netlabel_kapi.c else if (++idx >= NETLBL_CATMAP_MAPCNT) { idx 702 net/netlabel/netlabel_kapi.c idx = 0; idx 731 net/netlabel/netlabel_kapi.c u32 idx; idx 758 net/netlabel/netlabel_kapi.c idx = off / NETLBL_CATMAP_MAPSIZE; idx 759 net/netlabel/netlabel_kapi.c *bitmap = iter->bitmap[idx] >> (off % NETLBL_CATMAP_MAPSIZE); idx 780 net/netlabel/netlabel_kapi.c u32 idx; idx 787 net/netlabel/netlabel_kapi.c idx = bit / NETLBL_CATMAP_MAPSIZE; idx 788 net/netlabel/netlabel_kapi.c iter->bitmap[idx] |= NETLBL_CATMAP_BIT << (bit % NETLBL_CATMAP_MAPSIZE); idx 848 net/netlabel/netlabel_kapi.c u32 idx; idx 859 net/netlabel/netlabel_kapi.c idx = offset / NETLBL_CATMAP_MAPSIZE; idx 860 net/netlabel/netlabel_kapi.c iter->bitmap[idx] |= bitmap << (offset % NETLBL_CATMAP_MAPSIZE); idx 1770 net/netlink/af_netlink.c int pos, idx, shift; idx 1778 net/netlink/af_netlink.c idx = pos / sizeof(unsigned long); idx 1780 net/netlink/af_netlink.c if (put_user((u32)(nlk->groups[idx] >> shift), idx 272 net/nfc/core.c if (dev->targets[i].idx == target_idx) idx 459 net/nfc/core.c if (dev->active_target->idx != target_idx) { idx 503 net/nfc/core.c if (dev->active_target->idx != target_idx) { idx 537 net/nfc/core.c if (se->idx == se_idx) idx 771 net/nfc/core.c targets[i].idx = dev->target_next_idx++; idx 831 net/nfc/core.c if (tg->idx == target_idx) idx 881 net/nfc/core.c se->idx = se_idx; idx 908 net/nfc/core.c if (se->idx == se_idx) { idx 968 net/nfc/core.c nfc_genl_se_removed(dev, se->idx); idx 973 net/nfc/core.c ida_simple_remove(&nfc_index_ida, dev->idx); idx 991 net/nfc/core.c u32 active_target_idx = dev->active_target->idx; idx 1022 net/nfc/core.c const unsigned int *idx = data; idx 1024 net/nfc/core.c return dev->idx == *idx; idx 1027 net/nfc/core.c struct nfc_dev *nfc_get_device(unsigned int idx) idx 1031 net/nfc/core.c d = class_find_device(&nfc_class, NULL, &idx, match_idx); idx 1065 net/nfc/core.c dev->idx = rc; idx 1068 net/nfc/core.c dev_set_name(&dev->dev, "nfc%d", dev->idx); idx 350 net/nfc/digital_dep.c nfc_dep_link_is_up(ddev->nfc_dev, target->idx, NFC_COMM_ACTIVE, idx 462 net/nfc/digital_dep.c rc = nfc_dep_link_is_up(ddev->nfc_dev, target->idx, NFC_COMM_ACTIVE, idx 144 net/nfc/hci/command.c int nfc_hci_set_param(struct nfc_hci_dev *hdev, u8 gate, u8 idx, idx 158 net/nfc/hci/command.c pr_debug("idx=%d to gate %d\n", idx, gate); idx 164 net/nfc/hci/command.c *tmp = idx; idx 176 net/nfc/hci/command.c int nfc_hci_get_param(struct nfc_hci_dev *hdev, u8 gate, u8 idx, idx 179 net/nfc/hci/command.c pr_debug("gate=%d regidx=%d\n", gate, idx); idx 182 net/nfc/hci/command.c &idx, 1, skb); idx 729 net/nfc/hci/core.c pr_debug("target_idx=%d\n", target->idx); idx 696 net/nfc/llcp_core.c data[0] = local->dev ? local->dev->idx : 0xFF; idx 979 net/nfc/llcp_core.c nfc_get_device(local->dev->idx); idx 513 net/nfc/llcp_sock.c llcp_addr->dev_idx = llcp_sock->dev->idx; idx 856 net/nfc/nci/core.c pr_debug("target_idx %d, protocol 0x%x\n", target->idx, protocol); idx 870 net/nfc/nci/core.c if (ndev->targets[i].idx == target->idx) { idx 947 net/nfc/nci/core.c pr_debug("target_idx %d, comm_mode %d\n", target->idx, comm_mode); idx 956 net/nfc/nci/core.c rc = nfc_dep_link_is_up(nfc_dev, target->idx, NFC_COMM_PASSIVE, idx 999 net/nfc/nci/core.c pr_debug("target_idx %d, len %d\n", target->idx, skb->len); idx 562 net/nfc/nci/hci.c int nci_hci_set_param(struct nci_dev *ndev, u8 gate, u8 idx, idx 572 net/nfc/nci/hci.c pr_debug("idx=%d to gate %d\n", idx, gate); idx 585 net/nfc/nci/hci.c *tmp = idx; idx 610 net/nfc/nci/hci.c int nci_hci_get_param(struct nci_dev *ndev, u8 gate, u8 idx, idx 619 net/nfc/nci/hci.c pr_debug("idx=%d to gate %d\n", idx, gate); idx 632 net/nfc/nci/hci.c data.data = &idx; idx 73 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_TARGET_INDEX, target->idx) || idx 112 net/nfc/netlink.c u32 idx; idx 124 net/nfc/netlink.c idx = nla_get_u32(attrbuf[NFC_ATTR_DEVICE_INDEX]); idx 126 net/nfc/netlink.c dev = nfc_get_device(idx); idx 194 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 251 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 282 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 300 net/nfc/netlink.c nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx) || idx 351 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 385 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 446 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx) || idx 477 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx) || idx 513 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx) || idx 558 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx) || idx 669 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 708 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 727 net/nfc/netlink.c u32 idx; idx 733 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 735 net/nfc/netlink.c dev = nfc_get_device(idx); idx 765 net/nfc/netlink.c u32 idx; idx 770 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 772 net/nfc/netlink.c dev = nfc_get_device(idx); idx 786 net/nfc/netlink.c u32 idx; idx 791 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 793 net/nfc/netlink.c dev = nfc_get_device(idx); idx 807 net/nfc/netlink.c u32 idx; idx 818 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 828 net/nfc/netlink.c dev = nfc_get_device(idx); idx 848 net/nfc/netlink.c u32 idx; idx 853 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 855 net/nfc/netlink.c dev = nfc_get_device(idx); idx 940 net/nfc/netlink.c u32 idx; idx 949 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 960 net/nfc/netlink.c dev = nfc_get_device(idx); idx 975 net/nfc/netlink.c u32 idx; idx 981 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 983 net/nfc/netlink.c dev = nfc_get_device(idx); idx 1004 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, local->dev->idx) || idx 1024 net/nfc/netlink.c u32 idx; idx 1030 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 1032 net/nfc/netlink.c dev = nfc_get_device(idx); idx 1073 net/nfc/netlink.c u32 idx; idx 1096 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 1098 net/nfc/netlink.c dev = nfc_get_device(idx); idx 1138 net/nfc/netlink.c u32 idx; idx 1150 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 1152 net/nfc/netlink.c dev = nfc_get_device(idx); idx 1225 net/nfc/netlink.c u32 idx; idx 1231 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 1233 net/nfc/netlink.c dev = nfc_get_device(idx); idx 1263 net/nfc/netlink.c nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 1282 net/nfc/netlink.c u32 idx, se_idx; idx 1288 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 1291 net/nfc/netlink.c dev = nfc_get_device(idx); idx 1305 net/nfc/netlink.c u32 idx, se_idx; idx 1311 net/nfc/netlink.c idx = nla_get_u32(info->attrs[NFC_ATTR_DEVICE_INDEX]); idx 1314 net/nfc/netlink.c dev = nfc_get_device(idx); idx 1341 net/nfc/netlink.c if (nla_put_u32(msg, NFC_ATTR_DEVICE_INDEX, dev->idx) || idx 1342 net/nfc/netlink.c nla_put_u32(msg, NFC_ATTR_SE_INDEX, se->idx) || idx 1612 net/nfc/netlink.c if (nla_put_u32(skb, NFC_ATTR_DEVICE_INDEX, dev->idx)) idx 98 net/nfc/nfc.h struct nfc_dev *nfc_get_device(unsigned int idx); idx 376 net/nfc/rawsock.c data[0] = dev ? dev->idx : 0xFF; idx 61 net/openvswitch/flow_table.h u32 *bucket, u32 *idx); idx 1088 net/packet/af_packet.c unsigned int idx, idx 1092 net/packet/af_packet.c struct tpacket_block_desc *pbd = GET_PBLOCK_DESC(pkc, idx); idx 1204 net/packet/af_packet.c int idx, len; idx 1207 net/packet/af_packet.c idx = READ_ONCE(po->rx_ring.head); idx 1209 net/packet/af_packet.c idx += len >> pow_off; idx 1210 net/packet/af_packet.c if (idx >= len) idx 1211 net/packet/af_packet.c idx -= len; idx 1212 net/packet/af_packet.c return packet_lookup_frame(po, &po->rx_ring, idx, TP_STATUS_KERNEL); idx 1217 net/packet/af_packet.c int idx, len; idx 1220 net/packet/af_packet.c idx = READ_ONCE(po->rx_ring.prb_bdqc.kactive_blk_num); idx 1222 net/packet/af_packet.c idx += len >> pow_off; idx 1223 net/packet/af_packet.c if (idx >= len) idx 1224 net/packet/af_packet.c idx -= len; idx 1225 net/packet/af_packet.c return prb_lookup_block(po, &po->rx_ring, idx, TP_STATUS_KERNEL); idx 1349 net/packet/af_packet.c unsigned int idx, bool try_self, idx 1355 net/packet/af_packet.c po = pkt_sk(f->arr[idx]); idx 1361 net/packet/af_packet.c return idx; idx 1383 net/packet/af_packet.c return idx; idx 1421 net/packet/af_packet.c unsigned int idx; idx 1436 net/packet/af_packet.c idx = fanout_demux_hash(f, skb, num); idx 1439 net/packet/af_packet.c idx = fanout_demux_lb(f, skb, num); idx 1442 net/packet/af_packet.c idx = fanout_demux_cpu(f, skb, num); idx 1445 net/packet/af_packet.c idx = fanout_demux_rnd(f, skb, num); idx 1448 net/packet/af_packet.c idx = fanout_demux_qm(f, skb, num); idx 1451 net/packet/af_packet.c idx = fanout_demux_rollover(f, skb, 0, false, num); idx 1455 net/packet/af_packet.c idx = fanout_demux_bpf(f, skb, num); idx 1460 net/packet/af_packet.c idx = fanout_demux_rollover(f, skb, idx, true, num); idx 1462 net/packet/af_packet.c po = pkt_sk(f->arr[idx]); idx 73 net/psample/psample.c int idx = 0; idx 80 net/psample/psample.c if (idx < start) { idx 81 net/psample/psample.c idx++; idx 89 net/psample/psample.c idx++; idx 93 net/psample/psample.c cb->args[0] = idx; idx 706 net/rds/ib_cm.c int idx = 0; idx 711 net/rds/ib_cm.c idx = dev->ifindex; idx 717 net/rds/ib_cm.c return idx; idx 44 net/rfkill/core.c u32 idx; idx 260 net/rfkill/core.c ev->idx = rfkill->idx; idx 692 net/rfkill/core.c return sprintf(buf, "%d\n", rfkill->idx); idx 1020 net/rfkill/core.c rfkill->idx = rfkill_no; idx 1242 net/rfkill/core.c if (rfkill->idx == ev.idx && idx 171 net/sched/ematch.c struct tcf_ematch *em, struct nlattr *nla, int idx) idx 198 net/sched/ematch.c if (ref <= idx) idx 306 net/sched/ematch.c int idx, list_len, matches_len, err; idx 348 net/sched/ematch.c for (idx = 0; nla_ok(rt_match, list_len); idx++) { idx 351 net/sched/ematch.c if (rt_match->nla_type != (idx + 1)) idx 354 net/sched/ematch.c if (idx >= tree_hdr->nmatches) idx 360 net/sched/ematch.c em = tcf_em_get_match(tree, idx); idx 362 net/sched/ematch.c err = tcf_em_validate(tp, tree_hdr, em, rt_match, idx); idx 374 net/sched/ematch.c if (idx != tree_hdr->nmatches) { idx 1730 net/sched/sch_api.c int idx, q_idx; idx 1740 net/sched/sch_api.c idx = 0; idx 1751 net/sched/sch_api.c if (idx < s_idx) idx 1753 net/sched/sch_api.c if (idx > s_idx) idx 1769 net/sched/sch_api.c idx++; idx 1773 net/sched/sch_api.c cb->args[0] = idx; idx 1465 net/sched/sch_cake.c u32 idx = 0, tin = 0, len; idx 1482 net/sched/sch_cake.c idx = qq.b; idx 1485 net/sched/sch_cake.c flow = &b->flows[idx]; idx 1490 net/sched/sch_cake.c return idx + (tin << 16); idx 1498 net/sched/sch_cake.c b->backlogs[idx] -= len; idx 1515 net/sched/sch_cake.c return idx + (tin << 16); idx 1640 net/sched/sch_cake.c u32 idx; idx 1643 net/sched/sch_cake.c idx = cake_classify(sch, &b, skb, q->flow_mode, &ret); idx 1644 net/sched/sch_cake.c if (idx == 0) { idx 1650 net/sched/sch_cake.c idx--; idx 1651 net/sched/sch_cake.c flow = &b->flows[idx]; idx 1705 net/sched/sch_cake.c b->backlogs[idx] += slen; idx 1740 net/sched/sch_cake.c b->backlogs[idx] += len; idx 1747 net/sched/sch_cake.c cake_heapify_up(q, b->overflow_idx[idx]); idx 2952 net/sched/sch_cake.c u32 idx = cl - 1; idx 2954 net/sched/sch_cake.c if (idx < CAKE_QUEUES * q->tin_cnt) { idx 2956 net/sched/sch_cake.c &q->tins[q->tin_order[idx / CAKE_QUEUES]]; idx 2959 net/sched/sch_cake.c flow = &b->flows[idx % CAKE_QUEUES]; idx 2970 net/sched/sch_cake.c qs.backlog = b->backlogs[idx % CAKE_QUEUES]; idx 114 net/sched/sch_choke.c static void choke_drop_by_idx(struct Qdisc *sch, unsigned int idx, idx 118 net/sched/sch_choke.c struct sk_buff *skb = q->tab[idx]; idx 120 net/sched/sch_choke.c q->tab[idx] = NULL; idx 122 net/sched/sch_choke.c if (idx == q->head) idx 124 net/sched/sch_choke.c if (idx == q->tail) idx 235 net/sched/sch_choke.c unsigned int idx; idx 238 net/sched/sch_choke.c if (choke_match_random(q, skb, &idx)) { idx 240 net/sched/sch_choke.c choke_drop_by_idx(sch, idx, to_free); idx 624 net/sched/sch_fq.c unsigned int idx; idx 634 net/sched/sch_fq.c for (idx = 0; idx < (1U << q->fq_trees_log); idx++) { idx 635 net/sched/sch_fq.c root = &q->fq_root[idx]; idx 661 net/sched/sch_fq.c u32 idx; idx 663 net/sched/sch_fq.c for (idx = 0; idx < (1U << old_log); idx++) { idx 664 net/sched/sch_fq.c oroot = &old_array[idx]; idx 708 net/sched/sch_fq.c u32 idx; idx 719 net/sched/sch_fq.c for (idx = 0; idx < (1U << log); idx++) idx 720 net/sched/sch_fq.c array[idx] = RB_ROOT; idx 143 net/sched/sch_fq_codel.c unsigned int maxbacklog = 0, idx = 0, i, len; idx 158 net/sched/sch_fq_codel.c idx = i; idx 165 net/sched/sch_fq_codel.c flow = &q->flows[idx]; idx 177 net/sched/sch_fq_codel.c q->backlogs[idx] -= len; idx 182 net/sched/sch_fq_codel.c return idx; idx 189 net/sched/sch_fq_codel.c unsigned int idx, prev_backlog, prev_qlen; idx 195 net/sched/sch_fq_codel.c idx = fq_codel_classify(skb, sch, &ret); idx 196 net/sched/sch_fq_codel.c if (idx == 0) { idx 202 net/sched/sch_fq_codel.c idx--; idx 205 net/sched/sch_fq_codel.c flow = &q->flows[idx]; idx 207 net/sched/sch_fq_codel.c q->backlogs[idx] += qdisc_pkt_len(skb); idx 243 net/sched/sch_fq_codel.c if (ret == idx) { idx 619 net/sched/sch_fq_codel.c u32 idx = cl - 1; idx 623 net/sched/sch_fq_codel.c if (idx < q->flows_cnt) { idx 624 net/sched/sch_fq_codel.c const struct fq_codel_flow *flow = &q->flows[idx]; idx 652 net/sched/sch_fq_codel.c qs.backlog = q->backlogs[idx]; idx 657 net/sched/sch_fq_codel.c if (idx < q->flows_cnt) idx 375 net/sched/sch_hhf.c enum wdrr_bucket_idx idx; idx 379 net/sched/sch_hhf.c idx = hhf_classify(skb, sch); idx 381 net/sched/sch_hhf.c bucket = &q->buckets[idx]; idx 392 net/sched/sch_hhf.c if (idx == WDRR_BUCKET_FOR_HH) { idx 410 net/sched/sch_hhf.c if (hhf_drop(sch, to_free) == idx) idx 866 net/sched/sch_sfq.c sfq_index idx = q->ht[cl - 1]; idx 870 net/sched/sch_sfq.c if (idx != SFQ_EMPTY_SLOT) { idx 871 net/sched/sch_sfq.c const struct sfq_slot *slot = &q->slots[idx]; idx 478 net/smc/smc_pnet.c int idx = 0; idx 490 net/smc/smc_pnet.c if (idx++ < start_idx) idx 498 net/smc/smc_pnet.c --idx; idx 506 net/smc/smc_pnet.c return idx; idx 517 net/smc/smc_pnet.c if (idx++ < start_idx) idx 528 net/smc/smc_pnet.c --idx; idx 542 net/smc/smc_pnet.c if (idx++ < start_idx) idx 550 net/smc/smc_pnet.c --idx; idx 557 net/smc/smc_pnet.c return idx; idx 563 net/smc/smc_pnet.c int idx; idx 565 net/smc/smc_pnet.c idx = _smc_pnet_dump(net, skb, NETLINK_CB(cb->skb).portid, idx 568 net/smc/smc_pnet.c cb->args[0] = idx; idx 45 net/smc/smc_wr.c u32 idx; idx 146 net/smc/smc_wr.c static inline int smc_wr_tx_get_free_slot_index(struct smc_link *link, u32 *idx) idx 148 net/smc/smc_wr.c *idx = link->wr_tx_cnt; idx 149 net/smc/smc_wr.c for_each_clear_bit(*idx, link->wr_tx_mask, link->wr_tx_cnt) { idx 150 net/smc/smc_wr.c if (!test_and_set_bit(*idx, link->wr_tx_mask)) idx 153 net/smc/smc_wr.c *idx = link->wr_tx_cnt; idx 175 net/smc/smc_wr.c u32 idx = link->wr_tx_cnt; idx 183 net/smc/smc_wr.c rc = smc_wr_tx_get_free_slot_index(link, &idx); idx 190 net/smc/smc_wr.c (smc_wr_tx_get_free_slot_index(link, &idx) != -EBUSY), idx 197 net/smc/smc_wr.c if (idx == link->wr_tx_cnt) idx 201 net/smc/smc_wr.c wr_pend = &link->wr_tx_pends[idx]; idx 205 net/smc/smc_wr.c wr_pend->idx = idx; idx 206 net/smc/smc_wr.c wr_ib = &link->wr_tx_ibs[idx]; idx 208 net/smc/smc_wr.c *wr_buf = &link->wr_tx_bufs[idx]; idx 210 net/smc/smc_wr.c *wr_rdma_buf = &link->wr_tx_rdmas[idx]; idx 221 net/smc/smc_wr.c if (pend->idx < link->wr_tx_cnt) { idx 222 net/smc/smc_wr.c u32 idx = pend->idx; idx 225 net/smc/smc_wr.c memset(&link->wr_tx_pends[idx], 0, idx 226 net/smc/smc_wr.c sizeof(link->wr_tx_pends[idx])); idx 227 net/smc/smc_wr.c memset(&link->wr_tx_bufs[idx], 0, idx 228 net/smc/smc_wr.c sizeof(link->wr_tx_bufs[idx])); idx 229 net/smc/smc_wr.c test_and_clear_bit(idx, link->wr_tx_mask); idx 247 net/smc/smc_wr.c rc = ib_post_send(link->roce_qp, &link->wr_tx_ibs[pend->idx], NULL); idx 1627 net/sunrpc/clnt.c int idx = task->tk_msg.rpc_proc->p_statidx; idx 1637 net/sunrpc/clnt.c clnt->cl_program->version[clnt->cl_vers]->counts[idx]++; idx 1029 net/wireless/nl80211.c int idx; idx 1056 net/wireless/nl80211.c k->idx = nla_get_u8(tb[NL80211_KEY_IDX]); idx 1108 net/wireless/nl80211.c k->idx = nla_get_u8(info->attrs[NL80211_ATTR_KEY_IDX]); idx 1148 net/wireless/nl80211.c k->idx = -1; idx 1171 net/wireless/nl80211.c if (k->idx != -1) { idx 1173 net/wireless/nl80211.c if (k->idx < 4 || k->idx > 5) { idx 1179 net/wireless/nl80211.c if (k->idx < 0 || k->idx > 3) { idx 1184 net/wireless/nl80211.c if (k->idx < 0 || k->idx > 5) { idx 1221 net/wireless/nl80211.c parse.idx = -1; idx 1229 net/wireless/nl80211.c if (parse.idx < 0 || parse.idx > 3) { idx 1240 net/wireless/nl80211.c result->def = parse.idx; idx 1246 net/wireless/nl80211.c parse.idx, false, NULL); idx 1255 net/wireless/nl80211.c result->params[parse.idx].cipher = parse.p.cipher; idx 1256 net/wireless/nl80211.c result->params[parse.idx].key_len = parse.p.key_len; idx 1257 net/wireless/nl80211.c result->params[parse.idx].key = result->data[parse.idx]; idx 1258 net/wireless/nl80211.c memcpy(result->data[parse.idx], parse.p.key, parse.p.key_len); idx 2512 net/wireless/nl80211.c int idx = 0, ret; idx 2536 net/wireless/nl80211.c if (++idx <= state->start) idx 2570 net/wireless/nl80211.c idx--; idx 2578 net/wireless/nl80211.c state->start = idx; idx 2969 net/wireless/nl80211.c int idx, mbm = 0; idx 2977 net/wireless/nl80211.c idx = NL80211_ATTR_WIPHY_TX_POWER_SETTING; idx 2978 net/wireless/nl80211.c type = nla_get_u32(info->attrs[idx]); idx 2985 net/wireless/nl80211.c idx = NL80211_ATTR_WIPHY_TX_POWER_LEVEL; idx 2986 net/wireless/nl80211.c mbm = nla_get_u32(info->attrs[idx]); idx 3710 net/wireless/nl80211.c int idx; idx 3744 net/wireless/nl80211.c if (nla_put_u8(cookie->msg, NL80211_KEY_IDX, cookie->idx)) idx 3800 net/wireless/nl80211.c cookie.idx = key_idx; idx 3839 net/wireless/nl80211.c if (key.idx < 0) idx 3861 net/wireless/nl80211.c err = rdev_set_default_key(rdev, dev, key.idx, idx 3868 net/wireless/nl80211.c dev->ieee80211_ptr->wext.default_key = key.idx; idx 3885 net/wireless/nl80211.c err = rdev_set_default_mgmt_key(rdev, dev, key.idx); idx 3890 net/wireless/nl80211.c dev->ieee80211_ptr->wext.default_mgmt_key = key.idx; idx 3900 net/wireless/nl80211.c if (!mac_addr || key.idx < 0 || key.idx > 1) { idx 3905 net/wireless/nl80211.c err = rdev_add_key(rdev, dev, key.idx, idx 3950 net/wireless/nl80211.c if (cfg80211_validate_key_settings(rdev, &key.p, key.idx, idx 3958 net/wireless/nl80211.c err = rdev_add_key(rdev, dev, key.idx, idx 4004 net/wireless/nl80211.c err = rdev_del_key(rdev, dev, key.idx, idx 4010 net/wireless/nl80211.c if (key.idx == dev->ieee80211_ptr->wext.default_key) idx 4012 net/wireless/nl80211.c else if (key.idx == dev->ieee80211_ptr->wext.default_mgmt_key) idx 5671 net/wireless/nl80211.c int idx; idx 5679 net/wireless/nl80211.c idx = NL80211_ATTR_STA_TX_POWER_SETTING; idx 5680 net/wireless/nl80211.c params->txpwr.type = nla_get_u8(info->attrs[idx]); idx 5683 net/wireless/nl80211.c idx = NL80211_ATTR_STA_TX_POWER; idx 5685 net/wireless/nl80211.c if (info->attrs[idx]) idx 5687 net/wireless/nl80211.c nla_get_s16(info->attrs[idx]); idx 8748 net/wireless/nl80211.c int start = cb->args[2], idx = 0; idx 8773 net/wireless/nl80211.c if (++idx <= start) idx 8778 net/wireless/nl80211.c idx--; idx 8786 net/wireless/nl80211.c cb->args[2] = idx; idx 8963 net/wireless/nl80211.c if (key.idx >= 0) { idx 8973 net/wireless/nl80211.c if (key.idx > 3) idx 8980 net/wireless/nl80211.c if (key.idx >= 0) { idx 9051 net/wireless/nl80211.c key.p.key, key.p.key_len, key.idx, idx 216 net/wireless/pmsr.c int count, rem, err, idx; idx 265 net/wireless/pmsr.c idx = 0; idx 268 net/wireless/pmsr.c err = pmsr_parse_peer(rdev, peer, &req->peers[idx], info); idx 271 net/wireless/pmsr.c idx++; idx 216 net/wireless/rdev-ops.h struct net_device *dev, int idx, u8 *mac, idx 220 net/wireless/rdev-ops.h trace_rdev_dump_station(&rdev->wiphy, dev, idx, mac); idx 221 net/wireless/rdev-ops.h ret = rdev->ops->dump_station(&rdev->wiphy, dev, idx, mac, sinfo); idx 282 net/wireless/rdev-ops.h struct net_device *dev, int idx, u8 *dst, idx 287 net/wireless/rdev-ops.h trace_rdev_dump_mpath(&rdev->wiphy, dev, idx, dst, next_hop); idx 288 net/wireless/rdev-ops.h ret = rdev->ops->dump_mpath(&rdev->wiphy, dev, idx, dst, next_hop, idx 295 net/wireless/rdev-ops.h struct net_device *dev, int idx, u8 *dst, idx 301 net/wireless/rdev-ops.h trace_rdev_dump_mpp(&rdev->wiphy, dev, idx, dst, mpp); idx 302 net/wireless/rdev-ops.h ret = rdev->ops->dump_mpp(&rdev->wiphy, dev, idx, dst, mpp, pinfo); idx 652 net/wireless/rdev-ops.h struct net_device *netdev, int idx, idx 656 net/wireless/rdev-ops.h trace_rdev_dump_survey(&rdev->wiphy, netdev, idx); idx 657 net/wireless/rdev-ops.h ret = rdev->ops->dump_survey(&rdev->wiphy, netdev, idx, info); idx 1153 net/wireless/reg.c u32 start_freq, end_freq, idx, no; idx 1155 net/wireless/reg.c for (idx = 0; idx < rd->n_reg_rules; idx++) idx 1156 net/wireless/reg.c if (rule == &rd->reg_rules[idx]) idx 1159 net/wireless/reg.c if (idx == rd->n_reg_rules) idx 1163 net/wireless/reg.c no = idx; idx 1179 net/wireless/reg.c no = idx; idx 1206 net/wireless/sme.c int idx; idx 1209 net/wireless/sme.c idx = connkeys->def; idx 1210 net/wireless/sme.c cipher = connkeys->params[idx].cipher; idx 1214 net/wireless/sme.c connect->key_idx = idx; idx 1215 net/wireless/sme.c connect->key = connkeys->params[idx].key; idx 1216 net/wireless/sme.c connect->key_len = connkeys->params[idx].key_len; idx 824 net/wireless/trace.h __field(int, idx) idx 830 net/wireless/trace.h __entry->idx = _idx; idx 834 net/wireless/trace.h __entry->idx) idx 902 net/wireless/trace.h __field(int, idx) idx 909 net/wireless/trace.h __entry->idx = _idx; idx 913 net/wireless/trace.h WIPHY_PR_ARG, NETDEV_PR_ARG, __entry->idx, MAC_PR_ARG(dst), idx 947 net/wireless/trace.h __field(int, idx) idx 954 net/wireless/trace.h __entry->idx = _idx; idx 958 net/wireless/trace.h WIPHY_PR_ARG, NETDEV_PR_ARG, __entry->idx, MAC_PR_ARG(dst), idx 1725 net/wireless/trace.h __field(int, idx) idx 1730 net/wireless/trace.h __entry->idx = _idx; idx 1733 net/wireless/trace.h WIPHY_PR_ARG, NETDEV_PR_ARG, __entry->idx) idx 1183 net/wireless/util.c int idx; idx 1190 net/wireless/util.c idx = 3; idx 1193 net/wireless/util.c idx = 2; idx 1196 net/wireless/util.c idx = 1; idx 1203 net/wireless/util.c idx = 0; idx 1206 net/wireless/util.c bitrate = base[idx][rate->mcs]; idx 399 net/wireless/wext-compat.c int idx, struct key_params *params) idx 433 net/wireless/wext-compat.c if (idx < 4 || idx > 5) idx 435 net/wireless/wext-compat.c } else if (idx < 0 || idx > 3) idx 445 net/wireless/wext-compat.c if (idx == wdev->wext.default_key && idx 455 net/wireless/wext-compat.c err = rdev_del_key(rdev, dev, idx, pairwise, idx 466 net/wireless/wext-compat.c if (!addr && idx < 4) { idx 467 net/wireless/wext-compat.c memset(wdev->wext.keys->data[idx], 0, idx 468 net/wireless/wext-compat.c sizeof(wdev->wext.keys->data[idx])); idx 469 net/wireless/wext-compat.c wdev->wext.keys->params[idx].key_len = 0; idx 470 net/wireless/wext-compat.c wdev->wext.keys->params[idx].cipher = 0; idx 472 net/wireless/wext-compat.c if (idx == wdev->wext.default_key) idx 474 net/wireless/wext-compat.c else if (idx == wdev->wext.default_mgmt_key) idx 487 net/wireless/wext-compat.c if (cfg80211_validate_key_settings(rdev, params, idx, pairwise, addr)) idx 492 net/wireless/wext-compat.c err = rdev_add_key(rdev, dev, idx, pairwise, addr, params); idx 506 net/wireless/wext-compat.c wdev->wext.keys->params[idx] = *params; idx 507 net/wireless/wext-compat.c memcpy(wdev->wext.keys->data[idx], idx 509 net/wireless/wext-compat.c wdev->wext.keys->params[idx].key = idx 510 net/wireless/wext-compat.c wdev->wext.keys->data[idx]; idx 527 net/wireless/wext-compat.c err = rdev_set_default_key(rdev, dev, idx, true, true); idx 530 net/wireless/wext-compat.c wdev->wext.default_key = idx; idx 540 net/wireless/wext-compat.c err = rdev_set_default_mgmt_key(rdev, dev, idx); idx 542 net/wireless/wext-compat.c wdev->wext.default_mgmt_key = idx; idx 552 net/wireless/wext-compat.c int idx, struct key_params *params) idx 558 net/wireless/wext-compat.c remove, tx_key, idx, params); idx 570 net/wireless/wext-compat.c int idx, err; idx 584 net/wireless/wext-compat.c idx = erq->flags & IW_ENCODE_INDEX; idx 585 net/wireless/wext-compat.c if (idx == 0) { idx 586 net/wireless/wext-compat.c idx = wdev->wext.default_key; idx 587 net/wireless/wext-compat.c if (idx < 0) idx 588 net/wireless/wext-compat.c idx = 0; idx 589 net/wireless/wext-compat.c } else if (idx < 1 || idx > 4) idx 592 net/wireless/wext-compat.c idx--; idx 601 net/wireless/wext-compat.c err = rdev_set_default_key(rdev, dev, idx, true, idx 604 net/wireless/wext-compat.c wdev->wext.default_key = idx; idx 621 net/wireless/wext-compat.c idx, ¶ms); idx 632 net/wireless/wext-compat.c int idx; idx 676 net/wireless/wext-compat.c idx = erq->flags & IW_ENCODE_INDEX; idx 678 net/wireless/wext-compat.c if (idx < 4 || idx > 5) { idx 679 net/wireless/wext-compat.c idx = wdev->wext.default_mgmt_key; idx 680 net/wireless/wext-compat.c if (idx < 0) idx 683 net/wireless/wext-compat.c idx--; idx 685 net/wireless/wext-compat.c if (idx < 1 || idx > 4) { idx 686 net/wireless/wext-compat.c idx = wdev->wext.default_key; idx 687 net/wireless/wext-compat.c if (idx < 0) idx 690 net/wireless/wext-compat.c idx--; idx 712 net/wireless/wext-compat.c idx, ¶ms); idx 720 net/wireless/wext-compat.c int idx; idx 726 net/wireless/wext-compat.c idx = erq->flags & IW_ENCODE_INDEX; idx 727 net/wireless/wext-compat.c if (idx == 0) { idx 728 net/wireless/wext-compat.c idx = wdev->wext.default_key; idx 729 net/wireless/wext-compat.c if (idx < 0) idx 730 net/wireless/wext-compat.c idx = 0; idx 731 net/wireless/wext-compat.c } else if (idx < 1 || idx > 4) idx 734 net/wireless/wext-compat.c idx--; idx 736 net/wireless/wext-compat.c erq->flags = idx + 1; idx 738 net/wireless/wext-compat.c if (!wdev->wext.keys || !wdev->wext.keys->params[idx].cipher) { idx 745 net/wireless/wext-compat.c wdev->wext.keys->params[idx].key_len); idx 746 net/wireless/wext-compat.c memcpy(keybuf, wdev->wext.keys->params[idx].key, erq->length); idx 179 net/xdp/xsk_queue.h unsigned int idx = q->cons_tail & q->ring_mask; idx 181 net/xdp/xsk_queue.h *addr = READ_ONCE(ring->desc[idx]) & q->chunk_mask; idx 305 net/xdp/xsk_queue.h unsigned int idx = q->cons_tail & q->ring_mask; idx 307 net/xdp/xsk_queue.h *desc = READ_ONCE(ring->desc[idx]); idx 342 net/xdp/xsk_queue.h unsigned int idx; idx 348 net/xdp/xsk_queue.h idx = (q->prod_head++) & q->ring_mask; idx 349 net/xdp/xsk_queue.h ring->desc[idx].addr = addr; idx 350 net/xdp/xsk_queue.h ring->desc[idx].len = len; idx 753 net/xfrm/xfrm_algo.c struct xfrm_algo_desc *xfrm_aalg_get_byidx(unsigned int idx) idx 755 net/xfrm/xfrm_algo.c if (idx >= aalg_entries()) idx 758 net/xfrm/xfrm_algo.c return &aalg_list[idx]; idx 762 net/xfrm/xfrm_algo.c struct xfrm_algo_desc *xfrm_ealg_get_byidx(unsigned int idx) idx 764 net/xfrm/xfrm_algo.c if (idx >= ealg_entries()) idx 767 net/xfrm/xfrm_algo.c return &ealg_list[idx]; idx 1365 net/xfrm/xfrm_policy.c u32 idx; idx 1369 net/xfrm/xfrm_policy.c idx = (idx_generator | dir); idx 1372 net/xfrm/xfrm_policy.c idx = index; idx 1376 net/xfrm/xfrm_policy.c if (idx == 0) idx 1377 net/xfrm/xfrm_policy.c idx = 8; idx 1378 net/xfrm/xfrm_policy.c list = net->xfrm.policy_byidx + idx_hash(net, idx); idx 1381 net/xfrm/xfrm_policy.c if (p->index == idx) { idx 1387 net/xfrm/xfrm_policy.c return idx; idx 3197 net/xfrm/xfrm_policy.c xfrm_secpath_reject(int idx, struct sk_buff *skb, const struct flowi *fl) idx 3202 net/xfrm/xfrm_policy.c if (!sp || idx < 0 || idx >= sp->len) idx 3204 net/xfrm/xfrm_policy.c x = sp->xvec[idx]; idx 3243 net/xfrm/xfrm_policy.c int idx = start; idx 3250 net/xfrm/xfrm_policy.c for (; idx < sp->len; idx++) { idx 3251 net/xfrm/xfrm_policy.c if (xfrm_state_ok(tmpl, sp->xvec[idx], family)) idx 3252 net/xfrm/xfrm_policy.c return ++idx; idx 3253 net/xfrm/xfrm_policy.c if (sp->xvec[idx]->props.mode != XFRM_MODE_TRANSPORT) { idx 3255 net/xfrm/xfrm_policy.c start = -2-idx; idx 27 samples/bpf/bpf_load.h typedef void (*fixup_map_cb)(struct bpf_map_data *map, int idx); idx 395 samples/bpf/map_perf_test_user.c static void fixup_map(struct bpf_map_data *map, int idx) idx 400 samples/bpf/map_perf_test_user.c inner_lru_hash_idx = idx; idx 410 samples/bpf/map_perf_test_user.c array_of_lru_hashs_idx = idx; idx 414 samples/bpf/map_perf_test_user.c lru_hash_lookup_idx = idx; idx 30 samples/bpf/test_cgrp2_sock.c static int prog_load(__u32 idx, __u32 mark, __u32 prio) idx 43 samples/bpf/test_cgrp2_sock.c BPF_MOV64_IMM(BPF_REG_3, idx), idx 80 samples/bpf/test_cgrp2_sock.c if (idx) idx 98 samples/bpf/test_cgrp2_sock.c if (idx) { idx 206 samples/bpf/test_cgrp2_sock.c __u32 idx = 0, mark = 0, prio = 0; idx 219 samples/bpf/test_cgrp2_sock.c idx = if_nametoindex(optarg); idx 220 samples/bpf/test_cgrp2_sock.c if (!idx) { idx 221 samples/bpf/test_cgrp2_sock.c idx = strtoumax(optarg, NULL, 0); idx 222 samples/bpf/test_cgrp2_sock.c if (!idx) { idx 251 samples/bpf/test_cgrp2_sock.c if (do_attach && !idx && !mark && !prio) { idx 264 samples/bpf/test_cgrp2_sock.c prog_fd = prog_load(idx, mark, prio); idx 33 samples/bpf/test_current_task_under_cgroup_kern.c int idx = 0; idx 38 samples/bpf/test_current_task_under_cgroup_kern.c bpf_map_update_elem(&perf_map, &idx, &pid, BPF_ANY); idx 18 samples/bpf/test_current_task_under_cgroup_user.c int cg2, idx = 0, rc = 0; idx 35 samples/bpf/test_current_task_under_cgroup_user.c if (bpf_map_update_elem(map_fd[0], &idx, &cg2, BPF_ANY)) { idx 49 samples/bpf/test_current_task_under_cgroup_user.c bpf_map_lookup_elem(map_fd[1], &idx, &remote_pid); idx 63 samples/bpf/test_current_task_under_cgroup_user.c bpf_map_update_elem(map_fd[1], &idx, &remote_pid, BPF_ANY); idx 66 samples/bpf/test_current_task_under_cgroup_user.c bpf_map_lookup_elem(map_fd[1], &idx, &remote_pid); idx 30 samples/bpf/xdp_fwd_user.c static int do_attach(int idx, int prog_fd, int map_fd, const char *name) idx 34 samples/bpf/xdp_fwd_user.c err = bpf_set_link_xdp_fd(idx, prog_fd, 0); idx 41 samples/bpf/xdp_fwd_user.c err = bpf_map_update_elem(map_fd, &idx, &idx, 0); idx 48 samples/bpf/xdp_fwd_user.c static int do_detach(int idx, const char *name) idx 52 samples/bpf/xdp_fwd_user.c err = bpf_set_link_xdp_fd(idx, -1, 0); idx 82 samples/bpf/xdp_fwd_user.c int opt, i, idx, err; idx 139 samples/bpf/xdp_fwd_user.c idx = if_nametoindex(argv[i]); idx 140 samples/bpf/xdp_fwd_user.c if (!idx) idx 141 samples/bpf/xdp_fwd_user.c idx = strtoul(argv[i], NULL, 0); idx 143 samples/bpf/xdp_fwd_user.c if (!idx) { idx 148 samples/bpf/xdp_fwd_user.c err = do_detach(idx, argv[i]); idx 152 samples/bpf/xdp_fwd_user.c err = do_attach(idx, prog_fd, map_fd, argv[i]); idx 28 samples/bpf/xdp_sample_pkts_user.c static int do_attach(int idx, int fd, const char *name) idx 34 samples/bpf/xdp_sample_pkts_user.c err = bpf_set_link_xdp_fd(idx, fd, xdp_flags); idx 50 samples/bpf/xdp_sample_pkts_user.c static int do_detach(int idx, const char *name) idx 55 samples/bpf/xdp_sample_pkts_user.c err = bpf_get_link_xdp_id(idx, &curr_prog_id, 0); idx 61 samples/bpf/xdp_sample_pkts_user.c err = bpf_set_link_xdp_fd(idx, -1, 0); idx 315 samples/bpf/xdpsock_user.c u32 idx; idx 339 samples/bpf/xdpsock_user.c &idx); idx 343 samples/bpf/xdpsock_user.c *xsk_ring_prod__fill_addr(&xsk->umem->fq, idx++) = idx 532 samples/bpf/xdpsock_user.c u32 idx; idx 540 samples/bpf/xdpsock_user.c rcvd = xsk_ring_cons__peek(&xsk->umem->cq, BATCH_SIZE, &idx); idx 613 samples/bpf/xdpsock_user.c u32 idx; idx 615 samples/bpf/xdpsock_user.c if (xsk_ring_prod__reserve(&xsk->tx, BATCH_SIZE, &idx) == BATCH_SIZE) { idx 619 samples/bpf/xdpsock_user.c xsk_ring_prod__tx_desc(&xsk->tx, idx + i)->addr = idx 621 samples/bpf/xdpsock_user.c xsk_ring_prod__tx_desc(&xsk->tx, idx + i)->len = idx 28 samples/connector/cn_test.c __func__, jiffies, msg->id.idx, msg->id.val, idx 69 samples/connector/cn_test.c msg->id.idx = -1; idx 87 samples/connector/cn_test.c req->first = cn_test_id.idx; idx 162 samples/connector/cn_test.c cn_test_id.idx, cn_test_id.val); idx 66 samples/connector/ucon.c __func__, msg->id.idx, msg->id.val, msg->len, msg->seq, msg->ack); idx 170 samples/connector/ucon.c data->id.idx = CN_TEST_IDX; idx 181 samples/connector/ucon.c ulog("%d messages have been sent to %08x.%08x.\n", i, data->id.idx, data->id.val); idx 226 samples/connector/ucon.c ctime(&tm), data->id.idx, data->id.val, data->seq, data->ack); idx 515 samples/mic/mpssd/mpssd.c while (avail_idx == le16toh(READ_ONCE(vr->vr.avail->idx))) { idx 519 samples/mic/mpssd/mpssd.c le16toh(vr->vr.avail->idx), vr->info->avail_idx); idx 657 samples/mic/mpssd/mpssd.c le16toh(rx_vr.vr.avail->idx)) { idx 856 samples/mic/mpssd/mpssd.c le16toh(rx_vr.vr.avail->idx)) { idx 1216 samples/mic/mpssd/mpssd.c le16toh(vring.vr.avail->idx)) { idx 66 samples/seccomp/bpf-helper.h #define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) idx 68 samples/seccomp/bpf-helper.h #define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32) idx 90 samples/seccomp/bpf-helper.h #define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32) idx 93 samples/seccomp/bpf-helper.h #define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) idx 139 samples/seccomp/bpf-helper.h #define ARG_32(idx) \ idx 140 samples/seccomp/bpf-helper.h BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx)) idx 143 samples/seccomp/bpf-helper.h #define ARG_64(idx) \ idx 144 samples/seccomp/bpf-helper.h BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx)), \ idx 146 samples/seccomp/bpf-helper.h BPF_STMT(BPF_LD+BPF_W+BPF_ABS, HI_ARG(idx)), \ idx 56 samples/vfio-mdev/mtty.c #define CIRCULAR_BUF_INC_IDX(idx) (idx = (idx + 1) & (MAX_FIFO_SIZE - 1)) idx 726 scripts/dtc/libfdt/fdt_ro.c int length, len, idx = 0; idx 744 scripts/dtc/libfdt/fdt_ro.c return idx; idx 747 scripts/dtc/libfdt/fdt_ro.c idx++; idx 754 scripts/dtc/libfdt/fdt_ro.c const char *property, int idx, idx 781 scripts/dtc/libfdt/fdt_ro.c if (idx == 0) { idx 789 scripts/dtc/libfdt/fdt_ro.c idx--; idx 15 scripts/dtc/libfdt/fdt_wip.c uint32_t idx, const void *val, idx 26 scripts/dtc/libfdt/fdt_wip.c if (proplen < (len + idx)) idx 29 scripts/dtc/libfdt/fdt_wip.c memcpy((char *)propval + idx, val, len); idx 1219 scripts/dtc/libfdt/libfdt.h uint32_t idx, const void *val, idx 620 scripts/gcc-plugins/randomize_layout_plugin.c unsigned HOST_WIDE_INT idx; idx 624 scripts/gcc-plugins/randomize_layout_plugin.c FOR_EACH_CONSTRUCTOR_ELT(CONSTRUCTOR_ELTS(init), idx, field, val) { idx 512 scripts/kallsyms.c static void compress_symbols(unsigned char *str, int idx) idx 532 scripts/kallsyms.c *p2 = idx; idx 775 scripts/kconfig/qconf.cc int idx, x; idx 782 scripts/kconfig/qconf.cc idx = header()->logicalIndexAt(x); idx 783 scripts/kconfig/qconf.cc switch (idx) { idx 92 scripts/kconfig/qconf.h void addColumn(colIdx idx) idx 94 scripts/kconfig/qconf.h showColumn(idx); idx 96 scripts/kconfig/qconf.h void removeColumn(colIdx idx) idx 98 scripts/kconfig/qconf.h hideColumn(idx); idx 168 scripts/kconfig/qconf.h void setText(colIdx idx, const QString& text) idx 170 scripts/kconfig/qconf.h Parent::setText(idx, text); idx 172 scripts/kconfig/qconf.h QString text(colIdx idx) const idx 174 scripts/kconfig/qconf.h return Parent::text(idx); idx 176 scripts/kconfig/qconf.h void setPixmap(colIdx idx, const QIcon &icon) idx 178 scripts/kconfig/qconf.h Parent::setIcon(idx, icon); idx 180 scripts/kconfig/qconf.h const QIcon pixmap(colIdx idx) const idx 182 scripts/kconfig/qconf.h return icon(idx); idx 102 scripts/recordmcount.c off_t idx = 0; idx 123 scripts/recordmcount.c idx = aoffset - count; idx 131 scripts/recordmcount.c memcpy(file_append + idx, buf + cnt, count - cnt); idx 111 scripts/sortextable.h int idx; idx 128 scripts/sortextable.h idx = r(&shdr[i].sh_name); idx 129 scripts/sortextable.h if (strcmp(secstrtab + idx, "__ex_table") == 0) { idx 139 scripts/sortextable.h if (strcmp(secstrtab + idx, ".symtab") == 0) idx 141 scripts/sortextable.h if (strcmp(secstrtab + idx, ".strtab") == 0) idx 183 scripts/sortextable.h idx = r(&sym->st_name); idx 184 scripts/sortextable.h if (strcmp(strtab + idx, "main_extable_sort_needed") == 0) { idx 238 security/device_cgroup.c int idx = 0; idx 241 security/device_cgroup.c acc[idx++] = 'r'; idx 243 security/device_cgroup.c acc[idx++] = 'w'; idx 245 security/device_cgroup.c acc[idx++] = 'm'; idx 648 security/integrity/ima/ima_crypto.c static void __init ima_pcrread(u32 idx, struct tpm_digest *d) idx 653 security/integrity/ima/ima_crypto.c if (tpm_pcr_read(ima_tpm_chip, idx, d) != 0) idx 73 security/selinux/ibpkey.c unsigned int idx; idx 76 security/selinux/ibpkey.c idx = sel_ib_pkey_hashfn(pkey_num); idx 77 security/selinux/ibpkey.c list_for_each_entry_rcu(pkey, &sel_ib_pkey_hash[idx].list, list) { idx 96 security/selinux/ibpkey.c unsigned int idx; idx 101 security/selinux/ibpkey.c idx = sel_ib_pkey_hashfn(pkey->psec.pkey); idx 102 security/selinux/ibpkey.c list_add_rcu(&pkey->list, &sel_ib_pkey_hash[idx].list); idx 103 security/selinux/ibpkey.c if (sel_ib_pkey_hash[idx].size == SEL_PKEY_HASH_BKT_LIMIT) { idx 108 security/selinux/ibpkey.c sel_ib_pkey_hash[idx].list.prev, idx 114 security/selinux/ibpkey.c sel_ib_pkey_hash[idx].size++; idx 205 security/selinux/ibpkey.c unsigned int idx; idx 210 security/selinux/ibpkey.c for (idx = 0; idx < SEL_PKEY_HASH_SIZE; idx++) { idx 212 security/selinux/ibpkey.c &sel_ib_pkey_hash[idx].list, list) { idx 216 security/selinux/ibpkey.c sel_ib_pkey_hash[idx].size = 0; idx 71 security/selinux/netif.c int idx = sel_netif_hashfn(ns, ifindex); idx 74 security/selinux/netif.c list_for_each_entry_rcu(netif, &sel_netif_hash[idx], list) idx 93 security/selinux/netif.c int idx; idx 98 security/selinux/netif.c idx = sel_netif_hashfn(netif->nsec.ns, netif->nsec.ifindex); idx 99 security/selinux/netif.c list_add_rcu(&netif->list, &sel_netif_hash[idx]); idx 240 security/selinux/netif.c int idx; idx 244 security/selinux/netif.c for (idx = 0; idx < SEL_NETIF_HASH_SIZE; idx++) idx 245 security/selinux/netif.c list_for_each_entry(netif, &sel_netif_hash[idx], list) idx 106 security/selinux/netnode.c unsigned int idx; idx 111 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv4(*(__be32 *)addr); idx 114 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv6(addr); idx 121 security/selinux/netnode.c list_for_each_entry_rcu(node, &sel_netnode_hash[idx].list, list) idx 148 security/selinux/netnode.c unsigned int idx; idx 152 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv4(node->nsec.addr.ipv4); idx 155 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv6(&node->nsec.addr.ipv6); idx 164 security/selinux/netnode.c list_add_rcu(&node->list, &sel_netnode_hash[idx].list); idx 165 security/selinux/netnode.c if (sel_netnode_hash[idx].size == SEL_NETNODE_HASH_BKT_LIMIT) { idx 168 security/selinux/netnode.c rcu_dereference_protected(sel_netnode_hash[idx].list.prev, idx 174 security/selinux/netnode.c sel_netnode_hash[idx].size++; idx 275 security/selinux/netnode.c unsigned int idx; idx 279 security/selinux/netnode.c for (idx = 0; idx < SEL_NETNODE_HASH_SIZE; idx++) { idx 281 security/selinux/netnode.c &sel_netnode_hash[idx].list, list) { idx 285 security/selinux/netnode.c sel_netnode_hash[idx].size = 0; idx 86 security/selinux/netport.c unsigned int idx; idx 89 security/selinux/netport.c idx = sel_netport_hashfn(pnum); idx 90 security/selinux/netport.c list_for_each_entry_rcu(port, &sel_netport_hash[idx].list, list) idx 107 security/selinux/netport.c unsigned int idx; idx 111 security/selinux/netport.c idx = sel_netport_hashfn(port->psec.port); idx 112 security/selinux/netport.c list_add_rcu(&port->list, &sel_netport_hash[idx].list); idx 113 security/selinux/netport.c if (sel_netport_hash[idx].size == SEL_NETPORT_HASH_BKT_LIMIT) { idx 117 security/selinux/netport.c sel_netport_hash[idx].list.prev, idx 123 security/selinux/netport.c sel_netport_hash[idx].size++; idx 209 security/selinux/netport.c unsigned int idx; idx 213 security/selinux/netport.c for (idx = 0; idx < SEL_NETPORT_HASH_SIZE; idx++) { idx 215 security/selinux/netport.c &sel_netport_hash[idx].list, list) { idx 219 security/selinux/netport.c sel_netport_hash[idx].size = 0; idx 1501 security/selinux/selinuxfs.c static struct avc_cache_stats *sel_avc_get_stat_idx(loff_t *idx) idx 1505 security/selinux/selinuxfs.c for (cpu = *idx; cpu < nr_cpu_ids; ++cpu) { idx 1508 security/selinux/selinuxfs.c *idx = cpu + 1; idx 149 security/selinux/ss/ebitmap.c u32 offset = 0, idx; idx 180 security/selinux/ss/ebitmap.c idx = EBITMAP_NODE_INDEX(e_iter, offset); idx 181 security/selinux/ss/ebitmap.c e_iter->maps[idx] = bitmap; idx 349 security/tomoyo/common.c unsigned int idx; idx 351 security/tomoyo/common.c for (idx = 0; idx < TOMOYO_MAX_ACL_GROUPS; idx++) idx 352 security/tomoyo/common.c INIT_LIST_HEAD(&ns->acl_group[idx]); idx 353 security/tomoyo/common.c for (idx = 0; idx < TOMOYO_MAX_GROUP; idx++) idx 354 security/tomoyo/common.c INIT_LIST_HEAD(&ns->group_list[idx]); idx 355 security/tomoyo/common.c for (idx = 0; idx < TOMOYO_MAX_POLICY; idx++) idx 356 security/tomoyo/common.c INIT_LIST_HEAD(&ns->policy_list[idx]); idx 1179 security/tomoyo/common.c unsigned int idx; idx 1197 security/tomoyo/common.c if (sscanf(data, "use_profile %u", &idx) == 1 idx 1198 security/tomoyo/common.c && idx < TOMOYO_MAX_PROFILES) { idx 1199 security/tomoyo/common.c if (!tomoyo_policy_loaded || ns->profile_ptr[idx]) idx 1201 security/tomoyo/common.c domain->profile = (u8) idx; idx 1204 security/tomoyo/common.c if (sscanf(data, "use_group %u\n", &idx) == 1 idx 1205 security/tomoyo/common.c && idx < TOMOYO_MAX_ACL_GROUPS) { idx 1207 security/tomoyo/common.c set_bit(idx, domain->group); idx 1209 security/tomoyo/common.c clear_bit(idx, domain->group); idx 1212 security/tomoyo/common.c for (idx = 0; idx < TOMOYO_MAX_DOMAIN_INFO_FLAGS; idx++) { idx 1213 security/tomoyo/common.c const char *cp = tomoyo_dif[idx]; idx 1217 security/tomoyo/common.c domain->flags[idx] = !is_delete; idx 1797 security/tomoyo/common.c static bool tomoyo_read_group(struct tomoyo_io_buffer *head, const int idx) idx 1801 security/tomoyo/common.c struct list_head *list = &ns->group_list[idx]; idx 1816 security/tomoyo/common.c tomoyo_set_string(head, tomoyo_group_name[idx]); idx 1818 security/tomoyo/common.c if (idx == TOMOYO_PATH_GROUP) { idx 1823 security/tomoyo/common.c } else if (idx == TOMOYO_NUMBER_GROUP) { idx 1828 security/tomoyo/common.c } else if (idx == TOMOYO_ADDRESS_GROUP) { idx 1856 security/tomoyo/common.c static bool tomoyo_read_policy(struct tomoyo_io_buffer *head, const int idx) idx 1860 security/tomoyo/common.c struct list_head *list = &ns->policy_list[idx]; idx 1869 security/tomoyo/common.c switch (idx) { idx 2586 security/tomoyo/common.c int idx; idx 2594 security/tomoyo/common.c idx = tomoyo_read_lock(); idx 2602 security/tomoyo/common.c tomoyo_read_unlock(idx); idx 2661 security/tomoyo/common.c int idx; idx 2670 security/tomoyo/common.c idx = tomoyo_read_lock(); idx 2748 security/tomoyo/common.c tomoyo_read_unlock(idx); idx 2776 security/tomoyo/common.c const int idx = tomoyo_read_lock(); idx 2802 security/tomoyo/common.c tomoyo_read_unlock(idx); idx 2832 security/tomoyo/common.c const int idx = tomoyo_read_lock(); idx 2877 security/tomoyo/common.c tomoyo_read_unlock(idx); idx 1036 security/tomoyo/common.h const u8 idx); idx 1121 security/tomoyo/common.h static inline void tomoyo_read_unlock(int idx) idx 1123 security/tomoyo/common.h srcu_read_unlock(&tomoyo_ss, idx); idx 717 security/tomoyo/file.c int idx; idx 722 security/tomoyo/file.c idx = tomoyo_read_lock(); idx 738 security/tomoyo/file.c tomoyo_read_unlock(idx); idx 763 security/tomoyo/file.c int idx; idx 767 security/tomoyo/file.c idx = tomoyo_read_lock(); idx 787 security/tomoyo/file.c tomoyo_read_unlock(idx); idx 813 security/tomoyo/file.c int idx; idx 821 security/tomoyo/file.c idx = tomoyo_read_lock(); idx 843 security/tomoyo/file.c tomoyo_read_unlock(idx); idx 868 security/tomoyo/file.c int idx; idx 873 security/tomoyo/file.c idx = tomoyo_read_lock(); idx 888 security/tomoyo/file.c tomoyo_read_unlock(idx); idx 914 security/tomoyo/file.c int idx; idx 921 security/tomoyo/file.c idx = tomoyo_read_lock(); idx 948 security/tomoyo/file.c tomoyo_read_unlock(idx); idx 96 security/tomoyo/memory.c const u8 idx) idx 104 security/tomoyo/memory.c if (!tomoyo_correct_word(group_name) || idx >= TOMOYO_MAX_GROUP) idx 111 security/tomoyo/memory.c list = ¶m->ns->group_list[idx]; idx 197 security/tomoyo/memory.c int idx; idx 199 security/tomoyo/memory.c for (idx = 0; idx < TOMOYO_MAX_HASH; idx++) idx 200 security/tomoyo/memory.c INIT_LIST_HEAD(&tomoyo_name_list[idx]); idx 197 security/tomoyo/mount.c int idx; idx 236 security/tomoyo/mount.c idx = tomoyo_read_lock(); idx 238 security/tomoyo/mount.c tomoyo_read_unlock(idx); idx 469 security/tomoyo/network.c const int idx = tomoyo_read_lock(); idx 487 security/tomoyo/network.c tomoyo_read_unlock(idx); idx 546 security/tomoyo/network.c const int idx = tomoyo_read_lock(); idx 580 security/tomoyo/network.c tomoyo_read_unlock(idx); idx 54 security/tomoyo/securityfs_if.c const int idx = tomoyo_read_lock(); idx 83 security/tomoyo/securityfs_if.c tomoyo_read_unlock(idx); idx 106 security/tomoyo/tomoyo.c const int idx = tomoyo_read_lock(); idx 109 security/tomoyo/tomoyo.c tomoyo_read_unlock(idx); idx 70 sound/ac97/bus.c ac97_of_get_child_device(struct ac97_controller *ac97_ctrl, int idx, idx 81 sound/ac97/bus.c if ((idx != of_property_read_u32(node, "reg", ®)) || idx 102 sound/ac97/bus.c static int ac97_codec_add(struct ac97_controller *ac97_ctrl, int idx, idx 111 sound/ac97/bus.c ac97_ctrl->codecs[idx] = codec; idx 116 sound/ac97/bus.c codec->num = idx; idx 120 sound/ac97/bus.c dev_set_name(&codec->dev, "%s:%u", dev_name(ac97_ctrl->parent), idx); idx 121 sound/ac97/bus.c codec->dev.of_node = ac97_of_get_child_device(ac97_ctrl, idx, idx 92 sound/aoa/codecs/tas-basstreble.h static inline u8 tas3004_treble(int idx) idx 94 sound/aoa/codecs/tas-basstreble.h return tas3004_treble_table[idx]; idx 128 sound/aoa/codecs/tas-basstreble.h static inline u8 tas3004_bass(int idx) idx 130 sound/aoa/codecs/tas-basstreble.h u8 result = tas3004_treble_table[idx]; idx 132 sound/aoa/codecs/tas-basstreble.h if (idx >= 50) idx 133 sound/aoa/codecs/tas-basstreble.h result += tas3004_bass_diff_to_treble[idx-50]; idx 339 sound/aoa/codecs/tas.c int idx = kcontrol->private_value; idx 342 sound/aoa/codecs/tas.c ucontrol->value.integer.value[0] = tas->mixer_l[idx]; idx 343 sound/aoa/codecs/tas.c ucontrol->value.integer.value[1] = tas->mixer_r[idx]; idx 353 sound/aoa/codecs/tas.c int idx = kcontrol->private_value; idx 356 sound/aoa/codecs/tas.c if (tas->mixer_l[idx] == ucontrol->value.integer.value[0] idx 357 sound/aoa/codecs/tas.c && tas->mixer_r[idx] == ucontrol->value.integer.value[1]) { idx 362 sound/aoa/codecs/tas.c tas->mixer_l[idx] = ucontrol->value.integer.value[0]; idx 363 sound/aoa/codecs/tas.c tas->mixer_r[idx] = ucontrol->value.integer.value[1]; idx 371 sound/aoa/codecs/tas.c #define MIXER_CONTROL(n,descr,idx) \ idx 379 sound/aoa/codecs/tas.c .private_value = idx, \ idx 109 sound/core/control.c unsigned int idx; idx 119 sound/core/control.c for (idx = 0; idx < control->count; idx++) idx 120 sound/core/control.c if (control->vd[idx].owner == ctl) idx 121 sound/core/control.c control->vd[idx].owner = NULL; idx 199 sound/core/control.c unsigned int idx; idx 208 sound/core/control.c for (idx = 0; idx < count; idx++) { idx 209 sound/core/control.c (*kctl)->vd[idx].access = access; idx 210 sound/core/control.c (*kctl)->vd[idx].owner = file; idx 342 sound/core/control.c unsigned int idx; idx 379 sound/core/control.c for (idx = 0; idx < count; idx++, id.index++, id.numid++) idx 464 sound/core/control.c unsigned int idx; idx 471 sound/core/control.c for (idx = 0; idx < kcontrol->count; idx++, id.index++, id.numid++) idx 520 sound/core/control.c int idx, ret; idx 532 sound/core/control.c for (idx = 0; idx < kctl->count; idx++) idx 533 sound/core/control.c if (kctl->vd[idx].owner != NULL && kctl->vd[idx].owner != file) { idx 53 sound/core/info_oss.c int idx, ok = -1; idx 58 sound/core/info_oss.c for (idx = 0; idx < SNDRV_CARDS; idx++) { idx 59 sound/core/info_oss.c str = snd_sndstat_strings[idx][dev]; idx 65 sound/core/info_oss.c snd_iprintf(buf, "%i: %s\n", idx, str); idx 48 sound/core/init.c static int module_slot_match(struct module *module, int idx) idx 54 sound/core/init.c if (!module || !*module->name || !slots[idx]) idx 58 sound/core/init.c s2 = slots[idx]; idx 160 sound/core/init.c int snd_card_new(struct device *parent, int idx, const char *xid, idx 182 sound/core/init.c if (idx < 0) /* first check the matching module-name slot */ idx 183 sound/core/init.c idx = get_slot_from_bitmask(idx, module_slot_match, module); idx 184 sound/core/init.c if (idx < 0) /* if not matched, assign an empty slot */ idx 185 sound/core/init.c idx = get_slot_from_bitmask(idx, check_empty_slot, module); idx 186 sound/core/init.c if (idx < 0) idx 188 sound/core/init.c else if (idx < snd_ecards_limit) { idx 189 sound/core/init.c if (test_bit(idx, snd_cards_lock)) idx 191 sound/core/init.c } else if (idx >= SNDRV_CARDS) idx 196 sound/core/init.c idx, snd_ecards_limit - 1, err); idx 200 sound/core/init.c set_bit(idx, snd_cards_lock); /* lock it */ idx 201 sound/core/init.c if (idx >= snd_ecards_limit) idx 202 sound/core/init.c snd_ecards_limit = idx + 1; /* increase the limit */ idx 205 sound/core/init.c card->number = idx; idx 225 sound/core/init.c err = kobject_set_name(&card->card_dev.kobj, "card%d", idx); idx 262 sound/core/init.c struct snd_card *snd_card_ref(int idx) idx 267 sound/core/init.c card = snd_cards[idx]; idx 662 sound/core/init.c size_t idx; idx 665 sound/core/init.c for (idx = 0; idx < copy; idx++) { idx 666 sound/core/init.c c = buf[idx]; idx 792 sound/core/init.c int idx, count; idx 795 sound/core/init.c for (idx = count = 0; idx < SNDRV_CARDS; idx++) { idx 797 sound/core/init.c if ((card = snd_cards[idx]) != NULL) { idx 800 sound/core/init.c idx, idx 816 sound/core/init.c int idx, count; idx 819 sound/core/init.c for (idx = count = 0; idx < SNDRV_CARDS; idx++) { idx 821 sound/core/init.c if ((card = snd_cards[idx]) != NULL) { idx 838 sound/core/init.c int idx; idx 841 sound/core/init.c for (idx = 0; idx < SNDRV_CARDS; idx++) { idx 843 sound/core/init.c if ((card = snd_cards[idx]) != NULL) idx 845 sound/core/init.c idx, card->module->name); idx 798 sound/core/oss/mixer_oss.c int err, idx; idx 816 sound/core/oss/mixer_oss.c for (idx = 0; idx < 32; idx++) { idx 817 sound/core/oss/mixer_oss.c if (!(mixer->mask_recsrc & (1 << idx))) idx 819 sound/core/oss/mixer_oss.c pslot = &mixer->slots[idx]; idx 826 sound/core/oss/mixer_oss.c *active_index = idx; idx 849 sound/core/oss/mixer_oss.c unsigned int idx; idx 865 sound/core/oss/mixer_oss.c for (idx = 0; idx < 32; idx++) { idx 866 sound/core/oss/mixer_oss.c if (!(mixer->mask_recsrc & (1 << idx))) idx 868 sound/core/oss/mixer_oss.c pslot = &mixer->slots[idx]; idx 874 sound/core/oss/mixer_oss.c if (idx == active_index) idx 880 sound/core/oss/mixer_oss.c for (idx = 0; idx < uinfo->count; idx++) idx 881 sound/core/oss/mixer_oss.c uctl->value.enumerated.item[idx] = slot->capture_item; idx 946 sound/core/oss/mixer_oss.c int idx = rslot->number; /* remember this */ idx 950 sound/core/oss/mixer_oss.c rslot->number = idx; idx 1172 sound/core/oss/mixer_oss.c unsigned int idx; idx 1196 sound/core/oss/mixer_oss.c idx = simple_strtoul(idxstr, NULL, 10); idx 1197 sound/core/oss/mixer_oss.c if (idx >= 0x4000) { /* too big */ idx 1198 sound/core/oss/mixer_oss.c pr_err("ALSA: mixer_oss: invalid index %d\n", idx); idx 1204 sound/core/oss/mixer_oss.c slot->assigned->index == idx && ! strcmp(slot->assigned->name, str)) idx 1216 sound/core/oss/mixer_oss.c tbl->index = idx; idx 1298 sound/core/oss/mixer_oss.c unsigned int idx; idx 1300 sound/core/oss/mixer_oss.c for (idx = 0; idx < ARRAY_SIZE(table); idx++) idx 1301 sound/core/oss/mixer_oss.c snd_mixer_oss_build_input(mixer, &table[idx], 0, 0); idx 1316 sound/core/oss/mixer_oss.c int idx; idx 1324 sound/core/oss/mixer_oss.c for (idx = 0; idx < SNDRV_OSS_MAX_MIXERS; idx++) { idx 1325 sound/core/oss/mixer_oss.c struct snd_mixer_oss_slot *chn = &mixer->slots[idx]; idx 1338 sound/core/oss/mixer_oss.c int idx, err; idx 1365 sound/core/oss/mixer_oss.c for (idx = 0; idx < SNDRV_OSS_MAX_MIXERS; idx++) idx 1366 sound/core/oss/mixer_oss.c mixer->slots[idx].number = idx; idx 1392 sound/core/oss/mixer_oss.c int idx; idx 1395 sound/core/oss/mixer_oss.c for (idx = 0; idx < SNDRV_CARDS; idx++) { idx 1396 sound/core/oss/mixer_oss.c card = snd_card_ref(idx); idx 1408 sound/core/oss/mixer_oss.c int idx; idx 1411 sound/core/oss/mixer_oss.c for (idx = 0; idx < SNDRV_CARDS; idx++) { idx 1412 sound/core/oss/mixer_oss.c card = snd_card_ref(idx); idx 1094 sound/core/oss/pcm_oss.c int idx, err; idx 1097 sound/core/oss/pcm_oss.c for (idx = 0; idx < 2; idx++) { idx 1098 sound/core/oss/pcm_oss.c substream = pcm_oss_file->streams[idx]; idx 1723 sound/core/oss/pcm_oss.c int idx; idx 1725 sound/core/oss/pcm_oss.c for (idx = 1; idx >= 0; --idx) { idx 1726 sound/core/oss/pcm_oss.c struct snd_pcm_substream *substream = pcm_oss_file->streams[idx]; idx 1761 sound/core/oss/pcm_oss.c int idx; idx 1766 sound/core/oss/pcm_oss.c for (idx = 1; idx >= 0; --idx) { idx 1767 sound/core/oss/pcm_oss.c struct snd_pcm_substream *substream = pcm_oss_file->streams[idx]; idx 1853 sound/core/oss/pcm_oss.c int formats, idx; idx 1862 sound/core/oss/pcm_oss.c for (idx = 1; idx >= 0; --idx) { idx 1863 sound/core/oss/pcm_oss.c struct snd_pcm_substream *substream = pcm_oss_file->streams[idx]; idx 1914 sound/core/oss/pcm_oss.c int err = -EINVAL, idx; idx 1916 sound/core/oss/pcm_oss.c for (idx = 1; idx >= 0; --idx) { idx 1917 sound/core/oss/pcm_oss.c struct snd_pcm_substream *substream = pcm_oss_file->streams[idx]; idx 1953 sound/core/oss/pcm_oss.c int err = -EINVAL, idx; idx 1955 sound/core/oss/pcm_oss.c for (idx = 1; idx >= 0; --idx) { idx 1956 sound/core/oss/pcm_oss.c struct snd_pcm_substream *substream = pcm_oss_file->streams[idx]; idx 2007 sound/core/oss/pcm_oss.c int result, idx; idx 2010 sound/core/oss/pcm_oss.c for (idx = 0; idx < 2; idx++) { idx 2011 sound/core/oss/pcm_oss.c struct snd_pcm_substream *substream = pcm_oss_file->streams[idx]; idx 2380 sound/core/oss/pcm_oss.c int idx, err; idx 2397 sound/core/oss/pcm_oss.c for (idx = 0; idx < 2; idx++) { idx 2398 sound/core/oss/pcm_oss.c if (setup[idx].disable) idx 2400 sound/core/oss/pcm_oss.c if (! pcm->streams[idx].substream_count) idx 2402 sound/core/oss/pcm_oss.c if (idx == SNDRV_PCM_STREAM_PLAYBACK) { idx 2409 sound/core/oss/pcm_oss.c err = snd_pcm_open_substream(pcm, idx, file, &substream); idx 2415 sound/core/oss/pcm_oss.c pcm_oss_file->streams[idx] = substream; idx 2416 sound/core/oss/pcm_oss.c snd_pcm_oss_init_substream(substream, &setup[idx], minor); idx 2433 sound/core/oss/pcm_oss.c unsigned int idx; idx 2437 sound/core/oss/pcm_oss.c for (idx = 0; idx < sizeof(task->comm) && idx + 1 < size; idx++) idx 2438 sound/core/oss/pcm_oss.c name[idx] = task->comm[idx]; idx 2439 sound/core/oss/pcm_oss.c name[idx] = '\0'; idx 2569 sound/core/oss/pcm_oss.c int idx; idx 2570 sound/core/oss/pcm_oss.c for (idx = 0; idx < 2; ++idx) { idx 2571 sound/core/oss/pcm_oss.c substream = pcm_oss_file->streams[idx]; idx 2575 sound/core/oss/pcm_oss.c if (snd_BUG_ON(idx >= 2)) idx 638 sound/core/pcm.c int idx, err; idx 665 sound/core/pcm.c for (idx = 0, prev = NULL; idx < substream_count; idx++) { idx 671 sound/core/pcm.c substream->number = idx; idx 673 sound/core/pcm.c sprintf(substream->name, "subdevice #%i", idx); idx 2326 sound/core/pcm_lib.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2332 sound/core/pcm_lib.c substream = snd_pcm_chmap_substream(info, idx); idx 268 sound/core/pcm_memory.c unsigned int idx = offset >> PAGE_SHIFT; idx 269 sound/core/pcm_memory.c if (idx >= (unsigned int)sgbuf->pages) idx 271 sound/core/pcm_memory.c return sgbuf->page_table[idx]; idx 1532 sound/core/rawmidi.c int idx; idx 1534 sound/core/rawmidi.c for (idx = 0; idx < count; idx++) { idx 1539 sound/core/rawmidi.c substream->number = idx; idx 127 sound/core/seq/seq_clientmgr.c int idx; idx 131 sound/core/seq/seq_clientmgr.c for (idx = 0; idx < 15; idx++) { idx 132 sound/core/seq/seq_clientmgr.c if (seq_client_load[idx] < 0) idx 134 sound/core/seq/seq_clientmgr.c if (seq_client_load[idx] == clientid) { idx 106 sound/core/seq/seq_dummy.c create_port(int idx, int type) idx 121 sound/core/seq/seq_dummy.c sprintf(pinfo.name, "Midi Through Port-%d:%c", idx, idx 124 sound/core/seq/seq_dummy.c sprintf(pinfo.name, "Midi Through Port-%d", idx); idx 387 sound/core/seq/seq_midi_event.c int idx = 0; idx 398 sound/core/seq/seq_midi_event.c buf[idx++] = dev->lastcmd = cmd; idx 400 sound/core/seq/seq_midi_event.c buf[idx++] = ev->data.control.param; idx 401 sound/core/seq/seq_midi_event.c buf[idx++] = (ev->data.control.value >> 7) & 0x7f; idx 403 sound/core/seq/seq_midi_event.c buf[idx++] = cmd; idx 404 sound/core/seq/seq_midi_event.c buf[idx++] = ev->data.control.param + 0x20; idx 405 sound/core/seq/seq_midi_event.c buf[idx++] = ev->data.control.value & 0x7f; idx 412 sound/core/seq/seq_midi_event.c buf[idx++] = dev->lastcmd = cmd; idx 414 sound/core/seq/seq_midi_event.c buf[idx++] = ev->data.control.param & 0x7f; idx 415 sound/core/seq/seq_midi_event.c buf[idx++] = ev->data.control.value & 0x7f; idx 417 sound/core/seq/seq_midi_event.c return idx; idx 435 sound/core/seq/seq_midi_event.c int idx = 0, i; idx 449 sound/core/seq/seq_midi_event.c buf[idx++] = dev->lastcmd = cmd; idx 454 sound/core/seq/seq_midi_event.c buf[idx++] = dev->lastcmd = cmd; idx 455 sound/core/seq/seq_midi_event.c buf[idx++] = cbytes[i]; idx 456 sound/core/seq/seq_midi_event.c buf[idx++] = bytes[i]; idx 458 sound/core/seq/seq_midi_event.c return idx; idx 465 sound/core/seq/seq_timer.c int idx; idx 471 sound/core/seq/seq_timer.c for (idx = 0; idx < SNDRV_SEQ_MAX_QUEUES; idx++) { idx 472 sound/core/seq/seq_timer.c q = queueptr(idx); idx 1016 sound/drivers/aloop.c int err, dev, substr, substr_count, idx; idx 1030 sound/drivers/aloop.c for (idx = 0; idx < ARRAY_SIZE(loopback_controls); idx 1031 sound/drivers/aloop.c idx++) { idx 1032 sound/drivers/aloop.c kctl = snd_ctl_new1(&loopback_controls[idx], idx 1038 sound/drivers/aloop.c switch (idx) { idx 888 sound/drivers/dummy.c unsigned int idx; idx 895 sound/drivers/dummy.c for (idx = 0; idx < ARRAY_SIZE(snd_dummy_controls); idx++) { idx 896 sound/drivers/dummy.c kcontrol = snd_ctl_new1(&snd_dummy_controls[idx], dummy); idx 1038 sound/drivers/dummy.c int idx, err; idx 1056 sound/drivers/dummy.c for (idx = 0; idx < MAX_PCM_DEVICES && idx < pcm_devs[dev]; idx++) { idx 1061 sound/drivers/dummy.c err = snd_card_dummy_pcm(dummy, idx, pcm_substreams[dev]); idx 140 sound/drivers/mts64.c u8 idx); idx 354 sound/drivers/mts64.c u8 idx) idx 369 sound/drivers/mts64.c mts64_write_command(p, fps[idx]); idx 516 sound/drivers/mts64.c int idx = kctl->private_value; idx 519 sound/drivers/mts64.c uctl->value.integer.value[0] = mts->time[idx]; idx 529 sound/drivers/mts64.c int idx = kctl->private_value; idx 534 sound/drivers/mts64.c if (mts->time[idx] != time) { idx 536 sound/drivers/mts64.c mts->time[idx] = time; idx 83 sound/drivers/opl3/opl3_midi.c int idx = (note % 12) + 2; idx 96 sound/drivers/opl3/opl3_midi.c freq = opl3_note_table[idx+segment]; idx 97 sound/drivers/opl3/opl3_midi.c freq += ((opl3_note_table[idx+segment+1] - freq) * idx 100 sound/drivers/opl3/opl3_midi.c freq = opl3_note_table[idx]; idx 40 sound/drivers/opl3/opl3_seq.c int idx; idx 53 sound/drivers/opl3/opl3_seq.c for (idx = 0; idx < MAX_OPL3_VOICES; idx++) { idx 54 sound/drivers/opl3/opl3_seq.c opl3->voices[idx].state = SNDRV_OPL3_ST_OFF; idx 55 sound/drivers/opl3/opl3_seq.c opl3->voices[idx].time = 0; idx 56 sound/drivers/opl3/opl3_seq.c opl3->voices[idx].keyon_reg = 0x00; idx 76 sound/drivers/virmidi.c int idx, err; idx 92 sound/drivers/virmidi.c for (idx = 0; idx < midi_devs[dev]; idx++) { idx 96 sound/drivers/virmidi.c err = snd_virmidi_new(card, idx, &rmidi); idx 100 sound/drivers/virmidi.c vmidi->midi[idx] = rmidi; idx 39 sound/firewire/digi00x/amdtp-dot.c u8 idx; idx 62 sound/firewire/digi00x/amdtp-dot.c static u8 dot_scrt(const u8 idx, const unsigned int off) idx 91 sound/firewire/digi00x/amdtp-dot.c const u8 ln = idx & 0xf; idx 92 sound/firewire/digi00x/amdtp-dot.c const u8 hn = (idx >> 4) & 0xf; idx 107 sound/firewire/digi00x/amdtp-dot.c state->idx = data[MAGIC_DOT_BYTE] ^ state->carry; idx 110 sound/firewire/digi00x/amdtp-dot.c state->carry = dot_scrt(state->idx, ++(state->off)); idx 419 sound/firewire/digi00x/amdtp-dot.c p->state.idx = 0x00; idx 45 sound/hda/ext/hdac_ext_bus.c bus->idx = 0; idx 83 sound/hda/ext/hdac_ext_bus.c snprintf(name, sizeof(name), "ehdaudio%dD%d", bus->idx, addr); idx 88 sound/hda/ext/hdac_ext_controller.c int idx; idx 96 sound/hda/ext/hdac_ext_controller.c for (idx = 0; idx < link_count; idx++) { idx 100 sound/hda/ext/hdac_ext_controller.c hlink->index = idx; idx 103 sound/hda/ext/hdac_ext_controller.c (AZX_ML_INTERVAL * idx); idx 149 sound/hda/ext/hdac_ext_controller.c if (bus->idx != bus_idx) idx 31 sound/hda/ext/hdac_ext_stream.c int idx, int direction, int tag) idx 35 sound/hda/ext/hdac_ext_stream.c AZX_PPHC_INTERVAL * idx; idx 39 sound/hda/ext/hdac_ext_stream.c AZX_PPLC_INTERVAL * idx; idx 44 sound/hda/ext/hdac_ext_stream.c AZX_SPB_INTERVAL * idx + idx 48 sound/hda/ext/hdac_ext_stream.c AZX_SPB_INTERVAL * idx + idx 54 sound/hda/ext/hdac_ext_stream.c AZX_DRSM_INTERVAL * idx; idx 57 sound/hda/ext/hdac_ext_stream.c snd_hdac_stream_init(bus, &stream->hstream, idx, direction, tag); idx 73 sound/hda/ext/hdac_ext_stream.c int i, tag, idx = start_idx; idx 81 sound/hda/ext/hdac_ext_stream.c snd_hdac_ext_stream_init(bus, stream, idx, dir, tag); idx 82 sound/hda/ext/hdac_ext_stream.c idx++; idx 66 sound/hda/hdac_component.c void snd_hdac_display_power(struct hdac_bus *bus, unsigned int idx, bool enable) idx 75 sound/hda/hdac_component.c set_bit(idx, &bus->display_power_status); idx 77 sound/hda/hdac_component.c clear_bit(idx, &bus->display_power_status); idx 61 sound/hda/hdac_stream.c int idx, int direction, int tag) idx 65 sound/hda/hdac_stream.c azx_dev->sd_addr = bus->remap_addr + (0x20 * idx + 0x80); idx 67 sound/hda/hdac_stream.c azx_dev->sd_int_sta_mask = 1 << idx; idx 68 sound/hda/hdac_stream.c azx_dev->index = idx; idx 107 sound/i2c/cs8427.c int err, idx; idx 126 sound/i2c/cs8427.c for (idx = 0; idx < count; idx++) idx 127 sound/i2c/cs8427.c data[idx + 1] = bitrev8(ndata[idx]); idx 521 sound/i2c/cs8427.c unsigned int idx; idx 526 sound/i2c/cs8427.c for (idx = 0; idx < ARRAY_SIZE(snd_cs8427_iec958_controls); idx++) { idx 527 sound/i2c/cs8427.c kctl = snd_ctl_new1(&snd_cs8427_iec958_controls[idx], cs8427); idx 488 sound/i2c/other/ak4113.c unsigned int idx; idx 494 sound/i2c/other/ak4113.c for (idx = 0; idx < AK4113_CONTROLS; idx++) { idx 495 sound/i2c/other/ak4113.c kctl = snd_ctl_new1(&snd_ak4113_iec958_controls[idx], ak4113); idx 503 sound/i2c/other/ak4113.c ak4113->kctls[idx] = kctl; idx 462 sound/i2c/other/ak4114.c unsigned int idx; idx 469 sound/i2c/other/ak4114.c for (idx = 0; idx < AK4114_CONTROLS; idx++) { idx 470 sound/i2c/other/ak4114.c kctl = snd_ctl_new1(&snd_ak4114_iec958_controls[idx], ak4114); idx 476 sound/i2c/other/ak4114.c ak4114->kctls[idx] = NULL; idx 488 sound/i2c/other/ak4114.c ak4114->kctls[idx] = kctl; idx 413 sound/i2c/other/ak4117.c unsigned int idx; idx 419 sound/i2c/other/ak4117.c for (idx = 0; idx < AK4117_CONTROLS; idx++) { idx 420 sound/i2c/other/ak4117.c kctl = snd_ctl_new1(&snd_ak4117_iec958_controls[idx], ak4117); idx 428 sound/i2c/other/ak4117.c ak4117->kctls[idx] = kctl; idx 605 sound/i2c/other/ak4xxx-adda.c int idx, err, mixer_ch, num_stereo; idx 609 sound/i2c/other/ak4xxx-adda.c for (idx = 0; idx < ak->num_dacs; ) { idx 625 sound/i2c/other/ak4xxx-adda.c AK_COMPOSE(idx/2, 1, 0, 0) | AK_INVERT; idx 656 sound/i2c/other/ak4xxx-adda.c AK_COMPOSE(idx/2, (idx%2) + 6, 0, 127) | idx 663 sound/i2c/other/ak4xxx-adda.c AK_COMPOSE(idx/2, (idx%2) + 4, 0, 127) | idx 669 sound/i2c/other/ak4xxx-adda.c int val = idx < 6 ? idx + 2 : (idx - 6) + 0xb; idx 677 sound/i2c/other/ak4xxx-adda.c knew.private_value = AK_COMPOSE(0, idx + 4, 0, 255); idx 682 sound/i2c/other/ak4xxx-adda.c int addr = idx < 6 ? idx + 4 : idx + 5; idx 691 sound/i2c/other/ak4xxx-adda.c AK_COMPOSE(idx/2, (idx%2) + 3, 0, 255); idx 697 sound/i2c/other/ak4xxx-adda.c AK_COMPOSE(idx/2, (idx%2) + 6, 0, 255); idx 708 sound/i2c/other/ak4xxx-adda.c idx += num_stereo; idx 716 sound/i2c/other/ak4xxx-adda.c int idx, err, mixer_ch, num_stereo, max_steps; idx 722 sound/i2c/other/ak4xxx-adda.c for (idx = 0; idx < ak->num_adcs;) { idx 751 sound/i2c/other/ak4xxx-adda.c AK_COMPOSE(idx/2, (idx%2) + 4, 0, max_steps) | idx 758 sound/i2c/other/ak4xxx-adda.c if (ak->type == SND_AK5365 && (idx % 2) == 0) { idx 772 sound/i2c/other/ak4xxx-adda.c AK_COMPOSE(idx/2, 2, 0, 0) | AK_INVERT; idx 793 sound/i2c/other/ak4xxx-adda.c = AK_COMPOSE(idx/2, 1, mixer_ch, 0x07); idx 799 sound/i2c/other/ak4xxx-adda.c idx += num_stereo; idx 807 sound/i2c/other/ak4xxx-adda.c int idx, err; idx 810 sound/i2c/other/ak4xxx-adda.c for (idx = 0; idx < num_emphs; idx++) { idx 813 sound/i2c/other/ak4xxx-adda.c knew.index = idx + ak->idx_offset; idx 824 sound/i2c/other/ak4xxx-adda.c knew.private_value = AK_COMPOSE(idx, 3, 0, 0); idx 827 sound/i2c/other/ak4xxx-adda.c int shift = idx == 3 ? 6 : (2 - idx) * 2; idx 834 sound/i2c/other/ak4xxx-adda.c knew.private_value = AK_COMPOSE(idx, 3, 0, 0); idx 837 sound/i2c/other/ak4xxx-adda.c knew.private_value = AK_COMPOSE(idx, 1, 1, 0); idx 282 sound/i2c/tea6330t.c unsigned int idx; idx 328 sound/i2c/tea6330t.c for (idx = 0; idx < 6; idx++) idx 329 sound/i2c/tea6330t.c bytes[idx+1] = tea->regs[idx]; idx 337 sound/i2c/tea6330t.c for (idx = 0; idx < ARRAY_SIZE(snd_tea6330t_controls); idx++) { idx 338 sound/i2c/tea6330t.c knew = &snd_tea6330t_controls[idx]; idx 953 sound/isa/ad1816a/ad1816a_lib.c unsigned int idx; idx 963 sound/isa/ad1816a/ad1816a_lib.c for (idx = 0; idx < ARRAY_SIZE(snd_ad1816a_controls); idx++) { idx 964 sound/isa/ad1816a/ad1816a_lib.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_ad1816a_controls[idx], chip))) < 0) idx 272 sound/isa/cmi8330.c int idx, err; idx 280 sound/isa/cmi8330.c for (idx = 0; idx < ARRAY_SIZE(cmi8330_sb_init_values); idx++) { idx 282 sound/isa/cmi8330.c snd_sbmixer_write(chip, cmi8330_sb_init_values[idx][0], idx 283 sound/isa/cmi8330.c cmi8330_sb_init_values[idx][1]); idx 287 sound/isa/cmi8330.c for (idx = 0; idx < ARRAY_SIZE(cmi8330_sb_mixers); idx++) { idx 288 sound/isa/cmi8330.c if ((err = snd_sbmixer_add_ctl_elem(chip, &cmi8330_sb_mixers[idx])) < 0) idx 297 sound/isa/cmi8330.c unsigned int idx; idx 302 sound/isa/cmi8330.c for (idx = 0; idx < ARRAY_SIZE(snd_cmi8330_controls); idx++) { idx 304 sound/isa/cmi8330.c snd_ctl_new1(&snd_cmi8330_controls[idx], idx 289 sound/isa/cs423x/cs4236_lib.c int idx; idx 290 sound/isa/cs423x/cs4236_lib.c for (idx = 0; idx < 8; idx++) idx 292 sound/isa/cs423x/cs4236_lib.c idx, inb(chip->cport + idx)); idx 293 sound/isa/cs423x/cs4236_lib.c for (idx = 0; idx < 9; idx++) idx 295 sound/isa/cs423x/cs4236_lib.c idx, snd_cs4236_ctrl_in(chip, idx)); idx 1021 sound/isa/cs423x/cs4236_lib.c unsigned int idx, count; idx 1032 sound/isa/cs423x/cs4236_lib.c for (idx = 0; idx < ARRAY_SIZE(snd_cs4235_controls); idx++) { idx 1033 sound/isa/cs423x/cs4236_lib.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_cs4235_controls[idx], chip))) < 0) idx 1037 sound/isa/cs423x/cs4236_lib.c for (idx = 0; idx < ARRAY_SIZE(snd_cs4236_controls); idx++) { idx 1038 sound/isa/cs423x/cs4236_lib.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_cs4236_controls[idx], chip))) < 0) idx 1060 sound/isa/cs423x/cs4236_lib.c for (idx = 0; idx < count; idx++, kcontrol++) { idx 1066 sound/isa/cs423x/cs4236_lib.c for (idx = 0; idx < ARRAY_SIZE(snd_cs4236_iec958_controls); idx++) { idx 1067 sound/isa/cs423x/cs4236_lib.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_cs4236_iec958_controls[idx], chip))) < 0) idx 987 sound/isa/es1688/es1688_lib.c unsigned int idx; idx 996 sound/isa/es1688/es1688_lib.c for (idx = 0; idx < ARRAY_SIZE(snd_es1688_controls); idx++) { idx 997 sound/isa/es1688/es1688_lib.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_es1688_controls[idx], chip))) < 0) idx 1000 sound/isa/es1688/es1688_lib.c for (idx = 0; idx < ES1688_INIT_TABLE_SIZE; idx++) { idx 1001 sound/isa/es1688/es1688_lib.c reg = snd_es1688_init_table[idx][0]; idx 1002 sound/isa/es1688/es1688_lib.c val = snd_es1688_init_table[idx][1]; idx 1831 sound/isa/es18xx.c unsigned int idx; idx 1835 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_base_controls); idx++) { idx 1837 sound/isa/es18xx.c kctl = snd_ctl_new1(&snd_es18xx_base_controls[idx], chip); idx 1839 sound/isa/es18xx.c switch (idx) { idx 1854 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_pcm2_controls); idx++) { idx 1855 sound/isa/es18xx.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_es18xx_pcm2_controls[idx], chip))) < 0) idx 1859 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_pcm1_controls); idx++) { idx 1860 sound/isa/es18xx.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_es18xx_pcm1_controls[idx], chip))) < 0) idx 1866 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_recmix_controls); idx++) { idx 1867 sound/isa/es18xx.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_es18xx_recmix_controls[idx], chip))) < 0) idx 1883 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_spatializer_controls); idx++) { idx 1884 sound/isa/es18xx.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_es18xx_spatializer_controls[idx], chip))) < 0) idx 1889 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_hw_volume_controls); idx++) { idx 1891 sound/isa/es18xx.c kctl = snd_ctl_new1(&snd_es18xx_hw_volume_controls[idx], chip); idx 1892 sound/isa/es18xx.c if (idx == 0) idx 1911 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_opt_1869); idx++) { idx 1913 sound/isa/es18xx.c snd_ctl_new1(&snd_es18xx_opt_1869[idx], idx 1924 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_opt_1879); idx++) { idx 1926 sound/isa/es18xx.c snd_ctl_new1(&snd_es18xx_opt_1879[idx], idx 1933 sound/isa/es18xx.c for (idx = 0; idx < ARRAY_SIZE(snd_es18xx_opt_gpo_2bit); idx++) { idx 1935 sound/isa/es18xx.c snd_ctl_new1(&snd_es18xx_opt_gpo_2bit[idx], idx 107 sound/isa/gus/gus_irq.c int idx; idx 117 sound/isa/gus/gus_irq.c for (idx = 0; idx < 32; idx++) { idx 118 sound/isa/gus/gus_irq.c pvoice = &gus->gf1.voices[idx]; idx 120 sound/isa/gus/gus_irq.c idx, idx 231 sound/isa/gus/gus_main.c int l, idx, local; idx 240 sound/isa/gus/gus_main.c for (idx = 1, d = 0xab; idx < 4; idx++, d++) { idx 241 sound/isa/gus/gus_main.c local = idx << 18; idx 250 sound/isa/gus/gus_main.c gus->gf1.memory = idx << 18; idx 126 sound/isa/gus/gus_mem.c unsigned int idx, boundary; idx 140 sound/isa/gus/gus_mem.c for (pblock = alloc->first, idx = 0; pblock; pblock = pblock->next) { idx 141 sound/isa/gus/gus_mem.c while (pblock->ptr >= (boundary = info[idx].address + info[idx].size)) idx 142 sound/isa/gus/gus_mem.c idx++; idx 143 sound/isa/gus/gus_mem.c while (pblock->ptr + pblock->size >= (boundary = info[idx].address + info[idx].size)) idx 144 sound/isa/gus/gus_mem.c idx++; idx 162 sound/isa/gus/gus_mem.c while (++idx < 4) { idx 163 sound/isa/gus/gus_mem.c if (size <= info[idx].size) { idx 165 sound/isa/gus/gus_mem.c block->ptr = info[idx].address; idx 46 sound/isa/gus/gus_mem_proc.c int idx; idx 51 sound/isa/gus/gus_mem_proc.c for (idx = 0; idx < 4; idx++) { idx 52 sound/isa/gus/gus_mem_proc.c if (gus->gf1.mem_alloc.banks_8[idx].size > 0) { idx 57 sound/isa/gus/gus_mem_proc.c sprintf(name, "gus-ram-%i", idx); idx 63 sound/isa/gus/gus_mem_proc.c priv->address = gus->gf1.mem_alloc.banks_8[idx].address; idx 64 sound/isa/gus/gus_mem_proc.c priv->size = entry->size = gus->gf1.mem_alloc.banks_8[idx].size; idx 68 sound/isa/gus/gus_mem_proc.c for (idx = 0; idx < 4; idx++) { idx 69 sound/isa/gus/gus_mem_proc.c if (gus->gf1.rom_present & (1 << idx)) { idx 75 sound/isa/gus/gus_mem_proc.c sprintf(name, "gus-rom-%i", idx); idx 81 sound/isa/gus/gus_mem_proc.c priv->address = idx * 4096 * 1024; idx 143 sound/isa/gus/gus_mixer.c unsigned int idx, max; idx 164 sound/isa/gus/gus_mixer.c for (idx = 0; idx < max; idx++) { idx 165 sound/isa/gus/gus_mixer.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_gf1_controls[idx], gus))) < 0) idx 169 sound/isa/gus/gus_mixer.c for (idx = 0; idx < ARRAY_SIZE(snd_ics_controls); idx++) { idx 170 sound/isa/gus/gus_mixer.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_ics_controls[idx], gus))) < 0) idx 177 sound/isa/gus/gus_pcm.c unsigned int idx; idx 222 sound/isa/gus/gus_pcm.c for (idx = 0; idx < pcmp->voices; idx++, end += step) { idx 223 sound/isa/gus/gus_pcm.c snd_gf1_select_voice(gus, pcmp->pvoices[idx]->number); idx 232 sound/isa/gus/gus_pcm.c for (idx = 0; idx < pcmp->voices; idx++) { idx 233 sound/isa/gus/gus_pcm.c snd_gf1_select_voice(gus, pcmp->pvoices[idx]->number); idx 775 sound/isa/gus/gus_pcm.c unsigned int idx; idx 792 sound/isa/gus/gus_pcm.c for (idx = 0; idx < 32; idx++) { idx 793 sound/isa/gus/gus_pcm.c pvoice = &gus->gf1.voices[idx]; idx 246 sound/isa/gus/gus_reset.c int idx; idx 255 sound/isa/gus/gus_reset.c for (idx = 0; idx < 32; idx++) { idx 256 sound/isa/gus/gus_reset.c pvoice = &gus->gf1.voices[idx]; idx 263 sound/isa/gus/gus_reset.c for (idx = 0; idx < 32; idx++) { idx 264 sound/isa/gus/gus_reset.c pvoice = &gus->gf1.voices[idx]; idx 316 sound/isa/gus/interwave.c unsigned int idx; idx 320 sound/isa/gus/interwave.c for (idx = 0; idx < 4; idx++) { idx 321 sound/isa/gus/interwave.c sizes[idx] = 0; idx 323 sound/isa/gus/interwave.c for (local = idx << 22; idx 324 sound/isa/gus/interwave.c local < (idx << 22) + 0x400000; idx 334 sound/isa/gus/interwave.c snd_gf1_peek(gus, idx << 22)); idx 338 sound/isa/gus/interwave.c snd_gf1_peek(gus, idx << 22) != 0x55) idx 340 sound/isa/gus/interwave.c sizes[idx]++; idx 493 sound/isa/gus/interwave.c unsigned int idx; idx 509 sound/isa/gus/interwave.c for (idx = 0; idx < ARRAY_SIZE(snd_interwave_controls); idx++) idx 510 sound/isa/gus/interwave.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_interwave_controls[idx], chip))) < 0) idx 879 sound/isa/msnd/msnd_pinnacle.c static int snd_msnd_isa_probe(struct device *pdev, unsigned int idx) idx 885 sound/isa/msnd/msnd_pinnacle.c if (has_isapnp(idx) idx 887 sound/isa/msnd/msnd_pinnacle.c || cfg[idx] == SNDRV_AUTO_PORT idx 894 sound/isa/msnd/msnd_pinnacle.c err = snd_card_new(pdev, index[idx], id[idx], THIS_MODULE, idx 903 sound/isa/msnd/msnd_pinnacle.c switch (irq[idx]) { idx 918 sound/isa/msnd/msnd_pinnacle.c switch (mem[idx]) { idx 934 sound/isa/msnd/msnd_pinnacle.c cfg[idx]); idx 936 sound/isa/msnd/msnd_pinnacle.c if (!request_region(cfg[idx], 2, "Pinnacle/Fiji Config")) { idx 938 sound/isa/msnd/msnd_pinnacle.c cfg[idx]); idx 942 sound/isa/msnd/msnd_pinnacle.c if (reset[idx]) idx 943 sound/isa/msnd/msnd_pinnacle.c if (snd_msnd_pinnacle_cfg_reset(cfg[idx])) { idx 949 sound/isa/msnd/msnd_pinnacle.c err = snd_msnd_write_cfg_logical(cfg[idx], 0, idx 950 sound/isa/msnd/msnd_pinnacle.c io[idx], 0, idx 951 sound/isa/msnd/msnd_pinnacle.c irq[idx], mem[idx]); idx 959 sound/isa/msnd/msnd_pinnacle.c if (mpu_io[idx] != SNDRV_AUTO_PORT idx 960 sound/isa/msnd/msnd_pinnacle.c && mpu_irq[idx] != SNDRV_AUTO_IRQ) { idx 963 sound/isa/msnd/msnd_pinnacle.c mpu_io[idx], mpu_irq[idx]); idx 964 sound/isa/msnd/msnd_pinnacle.c err = snd_msnd_write_cfg_logical(cfg[idx], 1, idx 965 sound/isa/msnd/msnd_pinnacle.c mpu_io[idx], 0, idx 966 sound/isa/msnd/msnd_pinnacle.c mpu_irq[idx], 0); idx 973 sound/isa/msnd/msnd_pinnacle.c if (ide_io0[idx] != SNDRV_AUTO_PORT idx 974 sound/isa/msnd/msnd_pinnacle.c && ide_io1[idx] != SNDRV_AUTO_PORT idx 975 sound/isa/msnd/msnd_pinnacle.c && ide_irq[idx] != SNDRV_AUTO_IRQ) { idx 978 sound/isa/msnd/msnd_pinnacle.c ide_io0[idx], ide_io1[idx], ide_irq[idx]); idx 979 sound/isa/msnd/msnd_pinnacle.c err = snd_msnd_write_cfg_logical(cfg[idx], 2, idx 980 sound/isa/msnd/msnd_pinnacle.c ide_io0[idx], ide_io1[idx], idx 981 sound/isa/msnd/msnd_pinnacle.c ide_irq[idx], 0); idx 988 sound/isa/msnd/msnd_pinnacle.c if (joystick_io[idx] != SNDRV_AUTO_PORT) { idx 991 sound/isa/msnd/msnd_pinnacle.c joystick_io[idx]); idx 992 sound/isa/msnd/msnd_pinnacle.c err = snd_msnd_write_cfg_logical(cfg[idx], 3, idx 993 sound/isa/msnd/msnd_pinnacle.c joystick_io[idx], 0, idx 999 sound/isa/msnd/msnd_pinnacle.c release_region(cfg[idx], 2); idx 1009 sound/isa/msnd/msnd_pinnacle.c chip->io = io[idx]; idx 1010 sound/isa/msnd/msnd_pinnacle.c chip->irq = irq[idx]; idx 1011 sound/isa/msnd/msnd_pinnacle.c chip->base = mem[idx]; idx 1017 sound/isa/msnd/msnd_pinnacle.c if (write_ndelay[idx]) idx 1022 sound/isa/msnd/msnd_pinnacle.c if (digital[idx]) idx 1045 sound/isa/msnd/msnd_pinnacle.c release_region(cfg[idx], 2); idx 1071 sound/isa/msnd/msnd_pinnacle.c static int idx; idx 1078 sound/isa/msnd/msnd_pinnacle.c for ( ; idx < SNDRV_CARDS; idx++) { idx 1079 sound/isa/msnd/msnd_pinnacle.c if (has_isapnp(idx)) idx 1082 sound/isa/msnd/msnd_pinnacle.c if (idx >= SNDRV_CARDS) idx 1111 sound/isa/msnd/msnd_pinnacle.c index[idx], id[idx], THIS_MODULE, idx 1122 sound/isa/msnd/msnd_pinnacle.c io[idx] = pnp_port_start(pnp_dev, 0); idx 1123 sound/isa/msnd/msnd_pinnacle.c irq[idx] = pnp_irq(pnp_dev, 0); idx 1124 sound/isa/msnd/msnd_pinnacle.c mem[idx] = pnp_mem_start(pnp_dev, 0); idx 1125 sound/isa/msnd/msnd_pinnacle.c mpu_io[idx] = pnp_port_start(mpu_dev, 0); idx 1126 sound/isa/msnd/msnd_pinnacle.c mpu_irq[idx] = pnp_irq(mpu_dev, 0); idx 1134 sound/isa/msnd/msnd_pinnacle.c chip->io = io[idx]; idx 1135 sound/isa/msnd/msnd_pinnacle.c chip->irq = irq[idx]; idx 1136 sound/isa/msnd/msnd_pinnacle.c chip->base = mem[idx]; idx 1142 sound/isa/msnd/msnd_pinnacle.c if (write_ndelay[idx]) idx 1147 sound/isa/msnd/msnd_pinnacle.c if (digital[idx]) idx 1164 sound/isa/msnd/msnd_pinnacle.c ++idx; idx 298 sound/isa/msnd/msnd_pinnacle_mixer.c unsigned int idx; idx 306 sound/isa/msnd/msnd_pinnacle_mixer.c for (idx = 0; idx < ARRAY_SIZE(snd_msnd_controls); idx++) { idx 308 sound/isa/msnd/msnd_pinnacle_mixer.c snd_ctl_new1(snd_msnd_controls + idx, chip)); idx 488 sound/isa/opl3sa2.c unsigned int idx; idx 521 sound/isa/opl3sa2.c for (idx = 0; idx < ARRAY_SIZE(snd_opl3sa2_controls); idx++) { idx 522 sound/isa/opl3sa2.c if ((err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_opl3sa2_controls[idx], chip))) < 0) idx 524 sound/isa/opl3sa2.c switch (idx) { idx 530 sound/isa/opl3sa2.c for (idx = 0; idx < ARRAY_SIZE(snd_opl3sa2_tone_controls); idx++) idx 531 sound/isa/opl3sa2.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_opl3sa2_tone_controls[idx], chip))) < 0) idx 665 sound/isa/opti9xx/miro.c int idx, error; idx 690 sound/isa/opti9xx/miro.c for (idx = 0; idx < ARRAY_SIZE(aci_init_values); idx++) { idx 691 sound/isa/opti9xx/miro.c error = aci_setvalue(aci, aci_init_values[idx][0], idx 692 sound/isa/opti9xx/miro.c aci_init_values[idx][1]); idx 695 sound/isa/opti9xx/miro.c aci_init_values[idx][0], error); idx 709 sound/isa/opti9xx/miro.c unsigned int idx; idx 727 sound/isa/opti9xx/miro.c for (idx = 0; idx < ARRAY_SIZE(snd_miro_controls); idx++) { idx 728 sound/isa/opti9xx/miro.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_miro_controls[idx], miro))) < 0) idx 755 sound/isa/opti9xx/miro.c for (idx = 0; idx < ARRAY_SIZE(snd_miro_eq_controls); idx++) { idx 756 sound/isa/opti9xx/miro.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_miro_eq_controls[idx], miro))) < 0) idx 591 sound/isa/opti9xx/opti92x-ad1848.c unsigned int idx; idx 627 sound/isa/opti9xx/opti92x-ad1848.c for (idx = 0; idx < ARRAY_SIZE(snd_opti93x_controls); idx++) { idx 628 sound/isa/opti9xx/opti92x-ad1848.c strcpy(id1.name, snd_opti93x_controls[idx].name); idx 632 sound/isa/opti9xx/opti92x-ad1848.c snd_ctl_new1(&snd_opti93x_controls[idx], chip)); idx 82 sound/isa/sb/jazz16.c unsigned long mpu_port, int idx) idx 90 sound/isa/sb/jazz16.c outb(SB_JAZZ16_WAKEUP - idx, 0x201); idx 92 sound/isa/sb/jazz16.c outb(SB_JAZZ16_SET_PORTS + idx, 0x201); idx 700 sound/isa/sb/sb_mixer.c int idx, err; idx 708 sound/isa/sb/sb_mixer.c for (idx = 0; idx < map_count; idx++) { idx 710 sound/isa/sb/sb_mixer.c snd_sbmixer_write(chip, map[idx][0], map[idx][1]); idx 714 sound/isa/sb/sb_mixer.c for (idx = 0; idx < controls_count; idx++) { idx 715 sound/isa/sb/sb_mixer.c err = snd_sbmixer_add_ctl_elem(chip, &controls[idx]); idx 1216 sound/isa/sscape.c static int idx = 0; idx 1226 sound/isa/sscape.c idx = get_next_autoindex(idx); idx 1227 sound/isa/sscape.c if (idx >= SNDRV_CARDS) idx 1249 sound/isa/sscape.c index[idx], id[idx], THIS_MODULE, idx 1267 sound/isa/sscape.c port[idx] = pnp_port_start(dev, 0); idx 1268 sound/isa/sscape.c irq[idx] = pnp_irq(dev, 0); idx 1269 sound/isa/sscape.c mpu_irq[idx] = pnp_irq(dev, 1); idx 1270 sound/isa/sscape.c dma[idx] = pnp_dma(dev, 0) & 0x03; idx 1272 sound/isa/sscape.c dma2[idx] = dma[idx]; idx 1273 sound/isa/sscape.c wss_port[idx] = CODEC_IO(port[idx]); idx 1275 sound/isa/sscape.c wss_port[idx] = pnp_port_start(dev, 1); idx 1276 sound/isa/sscape.c dma2[idx] = pnp_dma(dev, 1); idx 1279 sound/isa/sscape.c ret = create_sscape(idx, card); idx 1290 sound/isa/sscape.c ++idx; idx 2230 sound/isa/wss/wss_lib.c unsigned int idx; idx 2248 sound/isa/wss/wss_lib.c for (idx = 0; idx < count; idx++) { idx 2250 sound/isa/wss/wss_lib.c snd_ctl_new1(&snd_wss_controls[idx], idx 13 sound/last.c int idx, ok = 0; idx 16 sound/last.c for (idx = 0; idx < SNDRV_CARDS; idx++) { idx 17 sound/last.c card = snd_card_ref(idx); idx 19 sound/last.c printk(KERN_INFO " #%i: %s\n", idx, card->longname); idx 72 sound/mips/sgio2audio.c int idx; idx 444 sound/mips/sgio2audio.c int ch = chan->idx; idx 465 sound/mips/sgio2audio.c writeq(0, &mace->perif.audio.chan[chan->idx].control); idx 478 sound/mips/sgio2audio.c ch = chan->idx; idx 498 sound/mips/sgio2audio.c ch = chan->idx; idx 601 sound/mips/sgio2audio.c int ch = chan->idx; idx 660 sound/mips/sgio2audio.c chip->channel[chan->idx].pos); idx 740 sound/mips/sgio2audio.c int idx; idx 746 sound/mips/sgio2audio.c .idx = 0, idx 751 sound/mips/sgio2audio.c .idx = 0, idx 756 sound/mips/sgio2audio.c .idx = 1, idx 761 sound/mips/sgio2audio.c .idx = 1, idx 766 sound/mips/sgio2audio.c .idx = 2, idx 771 sound/mips/sgio2audio.c .idx = 2, idx 792 sound/mips/sgio2audio.c &chip->channel[snd_sgio2_isr_table[i].idx]); idx 847 sound/mips/sgio2audio.c chip->channel[i].idx = i; idx 856 sound/mips/sgio2audio.c &chip->channel[snd_sgio2_isr_table[i].idx])) { idx 909 sound/oss/dmasound/dmasound_atari.c int mode, i, idx; idx 914 sound/oss/dmasound/dmasound_atari.c idx = -1; idx 920 sound/oss/dmasound/dmasound_atari.c idx = i; idx 921 sound/oss/dmasound/dmasound_atari.c if (idx > -1) { idx 922 sound/oss/dmasound/dmasound_atari.c dmasound.soft.speed = freq[idx]; idx 1035 sound/oss/dmasound/dmasound_atari.c int divider, i, idx; idx 1040 sound/oss/dmasound/dmasound_atari.c idx = -1; idx 1047 sound/oss/dmasound/dmasound_atari.c idx = i; idx 1048 sound/oss/dmasound/dmasound_atari.c if (idx > -1) { idx 1049 sound/oss/dmasound/dmasound_atari.c dmasound.soft.speed = freq[idx]; idx 510 sound/oss/dmasound/dmasound_q40.c int i, idx; idx 515 sound/oss/dmasound/dmasound_q40.c idx = -1; idx 518 sound/oss/dmasound/dmasound_q40.c idx = i; idx 524 sound/oss/dmasound/dmasound_q40.c if (idx > -1) { idx 525 sound/oss/dmasound/dmasound_q40.c dmasound.soft.speed = freq[idx]; idx 847 sound/parisc/harmony.c int idx, err; idx 854 sound/parisc/harmony.c for (idx = 0; idx < HARMONY_CONTROLS; idx++) { idx 856 sound/parisc/harmony.c snd_ctl_new1(&snd_harmony_controls[idx], h)); idx 1315 sound/pci/ac97/ac97_codec.c unsigned int idx; idx 1388 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 2; idx++) { idx 1389 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, kctl = snd_ac97_cnew(&snd_ac97_controls_tone[idx], ac97))) < 0) idx 1405 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 2; idx++) idx 1406 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, kctl = snd_ac97_cnew(&snd_ac97_controls_pc_beep[idx], ac97))) < 0) idx 1478 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 2; idx++) idx 1479 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, kctl = snd_ac97_cnew(&snd_ac97_controls_ad18xx_pcm[idx], ac97))) < 0) idx 1484 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 2; idx++) idx 1485 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, kctl = snd_ac97_cnew(&snd_ac97_controls_ad18xx_surround[idx], ac97))) < 0) idx 1491 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 2; idx++) idx 1492 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, kctl = snd_ac97_cnew(&snd_ac97_controls_ad18xx_center[idx], ac97))) < 0) idx 1495 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 2; idx++) idx 1496 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, kctl = snd_ac97_cnew(&snd_ac97_controls_ad18xx_lfe[idx], ac97))) < 0) idx 1534 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 2; idx++) idx 1535 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, kctl = snd_ac97_cnew(&snd_ac97_controls_mic_capture[idx], ac97))) < 0) idx 1619 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < 5; idx++) idx 1620 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, snd_ac97_cnew(&snd_ac97_controls_spdif[idx], ac97))) < 0) idx 1654 sound/pci/ac97/ac97_codec.c int err, idx; idx 1667 sound/pci/ac97/ac97_codec.c for (idx = 0; idx < ARRAY_SIZE(snd_ac97_controls_modem_switches); idx++) idx 1668 sound/pci/ac97/ac97_codec.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_ac97_controls_modem_switches[idx], ac97))) < 0) idx 30 sound/pci/ac97/ac97_patch.c int idx, err; idx 32 sound/pci/ac97/ac97_patch.c for (idx = 0; idx < count; idx++) idx 33 sound/pci/ac97/ac97_patch.c if ((err = snd_ctl_add(ac97->bus->card, snd_ac97_cnew(&controls[idx], ac97))) < 0) idx 1455 sound/pci/ac97/ac97_patch.c static unsigned short patch_ad1881_unchained(struct snd_ac97 * ac97, int idx, unsigned short mask) idx 1465 sound/pci/ac97/ac97_patch.c ac97->spec.ad18xx.unchained[idx] = mask; idx 1466 sound/pci/ac97/ac97_patch.c ac97->spec.ad18xx.id[idx] = val; idx 1467 sound/pci/ac97/ac97_patch.c ac97->spec.ad18xx.codec_cfg[idx] = 0x0000; idx 1471 sound/pci/ac97/ac97_patch.c static int patch_ad1881_chained1(struct snd_ac97 * ac97, int idx, unsigned short codec_bits) idx 1476 sound/pci/ac97/ac97_patch.c snd_ac97_update_bits(ac97, AC97_AD_SERIAL_CFG, 0x7000, cfg_bits[idx]); idx 1483 sound/pci/ac97/ac97_patch.c ac97->spec.ad18xx.chained[idx] = cfg_bits[idx]; idx 1484 sound/pci/ac97/ac97_patch.c ac97->spec.ad18xx.id[idx] = val; idx 1485 sound/pci/ac97/ac97_patch.c ac97->spec.ad18xx.codec_cfg[idx] = codec_bits ? codec_bits : 0x0004; idx 1532 sound/pci/ac97/ac97_patch.c int idx, num; idx 1543 sound/pci/ac97/ac97_patch.c for (idx = 0; idx < 3; idx++) idx 1544 sound/pci/ac97/ac97_patch.c if (ac97->spec.ad18xx.unchained[idx]) idx 1545 sound/pci/ac97/ac97_patch.c patch_ad1881_chained(ac97, idx, cfg_idxs[idx][0], cfg_idxs[idx][1]); idx 1560 sound/pci/ac97/ac97_patch.c for (idx = num = 0; idx < 3; idx++) idx 1561 sound/pci/ac97/ac97_patch.c if (ac97->spec.ad18xx.id[idx]) idx 3847 sound/pci/ac97/ac97_patch.c int idx, err; idx 3848 sound/pci/ac97/ac97_patch.c for (idx = 0; idx < ARRAY_SIZE(snd_ac97_controls_si3036); idx++) idx 3849 sound/pci/ac97/ac97_patch.c if ((err = snd_ctl_add(ac97->bus->card, snd_ctl_new1(&snd_ac97_controls_si3036[idx], ac97))) < 0) idx 3914 sound/pci/ac97/ac97_patch.c int idx, err; idx 3915 sound/pci/ac97/ac97_patch.c for (idx = 0; idx < ARRAY_SIZE(snd_ac97_controls_ucb1400); idx++) idx 3916 sound/pci/ac97/ac97_patch.c if ((err = snd_ctl_add(ac97->bus->card, snd_ctl_new1(&snd_ac97_controls_ucb1400[idx], ac97))) < 0) idx 395 sound/pci/ac97/ac97_pcm.c int i, idx; idx 404 sound/pci/ac97/ac97_pcm.c case AC97_PCM_FRONT_DAC_RATE: idx = AC97_RATES_FRONT_DAC; break; idx 405 sound/pci/ac97/ac97_pcm.c case AC97_PCM_SURR_DAC_RATE: idx = AC97_RATES_SURR_DAC; break; idx 406 sound/pci/ac97/ac97_pcm.c case AC97_PCM_LFE_DAC_RATE: idx = AC97_RATES_LFE_DAC; break; idx 407 sound/pci/ac97/ac97_pcm.c case AC97_PCM_LR_ADC_RATE: idx = AC97_RATES_ADC; break; idx 408 sound/pci/ac97/ac97_pcm.c case AC97_PCM_MIC_ADC_RATE: idx = AC97_RATES_MIC_ADC; break; idx 409 sound/pci/ac97/ac97_pcm.c default: idx = AC97_RATES_SPDIF; break; idx 411 sound/pci/ac97/ac97_pcm.c rates &= pcm->r[dbl].codec[cidx]->rates[idx]; idx 334 sound/pci/ac97/ac97_proc.c int idx; idx 335 sound/pci/ac97/ac97_proc.c for (idx = 0; idx < 3; idx++) idx 336 sound/pci/ac97/ac97_proc.c if (ac97->spec.ad18xx.id[idx]) { idx 339 sound/pci/ac97/ac97_proc.c ac97->spec.ad18xx.unchained[idx] | ac97->spec.ad18xx.chained[idx]); idx 340 sound/pci/ac97/ac97_proc.c snd_ac97_proc_read_main(ac97, buffer, idx); idx 398 sound/pci/ac97/ac97_proc.c int idx; idx 399 sound/pci/ac97/ac97_proc.c for (idx = 0; idx < 3; idx++) idx 400 sound/pci/ac97/ac97_proc.c if (ac97->spec.ad18xx.id[idx]) { idx 403 sound/pci/ac97/ac97_proc.c ac97->spec.ad18xx.unchained[idx] | ac97->spec.ad18xx.chained[idx]); idx 404 sound/pci/ac97/ac97_proc.c snd_ac97_proc_regs_read_main(ac97, buffer, idx); idx 33 sound/pci/ak4531_codec.c int idx; idx 35 sound/pci/ak4531_codec.c for (idx = 0; idx < 0x19; idx++) idx 37 sound/pci/ak4531_codec.c idx, ak4531->regs[idx]); idx 371 sound/pci/ak4531_codec.c unsigned int idx; idx 395 sound/pci/ak4531_codec.c for (idx = 0; idx <= 0x19; idx++) { idx 396 sound/pci/ak4531_codec.c if (idx == AK4531_RESET || idx == AK4531_CLOCK) idx 398 sound/pci/ak4531_codec.c ak4531->write(ak4531, idx, ak4531->regs[idx] = snd_ak4531_initial_map[idx]); /* recording source is mixer */ idx 400 sound/pci/ak4531_codec.c for (idx = 0; idx < ARRAY_SIZE(snd_ak4531_controls); idx++) { idx 401 sound/pci/ak4531_codec.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_ak4531_controls[idx], ak4531))) < 0) { idx 435 sound/pci/ak4531_codec.c int idx; idx 442 sound/pci/ak4531_codec.c for (idx = 0; idx <= 0x19; idx++) { idx 443 sound/pci/ak4531_codec.c if (idx == AK4531_RESET || idx == AK4531_CLOCK) idx 445 sound/pci/ak4531_codec.c ak4531->write(ak4531, idx, ak4531->regs[idx]); idx 517 sound/pci/ali5451/ali5451.c unsigned int idx = channel & 0x1f; idx 525 sound/pci/ali5451/ali5451.c if (!(codec->synth.chmap & (1 << idx))) { idx 526 sound/pci/ali5451/ali5451.c codec->synth.chmap |= 1 << idx; idx 528 sound/pci/ali5451/ali5451.c dev_dbg(codec->card->dev, "alloc_pcm_channel no. %d.\n", idx); idx 529 sound/pci/ali5451/ali5451.c return idx; idx 536 sound/pci/ali5451/ali5451.c int idx; idx 547 sound/pci/ali5451/ali5451.c idx = ALI_SPDIF_IN_CHANNEL; idx 549 sound/pci/ali5451/ali5451.c idx = ALI_PCM_IN_CHANNEL; idx 551 sound/pci/ali5451/ali5451.c result = snd_ali_alloc_pcm_channel(codec, idx); idx 565 sound/pci/ali5451/ali5451.c idx = ALI_SPDIF_OUT_CHANNEL; idx 566 sound/pci/ali5451/ali5451.c result = snd_ali_alloc_pcm_channel(codec, idx); idx 574 sound/pci/ali5451/ali5451.c for (idx = 0; idx < ALI_CHANNELS; idx++) { idx 575 sound/pci/ali5451/ali5451.c result = snd_ali_alloc_pcm_channel(codec, idx); idx 585 sound/pci/ali5451/ali5451.c unsigned int idx = channel & 0x0000001f; idx 592 sound/pci/ali5451/ali5451.c if (!(codec->synth.chmap & (1 << idx))) { idx 598 sound/pci/ali5451/ali5451.c codec->synth.chmap &= ~(1 << idx); idx 913 sound/pci/ali5451/ali5451.c int idx; idx 919 sound/pci/ali5451/ali5451.c idx = channel > 0 ? snd_ali_alloc_pcm_channel(codec, channel) : idx 921 sound/pci/ali5451/ali5451.c if (idx < 0) { idx 926 sound/pci/ali5451/ali5451.c pvoice = &(codec->synth.voices[idx]); idx 1820 sound/pci/ali5451/ali5451.c unsigned int idx; idx 1848 sound/pci/ali5451/ali5451.c for (idx = 0; idx < ARRAY_SIZE(snd_ali5451_mixer_spdif); idx++) { idx 1850 sound/pci/ali5451/ali5451.c snd_ctl_new1(&snd_ali5451_mixer_spdif[idx], codec)); idx 355 sound/pci/asihpi/asihpi.c int idx; idx 376 sound/pci/asihpi/asihpi.c for (idx = -1; idx < 100; idx++) { idx 377 sound/pci/asihpi/asihpi.c if (idx == -1) { idx 382 sound/pci/asihpi/asihpi.c idx, &sample_rate)) { idx 1835 sound/pci/asihpi/asihpi.c short idx; idx 1838 sound/pci/asihpi/asihpi.c for (idx = 0; idx < 3; idx++) { idx 1840 sound/pci/asihpi/asihpi.c idx, &gain_range[idx]); idx 1926 sound/pci/asihpi/asihpi.c u16 band, idx; idx 1936 sound/pci/asihpi/asihpi.c for (idx = 0; idx < HPI_TUNER_BAND_LAST; idx++) idx 1937 sound/pci/asihpi/asihpi.c if (tuner_bands[idx] == band) { idx 1938 sound/pci/asihpi/asihpi.c ucontrol->value.enumerated.item[0] = idx; idx 1952 sound/pci/asihpi/asihpi.c unsigned int idx; idx 1960 sound/pci/asihpi/asihpi.c idx = ucontrol->value.enumerated.item[0]; idx 1961 sound/pci/asihpi/asihpi.c if (idx >= ARRAY_SIZE(tuner_bands)) idx 1962 sound/pci/asihpi/asihpi.c idx = ARRAY_SIZE(tuner_bands) - 1; idx 1963 sound/pci/asihpi/asihpi.c band = tuner_bands[idx]; idx 1977 sound/pci/asihpi/asihpi.c u16 num_bands = 0, band_iter, idx; idx 1988 sound/pci/asihpi/asihpi.c for (idx = 0; idx < 3; idx++) { idx 1990 sound/pci/asihpi/asihpi.c idx, tuner_bands[band_iter], idx 1991 sound/pci/asihpi/asihpi.c &temp_freq_range[idx]); idx 2596 sound/pci/asihpi/asihpi.c unsigned int idx = 0; idx 2616 sound/pci/asihpi/asihpi.c for (idx = 0; idx < 2000; idx++) { idx 2619 sound/pci/asihpi/asihpi.c idx, idx 2631 sound/pci/asihpi/asihpi.c idx); idx 2696 sound/pci/asihpi/asihpi.c idx, idx 2710 sound/pci/asihpi/asihpi.c dev_info(&asihpi->pci->dev, "%d mixer controls found\n", idx); idx 340 sound/pci/asihpi/hpioctl.c int idx, nm, low_latency_mode = 0, irq_supported = 0; idx 372 sound/pci/asihpi/hpioctl.c for (idx = 0; idx < nm; idx++) { idx 373 sound/pci/asihpi/hpioctl.c HPI_DEBUG_LOG(INFO, "resource %d %pR\n", idx, idx 374 sound/pci/asihpi/hpioctl.c &pci_dev->resource[idx]); idx 376 sound/pci/asihpi/hpioctl.c if (pci_resource_flags(pci_dev, idx) & IORESOURCE_MEM) { idx 377 sound/pci/asihpi/hpioctl.c memlen = pci_resource_len(pci_dev, idx); idx 378 sound/pci/asihpi/hpioctl.c pci.ap_mem_base[idx] = idx 379 sound/pci/asihpi/hpioctl.c ioremap(pci_resource_start(pci_dev, idx), idx 381 sound/pci/asihpi/hpioctl.c if (!pci.ap_mem_base[idx]) { idx 502 sound/pci/asihpi/hpioctl.c for (idx = 0; idx < HPI_MAX_ADAPTER_MEM_SPACES; idx++) { idx 503 sound/pci/asihpi/hpioctl.c if (pci.ap_mem_base[idx]) { idx 504 sound/pci/asihpi/hpioctl.c iounmap(pci.ap_mem_base[idx]); idx 505 sound/pci/asihpi/hpioctl.c pci.ap_mem_base[idx] = NULL; idx 520 sound/pci/asihpi/hpioctl.c int idx; idx 544 sound/pci/asihpi/hpioctl.c for (idx = 0; idx < HPI_MAX_ADAPTER_MEM_SPACES; ++idx) idx 545 sound/pci/asihpi/hpioctl.c iounmap(pci.ap_mem_base[idx]); idx 278 sound/pci/au88x0/au88x0.h static int snd_vortex_new_pcm(vortex_t * vortex, int idx, int nr); idx 605 sound/pci/au88x0/au88x0_pcm.c static int snd_vortex_new_pcm(vortex_t *chip, int idx, int nr) idx 612 sound/pci/au88x0/au88x0_pcm.c if (!chip || idx < 0 || idx >= VORTEX_PCM_LAST) idx 617 sound/pci/au88x0/au88x0_pcm.c if (idx == VORTEX_PCM_ADB) idx 621 sound/pci/au88x0/au88x0_pcm.c err = snd_pcm_new(chip->card, vortex_pcm_prettyname[idx], idx, nr, idx 626 sound/pci/au88x0/au88x0_pcm.c "%s %s", CARD_NAME_SHORT, vortex_pcm_name[idx]); idx 627 sound/pci/au88x0/au88x0_pcm.c chip->pcm[idx] = pcm; idx 629 sound/pci/au88x0/au88x0_pcm.c VORTEX_PCM_TYPE(pcm) = idx; idx 634 sound/pci/au88x0/au88x0_pcm.c if (idx == VORTEX_PCM_ADB) idx 1176 sound/pci/azt3328.c unsigned int idx; idx 1188 sound/pci/azt3328.c for (idx = 0; idx < ARRAY_SIZE(snd_azf3328_init_values); ++idx) { idx 1190 sound/pci/azt3328.c snd_azf3328_init_values[idx][0], idx 1191 sound/pci/azt3328.c snd_azf3328_init_values[idx][1]); idx 1196 sound/pci/azt3328.c for (idx = 0; idx < ARRAY_SIZE(snd_azf3328_mixer_controls); idx 1197 sound/pci/azt3328.c ++idx, ++sw) { idx 496 sound/pci/ca0106/ca0106_main.c static void restore_spdif_bits(struct snd_ca0106 *chip, int idx) idx 498 sound/pci/ca0106/ca0106_main.c if (chip->spdif_str_bits[idx] != chip->spdif_bits[idx]) { idx 499 sound/pci/ca0106/ca0106_main.c chip->spdif_str_bits[idx] = chip->spdif_bits[idx]; idx 500 sound/pci/ca0106/ca0106_main.c snd_ca0106_ptr_write(chip, SPCS0 + idx, 0, idx 501 sound/pci/ca0106/ca0106_main.c chip->spdif_str_bits[idx]); idx 1749 sound/pci/ca0106/ca0106_main.c static unsigned char ca0106_midi_read(struct snd_ca_midi *midi, int idx) idx 1752 sound/pci/ca0106/ca0106_main.c midi->port + idx, 0); idx 1755 sound/pci/ca0106/ca0106_main.c static void ca0106_midi_write(struct snd_ca_midi *midi, int data, int idx) idx 1757 sound/pci/ca0106/ca0106_main.c snd_ca0106_ptr_write((struct snd_ca0106 *)(midi->dev_id), midi->port + idx, 0, data); idx 134 sound/pci/ca0106/ca0106_mixer.c static void ca0106_set_spdif_bits(struct snd_ca0106 *emu, int idx) idx 136 sound/pci/ca0106/ca0106_mixer.c snd_ca0106_ptr_write(emu, SPCS0 + idx, 0, emu->spdif_str_bits[idx]); idx 329 sound/pci/ca0106/ca0106_mixer.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 332 sound/pci/ca0106/ca0106_mixer.c emu->spdif_bits[idx]); idx 340 sound/pci/ca0106/ca0106_mixer.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 343 sound/pci/ca0106/ca0106_mixer.c emu->spdif_str_bits[idx]); idx 369 sound/pci/ca0106/ca0106_mixer.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 373 sound/pci/ca0106/ca0106_mixer.c if (val != emu->spdif_bits[idx]) { idx 374 sound/pci/ca0106/ca0106_mixer.c emu->spdif_bits[idx] = val; idx 378 sound/pci/ca0106/ca0106_mixer.c emu->spdif_str_bits[idx] = val; idx 379 sound/pci/ca0106/ca0106_mixer.c ca0106_set_spdif_bits(emu, idx); idx 389 sound/pci/ca0106/ca0106_mixer.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 393 sound/pci/ca0106/ca0106_mixer.c if (val != emu->spdif_str_bits[idx]) { idx 394 sound/pci/ca0106/ca0106_mixer.c emu->spdif_str_bits[idx] = val; idx 395 sound/pci/ca0106/ca0106_mixer.c ca0106_set_spdif_bits(emu, idx); idx 44 sound/pci/ca0106/ca_midi.h unsigned char (*read)(struct snd_ca_midi *midi, int idx); idx 45 sound/pci/ca0106/ca_midi.h void (*write)(struct snd_ca_midi *midi, int data, int idx); idx 1970 sound/pci/cmipci.c static void snd_cmipci_mixer_write(struct cmipci *s, unsigned char idx, unsigned char data) idx 1972 sound/pci/cmipci.c outb(idx, s->iobase + CM_REG_SB16_ADDR); idx 1976 sound/pci/cmipci.c static unsigned char snd_cmipci_mixer_read(struct cmipci *s, unsigned char idx) idx 1980 sound/pci/cmipci.c outb(idx, s->iobase + CM_REG_SB16_ADDR); idx 2658 sound/pci/cmipci.c unsigned int idx; idx 2672 sound/pci/cmipci.c for (idx = 0; idx < ARRAY_SIZE(snd_cmipci_mixers); idx++) { idx 2674 sound/pci/cmipci.c if (!strcmp(snd_cmipci_mixers[idx].name, idx 2678 sound/pci/cmipci.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_cmipci_mixers[idx], cm))) < 0) idx 2684 sound/pci/cmipci.c for (idx = 0; idx < ARRAY_SIZE(snd_cmipci_mixer_switches); idx++, sw++) { idx 2697 sound/pci/cmipci.c for (idx = 0; idx < ARRAY_SIZE(snd_cmipci_8738_mixer_switches); idx++, sw++) { idx 2715 sound/pci/cmipci.c for (idx = 0; idx < ARRAY_SIZE(snd_cmipci_old_mixer_switches); idx++, sw++) { idx 2724 sound/pci/cmipci.c for (idx = 0; idx < ARRAY_SIZE(snd_cmipci_extra_mixer_switches); idx++, sw++) { idx 2743 sound/pci/cmipci.c for (idx = 0; idx < CM_SAVED_MIXERS; idx++) { idx 2748 sound/pci/cmipci.c strcpy(elem_id.name, cm_saved_mixer[idx].name); idx 2751 sound/pci/cmipci.c cm->mixer_res_ctl[idx] = ctl; idx 1653 sound/pci/cs46xx/cs46xx.h struct snd_cs46xx_region idx[5]; idx 315 sound/pci/cs46xx/cs46xx_lib.c dst = chip->region.idx[bank+1].remap_addr + offset; idx 478 sound/pci/cs46xx/cs46xx_lib.c dst = chip->region.idx[bank+1].remap_addr + offset; idx 536 sound/pci/cs46xx/cs46xx_lib.c int idx, err; idx 540 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < BA1_MEMORY_COUNT; idx++) { idx 543 sound/pci/cs46xx/cs46xx_lib.c ba1->memory[idx].offset, idx 544 sound/pci/cs46xx/cs46xx_lib.c ba1->memory[idx].size); idx 547 sound/pci/cs46xx/cs46xx_lib.c offset += ba1->memory[idx].size >> 2; idx 559 sound/pci/cs46xx/cs46xx_lib.c int idx; idx 574 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < 8; idx++) { idx 575 sound/pci/cs46xx/cs46xx_lib.c snd_cs46xx_poke(chip, BA1_DREG, DREG_REGID_TRAP_SELECT + idx); idx 612 sound/pci/cs46xx/cs46xx_lib.c int idx, powerdown = 0; idx 635 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < 0xFF; idx++) { idx 642 sound/pci/cs46xx/cs46xx_lib.c idx); idx 652 sound/pci/cs46xx/cs46xx_lib.c snd_cs46xx_pokeBA0(chip, BA0_SERBAD, idx); idx 2436 sound/pci/cs46xx/cs46xx_lib.c int idx, err; idx 2457 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < 100; ++idx) { idx 2473 sound/pci/cs46xx/cs46xx_lib.c unsigned int idx; idx 2501 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < ARRAY_SIZE(snd_cs46xx_controls); idx++) { idx 2503 sound/pci/cs46xx/cs46xx_lib.c kctl = snd_ctl_new1(&snd_cs46xx_controls[idx], chip); idx 2831 sound/pci/cs46xx/cs46xx_lib.c int idx; idx 2833 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < 5; idx++) { idx 2834 sound/pci/cs46xx/cs46xx_lib.c struct snd_cs46xx_region *region = &chip->region.idx[idx]; idx 2913 sound/pci/cs46xx/cs46xx_lib.c int idx; idx 2928 sound/pci/cs46xx/cs46xx_lib.c if (chip->region.idx[0].resource) idx 2937 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < 5; idx++) { idx 2938 sound/pci/cs46xx/cs46xx_lib.c struct snd_cs46xx_region *region = &chip->region.idx[idx]; idx 2949 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < CS46XX_DSP_MODULES; idx++) idx 2950 sound/pci/cs46xx/cs46xx_lib.c free_module_desc(chip->modules[idx]); idx 3317 sound/pci/cs46xx/cs46xx_lib.c u32 idx, valid_slots,tmp,powerdown = 0; idx 3388 sound/pci/cs46xx/cs46xx_lib.c for(idx = 0x90; idx <= 0x9F; idx++) { idx 3403 sound/pci/cs46xx/cs46xx_lib.c idx); idx 3411 sound/pci/cs46xx/cs46xx_lib.c snd_cs46xx_pokeBA0(chip, BA0_SERBAD, idx); idx 3517 sound/pci/cs46xx/cs46xx_lib.c unsigned int idx; idx 3531 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0 ; idx < ARRAY_SIZE(snd_hercules_controls); idx++) { idx 3534 sound/pci/cs46xx/cs46xx_lib.c kctl = snd_ctl_new1(&snd_hercules_controls[idx], chip); idx 3869 sound/pci/cs46xx/cs46xx_lib.c int err, idx; idx 3972 sound/pci/cs46xx/cs46xx_lib.c for (idx = 0; idx < 5; idx++) { idx 3973 sound/pci/cs46xx/cs46xx_lib.c region = &chip->region.idx[idx]; idx 55 sound/pci/cs46xx/cs46xx_lib.h writel(val, chip->region.idx[bank+1].remap_addr + offset); idx 62 sound/pci/cs46xx/cs46xx_lib.h return readl(chip->region.idx[bank+1].remap_addr + offset); idx 550 sound/pci/cs46xx/dsp_spos.c void __iomem *dst = chip->region.idx[1].remap_addr + DSP_PARAMETER_BYTE_OFFSET; idx 611 sound/pci/cs46xx/dsp_spos.c void __iomem *dst = chip->region.idx[1].remap_addr + DSP_PARAMETER_BYTE_OFFSET; idx 638 sound/pci/cs46xx/dsp_spos.c void __iomem *dst = chip->region.idx[2].remap_addr; idx 870 sound/pci/cs46xx/dsp_spos.c void __iomem *spdst = chip->region.idx[1].remap_addr + idx 884 sound/pci/cs46xx/dsp_spos.c void __iomem *spdst = chip->region.idx[1].remap_addr + idx 64 sound/pci/cs46xx/dsp_spos_scb_lib.c void __iomem *dst = chip->region.idx[1].remap_addr + DSP_PARAMETER_BYTE_OFFSET; idx 151 sound/pci/cs46xx/dsp_spos_scb_lib.c void __iomem *dst = chip->region.idx[2].remap_addr + sample_buffer_addr; idx 29 sound/pci/ctxfi/ctamixer.c return rsc->idx = container_of(rsc, struct amixer, rsc)->idx[0]; idx 35 sound/pci/ctxfi/ctamixer.c return container_of(rsc, struct amixer, rsc)->idx[rsc->conj]; idx 40 sound/pci/ctxfi/ctamixer.c return container_of(rsc, struct amixer, rsc)->idx[rsc->conj]; idx 202 sound/pci/ctxfi/ctamixer.c err = rsc_init(&amixer->rsc, amixer->idx[0], idx 233 sound/pci/ctxfi/ctamixer.c unsigned int idx; idx 249 sound/pci/ctxfi/ctamixer.c err = mgr_get_resource(&mgr->mgr, 1, &idx); idx 253 sound/pci/ctxfi/ctamixer.c amixer->idx[i] = idx; idx 273 sound/pci/ctxfi/ctamixer.c mgr_put_resource(&mgr->mgr, 1, amixer->idx[i]); idx 287 sound/pci/ctxfi/ctamixer.c mgr_put_resource(&mgr->mgr, 1, amixer->idx[i]); idx 337 sound/pci/ctxfi/ctamixer.c return rsc->idx = container_of(rsc, struct sum, rsc)->idx[0]; idx 343 sound/pci/ctxfi/ctamixer.c return container_of(rsc, struct sum, rsc)->idx[rsc->conj]; idx 348 sound/pci/ctxfi/ctamixer.c return container_of(rsc, struct sum, rsc)->idx[rsc->conj]; idx 369 sound/pci/ctxfi/ctamixer.c err = rsc_init(&sum->rsc, sum->idx[0], SUM, desc->msr, mgr->mgr.hw); idx 389 sound/pci/ctxfi/ctamixer.c unsigned int idx; idx 404 sound/pci/ctxfi/ctamixer.c err = mgr_get_resource(&mgr->mgr, 1, &idx); idx 408 sound/pci/ctxfi/ctamixer.c sum->idx[i] = idx; idx 428 sound/pci/ctxfi/ctamixer.c mgr_put_resource(&mgr->mgr, 1, sum->idx[i]); idx 442 sound/pci/ctxfi/ctamixer.c mgr_put_resource(&mgr->mgr, 1, sum->idx[i]); idx 25 sound/pci/ctxfi/ctamixer.h unsigned char idx[8]; idx 54 sound/pci/ctxfi/ctamixer.h unsigned char idx[8]; idx 59 sound/pci/ctxfi/ctdaio.c return rsc->conj = rsc->idx; idx 107 sound/pci/ctxfi/cthardware.h int (*src_commit_write)(struct hw *hw, unsigned int idx, void *blk); idx 108 sound/pci/ctxfi/cthardware.h int (*src_get_ca)(struct hw *hw, unsigned int idx, void *blk); idx 114 sound/pci/ctxfi/cthardware.h int (*src_mgr_enbs_src)(void *blk, unsigned int idx); idx 116 sound/pci/ctxfi/cthardware.h int (*src_mgr_enb_src)(void *blk, unsigned int idx); idx 118 sound/pci/ctxfi/cthardware.h int (*src_mgr_dsb_src)(void *blk, unsigned int idx); idx 143 sound/pci/ctxfi/cthardware.h int (*amixer_commit_write)(struct hw *hw, unsigned int idx, void *blk); idx 156 sound/pci/ctxfi/cthardware.h int (*dai_commit_write)(struct hw *hw, unsigned int idx, void *blk); idx 160 sound/pci/ctxfi/cthardware.h int (*dao_commit_write)(struct hw *hw, unsigned int idx, void *blk); idx 165 sound/pci/ctxfi/cthardware.h int (*daio_mgr_enb_dai)(void *blk, unsigned int idx); idx 166 sound/pci/ctxfi/cthardware.h int (*daio_mgr_dsb_dai)(void *blk, unsigned int idx); idx 167 sound/pci/ctxfi/cthardware.h int (*daio_mgr_enb_dao)(void *blk, unsigned int idx); idx 168 sound/pci/ctxfi/cthardware.h int (*daio_mgr_dsb_dao)(void *blk, unsigned int idx); idx 169 sound/pci/ctxfi/cthardware.h int (*daio_mgr_dao_init)(void *blk, unsigned int idx, idx 134 sound/pci/ctxfi/cthw20k1.c unsigned int idx; idx 359 sound/pci/ctxfi/cthw20k1.c static int src_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 367 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCUPZ+idx*0x100+i*0x4, 0); idx 370 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCDN0Z+idx*0x100+i*0x4, 0); idx 373 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCDN1Z+idx*0x100+i*0x4, 0); idx 382 sound/pci/ctxfi/cthw20k1.c unsigned int pm_idx = src_param_pitch_mixer(idx); idx 389 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCSA+idx*0x100, ctl->sa); idx 393 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCLA+idx*0x100, ctl->la); idx 397 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCCA+idx*0x100, ctl->ca); idx 402 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCCF+idx*0x100, 0x0); idx 405 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCCCR+idx*0x100, ctl->ccr); idx 409 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCCTL+idx*0x100, ctl->ctl); idx 416 sound/pci/ctxfi/cthw20k1.c static int src_get_ca(struct hw *hw, unsigned int idx, void *blk) idx 420 sound/pci/ctxfi/cthw20k1.c ctl->ca = hw_read_20kx(hw, SRCCA+idx*0x100); idx 436 sound/pci/ctxfi/cthw20k1.c static int src_mgr_enbs_src(void *blk, unsigned int idx) idx 440 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); idx 444 sound/pci/ctxfi/cthw20k1.c static int src_mgr_enb_src(void *blk, unsigned int idx) idx 446 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); idx 447 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); idx 451 sound/pci/ctxfi/cthw20k1.c static int src_mgr_dsb_src(void *blk, unsigned int idx) idx 453 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] &= ~(0x1 << (idx%32)); idx 454 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); idx 554 sound/pci/ctxfi/cthw20k1.c ctl->srcimap.idx = addr; idx 564 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRCIMAP+ctl->srcimap.idx*0x100, idx 663 sound/pci/ctxfi/cthw20k1.c static int amixer_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 668 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, AMOPLO+idx*8, ctl->amoplo); idx 670 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, AMOPHI+idx*8, ctl->amophi); idx 781 sound/pci/ctxfi/cthw20k1.c unsigned int idx; idx 878 sound/pci/ctxfi/cthw20k1.c static int dai_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 883 sound/pci/ctxfi/cthw20k1.c if (idx < 4) { idx 885 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SRTSCTL+0x4*idx, ctl->srtctl); idx 924 sound/pci/ctxfi/cthw20k1.c static int dao_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 929 sound/pci/ctxfi/cthw20k1.c if (idx < 4) { idx 931 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, SPOS+0x4*idx, ctl->spos); idx 966 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_enb_dai(void *blk, unsigned int idx) idx 970 sound/pci/ctxfi/cthw20k1.c if (idx < 4) { idx 972 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spictl, SPICTL_EN << (idx*8), 1); idx 973 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.spictl |= (0x1 << idx); idx 976 sound/pci/ctxfi/cthw20k1.c idx %= 4; idx 977 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->i2sctl, I2SCTL_EI << (idx*8), 1); idx 978 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.i2sictl |= (0x1 << idx); idx 983 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_dsb_dai(void *blk, unsigned int idx) idx 987 sound/pci/ctxfi/cthw20k1.c if (idx < 4) { idx 989 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spictl, SPICTL_EN << (idx*8), 0); idx 990 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.spictl |= (0x1 << idx); idx 993 sound/pci/ctxfi/cthw20k1.c idx %= 4; idx 994 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->i2sctl, I2SCTL_EI << (idx*8), 0); idx 995 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.i2sictl |= (0x1 << idx); idx 1000 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_enb_dao(void *blk, unsigned int idx) idx 1004 sound/pci/ctxfi/cthw20k1.c if (idx < 4) { idx 1006 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_OE << (idx*8), 1); idx 1007 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.spoctl |= (0x1 << idx); idx 1010 sound/pci/ctxfi/cthw20k1.c idx %= 4; idx 1011 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->i2sctl, I2SCTL_EA << (idx*8), 1); idx 1012 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.i2soctl |= (0x1 << idx); idx 1017 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_dsb_dao(void *blk, unsigned int idx) idx 1021 sound/pci/ctxfi/cthw20k1.c if (idx < 4) { idx 1023 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_OE << (idx*8), 0); idx 1024 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.spoctl |= (0x1 << idx); idx 1027 sound/pci/ctxfi/cthw20k1.c idx %= 4; idx 1028 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->i2sctl, I2SCTL_EA << (idx*8), 0); idx 1029 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.i2soctl |= (0x1 << idx); idx 1034 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_dao_init(void *blk, unsigned int idx, unsigned int conf) idx 1038 sound/pci/ctxfi/cthw20k1.c if (idx < 4) { idx 1042 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_SR << (idx*8), 3); idx 1045 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_SR << (idx*8), 0); idx 1048 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_SR << (idx*8), 1); idx 1051 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_SR << (idx*8), 2); idx 1056 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_LIV << (idx*8), idx 1058 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_RIV << (idx*8), idx 1060 sound/pci/ctxfi/cthw20k1.c set_field(&ctl->spoctl, SPOCTL_OS << (idx*8), idx 1063 sound/pci/ctxfi/cthw20k1.c ctl->dirty.bf.spoctl |= (0x1 << idx); idx 1093 sound/pci/ctxfi/cthw20k1.c ctl->daoimap.idx = addr; idx 1131 sound/pci/ctxfi/cthw20k1.c hw_write_20kx(hw, DAOIMAP+ctl->daoimap.idx*4, idx 134 sound/pci/ctxfi/cthw20k2.c unsigned int idx; idx 359 sound/pci/ctxfi/cthw20k2.c static int src_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 367 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_UPZ+idx*0x100+i*0x4, 0); idx 370 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_DN0Z+idx*0x100+i*0x4, 0); idx 373 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_DN1Z+idx*0x100+i*0x4, 0); idx 382 sound/pci/ctxfi/cthw20k2.c unsigned int pm_idx = src_param_pitch_mixer(idx); idx 389 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_SA+idx*0x100, ctl->sa); idx 393 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_LA+idx*0x100, ctl->la); idx 397 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_CA+idx*0x100, ctl->ca); idx 402 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_CF+idx*0x100, 0x0); idx 405 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_CCR+idx*0x100, ctl->ccr); idx 409 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_CTL+idx*0x100, ctl->ctl); idx 416 sound/pci/ctxfi/cthw20k2.c static int src_get_ca(struct hw *hw, unsigned int idx, void *blk) idx 420 sound/pci/ctxfi/cthw20k2.c ctl->ca = hw_read_20kx(hw, SRC_CA+idx*0x100); idx 436 sound/pci/ctxfi/cthw20k2.c static int src_mgr_enbs_src(void *blk, unsigned int idx) idx 438 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enbsa |= (0x1 << ((idx%128)/4)); idx 440 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); idx 444 sound/pci/ctxfi/cthw20k2.c static int src_mgr_enb_src(void *blk, unsigned int idx) idx 446 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); idx 447 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); idx 451 sound/pci/ctxfi/cthw20k2.c static int src_mgr_dsb_src(void *blk, unsigned int idx) idx 453 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] &= ~(0x1 << (idx%32)); idx 454 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); idx 552 sound/pci/ctxfi/cthw20k2.c ((struct srcimp_mgr_ctrl_blk *)blk)->srcimap.idx = addr; idx 562 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, SRC_IMAP+ctl->srcimap.idx*0x100, idx 665 sound/pci/ctxfi/cthw20k2.c static int amixer_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 670 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, MIXER_AMOPLO+idx*8, ctl->amoplo); idx 672 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, MIXER_AMOPHI+idx*8, ctl->amophi); idx 757 sound/pci/ctxfi/cthw20k2.c unsigned int idx; idx 868 sound/pci/ctxfi/cthw20k2.c static int dai_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 873 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, AUDIO_IO_RX_SRT_CTL+0x40*idx, ctl->srt); idx 908 sound/pci/ctxfi/cthw20k2.c static int dao_commit_write(struct hw *hw, unsigned int idx, void *blk) idx 913 sound/pci/ctxfi/cthw20k2.c if (idx < 4) { idx 915 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, AUDIO_IO_TX_CSTAT_L+0x40*idx, idx 951 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_enb_dai(void *blk, unsigned int idx) idx 955 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->rxctl[idx], ARXCTL_EN, 1); idx 956 sound/pci/ctxfi/cthw20k2.c ctl->dirty.bf.arxctl |= (0x1 << idx); idx 960 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_dsb_dai(void *blk, unsigned int idx) idx 964 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->rxctl[idx], ARXCTL_EN, 0); idx 966 sound/pci/ctxfi/cthw20k2.c ctl->dirty.bf.arxctl |= (0x1 << idx); idx 970 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_enb_dao(void *blk, unsigned int idx) idx 974 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_EN, 1); idx 975 sound/pci/ctxfi/cthw20k2.c ctl->dirty.bf.atxctl |= (0x1 << idx); idx 979 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_dsb_dao(void *blk, unsigned int idx) idx 983 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_EN, 0); idx 984 sound/pci/ctxfi/cthw20k2.c ctl->dirty.bf.atxctl |= (0x1 << idx); idx 988 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_dao_init(void *blk, unsigned int idx, unsigned int conf) idx 992 sound/pci/ctxfi/cthw20k2.c if (idx < 4) { idx 996 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_NUC, 0); idx 999 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_NUC, 1); idx 1002 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_NUC, 2); idx 1005 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_NUC, 3); idx 1011 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_CD, (!(conf & 0x7))); idx 1013 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_LIV, (conf >> 4) & 0x1); idx 1015 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_RIV, (conf >> 4) & 0x1); idx 1016 sound/pci/ctxfi/cthw20k2.c set_field(&ctl->txctl[idx], ATXCTL_RAW, idx 1018 sound/pci/ctxfi/cthw20k2.c ctl->dirty.bf.atxctl |= (0x1 << idx); idx 1046 sound/pci/ctxfi/cthw20k2.c ((struct daio_mgr_ctrl_blk *)blk)->daoimap.idx = addr; idx 1072 sound/pci/ctxfi/cthw20k2.c hw_write_20kx(hw, AUDIO_IO_AIM+ctl->daoimap.idx*4, idx 58 sound/pci/ctxfi/ctresource.c static int put_resource(u8 *rscs, unsigned int multi, unsigned int idx) idx 63 sound/pci/ctxfi/ctresource.c for (n = multi, i = idx; n > 0; n--) { idx 87 sound/pci/ctxfi/ctresource.c int mgr_put_resource(struct rsc_mgr *mgr, unsigned int n, unsigned int idx) idx 89 sound/pci/ctxfi/ctresource.c put_resource(mgr->rscs, n, idx); idx 123 sound/pci/ctxfi/ctresource.c return rsc->conj = rsc->idx; idx 134 sound/pci/ctxfi/ctresource.c rsc_init(struct rsc *rsc, u32 idx, enum RSCTYP type, u32 msr, struct hw *hw) idx 138 sound/pci/ctxfi/ctresource.c rsc->idx = idx; idx 139 sound/pci/ctxfi/ctresource.c rsc->conj = idx; idx 198 sound/pci/ctxfi/ctresource.c rsc->idx = rsc->conj = 0; idx 32 sound/pci/ctxfi/ctresource.h u32 idx:12; /* The index of a resource */ idx 50 sound/pci/ctxfi/ctresource.h rsc_init(struct rsc *rsc, u32 idx, enum RSCTYP type, u32 msr, struct hw *hw); idx 67 sound/pci/ctxfi/ctresource.h int mgr_put_resource(struct rsc_mgr *mgr, unsigned int n, unsigned int idx); idx 356 sound/pci/ctxfi/ctsrc.c src_rsc_init(struct src *src, u32 idx, idx 365 sound/pci/ctxfi/ctsrc.c err = rsc_init(&p->rsc, idx + i, SRC, desc->msr, mgr->mgr.hw); idx 414 sound/pci/ctxfi/ctsrc.c unsigned int idx = SRC_RESOURCE_NUM; idx 424 sound/pci/ctxfi/ctsrc.c err = mgr_get_resource(&mgr->mgr, desc->multi, &idx); idx 426 sound/pci/ctxfi/ctsrc.c err = mgr_get_resource(&mgr->mgr, 1, &idx); idx 446 sound/pci/ctxfi/ctsrc.c err = src_rsc_init(src, idx, desc, mgr); idx 459 sound/pci/ctxfi/ctsrc.c mgr_put_resource(&mgr->mgr, desc->multi, idx); idx 461 sound/pci/ctxfi/ctsrc.c mgr_put_resource(&mgr->mgr, 1, idx); idx 596 sound/pci/ctxfi/ctsrc.c return rsc->idx = container_of(rsc, struct srcimp, rsc)->idx[0]; idx 602 sound/pci/ctxfi/ctsrc.c return container_of(rsc, struct srcimp, rsc)->idx[rsc->conj]; idx 607 sound/pci/ctxfi/ctsrc.c return container_of(rsc, struct srcimp, rsc)->idx[rsc->conj]; idx 672 sound/pci/ctxfi/ctsrc.c err = rsc_init(&srcimp->rsc, srcimp->idx[0], idx 715 sound/pci/ctxfi/ctsrc.c unsigned int idx; idx 730 sound/pci/ctxfi/ctsrc.c err = mgr_get_resource(&mgr->mgr, 1, &idx); idx 734 sound/pci/ctxfi/ctsrc.c srcimp->idx[i] = idx; idx 754 sound/pci/ctxfi/ctsrc.c mgr_put_resource(&mgr->mgr, 1, srcimp->idx[i]); idx 768 sound/pci/ctxfi/ctsrc.c mgr_put_resource(&mgr->mgr, 1, srcimp->idx[i]); idx 105 sound/pci/ctxfi/ctsrc.h unsigned char idx[8]; idx 1789 sound/pci/emu10k1/emu10k1_main.c int idx, err; idx 2011 sound/pci/emu10k1/emu10k1_main.c for (idx = 0; idx < (emu->address_mode ? MAXPAGES1 : MAXPAGES0); idx++) idx 2012 sound/pci/emu10k1/emu10k1_main.c ((u32 *)emu->ptb_pages.area)[idx] = cpu_to_le32(silent_page | idx); idx 2015 sound/pci/emu10k1/emu10k1_main.c for (idx = 0; idx < NUM_G; idx++) { idx 2016 sound/pci/emu10k1/emu10k1_main.c emu->voices[idx].emu = emu; idx 2017 sound/pci/emu10k1/emu10k1_main.c emu->voices[idx].number = idx; idx 1114 sound/pci/emu10k1/emu10k1x.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 1116 sound/pci/emu10k1/emu10k1x.c ucontrol->value.iec958.status[0] = (emu->spdif_bits[idx] >> 0) & 0xff; idx 1117 sound/pci/emu10k1/emu10k1x.c ucontrol->value.iec958.status[1] = (emu->spdif_bits[idx] >> 8) & 0xff; idx 1118 sound/pci/emu10k1/emu10k1x.c ucontrol->value.iec958.status[2] = (emu->spdif_bits[idx] >> 16) & 0xff; idx 1119 sound/pci/emu10k1/emu10k1x.c ucontrol->value.iec958.status[3] = (emu->spdif_bits[idx] >> 24) & 0xff; idx 1137 sound/pci/emu10k1/emu10k1x.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 1145 sound/pci/emu10k1/emu10k1x.c change = val != emu->spdif_bits[idx]; idx 1147 sound/pci/emu10k1/emu10k1x.c snd_emu10k1x_ptr_write(emu, SPCS0 + idx, 0, val); idx 1148 sound/pci/emu10k1/emu10k1x.c emu->spdif_bits[idx] = val; idx 1198 sound/pci/emu10k1/emu10k1x.c static inline unsigned char mpu401_read(struct emu10k1x *emu, struct emu10k1x_midi *mpu, int idx) idx 1200 sound/pci/emu10k1/emu10k1x.c return (unsigned char)snd_emu10k1x_ptr_read(emu, mpu->port + idx, 0); idx 1203 sound/pci/emu10k1/emu10k1x.c static inline void mpu401_write(struct emu10k1x *emu, struct emu10k1x_midi *mpu, int data, int idx) idx 1205 sound/pci/emu10k1/emu10k1x.c snd_emu10k1x_ptr_write(emu, mpu->port + idx, 0, data); idx 677 sound/pci/emu10k1/emufx.c int idx, bool in_kernel) idx 683 sound/pci/emu10k1/emufx.c memcpy(gctl, (__force void *)&_gctl[idx], sizeof(*gctl)); idx 684 sound/pci/emu10k1/emufx.c else if (copy_from_user(gctl, &_gctl[idx], sizeof(*gctl))) idx 691 sound/pci/emu10k1/emufx.c memcpy(gctl, (__force void *)&octl[idx], sizeof(*octl)); idx 692 sound/pci/emu10k1/emufx.c else if (copy_from_user(gctl, &octl[idx], sizeof(*octl))) idx 701 sound/pci/emu10k1/emufx.c int idx) idx 706 sound/pci/emu10k1/emufx.c return copy_to_user(&_gctl[idx], gctl, sizeof(*gctl)); idx 709 sound/pci/emu10k1/emufx.c return copy_to_user(&octl[idx], gctl, sizeof(*octl)); idx 2486 sound/pci/emu10k1/emufx.c static void copy_string(char *dst, char *src, char *null, int idx) idx 2489 sound/pci/emu10k1/emufx.c sprintf(dst, "%s %02X", null, idx); idx 43 sound/pci/emu10k1/emumixer.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 47 sound/pci/emu10k1/emumixer.c if (idx >= 3) idx 50 sound/pci/emu10k1/emumixer.c ucontrol->value.iec958.status[0] = (emu->spdif_bits[idx] >> 0) & 0xff; idx 51 sound/pci/emu10k1/emumixer.c ucontrol->value.iec958.status[1] = (emu->spdif_bits[idx] >> 8) & 0xff; idx 52 sound/pci/emu10k1/emumixer.c ucontrol->value.iec958.status[2] = (emu->spdif_bits[idx] >> 16) & 0xff; idx 53 sound/pci/emu10k1/emumixer.c ucontrol->value.iec958.status[3] = (emu->spdif_bits[idx] >> 24) & 0xff; idx 1144 sound/pci/emu10k1/emumixer.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 1150 sound/pci/emu10k1/emumixer.c if (idx >= 3) idx 1157 sound/pci/emu10k1/emumixer.c change = val != emu->spdif_bits[idx]; idx 1159 sound/pci/emu10k1/emumixer.c snd_emu10k1_ptr_write(emu, SPCS0 + idx, 0, val); idx 1160 sound/pci/emu10k1/emumixer.c emu->spdif_bits[idx] = val; idx 1234 sound/pci/emu10k1/emumixer.c int voice, idx; idx 1240 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < num_efx; idx++) idx 1241 sound/pci/emu10k1/emumixer.c ucontrol->value.integer.value[(voice * num_efx) + idx] = idx 1242 sound/pci/emu10k1/emumixer.c mix->send_routing[voice][idx] & mask; idx 1254 sound/pci/emu10k1/emumixer.c int change = 0, voice, idx, val; idx 1260 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < num_efx; idx++) { idx 1261 sound/pci/emu10k1/emumixer.c val = ucontrol->value.integer.value[(voice * num_efx) + idx] & mask; idx 1262 sound/pci/emu10k1/emumixer.c if (mix->send_routing[voice][idx] != val) { idx 1263 sound/pci/emu10k1/emumixer.c mix->send_routing[voice][idx] = val; idx 1310 sound/pci/emu10k1/emumixer.c int idx; idx 1314 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < 3*num_efx; idx++) idx 1315 sound/pci/emu10k1/emumixer.c ucontrol->value.integer.value[idx] = mix->send_volume[idx/num_efx][idx%num_efx]; idx 1327 sound/pci/emu10k1/emumixer.c int change = 0, idx, val; idx 1331 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < 3*num_efx; idx++) { idx 1332 sound/pci/emu10k1/emumixer.c val = ucontrol->value.integer.value[idx] & 255; idx 1333 sound/pci/emu10k1/emumixer.c if (mix->send_volume[idx/num_efx][idx%num_efx] != val) { idx 1334 sound/pci/emu10k1/emumixer.c mix->send_volume[idx/num_efx][idx%num_efx] = val; idx 1380 sound/pci/emu10k1/emumixer.c int idx; idx 1383 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < 3; idx++) idx 1384 sound/pci/emu10k1/emumixer.c ucontrol->value.integer.value[idx] = mix->attn[idx]; idx 1396 sound/pci/emu10k1/emumixer.c int change = 0, idx, val; idx 1399 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < 3; idx++) { idx 1400 sound/pci/emu10k1/emumixer.c val = ucontrol->value.integer.value[idx] & 0xffff; idx 1401 sound/pci/emu10k1/emumixer.c if (mix->attn[idx] != val) { idx 1402 sound/pci/emu10k1/emumixer.c mix->attn[idx] = val; idx 1448 sound/pci/emu10k1/emumixer.c int idx; idx 1453 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < num_efx; idx++) idx 1454 sound/pci/emu10k1/emumixer.c ucontrol->value.integer.value[idx] = idx 1455 sound/pci/emu10k1/emumixer.c mix->send_routing[0][idx] & mask; idx 1467 sound/pci/emu10k1/emumixer.c int change = 0, idx, val; idx 1472 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < num_efx; idx++) { idx 1473 sound/pci/emu10k1/emumixer.c val = ucontrol->value.integer.value[idx] & mask; idx 1474 sound/pci/emu10k1/emumixer.c if (mix->send_routing[0][idx] != val) { idx 1475 sound/pci/emu10k1/emumixer.c mix->send_routing[0][idx] = val; idx 1518 sound/pci/emu10k1/emumixer.c int idx; idx 1522 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < num_efx; idx++) idx 1523 sound/pci/emu10k1/emumixer.c ucontrol->value.integer.value[idx] = mix->send_volume[0][idx]; idx 1535 sound/pci/emu10k1/emumixer.c int change = 0, idx, val; idx 1539 sound/pci/emu10k1/emumixer.c for (idx = 0; idx < num_efx; idx++) { idx 1540 sound/pci/emu10k1/emumixer.c val = ucontrol->value.integer.value[idx] & 255; idx 1541 sound/pci/emu10k1/emumixer.c if (mix->send_volume[0][idx] != val) { idx 1542 sound/pci/emu10k1/emumixer.c mix->send_volume[0][idx] = val; idx 16 sound/pci/emu10k1/emumpu401.c struct snd_emu10k1_midi *mpu, int idx) idx 19 sound/pci/emu10k1/emumpu401.c return (unsigned char)snd_emu10k1_ptr_read(emu, mpu->port + idx, 0); idx 21 sound/pci/emu10k1/emumpu401.c return inb(emu->port + mpu->port + idx); idx 25 sound/pci/emu10k1/emumpu401.c struct snd_emu10k1_midi *mpu, int data, int idx) idx 28 sound/pci/emu10k1/emumpu401.c snd_emu10k1_ptr_write(emu, mpu->port + idx, 0, data); idx 30 sound/pci/emu10k1/emumpu401.c outb(data, emu->port + mpu->port + idx); idx 588 sound/pci/emu10k1/emupcm.c int idx; idx 609 sound/pci/emu10k1/emupcm.c for (idx = 0; idx < 31; idx++) { idx 610 sound/pci/emu10k1/emupcm.c if (capture_period_sizes[idx] == epcm->capture_bufsize) { idx 611 sound/pci/emu10k1/emupcm.c epcm->capture_bs_val = idx + 1; idx 1041 sound/pci/emu10k1/emupcm.c static void snd_emu10k1_pcm_mixer_notify1(struct snd_emu10k1 *emu, struct snd_kcontrol *kctl, int idx, int activate) idx 1048 sound/pci/emu10k1/emupcm.c kctl->vd[idx].access &= ~SNDRV_CTL_ELEM_ACCESS_INACTIVE; idx 1050 sound/pci/emu10k1/emupcm.c kctl->vd[idx].access |= SNDRV_CTL_ELEM_ACCESS_INACTIVE; idx 1053 sound/pci/emu10k1/emupcm.c snd_ctl_build_ioff(&id, kctl, idx)); idx 1056 sound/pci/emu10k1/emupcm.c static void snd_emu10k1_pcm_mixer_notify(struct snd_emu10k1 *emu, int idx, int activate) idx 1058 sound/pci/emu10k1/emupcm.c snd_emu10k1_pcm_mixer_notify1(emu, emu->ctl_send_routing, idx, activate); idx 1059 sound/pci/emu10k1/emupcm.c snd_emu10k1_pcm_mixer_notify1(emu, emu->ctl_send_volume, idx, activate); idx 1060 sound/pci/emu10k1/emupcm.c snd_emu10k1_pcm_mixer_notify1(emu, emu->ctl_attn, idx, activate); idx 1063 sound/pci/emu10k1/emupcm.c static void snd_emu10k1_pcm_efx_mixer_notify(struct snd_emu10k1 *emu, int idx, int activate) idx 1065 sound/pci/emu10k1/emupcm.c snd_emu10k1_pcm_mixer_notify1(emu, emu->ctl_efx_send_routing, idx, activate); idx 1066 sound/pci/emu10k1/emupcm.c snd_emu10k1_pcm_mixer_notify1(emu, emu->ctl_efx_send_volume, idx, activate); idx 1067 sound/pci/emu10k1/emupcm.c snd_emu10k1_pcm_mixer_notify1(emu, emu->ctl_efx_attn, idx, activate); idx 1258 sound/pci/emu10k1/emupcm.c int idx; idx 1335 sound/pci/emu10k1/emupcm.c for (idx = 0; idx < nefx; idx++) { idx 1336 sound/pci/emu10k1/emupcm.c if (emu->efx_voices_mask[idx/32] & (1 << (idx%32))) { idx 1504 sound/pci/emu10k1/emupcm.c int idx; idx 1507 sound/pci/emu10k1/emupcm.c for (idx = 0; idx < nefx; idx++) idx 1508 sound/pci/emu10k1/emupcm.c ucontrol->value.integer.value[idx] = (emu->efx_voices_mask[idx / 32] & (1 << (idx % 32))) ? 1 : 0; idx 1519 sound/pci/emu10k1/emupcm.c int change, idx; idx 1522 sound/pci/emu10k1/emupcm.c for (idx = 0, bits = 0; idx < nefx; idx++) idx 1523 sound/pci/emu10k1/emupcm.c if (ucontrol->value.integer.value[idx]) { idx 1524 sound/pci/emu10k1/emupcm.c nval[idx / 32] |= 1 << (idx % 32); idx 1528 sound/pci/emu10k1/emupcm.c for (idx = 0; idx < nefxb; idx++) idx 1529 sound/pci/emu10k1/emupcm.c if (1 << idx == bits) idx 1532 sound/pci/emu10k1/emupcm.c if (idx >= nefxb) idx 176 sound/pci/emu10k1/emuproc.c int idx; idx 185 sound/pci/emu10k1/emuproc.c for (idx = 0; idx < NUM_G; idx++) { idx 187 sound/pci/emu10k1/emuproc.c snd_emu10k1_ptr_read(emu, A_FXRT1, idx) : idx 188 sound/pci/emu10k1/emuproc.c snd_emu10k1_ptr_read(emu, FXRT, idx); idx 190 sound/pci/emu10k1/emuproc.c snd_emu10k1_ptr_read(emu, A_FXRT2, idx) : idx 194 sound/pci/emu10k1/emuproc.c idx, idx 206 sound/pci/emu10k1/emuproc.c idx, idx 214 sound/pci/emu10k1/emuproc.c for (idx = 0; idx < nefx; idx++) { idx 215 sound/pci/emu10k1/emuproc.c if (emu->efx_voices_mask[idx/32] & (1 << (idx%32))) idx 216 sound/pci/emu10k1/emuproc.c snd_iprintf(buffer, " Output %02i [%s]\n", idx, outputs[idx]); idx 219 sound/pci/emu10k1/emuproc.c for (idx = 0; idx < (emu->audigy ? 64 : 32); idx++) idx 220 sound/pci/emu10k1/emuproc.c snd_iprintf(buffer, " Output %02i [%s]\n", idx, outputs[idx]); idx 329 sound/pci/emu10k1/emuproc.c unsigned int idx; idx 345 sound/pci/emu10k1/emuproc.c for (idx = 0; idx < ((pos & 3) + count + 3) >> 2; idx++) { idx 347 sound/pci/emu10k1/emuproc.c val = snd_emu10k1_ptr_read(emu, offset + idx + (pos >> 2), 0); idx 350 sound/pci/emu10k1/emuproc.c val |= snd_emu10k1_ptr_read(emu, 0x100 + idx + (pos >> 2), 0) << 20; idx 352 sound/pci/emu10k1/emuproc.c tmp[idx] = val; idx 367 sound/pci/emu10k1/emuproc.c int idx; idx 370 sound/pci/emu10k1/emuproc.c for (idx = 0; idx < NUM_G; idx++) { idx 371 sound/pci/emu10k1/emuproc.c voice = &emu->voices[idx]; idx 373 sound/pci/emu10k1/emuproc.c idx, idx 302 sound/pci/emu10k1/memory.c int page, err, idx; idx 313 sound/pci/emu10k1/memory.c idx = runtime->period_size >= runtime->buffer_size ? idx 316 sound/pci/emu10k1/memory.c blk = search_empty(emu, runtime->dma_bytes + idx); idx 324 sound/pci/emu10k1/memory.c idx = 0; idx 325 sound/pci/emu10k1/memory.c for (page = blk->first_page; page <= blk->last_page; page++, idx++) { idx 326 sound/pci/emu10k1/memory.c unsigned long ofs = idx << PAGE_SHIFT; idx 334 sound/pci/emu10k1/memory.c "emu: failure page = %d\n", idx); idx 1738 sound/pci/ens1370.c unsigned int idx; idx 1755 sound/pci/ens1370.c for (idx = 0; idx < ES1370_CONTROLS; idx++) { idx 1756 sound/pci/ens1370.c err = snd_ctl_add(card, snd_ctl_new1(&snd_es1370_controls[idx], ensoniq)); idx 1955 sound/pci/ens1370.c int idx; idx 1984 sound/pci/ens1370.c for (idx = 0; idx < 0x80; idx++) idx 1985 sound/pci/ens1370.c snd_es1371_src_write(ensoniq, idx, 0); idx 2250 sound/pci/ens1370.c int idx; idx 2256 sound/pci/ens1370.c for (idx = 0; idx < 32; idx++) idx 1730 sound/pci/es1938.c unsigned int idx; idx 1737 sound/pci/es1938.c for (idx = 0; idx < ARRAY_SIZE(snd_es1938_controls); idx++) { idx 1739 sound/pci/es1938.c kctl = snd_ctl_new1(&snd_es1938_controls[idx], chip); idx 1740 sound/pci/es1938.c switch (idx) { idx 1772 sound/pci/es1938.c int idx, err; idx 1785 sound/pci/es1938.c for (idx = 0; idx < 5; idx++) { idx 1786 sound/pci/es1938.c if (pci_resource_start(pci, idx) == 0 || idx 1787 sound/pci/es1938.c !(pci_resource_flags(pci, idx) & IORESOURCE_IO)) { idx 233 sound/pci/fm801.c unsigned int idx; idx 235 sound/pci/fm801.c for (idx = 0; idx < iterations; idx++) { idx 245 sound/pci/fm801.c unsigned int idx; idx 247 sound/pci/fm801.c for (idx = 0; idx < iterations; idx++) { idx 355 sound/pci/fm801.c unsigned int idx; idx 357 sound/pci/fm801.c for (idx = 0; idx < ARRAY_SIZE(rates); idx++) idx 358 sound/pci/fm801.c if (rates[idx] == rate) idx 359 sound/pci/fm801.c return idx; idx 636 sound/pci/hda/hda_auto_parser.c int i, j, idx = 0; idx 645 sound/pci/hda/hda_auto_parser.c idx++; idx 647 sound/pci/hda/hda_auto_parser.c return idx; idx 673 sound/pci/hda/hda_auto_parser.c int idx = get_hp_label_index(codec, nid, cfg->hp_pins, idx 675 sound/pci/hda/hda_auto_parser.c if (idx >= 0 && indexp) idx 676 sound/pci/hda/hda_auto_parser.c *indexp = idx; idx 1271 sound/pci/hda/hda_codec.c int ch, int dir, int idx) idx 1273 sound/pci/hda/hda_codec.c unsigned int cmd = snd_hdac_regmap_encode_amp(nid, ch, dir, idx); idx 1295 sound/pci/hda/hda_codec.c int ch, int dir, int idx, int mask, int val) idx 1297 sound/pci/hda/hda_codec.c unsigned int cmd = encode_amp(codec, nid, ch, dir, idx); idx 1316 sound/pci/hda/hda_codec.c int direction, int idx, int mask, int val) idx 1324 sound/pci/hda/hda_codec.c idx, mask, val); idx 1344 sound/pci/hda/hda_codec.c int dir, int idx, int mask, int val) idx 1346 sound/pci/hda/hda_codec.c unsigned int cmd = encode_amp(codec, nid, ch, dir, idx); idx 1366 sound/pci/hda/hda_codec.c int dir, int idx, int mask, int val) idx 1374 sound/pci/hda/hda_codec.c idx, mask, val); idx 1424 sound/pci/hda/hda_codec.c int ch, int dir, int idx, unsigned int ofs) idx 1427 sound/pci/hda/hda_codec.c val = snd_hda_codec_amp_read(codec, nid, ch, dir, idx); idx 1438 sound/pci/hda/hda_codec.c int ch, int dir, int idx, unsigned int ofs, idx 1449 sound/pci/hda/hda_codec.c return snd_hda_codec_amp_update(codec, nid, ch, dir, idx, idx 1468 sound/pci/hda/hda_codec.c int idx = get_amp_index(kcontrol); idx 1473 sound/pci/hda/hda_codec.c *valp++ = read_amp_value(codec, nid, 0, dir, idx, ofs); idx 1475 sound/pci/hda/hda_codec.c *valp = read_amp_value(codec, nid, 1, dir, idx, ofs); idx 1495 sound/pci/hda/hda_codec.c int idx = get_amp_index(kcontrol); idx 1501 sound/pci/hda/hda_codec.c change = update_amp_value(codec, nid, 0, dir, idx, ofs, *valp); idx 1505 sound/pci/hda/hda_codec.c change |= update_amp_value(codec, nid, 1, dir, idx, ofs, *valp); idx 1588 sound/pci/hda/hda_codec.c find_mixer_ctl(struct hda_codec *codec, const char *name, int dev, int idx) idx 1594 sound/pci/hda/hda_codec.c id.index = idx; idx 1618 sound/pci/hda/hda_codec.c int i, idx; idx 1620 sound/pci/hda/hda_codec.c for (i = 0, idx = start_idx; i < 16; i++, idx++) { idx 1621 sound/pci/hda/hda_codec.c if (!find_mixer_ctl(codec, name, 0, idx)) idx 1622 sound/pci/hda/hda_codec.c return idx; idx 2123 sound/pci/hda/hda_codec.c int idx = get_amp_index(kcontrol); idx 2127 sound/pci/hda/hda_codec.c *valp++ = (snd_hda_codec_amp_read(codec, nid, 0, dir, idx) & idx 2130 sound/pci/hda/hda_codec.c *valp = (snd_hda_codec_amp_read(codec, nid, 1, dir, idx) & idx 2151 sound/pci/hda/hda_codec.c int idx = get_amp_index(kcontrol); idx 2156 sound/pci/hda/hda_codec.c change = snd_hda_codec_amp_update(codec, nid, 0, dir, idx, idx 2162 sound/pci/hda/hda_codec.c change |= snd_hda_codec_amp_update(codec, nid, 1, dir, idx, idx 2207 sound/pci/hda/hda_codec.c int idx = kcontrol->private_value; idx 2210 sound/pci/hda/hda_codec.c if (WARN_ON(codec->spdif_out.used <= idx)) idx 2213 sound/pci/hda/hda_codec.c spdif = snd_array_elem(&codec->spdif_out, idx); idx 2313 sound/pci/hda/hda_codec.c int idx = kcontrol->private_value; idx 2319 sound/pci/hda/hda_codec.c if (WARN_ON(codec->spdif_out.used <= idx)) idx 2322 sound/pci/hda/hda_codec.c spdif = snd_array_elem(&codec->spdif_out, idx); idx 2344 sound/pci/hda/hda_codec.c int idx = kcontrol->private_value; idx 2347 sound/pci/hda/hda_codec.c if (WARN_ON(codec->spdif_out.used <= idx)) idx 2350 sound/pci/hda/hda_codec.c spdif = snd_array_elem(&codec->spdif_out, idx); idx 2371 sound/pci/hda/hda_codec.c int idx = kcontrol->private_value; idx 2377 sound/pci/hda/hda_codec.c if (WARN_ON(codec->spdif_out.used <= idx)) idx 2380 sound/pci/hda/hda_codec.c spdif = snd_array_elem(&codec->spdif_out, idx); idx 2444 sound/pci/hda/hda_codec.c int idx = 0; idx 2452 sound/pci/hda/hda_codec.c idx = spdif_index; idx 2467 sound/pci/hda/hda_codec.c idx = find_empty_mixer_ctl_idx(codec, "IEC958 Playback Switch", idx); idx 2468 sound/pci/hda/hda_codec.c if (idx < 0) { idx 2479 sound/pci/hda/hda_codec.c kctl->id.index = idx; idx 2522 sound/pci/hda/hda_codec.c void snd_hda_spdif_ctls_unassign(struct hda_codec *codec, int idx) idx 2526 sound/pci/hda/hda_codec.c if (WARN_ON(codec->spdif_out.used <= idx)) idx 2529 sound/pci/hda/hda_codec.c spdif = snd_array_elem(&codec->spdif_out, idx); idx 2543 sound/pci/hda/hda_codec.c void snd_hda_spdif_ctls_assign(struct hda_codec *codec, int idx, hda_nid_t nid) idx 2548 sound/pci/hda/hda_codec.c if (WARN_ON(codec->spdif_out.used <= idx)) idx 2551 sound/pci/hda/hda_codec.c spdif = snd_array_elem(&codec->spdif_out, idx); idx 2694 sound/pci/hda/hda_codec.c int idx; idx 2696 sound/pci/hda/hda_codec.c idx = find_empty_mixer_ctl_idx(codec, "IEC958 Capture Switch", 0); idx 2697 sound/pci/hda/hda_codec.c if (idx < 0) { idx 3339 sound/pci/hda/hda_codec.c int addr = 0, idx = 0; idx 3348 sound/pci/hda/hda_codec.c if (idx > 0) idx 3349 sound/pci/hda/hda_codec.c kctl->id.index = idx; idx 3359 sound/pci/hda/hda_codec.c else if (!idx && !knew->index) { idx 3360 sound/pci/hda/hda_codec.c idx = find_empty_mixer_ctl_idx(codec, idx 3362 sound/pci/hda/hda_codec.c if (idx <= 0) idx 3440 sound/pci/hda/hda_codec.c p->idx); idx 3500 sound/pci/hda/hda_codec.c unsigned int idx; idx 3504 sound/pci/hda/hda_codec.c idx = ucontrol->value.enumerated.item[0]; idx 3505 sound/pci/hda/hda_codec.c if (idx >= imux->num_items) idx 3506 sound/pci/hda/hda_codec.c idx = imux->num_items - 1; idx 3507 sound/pci/hda/hda_codec.c if (*cur_val == idx) idx 3510 sound/pci/hda/hda_codec.c imux->items[idx].index); idx 3511 sound/pci/hda/hda_codec.c *cur_val = idx; idx 22 sound/pci/hda/hda_controller_trace.h __field( int, idx ) idx 28 sound/pci/hda/hda_controller_trace.h __entry->idx = (dev)->core.index; idx 32 sound/pci/hda/hda_controller_trace.h TP_printk("[%d:%d] cmd=%d", __entry->card, __entry->idx, __entry->cmd) idx 43 sound/pci/hda/hda_controller_trace.h __field( int, idx ) idx 50 sound/pci/hda/hda_controller_trace.h __entry->idx = (dev)->core.index; idx 55 sound/pci/hda/hda_controller_trace.h TP_printk("[%d:%d] pos=%u, delay=%u", __entry->card, __entry->idx, __entry->pos, __entry->delay) idx 285 sound/pci/hda/hda_generic.c ssize_t idx; idx 289 sound/pci/hda/hda_generic.c idx = path - array; idx 290 sound/pci/hda/hda_generic.c if (idx < 0 || idx >= spec->paths.used) idx 292 sound/pci/hda/hda_generic.c return idx + 1; idx 302 sound/pci/hda/hda_generic.c struct nid_path *snd_hda_get_path_from_idx(struct hda_codec *codec, int idx) idx 306 sound/pci/hda/hda_generic.c if (idx <= 0 || idx > spec->paths.used) idx 308 sound/pci/hda/hda_generic.c return snd_array_elem(&spec->paths, idx - 1); idx 355 sound/pci/hda/hda_generic.c int dir, int idx, int type) idx 357 sound/pci/hda/hda_generic.c unsigned int val = HDA_COMPOSE_AMP_VAL(nid, 3, idx, dir); idx 422 sound/pci/hda/hda_generic.c path->idx[path->depth + 1] = i; idx 498 sound/pci/hda/hda_generic.c static void invalidate_nid_path(struct hda_codec *codec, int idx) idx 500 sound/pci/hda/hda_generic.c struct nid_path *path = snd_hda_get_path_from_idx(codec, idx); idx 596 sound/pci/hda/hda_generic.c static bool has_amp_in(struct hda_codec *codec, struct nid_path *path, int idx) idx 598 sound/pci/hda/hda_generic.c hda_nid_t nid = path->path[idx]; idx 604 sound/pci/hda/hda_generic.c if (type == AC_WID_PIN && idx > 0) /* only for input pins */ idx 610 sound/pci/hda/hda_generic.c static bool has_amp_out(struct hda_codec *codec, struct nid_path *path, int idx) idx 612 sound/pci/hda/hda_generic.c hda_nid_t nid = path->path[idx]; idx 618 sound/pci/hda/hda_generic.c if (type == AC_WID_PIN && !idx) /* only for output pins */ idx 625 sound/pci/hda/hda_generic.c unsigned int dir, unsigned int idx) idx 648 sound/pci/hda/hda_generic.c if (dir == HDA_OUTPUT || idx == -1 || idx 649 sound/pci/hda/hda_generic.c path->idx[i] == idx) idx 698 sound/pci/hda/hda_generic.c static void init_amp(struct hda_codec *codec, hda_nid_t nid, int dir, int idx) idx 704 sound/pci/hda/hda_generic.c snd_hda_codec_amp_init_stereo(codec, nid, dir, idx, 0xff, val); idx 706 sound/pci/hda/hda_generic.c snd_hda_codec_amp_init(codec, nid, 0, dir, idx, 0xff, val); idx 710 sound/pci/hda/hda_generic.c static int update_amp(struct hda_codec *codec, hda_nid_t nid, int dir, int idx, idx 714 sound/pci/hda/hda_generic.c return snd_hda_codec_amp_stereo(codec, nid, dir, idx, idx 717 sound/pci/hda/hda_generic.c return snd_hda_codec_amp_update(codec, nid, 0, dir, idx, idx 725 sound/pci/hda/hda_generic.c hda_nid_t nid, int dir, int idx, idx 731 sound/pci/hda/hda_generic.c if (is_ctl_associated(codec, nid, dir, idx, NID_PATH_MUTE_CTL)) idx 735 sound/pci/hda/hda_generic.c if (is_ctl_associated(codec, nid, dir, idx, NID_PATH_VOL_CTL) || idx 736 sound/pci/hda/hda_generic.c is_ctl_associated(codec, nid, dir, idx, NID_PATH_BOOST_CTL)) idx 743 sound/pci/hda/hda_generic.c int idx, int idx_to_check, bool enable) idx 755 sound/pci/hda/hda_generic.c update_amp(codec, nid, dir, idx, mask, val); idx 759 sound/pci/hda/hda_generic.c int dir, int idx, int idx_to_check, idx 765 sound/pci/hda/hda_generic.c activate_amp(codec, nid, dir, idx, idx_to_check, enable); idx 781 sound/pci/hda/hda_generic.c int n, nums, idx; idx 792 sound/pci/hda/hda_generic.c idx = 0; idx 794 sound/pci/hda/hda_generic.c idx = path->idx[i]; idx 803 sound/pci/hda/hda_generic.c if (n != idx) { idx 812 sound/pci/hda/hda_generic.c check_and_activate_amp(codec, nid, HDA_INPUT, n, idx, enable); idx 891 sound/pci/hda/hda_generic.c path->idx[i]); idx 1329 sound/pci/hda/hda_generic.c static hda_nid_t get_primary_out(struct hda_codec *codec, int idx) idx 1334 sound/pci/hda/hda_generic.c if (cfg->line_outs > idx) idx 1335 sound/pci/hda/hda_generic.c return spec->private_dac_nids[idx]; idx 1336 sound/pci/hda/hda_generic.c idx -= cfg->line_outs; idx 1337 sound/pci/hda/hda_generic.c if (spec->multi_ios > idx) idx 1338 sound/pci/hda/hda_generic.c return spec->multi_io[idx].dac; idx 1667 sound/pci/hda/hda_generic.c int i, idx; idx 1670 sound/pci/hda/hda_generic.c idx = spec->out_paths[0]; idx 1672 sound/pci/hda/hda_generic.c idx = spec->hp_paths[0]; idx 1673 sound/pci/hda/hda_generic.c path = snd_hda_get_path_from_idx(codec, idx); idx 1683 sound/pci/hda/hda_generic.c if (spec->out_paths[i] == idx) idx 1882 sound/pci/hda/hda_generic.c const char *pfx, int idx) idx 1886 sound/pci/hda/hda_generic.c path = snd_hda_get_path_from_idx(codec, idx); idx 2161 sound/pci/hda/hda_generic.c int err, idx = 0; idx 2171 sound/pci/hda/hda_generic.c idx = i; idx 2173 sound/pci/hda/hda_generic.c err = create_extra_out(codec, paths[i], name, idx); idx 2334 sound/pci/hda/hda_generic.c get_multiio_path(struct hda_codec *codec, int idx) idx 2338 sound/pci/hda/hda_generic.c spec->out_paths[spec->autocfg.line_outs + idx]); idx 2351 sound/pci/hda/hda_generic.c static int set_multi_io(struct hda_codec *codec, int idx, bool output) idx 2354 sound/pci/hda/hda_generic.c hda_nid_t nid = spec->multi_io[idx].pin; idx 2357 sound/pci/hda/hda_generic.c path = get_multiio_path(codec, idx); idx 2371 sound/pci/hda/hda_generic.c set_pin_target(codec, nid, spec->multi_io[idx].ctl_in, true); idx 2684 sound/pci/hda/hda_generic.c static bool find_kctl_name(struct hda_codec *codec, const char *name, int idx) idx 2691 sound/pci/hda/hda_generic.c if (!strcmp(kctl->name, name) && kctl->index == idx) idx 2701 sound/pci/hda/hda_generic.c int idx = 0; idx 2703 sound/pci/hda/hda_generic.c snd_hda_get_pin_label(codec, pin, &spec->autocfg, name, name_len, &idx); idx 2706 sound/pci/hda/hda_generic.c for (; find_kctl_name(codec, name, idx); idx++) idx 2785 sound/pci/hda/hda_generic.c static int cvt_from_vref_idx(unsigned int vref_caps, unsigned int idx) idx 2790 sound/pci/hda/hda_generic.c if (i == idx) idx 2819 sound/pci/hda/hda_generic.c unsigned int idx; idx 2821 sound/pci/hda/hda_generic.c idx = snd_hda_codec_get_pin_target(codec, nid) & AC_PINCTL_VREFEN; idx 2822 sound/pci/hda/hda_generic.c ucontrol->value.enumerated.item[0] = cvt_from_vref_idx(vref_caps, idx); idx 2832 sound/pci/hda/hda_generic.c unsigned int val, idx; idx 2835 sound/pci/hda/hda_generic.c idx = cvt_from_vref_idx(vref_caps, val & AC_PINCTL_VREFEN); idx 2836 sound/pci/hda/hda_generic.c if (idx == ucontrol->value.enumerated.item[0]) idx 2899 sound/pci/hda/hda_generic.c int idx; idx 2906 sound/pci/hda/hda_generic.c idx = uinfo->value.enumerated.item; idx 2907 sound/pci/hda/hda_generic.c if (idx < out_jacks) { idx 2909 sound/pci/hda/hda_generic.c text = out_jack_texts[idx]; idx 2913 sound/pci/hda/hda_generic.c idx -= out_jacks; idx 2916 sound/pci/hda/hda_generic.c text = vref_texts[get_vref_idx(vref_caps, idx)]; idx 2930 sound/pci/hda/hda_generic.c int idx = 0; idx 2934 sound/pci/hda/hda_generic.c idx = 1; idx 2936 sound/pci/hda/hda_generic.c idx = out_jacks; idx 2940 sound/pci/hda/hda_generic.c idx += cvt_from_vref_idx(vref_caps, val); idx 2943 sound/pci/hda/hda_generic.c return idx; idx 2963 sound/pci/hda/hda_generic.c unsigned int val, oldval, idx; idx 2966 sound/pci/hda/hda_generic.c idx = ucontrol->value.enumerated.item[0]; idx 2967 sound/pci/hda/hda_generic.c if (oldval == idx) idx 2970 sound/pci/hda/hda_generic.c if (idx < out_jacks) { idx 2972 sound/pci/hda/hda_generic.c val = idx ? PIN_HP : PIN_OUT; idx 2976 sound/pci/hda/hda_generic.c idx -= out_jacks; idx 2981 sound/pci/hda/hda_generic.c val |= get_vref_idx(vref_caps, idx) | PIN_IN; idx 3017 sound/pci/hda/hda_generic.c static int add_loopback_list(struct hda_gen_spec *spec, hda_nid_t mix, int idx) idx 3026 sound/pci/hda/hda_generic.c list->idx = idx; idx 3038 sound/pci/hda/hda_generic.c int idx, num_conns; idx 3042 sound/pci/hda/hda_generic.c idx = snd_hda_get_conn_index(codec, mix_nid, pin, true); idx 3043 sound/pci/hda/hda_generic.c if (idx < 0) idx 3048 sound/pci/hda/hda_generic.c *mix_val = HDA_COMPOSE_AMP_VAL(mix_nid, 3, idx, HDA_INPUT); idx 3050 sound/pci/hda/hda_generic.c *mute_val = HDA_COMPOSE_AMP_VAL(mix_nid, 3, idx, HDA_INPUT); idx 3056 sound/pci/hda/hda_generic.c if (num_conns < idx) idx 3058 sound/pci/hda/hda_generic.c nid = list[idx]; idx 3077 sound/pci/hda/hda_generic.c int err, idx; idx 3088 sound/pci/hda/hda_generic.c idx = path->idx[path->depth - 1]; idx 3105 sound/pci/hda/hda_generic.c err = add_loopback_list(spec, mix_nid, idx); idx 3282 sound/pci/hda/hda_generic.c int j, idx; idx 3288 sound/pci/hda/hda_generic.c idx = 0; idx 3292 sound/pci/hda/hda_generic.c idx = spec->input_label_idxs[j] + 1; idx 3298 sound/pci/hda/hda_generic.c spec->input_label_idxs[i] = idx; idx 3396 sound/pci/hda/hda_generic.c unsigned int idx); idx 3532 sound/pci/hda/hda_generic.c int idx = path->idx[i]; idx 3534 sound/pci/hda/hda_generic.c idx = 0; idx 3536 sound/pci/hda/hda_generic.c HDA_COMPOSE_AMP_VAL(nid, 3, idx, HDA_INPUT); idx 3544 sound/pci/hda/hda_generic.c int idx = path->idx[i]; idx 3546 sound/pci/hda/hda_generic.c idx = 0; idx 3548 sound/pci/hda/hda_generic.c HDA_COMPOSE_AMP_VAL(nid, 3, idx, HDA_INPUT); idx 3594 sound/pci/hda/hda_generic.c int idx, bool is_switch, unsigned int ctl, idx 3613 sound/pci/hda/hda_generic.c knew = add_control(spec, type, tmpname, idx, idx 3629 sound/pci/hda/hda_generic.c knew = add_control(spec, type, tmpname, idx, idx 3639 sound/pci/hda/hda_generic.c static int create_single_cap_vol_ctl(struct hda_codec *codec, int idx, idx 3644 sound/pci/hda/hda_generic.c err = add_single_cap_ctl(codec, NULL, idx, false, vol_ctl, inv_dmic); idx 3647 sound/pci/hda/hda_generic.c err = add_single_cap_ctl(codec, NULL, idx, true, sw_ctl, inv_dmic); idx 3654 sound/pci/hda/hda_generic.c static int create_bind_cap_vol_ctl(struct hda_codec *codec, int idx, idx 3664 sound/pci/hda/hda_generic.c knew->index = idx; idx 3672 sound/pci/hda/hda_generic.c knew->index = idx; idx 3680 sound/pci/hda/hda_generic.c static unsigned int get_first_cap_ctl(struct hda_codec *codec, int idx, int type) idx 3686 sound/pci/hda/hda_generic.c path = get_input_path(codec, 0, idx); idx 3692 sound/pci/hda/hda_generic.c for (i = 0; i < idx - 1; i++) { idx 3709 sound/pci/hda/hda_generic.c int idx; idx 3711 sound/pci/hda/hda_generic.c idx = imux->items[i].index; idx 3712 sound/pci/hda/hda_generic.c if (idx >= spec->autocfg.num_inputs) idx 3718 sound/pci/hda/hda_generic.c spec->input_labels[idx], idx 3719 sound/pci/hda/hda_generic.c spec->input_label_idxs[idx], idx 3804 sound/pci/hda/hda_generic.c int dir, int idx) idx 3809 sound/pci/hda/hda_generic.c is_ctl_associated(codec, nid, dir, idx, NID_PATH_VOL_CTL) || idx 3810 sound/pci/hda/hda_generic.c is_ctl_associated(codec, nid, dir, idx, NID_PATH_BOOST_CTL)) idx 3836 sound/pci/hda/hda_generic.c path->idx[depth])) { idx 3837 sound/pci/hda/hda_generic.c val = HDA_COMPOSE_AMP_VAL(nid, 3, path->idx[depth], idx 3859 sound/pci/hda/hda_generic.c int idx; idx 3862 sound/pci/hda/hda_generic.c idx = imux->items[i].index; idx 3863 sound/pci/hda/hda_generic.c if (idx >= imux->num_items) idx 3867 sound/pci/hda/hda_generic.c if (cfg->inputs[idx].type > AUTO_PIN_LINE_IN) idx 3880 sound/pci/hda/hda_generic.c "%s Boost Volume", spec->input_labels[idx]); idx 3882 sound/pci/hda/hda_generic.c spec->input_label_idxs[idx], val)) idx 4124 sound/pci/hda/hda_generic.c unsigned int idx) idx 4134 sound/pci/hda/hda_generic.c if (idx >= imux->num_items) idx 4135 sound/pci/hda/hda_generic.c idx = imux->num_items - 1; idx 4136 sound/pci/hda/hda_generic.c if (spec->cur_mux[adc_idx] == idx) idx 4145 sound/pci/hda/hda_generic.c spec->cur_mux[adc_idx] = idx; idx 4151 sound/pci/hda/hda_generic.c dyn_adc_pcm_resetup(codec, idx); idx 4153 sound/pci/hda/hda_generic.c path = get_input_path(codec, adc_idx, idx); idx 4618 sound/pci/hda/hda_generic.c mux_select(codec, 0, spec->am_entry[i].idx); idx 4622 sound/pci/hda/hda_generic.c mux_select(codec, 0, spec->am_entry[0].idx); idx 4840 sound/pci/hda/hda_generic.c spec->am_entry[i].idx = idx 4843 sound/pci/hda/hda_generic.c if (spec->am_entry[i].idx < 0) idx 4925 sound/pci/hda/hda_generic.c spec->cur_mux[0] = spec->am_entry[0].idx; idx 42 sound/pci/hda/hda_generic.h unsigned char idx[MAX_NID_PATH_DEPTH]; idx 57 sound/pci/hda/hda_generic.h int idx; /* imux index, -1 = invalid */ idx 321 sound/pci/hda/hda_generic.h struct nid_path *snd_hda_get_path_from_idx(struct hda_codec *codec, int idx); idx 26 sound/pci/hda/hda_local.h #define HDA_COMPOSE_AMP_VAL_OFS(nid,chs,idx,dir,ofs) \ idx 27 sound/pci/hda/hda_local.h ((nid) | ((chs)<<16) | ((dir)<<18) | ((idx)<<19) | ((ofs)<<23)) idx 29 sound/pci/hda/hda_local.h #define HDA_COMPOSE_AMP_VAL(nid,chs,idx,dir) \ idx 30 sound/pci/hda/hda_local.h HDA_COMPOSE_AMP_VAL_OFS(nid, chs, idx, dir, 0) idx 117 sound/pci/hda/hda_local.h #define snd_hda_codec_amp_read(codec, nid, ch, dir, idx) \ idx 118 sound/pci/hda/hda_local.h snd_hdac_regmap_get_amp(&(codec)->core, nid, ch, dir, idx) idx 120 sound/pci/hda/hda_local.h int ch, int dir, int idx, int mask, int val); idx 122 sound/pci/hda/hda_local.h int dir, int idx, int mask, int val); idx 124 sound/pci/hda/hda_local.h int direction, int idx, int mask, int val); idx 126 sound/pci/hda/hda_local.h int dir, int idx, int mask, int val); idx 384 sound/pci/hda/hda_local.h #define AMP_IN_MUTE(idx) (0x7080 | ((idx)<<8)) idx 385 sound/pci/hda/hda_local.h #define AMP_IN_UNMUTE(idx) (0x7000 | ((idx)<<8)) idx 596 sound/pci/hda/hda_local.h unsigned char idx; idx 42 sound/pci/hda/patch_analog.c #define set_beep_amp(spec, nid, idx, dir) \ idx 43 sound/pci/hda/patch_analog.c ((spec)->beep_amp = HDA_COMPOSE_AMP_VAL(nid, 1, idx, dir)) /* mono */ idx 45 sound/pci/hda/patch_analog.c #define set_beep_amp(spec, nid, idx, dir) /* NOP */ idx 819 sound/pci/hda/patch_analog.c .idx = { 0, 0, 0 }, idx 825 sound/pci/hda/patch_analog.c .idx = { 0, 0, 1, 0 }, idx 831 sound/pci/hda/patch_analog.c .idx = { 0, 1, 1, 0 }, idx 837 sound/pci/hda/patch_analog.c .idx = { 0, 2, 1, 0 }, idx 3835 sound/pci/hda/patch_ca0132.c unsigned int *lookup, int idx) idx 3846 sound/pci/hda/patch_ca0132.c &(lookup[idx]), sizeof(unsigned int)); idx 3859 sound/pci/hda/patch_ca0132.c int idx = nid - TUNING_CTL_START_NID; idx 3861 sound/pci/hda/patch_ca0132.c *valp = spec->cur_ctl_vals[idx]; idx 3885 sound/pci/hda/patch_ca0132.c int idx; idx 3887 sound/pci/hda/patch_ca0132.c idx = nid - TUNING_CTL_START_NID; idx 3889 sound/pci/hda/patch_ca0132.c if (spec->cur_ctl_vals[idx] == *valp) idx 3892 sound/pci/hda/patch_ca0132.c spec->cur_ctl_vals[idx] = *valp; idx 3894 sound/pci/hda/patch_ca0132.c idx = *valp - 20; idx 3895 sound/pci/hda/patch_ca0132.c tuning_ctl_set(codec, nid, voice_focus_vals_lookup, idx); idx 3920 sound/pci/hda/patch_ca0132.c int idx; idx 3922 sound/pci/hda/patch_ca0132.c idx = nid - TUNING_CTL_START_NID; idx 3924 sound/pci/hda/patch_ca0132.c if (spec->cur_ctl_vals[idx] == *valp) idx 3927 sound/pci/hda/patch_ca0132.c spec->cur_ctl_vals[idx] = *valp; idx 3929 sound/pci/hda/patch_ca0132.c idx = *valp; idx 3930 sound/pci/hda/patch_ca0132.c tuning_ctl_set(codec, nid, mic_svm_vals_lookup, idx); idx 3955 sound/pci/hda/patch_ca0132.c int idx; idx 3957 sound/pci/hda/patch_ca0132.c idx = nid - TUNING_CTL_START_NID; idx 3959 sound/pci/hda/patch_ca0132.c if (spec->cur_ctl_vals[idx] == *valp) idx 3962 sound/pci/hda/patch_ca0132.c spec->cur_ctl_vals[idx] = *valp; idx 3964 sound/pci/hda/patch_ca0132.c idx = *valp; idx 3965 sound/pci/hda/patch_ca0132.c tuning_ctl_set(codec, nid, equalizer_vals_lookup, idx); idx 4856 sound/pci/hda/patch_ca0132.c int idx = nid - EFFECT_START_NID; idx 4858 sound/pci/hda/patch_ca0132.c if ((idx < 0) || (idx >= num_fx)) idx 4925 sound/pci/hda/patch_ca0132.c err = dspio_set_uint_param(codec, ca0132_effects[idx].mid, idx 4926 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].reqs[0], on); idx 5132 sound/pci/hda/patch_ca0132.c const unsigned int *lookup, int idx) idx 5153 sound/pci/hda/patch_ca0132.c &(lookup[idx - 1]), sizeof(unsigned int)); idx 5162 sound/pci/hda/patch_ca0132.c &(lookup[idx]), sizeof(unsigned int)); idx 5188 sound/pci/hda/patch_ca0132.c int idx = nid - OUT_EFFECT_START_NID; idx 5190 sound/pci/hda/patch_ca0132.c *valp = spec->fx_ctl_val[idx]; idx 5231 sound/pci/hda/patch_ca0132.c int idx; idx 5239 sound/pci/hda/patch_ca0132.c idx = *valp; idx 5240 sound/pci/hda/patch_ca0132.c ca0132_alt_slider_ctl_set(codec, nid, float_xbass_xover_lookup, idx); idx 5252 sound/pci/hda/patch_ca0132.c int idx; idx 5254 sound/pci/hda/patch_ca0132.c idx = nid - EFFECT_START_NID; idx 5256 sound/pci/hda/patch_ca0132.c if (spec->fx_ctl_val[idx] == *valp) idx 5259 sound/pci/hda/patch_ca0132.c spec->fx_ctl_val[idx] = *valp; idx 5261 sound/pci/hda/patch_ca0132.c idx = *valp; idx 5262 sound/pci/hda/patch_ca0132.c ca0132_alt_slider_ctl_set(codec, nid, float_zero_to_one_lookup, idx); idx 5567 sound/pci/hda/patch_ca0132.c unsigned int idx = SMART_VOLUME - EFFECT_START_NID; idx 5593 sound/pci/hda/patch_ca0132.c dspio_set_uint_param(codec, ca0132_effects[idx].mid, idx 5594 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].reqs[2], tmp); idx 7234 sound/pci/hda/patch_ca0132.c int idx, i; idx 7241 sound/pci/hda/patch_ca0132.c for (idx = 0; idx < num_fx; idx++) { idx 7242 sound/pci/hda/patch_ca0132.c for (i = 0; i <= ca0132_effects[idx].params; i++) { idx 7243 sound/pci/hda/patch_ca0132.c dspio_set_uint_param(codec, ca0132_effects[idx].mid, idx 7244 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].reqs[i], idx 7245 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].def_vals[i]); idx 7279 sound/pci/hda/patch_ca0132.c int idx, i; idx 7304 sound/pci/hda/patch_ca0132.c for (idx = 0; idx < num_fx; idx++) { idx 7305 sound/pci/hda/patch_ca0132.c for (i = 0; i <= ca0132_effects[idx].params; i++) { idx 7307 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].mid, idx 7308 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].reqs[i], idx 7309 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].def_vals[i]); idx 7323 sound/pci/hda/patch_ca0132.c int idx, i; idx 7360 sound/pci/hda/patch_ca0132.c for (idx = 0; idx < num_fx; idx++) { idx 7361 sound/pci/hda/patch_ca0132.c for (i = 0; i <= ca0132_effects[idx].params; i++) { idx 7363 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].mid, idx 7364 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].reqs[i], idx 7365 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].def_vals[i]); idx 7380 sound/pci/hda/patch_ca0132.c int idx, i; idx 7427 sound/pci/hda/patch_ca0132.c for (idx = 0; idx < num_fx; idx++) { idx 7428 sound/pci/hda/patch_ca0132.c for (i = 0; i <= ca0132_effects[idx].params; i++) { idx 7430 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].mid, idx 7431 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].reqs[i], idx 7432 sound/pci/hda/patch_ca0132.c ca0132_effects[idx].def_vals[i]); idx 129 sound/pci/hda/patch_cirrus.c static inline int cs_vendor_coef_get(struct hda_codec *codec, unsigned int idx) idx 133 sound/pci/hda/patch_cirrus.c AC_VERB_SET_COEF_INDEX, idx); idx 138 sound/pci/hda/patch_cirrus.c static inline void cs_vendor_coef_set(struct hda_codec *codec, unsigned int idx, idx 143 sound/pci/hda/patch_cirrus.c AC_VERB_SET_COEF_INDEX, idx); idx 365 sound/pci/hda/patch_cirrus.c int idx = spec->gen.dyn_adc_idx[i]; idx 366 sound/pci/hda/patch_cirrus.c if (done & (1 << idx)) idx 369 sound/pci/hda/patch_cirrus.c spec->gen.adc_nids[idx]); idx 370 sound/pci/hda/patch_cirrus.c done |= 1 << idx; idx 57 sound/pci/hda/patch_conexant.c int idx, int dir) idx 60 sound/pci/hda/patch_conexant.c unsigned int beep_amp = HDA_COMPOSE_AMP_VAL(nid, 1, idx, dir); idx 484 sound/pci/hda/patch_conexant.c unsigned int idx; idx 486 sound/pci/hda/patch_conexant.c idx = ucontrol->value.enumerated.item[0]; idx 487 sound/pci/hda/patch_conexant.c if (idx >= imux->num_items) idx 488 sound/pci/hda/patch_conexant.c idx = imux->num_items - 1; idx 489 sound/pci/hda/patch_conexant.c if (spec->dc_input_bias == idx) idx 492 sound/pci/hda/patch_conexant.c spec->dc_input_bias = idx; idx 234 sound/pci/hda/patch_hdmi.c #define get_pin(spec, idx) \ idx 235 sound/pci/hda/patch_hdmi.c ((struct hdmi_spec_per_pin *)snd_array_elem(&spec->pins, idx)) idx 236 sound/pci/hda/patch_hdmi.c #define get_cvt(spec, idx) \ idx 237 sound/pci/hda/patch_hdmi.c ((struct hdmi_spec_per_cvt *)snd_array_elem(&spec->cvts, idx)) idx 239 sound/pci/hda/patch_hdmi.c #define get_hdmi_pcm(spec, idx) (&(spec)->pcm_rec[idx]) idx 241 sound/pci/hda/patch_hdmi.c #define get_pcm_rec(spec, idx) (get_hdmi_pcm(spec, idx)->pcm) idx 1346 sound/pci/hda/patch_hdmi.c int idx; idx 1351 sound/pci/hda/patch_hdmi.c idx = hdmi_find_pcm_slot(spec, per_pin); idx 1352 sound/pci/hda/patch_hdmi.c if (idx == -EBUSY) idx 1354 sound/pci/hda/patch_hdmi.c per_pin->pcm_idx = idx; idx 1355 sound/pci/hda/patch_hdmi.c per_pin->pcm = get_hdmi_pcm(spec, idx); idx 1356 sound/pci/hda/patch_hdmi.c set_bit(idx, &spec->pcm_bitmap); idx 1362 sound/pci/hda/patch_hdmi.c int idx; idx 1367 sound/pci/hda/patch_hdmi.c idx = per_pin->pcm_idx; idx 1370 sound/pci/hda/patch_hdmi.c if (idx >= 0 && idx < spec->pcm_used) idx 1371 sound/pci/hda/patch_hdmi.c clear_bit(idx, &spec->pcm_bitmap); idx 2091 sound/pci/hda/patch_hdmi.c int idx; idx 2099 sound/pci/hda/patch_hdmi.c for (idx = 0; idx < spec->num_nids + spec->dev_num - 1; idx++) { idx 2103 sound/pci/hda/patch_hdmi.c info = snd_hda_codec_pcm_new(codec, "HDMI %d", idx); idx 2107 sound/pci/hda/patch_hdmi.c spec->pcm_rec[idx].pcm = info; idx 182 sound/pci/hda/patch_realtek.c unsigned char idx; idx 188 sound/pci/hda/patch_realtek.c { .nid = (_nid), .idx = (_idx), .mask = (_mask), .val = (_val) } idx 198 sound/pci/hda/patch_realtek.c alc_write_coefex_idx(codec, fw->nid, fw->idx, fw->val); idx 200 sound/pci/hda/patch_realtek.c alc_update_coefex_idx(codec, fw->nid, fw->idx, idx 1055 sound/pci/hda/patch_realtek.c int idx, int dir) idx 1058 sound/pci/hda/patch_realtek.c unsigned int beep_amp = HDA_COMPOSE_AMP_VAL(nid, 3, idx, dir); idx 1097 sound/pci/hda/patch_realtek.c #define set_beep_amp(spec, nid, idx, dir) 0 idx 256 sound/pci/hda/patch_sigmatel.c int i, idx = 0; idx 263 sound/pci/hda/patch_sigmatel.c idx = i; idx 273 sound/pci/hda/patch_sigmatel.c spec->active_adcs |= (1 << idx); idx 278 sound/pci/hda/patch_sigmatel.c spec->active_adcs &= ~(1 << idx); idx 431 sound/pci/hda/patch_sigmatel.c unsigned int idx, val; idx 433 sound/pci/hda/patch_sigmatel.c for (idx = 0; idx < spec->num_pwrs; idx++) { idx 434 sound/pci/hda/patch_sigmatel.c if (spec->pwr_nids[idx] == nid) idx 437 sound/pci/hda/patch_sigmatel.c if (idx >= spec->num_pwrs) idx 440 sound/pci/hda/patch_sigmatel.c idx = 1 << idx; idx 444 sound/pci/hda/patch_sigmatel.c val &= ~idx; idx 446 sound/pci/hda/patch_sigmatel.c val |= idx; idx 569 sound/pci/hda/patch_sigmatel.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 573 sound/pci/hda/patch_sigmatel.c (spec->aloopback_mask << idx)); idx 582 sound/pci/hda/patch_sigmatel.c unsigned int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 586 sound/pci/hda/patch_sigmatel.c idx_val = spec->aloopback_mask << idx; idx 259 sound/pci/hda/patch_via.c int idx, int dir) idx 262 sound/pci/hda/patch_via.c unsigned int beep_amp = HDA_COMPOSE_AMP_VAL(nid, 1, idx, dir); idx 303 sound/pci/hda/patch_via.c p->idx); idx 45 sound/pci/ice1712/ak4xxx.c int idx; idx 75 sound/pci/ice1712/ak4xxx.c for (idx = 15; idx >= 0; idx--) { idx 81 sound/pci/ice1712/ak4xxx.c if (addrdata & (1 << idx)) idx 810 sound/pci/ice1712/aureon.c int i, idx, ofs, voices; idx 823 sound/pci/ice1712/aureon.c idx = WM_DAC_ATTEN + ofs + i; idx 824 sound/pci/ice1712/aureon.c wm_set_vol(ice, idx, spec->vol[ofs + i], idx 1031 sound/pci/ice1712/aureon.c int i, idx; idx 1036 sound/pci/ice1712/aureon.c idx = WM_ADC_GAIN + i; idx 1037 sound/pci/ice1712/aureon.c vol = wm_get(ice, idx) & 0x1f; idx 1047 sound/pci/ice1712/aureon.c int i, idx; idx 1053 sound/pci/ice1712/aureon.c idx = WM_ADC_GAIN + i; idx 1055 sound/pci/ice1712/aureon.c ovol = wm_get(ice, idx); idx 1057 sound/pci/ice1712/aureon.c wm_put(ice, idx, nvol | (ovol & ~0x1f)); idx 35 sound/pci/ice1712/delta.c int idx; idx 37 sound/pci/ice1712/delta.c for (idx = 7; idx >= 0; idx--) { idx 39 sound/pci/ice1712/delta.c if (data & (1 << idx)) idx 53 sound/pci/ice1712/delta.c int idx; idx 55 sound/pci/ice1712/delta.c for (idx = 7; idx >= 0; idx--) { idx 60 sound/pci/ice1712/delta.c data |= 1 << idx; idx 175 sound/pci/ice1712/delta.c int idx; idx 181 sound/pci/ice1712/delta.c for (idx = 7; idx >= 0; idx--) { idx 183 sound/pci/ice1712/delta.c if (bits & (1 << idx)) idx 949 sound/pci/ice1712/ews.c unsigned int idx; idx 976 sound/pci/ice1712/ews.c for (idx = 0; idx < ARRAY_SIZE(snd_ice1712_ewx2496_controls); idx++) { idx 977 sound/pci/ice1712/ews.c err = snd_ctl_add(ice->card, snd_ctl_new1(&snd_ice1712_ewx2496_controls[idx], ice)); idx 994 sound/pci/ice1712/ews.c for (idx = 0; idx < ARRAY_SIZE(snd_ice1712_ews88d_controls); idx++) { idx 995 sound/pci/ice1712/ews.c err = snd_ctl_add(ice->card, snd_ctl_new1(&snd_ice1712_ews88d_controls[idx], ice)); idx 1001 sound/pci/ice1712/ews.c for (idx = 0; idx < ARRAY_SIZE(snd_ice1712_6fire_controls); idx++) { idx 1002 sound/pci/ice1712/ews.c err = snd_ctl_add(ice->card, snd_ctl_new1(&snd_ice1712_6fire_controls[idx], ice)); idx 452 sound/pci/ice1712/ice1712.c u32 idx; idx 457 sound/pci/ice1712/ice1712.c for (idx = 0; idx < 6; idx++) { idx 458 sound/pci/ice1712/ice1712.c if ((pbkstatus & (3 << (idx * 2))) == 0) idx 460 sound/pci/ice1712/ice1712.c substream = ice->playback_con_substream_ds[idx]; idx 463 sound/pci/ice1712/ice1712.c outw(3 << (idx * 2), ICEDS(ice, INTSTAT)); idx 1434 sound/pci/ice1712/ice1712.c unsigned int idx; idx 1438 sound/pci/ice1712/ice1712.c for (idx = 0; idx < ARRAY_SIZE(snd_ice1712_multi_playback_ctrls); idx++) { idx 1439 sound/pci/ice1712/ice1712.c err = snd_ctl_add(card, snd_ctl_new1(&snd_ice1712_multi_playback_ctrls[idx], ice)); idx 1469 sound/pci/ice1712/ice1712.c for (idx = 0; idx < 10; idx++) { idx 1470 sound/pci/ice1712/ice1712.c ice->pro_volumes[idx] = 0x80008000; /* mute */ idx 1471 sound/pci/ice1712/ice1712.c snd_ice1712_update_volume(ice, idx); idx 1473 sound/pci/ice1712/ice1712.c for (idx = 10; idx < 10 + ice->num_total_adcs; idx++) { idx 1474 sound/pci/ice1712/ice1712.c ice->pro_volumes[idx] = 0x80008000; /* mute */ idx 1475 sound/pci/ice1712/ice1712.c snd_ice1712_update_volume(ice, idx); idx 1477 sound/pci/ice1712/ice1712.c for (idx = 18; idx < 20; idx++) { idx 1478 sound/pci/ice1712/ice1712.c ice->pro_volumes[idx] = 0x80008000; /* mute */ idx 1479 sound/pci/ice1712/ice1712.c snd_ice1712_update_volume(ice, idx); idx 1545 sound/pci/ice1712/ice1712.c static inline unsigned int eeprom_double(struct snd_ice1712 *ice, int idx) idx 1547 sound/pci/ice1712/ice1712.c return (unsigned int)ice->eeprom.data[idx] | ((unsigned int)ice->eeprom.data[idx + 1] << 8); idx 1554 sound/pci/ice1712/ice1712.c unsigned int idx; idx 1573 sound/pci/ice1712/ice1712.c for (idx = 0; idx < 4; idx++) idx 1574 sound/pci/ice1712/ice1712.c snd_iprintf(buffer, " DAC ID #%i : 0x%x\n", idx, ice->eeprom.data[ICE_EEP1_DAC_ID + idx]); idx 1575 sound/pci/ice1712/ice1712.c for (idx = 0; idx < 4; idx++) idx 1576 sound/pci/ice1712/ice1712.c snd_iprintf(buffer, " ADC ID #%i : 0x%x\n", idx, ice->eeprom.data[ICE_EEP1_ADC_ID + idx]); idx 1577 sound/pci/ice1712/ice1712.c for (idx = 0x1c; idx < ice->eeprom.size; idx++) idx 1578 sound/pci/ice1712/ice1712.c snd_iprintf(buffer, " Extra #%02i : 0x%x\n", idx, ice->eeprom.data[idx]); idx 2020 sound/pci/ice1712/ice1712.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2028 sound/pci/ice1712/ice1712.c val >>= ((idx % 2) * 8) + ((idx / 2) * 2); idx 2030 sound/pci/ice1712/ice1712.c cval >>= ((idx / 2) * 8) + ((idx % 2) * 4); idx 2031 sound/pci/ice1712/ice1712.c if (val == 1 && idx < 2) idx 2047 sound/pci/ice1712/ice1712.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2052 sound/pci/ice1712/ice1712.c nval = idx < 2 ? 1 : 0; /* dig mixer (or pcm) */ idx 2059 sound/pci/ice1712/ice1712.c shift = ((idx % 2) * 8) + ((idx / 2) * 2); idx 2074 sound/pci/ice1712/ice1712.c shift = ((idx / 2) * 8) + ((idx % 2) * 4); idx 2096 sound/pci/ice1712/ice1712.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2099 sound/pci/ice1712/ice1712.c cval = (val >> (idx * 4 + 8)) & 0x0f; idx 2100 sound/pci/ice1712/ice1712.c val = (val >> (idx * 2)) & 0x03; idx 2117 sound/pci/ice1712/ice1712.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2131 sound/pci/ice1712/ice1712.c shift = idx * 2; idx 2134 sound/pci/ice1712/ice1712.c shift = idx * 4 + 8; idx 2223 sound/pci/ice1712/ice1712.c int idx; idx 2226 sound/pci/ice1712/ice1712.c for (idx = 0; idx < 22; idx++) { idx 2227 sound/pci/ice1712/ice1712.c outb(idx, ICEMT(ice, MONITOR_PEAKINDEX)); idx 2228 sound/pci/ice1712/ice1712.c ucontrol->value.integer.value[idx] = inb(ICEMT(ice, MONITOR_PEAKDATA)); idx 1512 sound/pci/ice1712/ice1724.c static inline unsigned int eeprom_triple(struct snd_ice1712 *ice, int idx) idx 1514 sound/pci/ice1712/ice1724.c return (unsigned int)ice->eeprom.data[idx] | \ idx 1515 sound/pci/ice1712/ice1724.c ((unsigned int)ice->eeprom.data[idx + 1] << 8) | \ idx 1516 sound/pci/ice1712/ice1724.c ((unsigned int)ice->eeprom.data[idx + 2] << 16); idx 1523 sound/pci/ice1712/ice1724.c unsigned int idx; idx 1545 sound/pci/ice1712/ice1724.c for (idx = 0x12; idx < ice->eeprom.size; idx++) idx 1547 sound/pci/ice1712/ice1724.c idx, ice->eeprom.data[idx]); idx 1553 sound/pci/ice1712/ice1724.c for (idx = 0x0; idx < 0x20 ; idx++) idx 1555 sound/pci/ice1712/ice1724.c idx, inb(ice->port+idx)); idx 1556 sound/pci/ice1712/ice1724.c for (idx = 0x0; idx < 0x30 ; idx++) idx 1558 sound/pci/ice1712/ice1724.c idx, inb(ice->profi_port+idx)); idx 2039 sound/pci/ice1712/ice1724.c static inline int analog_route_shift(int idx) idx 2041 sound/pci/ice1712/ice1724.c return (idx % 2) * 12 + ((idx / 2) * 3) + 8; idx 2044 sound/pci/ice1712/ice1724.c static inline int digital_route_shift(int idx) idx 2046 sound/pci/ice1712/ice1724.c return idx * 3; idx 2095 sound/pci/ice1712/ice1724.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2097 sound/pci/ice1712/ice1724.c snd_ice1724_get_route_val(ice, analog_route_shift(idx)); idx 2105 sound/pci/ice1712/ice1724.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2108 sound/pci/ice1712/ice1724.c analog_route_shift(idx)); idx 2115 sound/pci/ice1712/ice1724.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2117 sound/pci/ice1712/ice1724.c snd_ice1724_get_route_val(ice, digital_route_shift(idx)); idx 2125 sound/pci/ice1712/ice1724.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 2128 sound/pci/ice1712/ice1724.c digital_route_shift(idx)); idx 2164 sound/pci/ice1712/ice1724.c int idx; idx 2167 sound/pci/ice1712/ice1724.c for (idx = 0; idx < 22; idx++) { idx 2168 sound/pci/ice1712/ice1724.c outb(idx, ICEMT1724(ice, MONITOR_PEAKINDEX)); idx 2169 sound/pci/ice1712/ice1724.c ucontrol->value.integer.value[idx] = idx 160 sound/pci/ice1712/maya44.c unsigned int idx = kcontrol->private_value; idx 161 sound/pci/ice1712/maya44.c struct maya_vol_info *vol = &vol_info[idx]; idx 176 sound/pci/ice1712/maya44.c unsigned int idx = kcontrol->private_value; idx 179 sound/pci/ice1712/maya44.c ucontrol->value.integer.value[0] = wm->volumes[idx][0]; idx 180 sound/pci/ice1712/maya44.c ucontrol->value.integer.value[1] = wm->volumes[idx][1]; idx 191 sound/pci/ice1712/maya44.c unsigned int idx = kcontrol->private_value; idx 192 sound/pci/ice1712/maya44.c struct maya_vol_info *vol = &vol_info[idx]; idx 201 sound/pci/ice1712/maya44.c if (val == wm->volumes[idx][ch]) idx 214 sound/pci/ice1712/maya44.c wm->volumes[idx][ch] = val; idx 224 sound/pci/ice1712/maya44.c #define COMPOSE_SW_VAL(idx, reg, mask) ((idx) | ((reg) << 8) | ((mask) << 16)) idx 237 sound/pci/ice1712/maya44.c unsigned int idx = GET_SW_VAL_IDX(kcontrol->private_value); idx 239 sound/pci/ice1712/maya44.c ucontrol->value.integer.value[0] = (wm->switch_bits >> idx) & 1; idx 249 sound/pci/ice1712/maya44.c unsigned int idx = GET_SW_VAL_IDX(kcontrol->private_value); idx 254 sound/pci/ice1712/maya44.c mask = 1 << idx; idx 337 sound/pci/ice1712/maya44.c static void wm8776_select_input(struct snd_maya44 *chip, int idx, int line) idx 339 sound/pci/ice1712/maya44.c wm8776_write_bits(chip->ice, &chip->wm[idx], WM8776_REG_ADC_MUX, idx 395 sound/pci/ice1712/maya44.c static int maya_pb_route_shift(int idx) idx 399 sound/pci/ice1712/maya44.c return shift[idx % 10]; idx 406 sound/pci/ice1712/maya44.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 408 sound/pci/ice1712/maya44.c snd_ice1724_get_route_val(chip->ice, maya_pb_route_shift(idx)); idx 416 sound/pci/ice1712/maya44.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); idx 419 sound/pci/ice1712/maya44.c maya_pb_route_shift(idx)); idx 495 sound/pci/ice1712/phase.c int i, idx, ofs, voices; idx 509 sound/pci/ice1712/phase.c idx = WM_DAC_ATTEN + ofs + i; idx 510 sound/pci/ice1712/phase.c wm_set_vol(ice, idx, spec->vol[ofs+i], idx 129 sound/pci/ice1712/pontis.c int i, idx, change = 0; idx 135 sound/pci/ice1712/pontis.c idx = WM_DAC_ATTEN_L + i; idx 136 sound/pci/ice1712/pontis.c oval = wm_get(ice, idx) & 0xff; idx 138 sound/pci/ice1712/pontis.c wm_put(ice, idx, nval); idx 139 sound/pci/ice1712/pontis.c wm_put_nocache(ice, idx, nval | 0x100); idx 184 sound/pci/ice1712/pontis.c int i, idx, change = 0; idx 190 sound/pci/ice1712/pontis.c idx = WM_ADC_ATTEN_L + i; idx 191 sound/pci/ice1712/pontis.c ovol = wm_get(ice, idx) & 0xff; idx 193 sound/pci/ice1712/pontis.c wm_put(ice, idx, nvol); idx 78 sound/pci/ice1712/prodigy192.c static int stac9460_dac_mute(struct snd_ice1712 *ice, int idx, idx 83 sound/pci/ice1712/prodigy192.c old = stac9460_get(ice, idx); idx 88 sound/pci/ice1712/prodigy192.c stac9460_put(ice, idx, new); idx 98 sound/pci/ice1712/prodigy192.c int idx; idx 101 sound/pci/ice1712/prodigy192.c idx = STAC946X_MASTER_VOLUME; idx 103 sound/pci/ice1712/prodigy192.c idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id) + STAC946X_LF_VOLUME; idx 104 sound/pci/ice1712/prodigy192.c val = stac9460_get(ice, idx); idx 113 sound/pci/ice1712/prodigy192.c int idx, change; idx 116 sound/pci/ice1712/prodigy192.c idx = STAC946X_MASTER_VOLUME; idx 118 sound/pci/ice1712/prodigy192.c idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id) + STAC946X_LF_VOLUME; idx 125 sound/pci/ice1712/prodigy192.c change = stac9460_dac_mute(ice, idx, ucontrol->value.integer.value[0]); idx 145 sound/pci/ice1712/prodigy192.c int idx; idx 149 sound/pci/ice1712/prodigy192.c idx = STAC946X_MASTER_VOLUME; idx 151 sound/pci/ice1712/prodigy192.c idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id) + STAC946X_LF_VOLUME; idx 152 sound/pci/ice1712/prodigy192.c vol = stac9460_get(ice, idx) & 0x7f; idx 161 sound/pci/ice1712/prodigy192.c int idx; idx 166 sound/pci/ice1712/prodigy192.c idx = STAC946X_MASTER_VOLUME; idx 168 sound/pci/ice1712/prodigy192.c idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id) + STAC946X_LF_VOLUME; idx 170 sound/pci/ice1712/prodigy192.c tmp = stac9460_get(ice, idx); idx 179 sound/pci/ice1712/prodigy192.c stac9460_put(ice, idx, (0x7f - nvol) | (tmp & 0x80)); idx 307 sound/pci/ice1712/prodigy192.c int idx; idx 326 sound/pci/ice1712/prodigy192.c for (idx = 0; idx < 7 ; ++idx) idx 327 sound/pci/ice1712/prodigy192.c changed[idx] = stac9460_dac_mute(ice, idx 328 sound/pci/ice1712/prodigy192.c STAC946X_MASTER_VOLUME + idx, 0); idx 334 sound/pci/ice1712/prodigy192.c for (idx = 0; idx < 7 ; ++idx) { idx 335 sound/pci/ice1712/prodigy192.c if (changed[idx]) idx 336 sound/pci/ice1712/prodigy192.c stac9460_dac_mute(ice, STAC946X_MASTER_VOLUME + idx, 1); idx 433 sound/pci/ice1712/prodigy192.c unsigned int data, int idx) idx 435 sound/pci/ice1712/prodigy192.c for (; idx >= 0; idx--) { idx 441 sound/pci/ice1712/prodigy192.c if (data & (1 << idx)) idx 458 sound/pci/ice1712/prodigy192.c int idx) idx 462 sound/pci/ice1712/prodigy192.c for (; idx >= 0; idx--) { idx 469 sound/pci/ice1712/prodigy192.c data |= (1 << idx); idx 377 sound/pci/ice1712/prodigy_hifi.c int i, idx, change = 0; idx 382 sound/pci/ice1712/prodigy_hifi.c idx = WM_DAC_ATTEN_L + i; idx 385 sound/pci/ice1712/prodigy_hifi.c wm_set_vol(ice, idx, spec->vol[2 + i], spec->master[i]); idx 426 sound/pci/ice1712/prodigy_hifi.c int i, idx, ofs, voices; idx 434 sound/pci/ice1712/prodigy_hifi.c idx = WM8766_LDA1 + ofs + i; idx 437 sound/pci/ice1712/prodigy_hifi.c wm8766_set_vol(ice, idx, idx 604 sound/pci/ice1712/prodigy_hifi.c int i, idx, change = 0; idx 610 sound/pci/ice1712/prodigy_hifi.c idx = WM_ADC_ATTEN_L + i; idx 611 sound/pci/ice1712/prodigy_hifi.c ovol = wm_get(ice, idx) & 0xff; idx 613 sound/pci/ice1712/prodigy_hifi.c wm_put(ice, idx, nvol); idx 261 sound/pci/ice1712/quartet.c int idx; idx 291 sound/pci/ice1712/quartet.c for (idx = 15; idx >= 0; idx--) { idx 297 sound/pci/ice1712/quartet.c if (addrdata & (1 << idx)) idx 365 sound/pci/ice1712/revo.c unsigned int data, int idx) idx 367 sound/pci/ice1712/revo.c for (; idx >= 0; idx--) { idx 373 sound/pci/ice1712/revo.c if (data & (1 << idx)) idx 387 sound/pci/ice1712/revo.c int idx) idx 391 sound/pci/ice1712/revo.c for (; idx >= 0; idx--) { idx 398 sound/pci/ice1712/revo.c data |= (1 << idx); idx 68 sound/pci/ice1712/wtm.c int id, idx, change; idx 74 sound/pci/ice1712/wtm.c idx = STAC946X_MASTER_VOLUME; idx 76 sound/pci/ice1712/wtm.c idx = STAC946X_LF_VOLUME - 1 + id; idx 77 sound/pci/ice1712/wtm.c old = stac9460_get(ice, idx); idx 81 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, new); idx 93 sound/pci/ice1712/wtm.c idx = STAC946X_MASTER_VOLUME; idx 95 sound/pci/ice1712/wtm.c idx = STAC946X_LF_VOLUME - 1 + id; idx 96 sound/pci/ice1712/wtm.c old = stac9460_2_get(ice, idx); idx 100 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx, new); idx 119 sound/pci/ice1712/wtm.c int idx, id; idx 124 sound/pci/ice1712/wtm.c idx = STAC946X_MASTER_VOLUME; idx 128 sound/pci/ice1712/wtm.c idx = id + STAC946X_LF_VOLUME; idx 131 sound/pci/ice1712/wtm.c val = stac9460_get(ice, idx); idx 133 sound/pci/ice1712/wtm.c val = stac9460_2_get(ice, idx - 6); idx 145 sound/pci/ice1712/wtm.c int id, idx; idx 149 sound/pci/ice1712/wtm.c idx = STAC946X_MASTER_VOLUME; idx 150 sound/pci/ice1712/wtm.c old = stac9460_get(ice, idx); idx 155 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, new); idx 156 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx, new); idx 160 sound/pci/ice1712/wtm.c idx = id + STAC946X_LF_VOLUME; idx 162 sound/pci/ice1712/wtm.c old = stac9460_get(ice, idx); idx 164 sound/pci/ice1712/wtm.c old = stac9460_2_get(ice, idx - 6); idx 170 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, new); idx 172 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx - 6, new); idx 195 sound/pci/ice1712/wtm.c int idx, id; idx 199 sound/pci/ice1712/wtm.c idx = STAC946X_MASTER_VOLUME; idx 203 sound/pci/ice1712/wtm.c idx = id + STAC946X_LF_VOLUME; idx 206 sound/pci/ice1712/wtm.c vol = stac9460_get(ice, idx) & 0x7f; idx 208 sound/pci/ice1712/wtm.c vol = stac9460_2_get(ice, idx - 6) & 0x7f; idx 217 sound/pci/ice1712/wtm.c int idx, id; idx 222 sound/pci/ice1712/wtm.c idx = STAC946X_MASTER_VOLUME; idx 224 sound/pci/ice1712/wtm.c tmp = stac9460_get(ice, idx); idx 228 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, (0x7f - nvol) | (tmp & 0x80)); idx 229 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx, (0x7f - nvol) | (tmp & 0x80)); idx 233 sound/pci/ice1712/wtm.c idx = id + STAC946X_LF_VOLUME; idx 236 sound/pci/ice1712/wtm.c tmp = stac9460_get(ice, idx); idx 238 sound/pci/ice1712/wtm.c tmp = stac9460_2_get(ice, idx - 6); idx 243 sound/pci/ice1712/wtm.c stac9460_put(ice, idx, (0x7f - nvol) | idx 246 sound/pci/ice1712/wtm.c stac9460_2_put(ice, idx-6, (0x7f - nvol) | idx 658 sound/pci/intel8x0.c int idx; idx 666 sound/pci/intel8x0.c for (idx = 0; idx < (ICH_REG_LVI_MASK + 1) * 2; idx += 4) { idx 667 sound/pci/intel8x0.c bdbar[idx + 0] = cpu_to_le32(ichdev->physbuf); idx 668 sound/pci/intel8x0.c bdbar[idx + 1] = cpu_to_le32(0x80000000 | /* interrupt on completion */ idx 670 sound/pci/intel8x0.c bdbar[idx + 2] = cpu_to_le32(ichdev->physbuf + (ichdev->size >> 1)); idx 671 sound/pci/intel8x0.c bdbar[idx + 3] = cpu_to_le32(0x80000000 | /* interrupt on completion */ idx 678 sound/pci/intel8x0.c for (idx = 0; idx < (ICH_REG_LVI_MASK + 1) * 2; idx += 2) { idx 679 sound/pci/intel8x0.c bdbar[idx + 0] = cpu_to_le32(ichdev->physbuf + idx 680 sound/pci/intel8x0.c (((idx >> 1) * ichdev->fragsize) % idx 682 sound/pci/intel8x0.c bdbar[idx + 1] = cpu_to_le32(0x80000000 | /* interrupt on completion */ idx 686 sound/pci/intel8x0.c idx + 0, bdbar[idx + 0], bdbar[idx + 1]); idx 1238 sound/pci/intel8x0.c int idx = chip->device_type == DEVICE_NFORCE ? NVD_SPBAR : ICHD_SPBAR; idx 1240 sound/pci/intel8x0.c return snd_intel8x0_pcm_open(substream, &chip->ichd[idx]); idx 1246 sound/pci/intel8x0.c int idx = chip->device_type == DEVICE_NFORCE ? NVD_SPBAR : ICHD_SPBAR; idx 1248 sound/pci/intel8x0.c chip->ichd[idx].substream = NULL; idx 382 sound/pci/intel8x0m.c int idx; idx 390 sound/pci/intel8x0m.c for (idx = 0; idx < (ICH_REG_LVI_MASK + 1) * 2; idx += 4) { idx 391 sound/pci/intel8x0m.c bdbar[idx + 0] = cpu_to_le32(ichdev->physbuf); idx 392 sound/pci/intel8x0m.c bdbar[idx + 1] = cpu_to_le32(0x80000000 | /* interrupt on completion */ idx 394 sound/pci/intel8x0m.c bdbar[idx + 2] = cpu_to_le32(ichdev->physbuf + (ichdev->size >> 1)); idx 395 sound/pci/intel8x0m.c bdbar[idx + 3] = cpu_to_le32(0x80000000 | /* interrupt on completion */ idx 402 sound/pci/intel8x0m.c for (idx = 0; idx < (ICH_REG_LVI_MASK + 1) * 2; idx += 2) { idx 403 sound/pci/intel8x0m.c bdbar[idx + 0] = cpu_to_le32(ichdev->physbuf + (((idx >> 1) * ichdev->fragsize) % ichdev->size)); idx 404 sound/pci/intel8x0m.c bdbar[idx + 1] = cpu_to_le32(0x80000000 | /* interrupt on completion */ idx 377 sound/pci/lola/lola.h #define lola_readl(chip, idx, name) \ idx 378 sound/pci/lola/lola.h readl((chip)->bar[idx].remap_addr + LOLA_##idx##_##name) idx 379 sound/pci/lola/lola.h #define lola_readw(chip, idx, name) \ idx 380 sound/pci/lola/lola.h readw((chip)->bar[idx].remap_addr + LOLA_##idx##_##name) idx 381 sound/pci/lola/lola.h #define lola_readb(chip, idx, name) \ idx 382 sound/pci/lola/lola.h readb((chip)->bar[idx].remap_addr + LOLA_##idx##_##name) idx 383 sound/pci/lola/lola.h #define lola_writel(chip, idx, name, val) \ idx 384 sound/pci/lola/lola.h writel((val), (chip)->bar[idx].remap_addr + LOLA_##idx##_##name) idx 385 sound/pci/lola/lola.h #define lola_writew(chip, idx, name, val) \ idx 386 sound/pci/lola/lola.h writew((val), (chip)->bar[idx].remap_addr + LOLA_##idx##_##name) idx 387 sound/pci/lola/lola.h #define lola_writeb(chip, idx, name, val) \ idx 388 sound/pci/lola/lola.h writeb((val), (chip)->bar[idx].remap_addr + LOLA_##idx##_##name) idx 491 sound/pci/lola/lola.h int lola_set_clock_index(struct lola *chip, unsigned int idx); idx 492 sound/pci/lola/lola.h int lola_set_clock(struct lola *chip, int idx); idx 113 sound/pci/lola/lola_clock.c int i, j, nitems, nb_verbs, idx, idx_list; idx 139 sound/pci/lola/lola_clock.c idx = 0; idx 146 sound/pci/lola/lola_clock.c idx, 0, &val, &res_ex); idx 188 sound/pci/lola/lola_clock.c chip->clock.idx_lookup[idx_list] = idx; idx 193 sound/pci/lola/lola_clock.c if (++idx >= nitems) idx 220 sound/pci/lola/lola_clock.c int lola_set_clock_index(struct lola *chip, unsigned int idx) idx 227 sound/pci/lola/lola_clock.c chip->clock.idx_lookup[idx], idx 260 sound/pci/lola/lola_clock.c int lola_set_clock(struct lola *chip, int idx) idx 265 sound/pci/lola/lola_clock.c if (idx == chip->clock.cur_index) { idx 269 sound/pci/lola/lola_clock.c } else if (chip->clock.sample_clock[idx].type == idx 272 sound/pci/lola/lola_clock.c freq = chip->clock.sample_clock[idx].freq; idx 282 sound/pci/lola/lola_clock.c if (idx != chip->clock.cur_index) { idx 283 sound/pci/lola/lola_clock.c int err = lola_set_clock_index(chip, idx); idx 287 sound/pci/lola/lola_clock.c chip->clock.cur_index = idx; idx 317 sound/pci/lola/lola_mixer.c unsigned int idx, unsigned int val, idx 323 sound/pci/lola/lola_mixer.c int idx, max_idx; idx 327 sound/pci/lola/lola_mixer.c for (idx = 0; idx < max_idx; idx++) { idx 328 sound/pci/lola/lola_mixer.c if (pin[idx].is_analog) { idx 329 sound/pci/lola/lola_mixer.c unsigned int val = mute ? 0 : pin[idx].cur_gain_step; idx 331 sound/pci/lola/lola_mixer.c set_analog_volume(chip, dir, idx, val, false); idx 384 sound/pci/lola/lola_mixer.c unsigned int idx, unsigned int val, idx 390 sound/pci/lola/lola_mixer.c if (idx >= chip->pin[dir].num_pins) idx 392 sound/pci/lola/lola_mixer.c pin = &chip->pin[dir].pins[idx]; idx 401 sound/pci/lola/lola_mixer.c dir, idx, val); idx 681 sound/pci/lola/lola_mixer.c unsigned int idx = ofs + i; idx 683 sound/pci/lola/lola_mixer.c if (!(chip->mixer.src_mask & (1 << idx))) idx 685 sound/pci/lola/lola_mixer.c if (mask & (1 << idx)) idx 686 sound/pci/lola/lola_mixer.c val = readw(&chip->mixer.array->src_gain[idx]) + 1; idx 703 sound/pci/lola/lola_mixer.c unsigned int idx = ofs + i; idx 707 sound/pci/lola/lola_mixer.c err = lola_mixer_set_src_gain(chip, idx, val, !!val); idx 30 sound/pci/lola/lola_pcm.c unsigned int idx = substream->number; idx 31 sound/pci/lola/lola_pcm.c return &pcm->streams[idx]; idx 630 sound/pci/lola/lola_pcm.c int idx, int nid, int dir) idx 636 sound/pci/lola/lola_pcm.c str->index = idx; idx 637 sound/pci/lola/lola_pcm.c str->dsd = idx; idx 665 sound/pci/lola/lola_pcm.c chip->input_src_caps_mask |= (1 << idx); idx 231 sound/pci/lx6464es/lx_core.c u8 idx = rmh->cmd_idx; idx 234 sound/pci/lx6464es/lx_core.c snd_printk(LXRMH "command %s\n", dsp_commands[idx].dcOpName); idx 942 sound/pci/mixart/mixart.c int idx = 0; idx 943 sound/pci/mixart/mixart.c for (subs = pcm->streams[stream].substream; subs; subs = subs->next, idx++) idx 1034 sound/pci/mixart/mixart.c static int snd_mixart_create(struct mixart_mgr *mgr, struct snd_card *card, int idx) idx 1047 sound/pci/mixart/mixart.c chip->chip_idx = idx; idx 1055 sound/pci/mixart/mixart.c mgr->chip[idx] = chip; idx 1317 sound/pci/mixart/mixart.c int idx; idx 1320 sound/pci/mixart/mixart.c idx = index[dev]; idx 1322 sound/pci/mixart/mixart.c idx = index[dev] + i; idx 1324 sound/pci/mixart/mixart.c err = snd_card_new(&pci->dev, idx, tmpid, THIS_MODULE, idx 710 sound/pci/mixart/mixart_mixer.c int mixart_update_playback_stream_level(struct snd_mixart* chip, int is_aes, int idx) idx 721 sound/pci/mixart/mixart_mixer.c set_level.stream_level.desc.stream_idx = idx; idx 725 sound/pci/mixart/mixart_mixer.c idx += MIXART_PLAYBACK_STREAMS; idx 737 sound/pci/mixart/mixart_mixer.c if(chip->digital_playback_active[idx][i]) idx 738 sound/pci/mixart/mixart_mixer.c volume[i] = chip->digital_playback_volume[idx][i]; idx 764 sound/pci/mixart/mixart_mixer.c int err, i, idx; idx 771 sound/pci/mixart/mixart_mixer.c idx = 1; idx 774 sound/pci/mixart/mixart_mixer.c idx = 0; idx 789 sound/pci/mixart/mixart_mixer.c set_level.level[i].digital_level = mixart_digital_level[chip->digital_capture_volume[idx][i]]; idx 824 sound/pci/mixart/mixart_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 833 sound/pci/mixart/mixart_mixer.c snd_BUG_ON(idx >= MIXART_PLAYBACK_STREAMS); idx 834 sound/pci/mixart/mixart_mixer.c if(is_aes) stored_volume = chip->digital_playback_volume[MIXART_PLAYBACK_STREAMS + idx]; /* AES playback */ idx 835 sound/pci/mixart/mixart_mixer.c else stored_volume = chip->digital_playback_volume[idx]; /* analog playback */ idx 846 sound/pci/mixart/mixart_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 859 sound/pci/mixart/mixart_mixer.c snd_BUG_ON(idx >= MIXART_PLAYBACK_STREAMS); idx 861 sound/pci/mixart/mixart_mixer.c stored_volume = chip->digital_playback_volume[MIXART_PLAYBACK_STREAMS + idx]; idx 863 sound/pci/mixart/mixart_mixer.c stored_volume = chip->digital_playback_volume[idx]; idx 879 sound/pci/mixart/mixart_mixer.c mixart_update_playback_stream_level(chip, is_aes, idx); idx 904 sound/pci/mixart/mixart_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 905 sound/pci/mixart/mixart_mixer.c snd_BUG_ON(idx >= MIXART_PLAYBACK_STREAMS); idx 908 sound/pci/mixart/mixart_mixer.c idx += MIXART_PLAYBACK_STREAMS; idx 909 sound/pci/mixart/mixart_mixer.c ucontrol->value.integer.value[0] = chip->digital_playback_active[idx][0]; idx 910 sound/pci/mixart/mixart_mixer.c ucontrol->value.integer.value[1] = chip->digital_playback_active[idx][1]; idx 920 sound/pci/mixart/mixart_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 922 sound/pci/mixart/mixart_mixer.c snd_BUG_ON(idx >= MIXART_PLAYBACK_STREAMS); idx 924 sound/pci/mixart/mixart_mixer.c j = idx; idx 936 sound/pci/mixart/mixart_mixer.c mixart_update_playback_stream_level(chip, is_aes, idx); idx 14 sound/pci/mixart/mixart_mixer.h int mixart_update_playback_stream_level(struct snd_mixart* chip, int is_aes, int idx); idx 1221 sound/pci/nm256/nm256.c int idx = nm256_ac97_idx(reg); idx 1223 sound/pci/nm256/nm256.c if (idx < 0) idx 1225 sound/pci/nm256/nm256.c return chip->ac97_regs[idx]; idx 1236 sound/pci/nm256/nm256.c int idx = nm256_ac97_idx(reg); idx 1239 sound/pci/nm256/nm256.c if (idx < 0) idx 1252 sound/pci/nm256/nm256.c chip->ac97_regs[idx] = val; idx 213 sound/pci/oxygen/xonar_dg_mixer.c unsigned int idx = ctl->private_value; idx 216 sound/pci/oxygen/xonar_dg_mixer.c value->value.integer.value[0] = data->input_vol[idx][0]; idx 217 sound/pci/oxygen/xonar_dg_mixer.c value->value.integer.value[1] = data->input_vol[idx][1]; idx 227 sound/pci/oxygen/xonar_dg_mixer.c unsigned int idx = ctl->private_value; idx 237 sound/pci/oxygen/xonar_dg_mixer.c changed = data->input_vol[idx][0] != value->value.integer.value[0] || idx 238 sound/pci/oxygen/xonar_dg_mixer.c data->input_vol[idx][1] != value->value.integer.value[1]; idx 240 sound/pci/oxygen/xonar_dg_mixer.c data->input_vol[idx][0] = value->value.integer.value[0]; idx 241 sound/pci/oxygen/xonar_dg_mixer.c data->input_vol[idx][1] = value->value.integer.value[1]; idx 242 sound/pci/oxygen/xonar_dg_mixer.c if (idx == data->input_sel) { idx 244 sound/pci/oxygen/xonar_dg_mixer.c data->input_vol[idx][0], idx 245 sound/pci/oxygen/xonar_dg_mixer.c data->input_vol[idx][1]); idx 1196 sound/pci/pcxhr/pcxhr.c struct snd_card *card, int idx) idx 1209 sound/pci/pcxhr/pcxhr.c chip->chip_idx = idx; idx 1212 sound/pci/pcxhr/pcxhr.c if (idx < mgr->playback_chips) idx 1216 sound/pci/pcxhr/pcxhr.c if (idx < mgr->capture_chips) { idx 1228 sound/pci/pcxhr/pcxhr.c mgr->chip[idx] = chip; idx 1601 sound/pci/pcxhr/pcxhr.c int idx; idx 1608 sound/pci/pcxhr/pcxhr.c idx = index[dev]; idx 1610 sound/pci/pcxhr/pcxhr.c idx = index[dev] + i; idx 1614 sound/pci/pcxhr/pcxhr.c err = snd_card_new(&pci->dev, idx, tmpid, THIS_MODULE, idx 223 sound/pci/pcxhr/pcxhr_mix22.c int idx, int level) idx 226 sound/pci/pcxhr/pcxhr_mix22.c if (idx > 1 || idx 231 sound/pci/pcxhr/pcxhr_mix22.c if (idx == 0) idx 626 sound/pci/pcxhr/pcxhr_mix22.c unsigned char idx = (unsigned char)(aes_idx * 8); idx 632 sound/pci/pcxhr/pcxhr_mix22.c PCXHR_OUTPB(chip->mgr, PCXHR_XLX_RUER, idx++); /* idx < 192 */ idx 650 sound/pci/pcxhr/pcxhr_mix22.c unsigned char idx = (unsigned char)(aes_idx * 8); idx 654 sound/pci/pcxhr/pcxhr_mix22.c PCXHR_OUTPB(chip->mgr, PCXHR_XLX_RUER, idx); idx 659 sound/pci/pcxhr/pcxhr_mix22.c idx++; idx 246 sound/pci/pcxhr/pcxhr_mixer.c static int pcxhr_update_playback_stream_level(struct snd_pcxhr* chip, int idx) idx 253 sound/pci/pcxhr/pcxhr_mixer.c if (chip->digital_playback_active[idx][0]) idx 254 sound/pci/pcxhr/pcxhr_mixer.c left = chip->digital_playback_volume[idx][0]; idx 257 sound/pci/pcxhr/pcxhr_mixer.c if (chip->digital_playback_active[idx][1]) idx 258 sound/pci/pcxhr/pcxhr_mixer.c right = chip->digital_playback_volume[idx][1]; idx 264 sound/pci/pcxhr/pcxhr_mixer.c pcxhr_set_pipe_cmd_params(&rmh, 0, pipe->first_audio, 0, 1<<idx); idx 351 sound/pci/pcxhr/pcxhr_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 359 sound/pci/pcxhr/pcxhr_mixer.c stored_volume = chip->digital_playback_volume[idx]; idx 370 sound/pci/pcxhr/pcxhr_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 380 sound/pci/pcxhr/pcxhr_mixer.c stored_volume = chip->digital_playback_volume[idx]; idx 394 sound/pci/pcxhr/pcxhr_mixer.c pcxhr_update_playback_stream_level(chip, idx); idx 417 sound/pci/pcxhr/pcxhr_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 420 sound/pci/pcxhr/pcxhr_mixer.c ucontrol->value.integer.value[0] = chip->digital_playback_active[idx][0]; idx 421 sound/pci/pcxhr/pcxhr_mixer.c ucontrol->value.integer.value[1] = chip->digital_playback_active[idx][1]; idx 431 sound/pci/pcxhr/pcxhr_mixer.c int idx = snd_ctl_get_ioffidx(kcontrol, &ucontrol->id); /* index */ idx 435 sound/pci/pcxhr/pcxhr_mixer.c j = idx; idx 445 sound/pci/pcxhr/pcxhr_mixer.c pcxhr_update_playback_stream_level(chip, idx); idx 1884 sound/pci/rme32.c int idx, err; idx 1887 sound/pci/rme32.c for (idx = 0; idx < (int)ARRAY_SIZE(snd_rme32_controls); idx++) { idx 1888 sound/pci/rme32.c if ((err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_rme32_controls[idx], rme32))) < 0) idx 1890 sound/pci/rme32.c if (idx == 1) /* IEC958 (S/PDIF) Stream */ idx 2343 sound/pci/rme96.c int idx, err; idx 2346 sound/pci/rme96.c for (idx = 0; idx < 7; idx++) { idx 2347 sound/pci/rme96.c if ((err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_rme96_controls[idx], rme96))) < 0) idx 2349 sound/pci/rme96.c if (idx == 1) /* IEC958 (S/PDIF) Stream */ idx 2354 sound/pci/rme96.c for (idx = 7; idx < 10; idx++) idx 2355 sound/pci/rme96.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_rme96_controls[idx], rme96))) < 0) idx 2773 sound/pci/rme9652/hdsp.c static int hdsp_adat_sync_check(struct hdsp *hdsp, int idx) idx 2777 sound/pci/rme9652/hdsp.c if (status & (HDSP_Lock0>>idx)) { idx 2778 sound/pci/rme9652/hdsp.c if (status & (HDSP_Sync0>>idx)) idx 3264 sound/pci/rme9652/hdsp.c unsigned int idx; idx 3270 sound/pci/rme9652/hdsp.c for (idx = 0; idx < ARRAY_SIZE(snd_hdsp_rpm_controls); idx++) { idx 3271 sound/pci/rme9652/hdsp.c err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_hdsp_rpm_controls[idx], hdsp)); idx 3278 sound/pci/rme9652/hdsp.c for (idx = 0; idx < ARRAY_SIZE(snd_hdsp_controls); idx++) { idx 3279 sound/pci/rme9652/hdsp.c if ((err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_hdsp_controls[idx], hdsp))) < 0) idx 3281 sound/pci/rme9652/hdsp.c if (idx == 1) /* IEC958 (S/PDIF) Stream */ idx 3291 sound/pci/rme9652/hdsp.c for (idx = 1; idx < 3; ++idx) { idx 3292 sound/pci/rme9652/hdsp.c snd_hdsp_adat_sync_check.index = idx+1; idx 3300 sound/pci/rme9652/hdsp.c for (idx = 0; idx < ARRAY_SIZE(snd_hdsp_9632_controls); idx++) { idx 3301 sound/pci/rme9652/hdsp.c if ((err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_hdsp_9632_controls[idx], hdsp))) < 0) idx 1101 sound/pci/rme9652/hdspm.c static int hdspm_aes_sync_check(struct hdspm *hdspm, int idx); idx 2368 sound/pci/rme9652/hdspm.c static int hdspm_get_s1_sample_rate(struct hdspm *hdspm, unsigned int idx) idx 2372 sound/pci/rme9652/hdspm.c return (status >> (idx*4)) & 0xF; idx 3891 sound/pci/rme9652/hdspm.c static int hdspm_s1_sync_check(struct hdspm *hdspm, int idx) idx 3897 sound/pci/rme9652/hdspm.c lock = (status & (0x1<<idx)) ? 1 : 0; idx 3898 sound/pci/rme9652/hdspm.c sync = (status & (0x100<<idx)) ? 1 : 0; idx 3944 sound/pci/rme9652/hdspm.c static int hdspm_aes_sync_check(struct hdspm *hdspm, int idx) idx 3949 sound/pci/rme9652/hdspm.c lock = (status2 & (0x0080 >> idx)) ? 1 : 0; idx 3950 sound/pci/rme9652/hdspm.c sync = (status2 & (0x8000 >> idx)) ? 1 : 0; idx 4671 sound/pci/rme9652/hdspm.c unsigned int idx, limit; idx 4700 sound/pci/rme9652/hdspm.c for (idx = 0; idx < limit; idx++) { idx 4702 sound/pci/rme9652/hdspm.c snd_ctl_new1(&list[idx], hdspm)); idx 4718 sound/pci/rme9652/hdspm.c for (idx = 0; idx < limit; ++idx) { idx 4719 sound/pci/rme9652/hdspm.c snd_hdspm_playback_mixer.index = idx + 1; idx 4724 sound/pci/rme9652/hdspm.c hdspm->playback_mixer_ctls[idx] = kctl; idx 4732 sound/pci/rme9652/hdspm.c for (idx = 0; idx < limit; idx++) { idx 4734 sound/pci/rme9652/hdspm.c snd_ctl_new1(&list[idx], hdspm)); idx 1525 sound/pci/rme9652/rme9652.c unsigned int idx; idx 1529 sound/pci/rme9652/rme9652.c for (idx = 0; idx < ARRAY_SIZE(snd_rme9652_controls); idx++) { idx 1530 sound/pci/rme9652/rme9652.c if ((err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_rme9652_controls[idx], rme9652))) < 0) idx 1532 sound/pci/rme9652/rme9652.c if (idx == 1) /* IEC958 (S/PDIF) Stream */ idx 1105 sound/pci/sonicvibes.c unsigned int idx; idx 1113 sound/pci/sonicvibes.c for (idx = 0; idx < ARRAY_SIZE(snd_sonicvibes_controls); idx++) { idx 1114 sound/pci/sonicvibes.c if ((err = snd_ctl_add(card, kctl = snd_ctl_new1(&snd_sonicvibes_controls[idx], sonic))) < 0) idx 1116 sound/pci/sonicvibes.c switch (idx) { idx 1420 sound/pci/sonicvibes.c unsigned int idx; idx 1426 sound/pci/sonicvibes.c for (idx = 0; idx < ARRAY_SIZE(snd_sonicvibes_midi_controls); idx++) idx 1427 sound/pci/sonicvibes.c if ((err = snd_ctl_add(card, snd_ctl_new1(&snd_sonicvibes_midi_controls[idx], sonic))) < 0) idx 1440 sound/pci/sonicvibes.c int idx, err; idx 1453 sound/pci/sonicvibes.c for (idx = 0; idx < 5; idx++) { idx 1454 sound/pci/sonicvibes.c if (pci_resource_start(pci, idx) == 0 || idx 1455 sound/pci/sonicvibes.c !(pci_resource_flags(pci, idx) & IORESOURCE_IO)) { idx 334 sound/pci/trident/trident_main.c int idx; idx 338 sound/pci/trident/trident_main.c for (idx = 31; idx >= 0; idx--) { idx 339 sound/pci/trident/trident_main.c if (!(trident->ChanMap[T4D_BANK_B] & (1 << idx))) { idx 340 sound/pci/trident/trident_main.c trident->ChanMap[T4D_BANK_B] |= 1 << idx; idx 342 sound/pci/trident/trident_main.c return idx + 32; idx 384 sound/pci/trident/trident_main.c int idx; idx 386 sound/pci/trident/trident_main.c for (idx = 31; idx >= 0; idx--) { idx 387 sound/pci/trident/trident_main.c if (!(trident->ChanMap[T4D_BANK_A] & (1 << idx))) { idx 388 sound/pci/trident/trident_main.c trident->ChanMap[T4D_BANK_A] |= 1 << idx; idx 390 sound/pci/trident/trident_main.c return idx; idx 2952 sound/pci/trident/trident_main.c int idx, err, retries = 2; idx 3014 sound/pci/trident/trident_main.c for (idx = 0; idx < 32; idx++) { idx 3017 sound/pci/trident/trident_main.c tmix = &trident->pcm_mixer[idx]; idx 3056 sound/pci/trident/trident_main.c idx = kctl->id.index; idx 3066 sound/pci/trident/trident_main.c kctl->id.index = idx; idx 3076 sound/pci/trident/trident_main.c kctl->id.index = idx; idx 3086 sound/pci/trident/trident_main.c kctl->id.index = idx; idx 3804 sound/pci/trident/trident_main.c int idx; idx 3808 sound/pci/trident/trident_main.c idx = snd_trident_allocate_pcm_channel(trident); idx 3809 sound/pci/trident/trident_main.c if(idx < 0) { idx 3813 sound/pci/trident/trident_main.c pvoice = &trident->synth.voices[idx]; idx 3824 sound/pci/trident/trident_main.c idx = snd_trident_allocate_synth_channel(trident); idx 3825 sound/pci/trident/trident_main.c if(idx < 0) { idx 3829 sound/pci/trident/trident_main.c pvoice = &trident->synth.voices[idx]; idx 180 sound/pci/trident/trident_memory.c int idx, page; idx 200 sound/pci/trident/trident_memory.c idx = 0; idx 201 sound/pci/trident/trident_memory.c for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) { idx 202 sound/pci/trident/trident_memory.c unsigned long ofs = idx << PAGE_SHIFT; idx 415 sound/pci/via82xx.c unsigned int i, idx, ofs, rest; idx 436 sound/pci/via82xx.c idx = 0; idx 449 sound/pci/via82xx.c if (idx >= VIA_TABLE_SIZE) { idx 454 sound/pci/via82xx.c ((u32 *)dev->table.area)[idx << 1] = cpu_to_le32(addr); idx 469 sound/pci/via82xx.c ((u32 *)dev->table.area)[(idx<<1) + 1] = cpu_to_le32(r | flag); idx 470 sound/pci/via82xx.c dev->idx_table[idx].offset = ofs; idx 471 sound/pci/via82xx.c dev->idx_table[idx].size = r; idx 473 sound/pci/via82xx.c idx++; idx 476 sound/pci/via82xx.c dev->tbl_entries = idx; idx 772 sound/pci/via82xx.c unsigned int idx, idx 777 sound/pci/via82xx.c size = viadev->idx_table[idx].size; idx 778 sound/pci/via82xx.c base = viadev->idx_table[idx].offset; idx 805 sound/pci/via82xx.c idx, viadev->tbl_entries, idx 807 sound/pci/via82xx.c viadev->idx_table[idx].offset, idx 808 sound/pci/via82xx.c viadev->idx_table[idx].size, count); idx 829 sound/pci/via82xx.c unsigned int idx, ptr, count, res; idx 843 sound/pci/via82xx.c idx = 0; idx 845 sound/pci/via82xx.c idx = ((ptr - (unsigned int)viadev->table.addr) / 8 - 1) % viadev->tbl_entries; idx 846 sound/pci/via82xx.c res = calc_linear_pos(chip, viadev, idx, count); idx 860 sound/pci/via82xx.c unsigned int idx, count, res; idx 882 sound/pci/via82xx.c idx = count >> 24; idx 883 sound/pci/via82xx.c if (idx >= viadev->tbl_entries) { idx 886 sound/pci/via82xx.c "fail: invalid idx = %i/%i\n", idx, idx 892 sound/pci/via82xx.c res = calc_linear_pos(chip, viadev, idx, count); idx 1196 sound/pci/via82xx.c int idx = viadev->direction ? AC97_RATES_ADC : AC97_RATES_FRONT_DAC; idx 1197 sound/pci/via82xx.c runtime->hw.rates = chip->ac97->rates[idx]; idx 1422 sound/pci/via82xx.c static void init_viadev(struct via82xx *chip, int idx, unsigned int reg_offset, idx 1425 sound/pci/via82xx.c chip->devs[idx].reg_offset = reg_offset; idx 1426 sound/pci/via82xx.c chip->devs[idx].shadow_shift = shadow_pos * 4; idx 1427 sound/pci/via82xx.c chip->devs[idx].direction = direction; idx 1428 sound/pci/via82xx.c chip->devs[idx].port = chip->port + reg_offset; idx 1696 sound/pci/via82xx.c unsigned int idx = kcontrol->id.subdevice; idx 1698 sound/pci/via82xx.c ucontrol->value.integer.value[0] = VIA_DXS_MAX_VOLUME - chip->playback_volume[idx][0]; idx 1699 sound/pci/via82xx.c ucontrol->value.integer.value[1] = VIA_DXS_MAX_VOLUME - chip->playback_volume[idx][1]; idx 1716 sound/pci/via82xx.c unsigned int idx = kcontrol->id.subdevice; idx 1717 sound/pci/via82xx.c unsigned long port = chip->port + 0x10 * idx; idx 1726 sound/pci/via82xx.c change |= val != chip->playback_volume[idx][i]; idx 1728 sound/pci/via82xx.c chip->playback_volume[idx][i] = val; idx 1739 sound/pci/via82xx.c unsigned int idx; idx 1751 sound/pci/via82xx.c for (idx = 0; idx < 4; idx++) { idx 1752 sound/pci/via82xx.c unsigned long port = chip->port + 0x10 * idx; idx 1753 sound/pci/via82xx.c chip->playback_volume[idx][i] = val; idx 2240 sound/pci/via82xx.c int i, idx; idx 2241 sound/pci/via82xx.c for (idx = 0; idx < 4; idx++) { idx 2242 sound/pci/via82xx.c unsigned long port = chip->port + 0x10 * idx; idx 2244 sound/pci/via82xx.c chip->playback_volume[idx][i]=chip->playback_volume_c[i]; idx 268 sound/pci/via82xx_modem.c unsigned int i, idx, ofs, rest; idx 289 sound/pci/via82xx_modem.c idx = 0; idx 302 sound/pci/via82xx_modem.c if (idx >= VIA_TABLE_SIZE) { idx 307 sound/pci/via82xx_modem.c ((u32 *)dev->table.area)[idx << 1] = cpu_to_le32(addr); idx 324 sound/pci/via82xx_modem.c ((u32 *)dev->table.area)[(idx<<1) + 1] = cpu_to_le32(r | flag); idx 325 sound/pci/via82xx_modem.c dev->idx_table[idx].offset = ofs; idx 326 sound/pci/via82xx_modem.c dev->idx_table[idx].size = r; idx 328 sound/pci/via82xx_modem.c idx++; idx 331 sound/pci/via82xx_modem.c dev->tbl_entries = idx; idx 555 sound/pci/via82xx_modem.c unsigned int idx, idx 560 sound/pci/via82xx_modem.c size = viadev->idx_table[idx].size; idx 561 sound/pci/via82xx_modem.c res = viadev->idx_table[idx].offset + size - count; idx 573 sound/pci/via82xx_modem.c idx, viadev->tbl_entries, viadev->lastpos, idx 574 sound/pci/via82xx_modem.c viadev->bufsize2, viadev->idx_table[idx].offset, idx 575 sound/pci/via82xx_modem.c viadev->idx_table[idx].size, count); idx 584 sound/pci/via82xx_modem.c res = viadev->idx_table[idx].offset; idx 589 sound/pci/via82xx_modem.c res = viadev->idx_table[idx].offset + size; idx 610 sound/pci/via82xx_modem.c unsigned int idx, ptr, count, res; idx 624 sound/pci/via82xx_modem.c idx = 0; idx 626 sound/pci/via82xx_modem.c idx = ((ptr - (unsigned int)viadev->table.addr) / 8 - 1) % idx 628 sound/pci/via82xx_modem.c res = calc_linear_pos(chip, viadev, idx, count); idx 821 sound/pci/via82xx_modem.c static void init_viadev(struct via82xx_modem *chip, int idx, unsigned int reg_offset, idx 824 sound/pci/via82xx_modem.c chip->devs[idx].reg_offset = reg_offset; idx 825 sound/pci/via82xx_modem.c chip->devs[idx].direction = direction; idx 826 sound/pci/via82xx_modem.c chip->devs[idx].port = chip->port + reg_offset; idx 210 sound/pci/ymfpci/ymfpci_main.c int idx; idx 213 sound/pci/ymfpci/ymfpci_main.c for (idx = 0; idx < YDSXG_PLAYBACK_VOICES; idx += pair ? 2 : 1) { idx 214 sound/pci/ymfpci/ymfpci_main.c voice = &chip->voices[idx]; idx 215 sound/pci/ymfpci/ymfpci_main.c voice2 = pair ? &chip->voices[idx+1] : NULL; idx 1797 sound/pci/ymfpci/ymfpci_main.c unsigned int idx; idx 1819 sound/pci/ymfpci/ymfpci_main.c for (idx = 0; idx < ARRAY_SIZE(snd_ymfpci_controls); idx++) { idx 1820 sound/pci/ymfpci/ymfpci_main.c if ((err = snd_ctl_add(chip->card, snd_ctl_new1(&snd_ymfpci_controls[idx], chip))) < 0) idx 1859 sound/pci/ymfpci/ymfpci_main.c for (idx = 0; idx < 32; ++idx) { idx 1864 sound/pci/ymfpci/ymfpci_main.c kctl->id.subdevice = idx; idx 1868 sound/pci/ymfpci/ymfpci_main.c chip->pcm_mixer[idx].left = 0x8000; idx 1869 sound/pci/ymfpci/ymfpci_main.c chip->pcm_mixer[idx].right = 0x8000; idx 1870 sound/pci/ymfpci/ymfpci_main.c chip->pcm_mixer[idx].ctl = kctl; idx 621 sound/ppc/tumbler.c static int snapper_set_mix_vol1(struct pmac_tumbler *mix, int idx, int ch, int reg) idx 626 sound/ppc/tumbler.c vol = mix->mix_vol[idx][ch]; idx 629 sound/ppc/tumbler.c mix->mix_vol[idx][ch] = vol; idx 646 sound/ppc/tumbler.c static int snapper_set_mix_vol(struct pmac_tumbler *mix, int idx) idx 650 sound/ppc/tumbler.c if (snapper_set_mix_vol1(mix, idx, 0, TAS_REG_LMIX) < 0 || idx 651 sound/ppc/tumbler.c snapper_set_mix_vol1(mix, idx, 1, TAS_REG_RMIX) < 0) idx 669 sound/ppc/tumbler.c int idx = (int)kcontrol->private_value; idx 674 sound/ppc/tumbler.c ucontrol->value.integer.value[0] = mix->mix_vol[idx][0]; idx 675 sound/ppc/tumbler.c ucontrol->value.integer.value[1] = mix->mix_vol[idx][1]; idx 682 sound/ppc/tumbler.c int idx = (int)kcontrol->private_value; idx 695 sound/ppc/tumbler.c change = mix->mix_vol[idx][0] != vol[0] || idx 696 sound/ppc/tumbler.c mix->mix_vol[idx][1] != vol[1]; idx 698 sound/ppc/tumbler.c mix->mix_vol[idx][0] = vol[0]; idx 699 sound/ppc/tumbler.c mix->mix_vol[idx][1] = vol[1]; idx 700 sound/ppc/tumbler.c snapper_set_mix_vol(mix, idx); idx 816 sound/ppc/tumbler.c #define DEFINE_SNAPPER_MIX(xname,idx,ofs) { \ idx 822 sound/ppc/tumbler.c .index = idx,\ idx 451 sound/soc/codecs/hdmi-codec.c int ret, idx; idx 472 sound/soc/codecs/hdmi-codec.c idx = hdmi_codec_get_ch_alloc_table_idx(hcp, hp.cea.channels); idx 473 sound/soc/codecs/hdmi-codec.c if (idx < 0) { idx 475 sound/soc/codecs/hdmi-codec.c idx); idx 477 sound/soc/codecs/hdmi-codec.c return idx; idx 479 sound/soc/codecs/hdmi-codec.c hp.cea.channel_allocation = hdmi_codec_channel_alloc[idx].ca_id; idx 480 sound/soc/codecs/hdmi-codec.c hcp->chmap_idx = hdmi_codec_channel_alloc[idx].ca_id; idx 328 sound/soc/codecs/mt6351.c int idx, old_idx, offset, reg_idx; idx 331 sound/soc/codecs/mt6351.c idx = 8; /* 0dB */ idx 334 sound/soc/codecs/mt6351.c idx = priv->ana_gain[AUDIO_ANALOG_VOLUME_HPOUTL]; idx 338 sound/soc/codecs/mt6351.c __func__, idx, old_idx); idx 340 sound/soc/codecs/mt6351.c if (idx > old_idx) idx 341 sound/soc/codecs/mt6351.c offset = idx - old_idx; idx 343 sound/soc/codecs/mt6351.c offset = old_idx - idx; idx 348 sound/soc/codecs/mt6351.c reg_idx = idx > old_idx ? reg_idx + 1 : reg_idx - 1; idx 520 sound/soc/codecs/rt5514.c int idx; idx 522 sound/soc/codecs/rt5514.c idx = rt5514_calc_dmic_clk(component, rt5514->sysclk); idx 523 sound/soc/codecs/rt5514.c if (idx < 0) idx 528 sound/soc/codecs/rt5514.c idx << RT5514_CLK_DMIC_OUT_SEL_SFT); idx 533 sound/soc/codecs/rt5514.c return idx; idx 464 sound/soc/codecs/rt5640.c int idx, rate; idx 468 sound/soc/codecs/rt5640.c idx = rl6231_calc_dmic_clk(rate); idx 469 sound/soc/codecs/rt5640.c if (idx < 0) idx 473 sound/soc/codecs/rt5640.c idx << RT5640_DMIC_CLK_SFT); idx 474 sound/soc/codecs/rt5640.c return idx; idx 850 sound/soc/codecs/rt5645.c int idx, rate; idx 854 sound/soc/codecs/rt5645.c idx = rl6231_calc_dmic_clk(rate); idx 855 sound/soc/codecs/rt5645.c if (idx < 0) idx 859 sound/soc/codecs/rt5645.c RT5645_DMIC_CLK_MASK, idx << RT5645_DMIC_CLK_SFT); idx 860 sound/soc/codecs/rt5645.c return idx; idx 381 sound/soc/codecs/rt5651.c int idx, rate; idx 385 sound/soc/codecs/rt5651.c idx = rl6231_calc_dmic_clk(rate); idx 386 sound/soc/codecs/rt5651.c if (idx < 0) idx 390 sound/soc/codecs/rt5651.c idx << RT5651_DMIC_CLK_SFT); idx 392 sound/soc/codecs/rt5651.c return idx; idx 1607 sound/soc/codecs/rt5659.c int pd, idx = -EINVAL; idx 1611 sound/soc/codecs/rt5659.c idx = rl6231_calc_dmic_clk(rt5659->sysclk / pd); idx 1613 sound/soc/codecs/rt5659.c if (idx < 0) idx 1617 sound/soc/codecs/rt5659.c RT5659_DMIC_CLK_MASK, idx << RT5659_DMIC_CLK_SFT); idx 1619 sound/soc/codecs/rt5659.c return idx; idx 356 sound/soc/codecs/rt5660.c int idx, rate; idx 360 sound/soc/codecs/rt5660.c idx = rl6231_calc_dmic_clk(rate); idx 361 sound/soc/codecs/rt5660.c if (idx < 0) idx 365 sound/soc/codecs/rt5660.c RT5660_DMIC_CLK_MASK, idx << RT5660_DMIC_CLK_SFT); idx 367 sound/soc/codecs/rt5660.c return idx; idx 1481 sound/soc/codecs/rt5665.c int pd, idx; idx 1485 sound/soc/codecs/rt5665.c idx = rl6231_calc_dmic_clk(rt5665->sysclk / pd); idx 1487 sound/soc/codecs/rt5665.c if (idx < 0) idx 1491 sound/soc/codecs/rt5665.c RT5665_DMIC_CLK_MASK, idx << RT5665_DMIC_CLK_SFT); idx 1493 sound/soc/codecs/rt5665.c return idx; idx 1174 sound/soc/codecs/rt5668.c int idx = -EINVAL; idx 1177 sound/soc/codecs/rt5668.c idx = rt5668_div_sel(rt5668, 1500000, div, ARRAY_SIZE(div)); idx 1180 sound/soc/codecs/rt5668.c RT5668_DMIC_CLK_MASK, idx << RT5668_DMIC_CLK_SFT); idx 1191 sound/soc/codecs/rt5668.c int ref, val, reg, idx = -EINVAL; idx 1202 sound/soc/codecs/rt5668.c idx = rt5668_div_sel(rt5668, ref, div, ARRAY_SIZE(div)); idx 1210 sound/soc/codecs/rt5668.c RT5668_FILTER_CLK_SEL_MASK, idx << RT5668_FILTER_CLK_SEL_SFT); idx 701 sound/soc/codecs/rt5670.c int idx, rate; idx 705 sound/soc/codecs/rt5670.c idx = rl6231_calc_dmic_clk(rate); idx 706 sound/soc/codecs/rt5670.c if (idx < 0) idx 710 sound/soc/codecs/rt5670.c RT5670_DMIC_CLK_MASK, idx << RT5670_DMIC_CLK_SFT); idx 711 sound/soc/codecs/rt5670.c return idx; idx 919 sound/soc/codecs/rt5677.c int idx, rate; idx 923 sound/soc/codecs/rt5677.c idx = rl6231_calc_dmic_clk(rate); idx 924 sound/soc/codecs/rt5677.c if (idx < 0) idx 928 sound/soc/codecs/rt5677.c RT5677_DMIC_CLK_MASK, idx << RT5677_DMIC_CLK_SFT); idx 929 sound/soc/codecs/rt5677.c return idx; idx 1200 sound/soc/codecs/rt5682.c int idx = -EINVAL; idx 1203 sound/soc/codecs/rt5682.c idx = rt5682_div_sel(rt5682, 1500000, div, ARRAY_SIZE(div)); idx 1206 sound/soc/codecs/rt5682.c RT5682_DMIC_CLK_MASK, idx << RT5682_DMIC_CLK_SFT); idx 1217 sound/soc/codecs/rt5682.c int ref, val, reg, idx = -EINVAL; idx 1229 sound/soc/codecs/rt5682.c idx = rt5682_div_sel(rt5682, ref, div_f, ARRAY_SIZE(div_f)); idx 1237 sound/soc/codecs/rt5682.c RT5682_FILTER_CLK_DIV_MASK, idx << RT5682_FILTER_CLK_DIV_SFT); idx 1240 sound/soc/codecs/rt5682.c for (idx = 0; idx < ARRAY_SIZE(div_o); idx++) { idx 1241 sound/soc/codecs/rt5682.c if (rt5682->sysclk <= 12288000 * div_o[idx]) idx 1247 sound/soc/codecs/rt5682.c (idx << RT5682_ADC_OSR_SFT) | (idx << RT5682_DAC_OSR_SFT)); idx 72 sound/soc/codecs/ts3a227e.c #define PRESS_MASK(idx) (0x01 << (2 * (idx))) idx 73 sound/soc/codecs/ts3a227e.c #define RELEASE_MASK(idx) (0x02 << (2 * (idx))) idx 522 sound/soc/codecs/wm8978.c int idx = wm8978_enum_mclk(f_256fs, f_mclk, &wm8978->f_pllout); idx 523 sound/soc/codecs/wm8978.c if (idx < 0) idx 524 sound/soc/codecs/wm8978.c return idx; idx 526 sound/soc/codecs/wm8978.c wm8978->mclk_idx = idx; idx 488 sound/soc/dwc/dwc-i2s.c u32 idx; idx 500 sound/soc/dwc/dwc-i2s.c idx = COMP1_TX_WORDSIZE_0(comp1); idx 501 sound/soc/dwc/dwc-i2s.c if (WARN_ON(idx >= ARRAY_SIZE(formats))) idx 504 sound/soc/dwc/dwc-i2s.c idx = 1; idx 508 sound/soc/dwc/dwc-i2s.c dw_i2s_dai->playback.formats = formats[idx]; idx 514 sound/soc/dwc/dwc-i2s.c idx = COMP2_RX_WORDSIZE_0(comp2); idx 515 sound/soc/dwc/dwc-i2s.c if (WARN_ON(idx >= ARRAY_SIZE(formats))) idx 518 sound/soc/dwc/dwc-i2s.c idx = 1; idx 522 sound/soc/dwc/dwc-i2s.c dw_i2s_dai->capture.formats = formats[idx]; idx 544 sound/soc/dwc/dwc-i2s.c u32 idx = COMP1_APB_DATA_WIDTH(comp1); idx 547 sound/soc/dwc/dwc-i2s.c if (WARN_ON(idx >= ARRAY_SIZE(bus_widths))) idx 555 sound/soc/dwc/dwc-i2s.c idx = 1; idx 563 sound/soc/dwc/dwc-i2s.c dev->play_dma_data.pd.addr_width = bus_widths[idx]; idx 564 sound/soc/dwc/dwc-i2s.c dev->capture_dma_data.pd.addr_width = bus_widths[idx]; idx 578 sound/soc/dwc/dwc-i2s.c u32 idx = COMP1_APB_DATA_WIDTH(comp1); idx 582 sound/soc/dwc/dwc-i2s.c if (WARN_ON(idx >= ARRAY_SIZE(bus_widths))) idx 594 sound/soc/dwc/dwc-i2s.c dev->play_dma_data.dt.addr_width = bus_widths[idx]; idx 604 sound/soc/dwc/dwc-i2s.c dev->capture_dma_data.dt.addr_width = bus_widths[idx]; idx 640 sound/soc/fsl/fsl-asoc-card.c u32 idx; idx 642 sound/soc/fsl/fsl-asoc-card.c ret = of_property_read_u32(cpu_np, "cell-index", &idx); idx 652 sound/soc/fsl/fsl-asoc-card.c (unsigned int)idx); idx 729 sound/soc/fsl/fsl_spdif.c int idx = (ctrl->ready_buf - 1) * SPDIF_UBITS_SIZE; idx 731 sound/soc/fsl/fsl_spdif.c &ctrl->subcode[idx], SPDIF_UBITS_SIZE); idx 761 sound/soc/fsl/fsl_spdif.c int idx = (ctrl->ready_buf - 1) * SPDIF_QSUB_SIZE; idx 763 sound/soc/fsl/fsl_spdif.c &ctrl->qsub[idx], SPDIF_QSUB_SIZE); idx 1416 sound/soc/intel/atom/sst-atom-controls.c char *idx; idx 1421 sound/soc/intel/atom/sst-atom-controls.c idx = strchr(kctl->id.name, ' '); idx 1422 sound/soc/intel/atom/sst-atom-controls.c if (idx == NULL) idx 1424 sound/soc/intel/atom/sst-atom-controls.c index = idx - (char*)kctl->id.name; idx 1814 sound/soc/intel/haswell/sst-haswell-ipc.c int ret, idx; idx 1822 sound/soc/intel/haswell/sst-haswell-ipc.c for (idx = 0; idx < hsw->param_idx_w; idx++) { idx 1824 sound/soc/intel/haswell/sst-haswell-ipc.c SST_HSW_MODULE_WAVES, 0, hsw->param_buf[idx][0], idx 1825 sound/soc/intel/haswell/sst-haswell-ipc.c WAVES_PARAM_COUNT, hsw->param_buf[idx]); idx 443 sound/soc/intel/haswell/sst-haswell-pcm.c u32 idx = (((i << 2) + i)) >> 1; idx 447 sound/soc/intel/haswell/sst-haswell-pcm.c dev_dbg(rtd->dev, "pfn i %i idx %d pfn %x\n", i, idx, pfn); idx 449 sound/soc/intel/haswell/sst-haswell-pcm.c pg_table = (u32 *)(pdata->dmab[pcm][stream].area + idx); idx 215 sound/soc/qcom/qdsp6/q6adm.c int idx; idx 217 sound/soc/qcom/qdsp6/q6adm.c idx = find_first_zero_bit(&adm->copp_bitmap[port_idx], idx 220 sound/soc/qcom/qdsp6/q6adm.c if (idx > MAX_COPPS_PER_PORT) idx 227 sound/soc/qcom/qdsp6/q6adm.c set_bit(idx, &adm->copp_bitmap[port_idx]); idx 228 sound/soc/qcom/qdsp6/q6adm.c c->copp_idx = idx; idx 390 sound/soc/qcom/qdsp6/q6routing.c int idx; idx 396 sound/soc/qcom/qdsp6/q6routing.c for_each_set_bit(idx, &session->copp_map, MAX_COPPS_PER_PORT) { idx 397 sound/soc/qcom/qdsp6/q6routing.c if (session->copps[idx]) { idx 398 sound/soc/qcom/qdsp6/q6routing.c q6adm_close(routing_data->dev, session->copps[idx]); idx 399 sound/soc/qcom/qdsp6/q6routing.c session->copps[idx] = NULL; idx 114 sound/soc/sh/rcar/adg.c int idx, sel, div, step; idx 129 sound/soc/sh/rcar/adg.c idx = 0; idx 138 sound/soc/sh/rcar/adg.c val = (sel << 8) | idx; idx 147 sound/soc/sh/rcar/adg.c if ((idx > 2) && (idx % 2)) idx 149 sound/soc/sh/rcar/adg.c if (idx == 0x1c) { idx 153 sound/soc/sh/rcar/adg.c idx++; idx 635 sound/soc/sh/rcar/dma.c int nr, i, idx; idx 686 sound/soc/sh/rcar/dma.c idx = 0; idx 687 sound/soc/sh/rcar/dma.c mod[idx++] = mod_start; idx 690 sound/soc/sh/rcar/dma.c mod[idx++] = src; idx 693 sound/soc/sh/rcar/dma.c mod[idx++] = ctu; idx 696 sound/soc/sh/rcar/dma.c mod[idx++] = mix; idx 699 sound/soc/sh/rcar/dma.c mod[idx++] = dvc; idx 703 sound/soc/sh/rcar/dma.c mod[idx] = mod_end; idx 712 sound/soc/sh/rcar/dma.c *mod_from = mod[idx - 1]; idx 713 sound/soc/sh/rcar/dma.c *mod_to = mod[idx]; idx 720 sound/soc/sh/rcar/dma.c for (i = 0; i <= idx; i++) { idx 37 sound/soc/sh/rcar/gen.c int idx; idx 45 sound/soc/sh/rcar/gen.c .idx = id, \ idx 202 sound/soc/sh/rcar/gen.c gen->regs[conf[i].idx] = regs; idx 203 sound/soc/sh/rcar/gen.c gen->reg_name[conf[i].idx] = conf[i].reg_name; idx 780 sound/soc/sh/rcar/rsnd.h int param1, int param2, int *idx); idx 207 sound/soc/sh/rcar/src.c int idx; idx 310 sound/soc/sh/rcar/src.c for (idx = 0; idx < ARRAY_SIZE(chan222222); idx++) idx 311 sound/soc/sh/rcar/src.c if (chptn[idx] & (1 << chan)) idx 315 sound/soc/sh/rcar/src.c idx >= ARRAY_SIZE(chan222222)) idx 330 sound/soc/sh/rcar/src.c rsnd_mod_write(mod, SRC_BSDSR, bsdsr_table[idx]); idx 331 sound/soc/sh/rcar/src.c rsnd_mod_write(mod, SRC_BSISR, bsisr_table[idx]); idx 230 sound/soc/sh/rcar/ssi.c int param1, int param2, int *idx) idx 257 sound/soc/sh/rcar/ssi.c if (idx) idx 258 sound/soc/sh/rcar/ssi.c *idx = j; idx 274 sound/soc/sh/rcar/ssi.c int idx, ret; idx 308 sound/soc/sh/rcar/ssi.c main_rate = rsnd_ssi_clk_query(rdai, rate, chan, &idx); idx 329 sound/soc/sh/rcar/ssi.c SCKD | SWSD | CKDV(idx); idx 151 sound/soc/sh/siu_dai.c u32 idx; idx 155 sound/soc/sh/siu_dai.c idx = 1; /* portA */ idx 157 sound/soc/sh/siu_dai.c idx = 2; /* portB */ idx 159 sound/soc/sh/siu_dai.c ydef[0] = (fw->spbpar[idx].ab1a << 16) | idx 160 sound/soc/sh/siu_dai.c (fw->spbpar[idx].ab0a << 8) | idx 161 sound/soc/sh/siu_dai.c (fw->spbpar[idx].dir << 7) | 3; idx 166 sound/soc/sh/siu_dai.c ydef[7] = fw->spbpar[idx].event; idx 167 sound/soc/sh/siu_dai.c port_info->stfifo |= fw->spbpar[idx].stfifo; idx 168 sound/soc/sh/siu_dai.c port_info->trdat |= fw->spbpar[idx].trdat; idx 176 sound/soc/sh/siu_dai.c u32 idx; idx 180 sound/soc/sh/siu_dai.c idx = 7; /* portA */ idx 182 sound/soc/sh/siu_dai.c idx = 8; /* portB */ idx 184 sound/soc/sh/siu_dai.c ydef[5] = (fw->spbpar[idx].ab1a << 16) | idx 185 sound/soc/sh/siu_dai.c (fw->spbpar[idx].ab0a << 8) | 1; idx 186 sound/soc/sh/siu_dai.c ydef[6] = fw->spbpar[idx].event; idx 187 sound/soc/sh/siu_dai.c port_info->stfifo |= fw->spbpar[idx].stfifo; idx 188 sound/soc/sh/siu_dai.c port_info->trdat |= fw->spbpar[idx].trdat; idx 123 sound/soc/soc-core.c struct attribute *attr, int idx) idx 378 sound/soc/soc-jack.c gpios[i].idx, GPIOD_IN); idx 224 sound/soc/sof/core.c u32 idx = (5 * i) >> 1; idx 228 sound/soc/sof/core.c dev_vdbg(sdev->dev, "pfn i %i idx %d pfn %x\n", i, idx, pfn); idx 230 sound/soc/sof/core.c pg_table = (u8 *)(page_table + idx); idx 41 sound/soc/sof/intel/hda-bus.c bus->idx = 0; idx 891 sound/sparc/amd7930.c int idx, err; idx 899 sound/sparc/amd7930.c for (idx = 0; idx < ARRAY_SIZE(amd7930_controls); idx++) { idx 901 sound/sparc/amd7930.c snd_ctl_new1(&amd7930_controls[idx], amd))) < 0) idx 1539 sound/sparc/cs4231.c int err, idx; idx 1546 sound/sparc/cs4231.c for (idx = 0; idx < ARRAY_SIZE(snd_cs4231_controls); idx++) { idx 1548 sound/sparc/cs4231.c snd_ctl_new1(&snd_cs4231_controls[idx], chip)); idx 2450 sound/sparc/dbri.c int idx, err; idx 2459 sound/sparc/dbri.c for (idx = 0; idx < ARRAY_SIZE(dbri_controls); idx++) { idx 2461 sound/sparc/dbri.c snd_ctl_new1(&dbri_controls[idx], dbri)); idx 2466 sound/sparc/dbri.c for (idx = DBRI_REC; idx < DBRI_NO_STREAMS; idx++) { idx 2467 sound/sparc/dbri.c dbri->stream_info[idx].left_gain = 0; idx 2468 sound/sparc/dbri.c dbri->stream_info[idx].right_gain = 0; idx 715 sound/spi/at73c213.c int errval, idx; idx 724 sound/spi/at73c213.c for (idx = 0; idx < ARRAY_SIZE(snd_at73c213_controls); idx++) { idx 726 sound/spi/at73c213.c snd_ctl_new1(&snd_at73c213_controls[idx], idx 735 sound/spi/at73c213.c for (idx = 1; idx < ARRAY_SIZE(snd_at73c213_controls) + 1; idx++) { idx 737 sound/spi/at73c213.c kctl = snd_ctl_find_numid(card, idx); idx 115 sound/usb/card.c int idx; idx 118 sound/usb/card.c for (idx = 0; idx < 2; idx++) { idx 119 sound/usb/card.c subs = &as->substream[idx]; idx 437 sound/usb/card.c struct usb_device *dev, int idx, idx 462 sound/usb/card.c err = snd_card_new(&intf->dev, index[idx], id[idx], THIS_MODULE, idx 465 sound/usb/card.c dev_err(&dev->dev, "cannot create card instance %d\n", idx); idx 472 sound/usb/card.c chip->index = idx; idx 475 sound/usb/card.c chip->setup = device_setup[idx]; idx 181 sound/usb/format.c int r, idx; idx 190 sound/usb/format.c for (r = 0, idx = offset + 1; r < nr_rates; r++, idx += 3) { idx 191 sound/usb/format.c unsigned int rate = combine_triple(&fmt[idx]); idx 64 sound/usb/hiface/chip.c struct usb_device *device, int idx, idx 76 sound/usb/hiface/chip.c ret = snd_card_new(&intf->dev, index[idx], id[idx], THIS_MODULE, idx 302 sound/usb/mixer.c int idx = 0, err; idx 309 sound/usb/mixer.c idx = snd_usb_ctrl_intf(chip) | (cval->head.id << 8); idx 312 sound/usb/mixer.c validx, idx, buf, val_len); idx 323 sound/usb/mixer.c request, validx, idx, cval->val_type); idx 338 sound/usb/mixer.c int idx = 0, ret, val_size, size; idx 357 sound/usb/mixer.c idx = snd_usb_ctrl_intf(chip) | (cval->head.id << 8); idx 360 sound/usb/mixer.c validx, idx, buf, size); idx 367 sound/usb/mixer.c request, validx, idx, cval->val_type); idx 452 sound/usb/mixer.c int idx = 0, val_len, err, timeout = 10; idx 482 sound/usb/mixer.c idx = snd_usb_ctrl_intf(chip) | (cval->head.id << 8); idx 486 sound/usb/mixer.c validx, idx, buf, val_len); idx 495 sound/usb/mixer.c request, validx, idx, cval->val_type, buf[0], buf[1]); idx 568 sound/usb/mixer.c int idx = ich * num_outs + och; idx 569 sound/usb/mixer.c return bmap[idx >> 3] & (0x80 >> (idx & 7)); idx 1425 sound/usb/mixer.c int idx = 0, validx, ret, val; idx 1433 sound/usb/mixer.c idx = snd_usb_ctrl_intf(chip) | (cval->head.id << 8); idx 1439 sound/usb/mixer.c validx, idx, &uac2_conn, sizeof(uac2_conn)); idx 1446 sound/usb/mixer.c validx, idx, &uac3_conn, sizeof(uac3_conn)); idx 1456 sound/usb/mixer.c UAC_GET_CUR, validx, idx, cval->val_type); idx 3582 sound/usb/mixer.c int c, err, idx; idx 3585 sound/usb/mixer.c idx = 0; idx 3590 sound/usb/mixer.c err = snd_usb_set_cur_mix_value(cval, c + 1, idx, idx 3591 sound/usb/mixer.c cval->cache_val[idx]); idx 3595 sound/usb/mixer.c idx++; idx 1909 sound/usb/mixer_quirks.c int idx; idx 1917 sound/usb/mixer_quirks.c idx = SND_RME_CLK_SYSTEM(status1); idx 1918 sound/usb/mixer_quirks.c if (idx < ARRAY_SIZE(snd_rme_rate_table)) idx 1919 sound/usb/mixer_quirks.c rate = snd_rme_rate_table[idx]; idx 1922 sound/usb/mixer_quirks.c idx = SND_RME_CLK_AES(status1); idx 1923 sound/usb/mixer_quirks.c if (idx < SND_RME_RATE_IDX_AES_SPDIF_NUM) idx 1924 sound/usb/mixer_quirks.c rate = snd_rme_rate_table[idx]; idx 1927 sound/usb/mixer_quirks.c idx = SND_RME_CLK_SPDIF(status1); idx 1928 sound/usb/mixer_quirks.c if (idx < SND_RME_RATE_IDX_AES_SPDIF_NUM) idx 1929 sound/usb/mixer_quirks.c rate = snd_rme_rate_table[idx]; idx 1942 sound/usb/mixer_quirks.c int idx = SND_RME_CLOCK_NOLOCK; idx 1951 sound/usb/mixer_quirks.c idx = SND_RME_CLOCK_SYNC; idx 1953 sound/usb/mixer_quirks.c idx = SND_RME_CLOCK_LOCK; idx 1957 sound/usb/mixer_quirks.c idx = SND_RME_CLOCK_SYNC; idx 1959 sound/usb/mixer_quirks.c idx = SND_RME_CLOCK_LOCK; idx 1964 sound/usb/mixer_quirks.c ucontrol->value.enumerated.item[0] = idx; idx 453 sound/usb/mixer_scarlett.c int idx = snd_usb_ctrl_intf(chip) | (elem->head.id << 8); idx 460 sound/usb/mixer_scarlett.c USB_DIR_IN, wValue, idx, buf, elem->channels); idx 1497 sound/usb/pcm.c int idx = (src_idx + subs->dsd_dop.byte_idx - 1) % wrap; idx 1500 sound/usb/pcm.c dst[dst_idx++] = bitrev8(src[idx]); idx 1502 sound/usb/pcm.c dst[dst_idx++] = src[idx]; idx 1628 sound/usb/pcm.c int idx = (subs->hwptr_done + i) idx 1630 sound/usb/pcm.c buf[i] = bitrev8(runtime->dma_area[idx]); idx 58 sound/usb/power.c int err, idx; idx 60 sound/usb/power.c idx = snd_usb_ctrl_intf(chip) | (pd->pd_id << 8); idx 65 sound/usb/power.c UAC3_AC_POWER_DOMAIN_CONTROL << 8, idx, idx 81 sound/usb/power.c UAC3_AC_POWER_DOMAIN_CONTROL << 8, idx, idx 872 sound/x86/intel_hdmi_audio.c int idx = intelhaddata->bd_head; idx 879 sound/x86/intel_hdmi_audio.c had_write_register(intelhaddata, AUD_BUF_ADDR(idx), addr); idx 880 sound/x86/intel_hdmi_audio.c had_write_register(intelhaddata, AUD_BUF_LEN(idx), idx 892 sound/x86/intel_hdmi_audio.c int idx) idx 894 sound/x86/intel_hdmi_audio.c had_write_register(intelhaddata, AUD_BUF_ADDR(idx), 0); idx 895 sound/x86/intel_hdmi_audio.c had_write_register(intelhaddata, AUD_BUF_LEN(idx), 0); idx 22 tools/bpf/bpftool/cfg.c int idx; idx 32 tools/bpf/bpftool/cfg.c int idx; idx 81 tools/bpf/bpftool/cfg.c new_func->idx = cfg->func_num; idx 209 tools/bpf/bpftool/cfg.c bb->idx = bb_idx++; idx 212 tools/bpf/bpftool/cfg.c last->idx = bb_idx++; idx 223 tools/bpf/bpftool/cfg.c bb->idx = ENTRY_BLOCK_INDEX; idx 228 tools/bpf/bpftool/cfg.c bb->idx = EXIT_BLOCK_INDEX; idx 387 tools/bpf/bpftool/cfg.c if (bb->idx == ENTRY_BLOCK_INDEX || bb->idx == EXIT_BLOCK_INDEX) idx 393 tools/bpf/bpftool/cfg.c func->idx, bb->idx, shape); idx 395 tools/bpf/bpftool/cfg.c if (bb->idx == ENTRY_BLOCK_INDEX) { idx 397 tools/bpf/bpftool/cfg.c } else if (bb->idx == EXIT_BLOCK_INDEX) { idx 418 tools/bpf/bpftool/cfg.c int func_idx = func->idx; idx 427 tools/bpf/bpftool/cfg.c func_idx, e->src->idx, func_idx, e->dst->idx, idx 444 tools/bpf/bpftool/cfg.c int func_idx = func->idx; idx 465 tools/bpf/bpftool/cfg.c func->idx, func->idx); idx 62 tools/bpf/bpftool/map_perf_ring.c int idx; idx 75 tools/bpf/bpftool/map_perf_ring.c int idx = ctx->all_cpus ? cpu : ctx->idx; idx 84 tools/bpf/bpftool/map_perf_ring.c jsonw_uint(json_wtr, idx); idx 104 tools/bpf/bpftool/map_perf_ring.c cpu, idx); idx 132 tools/bpf/bpftool/map_perf_ring.c .idx = -1, idx 169 tools/bpf/bpftool/map_perf_ring.c ctx.idx = strtoul(*argv, &endptr, 0); idx 185 tools/bpf/bpftool/map_perf_ring.c if (ctx.idx == -1 || ctx.cpu == -1) { idx 191 tools/bpf/bpftool/map_perf_ring.c ctx.idx = 0; idx 199 tools/bpf/bpftool/map_perf_ring.c opts.map_keys = &ctx.idx; idx 665 tools/bpf/bpftool/prog.c int idx; idx 674 tools/bpf/bpftool/prog.c return a->idx - b->idx; idx 1108 tools/bpf/bpftool/prog.c int idx, err; idx 1159 tools/bpf/bpftool/prog.c idx = strtoul(*argv, &endptr, 0); idx 1169 tools/bpf/bpftool/prog.c idx = -1; idx 1190 tools/bpf/bpftool/prog.c map_replace[old_map_fds].idx = idx; idx 1261 tools/bpf/bpftool/prog.c map_replace[j].idx = i; idx 1266 tools/bpf/bpftool/prog.c if (map_replace[j].idx == -1) { idx 1279 tools/bpf/bpftool/prog.c idx = 0; idx 1284 tools/bpf/bpftool/prog.c if (j < old_map_fds && idx == map_replace[j].idx) { idx 1292 tools/bpf/bpftool/prog.c if (j < old_map_fds && map_replace[j].idx == idx) { idx 1294 tools/bpf/bpftool/prog.c idx); idx 1299 tools/bpf/bpftool/prog.c idx++; idx 1302 tools/bpf/bpftool/prog.c p_err("map idx '%d' not used", map_replace[j].idx); idx 238 tools/include/uapi/drm/drm.h int idx; /**< Which client desired? */ idx 376 tools/include/uapi/drm/drm.h int idx; /**< Index into the master buffer list */ idx 25 tools/lib/api/fd/array.h int idx; idx 242 tools/lib/api/fs/fs.c static const char *fs__mountpoint(int idx) idx 244 tools/lib/api/fs/fs.c struct fs *fs = &fs__entries[idx]; idx 264 tools/lib/api/fs/fs.c static const char *fs__mount(int idx) idx 266 tools/lib/api/fs/fs.c struct fs *fs = &fs__entries[idx]; idx 269 tools/lib/api/fs/fs.c if (fs__mountpoint(idx)) idx 280 tools/lib/api/fs/fs.c #define FS(name, idx) \ idx 283 tools/lib/api/fs/fs.c return fs__mountpoint(idx); \ idx 288 tools/lib/api/fs/fs.c return fs__mount(idx); \ idx 385 tools/lib/bpf/btf.c int err = 0, fd = -1, idx = 0; idx 427 tools/lib/bpf/btf.c idx++; idx 430 tools/lib/bpf/btf.c idx, path); idx 436 tools/lib/bpf/btf.c idx, path); idx 443 tools/lib/bpf/btf.c idx, name, path); idx 451 tools/lib/bpf/btf.c idx, name, path); idx 152 tools/lib/bpf/libbpf.c int idx; idx 344 tools/lib/bpf/libbpf.c prog->idx = -1; idx 359 tools/lib/bpf/libbpf.c bpf_program__init(void *data, size_t size, char *section_name, int idx, idx 375 tools/lib/bpf/libbpf.c idx, section_name); idx 382 tools/lib/bpf/libbpf.c idx, section_name); idx 394 tools/lib/bpf/libbpf.c prog->idx = idx; idx 407 tools/lib/bpf/libbpf.c char *section_name, int idx) idx 412 tools/lib/bpf/libbpf.c err = bpf_program__init(data, size, section_name, idx, &prog); idx 459 tools/lib/bpf/libbpf.c if (sym.st_shndx != prog->idx) idx 474 tools/lib/bpf/libbpf.c if (!name && prog->idx == obj->efile.text_shndx) idx 678 tools/lib/bpf/libbpf.c int idx = 0; idx 685 tools/lib/bpf/libbpf.c idx++; idx 688 tools/lib/bpf/libbpf.c idx, obj->path); idx 695 tools/lib/bpf/libbpf.c idx, obj->path); idx 705 tools/lib/bpf/libbpf.c idx, name, obj->path); idx 1344 tools/lib/bpf/libbpf.c static bool section_have_execinstr(struct bpf_object *obj, int idx) idx 1349 tools/lib/bpf/libbpf.c scn = elf_getscn(obj->efile.elf, idx); idx 1526 tools/lib/bpf/libbpf.c int idx = 0, err = 0; idx 1539 tools/lib/bpf/libbpf.c idx++; idx 1542 tools/lib/bpf/libbpf.c idx, obj->path); idx 1549 tools/lib/bpf/libbpf.c idx, obj->path); idx 1556 tools/lib/bpf/libbpf.c idx, name, obj->path); idx 1560 tools/lib/bpf/libbpf.c idx, name, (unsigned long)data->d_size, idx 1577 tools/lib/bpf/libbpf.c obj->efile.maps_shndx = idx; idx 1579 tools/lib/bpf/libbpf.c obj->efile.btf_maps_shndx = idx; idx 1595 tools/lib/bpf/libbpf.c obj->efile.text_shndx = idx; idx 1597 tools/lib/bpf/libbpf.c data->d_size, name, idx); idx 1609 tools/lib/bpf/libbpf.c obj->efile.data_shndx = idx; idx 1612 tools/lib/bpf/libbpf.c obj->efile.rodata_shndx = idx; idx 1614 tools/lib/bpf/libbpf.c pr_debug("skip section(%d) %s\n", idx, name); idx 1624 tools/lib/bpf/libbpf.c name, idx, sec); idx 1642 tools/lib/bpf/libbpf.c obj->efile.bss_shndx = idx; idx 1644 tools/lib/bpf/libbpf.c pr_debug("skip section(%d) %s\n", idx, name); idx 1648 tools/lib/bpf/libbpf.c if (!obj->efile.strtabidx || obj->efile.strtabidx >= idx) { idx 1663 tools/lib/bpf/libbpf.c bpf_object__find_prog_by_idx(struct bpf_object *obj, int idx) idx 1670 tools/lib/bpf/libbpf.c if (prog->idx == idx) idx 2310 tools/lib/bpf/libbpf.c __u32 idx; /* field index or array index */ idx 2403 tools/lib/bpf/libbpf.c spec->spec[0].idx = access_idx; idx 2439 tools/lib/bpf/libbpf.c spec->spec[spec->len].idx = access_idx; idx 2453 tools/lib/bpf/libbpf.c spec->spec[spec->len].idx = access_idx; idx 2651 tools/lib/bpf/libbpf.c local_member = btf_members(local_type) + local_acc->idx; idx 2688 tools/lib/bpf/libbpf.c targ_acc->idx = i; idx 2751 tools/lib/bpf/libbpf.c if (local_acc->idx >= a->nelems) idx 2763 tools/lib/bpf/libbpf.c targ_acc->idx = local_acc->idx; idx 2766 tools/lib/bpf/libbpf.c targ_spec->raw_spec[targ_spec->raw_len] = targ_acc->idx; idx 2772 tools/lib/bpf/libbpf.c targ_spec->offset += local_acc->idx * sz; idx 2936 tools/lib/bpf/libbpf.c libbpf_print(level, "[%u]", spec->spec[i].idx); idx 3207 tools/lib/bpf/libbpf.c if (prog->idx == obj->efile.text_shndx) { idx 3342 tools/lib/bpf/libbpf.c int idx = shdr->sh_info; idx 3350 tools/lib/bpf/libbpf.c prog = bpf_object__find_prog_by_idx(obj, idx); idx 3352 tools/lib/bpf/libbpf.c pr_warning("relocation failed: no section(%d)\n", idx); idx 3540 tools/lib/bpf/libbpf.c return prog->idx == obj->efile.text_shndx && obj->has_pseudo_calls; idx 4286 tools/lib/bpf/libbpf.c ssize_t idx; idx 4301 tools/lib/bpf/libbpf.c idx = (p - obj->programs) + (forward ? 1 : -1); idx 4302 tools/lib/bpf/libbpf.c if (idx >= obj->nr_programs || idx < 0) idx 4304 tools/lib/bpf/libbpf.c return &obj->programs[idx]; idx 4728 tools/lib/bpf/libbpf.c ssize_t idx; idx 4743 tools/lib/bpf/libbpf.c idx = (m - obj->maps) + i; idx 4744 tools/lib/bpf/libbpf.c if (idx >= obj->nr_maps || idx < 0) idx 4746 tools/lib/bpf/libbpf.c return &obj->maps[idx]; idx 49 tools/lib/bpf/xsk.h __u32 idx) idx 53 tools/lib/bpf/xsk.h return &addrs[idx & fill->mask]; idx 57 tools/lib/bpf/xsk.h xsk_ring_cons__comp_addr(const struct xsk_ring_cons *comp, __u32 idx) idx 61 tools/lib/bpf/xsk.h return &addrs[idx & comp->mask]; idx 65 tools/lib/bpf/xsk.h __u32 idx) idx 69 tools/lib/bpf/xsk.h return &descs[idx & tx->mask]; idx 73 tools/lib/bpf/xsk.h xsk_ring_cons__rx_desc(const struct xsk_ring_cons *rx, __u32 idx) idx 77 tools/lib/bpf/xsk.h return &descs[idx & rx->mask]; idx 117 tools/lib/bpf/xsk.h size_t nb, __u32 *idx) idx 122 tools/lib/bpf/xsk.h *idx = prod->cached_prod; idx 139 tools/lib/bpf/xsk.h size_t nb, __u32 *idx) idx 149 tools/lib/bpf/xsk.h *idx = cons->cached_cons; idx 81 tools/lib/find_bit.c unsigned long idx; idx 83 tools/lib/find_bit.c for (idx = 0; idx * BITS_PER_LONG < size; idx++) { idx 84 tools/lib/find_bit.c if (addr[idx]) idx 85 tools/lib/find_bit.c return min(idx * BITS_PER_LONG + __ffs(addr[idx]), size); idx 98 tools/lib/find_bit.c unsigned long idx; idx 100 tools/lib/find_bit.c for (idx = 0; idx * BITS_PER_LONG < size; idx++) { idx 101 tools/lib/find_bit.c if (addr[idx] != ~0UL) idx 102 tools/lib/find_bit.c return min(idx * BITS_PER_LONG + ffz(addr[idx]), size); idx 150 tools/lib/lockdep/preload.c int idx = __locks_nr++; idx 151 tools/lib/lockdep/preload.c if (idx >= ARRAY_SIZE(__locks)) { idx 156 tools/lib/lockdep/preload.c return __locks + idx; idx 33 tools/lib/traceevent/tep_strerror.c int idx; idx 48 tools/lib/traceevent/tep_strerror.c idx = errnum - __TEP_ERRNO__START - 1; idx 49 tools/lib/traceevent/tep_strerror.c msg = tep_error_str[idx]; idx 570 tools/objtool/check.c } else if (rela->sym->sec->idx) { idx 36 tools/objtool/elf.c unsigned int idx) idx 41 tools/objtool/elf.c if (sec->idx == idx) idx 47 tools/objtool/elf.c static struct symbol *find_symbol_by_index(struct elf *elf, unsigned int idx) idx 53 tools/objtool/elf.c hash_for_each_possible(sec->symbol_hash, sym, hash, idx) idx 54 tools/objtool/elf.c if (sym->idx == idx) idx 169 tools/objtool/elf.c sec->idx = elf_ndxscn(s); idx 232 tools/objtool/elf.c sym->idx = i; idx 292 tools/objtool/elf.c hash_add(sym->sec->symbol_hash, &sym->hash, sym->idx); idx 495 tools/objtool/elf.c sec->idx = elf_ndxscn(s); idx 538 tools/objtool/elf.c s = elf_getscn(elf->elf, shstrtab->idx); idx 585 tools/objtool/elf.c sec->sh.sh_link = find_section_by_name(elf, ".symtab")->idx; idx 586 tools/objtool/elf.c sec->sh.sh_info = base->idx; idx 595 tools/objtool/elf.c int nr, idx = 0, size; idx 614 tools/objtool/elf.c idx = 0; idx 616 tools/objtool/elf.c relas[idx].r_offset = rela->offset; idx 617 tools/objtool/elf.c relas[idx].r_addend = rela->addend; idx 618 tools/objtool/elf.c relas[idx].r_info = GELF_R_INFO(rela->sym->idx, rela->type); idx 619 tools/objtool/elf.c idx++; idx 633 tools/objtool/elf.c s = elf_getscn(elf->elf, sec->idx); idx 37 tools/objtool/elf.h int idx; idx 48 tools/objtool/elf.h unsigned int idx; idx 85 tools/objtool/orc_gen.c unsigned int idx, struct section *insn_sec, idx 97 tools/objtool/orc_gen.c orc = (struct orc_entry *)u_sec->data->d_buf + idx; idx 111 tools/objtool/orc_gen.c rela->offset = idx * sizeof(int); idx 123 tools/objtool/orc_gen.c unsigned int idx; idx 138 tools/objtool/orc_gen.c idx = 0; idx 148 tools/objtool/orc_gen.c idx++; idx 155 tools/objtool/orc_gen.c idx++; idx 157 tools/objtool/orc_gen.c if (!idx) idx 162 tools/objtool/orc_gen.c sec = elf_create_section(file->elf, ".orc_unwind_ip", sizeof(int), idx); idx 172 tools/objtool/orc_gen.c sizeof(struct orc_entry), idx); idx 175 tools/objtool/orc_gen.c idx = 0; idx 185 tools/objtool/orc_gen.c if (create_orc_entry(u_sec, ip_relasec, idx, idx 190 tools/objtool/orc_gen.c idx++; idx 197 tools/objtool/orc_gen.c if (create_orc_entry(u_sec, ip_relasec, idx, idx 203 tools/objtool/orc_gen.c idx++; idx 72 tools/objtool/special.c struct section *sec, int idx, idx 78 tools/objtool/special.c offset = idx * entry->size; idx 165 tools/objtool/special.c int idx, ret; idx 182 tools/objtool/special.c for (idx = 0; idx < nr_entries; idx++) { idx 190 tools/objtool/special.c ret = get_alt_entry(elf, entry, sec, idx, alt); idx 691 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_alloc_wrapped_array(struct cs_etm_recording *ptr, int idx) idx 697 tools/perf/arch/arm/util/cs-etm.c while (cnt <= idx) idx 768 tools/perf/arch/arm/util/cs-etm.c int idx, struct auxtrace_mmap *mm, idx 781 tools/perf/arch/arm/util/cs-etm.c if (idx >= ptr->wrapped_cnt) { idx 782 tools/perf/arch/arm/util/cs-etm.c err = cs_etm_alloc_wrapped_array(ptr, idx); idx 793 tools/perf/arch/arm/util/cs-etm.c wrapped = ptr->wrapped[idx]; idx 796 tools/perf/arch/arm/util/cs-etm.c ptr->wrapped[idx] = true; idx 800 tools/perf/arch/arm/util/cs-etm.c __func__, idx, (size_t)*old, (size_t)*head, mm->len); idx 861 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_read_finish(struct auxtrace_record *itr, int idx) idx 872 tools/perf/arch/arm/util/cs-etm.c evsel, idx); idx 161 tools/perf/arch/arm64/util/arm-spe.c static int arm_spe_read_finish(struct auxtrace_record *itr, int idx) idx 172 tools/perf/arch/arm64/util/arm-spe.c evsel, idx); idx 135 tools/perf/arch/common.c int idx; idx 184 tools/perf/arch/common.c idx = lookup_triplets(path_list, name); idx 185 tools/perf/arch/common.c if (idx < 0) { idx 193 tools/perf/arch/common.c if (asprintf(&buf, "%s%s", path_list[idx], name) < 0) idx 10 tools/perf/arch/powerpc/include/dwarf-regs-table.h #define REG_DWARFNUM_NAME(reg, idx) [idx] = "%" #reg idx 5 tools/perf/arch/s390/include/dwarf-regs-table.h #define REG_DWARFNUM_NAME(reg, idx) [idx] = "%" #reg idx 40 tools/perf/arch/x86/tests/rdpmc.c u32 seq, idx, time_mult = 0, time_shift = 0; idx 57 tools/perf/arch/x86/tests/rdpmc.c idx = pc->index; idx 59 tools/perf/arch/x86/tests/rdpmc.c if (idx) idx 60 tools/perf/arch/x86/tests/rdpmc.c count += rdpmc(idx - 1); idx 74 tools/perf/arch/x86/tests/rdpmc.c if (idx) idx 274 tools/perf/arch/x86/util/intel-bts.c int idx) idx 283 tools/perf/arch/x86/util/intel-bts.c while (new_cnt <= idx) idx 359 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_find_snapshot(struct auxtrace_record *itr, int idx, idx 369 tools/perf/arch/x86/util/intel-bts.c __func__, idx, (size_t)*old, (size_t)*head); idx 371 tools/perf/arch/x86/util/intel-bts.c if (idx >= btsr->snapshot_ref_cnt) { idx 372 tools/perf/arch/x86/util/intel-bts.c err = intel_bts_alloc_snapshot_refs(btsr, idx); idx 377 tools/perf/arch/x86/util/intel-bts.c wrapped = btsr->snapshot_refs[idx].wrapped; idx 379 tools/perf/arch/x86/util/intel-bts.c btsr->snapshot_refs[idx].wrapped = true; idx 411 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_read_finish(struct auxtrace_record *itr, int idx) idx 422 tools/perf/arch/x86/util/intel-bts.c evsel, idx); idx 835 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_alloc_snapshot_refs(struct intel_pt_recording *ptr, int idx) idx 844 tools/perf/arch/x86/util/intel-pt.c while (new_cnt <= idx) idx 877 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_alloc_snapshot_ref(struct intel_pt_recording *ptr, int idx, idx 887 tools/perf/arch/x86/util/intel-pt.c ptr->snapshot_refs[idx].ref_buf = ref_buf; idx 888 tools/perf/arch/x86/util/intel-pt.c ptr->snapshot_refs[idx].ref_offset = snapshot_buf_size - ref_buf_size; idx 991 tools/perf/arch/x86/util/intel-pt.c static bool intel_pt_wrapped(struct intel_pt_recording *ptr, int idx, idx 995 tools/perf/arch/x86/util/intel-pt.c struct intel_pt_snapshot_ref *ref = &ptr->snapshot_refs[idx]; idx 1025 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_find_snapshot(struct auxtrace_record *itr, int idx, idx 1035 tools/perf/arch/x86/util/intel-pt.c __func__, idx, (size_t)*old, (size_t)*head); idx 1041 tools/perf/arch/x86/util/intel-pt.c if (idx >= ptr->snapshot_ref_cnt) { idx 1042 tools/perf/arch/x86/util/intel-pt.c err = intel_pt_alloc_snapshot_refs(ptr, idx); idx 1048 tools/perf/arch/x86/util/intel-pt.c if (!ptr->snapshot_refs[idx].ref_buf) { idx 1049 tools/perf/arch/x86/util/intel-pt.c err = intel_pt_alloc_snapshot_ref(ptr, idx, mm->len); idx 1053 tools/perf/arch/x86/util/intel-pt.c wrapped = intel_pt_wrapped(ptr, idx, mm, data, *head); idx 1055 tools/perf/arch/x86/util/intel-pt.c wrapped = ptr->snapshot_refs[idx].wrapped; idx 1057 tools/perf/arch/x86/util/intel-pt.c ptr->snapshot_refs[idx].wrapped = true; idx 1095 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_read_finish(struct auxtrace_record *itr, int idx) idx 1106 tools/perf/arch/x86/util/intel-pt.c idx); idx 2010 tools/perf/builtin-c2c.c static unsigned int idx; idx 2012 tools/perf/builtin-c2c.c c2c_he->cacheline_idx = idx++; idx 62 tools/perf/builtin-diff.c int idx; idx 70 tools/perf/builtin-diff.c int idx; idx 483 tools/perf/builtin-diff.c void *ptr = dfmt - dfmt->idx; idx 923 tools/perf/builtin-diff.c return hist_entry__cmp_compute(right, left, COMPUTE_DELTA, d->idx); idx 932 tools/perf/builtin-diff.c return hist_entry__cmp_compute(right, left, COMPUTE_DELTA_ABS, d->idx); idx 941 tools/perf/builtin-diff.c return hist_entry__cmp_compute(right, left, COMPUTE_RATIO, d->idx); idx 950 tools/perf/builtin-diff.c return hist_entry__cmp_compute(right, left, COMPUTE_WEIGHTED_DIFF, d->idx); idx 1009 tools/perf/builtin-diff.c d->idx, d->data.path, idx 1010 tools/perf/builtin-diff.c !d->idx ? "(Baseline)" : ""); idx 1466 tools/perf/builtin-diff.c hpp__entry_unpair(struct hist_entry *he, int idx, char *buf, size_t size) idx 1468 tools/perf/builtin-diff.c switch (idx) { idx 1480 tools/perf/builtin-diff.c int idx, char *buf, size_t size) idx 1486 tools/perf/builtin-diff.c switch (idx) { idx 1547 tools/perf/builtin-diff.c int idx = dfmt->idx; idx 1550 tools/perf/builtin-diff.c if (idx == PERF_HPP_DIFF__BASELINE) idx 1554 tools/perf/builtin-diff.c hpp__entry_pair(he, pair, idx, buf, size); idx 1556 tools/perf/builtin-diff.c hpp__entry_unpair(he, idx, buf, size); idx 1607 tools/perf/builtin-diff.c BUG_ON(dfmt->idx >= PERF_HPP_DIFF__MAX_INDEX); idx 1608 tools/perf/builtin-diff.c header = columns[dfmt->idx].name; idx 1609 tools/perf/builtin-diff.c width = columns[dfmt->idx].width; idx 1615 tools/perf/builtin-diff.c scnprintf(buf, MAX_HEADER_NAME, "%s/%d", header, d->idx); idx 1631 tools/perf/builtin-diff.c static void data__hpp_register(struct data__file *d, int idx) idx 1633 tools/perf/builtin-diff.c struct diff_hpp_fmt *dfmt = &d->fmt[idx]; idx 1636 tools/perf/builtin-diff.c dfmt->idx = idx; idx 1645 tools/perf/builtin-diff.c switch (idx) { idx 1805 tools/perf/builtin-diff.c d->idx = i; idx 750 tools/perf/builtin-kvm.c static s64 perf_kvm__mmap_read_idx(struct perf_kvm_stat *kvm, int idx, idx 761 tools/perf/builtin-kvm.c md = &evlist->mmap[idx]; idx 306 tools/perf/builtin-record.c int ret, idx; idx 315 tools/perf/builtin-record.c idx = record__aio_sync(map, false); idx 316 tools/perf/builtin-record.c aio.data = map->aio.data[idx]; idx 322 tools/perf/builtin-record.c ret = record__aio_write(&(map->aio.cblocks[idx]), trace_fd, aio.data, aio.size, *off); idx 317 tools/perf/builtin-report.c evsel->idx, idx 617 tools/perf/builtin-report.c if (pos->idx == 0) idx 358 tools/perf/builtin-sched.c unsigned long idx = task->nr_events; idx 362 tools/perf/builtin-sched.c event->nr = idx; idx 369 tools/perf/builtin-sched.c task->atoms[idx] = event; idx 828 tools/perf/builtin-stat.c aggr_get_id_t get_id, struct perf_cpu_map *map, int idx) idx 832 tools/perf/builtin-stat.c if (idx >= map->nr) idx 835 tools/perf/builtin-stat.c cpu = map->map[idx]; idx 838 tools/perf/builtin-stat.c config->cpus_aggr_map->map[cpu] = get_id(config, map, idx); idx 844 tools/perf/builtin-stat.c struct perf_cpu_map *map, int idx) idx 846 tools/perf/builtin-stat.c return perf_stat__get_aggr(config, perf_stat__get_socket, map, idx); idx 850 tools/perf/builtin-stat.c struct perf_cpu_map *map, int idx) idx 852 tools/perf/builtin-stat.c return perf_stat__get_aggr(config, perf_stat__get_die, map, idx); idx 856 tools/perf/builtin-stat.c struct perf_cpu_map *map, int idx) idx 858 tools/perf/builtin-stat.c return perf_stat__get_aggr(config, perf_stat__get_core, map, idx); idx 934 tools/perf/builtin-stat.c static inline int perf_env__get_cpu(struct perf_env *env, struct perf_cpu_map *map, int idx) idx 938 tools/perf/builtin-stat.c if (idx > map->nr) idx 941 tools/perf/builtin-stat.c cpu = map->map[idx]; idx 949 tools/perf/builtin-stat.c static int perf_env__get_socket(struct perf_cpu_map *map, int idx, void *data) idx 952 tools/perf/builtin-stat.c int cpu = perf_env__get_cpu(env, map, idx); idx 957 tools/perf/builtin-stat.c static int perf_env__get_die(struct perf_cpu_map *map, int idx, void *data) idx 960 tools/perf/builtin-stat.c int die_id = -1, cpu = perf_env__get_cpu(env, map, idx); idx 981 tools/perf/builtin-stat.c static int perf_env__get_core(struct perf_cpu_map *map, int idx, void *data) idx 984 tools/perf/builtin-stat.c int core = -1, cpu = perf_env__get_cpu(env, map, idx); idx 1030 tools/perf/builtin-stat.c struct perf_cpu_map *map, int idx) idx 1032 tools/perf/builtin-stat.c return perf_env__get_socket(map, idx, &perf_stat.session->header.env); idx 1035 tools/perf/builtin-stat.c struct perf_cpu_map *map, int idx) idx 1037 tools/perf/builtin-stat.c return perf_env__get_die(map, idx, &perf_stat.session->header.env); idx 1041 tools/perf/builtin-stat.c struct perf_cpu_map *map, int idx) idx 1043 tools/perf/builtin-stat.c return perf_env__get_core(map, idx, &perf_stat.session->header.env); idx 263 tools/perf/builtin-top.c symbol__annotate_zero_histogram(symbol, top->sym_evsel->idx); idx 265 tools/perf/builtin-top.c symbol__annotate_decay_histogram(symbol, top->sym_evsel->idx); idx 530 tools/perf/builtin-top.c fprintf(stderr, "\n\t%d %s", top->sym_evsel->idx, perf_evsel__name(top->sym_evsel)); idx 541 tools/perf/builtin-top.c if (top->sym_evsel->idx == counter) idx 864 tools/perf/builtin-top.c static void perf_top__mmap_read_idx(struct perf_top *top, int idx) idx 871 tools/perf/builtin-top.c md = opts->overwrite ? &evlist->overwrite_mmap[idx] : &evlist->mmap[idx]; idx 397 tools/perf/builtin-trace.c int idx = val - sa->offset; idx 399 tools/perf/builtin-trace.c if (idx < 0 || idx >= sa->nr_entries || sa->entries[idx] == NULL) { idx 406 tools/perf/builtin-trace.c return scnprintf(bf, size, "%s%s", show_prefix ? sa->prefix : "", sa->entries[idx]); idx 436 tools/perf/builtin-trace.c int idx = val - sa->offset; idx 438 tools/perf/builtin-trace.c if (idx >= 0 && idx < sa->nr_entries) { idx 439 tools/perf/builtin-trace.c if (sa->entries[idx] == NULL) idx 441 tools/perf/builtin-trace.c return scnprintf(bf, size, "%s%s", show_prefix ? sa->prefix : "", sa->entries[idx]); idx 1438 tools/perf/builtin-trace.c int idx; idx 1447 tools/perf/builtin-trace.c for (idx = 0; idx < nr_args; ++idx) { idx 1449 tools/perf/builtin-trace.c sc->arg_fmt[idx] = sc->fmt->arg[idx]; idx 1459 tools/perf/builtin-trace.c int idx = 0, len; idx 1461 tools/perf/builtin-trace.c for (field = sc->args; field; field = field->next, ++idx) { idx 1464 tools/perf/builtin-trace.c if (sc->fmt && sc->fmt->arg[idx].scnprintf) idx 1472 tools/perf/builtin-trace.c sc->arg_fmt[idx].scnprintf = SCA_FILENAME; idx 1474 tools/perf/builtin-trace.c sc->arg_fmt[idx].scnprintf = SCA_PTR; idx 1476 tools/perf/builtin-trace.c sc->arg_fmt[idx].scnprintf = SCA_PID; idx 1478 tools/perf/builtin-trace.c sc->arg_fmt[idx].scnprintf = SCA_MODE_T; idx 1490 tools/perf/builtin-trace.c sc->arg_fmt[idx].scnprintf = SCA_FD; idx 1660 tools/perf/builtin-trace.c unsigned long syscall_arg__val(struct syscall_arg *arg, u8 idx) idx 1663 tools/perf/builtin-trace.c unsigned char *p = arg->args + sizeof(unsigned long) * idx; idx 1672 tools/perf/builtin-trace.c if (sc->arg_fmt && sc->arg_fmt[arg->idx].name) idx 1673 tools/perf/builtin-trace.c return scnprintf(bf, size, "%s: ", sc->arg_fmt[arg->idx].name); idx 1675 tools/perf/builtin-trace.c return scnprintf(bf, size, "arg%d: ", arg->idx); idx 1685 tools/perf/builtin-trace.c if (sc->arg_fmt && sc->arg_fmt[arg->idx].mask_val) idx 1686 tools/perf/builtin-trace.c return sc->arg_fmt[arg->idx].mask_val(arg, val); idx 1694 tools/perf/builtin-trace.c if (sc->arg_fmt && sc->arg_fmt[arg->idx].scnprintf) { idx 1696 tools/perf/builtin-trace.c if (sc->arg_fmt[arg->idx].parm) idx 1697 tools/perf/builtin-trace.c arg->parm = sc->arg_fmt[arg->idx].parm; idx 1698 tools/perf/builtin-trace.c return sc->arg_fmt[arg->idx].scnprintf(bf, size, arg); idx 1716 tools/perf/builtin-trace.c .idx = 0, idx 1735 tools/perf/builtin-trace.c field = field->next, ++arg.idx, bit <<= 1) { idx 1739 tools/perf/builtin-trace.c val = syscall_arg__val(&arg, arg.idx); idx 1754 tools/perf/builtin-trace.c (sc->arg_fmt[arg.idx].show_zero || idx 1755 tools/perf/builtin-trace.c sc->arg_fmt[arg.idx].scnprintf == SCA_STRARRAY || idx 1756 tools/perf/builtin-trace.c sc->arg_fmt[arg.idx].scnprintf == SCA_STRARRAYS) && idx 1757 tools/perf/builtin-trace.c sc->arg_fmt[arg.idx].parm)) idx 1773 tools/perf/builtin-trace.c while (arg.idx < sc->nr_args) { idx 1776 tools/perf/builtin-trace.c val = syscall_arg__val(&arg, arg.idx); idx 1782 tools/perf/builtin-trace.c ++arg.idx; idx 2635 tools/perf/builtin-trace.c struct parse_events_error err = { .idx = 0, }; idx 3913 tools/perf/builtin-trace.c int len = strlen(str) + 1, err = -1, list, idx; idx 3932 tools/perf/builtin-trace.c syscalltbl__strglobmatch_first(trace->sctbl, s, &idx) >= 0) { idx 234 tools/perf/lib/cpumap.c int perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx) idx 236 tools/perf/lib/cpumap.c if (idx < cpus->nr) idx 237 tools/perf/lib/cpumap.c return cpus->map[idx]; idx 29 tools/perf/lib/include/internal/evsel.h int idx; idx 16 tools/perf/lib/include/perf/cpumap.h LIBPERF_API int perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx); idx 21 tools/perf/lib/include/perf/cpumap.h #define perf_cpu_map__for_each_cpu(cpu, idx, cpus) \ idx 22 tools/perf/lib/include/perf/cpumap.h for ((idx) = 0, (cpu) = perf_cpu_map__cpu(cpus, idx); \ idx 23 tools/perf/lib/include/perf/cpumap.h (idx) < perf_cpu_map__nr(cpus); \ idx 24 tools/perf/lib/include/perf/cpumap.h (idx)++, (cpu) = perf_cpu_map__cpu(cpus, idx)) idx 199 tools/perf/lib/include/perf/event.h __u64 idx; idx 222 tools/perf/lib/include/perf/event.h __u32 idx; idx 224 tools/perf/tests/bpf.c static int __test__bpf(int idx) idx 232 tools/perf/tests/bpf.c bpf_testcase_table[idx].prog_id, idx 236 tools/perf/tests/bpf.c bpf_testcase_table[idx].msg_compile_fail); idx 237 tools/perf/tests/bpf.c if (idx == 0) idx 244 tools/perf/tests/bpf.c bpf_testcase_table[idx].name); idx 245 tools/perf/tests/bpf.c if ((!!bpf_testcase_table[idx].target_func) != (!!obj)) { idx 248 tools/perf/tests/bpf.c bpf_testcase_table[idx].msg_load_fail); idx 251 tools/perf/tests/bpf.c bpf_testcase_table[idx].msg_load_fail); idx 258 tools/perf/tests/bpf.c bpf_testcase_table[idx].target_func, idx 259 tools/perf/tests/bpf.c bpf_testcase_table[idx].expect_result); idx 262 tools/perf/tests/bpf.c if (bpf_testcase_table[idx].pin) { idx 77 tools/perf/tests/dwarf-unwind.c int idx = callchain_param.order == ORDER_CALLER ? idx 93 tools/perf/tests/dwarf-unwind.c symbol, entry->ip, funcs[idx]); idx 94 tools/perf/tests/dwarf-unwind.c return strcmp((const char *) symbol, funcs[idx]); idx 13 tools/perf/tests/evsel-roundtrip-name.c int type, op, err = 0, ret = 0, i, idx; idx 36 tools/perf/tests/evsel-roundtrip-name.c idx = 0; idx 48 tools/perf/tests/evsel-roundtrip-name.c if (evsel->idx != idx) idx 51 tools/perf/tests/evsel-roundtrip-name.c ++idx; idx 87 tools/perf/tests/evsel-roundtrip-name.c if (strcmp(perf_evsel__name(evsel), names[evsel->idx])) { idx 89 tools/perf/tests/evsel-roundtrip-name.c pr_debug("%s != %s\n", perf_evsel__name(evsel), names[evsel->idx]); idx 189 tools/perf/tests/hists_link.c static int __validate_link(struct hists *hists, int idx) idx 217 tools/perf/tests/hists_link.c !find_sample(fake_samples[idx], idx 218 tools/perf/tests/hists_link.c ARRAY_SIZE(fake_samples[idx]), idx 223 tools/perf/tests/hists_link.c } else if (idx) { idx 235 tools/perf/tests/hists_link.c if (idx == 0) { idx 58 tools/perf/tests/llvm.c enum test_llvm__testcase idx, idx 68 tools/perf/tests/llvm.c if (idx >= __LLVM_TESTCASE_MAX) idx 71 tools/perf/tests/llvm.c source = bpf_source_table[idx].source; idx 72 tools/perf/tests/llvm.c desc = bpf_source_table[idx].desc; idx 74 tools/perf/tests/llvm.c *should_load_fail = bpf_source_table[idx].should_load_fail; idx 141 tools/perf/tests/mmap-basic.c nr_events[evsel->idx]++; idx 149 tools/perf/tests/mmap-basic.c if (nr_events[evsel->idx] != expected_nr_events[evsel->idx]) { idx 151 tools/perf/tests/mmap-basic.c expected_nr_events[evsel->idx], idx 152 tools/perf/tests/mmap-basic.c perf_evsel__name(evsel), nr_events[evsel->idx]); idx 1771 tools/perf/tests/parse-events.c struct parse_events_error err = { .idx = 0, }; idx 104 tools/perf/trace/beauty/beauty.h u8 idx; idx 109 tools/perf/trace/beauty/beauty.h unsigned long syscall_arg__val(struct syscall_arg *arg, u8 idx); idx 25 tools/perf/trace/beauty/kcmp.c pid = syscall_arg__val(arg, arg->idx == 3 ? 0 : 1); /* idx1 -> pid1, idx2 -> pid2 */ idx 85 tools/perf/trace/beauty/open_flags.c arg->mask |= 1 << (arg->idx + 1); /* Mask the mode parm */ idx 629 tools/perf/ui/browser.c unsigned int row = 0, idx = browser->top_idx; idx 636 tools/perf/ui/browser.c while (idx < browser->nr_entries && idx 646 tools/perf/ui/browser.c ++idx; idx 197 tools/perf/ui/browsers/annotate.c from = (u64)cursor->al.idx; idx 198 tools/perf/ui/browsers/annotate.c to = (u64)target->idx; idx 266 tools/perf/ui/browsers/annotate.c struct annotation_line *pos, u32 idx) idx 273 tools/perf/ui/browsers/annotate.c browser->b.top_idx = browser->b.index = idx; idx 294 tools/perf/ui/browsers/annotate.c u32 idx = pos->idx; idx 297 tools/perf/ui/browsers/annotate.c idx = pos->idx_asm; idx 298 tools/perf/ui/browsers/annotate.c annotate_browser__set_top(browser, pos, idx); idx 357 tools/perf/ui/browsers/annotate.c offset = al->idx; idx 362 tools/perf/ui/browsers/annotate.c browser->b.top_idx = al->idx - offset; idx 363 tools/perf/ui/browsers/annotate.c browser->b.index = al->idx; idx 442 tools/perf/ui/browsers/annotate.c s64 offset, s64 *idx) idx 447 tools/perf/ui/browsers/annotate.c *idx = 0; idx 452 tools/perf/ui/browsers/annotate.c ++*idx; idx 464 tools/perf/ui/browsers/annotate.c s64 idx; idx 475 tools/perf/ui/browsers/annotate.c dl = annotate_browser__find_offset(browser, offset, &idx); idx 481 tools/perf/ui/browsers/annotate.c annotate_browser__set_top(browser, &dl->al, idx); idx 488 tools/perf/ui/browsers/annotate.c char *s, s64 *idx) idx 493 tools/perf/ui/browsers/annotate.c *idx = browser->b.index; idx 498 tools/perf/ui/browsers/annotate.c ++*idx; idx 510 tools/perf/ui/browsers/annotate.c s64 idx; idx 512 tools/perf/ui/browsers/annotate.c al = annotate_browser__find_string(browser, browser->search_bf, &idx); idx 518 tools/perf/ui/browsers/annotate.c annotate_browser__set_top(browser, al, idx); idx 525 tools/perf/ui/browsers/annotate.c char *s, s64 *idx) idx 530 tools/perf/ui/browsers/annotate.c *idx = browser->b.index; idx 535 tools/perf/ui/browsers/annotate.c --*idx; idx 547 tools/perf/ui/browsers/annotate.c s64 idx; idx 549 tools/perf/ui/browsers/annotate.c al = annotate_browser__find_string_reverse(browser, browser->search_bf, &idx); idx 555 tools/perf/ui/browsers/annotate.c annotate_browser__set_top(browser, al, idx); idx 706 tools/perf/ui/browsers/annotate.c symbol__annotate_decay_histogram(sym, evsel->idx); idx 65 tools/perf/ui/browsers/map.c u32 *idx = symbol__browser_index(sym); idx 68 tools/perf/ui/browsers/map.c browser->b.index = browser->b.top_idx = *idx; idx 126 tools/perf/ui/browsers/map.c u32 *idx = symbol__browser_index(pos); idx 127 tools/perf/ui/browsers/map.c *idx = mb.b.nr_entries; idx 137 tools/perf/ui/gtk/annotate.c evsel->idx + i); idx 142 tools/perf/ui/gtk/annotate.c evsel->idx); idx 395 tools/perf/ui/hist.c return a->idx == b->idx; idx 408 tools/perf/ui/hist.c .idx = PERF_HPP__ ## _idx, \ idx 422 tools/perf/ui/hist.c .idx = PERF_HPP__ ## _idx, \ idx 435 tools/perf/ui/hist.c .idx = PERF_HPP__ ## _idx, \ idx 702 tools/perf/ui/hist.c BUG_ON(fmt->idx >= PERF_HPP__MAX_INDEX); idx 704 tools/perf/ui/hist.c switch (fmt->idx) { idx 946 tools/perf/util/annotate.c return (src) ? __symbol__inc_addr_samples(sym, map, src, evsel->idx, idx 2059 tools/perf/util/annotate.c sym_hist = annotation__histogram(notes, evsel->idx); idx 2243 tools/perf/util/annotate.c struct sym_hist *h = annotation__histogram(notes, evsel->idx); idx 2275 tools/perf/util/annotate.c struct sym_hist *h = annotation__histogram(notes, evsel->idx); idx 2576 tools/perf/util/annotate.c al->idx = notes->nr_entries++; idx 143 tools/perf/util/annotate.h u32 idx; idx 322 tools/perf/util/annotate.h static inline struct sym_hist *annotated_source__histogram(struct annotated_source *src, int idx) idx 324 tools/perf/util/annotate.h return ((void *)src->histograms) + (src->sizeof_sym_hist * idx); idx 327 tools/perf/util/annotate.h static inline struct sym_hist *annotation__histogram(struct annotation *notes, int idx) idx 329 tools/perf/util/annotate.h return annotated_source__histogram(notes->src, idx); idx 277 tools/perf/util/arm-spe-pkt-decoder.c int ret, ns, el, idx = packet->index; idx 333 tools/perf/util/arm-spe-pkt-decoder.c if (idx > 1) { idx 356 tools/perf/util/arm-spe-pkt-decoder.c switch (idx) { idx 421 tools/perf/util/arm-spe-pkt-decoder.c switch (idx) { idx 427 tools/perf/util/arm-spe-pkt-decoder.c (idx == 1) ? "TGT" : "PC", payload, el, ns); idx 437 tools/perf/util/arm-spe-pkt-decoder.c (unsigned long)payload, idx + 1); idx 445 tools/perf/util/arm-spe-pkt-decoder.c switch (idx) { idx 78 tools/perf/util/auxtrace.c mm->idx = mp->idx; idx 130 tools/perf/util/auxtrace.c struct evlist *evlist, int idx, idx 133 tools/perf/util/auxtrace.c mp->idx = idx; idx 136 tools/perf/util/auxtrace.c mp->cpu = evlist->core.cpus->map[idx]; idx 143 tools/perf/util/auxtrace.c mp->tid = perf_thread_map__pid(evlist->core.threads, idx); idx 237 tools/perf/util/auxtrace.c unsigned int idx, idx 243 tools/perf/util/auxtrace.c if (idx >= queues->nr_queues) { idx 244 tools/perf/util/auxtrace.c err = auxtrace_queues__grow(queues, idx + 1); idx 249 tools/perf/util/auxtrace.c queue = &queues->queue_array[idx]; idx 275 tools/perf/util/auxtrace.c unsigned int idx, idx 289 tools/perf/util/auxtrace.c err = auxtrace_queues__queue_buffer(queues, idx, b); idx 314 tools/perf/util/auxtrace.c unsigned int idx, idx 337 tools/perf/util/auxtrace.c err = auxtrace_queues__split_buffer(queues, idx, buffer); idx 342 tools/perf/util/auxtrace.c err = auxtrace_queues__queue_buffer(queues, idx, buffer); idx 371 tools/perf/util/auxtrace.c unsigned int idx = event->auxtrace.idx; idx 373 tools/perf/util/auxtrace.c return auxtrace_queues__add_buffer(queues, session, idx, &buffer, idx 552 tools/perf/util/auxtrace.c int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx, idx 557 tools/perf/util/auxtrace.c return itr->find_snapshot(itr, idx, mm, data, head, old); idx 950 tools/perf/util/auxtrace.c event->auxtrace.reference, event->auxtrace.idx, idx 1245 tools/perf/util/auxtrace.c if (auxtrace_record__find_snapshot(itr, mm->idx, mm, data, idx 1256 tools/perf/util/auxtrace.c mm->idx, old, head, head - old); idx 1319 tools/perf/util/auxtrace.c ev.auxtrace.idx = mm->idx; idx 1333 tools/perf/util/auxtrace.c err = itr->read_finish(itr, mm->idx); idx 1588 tools/perf/util/auxtrace.c static int parse_sym_idx(char **inp, int *idx) idx 1590 tools/perf/util/auxtrace.c *idx = -1; idx 1601 tools/perf/util/auxtrace.c *idx = 0; idx 1613 tools/perf/util/auxtrace.c *idx = num; idx 1619 tools/perf/util/auxtrace.c static int parse_addr_size(char **inp, u64 *num, const char **str, int *idx) idx 1624 tools/perf/util/auxtrace.c err = parse_sym_idx(inp, idx); idx 1708 tools/perf/util/auxtrace.c int idx; idx 1723 tools/perf/util/auxtrace.c (args->selected && ++(args->cnt) == args->idx) || idx 1765 tools/perf/util/auxtrace.c static int sym_not_found_error(const char *sym_name, int idx) idx 1767 tools/perf/util/auxtrace.c if (idx > 0) { idx 1769 tools/perf/util/auxtrace.c idx, sym_name); idx 1770 tools/perf/util/auxtrace.c } else if (!idx) { idx 1780 tools/perf/util/auxtrace.c static int find_kern_sym(const char *sym_name, u64 *start, u64 *size, int idx) idx 1784 tools/perf/util/auxtrace.c .idx = idx, idx 1785 tools/perf/util/auxtrace.c .global = !idx, idx 1786 tools/perf/util/auxtrace.c .selected = idx > 0, idx 1811 tools/perf/util/auxtrace.c return sym_not_found_error(sym_name, idx); idx 1940 tools/perf/util/auxtrace.c int idx) idx 1944 tools/perf/util/auxtrace.c ((!idx && sym->binding == STB_GLOBAL) || idx 1945 tools/perf/util/auxtrace.c (idx > 0 && ++*cnt == idx) || idx 1946 tools/perf/util/auxtrace.c idx < 0); idx 1979 tools/perf/util/auxtrace.c u64 *size, int idx) idx 1992 tools/perf/util/auxtrace.c if (idx > 0) { idx 1995 tools/perf/util/auxtrace.c } else if (dso_sym_match(sym, sym_name, &cnt, idx)) { idx 1999 tools/perf/util/auxtrace.c } else if (dso_sym_match(sym, sym_name, &cnt, idx)) { idx 2007 tools/perf/util/auxtrace.c return sym_not_found_error(sym_name, idx); idx 277 tools/perf/util/auxtrace.h int idx; idx 298 tools/perf/util/auxtrace.h int idx; idx 330 tools/perf/util/auxtrace.h int (*find_snapshot)(struct auxtrace_record *itr, int idx, idx 337 tools/perf/util/auxtrace.h int (*read_finish)(struct auxtrace_record *itr, int idx); idx 444 tools/perf/util/auxtrace.h struct evlist *evlist, int idx, idx 512 tools/perf/util/auxtrace.h int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx, idx 717 tools/perf/util/auxtrace.h struct evlist *evlist, int idx, idx 1162 tools/perf/util/bpf-loader.c unsigned int idx = start + length - 1; idx 1164 tools/perf/util/bpf-loader.c if (idx >= def->max_entries) { idx 1165 tools/perf/util/bpf-loader.c pr_debug("ERROR: index %d too large\n", idx); idx 1295 tools/perf/util/bpf-loader.c unsigned int idx = start + j; idx 1297 tools/perf/util/bpf-loader.c err = func(name, map_fd, pdef, op, &idx, arg); idx 1300 tools/perf/util/bpf-loader.c name, idx); idx 109 tools/perf/util/branch.c static int count_str_scnprintf(int idx, const char *str, char *bf, int size) idx 111 tools/perf/util/branch.c return scnprintf(bf, size, "%s%s", (idx) ? " " : " (", str); idx 1302 tools/perf/util/callchain.c static int count_pri64_printf(int idx, const char *str, u64 value, char *bf, int bfsize) idx 1306 tools/perf/util/callchain.c printed = scnprintf(bf, bfsize, "%s%s:%" PRId64 "", (idx) ? " " : " (", str, value); idx 1311 tools/perf/util/callchain.c static int count_float_printf(int idx, const char *str, float value, idx 1319 tools/perf/util/callchain.c printed = scnprintf(bf, bfsize, "%s%s:%.1f%%", (idx) ? " " : " (", str, value); idx 114 tools/perf/util/cpumap.c int cpu_map__get_socket(struct perf_cpu_map *map, int idx, void *data __maybe_unused) idx 118 tools/perf/util/cpumap.c if (idx > map->nr) idx 121 tools/perf/util/cpumap.c cpu = map->map[idx]; idx 170 tools/perf/util/cpumap.c int cpu_map__get_die(struct perf_cpu_map *map, int idx, void *data) idx 174 tools/perf/util/cpumap.c if (idx > map->nr) idx 177 tools/perf/util/cpumap.c cpu = map->map[idx]; idx 184 tools/perf/util/cpumap.c s = cpu_map__get_socket(map, idx, data); idx 209 tools/perf/util/cpumap.c int cpu_map__get_core(struct perf_cpu_map *map, int idx, void *data) idx 213 tools/perf/util/cpumap.c if (idx > map->nr) idx 216 tools/perf/util/cpumap.c cpu = map->map[idx]; idx 221 tools/perf/util/cpumap.c s_die = cpu_map__get_die(map, idx, data); idx 465 tools/perf/util/cpumap.c int cpu_map__cpu(struct perf_cpu_map *cpus, int idx) idx 467 tools/perf/util/cpumap.c return cpus->map[idx]; idx 18 tools/perf/util/cpumap.h int cpu_map__get_socket(struct perf_cpu_map *map, int idx, void *data); idx 20 tools/perf/util/cpumap.h int cpu_map__get_die(struct perf_cpu_map *map, int idx, void *data); idx 22 tools/perf/util/cpumap.h int cpu_map__get_core(struct perf_cpu_map *map, int idx, void *data); idx 61 tools/perf/util/cpumap.h int cpu_map__cpu(struct perf_cpu_map *cpus, int idx); idx 218 tools/perf/util/cs-etm.c int idx; idx 224 tools/perf/util/cs-etm.c idx = (int)(intptr_t)inode->priv; idx 225 tools/perf/util/cs-etm.c tidq = etmq->traceid_queues[idx]; idx 284 tools/perf/util/cs-etm.c int idx; idx 301 tools/perf/util/cs-etm.c idx = (int)(intptr_t)inode->priv; idx 302 tools/perf/util/cs-etm.c return etmq->traceid_queues[idx]; idx 313 tools/perf/util/cs-etm.c idx = intlist__nr_entries(traceid_queues_list); idx 320 tools/perf/util/cs-etm.c inode->priv = (void *)(intptr_t)idx; idx 328 tools/perf/util/cs-etm.c idx + 1, idx 338 tools/perf/util/cs-etm.c traceid_queues[idx] = tidq; idx 341 tools/perf/util/cs-etm.c return etmq->traceid_queues[idx]; idx 380 tools/perf/util/cs-etm.c struct cs_etm_auxtrace *etm, int idx, idx 385 tools/perf/util/cs-etm.c t_params[idx].protocol = cs_etm__get_v7_protocol_version(etmidr); idx 386 tools/perf/util/cs-etm.c t_params[idx].etmv3.reg_ctrl = metadata[idx][CS_ETM_ETMCR]; idx 387 tools/perf/util/cs-etm.c t_params[idx].etmv3.reg_trc_id = metadata[idx][CS_ETM_ETMTRACEIDR]; idx 391 tools/perf/util/cs-etm.c struct cs_etm_auxtrace *etm, int idx) idx 395 tools/perf/util/cs-etm.c t_params[idx].protocol = CS_ETM_PROTO_ETMV4i; idx 396 tools/perf/util/cs-etm.c t_params[idx].etmv4.reg_idr0 = metadata[idx][CS_ETMV4_TRCIDR0]; idx 397 tools/perf/util/cs-etm.c t_params[idx].etmv4.reg_idr1 = metadata[idx][CS_ETMV4_TRCIDR1]; idx 398 tools/perf/util/cs-etm.c t_params[idx].etmv4.reg_idr2 = metadata[idx][CS_ETMV4_TRCIDR2]; idx 399 tools/perf/util/cs-etm.c t_params[idx].etmv4.reg_idr8 = metadata[idx][CS_ETMV4_TRCIDR8]; idx 400 tools/perf/util/cs-etm.c t_params[idx].etmv4.reg_configr = metadata[idx][CS_ETMV4_TRCCONFIGR]; idx 401 tools/perf/util/cs-etm.c t_params[idx].etmv4.reg_traceidr = metadata[idx][CS_ETMV4_TRCTRACEIDR]; idx 530 tools/perf/util/cs-etm.c int idx; idx 538 tools/perf/util/cs-etm.c idx = priv; idx 541 tools/perf/util/cs-etm.c tidq = etmq->traceid_queues[idx]; idx 2003 tools/perf/util/cs-etm.c int idx; idx 2009 tools/perf/util/cs-etm.c idx = (int)(intptr_t)inode->priv; idx 2010 tools/perf/util/cs-etm.c tidq = etmq->traceid_queues[idx]; idx 2409 tools/perf/util/cs-etm.c int err = 0, idx = -1; idx 2478 tools/perf/util/cs-etm.c idx = metadata[j][CS_ETM_ETMTRACEIDR]; idx 2491 tools/perf/util/cs-etm.c idx = metadata[j][CS_ETMV4_TRCTRACEIDR]; idx 2496 tools/perf/util/cs-etm.c inode = intlist__findnew(traceid_list, idx); idx 1332 tools/perf/util/dso.c int idx, errnum = dso->load_errno; idx 1358 tools/perf/util/dso.c idx = errnum - __DSO_LOAD_ERRNO__START; idx 1359 tools/perf/util/dso.c scnprintf(buf, buflen, "%s", dso_load__error_str[idx]); idx 392 tools/perf/util/dwarf-aux.c Dwarf_Sword idx; idx 394 tools/perf/util/dwarf-aux.c if (die_get_attr_sdata(in_die, DW_AT_call_file, &idx) == 0) idx 395 tools/perf/util/dwarf-aux.c return (int)idx; idx 403 tools/perf/util/dwarf-aux.c Dwarf_Sword idx; idx 405 tools/perf/util/dwarf-aux.c if (die_get_attr_sdata(pdie, DW_AT_decl_file, &idx) == 0) idx 406 tools/perf/util/dwarf-aux.c return (int)idx; idx 422 tools/perf/util/dwarf-aux.c int idx; idx 424 tools/perf/util/dwarf-aux.c idx = die_get_call_fileno(in_die); idx 425 tools/perf/util/dwarf-aux.c if (idx < 0 || !dwarf_diecu(in_die, &cu_die, NULL, NULL) || idx 429 tools/perf/util/dwarf-aux.c return dwarf_filesrc(files, idx, NULL, NULL); idx 1212 tools/perf/util/dwarf-aux.c Dwarf_Addr addr, unsigned long *idx) idx 1222 tools/perf/util/dwarf-aux.c *idx = i; idx 160 tools/perf/util/event.c u32 nr_namespaces, idx; idx 170 tools/perf/util/event.c for (idx = 0; idx < nr_namespaces; idx++) { idx 171 tools/perf/util/event.c if (idx && (idx % 4 == 0)) idx 174 tools/perf/util/event.c ret += fprintf(fp, "%u/%s: %" PRIu64 "/%#" PRIx64 "%s", idx, idx 175 tools/perf/util/event.c perf_ns__name(idx), (u64)ns_link_info[idx].dev, idx 176 tools/perf/util/event.c (u64)ns_link_info[idx].ino, idx 177 tools/perf/util/event.c ((idx + 1) != nr_namespaces) ? ", " : "]\n"); idx 163 tools/perf/util/evlist.c entry->idx = evlist->core.nr_entries; idx 164 tools/perf/util/evlist.c entry->tracking = !entry->idx; idx 196 tools/perf/util/evlist.c leader->core.nr_members = evsel->idx - leader->idx + 1; idx 391 tools/perf/util/evlist.c struct evsel *evsel, int idx) idx 396 tools/perf/util/evlist.c return perf_evlist__enable_event_cpu(evlist, evsel, idx); idx 398 tools/perf/util/evlist.c return perf_evlist__enable_event_thread(evlist, evsel, idx); idx 427 tools/perf/util/evlist.c struct evsel *evsel, int idx, int cpu, idx 431 tools/perf/util/evlist.c sid->idx = idx; idx 633 tools/perf/util/evlist.c static int evlist__mmap_per_evsel(struct evlist *evlist, int idx, idx 675 tools/perf/util/evlist.c if (perf_mmap__mmap(&maps[idx], mp, *output, evlist_cpu) < 0) idx 681 tools/perf/util/evlist.c perf_mmap__get(&maps[idx]); idx 694 tools/perf/util/evlist.c perf_evlist__add_pollfd(&evlist->core, fd, &maps[idx], revent) < 0) { idx 695 tools/perf/util/evlist.c perf_mmap__put(&maps[idx]); idx 703 tools/perf/util/evlist.c perf_evlist__set_sid_idx(evlist, evsel, idx, cpu, idx 156 tools/perf/util/evlist.h void evlist__mmap_consume(struct evlist *evlist, int idx); idx 201 tools/perf/util/evlist.h struct evsel *evsel, int idx); idx 131 tools/perf/util/evsel.c int idx = 0; idx 140 tools/perf/util/evsel.c idx += 1; idx 143 tools/perf/util/evsel.c idx += 1; idx 146 tools/perf/util/evsel.c idx += 1; idx 149 tools/perf/util/evsel.c idx += 1; idx 151 tools/perf/util/evsel.c return idx; idx 164 tools/perf/util/evsel.c int idx = 1; idx 173 tools/perf/util/evsel.c idx += 1; idx 176 tools/perf/util/evsel.c idx += 1; idx 178 tools/perf/util/evsel.c return idx; idx 238 tools/perf/util/evsel.c struct perf_event_attr *attr, int idx) idx 241 tools/perf/util/evsel.c evsel->idx = idx; idx 242 tools/perf/util/evsel.c evsel->tracking = !idx; idx 262 tools/perf/util/evsel.c struct evsel *perf_evsel__new_idx(struct perf_event_attr *attr, int idx) idx 268 tools/perf/util/evsel.c evsel__init(evsel, attr, idx); idx 337 tools/perf/util/evsel.c struct evsel *perf_evsel__newtp_idx(const char *sys, const char *name, int idx) idx 363 tools/perf/util/evsel.c evsel__init(evsel, &attr, idx); idx 49 tools/perf/util/evsel.h int idx; idx 147 tools/perf/util/evsel.h struct evsel *perf_evsel__new_idx(struct perf_event_attr *attr, int idx); idx 154 tools/perf/util/evsel.h struct evsel *perf_evsel__newtp_idx(const char *sys, const char *name, int idx); idx 168 tools/perf/util/evsel.h void evsel__init(struct evsel *evsel, struct perf_event_attr *attr, int idx); idx 365 tools/perf/util/evsel.h return evsel->idx - evsel->leader->idx; idx 174 tools/perf/util/expr.y int idx; idx 176 tools/perf/util/expr.y idx = ctx->num_ids++; idx 177 tools/perf/util/expr.y ctx->ids[idx].name = name; idx 178 tools/perf/util/expr.y ctx->ids[idx].val = val; idx 265 tools/perf/util/genelf_debug.c static void emit_set_file(struct buffer_ext *be, unsigned long idx) idx 267 tools/perf/util/genelf_debug.c emit_opcode_unsigned(be, DW_LNS_set_file, idx); idx 789 tools/perf/util/header.c u32 leader_idx = evsel->idx; idx 1203 tools/perf/util/header.c static int memory_node__read(struct memory_node *n, unsigned long idx) idx 1218 tools/perf/util/header.c sysfs__mountpoint(), idx); idx 1238 tools/perf/util/header.c n->node = idx; idx 1278 tools/perf/util/header.c unsigned int idx; idx 1285 tools/perf/util/header.c r = sscanf(ent->d_name, "node%u", &idx); idx 1295 tools/perf/util/header.c ret = memory_node__read(&nodes[cnt++], idx); idx 1623 tools/perf/util/header.c evsel->idx = i; idx 2100 tools/perf/util/header.c perf_evlist__find_by_index(struct evlist *evlist, int idx) idx 2105 tools/perf/util/header.c if (evsel->idx == idx) idx 2121 tools/perf/util/header.c evsel = perf_evlist__find_by_index(evlist, event->idx); idx 2467 tools/perf/util/header.c if (evsel->idx == (int) desc[i].leader_idx) { idx 389 tools/perf/util/hist.c struct hist_entry *hists__get_entry(struct hists *hists, int idx) idx 397 tools/perf/util/hist.c if (i == idx) idx 187 tools/perf/util/hist.h struct hist_entry *hists__get_entry(struct hists *hists, int idx); idx 280 tools/perf/util/hist.h int idx; idx 135 tools/perf/util/intlist.c struct int_node *intlist__entry(const struct intlist *ilist, unsigned int idx) idx 140 tools/perf/util/intlist.c rb_node = rblist__entry(&ilist->rblist, idx); idx 26 tools/perf/util/intlist.h struct int_node *intlist__entry(const struct intlist *ilist, unsigned int idx); idx 159 tools/perf/util/mmap.c int idx __maybe_unused, idx 171 tools/perf/util/mmap.c static int perf_mmap__aio_alloc(struct mmap *map, int idx) idx 173 tools/perf/util/mmap.c map->aio.data[idx] = mmap(NULL, perf_mmap__mmap_len(map), PROT_READ|PROT_WRITE, idx 175 tools/perf/util/mmap.c if (map->aio.data[idx] == MAP_FAILED) { idx 176 tools/perf/util/mmap.c map->aio.data[idx] = NULL; idx 183 tools/perf/util/mmap.c static void perf_mmap__aio_free(struct mmap *map, int idx) idx 185 tools/perf/util/mmap.c if (map->aio.data[idx]) { idx 186 tools/perf/util/mmap.c munmap(map->aio.data[idx], perf_mmap__mmap_len(map)); idx 187 tools/perf/util/mmap.c map->aio.data[idx] = NULL; idx 191 tools/perf/util/mmap.c static int perf_mmap__aio_bind(struct mmap *map, int idx, int cpu, int affinity) idx 198 tools/perf/util/mmap.c data = map->aio.data[idx]; idx 211 tools/perf/util/mmap.c static int perf_mmap__aio_alloc(struct mmap *map, int idx) idx 213 tools/perf/util/mmap.c map->aio.data[idx] = malloc(perf_mmap__mmap_len(map)); idx 214 tools/perf/util/mmap.c if (map->aio.data[idx] == NULL) idx 220 tools/perf/util/mmap.c static void perf_mmap__aio_free(struct mmap *map, int idx) idx 222 tools/perf/util/mmap.c zfree(&(map->aio.data[idx])); idx 225 tools/perf/util/mmap.c static int perf_mmap__aio_bind(struct mmap *map __maybe_unused, int idx __maybe_unused, idx 320 tools/perf/util/parse-events.c __add_event(struct list_head *list, int *idx, idx 332 tools/perf/util/parse-events.c evsel = perf_evsel__new_idx(attr, *idx); idx 336 tools/perf/util/parse-events.c (*idx)++; idx 352 tools/perf/util/parse-events.c static int add_event(struct list_head *list, int *idx, idx 356 tools/perf/util/parse-events.c return __add_event(list, idx, attr, name, NULL, config_terms, false, NULL) ? 0 : -ENOMEM; idx 359 tools/perf/util/parse-events.c static int add_event_tool(struct list_head *list, int *idx, idx 368 tools/perf/util/parse-events.c evsel = __add_event(list, idx, &attr, NULL, NULL, NULL, false, "0"); idx 406 tools/perf/util/parse-events.c int parse_events_add_cache(struct list_head *list, int *idx, idx 477 tools/perf/util/parse-events.c return add_event(list, idx, &attr, config_name ? : name, &config_terms); idx 510 tools/perf/util/parse-events.c static int add_tracepoint(struct list_head *list, int *idx, idx 517 tools/perf/util/parse-events.c evsel = perf_evsel__newtp_idx(sys_name, evt_name, (*idx)++); idx 535 tools/perf/util/parse-events.c static int add_tracepoint_multi_event(struct list_head *list, int *idx, idx 569 tools/perf/util/parse-events.c ret = add_tracepoint(list, idx, sys_name, evt_ent->d_name, idx 583 tools/perf/util/parse-events.c static int add_tracepoint_event(struct list_head *list, int *idx, idx 589 tools/perf/util/parse-events.c add_tracepoint_multi_event(list, idx, sys_name, evt_name, idx 591 tools/perf/util/parse-events.c add_tracepoint(list, idx, sys_name, evt_name, idx 595 tools/perf/util/parse-events.c static int add_tracepoint_multi_sys(struct list_head *list, int *idx, idx 621 tools/perf/util/parse-events.c ret = add_tracepoint_event(list, idx, events_ent->d_name, idx 657 tools/perf/util/parse-events.c err = parse_events_add_tracepoint(&new_evsels, &parse_state->idx, group, idx 756 tools/perf/util/parse-events.c parse_state->error->idx = term->err_term; idx 775 tools/perf/util/parse-events.c parse_state->error->idx = term->err_val; idx 777 tools/perf/util/parse-events.c parse_state->error->idx = term->err_term + error_pos; idx 900 tools/perf/util/parse-events.c int parse_events_add_breakpoint(struct list_head *list, int *idx, idx 924 tools/perf/util/parse-events.c return add_event(list, idx, &attr, NULL, NULL); idx 935 tools/perf/util/parse-events.c err->idx = term->err_val; idx 1041 tools/perf/util/parse-events.c err->idx = term->err_val; idx 1049 tools/perf/util/parse-events.c err->idx = term->err_val; idx 1084 tools/perf/util/parse-events.c err->idx = term->err_val; idx 1093 tools/perf/util/parse-events.c err->idx = term->err_term; idx 1145 tools/perf/util/parse-events.c err->idx = term->err_term; idx 1245 tools/perf/util/parse-events.c int parse_events_add_tracepoint(struct list_head *list, int *idx, idx 1259 tools/perf/util/parse-events.c return add_tracepoint_multi_sys(list, idx, sys, event, idx 1262 tools/perf/util/parse-events.c return add_tracepoint_event(list, idx, sys, event, idx 1287 tools/perf/util/parse-events.c return add_event(list, &parse_state->idx, &attr, idx 1295 tools/perf/util/parse-events.c return add_event_tool(list, &parse_state->idx, tool_event); idx 1344 tools/perf/util/parse-events.c evsel = __add_event(list, &parse_state->idx, &attr, NULL, pmu, NULL, idx 1378 tools/perf/util/parse-events.c evsel = __add_event(list, &parse_state->idx, &attr, idx 1480 tools/perf/util/parse-events.c total_members = evsel->idx - leader->idx + 1; idx 1929 tools/perf/util/parse-events.c .idx = evlist->core.nr_entries, idx 1981 tools/perf/util/parse-events.c int idx = 0; idx 2006 tools/perf/util/parse-events.c if (err->idx > max_err_idx) idx 2007 tools/perf/util/parse-events.c cut = err->idx - max_err_idx; idx 2020 tools/perf/util/parse-events.c idx = len_str + err->idx - cut; idx 2024 tools/perf/util/parse-events.c if (idx) { idx 2025 tools/perf/util/parse-events.c fprintf(stderr, "%*s\\___ %s\n", idx + 1, "", err->str); idx 2039 tools/perf/util/parse-events.c struct parse_events_error err = { .idx = 0, }; idx 2730 tools/perf/util/parse-events.c char *config, unsigned idx) idx 2739 tools/perf/util/parse-events.c BUG_ON(idx >= PERF_COUNT_HW_MAX); idx 2740 tools/perf/util/parse-events.c sym = &event_symbols_hw[idx]; idx 2810 tools/perf/util/parse-events.c int idx, const char *str) idx 2816 tools/perf/util/parse-events.c err->idx = idx; idx 113 tools/perf/util/parse-events.h int idx; /* index in the parsed string */ idx 120 tools/perf/util/parse-events.h int idx; idx 137 tools/perf/util/parse-events.h char *config, unsigned idx); idx 146 tools/perf/util/parse-events.h int parse_events_add_tracepoint(struct list_head *list, int *idx, idx 169 tools/perf/util/parse-events.h int parse_events_add_cache(struct list_head *list, int *idx, idx 173 tools/perf/util/parse-events.h int parse_events_add_breakpoint(struct list_head *list, int *idx, idx 195 tools/perf/util/parse-events.h int idx, const char *str); idx 239 tools/perf/util/parse-events.y error->idx = @1.first_column; idx 343 tools/perf/util/parse-events.y ABORT_ON(parse_events_add_cache(list, &parse_state->idx, $1, $3, $5, error, $6)); idx 355 tools/perf/util/parse-events.y ABORT_ON(parse_events_add_cache(list, &parse_state->idx, $1, $3, NULL, error, $4)); idx 367 tools/perf/util/parse-events.y ABORT_ON(parse_events_add_cache(list, &parse_state->idx, $1, NULL, NULL, error, $2)); idx 379 tools/perf/util/parse-events.y ABORT_ON(parse_events_add_breakpoint(list, &parse_state->idx, idx 390 tools/perf/util/parse-events.y ABORT_ON(parse_events_add_breakpoint(list, &parse_state->idx, idx 401 tools/perf/util/parse-events.y ABORT_ON(parse_events_add_breakpoint(list, &parse_state->idx, idx 412 tools/perf/util/parse-events.y ABORT_ON(parse_events_add_breakpoint(list, &parse_state->idx, idx 426 tools/perf/util/parse-events.y error->idx = @1.first_column; idx 428 tools/perf/util/parse-events.y if (parse_events_add_tracepoint(list, &parse_state->idx, $1.sys, $1.event, idx 32 tools/perf/util/path.c static int idx; idx 34 tools/perf/util/path.c return pathname_array[3 & ++idx]; idx 29 tools/perf/util/perf_regs.c int i, idx = 0; idx 40 tools/perf/util/perf_regs.c idx++; idx 44 tools/perf/util/perf_regs.c regs->cache_regs[id] = regs->regs[idx]; idx 1047 tools/perf/util/pmu.c err->idx = term->err_term; idx 1077 tools/perf/util/pmu.c err->idx = term->err_val; idx 1091 tools/perf/util/pmu.c err->idx = term->err_val; idx 1105 tools/perf/util/pmu.c err->idx = term->err_val; idx 135 tools/perf/util/probe-file.c int ret, idx, fddup; idx 161 tools/perf/util/probe-file.c idx = strlen(p) - 1; idx 162 tools/perf/util/probe-file.c if (p[idx] == '\n') idx 163 tools/perf/util/probe-file.c p[idx] = '\0'; idx 1020 tools/perf/util/probe-file.c #define DEFINE_TYPE(idx, pat) \ idx 1021 tools/perf/util/probe-file.c [idx] = {.pattern = pat, .avail = false} idx 751 tools/perf/util/python.c int idx = 0; idx 764 tools/perf/util/python.c &attr.bp_addr, &attr.bp_len, &idx)) idx 796 tools/perf/util/python.c evsel__init(&pevsel->evsel, &attr, idx); idx 987 tools/perf/util/python.c evsel->idx = evlist->core.nr_entries; idx 129 tools/perf/util/rblist.c struct rb_node *rblist__entry(const struct rblist *rblist, unsigned int idx) idx 135 tools/perf/util/rblist.c if (!idx--) idx 38 tools/perf/util/rblist.h struct rb_node *rblist__entry(const struct rblist *rblist, unsigned int idx); idx 310 tools/perf/util/s390-cpumsf.c int idx) idx 317 tools/perf/util/s390-cpumsf.c memcpy(&ts, &te->timestamp[idx], sizeof(ts)); idx 815 tools/perf/util/session.c event->auxtrace.idx = bswap_32(event->auxtrace.idx); idx 2419 tools/perf/util/session.c fprintf(stdout, " idx: %"PRI_lu64, e->idx); idx 2427 tools/perf/util/session.c sid->idx = e->idx; idx 2856 tools/perf/util/sort.c void perf_hpp__set_elide(int idx, bool elide) idx 2866 tools/perf/util/sort.c if (hse->se->se_width_idx == idx) { idx 2884 tools/perf/util/sort.c static bool get_elide(int idx, FILE *output) idx 2886 tools/perf/util/sort.c switch (idx) { idx 2900 tools/perf/util/sort.c switch (idx) { idx 286 tools/perf/util/sort.h void perf_hpp__set_elide(int idx, bool elide); idx 1053 tools/perf/util/stat-display.c int idx, indent = 0; idx 1061 tools/perf/util/stat-display.c for (idx = 0; idx < config->run_count; idx++) { idx 1062 tools/perf/util/stat-display.c double run = (double) config->walltime_run[idx] / NSEC_PER_SEC; idx 198 tools/perf/util/strlist.c struct str_node *strlist__entry(const struct strlist *slist, unsigned int idx) idx 203 tools/perf/util/strlist.c rb_node = rblist__entry(&slist->rblist, idx); idx 39 tools/perf/util/strlist.h struct str_node *strlist__entry(const struct strlist *slist, unsigned int idx); idx 109 tools/perf/util/symbol-elf.c #define elf_symtab__for_each_symbol(syms, nr_syms, idx, sym) \ idx 110 tools/perf/util/symbol-elf.c for (idx = 0, gelf_getsym(syms, idx, &sym);\ idx 111 tools/perf/util/symbol-elf.c idx < nr_syms; \ idx 112 tools/perf/util/symbol-elf.c idx++, gelf_getsym(syms, idx, &sym)) idx 208 tools/perf/util/symbol-elf.c GElf_Shdr *shp, const char *name, size_t *idx) idx 223 tools/perf/util/symbol-elf.c if (idx) idx 224 tools/perf/util/symbol-elf.c *idx = cnt; idx 264 tools/perf/util/symbol-elf.c #define elf_section__for_each_rel(reldata, pos, pos_mem, idx, nr_entries) \ idx 265 tools/perf/util/symbol-elf.c for (idx = 0, pos = gelf_getrel(reldata, 0, &pos_mem); \ idx 266 tools/perf/util/symbol-elf.c idx < nr_entries; \ idx 267 tools/perf/util/symbol-elf.c ++idx, pos = gelf_getrel(reldata, idx, &pos_mem)) idx 269 tools/perf/util/symbol-elf.c #define elf_section__for_each_rela(reldata, pos, pos_mem, idx, nr_entries) \ idx 270 tools/perf/util/symbol-elf.c for (idx = 0, pos = gelf_getrela(reldata, 0, &pos_mem); \ idx 271 tools/perf/util/symbol-elf.c idx < nr_entries; \ idx 272 tools/perf/util/symbol-elf.c ++idx, pos = gelf_getrela(reldata, idx, &pos_mem)) idx 283 tools/perf/util/symbol-elf.c uint32_t nr_rel_entries, idx; idx 383 tools/perf/util/symbol-elf.c elf_section__for_each_rela(reldata, pos, pos_mem, idx, idx 409 tools/perf/util/symbol-elf.c elf_section__for_each_rel(reldata, pos, pos_mem, idx, idx 959 tools/perf/util/symbol-elf.c uint32_t idx; idx 1038 tools/perf/util/symbol-elf.c elf_symtab__for_each_symbol(syms, nr_syms, idx, sym) { idx 1066 tools/perf/util/symbol-elf.c elf_symtab__for_each_symbol(syms, nr_syms, idx, sym) { idx 1407 tools/perf/util/symbol-elf.c static int kcore__add_phdr(struct kcore *kcore, int idx, off_t offset, idx 1421 tools/perf/util/symbol-elf.c if (!gelf_update_phdr(kcore->elf, idx, &phdr)) idx 1867 tools/perf/util/symbol-elf.c int idx = 0, err = -1; idx 1905 tools/perf/util/symbol-elf.c if (kcore__add_phdr(&extract, idx++, offs, p->addr, p->len)) idx 1954 tools/perf/util/symbol-elf.c int idx = 0, err = -1; idx 1967 tools/perf/util/symbol-elf.c if (kcore__add_phdr(&extract, idx, offset, kce->addr, kce->len)) idx 38 tools/perf/util/symbol.h GElf_Shdr *shp, const char *name, size_t *idx); idx 210 tools/perf/util/synthetic-events.c u32 idx; idx 227 tools/perf/util/synthetic-events.c for (idx = 0; idx < event->namespaces.nr_namespaces; idx++) idx 228 tools/perf/util/synthetic-events.c perf_event__get_ns_link_info(pid, perf_ns__name(idx), idx 229 tools/perf/util/synthetic-events.c &ns_link_info[idx]); idx 1453 tools/perf/util/synthetic-events.c e->idx = sid->idx; idx 118 tools/perf/util/syscalltbl.c int syscalltbl__strglobmatch_next(struct syscalltbl *tbl, const char *syscall_glob, int *idx) idx 123 tools/perf/util/syscalltbl.c for (i = *idx + 1; i < tbl->syscalls.nr_entries; ++i) { idx 125 tools/perf/util/syscalltbl.c *idx = i; idx 133 tools/perf/util/syscalltbl.c int syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) idx 135 tools/perf/util/syscalltbl.c *idx = -1; idx 136 tools/perf/util/syscalltbl.c return syscalltbl__strglobmatch_next(tbl, syscall_glob, idx); idx 167 tools/perf/util/syscalltbl.c const char *syscall_glob __maybe_unused, int *idx __maybe_unused) idx 172 tools/perf/util/syscalltbl.c int syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx) idx 174 tools/perf/util/syscalltbl.c return syscalltbl__strglobmatch_next(tbl, syscall_glob, idx); idx 22 tools/perf/util/syscalltbl.h int syscalltbl__strglobmatch_first(struct syscalltbl *tbl, const char *syscall_glob, int *idx); idx 23 tools/perf/util/syscalltbl.h int syscalltbl__strglobmatch_next(struct syscalltbl *tbl, const char *syscall_glob, int *idx); idx 119 tools/perf/util/target.c int idx; idx 132 tools/perf/util/target.c idx = errnum - __TARGET_ERRNO__START; idx 133 tools/perf/util/target.c msg = target__error_str[idx]; idx 276 tools/perf/util/thread-stack.c struct thread_stack *ts, size_t idx, idx 288 tools/perf/util/thread-stack.c tse = &ts->stack[idx]; idx 310 tools/perf/util/thread-stack.c parent_db_id = idx ? &(tse - 1)->db_id : NULL; idx 409 tools/perf/util/thread_map.c int thread_map__remove(struct perf_thread_map *threads, int idx) idx 416 tools/perf/util/thread_map.c if (idx >= threads->nr) idx 422 tools/perf/util/thread_map.c zfree(&threads->map[idx].comm); idx 424 tools/perf/util/thread_map.c for (i = idx; i < threads->nr - 1; i++) idx 30 tools/perf/util/thread_map.h int thread_map__remove(struct perf_thread_map *threads, int idx); idx 77 tools/perf/util/unwind-libdw.c struct unwind_entry *e = &ui->entries[ui->idx++]; idx 243 tools/perf/util/unwind-libdw.c for (i = 0; i < ui->idx && !err; i++) { idx 247 tools/perf/util/unwind-libdw.c j = ui->idx - i - 1; idx 22 tools/perf/util/unwind-libdw.h int idx; idx 104 tools/power/cpupower/utils/helpers/helpers.h extern int read_msr(int cpu, unsigned int idx, unsigned long long *val); idx 105 tools/power/cpupower/utils/helpers/helpers.h extern int write_msr(int cpu, unsigned int idx, unsigned long long val); idx 148 tools/power/cpupower/utils/helpers/helpers.h static inline int read_msr(int cpu, unsigned int idx, unsigned long long *val) idx 150 tools/power/cpupower/utils/helpers/helpers.h static inline int write_msr(int cpu, unsigned int idx, unsigned long long val) idx 27 tools/power/cpupower/utils/helpers/msr.c int read_msr(int cpu, unsigned int idx, unsigned long long *val) idx 36 tools/power/cpupower/utils/helpers/msr.c if (lseek(fd, idx, SEEK_CUR) == -1) idx 56 tools/power/cpupower/utils/helpers/msr.c int write_msr(int cpu, unsigned int idx, unsigned long long val) idx 65 tools/power/cpupower/utils/helpers/msr.c if (lseek(fd, idx, SEEK_CUR) == -1) idx 232 tools/testing/nvdimm/test/nfit.c int idx) idx 235 tools/testing/nvdimm/test/nfit.c struct nfit_test_fw *fw = &t->fw[idx]; idx 238 tools/testing/nvdimm/test/nfit.c __func__, t, nd_cmd, buf_len, idx); idx 258 tools/testing/nvdimm/test/nfit.c int idx) idx 261 tools/testing/nvdimm/test/nfit.c struct nfit_test_fw *fw = &t->fw[idx]; idx 264 tools/testing/nvdimm/test/nfit.c __func__, t, nd_cmd, buf_len, idx); idx 288 tools/testing/nvdimm/test/nfit.c int idx) idx 291 tools/testing/nvdimm/test/nfit.c struct nfit_test_fw *fw = &t->fw[idx]; idx 295 tools/testing/nvdimm/test/nfit.c __func__, t, nd_cmd, buf_len, idx); idx 339 tools/testing/nvdimm/test/nfit.c unsigned int buf_len, int idx) idx 342 tools/testing/nvdimm/test/nfit.c struct nfit_test_fw *fw = &t->fw[idx]; idx 345 tools/testing/nvdimm/test/nfit.c __func__, t, nd_cmd, buf_len, idx); idx 390 tools/testing/nvdimm/test/nfit.c unsigned int buf_len, int idx) idx 393 tools/testing/nvdimm/test/nfit.c struct nfit_test_fw *fw = &t->fw[idx]; idx 396 tools/testing/nvdimm/test/nfit.c __func__, t, nd_cmd, buf_len, idx); idx 17 tools/testing/radix-tree/main.c long idx; idx 22 tools/testing/radix-tree/main.c for (idx = -down; idx < up; idx++) idx 23 tools/testing/radix-tree/main.c item_insert(&tree, middle + idx); idx 26 tools/testing/radix-tree/main.c for (idx = -down; idx < up; idx++) idx 27 tools/testing/radix-tree/main.c item_check_present(&tree, middle + idx); idx 126 tools/testing/radix-tree/main.c void check_copied_tags(struct radix_tree_root *tree, unsigned long start, unsigned long end, unsigned long *idx, int count, int fromtag, int totag) idx 133 tools/testing/radix-tree/main.c if (idx[i] < start || idx[i] > end) { idx 134 tools/testing/radix-tree/main.c if (item_tag_get(tree, idx[i], totag)) { idx 136 tools/testing/radix-tree/main.c end, idx[i], item_tag_get(tree, idx[i], idx 138 tools/testing/radix-tree/main.c item_tag_get(tree, idx[i], totag)); idx 140 tools/testing/radix-tree/main.c assert(!item_tag_get(tree, idx[i], totag)); idx 143 tools/testing/radix-tree/main.c if (item_tag_get(tree, idx[i], fromtag) ^ idx 144 tools/testing/radix-tree/main.c item_tag_get(tree, idx[i], totag)) { idx 146 tools/testing/radix-tree/main.c idx[i], item_tag_get(tree, idx[i], fromtag), idx 147 tools/testing/radix-tree/main.c item_tag_get(tree, idx[i], totag)); idx 149 tools/testing/radix-tree/main.c assert(!(item_tag_get(tree, idx[i], fromtag) ^ idx 150 tools/testing/radix-tree/main.c item_tag_get(tree, idx[i], totag))); idx 159 tools/testing/radix-tree/main.c unsigned long idx[ITEMS]; idx 203 tools/testing/radix-tree/main.c idx[i] = rand(); idx 204 tools/testing/radix-tree/main.c } while (item_lookup(&tree, idx[i])); idx 206 tools/testing/radix-tree/main.c item_insert(&tree, idx[i]); idx 208 tools/testing/radix-tree/main.c item_tag_set(&tree, idx[i], 0); idx 209 tools/testing/radix-tree/main.c if (idx[i] >= start && idx[i] <= end) idx 221 tools/testing/radix-tree/main.c check_copied_tags(&tree, start, end, idx, ITEMS, 0, 1); idx 231 tools/testing/radix-tree/main.c check_copied_tags(&tree, start, end, idx, ITEMS, 0, 2); idx 331 tools/testing/selftests/bpf/progs/strobemeta.h size_t idx, void *tls_base, idx 335 tools/testing/selftests/bpf/progs/strobemeta.h void *location = calc_location(&cfg->int_locs[idx], tls_base); idx 340 tools/testing/selftests/bpf/progs/strobemeta.h data->int_vals[idx] = value->val; idx 342 tools/testing/selftests/bpf/progs/strobemeta.h data->int_vals_set_mask |= (1 << idx); idx 346 tools/testing/selftests/bpf/progs/strobemeta.h size_t idx, void *tls_base, idx 354 tools/testing/selftests/bpf/progs/strobemeta.h data->str_lens[idx] = 0; idx 355 tools/testing/selftests/bpf/progs/strobemeta.h location = calc_location(&cfg->str_locs[idx], tls_base); idx 371 tools/testing/selftests/bpf/progs/strobemeta.h data->str_lens[idx] = len; idx 376 tools/testing/selftests/bpf/progs/strobemeta.h size_t idx, void *tls_base, idx 381 tools/testing/selftests/bpf/progs/strobemeta.h struct strobe_map_descr* descr = &data->map_descrs[idx]; idx 390 tools/testing/selftests/bpf/progs/strobemeta.h location = calc_location(&cfg->map_locs[idx], tls_base); idx 400 tools/testing/selftests/bpf/progs/strobemeta.h if (cfg->req_meta_idx == idx) { idx 3586 tools/testing/selftests/bpf/test_btf.c int idx = GET_NAME_NTH_IDX(raw_types[i]); idx 3588 tools/testing/selftests/bpf/test_btf.c if (CHECK(idx <= 0 || idx > strs_cnt, idx 3590 tools/testing/selftests/bpf/test_btf.c idx, strs_cnt)) { idx 3594 tools/testing/selftests/bpf/test_btf.c ret_types[i] = strs_idx[idx-1] - str; idx 422 tools/testing/selftests/bpf/test_verifier.c static int create_prog_dummy2(enum bpf_prog_type prog_type, int mfd, int idx) idx 425 tools/testing/selftests/bpf/test_verifier.c BPF_MOV64_IMM(BPF_REG_3, idx), idx 35 tools/testing/selftests/kvm/include/sparsebit.h bool sparsebit_is_set(struct sparsebit *sbit, sparsebit_idx_t idx); idx 37 tools/testing/selftests/kvm/include/sparsebit.h sparsebit_idx_t idx, sparsebit_num_t num); idx 38 tools/testing/selftests/kvm/include/sparsebit.h bool sparsebit_is_clear(struct sparsebit *sbit, sparsebit_idx_t idx); idx 40 tools/testing/selftests/kvm/include/sparsebit.h sparsebit_idx_t idx, sparsebit_num_t num); idx 55 tools/testing/selftests/kvm/include/sparsebit.h void sparsebit_set(struct sparsebit *sbitp, sparsebit_idx_t idx); idx 60 tools/testing/selftests/kvm/include/sparsebit.h void sparsebit_clear(struct sparsebit *sbitp, sparsebit_idx_t idx); idx 73 tools/testing/selftests/kvm/lib/s390x/processor.c int ri, idx; idx 96 tools/testing/selftests/kvm/lib/s390x/processor.c idx = (gva >> (64 - 11 * ri)) & 0x7ffu; idx 97 tools/testing/selftests/kvm/lib/s390x/processor.c if (entry[idx] & REGION_ENTRY_INVALID) idx 98 tools/testing/selftests/kvm/lib/s390x/processor.c entry[idx] = virt_alloc_region(vm, ri, memslot); idx 99 tools/testing/selftests/kvm/lib/s390x/processor.c entry = addr_gpa2hva(vm, entry[idx] & REGION_ENTRY_ORIGIN); idx 103 tools/testing/selftests/kvm/lib/s390x/processor.c idx = (gva >> 12) & 0x0ffu; /* page index */ idx 104 tools/testing/selftests/kvm/lib/s390x/processor.c if (!(entry[idx] & PAGE_INVALID)) idx 107 tools/testing/selftests/kvm/lib/s390x/processor.c entry[idx] = gpa; idx 132 tools/testing/selftests/kvm/lib/s390x/processor.c int ri, idx; idx 140 tools/testing/selftests/kvm/lib/s390x/processor.c idx = (gva >> (64 - 11 * ri)) & 0x7ffu; idx 141 tools/testing/selftests/kvm/lib/s390x/processor.c TEST_ASSERT(!(entry[idx] & REGION_ENTRY_INVALID), idx 144 tools/testing/selftests/kvm/lib/s390x/processor.c entry = addr_gpa2hva(vm, entry[idx] & REGION_ENTRY_ORIGIN); idx 147 tools/testing/selftests/kvm/lib/s390x/processor.c idx = (gva >> 12) & 0x0ffu; /* page index */ idx 149 tools/testing/selftests/kvm/lib/s390x/processor.c TEST_ASSERT(!(entry[idx] & PAGE_INVALID), idx 152 tools/testing/selftests/kvm/lib/s390x/processor.c return (entry[idx] & ~0xffful) + (gva & 0xffful); idx 172 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx; /* index of least-significant bit in mask */ idx 287 tools/testing/selftests/kvm/lib/sparsebit.c root->idx = subtree->idx; idx 310 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_find(struct sparsebit *s, sparsebit_idx_t idx) idx 316 tools/testing/selftests/kvm/lib/sparsebit.c nodep = nodep->idx > idx ? nodep->left : nodep->right) { idx 317 tools/testing/selftests/kvm/lib/sparsebit.c if (idx >= nodep->idx && idx 318 tools/testing/selftests/kvm/lib/sparsebit.c idx <= nodep->idx + MASK_BITS + nodep->num_after - 1) idx 333 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_add(struct sparsebit *s, sparsebit_idx_t idx) idx 344 tools/testing/selftests/kvm/lib/sparsebit.c nodep->idx = idx & -MASK_BITS; idx 358 tools/testing/selftests/kvm/lib/sparsebit.c if (idx < parentp->idx) { idx 366 tools/testing/selftests/kvm/lib/sparsebit.c assert(idx > parentp->idx + MASK_BITS + parentp->num_after - 1); idx 382 tools/testing/selftests/kvm/lib/sparsebit.c while (prev && prev->idx + MASK_BITS + prev->num_after - 1 >= nodep->idx) { idx 383 tools/testing/selftests/kvm/lib/sparsebit.c unsigned int n1 = (prev->idx + MASK_BITS + prev->num_after - 1) idx 384 tools/testing/selftests/kvm/lib/sparsebit.c - nodep->idx; idx 498 tools/testing/selftests/kvm/lib/sparsebit.c static struct node *node_split(struct sparsebit *s, sparsebit_idx_t idx) idx 504 tools/testing/selftests/kvm/lib/sparsebit.c assert(!(idx % MASK_BITS)); idx 510 tools/testing/selftests/kvm/lib/sparsebit.c nodep1 = node_find(s, idx); idx 512 tools/testing/selftests/kvm/lib/sparsebit.c return node_add(s, idx); idx 518 tools/testing/selftests/kvm/lib/sparsebit.c if (nodep1->idx == idx) idx 530 tools/testing/selftests/kvm/lib/sparsebit.c offset = idx - (nodep1->idx + MASK_BITS); idx 538 tools/testing/selftests/kvm/lib/sparsebit.c nodep2 = node_add(s, idx); idx 650 tools/testing/selftests/kvm/lib/sparsebit.c assert(nodep->idx + MASK_BITS > nodep->idx); idx 652 tools/testing/selftests/kvm/lib/sparsebit.c nodep->idx += MASK_BITS; idx 687 tools/testing/selftests/kvm/lib/sparsebit.c prev->idx + MASK_BITS == nodep->idx) { idx 701 tools/testing/selftests/kvm/lib/sparsebit.c prev_highest_bit = prev->idx + MASK_BITS - 1 + prev->num_after; idx 702 tools/testing/selftests/kvm/lib/sparsebit.c if (prev_highest_bit + 1 == nodep->idx && idx 759 tools/testing/selftests/kvm/lib/sparsebit.c if (next->idx == nodep->idx + MASK_BITS + nodep->num_after && idx 779 tools/testing/selftests/kvm/lib/sparsebit.c bool sparsebit_is_set(struct sparsebit *s, sparsebit_idx_t idx) idx 785 tools/testing/selftests/kvm/lib/sparsebit.c nodep = nodep->idx > idx ? nodep->left : nodep->right) idx 786 tools/testing/selftests/kvm/lib/sparsebit.c if (idx >= nodep->idx && idx 787 tools/testing/selftests/kvm/lib/sparsebit.c idx <= nodep->idx + MASK_BITS + nodep->num_after - 1) idx 794 tools/testing/selftests/kvm/lib/sparsebit.c if (nodep->num_after && idx >= nodep->idx + MASK_BITS) idx 798 tools/testing/selftests/kvm/lib/sparsebit.c assert(idx >= nodep->idx && idx - nodep->idx < MASK_BITS); idx 799 tools/testing/selftests/kvm/lib/sparsebit.c return !!(nodep->mask & (1 << (idx - nodep->idx))); idx 805 tools/testing/selftests/kvm/lib/sparsebit.c static void bit_set(struct sparsebit *s, sparsebit_idx_t idx) idx 810 tools/testing/selftests/kvm/lib/sparsebit.c if (sparsebit_is_set(s, idx)) idx 818 tools/testing/selftests/kvm/lib/sparsebit.c nodep = node_split(s, idx & -MASK_BITS); idx 821 tools/testing/selftests/kvm/lib/sparsebit.c assert(idx >= nodep->idx && idx <= nodep->idx + MASK_BITS - 1); idx 822 tools/testing/selftests/kvm/lib/sparsebit.c assert(!(nodep->mask & (1 << (idx - nodep->idx)))); idx 823 tools/testing/selftests/kvm/lib/sparsebit.c nodep->mask |= 1 << (idx - nodep->idx); idx 832 tools/testing/selftests/kvm/lib/sparsebit.c static void bit_clear(struct sparsebit *s, sparsebit_idx_t idx) idx 837 tools/testing/selftests/kvm/lib/sparsebit.c if (!sparsebit_is_set(s, idx)) idx 841 tools/testing/selftests/kvm/lib/sparsebit.c nodep = node_find(s, idx); idx 849 tools/testing/selftests/kvm/lib/sparsebit.c if (idx >= nodep->idx + MASK_BITS) idx 850 tools/testing/selftests/kvm/lib/sparsebit.c nodep = node_split(s, idx & -MASK_BITS); idx 856 tools/testing/selftests/kvm/lib/sparsebit.c assert(idx >= nodep->idx && idx <= nodep->idx + MASK_BITS - 1); idx 857 tools/testing/selftests/kvm/lib/sparsebit.c assert(nodep->mask & (1 << (idx - nodep->idx))); idx 858 tools/testing/selftests/kvm/lib/sparsebit.c nodep->mask &= ~(1 << (idx - nodep->idx)); idx 890 tools/testing/selftests/kvm/lib/sparsebit.c indent, "", nodep->idx, nodep->mask, nodep->num_after); idx 906 tools/testing/selftests/kvm/lib/sparsebit.c return nodep->idx + n1; idx 914 tools/testing/selftests/kvm/lib/sparsebit.c return nodep->idx + n1; idx 986 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx, sparsebit_num_t num) idx 991 tools/testing/selftests/kvm/lib/sparsebit.c assert(idx + num - 1 >= idx); idx 994 tools/testing/selftests/kvm/lib/sparsebit.c if (!sparsebit_is_set(s, idx)) idx 998 tools/testing/selftests/kvm/lib/sparsebit.c next_cleared = sparsebit_next_clear(s, idx); idx 1005 tools/testing/selftests/kvm/lib/sparsebit.c return next_cleared == 0 || next_cleared - idx >= num; idx 1010 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx) idx 1012 tools/testing/selftests/kvm/lib/sparsebit.c return !sparsebit_is_set(s, idx); idx 1017 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx, sparsebit_num_t num) idx 1022 tools/testing/selftests/kvm/lib/sparsebit.c assert(idx + num - 1 >= idx); idx 1025 tools/testing/selftests/kvm/lib/sparsebit.c if (!sparsebit_is_clear(s, idx)) idx 1029 tools/testing/selftests/kvm/lib/sparsebit.c next_set = sparsebit_next_set(s, idx); idx 1036 tools/testing/selftests/kvm/lib/sparsebit.c return next_set == 0 || next_set - idx >= num; idx 1110 tools/testing/selftests/kvm/lib/sparsebit.c if (!nodep1 || nodep1->idx > 0) idx 1129 tools/testing/selftests/kvm/lib/sparsebit.c assert(nodep1->idx + MASK_BITS + nodep1->num_after != (sparsebit_idx_t) 0); idx 1130 tools/testing/selftests/kvm/lib/sparsebit.c return nodep1->idx + MASK_BITS + nodep1->num_after; idx 1139 tools/testing/selftests/kvm/lib/sparsebit.c if (nodep1->idx + MASK_BITS + nodep1->num_after != nodep2->idx) idx 1140 tools/testing/selftests/kvm/lib/sparsebit.c return nodep1->idx + MASK_BITS + nodep1->num_after; idx 1181 tools/testing/selftests/kvm/lib/sparsebit.c if ((nodep->idx + MASK_BITS + nodep->num_after - 1) idx 1184 tools/testing/selftests/kvm/lib/sparsebit.c if (candidate->idx <= lowest_possible) { idx 1205 tools/testing/selftests/kvm/lib/sparsebit.c assert(candidate->idx > lowest_possible); idx 1218 tools/testing/selftests/kvm/lib/sparsebit.c start = lowest_possible - candidate->idx; idx 1224 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t first_num_after_idx = candidate->idx + MASK_BITS; idx 1252 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx; idx 1268 tools/testing/selftests/kvm/lib/sparsebit.c for (idx = lowest_possible - nodep1->idx; idx < MASK_BITS; idx++) idx 1269 tools/testing/selftests/kvm/lib/sparsebit.c if (!(nodep1->mask & (1 << idx))) idx 1270 tools/testing/selftests/kvm/lib/sparsebit.c return nodep1->idx + idx; idx 1279 tools/testing/selftests/kvm/lib/sparsebit.c return nodep1->idx + MASK_BITS + nodep1->num_after; idx 1287 tools/testing/selftests/kvm/lib/sparsebit.c if (nodep1->idx + MASK_BITS + nodep1->num_after != nodep2->idx) idx 1288 tools/testing/selftests/kvm/lib/sparsebit.c return nodep1->idx + MASK_BITS + nodep1->num_after; idx 1307 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx; idx 1311 tools/testing/selftests/kvm/lib/sparsebit.c for (idx = sparsebit_next_set(s, start); idx 1312 tools/testing/selftests/kvm/lib/sparsebit.c idx != 0 && idx + num - 1 >= idx; idx 1313 tools/testing/selftests/kvm/lib/sparsebit.c idx = sparsebit_next_set(s, idx)) { idx 1314 tools/testing/selftests/kvm/lib/sparsebit.c assert(sparsebit_is_set(s, idx)); idx 1320 tools/testing/selftests/kvm/lib/sparsebit.c if (sparsebit_is_set_num(s, idx, num)) idx 1321 tools/testing/selftests/kvm/lib/sparsebit.c return idx; idx 1327 tools/testing/selftests/kvm/lib/sparsebit.c idx = sparsebit_next_clear(s, idx); idx 1328 tools/testing/selftests/kvm/lib/sparsebit.c if (idx == 0) idx 1342 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx; idx 1346 tools/testing/selftests/kvm/lib/sparsebit.c for (idx = sparsebit_next_clear(s, start); idx 1347 tools/testing/selftests/kvm/lib/sparsebit.c idx != 0 && idx + num - 1 >= idx; idx 1348 tools/testing/selftests/kvm/lib/sparsebit.c idx = sparsebit_next_clear(s, idx)) { idx 1349 tools/testing/selftests/kvm/lib/sparsebit.c assert(sparsebit_is_clear(s, idx)); idx 1355 tools/testing/selftests/kvm/lib/sparsebit.c if (sparsebit_is_clear_num(s, idx, num)) idx 1356 tools/testing/selftests/kvm/lib/sparsebit.c return idx; idx 1362 tools/testing/selftests/kvm/lib/sparsebit.c idx = sparsebit_next_set(s, idx); idx 1363 tools/testing/selftests/kvm/lib/sparsebit.c if (idx == 0) idx 1376 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx; idx 1403 tools/testing/selftests/kvm/lib/sparsebit.c for (idx = start, n = num; n > 0 && idx % MASK_BITS != 0; idx++, n--) idx 1404 tools/testing/selftests/kvm/lib/sparsebit.c bit_set(s, idx); idx 1407 tools/testing/selftests/kvm/lib/sparsebit.c middle_start = idx; idx 1422 tools/testing/selftests/kvm/lib/sparsebit.c next && (next->idx < middle_end); idx 1424 tools/testing/selftests/kvm/lib/sparsebit.c assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end); idx 1443 tools/testing/selftests/kvm/lib/sparsebit.c idx = middle_end + 1; idx 1448 tools/testing/selftests/kvm/lib/sparsebit.c for (; n > 0; idx++, n--) idx 1449 tools/testing/selftests/kvm/lib/sparsebit.c bit_set(s, idx); idx 1458 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_idx_t idx; idx 1466 tools/testing/selftests/kvm/lib/sparsebit.c for (idx = start, n = num; n > 0 && idx % MASK_BITS != 0; idx++, n--) idx 1467 tools/testing/selftests/kvm/lib/sparsebit.c bit_clear(s, idx); idx 1470 tools/testing/selftests/kvm/lib/sparsebit.c middle_start = idx; idx 1485 tools/testing/selftests/kvm/lib/sparsebit.c next && (next->idx < middle_end); idx 1487 tools/testing/selftests/kvm/lib/sparsebit.c assert(next->idx + MASK_BITS + next->num_after - 1 <= middle_end); idx 1512 tools/testing/selftests/kvm/lib/sparsebit.c idx = middle_end + 1; idx 1517 tools/testing/selftests/kvm/lib/sparsebit.c for (; n > 0; idx++, n--) idx 1518 tools/testing/selftests/kvm/lib/sparsebit.c bit_clear(s, idx); idx 1522 tools/testing/selftests/kvm/lib/sparsebit.c void sparsebit_set(struct sparsebit *s, sparsebit_idx_t idx) idx 1524 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_set_num(s, idx, 1); idx 1528 tools/testing/selftests/kvm/lib/sparsebit.c void sparsebit_clear(struct sparsebit *s, sparsebit_idx_t idx) idx 1530 tools/testing/selftests/kvm/lib/sparsebit.c sparsebit_clear_num(s, idx, 1); idx 1608 tools/testing/selftests/kvm/lib/sparsebit.c low = high = nodep->idx + n1; idx 1612 tools/testing/selftests/kvm/lib/sparsebit.c high = nodep->idx + n1; idx 1651 tools/testing/selftests/kvm/lib/sparsebit.c low = nodep->idx + MASK_BITS; idx 1652 tools/testing/selftests/kvm/lib/sparsebit.c high = nodep->idx + MASK_BITS + nodep->num_after - 1; idx 1742 tools/testing/selftests/kvm/lib/sparsebit.c if (nodep->idx % MASK_BITS) { idx 1747 tools/testing/selftests/kvm/lib/sparsebit.c nodep, nodep->idx, MASK_BITS); idx 1756 tools/testing/selftests/kvm/lib/sparsebit.c if ((nodep->idx + MASK_BITS + nodep->num_after - 1) < nodep->idx) { idx 1761 tools/testing/selftests/kvm/lib/sparsebit.c nodep, nodep->idx, MASK_BITS, nodep->num_after); idx 1808 tools/testing/selftests/kvm/lib/sparsebit.c if (prev->idx >= nodep->idx) { idx 1813 tools/testing/selftests/kvm/lib/sparsebit.c prev, prev->idx, nodep, nodep->idx); idx 1822 tools/testing/selftests/kvm/lib/sparsebit.c if ((prev->idx + MASK_BITS + prev->num_after - 1) idx 1823 tools/testing/selftests/kvm/lib/sparsebit.c >= nodep->idx) { idx 1831 tools/testing/selftests/kvm/lib/sparsebit.c prev, prev->idx, prev->num_after, idx 1832 tools/testing/selftests/kvm/lib/sparsebit.c nodep, nodep->idx, nodep->num_after, idx 1844 tools/testing/selftests/kvm/lib/sparsebit.c prev->idx + MASK_BITS + prev->num_after == nodep->idx) { idx 1853 tools/testing/selftests/kvm/lib/sparsebit.c prev, prev->idx, prev->num_after, idx 1854 tools/testing/selftests/kvm/lib/sparsebit.c nodep, nodep->idx, nodep->num_after, idx 1904 tools/testing/selftests/kvm/lib/sparsebit.c static bool get_value(sparsebit_idx_t idx) idx 1909 tools/testing/selftests/kvm/lib/sparsebit.c if (ranges[i].first <= idx && idx <= ranges[i].last) idx 342 tools/testing/selftests/net/ip_defrag.c int idx; idx 346 tools/testing/selftests/net/ip_defrag.c for (idx = 0; idx < MSG_LEN_MAX; ++idx) idx 347 tools/testing/selftests/net/ip_defrag.c udp_payload[idx] = idx % 256; idx 25 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/simple_sync_srcu.c bool try_check_zero(struct srcu_struct *sp, int idx, int trycount); idx 31 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/simple_sync_srcu.c int idx; idx 42 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/simple_sync_srcu.c idx = 1 ^ (sp->completed & 1); idx 46 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/simple_sync_srcu.c assume(try_check_zero(sp, idx, trycount)); idx 50 tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/simple_sync_srcu.c assume(try_check_zero(sp, idx^1, trycount)); idx 14 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c int idx; idx 17 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c idx = srcu_read_lock(&ss); idx 23 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c srcu_read_unlock(&ss, idx); idx 24 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c idx = srcu_read_lock(&ss); idx 29 tools/testing/selftests/rcutorture/formal/srcu-cbmc/tests/store_buffering/test.c srcu_read_unlock(&ss, idx); idx 117 tools/testing/selftests/vm/transhuge-stress.c size_t idx = pfn >> (HPAGE_SHIFT - PAGE_SHIFT); idx 120 tools/testing/selftests/vm/transhuge-stress.c if (idx >= map_len) { idx 121 tools/testing/selftests/vm/transhuge-stress.c map = realloc(map, idx + 1); idx 124 tools/testing/selftests/vm/transhuge-stress.c memset(map + map_len, 0, idx + 1 - map_len); idx 125 tools/testing/selftests/vm/transhuge-stress.c map_len = idx + 1; idx 127 tools/testing/selftests/vm/transhuge-stress.c if (!map[idx]) idx 129 tools/testing/selftests/vm/transhuge-stress.c map[idx] = 1; idx 131 tools/testing/selftests/x86/sigreturn.c static unsigned short GDT3(int idx) idx 133 tools/testing/selftests/x86/sigreturn.c return (idx << 3) | 3; idx 136 tools/testing/selftests/x86/sigreturn.c static unsigned short LDT3(int idx) idx 138 tools/testing/selftests/x86/sigreturn.c return (idx << 3) | 7; idx 153 tools/testing/vsock/control.c size_t idx = 0; idx 161 tools/testing/vsock/control.c if (idx >= buflen) { idx 175 tools/testing/vsock/control.c ret = recv(control_fd, &buf[idx], 1, 0); idx 189 tools/testing/vsock/control.c if (buf[idx] == '\n') { idx 190 tools/testing/vsock/control.c buf[idx] = '\0'; idx 194 tools/testing/vsock/control.c idx++; idx 148 tools/virtio/ringtest/virtio_ring_0_9.c ring.avail->idx = guest.avail_idx; idx 167 tools/virtio/ringtest/virtio_ring_0_9.c if (ring.used->idx == guest.last_used_idx) idx 206 tools/virtio/ringtest/virtio_ring_0_9.c return ring.used->idx == last_used_idx; idx 266 tools/virtio/ringtest/virtio_ring_0_9.c return head == ring.avail->idx; idx 286 tools/virtio/ringtest/virtio_ring_0_9.c if (used_idx == ring.avail->idx) idx 314 tools/virtio/ringtest/virtio_ring_0_9.c ring.used->idx = host.used_idx; idx 28 tools/virtio/virtio_test.c int idx; idx 63 tools/virtio/virtio_test.c struct vhost_vring_state state = { .index = info->idx }; idx 64 tools/virtio/virtio_test.c struct vhost_vring_file file = { .index = info->idx }; idx 67 tools/virtio/virtio_test.c .index = info->idx, idx 95 tools/virtio/virtio_test.c info->idx = dev->nvqs; idx 102 tools/virtio/virtio_test.c info->vq = vring_new_virtqueue(info->idx, idx 109 tools/virtio/virtio_test.c dev->fds[info->idx].fd = info->call; idx 110 tools/virtio/virtio_test.c dev->fds[info->idx].events = POLLIN; idx 116 tools/virtio/vringh_test.c err = get_user(avail_idx, &vrh->vring.avail->idx); idx 648 tools/virtio/vringh_test.c assert(vrh.vring.used->idx % RINGSIZE != 0); idx 440 virt/kvm/arm/mmu.c int idx; idx 442 virt/kvm/arm/mmu.c idx = srcu_read_lock(&kvm->srcu); idx 450 virt/kvm/arm/mmu.c srcu_read_unlock(&kvm->srcu, idx); idx 975 virt/kvm/arm/mmu.c int idx; idx 977 virt/kvm/arm/mmu.c idx = srcu_read_lock(&kvm->srcu); idx 987 virt/kvm/arm/mmu.c srcu_read_unlock(&kvm->srcu, idx); idx 1916 virt/kvm/arm/mmu.c int ret, idx; idx 1951 virt/kvm/arm/mmu.c idx = srcu_read_lock(&vcpu->kvm->srcu); idx 2009 virt/kvm/arm/mmu.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 37 virt/kvm/arm/pmu.c pmc -= pmc->idx; idx 51 virt/kvm/arm/pmu.c return test_bit(pmc->idx >> 1, vcpu->arch.pmu.chained); idx 73 virt/kvm/arm/pmu.c kvm_pmu_idx_is_high_counter(pmc->idx)) idx 111 virt/kvm/arm/pmu.c reg = PMEVCNTR0_EL0 + pmc->idx; idx 118 virt/kvm/arm/pmu.c reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX) idx 119 virt/kvm/arm/pmu.c ? PMCCNTR_EL0 : PMEVCNTR0_EL0 + pmc->idx; idx 204 virt/kvm/arm/pmu.c if (pmc->idx == ARMV8_PMU_CYCLE_IDX) { idx 208 virt/kvm/arm/pmu.c reg = PMEVCNTR0_EL0 + pmc->idx; idx 231 virt/kvm/arm/pmu.c pmu->pmc[i].idx = i; idx 448 virt/kvm/arm/pmu.c int idx = pmc->idx; idx 459 virt/kvm/arm/pmu.c if (!kvm_pmu_idx_is_64bit(vcpu, pmc->idx)) idx 466 virt/kvm/arm/pmu.c __vcpu_sys_reg(vcpu, PMOVSSET_EL0) |= BIT(idx); idx 579 virt/kvm/arm/pmu.c reg = (pmc->idx == ARMV8_PMU_CYCLE_IDX) idx 580 virt/kvm/arm/pmu.c ? PMCCFILTR_EL0 : PMEVTYPER0_EL0 + pmc->idx; idx 588 virt/kvm/arm/pmu.c pmc->idx != ARMV8_PMU_CYCLE_IDX) idx 595 virt/kvm/arm/pmu.c attr.disabled = !kvm_pmu_counter_is_enabled(vcpu, pmc->idx); idx 600 virt/kvm/arm/pmu.c attr.config = (pmc->idx == ARMV8_PMU_CYCLE_IDX) ? idx 605 virt/kvm/arm/pmu.c if (kvm_pmu_idx_has_chain_evtype(vcpu, pmc->idx)) { idx 612 virt/kvm/arm/pmu.c if (kvm_pmu_counter_is_enabled(vcpu, pmc->idx + 1)) idx 620 virt/kvm/arm/pmu.c if (kvm_pmu_idx_is_64bit(vcpu, pmc->idx)) idx 651 virt/kvm/arm/pmu.c if (kvm_pmu_idx_has_chain_evtype(vcpu, pmc->idx)) { idx 659 virt/kvm/arm/pmu.c set_bit(pmc->idx >> 1, vcpu->arch.pmu.chained); idx 661 virt/kvm/arm/pmu.c clear_bit(pmc->idx >> 1, vcpu->arch.pmu.chained); idx 267 virt/kvm/arm/vgic/vgic-init.c int ret = 0, i, idx; idx 285 virt/kvm/arm/vgic/vgic-init.c kvm_for_each_vcpu(idx, vcpu, kvm) { idx 297 virt/kvm/arm/vgic/vgic-init.c irq->targets = 1U << idx; idx 903 virt/kvm/arm/vgic/vgic-its.c int index, idx; idx 965 virt/kvm/arm/vgic/vgic-its.c idx = srcu_read_lock(&its->dev->kvm->srcu); idx 967 virt/kvm/arm/vgic/vgic-its.c srcu_read_unlock(&its->dev->kvm->srcu, idx); idx 69 virt/kvm/eventfd.c int idx; idx 78 virt/kvm/eventfd.c idx = srcu_read_lock(&kvm->irq_srcu); idx 83 virt/kvm/eventfd.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 193 virt/kvm/eventfd.c int idx; idx 196 virt/kvm/eventfd.c idx = srcu_read_lock(&kvm->irq_srcu); idx 206 virt/kvm/eventfd.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 289 virt/kvm/eventfd.c int idx; idx 390 virt/kvm/eventfd.c idx = srcu_read_lock(&kvm->irq_srcu); idx 420 virt/kvm/eventfd.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 449 virt/kvm/eventfd.c int gsi, idx; idx 451 virt/kvm/eventfd.c idx = srcu_read_lock(&kvm->irq_srcu); idx 457 virt/kvm/eventfd.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 461 virt/kvm/eventfd.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 479 virt/kvm/eventfd.c int gsi, idx; idx 483 virt/kvm/eventfd.c idx = srcu_read_lock(&kvm->irq_srcu); idx 487 virt/kvm/eventfd.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 75 virt/kvm/irqchip.c int ret = -1, i, idx; idx 83 virt/kvm/irqchip.c idx = srcu_read_lock(&kvm->irq_srcu); idx 85 virt/kvm/irqchip.c srcu_read_unlock(&kvm->irq_srcu, idx); idx 388 virt/kvm/kvm_main.c int idx; idx 390 virt/kvm/kvm_main.c idx = srcu_read_lock(&kvm->srcu); idx 392 virt/kvm/kvm_main.c srcu_read_unlock(&kvm->srcu, idx); idx 401 virt/kvm/kvm_main.c int idx; idx 403 virt/kvm/kvm_main.c idx = srcu_read_lock(&kvm->srcu); idx 411 virt/kvm/kvm_main.c srcu_read_unlock(&kvm->srcu, idx); idx 418 virt/kvm/kvm_main.c int need_tlb_flush = 0, idx; idx 420 virt/kvm/kvm_main.c idx = srcu_read_lock(&kvm->srcu); idx 435 virt/kvm/kvm_main.c srcu_read_unlock(&kvm->srcu, idx); idx 470 virt/kvm/kvm_main.c int young, idx; idx 472 virt/kvm/kvm_main.c idx = srcu_read_lock(&kvm->srcu); idx 480 virt/kvm/kvm_main.c srcu_read_unlock(&kvm->srcu, idx); idx 491 virt/kvm/kvm_main.c int young, idx; idx 493 virt/kvm/kvm_main.c idx = srcu_read_lock(&kvm->srcu); idx 510 virt/kvm/kvm_main.c srcu_read_unlock(&kvm->srcu, idx); idx 520 virt/kvm/kvm_main.c int young, idx; idx 522 virt/kvm/kvm_main.c idx = srcu_read_lock(&kvm->srcu); idx 526 virt/kvm/kvm_main.c srcu_read_unlock(&kvm->srcu, idx); idx 535 virt/kvm/kvm_main.c int idx; idx 537 virt/kvm/kvm_main.c idx = srcu_read_lock(&kvm->srcu); idx 539 virt/kvm/kvm_main.c srcu_read_unlock(&kvm->srcu, idx); idx 2449 virt/kvm/kvm_main.c int idx = srcu_read_lock(&vcpu->kvm->srcu); idx 2462 virt/kvm/kvm_main.c srcu_read_unlock(&vcpu->kvm->srcu, idx); idx 3857 virt/kvm/kvm_main.c int idx; idx 3859 virt/kvm/kvm_main.c idx = kvm_io_bus_get_first_dev(bus, range->addr, range->len); idx 3860 virt/kvm/kvm_main.c if (idx < 0) idx 3863 virt/kvm/kvm_main.c while (idx < bus->dev_count && idx 3864 virt/kvm/kvm_main.c kvm_io_bus_cmp(range, &bus->range[idx]) == 0) { idx 3865 virt/kvm/kvm_main.c if (!kvm_iodevice_write(vcpu, bus->range[idx].dev, range->addr, idx 3867 virt/kvm/kvm_main.c return idx; idx 3868 virt/kvm/kvm_main.c idx++; idx 3928 virt/kvm/kvm_main.c int idx; idx 3930 virt/kvm/kvm_main.c idx = kvm_io_bus_get_first_dev(bus, range->addr, range->len); idx 3931 virt/kvm/kvm_main.c if (idx < 0) idx 3934 virt/kvm/kvm_main.c while (idx < bus->dev_count && idx 3935 virt/kvm/kvm_main.c kvm_io_bus_cmp(range, &bus->range[idx]) == 0) { idx 3936 virt/kvm/kvm_main.c if (!kvm_iodevice_read(vcpu, bus->range[idx].dev, range->addr, idx 3938 virt/kvm/kvm_main.c return idx; idx 3939 virt/kvm/kvm_main.c idx++;