Lines Matching refs:i
30 struct btree_iter_set *i) in sort_key_next() argument
32 i->k = bkey_next(i->k); in sort_key_next()
34 if (i->k == i->end) in sort_key_next()
35 *i = iter->data[--iter->used]; in sort_key_next()
48 unsigned i; in __ptr_invalid() local
50 for (i = 0; i < KEY_PTRS(k); i++) in __ptr_invalid()
51 if (ptr_available(c, k, i)) { in __ptr_invalid()
52 struct cache *ca = PTR_CACHE(c, k, i); in __ptr_invalid()
53 size_t bucket = PTR_BUCKET_NR(c, k, i); in __ptr_invalid()
54 size_t r = bucket_remainder(c, PTR_OFFSET(k, i)); in __ptr_invalid()
69 unsigned i; in bch_ptr_status() local
71 for (i = 0; i < KEY_PTRS(k); i++) in bch_ptr_status()
72 if (ptr_available(c, k, i)) { in bch_ptr_status()
73 struct cache *ca = PTR_CACHE(c, k, i); in bch_ptr_status()
74 size_t bucket = PTR_BUCKET_NR(c, k, i); in bch_ptr_status()
75 size_t r = bucket_remainder(c, PTR_OFFSET(k, i)); in bch_ptr_status()
83 if (ptr_stale(c, k, i)) in bch_ptr_status()
98 unsigned i = 0; in bch_extent_to_text() local
105 for (i = 0; i < KEY_PTRS(k); i++) { in bch_extent_to_text()
106 if (i) in bch_extent_to_text()
109 if (PTR_DEV(k, i) == PTR_CHECK_DEV) in bch_extent_to_text()
112 p("%llu:%llu gen %llu", PTR_DEV(k, i), in bch_extent_to_text()
113 PTR_OFFSET(k, i), PTR_GEN(k, i)); in bch_extent_to_text()
173 unsigned i; in btree_ptr_bad_expensive() local
178 for (i = 0; i < KEY_PTRS(k); i++) in btree_ptr_bad_expensive()
179 if (ptr_available(b->c, k, i)) { in btree_ptr_bad_expensive()
180 g = PTR_BUCKET(b->c, k, i); in btree_ptr_bad_expensive()
198 buf, PTR_BUCKET_NR(b->c, k, i), atomic_read(&g->pin), in btree_ptr_bad_expensive()
206 unsigned i; in bch_btree_ptr_bad() local
213 for (i = 0; i < KEY_PTRS(k); i++) in bch_btree_ptr_bad()
214 if (!ptr_available(b->c, k, i) || in bch_btree_ptr_bad()
215 ptr_stale(b->c, k, i)) in bch_btree_ptr_bad()
268 struct btree_iter_set *top = iter->data, *i = top + 1; in bch_extent_sort_fixup() local
271 bch_extent_sort_cmp(i[0], i[1])) in bch_extent_sort_fixup()
272 i++; in bch_extent_sort_fixup()
274 if (bkey_cmp(top->k, &START_KEY(i->k)) <= 0) in bch_extent_sort_fixup()
277 if (!KEY_SIZE(i->k)) { in bch_extent_sort_fixup()
278 sort_key_next(iter, i); in bch_extent_sort_fixup()
279 heap_sift(iter, i - top, bch_extent_sort_cmp); in bch_extent_sort_fixup()
283 if (top->k > i->k) { in bch_extent_sort_fixup()
284 if (bkey_cmp(top->k, i->k) >= 0) in bch_extent_sort_fixup()
285 sort_key_next(iter, i); in bch_extent_sort_fixup()
287 bch_cut_front(top->k, i->k); in bch_extent_sort_fixup()
289 heap_sift(iter, i - top, bch_extent_sort_cmp); in bch_extent_sort_fixup()
292 BUG_ON(!bkey_cmp(&START_KEY(top->k), &START_KEY(i->k))); in bch_extent_sort_fixup()
294 if (bkey_cmp(i->k, top->k) < 0) { in bch_extent_sort_fixup()
297 bch_cut_back(&START_KEY(i->k), tmp); in bch_extent_sort_fixup()
298 bch_cut_front(i->k, top->k); in bch_extent_sort_fixup()
303 bch_cut_back(&START_KEY(i->k), top->k); in bch_extent_sort_fixup()
365 unsigned i; in bch_extent_insert_fixup() local
386 for (i = 0; i < KEY_PTRS(replace_key); i++) in bch_extent_insert_fixup()
387 if (k->ptr[i] != replace_key->ptr[i] + offset) in bch_extent_insert_fixup()
537 unsigned i, stale; in bch_extent_bad() local
543 for (i = 0; i < KEY_PTRS(k); i++) in bch_extent_bad()
544 if (!ptr_available(b->c, k, i)) in bch_extent_bad()
550 for (i = 0; i < KEY_PTRS(k); i++) { in bch_extent_bad()
551 g = PTR_BUCKET(b->c, k, i); in bch_extent_bad()
552 stale = ptr_stale(b->c, k, i); in bch_extent_bad()
565 bch_extent_bad_expensive(b, k, i)) in bch_extent_bad()
581 unsigned i; in bch_extent_merge() local
586 for (i = 0; i < KEY_PTRS(l); i++) in bch_extent_merge()
587 if (l->ptr[i] + PTR(0, KEY_SIZE(l), 0) != r->ptr[i] || in bch_extent_merge()
588 PTR_BUCKET_NR(b->c, l, i) != PTR_BUCKET_NR(b->c, r, i)) in bch_extent_merge()