ref2 711 arch/x86/kernel/tsc.c u64 tsc1, tsc2, delta, ref1, ref2; ref2 758 arch/x86/kernel/tsc.c tsc2 = tsc_read_refs(&ref2, hpet); ref2 765 arch/x86/kernel/tsc.c if (ref1 == ref2) ref2 774 arch/x86/kernel/tsc.c tsc2 = calc_hpet_ref(tsc2, ref1, ref2); ref2 776 arch/x86/kernel/tsc.c tsc2 = calc_pmtimer_ref(tsc2, ref1, ref2); ref2 817 arch/x86/kernel/tsc.c if (!hpet && !ref1 && !ref2) { ref2 836 arch/x86/kernel/tsc.c if (!hpet && !ref1 && !ref2) { ref2 176 fs/btrfs/backref.c struct prelim_ref *ref2) ref2 178 fs/btrfs/backref.c if (ref1->level < ref2->level) ref2 180 fs/btrfs/backref.c if (ref1->level > ref2->level) ref2 182 fs/btrfs/backref.c if (ref1->root_id < ref2->root_id) ref2 184 fs/btrfs/backref.c if (ref1->root_id > ref2->root_id) ref2 186 fs/btrfs/backref.c if (ref1->key_for_search.type < ref2->key_for_search.type) ref2 188 fs/btrfs/backref.c if (ref1->key_for_search.type > ref2->key_for_search.type) ref2 190 fs/btrfs/backref.c if (ref1->key_for_search.objectid < ref2->key_for_search.objectid) ref2 192 fs/btrfs/backref.c if (ref1->key_for_search.objectid > ref2->key_for_search.objectid) ref2 194 fs/btrfs/backref.c if (ref1->key_for_search.offset < ref2->key_for_search.offset) ref2 196 fs/btrfs/backref.c if (ref1->key_for_search.offset > ref2->key_for_search.offset) ref2 198 fs/btrfs/backref.c if (ref1->parent < ref2->parent) ref2 200 fs/btrfs/backref.c if (ref1->parent > ref2->parent) ref2 205 fs/btrfs/delayed-ref.c struct btrfs_delayed_tree_ref *ref2) ref2 208 fs/btrfs/delayed-ref.c if (ref1->root < ref2->root) ref2 210 fs/btrfs/delayed-ref.c if (ref1->root > ref2->root) ref2 213 fs/btrfs/delayed-ref.c if (ref1->parent < ref2->parent) ref2 215 fs/btrfs/delayed-ref.c if (ref1->parent > ref2->parent) ref2 225 fs/btrfs/delayed-ref.c struct btrfs_delayed_data_ref *ref2) ref2 228 fs/btrfs/delayed-ref.c if (ref1->root < ref2->root) ref2 230 fs/btrfs/delayed-ref.c if (ref1->root > ref2->root) ref2 232 fs/btrfs/delayed-ref.c if (ref1->objectid < ref2->objectid) ref2 234 fs/btrfs/delayed-ref.c if (ref1->objectid > ref2->objectid) ref2 236 fs/btrfs/delayed-ref.c if (ref1->offset < ref2->offset) ref2 238 fs/btrfs/delayed-ref.c if (ref1->offset > ref2->offset) ref2 241 fs/btrfs/delayed-ref.c if (ref1->parent < ref2->parent) ref2 243 fs/btrfs/delayed-ref.c if (ref1->parent > ref2->parent) ref2 250 fs/btrfs/delayed-ref.c struct btrfs_delayed_ref_node *ref2, ref2 255 fs/btrfs/delayed-ref.c if (ref1->type < ref2->type) ref2 257 fs/btrfs/delayed-ref.c if (ref1->type > ref2->type) ref2 262 fs/btrfs/delayed-ref.c btrfs_delayed_node_to_tree_ref(ref2)); ref2 265 fs/btrfs/delayed-ref.c btrfs_delayed_node_to_data_ref(ref2)); ref2 269 fs/btrfs/delayed-ref.c if (ref1->seq < ref2->seq) ref2 271 fs/btrfs/delayed-ref.c if (ref1->seq > ref2->seq) ref2 637 fs/btrfs/extent-tree.c struct btrfs_shared_data_ref *ref2 = NULL; ref2 650 fs/btrfs/extent-tree.c ref2 = btrfs_item_ptr(leaf, path->slots[0], ref2 652 fs/btrfs/extent-tree.c num_refs = btrfs_shared_data_ref_count(leaf, ref2); ref2 671 fs/btrfs/extent-tree.c btrfs_set_shared_data_ref_count(leaf, ref2, num_refs); ref2 683 fs/btrfs/extent-tree.c struct btrfs_shared_data_ref *ref2; ref2 702 fs/btrfs/extent-tree.c ref2 = (struct btrfs_shared_data_ref *)(iref + 1); ref2 703 fs/btrfs/extent-tree.c num_refs = btrfs_shared_data_ref_count(leaf, ref2); ref2 710 fs/btrfs/extent-tree.c ref2 = btrfs_item_ptr(leaf, path->slots[0], ref2 712 fs/btrfs/extent-tree.c num_refs = btrfs_shared_data_ref_count(leaf, ref2); ref2 140 fs/btrfs/ref-verify.c static int comp_refs(struct ref_entry *ref1, struct ref_entry *ref2) ref2 142 fs/btrfs/ref-verify.c if (ref1->root_objectid < ref2->root_objectid) ref2 144 fs/btrfs/ref-verify.c if (ref1->root_objectid > ref2->root_objectid) ref2 146 fs/btrfs/ref-verify.c if (ref1->parent < ref2->parent) ref2 148 fs/btrfs/ref-verify.c if (ref1->parent > ref2->parent) ref2 150 fs/btrfs/ref-verify.c if (ref1->owner < ref2->owner) ref2 152 fs/btrfs/ref-verify.c if (ref1->owner > ref2->owner) ref2 154 fs/btrfs/ref-verify.c if (ref1->offset < ref2->offset) ref2 156 fs/btrfs/ref-verify.c if (ref1->offset > ref2->offset) ref2 316 fs/jffs2/debug.c struct jffs2_raw_node_ref *ref2 = jeb->first_node; ref2 318 fs/jffs2/debug.c while (ref2) { ref2 319 fs/jffs2/debug.c uint32_t totlen = ref_totlen(c, jeb, ref2); ref2 321 fs/jffs2/debug.c if (ref_offset(ref2) < jeb->offset || ref2 322 fs/jffs2/debug.c ref_offset(ref2) > jeb->offset + c->sector_size) { ref2 324 fs/jffs2/debug.c ref_offset(ref2), jeb->offset); ref2 328 fs/jffs2/debug.c if (ref_flags(ref2) == REF_UNCHECKED) ref2 330 fs/jffs2/debug.c else if (!ref_obsolete(ref2)) ref2 335 fs/jffs2/debug.c if ((!ref_next(ref2)) != (ref2 == jeb->last_node)) { ref2 337 fs/jffs2/debug.c ref_offset(ref2), ref2, ref_offset(ref_next(ref2)), ref_next(ref2), ref2 341 fs/jffs2/debug.c ref2 = ref_next(ref2); ref2 361 lib/lz4/lz4hc_compress.c const BYTE *ref2 = NULL; ref2 395 lib/lz4/lz4hc_compress.c matchlimit, ml, &ref2, ref2 422 lib/lz4/lz4hc_compress.c ref = ref2; ref2 445 lib/lz4/lz4hc_compress.c ref2 += correction; ref2 473 lib/lz4/lz4hc_compress.c ml2, ref2, limit, oend)) ref2 489 lib/lz4/lz4hc_compress.c ref2 += correction; ref2 493 lib/lz4/lz4hc_compress.c ref2 = ref3; ref2 506 lib/lz4/lz4hc_compress.c ref0 = ref2; ref2 512 lib/lz4/lz4hc_compress.c ref2 = ref3; ref2 533 lib/lz4/lz4hc_compress.c ref2 += correction; ref2 544 lib/lz4/lz4hc_compress.c ref = ref2; ref2 548 lib/lz4/lz4hc_compress.c ref2 = ref3; ref2 497 tools/perf/util/bpf-loader.c struct probe_trace_arg_ref *ref1, *ref2; ref2 507 tools/perf/util/bpf-loader.c ref2 = arg2->ref; ref2 509 tools/perf/util/bpf-loader.c while (ref1 && ref2) { ref2 510 tools/perf/util/bpf-loader.c ret = ref2->offset - ref1->offset; ref2 515 tools/perf/util/bpf-loader.c ref2 = ref2->next; ref2 518 tools/perf/util/bpf-loader.c if (ref1 || ref2) ref2 519 tools/perf/util/bpf-loader.c return ref2 ? 1 : -1;