btp               571 fs/xfs/libxfs/xfs_da_format.h xfs_dir2_block_leaf_p(struct xfs_dir2_block_tail *btp)
btp               573 fs/xfs/libxfs/xfs_da_format.h 	return ((struct xfs_dir2_leaf_entry *)btp) - be32_to_cpu(btp->count);
btp               161 fs/xfs/libxfs/xfs_dir2_block.c 	struct xfs_dir2_block_tail	*btp,
btp               180 fs/xfs/libxfs/xfs_dir2_block.c 	if (btp->stale) {
btp               205 fs/xfs/libxfs/xfs_dir2_block.c 			if (be16_to_cpu(dup->length) + (be32_to_cpu(btp->stale) - 1) *
btp               208 fs/xfs/libxfs/xfs_dir2_block.c 		} else if ((be32_to_cpu(btp->stale) - 1) * (uint)sizeof(*blp) < len)
btp               274 fs/xfs/libxfs/xfs_dir2_block.c 	struct xfs_dir2_block_tail	*btp,
btp               285 fs/xfs/libxfs/xfs_dir2_block.c 	fromidx = toidx = be32_to_cpu(btp->count) - 1;
btp               301 fs/xfs/libxfs/xfs_dir2_block.c 	*lfloglow = toidx + 1 - (be32_to_cpu(btp->stale) - 1);
btp               302 fs/xfs/libxfs/xfs_dir2_block.c 	*lfloghigh -= be32_to_cpu(btp->stale) - 1;
btp               303 fs/xfs/libxfs/xfs_dir2_block.c 	be32_add_cpu(&btp->count, -(be32_to_cpu(btp->stale) - 1));
btp               306 fs/xfs/libxfs/xfs_dir2_block.c 		(xfs_dir2_data_aoff_t)((be32_to_cpu(btp->stale) - 1) * sizeof(*blp)),
btp               308 fs/xfs/libxfs/xfs_dir2_block.c 	btp->stale = cpu_to_be32(1);
btp               327 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;		/* block tail */
btp               364 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp               365 fs/xfs/libxfs/xfs_dir2_block.c 	blp = xfs_dir2_block_leaf_p(btp);
btp               371 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_need_space(dp, hdr, btp, blp, &tagp, &dup,
btp               407 fs/xfs/libxfs/xfs_dir2_block.c 		xfs_dir2_block_compact(args, bp, hdr, btp, blp, &needlog,
btp               410 fs/xfs/libxfs/xfs_dir2_block.c 		blp = xfs_dir2_block_leaf_p(btp);
btp               411 fs/xfs/libxfs/xfs_dir2_block.c 	} else if (btp->stale) {
btp               416 fs/xfs/libxfs/xfs_dir2_block.c 		lfloglow = be32_to_cpu(btp->count);
btp               423 fs/xfs/libxfs/xfs_dir2_block.c 	for (low = 0, high = be32_to_cpu(btp->count) - 1; low <= high; ) {
btp               438 fs/xfs/libxfs/xfs_dir2_block.c 	if (!btp->stale) {
btp               455 fs/xfs/libxfs/xfs_dir2_block.c 		be32_add_cpu(&btp->count, 1);
btp               487 fs/xfs/libxfs/xfs_dir2_block.c 		     highstale < be32_to_cpu(btp->count) &&
btp               497 fs/xfs/libxfs/xfs_dir2_block.c 		    (highstale == be32_to_cpu(btp->count) ||
btp               509 fs/xfs/libxfs/xfs_dir2_block.c 			ASSERT(highstale < be32_to_cpu(btp->count));
btp               517 fs/xfs/libxfs/xfs_dir2_block.c 		be32_add_cpu(&btp->stale, -1);
btp               572 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;
btp               574 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(tp->t_mountp->m_dir_geo, hdr);
btp               575 fs/xfs/libxfs/xfs_dir2_block.c 	blp = xfs_dir2_block_leaf_p(btp);
btp               589 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;
btp               591 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(tp->t_mountp->m_dir_geo, hdr);
btp               592 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_trans_log_buf(tp, bp, (uint)((char *)btp - (char *)hdr),
btp               593 fs/xfs/libxfs/xfs_dir2_block.c 		(uint)((char *)(btp + 1) - (char *)hdr - 1));
btp               607 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;		/* block tail */
btp               624 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp               625 fs/xfs/libxfs/xfs_dir2_block.c 	blp = xfs_dir2_block_leaf_p(btp);
btp               655 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;		/* block tail */
btp               677 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp               678 fs/xfs/libxfs/xfs_dir2_block.c 	blp = xfs_dir2_block_leaf_p(btp);
btp               683 fs/xfs/libxfs/xfs_dir2_block.c 	for (low = 0, high = be32_to_cpu(btp->count) - 1; ; ) {
btp               729 fs/xfs/libxfs/xfs_dir2_block.c 	} while (++mid < be32_to_cpu(btp->count) &&
btp               757 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;		/* block tail */
btp               780 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp               781 fs/xfs/libxfs/xfs_dir2_block.c 	blp = xfs_dir2_block_leaf_p(btp);
btp               798 fs/xfs/libxfs/xfs_dir2_block.c 	be32_add_cpu(&btp->stale, 1);
btp               837 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;		/* block tail */
btp               854 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp               855 fs/xfs/libxfs/xfs_dir2_block.c 	blp = xfs_dir2_block_leaf_p(btp);
btp               901 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;		/* block tail */
btp               998 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp               999 fs/xfs/libxfs/xfs_dir2_block.c 	btp->count = cpu_to_be32(leafhdr.count - leafhdr.stale);
btp              1000 fs/xfs/libxfs/xfs_dir2_block.c 	btp->stale = 0;
btp              1005 fs/xfs/libxfs/xfs_dir2_block.c 	lep = xfs_dir2_block_leaf_p(btp);
btp              1011 fs/xfs/libxfs/xfs_dir2_block.c 	ASSERT(to == be32_to_cpu(btp->count));
btp              1012 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_log_leaf(tp, dbp, 0, be32_to_cpu(btp->count) - 1);
btp              1048 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_tail_t	*btp;		/* block tail pointer */
btp              1120 fs/xfs/libxfs/xfs_dir2_block.c 	i = (uint)sizeof(*btp) +
btp              1136 fs/xfs/libxfs/xfs_dir2_block.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp              1137 fs/xfs/libxfs/xfs_dir2_block.c 	btp->count = cpu_to_be32(sfp->count + 2);	/* ., .. */
btp              1138 fs/xfs/libxfs/xfs_dir2_block.c 	btp->stale = 0;
btp              1139 fs/xfs/libxfs/xfs_dir2_block.c 	blp = xfs_dir2_block_leaf_p(btp);
btp              1242 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_sort(blp, be32_to_cpu(btp->count), sizeof(*blp), xfs_dir2_block_sort);
btp              1248 fs/xfs/libxfs/xfs_dir2_block.c 	xfs_dir2_block_log_leaf(tp, bp, 0, be32_to_cpu(btp->count) - 1);
btp                38 fs/xfs/libxfs/xfs_dir2_data.c 	xfs_dir2_block_tail_t	*btp=NULL;	/* block tail */
btp                79 fs/xfs/libxfs/xfs_dir2_data.c 		btp = xfs_dir2_block_tail_p(geo, hdr);
btp                80 fs/xfs/libxfs/xfs_dir2_data.c 		lep = xfs_dir2_block_leaf_p(btp);
btp                89 fs/xfs/libxfs/xfs_dir2_data.c 		if (be32_to_cpu(btp->count) >=
btp                90 fs/xfs/libxfs/xfs_dir2_data.c 		    ((char *)btp - p) / sizeof(struct xfs_dir2_leaf_entry))
btp               193 fs/xfs/libxfs/xfs_dir2_data.c 			for (i = 0; i < be32_to_cpu(btp->count); i++) {
btp               198 fs/xfs/libxfs/xfs_dir2_data.c 			if (i >= be32_to_cpu(btp->count))
btp               210 fs/xfs/libxfs/xfs_dir2_data.c 		for (i = stale = 0; i < be32_to_cpu(btp->count); i++) {
btp               218 fs/xfs/libxfs/xfs_dir2_data.c 		if (count != be32_to_cpu(btp->count) - be32_to_cpu(btp->stale))
btp               220 fs/xfs/libxfs/xfs_dir2_data.c 		if (stale != be32_to_cpu(btp->stale))
btp               338 fs/xfs/libxfs/xfs_dir2_leaf.c 	xfs_dir2_block_tail_t	*btp;		/* block's tail */
btp               376 fs/xfs/libxfs/xfs_dir2_leaf.c 	btp = xfs_dir2_block_tail_p(args->geo, hdr);
btp               377 fs/xfs/libxfs/xfs_dir2_leaf.c 	blp = xfs_dir2_block_leaf_p(btp);
btp               385 fs/xfs/libxfs/xfs_dir2_leaf.c 	leafhdr.count = be32_to_cpu(btp->count);
btp               386 fs/xfs/libxfs/xfs_dir2_leaf.c 	leafhdr.stale = be32_to_cpu(btp->stale);
btp               394 fs/xfs/libxfs/xfs_dir2_leaf.c 	memcpy(ents, blp, be32_to_cpu(btp->count) * sizeof(xfs_dir2_leaf_entry_t));
btp                54 fs/xfs/libxfs/xfs_dir2_sf.c 	xfs_dir2_block_tail_t	*btp;		/* tail area of the block */
btp                78 fs/xfs/libxfs/xfs_dir2_sf.c 	btp = xfs_dir2_block_tail_p(geo, hdr);
btp                79 fs/xfs/libxfs/xfs_dir2_sf.c 	blp = xfs_dir2_block_leaf_p(btp);
btp                84 fs/xfs/libxfs/xfs_dir2_sf.c 	for (i = 0; i < be32_to_cpu(btp->count); i++) {
btp               569 fs/xfs/xfs_buf.c 	struct xfs_buftarg	*btp,
btp               588 fs/xfs/xfs_buf.c 	ASSERT(!(BBTOB(cmap.bm_len) < btp->bt_meta_sectorsize));
btp               589 fs/xfs/xfs_buf.c 	ASSERT(!(BBTOB(cmap.bm_bn) & (xfs_off_t)btp->bt_meta_sectormask));
btp               595 fs/xfs/xfs_buf.c 	eofs = XFS_FSB_TO_BB(btp->bt_mount, btp->bt_mount->m_sb.sb_dblocks);
btp               597 fs/xfs/xfs_buf.c 		xfs_alert(btp->bt_mount,
btp               604 fs/xfs/xfs_buf.c 	pag = xfs_perag_get(btp->bt_mount,
btp               605 fs/xfs/xfs_buf.c 			    xfs_daddr_to_agno(btp->bt_mount, cmap.bm_bn));
btp               617 fs/xfs/xfs_buf.c 		XFS_STATS_INC(btp->bt_mount, xb_miss_locked);
btp               638 fs/xfs/xfs_buf.c 			XFS_STATS_INC(btp->bt_mount, xb_busy_locked);
btp               642 fs/xfs/xfs_buf.c 		XFS_STATS_INC(btp->bt_mount, xb_get_locked_waited);
btp               658 fs/xfs/xfs_buf.c 	XFS_STATS_INC(btp->bt_mount, xb_get_locked);
btp              1588 fs/xfs/xfs_buf.c 	struct xfs_buftarg	*btp)
btp              1605 fs/xfs/xfs_buf.c 	while (percpu_counter_sum(&btp->bt_io_count))
btp              1607 fs/xfs/xfs_buf.c 	flush_workqueue(btp->bt_mount->m_buf_workqueue);
btp              1610 fs/xfs/xfs_buf.c 	while (list_lru_count(&btp->bt_lru)) {
btp              1611 fs/xfs/xfs_buf.c 		list_lru_walk(&btp->bt_lru, xfs_buftarg_wait_rele,
btp              1619 fs/xfs/xfs_buf.c 				xfs_alert(btp->bt_mount,
btp              1622 fs/xfs/xfs_buf.c 				xfs_alert(btp->bt_mount,
btp              1669 fs/xfs/xfs_buf.c 	struct xfs_buftarg	*btp = container_of(shrink,
btp              1674 fs/xfs/xfs_buf.c 	freed = list_lru_shrink_walk(&btp->bt_lru, sc,
btp              1692 fs/xfs/xfs_buf.c 	struct xfs_buftarg	*btp = container_of(shrink,
btp              1694 fs/xfs/xfs_buf.c 	return list_lru_shrink_count(&btp->bt_lru, sc);
btp              1699 fs/xfs/xfs_buf.c 	struct xfs_buftarg	*btp)
btp              1701 fs/xfs/xfs_buf.c 	unregister_shrinker(&btp->bt_shrinker);
btp              1702 fs/xfs/xfs_buf.c 	ASSERT(percpu_counter_sum(&btp->bt_io_count) == 0);
btp              1703 fs/xfs/xfs_buf.c 	percpu_counter_destroy(&btp->bt_io_count);
btp              1704 fs/xfs/xfs_buf.c 	list_lru_destroy(&btp->bt_lru);
btp              1706 fs/xfs/xfs_buf.c 	xfs_blkdev_issue_flush(btp);
btp              1708 fs/xfs/xfs_buf.c 	kmem_free(btp);
btp              1713 fs/xfs/xfs_buf.c 	xfs_buftarg_t		*btp,
btp              1717 fs/xfs/xfs_buf.c 	btp->bt_meta_sectorsize = sectorsize;
btp              1718 fs/xfs/xfs_buf.c 	btp->bt_meta_sectormask = sectorsize - 1;
btp              1720 fs/xfs/xfs_buf.c 	if (set_blocksize(btp->bt_bdev, sectorsize)) {
btp              1721 fs/xfs/xfs_buf.c 		xfs_warn(btp->bt_mount,
btp              1723 fs/xfs/xfs_buf.c 			sectorsize, btp->bt_bdev);
btp              1728 fs/xfs/xfs_buf.c 	btp->bt_logical_sectorsize = bdev_logical_block_size(btp->bt_bdev);
btp              1729 fs/xfs/xfs_buf.c 	btp->bt_logical_sectormask = bdev_logical_block_size(btp->bt_bdev) - 1;
btp              1741 fs/xfs/xfs_buf.c 	xfs_buftarg_t		*btp,
btp              1744 fs/xfs/xfs_buf.c 	return xfs_setsize_buftarg(btp, bdev_logical_block_size(bdev));
btp              1753 fs/xfs/xfs_buf.c 	xfs_buftarg_t		*btp;
btp              1755 fs/xfs/xfs_buf.c 	btp = kmem_zalloc(sizeof(*btp), KM_NOFS);
btp              1757 fs/xfs/xfs_buf.c 	btp->bt_mount = mp;
btp              1758 fs/xfs/xfs_buf.c 	btp->bt_dev =  bdev->bd_dev;
btp              1759 fs/xfs/xfs_buf.c 	btp->bt_bdev = bdev;
btp              1760 fs/xfs/xfs_buf.c 	btp->bt_daxdev = dax_dev;
btp              1762 fs/xfs/xfs_buf.c 	if (xfs_setsize_buftarg_early(btp, bdev))
btp              1765 fs/xfs/xfs_buf.c 	if (list_lru_init(&btp->bt_lru))
btp              1768 fs/xfs/xfs_buf.c 	if (percpu_counter_init(&btp->bt_io_count, 0, GFP_KERNEL))
btp              1771 fs/xfs/xfs_buf.c 	btp->bt_shrinker.count_objects = xfs_buftarg_shrink_count;
btp              1772 fs/xfs/xfs_buf.c 	btp->bt_shrinker.scan_objects = xfs_buftarg_shrink_scan;
btp              1773 fs/xfs/xfs_buf.c 	btp->bt_shrinker.seeks = DEFAULT_SEEKS;
btp              1774 fs/xfs/xfs_buf.c 	btp->bt_shrinker.flags = SHRINKER_NUMA_AWARE;
btp              1775 fs/xfs/xfs_buf.c 	if (register_shrinker(&btp->bt_shrinker))
btp              1777 fs/xfs/xfs_buf.c 	return btp;
btp              1780 fs/xfs/xfs_buf.c 	percpu_counter_destroy(&btp->bt_io_count);
btp              1782 fs/xfs/xfs_buf.c 	list_lru_destroy(&btp->bt_lru);
btp              1784 fs/xfs/xfs_buf.c 	kmem_free(btp);
btp               129 fs/xfs/xfs_pwork.c 	struct xfs_buftarg	*btp = mp->m_ddev_targp;
btp               135 fs/xfs/xfs_pwork.c 	return blk_queue_nonrot(btp->bt_bdev->bd_queue) ? 2 : 1;
btp               481 include/linux/trace_events.h int bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *prog);
btp               482 include/linux/trace_events.h int bpf_probe_unregister(struct bpf_raw_event_map *btp, struct bpf_prog *prog);
btp               484 include/linux/trace_events.h void bpf_put_raw_tracepoint(struct bpf_raw_event_map *btp);
btp               507 include/linux/trace_events.h static inline int bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *p)
btp               511 include/linux/trace_events.h static inline int bpf_probe_unregister(struct bpf_raw_event_map *btp, struct bpf_prog *p)
btp               519 include/linux/trace_events.h static inline void bpf_put_raw_tracepoint(struct bpf_raw_event_map *btp)
btp              1793 kernel/bpf/syscall.c 	struct bpf_raw_event_map *btp;
btp              1802 kernel/bpf/syscall.c 		bpf_probe_unregister(raw_tp->btp, raw_tp->prog);
btp              1805 kernel/bpf/syscall.c 	bpf_put_raw_tracepoint(raw_tp->btp);
btp              1821 kernel/bpf/syscall.c 	struct bpf_raw_event_map *btp;
btp              1831 kernel/bpf/syscall.c 	btp = bpf_get_raw_tracepoint(tp_name);
btp              1832 kernel/bpf/syscall.c 	if (!btp)
btp              1840 kernel/bpf/syscall.c 	raw_tp->btp = btp;
btp              1853 kernel/bpf/syscall.c 	err = bpf_probe_register(raw_tp->btp, prog);
btp              1861 kernel/bpf/syscall.c 		bpf_probe_unregister(raw_tp->btp, prog);
btp              1872 kernel/bpf/syscall.c 	bpf_put_raw_tracepoint(btp);
btp              2808 kernel/bpf/syscall.c 		struct bpf_raw_event_map *btp = raw_tp->btp;
btp              2813 kernel/bpf/syscall.c 					     btp->tp->name, 0, 0);
btp                36 kernel/trace/bpf_trace.c 	struct bpf_raw_event_map *btp, *ret = NULL;
btp                43 kernel/trace/bpf_trace.c 			btp = &btm->module->bpf_raw_events[i];
btp                44 kernel/trace/bpf_trace.c 			if (!strcmp(btp->tp->name, name)) {
btp                46 kernel/trace/bpf_trace.c 					ret = btp;
btp              1311 kernel/trace/bpf_trace.c 	struct bpf_raw_event_map *btp = __start__bpf_raw_tp;
btp              1313 kernel/trace/bpf_trace.c 	for (; btp < __stop__bpf_raw_tp; btp++) {
btp              1314 kernel/trace/bpf_trace.c 		if (!strcmp(btp->tp->name, name))
btp              1315 kernel/trace/bpf_trace.c 			return btp;
btp              1321 kernel/trace/bpf_trace.c void bpf_put_raw_tracepoint(struct bpf_raw_event_map *btp)
btp              1323 kernel/trace/bpf_trace.c 	struct module *mod = __module_address((unsigned long)btp);
btp              1384 kernel/trace/bpf_trace.c static int __bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *prog)
btp              1386 kernel/trace/bpf_trace.c 	struct tracepoint *tp = btp->tp;
btp              1392 kernel/trace/bpf_trace.c 	if (prog->aux->max_ctx_offset > btp->num_args * sizeof(u64))
btp              1395 kernel/trace/bpf_trace.c 	if (prog->aux->max_tp_access > btp->writable_size)
btp              1398 kernel/trace/bpf_trace.c 	return tracepoint_probe_register(tp, (void *)btp->bpf_func, prog);
btp              1401 kernel/trace/bpf_trace.c int bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *prog)
btp              1403 kernel/trace/bpf_trace.c 	return __bpf_probe_register(btp, prog);
btp              1406 kernel/trace/bpf_trace.c int bpf_probe_unregister(struct bpf_raw_event_map *btp, struct bpf_prog *prog)
btp              1408 kernel/trace/bpf_trace.c 	return tracepoint_probe_unregister(btp->tp, (void *)btp->bpf_func, prog);