rtp               508 drivers/gpu/drm/amd/amdgpu/atombios_dp.c 	int rtp = 0;
rtp               513 drivers/gpu/drm/amd/amdgpu/atombios_dp.c 		rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN1;
rtp               516 drivers/gpu/drm/amd/amdgpu/atombios_dp.c 		rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN2;
rtp               519 drivers/gpu/drm/amd/amdgpu/atombios_dp.c 		rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN3;
rtp               522 drivers/gpu/drm/amd/amdgpu/atombios_dp.c 	amdgpu_atombios_encoder_setup_dig_encoder(dp_info->encoder, rtp, 0);
rtp               569 drivers/gpu/drm/radeon/atombios_dp.c 	int rtp = 0;
rtp               575 drivers/gpu/drm/radeon/atombios_dp.c 			rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN1;
rtp               578 drivers/gpu/drm/radeon/atombios_dp.c 			rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN2;
rtp               581 drivers/gpu/drm/radeon/atombios_dp.c 			rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN3;
rtp               584 drivers/gpu/drm/radeon/atombios_dp.c 		atombios_dig_encoder_setup(dp_info->encoder, rtp, 0);
rtp               588 drivers/gpu/drm/radeon/atombios_dp.c 			rtp = 0;
rtp               591 drivers/gpu/drm/radeon/atombios_dp.c 			rtp = 1;
rtp               595 drivers/gpu/drm/radeon/atombios_dp.c 					  dp_info->dp_clock, dp_info->enc_id, rtp);
rtp               631 drivers/md/dm-raid.c static bool __got_raid10(struct raid_type *rtp, const int layout)
rtp               633 drivers/md/dm-raid.c 	if (rtp->level == 10) {
rtp               634 drivers/md/dm-raid.c 		switch (rtp->algorithm) {
rtp               653 drivers/md/dm-raid.c 	struct raid_type *rtp = raid_types + ARRAY_SIZE(raid_types);
rtp               655 drivers/md/dm-raid.c 	while (rtp-- > raid_types)
rtp               656 drivers/md/dm-raid.c 		if (!strcasecmp(rtp->name, name))
rtp               657 drivers/md/dm-raid.c 			return rtp;
rtp               665 drivers/md/dm-raid.c 	struct raid_type *rtp = raid_types + ARRAY_SIZE(raid_types);
rtp               667 drivers/md/dm-raid.c 	while (rtp-- > raid_types) {
rtp               669 drivers/md/dm-raid.c 		if (rtp->level == level &&
rtp               670 drivers/md/dm-raid.c 		    (__got_raid10(rtp, layout) || rtp->algorithm == layout))
rtp               671 drivers/md/dm-raid.c 			return rtp;
rtp               333 drivers/net/ethernet/amd/declance.c 	unsigned char *rtp;
rtp               349 drivers/net/ethernet/amd/declance.c 		rtp = (unsigned char *)tp;
rtp               352 drivers/net/ethernet/amd/declance.c 			*rtp++ = *rfp++;
rtp               377 drivers/net/ethernet/amd/declance.c 		rtp = (unsigned char *)tp;
rtp               380 drivers/net/ethernet/amd/declance.c 			*rtp++ = *rfp++;
rtp               392 drivers/net/ethernet/amd/declance.c 	unsigned char *rtp;
rtp               408 drivers/net/ethernet/amd/declance.c 		rtp = (unsigned char *)tp;
rtp               412 drivers/net/ethernet/amd/declance.c 			*rtp++ = *rfp++;
rtp               438 drivers/net/ethernet/amd/declance.c 		rtp = (unsigned char *)tp;
rtp               441 drivers/net/ethernet/amd/declance.c 			*rtp++ = *rfp++;
rtp               231 include/net/ip.h 			    struct ipcm_cookie *ipc, struct rtable **rtp,
rtp               255 lib/radix-tree.c 		struct radix_tree_preload *rtp;
rtp               272 lib/radix-tree.c 		rtp = this_cpu_ptr(&radix_tree_preloads);
rtp               273 lib/radix-tree.c 		if (rtp->nr) {
rtp               274 lib/radix-tree.c 			ret = rtp->nodes;
rtp               275 lib/radix-tree.c 			rtp->nodes = ret->parent;
rtp               276 lib/radix-tree.c 			rtp->nr--;
rtp               333 lib/radix-tree.c 	struct radix_tree_preload *rtp;
rtp               344 lib/radix-tree.c 	rtp = this_cpu_ptr(&radix_tree_preloads);
rtp               345 lib/radix-tree.c 	while (rtp->nr < nr) {
rtp               351 lib/radix-tree.c 		rtp = this_cpu_ptr(&radix_tree_preloads);
rtp               352 lib/radix-tree.c 		if (rtp->nr < nr) {
rtp               353 lib/radix-tree.c 			node->parent = rtp->nodes;
rtp               354 lib/radix-tree.c 			rtp->nodes = node;
rtp               355 lib/radix-tree.c 			rtp->nr++;
rtp              1589 lib/radix-tree.c 	struct radix_tree_preload *rtp;
rtp              1593 lib/radix-tree.c 	rtp = &per_cpu(radix_tree_preloads, cpu);
rtp              1594 lib/radix-tree.c 	while (rtp->nr) {
rtp              1595 lib/radix-tree.c 		node = rtp->nodes;
rtp              1596 lib/radix-tree.c 		rtp->nodes = node->parent;
rtp              1598 lib/radix-tree.c 		rtp->nr--;
rtp               183 net/decnet/dn_route.c 	struct dn_route __rcu **rtp;
rtp               188 net/decnet/dn_route.c 		rtp = &dn_rt_hash_table[i].chain;
rtp               191 net/decnet/dn_route.c 		while ((rt = rcu_dereference_protected(*rtp,
rtp               195 net/decnet/dn_route.c 				rtp = &rt->dn_next;
rtp               198 net/decnet/dn_route.c 			*rtp = rt->dn_next;
rtp               215 net/decnet/dn_route.c 	struct dn_route __rcu **rtp;
rtp               223 net/decnet/dn_route.c 		rtp = &dn_rt_hash_table[i].chain;
rtp               225 net/decnet/dn_route.c 		while ((rt = rcu_dereference_protected(*rtp,
rtp               229 net/decnet/dn_route.c 				rtp = &rt->dn_next;
rtp               232 net/decnet/dn_route.c 			*rtp = rt->dn_next;
rtp               146 net/ipv4/fib_semantics.c static void rt_fibinfo_free(struct rtable __rcu **rtp)
rtp               148 net/ipv4/fib_semantics.c 	struct rtable *rt = rcu_dereference_protected(*rtp, 1);
rtp               190 net/ipv4/fib_semantics.c static void rt_fibinfo_free_cpus(struct rtable __rcu * __percpu *rtp)
rtp               194 net/ipv4/fib_semantics.c 	if (!rtp)
rtp               200 net/ipv4/fib_semantics.c 		rt = rcu_dereference_protected(*per_cpu_ptr(rtp, cpu), 1);
rtp               206 net/ipv4/fib_semantics.c 	free_percpu(rtp);
rtp              1236 net/ipv4/ip_output.c 			 struct ipcm_cookie *ipc, struct rtable **rtp)
rtp              1241 net/ipv4/ip_output.c 	rt = *rtp;
rtp              1271 net/ipv4/ip_output.c 	*rtp = NULL;
rtp              1300 net/ipv4/ip_output.c 		   struct ipcm_cookie *ipc, struct rtable **rtp,
rtp              1310 net/ipv4/ip_output.c 		err = ip_setup_cork(sk, &inet->cork.base, ipc, rtp);
rtp              1610 net/ipv4/ip_output.c 			    struct ipcm_cookie *ipc, struct rtable **rtp,
rtp              1624 net/ipv4/ip_output.c 	err = ip_setup_cork(sk, cork, ipc, rtp);
rtp               345 net/ipv4/raw.c 			   struct rtable **rtp, unsigned int flags,
rtp               354 net/ipv4/raw.c 	struct rtable *rt = *rtp;
rtp               381 net/ipv4/raw.c 	*rtp = NULL;
rtp              6407 net/ipv6/addrconf.c 					struct rt6_info **rtp;
rtp              6409 net/ipv6/addrconf.c 					rtp = per_cpu_ptr(nh->rt6i_pcpu, cpu);
rtp              6410 net/ipv6/addrconf.c 					addrconf_set_nopolicy(*rtp, val);
rtp              1827 net/ipv6/ip6_fib.c 			   struct fib6_info __rcu **rtp, struct nl_info *info)
rtp              1830 net/ipv6/ip6_fib.c 	struct fib6_info *rt = rcu_dereference_protected(*rtp,
rtp              1837 net/ipv6/ip6_fib.c 	*rtp = rt->fib6_next;
rtp              1903 net/ipv6/ip6_fib.c 	struct fib6_info __rcu **rtp;
rtp              1915 net/ipv6/ip6_fib.c 	for (rtp = &fn->leaf; *rtp; rtp = rtp_next) {
rtp              1916 net/ipv6/ip6_fib.c 		struct fib6_info *cur = rcu_dereference_protected(*rtp,
rtp              1919 net/ipv6/ip6_fib.c 			fib6_del_route(table, fn, rtp, info);