rtp 508 drivers/gpu/drm/amd/amdgpu/atombios_dp.c int rtp = 0; rtp 513 drivers/gpu/drm/amd/amdgpu/atombios_dp.c rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN1; rtp 516 drivers/gpu/drm/amd/amdgpu/atombios_dp.c rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN2; rtp 519 drivers/gpu/drm/amd/amdgpu/atombios_dp.c rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN3; rtp 522 drivers/gpu/drm/amd/amdgpu/atombios_dp.c amdgpu_atombios_encoder_setup_dig_encoder(dp_info->encoder, rtp, 0); rtp 569 drivers/gpu/drm/radeon/atombios_dp.c int rtp = 0; rtp 575 drivers/gpu/drm/radeon/atombios_dp.c rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN1; rtp 578 drivers/gpu/drm/radeon/atombios_dp.c rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN2; rtp 581 drivers/gpu/drm/radeon/atombios_dp.c rtp = ATOM_ENCODER_CMD_DP_LINK_TRAINING_PATTERN3; rtp 584 drivers/gpu/drm/radeon/atombios_dp.c atombios_dig_encoder_setup(dp_info->encoder, rtp, 0); rtp 588 drivers/gpu/drm/radeon/atombios_dp.c rtp = 0; rtp 591 drivers/gpu/drm/radeon/atombios_dp.c rtp = 1; rtp 595 drivers/gpu/drm/radeon/atombios_dp.c dp_info->dp_clock, dp_info->enc_id, rtp); rtp 631 drivers/md/dm-raid.c static bool __got_raid10(struct raid_type *rtp, const int layout) rtp 633 drivers/md/dm-raid.c if (rtp->level == 10) { rtp 634 drivers/md/dm-raid.c switch (rtp->algorithm) { rtp 653 drivers/md/dm-raid.c struct raid_type *rtp = raid_types + ARRAY_SIZE(raid_types); rtp 655 drivers/md/dm-raid.c while (rtp-- > raid_types) rtp 656 drivers/md/dm-raid.c if (!strcasecmp(rtp->name, name)) rtp 657 drivers/md/dm-raid.c return rtp; rtp 665 drivers/md/dm-raid.c struct raid_type *rtp = raid_types + ARRAY_SIZE(raid_types); rtp 667 drivers/md/dm-raid.c while (rtp-- > raid_types) { rtp 669 drivers/md/dm-raid.c if (rtp->level == level && rtp 670 drivers/md/dm-raid.c (__got_raid10(rtp, layout) || rtp->algorithm == layout)) rtp 671 drivers/md/dm-raid.c return rtp; rtp 333 drivers/net/ethernet/amd/declance.c unsigned char *rtp; rtp 349 drivers/net/ethernet/amd/declance.c rtp = (unsigned char *)tp; rtp 352 drivers/net/ethernet/amd/declance.c *rtp++ = *rfp++; rtp 377 drivers/net/ethernet/amd/declance.c rtp = (unsigned char *)tp; rtp 380 drivers/net/ethernet/amd/declance.c *rtp++ = *rfp++; rtp 392 drivers/net/ethernet/amd/declance.c unsigned char *rtp; rtp 408 drivers/net/ethernet/amd/declance.c rtp = (unsigned char *)tp; rtp 412 drivers/net/ethernet/amd/declance.c *rtp++ = *rfp++; rtp 438 drivers/net/ethernet/amd/declance.c rtp = (unsigned char *)tp; rtp 441 drivers/net/ethernet/amd/declance.c *rtp++ = *rfp++; rtp 231 include/net/ip.h struct ipcm_cookie *ipc, struct rtable **rtp, rtp 255 lib/radix-tree.c struct radix_tree_preload *rtp; rtp 272 lib/radix-tree.c rtp = this_cpu_ptr(&radix_tree_preloads); rtp 273 lib/radix-tree.c if (rtp->nr) { rtp 274 lib/radix-tree.c ret = rtp->nodes; rtp 275 lib/radix-tree.c rtp->nodes = ret->parent; rtp 276 lib/radix-tree.c rtp->nr--; rtp 333 lib/radix-tree.c struct radix_tree_preload *rtp; rtp 344 lib/radix-tree.c rtp = this_cpu_ptr(&radix_tree_preloads); rtp 345 lib/radix-tree.c while (rtp->nr < nr) { rtp 351 lib/radix-tree.c rtp = this_cpu_ptr(&radix_tree_preloads); rtp 352 lib/radix-tree.c if (rtp->nr < nr) { rtp 353 lib/radix-tree.c node->parent = rtp->nodes; rtp 354 lib/radix-tree.c rtp->nodes = node; rtp 355 lib/radix-tree.c rtp->nr++; rtp 1589 lib/radix-tree.c struct radix_tree_preload *rtp; rtp 1593 lib/radix-tree.c rtp = &per_cpu(radix_tree_preloads, cpu); rtp 1594 lib/radix-tree.c while (rtp->nr) { rtp 1595 lib/radix-tree.c node = rtp->nodes; rtp 1596 lib/radix-tree.c rtp->nodes = node->parent; rtp 1598 lib/radix-tree.c rtp->nr--; rtp 183 net/decnet/dn_route.c struct dn_route __rcu **rtp; rtp 188 net/decnet/dn_route.c rtp = &dn_rt_hash_table[i].chain; rtp 191 net/decnet/dn_route.c while ((rt = rcu_dereference_protected(*rtp, rtp 195 net/decnet/dn_route.c rtp = &rt->dn_next; rtp 198 net/decnet/dn_route.c *rtp = rt->dn_next; rtp 215 net/decnet/dn_route.c struct dn_route __rcu **rtp; rtp 223 net/decnet/dn_route.c rtp = &dn_rt_hash_table[i].chain; rtp 225 net/decnet/dn_route.c while ((rt = rcu_dereference_protected(*rtp, rtp 229 net/decnet/dn_route.c rtp = &rt->dn_next; rtp 232 net/decnet/dn_route.c *rtp = rt->dn_next; rtp 146 net/ipv4/fib_semantics.c static void rt_fibinfo_free(struct rtable __rcu **rtp) rtp 148 net/ipv4/fib_semantics.c struct rtable *rt = rcu_dereference_protected(*rtp, 1); rtp 190 net/ipv4/fib_semantics.c static void rt_fibinfo_free_cpus(struct rtable __rcu * __percpu *rtp) rtp 194 net/ipv4/fib_semantics.c if (!rtp) rtp 200 net/ipv4/fib_semantics.c rt = rcu_dereference_protected(*per_cpu_ptr(rtp, cpu), 1); rtp 206 net/ipv4/fib_semantics.c free_percpu(rtp); rtp 1236 net/ipv4/ip_output.c struct ipcm_cookie *ipc, struct rtable **rtp) rtp 1241 net/ipv4/ip_output.c rt = *rtp; rtp 1271 net/ipv4/ip_output.c *rtp = NULL; rtp 1300 net/ipv4/ip_output.c struct ipcm_cookie *ipc, struct rtable **rtp, rtp 1310 net/ipv4/ip_output.c err = ip_setup_cork(sk, &inet->cork.base, ipc, rtp); rtp 1610 net/ipv4/ip_output.c struct ipcm_cookie *ipc, struct rtable **rtp, rtp 1624 net/ipv4/ip_output.c err = ip_setup_cork(sk, cork, ipc, rtp); rtp 345 net/ipv4/raw.c struct rtable **rtp, unsigned int flags, rtp 354 net/ipv4/raw.c struct rtable *rt = *rtp; rtp 381 net/ipv4/raw.c *rtp = NULL; rtp 6407 net/ipv6/addrconf.c struct rt6_info **rtp; rtp 6409 net/ipv6/addrconf.c rtp = per_cpu_ptr(nh->rt6i_pcpu, cpu); rtp 6410 net/ipv6/addrconf.c addrconf_set_nopolicy(*rtp, val); rtp 1827 net/ipv6/ip6_fib.c struct fib6_info __rcu **rtp, struct nl_info *info) rtp 1830 net/ipv6/ip6_fib.c struct fib6_info *rt = rcu_dereference_protected(*rtp, rtp 1837 net/ipv6/ip6_fib.c *rtp = rt->fib6_next; rtp 1903 net/ipv6/ip6_fib.c struct fib6_info __rcu **rtp; rtp 1915 net/ipv6/ip6_fib.c for (rtp = &fn->leaf; *rtp; rtp = rtp_next) { rtp 1916 net/ipv6/ip6_fib.c struct fib6_info *cur = rcu_dereference_protected(*rtp, rtp 1919 net/ipv6/ip6_fib.c fib6_del_route(table, fn, rtp, info);