/linux-4.1.27/net/core/ |
D | skbuff.c | 258 shinfo = skb_shinfo(skb); in __alloc_skb() 326 shinfo = skb_shinfo(skb); in __build_skb() 586 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag() 603 skb_drop_list(&skb_shinfo(skb)->frag_list); in skb_drop_fraglist() 624 struct skb_shared_info *shinfo = skb_shinfo(skb); in skb_release_data() 768 if (skb_shinfo(skb)->tx_flags & SKBTX_DEV_ZEROCOPY) { in skb_tx_error() 771 uarg = skb_shinfo(skb)->destructor_arg; in skb_tx_error() 774 skb_shinfo(skb)->tx_flags &= ~SKBTX_DEV_ZEROCOPY; in skb_tx_error() 887 atomic_inc(&(skb_shinfo(skb)->dataref)); in __skb_clone() 929 int num_frags = skb_shinfo(skb)->nr_frags; in skb_copy_ubufs() [all …]
|
D | tso.c | 10 return skb_shinfo(skb)->gso_segs * 2 + skb_shinfo(skb)->nr_frags; in tso_count_descs() 46 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_build_data() 47 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data() 69 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_start() 70 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
|
D | datagram.c | 376 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_datagram_iter() 378 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_iter() 463 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_datagram_from_iter() 465 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_datagram_from_iter() 592 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skb_copy_and_csum_datagram() 594 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_copy_and_csum_datagram()
|
D | dev.c | 2358 skb->len, skb->data_len, skb_shinfo(skb)->gso_size, in skb_warn_bad_offload() 2359 skb_shinfo(skb)->gso_type, skb->ip_summed); in skb_warn_bad_offload() 2374 if (unlikely(skb_shinfo(skb)->gso_size)) { in skb_checksum_help() 2533 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in illegal_highdma() 2534 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma() 2545 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in illegal_highdma() 2546 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in illegal_highdma() 2616 u16 gso_segs = skb_shinfo(skb)->gso_segs; in netif_skb_features() 2779 const struct skb_shared_info *shinfo = skb_shinfo(skb); in qdisc_pkt_len_init() 2940 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_SCHED_TSTAMP)) in __dev_queue_xmit() [all …]
|
D | pktgen.c | 2692 skb_shinfo(skb)->frags[i].page_offset = 0; in pktgen_finalize_skb() 2695 skb_frag_size_set(&skb_shinfo(skb)->frags[i], in pktgen_finalize_skb() 2698 skb_frag_size_set(&skb_shinfo(skb)->frags[i], frag_len); in pktgen_finalize_skb() 2699 datalen -= skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb() 2700 skb->len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb() 2701 skb->data_len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in pktgen_finalize_skb() 2703 skb_shinfo(skb)->nr_frags = i; in pktgen_finalize_skb() 3190 int nr_frags = pkt_dev->skb ? skb_shinfo(pkt_dev->skb)->nr_frags : -1; in pktgen_stop_device()
|
/linux-4.1.27/net/ipv6/ |
D | udp_offload.c | 33 mss = skb_shinfo(skb)->gso_size; in udp6_ufo_fragment() 39 int type = skb_shinfo(skb)->gso_type; in udp6_ufo_fragment() 53 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in udp6_ufo_fragment() 56 if (!skb_shinfo(skb)->ip6_frag_id) in udp6_ufo_fragment() 63 if (skb->encapsulation && skb_shinfo(skb)->gso_type & in udp6_ufo_fragment() 115 if (!skb_shinfo(skb)->ip6_frag_id) in udp6_ufo_fragment() 117 fptr->identification = skb_shinfo(skb)->ip6_frag_id; in udp6_ufo_fragment() 163 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL_CSUM; in udp6_gro_complete() 167 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL; in udp6_gro_complete()
|
D | reassembly.c | 441 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in ip6_frag_reasm() 443 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in ip6_frag_reasm() 444 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in ip6_frag_reasm() 480 if (!skb_shinfo(head)->frag_list) in ip6_frag_reasm() 481 skb_shinfo(head)->frag_list = fp; in ip6_frag_reasm()
|
D | tcpv6_offload.c | 39 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV6; in tcp6_gro_complete()
|
D | ip6_output.c | 620 frag = skb_shinfo(skb)->frag_list; in ip6_fragment() 1113 skb_shinfo(skb)->gso_size = (mtu - fragheaderlen - in ip6_ufo_append_data() 1115 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in ip6_ufo_append_data() 1117 skb_shinfo(skb)->ip6_frag_id = fhdr.identification; in ip6_ufo_append_data() 1438 skb_shinfo(skb)->tx_flags = tx_flags; in __ip6_append_data() 1440 skb_shinfo(skb)->tskey = tskey; in __ip6_append_data() 1499 int i = skb_shinfo(skb)->nr_frags; in __ip6_append_data() 1513 skb_shinfo(skb)->nr_frags = ++i; in __ip6_append_data() 1523 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in __ip6_append_data() 1619 tail_skb = &(skb_shinfo(skb)->frag_list); in __ip6_make_skb()
|
D | ip6_offload.c | 71 if (unlikely(skb_shinfo(skb)->gso_type & in ipv6_gso_segment() 104 skb_shinfo(skb)->gso_type & (SKB_GSO_SIT|SKB_GSO_IPIP)) in ipv6_gso_segment()
|
D | output_core.c | 59 skb_shinfo(skb)->ip6_frag_id = htonl(id); in ipv6_proxy_select_ident()
|
D | ip6_input.c | 114 max_t(unsigned short, 1, skb_shinfo(skb)->gso_segs)); in ipv6_rcv()
|
D | udp.c | 990 struct sk_buff *frags = skb_shinfo(skb)->frag_list; in udp6_hwcsum_outgoing()
|
/linux-4.1.27/net/ipv4/ |
D | tcp_offload.c | 22 skb_shinfo(skb)->tx_flags |= SKBTX_SW_TSTAMP; in tcp_gso_tstamp() 23 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp() 86 int type = skb_shinfo(skb)->gso_type; in tcp_gso_segment() 104 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in tcp_gso_segment() 128 if (unlikely(skb_shinfo(gso_skb)->tx_flags & SKBTX_SW_TSTAMP)) in tcp_gso_segment() 129 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment() 282 skb_shinfo(skb)->gso_segs = NAPI_GRO_CB(skb)->count; in tcp_gro_complete() 285 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tcp_gro_complete() 311 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV4; in tcp4_gro_complete()
|
D | udp_offload.c | 42 bool need_csum = !!(skb_shinfo(skb)->gso_type & in __skb_udp_tunnel_segment() 44 bool remcsum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_TUNNEL_REMCSUM); in __skb_udp_tunnel_segment() 188 (skb_shinfo(skb)->gso_type & in udp4_ufo_fragment() 197 mss = skb_shinfo(skb)->gso_size; in udp4_ufo_fragment() 203 int type = skb_shinfo(skb)->gso_type; in udp4_ufo_fragment() 214 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in udp4_ufo_fragment() 407 skb_shinfo(skb)->gso_type |= SKB_GSO_TUNNEL_REMCSUM; in udp_gro_complete() 421 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL_CSUM; in udp4_gro_complete() 425 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL; in udp4_gro_complete()
|
D | ip_output.c | 441 skb_shinfo(skb)->gso_segs ?: 1); in ip_queue_xmit() 573 frag = skb_shinfo(skb)->frag_list; in ip_fragment() 830 skb_shinfo(skb)->gso_size = maxfraglen - fragheaderlen; in ip_ufo_append_data() 831 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in ip_ufo_append_data() 985 skb_shinfo(skb)->tx_flags = cork->tx_flags; in __ip_append_data() 987 skb_shinfo(skb)->tskey = tskey; in __ip_append_data() 1043 int i = skb_shinfo(skb)->nr_frags; in __ip_append_data() 1057 skb_shinfo(skb)->nr_frags = ++i; in __ip_append_data() 1067 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in __ip_append_data() 1216 skb_shinfo(skb)->gso_size = mtu - fragheaderlen; in ip_append_page() [all …]
|
D | inet_lro.c | 154 nr_frags = skb_shinfo(skb)->nr_frags; in lro_init_desc() 156 lro_desc->next_frag = &(skb_shinfo(skb)->frags[nr_frags]); in lro_init_desc() 228 skb_shinfo(parent)->frag_list = skb; in lro_add_packet() 283 skb_shinfo(lro_desc->parent)->gso_size = lro_desc->mss; in lro_flush()
|
D | ip_tunnel_core.c | 78 skb_shinfo(skb)->gso_segs ?: 1); in iptunnel_xmit() 136 skb_shinfo(skb)->gso_type |= gso_type_mask; in iptunnel_handle_offloads()
|
D | ip_fragment.c | 567 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in ip_frag_reasm() 569 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in ip_frag_reasm() 570 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in ip_frag_reasm() 596 if (!skb_shinfo(head)->frag_list) in ip_frag_reasm() 597 skb_shinfo(head)->frag_list = fp; in ip_frag_reasm()
|
D | gre_offload.c | 31 if (unlikely(skb_shinfo(skb)->gso_type & in gre_gso_segment() 236 skb_shinfo(skb)->gso_type = SKB_GSO_GRE; in gre_gro_complete()
|
D | tcp_output.c | 380 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tcp_ecn_send() 396 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_init_nondata_skb() 1022 skb_shinfo(skb)->gso_segs = tcp_skb_pcount(skb); in tcp_transmit_skb() 1062 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_set_skb_tso_segs() 1128 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_fragment_tstamp() 1132 struct skb_shared_info *shinfo2 = skb_shinfo(skb2); in tcp_fragment_tstamp() 1187 if (!skb_shinfo(skb)->nr_frags && skb->ip_summed != CHECKSUM_PARTIAL) { in tcp_fragment() 1248 shinfo = skb_shinfo(skb); in __pskb_trim_head() 1980 if (!skb_shinfo(skb)->nr_frags) { in tcp_mtu_probe() 2503 if (skb_shinfo(skb)->nr_frags != 0) in tcp_can_collapse()
|
D | tcp.c | 434 struct skb_shared_info *shinfo = skb_shinfo(skb); in tcp_tx_timestamp() 923 i = skb_shinfo(skb)->nr_frags; in do_tcp_sendpages() 933 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in do_tcp_sendpages() 938 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG; in do_tcp_sendpages() 1183 int i = skb_shinfo(skb)->nr_frags; in tcp_sendmsg() 1212 skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy); in tcp_sendmsg() 2955 const struct skb_shared_info *shi = skb_shinfo(skb); in tcp_md5_hash_skb_data()
|
D | gre_demux.c | 88 !(skb_shinfo(skb)->gso_type & in gre_build_header()
|
D | ip_input.c | 421 max_t(unsigned short, 1, skb_shinfo(skb)->gso_segs)); in ip_rcv()
|
D | tcp_input.c | 141 len = skb_shinfo(skb)->gso_size ? : skb->len; in tcp_measure_rcv_mss() 1319 if (!skb_shinfo(prev)->gso_size) { in tcp_shifted_skb() 1320 skb_shinfo(prev)->gso_size = mss; in tcp_shifted_skb() 1321 skb_shinfo(prev)->gso_type = sk->sk_gso_type; in tcp_shifted_skb() 1326 skb_shinfo(skb)->gso_size = 0; in tcp_shifted_skb() 1327 skb_shinfo(skb)->gso_type = 0; in tcp_shifted_skb() 2258 mss = skb_shinfo(skb)->gso_size; in tcp_mark_head_lost() 3047 shinfo = skb_shinfo(skb); in tcp_ack_tstamp()
|
D | af_inet.c | 1215 if (unlikely(skb_shinfo(skb)->gso_type & in inet_gso_segment() 1259 skb_shinfo(skb)->gso_type & (SKB_GSO_SIT|SKB_GSO_IPIP)) in inet_gso_segment()
|
D | raw.c | 380 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in raw_send_hdrinc()
|
/linux-4.1.27/drivers/net/xen-netback/ |
D | netback.c | 340 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) in xenvif_gop_frag_copy() 342 else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in xenvif_gop_frag_copy() 371 int nr_frags = skb_shinfo(skb)->nr_frags; in xenvif_gop_skb() 384 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) in xenvif_gop_skb() 386 else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in xenvif_gop_skb() 395 meta->gso_size = skb_shinfo(skb)->gso_size; in xenvif_gop_skb() 405 meta->gso_size = skb_shinfo(skb)->gso_size; in xenvif_gop_skb() 431 skb_frag_page(&skb_shinfo(skb)->frags[i]), in xenvif_gop_skb() 432 skb_frag_size(&skb_shinfo(skb)->frags[i]), in xenvif_gop_skb() 433 skb_shinfo(skb)->frags[i].page_offset, in xenvif_gop_skb() [all …]
|
D | interface.c | 57 skb_shinfo(skb)->tx_flags |= SKBTX_DEV_ZEROCOPY; in xenvif_skb_zerocopy_prepare()
|
/linux-4.1.27/include/trace/events/ |
D | net.h | 54 __entry->tx_flags = skb_shinfo(skb)->tx_flags; 55 __entry->gso_size = skb_shinfo(skb)->gso_size; 56 __entry->gso_segs = skb_shinfo(skb)->gso_segs; 57 __entry->gso_type = skb_shinfo(skb)->gso_type; 189 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 190 __entry->gso_size = skb_shinfo(skb)->gso_size; 191 __entry->gso_type = skb_shinfo(skb)->gso_type;
|
/linux-4.1.27/include/linux/ |
D | skbuff.h | 987 #define skb_shinfo(SKB) ((struct skb_shared_info *)(skb_end_pointer(SKB))) macro 991 return &skb_shinfo(skb)->hwtstamps; in skb_hwtstamps() 1096 (atomic_read(&skb_shinfo(skb)->dataref) & SKB_DATAREF_MASK) != 1; in skb_cloned() 1123 dataref = atomic_read(&skb_shinfo(skb)->dataref); in skb_header_cloned() 1141 atomic_add(1 << SKB_DATAREF_SHIFT, &skb_shinfo(skb)->dataref); in skb_header_release() 1154 atomic_set(&skb_shinfo(skb)->dataref, 1 + (1 << SKB_DATAREF_SHIFT)); in __skb_header_release() 1570 for (i = (int)skb_shinfo(skb)->nr_frags - 1; i >= 0; i--) in skb_pagelen() 1571 len += skb_frag_size(&skb_shinfo(skb)->frags[i]); in skb_pagelen() 1591 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in __skb_fill_page_desc() 1625 skb_shinfo(skb)->nr_frags = i + 1; in skb_fill_page_desc() [all …]
|
D | netdevice.h | 3715 return net_gso_ok(features, skb_shinfo(skb)->gso_type) && in skb_gso_ok()
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
D | tx.c | 256 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in efx_skb_copy_bits_to_pio() 257 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in efx_skb_copy_bits_to_pio() 267 EFX_BUG_ON_PARANOID(skb_shinfo(skb)->frag_list); in efx_skb_copy_bits_to_pio() 282 if (skb_shinfo(skb)->nr_frags) { in efx_enqueue_skb_pio() 346 if (skb_shinfo(skb)->gso_size) in efx_enqueue_skb() 411 if (i >= skb_shinfo(skb)->nr_frags) in efx_enqueue_skb() 413 fragment = &skb_shinfo(skb)->frags[i]; in efx_enqueue_skb() 457 skb_shinfo(skb)->nr_frags + 1); in efx_enqueue_skb() 1151 bool is_last = st->out_len <= skb_shinfo(skb)->gso_size; in tso_start_new_packet() 1155 st->packet_space = skb_shinfo(skb)->gso_size; in tso_start_new_packet() [all …]
|
D | rx.c | 450 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in efx_rx_packet_gro() 455 if (skb_shinfo(skb)->nr_frags == n_frags) in efx_rx_packet_gro() 503 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in efx_rx_mk_skb() 509 if (skb_shinfo(skb)->nr_frags == n_frags) in efx_rx_mk_skb()
|
D | net_driver.h | 1499 return skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP; in efx_xmit_with_hwtstamp() 1503 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in efx_xmit_hwtstamp_pending()
|
D | ptp.c | 880 if (skb_shinfo(skb)->nr_frags != 0) { in efx_ptp_xmit_skb()
|
/linux-4.1.27/drivers/staging/octeon/ |
D | ethernet-tx.c | 220 if (unlikely(skb_shinfo(skb)->nr_frags > 5)) { in cvm_oct_xmit() 290 if (skb_shinfo(skb)->nr_frags == 0) { in cvm_oct_xmit() 299 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in cvm_oct_xmit() 300 struct skb_frag_struct *fs = skb_shinfo(skb)->frags + i; in cvm_oct_xmit() 309 hw_buffer.s.size = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit() 310 pko_command.s.segs = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit() 365 if (unlikely(skb_shinfo(skb)->nr_frags)) { in cvm_oct_xmit()
|
/linux-4.1.27/net/xfrm/ |
D | xfrm_ipcomp.c | 76 if (WARN_ON(skb_shinfo(skb)->nr_frags >= MAX_SKB_FRAGS)) in ipcomp_decompress() 79 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; in ipcomp_decompress() 100 skb_shinfo(skb)->nr_frags++; in ipcomp_decompress()
|
/linux-4.1.27/drivers/net/ethernet/sun/ |
D | sunvnet.c | 1020 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_map() 1021 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_map() 1060 docopy = skb_shinfo(skb)->nr_frags >= ncookies; in vnet_skb_shape() 1061 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_shape() 1062 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_shape() 1123 skb_shinfo(nskb)->gso_size = skb_shinfo(skb)->gso_size; in vnet_skb_shape() 1124 skb_shinfo(nskb)->gso_type = skb_shinfo(skb)->gso_type; in vnet_skb_shape() 1174 gso_size = skb_shinfo(skb)->gso_size; in vnet_handle_offloads() 1175 gso_type = skb_shinfo(skb)->gso_type; in vnet_handle_offloads() 1176 gso_segs = skb_shinfo(skb)->gso_segs; in vnet_handle_offloads() [all …]
|
D | sungem.c | 661 if (skb_shinfo(skb)->nr_frags) { in gem_tx() 662 int last = entry + skb_shinfo(skb)->nr_frags; in gem_tx() 680 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in gem_tx() 1015 if (unlikely(TX_BUFFS_AVAIL(gp) <= (skb_shinfo(skb)->nr_frags + 1))) { in gem_start_xmit() 1027 if (skb_shinfo(skb)->nr_frags == 0) { in gem_start_xmit() 1064 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in gem_start_xmit() 1065 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in gem_start_xmit() 1074 if (frag == skb_shinfo(skb)->nr_frags - 1) in gem_start_xmit() 1600 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in gem_clean_rings() 1609 if (frag != skb_shinfo(skb)->nr_frags) in gem_clean_rings()
|
D | sunhme.c | 1225 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_clean_rings() 1239 if (frag != skb_shinfo(skb)->nr_frags) in happy_meal_clean_rings() 1939 if (skb_shinfo(skb)->nr_frags) { in happy_meal_tx() 1942 last = elem + skb_shinfo(skb)->nr_frags; in happy_meal_tx() 1951 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_tx() 2305 if (TX_BUFFS_AVAIL(hp) <= (skb_shinfo(skb)->nr_frags + 1)) { in happy_meal_start_xmit() 2317 if (skb_shinfo(skb)->nr_frags == 0) { in happy_meal_start_xmit() 2343 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in happy_meal_start_xmit() 2344 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in happy_meal_start_xmit() 2356 if (frag == skb_shinfo(skb)->nr_frags - 1) in happy_meal_start_xmit()
|
D | cassini.c | 1880 count -= skb_shinfo(skb)->nr_frags + in cas_tx_ringN() 1891 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in cas_tx_ringN() 2005 skb_frag_t *frag = skb_shinfo(skb)->frags; in cas_rx_process_pkt() 2041 skb_shinfo(skb)->nr_frags++; in cas_rx_process_pkt() 2065 skb_shinfo(skb)->nr_frags++; in cas_rx_process_pkt() 2778 CAS_TABORT(cp)*(skb_shinfo(skb)->nr_frags + 1)) { in cas_xmit_tx_ringN() 2798 nr_frags = skb_shinfo(skb)->nr_frags; in cas_xmit_tx_ringN() 2824 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; in cas_xmit_tx_ringN() 3911 for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { in cas_clean_txd() 3923 if (frag != skb_shinfo(skb)->nr_frags) { in cas_clean_txd()
|
D | niu.c | 3280 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, offset, size); in niu_rx_skb_append() 3586 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in release_tx_packet() 3590 skb_frag_size(&skb_shinfo(skb)->frags[i]), in release_tx_packet() 6646 if (niu_tx_avail(rp) <= (skb_shinfo(skb)->nr_frags + 1)) { in niu_start_xmit() 6697 nfg = skb_shinfo(skb)->nr_frags; in niu_start_xmit() 6717 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in niu_start_xmit() 6718 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in niu_start_xmit()
|
/linux-4.1.27/net/ieee802154/6lowpan/ |
D | reassembly.c | 282 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in lowpan_frag_reasm() 284 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in lowpan_frag_reasm() 285 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in lowpan_frag_reasm() 305 if (!skb_shinfo(head)->frag_list) in lowpan_frag_reasm() 306 skb_shinfo(head)->frag_list = fp; in lowpan_frag_reasm()
|
/linux-4.1.27/drivers/net/ |
D | xen-netfront.c | 311 pfn = page_to_pfn(skb_frag_page(&skb_shinfo(skb)->frags[0])); in xennet_alloc_rx_buffers() 477 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_slots() 483 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_slots() 583 if (skb_shinfo(skb)->gso_size) { in xennet_start_xmit() 591 gso->u.gso.size = skb_shinfo(skb)->gso_size; in xennet_start_xmit() 592 gso->u.gso.type = (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) ? in xennet_start_xmit() 606 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xennet_start_xmit() 607 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xennet_start_xmit() 808 skb_shinfo(skb)->gso_size = gso->u.gso.size; in xennet_set_skb_gso() 809 skb_shinfo(skb)->gso_type = in xennet_set_skb_gso() [all …]
|
D | virtio_net.c | 301 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, offset, in page_to_skb() 379 num_skb_frags = skb_shinfo(curr_skb)->nr_frags; in receive_mergeable() 386 skb_shinfo(curr_skb)->frag_list = nskb; in receive_mergeable() 490 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in receive_buf() 493 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in receive_buf() 496 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in receive_buf() 505 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in receive_buf() 507 skb_shinfo(skb)->gso_size = virtio16_to_cpu(vi->vdev, in receive_buf() 509 if (skb_shinfo(skb)->gso_size == 0) { in receive_buf() 515 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in receive_buf() [all …]
|
D | macvtap.c | 608 skb_shinfo(skb)->gso_size = macvtap16_to_cpu(q, vnet_hdr->gso_size); in macvtap_skb_from_vnet_hdr() 609 skb_shinfo(skb)->gso_type = gso_type; in macvtap_skb_from_vnet_hdr() 612 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in macvtap_skb_from_vnet_hdr() 613 skb_shinfo(skb)->gso_segs = 0; in macvtap_skb_from_vnet_hdr() 625 struct skb_shared_info *sinfo = skb_shinfo(skb); in macvtap_skb_to_vnet_hdr() 762 skb_shinfo(skb)->destructor_arg = m->msg_control; in macvtap_get_user() 763 skb_shinfo(skb)->tx_flags |= SKBTX_DEV_ZEROCOPY; in macvtap_get_user() 764 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG; in macvtap_get_user()
|
D | tun.c | 811 sock_tx_timestamp(skb->sk, &skb_shinfo(skb)->tx_flags); in tun_net_xmit() 1169 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in tun_get_user() 1172 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in tun_get_user() 1175 skb_shinfo(skb)->gso_type = SKB_GSO_UDP; in tun_get_user() 1184 skb_shinfo(skb)->gso_type |= SKB_GSO_TCP_ECN; in tun_get_user() 1186 skb_shinfo(skb)->gso_size = tun16_to_cpu(tun, gso.gso_size); in tun_get_user() 1187 if (skb_shinfo(skb)->gso_size == 0) { in tun_get_user() 1194 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in tun_get_user() 1195 skb_shinfo(skb)->gso_segs = 0; in tun_get_user() 1200 skb_shinfo(skb)->destructor_arg = msg_control; in tun_get_user() [all …]
|
/linux-4.1.27/net/ipv6/netfilter/ |
D | nf_conntrack_reasm.c | 423 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in nf_ct_frag6_reasm() 425 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in nf_ct_frag6_reasm() 426 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in nf_ct_frag6_reasm() 445 skb_shinfo(head)->frag_list = head->next; in nf_ct_frag6_reasm() 478 fp = skb_shinfo(head)->frag_list; in nf_ct_frag6_reasm()
|
/linux-4.1.27/drivers/net/ethernet/neterion/vxge/ |
D | vxge-main.h | 515 #define vxge_tcp_mss(skb) (skb_shinfo(skb)->gso_size) 516 #define vxge_udp_mss(skb) (skb_shinfo(skb)->gso_size) 517 #define vxge_offload_type(skb) (skb_shinfo(skb)->gso_type)
|
D | vxge-main.c | 565 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_xmit_compl() 566 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit_compl() 913 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_xmit() 923 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit() 978 frag = &skb_shinfo(skb)->frags[0]; in vxge_xmit() 1047 frg_cnt = skb_shinfo(skb)->nr_frags; in vxge_tx_term() 1048 frag = &skb_shinfo(skb)->frags[0]; in vxge_tx_term()
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
D | ipoib_ib.c | 284 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in ipoib_dma_map_tx() 285 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_map_tx() 297 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in ipoib_dma_map_tx() 322 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) { in ipoib_dma_unmap_tx() 323 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ipoib_dma_unmap_tx() 519 skb_frag_t *frags = skb_shinfo(skb)->frags; in post_send() 520 int nr_frags = skb_shinfo(skb)->nr_frags; in post_send() 540 priv->tx_wr.wr.ud.mss = skb_shinfo(skb)->gso_size; in post_send()
|
D | ipoib_cm.c | 537 num_frags = skb_shinfo(skb)->nr_frags; in skb_put_frags() 539 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_put_frags() 545 --skb_shinfo(skb)->nr_frags; in skb_put_frags()
|
/linux-4.1.27/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_main.c | 293 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in fm10k_add_rx_frag() 533 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in fm10k_pull_tail() 822 first->gso_segs = skb_shinfo(skb)->gso_segs; in fm10k_tso() 828 tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in fm10k_tso() 926 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in fm10k_tx_desc_flags() 927 likely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in fm10k_tx_desc_flags() 1008 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in fm10k_tx_map() 1119 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in fm10k_xmit_frame_ring() 1120 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in fm10k_xmit_frame_ring() 1122 count += skb_shinfo(skb)->nr_frags; in fm10k_xmit_frame_ring()
|
D | fm10k_ptp.c | 82 skb_shinfo(clone)->tx_flags |= SKBTX_IN_PROGRESS; in fm10k_ts_tx_enqueue()
|
D | fm10k_netdev.c | 662 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) in fm10k_xmit_frame()
|
/linux-4.1.27/drivers/infiniband/hw/nes/ |
D | nes_nic.c | 394 wqe_misc |= NES_NIC_SQ_WQE_LSO_ENABLE | (u16)skb_shinfo(skb)->gso_size; in nes_nic_send() 411 if ((skb_shinfo(skb)->nr_frags + 1) > 4) { in nes_nic_send() 413 netdev->name, skb_shinfo(skb)->nr_frags + 2, skb_headlen(skb)); in nes_nic_send() 437 for (skb_fragment_index = 0; skb_fragment_index < skb_shinfo(skb)->nr_frags; in nes_nic_send() 440 &skb_shinfo(skb)->frags[skb_fragment_index]; in nes_nic_send() 445 cpu_to_le16(skb_frag_size(&skb_shinfo(skb)->frags[skb_fragment_index])); in nes_nic_send() 524 nr_frags = skb_shinfo(skb)->nr_frags; in nes_netdev_start_xmit() 557 for (tso_frag_count=0; tso_frag_count < skb_shinfo(skb)->nr_frags; in nes_netdev_start_xmit() 560 &skb_shinfo(skb)->frags[tso_frag_count]; in nes_netdev_start_xmit() 599 skb_shinfo(skb)->nr_frags, skb_is_gso(skb), skb_shinfo(skb)->gso_size); in nes_netdev_start_xmit() [all …]
|
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
D | i40e_txrx.c | 1095 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in i40e_clean_rx_irq_ps() 1460 cd_mss = skb_shinfo(skb)->gso_size; in i40e_tso() 1635 num_frags = skb_shinfo(skb)->nr_frags; in i40e_chk_linearize() 1636 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_chk_linearize() 1649 frag = &skb_shinfo(skb)->frags[0]; in i40e_chk_linearize() 1654 if ((size >= skb_shinfo(skb)->gso_size) && in i40e_chk_linearize() 1656 size = (size % skb_shinfo(skb)->gso_size); in i40e_chk_linearize() 1705 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_tx_map() 1720 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in i40e_tx_map() 1891 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in i40e_xmit_descriptor_count() [all …]
|
/linux-4.1.27/drivers/net/ethernet/aeroflex/ |
D | greth.c | 104 if (skb_shinfo(skb)->nr_frags == 0) in greth_print_tx_packet() 112 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in greth_print_tx_packet() 115 skb_frag_address(&skb_shinfo(skb)->frags[i]), in greth_print_tx_packet() 116 skb_shinfo(skb)->frags[i].size, true); in greth_print_tx_packet() 191 int nr_frags = skb_shinfo(skb)->nr_frags; in greth_clean_rings() 201 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_rings() 474 nr_frags = skb_shinfo(skb)->nr_frags; in greth_start_xmit_gbit() 521 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_start_xmit_gbit() 691 nr_frags = skb_shinfo(skb)->nr_frags; in greth_clean_tx_gbit() 718 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in greth_clean_tx_gbit()
|
/linux-4.1.27/net/mpls/ |
D | mpls_gso.c | 29 if (unlikely(skb_shinfo(skb)->gso_type & in mpls_gso_segment()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_io.c | 397 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in qlcnic_tx_encap_pkt() 500 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in qlcnic_tx_pkt() 587 nr_frags = skb_shinfo(skb)->nr_frags; in qlcnic_map_tx_skb() 599 frag = &skb_shinfo(skb)->frags[i]; in qlcnic_map_tx_skb() 629 int i, nr_frags = skb_shinfo(skb)->nr_frags; in qlcnic_unmap_buffers() 676 frag_count = skb_shinfo(skb)->nr_frags + 1; in qlcnic_xmit_frame() 683 delta += skb_frag_size(&skb_shinfo(skb)->frags[i]); in qlcnic_xmit_frame() 688 frag_count = 1 + skb_shinfo(skb)->nr_frags; in qlcnic_xmit_frame() 1337 skb_shinfo(skb)->gso_size = qlcnic_get_lro_sts_mss(sts_data1); in qlcnic_process_lro() 1339 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in qlcnic_process_lro() [all …]
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 1620 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in i40e_clean_rx_irq_ps() 2171 cd_mss = skb_shinfo(skb)->gso_size; in i40e_tso() 2192 if (likely(!(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP))) in i40e_tsyn() 2208 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in i40e_tsyn() 2426 num_frags = skb_shinfo(skb)->nr_frags; in i40e_chk_linearize() 2427 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_chk_linearize() 2440 frag = &skb_shinfo(skb)->frags[0]; in i40e_chk_linearize() 2445 if ((size >= skb_shinfo(skb)->gso_size) && in i40e_chk_linearize() 2447 size = (size % skb_shinfo(skb)->gso_size); in i40e_chk_linearize() 2502 gso_segs = skb_shinfo(skb)->gso_segs; in i40e_tx_map() [all …]
|
D | i40e_fcoe.c | 1218 if (skb_shinfo(skb)->gso_type != SKB_GSO_FCOE) { in i40e_fcoe_tso() 1221 skb_shinfo(skb)->gso_type); in i40e_fcoe_tso() 1243 cd_mss = skb_shinfo(skb)->gso_size; in i40e_fcoe_tso()
|
/linux-4.1.27/net/phonet/ |
D | pep-gprs.c | 112 skb_shinfo(rskb)->frag_list = skb; in gprs_recv() 120 skb->next = skb_shinfo(skb)->frag_list; in gprs_recv()
|
D | pep.c | 1221 skb_shinfo(rskb)->frag_list = skb; in pep_write() 1229 skb->next = skb_shinfo(skb)->frag_list; in pep_write()
|
/linux-4.1.27/drivers/net/hyperv/ |
D | netvsc_drv.c | 285 int frags = skb_shinfo(skb)->nr_frags; in init_page_array() 306 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in init_page_array() 317 int i, frags = skb_shinfo(skb)->nr_frags; in count_skb_frag_slots() 321 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in count_skb_frag_slots() 570 lso_info->lso_v2_transmit.mss = skb_shinfo(skb)->gso_size; in netvsc_start_xmit()
|
/linux-4.1.27/drivers/target/tcm_fc/ |
D | tfc_io.c | 153 skb_shinfo(fp_skb(fp))->nr_frags, in ft_queue_data_in() 177 (skb_shinfo(fp_skb(fp))->nr_frags < FC_FRAME_SG_LEN)) in ft_queue_data_in()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_cmn.c | 522 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in bnx2x_set_gro_params() 525 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in bnx2x_set_gro_params() 536 skb_shinfo(skb)->gso_size = len_on_bd - hdrs_len; in bnx2x_set_gro_params() 721 if (skb_shinfo(skb)->gso_size) { in bnx2x_gro_receive() 3389 if (skb_shinfo(skb)->nr_frags >= (MAX_FETCH_BD - 3)) { in bnx2x_pkt_req_lin() 3392 unsigned short lso_mss = skb_shinfo(skb)->gso_size; in bnx2x_pkt_req_lin() 3397 int num_wnds = skb_shinfo(skb)->nr_frags - wnd_size; in bnx2x_pkt_req_lin() 3414 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]); in bnx2x_pkt_req_lin() 3430 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx + wnd_size - 1]); in bnx2x_pkt_req_lin() 3437 skb_frag_size(&skb_shinfo(skb)->frags[wnd_idx]); in bnx2x_pkt_req_lin() [all …]
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 253 nfrags = skb_shinfo(skb)->nr_frags; in unmap_skb() 257 skb_frag_size(&skb_shinfo(skb)->frags[frag_idx]), in unmap_skb() 899 skb_fill_page_desc(newskb, skb_shinfo(newskb)->nr_frags, in get_packet_pg() 946 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1) + 2; in calc_tx_descs() 947 if (skb_shinfo(skb)->gso_size) in calc_tx_descs() 978 nfrags = skb_shinfo(skb)->nr_frags; in make_sgl() 980 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in make_sgl() 1154 tso_info = V_LSO_MSS(skb_shinfo(skb)->gso_size); in write_tx_pkt_wr() 1283 if (skb_shinfo(skb)->gso_size) in t3_eth_xmit() 1544 si = skb_shinfo(skb); in deferred_unmap_destructor() [all …]
|
/linux-4.1.27/include/net/ |
D | udp_tunnel.h | 109 skb_shinfo(skb)->gso_type |= uh->check ? in udp_tunnel_gro_complete()
|
D | udp.h | 146 for (skb = skb_shinfo(skb)->frag_list; skb; skb = skb->next) { in udp_csum()
|
D | sch_generic.h | 522 bstats->packets += skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 1; in bstats_update()
|
D | tcp.h | 771 return skb_shinfo(skb)->gso_size; in tcp_skb_mss()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/ |
D | sge.c | 1117 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in compute_large_page_tx_descs() 1124 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in compute_large_page_tx_descs() 1208 nfrags = skb_shinfo(skb)->nr_frags; in write_tx_descs() 1256 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in write_tx_descs() 1320 count = 1 + skb_shinfo(skb)->nr_frags; in restart_sched() 1674 count = 1 + skb_shinfo(skb)->nr_frags; in t1_sge_tx() 1711 count = 1 + skb_shinfo(skb)->nr_frags; in t1_sge_tx() 1796 if (skb_shinfo(skb)->gso_size) { in t1_start_xmit() 1811 skb_shinfo(skb)->gso_size)); in t1_start_xmit()
|
/linux-4.1.27/net/sctp/ |
D | ulpqueue.c | 335 struct sk_buff *list = skb_shinfo(f_frag)->frag_list; in sctp_make_reassembled_event() 365 skb_shinfo(new)->frag_list = pos; in sctp_make_reassembled_event() 367 skb_shinfo(f_frag)->frag_list = pos; in sctp_make_reassembled_event() 997 flist = skb_shinfo(skb)->frag_list; in sctp_ulpq_renege_list()
|
D | socket.c | 160 skb_shinfo(chunk->skb)->destructor_arg = chunk; in sctp_set_owner_w() 6891 struct sctp_chunk *chunk = skb_shinfo(skb)->destructor_arg; in sctp_wfree()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_fcoe.c | 522 if (skb_is_gso(skb) && (skb_shinfo(skb)->gso_type != SKB_GSO_FCOE)) { in ixgbe_fso() 524 skb_shinfo(skb)->gso_type); in ixgbe_fso() 593 skb_shinfo(skb)->gso_size); in ixgbe_fso() 602 mss_l4len_idx = skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbe_fso()
|
D | ixgbe_main.c | 1562 skb_shinfo(skb)->gso_size = DIV_ROUND_UP((skb->len - hdr_len), in ixgbe_set_rsc_gso_size() 1564 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in ixgbe_set_rsc_gso_size() 1695 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbe_pull_tail() 1741 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbe_dma_sync_frag() 1879 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in ixgbe_add_rx_frag() 6826 first->gso_segs = skb_shinfo(skb)->gso_segs; in ixgbe_tso() 6831 mss_l4len_idx |= skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbe_tso() 7038 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in ixgbe_tx_map() 7300 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in ixgbe_xmit_frame_ring() 7301 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in ixgbe_xmit_frame_ring() [all …]
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
D | en_rx.c | 550 struct skb_frag_struct *skb_frags_rx = skb_shinfo(skb)->frags; in mlx4_en_complete_rx_desc() 629 skb_shinfo(skb)->nr_frags = used_frags; in mlx4_en_rx_skb() 637 skb_shinfo(skb)->frags[0].page_offset += pull_len; in mlx4_en_rx_skb() 640 skb_frag_size_sub(&skb_shinfo(skb)->frags[0], pull_len); in mlx4_en_rx_skb() 897 void *va = skb_frag_address(skb_shinfo(gro_skb)->frags); in mlx4_en_process_rx_cq() 906 skb_shinfo(gro_skb)->nr_frags = nr; in mlx4_en_process_rx_cq()
|
D | en_tx.c | 707 struct skb_shared_info *shinfo = skb_shinfo(skb); in mlx4_en_xmit()
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
D | c2.c | 769 if (unlikely(c2_port->tx_avail < (skb_shinfo(skb)->nr_frags + 1))) { in c2_xmit_frame() 799 if (skb_shinfo(skb)->nr_frags) { in c2_xmit_frame() 800 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in c2_xmit_frame() 801 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in c2_xmit_frame() 824 c2_port->tx_avail -= (skb_shinfo(skb)->nr_frags + 1); in c2_xmit_frame()
|
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/ |
D | xgbe-drv.c | 1235 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in xgbe_prep_tx_tstamp() 1265 packet->mss = skb_shinfo(skb)->gso_size; in xgbe_prep_tso() 1274 packet->tx_packets = skb_shinfo(skb)->gso_segs; in xgbe_prep_tso() 1312 if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) { in xgbe_packet_info() 1341 if ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in xgbe_packet_info() 1351 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgbe_packet_info() 1352 frag = &skb_shinfo(skb)->frags[i]; in xgbe_packet_info() 2014 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in xgbe_rx_poll()
|
D | xgbe-desc.c | 565 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xgbe_map_tx_skb() 568 frag = &skb_shinfo(skb)->frags[i]; in xgbe_map_tx_skb()
|
/linux-4.1.27/drivers/net/usb/ |
D | cdc-phonet.c | 166 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in rx_complete() 172 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in rx_complete()
|
D | usbnet.c | 1257 num_sgs = skb_shinfo(skb)->nr_frags + 1; in build_dma_sg() 1273 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in build_dma_sg() 1274 struct skb_frag_struct *f = &skb_shinfo(skb)->frags[i]; in build_dma_sg()
|
D | r8152.c | 1371 if (skb_shinfo(skb)->gso_size) { in r8152_csum_workaround() 1452 u32 mss = skb_shinfo(skb)->gso_size; in r8152_tx_csum() 1971 u32 mss = skb_shinfo(skb)->gso_size; in rtl8152_features_check()
|
D | ax88179_178a.c | 1428 int mss = skb_shinfo(skb)->gso_size; in ax88179_tx_fixup()
|
D | smsc95xx.c | 1841 BUG_ON(skb_shinfo(skb)->nr_frags); in smsc95xx_tx_fixup()
|
D | smsc75xx.c | 2216 u16 mss = max(skb_shinfo(skb)->gso_size, TX_MSS_MIN); in smsc75xx_tx_fixup()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 272 si = skb_shinfo(skb); in map_skb() 301 si = skb_shinfo(skb); in unmap_skb() 325 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in unmap_sgl() 775 int hdrlen = skb_shinfo(skb)->gso_size ? in is_eth_imm() 812 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1) + 4; in calc_tx_flits() 813 if (skb_shinfo(skb)->gso_size) in calc_tx_flits() 858 const struct skb_shared_info *si = skb_shinfo(skb); in write_sgl() 1206 ssi = skb_shinfo(skb); in t4_eth_xmit() 1477 cnt = skb_shinfo(skb)->nr_frags; in calc_tx_flits_ofld() 1702 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
|
/linux-4.1.27/drivers/block/aoe/ |
D | aoedev.c | 405 while (atomic_read(&skb_shinfo(skb)->dataref) != 1 && i-- > 0) in skbfree() 415 skb_shinfo(skb)->nr_frags = skb->data_len = 0; in skbfree()
|
D | aoecmd.c | 181 if (skb && atomic_read(&skb_shinfo(skb)->dataref) == 1) { in skb_pool_get() 235 if (atomic_read(&skb_shinfo(skb)->dataref) != 1) { in newtframe() 244 skb_shinfo(skb)->nr_frags = skb->data_len = 0; in newtframe()
|
/linux-4.1.27/drivers/scsi/fcoe/ |
D | fcoe_transport.c | 308 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in fcoe_fc_crc() 309 frag = &skb_shinfo(skb)->frags[i]; in fcoe_fc_crc() 471 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, in fcoe_get_paged_crc_eof()
|
D | fcoe.c | 1645 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; in fcoe_xmit() 1698 skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; in fcoe_xmit() 1699 skb_shinfo(skb)->gso_size = fr_max_payload(fp); in fcoe_xmit() 1701 skb_shinfo(skb)->gso_type = 0; in fcoe_xmit() 1702 skb_shinfo(skb)->gso_size = 0; in fcoe_xmit()
|
/linux-4.1.27/drivers/net/ethernet/realtek/ |
D | 8139cp.c | 725 this_frag = &skb_shinfo(skb)->frags[frag]; in unwind_tx_frag_mapping() 744 if (TX_BUFFS_AVAIL(cp) <= (skb_shinfo(skb)->nr_frags + 1)) { in cp_start_xmit() 753 mss = skb_shinfo(skb)->gso_size; in cp_start_xmit() 757 if (skb_shinfo(skb)->nr_frags == 0) { in cp_start_xmit() 810 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in cp_start_xmit() 811 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag]; in cp_start_xmit() 841 if (frag == skb_shinfo(skb)->nr_frags - 1) in cp_start_xmit()
|
D | r8169.c | 6809 struct skb_shared_info *info = skb_shinfo(skb); in rtl8169_xmit_frags() 6871 if (skb_shinfo(skb)->gso_size) { in r8169_csum_workaround() 6941 u32 mss = skb_shinfo(skb)->gso_size; in rtl8169_tso_csum_v1() 6964 u32 mss = skb_shinfo(skb)->gso_size; in rtl8169_tso_csum_v2() 7051 if (unlikely(!TX_FRAGS_READY_FOR(tp, skb_shinfo(skb)->nr_frags))) { in rtl8169_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 287 si = skb_shinfo(skb); in map_skb() 310 unsigned int nfrags = skb_shinfo(skb)->nr_frags; in unmap_sgl() 880 flits = sgl_len(skb_shinfo(skb)->nr_frags + 1); in calc_tx_flits() 881 if (skb_shinfo(skb)->gso_size) in calc_tx_flits() 914 const struct skb_shared_info *si = skb_shinfo(skb); in write_sgl() 1269 ssi = skb_shinfo(skb); in t4vf_eth_xmit() 1466 skb_shinfo(skb)->nr_frags = gl->nfrags; in copy_frags()
|
/linux-4.1.27/drivers/hsi/clients/ |
D | ssi_protocol.c | 186 BUG_ON(msg->sgt.nents != (unsigned int)(skb_shinfo(skb)->nr_frags + 1)); in ssip_skb_to_msg() 190 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ssip_skb_to_msg() 193 frag = &skb_shinfo(skb)->frags[i]; in ssip_skb_to_msg() 215 msg = hsi_alloc_msg(skb_shinfo(skb)->nr_frags + 1, flags); in ssip_alloc_data()
|
/linux-4.1.27/drivers/atm/ |
D | eni.c | 1101 DPRINTK("iovcnt = %d\n",skb_shinfo(skb)->nr_frags); in do_tx() 1102 if (!skb_shinfo(skb)->nr_frags) dma_size += 5; in do_tx() 1103 else dma_size += 5*(skb_shinfo(skb)->nr_frags+1); in do_tx() 1124 if (!skb_shinfo(skb)->nr_frags) in do_tx() 1129 for (i = -1; i < skb_shinfo(skb)->nr_frags; i++) in do_tx() 1136 skb_frag_page(&skb_shinfo(skb)->frags[i]) + in do_tx() 1137 skb_shinfo(skb)->frags[i].page_offset, in do_tx() 1138 skb_frag_size(&skb_shinfo(skb)->frags[i])); in do_tx()
|
D | he.c | 2517 if (skb_shinfo(skb)->nr_frags) { in he_send() 2561 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in he_send() 2562 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in he_send()
|
D | zatm.c | 642 EVENT("iovcnt=%d\n",skb_shinfo(skb)->nr_frags,0); in do_tx() 644 if (!skb_shinfo(skb)->nr_frags) { in do_tx()
|
D | horizon.c | 1700 unsigned int tx_iovcnt = skb_shinfo(skb)->nr_frags; in hrz_send()
|
/linux-4.1.27/drivers/net/ethernet/neterion/ |
D | s2io.h | 1139 #define s2io_tcp_mss(skb) skb_shinfo(skb)->gso_size 1140 #define s2io_udp_mss(skb) skb_shinfo(skb)->gso_size 1141 #define s2io_offload_type(skb) skb_shinfo(skb)->gso_type
|
D | s2io.c | 2340 frg_cnt = skb_shinfo(skb)->nr_frags; in s2io_txdl_getskb() 2344 const skb_frag_t *frag = &skb_shinfo(skb)->frags[j]; in s2io_txdl_getskb() 4154 (__force u64)skb_shinfo(skb)->ip6_frag_id; in s2io_xmit() 4157 (__force u64)skb_shinfo(skb)->ip6_frag_id << 32; in s2io_xmit() 4179 frg_cnt = skb_shinfo(skb)->nr_frags; in s2io_xmit() 4182 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in s2io_xmit() 8572 if (skb_shinfo(first)->frag_list) in lro_append_pkt() 8575 skb_shinfo(first)->frag_list = skb; in lro_append_pkt()
|
/linux-4.1.27/Documentation/networking/ |
D | driver.txt | 21 if (TX_BUFFS_AVAIL(dp) <= (skb_shinfo(skb)->nr_frags + 1)) {
|
D | timestamping.txt | 225 relevant sequence number in skb_shinfo(skb)->tskey. Because an skbuff 432 - In hard_start_xmit(), check if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) 437 SKBTX_IN_PROGRESS in skb_shinfo(skb)->tx_flags , e.g. with 439 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
|
D | netdev-features.txt | 124 NETIF_F_SG --- paged skbs (skb_shinfo()->frags), NETIF_F_FRAGLIST ---
|
/linux-4.1.27/drivers/net/ethernet/freescale/ |
D | fec_main.c | 371 int nr_frags = skb_shinfo(skb)->nr_frags; in fec_enet_txq_submit_frag_skb() 383 this_frag = &skb_shinfo(skb)->frags[frag]; in fec_enet_txq_submit_frag_skb() 390 frag_len = skb_shinfo(skb)->frags[frag].size; in fec_enet_txq_submit_frag_skb() 397 if (unlikely(skb_shinfo(skb)->tx_flags & in fec_enet_txq_submit_frag_skb() 456 int nr_frags = skb_shinfo(skb)->nr_frags; in fec_enet_txq_submit_skb() 519 if (unlikely(skb_shinfo(skb)->tx_flags & in fec_enet_txq_submit_skb() 529 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP && in fec_enet_txq_submit_skb() 531 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in fec_enet_txq_submit_skb() 720 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in fec_enet_txq_submit_tso() 1254 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS) && in fec_enet_tx_queue()
|
D | gianfar.c | 1908 for (j = 0; j < skb_shinfo(tx_queue->tx_skbuff[i])->nr_frags; in free_skb_tx_queue() 2268 do_tstamp = (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in gfar_start_xmit() 2296 nr_frags = skb_shinfo(skb)->nr_frags; in gfar_start_xmit() 2343 frag_len = skb_shinfo(skb)->frags[i].size; in gfar_start_xmit() 2353 &skb_shinfo(skb)->frags[i], in gfar_start_xmit() 2404 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in gfar_start_xmit() 2627 frags = skb_shinfo(skb)->nr_frags; in gfar_clean_tx_ring() 2632 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in gfar_clean_tx_ring() 2646 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) { in gfar_clean_tx_ring() 2656 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) { in gfar_clean_tx_ring()
|
/linux-4.1.27/drivers/net/ethernet/emulex/benet/ |
D | be_main.c | 674 stats->tx_pkts += (skb_shinfo(skb)->gso_segs ? : 1); in be_tx_stats_update() 682 return 1 + (skb_headlen(skb) ? 1 : 0) + skb_shinfo(skb)->nr_frags; in skb_wrb_cnt() 756 wrb_params->lso_mss = skb_shinfo(skb)->gso_size; in be_get_wrb_params_from_skb() 931 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in be_xmit_enqueue() 932 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in be_xmit_enqueue() 1813 skb_shinfo(skb)->nr_frags = 1; in skb_fill_rx_data() 1815 skb_shinfo(skb)->frags[0].page_offset = in skb_fill_rx_data() 1817 skb_frag_size_set(&skb_shinfo(skb)->frags[0], in skb_fill_rx_data() 1841 skb_shinfo(skb)->frags[j].page_offset = in skb_fill_rx_data() 1843 skb_frag_size_set(&skb_shinfo(skb)->frags[j], 0); in skb_fill_rx_data() [all …]
|
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_main.c | 1291 int nr_frags = skb_shinfo(skb)->nr_frags; in sxgbe_xmit() 1294 u16 cur_mss = skb_shinfo(skb)->gso_size; in sxgbe_xmit() 1304 ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in sxgbe_xmit() 1369 const skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_num]; in sxgbe_xmit() 1428 if (unlikely((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in sxgbe_xmit() 1431 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in sxgbe_xmit()
|
/linux-4.1.27/drivers/net/ethernet/intel/igbvf/ |
D | netdev.c | 318 if (!skb_shinfo(skb)->nr_frags) { in igbvf_clean_rx_irq() 331 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in igbvf_clean_rx_irq() 829 segs = skb_shinfo(skb)->gso_segs ?: 1; in igbvf_clean_tx_irq() 1972 mss_l4len_idx = (skb_shinfo(skb)->gso_size << E1000_ADVTXD_MSS_SHIFT); in igbvf_tso() 2102 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) { in igbvf_tx_map_adv() 2110 frag = &skb_shinfo(skb)->frags[f]; in igbvf_tx_map_adv() 2241 if (igbvf_maybe_stop_tx(netdev, skb_shinfo(skb)->nr_frags + 4)) { in igbvf_xmit_frame_ring_adv()
|
/linux-4.1.27/drivers/net/ethernet/atheros/atl1e/ |
D | atl1e_main.c | 1617 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in atl1e_cal_tdp_req() 1618 fg_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in atl1e_cal_tdp_req() 1624 (skb_shinfo(skb)->gso_type == SKB_GSO_TCPV6)) { in atl1e_cal_tdp_req() 1652 offload_type = skb_shinfo(skb)->gso_type; in atl1e_tso_csum() 1680 tpd->word3 |= ((skb_shinfo(skb)->gso_size) & in atl1e_tso_csum() 1725 nr_frags = skb_shinfo(skb)->nr_frags; in atl1e_tx_map() 1794 frag = &skb_shinfo(skb)->frags[f]; in atl1e_tx_map()
|
/linux-4.1.27/net/packet/ |
D | af_packet.c | 1713 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in packet_sendmsg_spkt() 2102 ph = skb_shinfo(skb)->destructor_arg; in tpacket_destruct_skb() 2150 sock_tx_timestamp(&po->sk, &skb_shinfo(skb)->tx_flags); in tpacket_fill_skb() 2151 skb_shinfo(skb)->destructor_arg = ph.raw; in tpacket_fill_skb() 2231 nr_frags = skb_shinfo(skb)->nr_frags; in tpacket_fill_skb() 2551 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in packet_snd() 2576 skb_shinfo(skb)->gso_size = in packet_snd() 2578 skb_shinfo(skb)->gso_type = gso_type; in packet_snd() 2581 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in packet_snd() 2582 skb_shinfo(skb)->gso_segs = 0; in packet_snd() [all …]
|
/linux-4.1.27/drivers/net/vmxnet3/ |
D | vmxnet3_drv.c | 644 struct skb_frag_struct *frag = skb_shinfo(skb)->frags + in vmxnet3_append_frag() 645 skb_shinfo(skb)->nr_frags; in vmxnet3_append_frag() 647 BUG_ON(skb_shinfo(skb)->nr_frags >= MAX_SKB_FRAGS); in vmxnet3_append_frag() 654 skb_shinfo(skb)->nr_frags++; in vmxnet3_append_frag() 739 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vmxnet3_map_pkt() 740 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in vmxnet3_map_pkt() 913 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in txd_estimate() 914 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in txd_estimate() 953 ctx.mss = skb_shinfo(skb)->gso_size; in vmxnet3_tq_xmit()
|
/linux-4.1.27/drivers/net/ethernet/tile/ |
D | tilegx.c | 460 struct skb_shared_info *shtx = skb_shinfo(skb); in tile_tx_timestamp() 1677 struct skb_shared_info *sh = skb_shinfo(skb); in tso_count_edescs() 1723 struct skb_shared_info *sh = skb_shinfo(skb); in tso_headers_prepare() 1832 struct skb_shared_info *sh = skb_shinfo(skb); in tso_egress() 1980 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_frags()
|
D | tilepro.c | 1607 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_frags() 1680 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx_tso() 1854 struct skb_shared_info *sh = skb_shinfo(skb); in tile_net_tx()
|
/linux-4.1.27/drivers/net/ethernet/pasemi/ |
D | pasemi_mac.c | 302 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; in pasemi_mac_unmap_tx_skb() 573 nfrags = skb_shinfo(info->skb)->nr_frags; in pasemi_mac_free_tx_resources() 1398 const int nfrags = skb_shinfo(skb)->nr_frags; in pasemi_mac_queue_csdesc() 1494 nfrags = skb_shinfo(skb)->nr_frags; in pasemi_mac_start_tx() 1503 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in pasemi_mac_start_tx()
|
/linux-4.1.27/drivers/net/ethernet/marvell/ |
D | sky2.c | 1215 for (i = 0; i < skb_shinfo(re->skb)->nr_frags; i++) in sky2_rx_submit() 1232 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in sky2_rx_map_skb() 1233 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in sky2_rx_map_skb() 1247 skb_frag_size(&skb_shinfo(skb)->frags[i]), in sky2_rx_map_skb() 1269 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) in sky2_rx_unmap_skb() 1271 skb_frag_size(&skb_shinfo(skb)->frags[i]), in sky2_rx_unmap_skb() 1807 count = (skb_shinfo(skb)->nr_frags + 1) in tx_le_req() 1877 mss = skb_shinfo(skb)->gso_size; in sky2_xmit_frame() 1949 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in sky2_xmit_frame() 1950 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in sky2_xmit_frame() [all …]
|
D | mv643xx_eth.c | 675 for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) { in has_tiny_unaligned_frags() 676 const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag]; in has_tiny_unaligned_frags() 836 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in txq_submit_tso() 881 int nr_frags = skb_shinfo(skb)->nr_frags; in txq_submit_frag_skb() 889 this_frag = &skb_shinfo(skb)->frags[frag]; in txq_submit_frag_skb() 920 int nr_frags = skb_shinfo(skb)->nr_frags; in txq_submit_skb()
|
D | mvneta.c | 1646 data_left = min_t(int, skb_shinfo(skb)->gso_size, total_len); in mvneta_tx_tso() 1696 int i, nr_frags = skb_shinfo(skb)->nr_frags; in mvneta_tx_frag_process() 1699 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvneta_tx_frag_process() 1764 frags = skb_shinfo(skb)->nr_frags + 1; in mvneta_tx()
|
D | skge.c | 2744 if (unlikely(skge_avail(&skge->tx_ring) < skb_shinfo(skb)->nr_frags + 1)) in skge_xmit_frame() 2780 if (!skb_shinfo(skb)->nr_frags) /* single buffer i.e. no fragments */ in skge_xmit_frame() 2786 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in skge_xmit_frame() 2787 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skge_xmit_frame()
|
D | mvpp2.c | 5125 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mvpp2_tx_frag_process() 5126 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mvpp2_tx_frag_process() 5144 if (i == (skb_shinfo(skb)->nr_frags - 1)) { in mvpp2_tx_frag_process() 5186 frags = skb_shinfo(skb)->nr_frags + 1; in mvpp2_tx()
|
/linux-4.1.27/drivers/scsi/bnx2fc/ |
D | bnx2fc_fcoe.c | 329 frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; in bnx2fc_xmit() 373 skb_shinfo(skb)->gso_type = SKB_GSO_FCOE; in bnx2fc_xmit() 374 skb_shinfo(skb)->gso_size = fr_max_payload(fp); in bnx2fc_xmit() 376 skb_shinfo(skb)->gso_type = 0; in bnx2fc_xmit() 377 skb_shinfo(skb)->gso_size = 0; in bnx2fc_xmit()
|
/linux-4.1.27/drivers/net/ethernet/adi/ |
D | bfin_mac.c | 836 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) { in bfin_tx_hwtstamp() 842 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in bfin_tx_hwtstamp() 1159 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) in bfin_mac_hard_start_xmit() 1169 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) in bfin_mac_hard_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ |
D | ibmveth.c | 957 if (skb_shinfo(skb)->nr_frags > 5 && __skb_linearize(skb)) { in ibmveth_start_xmit() 1025 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ibmveth_start_xmit() 1026 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ibmveth_start_xmit() 1051 for (i = 1; i < skb_shinfo(skb)->nr_frags + 1; i++) in ibmveth_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/cisco/enic/ |
D | enic_main.c | 363 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { in enic_queue_wq_skb_cont() 489 for (frag = skb_shinfo(skb)->frags; len_left; frag++) { in enic_queue_wq_skb_tso() 517 unsigned int mss = skb_shinfo(skb)->gso_size; in enic_queue_wq_skb() 582 if (skb_shinfo(skb)->gso_size == 0 && in enic_hard_start_xmit() 583 skb_shinfo(skb)->nr_frags + 1 > ENIC_NON_TSO_MAX_DESC && in enic_hard_start_xmit() 592 skb_shinfo(skb)->nr_frags + ENIC_DESC_MAX_SPLITS) { in enic_hard_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgbevf/ |
D | ixgbevf_main.c | 666 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbevf_pull_tail() 813 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in ixgbevf_add_rx_frag() 3317 first->gso_segs = skb_shinfo(skb)->gso_segs; in ixgbevf_tso() 3322 mss_l4len_idx |= skb_shinfo(skb)->gso_size << IXGBE_ADVTXD_MSS_SHIFT; in ixgbevf_tso() 3453 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in ixgbevf_tx_map() 3628 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in ixgbevf_xmit_frame() 3629 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in ixgbevf_xmit_frame() 3631 count += skb_shinfo(skb)->nr_frags; in ixgbevf_xmit_frame()
|
/linux-4.1.27/net/openvswitch/ |
D | flow.c | 521 skb_shinfo(skb)->gso_type & SKB_GSO_UDP) in key_extract() 638 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP) in key_extract()
|
D | datapath.c | 337 unsigned short gso_type = skb_shinfo(skb)->gso_type; in queue_gso_packets()
|
/linux-4.1.27/drivers/net/ethernet/ti/ |
D | netcp_core.c | 665 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in netcp_process_one_rx_packet() 1028 prefetchw(skb_shinfo(skb)); in netcp_tx_map_skb() 1037 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in netcp_tx_map_skb() 1038 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in netcp_tx_map_skb() 1077 if (skb_shinfo(skb)->frag_list) { in netcp_tx_map_skb()
|
D | cpts.c | 343 if (!(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS)) in cpts_tx_timestamp()
|
D | cpsw.c | 1389 if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP && in cpsw_ndo_start_xmit() 1391 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in cpsw_ndo_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/tehuti/ |
D | tehuti.c | 1503 int nr_frags = skb_shinfo(skb)->nr_frags; in bdx_tx_map_skb() 1520 frag = &skb_shinfo(skb)->frags[i]; in bdx_tx_map_skb() 1626 int nr_frags = skb_shinfo(skb)->nr_frags; in bdx_tx_transmit() 1646 if (skb_shinfo(skb)->gso_size) { in bdx_tx_transmit() 1647 txd_mss = skb_shinfo(skb)->gso_size; in bdx_tx_transmit()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/netxen/ |
D | netxen_nic_main.c | 1902 skb_shinfo(skb)->gso_size > 0) { in netxen_tso_check() 1906 first_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in netxen_tso_check() 2006 nr_frags = skb_shinfo(skb)->nr_frags; in netxen_map_tx_skb() 2018 frag = &skb_shinfo(skb)->frags[i]; in netxen_map_tx_skb() 2071 frag_count = skb_shinfo(skb)->nr_frags + 1; in netxen_nic_xmit_frame() 2079 frag = &skb_shinfo(skb)->frags[i]; in netxen_nic_xmit_frame() 2086 frag_count = 1 + skb_shinfo(skb)->nr_frags; in netxen_nic_xmit_frame()
|
D | netxen_nic_init.c | 1653 skb_shinfo(skb)->gso_size = netxen_get_lro_sts_mss(sts_data1); in netxen_process_lro()
|
/linux-4.1.27/drivers/net/ethernet/atheros/atl1c/ |
D | atl1c_main.c | 1957 tpd_req = skb_shinfo(skb)->nr_frags + 1; in atl1c_cal_tpd_req() 1963 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) in atl1c_cal_tpd_req() 1986 offload_type = skb_shinfo(skb)->gso_type; in atl1c_tso_csum() 2041 (*tpd)->word1 |= (skb_shinfo(skb)->gso_size & TPD_MSS_MASK) << in atl1c_tso_csum() 2105 nr_frags = skb_shinfo(skb)->nr_frags; in atl1c_tx_map() 2155 frag = &skb_shinfo(skb)->frags[f]; in atl1c_tx_map()
|
/linux-4.1.27/drivers/net/ethernet/3com/ |
D | typhoon.c | 325 #define skb_tso_size(x) (skb_shinfo(x)->gso_size) 740 numDesc = skb_shinfo(skb)->nr_frags + 1; in typhoon_start_tx() 793 if(skb_shinfo(skb)->nr_frags == 0) { in typhoon_start_tx() 813 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in typhoon_start_tx() 814 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in typhoon_start_tx()
|
D | 3c59x.c | 2165 if (!skb_shinfo(skb)->nr_frags) { in boomerang_start_xmit() 2184 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in boomerang_start_xmit() 2185 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in boomerang_start_xmit() 2209 if (i == skb_shinfo(skb)->nr_frags-1) in boomerang_start_xmit() 2448 for (i=0; i<=skb_shinfo(skb)->nr_frags; i++) in boomerang_interrupt() 2777 for (k=0; k<=skb_shinfo(skb)->nr_frags; k++) in vortex_close()
|
/linux-4.1.27/drivers/net/ethernet/atheros/atlx/ |
D | atl1.c | 2125 if (skb_shinfo(skb)->gso_size) { in atl1_tso() 2170 ptpd->word3 |= (skb_shinfo(skb)->gso_size & in atl1_tso() 2220 nr_frags = skb_shinfo(skb)->nr_frags; in atl1_tx_map() 2282 frag = &skb_shinfo(skb)->frags[f]; in atl1_tx_map() 2381 nr_frags = skb_shinfo(skb)->nr_frags; in atl1_xmit_frame() 2383 unsigned int f_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); in atl1_xmit_frame() 2388 mss = skb_shinfo(skb)->gso_size; in atl1_xmit_frame()
|
/linux-4.1.27/drivers/net/wireless/ath/wil6210/ |
D | txrx.c | 1110 int nr_frags = skb_shinfo(skb)->nr_frags; in __wil_tx_vring() 1168 &skb_shinfo(skb)->frags[f]; in __wil_tx_vring() 1326 (skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS); in wil_need_txstat()
|
D | debugfs.c | 805 int nr_frags = skb_shinfo(skb)->nr_frags; in wil_seq_print_skb() 814 &skb_shinfo(skb)->frags[i]; in wil_seq_print_skb()
|
/linux-4.1.27/drivers/net/ethernet/adaptec/ |
D | starfire.c | 162 #define skb_num_frags(skb) (skb_shinfo(skb)->nr_frags + 1) 1230 const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[i - 1]; in start_tx() 1349 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in intr_handler() 1352 skb_frag_size(&skb_shinfo(skb)->frags[i]), in intr_handler()
|
/linux-4.1.27/drivers/net/ethernet/ |
D | jme.c | 1979 nr_alloc = skb_shinfo(skb)->nr_frags + 2; in jme_alloc_txdesc() 2057 int i, nr_frags = skb_shinfo(skb)->nr_frags; in jme_map_tx_skb() 2064 frag = &skb_shinfo(skb)->frags[i]; in jme_map_tx_skb() 2095 *mss = cpu_to_le16(skb_shinfo(skb)->gso_size << TXDESC_MSS_SHIFT); in jme_tx_tso() 2205 txbi->nr_desc = skb_shinfo(skb)->nr_frags + 2; in jme_fill_tx_desc() 2280 idx, skb_shinfo(skb)->nr_frags + 2, jiffies); in jme_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/atheros/alx/ |
D | main.c | 1026 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) { in alx_map_tx_skb() 1029 frag = &skb_shinfo(skb)->frags[f]; in alx_map_tx_skb() 1074 int tpdreq = skb_shinfo(skb)->nr_frags + 1; in alx_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/ |
D | fs_enet-main.c | 518 int nr_frags = skb_shinfo(skb)->nr_frags; in fs_enet_start_xmit() 569 frag = skb_shinfo(skb)->frags; in fs_enet_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgb/ |
D | ixgb_main.c | 1235 mss = skb_shinfo(skb)->gso_size; in ixgb_tso() 1331 unsigned int mss = skb_shinfo(skb)->gso_size; in ixgb_tx_map() 1332 unsigned int nr_frags = skb_shinfo(skb)->nr_frags; in ixgb_tx_map() 1369 frag = &skb_shinfo(skb)->frags[f]; in ixgb_tx_map()
|
/linux-4.1.27/drivers/net/ethernet/stmicro/stmmac/ |
D | stmmac_main.c | 354 if (likely(!skb || !(skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS))) in stmmac_get_tx_hwtstamp() 1944 int nfrags = skb_shinfo(skb)->nr_frags; in stmmac_xmit() 1995 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in stmmac_xmit() 2065 if (unlikely((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in stmmac_xmit() 2068 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in stmmac_xmit()
|
/linux-4.1.27/drivers/net/ethernet/intel/e1000/ |
D | e1000_main.c | 2724 mss = skb_shinfo(skb)->gso_size; in e1000_tso() 2908 frag = &skb_shinfo(skb)->frags[f]; in e1000_tx_map() 2953 segs = skb_shinfo(skb)->gso_segs ?: 1; in e1000_tx_map() 3141 mss = skb_shinfo(skb)->gso_size; in e1000_xmit_frame() 3207 nr_frags = skb_shinfo(skb)->nr_frags; in e1000_xmit_frame() 3209 count += TXD_USE_COUNT(skb_frag_size(&skb_shinfo(skb)->frags[f]), in e1000_xmit_frame() 4213 skb_shinfo(rxtop)->nr_frags, in e1000_clean_jumbo_rx_irq() 4222 skb_shinfo(rxtop)->nr_frags, in e1000_clean_jumbo_rx_irq()
|
/linux-4.1.27/drivers/net/ethernet/xilinx/ |
D | ll_temac_main.c | 685 num_frag = skb_shinfo(skb)->nr_frags; in temac_start_xmit() 686 frag = &skb_shinfo(skb)->frags[0]; in temac_start_xmit()
|
D | xilinx_axienet_main.c | 662 num_frag = skb_shinfo(skb)->nr_frags; in axienet_start_xmit() 694 frag = &skb_shinfo(skb)->frags[ii]; in axienet_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/intel/e1000e/ |
D | netdev.c | 1582 shinfo = skb_shinfo(rxtop); in e1000_clean_jumbo_rx_irq() 1594 shinfo = skb_shinfo(rxtop); in e1000_clean_jumbo_rx_irq() 5208 mss = skb_shinfo(skb)->gso_size; in e1000_tso() 5352 frag = &skb_shinfo(skb)->frags[f]; in e1000_tx_map() 5380 segs = skb_shinfo(skb)->gso_segs ? : 1; in e1000_tx_map() 5575 mss = skb_shinfo(skb)->gso_size; in e1000_xmit_frame() 5607 nr_frags = skb_shinfo(skb)->nr_frags; in e1000_xmit_frame() 5609 count += DIV_ROUND_UP(skb_frag_size(&skb_shinfo(skb)->frags[f]), in e1000_xmit_frame() 5654 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) && in e1000_xmit_frame() 5657 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in e1000_xmit_frame()
|
/linux-4.1.27/net/tipc/ |
D | msg.c | 164 skb_shinfo(head)->frag_list = frag; in tipc_buf_append()
|
/linux-4.1.27/drivers/scsi/cxgbi/ |
D | libcxgbi.c | 1139 int frags = skb_shinfo(skb)->nr_frags + in cxgbi_sock_send_pdus() 1151 csk, skb_shinfo(skb)->nr_frags, skb->len, in cxgbi_sock_send_pdus() 2221 skb_shinfo(skb)->nr_frags = tdata->nr_frags; in cxgbi_conn_init_pdu() 2239 i = skb_shinfo(skb)->nr_frags; in cxgbi_conn_init_pdu() 2240 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in cxgbi_conn_init_pdu()
|
/linux-4.1.27/drivers/net/ethernet/via/ |
D | via-velocity.c | 1746 if (skb_shinfo(skb)->nr_frags > 0) in velocity_free_tx_buf() 2567 if (skb_shinfo(skb)->nr_frags > 6 && __skb_linearize(skb)) { in velocity_xmit() 2572 pktlen = skb_shinfo(skb)->nr_frags == 0 ? in velocity_xmit() 2598 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in velocity_xmit() 2599 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in velocity_xmit()
|
D | via-rhine.c | 1753 …(((unsigned long)skb->data & 3) || skb_shinfo(skb)->nr_frags != 0 || skb->ip_summed == CHECKSUM_PA… in rhine_start_tx()
|
/linux-4.1.27/drivers/net/ethernet/intel/igb/ |
D | igb_main.c | 4695 first->gso_segs = skb_shinfo(skb)->gso_segs; in igb_tso() 4700 mss_l4len_idx |= skb_shinfo(skb)->gso_size << E1000_ADVTXD_MSS_SHIFT; in igb_tso() 4891 for (frag = &skb_shinfo(skb)->frags[0];; frag++) { in igb_tx_map() 5017 for (f = 0; f < skb_shinfo(skb)->nr_frags; f++) in igb_xmit_frame_ring() 5018 count += TXD_USE_COUNT(skb_shinfo(skb)->frags[f].size); in igb_xmit_frame_ring() 5020 count += skb_shinfo(skb)->nr_frags; in igb_xmit_frame_ring() 5034 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) { in igb_xmit_frame_ring() 5039 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in igb_xmit_frame_ring() 6681 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in igb_add_rx_frag() 6840 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0]; in igb_pull_tail()
|
/linux-4.1.27/drivers/net/ethernet/cadence/ |
D | macb.c | 1123 unsigned int f, nr_frags = skb_shinfo(skb)->nr_frags; in macb_tx_map() 1155 const skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; in macb_tx_map() 1265 nr_frags = skb_shinfo(skb)->nr_frags; in macb_start_xmit() 1267 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[f]); in macb_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/alteon/ |
D | acenic.c | 2424 if (!skb_shinfo(skb)->nr_frags) { in ace_start_xmit() 2462 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in ace_start_xmit() 2463 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in ace_start_xmit() 2479 if (i == skb_shinfo(skb)->nr_frags - 1) { in ace_start_xmit()
|
/linux-4.1.27/drivers/net/wireless/b43legacy/ |
D | pio.c | 197 B43legacy_WARN_ON(skb_shinfo(skb)->nr_frags != 0); in pio_tx_write_fragment()
|
D | dma.c | 1024 B43legacy_WARN_ON(skb_shinfo(skb)->nr_frags != 0); in dma_tx_fragment()
|
/linux-4.1.27/drivers/net/ethernet/tundra/ |
D | tsi108_eth.c | 655 int frags = skb_shinfo(skb)->nr_frags + 1; in tsi108_send_packet() 709 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1]; in tsi108_send_packet()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
D | bnx2.c | 2883 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnx2_tx_int() 2942 shinfo = skb_shinfo(skb); in bnx2_reuse_rx_skb_pages() 3069 &skb_shinfo(skb)->frags[i - 1]; in bnx2_rx_skb() 5436 skb_frag_size(&skb_shinfo(skb)->frags[k]), in bnx2_free_tx_skbs() 6583 (skb_shinfo(skb)->nr_frags + 1))) { in bnx2_start_xmit() 6603 if ((mss = skb_shinfo(skb)->gso_size)) { in bnx2_start_xmit() 6611 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) { in bnx2_start_xmit() 6654 last_frag = skb_shinfo(skb)->nr_frags; in bnx2_start_xmit() 6659 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in bnx2_start_xmit() 6728 skb_frag_size(&skb_shinfo(skb)->frags[i]), in bnx2_start_xmit()
|
D | bgmac.c | 159 nr_frags = skb_shinfo(skb)->nr_frags; in bgmac_dma_tx_add() 183 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in bgmac_dma_tx_add()
|
D | tg3.c | 6567 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in tg3_tx() 6574 skb_frag_size(&skb_shinfo(skb)->frags[i]), in tg3_tx() 7764 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in tg3_tx_skb_unmap() 7841 return skb_shinfo(skb)->gso_segs < tnapi->tx_pending / 3; in tg3_tso_bug_gso_check() 7853 u32 frag_cnt_est = skb_shinfo(skb)->gso_segs * 3; in tg3_tso_bug() 7917 if (unlikely(budget <= (skb_shinfo(skb)->nr_frags + 1))) { in tg3_start_xmit() 7931 mss = skb_shinfo(skb)->gso_size; in tg3_start_xmit() 8027 if ((unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) && in tg3_start_xmit() 8029 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in tg3_start_xmit() 8049 ((skb_shinfo(skb)->nr_frags == 0) ? TXD_FLAG_END : 0), in tg3_start_xmit() [all …]
|
/linux-4.1.27/net/can/ |
D | raw.c | 758 sock_tx_timestamp(sk, &skb_shinfo(skb)->tx_flags); in raw_sendmsg()
|
/linux-4.1.27/drivers/usb/gadget/function/ |
D | f_phonet.c | 349 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in pn_rx_complete()
|
/linux-4.1.27/net/appletalk/ |
D | ddp.c | 950 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in atalk_sum_skb() 952 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in atalk_sum_skb()
|
/linux-4.1.27/drivers/net/ethernet/nvidia/ |
D | forcedeth.c | 2190 unsigned int fragments = skb_shinfo(skb)->nr_frags; in nv_start_xmit() 2206 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit() 2256 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in nv_start_xmit() 2309 tx_flags_extra = NV_TX2_TSO | (skb_shinfo(skb)->gso_size << NV_TX2_TSO_SHIFT); in nv_start_xmit() 2337 unsigned int fragments = skb_shinfo(skb)->nr_frags; in nv_start_xmit_optimized() 2354 u32 frag_size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in nv_start_xmit_optimized() 2406 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in nv_start_xmit_optimized() 2459 tx_flags_extra = NV_TX2_TSO | (skb_shinfo(skb)->gso_size << NV_TX2_TSO_SHIFT); in nv_start_xmit_optimized()
|
/linux-4.1.27/drivers/net/ethernet/ibm/ehea/ |
D | ehea_main.c | 1644 swqe->mss = skb_shinfo(skb)->gso_size; in write_swqe2_immediate() 1676 nfrags = skb_shinfo(skb)->nr_frags; in write_swqe2_data() 1687 frag = &skb_shinfo(skb)->frags[0]; in write_swqe2_data() 1700 frag = &skb_shinfo(skb)->frags[i]; in write_swqe2_data()
|
/linux-4.1.27/drivers/net/ethernet/myricom/myri10ge/ |
D | myri10ge.c | 1410 frag = skb_shinfo(skb)->frags; in myri10ge_vlan_rx() 1464 rx_frags = skb_shinfo(skb)->frags; in myri10ge_rx_done() 2846 mss = skb_shinfo(skb)->gso_size; in myri10ge_xmit() 2929 frag_cnt = skb_shinfo(skb)->nr_frags; in myri10ge_xmit() 3024 frag = &skb_shinfo(skb)->frags[frag_idx]; in myri10ge_xmit()
|
/linux-4.1.27/drivers/net/ethernet/calxeda/ |
D | xgmac.c | 1090 int nfrags = skb_shinfo(skb)->nr_frags; in xgmac_xmit() 1115 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xgmac_xmit()
|
/linux-4.1.27/drivers/scsi/libfc/ |
D | fc_fcp.c | 651 skb_shinfo(fp_skb(fp))->nr_frags, in fc_fcp_send_data() 672 if ((skb_shinfo(fp_skb(fp))->nr_frags < FC_FRAME_SG_LEN) && in fc_fcp_send_data()
|
/linux-4.1.27/net/wireless/ |
D | wext-core.c | 628 skb_shinfo(skb)->frag_list = compskb; in wireless_send_event()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/ |
D | qlge_main.c | 1318 int frag_cnt = skb_shinfo(skb)->nr_frags; in ql_map_send() 1351 skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_idx]; in ql_map_send() 1519 __skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in ql_process_mac_rx_gro_page() 1527 skb_shinfo(skb)->nr_frags++; in ql_process_mac_rx_gro_page() 2580 mac_iocb_ptr->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in ql_tso()
|
/linux-4.1.27/drivers/net/ethernet/brocade/bna/ |
D | bnad.c | 559 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, in bnad_cq_setup_skb_frags() 2840 gso_size = skb_shinfo(skb)->gso_size; in bnad_txq_wi_prepare() 2980 vectors = 1 + skb_shinfo(skb)->nr_frags; in bnad_start_xmit() 3043 const struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in bnad_start_xmit()
|
/linux-4.1.27/drivers/s390/net/ |
D | qeth_core_main.c | 3869 for (cnt = 0; cnt < skb_shinfo(skb)->nr_frags; cnt++) { in qeth_get_elements_for_frags() 3870 frag = &skb_shinfo(skb)->frags[cnt]; in qeth_get_elements_for_frags() 3954 if (skb_shinfo(skb)->nr_frags) in __qeth_fill_buffer() 3975 for (cnt = 0; cnt < skb_shinfo(skb)->nr_frags; cnt++) { in __qeth_fill_buffer() 3976 frag = &skb_shinfo(skb)->frags[cnt]; in __qeth_fill_buffer() 5307 card->perf_stats.sg_frags_rx += skb_shinfo(skb)->nr_frags; in qeth_core_get_next_skb()
|
D | qeth_l3_main.c | 2909 hdr->ext.mss = skb_shinfo(skb)->gso_size; in qeth_tso_fill_header() 2984 (skb_shinfo(skb)->nr_frags == 0)) { in qeth_l3_hard_start_xmit() 3071 nr_frags = skb_shinfo(new_skb)->nr_frags; in qeth_l3_hard_start_xmit()
|
/linux-4.1.27/net/netlink/ |
D | af_netlink.c | 895 !atomic_dec_return(&(skb_shinfo(skb)->dataref))) in netlink_skb_destructor() 2438 if (unlikely(skb_shinfo(skb)->frag_list)) { in netlink_recvmsg() 2450 data_skb = skb_shinfo(skb)->frag_list; in netlink_recvmsg()
|
/linux-4.1.27/net/ |
D | Kconfig | 41 achieve this, you need to set skb_shinfo(skb)->frag_list to the
|
/linux-4.1.27/drivers/net/ethernet/natsemi/ |
D | ns83820.c | 1074 nr_frags = skb_shinfo(skb)->nr_frags; 1112 frag = skb_shinfo(skb)->frags;
|
/linux-4.1.27/drivers/net/ethernet/qlogic/ |
D | qla3xxx.c | 2317 int frag_cnt = (int)skb_shinfo(skb)->nr_frags; in ql_send_map() 2350 skb_frag_t *frag = &skb_shinfo(skb)->frags[completed_segs]; in ql_send_map() 2476 skb_shinfo(skb)->nr_frags); in ql3xxx_send()
|
/linux-4.1.27/drivers/net/wireless/mwifiex/ |
D | main.c | 760 skb_shinfo(skb)->tx_flags & SKBTX_WIFI_STATUS && in mwifiex_hard_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/faraday/ |
D | ftgmac100.c | 466 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, 0, size); in ftgmac100_rx_packet()
|
/linux-4.1.27/net/iucv/ |
D | af_iucv.c | 2247 if (skb_shinfo(list_skb) == skb_shinfo(skb)) { in afiucv_hs_callback_txnotify()
|
/linux-4.1.27/drivers/net/ethernet/altera/ |
D | altera_tse_main.c | 572 int nfrags = skb_shinfo(skb)->nr_frags; in tse_start_xmit()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/ |
D | bcmgenet.c | 1301 nr_frags = skb_shinfo(skb)->nr_frags; in bcmgenet_xmit() 1342 &skb_shinfo(skb)->frags[i], in bcmgenet_xmit()
|
/linux-4.1.27/drivers/net/ethernet/ibm/emac/ |
D | core.c | 1449 int nr_frags = skb_shinfo(skb)->nr_frags; in emac_start_xmit_sg() 1484 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i]; in emac_start_xmit_sg()
|
/linux-4.1.27/drivers/net/ethernet/xscale/ |
D | ixp4xx_eth.c | 334 shtx = skb_shinfo(skb); in ixp_tx_timestamp()
|
/linux-4.1.27/drivers/net/phy/ |
D | dp83640.c | 1442 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in dp83640_txtstamp()
|
/linux-4.1.27/drivers/scsi/cxgbi/cxgb3i/ |
D | cxgb3i.c | 408 int frags = skb_shinfo(skb)->nr_frags + (len != skb->data_len); in push_tx_frames()
|
/linux-4.1.27/drivers/net/ethernet/octeon/ |
D | octeon_mgmt.c | 1282 re.s.tstamp = ((skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) != 0); in octeon_mgmt_xmit()
|
/linux-4.1.27/drivers/net/ethernet/rocker/ |
D | rocker.c | 3963 if (skb_shinfo(skb)->nr_frags > ROCKER_TX_FRAGS_MAX) in rocker_port_xmit() 3966 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in rocker_port_xmit() 3967 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in rocker_port_xmit()
|
/linux-4.1.27/drivers/net/ethernet/agere/ |
D | et131x.c | 2427 u32 nr_frags = skb_shinfo(skb)->nr_frags + 1; in nic_send_packet() 2428 struct skb_frag_struct *frags = &skb_shinfo(skb)->frags[0]; in nic_send_packet()
|
/linux-4.1.27/drivers/net/ethernet/micrel/ |
D | ksz884x.c | 4648 int last_frag = skb_shinfo(skb)->nr_frags; in send_packet() 4679 this_frag = &skb_shinfo(skb)->frags[frag]; in send_packet() 4869 num = skb_shinfo(skb)->nr_frags + 1; in netdev_tx()
|
/linux-4.1.27/net/bluetooth/ |
D | hci_core.c | 3456 list = skb_shinfo(skb)->frag_list; in hci_queue_acl() 3466 skb_shinfo(skb)->frag_list = NULL; in hci_queue_acl()
|
D | l2cap_core.c | 2130 frag = &skb_shinfo(skb)->frag_list; in l2cap_skbuff_fromiovec() 2916 frag = &skb_shinfo(skb)->frag_list; in l2cap_build_cmd() 5789 skb_shinfo(skb)->frag_list = new_frag; in append_skb_frag()
|
/linux-4.1.27/drivers/scsi/cxgbi/cxgb4i/ |
D | cxgb4i.c | 494 cnt = skb_shinfo(skb)->nr_frags; in calc_tx_flits_ofld()
|
/linux-4.1.27/drivers/net/ppp/ |
D | ppp_generic.c | 2196 struct sk_buff **fragpp = &skb_shinfo(skb)->frag_list; in ppp_mp_reconstruct()
|
/linux-4.1.27/drivers/net/ethernet/oki-semi/pch_gbe/ |
D | pch_gbe_main.c | 195 shtx = skb_shinfo(skb); in pch_tx_timestamp()
|