Lines Matching refs:vf
160 #define vf_rxq_count(vf) ((vf)->alloc_resc.num_rxqs) argument
161 #define vf_txq_count(vf) ((vf)->alloc_resc.num_txqs) argument
162 #define vf_sb_count(vf) ((vf)->alloc_resc.num_sbs) argument
163 #define vf_mac_rules_cnt(vf) ((vf)->alloc_resc.num_mac_filters) argument
164 #define vf_vlan_rules_cnt(vf) ((vf)->alloc_resc.num_vlan_filters) argument
165 #define vf_mc_rules_cnt(vf) ((vf)->alloc_resc.num_mc_filters) argument
167 #define vf_vlan_rules_visible_cnt(vf) (vf_vlan_rules_cnt(vf) - 1) argument
175 #define bnx2x_vfq(vf, nr, var) ((vf)->vfqs[(nr)].var) argument
176 #define bnx2x_leading_vfq(vf, var) ((vf)->vfqs[LEADING_IDX].var) argument
217 #define for_each_vfq(vf, var) \ argument
218 for ((var) = 0; (var) < vf_rxq_count(vf); (var)++)
220 #define for_each_vf_sb(vf, var) \ argument
221 for ((var) = 0; (var) < vf_sb_count(vf); (var)++)
223 #define is_vf_multi(vf) (vf_rxq_count(vf) > 1) argument
234 void bnx2x_lock_vf_pf_channel(struct bnx2x *bp, struct bnx2x_virtf *vf,
237 void bnx2x_unlock_vf_pf_channel(struct bnx2x *bp, struct bnx2x_virtf *vf,
324 #define BP_VF_BULLETIN(bp, vf) \ argument
326 + (vf))
329 #define bnx2x_vf_sp(bp, vf, field) ((bp)->vfdb->sp_dma.addr + \ argument
330 (vf)->index * sizeof(struct bnx2x_vf_sp) + \
332 #define bnx2x_vf_sp_map(bp, vf, field) ((bp)->vfdb->sp_dma.mapping + \ argument
333 (vf)->index * sizeof(struct bnx2x_vf_sp) + \
352 static inline struct bnx2x_vf_queue *vfq_get(struct bnx2x_virtf *vf, u8 index) in vfq_get() argument
354 return &(vf->vfqs[index]); in vfq_get()
358 static inline u8 vf_igu_sb(struct bnx2x_virtf *vf, u16 sb_idx) in vf_igu_sb() argument
360 return vf->igu_base_id + sb_idx; in vf_igu_sb()
363 static inline u8 vf_hc_qzone(struct bnx2x_virtf *vf, u16 sb_idx) in vf_hc_qzone() argument
365 return vf_igu_sb(vf, sb_idx); in vf_hc_qzone()
368 static u8 vfq_cl_id(struct bnx2x_virtf *vf, struct bnx2x_vf_queue *q) in vfq_cl_id() argument
370 return vf->igu_base_id + q->index; in vfq_cl_id()
373 static inline u8 vfq_stat_id(struct bnx2x_virtf *vf, struct bnx2x_vf_queue *q) in vfq_stat_id() argument
375 if (vf->cfg_flags & VF_CFG_STATS_COALESCE) in vfq_stat_id()
376 return vf->leading_rss; in vfq_stat_id()
378 return vfq_cl_id(vf, q); in vfq_stat_id()
381 static inline u8 vfq_qzone_id(struct bnx2x_virtf *vf, struct bnx2x_vf_queue *q) in vfq_qzone_id() argument
383 return vfq_cl_id(vf, q); in vfq_qzone_id()
411 int bnx2x_vf_acquire(struct bnx2x *bp, struct bnx2x_virtf *vf,
414 int bnx2x_vf_init(struct bnx2x *bp, struct bnx2x_virtf *vf,
418 void bnx2x_vfop_qctor_dump_tx(struct bnx2x *bp, struct bnx2x_virtf *vf,
423 void bnx2x_vfop_qctor_dump_rx(struct bnx2x *bp, struct bnx2x_virtf *vf,
429 struct bnx2x_virtf *vf,
434 int bnx2x_vf_mac_vlan_config_list(struct bnx2x *bp, struct bnx2x_virtf *vf,
438 int bnx2x_vf_queue_setup(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid,
441 int bnx2x_vf_queue_teardown(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid);
443 int bnx2x_vf_mcast(struct bnx2x *bp, struct bnx2x_virtf *vf,
446 int bnx2x_vf_rxmode(struct bnx2x *bp, struct bnx2x_virtf *vf,
449 int bnx2x_vf_close(struct bnx2x *bp, struct bnx2x_virtf *vf);
451 int bnx2x_vf_free(struct bnx2x *bp, struct bnx2x_virtf *vf);
453 int bnx2x_vf_rss_update(struct bnx2x *bp, struct bnx2x_virtf *vf,
456 int bnx2x_vf_tpa_update(struct bnx2x *bp, struct bnx2x_virtf *vf,
465 int bnx2x_vf_release(struct bnx2x *bp, struct bnx2x_virtf *vf);
467 u8 bnx2x_vf_max_queue_cnt(struct bnx2x *bp, struct bnx2x_virtf *vf);
481 int bnx2x_post_vf_bulletin(struct bnx2x *bp, int vf);
537 int bnx2x_set_vf_link_state(struct net_device *dev, int vf, int link_state);
601 static inline int bnx2x_set_vf_link_state(struct net_device *dev, int vf, in bnx2x_set_vf_link_state() argument