Lines Matching refs:vpaths

127 		VXGE_COMPLETE_VPATH_TX(&vdev->vpaths[i].fifo);  in VXGE_COMPLETE_ALL_TX()
137 ring = &vdev->vpaths[i].ring; in VXGE_COMPLETE_ALL_RX()
711 vpath = &vdev->vpaths[mac->vpath_no]; in vxge_add_mac_addr()
739 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
749 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
765 vpath = &vdev->vpaths[vpath_idx]; in vxge_learn_mac()
861 fifo = &vdev->vpaths[vpath_no].fifo; in vxge_xmit()
1098 vpath = &vdev->vpaths[mac->vpath_no]; in vxge_del_mac_addr()
1145 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1155 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1168 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1187 mcast_cnt = vdev->vpaths[0].mcast_addr_cnt; in vxge_set_multicast()
1188 list_head = &vdev->vpaths[0].mac_addr_list; in vxge_set_multicast()
1190 (vdev->vpaths[0].mac_addr_cnt - mcast_cnt)) > in vxge_set_multicast()
1191 vdev->vpaths[0].max_mac_addr_cnt) in vxge_set_multicast()
1235 mcast_cnt = vdev->vpaths[0].mcast_addr_cnt; in vxge_set_multicast()
1257 vpath = &vdev->vpaths[i]; in vxge_set_multicast()
1312 struct vxge_vpath *vpath = &vdev->vpaths[vpath_idx]; in vxge_set_mac_addr()
1358 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_vpath_intr_enable()
1391 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_vpath_intr_disable()
1510 struct vxge_vpath *vpath = &vdev->vpaths[vp_id]; in vxge_reset_vpath()
1587 hw_ring = vdev->vpaths[i].ring.handle; in vxge_config_ci_for_tti_rti()
1594 struct __vxge_hw_fifo *hw_fifo = vdev->vpaths[i].fifo.handle; in vxge_config_ci_for_tti_rti()
1724 if (vdev->vpaths[i].handle) { in do_vxge_reset()
1726 vdev->vpaths[i].handle) in do_vxge_reset()
1747 vxge_restore_vpath_mac_addr(&vdev->vpaths[vp_id]); in do_vxge_reset()
1748 vxge_restore_vpath_vid_table(&vdev->vpaths[vp_id]); in do_vxge_reset()
1764 vxge_hw_vpath_enable(vdev->vpaths[i].handle); in do_vxge_reset()
1766 vxge_hw_vpath_rx_doorbell_init(vdev->vpaths[i].handle); in do_vxge_reset()
1850 ring = &vdev->vpaths[i].ring; in vxge_poll_inta()
1935 "for vpath:%d", vdev->vpaths[0].device_id); in vxge_rth_configure()
1956 vdev->vpaths[index].handle, in vxge_rth_configure()
1963 vdev->vpaths[index].device_id); in vxge_rth_configure()
1979 vpath = &vdev->vpaths[i]; in vxge_reset_all_vpaths()
2010 vpath = &vdev->vpaths[i]; in vxge_close_vpaths()
2031 vpath = &vdev->vpaths[i]; in vxge_open_vpaths()
2272 vxge_hw_vpath_msix_mask(vdev->vpaths[i].handle, msix_id); in vxge_alarm_msix_handle()
2273 vxge_hw_vpath_msix_clear(vdev->vpaths[i].handle, msix_id); in vxge_alarm_msix_handle()
2275 status = vxge_hw_vpath_alarm_process(vdev->vpaths[i].handle, in vxge_alarm_msix_handle()
2278 vxge_hw_vpath_msix_unmask(vdev->vpaths[i].handle, in vxge_alarm_msix_handle()
2393 struct vxge_vpath *vpath = &vdev->vpaths[i]; in vxge_enable_msix()
2482 &vdev->vpaths[vp_idx].fifo); in vxge_add_isr()
2484 &vdev->vpaths[vp_idx].fifo; in vxge_add_isr()
2497 &vdev->vpaths[vp_idx].ring); in vxge_add_isr()
2499 &vdev->vpaths[vp_idx].ring; in vxge_add_isr()
2519 msix_idx += vdev->vpaths[vp_idx].device_id * in vxge_add_isr()
2522 vdev->vpaths[vp_idx].handle, in vxge_add_isr()
2543 &vdev->vpaths[0]); in vxge_add_isr()
2556 msix_idx = (vdev->vpaths[0].handle->vpath->vp_id * in vxge_add_isr()
2558 vxge_hw_vpath_msix_unmask(vdev->vpaths[vp_idx].handle, in vxge_add_isr()
2561 vdev->vxge_entries[intr_cnt].arg = &vdev->vpaths[0]; in vxge_add_isr()
2571 vxge_hw_vpath_tti_ci_set(vdev->vpaths[0].fifo.handle); in vxge_add_isr()
2619 ring = &vdev->vpaths[i].ring; in vxge_poll_vp_lockup()
2634 vpath = &vdev->vpaths[i]; in vxge_poll_vp_lockup()
2742 vpath = &vdev->vpaths[i]; in vxge_open()
2747 vpath = &vdev->vpaths[i]; in vxge_open()
2770 vpath = &vdev->vpaths[i]; in vxge_open()
2792 vpath = &vdev->vpaths[i]; in vxge_open()
2830 vpath = &vdev->vpaths[i]; in vxge_open()
2873 vpath = &vdev->vpaths[i]; in vxge_open()
2895 napi_disable(&vdev->vpaths[i].ring.napi); in vxge_open()
2928 netif_napi_del(&vdev->vpaths[i].ring.napi); in vxge_napi_del_all()
2955 vpath_vector = vxge_mBIT(vdev->vpaths[0].device_id); in do_vxge_close()
3007 napi_disable(&vdev->vpaths[i].ring.napi); in do_vxge_close()
3111 struct vxge_ring_stats *rxstats = &vdev->vpaths[k].ring.stats; in vxge_get_stats64()
3112 struct vxge_fifo_stats *txstats = &vdev->vpaths[k].fifo.stats; in vxge_get_stats64()
3223 vdev->vpaths[i].ring.rx_hwts = vdev->rx_hwts; in vxge_hwtstamp_set()
3311 vpath = &vdev->vpaths[vp_id]; in vxge_vlan_rx_add_vid()
3339 vpath = &vdev->vpaths[vp_id]; in vxge_vlan_rx_kill_vid()
3430 vdev->vpaths = kcalloc(no_of_vpath, sizeof(struct vxge_vpath), in vxge_device_register()
3432 if (!vdev->vpaths) { in vxge_device_register()
3497 kfree(vdev->vpaths); in vxge_device_register()
3528 kfree(vdev->vpaths); in vxge_device_unregister()
3563 vpath = &vdev->vpaths[vpath_idx]; in vxge_callback_crit_err()
4606 vdev->vpaths[j].is_configured = 1; in vxge_probe()
4607 vdev->vpaths[j].device_id = i; in vxge_probe()
4608 vdev->vpaths[j].ring.driver_id = j; in vxge_probe()
4609 vdev->vpaths[j].vdev = vdev; in vxge_probe()
4610 vdev->vpaths[j].max_mac_addr_cnt = max_mac_vpath; in vxge_probe()
4611 memcpy((u8 *)vdev->vpaths[j].macaddr, in vxge_probe()
4616 INIT_LIST_HEAD(&vdev->vpaths[j].mac_addr_list); in vxge_probe()
4618 vdev->vpaths[j].mac_addr_cnt = 0; in vxge_probe()
4619 vdev->vpaths[j].mcast_addr_cnt = 0; in vxge_probe()
4631 macaddr = (u8 *)vdev->vpaths[0].macaddr; in vxge_probe()
4682 memcpy(vdev->ndev->dev_addr, (u8 *)vdev->vpaths[0].macaddr, ETH_ALEN); in vxge_probe()
4696 list_add(&entry->item, &vdev->vpaths[i].mac_addr_list); in vxge_probe()
4697 vdev->vpaths[i].mac_addr_cnt = 1; in vxge_probe()
4734 vxge_free_mac_add_list(&vdev->vpaths[i]); in vxge_probe()
4777 vxge_free_mac_add_list(&vdev->vpaths[i]); in vxge_remove()