Lines Matching refs:hldev

130 	if (vpath->hldev->first_vp_id != vpath->vp_id)  in vxge_hw_vpath_intr_enable()
337 u32 vxge_hw_device_set_intr_type(struct __vxge_hw_device *hldev, u32 intr_mode) in vxge_hw_device_set_intr_type() argument
346 hldev->config.intr_mode = intr_mode; in vxge_hw_device_set_intr_type()
359 void vxge_hw_device_intr_enable(struct __vxge_hw_device *hldev) in vxge_hw_device_intr_enable() argument
365 vxge_hw_device_mask_all(hldev); in vxge_hw_device_intr_enable()
369 if (!(hldev->vpaths_deployed & vxge_mBIT(i))) in vxge_hw_device_intr_enable()
373 VXGE_HW_VIRTUAL_PATH_HANDLE(&hldev->virtual_paths[i])); in vxge_hw_device_intr_enable()
376 if (hldev->config.intr_mode == VXGE_HW_INTR_MODE_IRQLINE) { in vxge_hw_device_intr_enable()
377 val64 = hldev->tim_int_mask0[VXGE_HW_VPATH_INTR_TX] | in vxge_hw_device_intr_enable()
378 hldev->tim_int_mask0[VXGE_HW_VPATH_INTR_RX]; in vxge_hw_device_intr_enable()
381 writeq(val64, &hldev->common_reg->tim_int_status0); in vxge_hw_device_intr_enable()
383 writeq(~val64, &hldev->common_reg->tim_int_mask0); in vxge_hw_device_intr_enable()
386 val32 = hldev->tim_int_mask1[VXGE_HW_VPATH_INTR_TX] | in vxge_hw_device_intr_enable()
387 hldev->tim_int_mask1[VXGE_HW_VPATH_INTR_RX]; in vxge_hw_device_intr_enable()
391 &hldev->common_reg->tim_int_status1); in vxge_hw_device_intr_enable()
394 &hldev->common_reg->tim_int_mask1); in vxge_hw_device_intr_enable()
398 val64 = readq(&hldev->common_reg->titan_general_int_status); in vxge_hw_device_intr_enable()
400 vxge_hw_device_unmask_all(hldev); in vxge_hw_device_intr_enable()
411 void vxge_hw_device_intr_disable(struct __vxge_hw_device *hldev) in vxge_hw_device_intr_disable() argument
415 vxge_hw_device_mask_all(hldev); in vxge_hw_device_intr_disable()
418 writeq(VXGE_HW_INTR_MASK_ALL, &hldev->common_reg->tim_int_mask0); in vxge_hw_device_intr_disable()
420 &hldev->common_reg->tim_int_mask1); in vxge_hw_device_intr_disable()
424 if (!(hldev->vpaths_deployed & vxge_mBIT(i))) in vxge_hw_device_intr_disable()
428 VXGE_HW_VIRTUAL_PATH_HANDLE(&hldev->virtual_paths[i])); in vxge_hw_device_intr_disable()
440 void vxge_hw_device_mask_all(struct __vxge_hw_device *hldev) in vxge_hw_device_mask_all() argument
448 &hldev->common_reg->titan_mask_all_int); in vxge_hw_device_mask_all()
459 void vxge_hw_device_unmask_all(struct __vxge_hw_device *hldev) in vxge_hw_device_unmask_all() argument
463 if (hldev->config.intr_mode == VXGE_HW_INTR_MODE_IRQLINE) in vxge_hw_device_unmask_all()
467 &hldev->common_reg->titan_mask_all_int); in vxge_hw_device_unmask_all()
478 void vxge_hw_device_flush_io(struct __vxge_hw_device *hldev) in vxge_hw_device_flush_io() argument
480 readl(&hldev->common_reg->titan_general_int_status); in vxge_hw_device_flush_io()
492 __vxge_hw_device_handle_error(struct __vxge_hw_device *hldev, u32 vp_id, in __vxge_hw_device_handle_error() argument
524 if (hldev->uld_callbacks->crit_err) in __vxge_hw_device_handle_error()
525 hldev->uld_callbacks->crit_err(hldev, in __vxge_hw_device_handle_error()
540 __vxge_hw_device_handle_link_down_ind(struct __vxge_hw_device *hldev) in __vxge_hw_device_handle_link_down_ind() argument
545 if (hldev->link_state == VXGE_HW_LINK_DOWN) in __vxge_hw_device_handle_link_down_ind()
548 hldev->link_state = VXGE_HW_LINK_DOWN; in __vxge_hw_device_handle_link_down_ind()
551 if (hldev->uld_callbacks->link_down) in __vxge_hw_device_handle_link_down_ind()
552 hldev->uld_callbacks->link_down(hldev); in __vxge_hw_device_handle_link_down_ind()
565 __vxge_hw_device_handle_link_up_ind(struct __vxge_hw_device *hldev) in __vxge_hw_device_handle_link_up_ind() argument
570 if (hldev->link_state == VXGE_HW_LINK_UP) in __vxge_hw_device_handle_link_up_ind()
573 hldev->link_state = VXGE_HW_LINK_UP; in __vxge_hw_device_handle_link_up_ind()
576 if (hldev->uld_callbacks->link_up) in __vxge_hw_device_handle_link_up_ind()
577 hldev->uld_callbacks->link_up(hldev); in __vxge_hw_device_handle_link_up_ind()
597 struct __vxge_hw_device *hldev = NULL; in __vxge_hw_vpath_alarm_process() local
609 hldev = vpath->hldev; in __vxge_hw_vpath_alarm_process()
657 __vxge_hw_device_handle_link_down_ind(hldev); in __vxge_hw_vpath_alarm_process()
678 __vxge_hw_device_handle_link_up_ind(hldev); in __vxge_hw_vpath_alarm_process()
843 hldev->stats.sw_dev_err_stats.vpath_alarms++; in __vxge_hw_vpath_alarm_process()
849 __vxge_hw_device_handle_error(hldev, vpath->vp_id, alarm_event); in __vxge_hw_vpath_alarm_process()
879 enum vxge_hw_status vxge_hw_device_begin_irq(struct __vxge_hw_device *hldev, in vxge_hw_device_begin_irq() argument
888 val64 = readq(&hldev->common_reg->titan_general_int_status); in vxge_hw_device_begin_irq()
899 adapter_status = readq(&hldev->common_reg->adapter_status); in vxge_hw_device_begin_irq()
903 __vxge_hw_device_handle_error(hldev, in vxge_hw_device_begin_irq()
911 hldev->stats.sw_dev_info_stats.total_intr_cnt++; in vxge_hw_device_begin_irq()
915 vpath_mask = hldev->vpaths_deployed >> in vxge_hw_device_begin_irq()
920 hldev->stats.sw_dev_info_stats.traffic_intr_cnt++; in vxge_hw_device_begin_irq()
925 hldev->stats.sw_dev_info_stats.not_traffic_intr_cnt++; in vxge_hw_device_begin_irq()
932 hldev->stats.sw_dev_err_stats.vpath_alarms++; in vxge_hw_device_begin_irq()
936 if (!(hldev->vpaths_deployed & vxge_mBIT(i))) in vxge_hw_device_begin_irq()
940 &hldev->virtual_paths[i], skip_alarms); in vxge_hw_device_begin_irq()
965 void vxge_hw_device_clear_tx_rx(struct __vxge_hw_device *hldev) in vxge_hw_device_clear_tx_rx() argument
968 if ((hldev->tim_int_mask0[VXGE_HW_VPATH_INTR_TX] != 0) || in vxge_hw_device_clear_tx_rx()
969 (hldev->tim_int_mask0[VXGE_HW_VPATH_INTR_RX] != 0)) { in vxge_hw_device_clear_tx_rx()
970 writeq((hldev->tim_int_mask0[VXGE_HW_VPATH_INTR_TX] | in vxge_hw_device_clear_tx_rx()
971 hldev->tim_int_mask0[VXGE_HW_VPATH_INTR_RX]), in vxge_hw_device_clear_tx_rx()
972 &hldev->common_reg->tim_int_status0); in vxge_hw_device_clear_tx_rx()
975 if ((hldev->tim_int_mask1[VXGE_HW_VPATH_INTR_TX] != 0) || in vxge_hw_device_clear_tx_rx()
976 (hldev->tim_int_mask1[VXGE_HW_VPATH_INTR_RX] != 0)) { in vxge_hw_device_clear_tx_rx()
978 (hldev->tim_int_mask1[VXGE_HW_VPATH_INTR_TX] | in vxge_hw_device_clear_tx_rx()
979 hldev->tim_int_mask1[VXGE_HW_VPATH_INTR_RX]), in vxge_hw_device_clear_tx_rx()
980 &hldev->common_reg->tim_int_status1); in vxge_hw_device_clear_tx_rx()
1979 if (!(vpath->hldev->access_rights & in vxge_hw_vpath_promisc_enable()
2182 (vpath->hldev->first_vp_id * 4) + alarm_msix_id), in vxge_hw_vpath_msix_set()
2185 if (vpath->hldev->config.intr_mode == in vxge_hw_vpath_msix_set()
2214 struct __vxge_hw_device *hldev = vp->vpath->hldev; in vxge_hw_vpath_msix_mask() local
2217 &hldev->common_reg->set_msix_mask_vect[msix_id % 4]); in vxge_hw_vpath_msix_mask()
2234 struct __vxge_hw_device *hldev = vp->vpath->hldev; in vxge_hw_vpath_msix_clear() local
2236 if (hldev->config.intr_mode == VXGE_HW_INTR_MODE_MSIX_ONE_SHOT) in vxge_hw_vpath_msix_clear()
2239 &hldev->common_reg->clr_msix_one_shot_vec[msix_id % 4]); in vxge_hw_vpath_msix_clear()
2243 &hldev->common_reg->clear_msix_mask_vect[msix_id % 4]); in vxge_hw_vpath_msix_clear()
2261 struct __vxge_hw_device *hldev = vp->vpath->hldev; in vxge_hw_vpath_msix_unmask() local
2264 &hldev->common_reg->clear_msix_mask_vect[msix_id%4]); in vxge_hw_vpath_msix_unmask()
2280 struct __vxge_hw_device *hldev = vp->vpath->hldev; in vxge_hw_vpath_inta_mask_tx_rx() local
2285 val64 = readq(&hldev->common_reg->tim_int_mask0); in vxge_hw_vpath_inta_mask_tx_rx()
2291 &hldev->common_reg->tim_int_mask0); in vxge_hw_vpath_inta_mask_tx_rx()
2294 val64 = readl(&hldev->common_reg->tim_int_mask1); in vxge_hw_vpath_inta_mask_tx_rx()
2301 &hldev->common_reg->tim_int_mask1); in vxge_hw_vpath_inta_mask_tx_rx()
2318 struct __vxge_hw_device *hldev = vp->vpath->hldev; in vxge_hw_vpath_inta_unmask_tx_rx() local
2323 val64 = readq(&hldev->common_reg->tim_int_mask0); in vxge_hw_vpath_inta_unmask_tx_rx()
2329 &hldev->common_reg->tim_int_mask0); in vxge_hw_vpath_inta_unmask_tx_rx()
2337 &hldev->common_reg->tim_int_mask1); in vxge_hw_vpath_inta_unmask_tx_rx()