Searched refs:softnet_data (Results 1 – 12 of 12) sorted by relevance
119 static u32 softnet_backlog_len(struct softnet_data *sd) in softnet_backlog_len()125 static struct softnet_data *softnet_get_online(loff_t *pos) in softnet_get_online()127 struct softnet_data *sd = NULL; in softnet_get_online()131 sd = &per_cpu(softnet_data, *pos); in softnet_get_online()155 struct softnet_data *sd = v; in softnet_seq_show()
222 static inline void rps_lock(struct softnet_data *sd) in rps_lock()229 static inline void rps_unlock(struct softnet_data *sd) in rps_unlock()403 DEFINE_PER_CPU_ALIGNED(struct softnet_data, softnet_data);404 EXPORT_PER_CPU_SYMBOL(softnet_data);3048 struct softnet_data *sd; in __netif_reschedule()3052 sd = this_cpu_ptr(&softnet_data); in __netif_reschedule()3116 skb->next = __this_cpu_read(softnet_data.completion_queue); in __dev_kfree_skb_irq()3117 __this_cpu_write(softnet_data.completion_queue, skb); in __dev_kfree_skb_irq()4277 static inline void ____napi_schedule(struct softnet_data *sd, in ____napi_schedule()4335 per_cpu(softnet_data, next_cpu).input_queue_head; in set_rps_cpu()[all …]
123 struct softnet_data *sd; in flow_limit_cpu_sysctl()138 sd = &per_cpu(softnet_data, i); in flow_limit_cpu_sysctl()170 sd = &per_cpu(softnet_data, i); in flow_limit_cpu_sysctl()
246 struct softnet_data *sd = &get_cpu_var(softnet_data); in zap_completion_queue()268 put_cpu_var(softnet_data); in zap_completion_queue()
117 struct softnet_data *sd; in validate_xmit_xfrm()140 sd = this_cpu_ptr(&softnet_data); in validate_xmit_xfrm()334 struct softnet_data *sd; in xfrm_dev_resume()347 sd = this_cpu_ptr(&softnet_data); in xfrm_dev_resume()356 void xfrm_dev_backlog(struct softnet_data *sd) in xfrm_dev_backlog()
3267 struct softnet_data { struct3276 struct softnet_data *rps_ipi_list; argument3300 struct softnet_data *rps_ipi_next; argument3310 static inline void input_queue_head_incr(struct softnet_data *sd) in input_queue_head_incr() argument3317 static inline void input_queue_tail_incr_save(struct softnet_data *sd, in input_queue_tail_incr_save()3325 DECLARE_PER_CPU_ALIGNED(struct softnet_data, softnet_data);3329 return this_cpu_read(softnet_data.xmit.recursion); in dev_recursion_level()3335 return unlikely(__this_cpu_read(softnet_data.xmit.recursion) > in dev_xmit_recursion()3341 __this_cpu_inc(softnet_data.xmit.recursion); in dev_xmit_recursion_inc()3346 __this_cpu_dec(softnet_data.xmit.recursion); in dev_xmit_recursion_dec()[all …]
1848 void xfrm_dev_backlog(struct softnet_data *sd);1905 static inline void xfrm_dev_backlog(struct softnet_data *sd) in xfrm_dev_backlog()
1380 softnet_data
1787 softnet_data
1847 softnet_data
2346 softnet_data
3832 softnet_data