| /OK3568_Linux_fs/kernel/drivers/base/regmap/ |
| H A D | regcache-rbtree.c | 41 struct regcache_rbtree_node *rbnode, in regcache_rbtree_get_base_top_reg() argument 44 *base = rbnode->base_reg; in regcache_rbtree_get_base_top_reg() 45 *top = rbnode->base_reg + ((rbnode->blklen - 1) * map->reg_stride); in regcache_rbtree_get_base_top_reg() 49 struct regcache_rbtree_node *rbnode, unsigned int idx) in regcache_rbtree_get_register() argument 51 return regcache_get_val(map, rbnode->block, idx); in regcache_rbtree_get_register() 55 struct regcache_rbtree_node *rbnode, in regcache_rbtree_set_register() argument 58 set_bit(idx, rbnode->cache_present); in regcache_rbtree_set_register() 59 regcache_set_val(map, rbnode->block, idx, val); in regcache_rbtree_set_register() 67 struct regcache_rbtree_node *rbnode; in regcache_rbtree_lookup() local 70 rbnode = rbtree_ctx->cached_rbnode; in regcache_rbtree_lookup() [all …]
|
| /OK3568_Linux_fs/kernel/net/ipv4/ |
| H A D | inet_fragment.c | 48 RB_CLEAR_NODE(&skb->rbnode); in fragcb_clear() 71 rb_link_node(&skb->rbnode, &q->last_run_head->rbnode, in fragrun_create() 72 &q->last_run_head->rbnode.rb_right); in fragrun_create() 74 rb_link_node(&skb->rbnode, NULL, &q->rb_fragments.rb_node); in fragrun_create() 75 rb_insert_color(&skb->rbnode, &q->rb_fragments); in fragrun_create() 238 struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); in inet_frag_rbtree_purge() 241 rb_erase(&skb->rbnode, root); in inet_frag_rbtree_purge() 400 rb_link_node(&skb->rbnode, parent, rbn); in inet_frag_queue_insert() 401 rb_insert_color(&skb->rbnode, &q->rb_fragments); in inet_frag_queue_insert() 422 if (RB_EMPTY_NODE(&skb->rbnode)) in inet_frag_reasm_prepare() [all …]
|
| H A D | tcp_input.c | 4705 rb_erase(&skb->rbnode, &tp->out_of_order_queue); in tcp_ofo_queue() 4785 rb_link_node(&skb->rbnode, NULL, p); in tcp_data_queue_ofo() 4786 rb_insert_color(&skb->rbnode, &tp->out_of_order_queue); in tcp_data_queue_ofo() 4808 parent = &tp->ooo_last_skb->rbnode; in tcp_data_queue_ofo() 4839 rb_replace_node(&skb1->rbnode, &skb->rbnode, in tcp_data_queue_ofo() 4857 rb_link_node(&skb->rbnode, parent, p); in tcp_data_queue_ofo() 4858 rb_insert_color(&skb->rbnode, &tp->out_of_order_queue); in tcp_data_queue_ofo() 4870 rb_erase(&skb1->rbnode, &tp->out_of_order_queue); in tcp_data_queue_ofo() 5095 rb_erase(&skb->rbnode, root); in tcp_collapse_one() 5118 rb_link_node(&skb->rbnode, parent, p); in tcp_rbtree_insert() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/infiniband/hw/hfi1/ |
| H A D | mmu_rb.c | 119 struct mmu_rb_node *rbnode; in hfi1_mmu_rb_unregister() local 140 rbnode = rb_entry(node, struct mmu_rb_node, node); in hfi1_mmu_rb_unregister() 143 list_move(&rbnode->list, &del_list); in hfi1_mmu_rb_unregister() 239 struct mmu_rb_node *rbnode, *ptr; in hfi1_mmu_rb_evict() local 250 list_for_each_entry_safe_reverse(rbnode, ptr, &handler->lru_list, in hfi1_mmu_rb_evict() 252 if (handler->ops->evict(handler->ops_arg, rbnode, evict_arg, in hfi1_mmu_rb_evict() 254 __mmu_int_rb_remove(rbnode, &handler->root); in hfi1_mmu_rb_evict() 256 list_move(&rbnode->list, &del_list); in hfi1_mmu_rb_evict() 264 rbnode = list_first_entry(&del_list, struct mmu_rb_node, list); in hfi1_mmu_rb_evict() 265 list_del(&rbnode->list); in hfi1_mmu_rb_evict() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_mem.c | 135 struct rb_node *rbnode = NULL; in kbase_region_tracker_find_region_enclosing_range_free() local 143 rbnode = rbtree->rb_node; in kbase_region_tracker_find_region_enclosing_range_free() 145 while (rbnode) { in kbase_region_tracker_find_region_enclosing_range_free() 148 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_region_tracker_find_region_enclosing_range_free() 154 rbnode = rbnode->rb_left; in kbase_region_tracker_find_region_enclosing_range_free() 157 rbnode = rbnode->rb_right; in kbase_region_tracker_find_region_enclosing_range_free() 168 struct rb_node *rbnode; in kbase_region_tracker_find_region_enclosing_address() local 179 rbnode = rbtree->rb_node; in kbase_region_tracker_find_region_enclosing_address() 181 while (rbnode) { in kbase_region_tracker_find_region_enclosing_address() 184 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_region_tracker_find_region_enclosing_address() [all …]
|
| /OK3568_Linux_fs/kernel/net/netfilter/ |
| H A D | nf_conncount.c | 309 struct rb_node **rbnode, *parent; in insert_tree() local 319 rbnode = &(root->rb_node); in insert_tree() 320 while (*rbnode) { in insert_tree() 322 rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node); in insert_tree() 324 parent = *rbnode; in insert_tree() 327 rbnode = &((*rbnode)->rb_left); in insert_tree() 329 rbnode = &((*rbnode)->rb_right); in insert_tree() 377 rb_link_node_rcu(&rbconn->node, parent, rbnode); in insert_tree()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_gwt.c | 30 struct rb_node *rbnode = node; in kbase_gpu_gwt_setup_page_permission() local 32 while (rbnode) { in kbase_gpu_gwt_setup_page_permission() 36 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_gpu_gwt_setup_page_permission() 48 rbnode = rb_next(rbnode); in kbase_gpu_gwt_setup_page_permission()
|
| H A D | mali_kbase_mem.c | 188 struct rb_node *rbnode; in find_region_enclosing_range_rbtree() local 192 rbnode = rbtree->rb_node; in find_region_enclosing_range_rbtree() 194 while (rbnode) { in find_region_enclosing_range_rbtree() 197 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in find_region_enclosing_range_rbtree() 203 rbnode = rbnode->rb_left; in find_region_enclosing_range_rbtree() 206 rbnode = rbnode->rb_right; in find_region_enclosing_range_rbtree() 218 struct rb_node *rbnode; in kbase_find_region_enclosing_address() local 221 rbnode = rbtree->rb_node; in kbase_find_region_enclosing_address() 223 while (rbnode) { in kbase_find_region_enclosing_address() 226 reg = rb_entry(rbnode, struct kbase_va_region, rblink); in kbase_find_region_enclosing_address() [all …]
|
| /OK3568_Linux_fs/kernel/mm/ |
| H A D | zswap.c | 162 struct rb_node rbnode; member 272 RB_CLEAR_NODE(&entry->rbnode); in zswap_entry_cache_alloc() 290 entry = rb_entry(node, struct zswap_entry, rbnode); in zswap_rb_search() 313 myentry = rb_entry(parent, struct zswap_entry, rbnode); in zswap_rb_insert() 323 rb_link_node(&entry->rbnode, parent, link); in zswap_rb_insert() 324 rb_insert_color(&entry->rbnode, root); in zswap_rb_insert() 330 if (!RB_EMPTY_NODE(&entry->rbnode)) { in zswap_rb_erase() 331 rb_erase(&entry->rbnode, root); in zswap_rb_erase() 332 RB_CLEAR_NODE(&entry->rbnode); in zswap_rb_erase() 1229 rbtree_postorder_for_each_entry_safe(entry, n, &tree->rbroot, rbnode) in zswap_frontswap_invalidate_area()
|
| /OK3568_Linux_fs/kernel/net/sched/ |
| H A D | sch_etf.c | 188 rb_link_node(&nskb->rbnode, parent, p); in etf_enqueue_timesortedlist() 189 rb_insert_color_cached(&nskb->rbnode, &q->head, leftmost); in etf_enqueue_timesortedlist() 211 rb_erase_cached(&skb->rbnode, &q->head); in timesortedlist_drop() 235 rb_erase_cached(&skb->rbnode, &q->head); in timesortedlist_remove() 430 rb_erase_cached(&skb->rbnode, &q->head); in timesortedlist_clear()
|
| H A D | sch_fq.c | 386 rb_erase(&skb->rbnode, &flow->t_root); in fq_erase_head() 432 rb_link_node(&skb->rbnode, parent, p); in flow_queue_add() 433 rb_insert_color(&skb->rbnode, &flow->t_root); in flow_queue_add() 654 rb_erase(&skb->rbnode, &flow->t_root); in fq_flow_purge()
|
| H A D | sch_netem.c | 369 rb_erase(&skb->rbnode, &q->t_root); in tfifo_reset() 402 rb_link_node(&nskb->rbnode, parent, p); in tfifo_enqueue() 403 rb_insert_color(&nskb->rbnode, &q->t_root); in tfifo_enqueue() 672 rb_erase(&skb->rbnode, &q->t_root); in netem_erase_head()
|
| /OK3568_Linux_fs/kernel/fs/btrfs/ |
| H A D | backref.c | 242 ref = rb_entry(parent, struct prelim_ref, rbnode); in prelim_ref_insert() 278 rb_link_node(&newref->rbnode, parent, p); in prelim_ref_insert() 279 rb_insert_color_cached(&newref->rbnode, root, leftmost); in prelim_ref_insert() 291 &preftree->root.rb_root, rbnode) { in prelim_release() 405 ref = rb_entry(parent, struct prelim_ref, rbnode); in is_shared_data_backref() 706 ref = rb_entry(rnode, struct prelim_ref, rbnode); in resolve_indirect_refs() 713 rb_erase_cached(&ref->rbnode, &preftrees->indirect.root); in resolve_indirect_refs() 798 ref = rb_entry(node, struct prelim_ref, rbnode); in add_missing_keys() 1341 ref = rb_entry(node, struct prelim_ref, rbnode); 1342 node = rb_next(&ref->rbnode);
|
| H A D | backref.h | 71 struct rb_node rbnode; member
|
| /OK3568_Linux_fs/kernel/net/mptcp/ |
| H A D | protocol.c | 179 rb_link_node(&skb->rbnode, NULL, p); in mptcp_data_queue_ofo() 180 rb_insert_color(&skb->rbnode, &msk->out_of_order_queue); in mptcp_data_queue_ofo() 197 parent = &msk->ooo_last_skb->rbnode; in mptcp_data_queue_ofo() 228 rb_replace_node(&skb1->rbnode, &skb->rbnode, in mptcp_data_queue_ofo() 243 rb_link_node(&skb->rbnode, parent, p); in mptcp_data_queue_ofo() 244 rb_insert_color(&skb->rbnode, &msk->out_of_order_queue); in mptcp_data_queue_ofo() 251 rb_erase(&skb1->rbnode, &msk->out_of_order_queue); in mptcp_data_queue_ofo() 562 rb_erase(&skb->rbnode, &msk->out_of_order_queue); in mptcp_ofo_queue()
|
| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | skbuff.h | 734 struct rb_node rbnode; /* used in netem, ip4 defrag, and tcp stack */ member 3498 #define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode) 3501 #define skb_rb_next(skb) rb_to_skb(rb_next(&(skb)->rbnode)) 3502 #define skb_rb_prev(skb) rb_to_skb(rb_prev(&(skb)->rbnode))
|
| /OK3568_Linux_fs/kernel/Documentation/networking/ |
| H A D | rds.rst | 318 wraps the raw congestion bitmap, contains rbnode, waitq, etc.
|
| /OK3568_Linux_fs/kernel/include/net/ |
| H A D | tcp.h | 1864 rb_erase(&skb->rbnode, &sk->tcp_rtx_queue); in tcp_rtx_queue_unlink()
|
| /OK3568_Linux_fs/kernel/net/core/ |
| H A D | skbuff.c | 3150 struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); in skb_rbtree_purge() 3153 rb_erase(&skb->rbnode, root); in skb_rbtree_purge()
|