| /OK3568_Linux_fs/kernel/drivers/hid/intel-ish-hid/ishtp/ |
| H A D | client.c | 16 int ishtp_cl_get_tx_free_buffer_size(struct ishtp_cl *cl) in ishtp_cl_get_tx_free_buffer_size() argument 21 spin_lock_irqsave(&cl->tx_free_list_spinlock, tx_free_flags); in ishtp_cl_get_tx_free_buffer_size() 22 size = cl->tx_ring_free_size * cl->device->fw_client->props.max_msg_length; in ishtp_cl_get_tx_free_buffer_size() 23 spin_unlock_irqrestore(&cl->tx_free_list_spinlock, tx_free_flags); in ishtp_cl_get_tx_free_buffer_size() 29 int ishtp_cl_get_tx_free_rings(struct ishtp_cl *cl) in ishtp_cl_get_tx_free_rings() argument 31 return cl->tx_ring_free_size; in ishtp_cl_get_tx_free_rings() 41 static void ishtp_read_list_flush(struct ishtp_cl *cl) in ishtp_read_list_flush() argument 47 spin_lock_irqsave(&cl->dev->read_list_spinlock, flags); in ishtp_read_list_flush() 48 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush() 49 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush() [all …]
|
| H A D | client-buffers.c | 19 int ishtp_cl_alloc_rx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_rx_ring() argument 21 size_t len = cl->device->fw_client->props.max_msg_length; in ishtp_cl_alloc_rx_ring() 27 for (j = 0; j < cl->rx_ring_size; ++j) { in ishtp_cl_alloc_rx_ring() 28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring() 36 spin_lock_irqsave(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring() 37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring() 38 spin_unlock_irqrestore(&cl->free_list_spinlock, flags); in ishtp_cl_alloc_rx_ring() 44 dev_err(&cl->device->dev, "error in allocating Rx buffers\n"); in ishtp_cl_alloc_rx_ring() 45 ishtp_cl_free_rx_ring(cl); in ishtp_cl_alloc_rx_ring() 57 int ishtp_cl_alloc_tx_ring(struct ishtp_cl *cl) in ishtp_cl_alloc_tx_ring() argument [all …]
|
| /OK3568_Linux_fs/kernel/net/sched/ |
| H A D | sch_hfsc.c | 186 eltree_insert(struct hfsc_class *cl) in eltree_insert() argument 188 struct rb_node **p = &cl->sched->eligible.rb_node; in eltree_insert() 195 if (cl->cl_e >= cl1->cl_e) in eltree_insert() 200 rb_link_node(&cl->el_node, parent, p); in eltree_insert() 201 rb_insert_color(&cl->el_node, &cl->sched->eligible); in eltree_insert() 205 eltree_remove(struct hfsc_class *cl) in eltree_remove() argument 207 rb_erase(&cl->el_node, &cl->sched->eligible); in eltree_remove() 211 eltree_update(struct hfsc_class *cl) in eltree_update() argument 213 eltree_remove(cl); in eltree_update() 214 eltree_insert(cl); in eltree_update() [all …]
|
| H A D | sch_cbq.c | 163 #define L2T(cl, len) qdisc_l2t((cl)->R_tab, len) argument 181 struct cbq_class *cl; in cbq_reclassify() local 183 for (cl = this->tparent; cl; cl = cl->tparent) { in cbq_reclassify() 184 struct cbq_class *new = cl->defaults[TC_PRIO_BESTEFFORT]; in cbq_reclassify() 210 struct cbq_class *cl = NULL; in cbq_classify() local 219 (cl = cbq_class_lookup(q, prio)) != NULL) in cbq_classify() 220 return cl; in cbq_classify() 235 cl = (void *)res.class; in cbq_classify() 236 if (!cl) { in cbq_classify() 238 cl = cbq_class_lookup(q, res.classid); in cbq_classify() [all …]
|
| H A D | sch_htb.c | 213 struct htb_class *cl; in htb_classify() local 224 cl = htb_find(skb->priority, sch); in htb_classify() 225 if (cl) { in htb_classify() 226 if (cl->level == 0) in htb_classify() 227 return cl; in htb_classify() 229 tcf = rcu_dereference_bh(cl->filter_list); in htb_classify() 247 cl = (void *)res.class; in htb_classify() 248 if (!cl) { in htb_classify() 251 cl = htb_find(res.classid, sch); in htb_classify() 252 if (!cl) in htb_classify() [all …]
|
| H A D | sch_drr.c | 59 struct drr_class *cl = (struct drr_class *)*arg; in drr_change_class() local 84 if (cl != NULL) { in drr_change_class() 86 err = gen_replace_estimator(&cl->bstats, NULL, in drr_change_class() 87 &cl->rate_est, in drr_change_class() 99 cl->quantum = quantum; in drr_change_class() 105 cl = kzalloc(sizeof(struct drr_class), GFP_KERNEL); in drr_change_class() 106 if (cl == NULL) in drr_change_class() 109 cl->common.classid = classid; in drr_change_class() 110 cl->quantum = quantum; in drr_change_class() 111 cl->qdisc = qdisc_create_dflt(sch->dev_queue, in drr_change_class() [all …]
|
| H A D | sch_qfq.c | 313 struct qfq_class *cl) in qfq_add_to_agg() argument 315 cl->agg = agg; in qfq_add_to_agg() 318 if (cl->qdisc->q.qlen > 0) { /* adding an active class */ in qfq_add_to_agg() 319 list_add_tail(&cl->alist, &agg->active); in qfq_add_to_agg() 321 cl && q->in_serv_agg != agg) /* agg was inactive */ in qfq_add_to_agg() 341 static void qfq_deactivate_class(struct qfq_sched *q, struct qfq_class *cl) in qfq_deactivate_class() argument 343 struct qfq_aggregate *agg = cl->agg; in qfq_deactivate_class() 346 list_del(&cl->alist); /* remove from RR queue of the aggregate */ in qfq_deactivate_class() 352 static void qfq_rm_from_agg(struct qfq_sched *q, struct qfq_class *cl) in qfq_rm_from_agg() argument 354 struct qfq_aggregate *agg = cl->agg; in qfq_rm_from_agg() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/misc/mei/ |
| H A D | client.c | 339 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue() 351 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue() 352 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue() 365 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument 370 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp() 387 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument 399 cb->cl = cl; in mei_io_cb_init() 414 const struct mei_cl *cl) in mei_io_list_flush_cl() argument 419 if (cl == cb->cl) { in mei_io_list_flush_cl() 435 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument [all …]
|
| H A D | client.h | 131 int mei_cl_link(struct mei_cl *cl); 132 int mei_cl_unlink(struct mei_cl *cl); 136 struct mei_cl_cb *mei_cl_read_cb(struct mei_cl *cl, const struct file *fp); 138 void mei_cl_add_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb); 139 void mei_cl_del_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb); 141 struct mei_cl_cb *mei_cl_alloc_cb(struct mei_cl *cl, size_t length, 144 struct mei_cl_cb *mei_cl_enqueue_ctrl_wr_cb(struct mei_cl *cl, size_t length, 147 int mei_cl_flush_queues(struct mei_cl *cl, const struct file *fp); 150 const struct file *mei_cl_fp_by_vtag(const struct mei_cl *cl, u8 vtag); 151 int mei_cl_vt_support_check(const struct mei_cl *cl); [all …]
|
| H A D | main.c | 48 struct mei_cl *cl; in mei_open() local 65 cl = mei_cl_alloc_linked(dev); in mei_open() 66 if (IS_ERR(cl)) { in mei_open() 67 err = PTR_ERR(cl); in mei_open() 71 cl->fp = file; in mei_open() 72 file->private_data = cl; in mei_open() 90 static void mei_cl_vtag_remove_by_fp(const struct mei_cl *cl, in mei_cl_vtag_remove_by_fp() argument 95 list_for_each_entry_safe(vtag_l, next, &cl->vtag_map, list) { in mei_cl_vtag_remove_by_fp() 114 struct mei_cl *cl = file->private_data; in mei_release() local 118 if (WARN_ON(!cl || !cl->dev)) in mei_release() [all …]
|
| H A D | interrupt.c | 32 struct mei_cl *cl; in mei_irq_compl_handler() local 35 cl = cb->cl; in mei_irq_compl_handler() 39 mei_cl_complete(cl, cb); in mei_irq_compl_handler() 52 static inline int mei_cl_hbm_equal(struct mei_cl *cl, in mei_cl_hbm_equal() argument 55 return mei_cl_host_addr(cl) == mei_hdr->host_addr && in mei_cl_hbm_equal() 56 mei_cl_me_id(cl) == mei_hdr->me_addr; in mei_cl_hbm_equal() 93 static int mei_cl_irq_read_msg(struct mei_cl *cl, in mei_cl_irq_read_msg() argument 98 struct mei_device *dev = cl->dev; in mei_cl_irq_read_msg() 112 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_irq_read_msg() 114 if (!mei_cl_is_fixed_address(cl)) { in mei_cl_irq_read_msg() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/md/bcache/ |
| H A D | closure.c | 16 static inline void closure_put_after_sub(struct closure *cl, int flags) in closure_put_after_sub() argument 24 if (cl->fn && !(flags & CLOSURE_DESTRUCTOR)) { in closure_put_after_sub() 25 atomic_set(&cl->remaining, in closure_put_after_sub() 27 closure_queue(cl); in closure_put_after_sub() 29 struct closure *parent = cl->parent; in closure_put_after_sub() 30 closure_fn *destructor = cl->fn; in closure_put_after_sub() 32 closure_debug_destroy(cl); in closure_put_after_sub() 35 destructor(cl); in closure_put_after_sub() 44 void closure_sub(struct closure *cl, int v) in closure_sub() argument 46 closure_put_after_sub(cl, atomic_sub_return(v, &cl->remaining)); in closure_sub() [all …]
|
| H A D | closure.h | 169 void closure_sub(struct closure *cl, int v); 170 void closure_put(struct closure *cl); 172 bool closure_wait(struct closure_waitlist *list, struct closure *cl); 173 void __closure_sync(struct closure *cl); 181 static inline void closure_sync(struct closure *cl) in closure_sync() argument 183 if ((atomic_read(&cl->remaining) & CLOSURE_REMAINING_MASK) != 1) in closure_sync() 184 __closure_sync(cl); in closure_sync() 190 void closure_debug_create(struct closure *cl); 191 void closure_debug_destroy(struct closure *cl); 196 static inline void closure_debug_create(struct closure *cl) {} in closure_debug_create() argument [all …]
|
| H A D | request.c | 28 static void bch_data_insert_start(struct closure *cl); 58 static void bch_data_insert_keys(struct closure *cl) in bch_data_insert_keys() argument 60 struct data_insert_op *op = container_of(cl, struct data_insert_op, cl); in bch_data_insert_keys() 67 op->flush_journal ? cl : NULL); in bch_data_insert_keys() 82 continue_at(cl, bch_data_insert_start, op->wq); in bch_data_insert_keys() 87 closure_return(cl); in bch_data_insert_keys() 108 static void bch_data_invalidate(struct closure *cl) in bch_data_invalidate() argument 110 struct data_insert_op *op = container_of(cl, struct data_insert_op, cl); in bch_data_invalidate() 136 continue_at(cl, bch_data_insert_keys, op->wq); in bch_data_invalidate() 139 static void bch_data_insert_error(struct closure *cl) in bch_data_insert_error() argument [all …]
|
| /OK3568_Linux_fs/kernel/drivers/clk/ |
| H A D | clkdev.c | 38 struct clk_lookup *p, *cl = NULL; in clk_find() local 62 cl = p; in clk_find() 69 return cl; in clk_find() 74 struct clk_lookup *cl; in clk_find_hw() local 78 cl = clk_find(dev_id, con_id); in clk_find_hw() 79 if (cl) in clk_find_hw() 80 hw = cl->clk_hw; in clk_find_hw() 121 static void __clkdev_add(struct clk_lookup *cl) in __clkdev_add() argument 124 list_add_tail(&cl->node, &clocks); in __clkdev_add() 128 void clkdev_add(struct clk_lookup *cl) in clkdev_add() argument [all …]
|
| /OK3568_Linux_fs/external/xserver/glx/ |
| H A D | glxcmdsswap.c | 57 __glXDispSwap_CreateContext(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_CreateContext() argument 69 return __glXDisp_CreateContext(cl, pc); in __glXDispSwap_CreateContext() 73 __glXDispSwap_CreateNewContext(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_CreateNewContext() argument 86 return __glXDisp_CreateNewContext(cl, pc); in __glXDispSwap_CreateNewContext() 90 __glXDispSwap_CreateContextWithConfigSGIX(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_CreateContextWithConfigSGIX() argument 92 ClientPtr client = cl->client; in __glXDispSwap_CreateContextWithConfigSGIX() 106 return __glXDisp_CreateContextWithConfigSGIX(cl, pc); in __glXDispSwap_CreateContextWithConfigSGIX() 110 __glXDispSwap_DestroyContext(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_DestroyContext() argument 119 return __glXDisp_DestroyContext(cl, pc); in __glXDispSwap_DestroyContext() 123 __glXDispSwap_MakeCurrent(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_MakeCurrent() argument [all …]
|
| H A D | vndservermapping.c | 93 GlxClientPriv *cl; in GlxAllocContextTag() local 100 cl = GlxGetClientData(client); in GlxAllocContextTag() 101 if (cl == NULL) { in GlxAllocContextTag() 106 for (index=0; index<cl->contextTagCount; index++) { in GlxAllocContextTag() 107 if (cl->contextTags[index].vendor == NULL) { in GlxAllocContextTag() 111 if (index >= cl->contextTagCount) { in GlxAllocContextTag() 114 unsigned int newSize = cl->contextTagCount * 2; in GlxAllocContextTag() 121 realloc(cl->contextTags, newSize * sizeof(GlxContextTagInfo)); in GlxAllocContextTag() 126 memset(&newTags[cl->contextTagCount], 0, in GlxAllocContextTag() 127 (newSize - cl->contextTagCount) * sizeof(GlxContextTagInfo)); in GlxAllocContextTag() [all …]
|
| H A D | singlepix.c | 43 __glXDisp_ReadPixels(__GLXclientState * cl, GLbyte * pc) in __glXDisp_ReadPixels() argument 50 ClientPtr client = cl->client; in __glXDisp_ReadPixels() 57 cx = __glXForceCurrent(cl, __GLX_GET_SINGLE_CONTEXT_TAG(pc), &error); in __glXDisp_ReadPixels() 75 __GLX_GET_ANSWER_BUFFER(answer, cl, compsize, 1); in __glXDisp_ReadPixels() 94 __glXDisp_GetTexImage(__GLXclientState * cl, GLbyte * pc) in __glXDisp_GetTexImage() argument 100 ClientPtr client = cl->client; in __glXDisp_GetTexImage() 108 cx = __glXForceCurrent(cl, __GLX_GET_SINGLE_CONTEXT_TAG(pc), &error); in __glXDisp_GetTexImage() 135 __GLX_GET_ANSWER_BUFFER(answer, cl, compsize, 1); in __glXDisp_GetTexImage() 156 __glXDisp_GetPolygonStipple(__GLXclientState * cl, GLbyte * pc) in __glXDisp_GetPolygonStipple() argument 160 ClientPtr client = cl->client; in __glXDisp_GetPolygonStipple() [all …]
|
| H A D | singlepixswap.c | 43 __glXDispSwap_ReadPixels(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_ReadPixels() argument 52 ClientPtr client = cl->client; in __glXDispSwap_ReadPixels() 60 cx = __glXForceCurrent(cl, __GLX_GET_SINGLE_CONTEXT_TAG(pc), &error); in __glXDispSwap_ReadPixels() 85 __GLX_GET_ANSWER_BUFFER(answer, cl, compsize, 1); in __glXDispSwap_ReadPixels() 106 __glXDispSwap_GetTexImage(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_GetTexImage() argument 114 ClientPtr client = cl->client; in __glXDispSwap_GetTexImage() 123 cx = __glXForceCurrent(cl, __GLX_GET_SINGLE_CONTEXT_TAG(pc), &error); in __glXDispSwap_GetTexImage() 155 __GLX_GET_ANSWER_BUFFER(answer, cl, compsize, 1); in __glXDispSwap_GetTexImage() 181 __glXDispSwap_GetPolygonStipple(__GLXclientState * cl, GLbyte * pc) in __glXDispSwap_GetPolygonStipple() argument 185 ClientPtr client = cl->client; in __glXDispSwap_GetPolygonStipple() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/hsi/ |
| H A D | hsi_core.c | 60 struct hsi_client *cl = to_hsi_client(dev); in hsi_client_release() local 62 kfree(cl->tx_cfg.channels); in hsi_client_release() 63 kfree(cl->rx_cfg.channels); in hsi_client_release() 64 kfree(cl); in hsi_client_release() 70 struct hsi_client *cl; in hsi_new_client() local 73 cl = kzalloc(sizeof(*cl), GFP_KERNEL); in hsi_new_client() 74 if (!cl) in hsi_new_client() 77 cl->tx_cfg = info->tx_cfg; in hsi_new_client() 78 if (cl->tx_cfg.channels) { in hsi_new_client() 79 size = cl->tx_cfg.num_channels * sizeof(*cl->tx_cfg.channels); in hsi_new_client() [all …]
|
| /OK3568_Linux_fs/external/xserver/hw/dmx/glxProxy/ |
| H A D | glxcmdsswap.c | 41 int __glXSwapGetFBConfigsSGIX(__GLXclientState * cl, GLbyte * pc); 53 __glXSwapCreateContext(__GLXclientState * cl, GLbyte * pc) in __glXSwapCreateContext() argument 65 return __glXCreateContext(cl, pc); in __glXSwapCreateContext() 69 __glXSwapCreateNewContext(__GLXclientState * cl, GLbyte * pc) in __glXSwapCreateNewContext() argument 81 return __glXCreateNewContext(cl, pc); in __glXSwapCreateNewContext() 85 __glXSwapCreateContextWithConfigSGIX(__GLXclientState * cl, GLbyte * pc) in __glXSwapCreateContextWithConfigSGIX() argument 97 return __glXCreateContextWithConfigSGIX(cl, pc); in __glXSwapCreateContextWithConfigSGIX() 101 __glXSwapQueryMaxSwapBarriersSGIX(__GLXclientState * cl, GLbyte * pc) in __glXSwapQueryMaxSwapBarriersSGIX() argument 110 return __glXQueryMaxSwapBarriersSGIX(cl, pc); in __glXSwapQueryMaxSwapBarriersSGIX() 114 __glXSwapBindSwapBarrierSGIX(__GLXclientState * cl, GLbyte * pc) in __glXSwapBindSwapBarrierSGIX() argument [all …]
|
| H A D | glxext.c | 73 __GLXclientState *cl = __glXClients[clientIndex]; in ResetClientState() local 77 free(cl->returnBuf); in ResetClientState() 78 free(cl->currentContexts); in ResetClientState() 79 free(cl->currentDrawables); in ResetClientState() 80 free(cl->largeCmdBuf); in ResetClientState() 83 if (cl->be_displays[i]) in ResetClientState() 84 XCloseDisplay(cl->be_displays[i]); in ResetClientState() 87 keep_be_displays = cl->be_displays; in ResetClientState() 88 memset(cl, 0, sizeof(__GLXclientState)); in ResetClientState() 89 cl->be_displays = keep_be_displays; in ResetClientState() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/hsi/clients/ |
| H A D | ssi_protocol.c | 35 void ssi_waketest(struct hsi_client *cl, unsigned int enable); 141 struct hsi_client *cl; member 219 struct ssi_protocol *ssi = hsi_client_drvdata(msg->cl); in ssip_release_cmd() 221 dev_dbg(&msg->cl->device, "Release cmd 0x%08x\n", ssip_get_cmd(msg)); in ssip_release_cmd() 333 if (slave->device.parent == ssi->cl->device.parent) { in ssip_slave_get_master() 334 master = ssi->cl; in ssip_slave_get_master() 387 static void ssip_reset(struct hsi_client *cl) in ssip_reset() argument 389 struct ssi_protocol *ssi = hsi_client_drvdata(cl); in ssip_reset() 395 hsi_flush(cl); in ssip_reset() 398 hsi_stop_tx(cl); in ssip_reset() [all …]
|
| /OK3568_Linux_fs/kernel/include/linux/hsi/ |
| H A D | hsi.h | 138 static inline void hsi_client_set_drvdata(struct hsi_client *cl, void *data) in hsi_client_set_drvdata() argument 140 dev_set_drvdata(&cl->device, data); in hsi_client_set_drvdata() 143 static inline void *hsi_client_drvdata(struct hsi_client *cl) in hsi_client_drvdata() argument 145 return dev_get_drvdata(&cl->device); in hsi_client_drvdata() 148 int hsi_register_port_event(struct hsi_client *cl, 150 int hsi_unregister_port_event(struct hsi_client *cl); 187 struct hsi_client *cl; member 230 int (*setup)(struct hsi_client *cl); 231 int (*flush)(struct hsi_client *cl); 232 int (*start_tx)(struct hsi_client *cl); [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/gt/ |
| H A D | intel_timeline.c | 126 struct intel_timeline_cacheline *cl = in __rcu_cacheline_free() local 127 container_of(rcu, typeof(*cl), rcu); in __rcu_cacheline_free() 130 i915_gem_object_unpin_map(cl->hwsp->vma->obj); in __rcu_cacheline_free() 131 __idle_hwsp_free(cl->hwsp, ptr_unmask_bits(cl->vaddr, CACHELINE_BITS)); in __rcu_cacheline_free() 133 i915_active_fini(&cl->active); in __rcu_cacheline_free() 134 kfree(cl); in __rcu_cacheline_free() 137 static void __idle_cacheline_free(struct intel_timeline_cacheline *cl) in __idle_cacheline_free() argument 139 GEM_BUG_ON(!i915_active_is_idle(&cl->active)); in __idle_cacheline_free() 140 call_rcu(&cl->rcu, __rcu_cacheline_free); in __idle_cacheline_free() 146 struct intel_timeline_cacheline *cl = in __cacheline_retire() local [all …]
|