xref: /OK3568_Linux_fs/kernel/net/sched/sch_cbs.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * net/sched/sch_cbs.c	Credit Based Shaper
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * Authors:	Vinicius Costa Gomes <vinicius.gomes@intel.com>
6*4882a593Smuzhiyun  */
7*4882a593Smuzhiyun 
8*4882a593Smuzhiyun /* Credit Based Shaper (CBS)
9*4882a593Smuzhiyun  * =========================
10*4882a593Smuzhiyun  *
11*4882a593Smuzhiyun  * This is a simple rate-limiting shaper aimed at TSN applications on
12*4882a593Smuzhiyun  * systems with known traffic workloads.
13*4882a593Smuzhiyun  *
14*4882a593Smuzhiyun  * Its algorithm is defined by the IEEE 802.1Q-2014 Specification,
15*4882a593Smuzhiyun  * Section 8.6.8.2, and explained in more detail in the Annex L of the
16*4882a593Smuzhiyun  * same specification.
17*4882a593Smuzhiyun  *
18*4882a593Smuzhiyun  * There are four tunables to be considered:
19*4882a593Smuzhiyun  *
20*4882a593Smuzhiyun  *	'idleslope': Idleslope is the rate of credits that is
21*4882a593Smuzhiyun  *	accumulated (in kilobits per second) when there is at least
22*4882a593Smuzhiyun  *	one packet waiting for transmission. Packets are transmitted
23*4882a593Smuzhiyun  *	when the current value of credits is equal or greater than
24*4882a593Smuzhiyun  *	zero. When there is no packet to be transmitted the amount of
25*4882a593Smuzhiyun  *	credits is set to zero. This is the main tunable of the CBS
26*4882a593Smuzhiyun  *	algorithm.
27*4882a593Smuzhiyun  *
28*4882a593Smuzhiyun  *	'sendslope':
29*4882a593Smuzhiyun  *	Sendslope is the rate of credits that is depleted (it should be a
30*4882a593Smuzhiyun  *	negative number of kilobits per second) when a transmission is
31*4882a593Smuzhiyun  *	ocurring. It can be calculated as follows, (IEEE 802.1Q-2014 Section
32*4882a593Smuzhiyun  *	8.6.8.2 item g):
33*4882a593Smuzhiyun  *
34*4882a593Smuzhiyun  *	sendslope = idleslope - port_transmit_rate
35*4882a593Smuzhiyun  *
36*4882a593Smuzhiyun  *	'hicredit': Hicredit defines the maximum amount of credits (in
37*4882a593Smuzhiyun  *	bytes) that can be accumulated. Hicredit depends on the
38*4882a593Smuzhiyun  *	characteristics of interfering traffic,
39*4882a593Smuzhiyun  *	'max_interference_size' is the maximum size of any burst of
40*4882a593Smuzhiyun  *	traffic that can delay the transmission of a frame that is
41*4882a593Smuzhiyun  *	available for transmission for this traffic class, (IEEE
42*4882a593Smuzhiyun  *	802.1Q-2014 Annex L, Equation L-3):
43*4882a593Smuzhiyun  *
44*4882a593Smuzhiyun  *	hicredit = max_interference_size * (idleslope / port_transmit_rate)
45*4882a593Smuzhiyun  *
46*4882a593Smuzhiyun  *	'locredit': Locredit is the minimum amount of credits that can
47*4882a593Smuzhiyun  *	be reached. It is a function of the traffic flowing through
48*4882a593Smuzhiyun  *	this qdisc (IEEE 802.1Q-2014 Annex L, Equation L-2):
49*4882a593Smuzhiyun  *
50*4882a593Smuzhiyun  *	locredit = max_frame_size * (sendslope / port_transmit_rate)
51*4882a593Smuzhiyun  */
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun #include <linux/module.h>
54*4882a593Smuzhiyun #include <linux/types.h>
55*4882a593Smuzhiyun #include <linux/kernel.h>
56*4882a593Smuzhiyun #include <linux/string.h>
57*4882a593Smuzhiyun #include <linux/errno.h>
58*4882a593Smuzhiyun #include <linux/skbuff.h>
59*4882a593Smuzhiyun #include <net/netevent.h>
60*4882a593Smuzhiyun #include <net/netlink.h>
61*4882a593Smuzhiyun #include <net/sch_generic.h>
62*4882a593Smuzhiyun #include <net/pkt_sched.h>
63*4882a593Smuzhiyun 
64*4882a593Smuzhiyun static LIST_HEAD(cbs_list);
65*4882a593Smuzhiyun static DEFINE_SPINLOCK(cbs_list_lock);
66*4882a593Smuzhiyun 
67*4882a593Smuzhiyun #define BYTES_PER_KBIT (1000LL / 8)
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun struct cbs_sched_data {
70*4882a593Smuzhiyun 	bool offload;
71*4882a593Smuzhiyun 	int queue;
72*4882a593Smuzhiyun 	atomic64_t port_rate; /* in bytes/s */
73*4882a593Smuzhiyun 	s64 last; /* timestamp in ns */
74*4882a593Smuzhiyun 	s64 credits; /* in bytes */
75*4882a593Smuzhiyun 	s32 locredit; /* in bytes */
76*4882a593Smuzhiyun 	s32 hicredit; /* in bytes */
77*4882a593Smuzhiyun 	s64 sendslope; /* in bytes/s */
78*4882a593Smuzhiyun 	s64 idleslope; /* in bytes/s */
79*4882a593Smuzhiyun 	struct qdisc_watchdog watchdog;
80*4882a593Smuzhiyun 	int (*enqueue)(struct sk_buff *skb, struct Qdisc *sch,
81*4882a593Smuzhiyun 		       struct sk_buff **to_free);
82*4882a593Smuzhiyun 	struct sk_buff *(*dequeue)(struct Qdisc *sch);
83*4882a593Smuzhiyun 	struct Qdisc *qdisc;
84*4882a593Smuzhiyun 	struct list_head cbs_list;
85*4882a593Smuzhiyun };
86*4882a593Smuzhiyun 
cbs_child_enqueue(struct sk_buff * skb,struct Qdisc * sch,struct Qdisc * child,struct sk_buff ** to_free)87*4882a593Smuzhiyun static int cbs_child_enqueue(struct sk_buff *skb, struct Qdisc *sch,
88*4882a593Smuzhiyun 			     struct Qdisc *child,
89*4882a593Smuzhiyun 			     struct sk_buff **to_free)
90*4882a593Smuzhiyun {
91*4882a593Smuzhiyun 	unsigned int len = qdisc_pkt_len(skb);
92*4882a593Smuzhiyun 	int err;
93*4882a593Smuzhiyun 
94*4882a593Smuzhiyun 	err = child->ops->enqueue(skb, child, to_free);
95*4882a593Smuzhiyun 	if (err != NET_XMIT_SUCCESS)
96*4882a593Smuzhiyun 		return err;
97*4882a593Smuzhiyun 
98*4882a593Smuzhiyun 	sch->qstats.backlog += len;
99*4882a593Smuzhiyun 	sch->q.qlen++;
100*4882a593Smuzhiyun 
101*4882a593Smuzhiyun 	return NET_XMIT_SUCCESS;
102*4882a593Smuzhiyun }
103*4882a593Smuzhiyun 
cbs_enqueue_offload(struct sk_buff * skb,struct Qdisc * sch,struct sk_buff ** to_free)104*4882a593Smuzhiyun static int cbs_enqueue_offload(struct sk_buff *skb, struct Qdisc *sch,
105*4882a593Smuzhiyun 			       struct sk_buff **to_free)
106*4882a593Smuzhiyun {
107*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
108*4882a593Smuzhiyun 	struct Qdisc *qdisc = q->qdisc;
109*4882a593Smuzhiyun 
110*4882a593Smuzhiyun 	return cbs_child_enqueue(skb, sch, qdisc, to_free);
111*4882a593Smuzhiyun }
112*4882a593Smuzhiyun 
cbs_enqueue_soft(struct sk_buff * skb,struct Qdisc * sch,struct sk_buff ** to_free)113*4882a593Smuzhiyun static int cbs_enqueue_soft(struct sk_buff *skb, struct Qdisc *sch,
114*4882a593Smuzhiyun 			    struct sk_buff **to_free)
115*4882a593Smuzhiyun {
116*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
117*4882a593Smuzhiyun 	struct Qdisc *qdisc = q->qdisc;
118*4882a593Smuzhiyun 
119*4882a593Smuzhiyun 	if (sch->q.qlen == 0 && q->credits > 0) {
120*4882a593Smuzhiyun 		/* We need to stop accumulating credits when there's
121*4882a593Smuzhiyun 		 * no enqueued packets and q->credits is positive.
122*4882a593Smuzhiyun 		 */
123*4882a593Smuzhiyun 		q->credits = 0;
124*4882a593Smuzhiyun 		q->last = ktime_get_ns();
125*4882a593Smuzhiyun 	}
126*4882a593Smuzhiyun 
127*4882a593Smuzhiyun 	return cbs_child_enqueue(skb, sch, qdisc, to_free);
128*4882a593Smuzhiyun }
129*4882a593Smuzhiyun 
cbs_enqueue(struct sk_buff * skb,struct Qdisc * sch,struct sk_buff ** to_free)130*4882a593Smuzhiyun static int cbs_enqueue(struct sk_buff *skb, struct Qdisc *sch,
131*4882a593Smuzhiyun 		       struct sk_buff **to_free)
132*4882a593Smuzhiyun {
133*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
134*4882a593Smuzhiyun 
135*4882a593Smuzhiyun 	return q->enqueue(skb, sch, to_free);
136*4882a593Smuzhiyun }
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun /* timediff is in ns, slope is in bytes/s */
timediff_to_credits(s64 timediff,s64 slope)139*4882a593Smuzhiyun static s64 timediff_to_credits(s64 timediff, s64 slope)
140*4882a593Smuzhiyun {
141*4882a593Smuzhiyun 	return div64_s64(timediff * slope, NSEC_PER_SEC);
142*4882a593Smuzhiyun }
143*4882a593Smuzhiyun 
delay_from_credits(s64 credits,s64 slope)144*4882a593Smuzhiyun static s64 delay_from_credits(s64 credits, s64 slope)
145*4882a593Smuzhiyun {
146*4882a593Smuzhiyun 	if (unlikely(slope == 0))
147*4882a593Smuzhiyun 		return S64_MAX;
148*4882a593Smuzhiyun 
149*4882a593Smuzhiyun 	return div64_s64(-credits * NSEC_PER_SEC, slope);
150*4882a593Smuzhiyun }
151*4882a593Smuzhiyun 
credits_from_len(unsigned int len,s64 slope,s64 port_rate)152*4882a593Smuzhiyun static s64 credits_from_len(unsigned int len, s64 slope, s64 port_rate)
153*4882a593Smuzhiyun {
154*4882a593Smuzhiyun 	if (unlikely(port_rate == 0))
155*4882a593Smuzhiyun 		return S64_MAX;
156*4882a593Smuzhiyun 
157*4882a593Smuzhiyun 	return div64_s64(len * slope, port_rate);
158*4882a593Smuzhiyun }
159*4882a593Smuzhiyun 
cbs_child_dequeue(struct Qdisc * sch,struct Qdisc * child)160*4882a593Smuzhiyun static struct sk_buff *cbs_child_dequeue(struct Qdisc *sch, struct Qdisc *child)
161*4882a593Smuzhiyun {
162*4882a593Smuzhiyun 	struct sk_buff *skb;
163*4882a593Smuzhiyun 
164*4882a593Smuzhiyun 	skb = child->ops->dequeue(child);
165*4882a593Smuzhiyun 	if (!skb)
166*4882a593Smuzhiyun 		return NULL;
167*4882a593Smuzhiyun 
168*4882a593Smuzhiyun 	qdisc_qstats_backlog_dec(sch, skb);
169*4882a593Smuzhiyun 	qdisc_bstats_update(sch, skb);
170*4882a593Smuzhiyun 	sch->q.qlen--;
171*4882a593Smuzhiyun 
172*4882a593Smuzhiyun 	return skb;
173*4882a593Smuzhiyun }
174*4882a593Smuzhiyun 
cbs_dequeue_soft(struct Qdisc * sch)175*4882a593Smuzhiyun static struct sk_buff *cbs_dequeue_soft(struct Qdisc *sch)
176*4882a593Smuzhiyun {
177*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
178*4882a593Smuzhiyun 	struct Qdisc *qdisc = q->qdisc;
179*4882a593Smuzhiyun 	s64 now = ktime_get_ns();
180*4882a593Smuzhiyun 	struct sk_buff *skb;
181*4882a593Smuzhiyun 	s64 credits;
182*4882a593Smuzhiyun 	int len;
183*4882a593Smuzhiyun 
184*4882a593Smuzhiyun 	/* The previous packet is still being sent */
185*4882a593Smuzhiyun 	if (now < q->last) {
186*4882a593Smuzhiyun 		qdisc_watchdog_schedule_ns(&q->watchdog, q->last);
187*4882a593Smuzhiyun 		return NULL;
188*4882a593Smuzhiyun 	}
189*4882a593Smuzhiyun 	if (q->credits < 0) {
190*4882a593Smuzhiyun 		credits = timediff_to_credits(now - q->last, q->idleslope);
191*4882a593Smuzhiyun 
192*4882a593Smuzhiyun 		credits = q->credits + credits;
193*4882a593Smuzhiyun 		q->credits = min_t(s64, credits, q->hicredit);
194*4882a593Smuzhiyun 
195*4882a593Smuzhiyun 		if (q->credits < 0) {
196*4882a593Smuzhiyun 			s64 delay;
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun 			delay = delay_from_credits(q->credits, q->idleslope);
199*4882a593Smuzhiyun 			qdisc_watchdog_schedule_ns(&q->watchdog, now + delay);
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 			q->last = now;
202*4882a593Smuzhiyun 
203*4882a593Smuzhiyun 			return NULL;
204*4882a593Smuzhiyun 		}
205*4882a593Smuzhiyun 	}
206*4882a593Smuzhiyun 	skb = cbs_child_dequeue(sch, qdisc);
207*4882a593Smuzhiyun 	if (!skb)
208*4882a593Smuzhiyun 		return NULL;
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun 	len = qdisc_pkt_len(skb);
211*4882a593Smuzhiyun 
212*4882a593Smuzhiyun 	/* As sendslope is a negative number, this will decrease the
213*4882a593Smuzhiyun 	 * amount of q->credits.
214*4882a593Smuzhiyun 	 */
215*4882a593Smuzhiyun 	credits = credits_from_len(len, q->sendslope,
216*4882a593Smuzhiyun 				   atomic64_read(&q->port_rate));
217*4882a593Smuzhiyun 	credits += q->credits;
218*4882a593Smuzhiyun 
219*4882a593Smuzhiyun 	q->credits = max_t(s64, credits, q->locredit);
220*4882a593Smuzhiyun 	/* Estimate of the transmission of the last byte of the packet in ns */
221*4882a593Smuzhiyun 	if (unlikely(atomic64_read(&q->port_rate) == 0))
222*4882a593Smuzhiyun 		q->last = now;
223*4882a593Smuzhiyun 	else
224*4882a593Smuzhiyun 		q->last = now + div64_s64(len * NSEC_PER_SEC,
225*4882a593Smuzhiyun 					  atomic64_read(&q->port_rate));
226*4882a593Smuzhiyun 
227*4882a593Smuzhiyun 	return skb;
228*4882a593Smuzhiyun }
229*4882a593Smuzhiyun 
cbs_dequeue_offload(struct Qdisc * sch)230*4882a593Smuzhiyun static struct sk_buff *cbs_dequeue_offload(struct Qdisc *sch)
231*4882a593Smuzhiyun {
232*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
233*4882a593Smuzhiyun 	struct Qdisc *qdisc = q->qdisc;
234*4882a593Smuzhiyun 
235*4882a593Smuzhiyun 	return cbs_child_dequeue(sch, qdisc);
236*4882a593Smuzhiyun }
237*4882a593Smuzhiyun 
cbs_dequeue(struct Qdisc * sch)238*4882a593Smuzhiyun static struct sk_buff *cbs_dequeue(struct Qdisc *sch)
239*4882a593Smuzhiyun {
240*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
241*4882a593Smuzhiyun 
242*4882a593Smuzhiyun 	return q->dequeue(sch);
243*4882a593Smuzhiyun }
244*4882a593Smuzhiyun 
245*4882a593Smuzhiyun static const struct nla_policy cbs_policy[TCA_CBS_MAX + 1] = {
246*4882a593Smuzhiyun 	[TCA_CBS_PARMS]	= { .len = sizeof(struct tc_cbs_qopt) },
247*4882a593Smuzhiyun };
248*4882a593Smuzhiyun 
cbs_disable_offload(struct net_device * dev,struct cbs_sched_data * q)249*4882a593Smuzhiyun static void cbs_disable_offload(struct net_device *dev,
250*4882a593Smuzhiyun 				struct cbs_sched_data *q)
251*4882a593Smuzhiyun {
252*4882a593Smuzhiyun 	struct tc_cbs_qopt_offload cbs = { };
253*4882a593Smuzhiyun 	const struct net_device_ops *ops;
254*4882a593Smuzhiyun 	int err;
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun 	if (!q->offload)
257*4882a593Smuzhiyun 		return;
258*4882a593Smuzhiyun 
259*4882a593Smuzhiyun 	q->enqueue = cbs_enqueue_soft;
260*4882a593Smuzhiyun 	q->dequeue = cbs_dequeue_soft;
261*4882a593Smuzhiyun 
262*4882a593Smuzhiyun 	ops = dev->netdev_ops;
263*4882a593Smuzhiyun 	if (!ops->ndo_setup_tc)
264*4882a593Smuzhiyun 		return;
265*4882a593Smuzhiyun 
266*4882a593Smuzhiyun 	cbs.queue = q->queue;
267*4882a593Smuzhiyun 	cbs.enable = 0;
268*4882a593Smuzhiyun 
269*4882a593Smuzhiyun 	err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs);
270*4882a593Smuzhiyun 	if (err < 0)
271*4882a593Smuzhiyun 		pr_warn("Couldn't disable CBS offload for queue %d\n",
272*4882a593Smuzhiyun 			cbs.queue);
273*4882a593Smuzhiyun }
274*4882a593Smuzhiyun 
cbs_enable_offload(struct net_device * dev,struct cbs_sched_data * q,const struct tc_cbs_qopt * opt,struct netlink_ext_ack * extack)275*4882a593Smuzhiyun static int cbs_enable_offload(struct net_device *dev, struct cbs_sched_data *q,
276*4882a593Smuzhiyun 			      const struct tc_cbs_qopt *opt,
277*4882a593Smuzhiyun 			      struct netlink_ext_ack *extack)
278*4882a593Smuzhiyun {
279*4882a593Smuzhiyun 	const struct net_device_ops *ops = dev->netdev_ops;
280*4882a593Smuzhiyun 	struct tc_cbs_qopt_offload cbs = { };
281*4882a593Smuzhiyun 	int err;
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun 	if (!ops->ndo_setup_tc) {
284*4882a593Smuzhiyun 		NL_SET_ERR_MSG(extack, "Specified device does not support cbs offload");
285*4882a593Smuzhiyun 		return -EOPNOTSUPP;
286*4882a593Smuzhiyun 	}
287*4882a593Smuzhiyun 
288*4882a593Smuzhiyun 	cbs.queue = q->queue;
289*4882a593Smuzhiyun 
290*4882a593Smuzhiyun 	cbs.enable = 1;
291*4882a593Smuzhiyun 	cbs.hicredit = opt->hicredit;
292*4882a593Smuzhiyun 	cbs.locredit = opt->locredit;
293*4882a593Smuzhiyun 	cbs.idleslope = opt->idleslope;
294*4882a593Smuzhiyun 	cbs.sendslope = opt->sendslope;
295*4882a593Smuzhiyun 
296*4882a593Smuzhiyun 	err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs);
297*4882a593Smuzhiyun 	if (err < 0) {
298*4882a593Smuzhiyun 		NL_SET_ERR_MSG(extack, "Specified device failed to setup cbs hardware offload");
299*4882a593Smuzhiyun 		return err;
300*4882a593Smuzhiyun 	}
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun 	q->enqueue = cbs_enqueue_offload;
303*4882a593Smuzhiyun 	q->dequeue = cbs_dequeue_offload;
304*4882a593Smuzhiyun 
305*4882a593Smuzhiyun 	return 0;
306*4882a593Smuzhiyun }
307*4882a593Smuzhiyun 
cbs_set_port_rate(struct net_device * dev,struct cbs_sched_data * q)308*4882a593Smuzhiyun static void cbs_set_port_rate(struct net_device *dev, struct cbs_sched_data *q)
309*4882a593Smuzhiyun {
310*4882a593Smuzhiyun 	struct ethtool_link_ksettings ecmd;
311*4882a593Smuzhiyun 	int speed = SPEED_10;
312*4882a593Smuzhiyun 	int port_rate;
313*4882a593Smuzhiyun 	int err;
314*4882a593Smuzhiyun 
315*4882a593Smuzhiyun 	err = __ethtool_get_link_ksettings(dev, &ecmd);
316*4882a593Smuzhiyun 	if (err < 0)
317*4882a593Smuzhiyun 		goto skip;
318*4882a593Smuzhiyun 
319*4882a593Smuzhiyun 	if (ecmd.base.speed && ecmd.base.speed != SPEED_UNKNOWN)
320*4882a593Smuzhiyun 		speed = ecmd.base.speed;
321*4882a593Smuzhiyun 
322*4882a593Smuzhiyun skip:
323*4882a593Smuzhiyun 	port_rate = speed * 1000 * BYTES_PER_KBIT;
324*4882a593Smuzhiyun 
325*4882a593Smuzhiyun 	atomic64_set(&q->port_rate, port_rate);
326*4882a593Smuzhiyun 	netdev_dbg(dev, "cbs: set %s's port_rate to: %lld, linkspeed: %d\n",
327*4882a593Smuzhiyun 		   dev->name, (long long)atomic64_read(&q->port_rate),
328*4882a593Smuzhiyun 		   ecmd.base.speed);
329*4882a593Smuzhiyun }
330*4882a593Smuzhiyun 
cbs_dev_notifier(struct notifier_block * nb,unsigned long event,void * ptr)331*4882a593Smuzhiyun static int cbs_dev_notifier(struct notifier_block *nb, unsigned long event,
332*4882a593Smuzhiyun 			    void *ptr)
333*4882a593Smuzhiyun {
334*4882a593Smuzhiyun 	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
335*4882a593Smuzhiyun 	struct cbs_sched_data *q;
336*4882a593Smuzhiyun 	struct net_device *qdev;
337*4882a593Smuzhiyun 	bool found = false;
338*4882a593Smuzhiyun 
339*4882a593Smuzhiyun 	ASSERT_RTNL();
340*4882a593Smuzhiyun 
341*4882a593Smuzhiyun 	if (event != NETDEV_UP && event != NETDEV_CHANGE)
342*4882a593Smuzhiyun 		return NOTIFY_DONE;
343*4882a593Smuzhiyun 
344*4882a593Smuzhiyun 	spin_lock(&cbs_list_lock);
345*4882a593Smuzhiyun 	list_for_each_entry(q, &cbs_list, cbs_list) {
346*4882a593Smuzhiyun 		qdev = qdisc_dev(q->qdisc);
347*4882a593Smuzhiyun 		if (qdev == dev) {
348*4882a593Smuzhiyun 			found = true;
349*4882a593Smuzhiyun 			break;
350*4882a593Smuzhiyun 		}
351*4882a593Smuzhiyun 	}
352*4882a593Smuzhiyun 	spin_unlock(&cbs_list_lock);
353*4882a593Smuzhiyun 
354*4882a593Smuzhiyun 	if (found)
355*4882a593Smuzhiyun 		cbs_set_port_rate(dev, q);
356*4882a593Smuzhiyun 
357*4882a593Smuzhiyun 	return NOTIFY_DONE;
358*4882a593Smuzhiyun }
359*4882a593Smuzhiyun 
cbs_change(struct Qdisc * sch,struct nlattr * opt,struct netlink_ext_ack * extack)360*4882a593Smuzhiyun static int cbs_change(struct Qdisc *sch, struct nlattr *opt,
361*4882a593Smuzhiyun 		      struct netlink_ext_ack *extack)
362*4882a593Smuzhiyun {
363*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
364*4882a593Smuzhiyun 	struct net_device *dev = qdisc_dev(sch);
365*4882a593Smuzhiyun 	struct nlattr *tb[TCA_CBS_MAX + 1];
366*4882a593Smuzhiyun 	struct tc_cbs_qopt *qopt;
367*4882a593Smuzhiyun 	int err;
368*4882a593Smuzhiyun 
369*4882a593Smuzhiyun 	err = nla_parse_nested_deprecated(tb, TCA_CBS_MAX, opt, cbs_policy,
370*4882a593Smuzhiyun 					  extack);
371*4882a593Smuzhiyun 	if (err < 0)
372*4882a593Smuzhiyun 		return err;
373*4882a593Smuzhiyun 
374*4882a593Smuzhiyun 	if (!tb[TCA_CBS_PARMS]) {
375*4882a593Smuzhiyun 		NL_SET_ERR_MSG(extack, "Missing CBS parameter which are mandatory");
376*4882a593Smuzhiyun 		return -EINVAL;
377*4882a593Smuzhiyun 	}
378*4882a593Smuzhiyun 
379*4882a593Smuzhiyun 	qopt = nla_data(tb[TCA_CBS_PARMS]);
380*4882a593Smuzhiyun 
381*4882a593Smuzhiyun 	if (!qopt->offload) {
382*4882a593Smuzhiyun 		cbs_set_port_rate(dev, q);
383*4882a593Smuzhiyun 		cbs_disable_offload(dev, q);
384*4882a593Smuzhiyun 	} else {
385*4882a593Smuzhiyun 		err = cbs_enable_offload(dev, q, qopt, extack);
386*4882a593Smuzhiyun 		if (err < 0)
387*4882a593Smuzhiyun 			return err;
388*4882a593Smuzhiyun 	}
389*4882a593Smuzhiyun 
390*4882a593Smuzhiyun 	/* Everything went OK, save the parameters used. */
391*4882a593Smuzhiyun 	q->hicredit = qopt->hicredit;
392*4882a593Smuzhiyun 	q->locredit = qopt->locredit;
393*4882a593Smuzhiyun 	q->idleslope = qopt->idleslope * BYTES_PER_KBIT;
394*4882a593Smuzhiyun 	q->sendslope = qopt->sendslope * BYTES_PER_KBIT;
395*4882a593Smuzhiyun 	q->offload = qopt->offload;
396*4882a593Smuzhiyun 
397*4882a593Smuzhiyun 	return 0;
398*4882a593Smuzhiyun }
399*4882a593Smuzhiyun 
cbs_init(struct Qdisc * sch,struct nlattr * opt,struct netlink_ext_ack * extack)400*4882a593Smuzhiyun static int cbs_init(struct Qdisc *sch, struct nlattr *opt,
401*4882a593Smuzhiyun 		    struct netlink_ext_ack *extack)
402*4882a593Smuzhiyun {
403*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
404*4882a593Smuzhiyun 	struct net_device *dev = qdisc_dev(sch);
405*4882a593Smuzhiyun 
406*4882a593Smuzhiyun 	if (!opt) {
407*4882a593Smuzhiyun 		NL_SET_ERR_MSG(extack, "Missing CBS qdisc options  which are mandatory");
408*4882a593Smuzhiyun 		return -EINVAL;
409*4882a593Smuzhiyun 	}
410*4882a593Smuzhiyun 
411*4882a593Smuzhiyun 	q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
412*4882a593Smuzhiyun 				     sch->handle, extack);
413*4882a593Smuzhiyun 	if (!q->qdisc)
414*4882a593Smuzhiyun 		return -ENOMEM;
415*4882a593Smuzhiyun 
416*4882a593Smuzhiyun 	spin_lock(&cbs_list_lock);
417*4882a593Smuzhiyun 	list_add(&q->cbs_list, &cbs_list);
418*4882a593Smuzhiyun 	spin_unlock(&cbs_list_lock);
419*4882a593Smuzhiyun 
420*4882a593Smuzhiyun 	qdisc_hash_add(q->qdisc, false);
421*4882a593Smuzhiyun 
422*4882a593Smuzhiyun 	q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);
423*4882a593Smuzhiyun 
424*4882a593Smuzhiyun 	q->enqueue = cbs_enqueue_soft;
425*4882a593Smuzhiyun 	q->dequeue = cbs_dequeue_soft;
426*4882a593Smuzhiyun 
427*4882a593Smuzhiyun 	qdisc_watchdog_init(&q->watchdog, sch);
428*4882a593Smuzhiyun 
429*4882a593Smuzhiyun 	return cbs_change(sch, opt, extack);
430*4882a593Smuzhiyun }
431*4882a593Smuzhiyun 
cbs_destroy(struct Qdisc * sch)432*4882a593Smuzhiyun static void cbs_destroy(struct Qdisc *sch)
433*4882a593Smuzhiyun {
434*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
435*4882a593Smuzhiyun 	struct net_device *dev = qdisc_dev(sch);
436*4882a593Smuzhiyun 
437*4882a593Smuzhiyun 	/* Nothing to do if we couldn't create the underlying qdisc */
438*4882a593Smuzhiyun 	if (!q->qdisc)
439*4882a593Smuzhiyun 		return;
440*4882a593Smuzhiyun 
441*4882a593Smuzhiyun 	qdisc_watchdog_cancel(&q->watchdog);
442*4882a593Smuzhiyun 	cbs_disable_offload(dev, q);
443*4882a593Smuzhiyun 
444*4882a593Smuzhiyun 	spin_lock(&cbs_list_lock);
445*4882a593Smuzhiyun 	list_del(&q->cbs_list);
446*4882a593Smuzhiyun 	spin_unlock(&cbs_list_lock);
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	qdisc_put(q->qdisc);
449*4882a593Smuzhiyun }
450*4882a593Smuzhiyun 
cbs_dump(struct Qdisc * sch,struct sk_buff * skb)451*4882a593Smuzhiyun static int cbs_dump(struct Qdisc *sch, struct sk_buff *skb)
452*4882a593Smuzhiyun {
453*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
454*4882a593Smuzhiyun 	struct tc_cbs_qopt opt = { };
455*4882a593Smuzhiyun 	struct nlattr *nest;
456*4882a593Smuzhiyun 
457*4882a593Smuzhiyun 	nest = nla_nest_start_noflag(skb, TCA_OPTIONS);
458*4882a593Smuzhiyun 	if (!nest)
459*4882a593Smuzhiyun 		goto nla_put_failure;
460*4882a593Smuzhiyun 
461*4882a593Smuzhiyun 	opt.hicredit = q->hicredit;
462*4882a593Smuzhiyun 	opt.locredit = q->locredit;
463*4882a593Smuzhiyun 	opt.sendslope = div64_s64(q->sendslope, BYTES_PER_KBIT);
464*4882a593Smuzhiyun 	opt.idleslope = div64_s64(q->idleslope, BYTES_PER_KBIT);
465*4882a593Smuzhiyun 	opt.offload = q->offload;
466*4882a593Smuzhiyun 
467*4882a593Smuzhiyun 	if (nla_put(skb, TCA_CBS_PARMS, sizeof(opt), &opt))
468*4882a593Smuzhiyun 		goto nla_put_failure;
469*4882a593Smuzhiyun 
470*4882a593Smuzhiyun 	return nla_nest_end(skb, nest);
471*4882a593Smuzhiyun 
472*4882a593Smuzhiyun nla_put_failure:
473*4882a593Smuzhiyun 	nla_nest_cancel(skb, nest);
474*4882a593Smuzhiyun 	return -1;
475*4882a593Smuzhiyun }
476*4882a593Smuzhiyun 
cbs_dump_class(struct Qdisc * sch,unsigned long cl,struct sk_buff * skb,struct tcmsg * tcm)477*4882a593Smuzhiyun static int cbs_dump_class(struct Qdisc *sch, unsigned long cl,
478*4882a593Smuzhiyun 			  struct sk_buff *skb, struct tcmsg *tcm)
479*4882a593Smuzhiyun {
480*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
481*4882a593Smuzhiyun 
482*4882a593Smuzhiyun 	if (cl != 1 || !q->qdisc)	/* only one class */
483*4882a593Smuzhiyun 		return -ENOENT;
484*4882a593Smuzhiyun 
485*4882a593Smuzhiyun 	tcm->tcm_handle |= TC_H_MIN(1);
486*4882a593Smuzhiyun 	tcm->tcm_info = q->qdisc->handle;
487*4882a593Smuzhiyun 
488*4882a593Smuzhiyun 	return 0;
489*4882a593Smuzhiyun }
490*4882a593Smuzhiyun 
cbs_graft(struct Qdisc * sch,unsigned long arg,struct Qdisc * new,struct Qdisc ** old,struct netlink_ext_ack * extack)491*4882a593Smuzhiyun static int cbs_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new,
492*4882a593Smuzhiyun 		     struct Qdisc **old, struct netlink_ext_ack *extack)
493*4882a593Smuzhiyun {
494*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
495*4882a593Smuzhiyun 
496*4882a593Smuzhiyun 	if (!new) {
497*4882a593Smuzhiyun 		new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
498*4882a593Smuzhiyun 					sch->handle, NULL);
499*4882a593Smuzhiyun 		if (!new)
500*4882a593Smuzhiyun 			new = &noop_qdisc;
501*4882a593Smuzhiyun 	}
502*4882a593Smuzhiyun 
503*4882a593Smuzhiyun 	*old = qdisc_replace(sch, new, &q->qdisc);
504*4882a593Smuzhiyun 	return 0;
505*4882a593Smuzhiyun }
506*4882a593Smuzhiyun 
cbs_leaf(struct Qdisc * sch,unsigned long arg)507*4882a593Smuzhiyun static struct Qdisc *cbs_leaf(struct Qdisc *sch, unsigned long arg)
508*4882a593Smuzhiyun {
509*4882a593Smuzhiyun 	struct cbs_sched_data *q = qdisc_priv(sch);
510*4882a593Smuzhiyun 
511*4882a593Smuzhiyun 	return q->qdisc;
512*4882a593Smuzhiyun }
513*4882a593Smuzhiyun 
cbs_find(struct Qdisc * sch,u32 classid)514*4882a593Smuzhiyun static unsigned long cbs_find(struct Qdisc *sch, u32 classid)
515*4882a593Smuzhiyun {
516*4882a593Smuzhiyun 	return 1;
517*4882a593Smuzhiyun }
518*4882a593Smuzhiyun 
cbs_walk(struct Qdisc * sch,struct qdisc_walker * walker)519*4882a593Smuzhiyun static void cbs_walk(struct Qdisc *sch, struct qdisc_walker *walker)
520*4882a593Smuzhiyun {
521*4882a593Smuzhiyun 	if (!walker->stop) {
522*4882a593Smuzhiyun 		if (walker->count >= walker->skip) {
523*4882a593Smuzhiyun 			if (walker->fn(sch, 1, walker) < 0) {
524*4882a593Smuzhiyun 				walker->stop = 1;
525*4882a593Smuzhiyun 				return;
526*4882a593Smuzhiyun 			}
527*4882a593Smuzhiyun 		}
528*4882a593Smuzhiyun 		walker->count++;
529*4882a593Smuzhiyun 	}
530*4882a593Smuzhiyun }
531*4882a593Smuzhiyun 
532*4882a593Smuzhiyun static const struct Qdisc_class_ops cbs_class_ops = {
533*4882a593Smuzhiyun 	.graft		=	cbs_graft,
534*4882a593Smuzhiyun 	.leaf		=	cbs_leaf,
535*4882a593Smuzhiyun 	.find		=	cbs_find,
536*4882a593Smuzhiyun 	.walk		=	cbs_walk,
537*4882a593Smuzhiyun 	.dump		=	cbs_dump_class,
538*4882a593Smuzhiyun };
539*4882a593Smuzhiyun 
540*4882a593Smuzhiyun static struct Qdisc_ops cbs_qdisc_ops __read_mostly = {
541*4882a593Smuzhiyun 	.id		=	"cbs",
542*4882a593Smuzhiyun 	.cl_ops		=	&cbs_class_ops,
543*4882a593Smuzhiyun 	.priv_size	=	sizeof(struct cbs_sched_data),
544*4882a593Smuzhiyun 	.enqueue	=	cbs_enqueue,
545*4882a593Smuzhiyun 	.dequeue	=	cbs_dequeue,
546*4882a593Smuzhiyun 	.peek		=	qdisc_peek_dequeued,
547*4882a593Smuzhiyun 	.init		=	cbs_init,
548*4882a593Smuzhiyun 	.reset		=	qdisc_reset_queue,
549*4882a593Smuzhiyun 	.destroy	=	cbs_destroy,
550*4882a593Smuzhiyun 	.change		=	cbs_change,
551*4882a593Smuzhiyun 	.dump		=	cbs_dump,
552*4882a593Smuzhiyun 	.owner		=	THIS_MODULE,
553*4882a593Smuzhiyun };
554*4882a593Smuzhiyun 
555*4882a593Smuzhiyun static struct notifier_block cbs_device_notifier = {
556*4882a593Smuzhiyun 	.notifier_call = cbs_dev_notifier,
557*4882a593Smuzhiyun };
558*4882a593Smuzhiyun 
cbs_module_init(void)559*4882a593Smuzhiyun static int __init cbs_module_init(void)
560*4882a593Smuzhiyun {
561*4882a593Smuzhiyun 	int err;
562*4882a593Smuzhiyun 
563*4882a593Smuzhiyun 	err = register_netdevice_notifier(&cbs_device_notifier);
564*4882a593Smuzhiyun 	if (err)
565*4882a593Smuzhiyun 		return err;
566*4882a593Smuzhiyun 
567*4882a593Smuzhiyun 	err = register_qdisc(&cbs_qdisc_ops);
568*4882a593Smuzhiyun 	if (err)
569*4882a593Smuzhiyun 		unregister_netdevice_notifier(&cbs_device_notifier);
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun 	return err;
572*4882a593Smuzhiyun }
573*4882a593Smuzhiyun 
cbs_module_exit(void)574*4882a593Smuzhiyun static void __exit cbs_module_exit(void)
575*4882a593Smuzhiyun {
576*4882a593Smuzhiyun 	unregister_qdisc(&cbs_qdisc_ops);
577*4882a593Smuzhiyun 	unregister_netdevice_notifier(&cbs_device_notifier);
578*4882a593Smuzhiyun }
579*4882a593Smuzhiyun module_init(cbs_module_init)
580*4882a593Smuzhiyun module_exit(cbs_module_exit)
581*4882a593Smuzhiyun MODULE_LICENSE("GPL");
582