1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * net/sched/sch_cbs.c Credit Based Shaper
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Authors: Vinicius Costa Gomes <vinicius.gomes@intel.com>
6*4882a593Smuzhiyun */
7*4882a593Smuzhiyun
8*4882a593Smuzhiyun /* Credit Based Shaper (CBS)
9*4882a593Smuzhiyun * =========================
10*4882a593Smuzhiyun *
11*4882a593Smuzhiyun * This is a simple rate-limiting shaper aimed at TSN applications on
12*4882a593Smuzhiyun * systems with known traffic workloads.
13*4882a593Smuzhiyun *
14*4882a593Smuzhiyun * Its algorithm is defined by the IEEE 802.1Q-2014 Specification,
15*4882a593Smuzhiyun * Section 8.6.8.2, and explained in more detail in the Annex L of the
16*4882a593Smuzhiyun * same specification.
17*4882a593Smuzhiyun *
18*4882a593Smuzhiyun * There are four tunables to be considered:
19*4882a593Smuzhiyun *
20*4882a593Smuzhiyun * 'idleslope': Idleslope is the rate of credits that is
21*4882a593Smuzhiyun * accumulated (in kilobits per second) when there is at least
22*4882a593Smuzhiyun * one packet waiting for transmission. Packets are transmitted
23*4882a593Smuzhiyun * when the current value of credits is equal or greater than
24*4882a593Smuzhiyun * zero. When there is no packet to be transmitted the amount of
25*4882a593Smuzhiyun * credits is set to zero. This is the main tunable of the CBS
26*4882a593Smuzhiyun * algorithm.
27*4882a593Smuzhiyun *
28*4882a593Smuzhiyun * 'sendslope':
29*4882a593Smuzhiyun * Sendslope is the rate of credits that is depleted (it should be a
30*4882a593Smuzhiyun * negative number of kilobits per second) when a transmission is
31*4882a593Smuzhiyun * ocurring. It can be calculated as follows, (IEEE 802.1Q-2014 Section
32*4882a593Smuzhiyun * 8.6.8.2 item g):
33*4882a593Smuzhiyun *
34*4882a593Smuzhiyun * sendslope = idleslope - port_transmit_rate
35*4882a593Smuzhiyun *
36*4882a593Smuzhiyun * 'hicredit': Hicredit defines the maximum amount of credits (in
37*4882a593Smuzhiyun * bytes) that can be accumulated. Hicredit depends on the
38*4882a593Smuzhiyun * characteristics of interfering traffic,
39*4882a593Smuzhiyun * 'max_interference_size' is the maximum size of any burst of
40*4882a593Smuzhiyun * traffic that can delay the transmission of a frame that is
41*4882a593Smuzhiyun * available for transmission for this traffic class, (IEEE
42*4882a593Smuzhiyun * 802.1Q-2014 Annex L, Equation L-3):
43*4882a593Smuzhiyun *
44*4882a593Smuzhiyun * hicredit = max_interference_size * (idleslope / port_transmit_rate)
45*4882a593Smuzhiyun *
46*4882a593Smuzhiyun * 'locredit': Locredit is the minimum amount of credits that can
47*4882a593Smuzhiyun * be reached. It is a function of the traffic flowing through
48*4882a593Smuzhiyun * this qdisc (IEEE 802.1Q-2014 Annex L, Equation L-2):
49*4882a593Smuzhiyun *
50*4882a593Smuzhiyun * locredit = max_frame_size * (sendslope / port_transmit_rate)
51*4882a593Smuzhiyun */
52*4882a593Smuzhiyun
53*4882a593Smuzhiyun #include <linux/module.h>
54*4882a593Smuzhiyun #include <linux/types.h>
55*4882a593Smuzhiyun #include <linux/kernel.h>
56*4882a593Smuzhiyun #include <linux/string.h>
57*4882a593Smuzhiyun #include <linux/errno.h>
58*4882a593Smuzhiyun #include <linux/skbuff.h>
59*4882a593Smuzhiyun #include <net/netevent.h>
60*4882a593Smuzhiyun #include <net/netlink.h>
61*4882a593Smuzhiyun #include <net/sch_generic.h>
62*4882a593Smuzhiyun #include <net/pkt_sched.h>
63*4882a593Smuzhiyun
64*4882a593Smuzhiyun static LIST_HEAD(cbs_list);
65*4882a593Smuzhiyun static DEFINE_SPINLOCK(cbs_list_lock);
66*4882a593Smuzhiyun
67*4882a593Smuzhiyun #define BYTES_PER_KBIT (1000LL / 8)
68*4882a593Smuzhiyun
69*4882a593Smuzhiyun struct cbs_sched_data {
70*4882a593Smuzhiyun bool offload;
71*4882a593Smuzhiyun int queue;
72*4882a593Smuzhiyun atomic64_t port_rate; /* in bytes/s */
73*4882a593Smuzhiyun s64 last; /* timestamp in ns */
74*4882a593Smuzhiyun s64 credits; /* in bytes */
75*4882a593Smuzhiyun s32 locredit; /* in bytes */
76*4882a593Smuzhiyun s32 hicredit; /* in bytes */
77*4882a593Smuzhiyun s64 sendslope; /* in bytes/s */
78*4882a593Smuzhiyun s64 idleslope; /* in bytes/s */
79*4882a593Smuzhiyun struct qdisc_watchdog watchdog;
80*4882a593Smuzhiyun int (*enqueue)(struct sk_buff *skb, struct Qdisc *sch,
81*4882a593Smuzhiyun struct sk_buff **to_free);
82*4882a593Smuzhiyun struct sk_buff *(*dequeue)(struct Qdisc *sch);
83*4882a593Smuzhiyun struct Qdisc *qdisc;
84*4882a593Smuzhiyun struct list_head cbs_list;
85*4882a593Smuzhiyun };
86*4882a593Smuzhiyun
cbs_child_enqueue(struct sk_buff * skb,struct Qdisc * sch,struct Qdisc * child,struct sk_buff ** to_free)87*4882a593Smuzhiyun static int cbs_child_enqueue(struct sk_buff *skb, struct Qdisc *sch,
88*4882a593Smuzhiyun struct Qdisc *child,
89*4882a593Smuzhiyun struct sk_buff **to_free)
90*4882a593Smuzhiyun {
91*4882a593Smuzhiyun unsigned int len = qdisc_pkt_len(skb);
92*4882a593Smuzhiyun int err;
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun err = child->ops->enqueue(skb, child, to_free);
95*4882a593Smuzhiyun if (err != NET_XMIT_SUCCESS)
96*4882a593Smuzhiyun return err;
97*4882a593Smuzhiyun
98*4882a593Smuzhiyun sch->qstats.backlog += len;
99*4882a593Smuzhiyun sch->q.qlen++;
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun return NET_XMIT_SUCCESS;
102*4882a593Smuzhiyun }
103*4882a593Smuzhiyun
cbs_enqueue_offload(struct sk_buff * skb,struct Qdisc * sch,struct sk_buff ** to_free)104*4882a593Smuzhiyun static int cbs_enqueue_offload(struct sk_buff *skb, struct Qdisc *sch,
105*4882a593Smuzhiyun struct sk_buff **to_free)
106*4882a593Smuzhiyun {
107*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
108*4882a593Smuzhiyun struct Qdisc *qdisc = q->qdisc;
109*4882a593Smuzhiyun
110*4882a593Smuzhiyun return cbs_child_enqueue(skb, sch, qdisc, to_free);
111*4882a593Smuzhiyun }
112*4882a593Smuzhiyun
cbs_enqueue_soft(struct sk_buff * skb,struct Qdisc * sch,struct sk_buff ** to_free)113*4882a593Smuzhiyun static int cbs_enqueue_soft(struct sk_buff *skb, struct Qdisc *sch,
114*4882a593Smuzhiyun struct sk_buff **to_free)
115*4882a593Smuzhiyun {
116*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
117*4882a593Smuzhiyun struct Qdisc *qdisc = q->qdisc;
118*4882a593Smuzhiyun
119*4882a593Smuzhiyun if (sch->q.qlen == 0 && q->credits > 0) {
120*4882a593Smuzhiyun /* We need to stop accumulating credits when there's
121*4882a593Smuzhiyun * no enqueued packets and q->credits is positive.
122*4882a593Smuzhiyun */
123*4882a593Smuzhiyun q->credits = 0;
124*4882a593Smuzhiyun q->last = ktime_get_ns();
125*4882a593Smuzhiyun }
126*4882a593Smuzhiyun
127*4882a593Smuzhiyun return cbs_child_enqueue(skb, sch, qdisc, to_free);
128*4882a593Smuzhiyun }
129*4882a593Smuzhiyun
cbs_enqueue(struct sk_buff * skb,struct Qdisc * sch,struct sk_buff ** to_free)130*4882a593Smuzhiyun static int cbs_enqueue(struct sk_buff *skb, struct Qdisc *sch,
131*4882a593Smuzhiyun struct sk_buff **to_free)
132*4882a593Smuzhiyun {
133*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
134*4882a593Smuzhiyun
135*4882a593Smuzhiyun return q->enqueue(skb, sch, to_free);
136*4882a593Smuzhiyun }
137*4882a593Smuzhiyun
138*4882a593Smuzhiyun /* timediff is in ns, slope is in bytes/s */
timediff_to_credits(s64 timediff,s64 slope)139*4882a593Smuzhiyun static s64 timediff_to_credits(s64 timediff, s64 slope)
140*4882a593Smuzhiyun {
141*4882a593Smuzhiyun return div64_s64(timediff * slope, NSEC_PER_SEC);
142*4882a593Smuzhiyun }
143*4882a593Smuzhiyun
delay_from_credits(s64 credits,s64 slope)144*4882a593Smuzhiyun static s64 delay_from_credits(s64 credits, s64 slope)
145*4882a593Smuzhiyun {
146*4882a593Smuzhiyun if (unlikely(slope == 0))
147*4882a593Smuzhiyun return S64_MAX;
148*4882a593Smuzhiyun
149*4882a593Smuzhiyun return div64_s64(-credits * NSEC_PER_SEC, slope);
150*4882a593Smuzhiyun }
151*4882a593Smuzhiyun
credits_from_len(unsigned int len,s64 slope,s64 port_rate)152*4882a593Smuzhiyun static s64 credits_from_len(unsigned int len, s64 slope, s64 port_rate)
153*4882a593Smuzhiyun {
154*4882a593Smuzhiyun if (unlikely(port_rate == 0))
155*4882a593Smuzhiyun return S64_MAX;
156*4882a593Smuzhiyun
157*4882a593Smuzhiyun return div64_s64(len * slope, port_rate);
158*4882a593Smuzhiyun }
159*4882a593Smuzhiyun
cbs_child_dequeue(struct Qdisc * sch,struct Qdisc * child)160*4882a593Smuzhiyun static struct sk_buff *cbs_child_dequeue(struct Qdisc *sch, struct Qdisc *child)
161*4882a593Smuzhiyun {
162*4882a593Smuzhiyun struct sk_buff *skb;
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun skb = child->ops->dequeue(child);
165*4882a593Smuzhiyun if (!skb)
166*4882a593Smuzhiyun return NULL;
167*4882a593Smuzhiyun
168*4882a593Smuzhiyun qdisc_qstats_backlog_dec(sch, skb);
169*4882a593Smuzhiyun qdisc_bstats_update(sch, skb);
170*4882a593Smuzhiyun sch->q.qlen--;
171*4882a593Smuzhiyun
172*4882a593Smuzhiyun return skb;
173*4882a593Smuzhiyun }
174*4882a593Smuzhiyun
cbs_dequeue_soft(struct Qdisc * sch)175*4882a593Smuzhiyun static struct sk_buff *cbs_dequeue_soft(struct Qdisc *sch)
176*4882a593Smuzhiyun {
177*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
178*4882a593Smuzhiyun struct Qdisc *qdisc = q->qdisc;
179*4882a593Smuzhiyun s64 now = ktime_get_ns();
180*4882a593Smuzhiyun struct sk_buff *skb;
181*4882a593Smuzhiyun s64 credits;
182*4882a593Smuzhiyun int len;
183*4882a593Smuzhiyun
184*4882a593Smuzhiyun /* The previous packet is still being sent */
185*4882a593Smuzhiyun if (now < q->last) {
186*4882a593Smuzhiyun qdisc_watchdog_schedule_ns(&q->watchdog, q->last);
187*4882a593Smuzhiyun return NULL;
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun if (q->credits < 0) {
190*4882a593Smuzhiyun credits = timediff_to_credits(now - q->last, q->idleslope);
191*4882a593Smuzhiyun
192*4882a593Smuzhiyun credits = q->credits + credits;
193*4882a593Smuzhiyun q->credits = min_t(s64, credits, q->hicredit);
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun if (q->credits < 0) {
196*4882a593Smuzhiyun s64 delay;
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun delay = delay_from_credits(q->credits, q->idleslope);
199*4882a593Smuzhiyun qdisc_watchdog_schedule_ns(&q->watchdog, now + delay);
200*4882a593Smuzhiyun
201*4882a593Smuzhiyun q->last = now;
202*4882a593Smuzhiyun
203*4882a593Smuzhiyun return NULL;
204*4882a593Smuzhiyun }
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun skb = cbs_child_dequeue(sch, qdisc);
207*4882a593Smuzhiyun if (!skb)
208*4882a593Smuzhiyun return NULL;
209*4882a593Smuzhiyun
210*4882a593Smuzhiyun len = qdisc_pkt_len(skb);
211*4882a593Smuzhiyun
212*4882a593Smuzhiyun /* As sendslope is a negative number, this will decrease the
213*4882a593Smuzhiyun * amount of q->credits.
214*4882a593Smuzhiyun */
215*4882a593Smuzhiyun credits = credits_from_len(len, q->sendslope,
216*4882a593Smuzhiyun atomic64_read(&q->port_rate));
217*4882a593Smuzhiyun credits += q->credits;
218*4882a593Smuzhiyun
219*4882a593Smuzhiyun q->credits = max_t(s64, credits, q->locredit);
220*4882a593Smuzhiyun /* Estimate of the transmission of the last byte of the packet in ns */
221*4882a593Smuzhiyun if (unlikely(atomic64_read(&q->port_rate) == 0))
222*4882a593Smuzhiyun q->last = now;
223*4882a593Smuzhiyun else
224*4882a593Smuzhiyun q->last = now + div64_s64(len * NSEC_PER_SEC,
225*4882a593Smuzhiyun atomic64_read(&q->port_rate));
226*4882a593Smuzhiyun
227*4882a593Smuzhiyun return skb;
228*4882a593Smuzhiyun }
229*4882a593Smuzhiyun
cbs_dequeue_offload(struct Qdisc * sch)230*4882a593Smuzhiyun static struct sk_buff *cbs_dequeue_offload(struct Qdisc *sch)
231*4882a593Smuzhiyun {
232*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
233*4882a593Smuzhiyun struct Qdisc *qdisc = q->qdisc;
234*4882a593Smuzhiyun
235*4882a593Smuzhiyun return cbs_child_dequeue(sch, qdisc);
236*4882a593Smuzhiyun }
237*4882a593Smuzhiyun
cbs_dequeue(struct Qdisc * sch)238*4882a593Smuzhiyun static struct sk_buff *cbs_dequeue(struct Qdisc *sch)
239*4882a593Smuzhiyun {
240*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
241*4882a593Smuzhiyun
242*4882a593Smuzhiyun return q->dequeue(sch);
243*4882a593Smuzhiyun }
244*4882a593Smuzhiyun
245*4882a593Smuzhiyun static const struct nla_policy cbs_policy[TCA_CBS_MAX + 1] = {
246*4882a593Smuzhiyun [TCA_CBS_PARMS] = { .len = sizeof(struct tc_cbs_qopt) },
247*4882a593Smuzhiyun };
248*4882a593Smuzhiyun
cbs_disable_offload(struct net_device * dev,struct cbs_sched_data * q)249*4882a593Smuzhiyun static void cbs_disable_offload(struct net_device *dev,
250*4882a593Smuzhiyun struct cbs_sched_data *q)
251*4882a593Smuzhiyun {
252*4882a593Smuzhiyun struct tc_cbs_qopt_offload cbs = { };
253*4882a593Smuzhiyun const struct net_device_ops *ops;
254*4882a593Smuzhiyun int err;
255*4882a593Smuzhiyun
256*4882a593Smuzhiyun if (!q->offload)
257*4882a593Smuzhiyun return;
258*4882a593Smuzhiyun
259*4882a593Smuzhiyun q->enqueue = cbs_enqueue_soft;
260*4882a593Smuzhiyun q->dequeue = cbs_dequeue_soft;
261*4882a593Smuzhiyun
262*4882a593Smuzhiyun ops = dev->netdev_ops;
263*4882a593Smuzhiyun if (!ops->ndo_setup_tc)
264*4882a593Smuzhiyun return;
265*4882a593Smuzhiyun
266*4882a593Smuzhiyun cbs.queue = q->queue;
267*4882a593Smuzhiyun cbs.enable = 0;
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs);
270*4882a593Smuzhiyun if (err < 0)
271*4882a593Smuzhiyun pr_warn("Couldn't disable CBS offload for queue %d\n",
272*4882a593Smuzhiyun cbs.queue);
273*4882a593Smuzhiyun }
274*4882a593Smuzhiyun
cbs_enable_offload(struct net_device * dev,struct cbs_sched_data * q,const struct tc_cbs_qopt * opt,struct netlink_ext_ack * extack)275*4882a593Smuzhiyun static int cbs_enable_offload(struct net_device *dev, struct cbs_sched_data *q,
276*4882a593Smuzhiyun const struct tc_cbs_qopt *opt,
277*4882a593Smuzhiyun struct netlink_ext_ack *extack)
278*4882a593Smuzhiyun {
279*4882a593Smuzhiyun const struct net_device_ops *ops = dev->netdev_ops;
280*4882a593Smuzhiyun struct tc_cbs_qopt_offload cbs = { };
281*4882a593Smuzhiyun int err;
282*4882a593Smuzhiyun
283*4882a593Smuzhiyun if (!ops->ndo_setup_tc) {
284*4882a593Smuzhiyun NL_SET_ERR_MSG(extack, "Specified device does not support cbs offload");
285*4882a593Smuzhiyun return -EOPNOTSUPP;
286*4882a593Smuzhiyun }
287*4882a593Smuzhiyun
288*4882a593Smuzhiyun cbs.queue = q->queue;
289*4882a593Smuzhiyun
290*4882a593Smuzhiyun cbs.enable = 1;
291*4882a593Smuzhiyun cbs.hicredit = opt->hicredit;
292*4882a593Smuzhiyun cbs.locredit = opt->locredit;
293*4882a593Smuzhiyun cbs.idleslope = opt->idleslope;
294*4882a593Smuzhiyun cbs.sendslope = opt->sendslope;
295*4882a593Smuzhiyun
296*4882a593Smuzhiyun err = ops->ndo_setup_tc(dev, TC_SETUP_QDISC_CBS, &cbs);
297*4882a593Smuzhiyun if (err < 0) {
298*4882a593Smuzhiyun NL_SET_ERR_MSG(extack, "Specified device failed to setup cbs hardware offload");
299*4882a593Smuzhiyun return err;
300*4882a593Smuzhiyun }
301*4882a593Smuzhiyun
302*4882a593Smuzhiyun q->enqueue = cbs_enqueue_offload;
303*4882a593Smuzhiyun q->dequeue = cbs_dequeue_offload;
304*4882a593Smuzhiyun
305*4882a593Smuzhiyun return 0;
306*4882a593Smuzhiyun }
307*4882a593Smuzhiyun
cbs_set_port_rate(struct net_device * dev,struct cbs_sched_data * q)308*4882a593Smuzhiyun static void cbs_set_port_rate(struct net_device *dev, struct cbs_sched_data *q)
309*4882a593Smuzhiyun {
310*4882a593Smuzhiyun struct ethtool_link_ksettings ecmd;
311*4882a593Smuzhiyun int speed = SPEED_10;
312*4882a593Smuzhiyun int port_rate;
313*4882a593Smuzhiyun int err;
314*4882a593Smuzhiyun
315*4882a593Smuzhiyun err = __ethtool_get_link_ksettings(dev, &ecmd);
316*4882a593Smuzhiyun if (err < 0)
317*4882a593Smuzhiyun goto skip;
318*4882a593Smuzhiyun
319*4882a593Smuzhiyun if (ecmd.base.speed && ecmd.base.speed != SPEED_UNKNOWN)
320*4882a593Smuzhiyun speed = ecmd.base.speed;
321*4882a593Smuzhiyun
322*4882a593Smuzhiyun skip:
323*4882a593Smuzhiyun port_rate = speed * 1000 * BYTES_PER_KBIT;
324*4882a593Smuzhiyun
325*4882a593Smuzhiyun atomic64_set(&q->port_rate, port_rate);
326*4882a593Smuzhiyun netdev_dbg(dev, "cbs: set %s's port_rate to: %lld, linkspeed: %d\n",
327*4882a593Smuzhiyun dev->name, (long long)atomic64_read(&q->port_rate),
328*4882a593Smuzhiyun ecmd.base.speed);
329*4882a593Smuzhiyun }
330*4882a593Smuzhiyun
cbs_dev_notifier(struct notifier_block * nb,unsigned long event,void * ptr)331*4882a593Smuzhiyun static int cbs_dev_notifier(struct notifier_block *nb, unsigned long event,
332*4882a593Smuzhiyun void *ptr)
333*4882a593Smuzhiyun {
334*4882a593Smuzhiyun struct net_device *dev = netdev_notifier_info_to_dev(ptr);
335*4882a593Smuzhiyun struct cbs_sched_data *q;
336*4882a593Smuzhiyun struct net_device *qdev;
337*4882a593Smuzhiyun bool found = false;
338*4882a593Smuzhiyun
339*4882a593Smuzhiyun ASSERT_RTNL();
340*4882a593Smuzhiyun
341*4882a593Smuzhiyun if (event != NETDEV_UP && event != NETDEV_CHANGE)
342*4882a593Smuzhiyun return NOTIFY_DONE;
343*4882a593Smuzhiyun
344*4882a593Smuzhiyun spin_lock(&cbs_list_lock);
345*4882a593Smuzhiyun list_for_each_entry(q, &cbs_list, cbs_list) {
346*4882a593Smuzhiyun qdev = qdisc_dev(q->qdisc);
347*4882a593Smuzhiyun if (qdev == dev) {
348*4882a593Smuzhiyun found = true;
349*4882a593Smuzhiyun break;
350*4882a593Smuzhiyun }
351*4882a593Smuzhiyun }
352*4882a593Smuzhiyun spin_unlock(&cbs_list_lock);
353*4882a593Smuzhiyun
354*4882a593Smuzhiyun if (found)
355*4882a593Smuzhiyun cbs_set_port_rate(dev, q);
356*4882a593Smuzhiyun
357*4882a593Smuzhiyun return NOTIFY_DONE;
358*4882a593Smuzhiyun }
359*4882a593Smuzhiyun
cbs_change(struct Qdisc * sch,struct nlattr * opt,struct netlink_ext_ack * extack)360*4882a593Smuzhiyun static int cbs_change(struct Qdisc *sch, struct nlattr *opt,
361*4882a593Smuzhiyun struct netlink_ext_ack *extack)
362*4882a593Smuzhiyun {
363*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
364*4882a593Smuzhiyun struct net_device *dev = qdisc_dev(sch);
365*4882a593Smuzhiyun struct nlattr *tb[TCA_CBS_MAX + 1];
366*4882a593Smuzhiyun struct tc_cbs_qopt *qopt;
367*4882a593Smuzhiyun int err;
368*4882a593Smuzhiyun
369*4882a593Smuzhiyun err = nla_parse_nested_deprecated(tb, TCA_CBS_MAX, opt, cbs_policy,
370*4882a593Smuzhiyun extack);
371*4882a593Smuzhiyun if (err < 0)
372*4882a593Smuzhiyun return err;
373*4882a593Smuzhiyun
374*4882a593Smuzhiyun if (!tb[TCA_CBS_PARMS]) {
375*4882a593Smuzhiyun NL_SET_ERR_MSG(extack, "Missing CBS parameter which are mandatory");
376*4882a593Smuzhiyun return -EINVAL;
377*4882a593Smuzhiyun }
378*4882a593Smuzhiyun
379*4882a593Smuzhiyun qopt = nla_data(tb[TCA_CBS_PARMS]);
380*4882a593Smuzhiyun
381*4882a593Smuzhiyun if (!qopt->offload) {
382*4882a593Smuzhiyun cbs_set_port_rate(dev, q);
383*4882a593Smuzhiyun cbs_disable_offload(dev, q);
384*4882a593Smuzhiyun } else {
385*4882a593Smuzhiyun err = cbs_enable_offload(dev, q, qopt, extack);
386*4882a593Smuzhiyun if (err < 0)
387*4882a593Smuzhiyun return err;
388*4882a593Smuzhiyun }
389*4882a593Smuzhiyun
390*4882a593Smuzhiyun /* Everything went OK, save the parameters used. */
391*4882a593Smuzhiyun q->hicredit = qopt->hicredit;
392*4882a593Smuzhiyun q->locredit = qopt->locredit;
393*4882a593Smuzhiyun q->idleslope = qopt->idleslope * BYTES_PER_KBIT;
394*4882a593Smuzhiyun q->sendslope = qopt->sendslope * BYTES_PER_KBIT;
395*4882a593Smuzhiyun q->offload = qopt->offload;
396*4882a593Smuzhiyun
397*4882a593Smuzhiyun return 0;
398*4882a593Smuzhiyun }
399*4882a593Smuzhiyun
cbs_init(struct Qdisc * sch,struct nlattr * opt,struct netlink_ext_ack * extack)400*4882a593Smuzhiyun static int cbs_init(struct Qdisc *sch, struct nlattr *opt,
401*4882a593Smuzhiyun struct netlink_ext_ack *extack)
402*4882a593Smuzhiyun {
403*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
404*4882a593Smuzhiyun struct net_device *dev = qdisc_dev(sch);
405*4882a593Smuzhiyun
406*4882a593Smuzhiyun if (!opt) {
407*4882a593Smuzhiyun NL_SET_ERR_MSG(extack, "Missing CBS qdisc options which are mandatory");
408*4882a593Smuzhiyun return -EINVAL;
409*4882a593Smuzhiyun }
410*4882a593Smuzhiyun
411*4882a593Smuzhiyun q->qdisc = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
412*4882a593Smuzhiyun sch->handle, extack);
413*4882a593Smuzhiyun if (!q->qdisc)
414*4882a593Smuzhiyun return -ENOMEM;
415*4882a593Smuzhiyun
416*4882a593Smuzhiyun spin_lock(&cbs_list_lock);
417*4882a593Smuzhiyun list_add(&q->cbs_list, &cbs_list);
418*4882a593Smuzhiyun spin_unlock(&cbs_list_lock);
419*4882a593Smuzhiyun
420*4882a593Smuzhiyun qdisc_hash_add(q->qdisc, false);
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);
423*4882a593Smuzhiyun
424*4882a593Smuzhiyun q->enqueue = cbs_enqueue_soft;
425*4882a593Smuzhiyun q->dequeue = cbs_dequeue_soft;
426*4882a593Smuzhiyun
427*4882a593Smuzhiyun qdisc_watchdog_init(&q->watchdog, sch);
428*4882a593Smuzhiyun
429*4882a593Smuzhiyun return cbs_change(sch, opt, extack);
430*4882a593Smuzhiyun }
431*4882a593Smuzhiyun
cbs_destroy(struct Qdisc * sch)432*4882a593Smuzhiyun static void cbs_destroy(struct Qdisc *sch)
433*4882a593Smuzhiyun {
434*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
435*4882a593Smuzhiyun struct net_device *dev = qdisc_dev(sch);
436*4882a593Smuzhiyun
437*4882a593Smuzhiyun /* Nothing to do if we couldn't create the underlying qdisc */
438*4882a593Smuzhiyun if (!q->qdisc)
439*4882a593Smuzhiyun return;
440*4882a593Smuzhiyun
441*4882a593Smuzhiyun qdisc_watchdog_cancel(&q->watchdog);
442*4882a593Smuzhiyun cbs_disable_offload(dev, q);
443*4882a593Smuzhiyun
444*4882a593Smuzhiyun spin_lock(&cbs_list_lock);
445*4882a593Smuzhiyun list_del(&q->cbs_list);
446*4882a593Smuzhiyun spin_unlock(&cbs_list_lock);
447*4882a593Smuzhiyun
448*4882a593Smuzhiyun qdisc_put(q->qdisc);
449*4882a593Smuzhiyun }
450*4882a593Smuzhiyun
cbs_dump(struct Qdisc * sch,struct sk_buff * skb)451*4882a593Smuzhiyun static int cbs_dump(struct Qdisc *sch, struct sk_buff *skb)
452*4882a593Smuzhiyun {
453*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
454*4882a593Smuzhiyun struct tc_cbs_qopt opt = { };
455*4882a593Smuzhiyun struct nlattr *nest;
456*4882a593Smuzhiyun
457*4882a593Smuzhiyun nest = nla_nest_start_noflag(skb, TCA_OPTIONS);
458*4882a593Smuzhiyun if (!nest)
459*4882a593Smuzhiyun goto nla_put_failure;
460*4882a593Smuzhiyun
461*4882a593Smuzhiyun opt.hicredit = q->hicredit;
462*4882a593Smuzhiyun opt.locredit = q->locredit;
463*4882a593Smuzhiyun opt.sendslope = div64_s64(q->sendslope, BYTES_PER_KBIT);
464*4882a593Smuzhiyun opt.idleslope = div64_s64(q->idleslope, BYTES_PER_KBIT);
465*4882a593Smuzhiyun opt.offload = q->offload;
466*4882a593Smuzhiyun
467*4882a593Smuzhiyun if (nla_put(skb, TCA_CBS_PARMS, sizeof(opt), &opt))
468*4882a593Smuzhiyun goto nla_put_failure;
469*4882a593Smuzhiyun
470*4882a593Smuzhiyun return nla_nest_end(skb, nest);
471*4882a593Smuzhiyun
472*4882a593Smuzhiyun nla_put_failure:
473*4882a593Smuzhiyun nla_nest_cancel(skb, nest);
474*4882a593Smuzhiyun return -1;
475*4882a593Smuzhiyun }
476*4882a593Smuzhiyun
cbs_dump_class(struct Qdisc * sch,unsigned long cl,struct sk_buff * skb,struct tcmsg * tcm)477*4882a593Smuzhiyun static int cbs_dump_class(struct Qdisc *sch, unsigned long cl,
478*4882a593Smuzhiyun struct sk_buff *skb, struct tcmsg *tcm)
479*4882a593Smuzhiyun {
480*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
481*4882a593Smuzhiyun
482*4882a593Smuzhiyun if (cl != 1 || !q->qdisc) /* only one class */
483*4882a593Smuzhiyun return -ENOENT;
484*4882a593Smuzhiyun
485*4882a593Smuzhiyun tcm->tcm_handle |= TC_H_MIN(1);
486*4882a593Smuzhiyun tcm->tcm_info = q->qdisc->handle;
487*4882a593Smuzhiyun
488*4882a593Smuzhiyun return 0;
489*4882a593Smuzhiyun }
490*4882a593Smuzhiyun
cbs_graft(struct Qdisc * sch,unsigned long arg,struct Qdisc * new,struct Qdisc ** old,struct netlink_ext_ack * extack)491*4882a593Smuzhiyun static int cbs_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new,
492*4882a593Smuzhiyun struct Qdisc **old, struct netlink_ext_ack *extack)
493*4882a593Smuzhiyun {
494*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
495*4882a593Smuzhiyun
496*4882a593Smuzhiyun if (!new) {
497*4882a593Smuzhiyun new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops,
498*4882a593Smuzhiyun sch->handle, NULL);
499*4882a593Smuzhiyun if (!new)
500*4882a593Smuzhiyun new = &noop_qdisc;
501*4882a593Smuzhiyun }
502*4882a593Smuzhiyun
503*4882a593Smuzhiyun *old = qdisc_replace(sch, new, &q->qdisc);
504*4882a593Smuzhiyun return 0;
505*4882a593Smuzhiyun }
506*4882a593Smuzhiyun
cbs_leaf(struct Qdisc * sch,unsigned long arg)507*4882a593Smuzhiyun static struct Qdisc *cbs_leaf(struct Qdisc *sch, unsigned long arg)
508*4882a593Smuzhiyun {
509*4882a593Smuzhiyun struct cbs_sched_data *q = qdisc_priv(sch);
510*4882a593Smuzhiyun
511*4882a593Smuzhiyun return q->qdisc;
512*4882a593Smuzhiyun }
513*4882a593Smuzhiyun
cbs_find(struct Qdisc * sch,u32 classid)514*4882a593Smuzhiyun static unsigned long cbs_find(struct Qdisc *sch, u32 classid)
515*4882a593Smuzhiyun {
516*4882a593Smuzhiyun return 1;
517*4882a593Smuzhiyun }
518*4882a593Smuzhiyun
cbs_walk(struct Qdisc * sch,struct qdisc_walker * walker)519*4882a593Smuzhiyun static void cbs_walk(struct Qdisc *sch, struct qdisc_walker *walker)
520*4882a593Smuzhiyun {
521*4882a593Smuzhiyun if (!walker->stop) {
522*4882a593Smuzhiyun if (walker->count >= walker->skip) {
523*4882a593Smuzhiyun if (walker->fn(sch, 1, walker) < 0) {
524*4882a593Smuzhiyun walker->stop = 1;
525*4882a593Smuzhiyun return;
526*4882a593Smuzhiyun }
527*4882a593Smuzhiyun }
528*4882a593Smuzhiyun walker->count++;
529*4882a593Smuzhiyun }
530*4882a593Smuzhiyun }
531*4882a593Smuzhiyun
532*4882a593Smuzhiyun static const struct Qdisc_class_ops cbs_class_ops = {
533*4882a593Smuzhiyun .graft = cbs_graft,
534*4882a593Smuzhiyun .leaf = cbs_leaf,
535*4882a593Smuzhiyun .find = cbs_find,
536*4882a593Smuzhiyun .walk = cbs_walk,
537*4882a593Smuzhiyun .dump = cbs_dump_class,
538*4882a593Smuzhiyun };
539*4882a593Smuzhiyun
540*4882a593Smuzhiyun static struct Qdisc_ops cbs_qdisc_ops __read_mostly = {
541*4882a593Smuzhiyun .id = "cbs",
542*4882a593Smuzhiyun .cl_ops = &cbs_class_ops,
543*4882a593Smuzhiyun .priv_size = sizeof(struct cbs_sched_data),
544*4882a593Smuzhiyun .enqueue = cbs_enqueue,
545*4882a593Smuzhiyun .dequeue = cbs_dequeue,
546*4882a593Smuzhiyun .peek = qdisc_peek_dequeued,
547*4882a593Smuzhiyun .init = cbs_init,
548*4882a593Smuzhiyun .reset = qdisc_reset_queue,
549*4882a593Smuzhiyun .destroy = cbs_destroy,
550*4882a593Smuzhiyun .change = cbs_change,
551*4882a593Smuzhiyun .dump = cbs_dump,
552*4882a593Smuzhiyun .owner = THIS_MODULE,
553*4882a593Smuzhiyun };
554*4882a593Smuzhiyun
555*4882a593Smuzhiyun static struct notifier_block cbs_device_notifier = {
556*4882a593Smuzhiyun .notifier_call = cbs_dev_notifier,
557*4882a593Smuzhiyun };
558*4882a593Smuzhiyun
cbs_module_init(void)559*4882a593Smuzhiyun static int __init cbs_module_init(void)
560*4882a593Smuzhiyun {
561*4882a593Smuzhiyun int err;
562*4882a593Smuzhiyun
563*4882a593Smuzhiyun err = register_netdevice_notifier(&cbs_device_notifier);
564*4882a593Smuzhiyun if (err)
565*4882a593Smuzhiyun return err;
566*4882a593Smuzhiyun
567*4882a593Smuzhiyun err = register_qdisc(&cbs_qdisc_ops);
568*4882a593Smuzhiyun if (err)
569*4882a593Smuzhiyun unregister_netdevice_notifier(&cbs_device_notifier);
570*4882a593Smuzhiyun
571*4882a593Smuzhiyun return err;
572*4882a593Smuzhiyun }
573*4882a593Smuzhiyun
cbs_module_exit(void)574*4882a593Smuzhiyun static void __exit cbs_module_exit(void)
575*4882a593Smuzhiyun {
576*4882a593Smuzhiyun unregister_qdisc(&cbs_qdisc_ops);
577*4882a593Smuzhiyun unregister_netdevice_notifier(&cbs_device_notifier);
578*4882a593Smuzhiyun }
579*4882a593Smuzhiyun module_init(cbs_module_init)
580*4882a593Smuzhiyun module_exit(cbs_module_exit)
581*4882a593Smuzhiyun MODULE_LICENSE("GPL");
582