blob: 416305e6d09327d637a71777962d75fc71238a04 [file] [log] [blame]
David Brazdil0f672f62019-12-10 10:32:29 +00001// SPDX-License-Identifier: GPL-2.0-or-later
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002/*
3 * Xenbus code for netif backend
4 *
5 * Copyright (C) 2005 Rusty Russell <rusty@rustcorp.com.au>
6 * Copyright (C) 2005 XenSource Ltd
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00007*/
8
9#include "common.h"
10#include <linux/vmalloc.h>
11#include <linux/rtnetlink.h>
12
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000013static int connect_data_rings(struct backend_info *be,
14 struct xenvif_queue *queue);
15static void connect(struct backend_info *be);
16static int read_xenbus_vif_flags(struct backend_info *be);
17static int backend_create_xenvif(struct backend_info *be);
18static void unregister_hotplug_status_watch(struct backend_info *be);
19static void xen_unregister_watchers(struct xenvif *vif);
20static void set_backend_state(struct backend_info *be,
21 enum xenbus_state state);
22
23#ifdef CONFIG_DEBUG_FS
24struct dentry *xen_netback_dbg_root = NULL;
25
26static int xenvif_read_io_ring(struct seq_file *m, void *v)
27{
28 struct xenvif_queue *queue = m->private;
29 struct xen_netif_tx_back_ring *tx_ring = &queue->tx;
30 struct xen_netif_rx_back_ring *rx_ring = &queue->rx;
31 struct netdev_queue *dev_queue;
32
33 if (tx_ring->sring) {
34 struct xen_netif_tx_sring *sring = tx_ring->sring;
35
36 seq_printf(m, "Queue %d\nTX: nr_ents %u\n", queue->id,
37 tx_ring->nr_ents);
38 seq_printf(m, "req prod %u (%d) cons %u (%d) event %u (%d)\n",
39 sring->req_prod,
40 sring->req_prod - sring->rsp_prod,
41 tx_ring->req_cons,
42 tx_ring->req_cons - sring->rsp_prod,
43 sring->req_event,
44 sring->req_event - sring->rsp_prod);
45 seq_printf(m, "rsp prod %u (base) pvt %u (%d) event %u (%d)\n",
46 sring->rsp_prod,
47 tx_ring->rsp_prod_pvt,
48 tx_ring->rsp_prod_pvt - sring->rsp_prod,
49 sring->rsp_event,
50 sring->rsp_event - sring->rsp_prod);
51 seq_printf(m, "pending prod %u pending cons %u nr_pending_reqs %u\n",
52 queue->pending_prod,
53 queue->pending_cons,
54 nr_pending_reqs(queue));
55 seq_printf(m, "dealloc prod %u dealloc cons %u dealloc_queue %u\n\n",
56 queue->dealloc_prod,
57 queue->dealloc_cons,
58 queue->dealloc_prod - queue->dealloc_cons);
59 }
60
61 if (rx_ring->sring) {
62 struct xen_netif_rx_sring *sring = rx_ring->sring;
63
64 seq_printf(m, "RX: nr_ents %u\n", rx_ring->nr_ents);
65 seq_printf(m, "req prod %u (%d) cons %u (%d) event %u (%d)\n",
66 sring->req_prod,
67 sring->req_prod - sring->rsp_prod,
68 rx_ring->req_cons,
69 rx_ring->req_cons - sring->rsp_prod,
70 sring->req_event,
71 sring->req_event - sring->rsp_prod);
72 seq_printf(m, "rsp prod %u (base) pvt %u (%d) event %u (%d)\n\n",
73 sring->rsp_prod,
74 rx_ring->rsp_prod_pvt,
75 rx_ring->rsp_prod_pvt - sring->rsp_prod,
76 sring->rsp_event,
77 sring->rsp_event - sring->rsp_prod);
78 }
79
80 seq_printf(m, "NAPI state: %lx NAPI weight: %d TX queue len %u\n"
81 "Credit timer_pending: %d, credit: %lu, usec: %lu\n"
82 "remaining: %lu, expires: %lu, now: %lu\n",
83 queue->napi.state, queue->napi.weight,
84 skb_queue_len(&queue->tx_queue),
85 timer_pending(&queue->credit_timeout),
86 queue->credit_bytes,
87 queue->credit_usec,
88 queue->remaining_credit,
89 queue->credit_timeout.expires,
90 jiffies);
91
92 dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id);
93
94 seq_printf(m, "\nRx internal queue: len %u max %u pkts %u %s\n",
95 queue->rx_queue_len, queue->rx_queue_max,
96 skb_queue_len(&queue->rx_queue),
97 netif_tx_queue_stopped(dev_queue) ? "stopped" : "running");
98
99 return 0;
100}
101
102#define XENVIF_KICK_STR "kick"
103#define BUFFER_SIZE 32
104
105static ssize_t
106xenvif_write_io_ring(struct file *filp, const char __user *buf, size_t count,
107 loff_t *ppos)
108{
109 struct xenvif_queue *queue =
110 ((struct seq_file *)filp->private_data)->private;
111 int len;
112 char write[BUFFER_SIZE];
113
114 /* don't allow partial writes and check the length */
115 if (*ppos != 0)
116 return 0;
117 if (count >= sizeof(write))
118 return -ENOSPC;
119
120 len = simple_write_to_buffer(write,
121 sizeof(write) - 1,
122 ppos,
123 buf,
124 count);
125 if (len < 0)
126 return len;
127
128 write[len] = '\0';
129
130 if (!strncmp(write, XENVIF_KICK_STR, sizeof(XENVIF_KICK_STR) - 1))
131 xenvif_interrupt(0, (void *)queue);
132 else {
133 pr_warn("Unknown command to io_ring_q%d. Available: kick\n",
134 queue->id);
135 count = -EINVAL;
136 }
137 return count;
138}
139
140static int xenvif_io_ring_open(struct inode *inode, struct file *filp)
141{
142 int ret;
143 void *queue = NULL;
144
145 if (inode->i_private)
146 queue = inode->i_private;
147 ret = single_open(filp, xenvif_read_io_ring, queue);
148 filp->f_mode |= FMODE_PWRITE;
149 return ret;
150}
151
152static const struct file_operations xenvif_dbg_io_ring_ops_fops = {
153 .owner = THIS_MODULE,
154 .open = xenvif_io_ring_open,
155 .read = seq_read,
156 .llseek = seq_lseek,
157 .release = single_release,
158 .write = xenvif_write_io_ring,
159};
160
David Brazdil0f672f62019-12-10 10:32:29 +0000161static int xenvif_ctrl_show(struct seq_file *m, void *v)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000162{
163 struct xenvif *vif = m->private;
164
165 xenvif_dump_hash_info(vif, m);
166
167 return 0;
168}
David Brazdil0f672f62019-12-10 10:32:29 +0000169DEFINE_SHOW_ATTRIBUTE(xenvif_ctrl);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000170
171static void xenvif_debugfs_addif(struct xenvif *vif)
172{
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000173 int i;
174
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000175 vif->xenvif_dbg_root = debugfs_create_dir(vif->dev->name,
176 xen_netback_dbg_root);
David Brazdil0f672f62019-12-10 10:32:29 +0000177 for (i = 0; i < vif->num_queues; ++i) {
178 char filename[sizeof("io_ring_q") + 4];
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000179
David Brazdil0f672f62019-12-10 10:32:29 +0000180 snprintf(filename, sizeof(filename), "io_ring_q%d", i);
181 debugfs_create_file(filename, 0600, vif->xenvif_dbg_root,
182 &vif->queues[i],
183 &xenvif_dbg_io_ring_ops_fops);
184 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000185
David Brazdil0f672f62019-12-10 10:32:29 +0000186 if (vif->ctrl_irq)
187 debugfs_create_file("ctrl", 0400, vif->xenvif_dbg_root, vif,
188 &xenvif_ctrl_fops);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000189}
190
191static void xenvif_debugfs_delif(struct xenvif *vif)
192{
David Brazdil0f672f62019-12-10 10:32:29 +0000193 debugfs_remove_recursive(vif->xenvif_dbg_root);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000194 vif->xenvif_dbg_root = NULL;
195}
196#endif /* CONFIG_DEBUG_FS */
197
198static int netback_remove(struct xenbus_device *dev)
199{
200 struct backend_info *be = dev_get_drvdata(&dev->dev);
201
202 set_backend_state(be, XenbusStateClosed);
203
204 unregister_hotplug_status_watch(be);
205 if (be->vif) {
206 kobject_uevent(&dev->dev.kobj, KOBJ_OFFLINE);
207 xen_unregister_watchers(be->vif);
208 xenbus_rm(XBT_NIL, dev->nodename, "hotplug-status");
209 xenvif_free(be->vif);
210 be->vif = NULL;
211 }
212 kfree(be->hotplug_script);
213 kfree(be);
214 dev_set_drvdata(&dev->dev, NULL);
215 return 0;
216}
217
218
219/**
220 * Entry point to this code when a new device is created. Allocate the basic
221 * structures and switch to InitWait.
222 */
223static int netback_probe(struct xenbus_device *dev,
224 const struct xenbus_device_id *id)
225{
226 const char *message;
227 struct xenbus_transaction xbt;
228 int err;
229 int sg;
230 const char *script;
231 struct backend_info *be = kzalloc(sizeof(struct backend_info),
232 GFP_KERNEL);
233 if (!be) {
234 xenbus_dev_fatal(dev, -ENOMEM,
235 "allocating backend structure");
236 return -ENOMEM;
237 }
238
239 be->dev = dev;
240 dev_set_drvdata(&dev->dev, be);
241
242 be->state = XenbusStateInitialising;
243 err = xenbus_switch_state(dev, XenbusStateInitialising);
244 if (err)
245 goto fail;
246
247 sg = 1;
248
249 do {
250 err = xenbus_transaction_start(&xbt);
251 if (err) {
252 xenbus_dev_fatal(dev, err, "starting transaction");
253 goto fail;
254 }
255
256 err = xenbus_printf(xbt, dev->nodename, "feature-sg", "%d", sg);
257 if (err) {
258 message = "writing feature-sg";
259 goto abort_transaction;
260 }
261
262 err = xenbus_printf(xbt, dev->nodename, "feature-gso-tcpv4",
263 "%d", sg);
264 if (err) {
265 message = "writing feature-gso-tcpv4";
266 goto abort_transaction;
267 }
268
269 err = xenbus_printf(xbt, dev->nodename, "feature-gso-tcpv6",
270 "%d", sg);
271 if (err) {
272 message = "writing feature-gso-tcpv6";
273 goto abort_transaction;
274 }
275
276 /* We support partial checksum setup for IPv6 packets */
277 err = xenbus_printf(xbt, dev->nodename,
278 "feature-ipv6-csum-offload",
279 "%d", 1);
280 if (err) {
281 message = "writing feature-ipv6-csum-offload";
282 goto abort_transaction;
283 }
284
285 /* We support rx-copy path. */
286 err = xenbus_printf(xbt, dev->nodename,
287 "feature-rx-copy", "%d", 1);
288 if (err) {
289 message = "writing feature-rx-copy";
290 goto abort_transaction;
291 }
292
293 /*
294 * We don't support rx-flip path (except old guests who don't
295 * grok this feature flag).
296 */
297 err = xenbus_printf(xbt, dev->nodename,
298 "feature-rx-flip", "%d", 0);
299 if (err) {
300 message = "writing feature-rx-flip";
301 goto abort_transaction;
302 }
303
304 /* We support dynamic multicast-control. */
305 err = xenbus_printf(xbt, dev->nodename,
306 "feature-multicast-control", "%d", 1);
307 if (err) {
308 message = "writing feature-multicast-control";
309 goto abort_transaction;
310 }
311
312 err = xenbus_printf(xbt, dev->nodename,
313 "feature-dynamic-multicast-control",
314 "%d", 1);
315 if (err) {
316 message = "writing feature-dynamic-multicast-control";
317 goto abort_transaction;
318 }
319
320 err = xenbus_transaction_end(xbt, 0);
321 } while (err == -EAGAIN);
322
323 if (err) {
324 xenbus_dev_fatal(dev, err, "completing transaction");
325 goto fail;
326 }
327
328 /*
329 * Split event channels support, this is optional so it is not
330 * put inside the above loop.
331 */
332 err = xenbus_printf(XBT_NIL, dev->nodename,
333 "feature-split-event-channels",
334 "%u", separate_tx_rx_irq);
335 if (err)
336 pr_debug("Error writing feature-split-event-channels\n");
337
338 /* Multi-queue support: This is an optional feature. */
339 err = xenbus_printf(XBT_NIL, dev->nodename,
340 "multi-queue-max-queues", "%u", xenvif_max_queues);
341 if (err)
342 pr_debug("Error writing multi-queue-max-queues\n");
343
344 err = xenbus_printf(XBT_NIL, dev->nodename,
345 "feature-ctrl-ring",
346 "%u", true);
347 if (err)
348 pr_debug("Error writing feature-ctrl-ring\n");
349
350 script = xenbus_read(XBT_NIL, dev->nodename, "script", NULL);
351 if (IS_ERR(script)) {
352 err = PTR_ERR(script);
353 xenbus_dev_fatal(dev, err, "reading script");
354 goto fail;
355 }
356
357 be->hotplug_script = script;
358
359
360 /* This kicks hotplug scripts, so do it immediately. */
361 err = backend_create_xenvif(be);
362 if (err)
363 goto fail;
364
365 return 0;
366
367abort_transaction:
368 xenbus_transaction_end(xbt, 1);
369 xenbus_dev_fatal(dev, err, "%s", message);
370fail:
371 pr_debug("failed\n");
372 netback_remove(dev);
373 return err;
374}
375
376
377/*
378 * Handle the creation of the hotplug script environment. We add the script
379 * and vif variables to the environment, for the benefit of the vif-* hotplug
380 * scripts.
381 */
382static int netback_uevent(struct xenbus_device *xdev,
383 struct kobj_uevent_env *env)
384{
385 struct backend_info *be = dev_get_drvdata(&xdev->dev);
386
387 if (!be)
388 return 0;
389
390 if (add_uevent_var(env, "script=%s", be->hotplug_script))
391 return -ENOMEM;
392
393 if (!be->vif)
394 return 0;
395
396 return add_uevent_var(env, "vif=%s", be->vif->dev->name);
397}
398
399
400static int backend_create_xenvif(struct backend_info *be)
401{
402 int err;
403 long handle;
404 struct xenbus_device *dev = be->dev;
405 struct xenvif *vif;
406
407 if (be->vif != NULL)
408 return 0;
409
410 err = xenbus_scanf(XBT_NIL, dev->nodename, "handle", "%li", &handle);
411 if (err != 1) {
412 xenbus_dev_fatal(dev, err, "reading handle");
413 return (err < 0) ? err : -EINVAL;
414 }
415
416 vif = xenvif_alloc(&dev->dev, dev->otherend_id, handle);
417 if (IS_ERR(vif)) {
418 err = PTR_ERR(vif);
419 xenbus_dev_fatal(dev, err, "creating interface");
420 return err;
421 }
422 be->vif = vif;
David Brazdil0f672f62019-12-10 10:32:29 +0000423 vif->be = be;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000424
425 kobject_uevent(&dev->dev.kobj, KOBJ_ONLINE);
426 return 0;
427}
428
429static void backend_disconnect(struct backend_info *be)
430{
431 struct xenvif *vif = be->vif;
432
433 if (vif) {
434 unsigned int num_queues = vif->num_queues;
435 unsigned int queue_index;
436
437 xen_unregister_watchers(vif);
438#ifdef CONFIG_DEBUG_FS
439 xenvif_debugfs_delif(vif);
440#endif /* CONFIG_DEBUG_FS */
441 xenvif_disconnect_data(vif);
442
443 /* At this point some of the handlers may still be active
444 * so we need to have additional synchronization here.
445 */
446 vif->num_queues = 0;
447 synchronize_net();
448
449 for (queue_index = 0; queue_index < num_queues; ++queue_index)
450 xenvif_deinit_queue(&vif->queues[queue_index]);
451
452 vfree(vif->queues);
453 vif->queues = NULL;
454
455 xenvif_disconnect_ctrl(vif);
456 }
457}
458
459static void backend_connect(struct backend_info *be)
460{
461 if (be->vif)
462 connect(be);
463}
464
465static inline void backend_switch_state(struct backend_info *be,
466 enum xenbus_state state)
467{
468 struct xenbus_device *dev = be->dev;
469
470 pr_debug("%s -> %s\n", dev->nodename, xenbus_strstate(state));
471 be->state = state;
472
473 /* If we are waiting for a hotplug script then defer the
474 * actual xenbus state change.
475 */
476 if (!be->have_hotplug_status_watch)
477 xenbus_switch_state(dev, state);
478}
479
480/* Handle backend state transitions:
481 *
482 * The backend state starts in Initialising and the following transitions are
483 * allowed.
484 *
485 * Initialising -> InitWait -> Connected
486 * \
487 * \ ^ \ |
488 * \ | \ |
489 * \ | \ |
490 * \ | \ |
491 * \ | \ |
492 * \ | \ |
493 * V | V V
494 *
495 * Closed <-> Closing
496 *
497 * The state argument specifies the eventual state of the backend and the
498 * function transitions to that state via the shortest path.
499 */
500static void set_backend_state(struct backend_info *be,
501 enum xenbus_state state)
502{
503 while (be->state != state) {
504 switch (be->state) {
505 case XenbusStateInitialising:
506 switch (state) {
507 case XenbusStateInitWait:
508 case XenbusStateConnected:
509 case XenbusStateClosing:
510 backend_switch_state(be, XenbusStateInitWait);
511 break;
512 case XenbusStateClosed:
513 backend_switch_state(be, XenbusStateClosed);
514 break;
515 default:
516 BUG();
517 }
518 break;
519 case XenbusStateClosed:
520 switch (state) {
521 case XenbusStateInitWait:
522 case XenbusStateConnected:
523 backend_switch_state(be, XenbusStateInitWait);
524 break;
525 case XenbusStateClosing:
526 backend_switch_state(be, XenbusStateClosing);
527 break;
528 default:
529 BUG();
530 }
531 break;
532 case XenbusStateInitWait:
533 switch (state) {
534 case XenbusStateConnected:
535 backend_connect(be);
536 backend_switch_state(be, XenbusStateConnected);
537 break;
538 case XenbusStateClosing:
539 case XenbusStateClosed:
540 backend_switch_state(be, XenbusStateClosing);
541 break;
542 default:
543 BUG();
544 }
545 break;
546 case XenbusStateConnected:
547 switch (state) {
548 case XenbusStateInitWait:
549 case XenbusStateClosing:
550 case XenbusStateClosed:
551 backend_disconnect(be);
552 backend_switch_state(be, XenbusStateClosing);
553 break;
554 default:
555 BUG();
556 }
557 break;
558 case XenbusStateClosing:
559 switch (state) {
560 case XenbusStateInitWait:
561 case XenbusStateConnected:
562 case XenbusStateClosed:
563 backend_switch_state(be, XenbusStateClosed);
564 break;
565 default:
566 BUG();
567 }
568 break;
569 default:
570 BUG();
571 }
572 }
573}
574
575/**
576 * Callback received when the frontend's state changes.
577 */
578static void frontend_changed(struct xenbus_device *dev,
579 enum xenbus_state frontend_state)
580{
581 struct backend_info *be = dev_get_drvdata(&dev->dev);
582
583 pr_debug("%s -> %s\n", dev->otherend, xenbus_strstate(frontend_state));
584
585 be->frontend_state = frontend_state;
586
587 switch (frontend_state) {
588 case XenbusStateInitialising:
589 set_backend_state(be, XenbusStateInitWait);
590 break;
591
592 case XenbusStateInitialised:
593 break;
594
595 case XenbusStateConnected:
596 set_backend_state(be, XenbusStateConnected);
597 break;
598
599 case XenbusStateClosing:
600 set_backend_state(be, XenbusStateClosing);
601 break;
602
603 case XenbusStateClosed:
604 set_backend_state(be, XenbusStateClosed);
605 if (xenbus_dev_is_online(dev))
606 break;
David Brazdil0f672f62019-12-10 10:32:29 +0000607 /* fall through - if not online */
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000608 case XenbusStateUnknown:
609 set_backend_state(be, XenbusStateClosed);
610 device_unregister(&dev->dev);
611 break;
612
613 default:
614 xenbus_dev_fatal(dev, -EINVAL, "saw state %d at frontend",
615 frontend_state);
616 break;
617 }
618}
619
620
621static void xen_net_read_rate(struct xenbus_device *dev,
622 unsigned long *bytes, unsigned long *usec)
623{
624 char *s, *e;
625 unsigned long b, u;
626 char *ratestr;
627
628 /* Default to unlimited bandwidth. */
629 *bytes = ~0UL;
630 *usec = 0;
631
632 ratestr = xenbus_read(XBT_NIL, dev->nodename, "rate", NULL);
633 if (IS_ERR(ratestr))
634 return;
635
636 s = ratestr;
637 b = simple_strtoul(s, &e, 10);
638 if ((s == e) || (*e != ','))
639 goto fail;
640
641 s = e + 1;
642 u = simple_strtoul(s, &e, 10);
643 if ((s == e) || (*e != '\0'))
644 goto fail;
645
646 *bytes = b;
647 *usec = u;
648
649 kfree(ratestr);
650 return;
651
652 fail:
653 pr_warn("Failed to parse network rate limit. Traffic unlimited.\n");
654 kfree(ratestr);
655}
656
657static int xen_net_read_mac(struct xenbus_device *dev, u8 mac[])
658{
659 char *s, *e, *macstr;
660 int i;
661
662 macstr = s = xenbus_read(XBT_NIL, dev->nodename, "mac", NULL);
663 if (IS_ERR(macstr))
664 return PTR_ERR(macstr);
665
666 for (i = 0; i < ETH_ALEN; i++) {
667 mac[i] = simple_strtoul(s, &e, 16);
668 if ((s == e) || (*e != ((i == ETH_ALEN-1) ? '\0' : ':'))) {
669 kfree(macstr);
670 return -ENOENT;
671 }
672 s = e+1;
673 }
674
675 kfree(macstr);
676 return 0;
677}
678
679static void xen_net_rate_changed(struct xenbus_watch *watch,
680 const char *path, const char *token)
681{
682 struct xenvif *vif = container_of(watch, struct xenvif, credit_watch);
683 struct xenbus_device *dev = xenvif_to_xenbus_device(vif);
684 unsigned long credit_bytes;
685 unsigned long credit_usec;
686 unsigned int queue_index;
687
688 xen_net_read_rate(dev, &credit_bytes, &credit_usec);
689 for (queue_index = 0; queue_index < vif->num_queues; queue_index++) {
690 struct xenvif_queue *queue = &vif->queues[queue_index];
691
692 queue->credit_bytes = credit_bytes;
693 queue->credit_usec = credit_usec;
694 if (!mod_timer_pending(&queue->credit_timeout, jiffies) &&
695 queue->remaining_credit > queue->credit_bytes) {
696 queue->remaining_credit = queue->credit_bytes;
697 }
698 }
699}
700
701static int xen_register_credit_watch(struct xenbus_device *dev,
702 struct xenvif *vif)
703{
704 int err = 0;
705 char *node;
706 unsigned maxlen = strlen(dev->nodename) + sizeof("/rate");
707
708 if (vif->credit_watch.node)
709 return -EADDRINUSE;
710
711 node = kmalloc(maxlen, GFP_KERNEL);
712 if (!node)
713 return -ENOMEM;
714 snprintf(node, maxlen, "%s/rate", dev->nodename);
715 vif->credit_watch.node = node;
Olivier Deprez0e641232021-09-23 10:07:05 +0200716 vif->credit_watch.will_handle = NULL;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000717 vif->credit_watch.callback = xen_net_rate_changed;
718 err = register_xenbus_watch(&vif->credit_watch);
719 if (err) {
720 pr_err("Failed to set watcher %s\n", vif->credit_watch.node);
721 kfree(node);
722 vif->credit_watch.node = NULL;
Olivier Deprez0e641232021-09-23 10:07:05 +0200723 vif->credit_watch.will_handle = NULL;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000724 vif->credit_watch.callback = NULL;
725 }
726 return err;
727}
728
729static void xen_unregister_credit_watch(struct xenvif *vif)
730{
731 if (vif->credit_watch.node) {
732 unregister_xenbus_watch(&vif->credit_watch);
733 kfree(vif->credit_watch.node);
734 vif->credit_watch.node = NULL;
735 }
736}
737
738static void xen_mcast_ctrl_changed(struct xenbus_watch *watch,
739 const char *path, const char *token)
740{
741 struct xenvif *vif = container_of(watch, struct xenvif,
742 mcast_ctrl_watch);
743 struct xenbus_device *dev = xenvif_to_xenbus_device(vif);
744
745 vif->multicast_control = !!xenbus_read_unsigned(dev->otherend,
746 "request-multicast-control", 0);
747}
748
749static int xen_register_mcast_ctrl_watch(struct xenbus_device *dev,
750 struct xenvif *vif)
751{
752 int err = 0;
753 char *node;
754 unsigned maxlen = strlen(dev->otherend) +
755 sizeof("/request-multicast-control");
756
757 if (vif->mcast_ctrl_watch.node) {
758 pr_err_ratelimited("Watch is already registered\n");
759 return -EADDRINUSE;
760 }
761
762 node = kmalloc(maxlen, GFP_KERNEL);
763 if (!node) {
764 pr_err("Failed to allocate memory for watch\n");
765 return -ENOMEM;
766 }
767 snprintf(node, maxlen, "%s/request-multicast-control",
768 dev->otherend);
769 vif->mcast_ctrl_watch.node = node;
Olivier Deprez0e641232021-09-23 10:07:05 +0200770 vif->mcast_ctrl_watch.will_handle = NULL;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000771 vif->mcast_ctrl_watch.callback = xen_mcast_ctrl_changed;
772 err = register_xenbus_watch(&vif->mcast_ctrl_watch);
773 if (err) {
774 pr_err("Failed to set watcher %s\n",
775 vif->mcast_ctrl_watch.node);
776 kfree(node);
777 vif->mcast_ctrl_watch.node = NULL;
Olivier Deprez0e641232021-09-23 10:07:05 +0200778 vif->mcast_ctrl_watch.will_handle = NULL;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000779 vif->mcast_ctrl_watch.callback = NULL;
780 }
781 return err;
782}
783
784static void xen_unregister_mcast_ctrl_watch(struct xenvif *vif)
785{
786 if (vif->mcast_ctrl_watch.node) {
787 unregister_xenbus_watch(&vif->mcast_ctrl_watch);
788 kfree(vif->mcast_ctrl_watch.node);
789 vif->mcast_ctrl_watch.node = NULL;
790 }
791}
792
793static void xen_register_watchers(struct xenbus_device *dev,
794 struct xenvif *vif)
795{
796 xen_register_credit_watch(dev, vif);
797 xen_register_mcast_ctrl_watch(dev, vif);
798}
799
800static void xen_unregister_watchers(struct xenvif *vif)
801{
802 xen_unregister_mcast_ctrl_watch(vif);
803 xen_unregister_credit_watch(vif);
804}
805
806static void unregister_hotplug_status_watch(struct backend_info *be)
807{
808 if (be->have_hotplug_status_watch) {
809 unregister_xenbus_watch(&be->hotplug_status_watch);
810 kfree(be->hotplug_status_watch.node);
811 }
812 be->have_hotplug_status_watch = 0;
813}
814
815static void hotplug_status_changed(struct xenbus_watch *watch,
816 const char *path,
817 const char *token)
818{
819 struct backend_info *be = container_of(watch,
820 struct backend_info,
821 hotplug_status_watch);
822 char *str;
823 unsigned int len;
824
825 str = xenbus_read(XBT_NIL, be->dev->nodename, "hotplug-status", &len);
826 if (IS_ERR(str))
827 return;
828 if (len == sizeof("connected")-1 && !memcmp(str, "connected", len)) {
829 /* Complete any pending state change */
830 xenbus_switch_state(be->dev, be->state);
831
832 /* Not interested in this watch anymore. */
833 unregister_hotplug_status_watch(be);
834 }
835 kfree(str);
836}
837
838static int connect_ctrl_ring(struct backend_info *be)
839{
840 struct xenbus_device *dev = be->dev;
841 struct xenvif *vif = be->vif;
842 unsigned int val;
843 grant_ref_t ring_ref;
844 unsigned int evtchn;
845 int err;
846
847 err = xenbus_scanf(XBT_NIL, dev->otherend,
848 "ctrl-ring-ref", "%u", &val);
849 if (err < 0)
850 goto done; /* The frontend does not have a control ring */
851
852 ring_ref = val;
853
854 err = xenbus_scanf(XBT_NIL, dev->otherend,
855 "event-channel-ctrl", "%u", &val);
856 if (err < 0) {
857 xenbus_dev_fatal(dev, err,
858 "reading %s/event-channel-ctrl",
859 dev->otherend);
860 goto fail;
861 }
862
863 evtchn = val;
864
865 err = xenvif_connect_ctrl(vif, ring_ref, evtchn);
866 if (err) {
867 xenbus_dev_fatal(dev, err,
868 "mapping shared-frame %u port %u",
869 ring_ref, evtchn);
870 goto fail;
871 }
872
873done:
874 return 0;
875
876fail:
877 return err;
878}
879
880static void connect(struct backend_info *be)
881{
882 int err;
883 struct xenbus_device *dev = be->dev;
884 unsigned long credit_bytes, credit_usec;
885 unsigned int queue_index;
886 unsigned int requested_num_queues;
887 struct xenvif_queue *queue;
888
889 /* Check whether the frontend requested multiple queues
890 * and read the number requested.
891 */
892 requested_num_queues = xenbus_read_unsigned(dev->otherend,
893 "multi-queue-num-queues", 1);
894 if (requested_num_queues > xenvif_max_queues) {
895 /* buggy or malicious guest */
896 xenbus_dev_fatal(dev, -EINVAL,
897 "guest requested %u queues, exceeding the maximum of %u.",
898 requested_num_queues, xenvif_max_queues);
899 return;
900 }
901
902 err = xen_net_read_mac(dev, be->vif->fe_dev_addr);
903 if (err) {
904 xenbus_dev_fatal(dev, err, "parsing %s/mac", dev->nodename);
905 return;
906 }
907
908 xen_net_read_rate(dev, &credit_bytes, &credit_usec);
909 xen_unregister_watchers(be->vif);
910 xen_register_watchers(dev, be->vif);
911 read_xenbus_vif_flags(be);
912
913 err = connect_ctrl_ring(be);
914 if (err) {
915 xenbus_dev_fatal(dev, err, "connecting control ring");
916 return;
917 }
918
919 /* Use the number of queues requested by the frontend */
920 be->vif->queues = vzalloc(array_size(requested_num_queues,
921 sizeof(struct xenvif_queue)));
922 if (!be->vif->queues) {
923 xenbus_dev_fatal(dev, -ENOMEM,
924 "allocating queues");
925 return;
926 }
927
928 be->vif->num_queues = requested_num_queues;
929 be->vif->stalled_queues = requested_num_queues;
930
931 for (queue_index = 0; queue_index < requested_num_queues; ++queue_index) {
932 queue = &be->vif->queues[queue_index];
933 queue->vif = be->vif;
934 queue->id = queue_index;
935 snprintf(queue->name, sizeof(queue->name), "%s-q%u",
936 be->vif->dev->name, queue->id);
937
938 err = xenvif_init_queue(queue);
939 if (err) {
940 /* xenvif_init_queue() cleans up after itself on
941 * failure, but we need to clean up any previously
942 * initialised queues. Set num_queues to i so that
943 * earlier queues can be destroyed using the regular
944 * disconnect logic.
945 */
946 be->vif->num_queues = queue_index;
947 goto err;
948 }
949
950 queue->credit_bytes = credit_bytes;
951 queue->remaining_credit = credit_bytes;
952 queue->credit_usec = credit_usec;
953
954 err = connect_data_rings(be, queue);
955 if (err) {
956 /* connect_data_rings() cleans up after itself on
957 * failure, but we need to clean up after
958 * xenvif_init_queue() here, and also clean up any
959 * previously initialised queues.
960 */
961 xenvif_deinit_queue(queue);
962 be->vif->num_queues = queue_index;
963 goto err;
964 }
965 }
966
967#ifdef CONFIG_DEBUG_FS
968 xenvif_debugfs_addif(be->vif);
969#endif /* CONFIG_DEBUG_FS */
970
971 /* Initialisation completed, tell core driver the number of
972 * active queues.
973 */
974 rtnl_lock();
975 netif_set_real_num_tx_queues(be->vif->dev, requested_num_queues);
976 netif_set_real_num_rx_queues(be->vif->dev, requested_num_queues);
977 rtnl_unlock();
978
979 xenvif_carrier_on(be->vif);
980
981 unregister_hotplug_status_watch(be);
Olivier Deprez0e641232021-09-23 10:07:05 +0200982 if (xenbus_exists(XBT_NIL, dev->nodename, "hotplug-status")) {
983 err = xenbus_watch_pathfmt(dev, &be->hotplug_status_watch,
984 NULL, hotplug_status_changed,
985 "%s/%s", dev->nodename,
986 "hotplug-status");
987 if (err)
988 goto err;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000989 be->have_hotplug_status_watch = 1;
Olivier Deprez0e641232021-09-23 10:07:05 +0200990 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000991
992 netif_tx_wake_all_queues(be->vif->dev);
993
994 return;
995
996err:
997 if (be->vif->num_queues > 0)
998 xenvif_disconnect_data(be->vif); /* Clean up existing queues */
999 for (queue_index = 0; queue_index < be->vif->num_queues; ++queue_index)
1000 xenvif_deinit_queue(&be->vif->queues[queue_index]);
1001 vfree(be->vif->queues);
1002 be->vif->queues = NULL;
1003 be->vif->num_queues = 0;
1004 xenvif_disconnect_ctrl(be->vif);
1005 return;
1006}
1007
1008
1009static int connect_data_rings(struct backend_info *be,
1010 struct xenvif_queue *queue)
1011{
1012 struct xenbus_device *dev = be->dev;
1013 unsigned int num_queues = queue->vif->num_queues;
1014 unsigned long tx_ring_ref, rx_ring_ref;
1015 unsigned int tx_evtchn, rx_evtchn;
1016 int err;
1017 char *xspath;
1018 size_t xspathsize;
1019 const size_t xenstore_path_ext_size = 11; /* sufficient for "/queue-NNN" */
1020
1021 /* If the frontend requested 1 queue, or we have fallen back
1022 * to single queue due to lack of frontend support for multi-
1023 * queue, expect the remaining XenStore keys in the toplevel
1024 * directory. Otherwise, expect them in a subdirectory called
1025 * queue-N.
1026 */
1027 if (num_queues == 1) {
1028 xspath = kzalloc(strlen(dev->otherend) + 1, GFP_KERNEL);
1029 if (!xspath) {
1030 xenbus_dev_fatal(dev, -ENOMEM,
1031 "reading ring references");
1032 return -ENOMEM;
1033 }
1034 strcpy(xspath, dev->otherend);
1035 } else {
1036 xspathsize = strlen(dev->otherend) + xenstore_path_ext_size;
1037 xspath = kzalloc(xspathsize, GFP_KERNEL);
1038 if (!xspath) {
1039 xenbus_dev_fatal(dev, -ENOMEM,
1040 "reading ring references");
1041 return -ENOMEM;
1042 }
1043 snprintf(xspath, xspathsize, "%s/queue-%u", dev->otherend,
1044 queue->id);
1045 }
1046
1047 err = xenbus_gather(XBT_NIL, xspath,
1048 "tx-ring-ref", "%lu", &tx_ring_ref,
1049 "rx-ring-ref", "%lu", &rx_ring_ref, NULL);
1050 if (err) {
1051 xenbus_dev_fatal(dev, err,
1052 "reading %s/ring-ref",
1053 xspath);
1054 goto err;
1055 }
1056
1057 /* Try split event channels first, then single event channel. */
1058 err = xenbus_gather(XBT_NIL, xspath,
1059 "event-channel-tx", "%u", &tx_evtchn,
1060 "event-channel-rx", "%u", &rx_evtchn, NULL);
1061 if (err < 0) {
1062 err = xenbus_scanf(XBT_NIL, xspath,
1063 "event-channel", "%u", &tx_evtchn);
1064 if (err < 0) {
1065 xenbus_dev_fatal(dev, err,
1066 "reading %s/event-channel(-tx/rx)",
1067 xspath);
1068 goto err;
1069 }
1070 rx_evtchn = tx_evtchn;
1071 }
1072
1073 /* Map the shared frame, irq etc. */
1074 err = xenvif_connect_data(queue, tx_ring_ref, rx_ring_ref,
1075 tx_evtchn, rx_evtchn);
1076 if (err) {
1077 xenbus_dev_fatal(dev, err,
1078 "mapping shared-frames %lu/%lu port tx %u rx %u",
1079 tx_ring_ref, rx_ring_ref,
1080 tx_evtchn, rx_evtchn);
1081 goto err;
1082 }
1083
1084 err = 0;
1085err: /* Regular return falls through with err == 0 */
1086 kfree(xspath);
1087 return err;
1088}
1089
1090static int read_xenbus_vif_flags(struct backend_info *be)
1091{
1092 struct xenvif *vif = be->vif;
1093 struct xenbus_device *dev = be->dev;
1094 unsigned int rx_copy;
1095 int err;
1096
1097 err = xenbus_scanf(XBT_NIL, dev->otherend, "request-rx-copy", "%u",
1098 &rx_copy);
1099 if (err == -ENOENT) {
1100 err = 0;
1101 rx_copy = 0;
1102 }
1103 if (err < 0) {
1104 xenbus_dev_fatal(dev, err, "reading %s/request-rx-copy",
1105 dev->otherend);
1106 return err;
1107 }
1108 if (!rx_copy)
1109 return -EOPNOTSUPP;
1110
1111 if (!xenbus_read_unsigned(dev->otherend, "feature-rx-notify", 0)) {
1112 /* - Reduce drain timeout to poll more frequently for
1113 * Rx requests.
1114 * - Disable Rx stall detection.
1115 */
1116 be->vif->drain_timeout = msecs_to_jiffies(30);
1117 be->vif->stall_timeout = 0;
1118 }
1119
1120 vif->can_sg = !!xenbus_read_unsigned(dev->otherend, "feature-sg", 0);
1121
1122 vif->gso_mask = 0;
1123
1124 if (xenbus_read_unsigned(dev->otherend, "feature-gso-tcpv4", 0))
1125 vif->gso_mask |= GSO_BIT(TCPV4);
1126
1127 if (xenbus_read_unsigned(dev->otherend, "feature-gso-tcpv6", 0))
1128 vif->gso_mask |= GSO_BIT(TCPV6);
1129
1130 vif->ip_csum = !xenbus_read_unsigned(dev->otherend,
1131 "feature-no-csum-offload", 0);
1132
1133 vif->ipv6_csum = !!xenbus_read_unsigned(dev->otherend,
1134 "feature-ipv6-csum-offload", 0);
1135
1136 return 0;
1137}
1138
1139static const struct xenbus_device_id netback_ids[] = {
1140 { "vif" },
1141 { "" }
1142};
1143
1144static struct xenbus_driver netback_driver = {
1145 .ids = netback_ids,
1146 .probe = netback_probe,
1147 .remove = netback_remove,
1148 .uevent = netback_uevent,
1149 .otherend_changed = frontend_changed,
1150};
1151
1152int xenvif_xenbus_init(void)
1153{
1154 return xenbus_register_backend(&netback_driver);
1155}
1156
1157void xenvif_xenbus_fini(void)
1158{
1159 return xenbus_unregister_driver(&netback_driver);
1160}