xref: /linux/drivers/xen/xenbus/xenbus_probe.c (revision 2da68a77)
1 /******************************************************************************
2  * Talks to Xen Store to figure out what devices we have.
3  *
4  * Copyright (C) 2005 Rusty Russell, IBM Corporation
5  * Copyright (C) 2005 Mike Wray, Hewlett-Packard
6  * Copyright (C) 2005, 2006 XenSource Ltd
7  *
8  * This program is free software; you can redistribute it and/or
9  * modify it under the terms of the GNU General Public License version 2
10  * as published by the Free Software Foundation; or, when distributed
11  * separately from the Linux kernel or incorporated into other
12  * software packages, subject to the following license:
13  *
14  * Permission is hereby granted, free of charge, to any person obtaining a copy
15  * of this source file (the "Software"), to deal in the Software without
16  * restriction, including without limitation the rights to use, copy, modify,
17  * merge, publish, distribute, sublicense, and/or sell copies of the Software,
18  * and to permit persons to whom the Software is furnished to do so, subject to
19  * the following conditions:
20  *
21  * The above copyright notice and this permission notice shall be included in
22  * all copies or substantial portions of the Software.
23  *
24  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
25  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
26  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
27  * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
28  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
29  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
30  * IN THE SOFTWARE.
31  */
32 
33 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
34 #define dev_fmt pr_fmt
35 
36 #define DPRINTK(fmt, args...)				\
37 	pr_debug("xenbus_probe (%s:%d) " fmt ".\n",	\
38 		 __func__, __LINE__, ##args)
39 
40 #include <linux/kernel.h>
41 #include <linux/err.h>
42 #include <linux/string.h>
43 #include <linux/ctype.h>
44 #include <linux/fcntl.h>
45 #include <linux/mm.h>
46 #include <linux/proc_fs.h>
47 #include <linux/notifier.h>
48 #include <linux/kthread.h>
49 #include <linux/mutex.h>
50 #include <linux/io.h>
51 #include <linux/slab.h>
52 #include <linux/module.h>
53 
54 #include <asm/page.h>
55 #include <asm/xen/hypervisor.h>
56 
57 #include <xen/xen.h>
58 #include <xen/xenbus.h>
59 #include <xen/events.h>
60 #include <xen/xen-ops.h>
61 #include <xen/page.h>
62 
63 #include <xen/hvm.h>
64 
65 #include "xenbus.h"
66 
67 
68 static int xs_init_irq;
69 int xen_store_evtchn;
70 EXPORT_SYMBOL_GPL(xen_store_evtchn);
71 
72 struct xenstore_domain_interface *xen_store_interface;
73 EXPORT_SYMBOL_GPL(xen_store_interface);
74 
75 enum xenstore_init xen_store_domain_type;
76 EXPORT_SYMBOL_GPL(xen_store_domain_type);
77 
78 static unsigned long xen_store_gfn;
79 
80 static BLOCKING_NOTIFIER_HEAD(xenstore_chain);
81 
82 /* If something in array of ids matches this device, return it. */
83 static const struct xenbus_device_id *
84 match_device(const struct xenbus_device_id *arr, struct xenbus_device *dev)
85 {
86 	for (; *arr->devicetype != '\0'; arr++) {
87 		if (!strcmp(arr->devicetype, dev->devicetype))
88 			return arr;
89 	}
90 	return NULL;
91 }
92 
93 int xenbus_match(struct device *_dev, struct device_driver *_drv)
94 {
95 	struct xenbus_driver *drv = to_xenbus_driver(_drv);
96 
97 	if (!drv->ids)
98 		return 0;
99 
100 	return match_device(drv->ids, to_xenbus_device(_dev)) != NULL;
101 }
102 EXPORT_SYMBOL_GPL(xenbus_match);
103 
104 
105 static void free_otherend_details(struct xenbus_device *dev)
106 {
107 	kfree(dev->otherend);
108 	dev->otherend = NULL;
109 }
110 
111 
112 static void free_otherend_watch(struct xenbus_device *dev)
113 {
114 	if (dev->otherend_watch.node) {
115 		unregister_xenbus_watch(&dev->otherend_watch);
116 		kfree(dev->otherend_watch.node);
117 		dev->otherend_watch.node = NULL;
118 	}
119 }
120 
121 
122 static int talk_to_otherend(struct xenbus_device *dev)
123 {
124 	struct xenbus_driver *drv = to_xenbus_driver(dev->dev.driver);
125 
126 	free_otherend_watch(dev);
127 	free_otherend_details(dev);
128 
129 	return drv->read_otherend_details(dev);
130 }
131 
132 
133 
134 static int watch_otherend(struct xenbus_device *dev)
135 {
136 	struct xen_bus_type *bus =
137 		container_of(dev->dev.bus, struct xen_bus_type, bus);
138 
139 	return xenbus_watch_pathfmt(dev, &dev->otherend_watch,
140 				    bus->otherend_will_handle,
141 				    bus->otherend_changed,
142 				    "%s/%s", dev->otherend, "state");
143 }
144 
145 
146 int xenbus_read_otherend_details(struct xenbus_device *xendev,
147 				 char *id_node, char *path_node)
148 {
149 	int err = xenbus_gather(XBT_NIL, xendev->nodename,
150 				id_node, "%i", &xendev->otherend_id,
151 				path_node, NULL, &xendev->otherend,
152 				NULL);
153 	if (err) {
154 		xenbus_dev_fatal(xendev, err,
155 				 "reading other end details from %s",
156 				 xendev->nodename);
157 		return err;
158 	}
159 	if (strlen(xendev->otherend) == 0 ||
160 	    !xenbus_exists(XBT_NIL, xendev->otherend, "")) {
161 		xenbus_dev_fatal(xendev, -ENOENT,
162 				 "unable to read other end from %s.  "
163 				 "missing or inaccessible.",
164 				 xendev->nodename);
165 		free_otherend_details(xendev);
166 		return -ENOENT;
167 	}
168 
169 	return 0;
170 }
171 EXPORT_SYMBOL_GPL(xenbus_read_otherend_details);
172 
173 void xenbus_otherend_changed(struct xenbus_watch *watch,
174 			     const char *path, const char *token,
175 			     int ignore_on_shutdown)
176 {
177 	struct xenbus_device *dev =
178 		container_of(watch, struct xenbus_device, otherend_watch);
179 	struct xenbus_driver *drv = to_xenbus_driver(dev->dev.driver);
180 	enum xenbus_state state;
181 
182 	/* Protect us against watches firing on old details when the otherend
183 	   details change, say immediately after a resume. */
184 	if (!dev->otherend ||
185 	    strncmp(dev->otherend, path, strlen(dev->otherend))) {
186 		dev_dbg(&dev->dev, "Ignoring watch at %s\n", path);
187 		return;
188 	}
189 
190 	state = xenbus_read_driver_state(dev->otherend);
191 
192 	dev_dbg(&dev->dev, "state is %d, (%s), %s, %s\n",
193 		state, xenbus_strstate(state), dev->otherend_watch.node, path);
194 
195 	/*
196 	 * Ignore xenbus transitions during shutdown. This prevents us doing
197 	 * work that can fail e.g., when the rootfs is gone.
198 	 */
199 	if (system_state > SYSTEM_RUNNING) {
200 		if (ignore_on_shutdown && (state == XenbusStateClosing))
201 			xenbus_frontend_closed(dev);
202 		return;
203 	}
204 
205 	if (drv->otherend_changed)
206 		drv->otherend_changed(dev, state);
207 }
208 EXPORT_SYMBOL_GPL(xenbus_otherend_changed);
209 
210 #define XENBUS_SHOW_STAT(name)						\
211 static ssize_t name##_show(struct device *_dev,				\
212 			   struct device_attribute *attr,		\
213 			   char *buf)					\
214 {									\
215 	struct xenbus_device *dev = to_xenbus_device(_dev);		\
216 									\
217 	return sprintf(buf, "%d\n", atomic_read(&dev->name));		\
218 }									\
219 static DEVICE_ATTR_RO(name)
220 
221 XENBUS_SHOW_STAT(event_channels);
222 XENBUS_SHOW_STAT(events);
223 XENBUS_SHOW_STAT(spurious_events);
224 XENBUS_SHOW_STAT(jiffies_eoi_delayed);
225 
226 static ssize_t spurious_threshold_show(struct device *_dev,
227 				       struct device_attribute *attr,
228 				       char *buf)
229 {
230 	struct xenbus_device *dev = to_xenbus_device(_dev);
231 
232 	return sprintf(buf, "%d\n", dev->spurious_threshold);
233 }
234 
235 static ssize_t spurious_threshold_store(struct device *_dev,
236 					struct device_attribute *attr,
237 					const char *buf, size_t count)
238 {
239 	struct xenbus_device *dev = to_xenbus_device(_dev);
240 	unsigned int val;
241 	ssize_t ret;
242 
243 	ret = kstrtouint(buf, 0, &val);
244 	if (ret)
245 		return ret;
246 
247 	dev->spurious_threshold = val;
248 
249 	return count;
250 }
251 
252 static DEVICE_ATTR_RW(spurious_threshold);
253 
254 static struct attribute *xenbus_attrs[] = {
255 	&dev_attr_event_channels.attr,
256 	&dev_attr_events.attr,
257 	&dev_attr_spurious_events.attr,
258 	&dev_attr_jiffies_eoi_delayed.attr,
259 	&dev_attr_spurious_threshold.attr,
260 	NULL
261 };
262 
263 static const struct attribute_group xenbus_group = {
264 	.name = "xenbus",
265 	.attrs = xenbus_attrs,
266 };
267 
268 int xenbus_dev_probe(struct device *_dev)
269 {
270 	struct xenbus_device *dev = to_xenbus_device(_dev);
271 	struct xenbus_driver *drv = to_xenbus_driver(_dev->driver);
272 	const struct xenbus_device_id *id;
273 	int err;
274 
275 	DPRINTK("%s", dev->nodename);
276 
277 	if (!drv->probe) {
278 		err = -ENODEV;
279 		goto fail;
280 	}
281 
282 	id = match_device(drv->ids, dev);
283 	if (!id) {
284 		err = -ENODEV;
285 		goto fail;
286 	}
287 
288 	err = talk_to_otherend(dev);
289 	if (err) {
290 		dev_warn(&dev->dev, "talk_to_otherend on %s failed.\n",
291 			 dev->nodename);
292 		return err;
293 	}
294 
295 	if (!try_module_get(drv->driver.owner)) {
296 		dev_warn(&dev->dev, "failed to acquire module reference on '%s'\n",
297 			 drv->driver.name);
298 		err = -ESRCH;
299 		goto fail;
300 	}
301 
302 	down(&dev->reclaim_sem);
303 	err = drv->probe(dev, id);
304 	up(&dev->reclaim_sem);
305 	if (err)
306 		goto fail_put;
307 
308 	err = watch_otherend(dev);
309 	if (err) {
310 		dev_warn(&dev->dev, "watch_otherend on %s failed.\n",
311 		       dev->nodename);
312 		return err;
313 	}
314 
315 	dev->spurious_threshold = 1;
316 	if (sysfs_create_group(&dev->dev.kobj, &xenbus_group))
317 		dev_warn(&dev->dev, "sysfs_create_group on %s failed.\n",
318 			 dev->nodename);
319 
320 	return 0;
321 fail_put:
322 	module_put(drv->driver.owner);
323 fail:
324 	xenbus_dev_error(dev, err, "xenbus_dev_probe on %s", dev->nodename);
325 	return err;
326 }
327 EXPORT_SYMBOL_GPL(xenbus_dev_probe);
328 
329 void xenbus_dev_remove(struct device *_dev)
330 {
331 	struct xenbus_device *dev = to_xenbus_device(_dev);
332 	struct xenbus_driver *drv = to_xenbus_driver(_dev->driver);
333 
334 	DPRINTK("%s", dev->nodename);
335 
336 	sysfs_remove_group(&dev->dev.kobj, &xenbus_group);
337 
338 	free_otherend_watch(dev);
339 
340 	if (drv->remove) {
341 		down(&dev->reclaim_sem);
342 		drv->remove(dev);
343 		up(&dev->reclaim_sem);
344 	}
345 
346 	module_put(drv->driver.owner);
347 
348 	free_otherend_details(dev);
349 
350 	/*
351 	 * If the toolstack has forced the device state to closing then set
352 	 * the state to closed now to allow it to be cleaned up.
353 	 * Similarly, if the driver does not support re-bind, set the
354 	 * closed.
355 	 */
356 	if (!drv->allow_rebind ||
357 	    xenbus_read_driver_state(dev->nodename) == XenbusStateClosing)
358 		xenbus_switch_state(dev, XenbusStateClosed);
359 }
360 EXPORT_SYMBOL_GPL(xenbus_dev_remove);
361 
362 int xenbus_register_driver_common(struct xenbus_driver *drv,
363 				  struct xen_bus_type *bus,
364 				  struct module *owner, const char *mod_name)
365 {
366 	drv->driver.name = drv->name ? drv->name : drv->ids[0].devicetype;
367 	drv->driver.bus = &bus->bus;
368 	drv->driver.owner = owner;
369 	drv->driver.mod_name = mod_name;
370 
371 	return driver_register(&drv->driver);
372 }
373 EXPORT_SYMBOL_GPL(xenbus_register_driver_common);
374 
375 void xenbus_unregister_driver(struct xenbus_driver *drv)
376 {
377 	driver_unregister(&drv->driver);
378 }
379 EXPORT_SYMBOL_GPL(xenbus_unregister_driver);
380 
381 struct xb_find_info {
382 	struct xenbus_device *dev;
383 	const char *nodename;
384 };
385 
386 static int cmp_dev(struct device *dev, void *data)
387 {
388 	struct xenbus_device *xendev = to_xenbus_device(dev);
389 	struct xb_find_info *info = data;
390 
391 	if (!strcmp(xendev->nodename, info->nodename)) {
392 		info->dev = xendev;
393 		get_device(dev);
394 		return 1;
395 	}
396 	return 0;
397 }
398 
399 static struct xenbus_device *xenbus_device_find(const char *nodename,
400 						struct bus_type *bus)
401 {
402 	struct xb_find_info info = { .dev = NULL, .nodename = nodename };
403 
404 	bus_for_each_dev(bus, NULL, &info, cmp_dev);
405 	return info.dev;
406 }
407 
408 static int cleanup_dev(struct device *dev, void *data)
409 {
410 	struct xenbus_device *xendev = to_xenbus_device(dev);
411 	struct xb_find_info *info = data;
412 	int len = strlen(info->nodename);
413 
414 	DPRINTK("%s", info->nodename);
415 
416 	/* Match the info->nodename path, or any subdirectory of that path. */
417 	if (strncmp(xendev->nodename, info->nodename, len))
418 		return 0;
419 
420 	/* If the node name is longer, ensure it really is a subdirectory. */
421 	if ((strlen(xendev->nodename) > len) && (xendev->nodename[len] != '/'))
422 		return 0;
423 
424 	info->dev = xendev;
425 	get_device(dev);
426 	return 1;
427 }
428 
429 static void xenbus_cleanup_devices(const char *path, struct bus_type *bus)
430 {
431 	struct xb_find_info info = { .nodename = path };
432 
433 	do {
434 		info.dev = NULL;
435 		bus_for_each_dev(bus, NULL, &info, cleanup_dev);
436 		if (info.dev) {
437 			device_unregister(&info.dev->dev);
438 			put_device(&info.dev->dev);
439 		}
440 	} while (info.dev);
441 }
442 
443 static void xenbus_dev_release(struct device *dev)
444 {
445 	if (dev)
446 		kfree(to_xenbus_device(dev));
447 }
448 
449 static ssize_t nodename_show(struct device *dev,
450 			     struct device_attribute *attr, char *buf)
451 {
452 	return sprintf(buf, "%s\n", to_xenbus_device(dev)->nodename);
453 }
454 static DEVICE_ATTR_RO(nodename);
455 
456 static ssize_t devtype_show(struct device *dev,
457 			    struct device_attribute *attr, char *buf)
458 {
459 	return sprintf(buf, "%s\n", to_xenbus_device(dev)->devicetype);
460 }
461 static DEVICE_ATTR_RO(devtype);
462 
463 static ssize_t modalias_show(struct device *dev,
464 			     struct device_attribute *attr, char *buf)
465 {
466 	return sprintf(buf, "%s:%s\n", dev->bus->name,
467 		       to_xenbus_device(dev)->devicetype);
468 }
469 static DEVICE_ATTR_RO(modalias);
470 
471 static ssize_t state_show(struct device *dev,
472 			    struct device_attribute *attr, char *buf)
473 {
474 	return sprintf(buf, "%s\n",
475 			xenbus_strstate(to_xenbus_device(dev)->state));
476 }
477 static DEVICE_ATTR_RO(state);
478 
479 static struct attribute *xenbus_dev_attrs[] = {
480 	&dev_attr_nodename.attr,
481 	&dev_attr_devtype.attr,
482 	&dev_attr_modalias.attr,
483 	&dev_attr_state.attr,
484 	NULL,
485 };
486 
487 static const struct attribute_group xenbus_dev_group = {
488 	.attrs = xenbus_dev_attrs,
489 };
490 
491 const struct attribute_group *xenbus_dev_groups[] = {
492 	&xenbus_dev_group,
493 	NULL,
494 };
495 EXPORT_SYMBOL_GPL(xenbus_dev_groups);
496 
497 int xenbus_probe_node(struct xen_bus_type *bus,
498 		      const char *type,
499 		      const char *nodename)
500 {
501 	char devname[XEN_BUS_ID_SIZE];
502 	int err;
503 	struct xenbus_device *xendev;
504 	size_t stringlen;
505 	char *tmpstring;
506 
507 	enum xenbus_state state = xenbus_read_driver_state(nodename);
508 
509 	if (state != XenbusStateInitialising) {
510 		/* Device is not new, so ignore it.  This can happen if a
511 		   device is going away after switching to Closed.  */
512 		return 0;
513 	}
514 
515 	stringlen = strlen(nodename) + 1 + strlen(type) + 1;
516 	xendev = kzalloc(sizeof(*xendev) + stringlen, GFP_KERNEL);
517 	if (!xendev)
518 		return -ENOMEM;
519 
520 	xendev->state = XenbusStateInitialising;
521 
522 	/* Copy the strings into the extra space. */
523 
524 	tmpstring = (char *)(xendev + 1);
525 	strcpy(tmpstring, nodename);
526 	xendev->nodename = tmpstring;
527 
528 	tmpstring += strlen(tmpstring) + 1;
529 	strcpy(tmpstring, type);
530 	xendev->devicetype = tmpstring;
531 	init_completion(&xendev->down);
532 
533 	xendev->dev.bus = &bus->bus;
534 	xendev->dev.release = xenbus_dev_release;
535 
536 	err = bus->get_bus_id(devname, xendev->nodename);
537 	if (err)
538 		goto fail;
539 
540 	dev_set_name(&xendev->dev, "%s", devname);
541 	sema_init(&xendev->reclaim_sem, 1);
542 
543 	/* Register with generic device framework. */
544 	err = device_register(&xendev->dev);
545 	if (err) {
546 		put_device(&xendev->dev);
547 		xendev = NULL;
548 		goto fail;
549 	}
550 
551 	return 0;
552 fail:
553 	kfree(xendev);
554 	return err;
555 }
556 EXPORT_SYMBOL_GPL(xenbus_probe_node);
557 
558 static int xenbus_probe_device_type(struct xen_bus_type *bus, const char *type)
559 {
560 	int err = 0;
561 	char **dir;
562 	unsigned int dir_n = 0;
563 	int i;
564 
565 	dir = xenbus_directory(XBT_NIL, bus->root, type, &dir_n);
566 	if (IS_ERR(dir))
567 		return PTR_ERR(dir);
568 
569 	for (i = 0; i < dir_n; i++) {
570 		err = bus->probe(bus, type, dir[i]);
571 		if (err)
572 			break;
573 	}
574 
575 	kfree(dir);
576 	return err;
577 }
578 
579 int xenbus_probe_devices(struct xen_bus_type *bus)
580 {
581 	int err = 0;
582 	char **dir;
583 	unsigned int i, dir_n;
584 
585 	dir = xenbus_directory(XBT_NIL, bus->root, "", &dir_n);
586 	if (IS_ERR(dir))
587 		return PTR_ERR(dir);
588 
589 	for (i = 0; i < dir_n; i++) {
590 		err = xenbus_probe_device_type(bus, dir[i]);
591 		if (err)
592 			break;
593 	}
594 
595 	kfree(dir);
596 	return err;
597 }
598 EXPORT_SYMBOL_GPL(xenbus_probe_devices);
599 
600 static unsigned int char_count(const char *str, char c)
601 {
602 	unsigned int i, ret = 0;
603 
604 	for (i = 0; str[i]; i++)
605 		if (str[i] == c)
606 			ret++;
607 	return ret;
608 }
609 
610 static int strsep_len(const char *str, char c, unsigned int len)
611 {
612 	unsigned int i;
613 
614 	for (i = 0; str[i]; i++)
615 		if (str[i] == c) {
616 			if (len == 0)
617 				return i;
618 			len--;
619 		}
620 	return (len == 0) ? i : -ERANGE;
621 }
622 
623 void xenbus_dev_changed(const char *node, struct xen_bus_type *bus)
624 {
625 	int exists, rootlen;
626 	struct xenbus_device *dev;
627 	char type[XEN_BUS_ID_SIZE];
628 	const char *p, *root;
629 
630 	if (char_count(node, '/') < 2)
631 		return;
632 
633 	exists = xenbus_exists(XBT_NIL, node, "");
634 	if (!exists) {
635 		xenbus_cleanup_devices(node, &bus->bus);
636 		return;
637 	}
638 
639 	/* backend/<type>/... or device/<type>/... */
640 	p = strchr(node, '/') + 1;
641 	snprintf(type, XEN_BUS_ID_SIZE, "%.*s", (int)strcspn(p, "/"), p);
642 	type[XEN_BUS_ID_SIZE-1] = '\0';
643 
644 	rootlen = strsep_len(node, '/', bus->levels);
645 	if (rootlen < 0)
646 		return;
647 	root = kasprintf(GFP_KERNEL, "%.*s", rootlen, node);
648 	if (!root)
649 		return;
650 
651 	dev = xenbus_device_find(root, &bus->bus);
652 	if (!dev)
653 		xenbus_probe_node(bus, type, root);
654 	else
655 		put_device(&dev->dev);
656 
657 	kfree(root);
658 }
659 EXPORT_SYMBOL_GPL(xenbus_dev_changed);
660 
661 int xenbus_dev_suspend(struct device *dev)
662 {
663 	int err = 0;
664 	struct xenbus_driver *drv;
665 	struct xenbus_device *xdev
666 		= container_of(dev, struct xenbus_device, dev);
667 
668 	DPRINTK("%s", xdev->nodename);
669 
670 	if (dev->driver == NULL)
671 		return 0;
672 	drv = to_xenbus_driver(dev->driver);
673 	if (drv->suspend)
674 		err = drv->suspend(xdev);
675 	if (err)
676 		dev_warn(dev, "suspend failed: %i\n", err);
677 	return 0;
678 }
679 EXPORT_SYMBOL_GPL(xenbus_dev_suspend);
680 
681 int xenbus_dev_resume(struct device *dev)
682 {
683 	int err;
684 	struct xenbus_driver *drv;
685 	struct xenbus_device *xdev
686 		= container_of(dev, struct xenbus_device, dev);
687 
688 	DPRINTK("%s", xdev->nodename);
689 
690 	if (dev->driver == NULL)
691 		return 0;
692 	drv = to_xenbus_driver(dev->driver);
693 	err = talk_to_otherend(xdev);
694 	if (err) {
695 		dev_warn(dev, "resume (talk_to_otherend) failed: %i\n", err);
696 		return err;
697 	}
698 
699 	xdev->state = XenbusStateInitialising;
700 
701 	if (drv->resume) {
702 		err = drv->resume(xdev);
703 		if (err) {
704 			dev_warn(dev, "resume failed: %i\n", err);
705 			return err;
706 		}
707 	}
708 
709 	err = watch_otherend(xdev);
710 	if (err) {
711 		dev_warn(dev, "resume (watch_otherend) failed: %d\n", err);
712 		return err;
713 	}
714 
715 	return 0;
716 }
717 EXPORT_SYMBOL_GPL(xenbus_dev_resume);
718 
719 int xenbus_dev_cancel(struct device *dev)
720 {
721 	/* Do nothing */
722 	DPRINTK("cancel");
723 	return 0;
724 }
725 EXPORT_SYMBOL_GPL(xenbus_dev_cancel);
726 
727 /* A flag to determine if xenstored is 'ready' (i.e. has started) */
728 int xenstored_ready;
729 
730 
731 int register_xenstore_notifier(struct notifier_block *nb)
732 {
733 	int ret = 0;
734 
735 	if (xenstored_ready > 0)
736 		ret = nb->notifier_call(nb, 0, NULL);
737 	else
738 		blocking_notifier_chain_register(&xenstore_chain, nb);
739 
740 	return ret;
741 }
742 EXPORT_SYMBOL_GPL(register_xenstore_notifier);
743 
744 void unregister_xenstore_notifier(struct notifier_block *nb)
745 {
746 	blocking_notifier_chain_unregister(&xenstore_chain, nb);
747 }
748 EXPORT_SYMBOL_GPL(unregister_xenstore_notifier);
749 
750 static void xenbus_probe(void)
751 {
752 	xenstored_ready = 1;
753 
754 	if (!xen_store_interface) {
755 		xen_store_interface = memremap(xen_store_gfn << XEN_PAGE_SHIFT,
756 					       XEN_PAGE_SIZE, MEMREMAP_WB);
757 		/*
758 		 * Now it is safe to free the IRQ used for xenstore late
759 		 * initialization. No need to unbind: it is about to be
760 		 * bound again from xb_init_comms. Note that calling
761 		 * unbind_from_irqhandler now would result in xen_evtchn_close()
762 		 * being called and the event channel not being enabled again
763 		 * afterwards, resulting in missed event notifications.
764 		 */
765 		free_irq(xs_init_irq, &xb_waitq);
766 	}
767 
768 	/*
769 	 * In the HVM case, xenbus_init() deferred its call to
770 	 * xs_init() in case callbacks were not operational yet.
771 	 * So do it now.
772 	 */
773 	if (xen_store_domain_type == XS_HVM)
774 		xs_init();
775 
776 	/* Notify others that xenstore is up */
777 	blocking_notifier_call_chain(&xenstore_chain, 0, NULL);
778 }
779 
780 /*
781  * Returns true when XenStore init must be deferred in order to
782  * allow the PCI platform device to be initialised, before we
783  * can actually have event channel interrupts working.
784  */
785 static bool xs_hvm_defer_init_for_callback(void)
786 {
787 #ifdef CONFIG_XEN_PVHVM
788 	return xen_store_domain_type == XS_HVM &&
789 		!xen_have_vector_callback;
790 #else
791 	return false;
792 #endif
793 }
794 
795 static int xenbus_probe_thread(void *unused)
796 {
797 	DEFINE_WAIT(w);
798 
799 	/*
800 	 * We actually just want to wait for *any* trigger of xb_waitq,
801 	 * and run xenbus_probe() the moment it occurs.
802 	 */
803 	prepare_to_wait(&xb_waitq, &w, TASK_INTERRUPTIBLE);
804 	schedule();
805 	finish_wait(&xb_waitq, &w);
806 
807 	DPRINTK("probing");
808 	xenbus_probe();
809 	return 0;
810 }
811 
812 static int __init xenbus_probe_initcall(void)
813 {
814 	/*
815 	 * Probe XenBus here in the XS_PV case, and also XS_HVM unless we
816 	 * need to wait for the platform PCI device to come up or
817 	 * xen_store_interface is not ready.
818 	 */
819 	if (xen_store_domain_type == XS_PV ||
820 	    (xen_store_domain_type == XS_HVM &&
821 	     !xs_hvm_defer_init_for_callback() &&
822 	     xen_store_interface != NULL))
823 		xenbus_probe();
824 
825 	/*
826 	 * For XS_LOCAL or when xen_store_interface is not ready, spawn a
827 	 * thread which will wait for xenstored or a xenstore-stubdom to be
828 	 * started, then probe.  It will be triggered when communication
829 	 * starts happening, by waiting on xb_waitq.
830 	 */
831 	if (xen_store_domain_type == XS_LOCAL || xen_store_interface == NULL) {
832 		struct task_struct *probe_task;
833 
834 		probe_task = kthread_run(xenbus_probe_thread, NULL,
835 					 "xenbus_probe");
836 		if (IS_ERR(probe_task))
837 			return PTR_ERR(probe_task);
838 	}
839 	return 0;
840 }
841 device_initcall(xenbus_probe_initcall);
842 
843 int xen_set_callback_via(uint64_t via)
844 {
845 	struct xen_hvm_param a;
846 	int ret;
847 
848 	a.domid = DOMID_SELF;
849 	a.index = HVM_PARAM_CALLBACK_IRQ;
850 	a.value = via;
851 
852 	ret = HYPERVISOR_hvm_op(HVMOP_set_param, &a);
853 	if (ret)
854 		return ret;
855 
856 	/*
857 	 * If xenbus_probe_initcall() deferred the xenbus_probe()
858 	 * due to the callback not functioning yet, we can do it now.
859 	 */
860 	if (!xenstored_ready && xs_hvm_defer_init_for_callback())
861 		xenbus_probe();
862 
863 	return ret;
864 }
865 EXPORT_SYMBOL_GPL(xen_set_callback_via);
866 
867 /* Set up event channel for xenstored which is run as a local process
868  * (this is normally used only in dom0)
869  */
870 static int __init xenstored_local_init(void)
871 {
872 	int err = -ENOMEM;
873 	unsigned long page = 0;
874 	struct evtchn_alloc_unbound alloc_unbound;
875 
876 	/* Allocate Xenstore page */
877 	page = get_zeroed_page(GFP_KERNEL);
878 	if (!page)
879 		goto out_err;
880 
881 	xen_store_gfn = virt_to_gfn((void *)page);
882 
883 	/* Next allocate a local port which xenstored can bind to */
884 	alloc_unbound.dom        = DOMID_SELF;
885 	alloc_unbound.remote_dom = DOMID_SELF;
886 
887 	err = HYPERVISOR_event_channel_op(EVTCHNOP_alloc_unbound,
888 					  &alloc_unbound);
889 	if (err == -ENOSYS)
890 		goto out_err;
891 
892 	BUG_ON(err);
893 	xen_store_evtchn = alloc_unbound.port;
894 
895 	return 0;
896 
897  out_err:
898 	if (page != 0)
899 		free_page(page);
900 	return err;
901 }
902 
903 static int xenbus_resume_cb(struct notifier_block *nb,
904 			    unsigned long action, void *data)
905 {
906 	int err = 0;
907 
908 	if (xen_hvm_domain()) {
909 		uint64_t v = 0;
910 
911 		err = hvm_get_parameter(HVM_PARAM_STORE_EVTCHN, &v);
912 		if (!err && v)
913 			xen_store_evtchn = v;
914 		else
915 			pr_warn("Cannot update xenstore event channel: %d\n",
916 				err);
917 	} else
918 		xen_store_evtchn = xen_start_info->store_evtchn;
919 
920 	return err;
921 }
922 
923 static struct notifier_block xenbus_resume_nb = {
924 	.notifier_call = xenbus_resume_cb,
925 };
926 
927 static irqreturn_t xenbus_late_init(int irq, void *unused)
928 {
929 	int err;
930 	uint64_t v = 0;
931 
932 	err = hvm_get_parameter(HVM_PARAM_STORE_PFN, &v);
933 	if (err || !v || !~v)
934 		return IRQ_HANDLED;
935 	xen_store_gfn = (unsigned long)v;
936 
937 	wake_up(&xb_waitq);
938 	return IRQ_HANDLED;
939 }
940 
941 static int __init xenbus_init(void)
942 {
943 	int err;
944 	uint64_t v = 0;
945 	bool wait = false;
946 	xen_store_domain_type = XS_UNKNOWN;
947 
948 	if (!xen_domain())
949 		return -ENODEV;
950 
951 	xenbus_ring_ops_init();
952 
953 	if (xen_pv_domain())
954 		xen_store_domain_type = XS_PV;
955 	if (xen_hvm_domain())
956 		xen_store_domain_type = XS_HVM;
957 	if (xen_hvm_domain() && xen_initial_domain())
958 		xen_store_domain_type = XS_LOCAL;
959 	if (xen_pv_domain() && !xen_start_info->store_evtchn)
960 		xen_store_domain_type = XS_LOCAL;
961 	if (xen_pv_domain() && xen_start_info->store_evtchn)
962 		xenstored_ready = 1;
963 
964 	switch (xen_store_domain_type) {
965 	case XS_LOCAL:
966 		err = xenstored_local_init();
967 		if (err)
968 			goto out_error;
969 		xen_store_interface = gfn_to_virt(xen_store_gfn);
970 		break;
971 	case XS_PV:
972 		xen_store_evtchn = xen_start_info->store_evtchn;
973 		xen_store_gfn = xen_start_info->store_mfn;
974 		xen_store_interface = gfn_to_virt(xen_store_gfn);
975 		break;
976 	case XS_HVM:
977 		err = hvm_get_parameter(HVM_PARAM_STORE_EVTCHN, &v);
978 		if (err)
979 			goto out_error;
980 		xen_store_evtchn = (int)v;
981 		err = hvm_get_parameter(HVM_PARAM_STORE_PFN, &v);
982 		if (err)
983 			goto out_error;
984 		/*
985 		 * Uninitialized hvm_params are zero and return no error.
986 		 * Although it is theoretically possible to have
987 		 * HVM_PARAM_STORE_PFN set to zero on purpose, in reality it is
988 		 * not zero when valid. If zero, it means that Xenstore hasn't
989 		 * been properly initialized. Instead of attempting to map a
990 		 * wrong guest physical address return error.
991 		 *
992 		 * Also recognize all bits set as an invalid/uninitialized value.
993 		 */
994 		if (!v) {
995 			err = -ENOENT;
996 			goto out_error;
997 		}
998 		if (v == ~0ULL) {
999 			wait = true;
1000 		} else {
1001 			/* Avoid truncation on 32-bit. */
1002 #if BITS_PER_LONG == 32
1003 			if (v > ULONG_MAX) {
1004 				pr_err("%s: cannot handle HVM_PARAM_STORE_PFN=%llx > ULONG_MAX\n",
1005 				       __func__, v);
1006 				err = -EINVAL;
1007 				goto out_error;
1008 			}
1009 #endif
1010 			xen_store_gfn = (unsigned long)v;
1011 			xen_store_interface =
1012 				memremap(xen_store_gfn << XEN_PAGE_SHIFT,
1013 					 XEN_PAGE_SIZE, MEMREMAP_WB);
1014 			if (xen_store_interface->connection != XENSTORE_CONNECTED)
1015 				wait = true;
1016 		}
1017 		if (wait) {
1018 			err = bind_evtchn_to_irqhandler(xen_store_evtchn,
1019 							xenbus_late_init,
1020 							0, "xenstore_late_init",
1021 							&xb_waitq);
1022 			if (err < 0) {
1023 				pr_err("xenstore_late_init couldn't bind irq err=%d\n",
1024 				       err);
1025 				return err;
1026 			}
1027 
1028 			xs_init_irq = err;
1029 		}
1030 		break;
1031 	default:
1032 		pr_warn("Xenstore state unknown\n");
1033 		break;
1034 	}
1035 
1036 	/*
1037 	 * HVM domains may not have a functional callback yet. In that
1038 	 * case let xs_init() be called from xenbus_probe(), which will
1039 	 * get invoked at an appropriate time.
1040 	 */
1041 	if (xen_store_domain_type != XS_HVM) {
1042 		err = xs_init();
1043 		if (err) {
1044 			pr_warn("Error initializing xenstore comms: %i\n", err);
1045 			goto out_error;
1046 		}
1047 	}
1048 
1049 	if ((xen_store_domain_type != XS_LOCAL) &&
1050 	    (xen_store_domain_type != XS_UNKNOWN))
1051 		xen_resume_notifier_register(&xenbus_resume_nb);
1052 
1053 #ifdef CONFIG_XEN_COMPAT_XENFS
1054 	/*
1055 	 * Create xenfs mountpoint in /proc for compatibility with
1056 	 * utilities that expect to find "xenbus" under "/proc/xen".
1057 	 */
1058 	proc_create_mount_point("xen");
1059 #endif
1060 	return 0;
1061 
1062 out_error:
1063 	xen_store_domain_type = XS_UNKNOWN;
1064 	return err;
1065 }
1066 
1067 postcore_initcall(xenbus_init);
1068 
1069 MODULE_LICENSE("GPL");
1070