| /* |
| * QEMU Xen emulation: Shared/overlay pages support |
| * |
| * Copyright © 2022 Amazon.com, Inc. or its affiliates. All Rights Reserved. |
| * |
| * Authors: David Woodhouse <dwmw2@infradead.org> |
| * |
| * This work is licensed under the terms of the GNU GPL, version 2 or later. |
| * See the COPYING file in the top-level directory. |
| */ |
| |
| #include "qemu/osdep.h" |
| |
| #include "qemu/host-utils.h" |
| #include "qemu/module.h" |
| #include "qemu/main-loop.h" |
| #include "qemu/cutils.h" |
| #include "qemu/error-report.h" |
| #include "qapi/error.h" |
| #include "qom/object.h" |
| #include "migration/vmstate.h" |
| |
| #include "hw/sysbus.h" |
| #include "hw/xen/xen.h" |
| #include "hw/xen/xen_backend_ops.h" |
| #include "xen_overlay.h" |
| #include "xen_evtchn.h" |
| #include "xen_primary_console.h" |
| #include "xen_xenstore.h" |
| |
| #include "sysemu/kvm.h" |
| #include "sysemu/kvm_xen.h" |
| |
| #include "trace.h" |
| |
| #include "xenstore_impl.h" |
| |
| #include "hw/xen/interface/io/xs_wire.h" |
| #include "hw/xen/interface/event_channel.h" |
| #include "hw/xen/interface/grant_table.h" |
| |
| #define TYPE_XEN_XENSTORE "xen-xenstore" |
| OBJECT_DECLARE_SIMPLE_TYPE(XenXenstoreState, XEN_XENSTORE) |
| |
| #define ENTRIES_PER_FRAME_V1 (XEN_PAGE_SIZE / sizeof(grant_entry_v1_t)) |
| #define ENTRIES_PER_FRAME_V2 (XEN_PAGE_SIZE / sizeof(grant_entry_v2_t)) |
| |
| #define XENSTORE_HEADER_SIZE ((unsigned int)sizeof(struct xsd_sockmsg)) |
| |
| struct XenXenstoreState { |
| /*< private >*/ |
| SysBusDevice busdev; |
| /*< public >*/ |
| |
| XenstoreImplState *impl; |
| GList *watch_events; /* for the guest */ |
| |
| MemoryRegion xenstore_page; |
| struct xenstore_domain_interface *xs; |
| uint8_t req_data[XENSTORE_HEADER_SIZE + XENSTORE_PAYLOAD_MAX]; |
| uint8_t rsp_data[XENSTORE_HEADER_SIZE + XENSTORE_PAYLOAD_MAX]; |
| uint32_t req_offset; |
| uint32_t rsp_offset; |
| bool rsp_pending; |
| bool fatal_error; |
| |
| evtchn_port_t guest_port; |
| evtchn_port_t be_port; |
| struct xenevtchn_handle *eh; |
| |
| uint8_t *impl_state; |
| uint32_t impl_state_size; |
| |
| struct xengntdev_handle *gt; |
| void *granted_xs; |
| }; |
| |
| struct XenXenstoreState *xen_xenstore_singleton; |
| |
| static void xen_xenstore_event(void *opaque); |
| static void fire_watch_cb(void *opaque, const char *path, const char *token); |
| |
| static struct xenstore_backend_ops emu_xenstore_backend_ops; |
| |
| static void G_GNUC_PRINTF (4, 5) relpath_printf(XenXenstoreState *s, |
| GList *perms, |
| const char *relpath, |
| const char *fmt, ...) |
| { |
| gchar *abspath; |
| gchar *value; |
| va_list args; |
| GByteArray *data; |
| int err; |
| |
| abspath = g_strdup_printf("/local/domain/%u/%s", xen_domid, relpath); |
| va_start(args, fmt); |
| value = g_strdup_vprintf(fmt, args); |
| va_end(args); |
| |
| data = g_byte_array_new_take((void *)value, strlen(value)); |
| |
| err = xs_impl_write(s->impl, DOMID_QEMU, XBT_NULL, abspath, data); |
| assert(!err); |
| |
| g_byte_array_unref(data); |
| |
| err = xs_impl_set_perms(s->impl, DOMID_QEMU, XBT_NULL, abspath, perms); |
| assert(!err); |
| |
| g_free(abspath); |
| } |
| |
| static void xen_xenstore_realize(DeviceState *dev, Error **errp) |
| { |
| XenXenstoreState *s = XEN_XENSTORE(dev); |
| GList *perms; |
| |
| if (xen_mode != XEN_EMULATE) { |
| error_setg(errp, "Xen xenstore support is for Xen emulation"); |
| return; |
| } |
| memory_region_init_ram(&s->xenstore_page, OBJECT(dev), "xen:xenstore_page", |
| XEN_PAGE_SIZE, &error_abort); |
| memory_region_set_enabled(&s->xenstore_page, true); |
| s->xs = memory_region_get_ram_ptr(&s->xenstore_page); |
| memset(s->xs, 0, XEN_PAGE_SIZE); |
| |
| /* We can't map it this early as KVM isn't ready */ |
| xen_xenstore_singleton = s; |
| |
| s->eh = xen_be_evtchn_open(); |
| if (!s->eh) { |
| error_setg(errp, "Xenstore evtchn port init failed"); |
| return; |
| } |
| aio_set_fd_handler(qemu_get_aio_context(), xen_be_evtchn_fd(s->eh), |
| xen_xenstore_event, NULL, NULL, NULL, s); |
| |
| s->impl = xs_impl_create(xen_domid); |
| |
| /* Populate the default nodes */ |
| |
| /* Nodes owned by 'dom0' but readable by the guest */ |
| perms = g_list_append(NULL, xs_perm_as_string(XS_PERM_NONE, DOMID_QEMU)); |
| perms = g_list_append(perms, xs_perm_as_string(XS_PERM_READ, xen_domid)); |
| |
| relpath_printf(s, perms, "", "%s", ""); |
| |
| relpath_printf(s, perms, "domid", "%u", xen_domid); |
| |
| relpath_printf(s, perms, "control/platform-feature-xs_reset_watches", "%u", 1); |
| relpath_printf(s, perms, "control/platform-feature-multiprocessor-suspend", "%u", 1); |
| |
| relpath_printf(s, perms, "platform/acpi", "%u", 1); |
| relpath_printf(s, perms, "platform/acpi_s3", "%u", 1); |
| relpath_printf(s, perms, "platform/acpi_s4", "%u", 1); |
| relpath_printf(s, perms, "platform/acpi_laptop_slate", "%u", 0); |
| |
| g_list_free_full(perms, g_free); |
| |
| /* Nodes owned by the guest */ |
| perms = g_list_append(NULL, xs_perm_as_string(XS_PERM_NONE, xen_domid)); |
| |
| relpath_printf(s, perms, "attr", "%s", ""); |
| |
| relpath_printf(s, perms, "control/shutdown", "%s", ""); |
| relpath_printf(s, perms, "control/feature-poweroff", "%u", 1); |
| relpath_printf(s, perms, "control/feature-reboot", "%u", 1); |
| relpath_printf(s, perms, "control/feature-suspend", "%u", 1); |
| relpath_printf(s, perms, "control/feature-s3", "%u", 1); |
| relpath_printf(s, perms, "control/feature-s4", "%u", 1); |
| |
| relpath_printf(s, perms, "data", "%s", ""); |
| relpath_printf(s, perms, "device", "%s", ""); |
| relpath_printf(s, perms, "drivers", "%s", ""); |
| relpath_printf(s, perms, "error", "%s", ""); |
| relpath_printf(s, perms, "feature", "%s", ""); |
| |
| g_list_free_full(perms, g_free); |
| |
| xen_xenstore_ops = &emu_xenstore_backend_ops; |
| } |
| |
| static bool xen_xenstore_is_needed(void *opaque) |
| { |
| return xen_mode == XEN_EMULATE; |
| } |
| |
| static int xen_xenstore_pre_save(void *opaque) |
| { |
| XenXenstoreState *s = opaque; |
| GByteArray *save; |
| |
| if (s->eh) { |
| s->guest_port = xen_be_evtchn_get_guest_port(s->eh); |
| } |
| |
| g_free(s->impl_state); |
| save = xs_impl_serialize(s->impl); |
| s->impl_state = save->data; |
| s->impl_state_size = save->len; |
| g_byte_array_free(save, false); |
| |
| return 0; |
| } |
| |
| static int xen_xenstore_post_load(void *opaque, int ver) |
| { |
| XenXenstoreState *s = opaque; |
| GByteArray *save; |
| int ret; |
| |
| /* |
| * As qemu/dom0, rebind to the guest's port. The Windows drivers may |
| * unbind the XenStore evtchn and rebind to it, having obtained the |
| * "remote" port through EVTCHNOP_status. In the case that migration |
| * occurs while it's unbound, the "remote" port needs to be the same |
| * as before so that the guest can find it, but should remain unbound. |
| */ |
| if (s->guest_port) { |
| int be_port = xen_be_evtchn_bind_interdomain(s->eh, xen_domid, |
| s->guest_port); |
| if (be_port < 0) { |
| return be_port; |
| } |
| s->be_port = be_port; |
| } |
| |
| save = g_byte_array_new_take(s->impl_state, s->impl_state_size); |
| s->impl_state = NULL; |
| s->impl_state_size = 0; |
| |
| ret = xs_impl_deserialize(s->impl, save, xen_domid, fire_watch_cb, s); |
| return ret; |
| } |
| |
| static const VMStateDescription xen_xenstore_vmstate = { |
| .name = "xen_xenstore", |
| .unmigratable = 1, /* The PV back ends don't migrate yet */ |
| .version_id = 1, |
| .minimum_version_id = 1, |
| .needed = xen_xenstore_is_needed, |
| .pre_save = xen_xenstore_pre_save, |
| .post_load = xen_xenstore_post_load, |
| .fields = (const VMStateField[]) { |
| VMSTATE_UINT8_ARRAY(req_data, XenXenstoreState, |
| sizeof_field(XenXenstoreState, req_data)), |
| VMSTATE_UINT8_ARRAY(rsp_data, XenXenstoreState, |
| sizeof_field(XenXenstoreState, rsp_data)), |
| VMSTATE_UINT32(req_offset, XenXenstoreState), |
| VMSTATE_UINT32(rsp_offset, XenXenstoreState), |
| VMSTATE_BOOL(rsp_pending, XenXenstoreState), |
| VMSTATE_UINT32(guest_port, XenXenstoreState), |
| VMSTATE_BOOL(fatal_error, XenXenstoreState), |
| VMSTATE_UINT32(impl_state_size, XenXenstoreState), |
| VMSTATE_VARRAY_UINT32_ALLOC(impl_state, XenXenstoreState, |
| impl_state_size, 0, |
| vmstate_info_uint8, uint8_t), |
| VMSTATE_END_OF_LIST() |
| } |
| }; |
| |
| static void xen_xenstore_class_init(ObjectClass *klass, void *data) |
| { |
| DeviceClass *dc = DEVICE_CLASS(klass); |
| |
| dc->realize = xen_xenstore_realize; |
| dc->vmsd = &xen_xenstore_vmstate; |
| } |
| |
| static const TypeInfo xen_xenstore_info = { |
| .name = TYPE_XEN_XENSTORE, |
| .parent = TYPE_SYS_BUS_DEVICE, |
| .instance_size = sizeof(XenXenstoreState), |
| .class_init = xen_xenstore_class_init, |
| }; |
| |
| void xen_xenstore_create(void) |
| { |
| DeviceState *dev = sysbus_create_simple(TYPE_XEN_XENSTORE, -1, NULL); |
| |
| xen_xenstore_singleton = XEN_XENSTORE(dev); |
| |
| /* |
| * Defer the init (xen_xenstore_reset()) until KVM is set up and the |
| * overlay page can be mapped. |
| */ |
| } |
| |
| static void xen_xenstore_register_types(void) |
| { |
| type_register_static(&xen_xenstore_info); |
| } |
| |
| type_init(xen_xenstore_register_types) |
| |
| uint16_t xen_xenstore_get_port(void) |
| { |
| XenXenstoreState *s = xen_xenstore_singleton; |
| if (!s) { |
| return 0; |
| } |
| return s->guest_port; |
| } |
| |
| static bool req_pending(XenXenstoreState *s) |
| { |
| struct xsd_sockmsg *req = (struct xsd_sockmsg *)s->req_data; |
| |
| return s->req_offset == XENSTORE_HEADER_SIZE + req->len; |
| } |
| |
| static void reset_req(XenXenstoreState *s) |
| { |
| memset(s->req_data, 0, sizeof(s->req_data)); |
| s->req_offset = 0; |
| } |
| |
| static void reset_rsp(XenXenstoreState *s) |
| { |
| s->rsp_pending = false; |
| |
| memset(s->rsp_data, 0, sizeof(s->rsp_data)); |
| s->rsp_offset = 0; |
| } |
| |
| static void xs_error(XenXenstoreState *s, unsigned int id, |
| xs_transaction_t tx_id, int errnum) |
| { |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| const char *errstr = NULL; |
| |
| for (unsigned int i = 0; i < ARRAY_SIZE(xsd_errors); i++) { |
| const struct xsd_errors *xsd_error = &xsd_errors[i]; |
| |
| if (xsd_error->errnum == errnum) { |
| errstr = xsd_error->errstring; |
| break; |
| } |
| } |
| assert(errstr); |
| |
| trace_xenstore_error(id, tx_id, errstr); |
| |
| rsp->type = XS_ERROR; |
| rsp->req_id = id; |
| rsp->tx_id = tx_id; |
| rsp->len = (uint32_t)strlen(errstr) + 1; |
| |
| memcpy(&rsp[1], errstr, rsp->len); |
| } |
| |
| static void xs_ok(XenXenstoreState *s, unsigned int type, unsigned int req_id, |
| xs_transaction_t tx_id) |
| { |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| const char *okstr = "OK"; |
| |
| rsp->type = type; |
| rsp->req_id = req_id; |
| rsp->tx_id = tx_id; |
| rsp->len = (uint32_t)strlen(okstr) + 1; |
| |
| memcpy(&rsp[1], okstr, rsp->len); |
| } |
| |
| /* |
| * The correct request and response formats are documented in xen.git: |
| * docs/misc/xenstore.txt. A summary is given below for convenience. |
| * The '|' symbol represents a NUL character. |
| * |
| * ---------- Database read, write and permissions operations ---------- |
| * |
| * READ <path>| <value|> |
| * WRITE <path>|<value|> |
| * Store and read the octet string <value> at <path>. |
| * WRITE creates any missing parent paths, with empty values. |
| * |
| * MKDIR <path>| |
| * Ensures that the <path> exists, by necessary by creating |
| * it and any missing parents with empty values. If <path> |
| * or any parent already exists, its value is left unchanged. |
| * |
| * RM <path>| |
| * Ensures that the <path> does not exist, by deleting |
| * it and all of its children. It is not an error if <path> does |
| * not exist, but it _is_ an error if <path>'s immediate parent |
| * does not exist either. |
| * |
| * DIRECTORY <path>| <child-leaf-name>|* |
| * Gives a list of the immediate children of <path>, as only the |
| * leafnames. The resulting children are each named |
| * <path>/<child-leaf-name>. |
| * |
| * DIRECTORY_PART <path>|<offset> <gencnt>|<child-leaf-name>|* |
| * Same as DIRECTORY, but to be used for children lists longer than |
| * XENSTORE_PAYLOAD_MAX. Input are <path> and the byte offset into |
| * the list of children to return. Return values are the generation |
| * count <gencnt> of the node (to be used to ensure the node hasn't |
| * changed between two reads: <gencnt> being the same for multiple |
| * reads guarantees the node hasn't changed) and the list of children |
| * starting at the specified <offset> of the complete list. |
| * |
| * GET_PERMS <path>| <perm-as-string>|+ |
| * SET_PERMS <path>|<perm-as-string>|+? |
| * <perm-as-string> is one of the following |
| * w<domid> write only |
| * r<domid> read only |
| * b<domid> both read and write |
| * n<domid> no access |
| * See https://wiki.xen.org/wiki/XenBus section |
| * `Permissions' for details of the permissions system. |
| * It is possible to set permissions for the special watch paths |
| * "@introduceDomain" and "@releaseDomain" to enable receiving those |
| * watches in unprivileged domains. |
| * |
| * ---------- Watches ---------- |
| * |
| * WATCH <wpath>|<token>|? |
| * Adds a watch. |
| * |
| * When a <path> is modified (including path creation, removal, |
| * contents change or permissions change) this generates an event |
| * on the changed <path>. Changes made in transactions cause an |
| * event only if and when committed. Each occurring event is |
| * matched against all the watches currently set up, and each |
| * matching watch results in a WATCH_EVENT message (see below). |
| * |
| * The event's path matches the watch's <wpath> if it is an child |
| * of <wpath>. |
| * |
| * <wpath> can be a <path> to watch or @<wspecial>. In the |
| * latter case <wspecial> may have any syntax but it matches |
| * (according to the rules above) only the following special |
| * events which are invented by xenstored: |
| * @introduceDomain occurs on INTRODUCE |
| * @releaseDomain occurs on any domain crash or |
| * shutdown, and also on RELEASE |
| * and domain destruction |
| * <wspecial> events are sent to privileged callers or explicitly |
| * via SET_PERMS enabled domains only. |
| * |
| * When a watch is first set up it is triggered once straight |
| * away, with <path> equal to <wpath>. Watches may be triggered |
| * spuriously. The tx_id in a WATCH request is ignored. |
| * |
| * Watches are supposed to be restricted by the permissions |
| * system but in practice the implementation is imperfect. |
| * Applications should not rely on being sent a notification for |
| * paths that they cannot read; however, an application may rely |
| * on being sent a watch when a path which it _is_ able to read |
| * is deleted even if that leaves only a nonexistent unreadable |
| * parent. A notification may omitted if a node's permissions |
| * are changed so as to make it unreadable, in which case future |
| * notifications may be suppressed (and if the node is later made |
| * readable, some notifications may have been lost). |
| * |
| * WATCH_EVENT <epath>|<token>| |
| * Unsolicited `reply' generated for matching modification events |
| * as described above. req_id and tx_id are both 0. |
| * |
| * <epath> is the event's path, ie the actual path that was |
| * modified; however if the event was the recursive removal of an |
| * parent of <wpath>, <epath> is just |
| * <wpath> (rather than the actual path which was removed). So |
| * <epath> is a child of <wpath>, regardless. |
| * |
| * Iff <wpath> for the watch was specified as a relative pathname, |
| * the <epath> path will also be relative (with the same base, |
| * obviously). |
| * |
| * UNWATCH <wpath>|<token>|? |
| * |
| * RESET_WATCHES | |
| * Reset all watches and transactions of the caller. |
| * |
| * ---------- Transactions ---------- |
| * |
| * TRANSACTION_START | <transid>| |
| * <transid> is an opaque uint32_t allocated by xenstored |
| * represented as unsigned decimal. After this, transaction may |
| * be referenced by using <transid> (as 32-bit binary) in the |
| * tx_id request header field. When transaction is started whole |
| * db is copied; reads and writes happen on the copy. |
| * It is not legal to send non-0 tx_id in TRANSACTION_START. |
| * |
| * TRANSACTION_END T| |
| * TRANSACTION_END F| |
| * tx_id must refer to existing transaction. After this |
| * request the tx_id is no longer valid and may be reused by |
| * xenstore. If F, the transaction is discarded. If T, |
| * it is committed: if there were any other intervening writes |
| * then our END gets get EAGAIN. |
| * |
| * The plan is that in the future only intervening `conflicting' |
| * writes cause EAGAIN, meaning only writes or other commits |
| * which changed paths which were read or written in the |
| * transaction at hand. |
| * |
| */ |
| |
| static void xs_read(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, unsigned int len) |
| { |
| const char *path = (const char *)req_data; |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| uint8_t *rsp_data = (uint8_t *)&rsp[1]; |
| g_autoptr(GByteArray) data = g_byte_array_new(); |
| int err; |
| |
| if (len == 0 || req_data[len - 1] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| trace_xenstore_read(tx_id, path); |
| err = xs_impl_read(s->impl, xen_domid, tx_id, path, data); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| rsp->type = XS_READ; |
| rsp->req_id = req_id; |
| rsp->tx_id = tx_id; |
| rsp->len = 0; |
| |
| len = data->len; |
| if (len > XENSTORE_PAYLOAD_MAX) { |
| xs_error(s, req_id, tx_id, E2BIG); |
| return; |
| } |
| |
| memcpy(&rsp_data[rsp->len], data->data, len); |
| rsp->len += len; |
| } |
| |
| static void xs_write(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| g_autoptr(GByteArray) data = g_byte_array_new(); |
| const char *path; |
| int err; |
| |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| path = (const char *)req_data; |
| |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| g_byte_array_append(data, req_data, len); |
| |
| trace_xenstore_write(tx_id, path); |
| err = xs_impl_write(s->impl, xen_domid, tx_id, path, data); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| xs_ok(s, XS_WRITE, req_id, tx_id); |
| } |
| |
| static void xs_mkdir(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| g_autoptr(GByteArray) data = g_byte_array_new(); |
| const char *path; |
| int err; |
| |
| if (len == 0 || req_data[len - 1] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| path = (const char *)req_data; |
| |
| trace_xenstore_mkdir(tx_id, path); |
| err = xs_impl_read(s->impl, xen_domid, tx_id, path, data); |
| if (err == ENOENT) { |
| err = xs_impl_write(s->impl, xen_domid, tx_id, path, data); |
| } |
| |
| if (!err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| xs_ok(s, XS_MKDIR, req_id, tx_id); |
| } |
| |
| static void xs_append_strings(XenXenstoreState *s, struct xsd_sockmsg *rsp, |
| GList *strings, unsigned int start, bool truncate) |
| { |
| uint8_t *rsp_data = (uint8_t *)&rsp[1]; |
| GList *l; |
| |
| for (l = strings; l; l = l->next) { |
| size_t len = strlen(l->data) + 1; /* Including the NUL termination */ |
| char *str = l->data; |
| |
| if (rsp->len + len > XENSTORE_PAYLOAD_MAX) { |
| if (truncate) { |
| len = XENSTORE_PAYLOAD_MAX - rsp->len; |
| if (!len) { |
| return; |
| } |
| } else { |
| xs_error(s, rsp->req_id, rsp->tx_id, E2BIG); |
| return; |
| } |
| } |
| |
| if (start) { |
| if (start >= len) { |
| start -= len; |
| continue; |
| } |
| |
| str += start; |
| len -= start; |
| start = 0; |
| } |
| |
| memcpy(&rsp_data[rsp->len], str, len); |
| rsp->len += len; |
| } |
| /* XS_DIRECTORY_PART wants an extra NUL to indicate the end */ |
| if (truncate && rsp->len < XENSTORE_PAYLOAD_MAX) { |
| rsp_data[rsp->len++] = '\0'; |
| } |
| } |
| |
| static void xs_directory(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| GList *items = NULL; |
| const char *path; |
| int err; |
| |
| if (len == 0 || req_data[len - 1] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| path = (const char *)req_data; |
| |
| trace_xenstore_directory(tx_id, path); |
| err = xs_impl_directory(s->impl, xen_domid, tx_id, path, NULL, &items); |
| if (err != 0) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| rsp->type = XS_DIRECTORY; |
| rsp->req_id = req_id; |
| rsp->tx_id = tx_id; |
| rsp->len = 0; |
| |
| xs_append_strings(s, rsp, items, 0, false); |
| |
| g_list_free_full(items, g_free); |
| } |
| |
| static void xs_directory_part(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| const char *offset_str, *path = (const char *)req_data; |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| char *rsp_data = (char *)&rsp[1]; |
| uint64_t gencnt = 0; |
| unsigned int offset; |
| GList *items = NULL; |
| int err; |
| |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| offset_str = (const char *)req_data; |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| if (len) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| if (qemu_strtoui(offset_str, NULL, 10, &offset) < 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| trace_xenstore_directory_part(tx_id, path, offset); |
| err = xs_impl_directory(s->impl, xen_domid, tx_id, path, &gencnt, &items); |
| if (err != 0) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| rsp->type = XS_DIRECTORY_PART; |
| rsp->req_id = req_id; |
| rsp->tx_id = tx_id; |
| rsp->len = snprintf(rsp_data, XENSTORE_PAYLOAD_MAX, "%" PRIu64, gencnt) + 1; |
| |
| xs_append_strings(s, rsp, items, offset, true); |
| |
| g_list_free_full(items, g_free); |
| } |
| |
| static void xs_transaction_start(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| char *rsp_data = (char *)&rsp[1]; |
| int err; |
| |
| if (len != 1 || req_data[0] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| rsp->type = XS_TRANSACTION_START; |
| rsp->req_id = req_id; |
| rsp->tx_id = tx_id; |
| rsp->len = 0; |
| |
| err = xs_impl_transaction_start(s->impl, xen_domid, &tx_id); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| trace_xenstore_transaction_start(tx_id); |
| |
| rsp->len = snprintf(rsp_data, XENSTORE_PAYLOAD_MAX, "%u", tx_id); |
| assert(rsp->len < XENSTORE_PAYLOAD_MAX); |
| rsp->len++; |
| } |
| |
| static void xs_transaction_end(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| bool commit; |
| int err; |
| |
| if (len != 2 || req_data[1] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| switch (req_data[0]) { |
| case 'T': |
| commit = true; |
| break; |
| case 'F': |
| commit = false; |
| break; |
| default: |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| trace_xenstore_transaction_end(tx_id, commit); |
| err = xs_impl_transaction_end(s->impl, xen_domid, tx_id, commit); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| xs_ok(s, XS_TRANSACTION_END, req_id, tx_id); |
| } |
| |
| static void xs_rm(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, unsigned int len) |
| { |
| const char *path = (const char *)req_data; |
| int err; |
| |
| if (len == 0 || req_data[len - 1] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| trace_xenstore_rm(tx_id, path); |
| err = xs_impl_rm(s->impl, xen_domid, tx_id, path); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| xs_ok(s, XS_RM, req_id, tx_id); |
| } |
| |
| static void xs_get_perms(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| const char *path = (const char *)req_data; |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| GList *perms = NULL; |
| int err; |
| |
| if (len == 0 || req_data[len - 1] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| trace_xenstore_get_perms(tx_id, path); |
| err = xs_impl_get_perms(s->impl, xen_domid, tx_id, path, &perms); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| rsp->type = XS_GET_PERMS; |
| rsp->req_id = req_id; |
| rsp->tx_id = tx_id; |
| rsp->len = 0; |
| |
| xs_append_strings(s, rsp, perms, 0, false); |
| |
| g_list_free_full(perms, g_free); |
| } |
| |
| static void xs_set_perms(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| const char *path = (const char *)req_data; |
| uint8_t *perm; |
| GList *perms = NULL; |
| int err; |
| |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| perm = req_data; |
| while (len--) { |
| if (*req_data++ == '\0') { |
| perms = g_list_append(perms, perm); |
| perm = req_data; |
| } |
| } |
| |
| /* |
| * Note that there may be trailing garbage at the end of the buffer. |
| * This is explicitly permitted by the '?' at the end of the definition: |
| * |
| * SET_PERMS <path>|<perm-as-string>|+? |
| */ |
| |
| trace_xenstore_set_perms(tx_id, path); |
| err = xs_impl_set_perms(s->impl, xen_domid, tx_id, path, perms); |
| g_list_free(perms); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| xs_ok(s, XS_SET_PERMS, req_id, tx_id); |
| } |
| |
| static void xs_watch(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| const char *token, *path = (const char *)req_data; |
| int err; |
| |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| token = (const char *)req_data; |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| /* |
| * Note that there may be trailing garbage at the end of the buffer. |
| * This is explicitly permitted by the '?' at the end of the definition: |
| * |
| * WATCH <wpath>|<token>|? |
| */ |
| |
| trace_xenstore_watch(path, token); |
| err = xs_impl_watch(s->impl, xen_domid, path, token, fire_watch_cb, s); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| xs_ok(s, XS_WATCH, req_id, tx_id); |
| } |
| |
| static void xs_unwatch(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| const char *token, *path = (const char *)req_data; |
| int err; |
| |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| token = (const char *)req_data; |
| while (len--) { |
| if (*req_data++ == '\0') { |
| break; |
| } |
| if (len == 0) { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| } |
| |
| trace_xenstore_unwatch(path, token); |
| err = xs_impl_unwatch(s->impl, xen_domid, path, token, fire_watch_cb, s); |
| if (err) { |
| xs_error(s, req_id, tx_id, err); |
| return; |
| } |
| |
| xs_ok(s, XS_UNWATCH, req_id, tx_id); |
| } |
| |
| static void xs_reset_watches(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *req_data, |
| unsigned int len) |
| { |
| if (len == 0 || req_data[len - 1] != '\0') { |
| xs_error(s, req_id, tx_id, EINVAL); |
| return; |
| } |
| |
| trace_xenstore_reset_watches(); |
| xs_impl_reset_watches(s->impl, xen_domid); |
| |
| xs_ok(s, XS_RESET_WATCHES, req_id, tx_id); |
| } |
| |
| static void xs_priv(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *data, |
| unsigned int len) |
| { |
| xs_error(s, req_id, tx_id, EACCES); |
| } |
| |
| static void xs_unimpl(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *data, |
| unsigned int len) |
| { |
| xs_error(s, req_id, tx_id, ENOSYS); |
| } |
| |
| typedef void (*xs_impl)(XenXenstoreState *s, unsigned int req_id, |
| xs_transaction_t tx_id, uint8_t *data, |
| unsigned int len); |
| |
| struct xsd_req { |
| const char *name; |
| xs_impl fn; |
| }; |
| #define XSD_REQ(_type, _fn) \ |
| [_type] = { .name = #_type, .fn = _fn } |
| |
| struct xsd_req xsd_reqs[] = { |
| XSD_REQ(XS_READ, xs_read), |
| XSD_REQ(XS_WRITE, xs_write), |
| XSD_REQ(XS_MKDIR, xs_mkdir), |
| XSD_REQ(XS_DIRECTORY, xs_directory), |
| XSD_REQ(XS_DIRECTORY_PART, xs_directory_part), |
| XSD_REQ(XS_TRANSACTION_START, xs_transaction_start), |
| XSD_REQ(XS_TRANSACTION_END, xs_transaction_end), |
| XSD_REQ(XS_RM, xs_rm), |
| XSD_REQ(XS_GET_PERMS, xs_get_perms), |
| XSD_REQ(XS_SET_PERMS, xs_set_perms), |
| XSD_REQ(XS_WATCH, xs_watch), |
| XSD_REQ(XS_UNWATCH, xs_unwatch), |
| XSD_REQ(XS_CONTROL, xs_priv), |
| XSD_REQ(XS_INTRODUCE, xs_priv), |
| XSD_REQ(XS_RELEASE, xs_priv), |
| XSD_REQ(XS_IS_DOMAIN_INTRODUCED, xs_priv), |
| XSD_REQ(XS_RESUME, xs_priv), |
| XSD_REQ(XS_SET_TARGET, xs_priv), |
| XSD_REQ(XS_RESET_WATCHES, xs_reset_watches), |
| }; |
| |
| static void process_req(XenXenstoreState *s) |
| { |
| struct xsd_sockmsg *req = (struct xsd_sockmsg *)s->req_data; |
| xs_impl handler = NULL; |
| |
| assert(req_pending(s)); |
| assert(!s->rsp_pending); |
| |
| if (req->type < ARRAY_SIZE(xsd_reqs)) { |
| handler = xsd_reqs[req->type].fn; |
| } |
| if (!handler) { |
| handler = &xs_unimpl; |
| } |
| |
| handler(s, req->req_id, req->tx_id, (uint8_t *)&req[1], req->len); |
| |
| s->rsp_pending = true; |
| reset_req(s); |
| } |
| |
| static unsigned int copy_from_ring(XenXenstoreState *s, uint8_t *ptr, |
| unsigned int len) |
| { |
| if (!len) { |
| return 0; |
| } |
| |
| XENSTORE_RING_IDX prod = qatomic_read(&s->xs->req_prod); |
| XENSTORE_RING_IDX cons = qatomic_read(&s->xs->req_cons); |
| unsigned int copied = 0; |
| |
| /* Ensure the ring contents don't cross the req_prod access. */ |
| smp_rmb(); |
| |
| while (len) { |
| unsigned int avail = prod - cons; |
| unsigned int offset = MASK_XENSTORE_IDX(cons); |
| unsigned int copylen = avail; |
| |
| if (avail > XENSTORE_RING_SIZE) { |
| error_report("XenStore ring handling error"); |
| s->fatal_error = true; |
| break; |
| } else if (avail == 0) { |
| break; |
| } |
| |
| if (copylen > len) { |
| copylen = len; |
| } |
| if (copylen > XENSTORE_RING_SIZE - offset) { |
| copylen = XENSTORE_RING_SIZE - offset; |
| } |
| |
| memcpy(ptr, &s->xs->req[offset], copylen); |
| copied += copylen; |
| |
| ptr += copylen; |
| len -= copylen; |
| |
| cons += copylen; |
| } |
| |
| /* |
| * Not sure this ever mattered except on Alpha, but this barrier |
| * is to ensure that the update to req_cons is globally visible |
| * only after we have consumed all the data from the ring, and we |
| * don't end up seeing data written to the ring *after* the other |
| * end sees the update and writes more to the ring. Xen's own |
| * xenstored has the same barrier here (although with no comment |
| * at all, obviously, because it's Xen code). |
| */ |
| smp_mb(); |
| |
| qatomic_set(&s->xs->req_cons, cons); |
| |
| return copied; |
| } |
| |
| static unsigned int copy_to_ring(XenXenstoreState *s, uint8_t *ptr, |
| unsigned int len) |
| { |
| if (!len) { |
| return 0; |
| } |
| |
| XENSTORE_RING_IDX cons = qatomic_read(&s->xs->rsp_cons); |
| XENSTORE_RING_IDX prod = qatomic_read(&s->xs->rsp_prod); |
| unsigned int copied = 0; |
| |
| /* |
| * This matches the barrier in copy_to_ring() (or the guest's |
| * equivalent) between writing the data to the ring and updating |
| * rsp_prod. It protects against the pathological case (which |
| * again I think never happened except on Alpha) where our |
| * subsequent writes to the ring could *cross* the read of |
| * rsp_cons and the guest could see the new data when it was |
| * intending to read the old. |
| */ |
| smp_mb(); |
| |
| while (len) { |
| unsigned int avail = cons + XENSTORE_RING_SIZE - prod; |
| unsigned int offset = MASK_XENSTORE_IDX(prod); |
| unsigned int copylen = len; |
| |
| if (avail > XENSTORE_RING_SIZE) { |
| error_report("XenStore ring handling error"); |
| s->fatal_error = true; |
| break; |
| } else if (avail == 0) { |
| break; |
| } |
| |
| if (copylen > avail) { |
| copylen = avail; |
| } |
| if (copylen > XENSTORE_RING_SIZE - offset) { |
| copylen = XENSTORE_RING_SIZE - offset; |
| } |
| |
| |
| memcpy(&s->xs->rsp[offset], ptr, copylen); |
| copied += copylen; |
| |
| ptr += copylen; |
| len -= copylen; |
| |
| prod += copylen; |
| } |
| |
| /* Ensure the ring contents are seen before rsp_prod update. */ |
| smp_wmb(); |
| |
| qatomic_set(&s->xs->rsp_prod, prod); |
| |
| return copied; |
| } |
| |
| static unsigned int get_req(XenXenstoreState *s) |
| { |
| unsigned int copied = 0; |
| |
| if (s->fatal_error) { |
| return 0; |
| } |
| |
| assert(!req_pending(s)); |
| |
| if (s->req_offset < XENSTORE_HEADER_SIZE) { |
| void *ptr = s->req_data + s->req_offset; |
| unsigned int len = XENSTORE_HEADER_SIZE; |
| unsigned int copylen = copy_from_ring(s, ptr, len); |
| |
| copied += copylen; |
| s->req_offset += copylen; |
| } |
| |
| if (s->req_offset >= XENSTORE_HEADER_SIZE) { |
| struct xsd_sockmsg *req = (struct xsd_sockmsg *)s->req_data; |
| |
| if (req->len > (uint32_t)XENSTORE_PAYLOAD_MAX) { |
| error_report("Illegal XenStore request"); |
| s->fatal_error = true; |
| return 0; |
| } |
| |
| void *ptr = s->req_data + s->req_offset; |
| unsigned int len = XENSTORE_HEADER_SIZE + req->len - s->req_offset; |
| unsigned int copylen = copy_from_ring(s, ptr, len); |
| |
| copied += copylen; |
| s->req_offset += copylen; |
| } |
| |
| return copied; |
| } |
| |
| static unsigned int put_rsp(XenXenstoreState *s) |
| { |
| if (s->fatal_error) { |
| return 0; |
| } |
| |
| assert(s->rsp_pending); |
| |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| assert(s->rsp_offset < XENSTORE_HEADER_SIZE + rsp->len); |
| |
| void *ptr = s->rsp_data + s->rsp_offset; |
| unsigned int len = XENSTORE_HEADER_SIZE + rsp->len - s->rsp_offset; |
| unsigned int copylen = copy_to_ring(s, ptr, len); |
| |
| s->rsp_offset += copylen; |
| |
| /* Have we produced a complete response? */ |
| if (s->rsp_offset == XENSTORE_HEADER_SIZE + rsp->len) { |
| reset_rsp(s); |
| } |
| |
| return copylen; |
| } |
| |
| static void deliver_watch(XenXenstoreState *s, const char *path, |
| const char *token) |
| { |
| struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data; |
| uint8_t *rsp_data = (uint8_t *)&rsp[1]; |
| unsigned int len; |
| |
| assert(!s->rsp_pending); |
| |
| trace_xenstore_watch_event(path, token); |
| |
| rsp->type = XS_WATCH_EVENT; |
| rsp->req_id = 0; |
| rsp->tx_id = 0; |
| rsp->len = 0; |
| |
| len = strlen(path); |
| |
| /* XENSTORE_ABS/REL_PATH_MAX should ensure there can be no overflow */ |
| assert(rsp->len + len < XENSTORE_PAYLOAD_MAX); |
| |
| memcpy(&rsp_data[rsp->len], path, len); |
| rsp->len += len; |
| rsp_data[rsp->len] = '\0'; |
| rsp->len++; |
| |
| len = strlen(token); |
| /* |
| * It is possible for the guest to have chosen a token that will |
| * not fit (along with the patch) into a watch event. We have no |
| * choice but to drop the event if this is the case. |
| */ |
| if (rsp->len + len >= XENSTORE_PAYLOAD_MAX) { |
| return; |
| } |
| |
| memcpy(&rsp_data[rsp->len], token, len); |
| rsp->len += len; |
| rsp_data[rsp->len] = '\0'; |
| rsp->len++; |
| |
| s->rsp_pending = true; |
| } |
| |
| struct watch_event { |
| char *path; |
| char *token; |
| }; |
| |
| static void free_watch_event(struct watch_event *ev) |
| { |
| if (ev) { |
| g_free(ev->path); |
| g_free(ev->token); |
| g_free(ev); |
| } |
| } |
| |
| static void queue_watch(XenXenstoreState *s, const char *path, |
| const char *token) |
| { |
| struct watch_event *ev = g_new0(struct watch_event, 1); |
| |
| ev->path = g_strdup(path); |
| ev->token = g_strdup(token); |
| |
| s->watch_events = g_list_append(s->watch_events, ev); |
| } |
| |
| static void fire_watch_cb(void *opaque, const char *path, const char *token) |
| { |
| XenXenstoreState *s = opaque; |
| |
| assert(bql_locked()); |
| |
| /* |
| * If there's a response pending, we obviously can't scribble over |
| * it. But if there's a request pending, it has dibs on the buffer |
| * too. |
| * |
| * In the common case of a watch firing due to backend activity |
| * when the ring was otherwise idle, we should be able to copy the |
| * strings directly into the rsp_data and thence the actual ring, |
| * without needing to perform any allocations and queue them. |
| */ |
| if (s->rsp_pending || req_pending(s)) { |
| queue_watch(s, path, token); |
| } else { |
| deliver_watch(s, path, token); |
| /* |
| * Attempt to queue the message into the actual ring, and send |
| * the event channel notification if any bytes are copied. |
| */ |
| if (s->rsp_pending && put_rsp(s) > 0) { |
| xen_be_evtchn_notify(s->eh, s->be_port); |
| } |
| } |
| } |
| |
| static void process_watch_events(XenXenstoreState *s) |
| { |
| struct watch_event *ev = s->watch_events->data; |
| |
| deliver_watch(s, ev->path, ev->token); |
| |
| s->watch_events = g_list_remove(s->watch_events, ev); |
| free_watch_event(ev); |
| } |
| |
| static void xen_xenstore_event(void *opaque) |
| { |
| XenXenstoreState *s = opaque; |
| evtchn_port_t port = xen_be_evtchn_pending(s->eh); |
| unsigned int copied_to, copied_from; |
| bool processed, notify = false; |
| |
| if (port != s->be_port) { |
| return; |
| } |
| |
| /* We know this is a no-op. */ |
| xen_be_evtchn_unmask(s->eh, port); |
| |
| do { |
| copied_to = copied_from = 0; |
| processed = false; |
| |
| if (!s->rsp_pending && s->watch_events) { |
| process_watch_events(s); |
| } |
| |
| if (s->rsp_pending) { |
| copied_to = put_rsp(s); |
| } |
| |
| if (!req_pending(s)) { |
| copied_from = get_req(s); |
| } |
| |
| if (req_pending(s) && !s->rsp_pending && !s->watch_events) { |
| process_req(s); |
| processed = true; |
| } |
| |
| notify |= copied_to || copied_from; |
| } while (copied_to || copied_from || processed); |
| |
| if (notify) { |
| xen_be_evtchn_notify(s->eh, s->be_port); |
| } |
| } |
| |
| static void alloc_guest_port(XenXenstoreState *s) |
| { |
| struct evtchn_alloc_unbound alloc = { |
| .dom = DOMID_SELF, |
| .remote_dom = DOMID_QEMU, |
| }; |
| |
| if (!xen_evtchn_alloc_unbound_op(&alloc)) { |
| s->guest_port = alloc.port; |
| } |
| } |
| |
| int xen_xenstore_reset(void) |
| { |
| XenXenstoreState *s = xen_xenstore_singleton; |
| int console_port; |
| GList *perms; |
| int err; |
| |
| if (!s) { |
| return -ENOTSUP; |
| } |
| |
| s->req_offset = s->rsp_offset = 0; |
| s->rsp_pending = false; |
| |
| if (!memory_region_is_mapped(&s->xenstore_page)) { |
| uint64_t gpa = XEN_SPECIAL_PFN(XENSTORE) << TARGET_PAGE_BITS; |
| xen_overlay_do_map_page(&s->xenstore_page, gpa); |
| } |
| |
| alloc_guest_port(s); |
| |
| /* |
| * As qemu/dom0, bind to the guest's port. For incoming migration, this |
| * will be unbound as the guest's evtchn table is overwritten. We then |
| * rebind to the correct guest port in xen_xenstore_post_load(). |
| */ |
| err = xen_be_evtchn_bind_interdomain(s->eh, xen_domid, s->guest_port); |
| if (err < 0) { |
| return err; |
| } |
| s->be_port = err; |
| |
| /* Create frontend store nodes */ |
| perms = g_list_append(NULL, xs_perm_as_string(XS_PERM_NONE, DOMID_QEMU)); |
| perms = g_list_append(perms, xs_perm_as_string(XS_PERM_READ, xen_domid)); |
| |
| relpath_printf(s, perms, "store/port", "%u", s->guest_port); |
| relpath_printf(s, perms, "store/ring-ref", "%lu", |
| XEN_SPECIAL_PFN(XENSTORE)); |
| |
| console_port = xen_primary_console_get_port(); |
| if (console_port) { |
| relpath_printf(s, perms, "console/ring-ref", "%lu", |
| XEN_SPECIAL_PFN(CONSOLE)); |
| relpath_printf(s, perms, "console/port", "%u", console_port); |
| relpath_printf(s, perms, "console/state", "%u", XenbusStateInitialised); |
| } |
| |
| g_list_free_full(perms, g_free); |
| |
| /* |
| * We don't actually access the guest's page through the grant, because |
| * this isn't real Xen, and we can just use the page we gave it in the |
| * first place. Map the grant anyway, mostly for cosmetic purposes so |
| * it *looks* like it's in use in the guest-visible grant table. |
| */ |
| s->gt = qemu_xen_gnttab_open(); |
| uint32_t xs_gntref = GNTTAB_RESERVED_XENSTORE; |
| s->granted_xs = qemu_xen_gnttab_map_refs(s->gt, 1, xen_domid, &xs_gntref, |
| PROT_READ | PROT_WRITE); |
| |
| return 0; |
| } |
| |
| struct qemu_xs_handle { |
| XenstoreImplState *impl; |
| GList *watches; |
| QEMUBH *watch_bh; |
| }; |
| |
| struct qemu_xs_watch { |
| struct qemu_xs_handle *h; |
| char *path; |
| xs_watch_fn fn; |
| void *opaque; |
| GList *events; |
| }; |
| |
| static char *xs_be_get_domain_path(struct qemu_xs_handle *h, unsigned int domid) |
| { |
| return g_strdup_printf("/local/domain/%u", domid); |
| } |
| |
| static char **xs_be_directory(struct qemu_xs_handle *h, xs_transaction_t t, |
| const char *path, unsigned int *num) |
| { |
| GList *items = NULL, *l; |
| unsigned int i = 0; |
| char **items_ret; |
| int err; |
| |
| err = xs_impl_directory(h->impl, DOMID_QEMU, t, path, NULL, &items); |
| if (err) { |
| errno = err; |
| return NULL; |
| } |
| |
| items_ret = g_new0(char *, g_list_length(items) + 1); |
| *num = 0; |
| for (l = items; l; l = l->next) { |
| items_ret[i++] = l->data; |
| (*num)++; |
| } |
| g_list_free(items); |
| return items_ret; |
| } |
| |
| static void *xs_be_read(struct qemu_xs_handle *h, xs_transaction_t t, |
| const char *path, unsigned int *len) |
| { |
| GByteArray *data = g_byte_array_new(); |
| bool free_segment = false; |
| int err; |
| |
| err = xs_impl_read(h->impl, DOMID_QEMU, t, path, data); |
| if (err) { |
| free_segment = true; |
| errno = err; |
| } else { |
| if (len) { |
| *len = data->len; |
| } |
| /* The xen-bus-helper code expects to get NUL terminated string! */ |
| g_byte_array_append(data, (void *)"", 1); |
| } |
| |
| return g_byte_array_free(data, free_segment); |
| } |
| |
| static bool xs_be_write(struct qemu_xs_handle *h, xs_transaction_t t, |
| const char *path, const void *data, unsigned int len) |
| { |
| GByteArray *gdata = g_byte_array_new(); |
| int err; |
| |
| g_byte_array_append(gdata, data, len); |
| err = xs_impl_write(h->impl, DOMID_QEMU, t, path, gdata); |
| g_byte_array_unref(gdata); |
| if (err) { |
| errno = err; |
| return false; |
| } |
| return true; |
| } |
| |
| static bool xs_be_create(struct qemu_xs_handle *h, xs_transaction_t t, |
| unsigned int owner, unsigned int domid, |
| unsigned int perms, const char *path) |
| { |
| g_autoptr(GByteArray) data = g_byte_array_new(); |
| GList *perms_list = NULL; |
| int err; |
| |
| /* mkdir does this */ |
| err = xs_impl_read(h->impl, DOMID_QEMU, t, path, data); |
| if (err == ENOENT) { |
| err = xs_impl_write(h->impl, DOMID_QEMU, t, path, data); |
| } |
| if (err) { |
| errno = err; |
| return false; |
| } |
| |
| perms_list = g_list_append(perms_list, |
| xs_perm_as_string(XS_PERM_NONE, owner)); |
| perms_list = g_list_append(perms_list, |
| xs_perm_as_string(perms, domid)); |
| |
| err = xs_impl_set_perms(h->impl, DOMID_QEMU, t, path, perms_list); |
| g_list_free_full(perms_list, g_free); |
| if (err) { |
| errno = err; |
| return false; |
| } |
| return true; |
| } |
| |
| static bool xs_be_destroy(struct qemu_xs_handle *h, xs_transaction_t t, |
| const char *path) |
| { |
| int err = xs_impl_rm(h->impl, DOMID_QEMU, t, path); |
| if (err) { |
| errno = err; |
| return false; |
| } |
| return true; |
| } |
| |
| static void be_watch_bh(void *_h) |
| { |
| struct qemu_xs_handle *h = _h; |
| GList *l; |
| |
| for (l = h->watches; l; l = l->next) { |
| struct qemu_xs_watch *w = l->data; |
| |
| while (w->events) { |
| struct watch_event *ev = w->events->data; |
| |
| w->fn(w->opaque, ev->path); |
| |
| w->events = g_list_remove(w->events, ev); |
| free_watch_event(ev); |
| } |
| } |
| } |
| |
| static void xs_be_watch_cb(void *opaque, const char *path, const char *token) |
| { |
| struct watch_event *ev = g_new0(struct watch_event, 1); |
| struct qemu_xs_watch *w = opaque; |
| |
| /* We don't care about the token */ |
| ev->path = g_strdup(path); |
| w->events = g_list_append(w->events, ev); |
| |
| qemu_bh_schedule(w->h->watch_bh); |
| } |
| |
| static struct qemu_xs_watch *xs_be_watch(struct qemu_xs_handle *h, |
| const char *path, xs_watch_fn fn, |
| void *opaque) |
| { |
| struct qemu_xs_watch *w = g_new0(struct qemu_xs_watch, 1); |
| int err; |
| |
| w->h = h; |
| w->fn = fn; |
| w->opaque = opaque; |
| |
| err = xs_impl_watch(h->impl, DOMID_QEMU, path, NULL, xs_be_watch_cb, w); |
| if (err) { |
| errno = err; |
| g_free(w); |
| return NULL; |
| } |
| |
| w->path = g_strdup(path); |
| h->watches = g_list_append(h->watches, w); |
| return w; |
| } |
| |
| static void xs_be_unwatch(struct qemu_xs_handle *h, struct qemu_xs_watch *w) |
| { |
| xs_impl_unwatch(h->impl, DOMID_QEMU, w->path, NULL, xs_be_watch_cb, w); |
| |
| h->watches = g_list_remove(h->watches, w); |
| g_list_free_full(w->events, (GDestroyNotify)free_watch_event); |
| g_free(w->path); |
| g_free(w); |
| } |
| |
| static xs_transaction_t xs_be_transaction_start(struct qemu_xs_handle *h) |
| { |
| unsigned int new_tx = XBT_NULL; |
| int err = xs_impl_transaction_start(h->impl, DOMID_QEMU, &new_tx); |
| if (err) { |
| errno = err; |
| return XBT_NULL; |
| } |
| return new_tx; |
| } |
| |
| static bool xs_be_transaction_end(struct qemu_xs_handle *h, xs_transaction_t t, |
| bool abort) |
| { |
| int err = xs_impl_transaction_end(h->impl, DOMID_QEMU, t, !abort); |
| if (err) { |
| errno = err; |
| return false; |
| } |
| return true; |
| } |
| |
| static struct qemu_xs_handle *xs_be_open(void) |
| { |
| XenXenstoreState *s = xen_xenstore_singleton; |
| struct qemu_xs_handle *h; |
| |
| if (!s || !s->impl) { |
| errno = -ENOSYS; |
| return NULL; |
| } |
| |
| h = g_new0(struct qemu_xs_handle, 1); |
| h->impl = s->impl; |
| |
| h->watch_bh = aio_bh_new(qemu_get_aio_context(), be_watch_bh, h); |
| |
| return h; |
| } |
| |
| static void xs_be_close(struct qemu_xs_handle *h) |
| { |
| while (h->watches) { |
| struct qemu_xs_watch *w = h->watches->data; |
| xs_be_unwatch(h, w); |
| } |
| |
| qemu_bh_delete(h->watch_bh); |
| g_free(h); |
| } |
| |
| static struct xenstore_backend_ops emu_xenstore_backend_ops = { |
| .open = xs_be_open, |
| .close = xs_be_close, |
| .get_domain_path = xs_be_get_domain_path, |
| .directory = xs_be_directory, |
| .read = xs_be_read, |
| .write = xs_be_write, |
| .create = xs_be_create, |
| .destroy = xs_be_destroy, |
| .watch = xs_be_watch, |
| .unwatch = xs_be_unwatch, |
| .transaction_start = xs_be_transaction_start, |
| .transaction_end = xs_be_transaction_end, |
| }; |