| /* |
| * QEMU paravirtual RDMA - Generic RDMA backend |
| * |
| * Copyright (C) 2018 Oracle |
| * Copyright (C) 2018 Red Hat Inc |
| * |
| * Authors: |
| * Yuval Shaia <yuval.shaia@oracle.com> |
| * Marcel Apfelbaum <marcel@redhat.com> |
| * |
| * This work is licensed under the terms of the GNU GPL, version 2 or later. |
| * See the COPYING file in the top-level directory. |
| * |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "trace.h" |
| #include "rdma_utils.h" |
| |
| void *rdma_pci_dma_map(PCIDevice *dev, dma_addr_t addr, dma_addr_t len) |
| { |
| void *p; |
| dma_addr_t pci_len = len; |
| |
| if (!addr) { |
| rdma_error_report("addr is NULL"); |
| return NULL; |
| } |
| |
| p = pci_dma_map(dev, addr, &pci_len, DMA_DIRECTION_TO_DEVICE); |
| if (!p) { |
| rdma_error_report("pci_dma_map fail, addr=0x%"PRIx64", len=%"PRId64, |
| addr, pci_len); |
| return NULL; |
| } |
| |
| if (pci_len != len) { |
| rdma_pci_dma_unmap(dev, p, pci_len); |
| return NULL; |
| } |
| |
| trace_rdma_pci_dma_map(addr, p, pci_len); |
| |
| return p; |
| } |
| |
| void rdma_pci_dma_unmap(PCIDevice *dev, void *buffer, dma_addr_t len) |
| { |
| trace_rdma_pci_dma_unmap(buffer); |
| if (buffer) { |
| pci_dma_unmap(dev, buffer, len, DMA_DIRECTION_TO_DEVICE, 0); |
| } |
| } |
| |
| void rdma_protected_gqueue_init(RdmaProtectedGQueue *list) |
| { |
| qemu_mutex_init(&list->lock); |
| list->list = g_queue_new(); |
| } |
| |
| void rdma_protected_gqueue_destroy(RdmaProtectedGQueue *list) |
| { |
| if (list->list) { |
| g_queue_free_full(list->list, g_free); |
| qemu_mutex_destroy(&list->lock); |
| list->list = NULL; |
| } |
| } |
| |
| void rdma_protected_gqueue_append_int64(RdmaProtectedGQueue *list, |
| int64_t value) |
| { |
| qemu_mutex_lock(&list->lock); |
| g_queue_push_tail(list->list, g_memdup(&value, sizeof(value))); |
| qemu_mutex_unlock(&list->lock); |
| } |
| |
| int64_t rdma_protected_gqueue_pop_int64(RdmaProtectedGQueue *list) |
| { |
| int64_t *valp; |
| int64_t val; |
| |
| qemu_mutex_lock(&list->lock); |
| |
| valp = g_queue_pop_head(list->list); |
| qemu_mutex_unlock(&list->lock); |
| |
| if (!valp) { |
| return -ENOENT; |
| } |
| |
| val = *valp; |
| g_free(valp); |
| return val; |
| } |
| |
| void rdma_protected_gslist_init(RdmaProtectedGSList *list) |
| { |
| qemu_mutex_init(&list->lock); |
| } |
| |
| void rdma_protected_gslist_destroy(RdmaProtectedGSList *list) |
| { |
| if (list->list) { |
| g_slist_free(list->list); |
| qemu_mutex_destroy(&list->lock); |
| list->list = NULL; |
| } |
| } |
| |
| void rdma_protected_gslist_append_int32(RdmaProtectedGSList *list, |
| int32_t value) |
| { |
| qemu_mutex_lock(&list->lock); |
| list->list = g_slist_prepend(list->list, GINT_TO_POINTER(value)); |
| qemu_mutex_unlock(&list->lock); |
| } |
| |
| void rdma_protected_gslist_remove_int32(RdmaProtectedGSList *list, |
| int32_t value) |
| { |
| qemu_mutex_lock(&list->lock); |
| list->list = g_slist_remove(list->list, GINT_TO_POINTER(value)); |
| qemu_mutex_unlock(&list->lock); |
| } |