| /* |
| * Copyright (c) 2003-2008 Fabrice Bellard |
| * Copyright (c) 2009 Red Hat, Inc. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a copy |
| * of this software and associated documentation files (the "Software"), to deal |
| * in the Software without restriction, including without limitation the rights |
| * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
| * copies of the Software, and to permit persons to whom the Software is |
| * furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in |
| * all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
| * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN |
| * THE SOFTWARE. |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "net/queue.h" |
| #include "qemu/queue.h" |
| #include "net/net.h" |
| |
| /* The delivery handler may only return zero if it will call |
| * qemu_net_queue_flush() when it determines that it is once again able |
| * to deliver packets. It must also call qemu_net_queue_purge() in its |
| * cleanup path. |
| * |
| * If a sent callback is provided to send(), the caller must handle a |
| * zero return from the delivery handler by not sending any more packets |
| * until we have invoked the callback. Only in that case will we queue |
| * the packet. |
| * |
| * If a sent callback isn't provided, we just drop the packet to avoid |
| * unbounded queueing. |
| */ |
| |
| struct NetPacket { |
| QTAILQ_ENTRY(NetPacket) entry; |
| NetClientState *sender; |
| unsigned flags; |
| int size; |
| NetPacketSent *sent_cb; |
| uint8_t data[]; |
| }; |
| |
| struct NetQueue { |
| void *opaque; |
| uint32_t nq_maxlen; |
| uint32_t nq_count; |
| NetQueueDeliverFunc *deliver; |
| |
| QTAILQ_HEAD(, NetPacket) packets; |
| |
| unsigned delivering : 1; |
| }; |
| |
| NetQueue *qemu_new_net_queue(NetQueueDeliverFunc *deliver, void *opaque) |
| { |
| NetQueue *queue; |
| |
| queue = g_new0(NetQueue, 1); |
| |
| queue->opaque = opaque; |
| queue->nq_maxlen = 10000; |
| queue->nq_count = 0; |
| queue->deliver = deliver; |
| |
| QTAILQ_INIT(&queue->packets); |
| |
| queue->delivering = 0; |
| |
| return queue; |
| } |
| |
| void qemu_del_net_queue(NetQueue *queue) |
| { |
| NetPacket *packet, *next; |
| |
| QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) { |
| QTAILQ_REMOVE(&queue->packets, packet, entry); |
| g_free(packet); |
| } |
| |
| g_free(queue); |
| } |
| |
| static void qemu_net_queue_append(NetQueue *queue, |
| NetClientState *sender, |
| unsigned flags, |
| const uint8_t *buf, |
| size_t size, |
| NetPacketSent *sent_cb) |
| { |
| NetPacket *packet; |
| |
| if (queue->nq_count >= queue->nq_maxlen && !sent_cb) { |
| return; /* drop if queue full and no callback */ |
| } |
| packet = g_malloc(sizeof(NetPacket) + size); |
| packet->sender = sender; |
| packet->flags = flags; |
| packet->size = size; |
| packet->sent_cb = sent_cb; |
| memcpy(packet->data, buf, size); |
| |
| queue->nq_count++; |
| QTAILQ_INSERT_TAIL(&queue->packets, packet, entry); |
| } |
| |
| void qemu_net_queue_append_iov(NetQueue *queue, |
| NetClientState *sender, |
| unsigned flags, |
| const struct iovec *iov, |
| int iovcnt, |
| NetPacketSent *sent_cb) |
| { |
| NetPacket *packet; |
| size_t max_len = 0; |
| int i; |
| |
| if (queue->nq_count >= queue->nq_maxlen && !sent_cb) { |
| return; /* drop if queue full and no callback */ |
| } |
| for (i = 0; i < iovcnt; i++) { |
| max_len += iov[i].iov_len; |
| } |
| |
| packet = g_malloc(sizeof(NetPacket) + max_len); |
| packet->sender = sender; |
| packet->sent_cb = sent_cb; |
| packet->flags = flags; |
| packet->size = 0; |
| |
| for (i = 0; i < iovcnt; i++) { |
| size_t len = iov[i].iov_len; |
| |
| memcpy(packet->data + packet->size, iov[i].iov_base, len); |
| packet->size += len; |
| } |
| |
| queue->nq_count++; |
| QTAILQ_INSERT_TAIL(&queue->packets, packet, entry); |
| } |
| |
| static ssize_t qemu_net_queue_deliver(NetQueue *queue, |
| NetClientState *sender, |
| unsigned flags, |
| const uint8_t *data, |
| size_t size) |
| { |
| ssize_t ret = -1; |
| struct iovec iov = { |
| .iov_base = (void *)data, |
| .iov_len = size |
| }; |
| |
| queue->delivering = 1; |
| ret = queue->deliver(sender, flags, &iov, 1, queue->opaque); |
| queue->delivering = 0; |
| |
| return ret; |
| } |
| |
| static ssize_t qemu_net_queue_deliver_iov(NetQueue *queue, |
| NetClientState *sender, |
| unsigned flags, |
| const struct iovec *iov, |
| int iovcnt) |
| { |
| ssize_t ret = -1; |
| |
| queue->delivering = 1; |
| ret = queue->deliver(sender, flags, iov, iovcnt, queue->opaque); |
| queue->delivering = 0; |
| |
| return ret; |
| } |
| |
| ssize_t qemu_net_queue_receive(NetQueue *queue, |
| const uint8_t *data, |
| size_t size) |
| { |
| if (queue->delivering) { |
| return 0; |
| } |
| |
| return qemu_net_queue_deliver(queue, NULL, 0, data, size); |
| } |
| |
| ssize_t qemu_net_queue_receive_iov(NetQueue *queue, |
| const struct iovec *iov, |
| int iovcnt) |
| { |
| if (queue->delivering) { |
| return 0; |
| } |
| |
| return qemu_net_queue_deliver_iov(queue, NULL, 0, iov, iovcnt); |
| } |
| |
| ssize_t qemu_net_queue_send(NetQueue *queue, |
| NetClientState *sender, |
| unsigned flags, |
| const uint8_t *data, |
| size_t size, |
| NetPacketSent *sent_cb) |
| { |
| ssize_t ret; |
| |
| if (queue->delivering || !qemu_can_send_packet(sender)) { |
| qemu_net_queue_append(queue, sender, flags, data, size, sent_cb); |
| return 0; |
| } |
| |
| ret = qemu_net_queue_deliver(queue, sender, flags, data, size); |
| if (ret == 0) { |
| qemu_net_queue_append(queue, sender, flags, data, size, sent_cb); |
| return 0; |
| } |
| |
| qemu_net_queue_flush(queue); |
| |
| return ret; |
| } |
| |
| ssize_t qemu_net_queue_send_iov(NetQueue *queue, |
| NetClientState *sender, |
| unsigned flags, |
| const struct iovec *iov, |
| int iovcnt, |
| NetPacketSent *sent_cb) |
| { |
| ssize_t ret; |
| |
| if (queue->delivering || !qemu_can_send_packet(sender)) { |
| qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, sent_cb); |
| return 0; |
| } |
| |
| ret = qemu_net_queue_deliver_iov(queue, sender, flags, iov, iovcnt); |
| if (ret == 0) { |
| qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, sent_cb); |
| return 0; |
| } |
| |
| qemu_net_queue_flush(queue); |
| |
| return ret; |
| } |
| |
| void qemu_net_queue_purge(NetQueue *queue, NetClientState *from) |
| { |
| NetPacket *packet, *next; |
| |
| QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) { |
| if (packet->sender == from) { |
| QTAILQ_REMOVE(&queue->packets, packet, entry); |
| queue->nq_count--; |
| if (packet->sent_cb) { |
| packet->sent_cb(packet->sender, 0); |
| } |
| g_free(packet); |
| } |
| } |
| } |
| |
| bool qemu_net_queue_flush(NetQueue *queue) |
| { |
| if (queue->delivering) |
| return false; |
| |
| while (!QTAILQ_EMPTY(&queue->packets)) { |
| NetPacket *packet; |
| int ret; |
| |
| packet = QTAILQ_FIRST(&queue->packets); |
| QTAILQ_REMOVE(&queue->packets, packet, entry); |
| queue->nq_count--; |
| |
| ret = qemu_net_queue_deliver(queue, |
| packet->sender, |
| packet->flags, |
| packet->data, |
| packet->size); |
| if (ret == 0) { |
| queue->nq_count++; |
| QTAILQ_INSERT_HEAD(&queue->packets, packet, entry); |
| return false; |
| } |
| |
| if (packet->sent_cb) { |
| packet->sent_cb(packet->sender, ret); |
| } |
| |
| g_free(packet); |
| } |
| return true; |
| } |