xref: /openbmc/qemu/net/queue.c (revision 8d3031fa)
1 /*
2  * Copyright (c) 2003-2008 Fabrice Bellard
3  * Copyright (c) 2009 Red Hat, Inc.
4  *
5  * Permission is hereby granted, free of charge, to any person obtaining a copy
6  * of this software and associated documentation files (the "Software"), to deal
7  * in the Software without restriction, including without limitation the rights
8  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9  * copies of the Software, and to permit persons to whom the Software is
10  * furnished to do so, subject to the following conditions:
11  *
12  * The above copyright notice and this permission notice shall be included in
13  * all copies or substantial portions of the Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
21  * THE SOFTWARE.
22  */
23 
24 #include "qemu/osdep.h"
25 #include "net/queue.h"
26 #include "qemu/queue.h"
27 #include "net/net.h"
28 
29 /* The delivery handler may only return zero if it will call
30  * qemu_net_queue_flush() when it determines that it is once again able
31  * to deliver packets. It must also call qemu_net_queue_purge() in its
32  * cleanup path.
33  *
34  * If a sent callback is provided to send(), the caller must handle a
35  * zero return from the delivery handler by not sending any more packets
36  * until we have invoked the callback. Only in that case will we queue
37  * the packet.
38  *
39  * If a sent callback isn't provided, we just drop the packet to avoid
40  * unbounded queueing.
41  */
42 
43 struct NetPacket {
44     QTAILQ_ENTRY(NetPacket) entry;
45     NetClientState *sender;
46     unsigned flags;
47     int size;
48     NetPacketSent *sent_cb;
49     uint8_t data[];
50 };
51 
52 struct NetQueue {
53     void *opaque;
54     uint32_t nq_maxlen;
55     uint32_t nq_count;
56     NetQueueDeliverFunc *deliver;
57 
58     QTAILQ_HEAD(, NetPacket) packets;
59 
60     unsigned delivering : 1;
61 };
62 
63 NetQueue *qemu_new_net_queue(NetQueueDeliverFunc *deliver, void *opaque)
64 {
65     NetQueue *queue;
66 
67     queue = g_new0(NetQueue, 1);
68 
69     queue->opaque = opaque;
70     queue->nq_maxlen = 10000;
71     queue->nq_count = 0;
72     queue->deliver = deliver;
73 
74     QTAILQ_INIT(&queue->packets);
75 
76     queue->delivering = 0;
77 
78     return queue;
79 }
80 
81 void qemu_del_net_queue(NetQueue *queue)
82 {
83     NetPacket *packet, *next;
84 
85     QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) {
86         QTAILQ_REMOVE(&queue->packets, packet, entry);
87         g_free(packet);
88     }
89 
90     g_free(queue);
91 }
92 
93 static void qemu_net_queue_append(NetQueue *queue,
94                                   NetClientState *sender,
95                                   unsigned flags,
96                                   const uint8_t *buf,
97                                   size_t size,
98                                   NetPacketSent *sent_cb)
99 {
100     NetPacket *packet;
101 
102     if (queue->nq_count >= queue->nq_maxlen && !sent_cb) {
103         return; /* drop if queue full and no callback */
104     }
105     packet = g_malloc(sizeof(NetPacket) + size);
106     packet->sender = sender;
107     packet->flags = flags;
108     packet->size = size;
109     packet->sent_cb = sent_cb;
110     memcpy(packet->data, buf, size);
111 
112     queue->nq_count++;
113     QTAILQ_INSERT_TAIL(&queue->packets, packet, entry);
114 }
115 
116 void qemu_net_queue_append_iov(NetQueue *queue,
117                                NetClientState *sender,
118                                unsigned flags,
119                                const struct iovec *iov,
120                                int iovcnt,
121                                NetPacketSent *sent_cb)
122 {
123     NetPacket *packet;
124     size_t max_len = 0;
125     int i;
126 
127     if (queue->nq_count >= queue->nq_maxlen && !sent_cb) {
128         return; /* drop if queue full and no callback */
129     }
130     for (i = 0; i < iovcnt; i++) {
131         max_len += iov[i].iov_len;
132     }
133 
134     packet = g_malloc(sizeof(NetPacket) + max_len);
135     packet->sender = sender;
136     packet->sent_cb = sent_cb;
137     packet->flags = flags;
138     packet->size = 0;
139 
140     for (i = 0; i < iovcnt; i++) {
141         size_t len = iov[i].iov_len;
142 
143         memcpy(packet->data + packet->size, iov[i].iov_base, len);
144         packet->size += len;
145     }
146 
147     queue->nq_count++;
148     QTAILQ_INSERT_TAIL(&queue->packets, packet, entry);
149 }
150 
151 static ssize_t qemu_net_queue_deliver(NetQueue *queue,
152                                       NetClientState *sender,
153                                       unsigned flags,
154                                       const uint8_t *data,
155                                       size_t size)
156 {
157     ssize_t ret = -1;
158     struct iovec iov = {
159         .iov_base = (void *)data,
160         .iov_len = size
161     };
162 
163     queue->delivering = 1;
164     ret = queue->deliver(sender, flags, &iov, 1, queue->opaque);
165     queue->delivering = 0;
166 
167     return ret;
168 }
169 
170 static ssize_t qemu_net_queue_deliver_iov(NetQueue *queue,
171                                           NetClientState *sender,
172                                           unsigned flags,
173                                           const struct iovec *iov,
174                                           int iovcnt)
175 {
176     ssize_t ret = -1;
177 
178     queue->delivering = 1;
179     ret = queue->deliver(sender, flags, iov, iovcnt, queue->opaque);
180     queue->delivering = 0;
181 
182     return ret;
183 }
184 
185 ssize_t qemu_net_queue_receive(NetQueue *queue,
186                                const uint8_t *data,
187                                size_t size)
188 {
189     if (queue->delivering) {
190         return 0;
191     }
192 
193     return qemu_net_queue_deliver(queue, NULL, 0, data, size);
194 }
195 
196 ssize_t qemu_net_queue_send(NetQueue *queue,
197                             NetClientState *sender,
198                             unsigned flags,
199                             const uint8_t *data,
200                             size_t size,
201                             NetPacketSent *sent_cb)
202 {
203     ssize_t ret;
204 
205     if (queue->delivering || !qemu_can_send_packet(sender)) {
206         qemu_net_queue_append(queue, sender, flags, data, size, sent_cb);
207         return 0;
208     }
209 
210     ret = qemu_net_queue_deliver(queue, sender, flags, data, size);
211     if (ret == 0) {
212         qemu_net_queue_append(queue, sender, flags, data, size, sent_cb);
213         return 0;
214     }
215 
216     qemu_net_queue_flush(queue);
217 
218     return ret;
219 }
220 
221 ssize_t qemu_net_queue_send_iov(NetQueue *queue,
222                                 NetClientState *sender,
223                                 unsigned flags,
224                                 const struct iovec *iov,
225                                 int iovcnt,
226                                 NetPacketSent *sent_cb)
227 {
228     ssize_t ret;
229 
230     if (queue->delivering || !qemu_can_send_packet(sender)) {
231         qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, sent_cb);
232         return 0;
233     }
234 
235     ret = qemu_net_queue_deliver_iov(queue, sender, flags, iov, iovcnt);
236     if (ret == 0) {
237         qemu_net_queue_append_iov(queue, sender, flags, iov, iovcnt, sent_cb);
238         return 0;
239     }
240 
241     qemu_net_queue_flush(queue);
242 
243     return ret;
244 }
245 
246 void qemu_net_queue_purge(NetQueue *queue, NetClientState *from)
247 {
248     NetPacket *packet, *next;
249 
250     QTAILQ_FOREACH_SAFE(packet, &queue->packets, entry, next) {
251         if (packet->sender == from) {
252             QTAILQ_REMOVE(&queue->packets, packet, entry);
253             queue->nq_count--;
254             if (packet->sent_cb) {
255                 packet->sent_cb(packet->sender, 0);
256             }
257             g_free(packet);
258         }
259     }
260 }
261 
262 bool qemu_net_queue_flush(NetQueue *queue)
263 {
264     if (queue->delivering)
265         return false;
266 
267     while (!QTAILQ_EMPTY(&queue->packets)) {
268         NetPacket *packet;
269         int ret;
270 
271         packet = QTAILQ_FIRST(&queue->packets);
272         QTAILQ_REMOVE(&queue->packets, packet, entry);
273         queue->nq_count--;
274 
275         ret = qemu_net_queue_deliver(queue,
276                                      packet->sender,
277                                      packet->flags,
278                                      packet->data,
279                                      packet->size);
280         if (ret == 0) {
281             queue->nq_count++;
282             QTAILQ_INSERT_HEAD(&queue->packets, packet, entry);
283             return false;
284         }
285 
286         if (packet->sent_cb) {
287             packet->sent_cb(packet->sender, ret);
288         }
289 
290         g_free(packet);
291     }
292     return true;
293 }
294