xref: /openbmc/qemu/hw/net/spapr_llan.c (revision 2b74dd91)
1 /*
2  * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
3  *
4  * PAPR Inter-VM Logical Lan, aka ibmveth
5  *
6  * Copyright (c) 2010,2011 David Gibson, IBM Corporation.
7  *
8  * Permission is hereby granted, free of charge, to any person obtaining a copy
9  * of this software and associated documentation files (the "Software"), to deal
10  * in the Software without restriction, including without limitation the rights
11  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12  * copies of the Software, and to permit persons to whom the Software is
13  * furnished to do so, subject to the following conditions:
14  *
15  * The above copyright notice and this permission notice shall be included in
16  * all copies or substantial portions of the Software.
17  *
18  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24  * THE SOFTWARE.
25  *
26  */
27 
28 #include "qemu/osdep.h"
29 #include "qemu/log.h"
30 #include "qemu/module.h"
31 #include "net/net.h"
32 #include "migration/vmstate.h"
33 #include "hw/ppc/spapr.h"
34 #include "hw/ppc/spapr_vio.h"
35 #include "hw/qdev-properties.h"
36 #include "sysemu/sysemu.h"
37 #include "trace.h"
38 
39 #include <libfdt.h>
40 #include "qom/object.h"
41 
42 #define ETH_ALEN        6
43 #define MAX_PACKET_SIZE 65536
44 
45 /* Compatibility flags for migration */
46 #define SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT  0
47 #define SPAPRVLAN_FLAG_RX_BUF_POOLS      (1 << SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT)
48 
49 /*
50  * Virtual LAN device
51  */
52 
53 typedef uint64_t vlan_bd_t;
54 
55 #define VLAN_BD_VALID        0x8000000000000000ULL
56 #define VLAN_BD_TOGGLE       0x4000000000000000ULL
57 #define VLAN_BD_NO_CSUM      0x0200000000000000ULL
58 #define VLAN_BD_CSUM_GOOD    0x0100000000000000ULL
59 #define VLAN_BD_LEN_MASK     0x00ffffff00000000ULL
60 #define VLAN_BD_LEN(bd)      (((bd) & VLAN_BD_LEN_MASK) >> 32)
61 #define VLAN_BD_ADDR_MASK    0x00000000ffffffffULL
62 #define VLAN_BD_ADDR(bd)     ((bd) & VLAN_BD_ADDR_MASK)
63 
64 #define VLAN_VALID_BD(addr, len) (VLAN_BD_VALID | \
65                                   (((len) << 32) & VLAN_BD_LEN_MASK) |  \
66                                   (addr & VLAN_BD_ADDR_MASK))
67 
68 #define VLAN_RXQC_TOGGLE     0x80
69 #define VLAN_RXQC_VALID      0x40
70 #define VLAN_RXQC_NO_CSUM    0x02
71 #define VLAN_RXQC_CSUM_GOOD  0x01
72 
73 #define VLAN_RQ_ALIGNMENT    16
74 #define VLAN_RXQ_BD_OFF      0
75 #define VLAN_FILTER_BD_OFF   8
76 #define VLAN_RX_BDS_OFF      16
77 /*
78  * The final 8 bytes of the buffer list is a counter of frames dropped
79  * because there was not a buffer in the buffer list capable of holding
80  * the frame. We must avoid it, or the operating system will report garbage
81  * for this statistic.
82  */
83 #define VLAN_RX_BDS_LEN      (SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF - 8)
84 #define VLAN_MAX_BUFS        (VLAN_RX_BDS_LEN / 8)
85 
86 #define TYPE_VIO_SPAPR_VLAN_DEVICE "spapr-vlan"
87 OBJECT_DECLARE_SIMPLE_TYPE(SpaprVioVlan, VIO_SPAPR_VLAN_DEVICE)
88 
89 #define RX_POOL_MAX_BDS 4096
90 #define RX_MAX_POOLS 5
91 
92 typedef struct {
93     int32_t bufsize;
94     int32_t count;
95     vlan_bd_t bds[RX_POOL_MAX_BDS];
96 } RxBufPool;
97 
98 struct SpaprVioVlan {
99     SpaprVioDevice sdev;
100     NICConf nicconf;
101     NICState *nic;
102     MACAddr perm_mac;
103     bool isopen;
104     hwaddr buf_list;
105     uint32_t add_buf_ptr, use_buf_ptr, rx_bufs;
106     hwaddr rxq_ptr;
107     QEMUTimer *rxp_timer;
108     uint32_t compat_flags;             /* Compatibility flags for migration */
109     RxBufPool *rx_pool[RX_MAX_POOLS];  /* Receive buffer descriptor pools */
110 };
111 
112 static bool spapr_vlan_can_receive(NetClientState *nc)
113 {
114     SpaprVioVlan *dev = qemu_get_nic_opaque(nc);
115 
116     return dev->isopen && dev->rx_bufs > 0;
117 }
118 
119 /**
120  * The last 8 bytes of the receive buffer list page (that has been
121  * supplied by the guest with the H_REGISTER_LOGICAL_LAN call) contain
122  * a counter for frames that have been dropped because there was no
123  * suitable receive buffer available. This function is used to increase
124  * this counter by one.
125  */
126 static void spapr_vlan_record_dropped_rx_frame(SpaprVioVlan *dev)
127 {
128     uint64_t cnt;
129 
130     cnt = vio_ldq(&dev->sdev, dev->buf_list + 4096 - 8);
131     vio_stq(&dev->sdev, dev->buf_list + 4096 - 8, cnt + 1);
132 }
133 
134 /**
135  * Get buffer descriptor from one of our receive buffer pools
136  */
137 static vlan_bd_t spapr_vlan_get_rx_bd_from_pool(SpaprVioVlan *dev,
138                                                 size_t size)
139 {
140     vlan_bd_t bd;
141     int pool;
142 
143     for (pool = 0; pool < RX_MAX_POOLS; pool++) {
144         if (dev->rx_pool[pool]->count > 0 &&
145             dev->rx_pool[pool]->bufsize >= size + 8) {
146             break;
147         }
148     }
149     if (pool == RX_MAX_POOLS) {
150         /* Failed to find a suitable buffer */
151         return 0;
152     }
153 
154 
155     trace_spapr_vlan_get_rx_bd_from_pool_found(pool,
156                                                dev->rx_pool[pool]->count,
157                                                dev->rx_bufs);
158 
159     /* Remove the buffer from the pool */
160     dev->rx_pool[pool]->count--;
161     bd = dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count];
162     dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count] = 0;
163 
164     return bd;
165 }
166 
167 /**
168  * Get buffer descriptor from the receive buffer list page that has been
169  * supplied by the guest with the H_REGISTER_LOGICAL_LAN call
170  */
171 static vlan_bd_t spapr_vlan_get_rx_bd_from_page(SpaprVioVlan *dev,
172                                                 size_t size)
173 {
174     int buf_ptr = dev->use_buf_ptr;
175     vlan_bd_t bd;
176 
177     do {
178         buf_ptr += 8;
179         if (buf_ptr >= VLAN_RX_BDS_LEN + VLAN_RX_BDS_OFF) {
180             buf_ptr = VLAN_RX_BDS_OFF;
181         }
182 
183         bd = vio_ldq(&dev->sdev, dev->buf_list + buf_ptr);
184 
185         trace_spapr_vlan_get_rx_bd_from_page(buf_ptr, (uint64_t)bd);
186     } while ((!(bd & VLAN_BD_VALID) || VLAN_BD_LEN(bd) < size + 8)
187              && buf_ptr != dev->use_buf_ptr);
188 
189     if (!(bd & VLAN_BD_VALID) || VLAN_BD_LEN(bd) < size + 8) {
190         /* Failed to find a suitable buffer */
191         return 0;
192     }
193 
194     /* Remove the buffer from the pool */
195     dev->use_buf_ptr = buf_ptr;
196     vio_stq(&dev->sdev, dev->buf_list + dev->use_buf_ptr, 0);
197 
198     trace_spapr_vlan_get_rx_bd_from_page_found(dev->use_buf_ptr, dev->rx_bufs);
199 
200     return bd;
201 }
202 
203 static ssize_t spapr_vlan_receive(NetClientState *nc, const uint8_t *buf,
204                                   size_t size)
205 {
206     SpaprVioVlan *dev = qemu_get_nic_opaque(nc);
207     SpaprVioDevice *sdev = VIO_SPAPR_DEVICE(dev);
208     vlan_bd_t rxq_bd = vio_ldq(sdev, dev->buf_list + VLAN_RXQ_BD_OFF);
209     vlan_bd_t bd;
210     uint64_t handle;
211     uint8_t control;
212 
213     trace_spapr_vlan_receive(sdev->qdev.id, dev->rx_bufs);
214 
215     if (!dev->isopen) {
216         return -1;
217     }
218 
219     if (!dev->rx_bufs) {
220         spapr_vlan_record_dropped_rx_frame(dev);
221         return 0;
222     }
223 
224     if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
225         bd = spapr_vlan_get_rx_bd_from_pool(dev, size);
226     } else {
227         bd = spapr_vlan_get_rx_bd_from_page(dev, size);
228     }
229     if (!bd) {
230         spapr_vlan_record_dropped_rx_frame(dev);
231         return 0;
232     }
233 
234     dev->rx_bufs--;
235 
236     /* Transfer the packet data */
237     if (spapr_vio_dma_write(sdev, VLAN_BD_ADDR(bd) + 8, buf, size) < 0) {
238         return -1;
239     }
240 
241     trace_spapr_vlan_receive_dma_completed();
242 
243     /* Update the receive queue */
244     control = VLAN_RXQC_TOGGLE | VLAN_RXQC_VALID;
245     if (rxq_bd & VLAN_BD_TOGGLE) {
246         control ^= VLAN_RXQC_TOGGLE;
247     }
248 
249     handle = vio_ldq(sdev, VLAN_BD_ADDR(bd));
250     vio_stq(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 8, handle);
251     vio_stl(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 4, size);
252     vio_sth(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr + 2, 8);
253     vio_stb(sdev, VLAN_BD_ADDR(rxq_bd) + dev->rxq_ptr, control);
254 
255     trace_spapr_vlan_receive_wrote(dev->rxq_ptr,
256                                    vio_ldq(sdev, VLAN_BD_ADDR(rxq_bd) +
257                                                  dev->rxq_ptr),
258                                    vio_ldq(sdev, VLAN_BD_ADDR(rxq_bd) +
259                                                  dev->rxq_ptr + 8));
260 
261     dev->rxq_ptr += 16;
262     if (dev->rxq_ptr >= VLAN_BD_LEN(rxq_bd)) {
263         dev->rxq_ptr = 0;
264         vio_stq(sdev, dev->buf_list + VLAN_RXQ_BD_OFF, rxq_bd ^ VLAN_BD_TOGGLE);
265     }
266 
267     if (sdev->signal_state & 1) {
268         spapr_vio_irq_pulse(sdev);
269     }
270 
271     return size;
272 }
273 
274 static NetClientInfo net_spapr_vlan_info = {
275     .type = NET_CLIENT_DRIVER_NIC,
276     .size = sizeof(NICState),
277     .can_receive = spapr_vlan_can_receive,
278     .receive = spapr_vlan_receive,
279 };
280 
281 static void spapr_vlan_flush_rx_queue(void *opaque)
282 {
283     SpaprVioVlan *dev = opaque;
284 
285     qemu_flush_queued_packets(qemu_get_queue(dev->nic));
286 }
287 
288 static void spapr_vlan_reset_rx_pool(RxBufPool *rxp)
289 {
290     /*
291      * Use INT_MAX as bufsize so that unused buffers are moved to the end
292      * of the list during the qsort in spapr_vlan_add_rxbuf_to_pool() later.
293      */
294     rxp->bufsize = INT_MAX;
295     rxp->count = 0;
296     memset(rxp->bds, 0, sizeof(rxp->bds));
297 }
298 
299 static void spapr_vlan_reset(SpaprVioDevice *sdev)
300 {
301     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
302     int i;
303 
304     dev->buf_list = 0;
305     dev->rx_bufs = 0;
306     dev->isopen = 0;
307 
308     if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
309         for (i = 0; i < RX_MAX_POOLS; i++) {
310             spapr_vlan_reset_rx_pool(dev->rx_pool[i]);
311         }
312     }
313 
314     memcpy(&dev->nicconf.macaddr.a, &dev->perm_mac.a,
315            sizeof(dev->nicconf.macaddr.a));
316     qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
317 }
318 
319 static void spapr_vlan_realize(SpaprVioDevice *sdev, Error **errp)
320 {
321     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
322 
323     qemu_macaddr_default_if_unset(&dev->nicconf.macaddr);
324 
325     memcpy(&dev->perm_mac.a, &dev->nicconf.macaddr.a, sizeof(dev->perm_mac.a));
326 
327     dev->nic = qemu_new_nic(&net_spapr_vlan_info, &dev->nicconf,
328                             object_get_typename(OBJECT(sdev)), sdev->qdev.id,
329                             &sdev->qdev.mem_reentrancy_guard, dev);
330     qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
331 
332     dev->rxp_timer = timer_new_us(QEMU_CLOCK_VIRTUAL, spapr_vlan_flush_rx_queue,
333                                   dev);
334 }
335 
336 static void spapr_vlan_instance_init(Object *obj)
337 {
338     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(obj);
339     int i;
340 
341     device_add_bootindex_property(obj, &dev->nicconf.bootindex,
342                                   "bootindex", "",
343                                   DEVICE(dev));
344 
345     if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
346         for (i = 0; i < RX_MAX_POOLS; i++) {
347             dev->rx_pool[i] = g_new(RxBufPool, 1);
348             spapr_vlan_reset_rx_pool(dev->rx_pool[i]);
349         }
350     }
351 }
352 
353 static void spapr_vlan_instance_finalize(Object *obj)
354 {
355     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(obj);
356     int i;
357 
358     if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
359         for (i = 0; i < RX_MAX_POOLS; i++) {
360             g_free(dev->rx_pool[i]);
361             dev->rx_pool[i] = NULL;
362         }
363     }
364 
365     if (dev->rxp_timer) {
366         timer_free(dev->rxp_timer);
367     }
368 }
369 
370 void spapr_vlan_create(SpaprVioBus *bus, NICInfo *nd)
371 {
372     DeviceState *dev;
373 
374     dev = qdev_new("spapr-vlan");
375 
376     qdev_set_nic_properties(dev, nd);
377 
378     qdev_realize_and_unref(dev, &bus->bus, &error_fatal);
379 }
380 
381 static int spapr_vlan_devnode(SpaprVioDevice *dev, void *fdt, int node_off)
382 {
383     SpaprVioVlan *vdev = VIO_SPAPR_VLAN_DEVICE(dev);
384     uint8_t padded_mac[8] = {0, 0};
385     int ret;
386 
387     /* Some old phyp versions give the mac address in an 8-byte
388      * property.  The kernel driver (before 3.10) has an insane workaround;
389      * rather than doing the obvious thing and checking the property
390      * length, it checks whether the first byte has 0b10 in the low
391      * bits.  If a correct 6-byte property has a different first byte
392      * the kernel will get the wrong mac address, overrunning its
393      * buffer in the process (read only, thank goodness).
394      *
395      * Here we return a 6-byte address unless that would break a pre-3.10
396      * driver.  In that case we return a padded 8-byte address to allow the old
397      * workaround to succeed. */
398     if ((vdev->nicconf.macaddr.a[0] & 0x3) == 0x2) {
399         ret = fdt_setprop(fdt, node_off, "local-mac-address",
400                           &vdev->nicconf.macaddr, ETH_ALEN);
401     } else {
402         memcpy(&padded_mac[2], &vdev->nicconf.macaddr, ETH_ALEN);
403         ret = fdt_setprop(fdt, node_off, "local-mac-address",
404                           padded_mac, sizeof(padded_mac));
405     }
406     if (ret < 0) {
407         return ret;
408     }
409 
410     ret = fdt_setprop_cell(fdt, node_off, "ibm,mac-address-filters", 0);
411     if (ret < 0) {
412         return ret;
413     }
414 
415     return 0;
416 }
417 
418 static int check_bd(SpaprVioVlan *dev, vlan_bd_t bd,
419                     target_ulong alignment)
420 {
421     if ((VLAN_BD_ADDR(bd) % alignment)
422         || (VLAN_BD_LEN(bd) % alignment)) {
423         return -1;
424     }
425 
426     if (!spapr_vio_dma_valid(&dev->sdev, VLAN_BD_ADDR(bd),
427                              VLAN_BD_LEN(bd), DMA_DIRECTION_FROM_DEVICE)
428         || !spapr_vio_dma_valid(&dev->sdev, VLAN_BD_ADDR(bd),
429                                 VLAN_BD_LEN(bd), DMA_DIRECTION_TO_DEVICE)) {
430         return -1;
431     }
432 
433     return 0;
434 }
435 
436 static target_ulong h_register_logical_lan(PowerPCCPU *cpu,
437                                            SpaprMachineState *spapr,
438                                            target_ulong opcode,
439                                            target_ulong *args)
440 {
441     target_ulong reg = args[0];
442     target_ulong buf_list = args[1];
443     target_ulong rec_queue = args[2];
444     target_ulong filter_list = args[3];
445     SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
446     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
447     vlan_bd_t filter_list_bd;
448 
449     if (!dev) {
450         return H_PARAMETER;
451     }
452 
453     if (dev->isopen) {
454         hcall_dprintf("H_REGISTER_LOGICAL_LAN called twice without "
455                       "H_FREE_LOGICAL_LAN\n");
456         return H_RESOURCE;
457     }
458 
459     if (check_bd(dev, VLAN_VALID_BD(buf_list, SPAPR_TCE_PAGE_SIZE),
460                  SPAPR_TCE_PAGE_SIZE) < 0) {
461         hcall_dprintf("Bad buf_list 0x" TARGET_FMT_lx "\n", buf_list);
462         return H_PARAMETER;
463     }
464 
465     filter_list_bd = VLAN_VALID_BD(filter_list, SPAPR_TCE_PAGE_SIZE);
466     if (check_bd(dev, filter_list_bd, SPAPR_TCE_PAGE_SIZE) < 0) {
467         hcall_dprintf("Bad filter_list 0x" TARGET_FMT_lx "\n", filter_list);
468         return H_PARAMETER;
469     }
470 
471     if (!(rec_queue & VLAN_BD_VALID)
472         || (check_bd(dev, rec_queue, VLAN_RQ_ALIGNMENT) < 0)) {
473         hcall_dprintf("Bad receive queue\n");
474         return H_PARAMETER;
475     }
476 
477     dev->buf_list = buf_list;
478     sdev->signal_state = 0;
479 
480     rec_queue &= ~VLAN_BD_TOGGLE;
481 
482     /* Initialize the buffer list */
483     vio_stq(sdev, buf_list, rec_queue);
484     vio_stq(sdev, buf_list + 8, filter_list_bd);
485     spapr_vio_dma_set(sdev, buf_list + VLAN_RX_BDS_OFF, 0,
486                       SPAPR_TCE_PAGE_SIZE - VLAN_RX_BDS_OFF);
487     dev->add_buf_ptr = VLAN_RX_BDS_OFF - 8;
488     dev->use_buf_ptr = VLAN_RX_BDS_OFF - 8;
489     dev->rx_bufs = 0;
490     dev->rxq_ptr = 0;
491 
492     /* Initialize the receive queue */
493     spapr_vio_dma_set(sdev, VLAN_BD_ADDR(rec_queue), 0, VLAN_BD_LEN(rec_queue));
494 
495     dev->isopen = 1;
496     qemu_flush_queued_packets(qemu_get_queue(dev->nic));
497 
498     return H_SUCCESS;
499 }
500 
501 
502 static target_ulong h_free_logical_lan(PowerPCCPU *cpu,
503                                        SpaprMachineState *spapr,
504                                        target_ulong opcode, target_ulong *args)
505 {
506     target_ulong reg = args[0];
507     SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
508     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
509 
510     if (!dev) {
511         return H_PARAMETER;
512     }
513 
514     if (!dev->isopen) {
515         hcall_dprintf("H_FREE_LOGICAL_LAN called without "
516                       "H_REGISTER_LOGICAL_LAN\n");
517         return H_RESOURCE;
518     }
519 
520     spapr_vlan_reset(sdev);
521     return H_SUCCESS;
522 }
523 
524 /**
525  * Used for qsort, this function compares two RxBufPools by size.
526  */
527 static int rx_pool_size_compare(const void *p1, const void *p2)
528 {
529     const RxBufPool *pool1 = *(RxBufPool **)p1;
530     const RxBufPool *pool2 = *(RxBufPool **)p2;
531 
532     if (pool1->bufsize < pool2->bufsize) {
533         return -1;
534     }
535     return pool1->bufsize > pool2->bufsize;
536 }
537 
538 /**
539  * Search for a matching buffer pool with exact matching size,
540  * or return -1 if no matching pool has been found.
541  */
542 static int spapr_vlan_get_rx_pool_id(SpaprVioVlan *dev, int size)
543 {
544     int pool;
545 
546     for (pool = 0; pool < RX_MAX_POOLS; pool++) {
547         if (dev->rx_pool[pool]->bufsize == size) {
548             return pool;
549         }
550     }
551 
552     return -1;
553 }
554 
555 /**
556  * Enqueuing receive buffer by adding it to one of our receive buffer pools
557  */
558 static target_long spapr_vlan_add_rxbuf_to_pool(SpaprVioVlan *dev,
559                                                 target_ulong buf)
560 {
561     int size = VLAN_BD_LEN(buf);
562     int pool;
563 
564     pool = spapr_vlan_get_rx_pool_id(dev, size);
565     if (pool < 0) {
566         /*
567          * No matching pool found? Try to use a new one. If the guest used all
568          * pools before, but changed the size of one pool in the meantime, we might
569          * need to recycle that pool here (if it's empty already). Thus scan
570          * all buffer pools now, starting with the last (likely empty) one.
571          */
572         for (pool = RX_MAX_POOLS - 1; pool >= 0 ; pool--) {
573             if (dev->rx_pool[pool]->count == 0) {
574                 dev->rx_pool[pool]->bufsize = size;
575                 /*
576                  * Sort pools by size so that spapr_vlan_receive()
577                  * can later find the smallest buffer pool easily.
578                  */
579                 qsort(dev->rx_pool, RX_MAX_POOLS, sizeof(dev->rx_pool[0]),
580                       rx_pool_size_compare);
581                 pool = spapr_vlan_get_rx_pool_id(dev, size);
582                 trace_spapr_vlan_add_rxbuf_to_pool_create(pool,
583                                                           VLAN_BD_LEN(buf));
584                 break;
585             }
586         }
587     }
588     /* Still no usable pool? Give up */
589     if (pool < 0 || dev->rx_pool[pool]->count >= RX_POOL_MAX_BDS) {
590         return H_RESOURCE;
591     }
592 
593     trace_spapr_vlan_add_rxbuf_to_pool(pool, VLAN_BD_LEN(buf),
594                                        dev->rx_pool[pool]->count);
595 
596     dev->rx_pool[pool]->bds[dev->rx_pool[pool]->count++] = buf;
597 
598     return 0;
599 }
600 
601 /**
602  * This is the old way of enqueuing receive buffers: Add it to the rx queue
603  * page that has been supplied by the guest (which is quite limited in size).
604  */
605 static target_long spapr_vlan_add_rxbuf_to_page(SpaprVioVlan *dev,
606                                                 target_ulong buf)
607 {
608     vlan_bd_t bd;
609 
610     if (dev->rx_bufs >= VLAN_MAX_BUFS) {
611         return H_RESOURCE;
612     }
613 
614     do {
615         dev->add_buf_ptr += 8;
616         if (dev->add_buf_ptr >= VLAN_RX_BDS_LEN + VLAN_RX_BDS_OFF) {
617             dev->add_buf_ptr = VLAN_RX_BDS_OFF;
618         }
619 
620         bd = vio_ldq(&dev->sdev, dev->buf_list + dev->add_buf_ptr);
621     } while (bd & VLAN_BD_VALID);
622 
623     vio_stq(&dev->sdev, dev->buf_list + dev->add_buf_ptr, buf);
624 
625     trace_spapr_vlan_add_rxbuf_to_page(dev->add_buf_ptr, dev->rx_bufs, buf);
626 
627     return 0;
628 }
629 
630 static target_ulong h_add_logical_lan_buffer(PowerPCCPU *cpu,
631                                              SpaprMachineState *spapr,
632                                              target_ulong opcode,
633                                              target_ulong *args)
634 {
635     target_ulong reg = args[0];
636     target_ulong buf = args[1];
637     SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
638     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
639     target_long ret;
640 
641     trace_spapr_vlan_h_add_logical_lan_buffer(reg, buf);
642 
643     if (!sdev) {
644         hcall_dprintf("Bad device\n");
645         return H_PARAMETER;
646     }
647 
648     if ((check_bd(dev, buf, 4) < 0)
649         || (VLAN_BD_LEN(buf) < 16)) {
650         hcall_dprintf("Bad buffer enqueued\n");
651         return H_PARAMETER;
652     }
653 
654     if (!dev->isopen) {
655         return H_RESOURCE;
656     }
657 
658     if (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) {
659         ret = spapr_vlan_add_rxbuf_to_pool(dev, buf);
660     } else {
661         ret = spapr_vlan_add_rxbuf_to_page(dev, buf);
662     }
663     if (ret) {
664         return ret;
665     }
666 
667     dev->rx_bufs++;
668 
669     /*
670      * Give guest some more time to add additional RX buffers before we
671      * flush the receive queue, so that e.g. fragmented IP packets can
672      * be passed to the guest in one go later (instead of passing single
673      * fragments if there is only one receive buffer available).
674      */
675     timer_mod(dev->rxp_timer, qemu_clock_get_us(QEMU_CLOCK_VIRTUAL) + 500);
676 
677     return H_SUCCESS;
678 }
679 
680 static target_ulong h_send_logical_lan(PowerPCCPU *cpu,
681                                        SpaprMachineState *spapr,
682                                        target_ulong opcode, target_ulong *args)
683 {
684     target_ulong reg = args[0];
685     target_ulong *bufs = args + 1;
686     target_ulong continue_token = args[7];
687     SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
688     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
689     unsigned total_len;
690     uint8_t *p;
691     g_autofree uint8_t *lbuf = NULL;
692     int i, nbufs;
693     int ret;
694 
695     trace_spapr_vlan_h_send_logical_lan(reg, continue_token);
696 
697     if (!sdev) {
698         return H_PARAMETER;
699     }
700 
701     trace_spapr_vlan_h_send_logical_lan_rxbufs(dev->rx_bufs);
702 
703     if (!dev->isopen) {
704         return H_DROPPED;
705     }
706 
707     if (continue_token) {
708         return H_HARDWARE; /* FIXME actually handle this */
709     }
710 
711     total_len = 0;
712     for (i = 0; i < 6; i++) {
713         trace_spapr_vlan_h_send_logical_lan_buf_desc(bufs[i]);
714         if (!(bufs[i] & VLAN_BD_VALID)) {
715             break;
716         }
717         total_len += VLAN_BD_LEN(bufs[i]);
718     }
719 
720     nbufs = i;
721     trace_spapr_vlan_h_send_logical_lan_total(nbufs, total_len);
722 
723     if (total_len == 0) {
724         return H_SUCCESS;
725     }
726 
727     if (total_len > MAX_PACKET_SIZE) {
728         /* Don't let the guest force too large an allocation */
729         return H_RESOURCE;
730     }
731 
732     lbuf = g_malloc(total_len);
733     p = lbuf;
734     for (i = 0; i < nbufs; i++) {
735         ret = spapr_vio_dma_read(sdev, VLAN_BD_ADDR(bufs[i]),
736                                  p, VLAN_BD_LEN(bufs[i]));
737         if (ret < 0) {
738             return ret;
739         }
740 
741         p += VLAN_BD_LEN(bufs[i]);
742     }
743 
744     qemu_send_packet(qemu_get_queue(dev->nic), lbuf, total_len);
745 
746     return H_SUCCESS;
747 }
748 
749 static target_ulong h_multicast_ctrl(PowerPCCPU *cpu, SpaprMachineState *spapr,
750                                      target_ulong opcode, target_ulong *args)
751 {
752     target_ulong reg = args[0];
753     SpaprVioDevice *dev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
754 
755     if (!dev) {
756         return H_PARAMETER;
757     }
758 
759     return H_SUCCESS;
760 }
761 
762 static target_ulong h_change_logical_lan_mac(PowerPCCPU *cpu,
763                                              SpaprMachineState *spapr,
764                                              target_ulong opcode,
765                                              target_ulong *args)
766 {
767     target_ulong reg = args[0];
768     target_ulong macaddr = args[1];
769     SpaprVioDevice *sdev = spapr_vio_find_by_reg(spapr->vio_bus, reg);
770     SpaprVioVlan *dev = VIO_SPAPR_VLAN_DEVICE(sdev);
771     int i;
772 
773     if (!dev) {
774         hcall_dprintf("H_CHANGE_LOGICAL_LAN_MAC called when "
775                       "no NIC is present\n");
776         return H_PARAMETER;
777     }
778 
779     for (i = 0; i < ETH_ALEN; i++) {
780         dev->nicconf.macaddr.a[ETH_ALEN - i - 1] = macaddr & 0xff;
781         macaddr >>= 8;
782     }
783 
784     qemu_format_nic_info_str(qemu_get_queue(dev->nic), dev->nicconf.macaddr.a);
785 
786     return H_SUCCESS;
787 }
788 
789 static Property spapr_vlan_properties[] = {
790     DEFINE_SPAPR_PROPERTIES(SpaprVioVlan, sdev),
791     DEFINE_NIC_PROPERTIES(SpaprVioVlan, nicconf),
792     DEFINE_PROP_BIT("use-rx-buffer-pools", SpaprVioVlan,
793                     compat_flags, SPAPRVLAN_FLAG_RX_BUF_POOLS_BIT, true),
794     DEFINE_PROP_END_OF_LIST(),
795 };
796 
797 static bool spapr_vlan_rx_buffer_pools_needed(void *opaque)
798 {
799     SpaprVioVlan *dev = opaque;
800 
801     return (dev->compat_flags & SPAPRVLAN_FLAG_RX_BUF_POOLS) != 0;
802 }
803 
804 static const VMStateDescription vmstate_rx_buffer_pool = {
805     .name = "spapr_llan/rx_buffer_pool",
806     .version_id = 1,
807     .minimum_version_id = 1,
808     .needed = spapr_vlan_rx_buffer_pools_needed,
809     .fields = (const VMStateField[]) {
810         VMSTATE_INT32(bufsize, RxBufPool),
811         VMSTATE_INT32(count, RxBufPool),
812         VMSTATE_UINT64_ARRAY(bds, RxBufPool, RX_POOL_MAX_BDS),
813         VMSTATE_END_OF_LIST()
814     }
815 };
816 
817 static const VMStateDescription vmstate_rx_pools = {
818     .name = "spapr_llan/rx_pools",
819     .version_id = 1,
820     .minimum_version_id = 1,
821     .needed = spapr_vlan_rx_buffer_pools_needed,
822     .fields = (const VMStateField[]) {
823         VMSTATE_ARRAY_OF_POINTER_TO_STRUCT(rx_pool, SpaprVioVlan,
824                                            RX_MAX_POOLS, 1,
825                                            vmstate_rx_buffer_pool, RxBufPool),
826         VMSTATE_END_OF_LIST()
827     }
828 };
829 
830 static const VMStateDescription vmstate_spapr_llan = {
831     .name = "spapr_llan",
832     .version_id = 1,
833     .minimum_version_id = 1,
834     .fields = (const VMStateField[]) {
835         VMSTATE_SPAPR_VIO(sdev, SpaprVioVlan),
836         /* LLAN state */
837         VMSTATE_BOOL(isopen, SpaprVioVlan),
838         VMSTATE_UINT64(buf_list, SpaprVioVlan),
839         VMSTATE_UINT32(add_buf_ptr, SpaprVioVlan),
840         VMSTATE_UINT32(use_buf_ptr, SpaprVioVlan),
841         VMSTATE_UINT32(rx_bufs, SpaprVioVlan),
842         VMSTATE_UINT64(rxq_ptr, SpaprVioVlan),
843 
844         VMSTATE_END_OF_LIST()
845     },
846     .subsections = (const VMStateDescription * const []) {
847         &vmstate_rx_pools,
848         NULL
849     }
850 };
851 
852 static void spapr_vlan_class_init(ObjectClass *klass, void *data)
853 {
854     DeviceClass *dc = DEVICE_CLASS(klass);
855     SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass);
856 
857     k->realize = spapr_vlan_realize;
858     k->reset = spapr_vlan_reset;
859     k->devnode = spapr_vlan_devnode;
860     k->dt_name = "l-lan";
861     k->dt_type = "network";
862     k->dt_compatible = "IBM,l-lan";
863     k->signal_mask = 0x1;
864     set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
865     device_class_set_props(dc, spapr_vlan_properties);
866     k->rtce_window_size = 0x10000000;
867     dc->vmsd = &vmstate_spapr_llan;
868 }
869 
870 static const TypeInfo spapr_vlan_info = {
871     .name          = TYPE_VIO_SPAPR_VLAN_DEVICE,
872     .parent        = TYPE_VIO_SPAPR_DEVICE,
873     .instance_size = sizeof(SpaprVioVlan),
874     .class_init    = spapr_vlan_class_init,
875     .instance_init = spapr_vlan_instance_init,
876     .instance_finalize = spapr_vlan_instance_finalize,
877 };
878 
879 static void spapr_vlan_register_types(void)
880 {
881     spapr_register_hypercall(H_REGISTER_LOGICAL_LAN, h_register_logical_lan);
882     spapr_register_hypercall(H_FREE_LOGICAL_LAN, h_free_logical_lan);
883     spapr_register_hypercall(H_SEND_LOGICAL_LAN, h_send_logical_lan);
884     spapr_register_hypercall(H_ADD_LOGICAL_LAN_BUFFER,
885                              h_add_logical_lan_buffer);
886     spapr_register_hypercall(H_MULTICAST_CTRL, h_multicast_ctrl);
887     spapr_register_hypercall(H_CHANGE_LOGICAL_LAN_MAC,
888                              h_change_logical_lan_mac);
889     type_register_static(&spapr_vlan_info);
890 }
891 
892 type_init(spapr_vlan_register_types)
893