1 /*
2 * QEMU Sun Happy Meal Ethernet emulation
3 *
4 * Copyright (c) 2017 Mark Cave-Ayland
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
24
25 #include "qemu/osdep.h"
26 #include "hw/pci/pci_device.h"
27 #include "hw/qdev-properties.h"
28 #include "migration/vmstate.h"
29 #include "hw/net/mii.h"
30 #include "net/net.h"
31 #include "qemu/module.h"
32 #include "net/checksum.h"
33 #include "net/eth.h"
34 #include "sysemu/sysemu.h"
35 #include "trace.h"
36 #include "qom/object.h"
37
38 #define HME_REG_SIZE 0x8000
39
40 #define HME_SEB_REG_SIZE 0x2000
41
42 #define HME_SEBI_RESET 0x0
43 #define HME_SEB_RESET_ETX 0x1
44 #define HME_SEB_RESET_ERX 0x2
45
46 #define HME_SEBI_STAT 0x100
47 #define HME_SEBI_STAT_LINUXBUG 0x108
48 #define HME_SEB_STAT_RXTOHOST 0x10000
49 #define HME_SEB_STAT_NORXD 0x20000
50 #define HME_SEB_STAT_MIFIRQ 0x800000
51 #define HME_SEB_STAT_HOSTTOTX 0x1000000
52 #define HME_SEB_STAT_TXALL 0x2000000
53
54 #define HME_SEBI_IMASK 0x104
55 #define HME_SEBI_IMASK_LINUXBUG 0x10c
56
57 #define HME_ETX_REG_SIZE 0x2000
58
59 #define HME_ETXI_PENDING 0x0
60
61 #define HME_ETXI_RING 0x8
62 #define HME_ETXI_RING_ADDR 0xffffff00
63 #define HME_ETXI_RING_OFFSET 0xff
64
65 #define HME_ETXI_RSIZE 0x2c
66
67 #define HME_ERX_REG_SIZE 0x2000
68
69 #define HME_ERXI_CFG 0x0
70 #define HME_ERX_CFG_RINGSIZE 0x600
71 #define HME_ERX_CFG_RINGSIZE_SHIFT 9
72 #define HME_ERX_CFG_BYTEOFFSET 0x38
73 #define HME_ERX_CFG_BYTEOFFSET_SHIFT 3
74 #define HME_ERX_CFG_CSUMSTART 0x7f0000
75 #define HME_ERX_CFG_CSUMSHIFT 16
76
77 #define HME_ERXI_RING 0x4
78 #define HME_ERXI_RING_ADDR 0xffffff00
79 #define HME_ERXI_RING_OFFSET 0xff
80
81 #define HME_MAC_REG_SIZE 0x1000
82
83 #define HME_MACI_TXCFG 0x20c
84 #define HME_MAC_TXCFG_ENABLE 0x1
85
86 #define HME_MACI_RXCFG 0x30c
87 #define HME_MAC_RXCFG_ENABLE 0x1
88 #define HME_MAC_RXCFG_PMISC 0x40
89 #define HME_MAC_RXCFG_HENABLE 0x800
90
91 #define HME_MACI_MACADDR2 0x318
92 #define HME_MACI_MACADDR1 0x31c
93 #define HME_MACI_MACADDR0 0x320
94
95 #define HME_MACI_HASHTAB3 0x340
96 #define HME_MACI_HASHTAB2 0x344
97 #define HME_MACI_HASHTAB1 0x348
98 #define HME_MACI_HASHTAB0 0x34c
99
100 #define HME_MIF_REG_SIZE 0x20
101
102 #define HME_MIFI_FO 0xc
103 #define HME_MIF_FO_ST 0xc0000000
104 #define HME_MIF_FO_ST_SHIFT 30
105 #define HME_MIF_FO_OPC 0x30000000
106 #define HME_MIF_FO_OPC_SHIFT 28
107 #define HME_MIF_FO_PHYAD 0x0f800000
108 #define HME_MIF_FO_PHYAD_SHIFT 23
109 #define HME_MIF_FO_REGAD 0x007c0000
110 #define HME_MIF_FO_REGAD_SHIFT 18
111 #define HME_MIF_FO_TAMSB 0x20000
112 #define HME_MIF_FO_TALSB 0x10000
113 #define HME_MIF_FO_DATA 0xffff
114
115 #define HME_MIFI_CFG 0x10
116 #define HME_MIF_CFG_MDI0 0x100
117 #define HME_MIF_CFG_MDI1 0x200
118
119 #define HME_MIFI_IMASK 0x14
120
121 #define HME_MIFI_STAT 0x18
122
123
124 /* Wired HME PHY addresses */
125 #define HME_PHYAD_INTERNAL 1
126 #define HME_PHYAD_EXTERNAL 0
127
128 #define MII_COMMAND_START 0x1
129 #define MII_COMMAND_READ 0x2
130 #define MII_COMMAND_WRITE 0x1
131
132 #define TYPE_SUNHME "sunhme"
133 OBJECT_DECLARE_SIMPLE_TYPE(SunHMEState, SUNHME)
134
135 /* Maximum size of buffer */
136 #define HME_FIFO_SIZE 0x800
137
138 /* Size of TX/RX descriptor */
139 #define HME_DESC_SIZE 0x8
140
141 #define HME_XD_OWN 0x80000000
142 #define HME_XD_OFL 0x40000000
143 #define HME_XD_SOP 0x40000000
144 #define HME_XD_EOP 0x20000000
145 #define HME_XD_RXLENMSK 0x3fff0000
146 #define HME_XD_RXLENSHIFT 16
147 #define HME_XD_RXCKSUM 0xffff
148 #define HME_XD_TXLENMSK 0x00001fff
149 #define HME_XD_TXCKSUM 0x10000000
150 #define HME_XD_TXCSSTUFF 0xff00000
151 #define HME_XD_TXCSSTUFFSHIFT 20
152 #define HME_XD_TXCSSTART 0xfc000
153 #define HME_XD_TXCSSTARTSHIFT 14
154
155 #define HME_MII_REGS_SIZE 0x20
156
157 struct SunHMEState {
158 /*< private >*/
159 PCIDevice parent_obj;
160
161 NICState *nic;
162 NICConf conf;
163
164 MemoryRegion hme;
165 MemoryRegion sebreg;
166 MemoryRegion etxreg;
167 MemoryRegion erxreg;
168 MemoryRegion macreg;
169 MemoryRegion mifreg;
170
171 uint32_t sebregs[HME_SEB_REG_SIZE >> 2];
172 uint32_t etxregs[HME_ETX_REG_SIZE >> 2];
173 uint32_t erxregs[HME_ERX_REG_SIZE >> 2];
174 uint32_t macregs[HME_MAC_REG_SIZE >> 2];
175 uint32_t mifregs[HME_MIF_REG_SIZE >> 2];
176
177 uint16_t miiregs[HME_MII_REGS_SIZE];
178 };
179
180 static Property sunhme_properties[] = {
181 DEFINE_NIC_PROPERTIES(SunHMEState, conf),
182 DEFINE_PROP_END_OF_LIST(),
183 };
184
sunhme_reset_tx(SunHMEState * s)185 static void sunhme_reset_tx(SunHMEState *s)
186 {
187 /* Indicate TX reset complete */
188 s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ETX;
189 }
190
sunhme_reset_rx(SunHMEState * s)191 static void sunhme_reset_rx(SunHMEState *s)
192 {
193 /* Indicate RX reset complete */
194 s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ERX;
195 }
196
sunhme_update_irq(SunHMEState * s)197 static void sunhme_update_irq(SunHMEState *s)
198 {
199 PCIDevice *d = PCI_DEVICE(s);
200 int level;
201
202 /* MIF interrupt mask (16-bit) */
203 uint32_t mifmask = ~(s->mifregs[HME_MIFI_IMASK >> 2]) & 0xffff;
204 uint32_t mif = s->mifregs[HME_MIFI_STAT >> 2] & mifmask;
205
206 /* Main SEB interrupt mask (include MIF status from above) */
207 uint32_t sebmask = ~(s->sebregs[HME_SEBI_IMASK >> 2]) &
208 ~HME_SEB_STAT_MIFIRQ;
209 uint32_t seb = s->sebregs[HME_SEBI_STAT >> 2] & sebmask;
210 if (mif) {
211 seb |= HME_SEB_STAT_MIFIRQ;
212 }
213
214 level = (seb ? 1 : 0);
215 trace_sunhme_update_irq(mifmask, mif, sebmask, seb, level);
216
217 pci_set_irq(d, level);
218 }
219
sunhme_seb_write(void * opaque,hwaddr addr,uint64_t val,unsigned size)220 static void sunhme_seb_write(void *opaque, hwaddr addr,
221 uint64_t val, unsigned size)
222 {
223 SunHMEState *s = SUNHME(opaque);
224
225 trace_sunhme_seb_write(addr, val);
226
227 /* Handly buggy Linux drivers before 4.13 which have
228 the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */
229 switch (addr) {
230 case HME_SEBI_STAT_LINUXBUG:
231 addr = HME_SEBI_STAT;
232 break;
233 case HME_SEBI_IMASK_LINUXBUG:
234 addr = HME_SEBI_IMASK;
235 break;
236 default:
237 break;
238 }
239
240 switch (addr) {
241 case HME_SEBI_RESET:
242 if (val & HME_SEB_RESET_ETX) {
243 sunhme_reset_tx(s);
244 }
245 if (val & HME_SEB_RESET_ERX) {
246 sunhme_reset_rx(s);
247 }
248 val = s->sebregs[HME_SEBI_RESET >> 2];
249 break;
250 }
251
252 s->sebregs[addr >> 2] = val;
253 }
254
sunhme_seb_read(void * opaque,hwaddr addr,unsigned size)255 static uint64_t sunhme_seb_read(void *opaque, hwaddr addr,
256 unsigned size)
257 {
258 SunHMEState *s = SUNHME(opaque);
259 uint64_t val;
260
261 /* Handly buggy Linux drivers before 4.13 which have
262 the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */
263 switch (addr) {
264 case HME_SEBI_STAT_LINUXBUG:
265 addr = HME_SEBI_STAT;
266 break;
267 case HME_SEBI_IMASK_LINUXBUG:
268 addr = HME_SEBI_IMASK;
269 break;
270 default:
271 break;
272 }
273
274 val = s->sebregs[addr >> 2];
275
276 switch (addr) {
277 case HME_SEBI_STAT:
278 /* Autoclear status (except MIF) */
279 s->sebregs[HME_SEBI_STAT >> 2] &= HME_SEB_STAT_MIFIRQ;
280 sunhme_update_irq(s);
281 break;
282 }
283
284 trace_sunhme_seb_read(addr, val);
285
286 return val;
287 }
288
289 static const MemoryRegionOps sunhme_seb_ops = {
290 .read = sunhme_seb_read,
291 .write = sunhme_seb_write,
292 .endianness = DEVICE_LITTLE_ENDIAN,
293 .valid = {
294 .min_access_size = 4,
295 .max_access_size = 4,
296 },
297 };
298
299 static void sunhme_transmit(SunHMEState *s);
300
sunhme_etx_write(void * opaque,hwaddr addr,uint64_t val,unsigned size)301 static void sunhme_etx_write(void *opaque, hwaddr addr,
302 uint64_t val, unsigned size)
303 {
304 SunHMEState *s = SUNHME(opaque);
305
306 trace_sunhme_etx_write(addr, val);
307
308 switch (addr) {
309 case HME_ETXI_PENDING:
310 if (val) {
311 sunhme_transmit(s);
312 }
313 break;
314 }
315
316 s->etxregs[addr >> 2] = val;
317 }
318
sunhme_etx_read(void * opaque,hwaddr addr,unsigned size)319 static uint64_t sunhme_etx_read(void *opaque, hwaddr addr,
320 unsigned size)
321 {
322 SunHMEState *s = SUNHME(opaque);
323 uint64_t val;
324
325 val = s->etxregs[addr >> 2];
326
327 trace_sunhme_etx_read(addr, val);
328
329 return val;
330 }
331
332 static const MemoryRegionOps sunhme_etx_ops = {
333 .read = sunhme_etx_read,
334 .write = sunhme_etx_write,
335 .endianness = DEVICE_LITTLE_ENDIAN,
336 .valid = {
337 .min_access_size = 4,
338 .max_access_size = 4,
339 },
340 };
341
sunhme_erx_write(void * opaque,hwaddr addr,uint64_t val,unsigned size)342 static void sunhme_erx_write(void *opaque, hwaddr addr,
343 uint64_t val, unsigned size)
344 {
345 SunHMEState *s = SUNHME(opaque);
346
347 trace_sunhme_erx_write(addr, val);
348
349 s->erxregs[addr >> 2] = val;
350 }
351
sunhme_erx_read(void * opaque,hwaddr addr,unsigned size)352 static uint64_t sunhme_erx_read(void *opaque, hwaddr addr,
353 unsigned size)
354 {
355 SunHMEState *s = SUNHME(opaque);
356 uint64_t val;
357
358 val = s->erxregs[addr >> 2];
359
360 trace_sunhme_erx_read(addr, val);
361
362 return val;
363 }
364
365 static const MemoryRegionOps sunhme_erx_ops = {
366 .read = sunhme_erx_read,
367 .write = sunhme_erx_write,
368 .endianness = DEVICE_LITTLE_ENDIAN,
369 .valid = {
370 .min_access_size = 4,
371 .max_access_size = 4,
372 },
373 };
374
sunhme_mac_write(void * opaque,hwaddr addr,uint64_t val,unsigned size)375 static void sunhme_mac_write(void *opaque, hwaddr addr,
376 uint64_t val, unsigned size)
377 {
378 SunHMEState *s = SUNHME(opaque);
379 uint64_t oldval = s->macregs[addr >> 2];
380
381 trace_sunhme_mac_write(addr, val);
382
383 s->macregs[addr >> 2] = val;
384
385 switch (addr) {
386 case HME_MACI_RXCFG:
387 if (!(oldval & HME_MAC_RXCFG_ENABLE) &&
388 (val & HME_MAC_RXCFG_ENABLE)) {
389 qemu_flush_queued_packets(qemu_get_queue(s->nic));
390 }
391 break;
392 }
393 }
394
sunhme_mac_read(void * opaque,hwaddr addr,unsigned size)395 static uint64_t sunhme_mac_read(void *opaque, hwaddr addr,
396 unsigned size)
397 {
398 SunHMEState *s = SUNHME(opaque);
399 uint64_t val;
400
401 val = s->macregs[addr >> 2];
402
403 trace_sunhme_mac_read(addr, val);
404
405 return val;
406 }
407
408 static const MemoryRegionOps sunhme_mac_ops = {
409 .read = sunhme_mac_read,
410 .write = sunhme_mac_write,
411 .endianness = DEVICE_LITTLE_ENDIAN,
412 .valid = {
413 .min_access_size = 4,
414 .max_access_size = 4,
415 },
416 };
417
sunhme_mii_write(SunHMEState * s,uint8_t reg,uint16_t data)418 static void sunhme_mii_write(SunHMEState *s, uint8_t reg, uint16_t data)
419 {
420 trace_sunhme_mii_write(reg, data);
421
422 switch (reg) {
423 case MII_BMCR:
424 if (data & MII_BMCR_RESET) {
425 /* Autoclear reset bit, enable auto negotiation */
426 data &= ~MII_BMCR_RESET;
427 data |= MII_BMCR_AUTOEN;
428 }
429 if (data & MII_BMCR_ANRESTART) {
430 /* Autoclear auto negotiation restart */
431 data &= ~MII_BMCR_ANRESTART;
432
433 /* Indicate negotiation complete */
434 s->miiregs[MII_BMSR] |= MII_BMSR_AN_COMP;
435
436 if (!qemu_get_queue(s->nic)->link_down) {
437 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD;
438 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST;
439 }
440 }
441 break;
442 }
443
444 s->miiregs[reg] = data;
445 }
446
sunhme_mii_read(SunHMEState * s,uint8_t reg)447 static uint16_t sunhme_mii_read(SunHMEState *s, uint8_t reg)
448 {
449 uint16_t data = s->miiregs[reg];
450
451 trace_sunhme_mii_read(reg, data);
452
453 return data;
454 }
455
sunhme_mif_write(void * opaque,hwaddr addr,uint64_t val,unsigned size)456 static void sunhme_mif_write(void *opaque, hwaddr addr,
457 uint64_t val, unsigned size)
458 {
459 SunHMEState *s = SUNHME(opaque);
460 uint8_t cmd, reg;
461 uint16_t data;
462
463 trace_sunhme_mif_write(addr, val);
464
465 switch (addr) {
466 case HME_MIFI_CFG:
467 /* Mask the read-only bits */
468 val &= ~(HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1);
469 val |= s->mifregs[HME_MIFI_CFG >> 2] &
470 (HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1);
471 break;
472 case HME_MIFI_FO:
473 /* Detect start of MII command */
474 if ((val & HME_MIF_FO_ST) >> HME_MIF_FO_ST_SHIFT
475 != MII_COMMAND_START) {
476 val |= HME_MIF_FO_TALSB;
477 break;
478 }
479
480 /* Internal phy only */
481 if ((val & HME_MIF_FO_PHYAD) >> HME_MIF_FO_PHYAD_SHIFT
482 != HME_PHYAD_INTERNAL) {
483 val |= HME_MIF_FO_TALSB;
484 break;
485 }
486
487 cmd = (val & HME_MIF_FO_OPC) >> HME_MIF_FO_OPC_SHIFT;
488 reg = (val & HME_MIF_FO_REGAD) >> HME_MIF_FO_REGAD_SHIFT;
489 data = (val & HME_MIF_FO_DATA);
490
491 switch (cmd) {
492 case MII_COMMAND_WRITE:
493 sunhme_mii_write(s, reg, data);
494 break;
495
496 case MII_COMMAND_READ:
497 val &= ~HME_MIF_FO_DATA;
498 val |= sunhme_mii_read(s, reg);
499 break;
500 }
501
502 val |= HME_MIF_FO_TALSB;
503 break;
504 }
505
506 s->mifregs[addr >> 2] = val;
507 }
508
sunhme_mif_read(void * opaque,hwaddr addr,unsigned size)509 static uint64_t sunhme_mif_read(void *opaque, hwaddr addr,
510 unsigned size)
511 {
512 SunHMEState *s = SUNHME(opaque);
513 uint64_t val;
514
515 val = s->mifregs[addr >> 2];
516
517 switch (addr) {
518 case HME_MIFI_STAT:
519 /* Autoclear MIF interrupt status */
520 s->mifregs[HME_MIFI_STAT >> 2] = 0;
521 sunhme_update_irq(s);
522 break;
523 }
524
525 trace_sunhme_mif_read(addr, val);
526
527 return val;
528 }
529
530 static const MemoryRegionOps sunhme_mif_ops = {
531 .read = sunhme_mif_read,
532 .write = sunhme_mif_write,
533 .endianness = DEVICE_LITTLE_ENDIAN,
534 .valid = {
535 .min_access_size = 4,
536 .max_access_size = 4,
537 },
538 };
539
sunhme_transmit_frame(SunHMEState * s,uint8_t * buf,int size)540 static void sunhme_transmit_frame(SunHMEState *s, uint8_t *buf, int size)
541 {
542 qemu_send_packet(qemu_get_queue(s->nic), buf, size);
543 }
544
sunhme_get_tx_ring_count(SunHMEState * s)545 static inline int sunhme_get_tx_ring_count(SunHMEState *s)
546 {
547 return (s->etxregs[HME_ETXI_RSIZE >> 2] + 1) << 4;
548 }
549
sunhme_get_tx_ring_nr(SunHMEState * s)550 static inline int sunhme_get_tx_ring_nr(SunHMEState *s)
551 {
552 return s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_OFFSET;
553 }
554
sunhme_set_tx_ring_nr(SunHMEState * s,int i)555 static inline void sunhme_set_tx_ring_nr(SunHMEState *s, int i)
556 {
557 uint32_t ring = s->etxregs[HME_ETXI_RING >> 2] & ~HME_ETXI_RING_OFFSET;
558 ring |= i & HME_ETXI_RING_OFFSET;
559
560 s->etxregs[HME_ETXI_RING >> 2] = ring;
561 }
562
sunhme_transmit(SunHMEState * s)563 static void sunhme_transmit(SunHMEState *s)
564 {
565 PCIDevice *d = PCI_DEVICE(s);
566 dma_addr_t tb, addr;
567 uint32_t intstatus, status, buffer, sum = 0;
568 int cr, nr, len, xmit_pos, csum_offset = 0, csum_stuff_offset = 0;
569 uint16_t csum = 0;
570 uint8_t xmit_buffer[HME_FIFO_SIZE];
571
572 tb = s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_ADDR;
573 nr = sunhme_get_tx_ring_count(s);
574 cr = sunhme_get_tx_ring_nr(s);
575
576 pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4);
577 pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4);
578
579 xmit_pos = 0;
580 while (status & HME_XD_OWN) {
581 trace_sunhme_tx_desc(buffer, status, cr, nr);
582
583 /* Copy data into transmit buffer */
584 addr = buffer;
585 len = status & HME_XD_TXLENMSK;
586
587 if (xmit_pos + len > HME_FIFO_SIZE) {
588 len = HME_FIFO_SIZE - xmit_pos;
589 }
590
591 pci_dma_read(d, addr, &xmit_buffer[xmit_pos], len);
592 xmit_pos += len;
593
594 /* Detect start of packet for TX checksum */
595 if (status & HME_XD_SOP) {
596 sum = 0;
597 csum_offset = (status & HME_XD_TXCSSTART) >> HME_XD_TXCSSTARTSHIFT;
598 csum_stuff_offset = (status & HME_XD_TXCSSTUFF) >>
599 HME_XD_TXCSSTUFFSHIFT;
600 }
601
602 if (status & HME_XD_TXCKSUM) {
603 /* Only start calculation from csum_offset */
604 if (xmit_pos - len <= csum_offset && xmit_pos > csum_offset) {
605 sum += net_checksum_add(xmit_pos - csum_offset,
606 xmit_buffer + csum_offset);
607 trace_sunhme_tx_xsum_add(csum_offset, xmit_pos - csum_offset);
608 } else {
609 sum += net_checksum_add(len, xmit_buffer + xmit_pos - len);
610 trace_sunhme_tx_xsum_add(xmit_pos - len, len);
611 }
612 }
613
614 /* Detect end of packet for TX checksum */
615 if (status & HME_XD_EOP) {
616 /* Stuff the checksum if required */
617 if (status & HME_XD_TXCKSUM) {
618 csum = net_checksum_finish(sum);
619 stw_be_p(xmit_buffer + csum_stuff_offset, csum);
620 trace_sunhme_tx_xsum_stuff(csum, csum_stuff_offset);
621 }
622
623 if (s->macregs[HME_MACI_TXCFG >> 2] & HME_MAC_TXCFG_ENABLE) {
624 sunhme_transmit_frame(s, xmit_buffer, xmit_pos);
625 trace_sunhme_tx_done(xmit_pos);
626 }
627 }
628
629 /* Update status */
630 status &= ~HME_XD_OWN;
631 pci_dma_write(d, tb + cr * HME_DESC_SIZE, &status, 4);
632
633 /* Move onto next descriptor */
634 cr++;
635 if (cr >= nr) {
636 cr = 0;
637 }
638 sunhme_set_tx_ring_nr(s, cr);
639
640 pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4);
641 pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4);
642
643 /* Indicate TX complete */
644 intstatus = s->sebregs[HME_SEBI_STAT >> 2];
645 intstatus |= HME_SEB_STAT_HOSTTOTX;
646 s->sebregs[HME_SEBI_STAT >> 2] = intstatus;
647
648 /* Autoclear TX pending */
649 s->etxregs[HME_ETXI_PENDING >> 2] = 0;
650
651 sunhme_update_irq(s);
652 }
653
654 /* TX FIFO now clear */
655 intstatus = s->sebregs[HME_SEBI_STAT >> 2];
656 intstatus |= HME_SEB_STAT_TXALL;
657 s->sebregs[HME_SEBI_STAT >> 2] = intstatus;
658 sunhme_update_irq(s);
659 }
660
sunhme_can_receive(NetClientState * nc)661 static bool sunhme_can_receive(NetClientState *nc)
662 {
663 SunHMEState *s = qemu_get_nic_opaque(nc);
664
665 return !!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE);
666 }
667
sunhme_link_status_changed(NetClientState * nc)668 static void sunhme_link_status_changed(NetClientState *nc)
669 {
670 SunHMEState *s = qemu_get_nic_opaque(nc);
671
672 if (nc->link_down) {
673 s->miiregs[MII_ANLPAR] &= ~MII_ANLPAR_TXFD;
674 s->miiregs[MII_BMSR] &= ~MII_BMSR_LINK_ST;
675 } else {
676 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD;
677 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST;
678 }
679
680 /* Exact bits unknown */
681 s->mifregs[HME_MIFI_STAT >> 2] = 0xffff;
682 sunhme_update_irq(s);
683 }
684
sunhme_get_rx_ring_count(SunHMEState * s)685 static inline int sunhme_get_rx_ring_count(SunHMEState *s)
686 {
687 uint32_t rings = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_RINGSIZE)
688 >> HME_ERX_CFG_RINGSIZE_SHIFT;
689
690 switch (rings) {
691 case 0:
692 return 32;
693 case 1:
694 return 64;
695 case 2:
696 return 128;
697 case 3:
698 return 256;
699 }
700
701 return 0;
702 }
703
sunhme_get_rx_ring_nr(SunHMEState * s)704 static inline int sunhme_get_rx_ring_nr(SunHMEState *s)
705 {
706 return s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_OFFSET;
707 }
708
sunhme_set_rx_ring_nr(SunHMEState * s,int i)709 static inline void sunhme_set_rx_ring_nr(SunHMEState *s, int i)
710 {
711 uint32_t ring = s->erxregs[HME_ERXI_RING >> 2] & ~HME_ERXI_RING_OFFSET;
712 ring |= i & HME_ERXI_RING_OFFSET;
713
714 s->erxregs[HME_ERXI_RING >> 2] = ring;
715 }
716
sunhme_receive(NetClientState * nc,const uint8_t * buf,size_t size)717 static ssize_t sunhme_receive(NetClientState *nc, const uint8_t *buf,
718 size_t size)
719 {
720 SunHMEState *s = qemu_get_nic_opaque(nc);
721 PCIDevice *d = PCI_DEVICE(s);
722 dma_addr_t rb, addr;
723 uint32_t intstatus, status, buffer, buffersize, sum;
724 uint16_t csum;
725 int nr, cr, len, rxoffset, csum_offset;
726
727 trace_sunhme_rx_incoming(size);
728
729 /* Do nothing if MAC RX disabled */
730 if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE)) {
731 return 0;
732 }
733
734 trace_sunhme_rx_filter_destmac(buf[0], buf[1], buf[2],
735 buf[3], buf[4], buf[5]);
736
737 /* Check destination MAC address */
738 if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_PMISC)) {
739 /* Try and match local MAC address */
740 if (((s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff00) >> 8) == buf[0] &&
741 (s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff) == buf[1] &&
742 ((s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff00) >> 8) == buf[2] &&
743 (s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff) == buf[3] &&
744 ((s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff00) >> 8) == buf[4] &&
745 (s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff) == buf[5]) {
746 /* Matched local MAC address */
747 trace_sunhme_rx_filter_local_match();
748 } else if (buf[0] == 0xff && buf[1] == 0xff && buf[2] == 0xff &&
749 buf[3] == 0xff && buf[4] == 0xff && buf[5] == 0xff) {
750 /* Matched broadcast address */
751 trace_sunhme_rx_filter_bcast_match();
752 } else if (s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_HENABLE) {
753 /* Didn't match local address, check hash filter */
754 int mcast_idx = net_crc32_le(buf, ETH_ALEN) >> 26;
755 if (!(s->macregs[(HME_MACI_HASHTAB0 >> 2) - (mcast_idx >> 4)] &
756 (1 << (mcast_idx & 0xf)))) {
757 /* Didn't match hash filter */
758 trace_sunhme_rx_filter_hash_nomatch();
759 trace_sunhme_rx_filter_reject();
760 return -1;
761 } else {
762 trace_sunhme_rx_filter_hash_match();
763 }
764 } else {
765 /* Not for us */
766 trace_sunhme_rx_filter_reject();
767 return -1;
768 }
769 } else {
770 trace_sunhme_rx_filter_promisc_match();
771 }
772
773 trace_sunhme_rx_filter_accept();
774
775 rb = s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_ADDR;
776 nr = sunhme_get_rx_ring_count(s);
777 cr = sunhme_get_rx_ring_nr(s);
778
779 pci_dma_read(d, rb + cr * HME_DESC_SIZE, &status, 4);
780 pci_dma_read(d, rb + cr * HME_DESC_SIZE + 4, &buffer, 4);
781
782 /* If we don't own the current descriptor then indicate overflow error */
783 if (!(status & HME_XD_OWN)) {
784 s->sebregs[HME_SEBI_STAT >> 2] |= HME_SEB_STAT_NORXD;
785 sunhme_update_irq(s);
786 trace_sunhme_rx_norxd();
787 return -1;
788 }
789
790 rxoffset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_BYTEOFFSET) >>
791 HME_ERX_CFG_BYTEOFFSET_SHIFT;
792
793 addr = buffer + rxoffset;
794 buffersize = (status & HME_XD_RXLENMSK) >> HME_XD_RXLENSHIFT;
795
796 /* Detect receive overflow */
797 len = size;
798 if (size > buffersize) {
799 status |= HME_XD_OFL;
800 len = buffersize;
801 }
802
803 pci_dma_write(d, addr, buf, len);
804
805 trace_sunhme_rx_desc(buffer, rxoffset, status, len, cr, nr);
806
807 /* Calculate the receive checksum */
808 csum_offset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_CSUMSTART) >>
809 HME_ERX_CFG_CSUMSHIFT << 1;
810 sum = 0;
811 sum += net_checksum_add(len - csum_offset, (uint8_t *)buf + csum_offset);
812 csum = net_checksum_finish(sum);
813
814 trace_sunhme_rx_xsum_calc(csum);
815
816 /* Update status */
817 status &= ~HME_XD_OWN;
818 status &= ~HME_XD_RXLENMSK;
819 status |= len << HME_XD_RXLENSHIFT;
820 status &= ~HME_XD_RXCKSUM;
821 status |= csum;
822
823 pci_dma_write(d, rb + cr * HME_DESC_SIZE, &status, 4);
824
825 cr++;
826 if (cr >= nr) {
827 cr = 0;
828 }
829
830 sunhme_set_rx_ring_nr(s, cr);
831
832 /* Indicate RX complete */
833 intstatus = s->sebregs[HME_SEBI_STAT >> 2];
834 intstatus |= HME_SEB_STAT_RXTOHOST;
835 s->sebregs[HME_SEBI_STAT >> 2] = intstatus;
836
837 sunhme_update_irq(s);
838
839 return len;
840 }
841
842 static NetClientInfo net_sunhme_info = {
843 .type = NET_CLIENT_DRIVER_NIC,
844 .size = sizeof(NICState),
845 .can_receive = sunhme_can_receive,
846 .receive = sunhme_receive,
847 .link_status_changed = sunhme_link_status_changed,
848 };
849
sunhme_realize(PCIDevice * pci_dev,Error ** errp)850 static void sunhme_realize(PCIDevice *pci_dev, Error **errp)
851 {
852 SunHMEState *s = SUNHME(pci_dev);
853 DeviceState *d = DEVICE(pci_dev);
854 uint8_t *pci_conf;
855
856 pci_conf = pci_dev->config;
857 pci_conf[PCI_INTERRUPT_PIN] = 1; /* interrupt pin A */
858
859 memory_region_init(&s->hme, OBJECT(pci_dev), "sunhme", HME_REG_SIZE);
860 pci_register_bar(pci_dev, 0, PCI_BASE_ADDRESS_SPACE_MEMORY, &s->hme);
861
862 memory_region_init_io(&s->sebreg, OBJECT(pci_dev), &sunhme_seb_ops, s,
863 "sunhme.seb", HME_SEB_REG_SIZE);
864 memory_region_add_subregion(&s->hme, 0, &s->sebreg);
865
866 memory_region_init_io(&s->etxreg, OBJECT(pci_dev), &sunhme_etx_ops, s,
867 "sunhme.etx", HME_ETX_REG_SIZE);
868 memory_region_add_subregion(&s->hme, 0x2000, &s->etxreg);
869
870 memory_region_init_io(&s->erxreg, OBJECT(pci_dev), &sunhme_erx_ops, s,
871 "sunhme.erx", HME_ERX_REG_SIZE);
872 memory_region_add_subregion(&s->hme, 0x4000, &s->erxreg);
873
874 memory_region_init_io(&s->macreg, OBJECT(pci_dev), &sunhme_mac_ops, s,
875 "sunhme.mac", HME_MAC_REG_SIZE);
876 memory_region_add_subregion(&s->hme, 0x6000, &s->macreg);
877
878 memory_region_init_io(&s->mifreg, OBJECT(pci_dev), &sunhme_mif_ops, s,
879 "sunhme.mif", HME_MIF_REG_SIZE);
880 memory_region_add_subregion(&s->hme, 0x7000, &s->mifreg);
881
882 qemu_macaddr_default_if_unset(&s->conf.macaddr);
883 s->nic = qemu_new_nic(&net_sunhme_info, &s->conf,
884 object_get_typename(OBJECT(d)), d->id,
885 &d->mem_reentrancy_guard, s);
886 qemu_format_nic_info_str(qemu_get_queue(s->nic), s->conf.macaddr.a);
887 }
888
sunhme_instance_init(Object * obj)889 static void sunhme_instance_init(Object *obj)
890 {
891 SunHMEState *s = SUNHME(obj);
892
893 device_add_bootindex_property(obj, &s->conf.bootindex,
894 "bootindex", "/ethernet-phy@0",
895 DEVICE(obj));
896 }
897
sunhme_reset(DeviceState * ds)898 static void sunhme_reset(DeviceState *ds)
899 {
900 SunHMEState *s = SUNHME(ds);
901
902 /* Configure internal transceiver */
903 s->mifregs[HME_MIFI_CFG >> 2] |= HME_MIF_CFG_MDI0;
904
905 /* Advertise auto, 100Mbps FD */
906 s->miiregs[MII_ANAR] = MII_ANAR_TXFD;
907 s->miiregs[MII_BMSR] = MII_BMSR_AUTONEG | MII_BMSR_100TX_FD |
908 MII_BMSR_AN_COMP;
909
910 if (!qemu_get_queue(s->nic)->link_down) {
911 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD;
912 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST;
913 }
914
915 /* Set manufacturer */
916 s->miiregs[MII_PHYID1] = DP83840_PHYID1;
917 s->miiregs[MII_PHYID2] = DP83840_PHYID2;
918
919 /* Configure default interrupt mask */
920 s->mifregs[HME_MIFI_IMASK >> 2] = 0xffff;
921 s->sebregs[HME_SEBI_IMASK >> 2] = 0xff7fffff;
922 }
923
924 static const VMStateDescription vmstate_hme = {
925 .name = "sunhme",
926 .version_id = 0,
927 .minimum_version_id = 0,
928 .fields = (const VMStateField[]) {
929 VMSTATE_PCI_DEVICE(parent_obj, SunHMEState),
930 VMSTATE_MACADDR(conf.macaddr, SunHMEState),
931 VMSTATE_UINT32_ARRAY(sebregs, SunHMEState, (HME_SEB_REG_SIZE >> 2)),
932 VMSTATE_UINT32_ARRAY(etxregs, SunHMEState, (HME_ETX_REG_SIZE >> 2)),
933 VMSTATE_UINT32_ARRAY(erxregs, SunHMEState, (HME_ERX_REG_SIZE >> 2)),
934 VMSTATE_UINT32_ARRAY(macregs, SunHMEState, (HME_MAC_REG_SIZE >> 2)),
935 VMSTATE_UINT32_ARRAY(mifregs, SunHMEState, (HME_MIF_REG_SIZE >> 2)),
936 VMSTATE_UINT16_ARRAY(miiregs, SunHMEState, HME_MII_REGS_SIZE),
937 VMSTATE_END_OF_LIST()
938 }
939 };
940
sunhme_class_init(ObjectClass * klass,void * data)941 static void sunhme_class_init(ObjectClass *klass, void *data)
942 {
943 DeviceClass *dc = DEVICE_CLASS(klass);
944 PCIDeviceClass *k = PCI_DEVICE_CLASS(klass);
945
946 k->realize = sunhme_realize;
947 k->vendor_id = PCI_VENDOR_ID_SUN;
948 k->device_id = PCI_DEVICE_ID_SUN_HME;
949 k->class_id = PCI_CLASS_NETWORK_ETHERNET;
950 dc->vmsd = &vmstate_hme;
951 dc->reset = sunhme_reset;
952 device_class_set_props(dc, sunhme_properties);
953 set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
954 }
955
956 static const TypeInfo sunhme_info = {
957 .name = TYPE_SUNHME,
958 .parent = TYPE_PCI_DEVICE,
959 .class_init = sunhme_class_init,
960 .instance_size = sizeof(SunHMEState),
961 .instance_init = sunhme_instance_init,
962 .interfaces = (InterfaceInfo[]) {
963 { INTERFACE_CONVENTIONAL_PCI_DEVICE },
964 { }
965 }
966 };
967
sunhme_register_types(void)968 static void sunhme_register_types(void)
969 {
970 type_register_static(&sunhme_info);
971 }
972
973 type_init(sunhme_register_types)
974