1 /* 2 * QEMU Sun Happy Meal Ethernet emulation 3 * 4 * Copyright (c) 2017 Mark Cave-Ayland 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a copy 7 * of this software and associated documentation files (the "Software"), to deal 8 * in the Software without restriction, including without limitation the rights 9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 10 * copies of the Software, and to permit persons to whom the Software is 11 * furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 22 * THE SOFTWARE. 23 */ 24 25 #include "qemu/osdep.h" 26 #include "hw/pci/pci.h" 27 #include "hw/qdev-properties.h" 28 #include "migration/vmstate.h" 29 #include "hw/net/mii.h" 30 #include "net/net.h" 31 #include "qemu/module.h" 32 #include "net/checksum.h" 33 #include "net/eth.h" 34 #include "sysemu/sysemu.h" 35 #include "trace.h" 36 37 #define HME_REG_SIZE 0x8000 38 39 #define HME_SEB_REG_SIZE 0x2000 40 41 #define HME_SEBI_RESET 0x0 42 #define HME_SEB_RESET_ETX 0x1 43 #define HME_SEB_RESET_ERX 0x2 44 45 #define HME_SEBI_STAT 0x100 46 #define HME_SEBI_STAT_LINUXBUG 0x108 47 #define HME_SEB_STAT_RXTOHOST 0x10000 48 #define HME_SEB_STAT_NORXD 0x20000 49 #define HME_SEB_STAT_MIFIRQ 0x800000 50 #define HME_SEB_STAT_HOSTTOTX 0x1000000 51 #define HME_SEB_STAT_TXALL 0x2000000 52 53 #define HME_SEBI_IMASK 0x104 54 #define HME_SEBI_IMASK_LINUXBUG 0x10c 55 56 #define HME_ETX_REG_SIZE 0x2000 57 58 #define HME_ETXI_PENDING 0x0 59 60 #define HME_ETXI_RING 0x8 61 #define HME_ETXI_RING_ADDR 0xffffff00 62 #define HME_ETXI_RING_OFFSET 0xff 63 64 #define HME_ETXI_RSIZE 0x2c 65 66 #define HME_ERX_REG_SIZE 0x2000 67 68 #define HME_ERXI_CFG 0x0 69 #define HME_ERX_CFG_RINGSIZE 0x600 70 #define HME_ERX_CFG_RINGSIZE_SHIFT 9 71 #define HME_ERX_CFG_BYTEOFFSET 0x38 72 #define HME_ERX_CFG_BYTEOFFSET_SHIFT 3 73 #define HME_ERX_CFG_CSUMSTART 0x7f0000 74 #define HME_ERX_CFG_CSUMSHIFT 16 75 76 #define HME_ERXI_RING 0x4 77 #define HME_ERXI_RING_ADDR 0xffffff00 78 #define HME_ERXI_RING_OFFSET 0xff 79 80 #define HME_MAC_REG_SIZE 0x1000 81 82 #define HME_MACI_TXCFG 0x20c 83 #define HME_MAC_TXCFG_ENABLE 0x1 84 85 #define HME_MACI_RXCFG 0x30c 86 #define HME_MAC_RXCFG_ENABLE 0x1 87 #define HME_MAC_RXCFG_PMISC 0x40 88 #define HME_MAC_RXCFG_HENABLE 0x800 89 90 #define HME_MACI_MACADDR2 0x318 91 #define HME_MACI_MACADDR1 0x31c 92 #define HME_MACI_MACADDR0 0x320 93 94 #define HME_MACI_HASHTAB3 0x340 95 #define HME_MACI_HASHTAB2 0x344 96 #define HME_MACI_HASHTAB1 0x348 97 #define HME_MACI_HASHTAB0 0x34c 98 99 #define HME_MIF_REG_SIZE 0x20 100 101 #define HME_MIFI_FO 0xc 102 #define HME_MIF_FO_ST 0xc0000000 103 #define HME_MIF_FO_ST_SHIFT 30 104 #define HME_MIF_FO_OPC 0x30000000 105 #define HME_MIF_FO_OPC_SHIFT 28 106 #define HME_MIF_FO_PHYAD 0x0f800000 107 #define HME_MIF_FO_PHYAD_SHIFT 23 108 #define HME_MIF_FO_REGAD 0x007c0000 109 #define HME_MIF_FO_REGAD_SHIFT 18 110 #define HME_MIF_FO_TAMSB 0x20000 111 #define HME_MIF_FO_TALSB 0x10000 112 #define HME_MIF_FO_DATA 0xffff 113 114 #define HME_MIFI_CFG 0x10 115 #define HME_MIF_CFG_MDI0 0x100 116 #define HME_MIF_CFG_MDI1 0x200 117 118 #define HME_MIFI_IMASK 0x14 119 120 #define HME_MIFI_STAT 0x18 121 122 123 /* Wired HME PHY addresses */ 124 #define HME_PHYAD_INTERNAL 1 125 #define HME_PHYAD_EXTERNAL 0 126 127 #define MII_COMMAND_START 0x1 128 #define MII_COMMAND_READ 0x2 129 #define MII_COMMAND_WRITE 0x1 130 131 #define TYPE_SUNHME "sunhme" 132 #define SUNHME(obj) OBJECT_CHECK(SunHMEState, (obj), TYPE_SUNHME) 133 134 /* Maximum size of buffer */ 135 #define HME_FIFO_SIZE 0x800 136 137 /* Size of TX/RX descriptor */ 138 #define HME_DESC_SIZE 0x8 139 140 #define HME_XD_OWN 0x80000000 141 #define HME_XD_OFL 0x40000000 142 #define HME_XD_SOP 0x40000000 143 #define HME_XD_EOP 0x20000000 144 #define HME_XD_RXLENMSK 0x3fff0000 145 #define HME_XD_RXLENSHIFT 16 146 #define HME_XD_RXCKSUM 0xffff 147 #define HME_XD_TXLENMSK 0x00001fff 148 #define HME_XD_TXCKSUM 0x10000000 149 #define HME_XD_TXCSSTUFF 0xff00000 150 #define HME_XD_TXCSSTUFFSHIFT 20 151 #define HME_XD_TXCSSTART 0xfc000 152 #define HME_XD_TXCSSTARTSHIFT 14 153 154 #define HME_MII_REGS_SIZE 0x20 155 156 typedef struct SunHMEState { 157 /*< private >*/ 158 PCIDevice parent_obj; 159 160 NICState *nic; 161 NICConf conf; 162 163 MemoryRegion hme; 164 MemoryRegion sebreg; 165 MemoryRegion etxreg; 166 MemoryRegion erxreg; 167 MemoryRegion macreg; 168 MemoryRegion mifreg; 169 170 uint32_t sebregs[HME_SEB_REG_SIZE >> 2]; 171 uint32_t etxregs[HME_ETX_REG_SIZE >> 2]; 172 uint32_t erxregs[HME_ERX_REG_SIZE >> 2]; 173 uint32_t macregs[HME_MAC_REG_SIZE >> 2]; 174 uint32_t mifregs[HME_MIF_REG_SIZE >> 2]; 175 176 uint16_t miiregs[HME_MII_REGS_SIZE]; 177 } SunHMEState; 178 179 static Property sunhme_properties[] = { 180 DEFINE_NIC_PROPERTIES(SunHMEState, conf), 181 DEFINE_PROP_END_OF_LIST(), 182 }; 183 184 static void sunhme_reset_tx(SunHMEState *s) 185 { 186 /* Indicate TX reset complete */ 187 s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ETX; 188 } 189 190 static void sunhme_reset_rx(SunHMEState *s) 191 { 192 /* Indicate RX reset complete */ 193 s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ERX; 194 } 195 196 static void sunhme_update_irq(SunHMEState *s) 197 { 198 PCIDevice *d = PCI_DEVICE(s); 199 int level; 200 201 /* MIF interrupt mask (16-bit) */ 202 uint32_t mifmask = ~(s->mifregs[HME_MIFI_IMASK >> 2]) & 0xffff; 203 uint32_t mif = s->mifregs[HME_MIFI_STAT >> 2] & mifmask; 204 205 /* Main SEB interrupt mask (include MIF status from above) */ 206 uint32_t sebmask = ~(s->sebregs[HME_SEBI_IMASK >> 2]) & 207 ~HME_SEB_STAT_MIFIRQ; 208 uint32_t seb = s->sebregs[HME_SEBI_STAT >> 2] & sebmask; 209 if (mif) { 210 seb |= HME_SEB_STAT_MIFIRQ; 211 } 212 213 level = (seb ? 1 : 0); 214 trace_sunhme_update_irq(mifmask, mif, sebmask, seb, level); 215 216 pci_set_irq(d, level); 217 } 218 219 static void sunhme_seb_write(void *opaque, hwaddr addr, 220 uint64_t val, unsigned size) 221 { 222 SunHMEState *s = SUNHME(opaque); 223 224 trace_sunhme_seb_write(addr, val); 225 226 /* Handly buggy Linux drivers before 4.13 which have 227 the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */ 228 switch (addr) { 229 case HME_SEBI_STAT_LINUXBUG: 230 addr = HME_SEBI_STAT; 231 break; 232 case HME_SEBI_IMASK_LINUXBUG: 233 addr = HME_SEBI_IMASK; 234 break; 235 default: 236 break; 237 } 238 239 switch (addr) { 240 case HME_SEBI_RESET: 241 if (val & HME_SEB_RESET_ETX) { 242 sunhme_reset_tx(s); 243 } 244 if (val & HME_SEB_RESET_ERX) { 245 sunhme_reset_rx(s); 246 } 247 val = s->sebregs[HME_SEBI_RESET >> 2]; 248 break; 249 } 250 251 s->sebregs[addr >> 2] = val; 252 } 253 254 static uint64_t sunhme_seb_read(void *opaque, hwaddr addr, 255 unsigned size) 256 { 257 SunHMEState *s = SUNHME(opaque); 258 uint64_t val; 259 260 /* Handly buggy Linux drivers before 4.13 which have 261 the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */ 262 switch (addr) { 263 case HME_SEBI_STAT_LINUXBUG: 264 addr = HME_SEBI_STAT; 265 break; 266 case HME_SEBI_IMASK_LINUXBUG: 267 addr = HME_SEBI_IMASK; 268 break; 269 default: 270 break; 271 } 272 273 val = s->sebregs[addr >> 2]; 274 275 switch (addr) { 276 case HME_SEBI_STAT: 277 /* Autoclear status (except MIF) */ 278 s->sebregs[HME_SEBI_STAT >> 2] &= HME_SEB_STAT_MIFIRQ; 279 sunhme_update_irq(s); 280 break; 281 } 282 283 trace_sunhme_seb_read(addr, val); 284 285 return val; 286 } 287 288 static const MemoryRegionOps sunhme_seb_ops = { 289 .read = sunhme_seb_read, 290 .write = sunhme_seb_write, 291 .endianness = DEVICE_LITTLE_ENDIAN, 292 .valid = { 293 .min_access_size = 4, 294 .max_access_size = 4, 295 }, 296 }; 297 298 static void sunhme_transmit(SunHMEState *s); 299 300 static void sunhme_etx_write(void *opaque, hwaddr addr, 301 uint64_t val, unsigned size) 302 { 303 SunHMEState *s = SUNHME(opaque); 304 305 trace_sunhme_etx_write(addr, val); 306 307 switch (addr) { 308 case HME_ETXI_PENDING: 309 if (val) { 310 sunhme_transmit(s); 311 } 312 break; 313 } 314 315 s->etxregs[addr >> 2] = val; 316 } 317 318 static uint64_t sunhme_etx_read(void *opaque, hwaddr addr, 319 unsigned size) 320 { 321 SunHMEState *s = SUNHME(opaque); 322 uint64_t val; 323 324 val = s->etxregs[addr >> 2]; 325 326 trace_sunhme_etx_read(addr, val); 327 328 return val; 329 } 330 331 static const MemoryRegionOps sunhme_etx_ops = { 332 .read = sunhme_etx_read, 333 .write = sunhme_etx_write, 334 .endianness = DEVICE_LITTLE_ENDIAN, 335 .valid = { 336 .min_access_size = 4, 337 .max_access_size = 4, 338 }, 339 }; 340 341 static void sunhme_erx_write(void *opaque, hwaddr addr, 342 uint64_t val, unsigned size) 343 { 344 SunHMEState *s = SUNHME(opaque); 345 346 trace_sunhme_erx_write(addr, val); 347 348 s->erxregs[addr >> 2] = val; 349 } 350 351 static uint64_t sunhme_erx_read(void *opaque, hwaddr addr, 352 unsigned size) 353 { 354 SunHMEState *s = SUNHME(opaque); 355 uint64_t val; 356 357 val = s->erxregs[addr >> 2]; 358 359 trace_sunhme_erx_read(addr, val); 360 361 return val; 362 } 363 364 static const MemoryRegionOps sunhme_erx_ops = { 365 .read = sunhme_erx_read, 366 .write = sunhme_erx_write, 367 .endianness = DEVICE_LITTLE_ENDIAN, 368 .valid = { 369 .min_access_size = 4, 370 .max_access_size = 4, 371 }, 372 }; 373 374 static void sunhme_mac_write(void *opaque, hwaddr addr, 375 uint64_t val, unsigned size) 376 { 377 SunHMEState *s = SUNHME(opaque); 378 uint64_t oldval = s->macregs[addr >> 2]; 379 380 trace_sunhme_mac_write(addr, val); 381 382 s->macregs[addr >> 2] = val; 383 384 switch (addr) { 385 case HME_MACI_RXCFG: 386 if (!(oldval & HME_MAC_RXCFG_ENABLE) && 387 (val & HME_MAC_RXCFG_ENABLE)) { 388 qemu_flush_queued_packets(qemu_get_queue(s->nic)); 389 } 390 break; 391 } 392 } 393 394 static uint64_t sunhme_mac_read(void *opaque, hwaddr addr, 395 unsigned size) 396 { 397 SunHMEState *s = SUNHME(opaque); 398 uint64_t val; 399 400 val = s->macregs[addr >> 2]; 401 402 trace_sunhme_mac_read(addr, val); 403 404 return val; 405 } 406 407 static const MemoryRegionOps sunhme_mac_ops = { 408 .read = sunhme_mac_read, 409 .write = sunhme_mac_write, 410 .endianness = DEVICE_LITTLE_ENDIAN, 411 .valid = { 412 .min_access_size = 4, 413 .max_access_size = 4, 414 }, 415 }; 416 417 static void sunhme_mii_write(SunHMEState *s, uint8_t reg, uint16_t data) 418 { 419 trace_sunhme_mii_write(reg, data); 420 421 switch (reg) { 422 case MII_BMCR: 423 if (data & MII_BMCR_RESET) { 424 /* Autoclear reset bit, enable auto negotiation */ 425 data &= ~MII_BMCR_RESET; 426 data |= MII_BMCR_AUTOEN; 427 } 428 if (data & MII_BMCR_ANRESTART) { 429 /* Autoclear auto negotiation restart */ 430 data &= ~MII_BMCR_ANRESTART; 431 432 /* Indicate negotiation complete */ 433 s->miiregs[MII_BMSR] |= MII_BMSR_AN_COMP; 434 435 if (!qemu_get_queue(s->nic)->link_down) { 436 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD; 437 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST; 438 } 439 } 440 break; 441 } 442 443 s->miiregs[reg] = data; 444 } 445 446 static uint16_t sunhme_mii_read(SunHMEState *s, uint8_t reg) 447 { 448 uint16_t data = s->miiregs[reg]; 449 450 trace_sunhme_mii_read(reg, data); 451 452 return data; 453 } 454 455 static void sunhme_mif_write(void *opaque, hwaddr addr, 456 uint64_t val, unsigned size) 457 { 458 SunHMEState *s = SUNHME(opaque); 459 uint8_t cmd, reg; 460 uint16_t data; 461 462 trace_sunhme_mif_write(addr, val); 463 464 switch (addr) { 465 case HME_MIFI_CFG: 466 /* Mask the read-only bits */ 467 val &= ~(HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1); 468 val |= s->mifregs[HME_MIFI_CFG >> 2] & 469 (HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1); 470 break; 471 case HME_MIFI_FO: 472 /* Detect start of MII command */ 473 if ((val & HME_MIF_FO_ST) >> HME_MIF_FO_ST_SHIFT 474 != MII_COMMAND_START) { 475 val |= HME_MIF_FO_TALSB; 476 break; 477 } 478 479 /* Internal phy only */ 480 if ((val & HME_MIF_FO_PHYAD) >> HME_MIF_FO_PHYAD_SHIFT 481 != HME_PHYAD_INTERNAL) { 482 val |= HME_MIF_FO_TALSB; 483 break; 484 } 485 486 cmd = (val & HME_MIF_FO_OPC) >> HME_MIF_FO_OPC_SHIFT; 487 reg = (val & HME_MIF_FO_REGAD) >> HME_MIF_FO_REGAD_SHIFT; 488 data = (val & HME_MIF_FO_DATA); 489 490 switch (cmd) { 491 case MII_COMMAND_WRITE: 492 sunhme_mii_write(s, reg, data); 493 break; 494 495 case MII_COMMAND_READ: 496 val &= ~HME_MIF_FO_DATA; 497 val |= sunhme_mii_read(s, reg); 498 break; 499 } 500 501 val |= HME_MIF_FO_TALSB; 502 break; 503 } 504 505 s->mifregs[addr >> 2] = val; 506 } 507 508 static uint64_t sunhme_mif_read(void *opaque, hwaddr addr, 509 unsigned size) 510 { 511 SunHMEState *s = SUNHME(opaque); 512 uint64_t val; 513 514 val = s->mifregs[addr >> 2]; 515 516 switch (addr) { 517 case HME_MIFI_STAT: 518 /* Autoclear MIF interrupt status */ 519 s->mifregs[HME_MIFI_STAT >> 2] = 0; 520 sunhme_update_irq(s); 521 break; 522 } 523 524 trace_sunhme_mif_read(addr, val); 525 526 return val; 527 } 528 529 static const MemoryRegionOps sunhme_mif_ops = { 530 .read = sunhme_mif_read, 531 .write = sunhme_mif_write, 532 .endianness = DEVICE_LITTLE_ENDIAN, 533 .valid = { 534 .min_access_size = 4, 535 .max_access_size = 4, 536 }, 537 }; 538 539 static void sunhme_transmit_frame(SunHMEState *s, uint8_t *buf, int size) 540 { 541 qemu_send_packet(qemu_get_queue(s->nic), buf, size); 542 } 543 544 static inline int sunhme_get_tx_ring_count(SunHMEState *s) 545 { 546 return (s->etxregs[HME_ETXI_RSIZE >> 2] + 1) << 4; 547 } 548 549 static inline int sunhme_get_tx_ring_nr(SunHMEState *s) 550 { 551 return s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_OFFSET; 552 } 553 554 static inline void sunhme_set_tx_ring_nr(SunHMEState *s, int i) 555 { 556 uint32_t ring = s->etxregs[HME_ETXI_RING >> 2] & ~HME_ETXI_RING_OFFSET; 557 ring |= i & HME_ETXI_RING_OFFSET; 558 559 s->etxregs[HME_ETXI_RING >> 2] = ring; 560 } 561 562 static void sunhme_transmit(SunHMEState *s) 563 { 564 PCIDevice *d = PCI_DEVICE(s); 565 dma_addr_t tb, addr; 566 uint32_t intstatus, status, buffer, sum = 0; 567 int cr, nr, len, xmit_pos, csum_offset = 0, csum_stuff_offset = 0; 568 uint16_t csum = 0; 569 uint8_t xmit_buffer[HME_FIFO_SIZE]; 570 571 tb = s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_ADDR; 572 nr = sunhme_get_tx_ring_count(s); 573 cr = sunhme_get_tx_ring_nr(s); 574 575 pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4); 576 pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4); 577 578 xmit_pos = 0; 579 while (status & HME_XD_OWN) { 580 trace_sunhme_tx_desc(buffer, status, cr, nr); 581 582 /* Copy data into transmit buffer */ 583 addr = buffer; 584 len = status & HME_XD_TXLENMSK; 585 586 if (xmit_pos + len > HME_FIFO_SIZE) { 587 len = HME_FIFO_SIZE - xmit_pos; 588 } 589 590 pci_dma_read(d, addr, &xmit_buffer[xmit_pos], len); 591 xmit_pos += len; 592 593 /* Detect start of packet for TX checksum */ 594 if (status & HME_XD_SOP) { 595 sum = 0; 596 csum_offset = (status & HME_XD_TXCSSTART) >> HME_XD_TXCSSTARTSHIFT; 597 csum_stuff_offset = (status & HME_XD_TXCSSTUFF) >> 598 HME_XD_TXCSSTUFFSHIFT; 599 } 600 601 if (status & HME_XD_TXCKSUM) { 602 /* Only start calculation from csum_offset */ 603 if (xmit_pos - len <= csum_offset && xmit_pos > csum_offset) { 604 sum += net_checksum_add(xmit_pos - csum_offset, 605 xmit_buffer + csum_offset); 606 trace_sunhme_tx_xsum_add(csum_offset, xmit_pos - csum_offset); 607 } else { 608 sum += net_checksum_add(len, xmit_buffer + xmit_pos - len); 609 trace_sunhme_tx_xsum_add(xmit_pos - len, len); 610 } 611 } 612 613 /* Detect end of packet for TX checksum */ 614 if (status & HME_XD_EOP) { 615 /* Stuff the checksum if required */ 616 if (status & HME_XD_TXCKSUM) { 617 csum = net_checksum_finish(sum); 618 stw_be_p(xmit_buffer + csum_stuff_offset, csum); 619 trace_sunhme_tx_xsum_stuff(csum, csum_stuff_offset); 620 } 621 622 if (s->macregs[HME_MACI_TXCFG >> 2] & HME_MAC_TXCFG_ENABLE) { 623 sunhme_transmit_frame(s, xmit_buffer, xmit_pos); 624 trace_sunhme_tx_done(xmit_pos); 625 } 626 } 627 628 /* Update status */ 629 status &= ~HME_XD_OWN; 630 pci_dma_write(d, tb + cr * HME_DESC_SIZE, &status, 4); 631 632 /* Move onto next descriptor */ 633 cr++; 634 if (cr >= nr) { 635 cr = 0; 636 } 637 sunhme_set_tx_ring_nr(s, cr); 638 639 pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4); 640 pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4); 641 642 /* Indicate TX complete */ 643 intstatus = s->sebregs[HME_SEBI_STAT >> 2]; 644 intstatus |= HME_SEB_STAT_HOSTTOTX; 645 s->sebregs[HME_SEBI_STAT >> 2] = intstatus; 646 647 /* Autoclear TX pending */ 648 s->etxregs[HME_ETXI_PENDING >> 2] = 0; 649 650 sunhme_update_irq(s); 651 } 652 653 /* TX FIFO now clear */ 654 intstatus = s->sebregs[HME_SEBI_STAT >> 2]; 655 intstatus |= HME_SEB_STAT_TXALL; 656 s->sebregs[HME_SEBI_STAT >> 2] = intstatus; 657 sunhme_update_irq(s); 658 } 659 660 static int sunhme_can_receive(NetClientState *nc) 661 { 662 SunHMEState *s = qemu_get_nic_opaque(nc); 663 664 return s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE; 665 } 666 667 static void sunhme_link_status_changed(NetClientState *nc) 668 { 669 SunHMEState *s = qemu_get_nic_opaque(nc); 670 671 if (nc->link_down) { 672 s->miiregs[MII_ANLPAR] &= ~MII_ANLPAR_TXFD; 673 s->miiregs[MII_BMSR] &= ~MII_BMSR_LINK_ST; 674 } else { 675 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD; 676 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST; 677 } 678 679 /* Exact bits unknown */ 680 s->mifregs[HME_MIFI_STAT >> 2] = 0xffff; 681 sunhme_update_irq(s); 682 } 683 684 static inline int sunhme_get_rx_ring_count(SunHMEState *s) 685 { 686 uint32_t rings = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_RINGSIZE) 687 >> HME_ERX_CFG_RINGSIZE_SHIFT; 688 689 switch (rings) { 690 case 0: 691 return 32; 692 case 1: 693 return 64; 694 case 2: 695 return 128; 696 case 3: 697 return 256; 698 } 699 700 return 0; 701 } 702 703 static inline int sunhme_get_rx_ring_nr(SunHMEState *s) 704 { 705 return s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_OFFSET; 706 } 707 708 static inline void sunhme_set_rx_ring_nr(SunHMEState *s, int i) 709 { 710 uint32_t ring = s->erxregs[HME_ERXI_RING >> 2] & ~HME_ERXI_RING_OFFSET; 711 ring |= i & HME_ERXI_RING_OFFSET; 712 713 s->erxregs[HME_ERXI_RING >> 2] = ring; 714 } 715 716 #define MIN_BUF_SIZE 60 717 718 static ssize_t sunhme_receive(NetClientState *nc, const uint8_t *buf, 719 size_t size) 720 { 721 SunHMEState *s = qemu_get_nic_opaque(nc); 722 PCIDevice *d = PCI_DEVICE(s); 723 dma_addr_t rb, addr; 724 uint32_t intstatus, status, buffer, buffersize, sum; 725 uint16_t csum; 726 uint8_t buf1[60]; 727 int nr, cr, len, rxoffset, csum_offset; 728 729 trace_sunhme_rx_incoming(size); 730 731 /* Do nothing if MAC RX disabled */ 732 if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE)) { 733 return 0; 734 } 735 736 trace_sunhme_rx_filter_destmac(buf[0], buf[1], buf[2], 737 buf[3], buf[4], buf[5]); 738 739 /* Check destination MAC address */ 740 if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_PMISC)) { 741 /* Try and match local MAC address */ 742 if (((s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff00) >> 8) == buf[0] && 743 (s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff) == buf[1] && 744 ((s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff00) >> 8) == buf[2] && 745 (s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff) == buf[3] && 746 ((s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff00) >> 8) == buf[4] && 747 (s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff) == buf[5]) { 748 /* Matched local MAC address */ 749 trace_sunhme_rx_filter_local_match(); 750 } else if (buf[0] == 0xff && buf[1] == 0xff && buf[2] == 0xff && 751 buf[3] == 0xff && buf[4] == 0xff && buf[5] == 0xff) { 752 /* Matched broadcast address */ 753 trace_sunhme_rx_filter_bcast_match(); 754 } else if (s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_HENABLE) { 755 /* Didn't match local address, check hash filter */ 756 int mcast_idx = net_crc32_le(buf, ETH_ALEN) >> 26; 757 if (!(s->macregs[(HME_MACI_HASHTAB0 >> 2) - (mcast_idx >> 4)] & 758 (1 << (mcast_idx & 0xf)))) { 759 /* Didn't match hash filter */ 760 trace_sunhme_rx_filter_hash_nomatch(); 761 trace_sunhme_rx_filter_reject(); 762 return -1; 763 } else { 764 trace_sunhme_rx_filter_hash_match(); 765 } 766 } else { 767 /* Not for us */ 768 trace_sunhme_rx_filter_reject(); 769 return -1; 770 } 771 } else { 772 trace_sunhme_rx_filter_promisc_match(); 773 } 774 775 trace_sunhme_rx_filter_accept(); 776 777 /* If too small buffer, then expand it */ 778 if (size < MIN_BUF_SIZE) { 779 memcpy(buf1, buf, size); 780 memset(buf1 + size, 0, MIN_BUF_SIZE - size); 781 buf = buf1; 782 size = MIN_BUF_SIZE; 783 } 784 785 rb = s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_ADDR; 786 nr = sunhme_get_rx_ring_count(s); 787 cr = sunhme_get_rx_ring_nr(s); 788 789 pci_dma_read(d, rb + cr * HME_DESC_SIZE, &status, 4); 790 pci_dma_read(d, rb + cr * HME_DESC_SIZE + 4, &buffer, 4); 791 792 /* If we don't own the current descriptor then indicate overflow error */ 793 if (!(status & HME_XD_OWN)) { 794 s->sebregs[HME_SEBI_STAT >> 2] |= HME_SEB_STAT_NORXD; 795 sunhme_update_irq(s); 796 trace_sunhme_rx_norxd(); 797 return -1; 798 } 799 800 rxoffset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_BYTEOFFSET) >> 801 HME_ERX_CFG_BYTEOFFSET_SHIFT; 802 803 addr = buffer + rxoffset; 804 buffersize = (status & HME_XD_RXLENMSK) >> HME_XD_RXLENSHIFT; 805 806 /* Detect receive overflow */ 807 len = size; 808 if (size > buffersize) { 809 status |= HME_XD_OFL; 810 len = buffersize; 811 } 812 813 pci_dma_write(d, addr, buf, len); 814 815 trace_sunhme_rx_desc(buffer, rxoffset, status, len, cr, nr); 816 817 /* Calculate the receive checksum */ 818 csum_offset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_CSUMSTART) >> 819 HME_ERX_CFG_CSUMSHIFT << 1; 820 sum = 0; 821 sum += net_checksum_add(len - csum_offset, (uint8_t *)buf + csum_offset); 822 csum = net_checksum_finish(sum); 823 824 trace_sunhme_rx_xsum_calc(csum); 825 826 /* Update status */ 827 status &= ~HME_XD_OWN; 828 status &= ~HME_XD_RXLENMSK; 829 status |= len << HME_XD_RXLENSHIFT; 830 status &= ~HME_XD_RXCKSUM; 831 status |= csum; 832 833 pci_dma_write(d, rb + cr * HME_DESC_SIZE, &status, 4); 834 835 cr++; 836 if (cr >= nr) { 837 cr = 0; 838 } 839 840 sunhme_set_rx_ring_nr(s, cr); 841 842 /* Indicate RX complete */ 843 intstatus = s->sebregs[HME_SEBI_STAT >> 2]; 844 intstatus |= HME_SEB_STAT_RXTOHOST; 845 s->sebregs[HME_SEBI_STAT >> 2] = intstatus; 846 847 sunhme_update_irq(s); 848 849 return len; 850 } 851 852 static NetClientInfo net_sunhme_info = { 853 .type = NET_CLIENT_DRIVER_NIC, 854 .size = sizeof(NICState), 855 .can_receive = sunhme_can_receive, 856 .receive = sunhme_receive, 857 .link_status_changed = sunhme_link_status_changed, 858 }; 859 860 static void sunhme_realize(PCIDevice *pci_dev, Error **errp) 861 { 862 SunHMEState *s = SUNHME(pci_dev); 863 DeviceState *d = DEVICE(pci_dev); 864 uint8_t *pci_conf; 865 866 pci_conf = pci_dev->config; 867 pci_conf[PCI_INTERRUPT_PIN] = 1; /* interrupt pin A */ 868 869 memory_region_init(&s->hme, OBJECT(pci_dev), "sunhme", HME_REG_SIZE); 870 pci_register_bar(pci_dev, 0, PCI_BASE_ADDRESS_SPACE_MEMORY, &s->hme); 871 872 memory_region_init_io(&s->sebreg, OBJECT(pci_dev), &sunhme_seb_ops, s, 873 "sunhme.seb", HME_SEB_REG_SIZE); 874 memory_region_add_subregion(&s->hme, 0, &s->sebreg); 875 876 memory_region_init_io(&s->etxreg, OBJECT(pci_dev), &sunhme_etx_ops, s, 877 "sunhme.etx", HME_ETX_REG_SIZE); 878 memory_region_add_subregion(&s->hme, 0x2000, &s->etxreg); 879 880 memory_region_init_io(&s->erxreg, OBJECT(pci_dev), &sunhme_erx_ops, s, 881 "sunhme.erx", HME_ERX_REG_SIZE); 882 memory_region_add_subregion(&s->hme, 0x4000, &s->erxreg); 883 884 memory_region_init_io(&s->macreg, OBJECT(pci_dev), &sunhme_mac_ops, s, 885 "sunhme.mac", HME_MAC_REG_SIZE); 886 memory_region_add_subregion(&s->hme, 0x6000, &s->macreg); 887 888 memory_region_init_io(&s->mifreg, OBJECT(pci_dev), &sunhme_mif_ops, s, 889 "sunhme.mif", HME_MIF_REG_SIZE); 890 memory_region_add_subregion(&s->hme, 0x7000, &s->mifreg); 891 892 qemu_macaddr_default_if_unset(&s->conf.macaddr); 893 s->nic = qemu_new_nic(&net_sunhme_info, &s->conf, 894 object_get_typename(OBJECT(d)), d->id, s); 895 qemu_format_nic_info_str(qemu_get_queue(s->nic), s->conf.macaddr.a); 896 } 897 898 static void sunhme_instance_init(Object *obj) 899 { 900 SunHMEState *s = SUNHME(obj); 901 902 device_add_bootindex_property(obj, &s->conf.bootindex, 903 "bootindex", "/ethernet-phy@0", 904 DEVICE(obj), NULL); 905 } 906 907 static void sunhme_reset(DeviceState *ds) 908 { 909 SunHMEState *s = SUNHME(ds); 910 911 /* Configure internal transceiver */ 912 s->mifregs[HME_MIFI_CFG >> 2] |= HME_MIF_CFG_MDI0; 913 914 /* Advetise auto, 100Mbps FD */ 915 s->miiregs[MII_ANAR] = MII_ANAR_TXFD; 916 s->miiregs[MII_BMSR] = MII_BMSR_AUTONEG | MII_BMSR_100TX_FD | 917 MII_BMSR_AN_COMP; 918 919 if (!qemu_get_queue(s->nic)->link_down) { 920 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD; 921 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST; 922 } 923 924 /* Set manufacturer */ 925 s->miiregs[MII_PHYID1] = DP83840_PHYID1; 926 s->miiregs[MII_PHYID2] = DP83840_PHYID2; 927 928 /* Configure default interrupt mask */ 929 s->mifregs[HME_MIFI_IMASK >> 2] = 0xffff; 930 s->sebregs[HME_SEBI_IMASK >> 2] = 0xff7fffff; 931 } 932 933 static const VMStateDescription vmstate_hme = { 934 .name = "sunhme", 935 .version_id = 0, 936 .minimum_version_id = 0, 937 .fields = (VMStateField[]) { 938 VMSTATE_PCI_DEVICE(parent_obj, SunHMEState), 939 VMSTATE_MACADDR(conf.macaddr, SunHMEState), 940 VMSTATE_UINT32_ARRAY(sebregs, SunHMEState, (HME_SEB_REG_SIZE >> 2)), 941 VMSTATE_UINT32_ARRAY(etxregs, SunHMEState, (HME_ETX_REG_SIZE >> 2)), 942 VMSTATE_UINT32_ARRAY(erxregs, SunHMEState, (HME_ERX_REG_SIZE >> 2)), 943 VMSTATE_UINT32_ARRAY(macregs, SunHMEState, (HME_MAC_REG_SIZE >> 2)), 944 VMSTATE_UINT32_ARRAY(mifregs, SunHMEState, (HME_MIF_REG_SIZE >> 2)), 945 VMSTATE_UINT16_ARRAY(miiregs, SunHMEState, HME_MII_REGS_SIZE), 946 VMSTATE_END_OF_LIST() 947 } 948 }; 949 950 static void sunhme_class_init(ObjectClass *klass, void *data) 951 { 952 DeviceClass *dc = DEVICE_CLASS(klass); 953 PCIDeviceClass *k = PCI_DEVICE_CLASS(klass); 954 955 k->realize = sunhme_realize; 956 k->vendor_id = PCI_VENDOR_ID_SUN; 957 k->device_id = PCI_DEVICE_ID_SUN_HME; 958 k->class_id = PCI_CLASS_NETWORK_ETHERNET; 959 dc->vmsd = &vmstate_hme; 960 dc->reset = sunhme_reset; 961 device_class_set_props(dc, sunhme_properties); 962 set_bit(DEVICE_CATEGORY_NETWORK, dc->categories); 963 } 964 965 static const TypeInfo sunhme_info = { 966 .name = TYPE_SUNHME, 967 .parent = TYPE_PCI_DEVICE, 968 .class_init = sunhme_class_init, 969 .instance_size = sizeof(SunHMEState), 970 .instance_init = sunhme_instance_init, 971 .interfaces = (InterfaceInfo[]) { 972 { INTERFACE_CONVENTIONAL_PCI_DEVICE }, 973 { } 974 } 975 }; 976 977 static void sunhme_register_types(void) 978 { 979 type_register_static(&sunhme_info); 980 } 981 982 type_init(sunhme_register_types) 983