1 // SPDX-License-Identifier: (GPL-2.0 OR MIT) 2 /* 3 * Microsemi SoCs FDMA driver 4 * 5 * Copyright (c) 2021 Microchip 6 * 7 * Page recycling code is mostly taken from gianfar driver. 8 */ 9 10 #include <linux/align.h> 11 #include <linux/bitops.h> 12 #include <linux/dmapool.h> 13 #include <linux/dsa/ocelot.h> 14 #include <linux/netdevice.h> 15 #include <linux/of_platform.h> 16 #include <linux/skbuff.h> 17 18 #include "ocelot_fdma.h" 19 #include "ocelot_qs.h" 20 21 DEFINE_STATIC_KEY_FALSE(ocelot_fdma_enabled); 22 23 static void ocelot_fdma_writel(struct ocelot *ocelot, u32 reg, u32 data) 24 { 25 regmap_write(ocelot->targets[FDMA], reg, data); 26 } 27 28 static u32 ocelot_fdma_readl(struct ocelot *ocelot, u32 reg) 29 { 30 u32 retval; 31 32 regmap_read(ocelot->targets[FDMA], reg, &retval); 33 34 return retval; 35 } 36 37 static dma_addr_t ocelot_fdma_idx_dma(dma_addr_t base, u16 idx) 38 { 39 return base + idx * sizeof(struct ocelot_fdma_dcb); 40 } 41 42 static u16 ocelot_fdma_dma_idx(dma_addr_t base, dma_addr_t dma) 43 { 44 return (dma - base) / sizeof(struct ocelot_fdma_dcb); 45 } 46 47 static u16 ocelot_fdma_idx_next(u16 idx, u16 ring_sz) 48 { 49 return unlikely(idx == ring_sz - 1) ? 0 : idx + 1; 50 } 51 52 static u16 ocelot_fdma_idx_prev(u16 idx, u16 ring_sz) 53 { 54 return unlikely(idx == 0) ? ring_sz - 1 : idx - 1; 55 } 56 57 static int ocelot_fdma_rx_ring_free(struct ocelot_fdma *fdma) 58 { 59 struct ocelot_fdma_rx_ring *rx_ring = &fdma->rx_ring; 60 61 if (rx_ring->next_to_use >= rx_ring->next_to_clean) 62 return OCELOT_FDMA_RX_RING_SIZE - 63 (rx_ring->next_to_use - rx_ring->next_to_clean) - 1; 64 else 65 return rx_ring->next_to_clean - rx_ring->next_to_use - 1; 66 } 67 68 static int ocelot_fdma_tx_ring_free(struct ocelot_fdma *fdma) 69 { 70 struct ocelot_fdma_tx_ring *tx_ring = &fdma->tx_ring; 71 72 if (tx_ring->next_to_use >= tx_ring->next_to_clean) 73 return OCELOT_FDMA_TX_RING_SIZE - 74 (tx_ring->next_to_use - tx_ring->next_to_clean) - 1; 75 else 76 return tx_ring->next_to_clean - tx_ring->next_to_use - 1; 77 } 78 79 static bool ocelot_fdma_tx_ring_empty(struct ocelot_fdma *fdma) 80 { 81 struct ocelot_fdma_tx_ring *tx_ring = &fdma->tx_ring; 82 83 return tx_ring->next_to_clean == tx_ring->next_to_use; 84 } 85 86 static void ocelot_fdma_activate_chan(struct ocelot *ocelot, dma_addr_t dma, 87 int chan) 88 { 89 ocelot_fdma_writel(ocelot, MSCC_FDMA_DCB_LLP(chan), dma); 90 /* Barrier to force memory writes to DCB to be completed before starting 91 * the channel. 92 */ 93 wmb(); 94 ocelot_fdma_writel(ocelot, MSCC_FDMA_CH_ACTIVATE, BIT(chan)); 95 } 96 97 static u32 ocelot_fdma_read_ch_safe(struct ocelot *ocelot) 98 { 99 return ocelot_fdma_readl(ocelot, MSCC_FDMA_CH_SAFE); 100 } 101 102 static int ocelot_fdma_wait_chan_safe(struct ocelot *ocelot, int chan) 103 { 104 u32 safe; 105 106 return readx_poll_timeout_atomic(ocelot_fdma_read_ch_safe, ocelot, safe, 107 safe & BIT(chan), 0, 108 OCELOT_FDMA_CH_SAFE_TIMEOUT_US); 109 } 110 111 static void ocelot_fdma_dcb_set_data(struct ocelot_fdma_dcb *dcb, 112 dma_addr_t dma_addr, 113 size_t size) 114 { 115 u32 offset = dma_addr & 0x3; 116 117 dcb->llp = 0; 118 dcb->datap = ALIGN_DOWN(dma_addr, 4); 119 dcb->datal = ALIGN_DOWN(size, 4); 120 dcb->stat = MSCC_FDMA_DCB_STAT_BLOCKO(offset); 121 } 122 123 static bool ocelot_fdma_rx_alloc_page(struct ocelot *ocelot, 124 struct ocelot_fdma_rx_buf *rxb) 125 { 126 dma_addr_t mapping; 127 struct page *page; 128 129 page = dev_alloc_page(); 130 if (unlikely(!page)) 131 return false; 132 133 mapping = dma_map_page(ocelot->dev, page, 0, PAGE_SIZE, 134 DMA_FROM_DEVICE); 135 if (unlikely(dma_mapping_error(ocelot->dev, mapping))) { 136 __free_page(page); 137 return false; 138 } 139 140 rxb->page = page; 141 rxb->page_offset = 0; 142 rxb->dma_addr = mapping; 143 144 return true; 145 } 146 147 static int ocelot_fdma_alloc_rx_buffs(struct ocelot *ocelot, u16 alloc_cnt) 148 { 149 struct ocelot_fdma *fdma = ocelot->fdma; 150 struct ocelot_fdma_rx_ring *rx_ring; 151 struct ocelot_fdma_rx_buf *rxb; 152 struct ocelot_fdma_dcb *dcb; 153 dma_addr_t dma_addr; 154 int ret = 0; 155 u16 idx; 156 157 rx_ring = &fdma->rx_ring; 158 idx = rx_ring->next_to_use; 159 160 while (alloc_cnt--) { 161 rxb = &rx_ring->bufs[idx]; 162 /* try reuse page */ 163 if (unlikely(!rxb->page)) { 164 if (unlikely(!ocelot_fdma_rx_alloc_page(ocelot, rxb))) { 165 dev_err_ratelimited(ocelot->dev, 166 "Failed to allocate rx\n"); 167 ret = -ENOMEM; 168 break; 169 } 170 } 171 172 dcb = &rx_ring->dcbs[idx]; 173 dma_addr = rxb->dma_addr + rxb->page_offset; 174 ocelot_fdma_dcb_set_data(dcb, dma_addr, OCELOT_FDMA_RXB_SIZE); 175 176 idx = ocelot_fdma_idx_next(idx, OCELOT_FDMA_RX_RING_SIZE); 177 /* Chain the DCB to the next one */ 178 dcb->llp = ocelot_fdma_idx_dma(rx_ring->dcbs_dma, idx); 179 } 180 181 rx_ring->next_to_use = idx; 182 rx_ring->next_to_alloc = idx; 183 184 return ret; 185 } 186 187 static bool ocelot_fdma_tx_dcb_set_skb(struct ocelot *ocelot, 188 struct ocelot_fdma_tx_buf *tx_buf, 189 struct ocelot_fdma_dcb *dcb, 190 struct sk_buff *skb) 191 { 192 dma_addr_t mapping; 193 194 mapping = dma_map_single(ocelot->dev, skb->data, skb->len, 195 DMA_TO_DEVICE); 196 if (unlikely(dma_mapping_error(ocelot->dev, mapping))) 197 return false; 198 199 dma_unmap_addr_set(tx_buf, dma_addr, mapping); 200 201 ocelot_fdma_dcb_set_data(dcb, mapping, OCELOT_FDMA_RX_SIZE); 202 tx_buf->skb = skb; 203 dcb->stat |= MSCC_FDMA_DCB_STAT_BLOCKL(skb->len); 204 dcb->stat |= MSCC_FDMA_DCB_STAT_SOF | MSCC_FDMA_DCB_STAT_EOF; 205 206 return true; 207 } 208 209 static bool ocelot_fdma_check_stop_rx(struct ocelot *ocelot) 210 { 211 u32 llp; 212 213 /* Check if the FDMA hits the DCB with LLP == NULL */ 214 llp = ocelot_fdma_readl(ocelot, MSCC_FDMA_DCB_LLP(MSCC_FDMA_XTR_CHAN)); 215 if (unlikely(llp)) 216 return false; 217 218 ocelot_fdma_writel(ocelot, MSCC_FDMA_CH_DISABLE, 219 BIT(MSCC_FDMA_XTR_CHAN)); 220 221 return true; 222 } 223 224 static void ocelot_fdma_rx_set_llp(struct ocelot_fdma_rx_ring *rx_ring) 225 { 226 struct ocelot_fdma_dcb *dcb; 227 unsigned int idx; 228 229 idx = ocelot_fdma_idx_prev(rx_ring->next_to_use, 230 OCELOT_FDMA_RX_RING_SIZE); 231 dcb = &rx_ring->dcbs[idx]; 232 dcb->llp = 0; 233 } 234 235 static void ocelot_fdma_rx_restart(struct ocelot *ocelot) 236 { 237 struct ocelot_fdma *fdma = ocelot->fdma; 238 struct ocelot_fdma_rx_ring *rx_ring; 239 const u8 chan = MSCC_FDMA_XTR_CHAN; 240 dma_addr_t new_llp, dma_base; 241 unsigned int idx; 242 u32 llp_prev; 243 int ret; 244 245 rx_ring = &fdma->rx_ring; 246 ret = ocelot_fdma_wait_chan_safe(ocelot, chan); 247 if (ret) { 248 dev_err_ratelimited(ocelot->dev, 249 "Unable to stop RX channel\n"); 250 return; 251 } 252 253 ocelot_fdma_rx_set_llp(rx_ring); 254 255 /* FDMA stopped on the last DCB that contained a NULL LLP, since 256 * we processed some DCBs in RX, there is free space, and we must set 257 * DCB_LLP to point to the next DCB 258 */ 259 llp_prev = ocelot_fdma_readl(ocelot, MSCC_FDMA_DCB_LLP_PREV(chan)); 260 dma_base = rx_ring->dcbs_dma; 261 262 /* Get the next DMA addr located after LLP == NULL DCB */ 263 idx = ocelot_fdma_dma_idx(dma_base, llp_prev); 264 idx = ocelot_fdma_idx_next(idx, OCELOT_FDMA_RX_RING_SIZE); 265 new_llp = ocelot_fdma_idx_dma(dma_base, idx); 266 267 /* Finally reactivate the channel */ 268 ocelot_fdma_activate_chan(ocelot, new_llp, chan); 269 } 270 271 static bool ocelot_fdma_add_rx_frag(struct ocelot_fdma_rx_buf *rxb, u32 stat, 272 struct sk_buff *skb, bool first) 273 { 274 int size = MSCC_FDMA_DCB_STAT_BLOCKL(stat); 275 struct page *page = rxb->page; 276 277 if (likely(first)) { 278 skb_put(skb, size); 279 } else { 280 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, 281 rxb->page_offset, size, OCELOT_FDMA_RX_SIZE); 282 } 283 284 /* Try to reuse page */ 285 if (unlikely(page_ref_count(page) != 1 || page_is_pfmemalloc(page))) 286 return false; 287 288 /* Change offset to the other half */ 289 rxb->page_offset ^= OCELOT_FDMA_RX_SIZE; 290 291 page_ref_inc(page); 292 293 return true; 294 } 295 296 static void ocelot_fdma_reuse_rx_page(struct ocelot *ocelot, 297 struct ocelot_fdma_rx_buf *old_rxb) 298 { 299 struct ocelot_fdma_rx_ring *rx_ring = &ocelot->fdma->rx_ring; 300 struct ocelot_fdma_rx_buf *new_rxb; 301 302 new_rxb = &rx_ring->bufs[rx_ring->next_to_alloc]; 303 rx_ring->next_to_alloc = ocelot_fdma_idx_next(rx_ring->next_to_alloc, 304 OCELOT_FDMA_RX_RING_SIZE); 305 306 /* Copy page reference */ 307 *new_rxb = *old_rxb; 308 309 /* Sync for use by the device */ 310 dma_sync_single_range_for_device(ocelot->dev, old_rxb->dma_addr, 311 old_rxb->page_offset, 312 OCELOT_FDMA_RX_SIZE, DMA_FROM_DEVICE); 313 } 314 315 static struct sk_buff *ocelot_fdma_get_skb(struct ocelot *ocelot, u32 stat, 316 struct ocelot_fdma_rx_buf *rxb, 317 struct sk_buff *skb) 318 { 319 bool first = false; 320 321 /* Allocate skb head and data */ 322 if (likely(!skb)) { 323 void *buff_addr = page_address(rxb->page) + 324 rxb->page_offset; 325 326 skb = build_skb(buff_addr, OCELOT_FDMA_SKBFRAG_SIZE); 327 if (unlikely(!skb)) { 328 dev_err_ratelimited(ocelot->dev, 329 "build_skb failed !\n"); 330 return NULL; 331 } 332 first = true; 333 } 334 335 dma_sync_single_range_for_cpu(ocelot->dev, rxb->dma_addr, 336 rxb->page_offset, OCELOT_FDMA_RX_SIZE, 337 DMA_FROM_DEVICE); 338 339 if (ocelot_fdma_add_rx_frag(rxb, stat, skb, first)) { 340 /* Reuse the free half of the page for the next_to_alloc DCB*/ 341 ocelot_fdma_reuse_rx_page(ocelot, rxb); 342 } else { 343 /* page cannot be reused, unmap it */ 344 dma_unmap_page(ocelot->dev, rxb->dma_addr, PAGE_SIZE, 345 DMA_FROM_DEVICE); 346 } 347 348 /* clear rx buff content */ 349 rxb->page = NULL; 350 351 return skb; 352 } 353 354 static bool ocelot_fdma_receive_skb(struct ocelot *ocelot, struct sk_buff *skb) 355 { 356 struct net_device *ndev; 357 void *xfh = skb->data; 358 u64 timestamp; 359 u64 src_port; 360 361 skb_pull(skb, OCELOT_TAG_LEN); 362 363 ocelot_xfh_get_src_port(xfh, &src_port); 364 if (unlikely(src_port >= ocelot->num_phys_ports)) 365 return false; 366 367 ndev = ocelot_port_to_netdev(ocelot, src_port); 368 if (unlikely(!ndev)) 369 return false; 370 371 if (pskb_trim(skb, skb->len - ETH_FCS_LEN)) 372 return false; 373 374 skb->dev = ndev; 375 skb->protocol = eth_type_trans(skb, skb->dev); 376 skb->dev->stats.rx_bytes += skb->len; 377 skb->dev->stats.rx_packets++; 378 379 if (ocelot->ptp) { 380 ocelot_xfh_get_rew_val(xfh, ×tamp); 381 ocelot_ptp_rx_timestamp(ocelot, skb, timestamp); 382 } 383 384 if (likely(!skb_defer_rx_timestamp(skb))) 385 netif_receive_skb(skb); 386 387 return true; 388 } 389 390 static int ocelot_fdma_rx_get(struct ocelot *ocelot, int budget) 391 { 392 struct ocelot_fdma *fdma = ocelot->fdma; 393 struct ocelot_fdma_rx_ring *rx_ring; 394 struct ocelot_fdma_rx_buf *rxb; 395 struct ocelot_fdma_dcb *dcb; 396 struct sk_buff *skb; 397 int work_done = 0; 398 int cleaned_cnt; 399 u32 stat; 400 u16 idx; 401 402 cleaned_cnt = ocelot_fdma_rx_ring_free(fdma); 403 rx_ring = &fdma->rx_ring; 404 skb = rx_ring->skb; 405 406 while (budget--) { 407 idx = rx_ring->next_to_clean; 408 dcb = &rx_ring->dcbs[idx]; 409 stat = dcb->stat; 410 if (MSCC_FDMA_DCB_STAT_BLOCKL(stat) == 0) 411 break; 412 413 /* New packet is a start of frame but we already got a skb set, 414 * we probably lost an EOF packet, free skb 415 */ 416 if (unlikely(skb && (stat & MSCC_FDMA_DCB_STAT_SOF))) { 417 dev_kfree_skb(skb); 418 skb = NULL; 419 } 420 421 rxb = &rx_ring->bufs[idx]; 422 /* Fetch next to clean buffer from the rx_ring */ 423 skb = ocelot_fdma_get_skb(ocelot, stat, rxb, skb); 424 if (unlikely(!skb)) 425 break; 426 427 work_done++; 428 cleaned_cnt++; 429 430 idx = ocelot_fdma_idx_next(idx, OCELOT_FDMA_RX_RING_SIZE); 431 rx_ring->next_to_clean = idx; 432 433 if (unlikely(stat & MSCC_FDMA_DCB_STAT_ABORT || 434 stat & MSCC_FDMA_DCB_STAT_PD)) { 435 dev_err_ratelimited(ocelot->dev, 436 "DCB aborted or pruned\n"); 437 dev_kfree_skb(skb); 438 skb = NULL; 439 continue; 440 } 441 442 /* We still need to process the other fragment of the packet 443 * before delivering it to the network stack 444 */ 445 if (!(stat & MSCC_FDMA_DCB_STAT_EOF)) 446 continue; 447 448 if (unlikely(!ocelot_fdma_receive_skb(ocelot, skb))) 449 dev_kfree_skb(skb); 450 451 skb = NULL; 452 } 453 454 rx_ring->skb = skb; 455 456 if (cleaned_cnt) 457 ocelot_fdma_alloc_rx_buffs(ocelot, cleaned_cnt); 458 459 return work_done; 460 } 461 462 static void ocelot_fdma_wakeup_netdev(struct ocelot *ocelot) 463 { 464 struct ocelot_port_private *priv; 465 struct ocelot_port *ocelot_port; 466 struct net_device *dev; 467 int port; 468 469 for (port = 0; port < ocelot->num_phys_ports; port++) { 470 ocelot_port = ocelot->ports[port]; 471 if (!ocelot_port) 472 continue; 473 priv = container_of(ocelot_port, struct ocelot_port_private, 474 port); 475 dev = priv->dev; 476 477 if (unlikely(netif_queue_stopped(dev))) 478 netif_wake_queue(dev); 479 } 480 } 481 482 static void ocelot_fdma_tx_cleanup(struct ocelot *ocelot, int budget) 483 { 484 struct ocelot_fdma *fdma = ocelot->fdma; 485 struct ocelot_fdma_tx_ring *tx_ring; 486 struct ocelot_fdma_tx_buf *buf; 487 unsigned int new_null_llp_idx; 488 struct ocelot_fdma_dcb *dcb; 489 bool end_of_list = false; 490 struct sk_buff *skb; 491 dma_addr_t dma; 492 u32 dcb_llp; 493 u16 ntc; 494 int ret; 495 496 tx_ring = &fdma->tx_ring; 497 498 /* Purge the TX packets that have been sent up to the NULL llp or the 499 * end of done list. 500 */ 501 while (!ocelot_fdma_tx_ring_empty(fdma)) { 502 ntc = tx_ring->next_to_clean; 503 dcb = &tx_ring->dcbs[ntc]; 504 if (!(dcb->stat & MSCC_FDMA_DCB_STAT_PD)) 505 break; 506 507 buf = &tx_ring->bufs[ntc]; 508 skb = buf->skb; 509 dma_unmap_single(ocelot->dev, dma_unmap_addr(buf, dma_addr), 510 skb->len, DMA_TO_DEVICE); 511 napi_consume_skb(skb, budget); 512 dcb_llp = dcb->llp; 513 514 /* Only update after accessing all dcb fields */ 515 tx_ring->next_to_clean = ocelot_fdma_idx_next(ntc, 516 OCELOT_FDMA_TX_RING_SIZE); 517 518 /* If we hit the NULL LLP, stop, we might need to reload FDMA */ 519 if (dcb_llp == 0) { 520 end_of_list = true; 521 break; 522 } 523 } 524 525 /* No need to try to wake if there were no TX cleaned_cnt up. */ 526 if (ocelot_fdma_tx_ring_free(fdma)) 527 ocelot_fdma_wakeup_netdev(ocelot); 528 529 /* If there is still some DCBs to be processed by the FDMA or if the 530 * pending list is empty, there is no need to restart the FDMA. 531 */ 532 if (!end_of_list || ocelot_fdma_tx_ring_empty(fdma)) 533 return; 534 535 ret = ocelot_fdma_wait_chan_safe(ocelot, MSCC_FDMA_INJ_CHAN); 536 if (ret) { 537 dev_warn(ocelot->dev, 538 "Failed to wait for TX channel to stop\n"); 539 return; 540 } 541 542 /* Set NULL LLP to be the last DCB used */ 543 new_null_llp_idx = ocelot_fdma_idx_prev(tx_ring->next_to_use, 544 OCELOT_FDMA_TX_RING_SIZE); 545 dcb = &tx_ring->dcbs[new_null_llp_idx]; 546 dcb->llp = 0; 547 548 dma = ocelot_fdma_idx_dma(tx_ring->dcbs_dma, tx_ring->next_to_clean); 549 ocelot_fdma_activate_chan(ocelot, dma, MSCC_FDMA_INJ_CHAN); 550 } 551 552 static int ocelot_fdma_napi_poll(struct napi_struct *napi, int budget) 553 { 554 struct ocelot_fdma *fdma = container_of(napi, struct ocelot_fdma, napi); 555 struct ocelot *ocelot = fdma->ocelot; 556 int work_done = 0; 557 bool rx_stopped; 558 559 ocelot_fdma_tx_cleanup(ocelot, budget); 560 561 rx_stopped = ocelot_fdma_check_stop_rx(ocelot); 562 563 work_done = ocelot_fdma_rx_get(ocelot, budget); 564 565 if (rx_stopped) 566 ocelot_fdma_rx_restart(ocelot); 567 568 if (work_done < budget) { 569 napi_complete_done(&fdma->napi, work_done); 570 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_ENA, 571 BIT(MSCC_FDMA_INJ_CHAN) | 572 BIT(MSCC_FDMA_XTR_CHAN)); 573 } 574 575 return work_done; 576 } 577 578 static irqreturn_t ocelot_fdma_interrupt(int irq, void *dev_id) 579 { 580 u32 ident, llp, frm, err, err_code; 581 struct ocelot *ocelot = dev_id; 582 583 ident = ocelot_fdma_readl(ocelot, MSCC_FDMA_INTR_IDENT); 584 frm = ocelot_fdma_readl(ocelot, MSCC_FDMA_INTR_FRM); 585 llp = ocelot_fdma_readl(ocelot, MSCC_FDMA_INTR_LLP); 586 587 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_LLP, llp & ident); 588 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_FRM, frm & ident); 589 if (frm || llp) { 590 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_ENA, 0); 591 napi_schedule(&ocelot->fdma->napi); 592 } 593 594 err = ocelot_fdma_readl(ocelot, MSCC_FDMA_EVT_ERR); 595 if (unlikely(err)) { 596 err_code = ocelot_fdma_readl(ocelot, MSCC_FDMA_EVT_ERR_CODE); 597 dev_err_ratelimited(ocelot->dev, 598 "Error ! chans mask: %#x, code: %#x\n", 599 err, err_code); 600 601 ocelot_fdma_writel(ocelot, MSCC_FDMA_EVT_ERR, err); 602 ocelot_fdma_writel(ocelot, MSCC_FDMA_EVT_ERR_CODE, err_code); 603 } 604 605 return IRQ_HANDLED; 606 } 607 608 static void ocelot_fdma_send_skb(struct ocelot *ocelot, 609 struct ocelot_fdma *fdma, struct sk_buff *skb) 610 { 611 struct ocelot_fdma_tx_ring *tx_ring = &fdma->tx_ring; 612 struct ocelot_fdma_tx_buf *tx_buf; 613 struct ocelot_fdma_dcb *dcb; 614 dma_addr_t dma; 615 u16 next_idx; 616 617 dcb = &tx_ring->dcbs[tx_ring->next_to_use]; 618 tx_buf = &tx_ring->bufs[tx_ring->next_to_use]; 619 if (!ocelot_fdma_tx_dcb_set_skb(ocelot, tx_buf, dcb, skb)) { 620 dev_kfree_skb_any(skb); 621 return; 622 } 623 624 next_idx = ocelot_fdma_idx_next(tx_ring->next_to_use, 625 OCELOT_FDMA_TX_RING_SIZE); 626 skb_tx_timestamp(skb); 627 628 /* If the FDMA TX chan is empty, then enqueue the DCB directly */ 629 if (ocelot_fdma_tx_ring_empty(fdma)) { 630 dma = ocelot_fdma_idx_dma(tx_ring->dcbs_dma, 631 tx_ring->next_to_use); 632 ocelot_fdma_activate_chan(ocelot, dma, MSCC_FDMA_INJ_CHAN); 633 } else { 634 /* Chain the DCBs */ 635 dcb->llp = ocelot_fdma_idx_dma(tx_ring->dcbs_dma, next_idx); 636 } 637 638 tx_ring->next_to_use = next_idx; 639 } 640 641 static int ocelot_fdma_prepare_skb(struct ocelot *ocelot, int port, u32 rew_op, 642 struct sk_buff *skb, struct net_device *dev) 643 { 644 int needed_headroom = max_t(int, OCELOT_TAG_LEN - skb_headroom(skb), 0); 645 int needed_tailroom = max_t(int, ETH_FCS_LEN - skb_tailroom(skb), 0); 646 void *ifh; 647 int err; 648 649 if (unlikely(needed_headroom || needed_tailroom || 650 skb_header_cloned(skb))) { 651 err = pskb_expand_head(skb, needed_headroom, needed_tailroom, 652 GFP_ATOMIC); 653 if (unlikely(err)) { 654 dev_kfree_skb_any(skb); 655 return 1; 656 } 657 } 658 659 err = skb_linearize(skb); 660 if (err) { 661 net_err_ratelimited("%s: skb_linearize error (%d)!\n", 662 dev->name, err); 663 dev_kfree_skb_any(skb); 664 return 1; 665 } 666 667 ifh = skb_push(skb, OCELOT_TAG_LEN); 668 skb_put(skb, ETH_FCS_LEN); 669 memset(ifh, 0, OCELOT_TAG_LEN); 670 ocelot_ifh_port_set(ifh, port, rew_op, skb_vlan_tag_get(skb)); 671 672 return 0; 673 } 674 675 int ocelot_fdma_inject_frame(struct ocelot *ocelot, int port, u32 rew_op, 676 struct sk_buff *skb, struct net_device *dev) 677 { 678 struct ocelot_fdma *fdma = ocelot->fdma; 679 int ret = NETDEV_TX_OK; 680 681 spin_lock(&fdma->tx_ring.xmit_lock); 682 683 if (ocelot_fdma_tx_ring_free(fdma) == 0) { 684 netif_stop_queue(dev); 685 ret = NETDEV_TX_BUSY; 686 goto out; 687 } 688 689 if (ocelot_fdma_prepare_skb(ocelot, port, rew_op, skb, dev)) 690 goto out; 691 692 ocelot_fdma_send_skb(ocelot, fdma, skb); 693 694 out: 695 spin_unlock(&fdma->tx_ring.xmit_lock); 696 697 return ret; 698 } 699 700 static void ocelot_fdma_free_rx_ring(struct ocelot *ocelot) 701 { 702 struct ocelot_fdma *fdma = ocelot->fdma; 703 struct ocelot_fdma_rx_ring *rx_ring; 704 struct ocelot_fdma_rx_buf *rxb; 705 u16 idx; 706 707 rx_ring = &fdma->rx_ring; 708 idx = rx_ring->next_to_clean; 709 710 /* Free the pages held in the RX ring */ 711 while (idx != rx_ring->next_to_use) { 712 rxb = &rx_ring->bufs[idx]; 713 dma_unmap_page(ocelot->dev, rxb->dma_addr, PAGE_SIZE, 714 DMA_FROM_DEVICE); 715 __free_page(rxb->page); 716 idx = ocelot_fdma_idx_next(idx, OCELOT_FDMA_RX_RING_SIZE); 717 } 718 719 if (fdma->rx_ring.skb) 720 dev_kfree_skb_any(fdma->rx_ring.skb); 721 } 722 723 static void ocelot_fdma_free_tx_ring(struct ocelot *ocelot) 724 { 725 struct ocelot_fdma *fdma = ocelot->fdma; 726 struct ocelot_fdma_tx_ring *tx_ring; 727 struct ocelot_fdma_tx_buf *txb; 728 struct sk_buff *skb; 729 u16 idx; 730 731 tx_ring = &fdma->tx_ring; 732 idx = tx_ring->next_to_clean; 733 734 while (idx != tx_ring->next_to_use) { 735 txb = &tx_ring->bufs[idx]; 736 skb = txb->skb; 737 dma_unmap_single(ocelot->dev, dma_unmap_addr(txb, dma_addr), 738 skb->len, DMA_TO_DEVICE); 739 dev_kfree_skb_any(skb); 740 idx = ocelot_fdma_idx_next(idx, OCELOT_FDMA_TX_RING_SIZE); 741 } 742 } 743 744 static int ocelot_fdma_rings_alloc(struct ocelot *ocelot) 745 { 746 struct ocelot_fdma *fdma = ocelot->fdma; 747 struct ocelot_fdma_dcb *dcbs; 748 unsigned int adjust; 749 dma_addr_t dcbs_dma; 750 int ret; 751 752 /* Create a pool of consistent memory blocks for hardware descriptors */ 753 fdma->dcbs_base = dmam_alloc_coherent(ocelot->dev, 754 OCELOT_DCBS_HW_ALLOC_SIZE, 755 &fdma->dcbs_dma_base, GFP_KERNEL); 756 if (!fdma->dcbs_base) 757 return -ENOMEM; 758 759 /* DCBs must be aligned on a 32bit boundary */ 760 dcbs = fdma->dcbs_base; 761 dcbs_dma = fdma->dcbs_dma_base; 762 if (!IS_ALIGNED(dcbs_dma, 4)) { 763 adjust = dcbs_dma & 0x3; 764 dcbs_dma = ALIGN(dcbs_dma, 4); 765 dcbs = (void *)dcbs + adjust; 766 } 767 768 /* TX queue */ 769 fdma->tx_ring.dcbs = dcbs; 770 fdma->tx_ring.dcbs_dma = dcbs_dma; 771 spin_lock_init(&fdma->tx_ring.xmit_lock); 772 773 /* RX queue */ 774 fdma->rx_ring.dcbs = dcbs + OCELOT_FDMA_TX_RING_SIZE; 775 fdma->rx_ring.dcbs_dma = dcbs_dma + OCELOT_FDMA_TX_DCB_SIZE; 776 ret = ocelot_fdma_alloc_rx_buffs(ocelot, 777 ocelot_fdma_tx_ring_free(fdma)); 778 if (ret) { 779 ocelot_fdma_free_rx_ring(ocelot); 780 return ret; 781 } 782 783 /* Set the last DCB LLP as NULL, this is normally done when restarting 784 * the RX chan, but this is for the first run 785 */ 786 ocelot_fdma_rx_set_llp(&fdma->rx_ring); 787 788 return 0; 789 } 790 791 void ocelot_fdma_netdev_init(struct ocelot *ocelot, struct net_device *dev) 792 { 793 struct ocelot_fdma *fdma = ocelot->fdma; 794 795 dev->needed_headroom = OCELOT_TAG_LEN; 796 dev->needed_tailroom = ETH_FCS_LEN; 797 798 if (fdma->ndev) 799 return; 800 801 fdma->ndev = dev; 802 netif_napi_add_weight(dev, &fdma->napi, ocelot_fdma_napi_poll, 803 OCELOT_FDMA_WEIGHT); 804 } 805 806 void ocelot_fdma_netdev_deinit(struct ocelot *ocelot, struct net_device *dev) 807 { 808 struct ocelot_fdma *fdma = ocelot->fdma; 809 810 if (fdma->ndev == dev) { 811 netif_napi_del(&fdma->napi); 812 fdma->ndev = NULL; 813 } 814 } 815 816 void ocelot_fdma_init(struct platform_device *pdev, struct ocelot *ocelot) 817 { 818 struct device *dev = ocelot->dev; 819 struct ocelot_fdma *fdma; 820 int ret; 821 822 fdma = devm_kzalloc(dev, sizeof(*fdma), GFP_KERNEL); 823 if (!fdma) 824 return; 825 826 ocelot->fdma = fdma; 827 ocelot->dev->coherent_dma_mask = DMA_BIT_MASK(32); 828 829 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_ENA, 0); 830 831 fdma->ocelot = ocelot; 832 fdma->irq = platform_get_irq_byname(pdev, "fdma"); 833 ret = devm_request_irq(dev, fdma->irq, ocelot_fdma_interrupt, 0, 834 dev_name(dev), ocelot); 835 if (ret) 836 goto err_free_fdma; 837 838 ret = ocelot_fdma_rings_alloc(ocelot); 839 if (ret) 840 goto err_free_irq; 841 842 static_branch_enable(&ocelot_fdma_enabled); 843 844 return; 845 846 err_free_irq: 847 devm_free_irq(dev, fdma->irq, fdma); 848 err_free_fdma: 849 devm_kfree(dev, fdma); 850 851 ocelot->fdma = NULL; 852 } 853 854 void ocelot_fdma_start(struct ocelot *ocelot) 855 { 856 struct ocelot_fdma *fdma = ocelot->fdma; 857 858 /* Reconfigure for extraction and injection using DMA */ 859 ocelot_write_rix(ocelot, QS_INJ_GRP_CFG_MODE(2), QS_INJ_GRP_CFG, 0); 860 ocelot_write_rix(ocelot, QS_INJ_CTRL_GAP_SIZE(0), QS_INJ_CTRL, 0); 861 862 ocelot_write_rix(ocelot, QS_XTR_GRP_CFG_MODE(2), QS_XTR_GRP_CFG, 0); 863 864 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_LLP, 0xffffffff); 865 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_FRM, 0xffffffff); 866 867 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_LLP_ENA, 868 BIT(MSCC_FDMA_INJ_CHAN) | BIT(MSCC_FDMA_XTR_CHAN)); 869 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_FRM_ENA, 870 BIT(MSCC_FDMA_XTR_CHAN)); 871 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_ENA, 872 BIT(MSCC_FDMA_INJ_CHAN) | BIT(MSCC_FDMA_XTR_CHAN)); 873 874 napi_enable(&fdma->napi); 875 876 ocelot_fdma_activate_chan(ocelot, ocelot->fdma->rx_ring.dcbs_dma, 877 MSCC_FDMA_XTR_CHAN); 878 } 879 880 void ocelot_fdma_deinit(struct ocelot *ocelot) 881 { 882 struct ocelot_fdma *fdma = ocelot->fdma; 883 884 ocelot_fdma_writel(ocelot, MSCC_FDMA_INTR_ENA, 0); 885 ocelot_fdma_writel(ocelot, MSCC_FDMA_CH_FORCEDIS, 886 BIT(MSCC_FDMA_XTR_CHAN)); 887 ocelot_fdma_writel(ocelot, MSCC_FDMA_CH_FORCEDIS, 888 BIT(MSCC_FDMA_INJ_CHAN)); 889 napi_synchronize(&fdma->napi); 890 napi_disable(&fdma->napi); 891 892 ocelot_fdma_free_rx_ring(ocelot); 893 ocelot_fdma_free_tx_ring(ocelot); 894 } 895