1 // SPDX-License-Identifier: GPL-2.0 2 // 3 // Driver for AT91 USART Controllers as SPI 4 // 5 // Copyright (C) 2018 Microchip Technology Inc. 6 // 7 // Author: Radu Pirea <radu.pirea@microchip.com> 8 9 #include <linux/clk.h> 10 #include <linux/delay.h> 11 #include <linux/dmaengine.h> 12 #include <linux/dma-direction.h> 13 #include <linux/interrupt.h> 14 #include <linux/kernel.h> 15 #include <linux/module.h> 16 #include <linux/of_platform.h> 17 #include <linux/of_gpio.h> 18 #include <linux/pinctrl/consumer.h> 19 #include <linux/platform_device.h> 20 #include <linux/pm_runtime.h> 21 22 #include <linux/spi/spi.h> 23 24 #define US_CR 0x00 25 #define US_MR 0x04 26 #define US_IER 0x08 27 #define US_IDR 0x0C 28 #define US_CSR 0x14 29 #define US_RHR 0x18 30 #define US_THR 0x1C 31 #define US_BRGR 0x20 32 #define US_VERSION 0xFC 33 34 #define US_CR_RSTRX BIT(2) 35 #define US_CR_RSTTX BIT(3) 36 #define US_CR_RXEN BIT(4) 37 #define US_CR_RXDIS BIT(5) 38 #define US_CR_TXEN BIT(6) 39 #define US_CR_TXDIS BIT(7) 40 41 #define US_MR_SPI_MASTER 0x0E 42 #define US_MR_CHRL GENMASK(7, 6) 43 #define US_MR_CPHA BIT(8) 44 #define US_MR_CPOL BIT(16) 45 #define US_MR_CLKO BIT(18) 46 #define US_MR_WRDBT BIT(20) 47 #define US_MR_LOOP BIT(15) 48 49 #define US_IR_RXRDY BIT(0) 50 #define US_IR_TXRDY BIT(1) 51 #define US_IR_OVRE BIT(5) 52 53 #define US_BRGR_SIZE BIT(16) 54 55 #define US_MIN_CLK_DIV 0x06 56 #define US_MAX_CLK_DIV BIT(16) 57 58 #define US_RESET (US_CR_RSTRX | US_CR_RSTTX) 59 #define US_DISABLE (US_CR_RXDIS | US_CR_TXDIS) 60 #define US_ENABLE (US_CR_RXEN | US_CR_TXEN) 61 #define US_OVRE_RXRDY_IRQS (US_IR_OVRE | US_IR_RXRDY) 62 63 #define US_INIT \ 64 (US_MR_SPI_MASTER | US_MR_CHRL | US_MR_CLKO | US_MR_WRDBT) 65 #define US_DMA_MIN_BYTES 16 66 #define US_DMA_TIMEOUT (msecs_to_jiffies(1000)) 67 68 /* Register access macros */ 69 #define at91_usart_spi_readl(port, reg) \ 70 readl_relaxed((port)->regs + US_##reg) 71 #define at91_usart_spi_writel(port, reg, value) \ 72 writel_relaxed((value), (port)->regs + US_##reg) 73 74 #define at91_usart_spi_readb(port, reg) \ 75 readb_relaxed((port)->regs + US_##reg) 76 #define at91_usart_spi_writeb(port, reg, value) \ 77 writeb_relaxed((value), (port)->regs + US_##reg) 78 79 struct at91_usart_spi { 80 struct platform_device *mpdev; 81 struct spi_transfer *current_transfer; 82 void __iomem *regs; 83 struct device *dev; 84 struct clk *clk; 85 86 struct completion xfer_completion; 87 88 /*used in interrupt to protect data reading*/ 89 spinlock_t lock; 90 91 phys_addr_t phybase; 92 93 int irq; 94 unsigned int current_tx_remaining_bytes; 95 unsigned int current_rx_remaining_bytes; 96 97 u32 spi_clk; 98 u32 status; 99 100 bool xfer_failed; 101 bool use_dma; 102 }; 103 104 static void dma_callback(void *data) 105 { 106 struct spi_controller *ctlr = data; 107 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); 108 109 at91_usart_spi_writel(aus, IER, US_IR_RXRDY); 110 aus->current_rx_remaining_bytes = 0; 111 complete(&aus->xfer_completion); 112 } 113 114 static bool at91_usart_spi_can_dma(struct spi_controller *ctrl, 115 struct spi_device *spi, 116 struct spi_transfer *xfer) 117 { 118 struct at91_usart_spi *aus = spi_master_get_devdata(ctrl); 119 120 return aus->use_dma && xfer->len >= US_DMA_MIN_BYTES; 121 } 122 123 static int at91_usart_spi_configure_dma(struct spi_controller *ctlr, 124 struct at91_usart_spi *aus) 125 { 126 struct dma_slave_config slave_config; 127 struct device *dev = &aus->mpdev->dev; 128 phys_addr_t phybase = aus->phybase; 129 dma_cap_mask_t mask; 130 int err = 0; 131 132 dma_cap_zero(mask); 133 dma_cap_set(DMA_SLAVE, mask); 134 135 ctlr->dma_tx = dma_request_slave_channel_reason(dev, "tx"); 136 if (IS_ERR_OR_NULL(ctlr->dma_tx)) { 137 if (IS_ERR(ctlr->dma_tx)) { 138 err = PTR_ERR(ctlr->dma_tx); 139 goto at91_usart_spi_error_clear; 140 } 141 142 dev_dbg(dev, 143 "DMA TX channel not available, SPI unable to use DMA\n"); 144 err = -EBUSY; 145 goto at91_usart_spi_error_clear; 146 } 147 148 ctlr->dma_rx = dma_request_slave_channel_reason(dev, "rx"); 149 if (IS_ERR_OR_NULL(ctlr->dma_rx)) { 150 if (IS_ERR(ctlr->dma_rx)) { 151 err = PTR_ERR(ctlr->dma_rx); 152 goto at91_usart_spi_error; 153 } 154 155 dev_dbg(dev, 156 "DMA RX channel not available, SPI unable to use DMA\n"); 157 err = -EBUSY; 158 goto at91_usart_spi_error; 159 } 160 161 slave_config.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE; 162 slave_config.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE; 163 slave_config.dst_addr = (dma_addr_t)phybase + US_THR; 164 slave_config.src_addr = (dma_addr_t)phybase + US_RHR; 165 slave_config.src_maxburst = 1; 166 slave_config.dst_maxburst = 1; 167 slave_config.device_fc = false; 168 169 slave_config.direction = DMA_DEV_TO_MEM; 170 if (dmaengine_slave_config(ctlr->dma_rx, &slave_config)) { 171 dev_err(&ctlr->dev, 172 "failed to configure rx dma channel\n"); 173 err = -EINVAL; 174 goto at91_usart_spi_error; 175 } 176 177 slave_config.direction = DMA_MEM_TO_DEV; 178 if (dmaengine_slave_config(ctlr->dma_tx, &slave_config)) { 179 dev_err(&ctlr->dev, 180 "failed to configure tx dma channel\n"); 181 err = -EINVAL; 182 goto at91_usart_spi_error; 183 } 184 185 aus->use_dma = true; 186 return 0; 187 188 at91_usart_spi_error: 189 if (!IS_ERR_OR_NULL(ctlr->dma_tx)) 190 dma_release_channel(ctlr->dma_tx); 191 if (!IS_ERR_OR_NULL(ctlr->dma_rx)) 192 dma_release_channel(ctlr->dma_rx); 193 ctlr->dma_tx = NULL; 194 ctlr->dma_rx = NULL; 195 196 at91_usart_spi_error_clear: 197 return err; 198 } 199 200 static void at91_usart_spi_release_dma(struct spi_controller *ctlr) 201 { 202 if (ctlr->dma_rx) 203 dma_release_channel(ctlr->dma_rx); 204 if (ctlr->dma_tx) 205 dma_release_channel(ctlr->dma_tx); 206 } 207 208 static void at91_usart_spi_stop_dma(struct spi_controller *ctlr) 209 { 210 if (ctlr->dma_rx) 211 dmaengine_terminate_all(ctlr->dma_rx); 212 if (ctlr->dma_tx) 213 dmaengine_terminate_all(ctlr->dma_tx); 214 } 215 216 static int at91_usart_spi_dma_transfer(struct spi_controller *ctlr, 217 struct spi_transfer *xfer) 218 { 219 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); 220 struct dma_chan *rxchan = ctlr->dma_rx; 221 struct dma_chan *txchan = ctlr->dma_tx; 222 struct dma_async_tx_descriptor *rxdesc; 223 struct dma_async_tx_descriptor *txdesc; 224 dma_cookie_t cookie; 225 226 /* Disable RX interrupt */ 227 at91_usart_spi_writel(aus, IDR, US_IR_RXRDY); 228 229 rxdesc = dmaengine_prep_slave_sg(rxchan, 230 xfer->rx_sg.sgl, 231 xfer->rx_sg.nents, 232 DMA_DEV_TO_MEM, 233 DMA_PREP_INTERRUPT | 234 DMA_CTRL_ACK); 235 if (!rxdesc) 236 goto at91_usart_spi_err_dma; 237 238 txdesc = dmaengine_prep_slave_sg(txchan, 239 xfer->tx_sg.sgl, 240 xfer->tx_sg.nents, 241 DMA_MEM_TO_DEV, 242 DMA_PREP_INTERRUPT | 243 DMA_CTRL_ACK); 244 if (!txdesc) 245 goto at91_usart_spi_err_dma; 246 247 rxdesc->callback = dma_callback; 248 rxdesc->callback_param = ctlr; 249 250 cookie = rxdesc->tx_submit(rxdesc); 251 if (dma_submit_error(cookie)) 252 goto at91_usart_spi_err_dma; 253 254 cookie = txdesc->tx_submit(txdesc); 255 if (dma_submit_error(cookie)) 256 goto at91_usart_spi_err_dma; 257 258 rxchan->device->device_issue_pending(rxchan); 259 txchan->device->device_issue_pending(txchan); 260 261 return 0; 262 263 at91_usart_spi_err_dma: 264 /* Enable RX interrupt if something fails and fallback to PIO */ 265 at91_usart_spi_writel(aus, IER, US_IR_RXRDY); 266 at91_usart_spi_stop_dma(ctlr); 267 268 return -ENOMEM; 269 } 270 271 static unsigned long at91_usart_spi_dma_timeout(struct at91_usart_spi *aus) 272 { 273 return wait_for_completion_timeout(&aus->xfer_completion, 274 US_DMA_TIMEOUT); 275 } 276 277 static inline u32 at91_usart_spi_tx_ready(struct at91_usart_spi *aus) 278 { 279 return aus->status & US_IR_TXRDY; 280 } 281 282 static inline u32 at91_usart_spi_rx_ready(struct at91_usart_spi *aus) 283 { 284 return aus->status & US_IR_RXRDY; 285 } 286 287 static inline u32 at91_usart_spi_check_overrun(struct at91_usart_spi *aus) 288 { 289 return aus->status & US_IR_OVRE; 290 } 291 292 static inline u32 at91_usart_spi_read_status(struct at91_usart_spi *aus) 293 { 294 aus->status = at91_usart_spi_readl(aus, CSR); 295 return aus->status; 296 } 297 298 static inline void at91_usart_spi_tx(struct at91_usart_spi *aus) 299 { 300 unsigned int len = aus->current_transfer->len; 301 unsigned int remaining = aus->current_tx_remaining_bytes; 302 const u8 *tx_buf = aus->current_transfer->tx_buf; 303 304 if (!remaining) 305 return; 306 307 if (at91_usart_spi_tx_ready(aus)) { 308 at91_usart_spi_writeb(aus, THR, tx_buf[len - remaining]); 309 aus->current_tx_remaining_bytes--; 310 } 311 } 312 313 static inline void at91_usart_spi_rx(struct at91_usart_spi *aus) 314 { 315 int len = aus->current_transfer->len; 316 int remaining = aus->current_rx_remaining_bytes; 317 u8 *rx_buf = aus->current_transfer->rx_buf; 318 319 if (!remaining) 320 return; 321 322 rx_buf[len - remaining] = at91_usart_spi_readb(aus, RHR); 323 aus->current_rx_remaining_bytes--; 324 } 325 326 static inline void 327 at91_usart_spi_set_xfer_speed(struct at91_usart_spi *aus, 328 struct spi_transfer *xfer) 329 { 330 at91_usart_spi_writel(aus, BRGR, 331 DIV_ROUND_UP(aus->spi_clk, xfer->speed_hz)); 332 } 333 334 static irqreturn_t at91_usart_spi_interrupt(int irq, void *dev_id) 335 { 336 struct spi_controller *controller = dev_id; 337 struct at91_usart_spi *aus = spi_master_get_devdata(controller); 338 339 spin_lock(&aus->lock); 340 at91_usart_spi_read_status(aus); 341 342 if (at91_usart_spi_check_overrun(aus)) { 343 aus->xfer_failed = true; 344 at91_usart_spi_writel(aus, IDR, US_IR_OVRE | US_IR_RXRDY); 345 spin_unlock(&aus->lock); 346 return IRQ_HANDLED; 347 } 348 349 if (at91_usart_spi_rx_ready(aus)) { 350 at91_usart_spi_rx(aus); 351 spin_unlock(&aus->lock); 352 return IRQ_HANDLED; 353 } 354 355 spin_unlock(&aus->lock); 356 357 return IRQ_NONE; 358 } 359 360 static int at91_usart_spi_setup(struct spi_device *spi) 361 { 362 struct at91_usart_spi *aus = spi_master_get_devdata(spi->controller); 363 u32 *ausd = spi->controller_state; 364 unsigned int mr = at91_usart_spi_readl(aus, MR); 365 366 if (spi->mode & SPI_CPOL) 367 mr |= US_MR_CPOL; 368 else 369 mr &= ~US_MR_CPOL; 370 371 if (spi->mode & SPI_CPHA) 372 mr |= US_MR_CPHA; 373 else 374 mr &= ~US_MR_CPHA; 375 376 if (spi->mode & SPI_LOOP) 377 mr |= US_MR_LOOP; 378 else 379 mr &= ~US_MR_LOOP; 380 381 if (!ausd) { 382 ausd = kzalloc(sizeof(*ausd), GFP_KERNEL); 383 if (!ausd) 384 return -ENOMEM; 385 386 spi->controller_state = ausd; 387 } 388 389 *ausd = mr; 390 391 dev_dbg(&spi->dev, 392 "setup: bpw %u mode 0x%x -> mr %d %08x\n", 393 spi->bits_per_word, spi->mode, spi->chip_select, mr); 394 395 return 0; 396 } 397 398 static int at91_usart_spi_transfer_one(struct spi_controller *ctlr, 399 struct spi_device *spi, 400 struct spi_transfer *xfer) 401 { 402 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); 403 unsigned long dma_timeout = 0; 404 int ret = 0; 405 406 at91_usart_spi_set_xfer_speed(aus, xfer); 407 aus->xfer_failed = false; 408 aus->current_transfer = xfer; 409 aus->current_tx_remaining_bytes = xfer->len; 410 aus->current_rx_remaining_bytes = xfer->len; 411 412 while ((aus->current_tx_remaining_bytes || 413 aus->current_rx_remaining_bytes) && !aus->xfer_failed) { 414 reinit_completion(&aus->xfer_completion); 415 if (at91_usart_spi_can_dma(ctlr, spi, xfer) && 416 !ret) { 417 ret = at91_usart_spi_dma_transfer(ctlr, xfer); 418 if (ret) 419 continue; 420 421 dma_timeout = at91_usart_spi_dma_timeout(aus); 422 423 if (WARN_ON(dma_timeout == 0)) { 424 dev_err(&spi->dev, "DMA transfer timeout\n"); 425 return -EIO; 426 } 427 aus->current_tx_remaining_bytes = 0; 428 } else { 429 at91_usart_spi_read_status(aus); 430 at91_usart_spi_tx(aus); 431 } 432 433 cpu_relax(); 434 } 435 436 if (aus->xfer_failed) { 437 dev_err(aus->dev, "Overrun!\n"); 438 return -EIO; 439 } 440 441 return 0; 442 } 443 444 static int at91_usart_spi_prepare_message(struct spi_controller *ctlr, 445 struct spi_message *message) 446 { 447 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); 448 struct spi_device *spi = message->spi; 449 u32 *ausd = spi->controller_state; 450 451 at91_usart_spi_writel(aus, CR, US_ENABLE); 452 at91_usart_spi_writel(aus, IER, US_OVRE_RXRDY_IRQS); 453 at91_usart_spi_writel(aus, MR, *ausd); 454 455 return 0; 456 } 457 458 static int at91_usart_spi_unprepare_message(struct spi_controller *ctlr, 459 struct spi_message *message) 460 { 461 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); 462 463 at91_usart_spi_writel(aus, CR, US_RESET | US_DISABLE); 464 at91_usart_spi_writel(aus, IDR, US_OVRE_RXRDY_IRQS); 465 466 return 0; 467 } 468 469 static void at91_usart_spi_cleanup(struct spi_device *spi) 470 { 471 struct at91_usart_spi_device *ausd = spi->controller_state; 472 473 spi->controller_state = NULL; 474 kfree(ausd); 475 } 476 477 static void at91_usart_spi_init(struct at91_usart_spi *aus) 478 { 479 at91_usart_spi_writel(aus, MR, US_INIT); 480 at91_usart_spi_writel(aus, CR, US_RESET | US_DISABLE); 481 } 482 483 static int at91_usart_gpio_setup(struct platform_device *pdev) 484 { 485 struct device_node *np = pdev->dev.parent->of_node; 486 int i; 487 int ret; 488 int nb; 489 490 if (!np) 491 return -EINVAL; 492 493 nb = of_gpio_named_count(np, "cs-gpios"); 494 for (i = 0; i < nb; i++) { 495 int cs_gpio = of_get_named_gpio(np, "cs-gpios", i); 496 497 if (cs_gpio < 0) 498 return cs_gpio; 499 500 if (gpio_is_valid(cs_gpio)) { 501 ret = devm_gpio_request_one(&pdev->dev, cs_gpio, 502 GPIOF_DIR_OUT, 503 dev_name(&pdev->dev)); 504 if (ret) 505 return ret; 506 } 507 } 508 509 return 0; 510 } 511 512 static int at91_usart_spi_probe(struct platform_device *pdev) 513 { 514 struct resource *regs; 515 struct spi_controller *controller; 516 struct at91_usart_spi *aus; 517 struct clk *clk; 518 int irq; 519 int ret; 520 521 regs = platform_get_resource(to_platform_device(pdev->dev.parent), 522 IORESOURCE_MEM, 0); 523 if (!regs) 524 return -EINVAL; 525 526 irq = platform_get_irq(to_platform_device(pdev->dev.parent), 0); 527 if (irq < 0) 528 return irq; 529 530 clk = devm_clk_get(pdev->dev.parent, "usart"); 531 if (IS_ERR(clk)) 532 return PTR_ERR(clk); 533 534 ret = -ENOMEM; 535 controller = spi_alloc_master(&pdev->dev, sizeof(*aus)); 536 if (!controller) 537 goto at91_usart_spi_probe_fail; 538 539 ret = at91_usart_gpio_setup(pdev); 540 if (ret) 541 goto at91_usart_spi_probe_fail; 542 543 controller->mode_bits = SPI_CPOL | SPI_CPHA | SPI_LOOP | SPI_CS_HIGH; 544 controller->dev.of_node = pdev->dev.parent->of_node; 545 controller->bits_per_word_mask = SPI_BPW_MASK(8); 546 controller->setup = at91_usart_spi_setup; 547 controller->flags = SPI_MASTER_MUST_RX | SPI_MASTER_MUST_TX; 548 controller->transfer_one = at91_usart_spi_transfer_one; 549 controller->prepare_message = at91_usart_spi_prepare_message; 550 controller->unprepare_message = at91_usart_spi_unprepare_message; 551 controller->can_dma = at91_usart_spi_can_dma; 552 controller->cleanup = at91_usart_spi_cleanup; 553 controller->max_speed_hz = DIV_ROUND_UP(clk_get_rate(clk), 554 US_MIN_CLK_DIV); 555 controller->min_speed_hz = DIV_ROUND_UP(clk_get_rate(clk), 556 US_MAX_CLK_DIV); 557 platform_set_drvdata(pdev, controller); 558 559 aus = spi_master_get_devdata(controller); 560 561 aus->dev = &pdev->dev; 562 aus->regs = devm_ioremap_resource(&pdev->dev, regs); 563 if (IS_ERR(aus->regs)) { 564 ret = PTR_ERR(aus->regs); 565 goto at91_usart_spi_probe_fail; 566 } 567 568 aus->irq = irq; 569 aus->clk = clk; 570 571 ret = devm_request_irq(&pdev->dev, irq, at91_usart_spi_interrupt, 0, 572 dev_name(&pdev->dev), controller); 573 if (ret) 574 goto at91_usart_spi_probe_fail; 575 576 ret = clk_prepare_enable(clk); 577 if (ret) 578 goto at91_usart_spi_probe_fail; 579 580 aus->spi_clk = clk_get_rate(clk); 581 at91_usart_spi_init(aus); 582 583 aus->phybase = regs->start; 584 585 aus->mpdev = to_platform_device(pdev->dev.parent); 586 587 ret = at91_usart_spi_configure_dma(controller, aus); 588 if (ret) 589 goto at91_usart_fail_dma; 590 591 spin_lock_init(&aus->lock); 592 init_completion(&aus->xfer_completion); 593 594 ret = devm_spi_register_master(&pdev->dev, controller); 595 if (ret) 596 goto at91_usart_fail_register_master; 597 598 dev_info(&pdev->dev, 599 "AT91 USART SPI Controller version 0x%x at %pa (irq %d)\n", 600 at91_usart_spi_readl(aus, VERSION), 601 ®s->start, irq); 602 603 return 0; 604 605 at91_usart_fail_register_master: 606 at91_usart_spi_release_dma(controller); 607 at91_usart_fail_dma: 608 clk_disable_unprepare(clk); 609 at91_usart_spi_probe_fail: 610 spi_master_put(controller); 611 return ret; 612 } 613 614 __maybe_unused static int at91_usart_spi_runtime_suspend(struct device *dev) 615 { 616 struct spi_controller *ctlr = dev_get_drvdata(dev); 617 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); 618 619 clk_disable_unprepare(aus->clk); 620 pinctrl_pm_select_sleep_state(dev); 621 622 return 0; 623 } 624 625 __maybe_unused static int at91_usart_spi_runtime_resume(struct device *dev) 626 { 627 struct spi_controller *ctrl = dev_get_drvdata(dev); 628 struct at91_usart_spi *aus = spi_master_get_devdata(ctrl); 629 630 pinctrl_pm_select_default_state(dev); 631 632 return clk_prepare_enable(aus->clk); 633 } 634 635 __maybe_unused static int at91_usart_spi_suspend(struct device *dev) 636 { 637 struct spi_controller *ctrl = dev_get_drvdata(dev); 638 int ret; 639 640 ret = spi_controller_suspend(ctrl); 641 if (ret) 642 return ret; 643 644 if (!pm_runtime_suspended(dev)) 645 at91_usart_spi_runtime_suspend(dev); 646 647 return 0; 648 } 649 650 __maybe_unused static int at91_usart_spi_resume(struct device *dev) 651 { 652 struct spi_controller *ctrl = dev_get_drvdata(dev); 653 struct at91_usart_spi *aus = spi_master_get_devdata(ctrl); 654 int ret; 655 656 if (!pm_runtime_suspended(dev)) { 657 ret = at91_usart_spi_runtime_resume(dev); 658 if (ret) 659 return ret; 660 } 661 662 at91_usart_spi_init(aus); 663 664 return spi_controller_resume(ctrl); 665 } 666 667 static int at91_usart_spi_remove(struct platform_device *pdev) 668 { 669 struct spi_controller *ctlr = platform_get_drvdata(pdev); 670 struct at91_usart_spi *aus = spi_master_get_devdata(ctlr); 671 672 at91_usart_spi_release_dma(ctlr); 673 clk_disable_unprepare(aus->clk); 674 675 return 0; 676 } 677 678 static const struct dev_pm_ops at91_usart_spi_pm_ops = { 679 SET_SYSTEM_SLEEP_PM_OPS(at91_usart_spi_suspend, at91_usart_spi_resume) 680 SET_RUNTIME_PM_OPS(at91_usart_spi_runtime_suspend, 681 at91_usart_spi_runtime_resume, NULL) 682 }; 683 684 static const struct of_device_id at91_usart_spi_dt_ids[] = { 685 { .compatible = "microchip,at91sam9g45-usart-spi"}, 686 { /* sentinel */} 687 }; 688 689 MODULE_DEVICE_TABLE(of, at91_usart_spi_dt_ids); 690 691 static struct platform_driver at91_usart_spi_driver = { 692 .driver = { 693 .name = "at91_usart_spi", 694 .pm = &at91_usart_spi_pm_ops, 695 }, 696 .probe = at91_usart_spi_probe, 697 .remove = at91_usart_spi_remove, 698 }; 699 700 module_platform_driver(at91_usart_spi_driver); 701 702 MODULE_DESCRIPTION("Microchip AT91 USART SPI Controller driver"); 703 MODULE_AUTHOR("Radu Pirea <radu.pirea@microchip.com>"); 704 MODULE_LICENSE("GPL v2"); 705 MODULE_ALIAS("platform:at91_usart_spi"); 706