1 /* Framework for configuring and reading PHY devices 2 * Based on code in sungem_phy.c and gianfar_phy.c 3 * 4 * Author: Andy Fleming 5 * 6 * Copyright (c) 2004 Freescale Semiconductor, Inc. 7 * Copyright (c) 2006, 2007 Maciej W. Rozycki 8 * 9 * This program is free software; you can redistribute it and/or modify it 10 * under the terms of the GNU General Public License as published by the 11 * Free Software Foundation; either version 2 of the License, or (at your 12 * option) any later version. 13 * 14 */ 15 16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 17 18 #include <linux/kernel.h> 19 #include <linux/string.h> 20 #include <linux/errno.h> 21 #include <linux/unistd.h> 22 #include <linux/interrupt.h> 23 #include <linux/delay.h> 24 #include <linux/netdevice.h> 25 #include <linux/etherdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/mm.h> 28 #include <linux/module.h> 29 #include <linux/mii.h> 30 #include <linux/ethtool.h> 31 #include <linux/phy.h> 32 #include <linux/phy_led_triggers.h> 33 #include <linux/workqueue.h> 34 #include <linux/mdio.h> 35 #include <linux/io.h> 36 #include <linux/uaccess.h> 37 #include <linux/atomic.h> 38 39 #include <asm/irq.h> 40 41 #define PHY_STATE_STR(_state) \ 42 case PHY_##_state: \ 43 return __stringify(_state); \ 44 45 static const char *phy_state_to_str(enum phy_state st) 46 { 47 switch (st) { 48 PHY_STATE_STR(DOWN) 49 PHY_STATE_STR(STARTING) 50 PHY_STATE_STR(READY) 51 PHY_STATE_STR(PENDING) 52 PHY_STATE_STR(UP) 53 PHY_STATE_STR(AN) 54 PHY_STATE_STR(RUNNING) 55 PHY_STATE_STR(NOLINK) 56 PHY_STATE_STR(FORCING) 57 PHY_STATE_STR(CHANGELINK) 58 PHY_STATE_STR(HALTED) 59 PHY_STATE_STR(RESUMING) 60 } 61 62 return NULL; 63 } 64 65 66 /** 67 * phy_print_status - Convenience function to print out the current phy status 68 * @phydev: the phy_device struct 69 */ 70 void phy_print_status(struct phy_device *phydev) 71 { 72 if (phydev->link) { 73 netdev_info(phydev->attached_dev, 74 "Link is Up - %s/%s - flow control %s\n", 75 phy_speed_to_str(phydev->speed), 76 phy_duplex_to_str(phydev->duplex), 77 phydev->pause ? "rx/tx" : "off"); 78 } else { 79 netdev_info(phydev->attached_dev, "Link is Down\n"); 80 } 81 } 82 EXPORT_SYMBOL(phy_print_status); 83 84 /** 85 * phy_clear_interrupt - Ack the phy device's interrupt 86 * @phydev: the phy_device struct 87 * 88 * If the @phydev driver has an ack_interrupt function, call it to 89 * ack and clear the phy device's interrupt. 90 * 91 * Returns 0 on success or < 0 on error. 92 */ 93 static int phy_clear_interrupt(struct phy_device *phydev) 94 { 95 if (phydev->drv->ack_interrupt) 96 return phydev->drv->ack_interrupt(phydev); 97 98 return 0; 99 } 100 101 /** 102 * phy_config_interrupt - configure the PHY device for the requested interrupts 103 * @phydev: the phy_device struct 104 * @interrupts: interrupt flags to configure for this @phydev 105 * 106 * Returns 0 on success or < 0 on error. 107 */ 108 static int phy_config_interrupt(struct phy_device *phydev, u32 interrupts) 109 { 110 phydev->interrupts = interrupts; 111 if (phydev->drv->config_intr) 112 return phydev->drv->config_intr(phydev); 113 114 return 0; 115 } 116 117 /** 118 * phy_restart_aneg - restart auto-negotiation 119 * @phydev: target phy_device struct 120 * 121 * Restart the autonegotiation on @phydev. Returns >= 0 on success or 122 * negative errno on error. 123 */ 124 int phy_restart_aneg(struct phy_device *phydev) 125 { 126 int ret; 127 128 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 129 ret = genphy_c45_restart_aneg(phydev); 130 else 131 ret = genphy_restart_aneg(phydev); 132 133 return ret; 134 } 135 EXPORT_SYMBOL_GPL(phy_restart_aneg); 136 137 /** 138 * phy_aneg_done - return auto-negotiation status 139 * @phydev: target phy_device struct 140 * 141 * Description: Return the auto-negotiation status from this @phydev 142 * Returns > 0 on success or < 0 on error. 0 means that auto-negotiation 143 * is still pending. 144 */ 145 int phy_aneg_done(struct phy_device *phydev) 146 { 147 if (phydev->drv && phydev->drv->aneg_done) 148 return phydev->drv->aneg_done(phydev); 149 150 /* Avoid genphy_aneg_done() if the Clause 45 PHY does not 151 * implement Clause 22 registers 152 */ 153 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 154 return -EINVAL; 155 156 return genphy_aneg_done(phydev); 157 } 158 EXPORT_SYMBOL(phy_aneg_done); 159 160 /** 161 * phy_find_valid - find a PHY setting that matches the requested parameters 162 * @speed: desired speed 163 * @duplex: desired duplex 164 * @supported: mask of supported link modes 165 * 166 * Locate a supported phy setting that is, in priority order: 167 * - an exact match for the specified speed and duplex mode 168 * - a match for the specified speed, or slower speed 169 * - the slowest supported speed 170 * Returns the matched phy_setting entry, or %NULL if no supported phy 171 * settings were found. 172 */ 173 static const struct phy_setting * 174 phy_find_valid(int speed, int duplex, u32 supported) 175 { 176 unsigned long mask = supported; 177 178 return phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, false); 179 } 180 181 /** 182 * phy_supported_speeds - return all speeds currently supported by a phy device 183 * @phy: The phy device to return supported speeds of. 184 * @speeds: buffer to store supported speeds in. 185 * @size: size of speeds buffer. 186 * 187 * Description: Returns the number of supported speeds, and fills the speeds 188 * buffer with the supported speeds. If speeds buffer is too small to contain 189 * all currently supported speeds, will return as many speeds as can fit. 190 */ 191 unsigned int phy_supported_speeds(struct phy_device *phy, 192 unsigned int *speeds, 193 unsigned int size) 194 { 195 unsigned long supported = phy->supported; 196 197 return phy_speeds(speeds, size, &supported, BITS_PER_LONG); 198 } 199 200 /** 201 * phy_check_valid - check if there is a valid PHY setting which matches 202 * speed, duplex, and feature mask 203 * @speed: speed to match 204 * @duplex: duplex to match 205 * @features: A mask of the valid settings 206 * 207 * Description: Returns true if there is a valid setting, false otherwise. 208 */ 209 static inline bool phy_check_valid(int speed, int duplex, u32 features) 210 { 211 unsigned long mask = features; 212 213 return !!phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, true); 214 } 215 216 /** 217 * phy_sanitize_settings - make sure the PHY is set to supported speed and duplex 218 * @phydev: the target phy_device struct 219 * 220 * Description: Make sure the PHY is set to supported speeds and 221 * duplexes. Drop down by one in this order: 1000/FULL, 222 * 1000/HALF, 100/FULL, 100/HALF, 10/FULL, 10/HALF. 223 */ 224 static void phy_sanitize_settings(struct phy_device *phydev) 225 { 226 const struct phy_setting *setting; 227 u32 features = phydev->supported; 228 229 /* Sanitize settings based on PHY capabilities */ 230 if ((features & SUPPORTED_Autoneg) == 0) 231 phydev->autoneg = AUTONEG_DISABLE; 232 233 setting = phy_find_valid(phydev->speed, phydev->duplex, features); 234 if (setting) { 235 phydev->speed = setting->speed; 236 phydev->duplex = setting->duplex; 237 } else { 238 /* We failed to find anything (no supported speeds?) */ 239 phydev->speed = SPEED_UNKNOWN; 240 phydev->duplex = DUPLEX_UNKNOWN; 241 } 242 } 243 244 /** 245 * phy_ethtool_sset - generic ethtool sset function, handles all the details 246 * @phydev: target phy_device struct 247 * @cmd: ethtool_cmd 248 * 249 * A few notes about parameter checking: 250 * 251 * - We don't set port or transceiver, so we don't care what they 252 * were set to. 253 * - phy_start_aneg() will make sure forced settings are sane, and 254 * choose the next best ones from the ones selected, so we don't 255 * care if ethtool tries to give us bad values. 256 */ 257 int phy_ethtool_sset(struct phy_device *phydev, struct ethtool_cmd *cmd) 258 { 259 u32 speed = ethtool_cmd_speed(cmd); 260 261 if (cmd->phy_address != phydev->mdio.addr) 262 return -EINVAL; 263 264 /* We make sure that we don't pass unsupported values in to the PHY */ 265 cmd->advertising &= phydev->supported; 266 267 /* Verify the settings we care about. */ 268 if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE) 269 return -EINVAL; 270 271 if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0) 272 return -EINVAL; 273 274 if (cmd->autoneg == AUTONEG_DISABLE && 275 ((speed != SPEED_1000 && 276 speed != SPEED_100 && 277 speed != SPEED_10) || 278 (cmd->duplex != DUPLEX_HALF && 279 cmd->duplex != DUPLEX_FULL))) 280 return -EINVAL; 281 282 phydev->autoneg = cmd->autoneg; 283 284 phydev->speed = speed; 285 286 phydev->advertising = cmd->advertising; 287 288 if (AUTONEG_ENABLE == cmd->autoneg) 289 phydev->advertising |= ADVERTISED_Autoneg; 290 else 291 phydev->advertising &= ~ADVERTISED_Autoneg; 292 293 phydev->duplex = cmd->duplex; 294 295 phydev->mdix_ctrl = cmd->eth_tp_mdix_ctrl; 296 297 /* Restart the PHY */ 298 phy_start_aneg(phydev); 299 300 return 0; 301 } 302 EXPORT_SYMBOL(phy_ethtool_sset); 303 304 int phy_ethtool_ksettings_set(struct phy_device *phydev, 305 const struct ethtool_link_ksettings *cmd) 306 { 307 u8 autoneg = cmd->base.autoneg; 308 u8 duplex = cmd->base.duplex; 309 u32 speed = cmd->base.speed; 310 u32 advertising; 311 312 if (cmd->base.phy_address != phydev->mdio.addr) 313 return -EINVAL; 314 315 ethtool_convert_link_mode_to_legacy_u32(&advertising, 316 cmd->link_modes.advertising); 317 318 /* We make sure that we don't pass unsupported values in to the PHY */ 319 advertising &= phydev->supported; 320 321 /* Verify the settings we care about. */ 322 if (autoneg != AUTONEG_ENABLE && autoneg != AUTONEG_DISABLE) 323 return -EINVAL; 324 325 if (autoneg == AUTONEG_ENABLE && advertising == 0) 326 return -EINVAL; 327 328 if (autoneg == AUTONEG_DISABLE && 329 ((speed != SPEED_1000 && 330 speed != SPEED_100 && 331 speed != SPEED_10) || 332 (duplex != DUPLEX_HALF && 333 duplex != DUPLEX_FULL))) 334 return -EINVAL; 335 336 phydev->autoneg = autoneg; 337 338 phydev->speed = speed; 339 340 phydev->advertising = advertising; 341 342 if (autoneg == AUTONEG_ENABLE) 343 phydev->advertising |= ADVERTISED_Autoneg; 344 else 345 phydev->advertising &= ~ADVERTISED_Autoneg; 346 347 phydev->duplex = duplex; 348 349 phydev->mdix_ctrl = cmd->base.eth_tp_mdix_ctrl; 350 351 /* Restart the PHY */ 352 phy_start_aneg(phydev); 353 354 return 0; 355 } 356 EXPORT_SYMBOL(phy_ethtool_ksettings_set); 357 358 void phy_ethtool_ksettings_get(struct phy_device *phydev, 359 struct ethtool_link_ksettings *cmd) 360 { 361 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, 362 phydev->supported); 363 364 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, 365 phydev->advertising); 366 367 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.lp_advertising, 368 phydev->lp_advertising); 369 370 cmd->base.speed = phydev->speed; 371 cmd->base.duplex = phydev->duplex; 372 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 373 cmd->base.port = PORT_BNC; 374 else 375 cmd->base.port = PORT_MII; 376 cmd->base.transceiver = phy_is_internal(phydev) ? 377 XCVR_INTERNAL : XCVR_EXTERNAL; 378 cmd->base.phy_address = phydev->mdio.addr; 379 cmd->base.autoneg = phydev->autoneg; 380 cmd->base.eth_tp_mdix_ctrl = phydev->mdix_ctrl; 381 cmd->base.eth_tp_mdix = phydev->mdix; 382 } 383 EXPORT_SYMBOL(phy_ethtool_ksettings_get); 384 385 /** 386 * phy_mii_ioctl - generic PHY MII ioctl interface 387 * @phydev: the phy_device struct 388 * @ifr: &struct ifreq for socket ioctl's 389 * @cmd: ioctl cmd to execute 390 * 391 * Note that this function is currently incompatible with the 392 * PHYCONTROL layer. It changes registers without regard to 393 * current state. Use at own risk. 394 */ 395 int phy_mii_ioctl(struct phy_device *phydev, struct ifreq *ifr, int cmd) 396 { 397 struct mii_ioctl_data *mii_data = if_mii(ifr); 398 u16 val = mii_data->val_in; 399 bool change_autoneg = false; 400 401 switch (cmd) { 402 case SIOCGMIIPHY: 403 mii_data->phy_id = phydev->mdio.addr; 404 /* fall through */ 405 406 case SIOCGMIIREG: 407 mii_data->val_out = mdiobus_read(phydev->mdio.bus, 408 mii_data->phy_id, 409 mii_data->reg_num); 410 return 0; 411 412 case SIOCSMIIREG: 413 if (mii_data->phy_id == phydev->mdio.addr) { 414 switch (mii_data->reg_num) { 415 case MII_BMCR: 416 if ((val & (BMCR_RESET | BMCR_ANENABLE)) == 0) { 417 if (phydev->autoneg == AUTONEG_ENABLE) 418 change_autoneg = true; 419 phydev->autoneg = AUTONEG_DISABLE; 420 if (val & BMCR_FULLDPLX) 421 phydev->duplex = DUPLEX_FULL; 422 else 423 phydev->duplex = DUPLEX_HALF; 424 if (val & BMCR_SPEED1000) 425 phydev->speed = SPEED_1000; 426 else if (val & BMCR_SPEED100) 427 phydev->speed = SPEED_100; 428 else phydev->speed = SPEED_10; 429 } 430 else { 431 if (phydev->autoneg == AUTONEG_DISABLE) 432 change_autoneg = true; 433 phydev->autoneg = AUTONEG_ENABLE; 434 } 435 break; 436 case MII_ADVERTISE: 437 phydev->advertising = mii_adv_to_ethtool_adv_t(val); 438 change_autoneg = true; 439 break; 440 default: 441 /* do nothing */ 442 break; 443 } 444 } 445 446 mdiobus_write(phydev->mdio.bus, mii_data->phy_id, 447 mii_data->reg_num, val); 448 449 if (mii_data->phy_id == phydev->mdio.addr && 450 mii_data->reg_num == MII_BMCR && 451 val & BMCR_RESET) 452 return phy_init_hw(phydev); 453 454 if (change_autoneg) 455 return phy_start_aneg(phydev); 456 457 return 0; 458 459 case SIOCSHWTSTAMP: 460 if (phydev->drv && phydev->drv->hwtstamp) 461 return phydev->drv->hwtstamp(phydev, ifr); 462 /* fall through */ 463 464 default: 465 return -EOPNOTSUPP; 466 } 467 } 468 EXPORT_SYMBOL(phy_mii_ioctl); 469 470 /** 471 * phy_start_aneg_priv - start auto-negotiation for this PHY device 472 * @phydev: the phy_device struct 473 * @sync: indicate whether we should wait for the workqueue cancelation 474 * 475 * Description: Sanitizes the settings (if we're not autonegotiating 476 * them), and then calls the driver's config_aneg function. 477 * If the PHYCONTROL Layer is operating, we change the state to 478 * reflect the beginning of Auto-negotiation or forcing. 479 */ 480 static int phy_start_aneg_priv(struct phy_device *phydev, bool sync) 481 { 482 bool trigger = 0; 483 int err; 484 485 if (!phydev->drv) 486 return -EIO; 487 488 mutex_lock(&phydev->lock); 489 490 if (AUTONEG_DISABLE == phydev->autoneg) 491 phy_sanitize_settings(phydev); 492 493 /* Invalidate LP advertising flags */ 494 phydev->lp_advertising = 0; 495 496 if (phydev->drv->config_aneg) 497 err = phydev->drv->config_aneg(phydev); 498 else 499 err = genphy_config_aneg(phydev); 500 if (err < 0) 501 goto out_unlock; 502 503 if (phydev->state != PHY_HALTED) { 504 if (AUTONEG_ENABLE == phydev->autoneg) { 505 phydev->state = PHY_AN; 506 phydev->link_timeout = PHY_AN_TIMEOUT; 507 } else { 508 phydev->state = PHY_FORCING; 509 phydev->link_timeout = PHY_FORCE_TIMEOUT; 510 } 511 } 512 513 /* Re-schedule a PHY state machine to check PHY status because 514 * negotiation may already be done and aneg interrupt may not be 515 * generated. 516 */ 517 if (phy_interrupt_is_valid(phydev) && (phydev->state == PHY_AN)) { 518 err = phy_aneg_done(phydev); 519 if (err > 0) { 520 trigger = true; 521 err = 0; 522 } 523 } 524 525 out_unlock: 526 mutex_unlock(&phydev->lock); 527 528 if (trigger) 529 phy_trigger_machine(phydev, sync); 530 531 return err; 532 } 533 534 /** 535 * phy_start_aneg - start auto-negotiation for this PHY device 536 * @phydev: the phy_device struct 537 * 538 * Description: Sanitizes the settings (if we're not autonegotiating 539 * them), and then calls the driver's config_aneg function. 540 * If the PHYCONTROL Layer is operating, we change the state to 541 * reflect the beginning of Auto-negotiation or forcing. 542 */ 543 int phy_start_aneg(struct phy_device *phydev) 544 { 545 return phy_start_aneg_priv(phydev, true); 546 } 547 EXPORT_SYMBOL(phy_start_aneg); 548 549 /** 550 * phy_start_machine - start PHY state machine tracking 551 * @phydev: the phy_device struct 552 * 553 * Description: The PHY infrastructure can run a state machine 554 * which tracks whether the PHY is starting up, negotiating, 555 * etc. This function starts the delayed workqueue which tracks 556 * the state of the PHY. If you want to maintain your own state machine, 557 * do not call this function. 558 */ 559 void phy_start_machine(struct phy_device *phydev) 560 { 561 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, HZ); 562 } 563 EXPORT_SYMBOL_GPL(phy_start_machine); 564 565 /** 566 * phy_trigger_machine - trigger the state machine to run 567 * 568 * @phydev: the phy_device struct 569 * @sync: indicate whether we should wait for the workqueue cancelation 570 * 571 * Description: There has been a change in state which requires that the 572 * state machine runs. 573 */ 574 575 void phy_trigger_machine(struct phy_device *phydev, bool sync) 576 { 577 if (sync) 578 cancel_delayed_work_sync(&phydev->state_queue); 579 else 580 cancel_delayed_work(&phydev->state_queue); 581 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 0); 582 } 583 584 /** 585 * phy_stop_machine - stop the PHY state machine tracking 586 * @phydev: target phy_device struct 587 * 588 * Description: Stops the state machine delayed workqueue, sets the 589 * state to UP (unless it wasn't up yet). This function must be 590 * called BEFORE phy_detach. 591 */ 592 void phy_stop_machine(struct phy_device *phydev) 593 { 594 cancel_delayed_work_sync(&phydev->state_queue); 595 596 mutex_lock(&phydev->lock); 597 if (phydev->state > PHY_UP && phydev->state != PHY_HALTED) 598 phydev->state = PHY_UP; 599 mutex_unlock(&phydev->lock); 600 } 601 602 /** 603 * phy_error - enter HALTED state for this PHY device 604 * @phydev: target phy_device struct 605 * 606 * Moves the PHY to the HALTED state in response to a read 607 * or write error, and tells the controller the link is down. 608 * Must not be called from interrupt context, or while the 609 * phydev->lock is held. 610 */ 611 static void phy_error(struct phy_device *phydev) 612 { 613 mutex_lock(&phydev->lock); 614 phydev->state = PHY_HALTED; 615 mutex_unlock(&phydev->lock); 616 617 phy_trigger_machine(phydev, false); 618 } 619 620 /** 621 * phy_disable_interrupts - Disable the PHY interrupts from the PHY side 622 * @phydev: target phy_device struct 623 */ 624 static int phy_disable_interrupts(struct phy_device *phydev) 625 { 626 int err; 627 628 /* Disable PHY interrupts */ 629 err = phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 630 if (err) 631 return err; 632 633 /* Clear the interrupt */ 634 return phy_clear_interrupt(phydev); 635 } 636 637 /** 638 * phy_change - Called by the phy_interrupt to handle PHY changes 639 * @phydev: phy_device struct that interrupted 640 */ 641 static irqreturn_t phy_change(struct phy_device *phydev) 642 { 643 if (phy_interrupt_is_valid(phydev)) { 644 if (phydev->drv->did_interrupt && 645 !phydev->drv->did_interrupt(phydev)) 646 return IRQ_NONE; 647 648 if (phydev->state == PHY_HALTED) 649 if (phy_disable_interrupts(phydev)) 650 goto phy_err; 651 } 652 653 mutex_lock(&phydev->lock); 654 if ((PHY_RUNNING == phydev->state) || (PHY_NOLINK == phydev->state)) 655 phydev->state = PHY_CHANGELINK; 656 mutex_unlock(&phydev->lock); 657 658 /* reschedule state queue work to run as soon as possible */ 659 phy_trigger_machine(phydev, true); 660 661 if (phy_interrupt_is_valid(phydev) && phy_clear_interrupt(phydev)) 662 goto phy_err; 663 return IRQ_HANDLED; 664 665 phy_err: 666 phy_error(phydev); 667 return IRQ_NONE; 668 } 669 670 /** 671 * phy_change_work - Scheduled by the phy_mac_interrupt to handle PHY changes 672 * @work: work_struct that describes the work to be done 673 */ 674 void phy_change_work(struct work_struct *work) 675 { 676 struct phy_device *phydev = 677 container_of(work, struct phy_device, phy_queue); 678 679 phy_change(phydev); 680 } 681 682 /** 683 * phy_interrupt - PHY interrupt handler 684 * @irq: interrupt line 685 * @phy_dat: phy_device pointer 686 * 687 * Description: When a PHY interrupt occurs, the handler disables 688 * interrupts, and uses phy_change to handle the interrupt. 689 */ 690 static irqreturn_t phy_interrupt(int irq, void *phy_dat) 691 { 692 struct phy_device *phydev = phy_dat; 693 694 if (PHY_HALTED == phydev->state) 695 return IRQ_NONE; /* It can't be ours. */ 696 697 return phy_change(phydev); 698 } 699 700 /** 701 * phy_enable_interrupts - Enable the interrupts from the PHY side 702 * @phydev: target phy_device struct 703 */ 704 static int phy_enable_interrupts(struct phy_device *phydev) 705 { 706 int err = phy_clear_interrupt(phydev); 707 708 if (err < 0) 709 return err; 710 711 return phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED); 712 } 713 714 /** 715 * phy_start_interrupts - request and enable interrupts for a PHY device 716 * @phydev: target phy_device struct 717 * 718 * Description: Request the interrupt for the given PHY. 719 * If this fails, then we set irq to PHY_POLL. 720 * Otherwise, we enable the interrupts in the PHY. 721 * This should only be called with a valid IRQ number. 722 * Returns 0 on success or < 0 on error. 723 */ 724 int phy_start_interrupts(struct phy_device *phydev) 725 { 726 if (request_threaded_irq(phydev->irq, NULL, phy_interrupt, 727 IRQF_ONESHOT | IRQF_SHARED, 728 phydev_name(phydev), phydev) < 0) { 729 pr_warn("%s: Can't get IRQ %d (PHY)\n", 730 phydev->mdio.bus->name, phydev->irq); 731 phydev->irq = PHY_POLL; 732 return 0; 733 } 734 735 return phy_enable_interrupts(phydev); 736 } 737 EXPORT_SYMBOL(phy_start_interrupts); 738 739 /** 740 * phy_stop_interrupts - disable interrupts from a PHY device 741 * @phydev: target phy_device struct 742 */ 743 int phy_stop_interrupts(struct phy_device *phydev) 744 { 745 int err = phy_disable_interrupts(phydev); 746 747 if (err) 748 phy_error(phydev); 749 750 free_irq(phydev->irq, phydev); 751 752 return err; 753 } 754 EXPORT_SYMBOL(phy_stop_interrupts); 755 756 /** 757 * phy_stop - Bring down the PHY link, and stop checking the status 758 * @phydev: target phy_device struct 759 */ 760 void phy_stop(struct phy_device *phydev) 761 { 762 mutex_lock(&phydev->lock); 763 764 if (PHY_HALTED == phydev->state) 765 goto out_unlock; 766 767 if (phy_interrupt_is_valid(phydev)) 768 phy_disable_interrupts(phydev); 769 770 phydev->state = PHY_HALTED; 771 772 out_unlock: 773 mutex_unlock(&phydev->lock); 774 775 /* Cannot call flush_scheduled_work() here as desired because 776 * of rtnl_lock(), but PHY_HALTED shall guarantee phy_change() 777 * will not reenable interrupts. 778 */ 779 } 780 EXPORT_SYMBOL(phy_stop); 781 782 /** 783 * phy_start - start or restart a PHY device 784 * @phydev: target phy_device struct 785 * 786 * Description: Indicates the attached device's readiness to 787 * handle PHY-related work. Used during startup to start the 788 * PHY, and after a call to phy_stop() to resume operation. 789 * Also used to indicate the MDIO bus has cleared an error 790 * condition. 791 */ 792 void phy_start(struct phy_device *phydev) 793 { 794 int err = 0; 795 796 mutex_lock(&phydev->lock); 797 798 switch (phydev->state) { 799 case PHY_STARTING: 800 phydev->state = PHY_PENDING; 801 break; 802 case PHY_READY: 803 phydev->state = PHY_UP; 804 break; 805 case PHY_HALTED: 806 /* if phy was suspended, bring the physical link up again */ 807 __phy_resume(phydev); 808 809 /* make sure interrupts are re-enabled for the PHY */ 810 if (phy_interrupt_is_valid(phydev)) { 811 err = phy_enable_interrupts(phydev); 812 if (err < 0) 813 break; 814 } 815 816 phydev->state = PHY_RESUMING; 817 break; 818 default: 819 break; 820 } 821 mutex_unlock(&phydev->lock); 822 823 phy_trigger_machine(phydev, true); 824 } 825 EXPORT_SYMBOL(phy_start); 826 827 static void phy_link_up(struct phy_device *phydev) 828 { 829 phydev->phy_link_change(phydev, true, true); 830 phy_led_trigger_change_speed(phydev); 831 } 832 833 static void phy_link_down(struct phy_device *phydev, bool do_carrier) 834 { 835 phydev->phy_link_change(phydev, false, do_carrier); 836 phy_led_trigger_change_speed(phydev); 837 } 838 839 /** 840 * phy_state_machine - Handle the state machine 841 * @work: work_struct that describes the work to be done 842 */ 843 void phy_state_machine(struct work_struct *work) 844 { 845 struct delayed_work *dwork = to_delayed_work(work); 846 struct phy_device *phydev = 847 container_of(dwork, struct phy_device, state_queue); 848 bool needs_aneg = false, do_suspend = false; 849 enum phy_state old_state; 850 int err = 0; 851 int old_link; 852 853 mutex_lock(&phydev->lock); 854 855 old_state = phydev->state; 856 857 if (phydev->drv && phydev->drv->link_change_notify) 858 phydev->drv->link_change_notify(phydev); 859 860 switch (phydev->state) { 861 case PHY_DOWN: 862 case PHY_STARTING: 863 case PHY_READY: 864 case PHY_PENDING: 865 break; 866 case PHY_UP: 867 needs_aneg = true; 868 869 phydev->link_timeout = PHY_AN_TIMEOUT; 870 871 break; 872 case PHY_AN: 873 err = phy_read_status(phydev); 874 if (err < 0) 875 break; 876 877 /* If the link is down, give up on negotiation for now */ 878 if (!phydev->link) { 879 phydev->state = PHY_NOLINK; 880 phy_link_down(phydev, true); 881 break; 882 } 883 884 /* Check if negotiation is done. Break if there's an error */ 885 err = phy_aneg_done(phydev); 886 if (err < 0) 887 break; 888 889 /* If AN is done, we're running */ 890 if (err > 0) { 891 phydev->state = PHY_RUNNING; 892 phy_link_up(phydev); 893 } else if (0 == phydev->link_timeout--) 894 needs_aneg = true; 895 break; 896 case PHY_NOLINK: 897 if (phydev->irq != PHY_POLL) 898 break; 899 900 err = phy_read_status(phydev); 901 if (err) 902 break; 903 904 if (phydev->link) { 905 if (AUTONEG_ENABLE == phydev->autoneg) { 906 err = phy_aneg_done(phydev); 907 if (err < 0) 908 break; 909 910 if (!err) { 911 phydev->state = PHY_AN; 912 phydev->link_timeout = PHY_AN_TIMEOUT; 913 break; 914 } 915 } 916 phydev->state = PHY_RUNNING; 917 phy_link_up(phydev); 918 } 919 break; 920 case PHY_FORCING: 921 err = genphy_update_link(phydev); 922 if (err) 923 break; 924 925 if (phydev->link) { 926 phydev->state = PHY_RUNNING; 927 phy_link_up(phydev); 928 } else { 929 if (0 == phydev->link_timeout--) 930 needs_aneg = true; 931 phy_link_down(phydev, false); 932 } 933 break; 934 case PHY_RUNNING: 935 /* Only register a CHANGE if we are polling and link changed 936 * since latest checking. 937 */ 938 if (phydev->irq == PHY_POLL) { 939 old_link = phydev->link; 940 err = phy_read_status(phydev); 941 if (err) 942 break; 943 944 if (old_link != phydev->link) 945 phydev->state = PHY_CHANGELINK; 946 } 947 /* 948 * Failsafe: check that nobody set phydev->link=0 between two 949 * poll cycles, otherwise we won't leave RUNNING state as long 950 * as link remains down. 951 */ 952 if (!phydev->link && phydev->state == PHY_RUNNING) { 953 phydev->state = PHY_CHANGELINK; 954 phydev_err(phydev, "no link in PHY_RUNNING\n"); 955 } 956 break; 957 case PHY_CHANGELINK: 958 err = phy_read_status(phydev); 959 if (err) 960 break; 961 962 if (phydev->link) { 963 phydev->state = PHY_RUNNING; 964 phy_link_up(phydev); 965 } else { 966 phydev->state = PHY_NOLINK; 967 phy_link_down(phydev, true); 968 } 969 break; 970 case PHY_HALTED: 971 if (phydev->link) { 972 phydev->link = 0; 973 phy_link_down(phydev, true); 974 do_suspend = true; 975 } 976 break; 977 case PHY_RESUMING: 978 if (AUTONEG_ENABLE == phydev->autoneg) { 979 err = phy_aneg_done(phydev); 980 if (err < 0) 981 break; 982 983 /* err > 0 if AN is done. 984 * Otherwise, it's 0, and we're still waiting for AN 985 */ 986 if (err > 0) { 987 err = phy_read_status(phydev); 988 if (err) 989 break; 990 991 if (phydev->link) { 992 phydev->state = PHY_RUNNING; 993 phy_link_up(phydev); 994 } else { 995 phydev->state = PHY_NOLINK; 996 phy_link_down(phydev, false); 997 } 998 } else { 999 phydev->state = PHY_AN; 1000 phydev->link_timeout = PHY_AN_TIMEOUT; 1001 } 1002 } else { 1003 err = phy_read_status(phydev); 1004 if (err) 1005 break; 1006 1007 if (phydev->link) { 1008 phydev->state = PHY_RUNNING; 1009 phy_link_up(phydev); 1010 } else { 1011 phydev->state = PHY_NOLINK; 1012 phy_link_down(phydev, false); 1013 } 1014 } 1015 break; 1016 } 1017 1018 mutex_unlock(&phydev->lock); 1019 1020 if (needs_aneg) 1021 err = phy_start_aneg_priv(phydev, false); 1022 else if (do_suspend) 1023 phy_suspend(phydev); 1024 1025 if (err < 0) 1026 phy_error(phydev); 1027 1028 if (old_state != phydev->state) 1029 phydev_dbg(phydev, "PHY state change %s -> %s\n", 1030 phy_state_to_str(old_state), 1031 phy_state_to_str(phydev->state)); 1032 1033 /* Only re-schedule a PHY state machine change if we are polling the 1034 * PHY, if PHY_IGNORE_INTERRUPT is set, then we will be moving 1035 * between states from phy_mac_interrupt() 1036 */ 1037 if (phydev->irq == PHY_POLL) 1038 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 1039 PHY_STATE_TIME * HZ); 1040 } 1041 1042 /** 1043 * phy_mac_interrupt - MAC says the link has changed 1044 * @phydev: phy_device struct with changed link 1045 * 1046 * The MAC layer is able to indicate there has been a change in the PHY link 1047 * status. Trigger the state machine and work a work queue. 1048 */ 1049 void phy_mac_interrupt(struct phy_device *phydev) 1050 { 1051 /* Trigger a state machine change */ 1052 queue_work(system_power_efficient_wq, &phydev->phy_queue); 1053 } 1054 EXPORT_SYMBOL(phy_mac_interrupt); 1055 1056 /** 1057 * phy_init_eee - init and check the EEE feature 1058 * @phydev: target phy_device struct 1059 * @clk_stop_enable: PHY may stop the clock during LPI 1060 * 1061 * Description: it checks if the Energy-Efficient Ethernet (EEE) 1062 * is supported by looking at the MMD registers 3.20 and 7.60/61 1063 * and it programs the MMD register 3.0 setting the "Clock stop enable" 1064 * bit if required. 1065 */ 1066 int phy_init_eee(struct phy_device *phydev, bool clk_stop_enable) 1067 { 1068 if (!phydev->drv) 1069 return -EIO; 1070 1071 /* According to 802.3az,the EEE is supported only in full duplex-mode. 1072 */ 1073 if (phydev->duplex == DUPLEX_FULL) { 1074 int eee_lp, eee_cap, eee_adv; 1075 u32 lp, cap, adv; 1076 int status; 1077 1078 /* Read phy status to properly get the right settings */ 1079 status = phy_read_status(phydev); 1080 if (status) 1081 return status; 1082 1083 /* First check if the EEE ability is supported */ 1084 eee_cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1085 if (eee_cap <= 0) 1086 goto eee_exit_err; 1087 1088 cap = mmd_eee_cap_to_ethtool_sup_t(eee_cap); 1089 if (!cap) 1090 goto eee_exit_err; 1091 1092 /* Check which link settings negotiated and verify it in 1093 * the EEE advertising registers. 1094 */ 1095 eee_lp = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1096 if (eee_lp <= 0) 1097 goto eee_exit_err; 1098 1099 eee_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1100 if (eee_adv <= 0) 1101 goto eee_exit_err; 1102 1103 adv = mmd_eee_adv_to_ethtool_adv_t(eee_adv); 1104 lp = mmd_eee_adv_to_ethtool_adv_t(eee_lp); 1105 if (!phy_check_valid(phydev->speed, phydev->duplex, lp & adv)) 1106 goto eee_exit_err; 1107 1108 if (clk_stop_enable) { 1109 /* Configure the PHY to stop receiving xMII 1110 * clock while it is signaling LPI. 1111 */ 1112 int val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1); 1113 if (val < 0) 1114 return val; 1115 1116 val |= MDIO_PCS_CTRL1_CLKSTOP_EN; 1117 phy_write_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1, val); 1118 } 1119 1120 return 0; /* EEE supported */ 1121 } 1122 eee_exit_err: 1123 return -EPROTONOSUPPORT; 1124 } 1125 EXPORT_SYMBOL(phy_init_eee); 1126 1127 /** 1128 * phy_get_eee_err - report the EEE wake error count 1129 * @phydev: target phy_device struct 1130 * 1131 * Description: it is to report the number of time where the PHY 1132 * failed to complete its normal wake sequence. 1133 */ 1134 int phy_get_eee_err(struct phy_device *phydev) 1135 { 1136 if (!phydev->drv) 1137 return -EIO; 1138 1139 return phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_WK_ERR); 1140 } 1141 EXPORT_SYMBOL(phy_get_eee_err); 1142 1143 /** 1144 * phy_ethtool_get_eee - get EEE supported and status 1145 * @phydev: target phy_device struct 1146 * @data: ethtool_eee data 1147 * 1148 * Description: it reportes the Supported/Advertisement/LP Advertisement 1149 * capabilities. 1150 */ 1151 int phy_ethtool_get_eee(struct phy_device *phydev, struct ethtool_eee *data) 1152 { 1153 int val; 1154 1155 if (!phydev->drv) 1156 return -EIO; 1157 1158 /* Get Supported EEE */ 1159 val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1160 if (val < 0) 1161 return val; 1162 data->supported = mmd_eee_cap_to_ethtool_sup_t(val); 1163 1164 /* Get advertisement EEE */ 1165 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1166 if (val < 0) 1167 return val; 1168 data->advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1169 1170 /* Get LP advertisement EEE */ 1171 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1172 if (val < 0) 1173 return val; 1174 data->lp_advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1175 1176 return 0; 1177 } 1178 EXPORT_SYMBOL(phy_ethtool_get_eee); 1179 1180 /** 1181 * phy_ethtool_set_eee - set EEE supported and status 1182 * @phydev: target phy_device struct 1183 * @data: ethtool_eee data 1184 * 1185 * Description: it is to program the Advertisement EEE register. 1186 */ 1187 int phy_ethtool_set_eee(struct phy_device *phydev, struct ethtool_eee *data) 1188 { 1189 int cap, old_adv, adv, ret; 1190 1191 if (!phydev->drv) 1192 return -EIO; 1193 1194 /* Get Supported EEE */ 1195 cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1196 if (cap < 0) 1197 return cap; 1198 1199 old_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1200 if (old_adv < 0) 1201 return old_adv; 1202 1203 adv = ethtool_adv_to_mmd_eee_adv_t(data->advertised) & cap; 1204 1205 /* Mask prohibited EEE modes */ 1206 adv &= ~phydev->eee_broken_modes; 1207 1208 if (old_adv != adv) { 1209 ret = phy_write_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV, adv); 1210 if (ret < 0) 1211 return ret; 1212 1213 /* Restart autonegotiation so the new modes get sent to the 1214 * link partner. 1215 */ 1216 ret = phy_restart_aneg(phydev); 1217 if (ret < 0) 1218 return ret; 1219 } 1220 1221 return 0; 1222 } 1223 EXPORT_SYMBOL(phy_ethtool_set_eee); 1224 1225 int phy_ethtool_set_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1226 { 1227 if (phydev->drv && phydev->drv->set_wol) 1228 return phydev->drv->set_wol(phydev, wol); 1229 1230 return -EOPNOTSUPP; 1231 } 1232 EXPORT_SYMBOL(phy_ethtool_set_wol); 1233 1234 void phy_ethtool_get_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1235 { 1236 if (phydev->drv && phydev->drv->get_wol) 1237 phydev->drv->get_wol(phydev, wol); 1238 } 1239 EXPORT_SYMBOL(phy_ethtool_get_wol); 1240 1241 int phy_ethtool_get_link_ksettings(struct net_device *ndev, 1242 struct ethtool_link_ksettings *cmd) 1243 { 1244 struct phy_device *phydev = ndev->phydev; 1245 1246 if (!phydev) 1247 return -ENODEV; 1248 1249 phy_ethtool_ksettings_get(phydev, cmd); 1250 1251 return 0; 1252 } 1253 EXPORT_SYMBOL(phy_ethtool_get_link_ksettings); 1254 1255 int phy_ethtool_set_link_ksettings(struct net_device *ndev, 1256 const struct ethtool_link_ksettings *cmd) 1257 { 1258 struct phy_device *phydev = ndev->phydev; 1259 1260 if (!phydev) 1261 return -ENODEV; 1262 1263 return phy_ethtool_ksettings_set(phydev, cmd); 1264 } 1265 EXPORT_SYMBOL(phy_ethtool_set_link_ksettings); 1266 1267 int phy_ethtool_nway_reset(struct net_device *ndev) 1268 { 1269 struct phy_device *phydev = ndev->phydev; 1270 1271 if (!phydev) 1272 return -ENODEV; 1273 1274 if (!phydev->drv) 1275 return -EIO; 1276 1277 return phy_restart_aneg(phydev); 1278 } 1279 EXPORT_SYMBOL(phy_ethtool_nway_reset); 1280