1 /* Framework for configuring and reading PHY devices 2 * Based on code in sungem_phy.c and gianfar_phy.c 3 * 4 * Author: Andy Fleming 5 * 6 * Copyright (c) 2004 Freescale Semiconductor, Inc. 7 * Copyright (c) 2006, 2007 Maciej W. Rozycki 8 * 9 * This program is free software; you can redistribute it and/or modify it 10 * under the terms of the GNU General Public License as published by the 11 * Free Software Foundation; either version 2 of the License, or (at your 12 * option) any later version. 13 * 14 */ 15 16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 17 18 #include <linux/kernel.h> 19 #include <linux/string.h> 20 #include <linux/errno.h> 21 #include <linux/unistd.h> 22 #include <linux/interrupt.h> 23 #include <linux/delay.h> 24 #include <linux/netdevice.h> 25 #include <linux/etherdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/mm.h> 28 #include <linux/module.h> 29 #include <linux/mii.h> 30 #include <linux/ethtool.h> 31 #include <linux/phy.h> 32 #include <linux/phy_led_triggers.h> 33 #include <linux/workqueue.h> 34 #include <linux/mdio.h> 35 #include <linux/io.h> 36 #include <linux/uaccess.h> 37 #include <linux/atomic.h> 38 39 #include <asm/irq.h> 40 41 #define PHY_STATE_STR(_state) \ 42 case PHY_##_state: \ 43 return __stringify(_state); \ 44 45 static const char *phy_state_to_str(enum phy_state st) 46 { 47 switch (st) { 48 PHY_STATE_STR(DOWN) 49 PHY_STATE_STR(STARTING) 50 PHY_STATE_STR(READY) 51 PHY_STATE_STR(PENDING) 52 PHY_STATE_STR(UP) 53 PHY_STATE_STR(AN) 54 PHY_STATE_STR(RUNNING) 55 PHY_STATE_STR(NOLINK) 56 PHY_STATE_STR(FORCING) 57 PHY_STATE_STR(CHANGELINK) 58 PHY_STATE_STR(HALTED) 59 PHY_STATE_STR(RESUMING) 60 } 61 62 return NULL; 63 } 64 65 66 /** 67 * phy_print_status - Convenience function to print out the current phy status 68 * @phydev: the phy_device struct 69 */ 70 void phy_print_status(struct phy_device *phydev) 71 { 72 if (phydev->link) { 73 netdev_info(phydev->attached_dev, 74 "Link is Up - %s/%s - flow control %s\n", 75 phy_speed_to_str(phydev->speed), 76 phy_duplex_to_str(phydev->duplex), 77 phydev->pause ? "rx/tx" : "off"); 78 } else { 79 netdev_info(phydev->attached_dev, "Link is Down\n"); 80 } 81 } 82 EXPORT_SYMBOL(phy_print_status); 83 84 /** 85 * phy_clear_interrupt - Ack the phy device's interrupt 86 * @phydev: the phy_device struct 87 * 88 * If the @phydev driver has an ack_interrupt function, call it to 89 * ack and clear the phy device's interrupt. 90 * 91 * Returns 0 on success or < 0 on error. 92 */ 93 static int phy_clear_interrupt(struct phy_device *phydev) 94 { 95 if (phydev->drv->ack_interrupt) 96 return phydev->drv->ack_interrupt(phydev); 97 98 return 0; 99 } 100 101 /** 102 * phy_config_interrupt - configure the PHY device for the requested interrupts 103 * @phydev: the phy_device struct 104 * @interrupts: interrupt flags to configure for this @phydev 105 * 106 * Returns 0 on success or < 0 on error. 107 */ 108 static int phy_config_interrupt(struct phy_device *phydev, u32 interrupts) 109 { 110 phydev->interrupts = interrupts; 111 if (phydev->drv->config_intr) 112 return phydev->drv->config_intr(phydev); 113 114 return 0; 115 } 116 117 /** 118 * phy_restart_aneg - restart auto-negotiation 119 * @phydev: target phy_device struct 120 * 121 * Restart the autonegotiation on @phydev. Returns >= 0 on success or 122 * negative errno on error. 123 */ 124 int phy_restart_aneg(struct phy_device *phydev) 125 { 126 int ret; 127 128 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 129 ret = genphy_c45_restart_aneg(phydev); 130 else 131 ret = genphy_restart_aneg(phydev); 132 133 return ret; 134 } 135 EXPORT_SYMBOL_GPL(phy_restart_aneg); 136 137 /** 138 * phy_aneg_done - return auto-negotiation status 139 * @phydev: target phy_device struct 140 * 141 * Description: Return the auto-negotiation status from this @phydev 142 * Returns > 0 on success or < 0 on error. 0 means that auto-negotiation 143 * is still pending. 144 */ 145 int phy_aneg_done(struct phy_device *phydev) 146 { 147 if (phydev->drv && phydev->drv->aneg_done) 148 return phydev->drv->aneg_done(phydev); 149 150 /* Avoid genphy_aneg_done() if the Clause 45 PHY does not 151 * implement Clause 22 registers 152 */ 153 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 154 return -EINVAL; 155 156 return genphy_aneg_done(phydev); 157 } 158 EXPORT_SYMBOL(phy_aneg_done); 159 160 /** 161 * phy_find_valid - find a PHY setting that matches the requested parameters 162 * @speed: desired speed 163 * @duplex: desired duplex 164 * @supported: mask of supported link modes 165 * 166 * Locate a supported phy setting that is, in priority order: 167 * - an exact match for the specified speed and duplex mode 168 * - a match for the specified speed, or slower speed 169 * - the slowest supported speed 170 * Returns the matched phy_setting entry, or %NULL if no supported phy 171 * settings were found. 172 */ 173 static const struct phy_setting * 174 phy_find_valid(int speed, int duplex, u32 supported) 175 { 176 unsigned long mask = supported; 177 178 return phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, false); 179 } 180 181 /** 182 * phy_supported_speeds - return all speeds currently supported by a phy device 183 * @phy: The phy device to return supported speeds of. 184 * @speeds: buffer to store supported speeds in. 185 * @size: size of speeds buffer. 186 * 187 * Description: Returns the number of supported speeds, and fills the speeds 188 * buffer with the supported speeds. If speeds buffer is too small to contain 189 * all currently supported speeds, will return as many speeds as can fit. 190 */ 191 unsigned int phy_supported_speeds(struct phy_device *phy, 192 unsigned int *speeds, 193 unsigned int size) 194 { 195 unsigned long supported = phy->supported; 196 197 return phy_speeds(speeds, size, &supported, BITS_PER_LONG); 198 } 199 200 /** 201 * phy_check_valid - check if there is a valid PHY setting which matches 202 * speed, duplex, and feature mask 203 * @speed: speed to match 204 * @duplex: duplex to match 205 * @features: A mask of the valid settings 206 * 207 * Description: Returns true if there is a valid setting, false otherwise. 208 */ 209 static inline bool phy_check_valid(int speed, int duplex, u32 features) 210 { 211 unsigned long mask = features; 212 213 return !!phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, true); 214 } 215 216 /** 217 * phy_sanitize_settings - make sure the PHY is set to supported speed and duplex 218 * @phydev: the target phy_device struct 219 * 220 * Description: Make sure the PHY is set to supported speeds and 221 * duplexes. Drop down by one in this order: 1000/FULL, 222 * 1000/HALF, 100/FULL, 100/HALF, 10/FULL, 10/HALF. 223 */ 224 static void phy_sanitize_settings(struct phy_device *phydev) 225 { 226 const struct phy_setting *setting; 227 u32 features = phydev->supported; 228 229 /* Sanitize settings based on PHY capabilities */ 230 if ((features & SUPPORTED_Autoneg) == 0) 231 phydev->autoneg = AUTONEG_DISABLE; 232 233 setting = phy_find_valid(phydev->speed, phydev->duplex, features); 234 if (setting) { 235 phydev->speed = setting->speed; 236 phydev->duplex = setting->duplex; 237 } else { 238 /* We failed to find anything (no supported speeds?) */ 239 phydev->speed = SPEED_UNKNOWN; 240 phydev->duplex = DUPLEX_UNKNOWN; 241 } 242 } 243 244 /** 245 * phy_ethtool_sset - generic ethtool sset function, handles all the details 246 * @phydev: target phy_device struct 247 * @cmd: ethtool_cmd 248 * 249 * A few notes about parameter checking: 250 * 251 * - We don't set port or transceiver, so we don't care what they 252 * were set to. 253 * - phy_start_aneg() will make sure forced settings are sane, and 254 * choose the next best ones from the ones selected, so we don't 255 * care if ethtool tries to give us bad values. 256 */ 257 int phy_ethtool_sset(struct phy_device *phydev, struct ethtool_cmd *cmd) 258 { 259 u32 speed = ethtool_cmd_speed(cmd); 260 261 if (cmd->phy_address != phydev->mdio.addr) 262 return -EINVAL; 263 264 /* We make sure that we don't pass unsupported values in to the PHY */ 265 cmd->advertising &= phydev->supported; 266 267 /* Verify the settings we care about. */ 268 if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE) 269 return -EINVAL; 270 271 if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0) 272 return -EINVAL; 273 274 if (cmd->autoneg == AUTONEG_DISABLE && 275 ((speed != SPEED_1000 && 276 speed != SPEED_100 && 277 speed != SPEED_10) || 278 (cmd->duplex != DUPLEX_HALF && 279 cmd->duplex != DUPLEX_FULL))) 280 return -EINVAL; 281 282 phydev->autoneg = cmd->autoneg; 283 284 phydev->speed = speed; 285 286 phydev->advertising = cmd->advertising; 287 288 if (AUTONEG_ENABLE == cmd->autoneg) 289 phydev->advertising |= ADVERTISED_Autoneg; 290 else 291 phydev->advertising &= ~ADVERTISED_Autoneg; 292 293 phydev->duplex = cmd->duplex; 294 295 phydev->mdix_ctrl = cmd->eth_tp_mdix_ctrl; 296 297 /* Restart the PHY */ 298 phy_start_aneg(phydev); 299 300 return 0; 301 } 302 EXPORT_SYMBOL(phy_ethtool_sset); 303 304 int phy_ethtool_ksettings_set(struct phy_device *phydev, 305 const struct ethtool_link_ksettings *cmd) 306 { 307 u8 autoneg = cmd->base.autoneg; 308 u8 duplex = cmd->base.duplex; 309 u32 speed = cmd->base.speed; 310 u32 advertising; 311 312 if (cmd->base.phy_address != phydev->mdio.addr) 313 return -EINVAL; 314 315 ethtool_convert_link_mode_to_legacy_u32(&advertising, 316 cmd->link_modes.advertising); 317 318 /* We make sure that we don't pass unsupported values in to the PHY */ 319 advertising &= phydev->supported; 320 321 /* Verify the settings we care about. */ 322 if (autoneg != AUTONEG_ENABLE && autoneg != AUTONEG_DISABLE) 323 return -EINVAL; 324 325 if (autoneg == AUTONEG_ENABLE && advertising == 0) 326 return -EINVAL; 327 328 if (autoneg == AUTONEG_DISABLE && 329 ((speed != SPEED_1000 && 330 speed != SPEED_100 && 331 speed != SPEED_10) || 332 (duplex != DUPLEX_HALF && 333 duplex != DUPLEX_FULL))) 334 return -EINVAL; 335 336 phydev->autoneg = autoneg; 337 338 phydev->speed = speed; 339 340 phydev->advertising = advertising; 341 342 if (autoneg == AUTONEG_ENABLE) 343 phydev->advertising |= ADVERTISED_Autoneg; 344 else 345 phydev->advertising &= ~ADVERTISED_Autoneg; 346 347 phydev->duplex = duplex; 348 349 phydev->mdix_ctrl = cmd->base.eth_tp_mdix_ctrl; 350 351 /* Restart the PHY */ 352 phy_start_aneg(phydev); 353 354 return 0; 355 } 356 EXPORT_SYMBOL(phy_ethtool_ksettings_set); 357 358 void phy_ethtool_ksettings_get(struct phy_device *phydev, 359 struct ethtool_link_ksettings *cmd) 360 { 361 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, 362 phydev->supported); 363 364 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, 365 phydev->advertising); 366 367 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.lp_advertising, 368 phydev->lp_advertising); 369 370 cmd->base.speed = phydev->speed; 371 cmd->base.duplex = phydev->duplex; 372 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 373 cmd->base.port = PORT_BNC; 374 else 375 cmd->base.port = PORT_MII; 376 cmd->base.transceiver = phy_is_internal(phydev) ? 377 XCVR_INTERNAL : XCVR_EXTERNAL; 378 cmd->base.phy_address = phydev->mdio.addr; 379 cmd->base.autoneg = phydev->autoneg; 380 cmd->base.eth_tp_mdix_ctrl = phydev->mdix_ctrl; 381 cmd->base.eth_tp_mdix = phydev->mdix; 382 } 383 EXPORT_SYMBOL(phy_ethtool_ksettings_get); 384 385 /** 386 * phy_mii_ioctl - generic PHY MII ioctl interface 387 * @phydev: the phy_device struct 388 * @ifr: &struct ifreq for socket ioctl's 389 * @cmd: ioctl cmd to execute 390 * 391 * Note that this function is currently incompatible with the 392 * PHYCONTROL layer. It changes registers without regard to 393 * current state. Use at own risk. 394 */ 395 int phy_mii_ioctl(struct phy_device *phydev, struct ifreq *ifr, int cmd) 396 { 397 struct mii_ioctl_data *mii_data = if_mii(ifr); 398 u16 val = mii_data->val_in; 399 bool change_autoneg = false; 400 401 switch (cmd) { 402 case SIOCGMIIPHY: 403 mii_data->phy_id = phydev->mdio.addr; 404 /* fall through */ 405 406 case SIOCGMIIREG: 407 mii_data->val_out = mdiobus_read(phydev->mdio.bus, 408 mii_data->phy_id, 409 mii_data->reg_num); 410 return 0; 411 412 case SIOCSMIIREG: 413 if (mii_data->phy_id == phydev->mdio.addr) { 414 switch (mii_data->reg_num) { 415 case MII_BMCR: 416 if ((val & (BMCR_RESET | BMCR_ANENABLE)) == 0) { 417 if (phydev->autoneg == AUTONEG_ENABLE) 418 change_autoneg = true; 419 phydev->autoneg = AUTONEG_DISABLE; 420 if (val & BMCR_FULLDPLX) 421 phydev->duplex = DUPLEX_FULL; 422 else 423 phydev->duplex = DUPLEX_HALF; 424 if (val & BMCR_SPEED1000) 425 phydev->speed = SPEED_1000; 426 else if (val & BMCR_SPEED100) 427 phydev->speed = SPEED_100; 428 else phydev->speed = SPEED_10; 429 } 430 else { 431 if (phydev->autoneg == AUTONEG_DISABLE) 432 change_autoneg = true; 433 phydev->autoneg = AUTONEG_ENABLE; 434 } 435 break; 436 case MII_ADVERTISE: 437 phydev->advertising = mii_adv_to_ethtool_adv_t(val); 438 change_autoneg = true; 439 break; 440 default: 441 /* do nothing */ 442 break; 443 } 444 } 445 446 mdiobus_write(phydev->mdio.bus, mii_data->phy_id, 447 mii_data->reg_num, val); 448 449 if (mii_data->phy_id == phydev->mdio.addr && 450 mii_data->reg_num == MII_BMCR && 451 val & BMCR_RESET) 452 return phy_init_hw(phydev); 453 454 if (change_autoneg) 455 return phy_start_aneg(phydev); 456 457 return 0; 458 459 case SIOCSHWTSTAMP: 460 if (phydev->drv && phydev->drv->hwtstamp) 461 return phydev->drv->hwtstamp(phydev, ifr); 462 /* fall through */ 463 464 default: 465 return -EOPNOTSUPP; 466 } 467 } 468 EXPORT_SYMBOL(phy_mii_ioctl); 469 470 static int phy_config_aneg(struct phy_device *phydev) 471 { 472 if (phydev->drv->config_aneg) 473 return phydev->drv->config_aneg(phydev); 474 else 475 return genphy_config_aneg(phydev); 476 } 477 478 /** 479 * phy_start_aneg_priv - start auto-negotiation for this PHY device 480 * @phydev: the phy_device struct 481 * @sync: indicate whether we should wait for the workqueue cancelation 482 * 483 * Description: Sanitizes the settings (if we're not autonegotiating 484 * them), and then calls the driver's config_aneg function. 485 * If the PHYCONTROL Layer is operating, we change the state to 486 * reflect the beginning of Auto-negotiation or forcing. 487 */ 488 static int phy_start_aneg_priv(struct phy_device *phydev, bool sync) 489 { 490 bool trigger = 0; 491 int err; 492 493 if (!phydev->drv) 494 return -EIO; 495 496 mutex_lock(&phydev->lock); 497 498 if (AUTONEG_DISABLE == phydev->autoneg) 499 phy_sanitize_settings(phydev); 500 501 /* Invalidate LP advertising flags */ 502 phydev->lp_advertising = 0; 503 504 err = phy_config_aneg(phydev); 505 if (err < 0) 506 goto out_unlock; 507 508 if (phydev->state != PHY_HALTED) { 509 if (AUTONEG_ENABLE == phydev->autoneg) { 510 phydev->state = PHY_AN; 511 phydev->link_timeout = PHY_AN_TIMEOUT; 512 } else { 513 phydev->state = PHY_FORCING; 514 phydev->link_timeout = PHY_FORCE_TIMEOUT; 515 } 516 } 517 518 /* Re-schedule a PHY state machine to check PHY status because 519 * negotiation may already be done and aneg interrupt may not be 520 * generated. 521 */ 522 if (phy_interrupt_is_valid(phydev) && (phydev->state == PHY_AN)) { 523 err = phy_aneg_done(phydev); 524 if (err > 0) { 525 trigger = true; 526 err = 0; 527 } 528 } 529 530 out_unlock: 531 mutex_unlock(&phydev->lock); 532 533 if (trigger) 534 phy_trigger_machine(phydev, sync); 535 536 return err; 537 } 538 539 /** 540 * phy_start_aneg - start auto-negotiation for this PHY device 541 * @phydev: the phy_device struct 542 * 543 * Description: Sanitizes the settings (if we're not autonegotiating 544 * them), and then calls the driver's config_aneg function. 545 * If the PHYCONTROL Layer is operating, we change the state to 546 * reflect the beginning of Auto-negotiation or forcing. 547 */ 548 int phy_start_aneg(struct phy_device *phydev) 549 { 550 return phy_start_aneg_priv(phydev, true); 551 } 552 EXPORT_SYMBOL(phy_start_aneg); 553 554 static int phy_poll_aneg_done(struct phy_device *phydev) 555 { 556 unsigned int retries = 100; 557 int ret; 558 559 do { 560 msleep(100); 561 ret = phy_aneg_done(phydev); 562 } while (!ret && --retries); 563 564 if (!ret) 565 return -ETIMEDOUT; 566 567 return ret < 0 ? ret : 0; 568 } 569 570 /** 571 * phy_speed_down - set speed to lowest speed supported by both link partners 572 * @phydev: the phy_device struct 573 * @sync: perform action synchronously 574 * 575 * Description: Typically used to save energy when waiting for a WoL packet 576 * 577 * WARNING: Setting sync to false may cause the system being unable to suspend 578 * in case the PHY generates an interrupt when finishing the autonegotiation. 579 * This interrupt may wake up the system immediately after suspend. 580 * Therefore use sync = false only if you're sure it's safe with the respective 581 * network chip. 582 */ 583 int phy_speed_down(struct phy_device *phydev, bool sync) 584 { 585 u32 adv = phydev->lp_advertising & phydev->supported; 586 u32 adv_old = phydev->advertising; 587 int ret; 588 589 if (phydev->autoneg != AUTONEG_ENABLE) 590 return 0; 591 592 if (adv & PHY_10BT_FEATURES) 593 phydev->advertising &= ~(PHY_100BT_FEATURES | 594 PHY_1000BT_FEATURES); 595 else if (adv & PHY_100BT_FEATURES) 596 phydev->advertising &= ~PHY_1000BT_FEATURES; 597 598 if (phydev->advertising == adv_old) 599 return 0; 600 601 ret = phy_config_aneg(phydev); 602 if (ret) 603 return ret; 604 605 return sync ? phy_poll_aneg_done(phydev) : 0; 606 } 607 EXPORT_SYMBOL_GPL(phy_speed_down); 608 609 /** 610 * phy_speed_up - (re)set advertised speeds to all supported speeds 611 * @phydev: the phy_device struct 612 * 613 * Description: Used to revert the effect of phy_speed_down 614 */ 615 int phy_speed_up(struct phy_device *phydev) 616 { 617 u32 mask = PHY_10BT_FEATURES | PHY_100BT_FEATURES | PHY_1000BT_FEATURES; 618 u32 adv_old = phydev->advertising; 619 620 if (phydev->autoneg != AUTONEG_ENABLE) 621 return 0; 622 623 phydev->advertising = (adv_old & ~mask) | (phydev->supported & mask); 624 625 if (phydev->advertising == adv_old) 626 return 0; 627 628 return phy_config_aneg(phydev); 629 } 630 EXPORT_SYMBOL_GPL(phy_speed_up); 631 632 /** 633 * phy_start_machine - start PHY state machine tracking 634 * @phydev: the phy_device struct 635 * 636 * Description: The PHY infrastructure can run a state machine 637 * which tracks whether the PHY is starting up, negotiating, 638 * etc. This function starts the delayed workqueue which tracks 639 * the state of the PHY. If you want to maintain your own state machine, 640 * do not call this function. 641 */ 642 void phy_start_machine(struct phy_device *phydev) 643 { 644 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, HZ); 645 } 646 EXPORT_SYMBOL_GPL(phy_start_machine); 647 648 /** 649 * phy_trigger_machine - trigger the state machine to run 650 * 651 * @phydev: the phy_device struct 652 * @sync: indicate whether we should wait for the workqueue cancelation 653 * 654 * Description: There has been a change in state which requires that the 655 * state machine runs. 656 */ 657 658 void phy_trigger_machine(struct phy_device *phydev, bool sync) 659 { 660 if (sync) 661 cancel_delayed_work_sync(&phydev->state_queue); 662 else 663 cancel_delayed_work(&phydev->state_queue); 664 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 0); 665 } 666 667 /** 668 * phy_stop_machine - stop the PHY state machine tracking 669 * @phydev: target phy_device struct 670 * 671 * Description: Stops the state machine delayed workqueue, sets the 672 * state to UP (unless it wasn't up yet). This function must be 673 * called BEFORE phy_detach. 674 */ 675 void phy_stop_machine(struct phy_device *phydev) 676 { 677 cancel_delayed_work_sync(&phydev->state_queue); 678 679 mutex_lock(&phydev->lock); 680 if (phydev->state > PHY_UP && phydev->state != PHY_HALTED) 681 phydev->state = PHY_UP; 682 mutex_unlock(&phydev->lock); 683 } 684 685 /** 686 * phy_error - enter HALTED state for this PHY device 687 * @phydev: target phy_device struct 688 * 689 * Moves the PHY to the HALTED state in response to a read 690 * or write error, and tells the controller the link is down. 691 * Must not be called from interrupt context, or while the 692 * phydev->lock is held. 693 */ 694 static void phy_error(struct phy_device *phydev) 695 { 696 mutex_lock(&phydev->lock); 697 phydev->state = PHY_HALTED; 698 mutex_unlock(&phydev->lock); 699 700 phy_trigger_machine(phydev, false); 701 } 702 703 /** 704 * phy_disable_interrupts - Disable the PHY interrupts from the PHY side 705 * @phydev: target phy_device struct 706 */ 707 static int phy_disable_interrupts(struct phy_device *phydev) 708 { 709 int err; 710 711 /* Disable PHY interrupts */ 712 err = phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 713 if (err) 714 return err; 715 716 /* Clear the interrupt */ 717 return phy_clear_interrupt(phydev); 718 } 719 720 /** 721 * phy_change - Called by the phy_interrupt to handle PHY changes 722 * @phydev: phy_device struct that interrupted 723 */ 724 static irqreturn_t phy_change(struct phy_device *phydev) 725 { 726 if (phy_interrupt_is_valid(phydev)) { 727 if (phydev->drv->did_interrupt && 728 !phydev->drv->did_interrupt(phydev)) 729 return IRQ_NONE; 730 731 if (phydev->state == PHY_HALTED) 732 if (phy_disable_interrupts(phydev)) 733 goto phy_err; 734 } 735 736 mutex_lock(&phydev->lock); 737 if ((PHY_RUNNING == phydev->state) || (PHY_NOLINK == phydev->state)) 738 phydev->state = PHY_CHANGELINK; 739 mutex_unlock(&phydev->lock); 740 741 /* reschedule state queue work to run as soon as possible */ 742 phy_trigger_machine(phydev, true); 743 744 if (phy_interrupt_is_valid(phydev) && phy_clear_interrupt(phydev)) 745 goto phy_err; 746 return IRQ_HANDLED; 747 748 phy_err: 749 phy_error(phydev); 750 return IRQ_NONE; 751 } 752 753 /** 754 * phy_change_work - Scheduled by the phy_mac_interrupt to handle PHY changes 755 * @work: work_struct that describes the work to be done 756 */ 757 void phy_change_work(struct work_struct *work) 758 { 759 struct phy_device *phydev = 760 container_of(work, struct phy_device, phy_queue); 761 762 phy_change(phydev); 763 } 764 765 /** 766 * phy_interrupt - PHY interrupt handler 767 * @irq: interrupt line 768 * @phy_dat: phy_device pointer 769 * 770 * Description: When a PHY interrupt occurs, the handler disables 771 * interrupts, and uses phy_change to handle the interrupt. 772 */ 773 static irqreturn_t phy_interrupt(int irq, void *phy_dat) 774 { 775 struct phy_device *phydev = phy_dat; 776 777 if (PHY_HALTED == phydev->state) 778 return IRQ_NONE; /* It can't be ours. */ 779 780 return phy_change(phydev); 781 } 782 783 /** 784 * phy_enable_interrupts - Enable the interrupts from the PHY side 785 * @phydev: target phy_device struct 786 */ 787 static int phy_enable_interrupts(struct phy_device *phydev) 788 { 789 int err = phy_clear_interrupt(phydev); 790 791 if (err < 0) 792 return err; 793 794 return phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED); 795 } 796 797 /** 798 * phy_start_interrupts - request and enable interrupts for a PHY device 799 * @phydev: target phy_device struct 800 * 801 * Description: Request the interrupt for the given PHY. 802 * If this fails, then we set irq to PHY_POLL. 803 * Otherwise, we enable the interrupts in the PHY. 804 * This should only be called with a valid IRQ number. 805 * Returns 0 on success or < 0 on error. 806 */ 807 int phy_start_interrupts(struct phy_device *phydev) 808 { 809 if (request_threaded_irq(phydev->irq, NULL, phy_interrupt, 810 IRQF_ONESHOT | IRQF_SHARED, 811 phydev_name(phydev), phydev) < 0) { 812 pr_warn("%s: Can't get IRQ %d (PHY)\n", 813 phydev->mdio.bus->name, phydev->irq); 814 phydev->irq = PHY_POLL; 815 return 0; 816 } 817 818 return phy_enable_interrupts(phydev); 819 } 820 EXPORT_SYMBOL(phy_start_interrupts); 821 822 /** 823 * phy_stop_interrupts - disable interrupts from a PHY device 824 * @phydev: target phy_device struct 825 */ 826 int phy_stop_interrupts(struct phy_device *phydev) 827 { 828 int err = phy_disable_interrupts(phydev); 829 830 if (err) 831 phy_error(phydev); 832 833 free_irq(phydev->irq, phydev); 834 835 return err; 836 } 837 EXPORT_SYMBOL(phy_stop_interrupts); 838 839 /** 840 * phy_stop - Bring down the PHY link, and stop checking the status 841 * @phydev: target phy_device struct 842 */ 843 void phy_stop(struct phy_device *phydev) 844 { 845 mutex_lock(&phydev->lock); 846 847 if (PHY_HALTED == phydev->state) 848 goto out_unlock; 849 850 if (phy_interrupt_is_valid(phydev)) 851 phy_disable_interrupts(phydev); 852 853 phydev->state = PHY_HALTED; 854 855 out_unlock: 856 mutex_unlock(&phydev->lock); 857 858 /* Cannot call flush_scheduled_work() here as desired because 859 * of rtnl_lock(), but PHY_HALTED shall guarantee phy_change() 860 * will not reenable interrupts. 861 */ 862 } 863 EXPORT_SYMBOL(phy_stop); 864 865 /** 866 * phy_start - start or restart a PHY device 867 * @phydev: target phy_device struct 868 * 869 * Description: Indicates the attached device's readiness to 870 * handle PHY-related work. Used during startup to start the 871 * PHY, and after a call to phy_stop() to resume operation. 872 * Also used to indicate the MDIO bus has cleared an error 873 * condition. 874 */ 875 void phy_start(struct phy_device *phydev) 876 { 877 int err = 0; 878 879 mutex_lock(&phydev->lock); 880 881 switch (phydev->state) { 882 case PHY_STARTING: 883 phydev->state = PHY_PENDING; 884 break; 885 case PHY_READY: 886 phydev->state = PHY_UP; 887 break; 888 case PHY_HALTED: 889 /* if phy was suspended, bring the physical link up again */ 890 __phy_resume(phydev); 891 892 /* make sure interrupts are re-enabled for the PHY */ 893 if (phy_interrupt_is_valid(phydev)) { 894 err = phy_enable_interrupts(phydev); 895 if (err < 0) 896 break; 897 } 898 899 phydev->state = PHY_RESUMING; 900 break; 901 default: 902 break; 903 } 904 mutex_unlock(&phydev->lock); 905 906 phy_trigger_machine(phydev, true); 907 } 908 EXPORT_SYMBOL(phy_start); 909 910 static void phy_link_up(struct phy_device *phydev) 911 { 912 phydev->phy_link_change(phydev, true, true); 913 phy_led_trigger_change_speed(phydev); 914 } 915 916 static void phy_link_down(struct phy_device *phydev, bool do_carrier) 917 { 918 phydev->phy_link_change(phydev, false, do_carrier); 919 phy_led_trigger_change_speed(phydev); 920 } 921 922 /** 923 * phy_state_machine - Handle the state machine 924 * @work: work_struct that describes the work to be done 925 */ 926 void phy_state_machine(struct work_struct *work) 927 { 928 struct delayed_work *dwork = to_delayed_work(work); 929 struct phy_device *phydev = 930 container_of(dwork, struct phy_device, state_queue); 931 bool needs_aneg = false, do_suspend = false; 932 enum phy_state old_state; 933 int err = 0; 934 int old_link; 935 936 mutex_lock(&phydev->lock); 937 938 old_state = phydev->state; 939 940 if (phydev->drv && phydev->drv->link_change_notify) 941 phydev->drv->link_change_notify(phydev); 942 943 switch (phydev->state) { 944 case PHY_DOWN: 945 case PHY_STARTING: 946 case PHY_READY: 947 case PHY_PENDING: 948 break; 949 case PHY_UP: 950 needs_aneg = true; 951 952 phydev->link_timeout = PHY_AN_TIMEOUT; 953 954 break; 955 case PHY_AN: 956 err = phy_read_status(phydev); 957 if (err < 0) 958 break; 959 960 /* If the link is down, give up on negotiation for now */ 961 if (!phydev->link) { 962 phydev->state = PHY_NOLINK; 963 phy_link_down(phydev, true); 964 break; 965 } 966 967 /* Check if negotiation is done. Break if there's an error */ 968 err = phy_aneg_done(phydev); 969 if (err < 0) 970 break; 971 972 /* If AN is done, we're running */ 973 if (err > 0) { 974 phydev->state = PHY_RUNNING; 975 phy_link_up(phydev); 976 } else if (0 == phydev->link_timeout--) 977 needs_aneg = true; 978 break; 979 case PHY_NOLINK: 980 if (phydev->irq != PHY_POLL) 981 break; 982 983 err = phy_read_status(phydev); 984 if (err) 985 break; 986 987 if (phydev->link) { 988 if (AUTONEG_ENABLE == phydev->autoneg) { 989 err = phy_aneg_done(phydev); 990 if (err < 0) 991 break; 992 993 if (!err) { 994 phydev->state = PHY_AN; 995 phydev->link_timeout = PHY_AN_TIMEOUT; 996 break; 997 } 998 } 999 phydev->state = PHY_RUNNING; 1000 phy_link_up(phydev); 1001 } 1002 break; 1003 case PHY_FORCING: 1004 err = genphy_update_link(phydev); 1005 if (err) 1006 break; 1007 1008 if (phydev->link) { 1009 phydev->state = PHY_RUNNING; 1010 phy_link_up(phydev); 1011 } else { 1012 if (0 == phydev->link_timeout--) 1013 needs_aneg = true; 1014 phy_link_down(phydev, false); 1015 } 1016 break; 1017 case PHY_RUNNING: 1018 /* Only register a CHANGE if we are polling and link changed 1019 * since latest checking. 1020 */ 1021 if (phydev->irq == PHY_POLL) { 1022 old_link = phydev->link; 1023 err = phy_read_status(phydev); 1024 if (err) 1025 break; 1026 1027 if (old_link != phydev->link) 1028 phydev->state = PHY_CHANGELINK; 1029 } 1030 /* 1031 * Failsafe: check that nobody set phydev->link=0 between two 1032 * poll cycles, otherwise we won't leave RUNNING state as long 1033 * as link remains down. 1034 */ 1035 if (!phydev->link && phydev->state == PHY_RUNNING) { 1036 phydev->state = PHY_CHANGELINK; 1037 phydev_err(phydev, "no link in PHY_RUNNING\n"); 1038 } 1039 break; 1040 case PHY_CHANGELINK: 1041 err = phy_read_status(phydev); 1042 if (err) 1043 break; 1044 1045 if (phydev->link) { 1046 phydev->state = PHY_RUNNING; 1047 phy_link_up(phydev); 1048 } else { 1049 phydev->state = PHY_NOLINK; 1050 phy_link_down(phydev, true); 1051 } 1052 break; 1053 case PHY_HALTED: 1054 if (phydev->link) { 1055 phydev->link = 0; 1056 phy_link_down(phydev, true); 1057 do_suspend = true; 1058 } 1059 break; 1060 case PHY_RESUMING: 1061 if (AUTONEG_ENABLE == phydev->autoneg) { 1062 err = phy_aneg_done(phydev); 1063 if (err < 0) 1064 break; 1065 1066 /* err > 0 if AN is done. 1067 * Otherwise, it's 0, and we're still waiting for AN 1068 */ 1069 if (err > 0) { 1070 err = phy_read_status(phydev); 1071 if (err) 1072 break; 1073 1074 if (phydev->link) { 1075 phydev->state = PHY_RUNNING; 1076 phy_link_up(phydev); 1077 } else { 1078 phydev->state = PHY_NOLINK; 1079 phy_link_down(phydev, false); 1080 } 1081 } else { 1082 phydev->state = PHY_AN; 1083 phydev->link_timeout = PHY_AN_TIMEOUT; 1084 } 1085 } else { 1086 err = phy_read_status(phydev); 1087 if (err) 1088 break; 1089 1090 if (phydev->link) { 1091 phydev->state = PHY_RUNNING; 1092 phy_link_up(phydev); 1093 } else { 1094 phydev->state = PHY_NOLINK; 1095 phy_link_down(phydev, false); 1096 } 1097 } 1098 break; 1099 } 1100 1101 mutex_unlock(&phydev->lock); 1102 1103 if (needs_aneg) 1104 err = phy_start_aneg_priv(phydev, false); 1105 else if (do_suspend) 1106 phy_suspend(phydev); 1107 1108 if (err < 0) 1109 phy_error(phydev); 1110 1111 if (old_state != phydev->state) 1112 phydev_dbg(phydev, "PHY state change %s -> %s\n", 1113 phy_state_to_str(old_state), 1114 phy_state_to_str(phydev->state)); 1115 1116 /* Only re-schedule a PHY state machine change if we are polling the 1117 * PHY, if PHY_IGNORE_INTERRUPT is set, then we will be moving 1118 * between states from phy_mac_interrupt() 1119 */ 1120 if (phydev->irq == PHY_POLL) 1121 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 1122 PHY_STATE_TIME * HZ); 1123 } 1124 1125 /** 1126 * phy_mac_interrupt - MAC says the link has changed 1127 * @phydev: phy_device struct with changed link 1128 * 1129 * The MAC layer is able to indicate there has been a change in the PHY link 1130 * status. Trigger the state machine and work a work queue. 1131 */ 1132 void phy_mac_interrupt(struct phy_device *phydev) 1133 { 1134 /* Trigger a state machine change */ 1135 queue_work(system_power_efficient_wq, &phydev->phy_queue); 1136 } 1137 EXPORT_SYMBOL(phy_mac_interrupt); 1138 1139 /** 1140 * phy_init_eee - init and check the EEE feature 1141 * @phydev: target phy_device struct 1142 * @clk_stop_enable: PHY may stop the clock during LPI 1143 * 1144 * Description: it checks if the Energy-Efficient Ethernet (EEE) 1145 * is supported by looking at the MMD registers 3.20 and 7.60/61 1146 * and it programs the MMD register 3.0 setting the "Clock stop enable" 1147 * bit if required. 1148 */ 1149 int phy_init_eee(struct phy_device *phydev, bool clk_stop_enable) 1150 { 1151 if (!phydev->drv) 1152 return -EIO; 1153 1154 /* According to 802.3az,the EEE is supported only in full duplex-mode. 1155 */ 1156 if (phydev->duplex == DUPLEX_FULL) { 1157 int eee_lp, eee_cap, eee_adv; 1158 u32 lp, cap, adv; 1159 int status; 1160 1161 /* Read phy status to properly get the right settings */ 1162 status = phy_read_status(phydev); 1163 if (status) 1164 return status; 1165 1166 /* First check if the EEE ability is supported */ 1167 eee_cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1168 if (eee_cap <= 0) 1169 goto eee_exit_err; 1170 1171 cap = mmd_eee_cap_to_ethtool_sup_t(eee_cap); 1172 if (!cap) 1173 goto eee_exit_err; 1174 1175 /* Check which link settings negotiated and verify it in 1176 * the EEE advertising registers. 1177 */ 1178 eee_lp = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1179 if (eee_lp <= 0) 1180 goto eee_exit_err; 1181 1182 eee_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1183 if (eee_adv <= 0) 1184 goto eee_exit_err; 1185 1186 adv = mmd_eee_adv_to_ethtool_adv_t(eee_adv); 1187 lp = mmd_eee_adv_to_ethtool_adv_t(eee_lp); 1188 if (!phy_check_valid(phydev->speed, phydev->duplex, lp & adv)) 1189 goto eee_exit_err; 1190 1191 if (clk_stop_enable) { 1192 /* Configure the PHY to stop receiving xMII 1193 * clock while it is signaling LPI. 1194 */ 1195 int val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1); 1196 if (val < 0) 1197 return val; 1198 1199 val |= MDIO_PCS_CTRL1_CLKSTOP_EN; 1200 phy_write_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1, val); 1201 } 1202 1203 return 0; /* EEE supported */ 1204 } 1205 eee_exit_err: 1206 return -EPROTONOSUPPORT; 1207 } 1208 EXPORT_SYMBOL(phy_init_eee); 1209 1210 /** 1211 * phy_get_eee_err - report the EEE wake error count 1212 * @phydev: target phy_device struct 1213 * 1214 * Description: it is to report the number of time where the PHY 1215 * failed to complete its normal wake sequence. 1216 */ 1217 int phy_get_eee_err(struct phy_device *phydev) 1218 { 1219 if (!phydev->drv) 1220 return -EIO; 1221 1222 return phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_WK_ERR); 1223 } 1224 EXPORT_SYMBOL(phy_get_eee_err); 1225 1226 /** 1227 * phy_ethtool_get_eee - get EEE supported and status 1228 * @phydev: target phy_device struct 1229 * @data: ethtool_eee data 1230 * 1231 * Description: it reportes the Supported/Advertisement/LP Advertisement 1232 * capabilities. 1233 */ 1234 int phy_ethtool_get_eee(struct phy_device *phydev, struct ethtool_eee *data) 1235 { 1236 int val; 1237 1238 if (!phydev->drv) 1239 return -EIO; 1240 1241 /* Get Supported EEE */ 1242 val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1243 if (val < 0) 1244 return val; 1245 data->supported = mmd_eee_cap_to_ethtool_sup_t(val); 1246 1247 /* Get advertisement EEE */ 1248 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1249 if (val < 0) 1250 return val; 1251 data->advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1252 1253 /* Get LP advertisement EEE */ 1254 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1255 if (val < 0) 1256 return val; 1257 data->lp_advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1258 1259 return 0; 1260 } 1261 EXPORT_SYMBOL(phy_ethtool_get_eee); 1262 1263 /** 1264 * phy_ethtool_set_eee - set EEE supported and status 1265 * @phydev: target phy_device struct 1266 * @data: ethtool_eee data 1267 * 1268 * Description: it is to program the Advertisement EEE register. 1269 */ 1270 int phy_ethtool_set_eee(struct phy_device *phydev, struct ethtool_eee *data) 1271 { 1272 int cap, old_adv, adv, ret; 1273 1274 if (!phydev->drv) 1275 return -EIO; 1276 1277 /* Get Supported EEE */ 1278 cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1279 if (cap < 0) 1280 return cap; 1281 1282 old_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1283 if (old_adv < 0) 1284 return old_adv; 1285 1286 adv = ethtool_adv_to_mmd_eee_adv_t(data->advertised) & cap; 1287 1288 /* Mask prohibited EEE modes */ 1289 adv &= ~phydev->eee_broken_modes; 1290 1291 if (old_adv != adv) { 1292 ret = phy_write_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV, adv); 1293 if (ret < 0) 1294 return ret; 1295 1296 /* Restart autonegotiation so the new modes get sent to the 1297 * link partner. 1298 */ 1299 ret = phy_restart_aneg(phydev); 1300 if (ret < 0) 1301 return ret; 1302 } 1303 1304 return 0; 1305 } 1306 EXPORT_SYMBOL(phy_ethtool_set_eee); 1307 1308 int phy_ethtool_set_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1309 { 1310 if (phydev->drv && phydev->drv->set_wol) 1311 return phydev->drv->set_wol(phydev, wol); 1312 1313 return -EOPNOTSUPP; 1314 } 1315 EXPORT_SYMBOL(phy_ethtool_set_wol); 1316 1317 void phy_ethtool_get_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1318 { 1319 if (phydev->drv && phydev->drv->get_wol) 1320 phydev->drv->get_wol(phydev, wol); 1321 } 1322 EXPORT_SYMBOL(phy_ethtool_get_wol); 1323 1324 int phy_ethtool_get_link_ksettings(struct net_device *ndev, 1325 struct ethtool_link_ksettings *cmd) 1326 { 1327 struct phy_device *phydev = ndev->phydev; 1328 1329 if (!phydev) 1330 return -ENODEV; 1331 1332 phy_ethtool_ksettings_get(phydev, cmd); 1333 1334 return 0; 1335 } 1336 EXPORT_SYMBOL(phy_ethtool_get_link_ksettings); 1337 1338 int phy_ethtool_set_link_ksettings(struct net_device *ndev, 1339 const struct ethtool_link_ksettings *cmd) 1340 { 1341 struct phy_device *phydev = ndev->phydev; 1342 1343 if (!phydev) 1344 return -ENODEV; 1345 1346 return phy_ethtool_ksettings_set(phydev, cmd); 1347 } 1348 EXPORT_SYMBOL(phy_ethtool_set_link_ksettings); 1349 1350 int phy_ethtool_nway_reset(struct net_device *ndev) 1351 { 1352 struct phy_device *phydev = ndev->phydev; 1353 1354 if (!phydev) 1355 return -ENODEV; 1356 1357 if (!phydev->drv) 1358 return -EIO; 1359 1360 return phy_restart_aneg(phydev); 1361 } 1362 EXPORT_SYMBOL(phy_ethtool_nway_reset); 1363