1 /* Framework for configuring and reading PHY devices 2 * Based on code in sungem_phy.c and gianfar_phy.c 3 * 4 * Author: Andy Fleming 5 * 6 * Copyright (c) 2004 Freescale Semiconductor, Inc. 7 * Copyright (c) 2006, 2007 Maciej W. Rozycki 8 * 9 * This program is free software; you can redistribute it and/or modify it 10 * under the terms of the GNU General Public License as published by the 11 * Free Software Foundation; either version 2 of the License, or (at your 12 * option) any later version. 13 * 14 */ 15 16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 17 18 #include <linux/kernel.h> 19 #include <linux/string.h> 20 #include <linux/errno.h> 21 #include <linux/unistd.h> 22 #include <linux/interrupt.h> 23 #include <linux/delay.h> 24 #include <linux/netdevice.h> 25 #include <linux/etherdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/mm.h> 28 #include <linux/module.h> 29 #include <linux/mii.h> 30 #include <linux/ethtool.h> 31 #include <linux/phy.h> 32 #include <linux/phy_led_triggers.h> 33 #include <linux/workqueue.h> 34 #include <linux/mdio.h> 35 #include <linux/io.h> 36 #include <linux/uaccess.h> 37 #include <linux/atomic.h> 38 39 #include <asm/irq.h> 40 41 #define PHY_STATE_STR(_state) \ 42 case PHY_##_state: \ 43 return __stringify(_state); \ 44 45 static const char *phy_state_to_str(enum phy_state st) 46 { 47 switch (st) { 48 PHY_STATE_STR(DOWN) 49 PHY_STATE_STR(STARTING) 50 PHY_STATE_STR(READY) 51 PHY_STATE_STR(PENDING) 52 PHY_STATE_STR(UP) 53 PHY_STATE_STR(AN) 54 PHY_STATE_STR(RUNNING) 55 PHY_STATE_STR(NOLINK) 56 PHY_STATE_STR(FORCING) 57 PHY_STATE_STR(CHANGELINK) 58 PHY_STATE_STR(HALTED) 59 PHY_STATE_STR(RESUMING) 60 } 61 62 return NULL; 63 } 64 65 66 /** 67 * phy_print_status - Convenience function to print out the current phy status 68 * @phydev: the phy_device struct 69 */ 70 void phy_print_status(struct phy_device *phydev) 71 { 72 if (phydev->link) { 73 netdev_info(phydev->attached_dev, 74 "Link is Up - %s/%s - flow control %s\n", 75 phy_speed_to_str(phydev->speed), 76 phy_duplex_to_str(phydev->duplex), 77 phydev->pause ? "rx/tx" : "off"); 78 } else { 79 netdev_info(phydev->attached_dev, "Link is Down\n"); 80 } 81 } 82 EXPORT_SYMBOL(phy_print_status); 83 84 /** 85 * phy_clear_interrupt - Ack the phy device's interrupt 86 * @phydev: the phy_device struct 87 * 88 * If the @phydev driver has an ack_interrupt function, call it to 89 * ack and clear the phy device's interrupt. 90 * 91 * Returns 0 on success or < 0 on error. 92 */ 93 static int phy_clear_interrupt(struct phy_device *phydev) 94 { 95 if (phydev->drv->ack_interrupt) 96 return phydev->drv->ack_interrupt(phydev); 97 98 return 0; 99 } 100 101 /** 102 * phy_config_interrupt - configure the PHY device for the requested interrupts 103 * @phydev: the phy_device struct 104 * @interrupts: interrupt flags to configure for this @phydev 105 * 106 * Returns 0 on success or < 0 on error. 107 */ 108 static int phy_config_interrupt(struct phy_device *phydev, u32 interrupts) 109 { 110 phydev->interrupts = interrupts; 111 if (phydev->drv->config_intr) 112 return phydev->drv->config_intr(phydev); 113 114 return 0; 115 } 116 117 /** 118 * phy_restart_aneg - restart auto-negotiation 119 * @phydev: target phy_device struct 120 * 121 * Restart the autonegotiation on @phydev. Returns >= 0 on success or 122 * negative errno on error. 123 */ 124 int phy_restart_aneg(struct phy_device *phydev) 125 { 126 int ret; 127 128 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 129 ret = genphy_c45_restart_aneg(phydev); 130 else 131 ret = genphy_restart_aneg(phydev); 132 133 return ret; 134 } 135 EXPORT_SYMBOL_GPL(phy_restart_aneg); 136 137 /** 138 * phy_aneg_done - return auto-negotiation status 139 * @phydev: target phy_device struct 140 * 141 * Description: Return the auto-negotiation status from this @phydev 142 * Returns > 0 on success or < 0 on error. 0 means that auto-negotiation 143 * is still pending. 144 */ 145 int phy_aneg_done(struct phy_device *phydev) 146 { 147 if (phydev->drv && phydev->drv->aneg_done) 148 return phydev->drv->aneg_done(phydev); 149 150 /* Avoid genphy_aneg_done() if the Clause 45 PHY does not 151 * implement Clause 22 registers 152 */ 153 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 154 return -EINVAL; 155 156 return genphy_aneg_done(phydev); 157 } 158 EXPORT_SYMBOL(phy_aneg_done); 159 160 /** 161 * phy_find_valid - find a PHY setting that matches the requested parameters 162 * @speed: desired speed 163 * @duplex: desired duplex 164 * @supported: mask of supported link modes 165 * 166 * Locate a supported phy setting that is, in priority order: 167 * - an exact match for the specified speed and duplex mode 168 * - a match for the specified speed, or slower speed 169 * - the slowest supported speed 170 * Returns the matched phy_setting entry, or %NULL if no supported phy 171 * settings were found. 172 */ 173 static const struct phy_setting * 174 phy_find_valid(int speed, int duplex, u32 supported) 175 { 176 unsigned long mask = supported; 177 178 return phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, false); 179 } 180 181 /** 182 * phy_supported_speeds - return all speeds currently supported by a phy device 183 * @phy: The phy device to return supported speeds of. 184 * @speeds: buffer to store supported speeds in. 185 * @size: size of speeds buffer. 186 * 187 * Description: Returns the number of supported speeds, and fills the speeds 188 * buffer with the supported speeds. If speeds buffer is too small to contain 189 * all currently supported speeds, will return as many speeds as can fit. 190 */ 191 unsigned int phy_supported_speeds(struct phy_device *phy, 192 unsigned int *speeds, 193 unsigned int size) 194 { 195 unsigned long supported = phy->supported; 196 197 return phy_speeds(speeds, size, &supported, BITS_PER_LONG); 198 } 199 200 /** 201 * phy_check_valid - check if there is a valid PHY setting which matches 202 * speed, duplex, and feature mask 203 * @speed: speed to match 204 * @duplex: duplex to match 205 * @features: A mask of the valid settings 206 * 207 * Description: Returns true if there is a valid setting, false otherwise. 208 */ 209 static inline bool phy_check_valid(int speed, int duplex, u32 features) 210 { 211 unsigned long mask = features; 212 213 return !!phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, true); 214 } 215 216 /** 217 * phy_sanitize_settings - make sure the PHY is set to supported speed and duplex 218 * @phydev: the target phy_device struct 219 * 220 * Description: Make sure the PHY is set to supported speeds and 221 * duplexes. Drop down by one in this order: 1000/FULL, 222 * 1000/HALF, 100/FULL, 100/HALF, 10/FULL, 10/HALF. 223 */ 224 static void phy_sanitize_settings(struct phy_device *phydev) 225 { 226 const struct phy_setting *setting; 227 u32 features = phydev->supported; 228 229 /* Sanitize settings based on PHY capabilities */ 230 if ((features & SUPPORTED_Autoneg) == 0) 231 phydev->autoneg = AUTONEG_DISABLE; 232 233 setting = phy_find_valid(phydev->speed, phydev->duplex, features); 234 if (setting) { 235 phydev->speed = setting->speed; 236 phydev->duplex = setting->duplex; 237 } else { 238 /* We failed to find anything (no supported speeds?) */ 239 phydev->speed = SPEED_UNKNOWN; 240 phydev->duplex = DUPLEX_UNKNOWN; 241 } 242 } 243 244 /** 245 * phy_ethtool_sset - generic ethtool sset function, handles all the details 246 * @phydev: target phy_device struct 247 * @cmd: ethtool_cmd 248 * 249 * A few notes about parameter checking: 250 * 251 * - We don't set port or transceiver, so we don't care what they 252 * were set to. 253 * - phy_start_aneg() will make sure forced settings are sane, and 254 * choose the next best ones from the ones selected, so we don't 255 * care if ethtool tries to give us bad values. 256 */ 257 int phy_ethtool_sset(struct phy_device *phydev, struct ethtool_cmd *cmd) 258 { 259 u32 speed = ethtool_cmd_speed(cmd); 260 261 if (cmd->phy_address != phydev->mdio.addr) 262 return -EINVAL; 263 264 /* We make sure that we don't pass unsupported values in to the PHY */ 265 cmd->advertising &= phydev->supported; 266 267 /* Verify the settings we care about. */ 268 if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE) 269 return -EINVAL; 270 271 if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0) 272 return -EINVAL; 273 274 if (cmd->autoneg == AUTONEG_DISABLE && 275 ((speed != SPEED_1000 && 276 speed != SPEED_100 && 277 speed != SPEED_10) || 278 (cmd->duplex != DUPLEX_HALF && 279 cmd->duplex != DUPLEX_FULL))) 280 return -EINVAL; 281 282 phydev->autoneg = cmd->autoneg; 283 284 phydev->speed = speed; 285 286 phydev->advertising = cmd->advertising; 287 288 if (AUTONEG_ENABLE == cmd->autoneg) 289 phydev->advertising |= ADVERTISED_Autoneg; 290 else 291 phydev->advertising &= ~ADVERTISED_Autoneg; 292 293 phydev->duplex = cmd->duplex; 294 295 phydev->mdix_ctrl = cmd->eth_tp_mdix_ctrl; 296 297 /* Restart the PHY */ 298 phy_start_aneg(phydev); 299 300 return 0; 301 } 302 EXPORT_SYMBOL(phy_ethtool_sset); 303 304 int phy_ethtool_ksettings_set(struct phy_device *phydev, 305 const struct ethtool_link_ksettings *cmd) 306 { 307 u8 autoneg = cmd->base.autoneg; 308 u8 duplex = cmd->base.duplex; 309 u32 speed = cmd->base.speed; 310 u32 advertising; 311 312 if (cmd->base.phy_address != phydev->mdio.addr) 313 return -EINVAL; 314 315 ethtool_convert_link_mode_to_legacy_u32(&advertising, 316 cmd->link_modes.advertising); 317 318 /* We make sure that we don't pass unsupported values in to the PHY */ 319 advertising &= phydev->supported; 320 321 /* Verify the settings we care about. */ 322 if (autoneg != AUTONEG_ENABLE && autoneg != AUTONEG_DISABLE) 323 return -EINVAL; 324 325 if (autoneg == AUTONEG_ENABLE && advertising == 0) 326 return -EINVAL; 327 328 if (autoneg == AUTONEG_DISABLE && 329 ((speed != SPEED_1000 && 330 speed != SPEED_100 && 331 speed != SPEED_10) || 332 (duplex != DUPLEX_HALF && 333 duplex != DUPLEX_FULL))) 334 return -EINVAL; 335 336 phydev->autoneg = autoneg; 337 338 phydev->speed = speed; 339 340 phydev->advertising = advertising; 341 342 if (autoneg == AUTONEG_ENABLE) 343 phydev->advertising |= ADVERTISED_Autoneg; 344 else 345 phydev->advertising &= ~ADVERTISED_Autoneg; 346 347 phydev->duplex = duplex; 348 349 phydev->mdix_ctrl = cmd->base.eth_tp_mdix_ctrl; 350 351 /* Restart the PHY */ 352 phy_start_aneg(phydev); 353 354 return 0; 355 } 356 EXPORT_SYMBOL(phy_ethtool_ksettings_set); 357 358 void phy_ethtool_ksettings_get(struct phy_device *phydev, 359 struct ethtool_link_ksettings *cmd) 360 { 361 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, 362 phydev->supported); 363 364 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, 365 phydev->advertising); 366 367 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.lp_advertising, 368 phydev->lp_advertising); 369 370 cmd->base.speed = phydev->speed; 371 cmd->base.duplex = phydev->duplex; 372 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 373 cmd->base.port = PORT_BNC; 374 else 375 cmd->base.port = PORT_MII; 376 cmd->base.transceiver = phy_is_internal(phydev) ? 377 XCVR_INTERNAL : XCVR_EXTERNAL; 378 cmd->base.phy_address = phydev->mdio.addr; 379 cmd->base.autoneg = phydev->autoneg; 380 cmd->base.eth_tp_mdix_ctrl = phydev->mdix_ctrl; 381 cmd->base.eth_tp_mdix = phydev->mdix; 382 } 383 EXPORT_SYMBOL(phy_ethtool_ksettings_get); 384 385 /** 386 * phy_mii_ioctl - generic PHY MII ioctl interface 387 * @phydev: the phy_device struct 388 * @ifr: &struct ifreq for socket ioctl's 389 * @cmd: ioctl cmd to execute 390 * 391 * Note that this function is currently incompatible with the 392 * PHYCONTROL layer. It changes registers without regard to 393 * current state. Use at own risk. 394 */ 395 int phy_mii_ioctl(struct phy_device *phydev, struct ifreq *ifr, int cmd) 396 { 397 struct mii_ioctl_data *mii_data = if_mii(ifr); 398 u16 val = mii_data->val_in; 399 bool change_autoneg = false; 400 401 switch (cmd) { 402 case SIOCGMIIPHY: 403 mii_data->phy_id = phydev->mdio.addr; 404 /* fall through */ 405 406 case SIOCGMIIREG: 407 mii_data->val_out = mdiobus_read(phydev->mdio.bus, 408 mii_data->phy_id, 409 mii_data->reg_num); 410 return 0; 411 412 case SIOCSMIIREG: 413 if (mii_data->phy_id == phydev->mdio.addr) { 414 switch (mii_data->reg_num) { 415 case MII_BMCR: 416 if ((val & (BMCR_RESET | BMCR_ANENABLE)) == 0) { 417 if (phydev->autoneg == AUTONEG_ENABLE) 418 change_autoneg = true; 419 phydev->autoneg = AUTONEG_DISABLE; 420 if (val & BMCR_FULLDPLX) 421 phydev->duplex = DUPLEX_FULL; 422 else 423 phydev->duplex = DUPLEX_HALF; 424 if (val & BMCR_SPEED1000) 425 phydev->speed = SPEED_1000; 426 else if (val & BMCR_SPEED100) 427 phydev->speed = SPEED_100; 428 else phydev->speed = SPEED_10; 429 } 430 else { 431 if (phydev->autoneg == AUTONEG_DISABLE) 432 change_autoneg = true; 433 phydev->autoneg = AUTONEG_ENABLE; 434 } 435 break; 436 case MII_ADVERTISE: 437 phydev->advertising = mii_adv_to_ethtool_adv_t(val); 438 change_autoneg = true; 439 break; 440 default: 441 /* do nothing */ 442 break; 443 } 444 } 445 446 mdiobus_write(phydev->mdio.bus, mii_data->phy_id, 447 mii_data->reg_num, val); 448 449 if (mii_data->phy_id == phydev->mdio.addr && 450 mii_data->reg_num == MII_BMCR && 451 val & BMCR_RESET) 452 return phy_init_hw(phydev); 453 454 if (change_autoneg) 455 return phy_start_aneg(phydev); 456 457 return 0; 458 459 case SIOCSHWTSTAMP: 460 if (phydev->drv && phydev->drv->hwtstamp) 461 return phydev->drv->hwtstamp(phydev, ifr); 462 /* fall through */ 463 464 default: 465 return -EOPNOTSUPP; 466 } 467 } 468 EXPORT_SYMBOL(phy_mii_ioctl); 469 470 /** 471 * phy_start_aneg_priv - start auto-negotiation for this PHY device 472 * @phydev: the phy_device struct 473 * @sync: indicate whether we should wait for the workqueue cancelation 474 * 475 * Description: Sanitizes the settings (if we're not autonegotiating 476 * them), and then calls the driver's config_aneg function. 477 * If the PHYCONTROL Layer is operating, we change the state to 478 * reflect the beginning of Auto-negotiation or forcing. 479 */ 480 static int phy_start_aneg_priv(struct phy_device *phydev, bool sync) 481 { 482 bool trigger = 0; 483 int err; 484 485 if (!phydev->drv) 486 return -EIO; 487 488 mutex_lock(&phydev->lock); 489 490 if (AUTONEG_DISABLE == phydev->autoneg) 491 phy_sanitize_settings(phydev); 492 493 /* Invalidate LP advertising flags */ 494 phydev->lp_advertising = 0; 495 496 if (phydev->drv->config_aneg) 497 err = phydev->drv->config_aneg(phydev); 498 else 499 err = genphy_config_aneg(phydev); 500 if (err < 0) 501 goto out_unlock; 502 503 if (phydev->state != PHY_HALTED) { 504 if (AUTONEG_ENABLE == phydev->autoneg) { 505 phydev->state = PHY_AN; 506 phydev->link_timeout = PHY_AN_TIMEOUT; 507 } else { 508 phydev->state = PHY_FORCING; 509 phydev->link_timeout = PHY_FORCE_TIMEOUT; 510 } 511 } 512 513 /* Re-schedule a PHY state machine to check PHY status because 514 * negotiation may already be done and aneg interrupt may not be 515 * generated. 516 */ 517 if (phy_interrupt_is_valid(phydev) && (phydev->state == PHY_AN)) { 518 err = phy_aneg_done(phydev); 519 if (err > 0) { 520 trigger = true; 521 err = 0; 522 } 523 } 524 525 out_unlock: 526 mutex_unlock(&phydev->lock); 527 528 if (trigger) 529 phy_trigger_machine(phydev, sync); 530 531 return err; 532 } 533 534 /** 535 * phy_start_aneg - start auto-negotiation for this PHY device 536 * @phydev: the phy_device struct 537 * 538 * Description: Sanitizes the settings (if we're not autonegotiating 539 * them), and then calls the driver's config_aneg function. 540 * If the PHYCONTROL Layer is operating, we change the state to 541 * reflect the beginning of Auto-negotiation or forcing. 542 */ 543 int phy_start_aneg(struct phy_device *phydev) 544 { 545 return phy_start_aneg_priv(phydev, true); 546 } 547 EXPORT_SYMBOL(phy_start_aneg); 548 549 /** 550 * phy_start_machine - start PHY state machine tracking 551 * @phydev: the phy_device struct 552 * 553 * Description: The PHY infrastructure can run a state machine 554 * which tracks whether the PHY is starting up, negotiating, 555 * etc. This function starts the delayed workqueue which tracks 556 * the state of the PHY. If you want to maintain your own state machine, 557 * do not call this function. 558 */ 559 void phy_start_machine(struct phy_device *phydev) 560 { 561 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, HZ); 562 } 563 EXPORT_SYMBOL_GPL(phy_start_machine); 564 565 /** 566 * phy_trigger_machine - trigger the state machine to run 567 * 568 * @phydev: the phy_device struct 569 * @sync: indicate whether we should wait for the workqueue cancelation 570 * 571 * Description: There has been a change in state which requires that the 572 * state machine runs. 573 */ 574 575 void phy_trigger_machine(struct phy_device *phydev, bool sync) 576 { 577 if (sync) 578 cancel_delayed_work_sync(&phydev->state_queue); 579 else 580 cancel_delayed_work(&phydev->state_queue); 581 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 0); 582 } 583 584 /** 585 * phy_stop_machine - stop the PHY state machine tracking 586 * @phydev: target phy_device struct 587 * 588 * Description: Stops the state machine delayed workqueue, sets the 589 * state to UP (unless it wasn't up yet). This function must be 590 * called BEFORE phy_detach. 591 */ 592 void phy_stop_machine(struct phy_device *phydev) 593 { 594 cancel_delayed_work_sync(&phydev->state_queue); 595 596 mutex_lock(&phydev->lock); 597 if (phydev->state > PHY_UP && phydev->state != PHY_HALTED) 598 phydev->state = PHY_UP; 599 mutex_unlock(&phydev->lock); 600 } 601 602 /** 603 * phy_error - enter HALTED state for this PHY device 604 * @phydev: target phy_device struct 605 * 606 * Moves the PHY to the HALTED state in response to a read 607 * or write error, and tells the controller the link is down. 608 * Must not be called from interrupt context, or while the 609 * phydev->lock is held. 610 */ 611 static void phy_error(struct phy_device *phydev) 612 { 613 mutex_lock(&phydev->lock); 614 phydev->state = PHY_HALTED; 615 mutex_unlock(&phydev->lock); 616 617 phy_trigger_machine(phydev, false); 618 } 619 620 /** 621 * phy_interrupt - PHY interrupt handler 622 * @irq: interrupt line 623 * @phy_dat: phy_device pointer 624 * 625 * Description: When a PHY interrupt occurs, the handler disables 626 * interrupts, and uses phy_change to handle the interrupt. 627 */ 628 static irqreturn_t phy_interrupt(int irq, void *phy_dat) 629 { 630 struct phy_device *phydev = phy_dat; 631 632 if (PHY_HALTED == phydev->state) 633 return IRQ_NONE; /* It can't be ours. */ 634 635 phy_change(phydev); 636 637 return IRQ_HANDLED; 638 } 639 640 /** 641 * phy_enable_interrupts - Enable the interrupts from the PHY side 642 * @phydev: target phy_device struct 643 */ 644 static int phy_enable_interrupts(struct phy_device *phydev) 645 { 646 int err = phy_clear_interrupt(phydev); 647 648 if (err < 0) 649 return err; 650 651 return phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED); 652 } 653 654 /** 655 * phy_disable_interrupts - Disable the PHY interrupts from the PHY side 656 * @phydev: target phy_device struct 657 */ 658 static int phy_disable_interrupts(struct phy_device *phydev) 659 { 660 int err; 661 662 /* Disable PHY interrupts */ 663 err = phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 664 if (err) 665 goto phy_err; 666 667 /* Clear the interrupt */ 668 err = phy_clear_interrupt(phydev); 669 if (err) 670 goto phy_err; 671 672 return 0; 673 674 phy_err: 675 phy_error(phydev); 676 677 return err; 678 } 679 680 /** 681 * phy_start_interrupts - request and enable interrupts for a PHY device 682 * @phydev: target phy_device struct 683 * 684 * Description: Request the interrupt for the given PHY. 685 * If this fails, then we set irq to PHY_POLL. 686 * Otherwise, we enable the interrupts in the PHY. 687 * This should only be called with a valid IRQ number. 688 * Returns 0 on success or < 0 on error. 689 */ 690 int phy_start_interrupts(struct phy_device *phydev) 691 { 692 if (request_threaded_irq(phydev->irq, NULL, phy_interrupt, 693 IRQF_ONESHOT | IRQF_SHARED, 694 phydev_name(phydev), phydev) < 0) { 695 pr_warn("%s: Can't get IRQ %d (PHY)\n", 696 phydev->mdio.bus->name, phydev->irq); 697 phydev->irq = PHY_POLL; 698 return 0; 699 } 700 701 return phy_enable_interrupts(phydev); 702 } 703 EXPORT_SYMBOL(phy_start_interrupts); 704 705 /** 706 * phy_stop_interrupts - disable interrupts from a PHY device 707 * @phydev: target phy_device struct 708 */ 709 int phy_stop_interrupts(struct phy_device *phydev) 710 { 711 int err = phy_disable_interrupts(phydev); 712 713 if (err) 714 phy_error(phydev); 715 716 free_irq(phydev->irq, phydev); 717 718 return err; 719 } 720 EXPORT_SYMBOL(phy_stop_interrupts); 721 722 /** 723 * phy_change - Called by the phy_interrupt to handle PHY changes 724 * @phydev: phy_device struct that interrupted 725 */ 726 void phy_change(struct phy_device *phydev) 727 { 728 if (phy_interrupt_is_valid(phydev)) { 729 if (phydev->drv->did_interrupt && 730 !phydev->drv->did_interrupt(phydev)) 731 return; 732 733 if (phydev->state == PHY_HALTED) 734 if (phy_disable_interrupts(phydev)) 735 goto phy_err; 736 } 737 738 mutex_lock(&phydev->lock); 739 if ((PHY_RUNNING == phydev->state) || (PHY_NOLINK == phydev->state)) 740 phydev->state = PHY_CHANGELINK; 741 mutex_unlock(&phydev->lock); 742 743 /* reschedule state queue work to run as soon as possible */ 744 phy_trigger_machine(phydev, true); 745 746 if (phy_interrupt_is_valid(phydev) && phy_clear_interrupt(phydev)) 747 goto phy_err; 748 return; 749 750 phy_err: 751 phy_error(phydev); 752 } 753 754 /** 755 * phy_change_work - Scheduled by the phy_mac_interrupt to handle PHY changes 756 * @work: work_struct that describes the work to be done 757 */ 758 void phy_change_work(struct work_struct *work) 759 { 760 struct phy_device *phydev = 761 container_of(work, struct phy_device, phy_queue); 762 763 phy_change(phydev); 764 } 765 766 /** 767 * phy_stop - Bring down the PHY link, and stop checking the status 768 * @phydev: target phy_device struct 769 */ 770 void phy_stop(struct phy_device *phydev) 771 { 772 mutex_lock(&phydev->lock); 773 774 if (PHY_HALTED == phydev->state) 775 goto out_unlock; 776 777 if (phy_interrupt_is_valid(phydev)) { 778 /* Disable PHY Interrupts */ 779 phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 780 781 /* Clear any pending interrupts */ 782 phy_clear_interrupt(phydev); 783 } 784 785 phydev->state = PHY_HALTED; 786 787 out_unlock: 788 mutex_unlock(&phydev->lock); 789 790 /* Cannot call flush_scheduled_work() here as desired because 791 * of rtnl_lock(), but PHY_HALTED shall guarantee phy_change() 792 * will not reenable interrupts. 793 */ 794 } 795 EXPORT_SYMBOL(phy_stop); 796 797 /** 798 * phy_start - start or restart a PHY device 799 * @phydev: target phy_device struct 800 * 801 * Description: Indicates the attached device's readiness to 802 * handle PHY-related work. Used during startup to start the 803 * PHY, and after a call to phy_stop() to resume operation. 804 * Also used to indicate the MDIO bus has cleared an error 805 * condition. 806 */ 807 void phy_start(struct phy_device *phydev) 808 { 809 int err = 0; 810 811 mutex_lock(&phydev->lock); 812 813 switch (phydev->state) { 814 case PHY_STARTING: 815 phydev->state = PHY_PENDING; 816 break; 817 case PHY_READY: 818 phydev->state = PHY_UP; 819 break; 820 case PHY_HALTED: 821 /* if phy was suspended, bring the physical link up again */ 822 phy_resume(phydev); 823 824 /* make sure interrupts are re-enabled for the PHY */ 825 if (phy_interrupt_is_valid(phydev)) { 826 err = phy_enable_interrupts(phydev); 827 if (err < 0) 828 break; 829 } 830 831 phydev->state = PHY_RESUMING; 832 break; 833 default: 834 break; 835 } 836 mutex_unlock(&phydev->lock); 837 838 phy_trigger_machine(phydev, true); 839 } 840 EXPORT_SYMBOL(phy_start); 841 842 static void phy_link_up(struct phy_device *phydev) 843 { 844 phydev->phy_link_change(phydev, true, true); 845 phy_led_trigger_change_speed(phydev); 846 } 847 848 static void phy_link_down(struct phy_device *phydev, bool do_carrier) 849 { 850 phydev->phy_link_change(phydev, false, do_carrier); 851 phy_led_trigger_change_speed(phydev); 852 } 853 854 /** 855 * phy_state_machine - Handle the state machine 856 * @work: work_struct that describes the work to be done 857 */ 858 void phy_state_machine(struct work_struct *work) 859 { 860 struct delayed_work *dwork = to_delayed_work(work); 861 struct phy_device *phydev = 862 container_of(dwork, struct phy_device, state_queue); 863 bool needs_aneg = false, do_suspend = false; 864 enum phy_state old_state; 865 int err = 0; 866 int old_link; 867 868 mutex_lock(&phydev->lock); 869 870 old_state = phydev->state; 871 872 if (phydev->drv && phydev->drv->link_change_notify) 873 phydev->drv->link_change_notify(phydev); 874 875 switch (phydev->state) { 876 case PHY_DOWN: 877 case PHY_STARTING: 878 case PHY_READY: 879 case PHY_PENDING: 880 break; 881 case PHY_UP: 882 needs_aneg = true; 883 884 phydev->link_timeout = PHY_AN_TIMEOUT; 885 886 break; 887 case PHY_AN: 888 err = phy_read_status(phydev); 889 if (err < 0) 890 break; 891 892 /* If the link is down, give up on negotiation for now */ 893 if (!phydev->link) { 894 phydev->state = PHY_NOLINK; 895 phy_link_down(phydev, true); 896 break; 897 } 898 899 /* Check if negotiation is done. Break if there's an error */ 900 err = phy_aneg_done(phydev); 901 if (err < 0) 902 break; 903 904 /* If AN is done, we're running */ 905 if (err > 0) { 906 phydev->state = PHY_RUNNING; 907 phy_link_up(phydev); 908 } else if (0 == phydev->link_timeout--) 909 needs_aneg = true; 910 break; 911 case PHY_NOLINK: 912 if (phy_interrupt_is_valid(phydev)) 913 break; 914 915 err = phy_read_status(phydev); 916 if (err) 917 break; 918 919 if (phydev->link) { 920 if (AUTONEG_ENABLE == phydev->autoneg) { 921 err = phy_aneg_done(phydev); 922 if (err < 0) 923 break; 924 925 if (!err) { 926 phydev->state = PHY_AN; 927 phydev->link_timeout = PHY_AN_TIMEOUT; 928 break; 929 } 930 } 931 phydev->state = PHY_RUNNING; 932 phy_link_up(phydev); 933 } 934 break; 935 case PHY_FORCING: 936 err = genphy_update_link(phydev); 937 if (err) 938 break; 939 940 if (phydev->link) { 941 phydev->state = PHY_RUNNING; 942 phy_link_up(phydev); 943 } else { 944 if (0 == phydev->link_timeout--) 945 needs_aneg = true; 946 phy_link_down(phydev, false); 947 } 948 break; 949 case PHY_RUNNING: 950 /* Only register a CHANGE if we are polling and link changed 951 * since latest checking. 952 */ 953 if (phydev->irq == PHY_POLL) { 954 old_link = phydev->link; 955 err = phy_read_status(phydev); 956 if (err) 957 break; 958 959 if (old_link != phydev->link) 960 phydev->state = PHY_CHANGELINK; 961 } 962 /* 963 * Failsafe: check that nobody set phydev->link=0 between two 964 * poll cycles, otherwise we won't leave RUNNING state as long 965 * as link remains down. 966 */ 967 if (!phydev->link && phydev->state == PHY_RUNNING) { 968 phydev->state = PHY_CHANGELINK; 969 phydev_err(phydev, "no link in PHY_RUNNING\n"); 970 } 971 break; 972 case PHY_CHANGELINK: 973 err = phy_read_status(phydev); 974 if (err) 975 break; 976 977 if (phydev->link) { 978 phydev->state = PHY_RUNNING; 979 phy_link_up(phydev); 980 } else { 981 phydev->state = PHY_NOLINK; 982 phy_link_down(phydev, true); 983 } 984 break; 985 case PHY_HALTED: 986 if (phydev->link) { 987 phydev->link = 0; 988 phy_link_down(phydev, true); 989 do_suspend = true; 990 } 991 break; 992 case PHY_RESUMING: 993 if (AUTONEG_ENABLE == phydev->autoneg) { 994 err = phy_aneg_done(phydev); 995 if (err < 0) 996 break; 997 998 /* err > 0 if AN is done. 999 * Otherwise, it's 0, and we're still waiting for AN 1000 */ 1001 if (err > 0) { 1002 err = phy_read_status(phydev); 1003 if (err) 1004 break; 1005 1006 if (phydev->link) { 1007 phydev->state = PHY_RUNNING; 1008 phy_link_up(phydev); 1009 } else { 1010 phydev->state = PHY_NOLINK; 1011 phy_link_down(phydev, false); 1012 } 1013 } else { 1014 phydev->state = PHY_AN; 1015 phydev->link_timeout = PHY_AN_TIMEOUT; 1016 } 1017 } else { 1018 err = phy_read_status(phydev); 1019 if (err) 1020 break; 1021 1022 if (phydev->link) { 1023 phydev->state = PHY_RUNNING; 1024 phy_link_up(phydev); 1025 } else { 1026 phydev->state = PHY_NOLINK; 1027 phy_link_down(phydev, false); 1028 } 1029 } 1030 break; 1031 } 1032 1033 mutex_unlock(&phydev->lock); 1034 1035 if (needs_aneg) 1036 err = phy_start_aneg_priv(phydev, false); 1037 else if (do_suspend) 1038 phy_suspend(phydev); 1039 1040 if (err < 0) 1041 phy_error(phydev); 1042 1043 if (old_state != phydev->state) 1044 phydev_dbg(phydev, "PHY state change %s -> %s\n", 1045 phy_state_to_str(old_state), 1046 phy_state_to_str(phydev->state)); 1047 1048 /* Only re-schedule a PHY state machine change if we are polling the 1049 * PHY, if PHY_IGNORE_INTERRUPT is set, then we will be moving 1050 * between states from phy_mac_interrupt() 1051 */ 1052 if (phydev->irq == PHY_POLL) 1053 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 1054 PHY_STATE_TIME * HZ); 1055 } 1056 1057 /** 1058 * phy_mac_interrupt - MAC says the link has changed 1059 * @phydev: phy_device struct with changed link 1060 * 1061 * The MAC layer is able to indicate there has been a change in the PHY link 1062 * status. Trigger the state machine and work a work queue. 1063 */ 1064 void phy_mac_interrupt(struct phy_device *phydev) 1065 { 1066 /* Trigger a state machine change */ 1067 queue_work(system_power_efficient_wq, &phydev->phy_queue); 1068 } 1069 EXPORT_SYMBOL(phy_mac_interrupt); 1070 1071 /** 1072 * phy_init_eee - init and check the EEE feature 1073 * @phydev: target phy_device struct 1074 * @clk_stop_enable: PHY may stop the clock during LPI 1075 * 1076 * Description: it checks if the Energy-Efficient Ethernet (EEE) 1077 * is supported by looking at the MMD registers 3.20 and 7.60/61 1078 * and it programs the MMD register 3.0 setting the "Clock stop enable" 1079 * bit if required. 1080 */ 1081 int phy_init_eee(struct phy_device *phydev, bool clk_stop_enable) 1082 { 1083 if (!phydev->drv) 1084 return -EIO; 1085 1086 /* According to 802.3az,the EEE is supported only in full duplex-mode. 1087 */ 1088 if (phydev->duplex == DUPLEX_FULL) { 1089 int eee_lp, eee_cap, eee_adv; 1090 u32 lp, cap, adv; 1091 int status; 1092 1093 /* Read phy status to properly get the right settings */ 1094 status = phy_read_status(phydev); 1095 if (status) 1096 return status; 1097 1098 /* First check if the EEE ability is supported */ 1099 eee_cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1100 if (eee_cap <= 0) 1101 goto eee_exit_err; 1102 1103 cap = mmd_eee_cap_to_ethtool_sup_t(eee_cap); 1104 if (!cap) 1105 goto eee_exit_err; 1106 1107 /* Check which link settings negotiated and verify it in 1108 * the EEE advertising registers. 1109 */ 1110 eee_lp = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1111 if (eee_lp <= 0) 1112 goto eee_exit_err; 1113 1114 eee_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1115 if (eee_adv <= 0) 1116 goto eee_exit_err; 1117 1118 adv = mmd_eee_adv_to_ethtool_adv_t(eee_adv); 1119 lp = mmd_eee_adv_to_ethtool_adv_t(eee_lp); 1120 if (!phy_check_valid(phydev->speed, phydev->duplex, lp & adv)) 1121 goto eee_exit_err; 1122 1123 if (clk_stop_enable) { 1124 /* Configure the PHY to stop receiving xMII 1125 * clock while it is signaling LPI. 1126 */ 1127 int val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1); 1128 if (val < 0) 1129 return val; 1130 1131 val |= MDIO_PCS_CTRL1_CLKSTOP_EN; 1132 phy_write_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1, val); 1133 } 1134 1135 return 0; /* EEE supported */ 1136 } 1137 eee_exit_err: 1138 return -EPROTONOSUPPORT; 1139 } 1140 EXPORT_SYMBOL(phy_init_eee); 1141 1142 /** 1143 * phy_get_eee_err - report the EEE wake error count 1144 * @phydev: target phy_device struct 1145 * 1146 * Description: it is to report the number of time where the PHY 1147 * failed to complete its normal wake sequence. 1148 */ 1149 int phy_get_eee_err(struct phy_device *phydev) 1150 { 1151 if (!phydev->drv) 1152 return -EIO; 1153 1154 return phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_WK_ERR); 1155 } 1156 EXPORT_SYMBOL(phy_get_eee_err); 1157 1158 /** 1159 * phy_ethtool_get_eee - get EEE supported and status 1160 * @phydev: target phy_device struct 1161 * @data: ethtool_eee data 1162 * 1163 * Description: it reportes the Supported/Advertisement/LP Advertisement 1164 * capabilities. 1165 */ 1166 int phy_ethtool_get_eee(struct phy_device *phydev, struct ethtool_eee *data) 1167 { 1168 int val; 1169 1170 if (!phydev->drv) 1171 return -EIO; 1172 1173 /* Get Supported EEE */ 1174 val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1175 if (val < 0) 1176 return val; 1177 data->supported = mmd_eee_cap_to_ethtool_sup_t(val); 1178 1179 /* Get advertisement EEE */ 1180 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1181 if (val < 0) 1182 return val; 1183 data->advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1184 1185 /* Get LP advertisement EEE */ 1186 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1187 if (val < 0) 1188 return val; 1189 data->lp_advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1190 1191 return 0; 1192 } 1193 EXPORT_SYMBOL(phy_ethtool_get_eee); 1194 1195 /** 1196 * phy_ethtool_set_eee - set EEE supported and status 1197 * @phydev: target phy_device struct 1198 * @data: ethtool_eee data 1199 * 1200 * Description: it is to program the Advertisement EEE register. 1201 */ 1202 int phy_ethtool_set_eee(struct phy_device *phydev, struct ethtool_eee *data) 1203 { 1204 int cap, old_adv, adv, ret; 1205 1206 if (!phydev->drv) 1207 return -EIO; 1208 1209 /* Get Supported EEE */ 1210 cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1211 if (cap < 0) 1212 return cap; 1213 1214 old_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1215 if (old_adv < 0) 1216 return old_adv; 1217 1218 adv = ethtool_adv_to_mmd_eee_adv_t(data->advertised) & cap; 1219 1220 /* Mask prohibited EEE modes */ 1221 adv &= ~phydev->eee_broken_modes; 1222 1223 if (old_adv != adv) { 1224 ret = phy_write_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV, adv); 1225 if (ret < 0) 1226 return ret; 1227 1228 /* Restart autonegotiation so the new modes get sent to the 1229 * link partner. 1230 */ 1231 ret = phy_restart_aneg(phydev); 1232 if (ret < 0) 1233 return ret; 1234 } 1235 1236 return 0; 1237 } 1238 EXPORT_SYMBOL(phy_ethtool_set_eee); 1239 1240 int phy_ethtool_set_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1241 { 1242 if (phydev->drv && phydev->drv->set_wol) 1243 return phydev->drv->set_wol(phydev, wol); 1244 1245 return -EOPNOTSUPP; 1246 } 1247 EXPORT_SYMBOL(phy_ethtool_set_wol); 1248 1249 void phy_ethtool_get_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1250 { 1251 if (phydev->drv && phydev->drv->get_wol) 1252 phydev->drv->get_wol(phydev, wol); 1253 } 1254 EXPORT_SYMBOL(phy_ethtool_get_wol); 1255 1256 int phy_ethtool_get_link_ksettings(struct net_device *ndev, 1257 struct ethtool_link_ksettings *cmd) 1258 { 1259 struct phy_device *phydev = ndev->phydev; 1260 1261 if (!phydev) 1262 return -ENODEV; 1263 1264 phy_ethtool_ksettings_get(phydev, cmd); 1265 1266 return 0; 1267 } 1268 EXPORT_SYMBOL(phy_ethtool_get_link_ksettings); 1269 1270 int phy_ethtool_set_link_ksettings(struct net_device *ndev, 1271 const struct ethtool_link_ksettings *cmd) 1272 { 1273 struct phy_device *phydev = ndev->phydev; 1274 1275 if (!phydev) 1276 return -ENODEV; 1277 1278 return phy_ethtool_ksettings_set(phydev, cmd); 1279 } 1280 EXPORT_SYMBOL(phy_ethtool_set_link_ksettings); 1281 1282 int phy_ethtool_nway_reset(struct net_device *ndev) 1283 { 1284 struct phy_device *phydev = ndev->phydev; 1285 1286 if (!phydev) 1287 return -ENODEV; 1288 1289 if (!phydev->drv) 1290 return -EIO; 1291 1292 return phy_restart_aneg(phydev); 1293 } 1294 EXPORT_SYMBOL(phy_ethtool_nway_reset); 1295