1 /* Framework for configuring and reading PHY devices 2 * Based on code in sungem_phy.c and gianfar_phy.c 3 * 4 * Author: Andy Fleming 5 * 6 * Copyright (c) 2004 Freescale Semiconductor, Inc. 7 * Copyright (c) 2006, 2007 Maciej W. Rozycki 8 * 9 * This program is free software; you can redistribute it and/or modify it 10 * under the terms of the GNU General Public License as published by the 11 * Free Software Foundation; either version 2 of the License, or (at your 12 * option) any later version. 13 * 14 */ 15 16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 17 18 #include <linux/kernel.h> 19 #include <linux/string.h> 20 #include <linux/errno.h> 21 #include <linux/unistd.h> 22 #include <linux/interrupt.h> 23 #include <linux/delay.h> 24 #include <linux/netdevice.h> 25 #include <linux/etherdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/mm.h> 28 #include <linux/module.h> 29 #include <linux/mii.h> 30 #include <linux/ethtool.h> 31 #include <linux/phy.h> 32 #include <linux/phy_led_triggers.h> 33 #include <linux/workqueue.h> 34 #include <linux/mdio.h> 35 #include <linux/io.h> 36 #include <linux/uaccess.h> 37 #include <linux/atomic.h> 38 39 #include <asm/irq.h> 40 41 #define PHY_STATE_STR(_state) \ 42 case PHY_##_state: \ 43 return __stringify(_state); \ 44 45 static const char *phy_state_to_str(enum phy_state st) 46 { 47 switch (st) { 48 PHY_STATE_STR(DOWN) 49 PHY_STATE_STR(STARTING) 50 PHY_STATE_STR(READY) 51 PHY_STATE_STR(PENDING) 52 PHY_STATE_STR(UP) 53 PHY_STATE_STR(AN) 54 PHY_STATE_STR(RUNNING) 55 PHY_STATE_STR(NOLINK) 56 PHY_STATE_STR(FORCING) 57 PHY_STATE_STR(CHANGELINK) 58 PHY_STATE_STR(HALTED) 59 PHY_STATE_STR(RESUMING) 60 } 61 62 return NULL; 63 } 64 65 66 /** 67 * phy_print_status - Convenience function to print out the current phy status 68 * @phydev: the phy_device struct 69 */ 70 void phy_print_status(struct phy_device *phydev) 71 { 72 if (phydev->link) { 73 netdev_info(phydev->attached_dev, 74 "Link is Up - %s/%s - flow control %s\n", 75 phy_speed_to_str(phydev->speed), 76 phy_duplex_to_str(phydev->duplex), 77 phydev->pause ? "rx/tx" : "off"); 78 } else { 79 netdev_info(phydev->attached_dev, "Link is Down\n"); 80 } 81 } 82 EXPORT_SYMBOL(phy_print_status); 83 84 /** 85 * phy_clear_interrupt - Ack the phy device's interrupt 86 * @phydev: the phy_device struct 87 * 88 * If the @phydev driver has an ack_interrupt function, call it to 89 * ack and clear the phy device's interrupt. 90 * 91 * Returns 0 on success or < 0 on error. 92 */ 93 static int phy_clear_interrupt(struct phy_device *phydev) 94 { 95 if (phydev->drv->ack_interrupt) 96 return phydev->drv->ack_interrupt(phydev); 97 98 return 0; 99 } 100 101 /** 102 * phy_config_interrupt - configure the PHY device for the requested interrupts 103 * @phydev: the phy_device struct 104 * @interrupts: interrupt flags to configure for this @phydev 105 * 106 * Returns 0 on success or < 0 on error. 107 */ 108 static int phy_config_interrupt(struct phy_device *phydev, u32 interrupts) 109 { 110 phydev->interrupts = interrupts; 111 if (phydev->drv->config_intr) 112 return phydev->drv->config_intr(phydev); 113 114 return 0; 115 } 116 117 /** 118 * phy_restart_aneg - restart auto-negotiation 119 * @phydev: target phy_device struct 120 * 121 * Restart the autonegotiation on @phydev. Returns >= 0 on success or 122 * negative errno on error. 123 */ 124 int phy_restart_aneg(struct phy_device *phydev) 125 { 126 int ret; 127 128 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 129 ret = genphy_c45_restart_aneg(phydev); 130 else 131 ret = genphy_restart_aneg(phydev); 132 133 return ret; 134 } 135 EXPORT_SYMBOL_GPL(phy_restart_aneg); 136 137 /** 138 * phy_aneg_done - return auto-negotiation status 139 * @phydev: target phy_device struct 140 * 141 * Description: Return the auto-negotiation status from this @phydev 142 * Returns > 0 on success or < 0 on error. 0 means that auto-negotiation 143 * is still pending. 144 */ 145 int phy_aneg_done(struct phy_device *phydev) 146 { 147 if (phydev->drv && phydev->drv->aneg_done) 148 return phydev->drv->aneg_done(phydev); 149 150 /* Avoid genphy_aneg_done() if the Clause 45 PHY does not 151 * implement Clause 22 registers 152 */ 153 if (phydev->is_c45 && !(phydev->c45_ids.devices_in_package & BIT(0))) 154 return -EINVAL; 155 156 return genphy_aneg_done(phydev); 157 } 158 EXPORT_SYMBOL(phy_aneg_done); 159 160 /** 161 * phy_find_valid - find a PHY setting that matches the requested parameters 162 * @speed: desired speed 163 * @duplex: desired duplex 164 * @supported: mask of supported link modes 165 * 166 * Locate a supported phy setting that is, in priority order: 167 * - an exact match for the specified speed and duplex mode 168 * - a match for the specified speed, or slower speed 169 * - the slowest supported speed 170 * Returns the matched phy_setting entry, or %NULL if no supported phy 171 * settings were found. 172 */ 173 static const struct phy_setting * 174 phy_find_valid(int speed, int duplex, u32 supported) 175 { 176 unsigned long mask = supported; 177 178 return phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, false); 179 } 180 181 /** 182 * phy_supported_speeds - return all speeds currently supported by a phy device 183 * @phy: The phy device to return supported speeds of. 184 * @speeds: buffer to store supported speeds in. 185 * @size: size of speeds buffer. 186 * 187 * Description: Returns the number of supported speeds, and fills the speeds 188 * buffer with the supported speeds. If speeds buffer is too small to contain 189 * all currently supported speeds, will return as many speeds as can fit. 190 */ 191 unsigned int phy_supported_speeds(struct phy_device *phy, 192 unsigned int *speeds, 193 unsigned int size) 194 { 195 unsigned long supported = phy->supported; 196 197 return phy_speeds(speeds, size, &supported, BITS_PER_LONG); 198 } 199 200 /** 201 * phy_check_valid - check if there is a valid PHY setting which matches 202 * speed, duplex, and feature mask 203 * @speed: speed to match 204 * @duplex: duplex to match 205 * @features: A mask of the valid settings 206 * 207 * Description: Returns true if there is a valid setting, false otherwise. 208 */ 209 static inline bool phy_check_valid(int speed, int duplex, u32 features) 210 { 211 unsigned long mask = features; 212 213 return !!phy_lookup_setting(speed, duplex, &mask, BITS_PER_LONG, true); 214 } 215 216 /** 217 * phy_sanitize_settings - make sure the PHY is set to supported speed and duplex 218 * @phydev: the target phy_device struct 219 * 220 * Description: Make sure the PHY is set to supported speeds and 221 * duplexes. Drop down by one in this order: 1000/FULL, 222 * 1000/HALF, 100/FULL, 100/HALF, 10/FULL, 10/HALF. 223 */ 224 static void phy_sanitize_settings(struct phy_device *phydev) 225 { 226 const struct phy_setting *setting; 227 u32 features = phydev->supported; 228 229 /* Sanitize settings based on PHY capabilities */ 230 if ((features & SUPPORTED_Autoneg) == 0) 231 phydev->autoneg = AUTONEG_DISABLE; 232 233 setting = phy_find_valid(phydev->speed, phydev->duplex, features); 234 if (setting) { 235 phydev->speed = setting->speed; 236 phydev->duplex = setting->duplex; 237 } else { 238 /* We failed to find anything (no supported speeds?) */ 239 phydev->speed = SPEED_UNKNOWN; 240 phydev->duplex = DUPLEX_UNKNOWN; 241 } 242 } 243 244 /** 245 * phy_ethtool_sset - generic ethtool sset function, handles all the details 246 * @phydev: target phy_device struct 247 * @cmd: ethtool_cmd 248 * 249 * A few notes about parameter checking: 250 * 251 * - We don't set port or transceiver, so we don't care what they 252 * were set to. 253 * - phy_start_aneg() will make sure forced settings are sane, and 254 * choose the next best ones from the ones selected, so we don't 255 * care if ethtool tries to give us bad values. 256 */ 257 int phy_ethtool_sset(struct phy_device *phydev, struct ethtool_cmd *cmd) 258 { 259 u32 speed = ethtool_cmd_speed(cmd); 260 261 if (cmd->phy_address != phydev->mdio.addr) 262 return -EINVAL; 263 264 /* We make sure that we don't pass unsupported values in to the PHY */ 265 cmd->advertising &= phydev->supported; 266 267 /* Verify the settings we care about. */ 268 if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE) 269 return -EINVAL; 270 271 if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0) 272 return -EINVAL; 273 274 if (cmd->autoneg == AUTONEG_DISABLE && 275 ((speed != SPEED_1000 && 276 speed != SPEED_100 && 277 speed != SPEED_10) || 278 (cmd->duplex != DUPLEX_HALF && 279 cmd->duplex != DUPLEX_FULL))) 280 return -EINVAL; 281 282 phydev->autoneg = cmd->autoneg; 283 284 phydev->speed = speed; 285 286 phydev->advertising = cmd->advertising; 287 288 if (AUTONEG_ENABLE == cmd->autoneg) 289 phydev->advertising |= ADVERTISED_Autoneg; 290 else 291 phydev->advertising &= ~ADVERTISED_Autoneg; 292 293 phydev->duplex = cmd->duplex; 294 295 phydev->mdix_ctrl = cmd->eth_tp_mdix_ctrl; 296 297 /* Restart the PHY */ 298 phy_start_aneg(phydev); 299 300 return 0; 301 } 302 EXPORT_SYMBOL(phy_ethtool_sset); 303 304 int phy_ethtool_ksettings_set(struct phy_device *phydev, 305 const struct ethtool_link_ksettings *cmd) 306 { 307 u8 autoneg = cmd->base.autoneg; 308 u8 duplex = cmd->base.duplex; 309 u32 speed = cmd->base.speed; 310 u32 advertising; 311 312 if (cmd->base.phy_address != phydev->mdio.addr) 313 return -EINVAL; 314 315 ethtool_convert_link_mode_to_legacy_u32(&advertising, 316 cmd->link_modes.advertising); 317 318 /* We make sure that we don't pass unsupported values in to the PHY */ 319 advertising &= phydev->supported; 320 321 /* Verify the settings we care about. */ 322 if (autoneg != AUTONEG_ENABLE && autoneg != AUTONEG_DISABLE) 323 return -EINVAL; 324 325 if (autoneg == AUTONEG_ENABLE && advertising == 0) 326 return -EINVAL; 327 328 if (autoneg == AUTONEG_DISABLE && 329 ((speed != SPEED_1000 && 330 speed != SPEED_100 && 331 speed != SPEED_10) || 332 (duplex != DUPLEX_HALF && 333 duplex != DUPLEX_FULL))) 334 return -EINVAL; 335 336 phydev->autoneg = autoneg; 337 338 phydev->speed = speed; 339 340 phydev->advertising = advertising; 341 342 if (autoneg == AUTONEG_ENABLE) 343 phydev->advertising |= ADVERTISED_Autoneg; 344 else 345 phydev->advertising &= ~ADVERTISED_Autoneg; 346 347 phydev->duplex = duplex; 348 349 phydev->mdix_ctrl = cmd->base.eth_tp_mdix_ctrl; 350 351 /* Restart the PHY */ 352 phy_start_aneg(phydev); 353 354 return 0; 355 } 356 EXPORT_SYMBOL(phy_ethtool_ksettings_set); 357 358 void phy_ethtool_ksettings_get(struct phy_device *phydev, 359 struct ethtool_link_ksettings *cmd) 360 { 361 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, 362 phydev->supported); 363 364 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, 365 phydev->advertising); 366 367 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.lp_advertising, 368 phydev->lp_advertising); 369 370 cmd->base.speed = phydev->speed; 371 cmd->base.duplex = phydev->duplex; 372 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 373 cmd->base.port = PORT_BNC; 374 else 375 cmd->base.port = PORT_MII; 376 cmd->base.transceiver = phy_is_internal(phydev) ? 377 XCVR_INTERNAL : XCVR_EXTERNAL; 378 cmd->base.phy_address = phydev->mdio.addr; 379 cmd->base.autoneg = phydev->autoneg; 380 cmd->base.eth_tp_mdix_ctrl = phydev->mdix_ctrl; 381 cmd->base.eth_tp_mdix = phydev->mdix; 382 } 383 EXPORT_SYMBOL(phy_ethtool_ksettings_get); 384 385 /** 386 * phy_mii_ioctl - generic PHY MII ioctl interface 387 * @phydev: the phy_device struct 388 * @ifr: &struct ifreq for socket ioctl's 389 * @cmd: ioctl cmd to execute 390 * 391 * Note that this function is currently incompatible with the 392 * PHYCONTROL layer. It changes registers without regard to 393 * current state. Use at own risk. 394 */ 395 int phy_mii_ioctl(struct phy_device *phydev, struct ifreq *ifr, int cmd) 396 { 397 struct mii_ioctl_data *mii_data = if_mii(ifr); 398 u16 val = mii_data->val_in; 399 bool change_autoneg = false; 400 401 switch (cmd) { 402 case SIOCGMIIPHY: 403 mii_data->phy_id = phydev->mdio.addr; 404 /* fall through */ 405 406 case SIOCGMIIREG: 407 mii_data->val_out = mdiobus_read(phydev->mdio.bus, 408 mii_data->phy_id, 409 mii_data->reg_num); 410 return 0; 411 412 case SIOCSMIIREG: 413 if (mii_data->phy_id == phydev->mdio.addr) { 414 switch (mii_data->reg_num) { 415 case MII_BMCR: 416 if ((val & (BMCR_RESET | BMCR_ANENABLE)) == 0) { 417 if (phydev->autoneg == AUTONEG_ENABLE) 418 change_autoneg = true; 419 phydev->autoneg = AUTONEG_DISABLE; 420 if (val & BMCR_FULLDPLX) 421 phydev->duplex = DUPLEX_FULL; 422 else 423 phydev->duplex = DUPLEX_HALF; 424 if (val & BMCR_SPEED1000) 425 phydev->speed = SPEED_1000; 426 else if (val & BMCR_SPEED100) 427 phydev->speed = SPEED_100; 428 else phydev->speed = SPEED_10; 429 } 430 else { 431 if (phydev->autoneg == AUTONEG_DISABLE) 432 change_autoneg = true; 433 phydev->autoneg = AUTONEG_ENABLE; 434 } 435 break; 436 case MII_ADVERTISE: 437 phydev->advertising = mii_adv_to_ethtool_adv_t(val); 438 change_autoneg = true; 439 break; 440 default: 441 /* do nothing */ 442 break; 443 } 444 } 445 446 mdiobus_write(phydev->mdio.bus, mii_data->phy_id, 447 mii_data->reg_num, val); 448 449 if (mii_data->phy_id == phydev->mdio.addr && 450 mii_data->reg_num == MII_BMCR && 451 val & BMCR_RESET) 452 return phy_init_hw(phydev); 453 454 if (change_autoneg) 455 return phy_start_aneg(phydev); 456 457 return 0; 458 459 case SIOCSHWTSTAMP: 460 if (phydev->drv && phydev->drv->hwtstamp) 461 return phydev->drv->hwtstamp(phydev, ifr); 462 /* fall through */ 463 464 default: 465 return -EOPNOTSUPP; 466 } 467 } 468 EXPORT_SYMBOL(phy_mii_ioctl); 469 470 /** 471 * phy_start_aneg_priv - start auto-negotiation for this PHY device 472 * @phydev: the phy_device struct 473 * @sync: indicate whether we should wait for the workqueue cancelation 474 * 475 * Description: Sanitizes the settings (if we're not autonegotiating 476 * them), and then calls the driver's config_aneg function. 477 * If the PHYCONTROL Layer is operating, we change the state to 478 * reflect the beginning of Auto-negotiation or forcing. 479 */ 480 static int phy_start_aneg_priv(struct phy_device *phydev, bool sync) 481 { 482 bool trigger = 0; 483 int err; 484 485 if (!phydev->drv) 486 return -EIO; 487 488 mutex_lock(&phydev->lock); 489 490 if (AUTONEG_DISABLE == phydev->autoneg) 491 phy_sanitize_settings(phydev); 492 493 /* Invalidate LP advertising flags */ 494 phydev->lp_advertising = 0; 495 496 err = phydev->drv->config_aneg(phydev); 497 if (err < 0) 498 goto out_unlock; 499 500 if (phydev->state != PHY_HALTED) { 501 if (AUTONEG_ENABLE == phydev->autoneg) { 502 phydev->state = PHY_AN; 503 phydev->link_timeout = PHY_AN_TIMEOUT; 504 } else { 505 phydev->state = PHY_FORCING; 506 phydev->link_timeout = PHY_FORCE_TIMEOUT; 507 } 508 } 509 510 /* Re-schedule a PHY state machine to check PHY status because 511 * negotiation may already be done and aneg interrupt may not be 512 * generated. 513 */ 514 if (phy_interrupt_is_valid(phydev) && (phydev->state == PHY_AN)) { 515 err = phy_aneg_done(phydev); 516 if (err > 0) { 517 trigger = true; 518 err = 0; 519 } 520 } 521 522 out_unlock: 523 mutex_unlock(&phydev->lock); 524 525 if (trigger) 526 phy_trigger_machine(phydev, sync); 527 528 return err; 529 } 530 531 /** 532 * phy_start_aneg - start auto-negotiation for this PHY device 533 * @phydev: the phy_device struct 534 * 535 * Description: Sanitizes the settings (if we're not autonegotiating 536 * them), and then calls the driver's config_aneg function. 537 * If the PHYCONTROL Layer is operating, we change the state to 538 * reflect the beginning of Auto-negotiation or forcing. 539 */ 540 int phy_start_aneg(struct phy_device *phydev) 541 { 542 return phy_start_aneg_priv(phydev, true); 543 } 544 EXPORT_SYMBOL(phy_start_aneg); 545 546 /** 547 * phy_start_machine - start PHY state machine tracking 548 * @phydev: the phy_device struct 549 * 550 * Description: The PHY infrastructure can run a state machine 551 * which tracks whether the PHY is starting up, negotiating, 552 * etc. This function starts the delayed workqueue which tracks 553 * the state of the PHY. If you want to maintain your own state machine, 554 * do not call this function. 555 */ 556 void phy_start_machine(struct phy_device *phydev) 557 { 558 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, HZ); 559 } 560 EXPORT_SYMBOL_GPL(phy_start_machine); 561 562 /** 563 * phy_trigger_machine - trigger the state machine to run 564 * 565 * @phydev: the phy_device struct 566 * @sync: indicate whether we should wait for the workqueue cancelation 567 * 568 * Description: There has been a change in state which requires that the 569 * state machine runs. 570 */ 571 572 void phy_trigger_machine(struct phy_device *phydev, bool sync) 573 { 574 if (sync) 575 cancel_delayed_work_sync(&phydev->state_queue); 576 else 577 cancel_delayed_work(&phydev->state_queue); 578 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 0); 579 } 580 581 /** 582 * phy_stop_machine - stop the PHY state machine tracking 583 * @phydev: target phy_device struct 584 * 585 * Description: Stops the state machine delayed workqueue, sets the 586 * state to UP (unless it wasn't up yet). This function must be 587 * called BEFORE phy_detach. 588 */ 589 void phy_stop_machine(struct phy_device *phydev) 590 { 591 cancel_delayed_work_sync(&phydev->state_queue); 592 593 mutex_lock(&phydev->lock); 594 if (phydev->state > PHY_UP && phydev->state != PHY_HALTED) 595 phydev->state = PHY_UP; 596 mutex_unlock(&phydev->lock); 597 } 598 599 /** 600 * phy_error - enter HALTED state for this PHY device 601 * @phydev: target phy_device struct 602 * 603 * Moves the PHY to the HALTED state in response to a read 604 * or write error, and tells the controller the link is down. 605 * Must not be called from interrupt context, or while the 606 * phydev->lock is held. 607 */ 608 static void phy_error(struct phy_device *phydev) 609 { 610 mutex_lock(&phydev->lock); 611 phydev->state = PHY_HALTED; 612 mutex_unlock(&phydev->lock); 613 614 phy_trigger_machine(phydev, false); 615 } 616 617 /** 618 * phy_interrupt - PHY interrupt handler 619 * @irq: interrupt line 620 * @phy_dat: phy_device pointer 621 * 622 * Description: When a PHY interrupt occurs, the handler disables 623 * interrupts, and uses phy_change to handle the interrupt. 624 */ 625 static irqreturn_t phy_interrupt(int irq, void *phy_dat) 626 { 627 struct phy_device *phydev = phy_dat; 628 629 if (PHY_HALTED == phydev->state) 630 return IRQ_NONE; /* It can't be ours. */ 631 632 disable_irq_nosync(irq); 633 atomic_inc(&phydev->irq_disable); 634 635 phy_change(phydev); 636 637 return IRQ_HANDLED; 638 } 639 640 /** 641 * phy_enable_interrupts - Enable the interrupts from the PHY side 642 * @phydev: target phy_device struct 643 */ 644 static int phy_enable_interrupts(struct phy_device *phydev) 645 { 646 int err = phy_clear_interrupt(phydev); 647 648 if (err < 0) 649 return err; 650 651 return phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED); 652 } 653 654 /** 655 * phy_disable_interrupts - Disable the PHY interrupts from the PHY side 656 * @phydev: target phy_device struct 657 */ 658 static int phy_disable_interrupts(struct phy_device *phydev) 659 { 660 int err; 661 662 /* Disable PHY interrupts */ 663 err = phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 664 if (err) 665 goto phy_err; 666 667 /* Clear the interrupt */ 668 err = phy_clear_interrupt(phydev); 669 if (err) 670 goto phy_err; 671 672 return 0; 673 674 phy_err: 675 phy_error(phydev); 676 677 return err; 678 } 679 680 /** 681 * phy_start_interrupts - request and enable interrupts for a PHY device 682 * @phydev: target phy_device struct 683 * 684 * Description: Request the interrupt for the given PHY. 685 * If this fails, then we set irq to PHY_POLL. 686 * Otherwise, we enable the interrupts in the PHY. 687 * This should only be called with a valid IRQ number. 688 * Returns 0 on success or < 0 on error. 689 */ 690 int phy_start_interrupts(struct phy_device *phydev) 691 { 692 atomic_set(&phydev->irq_disable, 0); 693 if (request_threaded_irq(phydev->irq, NULL, phy_interrupt, 694 IRQF_ONESHOT | IRQF_SHARED, 695 phydev_name(phydev), phydev) < 0) { 696 pr_warn("%s: Can't get IRQ %d (PHY)\n", 697 phydev->mdio.bus->name, phydev->irq); 698 phydev->irq = PHY_POLL; 699 return 0; 700 } 701 702 return phy_enable_interrupts(phydev); 703 } 704 EXPORT_SYMBOL(phy_start_interrupts); 705 706 /** 707 * phy_stop_interrupts - disable interrupts from a PHY device 708 * @phydev: target phy_device struct 709 */ 710 int phy_stop_interrupts(struct phy_device *phydev) 711 { 712 int err = phy_disable_interrupts(phydev); 713 714 if (err) 715 phy_error(phydev); 716 717 free_irq(phydev->irq, phydev); 718 719 /* If work indeed has been cancelled, disable_irq() will have 720 * been left unbalanced from phy_interrupt() and enable_irq() 721 * has to be called so that other devices on the line work. 722 */ 723 while (atomic_dec_return(&phydev->irq_disable) >= 0) 724 enable_irq(phydev->irq); 725 726 return err; 727 } 728 EXPORT_SYMBOL(phy_stop_interrupts); 729 730 /** 731 * phy_change - Called by the phy_interrupt to handle PHY changes 732 * @phydev: phy_device struct that interrupted 733 */ 734 void phy_change(struct phy_device *phydev) 735 { 736 if (phy_interrupt_is_valid(phydev)) { 737 if (phydev->drv->did_interrupt && 738 !phydev->drv->did_interrupt(phydev)) 739 goto ignore; 740 741 if (phy_disable_interrupts(phydev)) 742 goto phy_err; 743 } 744 745 mutex_lock(&phydev->lock); 746 if ((PHY_RUNNING == phydev->state) || (PHY_NOLINK == phydev->state)) 747 phydev->state = PHY_CHANGELINK; 748 mutex_unlock(&phydev->lock); 749 750 if (phy_interrupt_is_valid(phydev)) { 751 atomic_dec(&phydev->irq_disable); 752 enable_irq(phydev->irq); 753 754 /* Reenable interrupts */ 755 if (PHY_HALTED != phydev->state && 756 phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED)) 757 goto irq_enable_err; 758 } 759 760 /* reschedule state queue work to run as soon as possible */ 761 phy_trigger_machine(phydev, true); 762 return; 763 764 ignore: 765 atomic_dec(&phydev->irq_disable); 766 enable_irq(phydev->irq); 767 return; 768 769 irq_enable_err: 770 disable_irq(phydev->irq); 771 atomic_inc(&phydev->irq_disable); 772 phy_err: 773 phy_error(phydev); 774 } 775 776 /** 777 * phy_change_work - Scheduled by the phy_mac_interrupt to handle PHY changes 778 * @work: work_struct that describes the work to be done 779 */ 780 void phy_change_work(struct work_struct *work) 781 { 782 struct phy_device *phydev = 783 container_of(work, struct phy_device, phy_queue); 784 785 phy_change(phydev); 786 } 787 788 /** 789 * phy_stop - Bring down the PHY link, and stop checking the status 790 * @phydev: target phy_device struct 791 */ 792 void phy_stop(struct phy_device *phydev) 793 { 794 mutex_lock(&phydev->lock); 795 796 if (PHY_HALTED == phydev->state) 797 goto out_unlock; 798 799 if (phy_interrupt_is_valid(phydev)) { 800 /* Disable PHY Interrupts */ 801 phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 802 803 /* Clear any pending interrupts */ 804 phy_clear_interrupt(phydev); 805 } 806 807 phydev->state = PHY_HALTED; 808 809 out_unlock: 810 mutex_unlock(&phydev->lock); 811 812 /* Cannot call flush_scheduled_work() here as desired because 813 * of rtnl_lock(), but PHY_HALTED shall guarantee phy_change() 814 * will not reenable interrupts. 815 */ 816 } 817 EXPORT_SYMBOL(phy_stop); 818 819 /** 820 * phy_start - start or restart a PHY device 821 * @phydev: target phy_device struct 822 * 823 * Description: Indicates the attached device's readiness to 824 * handle PHY-related work. Used during startup to start the 825 * PHY, and after a call to phy_stop() to resume operation. 826 * Also used to indicate the MDIO bus has cleared an error 827 * condition. 828 */ 829 void phy_start(struct phy_device *phydev) 830 { 831 int err = 0; 832 833 mutex_lock(&phydev->lock); 834 835 switch (phydev->state) { 836 case PHY_STARTING: 837 phydev->state = PHY_PENDING; 838 break; 839 case PHY_READY: 840 phydev->state = PHY_UP; 841 break; 842 case PHY_HALTED: 843 /* if phy was suspended, bring the physical link up again */ 844 phy_resume(phydev); 845 846 /* make sure interrupts are re-enabled for the PHY */ 847 if (phydev->irq != PHY_POLL) { 848 err = phy_enable_interrupts(phydev); 849 if (err < 0) 850 break; 851 } 852 853 phydev->state = PHY_RESUMING; 854 break; 855 default: 856 break; 857 } 858 mutex_unlock(&phydev->lock); 859 860 phy_trigger_machine(phydev, true); 861 } 862 EXPORT_SYMBOL(phy_start); 863 864 static void phy_link_up(struct phy_device *phydev) 865 { 866 phydev->phy_link_change(phydev, true, true); 867 phy_led_trigger_change_speed(phydev); 868 } 869 870 static void phy_link_down(struct phy_device *phydev, bool do_carrier) 871 { 872 phydev->phy_link_change(phydev, false, do_carrier); 873 phy_led_trigger_change_speed(phydev); 874 } 875 876 /** 877 * phy_state_machine - Handle the state machine 878 * @work: work_struct that describes the work to be done 879 */ 880 void phy_state_machine(struct work_struct *work) 881 { 882 struct delayed_work *dwork = to_delayed_work(work); 883 struct phy_device *phydev = 884 container_of(dwork, struct phy_device, state_queue); 885 bool needs_aneg = false, do_suspend = false; 886 enum phy_state old_state; 887 int err = 0; 888 int old_link; 889 890 mutex_lock(&phydev->lock); 891 892 old_state = phydev->state; 893 894 if (phydev->drv && phydev->drv->link_change_notify) 895 phydev->drv->link_change_notify(phydev); 896 897 switch (phydev->state) { 898 case PHY_DOWN: 899 case PHY_STARTING: 900 case PHY_READY: 901 case PHY_PENDING: 902 break; 903 case PHY_UP: 904 needs_aneg = true; 905 906 phydev->link_timeout = PHY_AN_TIMEOUT; 907 908 break; 909 case PHY_AN: 910 err = phy_read_status(phydev); 911 if (err < 0) 912 break; 913 914 /* If the link is down, give up on negotiation for now */ 915 if (!phydev->link) { 916 phydev->state = PHY_NOLINK; 917 phy_link_down(phydev, true); 918 break; 919 } 920 921 /* Check if negotiation is done. Break if there's an error */ 922 err = phy_aneg_done(phydev); 923 if (err < 0) 924 break; 925 926 /* If AN is done, we're running */ 927 if (err > 0) { 928 phydev->state = PHY_RUNNING; 929 phy_link_up(phydev); 930 } else if (0 == phydev->link_timeout--) 931 needs_aneg = true; 932 break; 933 case PHY_NOLINK: 934 if (phy_interrupt_is_valid(phydev)) 935 break; 936 937 err = phy_read_status(phydev); 938 if (err) 939 break; 940 941 if (phydev->link) { 942 if (AUTONEG_ENABLE == phydev->autoneg) { 943 err = phy_aneg_done(phydev); 944 if (err < 0) 945 break; 946 947 if (!err) { 948 phydev->state = PHY_AN; 949 phydev->link_timeout = PHY_AN_TIMEOUT; 950 break; 951 } 952 } 953 phydev->state = PHY_RUNNING; 954 phy_link_up(phydev); 955 } 956 break; 957 case PHY_FORCING: 958 err = genphy_update_link(phydev); 959 if (err) 960 break; 961 962 if (phydev->link) { 963 phydev->state = PHY_RUNNING; 964 phy_link_up(phydev); 965 } else { 966 if (0 == phydev->link_timeout--) 967 needs_aneg = true; 968 phy_link_down(phydev, false); 969 } 970 break; 971 case PHY_RUNNING: 972 /* Only register a CHANGE if we are polling and link changed 973 * since latest checking. 974 */ 975 if (phydev->irq == PHY_POLL) { 976 old_link = phydev->link; 977 err = phy_read_status(phydev); 978 if (err) 979 break; 980 981 if (old_link != phydev->link) 982 phydev->state = PHY_CHANGELINK; 983 } 984 /* 985 * Failsafe: check that nobody set phydev->link=0 between two 986 * poll cycles, otherwise we won't leave RUNNING state as long 987 * as link remains down. 988 */ 989 if (!phydev->link && phydev->state == PHY_RUNNING) { 990 phydev->state = PHY_CHANGELINK; 991 phydev_err(phydev, "no link in PHY_RUNNING\n"); 992 } 993 break; 994 case PHY_CHANGELINK: 995 err = phy_read_status(phydev); 996 if (err) 997 break; 998 999 if (phydev->link) { 1000 phydev->state = PHY_RUNNING; 1001 phy_link_up(phydev); 1002 } else { 1003 phydev->state = PHY_NOLINK; 1004 phy_link_down(phydev, true); 1005 } 1006 1007 if (phy_interrupt_is_valid(phydev)) 1008 err = phy_config_interrupt(phydev, 1009 PHY_INTERRUPT_ENABLED); 1010 break; 1011 case PHY_HALTED: 1012 if (phydev->link) { 1013 phydev->link = 0; 1014 phy_link_down(phydev, true); 1015 do_suspend = true; 1016 } 1017 break; 1018 case PHY_RESUMING: 1019 if (AUTONEG_ENABLE == phydev->autoneg) { 1020 err = phy_aneg_done(phydev); 1021 if (err < 0) 1022 break; 1023 1024 /* err > 0 if AN is done. 1025 * Otherwise, it's 0, and we're still waiting for AN 1026 */ 1027 if (err > 0) { 1028 err = phy_read_status(phydev); 1029 if (err) 1030 break; 1031 1032 if (phydev->link) { 1033 phydev->state = PHY_RUNNING; 1034 phy_link_up(phydev); 1035 } else { 1036 phydev->state = PHY_NOLINK; 1037 phy_link_down(phydev, false); 1038 } 1039 } else { 1040 phydev->state = PHY_AN; 1041 phydev->link_timeout = PHY_AN_TIMEOUT; 1042 } 1043 } else { 1044 err = phy_read_status(phydev); 1045 if (err) 1046 break; 1047 1048 if (phydev->link) { 1049 phydev->state = PHY_RUNNING; 1050 phy_link_up(phydev); 1051 } else { 1052 phydev->state = PHY_NOLINK; 1053 phy_link_down(phydev, false); 1054 } 1055 } 1056 break; 1057 } 1058 1059 mutex_unlock(&phydev->lock); 1060 1061 if (needs_aneg) 1062 err = phy_start_aneg_priv(phydev, false); 1063 else if (do_suspend) 1064 phy_suspend(phydev); 1065 1066 if (err < 0) 1067 phy_error(phydev); 1068 1069 if (old_state != phydev->state) 1070 phydev_dbg(phydev, "PHY state change %s -> %s\n", 1071 phy_state_to_str(old_state), 1072 phy_state_to_str(phydev->state)); 1073 1074 /* Only re-schedule a PHY state machine change if we are polling the 1075 * PHY, if PHY_IGNORE_INTERRUPT is set, then we will be moving 1076 * between states from phy_mac_interrupt() 1077 */ 1078 if (phydev->irq == PHY_POLL) 1079 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 1080 PHY_STATE_TIME * HZ); 1081 } 1082 1083 /** 1084 * phy_mac_interrupt - MAC says the link has changed 1085 * @phydev: phy_device struct with changed link 1086 * @new_link: Link is Up/Down. 1087 * 1088 * Description: The MAC layer is able indicate there has been a change 1089 * in the PHY link status. Set the new link status, and trigger the 1090 * state machine, work a work queue. 1091 */ 1092 void phy_mac_interrupt(struct phy_device *phydev, int new_link) 1093 { 1094 phydev->link = new_link; 1095 1096 /* Trigger a state machine change */ 1097 queue_work(system_power_efficient_wq, &phydev->phy_queue); 1098 } 1099 EXPORT_SYMBOL(phy_mac_interrupt); 1100 1101 /** 1102 * phy_init_eee - init and check the EEE feature 1103 * @phydev: target phy_device struct 1104 * @clk_stop_enable: PHY may stop the clock during LPI 1105 * 1106 * Description: it checks if the Energy-Efficient Ethernet (EEE) 1107 * is supported by looking at the MMD registers 3.20 and 7.60/61 1108 * and it programs the MMD register 3.0 setting the "Clock stop enable" 1109 * bit if required. 1110 */ 1111 int phy_init_eee(struct phy_device *phydev, bool clk_stop_enable) 1112 { 1113 if (!phydev->drv) 1114 return -EIO; 1115 1116 /* According to 802.3az,the EEE is supported only in full duplex-mode. 1117 */ 1118 if (phydev->duplex == DUPLEX_FULL) { 1119 int eee_lp, eee_cap, eee_adv; 1120 u32 lp, cap, adv; 1121 int status; 1122 1123 /* Read phy status to properly get the right settings */ 1124 status = phy_read_status(phydev); 1125 if (status) 1126 return status; 1127 1128 /* First check if the EEE ability is supported */ 1129 eee_cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1130 if (eee_cap <= 0) 1131 goto eee_exit_err; 1132 1133 cap = mmd_eee_cap_to_ethtool_sup_t(eee_cap); 1134 if (!cap) 1135 goto eee_exit_err; 1136 1137 /* Check which link settings negotiated and verify it in 1138 * the EEE advertising registers. 1139 */ 1140 eee_lp = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1141 if (eee_lp <= 0) 1142 goto eee_exit_err; 1143 1144 eee_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1145 if (eee_adv <= 0) 1146 goto eee_exit_err; 1147 1148 adv = mmd_eee_adv_to_ethtool_adv_t(eee_adv); 1149 lp = mmd_eee_adv_to_ethtool_adv_t(eee_lp); 1150 if (!phy_check_valid(phydev->speed, phydev->duplex, lp & adv)) 1151 goto eee_exit_err; 1152 1153 if (clk_stop_enable) { 1154 /* Configure the PHY to stop receiving xMII 1155 * clock while it is signaling LPI. 1156 */ 1157 int val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1); 1158 if (val < 0) 1159 return val; 1160 1161 val |= MDIO_PCS_CTRL1_CLKSTOP_EN; 1162 phy_write_mmd(phydev, MDIO_MMD_PCS, MDIO_CTRL1, val); 1163 } 1164 1165 return 0; /* EEE supported */ 1166 } 1167 eee_exit_err: 1168 return -EPROTONOSUPPORT; 1169 } 1170 EXPORT_SYMBOL(phy_init_eee); 1171 1172 /** 1173 * phy_get_eee_err - report the EEE wake error count 1174 * @phydev: target phy_device struct 1175 * 1176 * Description: it is to report the number of time where the PHY 1177 * failed to complete its normal wake sequence. 1178 */ 1179 int phy_get_eee_err(struct phy_device *phydev) 1180 { 1181 if (!phydev->drv) 1182 return -EIO; 1183 1184 return phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_WK_ERR); 1185 } 1186 EXPORT_SYMBOL(phy_get_eee_err); 1187 1188 /** 1189 * phy_ethtool_get_eee - get EEE supported and status 1190 * @phydev: target phy_device struct 1191 * @data: ethtool_eee data 1192 * 1193 * Description: it reportes the Supported/Advertisement/LP Advertisement 1194 * capabilities. 1195 */ 1196 int phy_ethtool_get_eee(struct phy_device *phydev, struct ethtool_eee *data) 1197 { 1198 int val; 1199 1200 if (!phydev->drv) 1201 return -EIO; 1202 1203 /* Get Supported EEE */ 1204 val = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1205 if (val < 0) 1206 return val; 1207 data->supported = mmd_eee_cap_to_ethtool_sup_t(val); 1208 1209 /* Get advertisement EEE */ 1210 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1211 if (val < 0) 1212 return val; 1213 data->advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1214 1215 /* Get LP advertisement EEE */ 1216 val = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_LPABLE); 1217 if (val < 0) 1218 return val; 1219 data->lp_advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1220 1221 return 0; 1222 } 1223 EXPORT_SYMBOL(phy_ethtool_get_eee); 1224 1225 /** 1226 * phy_ethtool_set_eee - set EEE supported and status 1227 * @phydev: target phy_device struct 1228 * @data: ethtool_eee data 1229 * 1230 * Description: it is to program the Advertisement EEE register. 1231 */ 1232 int phy_ethtool_set_eee(struct phy_device *phydev, struct ethtool_eee *data) 1233 { 1234 int cap, old_adv, adv, ret; 1235 1236 if (!phydev->drv) 1237 return -EIO; 1238 1239 /* Get Supported EEE */ 1240 cap = phy_read_mmd(phydev, MDIO_MMD_PCS, MDIO_PCS_EEE_ABLE); 1241 if (cap < 0) 1242 return cap; 1243 1244 old_adv = phy_read_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV); 1245 if (old_adv < 0) 1246 return old_adv; 1247 1248 adv = ethtool_adv_to_mmd_eee_adv_t(data->advertised) & cap; 1249 1250 /* Mask prohibited EEE modes */ 1251 adv &= ~phydev->eee_broken_modes; 1252 1253 if (old_adv != adv) { 1254 ret = phy_write_mmd(phydev, MDIO_MMD_AN, MDIO_AN_EEE_ADV, adv); 1255 if (ret < 0) 1256 return ret; 1257 1258 /* Restart autonegotiation so the new modes get sent to the 1259 * link partner. 1260 */ 1261 ret = phy_restart_aneg(phydev); 1262 if (ret < 0) 1263 return ret; 1264 } 1265 1266 return 0; 1267 } 1268 EXPORT_SYMBOL(phy_ethtool_set_eee); 1269 1270 int phy_ethtool_set_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1271 { 1272 if (phydev->drv && phydev->drv->set_wol) 1273 return phydev->drv->set_wol(phydev, wol); 1274 1275 return -EOPNOTSUPP; 1276 } 1277 EXPORT_SYMBOL(phy_ethtool_set_wol); 1278 1279 void phy_ethtool_get_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1280 { 1281 if (phydev->drv && phydev->drv->get_wol) 1282 phydev->drv->get_wol(phydev, wol); 1283 } 1284 EXPORT_SYMBOL(phy_ethtool_get_wol); 1285 1286 int phy_ethtool_get_link_ksettings(struct net_device *ndev, 1287 struct ethtool_link_ksettings *cmd) 1288 { 1289 struct phy_device *phydev = ndev->phydev; 1290 1291 if (!phydev) 1292 return -ENODEV; 1293 1294 phy_ethtool_ksettings_get(phydev, cmd); 1295 1296 return 0; 1297 } 1298 EXPORT_SYMBOL(phy_ethtool_get_link_ksettings); 1299 1300 int phy_ethtool_set_link_ksettings(struct net_device *ndev, 1301 const struct ethtool_link_ksettings *cmd) 1302 { 1303 struct phy_device *phydev = ndev->phydev; 1304 1305 if (!phydev) 1306 return -ENODEV; 1307 1308 return phy_ethtool_ksettings_set(phydev, cmd); 1309 } 1310 EXPORT_SYMBOL(phy_ethtool_set_link_ksettings); 1311 1312 int phy_ethtool_nway_reset(struct net_device *ndev) 1313 { 1314 struct phy_device *phydev = ndev->phydev; 1315 1316 if (!phydev) 1317 return -ENODEV; 1318 1319 if (!phydev->drv) 1320 return -EIO; 1321 1322 return phy_restart_aneg(phydev); 1323 } 1324 EXPORT_SYMBOL(phy_ethtool_nway_reset); 1325