1 /* Framework for configuring and reading PHY devices 2 * Based on code in sungem_phy.c and gianfar_phy.c 3 * 4 * Author: Andy Fleming 5 * 6 * Copyright (c) 2004 Freescale Semiconductor, Inc. 7 * Copyright (c) 2006, 2007 Maciej W. Rozycki 8 * 9 * This program is free software; you can redistribute it and/or modify it 10 * under the terms of the GNU General Public License as published by the 11 * Free Software Foundation; either version 2 of the License, or (at your 12 * option) any later version. 13 * 14 */ 15 16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 17 18 #include <linux/kernel.h> 19 #include <linux/string.h> 20 #include <linux/errno.h> 21 #include <linux/unistd.h> 22 #include <linux/interrupt.h> 23 #include <linux/delay.h> 24 #include <linux/netdevice.h> 25 #include <linux/etherdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/mm.h> 28 #include <linux/module.h> 29 #include <linux/mii.h> 30 #include <linux/ethtool.h> 31 #include <linux/phy.h> 32 #include <linux/phy_led_triggers.h> 33 #include <linux/timer.h> 34 #include <linux/workqueue.h> 35 #include <linux/mdio.h> 36 #include <linux/io.h> 37 #include <linux/uaccess.h> 38 #include <linux/atomic.h> 39 40 #include <asm/irq.h> 41 42 static const char *phy_speed_to_str(int speed) 43 { 44 switch (speed) { 45 case SPEED_10: 46 return "10Mbps"; 47 case SPEED_100: 48 return "100Mbps"; 49 case SPEED_1000: 50 return "1Gbps"; 51 case SPEED_2500: 52 return "2.5Gbps"; 53 case SPEED_10000: 54 return "10Gbps"; 55 case SPEED_UNKNOWN: 56 return "Unknown"; 57 default: 58 return "Unsupported (update phy.c)"; 59 } 60 } 61 62 #define PHY_STATE_STR(_state) \ 63 case PHY_##_state: \ 64 return __stringify(_state); \ 65 66 static const char *phy_state_to_str(enum phy_state st) 67 { 68 switch (st) { 69 PHY_STATE_STR(DOWN) 70 PHY_STATE_STR(STARTING) 71 PHY_STATE_STR(READY) 72 PHY_STATE_STR(PENDING) 73 PHY_STATE_STR(UP) 74 PHY_STATE_STR(AN) 75 PHY_STATE_STR(RUNNING) 76 PHY_STATE_STR(NOLINK) 77 PHY_STATE_STR(FORCING) 78 PHY_STATE_STR(CHANGELINK) 79 PHY_STATE_STR(HALTED) 80 PHY_STATE_STR(RESUMING) 81 } 82 83 return NULL; 84 } 85 86 87 /** 88 * phy_print_status - Convenience function to print out the current phy status 89 * @phydev: the phy_device struct 90 */ 91 void phy_print_status(struct phy_device *phydev) 92 { 93 if (phydev->link) { 94 netdev_info(phydev->attached_dev, 95 "Link is Up - %s/%s - flow control %s\n", 96 phy_speed_to_str(phydev->speed), 97 DUPLEX_FULL == phydev->duplex ? "Full" : "Half", 98 phydev->pause ? "rx/tx" : "off"); 99 } else { 100 netdev_info(phydev->attached_dev, "Link is Down\n"); 101 } 102 } 103 EXPORT_SYMBOL(phy_print_status); 104 105 /** 106 * phy_clear_interrupt - Ack the phy device's interrupt 107 * @phydev: the phy_device struct 108 * 109 * If the @phydev driver has an ack_interrupt function, call it to 110 * ack and clear the phy device's interrupt. 111 * 112 * Returns 0 on success or < 0 on error. 113 */ 114 static int phy_clear_interrupt(struct phy_device *phydev) 115 { 116 if (phydev->drv->ack_interrupt) 117 return phydev->drv->ack_interrupt(phydev); 118 119 return 0; 120 } 121 122 /** 123 * phy_config_interrupt - configure the PHY device for the requested interrupts 124 * @phydev: the phy_device struct 125 * @interrupts: interrupt flags to configure for this @phydev 126 * 127 * Returns 0 on success or < 0 on error. 128 */ 129 static int phy_config_interrupt(struct phy_device *phydev, u32 interrupts) 130 { 131 phydev->interrupts = interrupts; 132 if (phydev->drv->config_intr) 133 return phydev->drv->config_intr(phydev); 134 135 return 0; 136 } 137 138 139 /** 140 * phy_aneg_done - return auto-negotiation status 141 * @phydev: target phy_device struct 142 * 143 * Description: Return the auto-negotiation status from this @phydev 144 * Returns > 0 on success or < 0 on error. 0 means that auto-negotiation 145 * is still pending. 146 */ 147 int phy_aneg_done(struct phy_device *phydev) 148 { 149 if (phydev->drv && phydev->drv->aneg_done) 150 return phydev->drv->aneg_done(phydev); 151 152 return genphy_aneg_done(phydev); 153 } 154 EXPORT_SYMBOL(phy_aneg_done); 155 156 /* A structure for mapping a particular speed and duplex 157 * combination to a particular SUPPORTED and ADVERTISED value 158 */ 159 struct phy_setting { 160 int speed; 161 int duplex; 162 u32 setting; 163 }; 164 165 /* A mapping of all SUPPORTED settings to speed/duplex */ 166 static const struct phy_setting settings[] = { 167 { 168 .speed = SPEED_10000, 169 .duplex = DUPLEX_FULL, 170 .setting = SUPPORTED_10000baseKR_Full, 171 }, 172 { 173 .speed = SPEED_10000, 174 .duplex = DUPLEX_FULL, 175 .setting = SUPPORTED_10000baseKX4_Full, 176 }, 177 { 178 .speed = SPEED_10000, 179 .duplex = DUPLEX_FULL, 180 .setting = SUPPORTED_10000baseT_Full, 181 }, 182 { 183 .speed = SPEED_2500, 184 .duplex = DUPLEX_FULL, 185 .setting = SUPPORTED_2500baseX_Full, 186 }, 187 { 188 .speed = SPEED_1000, 189 .duplex = DUPLEX_FULL, 190 .setting = SUPPORTED_1000baseKX_Full, 191 }, 192 { 193 .speed = SPEED_1000, 194 .duplex = DUPLEX_FULL, 195 .setting = SUPPORTED_1000baseT_Full, 196 }, 197 { 198 .speed = SPEED_1000, 199 .duplex = DUPLEX_HALF, 200 .setting = SUPPORTED_1000baseT_Half, 201 }, 202 { 203 .speed = SPEED_100, 204 .duplex = DUPLEX_FULL, 205 .setting = SUPPORTED_100baseT_Full, 206 }, 207 { 208 .speed = SPEED_100, 209 .duplex = DUPLEX_HALF, 210 .setting = SUPPORTED_100baseT_Half, 211 }, 212 { 213 .speed = SPEED_10, 214 .duplex = DUPLEX_FULL, 215 .setting = SUPPORTED_10baseT_Full, 216 }, 217 { 218 .speed = SPEED_10, 219 .duplex = DUPLEX_HALF, 220 .setting = SUPPORTED_10baseT_Half, 221 }, 222 }; 223 224 #define MAX_NUM_SETTINGS ARRAY_SIZE(settings) 225 226 /** 227 * phy_find_setting - find a PHY settings array entry that matches speed & duplex 228 * @speed: speed to match 229 * @duplex: duplex to match 230 * 231 * Description: Searches the settings array for the setting which 232 * matches the desired speed and duplex, and returns the index 233 * of that setting. Returns the index of the last setting if 234 * none of the others match. 235 */ 236 static inline unsigned int phy_find_setting(int speed, int duplex) 237 { 238 unsigned int idx = 0; 239 240 while (idx < ARRAY_SIZE(settings) && 241 (settings[idx].speed != speed || settings[idx].duplex != duplex)) 242 idx++; 243 244 return idx < MAX_NUM_SETTINGS ? idx : MAX_NUM_SETTINGS - 1; 245 } 246 247 /** 248 * phy_find_valid - find a PHY setting that matches the requested features mask 249 * @idx: The first index in settings[] to search 250 * @features: A mask of the valid settings 251 * 252 * Description: Returns the index of the first valid setting less 253 * than or equal to the one pointed to by idx, as determined by 254 * the mask in features. Returns the index of the last setting 255 * if nothing else matches. 256 */ 257 static inline unsigned int phy_find_valid(unsigned int idx, u32 features) 258 { 259 while (idx < MAX_NUM_SETTINGS && !(settings[idx].setting & features)) 260 idx++; 261 262 return idx < MAX_NUM_SETTINGS ? idx : MAX_NUM_SETTINGS - 1; 263 } 264 265 /** 266 * phy_supported_speeds - return all speeds currently supported by a phy device 267 * @phy: The phy device to return supported speeds of. 268 * @speeds: buffer to store supported speeds in. 269 * @size: size of speeds buffer. 270 * 271 * Description: Returns the number of supported speeds, and fills the speeds 272 * buffer with the supported speeds. If speeds buffer is too small to contain 273 * all currently supported speeds, will return as many speeds as can fit. 274 */ 275 unsigned int phy_supported_speeds(struct phy_device *phy, 276 unsigned int *speeds, 277 unsigned int size) 278 { 279 unsigned int count = 0; 280 unsigned int idx = 0; 281 282 while (idx < MAX_NUM_SETTINGS && count < size) { 283 idx = phy_find_valid(idx, phy->supported); 284 285 if (!(settings[idx].setting & phy->supported)) 286 break; 287 288 /* Assumes settings are grouped by speed */ 289 if ((count == 0) || 290 (speeds[count - 1] != settings[idx].speed)) { 291 speeds[count] = settings[idx].speed; 292 count++; 293 } 294 idx++; 295 } 296 297 return count; 298 } 299 300 /** 301 * phy_check_valid - check if there is a valid PHY setting which matches 302 * speed, duplex, and feature mask 303 * @speed: speed to match 304 * @duplex: duplex to match 305 * @features: A mask of the valid settings 306 * 307 * Description: Returns true if there is a valid setting, false otherwise. 308 */ 309 static inline bool phy_check_valid(int speed, int duplex, u32 features) 310 { 311 unsigned int idx; 312 313 idx = phy_find_valid(phy_find_setting(speed, duplex), features); 314 315 return settings[idx].speed == speed && settings[idx].duplex == duplex && 316 (settings[idx].setting & features); 317 } 318 319 /** 320 * phy_sanitize_settings - make sure the PHY is set to supported speed and duplex 321 * @phydev: the target phy_device struct 322 * 323 * Description: Make sure the PHY is set to supported speeds and 324 * duplexes. Drop down by one in this order: 1000/FULL, 325 * 1000/HALF, 100/FULL, 100/HALF, 10/FULL, 10/HALF. 326 */ 327 static void phy_sanitize_settings(struct phy_device *phydev) 328 { 329 u32 features = phydev->supported; 330 unsigned int idx; 331 332 /* Sanitize settings based on PHY capabilities */ 333 if ((features & SUPPORTED_Autoneg) == 0) 334 phydev->autoneg = AUTONEG_DISABLE; 335 336 idx = phy_find_valid(phy_find_setting(phydev->speed, phydev->duplex), 337 features); 338 339 phydev->speed = settings[idx].speed; 340 phydev->duplex = settings[idx].duplex; 341 } 342 343 /** 344 * phy_ethtool_sset - generic ethtool sset function, handles all the details 345 * @phydev: target phy_device struct 346 * @cmd: ethtool_cmd 347 * 348 * A few notes about parameter checking: 349 * - We don't set port or transceiver, so we don't care what they 350 * were set to. 351 * - phy_start_aneg() will make sure forced settings are sane, and 352 * choose the next best ones from the ones selected, so we don't 353 * care if ethtool tries to give us bad values. 354 */ 355 int phy_ethtool_sset(struct phy_device *phydev, struct ethtool_cmd *cmd) 356 { 357 u32 speed = ethtool_cmd_speed(cmd); 358 359 if (cmd->phy_address != phydev->mdio.addr) 360 return -EINVAL; 361 362 /* We make sure that we don't pass unsupported values in to the PHY */ 363 cmd->advertising &= phydev->supported; 364 365 /* Verify the settings we care about. */ 366 if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE) 367 return -EINVAL; 368 369 if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0) 370 return -EINVAL; 371 372 if (cmd->autoneg == AUTONEG_DISABLE && 373 ((speed != SPEED_1000 && 374 speed != SPEED_100 && 375 speed != SPEED_10) || 376 (cmd->duplex != DUPLEX_HALF && 377 cmd->duplex != DUPLEX_FULL))) 378 return -EINVAL; 379 380 phydev->autoneg = cmd->autoneg; 381 382 phydev->speed = speed; 383 384 phydev->advertising = cmd->advertising; 385 386 if (AUTONEG_ENABLE == cmd->autoneg) 387 phydev->advertising |= ADVERTISED_Autoneg; 388 else 389 phydev->advertising &= ~ADVERTISED_Autoneg; 390 391 phydev->duplex = cmd->duplex; 392 393 phydev->mdix_ctrl = cmd->eth_tp_mdix_ctrl; 394 395 /* Restart the PHY */ 396 phy_start_aneg(phydev); 397 398 return 0; 399 } 400 EXPORT_SYMBOL(phy_ethtool_sset); 401 402 int phy_ethtool_ksettings_set(struct phy_device *phydev, 403 const struct ethtool_link_ksettings *cmd) 404 { 405 u8 autoneg = cmd->base.autoneg; 406 u8 duplex = cmd->base.duplex; 407 u32 speed = cmd->base.speed; 408 u32 advertising; 409 410 if (cmd->base.phy_address != phydev->mdio.addr) 411 return -EINVAL; 412 413 ethtool_convert_link_mode_to_legacy_u32(&advertising, 414 cmd->link_modes.advertising); 415 416 /* We make sure that we don't pass unsupported values in to the PHY */ 417 advertising &= phydev->supported; 418 419 /* Verify the settings we care about. */ 420 if (autoneg != AUTONEG_ENABLE && autoneg != AUTONEG_DISABLE) 421 return -EINVAL; 422 423 if (autoneg == AUTONEG_ENABLE && advertising == 0) 424 return -EINVAL; 425 426 if (autoneg == AUTONEG_DISABLE && 427 ((speed != SPEED_1000 && 428 speed != SPEED_100 && 429 speed != SPEED_10) || 430 (duplex != DUPLEX_HALF && 431 duplex != DUPLEX_FULL))) 432 return -EINVAL; 433 434 phydev->autoneg = autoneg; 435 436 phydev->speed = speed; 437 438 phydev->advertising = advertising; 439 440 if (autoneg == AUTONEG_ENABLE) 441 phydev->advertising |= ADVERTISED_Autoneg; 442 else 443 phydev->advertising &= ~ADVERTISED_Autoneg; 444 445 phydev->duplex = duplex; 446 447 phydev->mdix_ctrl = cmd->base.eth_tp_mdix_ctrl; 448 449 /* Restart the PHY */ 450 phy_start_aneg(phydev); 451 452 return 0; 453 } 454 EXPORT_SYMBOL(phy_ethtool_ksettings_set); 455 456 int phy_ethtool_gset(struct phy_device *phydev, struct ethtool_cmd *cmd) 457 { 458 cmd->supported = phydev->supported; 459 460 cmd->advertising = phydev->advertising; 461 cmd->lp_advertising = phydev->lp_advertising; 462 463 ethtool_cmd_speed_set(cmd, phydev->speed); 464 cmd->duplex = phydev->duplex; 465 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 466 cmd->port = PORT_BNC; 467 else 468 cmd->port = PORT_MII; 469 cmd->phy_address = phydev->mdio.addr; 470 cmd->transceiver = phy_is_internal(phydev) ? 471 XCVR_INTERNAL : XCVR_EXTERNAL; 472 cmd->autoneg = phydev->autoneg; 473 cmd->eth_tp_mdix_ctrl = phydev->mdix_ctrl; 474 cmd->eth_tp_mdix = phydev->mdix; 475 476 return 0; 477 } 478 EXPORT_SYMBOL(phy_ethtool_gset); 479 480 int phy_ethtool_ksettings_get(struct phy_device *phydev, 481 struct ethtool_link_ksettings *cmd) 482 { 483 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported, 484 phydev->supported); 485 486 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising, 487 phydev->advertising); 488 489 ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.lp_advertising, 490 phydev->lp_advertising); 491 492 cmd->base.speed = phydev->speed; 493 cmd->base.duplex = phydev->duplex; 494 if (phydev->interface == PHY_INTERFACE_MODE_MOCA) 495 cmd->base.port = PORT_BNC; 496 else 497 cmd->base.port = PORT_MII; 498 499 cmd->base.phy_address = phydev->mdio.addr; 500 cmd->base.autoneg = phydev->autoneg; 501 cmd->base.eth_tp_mdix_ctrl = phydev->mdix_ctrl; 502 cmd->base.eth_tp_mdix = phydev->mdix; 503 504 return 0; 505 } 506 EXPORT_SYMBOL(phy_ethtool_ksettings_get); 507 508 /** 509 * phy_mii_ioctl - generic PHY MII ioctl interface 510 * @phydev: the phy_device struct 511 * @ifr: &struct ifreq for socket ioctl's 512 * @cmd: ioctl cmd to execute 513 * 514 * Note that this function is currently incompatible with the 515 * PHYCONTROL layer. It changes registers without regard to 516 * current state. Use at own risk. 517 */ 518 int phy_mii_ioctl(struct phy_device *phydev, struct ifreq *ifr, int cmd) 519 { 520 struct mii_ioctl_data *mii_data = if_mii(ifr); 521 u16 val = mii_data->val_in; 522 bool change_autoneg = false; 523 524 switch (cmd) { 525 case SIOCGMIIPHY: 526 mii_data->phy_id = phydev->mdio.addr; 527 /* fall through */ 528 529 case SIOCGMIIREG: 530 mii_data->val_out = mdiobus_read(phydev->mdio.bus, 531 mii_data->phy_id, 532 mii_data->reg_num); 533 return 0; 534 535 case SIOCSMIIREG: 536 if (mii_data->phy_id == phydev->mdio.addr) { 537 switch (mii_data->reg_num) { 538 case MII_BMCR: 539 if ((val & (BMCR_RESET | BMCR_ANENABLE)) == 0) { 540 if (phydev->autoneg == AUTONEG_ENABLE) 541 change_autoneg = true; 542 phydev->autoneg = AUTONEG_DISABLE; 543 if (val & BMCR_FULLDPLX) 544 phydev->duplex = DUPLEX_FULL; 545 else 546 phydev->duplex = DUPLEX_HALF; 547 if (val & BMCR_SPEED1000) 548 phydev->speed = SPEED_1000; 549 else if (val & BMCR_SPEED100) 550 phydev->speed = SPEED_100; 551 else phydev->speed = SPEED_10; 552 } 553 else { 554 if (phydev->autoneg == AUTONEG_DISABLE) 555 change_autoneg = true; 556 phydev->autoneg = AUTONEG_ENABLE; 557 } 558 break; 559 case MII_ADVERTISE: 560 phydev->advertising = mii_adv_to_ethtool_adv_t(val); 561 change_autoneg = true; 562 break; 563 default: 564 /* do nothing */ 565 break; 566 } 567 } 568 569 mdiobus_write(phydev->mdio.bus, mii_data->phy_id, 570 mii_data->reg_num, val); 571 572 if (mii_data->phy_id == phydev->mdio.addr && 573 mii_data->reg_num == MII_BMCR && 574 val & BMCR_RESET) 575 return phy_init_hw(phydev); 576 577 if (change_autoneg) 578 return phy_start_aneg(phydev); 579 580 return 0; 581 582 case SIOCSHWTSTAMP: 583 if (phydev->drv && phydev->drv->hwtstamp) 584 return phydev->drv->hwtstamp(phydev, ifr); 585 /* fall through */ 586 587 default: 588 return -EOPNOTSUPP; 589 } 590 } 591 EXPORT_SYMBOL(phy_mii_ioctl); 592 593 /** 594 * phy_start_aneg - start auto-negotiation for this PHY device 595 * @phydev: the phy_device struct 596 * 597 * Description: Sanitizes the settings (if we're not autonegotiating 598 * them), and then calls the driver's config_aneg function. 599 * If the PHYCONTROL Layer is operating, we change the state to 600 * reflect the beginning of Auto-negotiation or forcing. 601 */ 602 int phy_start_aneg(struct phy_device *phydev) 603 { 604 int err; 605 606 if (!phydev->drv) 607 return -EIO; 608 609 mutex_lock(&phydev->lock); 610 611 if (AUTONEG_DISABLE == phydev->autoneg) 612 phy_sanitize_settings(phydev); 613 614 /* Invalidate LP advertising flags */ 615 phydev->lp_advertising = 0; 616 617 err = phydev->drv->config_aneg(phydev); 618 if (err < 0) 619 goto out_unlock; 620 621 if (phydev->state != PHY_HALTED) { 622 if (AUTONEG_ENABLE == phydev->autoneg) { 623 phydev->state = PHY_AN; 624 phydev->link_timeout = PHY_AN_TIMEOUT; 625 } else { 626 phydev->state = PHY_FORCING; 627 phydev->link_timeout = PHY_FORCE_TIMEOUT; 628 } 629 } 630 631 out_unlock: 632 mutex_unlock(&phydev->lock); 633 return err; 634 } 635 EXPORT_SYMBOL(phy_start_aneg); 636 637 /** 638 * phy_start_machine - start PHY state machine tracking 639 * @phydev: the phy_device struct 640 * 641 * Description: The PHY infrastructure can run a state machine 642 * which tracks whether the PHY is starting up, negotiating, 643 * etc. This function starts the timer which tracks the state 644 * of the PHY. If you want to maintain your own state machine, 645 * do not call this function. 646 */ 647 void phy_start_machine(struct phy_device *phydev) 648 { 649 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, HZ); 650 } 651 652 /** 653 * phy_trigger_machine - trigger the state machine to run 654 * 655 * @phydev: the phy_device struct 656 * @sync: indicate whether we should wait for the workqueue cancelation 657 * 658 * Description: There has been a change in state which requires that the 659 * state machine runs. 660 */ 661 662 static void phy_trigger_machine(struct phy_device *phydev, bool sync) 663 { 664 if (sync) 665 cancel_delayed_work_sync(&phydev->state_queue); 666 else 667 cancel_delayed_work(&phydev->state_queue); 668 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 0); 669 } 670 671 /** 672 * phy_stop_machine - stop the PHY state machine tracking 673 * @phydev: target phy_device struct 674 * 675 * Description: Stops the state machine timer, sets the state to UP 676 * (unless it wasn't up yet). This function must be called BEFORE 677 * phy_detach. 678 */ 679 void phy_stop_machine(struct phy_device *phydev) 680 { 681 cancel_delayed_work_sync(&phydev->state_queue); 682 683 mutex_lock(&phydev->lock); 684 if (phydev->state > PHY_UP) 685 phydev->state = PHY_UP; 686 mutex_unlock(&phydev->lock); 687 } 688 689 /** 690 * phy_error - enter HALTED state for this PHY device 691 * @phydev: target phy_device struct 692 * 693 * Moves the PHY to the HALTED state in response to a read 694 * or write error, and tells the controller the link is down. 695 * Must not be called from interrupt context, or while the 696 * phydev->lock is held. 697 */ 698 static void phy_error(struct phy_device *phydev) 699 { 700 mutex_lock(&phydev->lock); 701 phydev->state = PHY_HALTED; 702 mutex_unlock(&phydev->lock); 703 704 phy_trigger_machine(phydev, false); 705 } 706 707 /** 708 * phy_interrupt - PHY interrupt handler 709 * @irq: interrupt line 710 * @phy_dat: phy_device pointer 711 * 712 * Description: When a PHY interrupt occurs, the handler disables 713 * interrupts, and uses phy_change to handle the interrupt. 714 */ 715 static irqreturn_t phy_interrupt(int irq, void *phy_dat) 716 { 717 struct phy_device *phydev = phy_dat; 718 719 if (PHY_HALTED == phydev->state) 720 return IRQ_NONE; /* It can't be ours. */ 721 722 disable_irq_nosync(irq); 723 atomic_inc(&phydev->irq_disable); 724 725 phy_change(phydev); 726 727 return IRQ_HANDLED; 728 } 729 730 /** 731 * phy_enable_interrupts - Enable the interrupts from the PHY side 732 * @phydev: target phy_device struct 733 */ 734 static int phy_enable_interrupts(struct phy_device *phydev) 735 { 736 int err = phy_clear_interrupt(phydev); 737 738 if (err < 0) 739 return err; 740 741 return phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED); 742 } 743 744 /** 745 * phy_disable_interrupts - Disable the PHY interrupts from the PHY side 746 * @phydev: target phy_device struct 747 */ 748 static int phy_disable_interrupts(struct phy_device *phydev) 749 { 750 int err; 751 752 /* Disable PHY interrupts */ 753 err = phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 754 if (err) 755 goto phy_err; 756 757 /* Clear the interrupt */ 758 err = phy_clear_interrupt(phydev); 759 if (err) 760 goto phy_err; 761 762 return 0; 763 764 phy_err: 765 phy_error(phydev); 766 767 return err; 768 } 769 770 /** 771 * phy_start_interrupts - request and enable interrupts for a PHY device 772 * @phydev: target phy_device struct 773 * 774 * Description: Request the interrupt for the given PHY. 775 * If this fails, then we set irq to PHY_POLL. 776 * Otherwise, we enable the interrupts in the PHY. 777 * This should only be called with a valid IRQ number. 778 * Returns 0 on success or < 0 on error. 779 */ 780 int phy_start_interrupts(struct phy_device *phydev) 781 { 782 atomic_set(&phydev->irq_disable, 0); 783 if (request_threaded_irq(phydev->irq, NULL, phy_interrupt, 784 IRQF_ONESHOT | IRQF_SHARED, 785 phydev_name(phydev), phydev) < 0) { 786 pr_warn("%s: Can't get IRQ %d (PHY)\n", 787 phydev->mdio.bus->name, phydev->irq); 788 phydev->irq = PHY_POLL; 789 return 0; 790 } 791 792 return phy_enable_interrupts(phydev); 793 } 794 EXPORT_SYMBOL(phy_start_interrupts); 795 796 /** 797 * phy_stop_interrupts - disable interrupts from a PHY device 798 * @phydev: target phy_device struct 799 */ 800 int phy_stop_interrupts(struct phy_device *phydev) 801 { 802 int err = phy_disable_interrupts(phydev); 803 804 if (err) 805 phy_error(phydev); 806 807 free_irq(phydev->irq, phydev); 808 809 /* If work indeed has been cancelled, disable_irq() will have 810 * been left unbalanced from phy_interrupt() and enable_irq() 811 * has to be called so that other devices on the line work. 812 */ 813 while (atomic_dec_return(&phydev->irq_disable) >= 0) 814 enable_irq(phydev->irq); 815 816 return err; 817 } 818 EXPORT_SYMBOL(phy_stop_interrupts); 819 820 /** 821 * phy_change - Called by the phy_interrupt to handle PHY changes 822 * @phydev: phy_device struct that interrupted 823 */ 824 void phy_change(struct phy_device *phydev) 825 { 826 if (phy_interrupt_is_valid(phydev)) { 827 if (phydev->drv->did_interrupt && 828 !phydev->drv->did_interrupt(phydev)) 829 goto ignore; 830 831 if (phy_disable_interrupts(phydev)) 832 goto phy_err; 833 } 834 835 mutex_lock(&phydev->lock); 836 if ((PHY_RUNNING == phydev->state) || (PHY_NOLINK == phydev->state)) 837 phydev->state = PHY_CHANGELINK; 838 mutex_unlock(&phydev->lock); 839 840 if (phy_interrupt_is_valid(phydev)) { 841 atomic_dec(&phydev->irq_disable); 842 enable_irq(phydev->irq); 843 844 /* Reenable interrupts */ 845 if (PHY_HALTED != phydev->state && 846 phy_config_interrupt(phydev, PHY_INTERRUPT_ENABLED)) 847 goto irq_enable_err; 848 } 849 850 /* reschedule state queue work to run as soon as possible */ 851 phy_trigger_machine(phydev, true); 852 return; 853 854 ignore: 855 atomic_dec(&phydev->irq_disable); 856 enable_irq(phydev->irq); 857 return; 858 859 irq_enable_err: 860 disable_irq(phydev->irq); 861 atomic_inc(&phydev->irq_disable); 862 phy_err: 863 phy_error(phydev); 864 } 865 866 /** 867 * phy_change_work - Scheduled by the phy_mac_interrupt to handle PHY changes 868 * @work: work_struct that describes the work to be done 869 */ 870 void phy_change_work(struct work_struct *work) 871 { 872 struct phy_device *phydev = 873 container_of(work, struct phy_device, phy_queue); 874 875 phy_change(phydev); 876 } 877 878 /** 879 * phy_stop - Bring down the PHY link, and stop checking the status 880 * @phydev: target phy_device struct 881 */ 882 void phy_stop(struct phy_device *phydev) 883 { 884 mutex_lock(&phydev->lock); 885 886 if (PHY_HALTED == phydev->state) 887 goto out_unlock; 888 889 if (phy_interrupt_is_valid(phydev)) { 890 /* Disable PHY Interrupts */ 891 phy_config_interrupt(phydev, PHY_INTERRUPT_DISABLED); 892 893 /* Clear any pending interrupts */ 894 phy_clear_interrupt(phydev); 895 } 896 897 phydev->state = PHY_HALTED; 898 899 out_unlock: 900 mutex_unlock(&phydev->lock); 901 902 /* Cannot call flush_scheduled_work() here as desired because 903 * of rtnl_lock(), but PHY_HALTED shall guarantee phy_change() 904 * will not reenable interrupts. 905 */ 906 } 907 EXPORT_SYMBOL(phy_stop); 908 909 /** 910 * phy_start - start or restart a PHY device 911 * @phydev: target phy_device struct 912 * 913 * Description: Indicates the attached device's readiness to 914 * handle PHY-related work. Used during startup to start the 915 * PHY, and after a call to phy_stop() to resume operation. 916 * Also used to indicate the MDIO bus has cleared an error 917 * condition. 918 */ 919 void phy_start(struct phy_device *phydev) 920 { 921 bool do_resume = false; 922 int err = 0; 923 924 mutex_lock(&phydev->lock); 925 926 switch (phydev->state) { 927 case PHY_STARTING: 928 phydev->state = PHY_PENDING; 929 break; 930 case PHY_READY: 931 phydev->state = PHY_UP; 932 break; 933 case PHY_HALTED: 934 /* make sure interrupts are re-enabled for the PHY */ 935 if (phydev->irq != PHY_POLL) { 936 err = phy_enable_interrupts(phydev); 937 if (err < 0) 938 break; 939 } 940 941 phydev->state = PHY_RESUMING; 942 do_resume = true; 943 break; 944 default: 945 break; 946 } 947 mutex_unlock(&phydev->lock); 948 949 /* if phy was suspended, bring the physical link up again */ 950 if (do_resume) 951 phy_resume(phydev); 952 953 phy_trigger_machine(phydev, true); 954 } 955 EXPORT_SYMBOL(phy_start); 956 957 static void phy_adjust_link(struct phy_device *phydev) 958 { 959 phydev->adjust_link(phydev->attached_dev); 960 phy_led_trigger_change_speed(phydev); 961 } 962 963 /** 964 * phy_state_machine - Handle the state machine 965 * @work: work_struct that describes the work to be done 966 */ 967 void phy_state_machine(struct work_struct *work) 968 { 969 struct delayed_work *dwork = to_delayed_work(work); 970 struct phy_device *phydev = 971 container_of(dwork, struct phy_device, state_queue); 972 bool needs_aneg = false, do_suspend = false; 973 enum phy_state old_state; 974 int err = 0; 975 int old_link; 976 977 mutex_lock(&phydev->lock); 978 979 old_state = phydev->state; 980 981 if (phydev->drv && phydev->drv->link_change_notify) 982 phydev->drv->link_change_notify(phydev); 983 984 switch (phydev->state) { 985 case PHY_DOWN: 986 case PHY_STARTING: 987 case PHY_READY: 988 case PHY_PENDING: 989 break; 990 case PHY_UP: 991 needs_aneg = true; 992 993 phydev->link_timeout = PHY_AN_TIMEOUT; 994 995 break; 996 case PHY_AN: 997 err = phy_read_status(phydev); 998 if (err < 0) 999 break; 1000 1001 /* If the link is down, give up on negotiation for now */ 1002 if (!phydev->link) { 1003 phydev->state = PHY_NOLINK; 1004 netif_carrier_off(phydev->attached_dev); 1005 phy_adjust_link(phydev); 1006 break; 1007 } 1008 1009 /* Check if negotiation is done. Break if there's an error */ 1010 err = phy_aneg_done(phydev); 1011 if (err < 0) 1012 break; 1013 1014 /* If AN is done, we're running */ 1015 if (err > 0) { 1016 phydev->state = PHY_RUNNING; 1017 netif_carrier_on(phydev->attached_dev); 1018 phy_adjust_link(phydev); 1019 1020 } else if (0 == phydev->link_timeout--) 1021 needs_aneg = true; 1022 break; 1023 case PHY_NOLINK: 1024 if (phy_interrupt_is_valid(phydev)) 1025 break; 1026 1027 err = phy_read_status(phydev); 1028 if (err) 1029 break; 1030 1031 if (phydev->link) { 1032 if (AUTONEG_ENABLE == phydev->autoneg) { 1033 err = phy_aneg_done(phydev); 1034 if (err < 0) 1035 break; 1036 1037 if (!err) { 1038 phydev->state = PHY_AN; 1039 phydev->link_timeout = PHY_AN_TIMEOUT; 1040 break; 1041 } 1042 } 1043 phydev->state = PHY_RUNNING; 1044 netif_carrier_on(phydev->attached_dev); 1045 phy_adjust_link(phydev); 1046 } 1047 break; 1048 case PHY_FORCING: 1049 err = genphy_update_link(phydev); 1050 if (err) 1051 break; 1052 1053 if (phydev->link) { 1054 phydev->state = PHY_RUNNING; 1055 netif_carrier_on(phydev->attached_dev); 1056 } else { 1057 if (0 == phydev->link_timeout--) 1058 needs_aneg = true; 1059 } 1060 1061 phy_adjust_link(phydev); 1062 break; 1063 case PHY_RUNNING: 1064 /* Only register a CHANGE if we are polling and link changed 1065 * since latest checking. 1066 */ 1067 if (phydev->irq == PHY_POLL) { 1068 old_link = phydev->link; 1069 err = phy_read_status(phydev); 1070 if (err) 1071 break; 1072 1073 if (old_link != phydev->link) 1074 phydev->state = PHY_CHANGELINK; 1075 } 1076 /* 1077 * Failsafe: check that nobody set phydev->link=0 between two 1078 * poll cycles, otherwise we won't leave RUNNING state as long 1079 * as link remains down. 1080 */ 1081 if (!phydev->link && phydev->state == PHY_RUNNING) { 1082 phydev->state = PHY_CHANGELINK; 1083 phydev_err(phydev, "no link in PHY_RUNNING\n"); 1084 } 1085 break; 1086 case PHY_CHANGELINK: 1087 err = phy_read_status(phydev); 1088 if (err) 1089 break; 1090 1091 if (phydev->link) { 1092 phydev->state = PHY_RUNNING; 1093 netif_carrier_on(phydev->attached_dev); 1094 } else { 1095 phydev->state = PHY_NOLINK; 1096 netif_carrier_off(phydev->attached_dev); 1097 } 1098 1099 phy_adjust_link(phydev); 1100 1101 if (phy_interrupt_is_valid(phydev)) 1102 err = phy_config_interrupt(phydev, 1103 PHY_INTERRUPT_ENABLED); 1104 break; 1105 case PHY_HALTED: 1106 if (phydev->link) { 1107 phydev->link = 0; 1108 netif_carrier_off(phydev->attached_dev); 1109 phy_adjust_link(phydev); 1110 do_suspend = true; 1111 } 1112 break; 1113 case PHY_RESUMING: 1114 if (AUTONEG_ENABLE == phydev->autoneg) { 1115 err = phy_aneg_done(phydev); 1116 if (err < 0) 1117 break; 1118 1119 /* err > 0 if AN is done. 1120 * Otherwise, it's 0, and we're still waiting for AN 1121 */ 1122 if (err > 0) { 1123 err = phy_read_status(phydev); 1124 if (err) 1125 break; 1126 1127 if (phydev->link) { 1128 phydev->state = PHY_RUNNING; 1129 netif_carrier_on(phydev->attached_dev); 1130 } else { 1131 phydev->state = PHY_NOLINK; 1132 } 1133 phy_adjust_link(phydev); 1134 } else { 1135 phydev->state = PHY_AN; 1136 phydev->link_timeout = PHY_AN_TIMEOUT; 1137 } 1138 } else { 1139 err = phy_read_status(phydev); 1140 if (err) 1141 break; 1142 1143 if (phydev->link) { 1144 phydev->state = PHY_RUNNING; 1145 netif_carrier_on(phydev->attached_dev); 1146 } else { 1147 phydev->state = PHY_NOLINK; 1148 } 1149 phy_adjust_link(phydev); 1150 } 1151 break; 1152 } 1153 1154 mutex_unlock(&phydev->lock); 1155 1156 if (needs_aneg) 1157 err = phy_start_aneg(phydev); 1158 else if (do_suspend) 1159 phy_suspend(phydev); 1160 1161 if (err < 0) 1162 phy_error(phydev); 1163 1164 phydev_dbg(phydev, "PHY state change %s -> %s\n", 1165 phy_state_to_str(old_state), 1166 phy_state_to_str(phydev->state)); 1167 1168 /* Only re-schedule a PHY state machine change if we are polling the 1169 * PHY, if PHY_IGNORE_INTERRUPT is set, then we will be moving 1170 * between states from phy_mac_interrupt() 1171 */ 1172 if (phydev->irq == PHY_POLL) 1173 queue_delayed_work(system_power_efficient_wq, &phydev->state_queue, 1174 PHY_STATE_TIME * HZ); 1175 } 1176 1177 /** 1178 * phy_mac_interrupt - MAC says the link has changed 1179 * @phydev: phy_device struct with changed link 1180 * @new_link: Link is Up/Down. 1181 * 1182 * Description: The MAC layer is able indicate there has been a change 1183 * in the PHY link status. Set the new link status, and trigger the 1184 * state machine, work a work queue. 1185 */ 1186 void phy_mac_interrupt(struct phy_device *phydev, int new_link) 1187 { 1188 phydev->link = new_link; 1189 1190 /* Trigger a state machine change */ 1191 queue_work(system_power_efficient_wq, &phydev->phy_queue); 1192 } 1193 EXPORT_SYMBOL(phy_mac_interrupt); 1194 1195 static inline void mmd_phy_indirect(struct mii_bus *bus, int prtad, int devad, 1196 int addr) 1197 { 1198 /* Write the desired MMD Devad */ 1199 bus->write(bus, addr, MII_MMD_CTRL, devad); 1200 1201 /* Write the desired MMD register address */ 1202 bus->write(bus, addr, MII_MMD_DATA, prtad); 1203 1204 /* Select the Function : DATA with no post increment */ 1205 bus->write(bus, addr, MII_MMD_CTRL, (devad | MII_MMD_CTRL_NOINCR)); 1206 } 1207 1208 /** 1209 * phy_read_mmd_indirect - reads data from the MMD registers 1210 * @phydev: The PHY device bus 1211 * @prtad: MMD Address 1212 * @devad: MMD DEVAD 1213 * 1214 * Description: it reads data from the MMD registers (clause 22 to access to 1215 * clause 45) of the specified phy address. 1216 * To read these register we have: 1217 * 1) Write reg 13 // DEVAD 1218 * 2) Write reg 14 // MMD Address 1219 * 3) Write reg 13 // MMD Data Command for MMD DEVAD 1220 * 3) Read reg 14 // Read MMD data 1221 */ 1222 int phy_read_mmd_indirect(struct phy_device *phydev, int prtad, int devad) 1223 { 1224 struct phy_driver *phydrv = phydev->drv; 1225 int addr = phydev->mdio.addr; 1226 int value = -1; 1227 1228 if (!phydrv->read_mmd_indirect) { 1229 struct mii_bus *bus = phydev->mdio.bus; 1230 1231 mutex_lock(&bus->mdio_lock); 1232 mmd_phy_indirect(bus, prtad, devad, addr); 1233 1234 /* Read the content of the MMD's selected register */ 1235 value = bus->read(bus, addr, MII_MMD_DATA); 1236 mutex_unlock(&bus->mdio_lock); 1237 } else { 1238 value = phydrv->read_mmd_indirect(phydev, prtad, devad, addr); 1239 } 1240 return value; 1241 } 1242 EXPORT_SYMBOL(phy_read_mmd_indirect); 1243 1244 /** 1245 * phy_write_mmd_indirect - writes data to the MMD registers 1246 * @phydev: The PHY device 1247 * @prtad: MMD Address 1248 * @devad: MMD DEVAD 1249 * @data: data to write in the MMD register 1250 * 1251 * Description: Write data from the MMD registers of the specified 1252 * phy address. 1253 * To write these register we have: 1254 * 1) Write reg 13 // DEVAD 1255 * 2) Write reg 14 // MMD Address 1256 * 3) Write reg 13 // MMD Data Command for MMD DEVAD 1257 * 3) Write reg 14 // Write MMD data 1258 */ 1259 void phy_write_mmd_indirect(struct phy_device *phydev, int prtad, 1260 int devad, u32 data) 1261 { 1262 struct phy_driver *phydrv = phydev->drv; 1263 int addr = phydev->mdio.addr; 1264 1265 if (!phydrv->write_mmd_indirect) { 1266 struct mii_bus *bus = phydev->mdio.bus; 1267 1268 mutex_lock(&bus->mdio_lock); 1269 mmd_phy_indirect(bus, prtad, devad, addr); 1270 1271 /* Write the data into MMD's selected register */ 1272 bus->write(bus, addr, MII_MMD_DATA, data); 1273 mutex_unlock(&bus->mdio_lock); 1274 } else { 1275 phydrv->write_mmd_indirect(phydev, prtad, devad, addr, data); 1276 } 1277 } 1278 EXPORT_SYMBOL(phy_write_mmd_indirect); 1279 1280 /** 1281 * phy_init_eee - init and check the EEE feature 1282 * @phydev: target phy_device struct 1283 * @clk_stop_enable: PHY may stop the clock during LPI 1284 * 1285 * Description: it checks if the Energy-Efficient Ethernet (EEE) 1286 * is supported by looking at the MMD registers 3.20 and 7.60/61 1287 * and it programs the MMD register 3.0 setting the "Clock stop enable" 1288 * bit if required. 1289 */ 1290 int phy_init_eee(struct phy_device *phydev, bool clk_stop_enable) 1291 { 1292 if (!phydev->drv) 1293 return -EIO; 1294 1295 /* According to 802.3az,the EEE is supported only in full duplex-mode. 1296 * Also EEE feature is active when core is operating with MII, GMII 1297 * or RGMII (all kinds). Internal PHYs are also allowed to proceed and 1298 * should return an error if they do not support EEE. 1299 */ 1300 if ((phydev->duplex == DUPLEX_FULL) && 1301 ((phydev->interface == PHY_INTERFACE_MODE_MII) || 1302 (phydev->interface == PHY_INTERFACE_MODE_GMII) || 1303 phy_interface_is_rgmii(phydev) || 1304 phy_is_internal(phydev))) { 1305 int eee_lp, eee_cap, eee_adv; 1306 u32 lp, cap, adv; 1307 int status; 1308 1309 /* Read phy status to properly get the right settings */ 1310 status = phy_read_status(phydev); 1311 if (status) 1312 return status; 1313 1314 /* First check if the EEE ability is supported */ 1315 eee_cap = phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_ABLE, 1316 MDIO_MMD_PCS); 1317 if (eee_cap <= 0) 1318 goto eee_exit_err; 1319 1320 cap = mmd_eee_cap_to_ethtool_sup_t(eee_cap); 1321 if (!cap) 1322 goto eee_exit_err; 1323 1324 /* Check which link settings negotiated and verify it in 1325 * the EEE advertising registers. 1326 */ 1327 eee_lp = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_LPABLE, 1328 MDIO_MMD_AN); 1329 if (eee_lp <= 0) 1330 goto eee_exit_err; 1331 1332 eee_adv = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_ADV, 1333 MDIO_MMD_AN); 1334 if (eee_adv <= 0) 1335 goto eee_exit_err; 1336 1337 adv = mmd_eee_adv_to_ethtool_adv_t(eee_adv); 1338 lp = mmd_eee_adv_to_ethtool_adv_t(eee_lp); 1339 if (!phy_check_valid(phydev->speed, phydev->duplex, lp & adv)) 1340 goto eee_exit_err; 1341 1342 if (clk_stop_enable) { 1343 /* Configure the PHY to stop receiving xMII 1344 * clock while it is signaling LPI. 1345 */ 1346 int val = phy_read_mmd_indirect(phydev, MDIO_CTRL1, 1347 MDIO_MMD_PCS); 1348 if (val < 0) 1349 return val; 1350 1351 val |= MDIO_PCS_CTRL1_CLKSTOP_EN; 1352 phy_write_mmd_indirect(phydev, MDIO_CTRL1, 1353 MDIO_MMD_PCS, val); 1354 } 1355 1356 return 0; /* EEE supported */ 1357 } 1358 eee_exit_err: 1359 return -EPROTONOSUPPORT; 1360 } 1361 EXPORT_SYMBOL(phy_init_eee); 1362 1363 /** 1364 * phy_get_eee_err - report the EEE wake error count 1365 * @phydev: target phy_device struct 1366 * 1367 * Description: it is to report the number of time where the PHY 1368 * failed to complete its normal wake sequence. 1369 */ 1370 int phy_get_eee_err(struct phy_device *phydev) 1371 { 1372 if (!phydev->drv) 1373 return -EIO; 1374 1375 return phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_WK_ERR, MDIO_MMD_PCS); 1376 } 1377 EXPORT_SYMBOL(phy_get_eee_err); 1378 1379 /** 1380 * phy_ethtool_get_eee - get EEE supported and status 1381 * @phydev: target phy_device struct 1382 * @data: ethtool_eee data 1383 * 1384 * Description: it reportes the Supported/Advertisement/LP Advertisement 1385 * capabilities. 1386 */ 1387 int phy_ethtool_get_eee(struct phy_device *phydev, struct ethtool_eee *data) 1388 { 1389 int val; 1390 1391 if (!phydev->drv) 1392 return -EIO; 1393 1394 /* Get Supported EEE */ 1395 val = phy_read_mmd_indirect(phydev, MDIO_PCS_EEE_ABLE, MDIO_MMD_PCS); 1396 if (val < 0) 1397 return val; 1398 data->supported = mmd_eee_cap_to_ethtool_sup_t(val); 1399 1400 /* Get advertisement EEE */ 1401 val = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_ADV, MDIO_MMD_AN); 1402 if (val < 0) 1403 return val; 1404 data->advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1405 1406 /* Get LP advertisement EEE */ 1407 val = phy_read_mmd_indirect(phydev, MDIO_AN_EEE_LPABLE, MDIO_MMD_AN); 1408 if (val < 0) 1409 return val; 1410 data->lp_advertised = mmd_eee_adv_to_ethtool_adv_t(val); 1411 1412 return 0; 1413 } 1414 EXPORT_SYMBOL(phy_ethtool_get_eee); 1415 1416 /** 1417 * phy_ethtool_set_eee - set EEE supported and status 1418 * @phydev: target phy_device struct 1419 * @data: ethtool_eee data 1420 * 1421 * Description: it is to program the Advertisement EEE register. 1422 */ 1423 int phy_ethtool_set_eee(struct phy_device *phydev, struct ethtool_eee *data) 1424 { 1425 int val = ethtool_adv_to_mmd_eee_adv_t(data->advertised); 1426 1427 if (!phydev->drv) 1428 return -EIO; 1429 1430 /* Mask prohibited EEE modes */ 1431 val &= ~phydev->eee_broken_modes; 1432 1433 phy_write_mmd_indirect(phydev, MDIO_AN_EEE_ADV, MDIO_MMD_AN, val); 1434 1435 return 0; 1436 } 1437 EXPORT_SYMBOL(phy_ethtool_set_eee); 1438 1439 int phy_ethtool_set_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1440 { 1441 if (phydev->drv && phydev->drv->set_wol) 1442 return phydev->drv->set_wol(phydev, wol); 1443 1444 return -EOPNOTSUPP; 1445 } 1446 EXPORT_SYMBOL(phy_ethtool_set_wol); 1447 1448 void phy_ethtool_get_wol(struct phy_device *phydev, struct ethtool_wolinfo *wol) 1449 { 1450 if (phydev->drv && phydev->drv->get_wol) 1451 phydev->drv->get_wol(phydev, wol); 1452 } 1453 EXPORT_SYMBOL(phy_ethtool_get_wol); 1454 1455 int phy_ethtool_get_link_ksettings(struct net_device *ndev, 1456 struct ethtool_link_ksettings *cmd) 1457 { 1458 struct phy_device *phydev = ndev->phydev; 1459 1460 if (!phydev) 1461 return -ENODEV; 1462 1463 return phy_ethtool_ksettings_get(phydev, cmd); 1464 } 1465 EXPORT_SYMBOL(phy_ethtool_get_link_ksettings); 1466 1467 int phy_ethtool_set_link_ksettings(struct net_device *ndev, 1468 const struct ethtool_link_ksettings *cmd) 1469 { 1470 struct phy_device *phydev = ndev->phydev; 1471 1472 if (!phydev) 1473 return -ENODEV; 1474 1475 return phy_ethtool_ksettings_set(phydev, cmd); 1476 } 1477 EXPORT_SYMBOL(phy_ethtool_set_link_ksettings); 1478 1479 int phy_ethtool_nway_reset(struct net_device *ndev) 1480 { 1481 struct phy_device *phydev = ndev->phydev; 1482 1483 if (!phydev) 1484 return -ENODEV; 1485 1486 if (!phydev->drv) 1487 return -EIO; 1488 1489 return genphy_restart_aneg(phydev); 1490 } 1491 EXPORT_SYMBOL(phy_ethtool_nway_reset); 1492