1.. |struct cpuidle_state| replace:: :c:type:`struct cpuidle_state <cpuidle_state>` 2.. |cpufreq| replace:: :doc:`CPU Performance Scaling <cpufreq>` 3 4======================== 5CPU Idle Time Management 6======================== 7 8:: 9 10 Copyright (c) 2018 Intel Corp., Rafael J. Wysocki <rafael.j.wysocki@intel.com> 11 12Concepts 13======== 14 15Modern processors are generally able to enter states in which the execution of 16a program is suspended and instructions belonging to it are not fetched from 17memory or executed. Those states are the *idle* states of the processor. 18 19Since part of the processor hardware is not used in idle states, entering them 20generally allows power drawn by the processor to be reduced and, in consequence, 21it is an opportunity to save energy. 22 23CPU idle time management is an energy-efficiency feature concerned about using 24the idle states of processors for this purpose. 25 26Logical CPUs 27------------ 28 29CPU idle time management operates on CPUs as seen by the *CPU scheduler* (that 30is the part of the kernel responsible for the distribution of computational 31work in the system). In its view, CPUs are *logical* units. That is, they need 32not be separate physical entities and may just be interfaces appearing to 33software as individual single-core processors. In other words, a CPU is an 34entity which appears to be fetching instructions that belong to one sequence 35(program) from memory and executing them, but it need not work this way 36physically. Generally, three different cases can be consider here. 37 38First, if the whole processor can only follow one sequence of instructions (one 39program) at a time, it is a CPU. In that case, if the hardware is asked to 40enter an idle state, that applies to the processor as a whole. 41 42Second, if the processor is multi-core, each core in it is able to follow at 43least one program at a time. The cores need not be entirely independent of each 44other (for example, they may share caches), but still most of the time they 45work physically in parallel with each other, so if each of them executes only 46one program, those programs run mostly independently of each other at the same 47time. The entire cores are CPUs in that case and if the hardware is asked to 48enter an idle state, that applies to the core that asked for it in the first 49place, but it also may apply to a larger unit (say a "package" or a "cluster") 50that the core belongs to (in fact, it may apply to an entire hierarchy of larger 51units containing the core). Namely, if all of the cores in the larger unit 52except for one have been put into idle states at the "core level" and the 53remaining core asks the processor to enter an idle state, that may trigger it 54to put the whole larger unit into an idle state which also will affect the 55other cores in that unit. 56 57Finally, each core in a multi-core processor may be able to follow more than one 58program in the same time frame (that is, each core may be able to fetch 59instructions from multiple locations in memory and execute them in the same time 60frame, but not necessarily entirely in parallel with each other). In that case 61the cores present themselves to software as "bundles" each consisting of 62multiple individual single-core "processors", referred to as *hardware threads* 63(or hyper-threads specifically on Intel hardware), that each can follow one 64sequence of instructions. Then, the hardware threads are CPUs from the CPU idle 65time management perspective and if the processor is asked to enter an idle state 66by one of them, the hardware thread (or CPU) that asked for it is stopped, but 67nothing more happens, unless all of the other hardware threads within the same 68core also have asked the processor to enter an idle state. In that situation, 69the core may be put into an idle state individually or a larger unit containing 70it may be put into an idle state as a whole (if the other cores within the 71larger unit are in idle states already). 72 73Idle CPUs 74--------- 75 76Logical CPUs, simply referred to as "CPUs" in what follows, are regarded as 77*idle* by the Linux kernel when there are no tasks to run on them except for the 78special "idle" task. 79 80Tasks are the CPU scheduler's representation of work. Each task consists of a 81sequence of instructions to execute, or code, data to be manipulated while 82running that code, and some context information that needs to be loaded into the 83processor every time the task's code is run by a CPU. The CPU scheduler 84distributes work by assigning tasks to run to the CPUs present in the system. 85 86Tasks can be in various states. In particular, they are *runnable* if there are 87no specific conditions preventing their code from being run by a CPU as long as 88there is a CPU available for that (for example, they are not waiting for any 89events to occur or similar). When a task becomes runnable, the CPU scheduler 90assigns it to one of the available CPUs to run and if there are no more runnable 91tasks assigned to it, the CPU will load the given task's context and run its 92code (from the instruction following the last one executed so far, possibly by 93another CPU). [If there are multiple runnable tasks assigned to one CPU 94simultaneously, they will be subject to prioritization and time sharing in order 95to allow them to make some progress over time.] 96 97The special "idle" task becomes runnable if there are no other runnable tasks 98assigned to the given CPU and the CPU is then regarded as idle. In other words, 99in Linux idle CPUs run the code of the "idle" task called *the idle loop*. That 100code may cause the processor to be put into one of its idle states, if they are 101supported, in order to save energy, but if the processor does not support any 102idle states, or there is not enough time to spend in an idle state before the 103next wakeup event, or there are strict latency constraints preventing any of the 104available idle states from being used, the CPU will simply execute more or less 105useless instructions in a loop until it is assigned a new task to run. 106 107 108.. _idle-loop: 109 110The Idle Loop 111============= 112 113The idle loop code takes two major steps in every iteration of it. First, it 114calls into a code module referred to as the *governor* that belongs to the CPU 115idle time management subsystem called ``CPUIdle`` to select an idle state for 116the CPU to ask the hardware to enter. Second, it invokes another code module 117from the ``CPUIdle`` subsystem, called the *driver*, to actually ask the 118processor hardware to enter the idle state selected by the governor. 119 120The role of the governor is to find an idle state most suitable for the 121conditions at hand. For this purpose, idle states that the hardware can be 122asked to enter by logical CPUs are represented in an abstract way independent of 123the platform or the processor architecture and organized in a one-dimensional 124(linear) array. That array has to be prepared and supplied by the ``CPUIdle`` 125driver matching the platform the kernel is running on at the initialization 126time. This allows ``CPUIdle`` governors to be independent of the underlying 127hardware and to work with any platforms that the Linux kernel can run on. 128 129Each idle state present in that array is characterized by two parameters to be 130taken into account by the governor, the *target residency* and the (worst-case) 131*exit latency*. The target residency is the minimum time the hardware must 132spend in the given state, including the time needed to enter it (which may be 133substantial), in order to save more energy than it would save by entering one of 134the shallower idle states instead. [The "depth" of an idle state roughly 135corresponds to the power drawn by the processor in that state.] The exit 136latency, in turn, is the maximum time it will take a CPU asking the processor 137hardware to enter an idle state to start executing the first instruction after a 138wakeup from that state. Note that in general the exit latency also must cover 139the time needed to enter the given state in case the wakeup occurs when the 140hardware is entering it and it must be entered completely to be exited in an 141ordered manner. 142 143There are two types of information that can influence the governor's decisions. 144First of all, the governor knows the time until the closest timer event. That 145time is known exactly, because the kernel programs timers and it knows exactly 146when they will trigger, and it is the maximum time the hardware that the given 147CPU depends on can spend in an idle state, including the time necessary to enter 148and exit it. However, the CPU may be woken up by a non-timer event at any time 149(in particular, before the closest timer triggers) and it generally is not known 150when that may happen. The governor can only see how much time the CPU actually 151was idle after it has been woken up (that time will be referred to as the *idle 152duration* from now on) and it can use that information somehow along with the 153time until the closest timer to estimate the idle duration in future. How the 154governor uses that information depends on what algorithm is implemented by it 155and that is the primary reason for having more than one governor in the 156``CPUIdle`` subsystem. 157 158There are two ``CPUIdle`` governors available, ``menu`` and ``ladder``. Which 159of them is used depends on the configuration of the kernel and in particular on 160whether or not the scheduler tick can be `stopped by the idle 161loop <idle-cpus-and-tick_>`_. It is possible to change the governor at run time 162if the ``cpuidle_sysfs_switch`` command line parameter has been passed to the 163kernel, but that is not safe in general, so it should not be done on production 164systems (that may change in the future, though). The name of the ``CPUIdle`` 165governor currently used by the kernel can be read from the 166:file:`current_governor_ro` (or :file:`current_governor` if 167``cpuidle_sysfs_switch`` is present in the kernel command line) file under 168:file:`/sys/devices/system/cpu/cpuidle/` in ``sysfs``. 169 170Which ``CPUIdle`` driver is used, on the other hand, usually depends on the 171platform the kernel is running on, but there are platforms with more than one 172matching driver. For example, there are two drivers that can work with the 173majority of Intel platforms, ``intel_idle`` and ``acpi_idle``, one with 174hardcoded idle states information and the other able to read that information 175from the system's ACPI tables, respectively. Still, even in those cases, the 176driver chosen at the system initialization time cannot be replaced later, so the 177decision on which one of them to use has to be made early (on Intel platforms 178the ``acpi_idle`` driver will be used if ``intel_idle`` is disabled for some 179reason or if it does not recognize the processor). The name of the ``CPUIdle`` 180driver currently used by the kernel can be read from the :file:`current_driver` 181file under :file:`/sys/devices/system/cpu/cpuidle/` in ``sysfs``. 182 183 184.. _idle-cpus-and-tick: 185 186Idle CPUs and The Scheduler Tick 187================================ 188 189The scheduler tick is a timer that triggers periodically in order to implement 190the time sharing strategy of the CPU scheduler. Of course, if there are 191multiple runnable tasks assigned to one CPU at the same time, the only way to 192allow them to make reasonable progress in a given time frame is to make them 193share the available CPU time. Namely, in rough approximation, each task is 194given a slice of the CPU time to run its code, subject to the scheduling class, 195prioritization and so on and when that time slice is used up, the CPU should be 196switched over to running (the code of) another task. The currently running task 197may not want to give the CPU away voluntarily, however, and the scheduler tick 198is there to make the switch happen regardless. That is not the only role of the 199tick, but it is the primary reason for using it. 200 201The scheduler tick is problematic from the CPU idle time management perspective, 202because it triggers periodically and relatively often (depending on the kernel 203configuration, the length of the tick period is between 1 ms and 10 ms). 204Thus, if the tick is allowed to trigger on idle CPUs, it will not make sense 205for them to ask the hardware to enter idle states with target residencies above 206the tick period length. Moreover, in that case the idle duration of any CPU 207will never exceed the tick period length and the energy used for entering and 208exiting idle states due to the tick wakeups on idle CPUs will be wasted. 209 210Fortunately, it is not really necessary to allow the tick to trigger on idle 211CPUs, because (by definition) they have no tasks to run except for the special 212"idle" one. In other words, from the CPU scheduler perspective, the only user 213of the CPU time on them is the idle loop. Since the time of an idle CPU need 214not be shared between multiple runnable tasks, the primary reason for using the 215tick goes away if the given CPU is idle. Consequently, it is possible to stop 216the scheduler tick entirely on idle CPUs in principle, even though that may not 217always be worth the effort. 218 219Whether or not it makes sense to stop the scheduler tick in the idle loop 220depends on what is expected by the governor. First, if there is another 221(non-tick) timer due to trigger within the tick range, stopping the tick clearly 222would be a waste of time, even though the timer hardware may not need to be 223reprogrammed in that case. Second, if the governor is expecting a non-timer 224wakeup within the tick range, stopping the tick is not necessary and it may even 225be harmful. Namely, in that case the governor will select an idle state with 226the target residency within the time until the expected wakeup, so that state is 227going to be relatively shallow. The governor really cannot select a deep idle 228state then, as that would contradict its own expectation of a wakeup in short 229order. Now, if the wakeup really occurs shortly, stopping the tick would be a 230waste of time and in this case the timer hardware would need to be reprogrammed, 231which is expensive. On the other hand, if the tick is stopped and the wakeup 232does not occur any time soon, the hardware may spend indefinite amount of time 233in the shallow idle state selected by the governor, which will be a waste of 234energy. Hence, if the governor is expecting a wakeup of any kind within the 235tick range, it is better to allow the tick trigger. Otherwise, however, the 236governor will select a relatively deep idle state, so the tick should be stopped 237so that it does not wake up the CPU too early. 238 239In any case, the governor knows what it is expecting and the decision on whether 240or not to stop the scheduler tick belongs to it. Still, if the tick has been 241stopped already (in one of the previous iterations of the loop), it is better 242to leave it as is and the governor needs to take that into account. 243 244The kernel can be configured to disable stopping the scheduler tick in the idle 245loop altogether. That can be done through the build-time configuration of it 246(by unsetting the ``CONFIG_NO_HZ_IDLE`` configuration option) or by passing 247``nohz=off`` to it in the command line. In both cases, as the stopping of the 248scheduler tick is disabled, the governor's decisions regarding it are simply 249ignored by the idle loop code and the tick is never stopped. 250 251The systems that run kernels configured to allow the scheduler tick to be 252stopped on idle CPUs are referred to as *tickless* systems and they are 253generally regarded as more energy-efficient than the systems running kernels in 254which the tick cannot be stopped. If the given system is tickless, it will use 255the ``menu`` governor by default and if it is not tickless, the default 256``CPUIdle`` governor on it will be ``ladder``. 257 258 259The ``menu`` Governor 260===================== 261 262The ``menu`` governor is the default ``CPUIdle`` governor for tickless systems. 263It is quite complex, but the basic principle of its design is straightforward. 264Namely, when invoked to select an idle state for a CPU (i.e. an idle state that 265the CPU will ask the processor hardware to enter), it attempts to predict the 266idle duration and uses the predicted value for idle state selection. 267 268It first obtains the time until the closest timer event with the assumption 269that the scheduler tick will be stopped. That time, referred to as the *sleep 270length* in what follows, is the upper bound on the time before the next CPU 271wakeup. It is used to determine the sleep length range, which in turn is needed 272to get the sleep length correction factor. 273 274The ``menu`` governor maintains two arrays of sleep length correction factors. 275One of them is used when tasks previously running on the given CPU are waiting 276for some I/O operations to complete and the other one is used when that is not 277the case. Each array contains several correction factor values that correspond 278to different sleep length ranges organized so that each range represented in the 279array is approximately 10 times wider than the previous one. 280 281The correction factor for the given sleep length range (determined before 282selecting the idle state for the CPU) is updated after the CPU has been woken 283up and the closer the sleep length is to the observed idle duration, the closer 284to 1 the correction factor becomes (it must fall between 0 and 1 inclusive). 285The sleep length is multiplied by the correction factor for the range that it 286falls into to obtain the first approximation of the predicted idle duration. 287 288Next, the governor uses a simple pattern recognition algorithm to refine its 289idle duration prediction. Namely, it saves the last 8 observed idle duration 290values and, when predicting the idle duration next time, it computes the average 291and variance of them. If the variance is small (smaller than 400 square 292milliseconds) or it is small relative to the average (the average is greater 293that 6 times the standard deviation), the average is regarded as the "typical 294interval" value. Otherwise, the longest of the saved observed idle duration 295values is discarded and the computation is repeated for the remaining ones. 296Again, if the variance of them is small (in the above sense), the average is 297taken as the "typical interval" value and so on, until either the "typical 298interval" is determined or too many data points are disregarded, in which case 299the "typical interval" is assumed to equal "infinity" (the maximum unsigned 300integer value). The "typical interval" computed this way is compared with the 301sleep length multiplied by the correction factor and the minimum of the two is 302taken as the predicted idle duration. 303 304Then, the governor computes an extra latency limit to help "interactive" 305workloads. It uses the observation that if the exit latency of the selected 306idle state is comparable with the predicted idle duration, the total time spent 307in that state probably will be very short and the amount of energy to save by 308entering it will be relatively small, so likely it is better to avoid the 309overhead related to entering that state and exiting it. Thus selecting a 310shallower state is likely to be a better option then. The first approximation 311of the extra latency limit is the predicted idle duration itself which 312additionally is divided by a value depending on the number of tasks that 313previously ran on the given CPU and now they are waiting for I/O operations to 314complete. The result of that division is compared with the latency limit coming 315from the power management quality of service, or `PM QoS <cpu-pm-qos_>`_, 316framework and the minimum of the two is taken as the limit for the idle states' 317exit latency. 318 319Now, the governor is ready to walk the list of idle states and choose one of 320them. For this purpose, it compares the target residency of each state with 321the predicted idle duration and the exit latency of it with the computed latency 322limit. It selects the state with the target residency closest to the predicted 323idle duration, but still below it, and exit latency that does not exceed the 324limit. 325 326In the final step the governor may still need to refine the idle state selection 327if it has not decided to `stop the scheduler tick <idle-cpus-and-tick_>`_. That 328happens if the idle duration predicted by it is less than the tick period and 329the tick has not been stopped already (in a previous iteration of the idle 330loop). Then, the sleep length used in the previous computations may not reflect 331the real time until the closest timer event and if it really is greater than 332that time, the governor may need to select a shallower state with a suitable 333target residency. 334 335 336.. _idle-states-representation: 337 338Representation of Idle States 339============================= 340 341For the CPU idle time management purposes all of the physical idle states 342supported by the processor have to be represented as a one-dimensional array of 343|struct cpuidle_state| objects each allowing an individual (logical) CPU to ask 344the processor hardware to enter an idle state of certain properties. If there 345is a hierarchy of units in the processor, one |struct cpuidle_state| object can 346cover a combination of idle states supported by the units at different levels of 347the hierarchy. In that case, the `target residency and exit latency parameters 348of it <idle-loop_>`_, must reflect the properties of the idle state at the 349deepest level (i.e. the idle state of the unit containing all of the other 350units). 351 352For example, take a processor with two cores in a larger unit referred to as 353a "module" and suppose that asking the hardware to enter a specific idle state 354(say "X") at the "core" level by one core will trigger the module to try to 355enter a specific idle state of its own (say "MX") if the other core is in idle 356state "X" already. In other words, asking for idle state "X" at the "core" 357level gives the hardware a license to go as deep as to idle state "MX" at the 358"module" level, but there is no guarantee that this is going to happen (the core 359asking for idle state "X" may just end up in that state by itself instead). 360Then, the target residency of the |struct cpuidle_state| object representing 361idle state "X" must reflect the minimum time to spend in idle state "MX" of 362the module (including the time needed to enter it), because that is the minimum 363time the CPU needs to be idle to save any energy in case the hardware enters 364that state. Analogously, the exit latency parameter of that object must cover 365the exit time of idle state "MX" of the module (and usually its entry time too), 366because that is the maximum delay between a wakeup signal and the time the CPU 367will start to execute the first new instruction (assuming that both cores in the 368module will always be ready to execute instructions as soon as the module 369becomes operational as a whole). 370 371There are processors without direct coordination between different levels of the 372hierarchy of units inside them, however. In those cases asking for an idle 373state at the "core" level does not automatically affect the "module" level, for 374example, in any way and the ``CPUIdle`` driver is responsible for the entire 375handling of the hierarchy. Then, the definition of the idle state objects is 376entirely up to the driver, but still the physical properties of the idle state 377that the processor hardware finally goes into must always follow the parameters 378used by the governor for idle state selection (for instance, the actual exit 379latency of that idle state must not exceed the exit latency parameter of the 380idle state object selected by the governor). 381 382In addition to the target residency and exit latency idle state parameters 383discussed above, the objects representing idle states each contain a few other 384parameters describing the idle state and a pointer to the function to run in 385order to ask the hardware to enter that state. Also, for each 386|struct cpuidle_state| object, there is a corresponding 387:c:type:`struct cpuidle_state_usage <cpuidle_state_usage>` one containing usage 388statistics of the given idle state. That information is exposed by the kernel 389via ``sysfs``. 390 391For each CPU in the system, there is a :file:`/sys/devices/system/cpu<N>/cpuidle/` 392directory in ``sysfs``, where the number ``<N>`` is assigned to the given 393CPU at the initialization time. That directory contains a set of subdirectories 394called :file:`state0`, :file:`state1` and so on, up to the number of idle state 395objects defined for the given CPU minus one. Each of these directories 396corresponds to one idle state object and the larger the number in its name, the 397deeper the (effective) idle state represented by it. Each of them contains 398a number of files (attributes) representing the properties of the idle state 399object corresponding to it, as follows: 400 401``above`` 402 Total number of times this idle state had been asked for, but the 403 observed idle duration was certainly too short to match its target 404 residency. 405 406``below`` 407 Total number of times this idle state had been asked for, but cerainly 408 a deeper idle state would have been a better match for the observed idle 409 duration. 410 411``desc`` 412 Description of the idle state. 413 414``disable`` 415 Whether or not this idle state is disabled. 416 417``latency`` 418 Exit latency of the idle state in microseconds. 419 420``name`` 421 Name of the idle state. 422 423``power`` 424 Power drawn by hardware in this idle state in milliwatts (if specified, 425 0 otherwise). 426 427``residency`` 428 Target residency of the idle state in microseconds. 429 430``time`` 431 Total time spent in this idle state by the given CPU (as measured by the 432 kernel) in microseconds. 433 434``usage`` 435 Total number of times the hardware has been asked by the given CPU to 436 enter this idle state. 437 438The :file:`desc` and :file:`name` files both contain strings. The difference 439between them is that the name is expected to be more concise, while the 440description may be longer and it may contain white space or special characters. 441The other files listed above contain integer numbers. 442 443The :file:`disable` attribute is the only writeable one. If it contains 1, the 444given idle state is disabled for this particular CPU, which means that the 445governor will never select it for this particular CPU and the ``CPUIdle`` 446driver will never ask the hardware to enter it for that CPU as a result. 447However, disabling an idle state for one CPU does not prevent it from being 448asked for by the other CPUs, so it must be disabled for all of them in order to 449never be asked for by any of them. [Note that, due to the way the ``ladder`` 450governor is implemented, disabling an idle state prevents that governor from 451selecting any idle states deeper than the disabled one too.] 452 453If the :file:`disable` attribute contains 0, the given idle state is enabled for 454this particular CPU, but it still may be disabled for some or all of the other 455CPUs in the system at the same time. Writing 1 to it causes the idle state to 456be disabled for this particular CPU and writing 0 to it allows the governor to 457take it into consideration for the given CPU and the driver to ask for it, 458unless that state was disabled globally in the driver (in which case it cannot 459be used at all). 460 461The :file:`power` attribute is not defined very well, especially for idle state 462objects representing combinations of idle states at different levels of the 463hierarchy of units in the processor, and it generally is hard to obtain idle 464state power numbers for complex hardware, so :file:`power` often contains 0 (not 465available) and if it contains a nonzero number, that number may not be very 466accurate and it should not be relied on for anything meaningful. 467 468The number in the :file:`time` file generally may be greater than the total time 469really spent by the given CPU in the given idle state, because it is measured by 470the kernel and it may not cover the cases in which the hardware refused to enter 471this idle state and entered a shallower one instead of it (or even it did not 472enter any idle state at all). The kernel can only measure the time span between 473asking the hardware to enter an idle state and the subsequent wakeup of the CPU 474and it cannot say what really happened in the meantime at the hardware level. 475Moreover, if the idle state object in question represents a combination of idle 476states at different levels of the hierarchy of units in the processor, 477the kernel can never say how deep the hardware went down the hierarchy in any 478particular case. For these reasons, the only reliable way to find out how 479much time has been spent by the hardware in different idle states supported by 480it is to use idle state residency counters in the hardware, if available. 481 482 483.. _cpu-pm-qos: 484 485Power Management Quality of Service for CPUs 486============================================ 487 488The power management quality of service (PM QoS) framework in the Linux kernel 489allows kernel code and user space processes to set constraints on various 490energy-efficiency features of the kernel to prevent performance from dropping 491below a required level. The PM QoS constraints can be set globally, in 492predefined categories referred to as PM QoS classes, or against individual 493devices. 494 495CPU idle time management can be affected by PM QoS in two ways, through the 496global constraint in the ``PM_QOS_CPU_DMA_LATENCY`` class and through the 497resume latency constraints for individual CPUs. Kernel code (e.g. device 498drivers) can set both of them with the help of special internal interfaces 499provided by the PM QoS framework. User space can modify the former by opening 500the :file:`cpu_dma_latency` special device file under :file:`/dev/` and writing 501a binary value (interpreted as a signed 32-bit integer) to it. In turn, the 502resume latency constraint for a CPU can be modified by user space by writing a 503string (representing a signed 32-bit integer) to the 504:file:`power/pm_qos_resume_latency_us` file under 505:file:`/sys/devices/system/cpu/cpu<N>/` in ``sysfs``, where the CPU number 506``<N>`` is allocated at the system initialization time. Negative values 507will be rejected in both cases and, also in both cases, the written integer 508number will be interpreted as a requested PM QoS constraint in microseconds. 509 510The requested value is not automatically applied as a new constraint, however, 511as it may be less restrictive (greater in this particular case) than another 512constraint previously requested by someone else. For this reason, the PM QoS 513framework maintains a list of requests that have been made so far in each 514global class and for each device, aggregates them and applies the effective 515(minimum in this particular case) value as the new constraint. 516 517In fact, opening the :file:`cpu_dma_latency` special device file causes a new 518PM QoS request to be created and added to the priority list of requests in the 519``PM_QOS_CPU_DMA_LATENCY`` class and the file descriptor coming from the 520"open" operation represents that request. If that file descriptor is then 521used for writing, the number written to it will be associated with the PM QoS 522request represented by it as a new requested constraint value. Next, the 523priority list mechanism will be used to determine the new effective value of 524the entire list of requests and that effective value will be set as a new 525constraint. Thus setting a new requested constraint value will only change the 526real constraint if the effective "list" value is affected by it. In particular, 527for the ``PM_QOS_CPU_DMA_LATENCY`` class it only affects the real constraint if 528it is the minimum of the requested constraints in the list. The process holding 529a file descriptor obtained by opening the :file:`cpu_dma_latency` special device 530file controls the PM QoS request associated with that file descriptor, but it 531controls this particular PM QoS request only. 532 533Closing the :file:`cpu_dma_latency` special device file or, more precisely, the 534file descriptor obtained while opening it, causes the PM QoS request associated 535with that file descriptor to be removed from the ``PM_QOS_CPU_DMA_LATENCY`` 536class priority list and destroyed. If that happens, the priority list mechanism 537will be used, again, to determine the new effective value for the whole list 538and that value will become the new real constraint. 539 540In turn, for each CPU there is only one resume latency PM QoS request 541associated with the :file:`power/pm_qos_resume_latency_us` file under 542:file:`/sys/devices/system/cpu/cpu<N>/` in ``sysfs`` and writing to it causes 543this single PM QoS request to be updated regardless of which user space 544process does that. In other words, this PM QoS request is shared by the entire 545user space, so access to the file associated with it needs to be arbitrated 546to avoid confusion. [Arguably, the only legitimate use of this mechanism in 547practice is to pin a process to the CPU in question and let it use the 548``sysfs`` interface to control the resume latency constraint for it.] It 549still only is a request, however. It is a member of a priority list used to 550determine the effective value to be set as the resume latency constraint for the 551CPU in question every time the list of requests is updated this way or another 552(there may be other requests coming from kernel code in that list). 553 554CPU idle time governors are expected to regard the minimum of the global 555effective ``PM_QOS_CPU_DMA_LATENCY`` class constraint and the effective 556resume latency constraint for the given CPU as the upper limit for the exit 557latency of the idle states they can select for that CPU. They should never 558select any idle states with exit latency beyond that limit. 559 560 561Idle States Control Via Kernel Command Line 562=========================================== 563 564In addition to the ``sysfs`` interface allowing individual idle states to be 565`disabled for individual CPUs <idle-states-representation_>`_, there are kernel 566command line parameters affecting CPU idle time management. 567 568The ``cpuidle.off=1`` kernel command line option can be used to disable the 569CPU idle time management entirely. It does not prevent the idle loop from 570running on idle CPUs, but it prevents the CPU idle time governors and drivers 571from being invoked. If it is added to the kernel command line, the idle loop 572will ask the hardware to enter idle states on idle CPUs via the CPU architecture 573support code that is expected to provide a default mechanism for this purpose. 574That default mechanism usually is the least common denominator for all of the 575processors implementing the architecture (i.e. CPU instruction set) in question, 576however, so it is rather crude and not very energy-efficient. For this reason, 577it is not recommended for production use. 578 579The ``cpuidle.governor=`` kernel command line switch allows the ``CPUIdle`` 580governor to use to be specified. It has to be appended with a string matching 581the name of an available governor (e.g. ``cpuidle.governor=menu``) and that 582governor will be used instead of the default one. It is possible to force 583the ``menu`` governor to be used on the systems that use the ``ladder`` governor 584by default this way, for example. 585 586The other kernel command line parameters controlling CPU idle time management 587described below are only relevant for the *x86* architecture and some of 588them affect Intel processors only. 589 590The *x86* architecture support code recognizes three kernel command line 591options related to CPU idle time management: ``idle=poll``, ``idle=halt``, 592and ``idle=nomwait``. The first two of them disable the ``acpi_idle`` and 593``intel_idle`` drivers altogether, which effectively causes the entire 594``CPUIdle`` subsystem to be disabled and makes the idle loop invoke the 595architecture support code to deal with idle CPUs. How it does that depends on 596which of the two parameters is added to the kernel command line. In the 597``idle=halt`` case, the architecture support code will use the ``HLT`` 598instruction of the CPUs (which, as a rule, suspends the execution of the program 599and causes the hardware to attempt to enter the shallowest available idle state) 600for this purpose, and if ``idle=poll`` is used, idle CPUs will execute a 601more or less ``lightweight'' sequence of instructions in a tight loop. [Note 602that using ``idle=poll`` is somewhat drastic in many cases, as preventing idle 603CPUs from saving almost any energy at all may not be the only effect of it. 604For example, on Intel hardware it effectively prevents CPUs from using 605P-states (see |cpufreq|) that require any number of CPUs in a package to be 606idle, so it very well may hurt single-thread computations performance as well as 607energy-efficiency. Thus using it for performance reasons may not be a good idea 608at all.] 609 610The ``idle=nomwait`` option disables the ``intel_idle`` driver and causes 611``acpi_idle`` to be used (as long as all of the information needed by it is 612there in the system's ACPI tables), but it is not allowed to use the 613``MWAIT`` instruction of the CPUs to ask the hardware to enter idle states. 614 615In addition to the architecture-level kernel command line options affecting CPU 616idle time management, there are parameters affecting individual ``CPUIdle`` 617drivers that can be passed to them via the kernel command line. Specifically, 618the ``intel_idle.max_cstate=<n>`` and ``processor.max_cstate=<n>`` parameters, 619where ``<n>`` is an idle state index also used in the name of the given 620state's directory in ``sysfs`` (see 621`Representation of Idle States <idle-states-representation_>`_), causes the 622``intel_idle`` and ``acpi_idle`` drivers, respectively, to discard all of the 623idle states deeper than idle state ``<n>``. In that case, they will never ask 624for any of those idle states or expose them to the governor. [The behavior of 625the two drivers is different for ``<n>`` equal to ``0``. Adding 626``intel_idle.max_cstate=0`` to the kernel command line disables the 627``intel_idle`` driver and allows ``acpi_idle`` to be used, whereas 628``processor.max_cstate=0`` is equivalent to ``processor.max_cstate=1``. 629Also, the ``acpi_idle`` driver is part of the ``processor`` kernel module that 630can be loaded separately and ``max_cstate=<n>`` can be passed to it as a module 631parameter when it is loaded.] 632