1.. SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause) 2 3.. _napi: 4 5==== 6NAPI 7==== 8 9NAPI is the event handling mechanism used by the Linux networking stack. 10The name NAPI no longer stands for anything in particular [#]_. 11 12In basic operation the device notifies the host about new events 13via an interrupt. 14The host then schedules a NAPI instance to process the events. 15The device may also be polled for events via NAPI without receiving 16interrupts first (:ref:`busy polling<poll>`). 17 18NAPI processing usually happens in the software interrupt context, 19but there is an option to use :ref:`separate kernel threads<threaded>` 20for NAPI processing. 21 22All in all NAPI abstracts away from the drivers the context and configuration 23of event (packet Rx and Tx) processing. 24 25Driver API 26========== 27 28The two most important elements of NAPI are the struct napi_struct 29and the associated poll method. struct napi_struct holds the state 30of the NAPI instance while the method is the driver-specific event 31handler. The method will typically free Tx packets that have been 32transmitted and process newly received packets. 33 34.. _drv_ctrl: 35 36Control API 37----------- 38 39netif_napi_add() and netif_napi_del() add/remove a NAPI instance 40from the system. The instances are attached to the netdevice passed 41as argument (and will be deleted automatically when netdevice is 42unregistered). Instances are added in a disabled state. 43 44napi_enable() and napi_disable() manage the disabled state. 45A disabled NAPI can't be scheduled and its poll method is guaranteed 46to not be invoked. napi_disable() waits for ownership of the NAPI 47instance to be released. 48 49The control APIs are not idempotent. Control API calls are safe against 50concurrent use of datapath APIs but an incorrect sequence of control API 51calls may result in crashes, deadlocks, or race conditions. For example, 52calling napi_disable() multiple times in a row will deadlock. 53 54Datapath API 55------------ 56 57napi_schedule() is the basic method of scheduling a NAPI poll. 58Drivers should call this function in their interrupt handler 59(see :ref:`drv_sched` for more info). A successful call to napi_schedule() 60will take ownership of the NAPI instance. 61 62Later, after NAPI is scheduled, the driver's poll method will be 63called to process the events/packets. The method takes a ``budget`` 64argument - drivers can process completions for any number of Tx 65packets but should only process up to ``budget`` number of 66Rx packets. Rx processing is usually much more expensive. 67 68In other words, it is recommended to ignore the budget argument when 69performing TX buffer reclamation to ensure that the reclamation is not 70arbitrarily bounded; however, it is required to honor the budget argument 71for RX processing. 72 73.. warning:: 74 75 The ``budget`` argument may be 0 if core tries to only process Tx completions 76 and no Rx packets. 77 78The poll method returns the amount of work done. If the driver still 79has outstanding work to do (e.g. ``budget`` was exhausted) 80the poll method should return exactly ``budget``. In that case, 81the NAPI instance will be serviced/polled again (without the 82need to be scheduled). 83 84If event processing has been completed (all outstanding packets 85processed) the poll method should call napi_complete_done() 86before returning. napi_complete_done() releases the ownership 87of the instance. 88 89.. warning:: 90 91 The case of finishing all events and using exactly ``budget`` 92 must be handled carefully. There is no way to report this 93 (rare) condition to the stack, so the driver must either 94 not call napi_complete_done() and wait to be called again, 95 or return ``budget - 1``. 96 97 If the ``budget`` is 0 napi_complete_done() should never be called. 98 99Call sequence 100------------- 101 102Drivers should not make assumptions about the exact sequencing 103of calls. The poll method may be called without the driver scheduling 104the instance (unless the instance is disabled). Similarly, 105it's not guaranteed that the poll method will be called, even 106if napi_schedule() succeeded (e.g. if the instance gets disabled). 107 108As mentioned in the :ref:`drv_ctrl` section - napi_disable() and subsequent 109calls to the poll method only wait for the ownership of the instance 110to be released, not for the poll method to exit. This means that 111drivers should avoid accessing any data structures after calling 112napi_complete_done(). 113 114.. _drv_sched: 115 116Scheduling and IRQ masking 117-------------------------- 118 119Drivers should keep the interrupts masked after scheduling 120the NAPI instance - until NAPI polling finishes any further 121interrupts are unnecessary. 122 123Drivers which have to mask the interrupts explicitly (as opposed 124to IRQ being auto-masked by the device) should use the napi_schedule_prep() 125and __napi_schedule() calls: 126 127.. code-block:: c 128 129 if (napi_schedule_prep(&v->napi)) { 130 mydrv_mask_rxtx_irq(v->idx); 131 /* schedule after masking to avoid races */ 132 __napi_schedule(&v->napi); 133 } 134 135IRQ should only be unmasked after a successful call to napi_complete_done(): 136 137.. code-block:: c 138 139 if (budget && napi_complete_done(&v->napi, work_done)) { 140 mydrv_unmask_rxtx_irq(v->idx); 141 return min(work_done, budget - 1); 142 } 143 144napi_schedule_irqoff() is a variant of napi_schedule() which takes advantage 145of guarantees given by being invoked in IRQ context (no need to 146mask interrupts). Note that PREEMPT_RT forces all interrupts 147to be threaded so the interrupt may need to be marked ``IRQF_NO_THREAD`` 148to avoid issues on real-time kernel configurations. 149 150Instance to queue mapping 151------------------------- 152 153Modern devices have multiple NAPI instances (struct napi_struct) per 154interface. There is no strong requirement on how the instances are 155mapped to queues and interrupts. NAPI is primarily a polling/processing 156abstraction without specific user-facing semantics. That said, most networking 157devices end up using NAPI in fairly similar ways. 158 159NAPI instances most often correspond 1:1:1 to interrupts and queue pairs 160(queue pair is a set of a single Rx and single Tx queue). 161 162In less common cases a NAPI instance may be used for multiple queues 163or Rx and Tx queues can be serviced by separate NAPI instances on a single 164core. Regardless of the queue assignment, however, there is usually still 165a 1:1 mapping between NAPI instances and interrupts. 166 167It's worth noting that the ethtool API uses a "channel" terminology where 168each channel can be either ``rx``, ``tx`` or ``combined``. It's not clear 169what constitutes a channel; the recommended interpretation is to understand 170a channel as an IRQ/NAPI which services queues of a given type. For example, 171a configuration of 1 ``rx``, 1 ``tx`` and 1 ``combined`` channel is expected 172to utilize 3 interrupts, 2 Rx and 2 Tx queues. 173 174User API 175======== 176 177User interactions with NAPI depend on NAPI instance ID. The instance IDs 178are only visible to the user thru the ``SO_INCOMING_NAPI_ID`` socket option. 179It's not currently possible to query IDs used by a given device. 180 181Software IRQ coalescing 182----------------------- 183 184NAPI does not perform any explicit event coalescing by default. 185In most scenarios batching happens due to IRQ coalescing which is done 186by the device. There are cases where software coalescing is helpful. 187 188NAPI can be configured to arm a repoll timer instead of unmasking 189the hardware interrupts as soon as all packets are processed. 190The ``gro_flush_timeout`` sysfs configuration of the netdevice 191is reused to control the delay of the timer, while 192``napi_defer_hard_irqs`` controls the number of consecutive empty polls 193before NAPI gives up and goes back to using hardware IRQs. 194 195.. _poll: 196 197Busy polling 198------------ 199 200Busy polling allows a user process to check for incoming packets before 201the device interrupt fires. As is the case with any busy polling it trades 202off CPU cycles for lower latency (production uses of NAPI busy polling 203are not well known). 204 205Busy polling is enabled by either setting ``SO_BUSY_POLL`` on 206selected sockets or using the global ``net.core.busy_poll`` and 207``net.core.busy_read`` sysctls. An io_uring API for NAPI busy polling 208also exists. 209 210IRQ mitigation 211--------------- 212 213While busy polling is supposed to be used by low latency applications, 214a similar mechanism can be used for IRQ mitigation. 215 216Very high request-per-second applications (especially routing/forwarding 217applications and especially applications using AF_XDP sockets) may not 218want to be interrupted until they finish processing a request or a batch 219of packets. 220 221Such applications can pledge to the kernel that they will perform a busy 222polling operation periodically, and the driver should keep the device IRQs 223permanently masked. This mode is enabled by using the ``SO_PREFER_BUSY_POLL`` 224socket option. To avoid system misbehavior the pledge is revoked 225if ``gro_flush_timeout`` passes without any busy poll call. 226 227The NAPI budget for busy polling is lower than the default (which makes 228sense given the low latency intention of normal busy polling). This is 229not the case with IRQ mitigation, however, so the budget can be adjusted 230with the ``SO_BUSY_POLL_BUDGET`` socket option. 231 232.. _threaded: 233 234Threaded NAPI 235------------- 236 237Threaded NAPI is an operating mode that uses dedicated kernel 238threads rather than software IRQ context for NAPI processing. 239The configuration is per netdevice and will affect all 240NAPI instances of that device. Each NAPI instance will spawn a separate 241thread (called ``napi/${ifc-name}-${napi-id}``). 242 243It is recommended to pin each kernel thread to a single CPU, the same 244CPU as the CPU which services the interrupt. Note that the mapping 245between IRQs and NAPI instances may not be trivial (and is driver 246dependent). The NAPI instance IDs will be assigned in the opposite 247order than the process IDs of the kernel threads. 248 249Threaded NAPI is controlled by writing 0/1 to the ``threaded`` file in 250netdev's sysfs directory. 251 252.. rubric:: Footnotes 253 254.. [#] NAPI was originally referred to as New API in 2.4 Linux. 255