1Runtime locking correctness validator 2===================================== 3 4started by Ingo Molnar <mingo@redhat.com> 5 6additions by Arjan van de Ven <arjan@linux.intel.com> 7 8Lock-class 9---------- 10 11The basic object the validator operates upon is a 'class' of locks. 12 13A class of locks is a group of locks that are logically the same with 14respect to locking rules, even if the locks may have multiple (possibly 15tens of thousands of) instantiations. For example a lock in the inode 16struct is one class, while each inode has its own instantiation of that 17lock class. 18 19The validator tracks the 'usage state' of lock-classes, and it tracks 20the dependencies between different lock-classes. Lock usage indicates 21how a lock is used with regard to its IRQ contexts, while lock 22dependency can be understood as lock order, where L1 -> L2 suggests that 23a task is attempting to acquire L2 while holding L1. From lockdep's 24perspective, the two locks (L1 and L2) are not necessarily related; that 25dependency just means the order ever happened. The validator maintains a 26continuing effort to prove lock usages and dependencies are correct or 27the validator will shoot a splat if incorrect. 28 29A lock-class's behavior is constructed by its instances collectively: 30when the first instance of a lock-class is used after bootup the class 31gets registered, then all (subsequent) instances will be mapped to the 32class and hence their usages and dependecies will contribute to those of 33the class. A lock-class does not go away when a lock instance does, but 34it can be removed if the memory space of the lock class (static or 35dynamic) is reclaimed, this happens for example when a module is 36unloaded or a workqueue is destroyed. 37 38State 39----- 40 41The validator tracks lock-class usage history and divides the usage into 42(4 usages * n STATEs + 1) categories: 43 44where the 4 usages can be: 45- 'ever held in STATE context' 46- 'ever held as readlock in STATE context' 47- 'ever held with STATE enabled' 48- 'ever held as readlock with STATE enabled' 49 50where the n STATEs are coded in kernel/locking/lockdep_states.h and as of 51now they include: 52- hardirq 53- softirq 54 55where the last 1 category is: 56- 'ever used' [ == !unused ] 57 58When locking rules are violated, these usage bits are presented in the 59locking error messages, inside curlies, with a total of 2 * n STATEs bits. 60A contrived example:: 61 62 modprobe/2287 is trying to acquire lock: 63 (&sio_locks[i].lock){-.-.}, at: [<c02867fd>] mutex_lock+0x21/0x24 64 65 but task is already holding lock: 66 (&sio_locks[i].lock){-.-.}, at: [<c02867fd>] mutex_lock+0x21/0x24 67 68 69For a given lock, the bit positions from left to right indicate the usage 70of the lock and readlock (if exists), for each of the n STATEs listed 71above respectively, and the character displayed at each bit position 72indicates: 73 74 === =================================================== 75 '.' acquired while irqs disabled and not in irq context 76 '-' acquired in irq context 77 '+' acquired with irqs enabled 78 '?' acquired in irq context with irqs enabled. 79 === =================================================== 80 81The bits are illustrated with an example:: 82 83 (&sio_locks[i].lock){-.-.}, at: [<c02867fd>] mutex_lock+0x21/0x24 84 |||| 85 ||| \-> softirq disabled and not in softirq context 86 || \--> acquired in softirq context 87 | \---> hardirq disabled and not in hardirq context 88 \----> acquired in hardirq context 89 90 91For a given STATE, whether the lock is ever acquired in that STATE 92context and whether that STATE is enabled yields four possible cases as 93shown in the table below. The bit character is able to indicate which 94exact case is for the lock as of the reporting time. 95 96 +--------------+-------------+--------------+ 97 | | irq enabled | irq disabled | 98 +--------------+-------------+--------------+ 99 | ever in irq | ? | - | 100 +--------------+-------------+--------------+ 101 | never in irq | + | . | 102 +--------------+-------------+--------------+ 103 104The character '-' suggests irq is disabled because if otherwise the 105charactor '?' would have been shown instead. Similar deduction can be 106applied for '+' too. 107 108Unused locks (e.g., mutexes) cannot be part of the cause of an error. 109 110 111Single-lock state rules: 112------------------------ 113 114A lock is irq-safe means it was ever used in an irq context, while a lock 115is irq-unsafe means it was ever acquired with irq enabled. 116 117A softirq-unsafe lock-class is automatically hardirq-unsafe as well. The 118following states must be exclusive: only one of them is allowed to be set 119for any lock-class based on its usage:: 120 121 <hardirq-safe> or <hardirq-unsafe> 122 <softirq-safe> or <softirq-unsafe> 123 124This is because if a lock can be used in irq context (irq-safe) then it 125cannot be ever acquired with irq enabled (irq-unsafe). Otherwise, a 126deadlock may happen. For example, in the scenario that after this lock 127was acquired but before released, if the context is interrupted this 128lock will be attempted to acquire twice, which creates a deadlock, 129referred to as lock recursion deadlock. 130 131The validator detects and reports lock usage that violates these 132single-lock state rules. 133 134Multi-lock dependency rules: 135---------------------------- 136 137The same lock-class must not be acquired twice, because this could lead 138to lock recursion deadlocks. 139 140Furthermore, two locks can not be taken in inverse order:: 141 142 <L1> -> <L2> 143 <L2> -> <L1> 144 145because this could lead to a deadlock - referred to as lock inversion 146deadlock - as attempts to acquire the two locks form a circle which 147could lead to the two contexts waiting for each other permanently. The 148validator will find such dependency circle in arbitrary complexity, 149i.e., there can be any other locking sequence between the acquire-lock 150operations; the validator will still find whether these locks can be 151acquired in a circular fashion. 152 153Furthermore, the following usage based lock dependencies are not allowed 154between any two lock-classes:: 155 156 <hardirq-safe> -> <hardirq-unsafe> 157 <softirq-safe> -> <softirq-unsafe> 158 159The first rule comes from the fact that a hardirq-safe lock could be 160taken by a hardirq context, interrupting a hardirq-unsafe lock - and 161thus could result in a lock inversion deadlock. Likewise, a softirq-safe 162lock could be taken by an softirq context, interrupting a softirq-unsafe 163lock. 164 165The above rules are enforced for any locking sequence that occurs in the 166kernel: when acquiring a new lock, the validator checks whether there is 167any rule violation between the new lock and any of the held locks. 168 169When a lock-class changes its state, the following aspects of the above 170dependency rules are enforced: 171 172- if a new hardirq-safe lock is discovered, we check whether it 173 took any hardirq-unsafe lock in the past. 174 175- if a new softirq-safe lock is discovered, we check whether it took 176 any softirq-unsafe lock in the past. 177 178- if a new hardirq-unsafe lock is discovered, we check whether any 179 hardirq-safe lock took it in the past. 180 181- if a new softirq-unsafe lock is discovered, we check whether any 182 softirq-safe lock took it in the past. 183 184(Again, we do these checks too on the basis that an interrupt context 185could interrupt _any_ of the irq-unsafe or hardirq-unsafe locks, which 186could lead to a lock inversion deadlock - even if that lock scenario did 187not trigger in practice yet.) 188 189Exception: Nested data dependencies leading to nested locking 190------------------------------------------------------------- 191 192There are a few cases where the Linux kernel acquires more than one 193instance of the same lock-class. Such cases typically happen when there 194is some sort of hierarchy within objects of the same type. In these 195cases there is an inherent "natural" ordering between the two objects 196(defined by the properties of the hierarchy), and the kernel grabs the 197locks in this fixed order on each of the objects. 198 199An example of such an object hierarchy that results in "nested locking" 200is that of a "whole disk" block-dev object and a "partition" block-dev 201object; the partition is "part of" the whole device and as long as one 202always takes the whole disk lock as a higher lock than the partition 203lock, the lock ordering is fully correct. The validator does not 204automatically detect this natural ordering, as the locking rule behind 205the ordering is not static. 206 207In order to teach the validator about this correct usage model, new 208versions of the various locking primitives were added that allow you to 209specify a "nesting level". An example call, for the block device mutex, 210looks like this:: 211 212 enum bdev_bd_mutex_lock_class 213 { 214 BD_MUTEX_NORMAL, 215 BD_MUTEX_WHOLE, 216 BD_MUTEX_PARTITION 217 }; 218 219mutex_lock_nested(&bdev->bd_contains->bd_mutex, BD_MUTEX_PARTITION); 220 221In this case the locking is done on a bdev object that is known to be a 222partition. 223 224The validator treats a lock that is taken in such a nested fashion as a 225separate (sub)class for the purposes of validation. 226 227Note: When changing code to use the _nested() primitives, be careful and 228check really thoroughly that the hierarchy is correctly mapped; otherwise 229you can get false positives or false negatives. 230 231Annotations 232----------- 233 234Two constructs can be used to annotate and check where and if certain locks 235must be held: lockdep_assert_held*(&lock) and lockdep_*pin_lock(&lock). 236 237As the name suggests, lockdep_assert_held* family of macros assert that a 238particular lock is held at a certain time (and generate a WARN() otherwise). 239This annotation is largely used all over the kernel, e.g. kernel/sched/ 240core.c:: 241 242 void update_rq_clock(struct rq *rq) 243 { 244 s64 delta; 245 246 lockdep_assert_held(&rq->lock); 247 [...] 248 } 249 250where holding rq->lock is required to safely update a rq's clock. 251 252The other family of macros is lockdep_*pin_lock(), which is admittedly only 253used for rq->lock ATM. Despite their limited adoption these annotations 254generate a WARN() if the lock of interest is "accidentally" unlocked. This turns 255out to be especially helpful to debug code with callbacks, where an upper 256layer assumes a lock remains taken, but a lower layer thinks it can maybe drop 257and reacquire the lock ("unwittingly" introducing races). lockdep_pin_lock() 258returns a 'struct pin_cookie' that is then used by lockdep_unpin_lock() to check 259that nobody tampered with the lock, e.g. kernel/sched/sched.h:: 260 261 static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf) 262 { 263 rf->cookie = lockdep_pin_lock(&rq->lock); 264 [...] 265 } 266 267 static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf) 268 { 269 [...] 270 lockdep_unpin_lock(&rq->lock, rf->cookie); 271 } 272 273While comments about locking requirements might provide useful information, 274the runtime checks performed by annotations are invaluable when debugging 275locking problems and they carry the same level of details when inspecting 276code. Always prefer annotations when in doubt! 277 278Proof of 100% correctness: 279-------------------------- 280 281The validator achieves perfect, mathematical 'closure' (proof of locking 282correctness) in the sense that for every simple, standalone single-task 283locking sequence that occurred at least once during the lifetime of the 284kernel, the validator proves it with a 100% certainty that no 285combination and timing of these locking sequences can cause any class of 286lock related deadlock. [1]_ 287 288I.e. complex multi-CPU and multi-task locking scenarios do not have to 289occur in practice to prove a deadlock: only the simple 'component' 290locking chains have to occur at least once (anytime, in any 291task/context) for the validator to be able to prove correctness. (For 292example, complex deadlocks that would normally need more than 3 CPUs and 293a very unlikely constellation of tasks, irq-contexts and timings to 294occur, can be detected on a plain, lightly loaded single-CPU system as 295well!) 296 297This radically decreases the complexity of locking related QA of the 298kernel: what has to be done during QA is to trigger as many "simple" 299single-task locking dependencies in the kernel as possible, at least 300once, to prove locking correctness - instead of having to trigger every 301possible combination of locking interaction between CPUs, combined with 302every possible hardirq and softirq nesting scenario (which is impossible 303to do in practice). 304 305.. [1] 306 307 assuming that the validator itself is 100% correct, and no other 308 part of the system corrupts the state of the validator in any way. 309 We also assume that all NMI/SMM paths [which could interrupt 310 even hardirq-disabled codepaths] are correct and do not interfere 311 with the validator. We also assume that the 64-bit 'chain hash' 312 value is unique for every lock-chain in the system. Also, lock 313 recursion must not be higher than 20. 314 315Performance: 316------------ 317 318The above rules require **massive** amounts of runtime checking. If we did 319that for every lock taken and for every irqs-enable event, it would 320render the system practically unusably slow. The complexity of checking 321is O(N^2), so even with just a few hundred lock-classes we'd have to do 322tens of thousands of checks for every event. 323 324This problem is solved by checking any given 'locking scenario' (unique 325sequence of locks taken after each other) only once. A simple stack of 326held locks is maintained, and a lightweight 64-bit hash value is 327calculated, which hash is unique for every lock chain. The hash value, 328when the chain is validated for the first time, is then put into a hash 329table, which hash-table can be checked in a lockfree manner. If the 330locking chain occurs again later on, the hash table tells us that we 331don't have to validate the chain again. 332 333Troubleshooting: 334---------------- 335 336The validator tracks a maximum of MAX_LOCKDEP_KEYS number of lock classes. 337Exceeding this number will trigger the following lockdep warning: 338 339 (DEBUG_LOCKS_WARN_ON(id >= MAX_LOCKDEP_KEYS)) 340 341By default, MAX_LOCKDEP_KEYS is currently set to 8191, and typical 342desktop systems have less than 1,000 lock classes, so this warning 343normally results from lock-class leakage or failure to properly 344initialize locks. These two problems are illustrated below: 345 3461. Repeated module loading and unloading while running the validator 347 will result in lock-class leakage. The issue here is that each 348 load of the module will create a new set of lock classes for 349 that module's locks, but module unloading does not remove old 350 classes (see below discussion of reuse of lock classes for why). 351 Therefore, if that module is loaded and unloaded repeatedly, 352 the number of lock classes will eventually reach the maximum. 353 3542. Using structures such as arrays that have large numbers of 355 locks that are not explicitly initialized. For example, 356 a hash table with 8192 buckets where each bucket has its own 357 spinlock_t will consume 8192 lock classes -unless- each spinlock 358 is explicitly initialized at runtime, for example, using the 359 run-time spin_lock_init() as opposed to compile-time initializers 360 such as __SPIN_LOCK_UNLOCKED(). Failure to properly initialize 361 the per-bucket spinlocks would guarantee lock-class overflow. 362 In contrast, a loop that called spin_lock_init() on each lock 363 would place all 8192 locks into a single lock class. 364 365 The moral of this story is that you should always explicitly 366 initialize your locks. 367 368One might argue that the validator should be modified to allow 369lock classes to be reused. However, if you are tempted to make this 370argument, first review the code and think through the changes that would 371be required, keeping in mind that the lock classes to be removed are 372likely to be linked into the lock-dependency graph. This turns out to 373be harder to do than to say. 374 375Of course, if you do run out of lock classes, the next thing to do is 376to find the offending lock classes. First, the following command gives 377you the number of lock classes currently in use along with the maximum:: 378 379 grep "lock-classes" /proc/lockdep_stats 380 381This command produces the following output on a modest system:: 382 383 lock-classes: 748 [max: 8191] 384 385If the number allocated (748 above) increases continually over time, 386then there is likely a leak. The following command can be used to 387identify the leaking lock classes:: 388 389 grep "BD" /proc/lockdep 390 391Run the command and save the output, then compare against the output from 392a later run of this command to identify the leakers. This same output 393can also help you find situations where runtime lock initialization has 394been omitted. 395