1Runtime locking correctness validator
2=====================================
3
4started by Ingo Molnar <mingo@redhat.com>
5
6additions by Arjan van de Ven <arjan@linux.intel.com>
7
8Lock-class
9----------
10
11The basic object the validator operates upon is a 'class' of locks.
12
13A class of locks is a group of locks that are logically the same with
14respect to locking rules, even if the locks may have multiple (possibly
15tens of thousands of) instantiations. For example a lock in the inode
16struct is one class, while each inode has its own instantiation of that
17lock class.
18
19The validator tracks the 'usage state' of lock-classes, and it tracks
20the dependencies between different lock-classes. Lock usage indicates
21how a lock is used with regard to its IRQ contexts, while lock
22dependency can be understood as lock order, where L1 -> L2 suggests that
23a task is attempting to acquire L2 while holding L1. From lockdep's
24perspective, the two locks (L1 and L2) are not necessarily related; that
25dependency just means the order ever happened. The validator maintains a
26continuing effort to prove lock usages and dependencies are correct or
27the validator will shoot a splat if incorrect.
28
29A lock-class's behavior is constructed by its instances collectively:
30when the first instance of a lock-class is used after bootup the class
31gets registered, then all (subsequent) instances will be mapped to the
32class and hence their usages and dependecies will contribute to those of
33the class. A lock-class does not go away when a lock instance does, but
34it can be removed if the memory space of the lock class (static or
35dynamic) is reclaimed, this happens for example when a module is
36unloaded or a workqueue is destroyed.
37
38State
39-----
40
41The validator tracks lock-class usage history and divides the usage into
42(4 usages * n STATEs + 1) categories:
43
44where the 4 usages can be:
45- 'ever held in STATE context'
46- 'ever held as readlock in STATE context'
47- 'ever held with STATE enabled'
48- 'ever held as readlock with STATE enabled'
49
50where the n STATEs are coded in kernel/locking/lockdep_states.h and as of
51now they include:
52- hardirq
53- softirq
54
55where the last 1 category is:
56- 'ever used'                                       [ == !unused        ]
57
58When locking rules are violated, these usage bits are presented in the
59locking error messages, inside curlies, with a total of 2 * n STATEs bits.
60A contrived example::
61
62   modprobe/2287 is trying to acquire lock:
63    (&sio_locks[i].lock){-.-.}, at: [<c02867fd>] mutex_lock+0x21/0x24
64
65   but task is already holding lock:
66    (&sio_locks[i].lock){-.-.}, at: [<c02867fd>] mutex_lock+0x21/0x24
67
68
69For a given lock, the bit positions from left to right indicate the usage
70of the lock and readlock (if exists), for each of the n STATEs listed
71above respectively, and the character displayed at each bit position
72indicates:
73
74   ===  ===================================================
75   '.'  acquired while irqs disabled and not in irq context
76   '-'  acquired in irq context
77   '+'  acquired with irqs enabled
78   '?'  acquired in irq context with irqs enabled.
79   ===  ===================================================
80
81The bits are illustrated with an example::
82
83    (&sio_locks[i].lock){-.-.}, at: [<c02867fd>] mutex_lock+0x21/0x24
84                         ||||
85                         ||| \-> softirq disabled and not in softirq context
86                         || \--> acquired in softirq context
87                         | \---> hardirq disabled and not in hardirq context
88                          \----> acquired in hardirq context
89
90
91For a given STATE, whether the lock is ever acquired in that STATE
92context and whether that STATE is enabled yields four possible cases as
93shown in the table below. The bit character is able to indicate which
94exact case is for the lock as of the reporting time.
95
96  +--------------+-------------+--------------+
97  |              | irq enabled | irq disabled |
98  +--------------+-------------+--------------+
99  | ever in irq  |      ?      |       -      |
100  +--------------+-------------+--------------+
101  | never in irq |      +      |       .      |
102  +--------------+-------------+--------------+
103
104The character '-' suggests irq is disabled because if otherwise the
105charactor '?' would have been shown instead. Similar deduction can be
106applied for '+' too.
107
108Unused locks (e.g., mutexes) cannot be part of the cause of an error.
109
110
111Single-lock state rules:
112------------------------
113
114A lock is irq-safe means it was ever used in an irq context, while a lock
115is irq-unsafe means it was ever acquired with irq enabled.
116
117A softirq-unsafe lock-class is automatically hardirq-unsafe as well. The
118following states must be exclusive: only one of them is allowed to be set
119for any lock-class based on its usage::
120
121 <hardirq-safe> or <hardirq-unsafe>
122 <softirq-safe> or <softirq-unsafe>
123
124This is because if a lock can be used in irq context (irq-safe) then it
125cannot be ever acquired with irq enabled (irq-unsafe). Otherwise, a
126deadlock may happen. For example, in the scenario that after this lock
127was acquired but before released, if the context is interrupted this
128lock will be attempted to acquire twice, which creates a deadlock,
129referred to as lock recursion deadlock.
130
131The validator detects and reports lock usage that violates these
132single-lock state rules.
133
134Multi-lock dependency rules:
135----------------------------
136
137The same lock-class must not be acquired twice, because this could lead
138to lock recursion deadlocks.
139
140Furthermore, two locks can not be taken in inverse order::
141
142 <L1> -> <L2>
143 <L2> -> <L1>
144
145because this could lead to a deadlock - referred to as lock inversion
146deadlock - as attempts to acquire the two locks form a circle which
147could lead to the two contexts waiting for each other permanently. The
148validator will find such dependency circle in arbitrary complexity,
149i.e., there can be any other locking sequence between the acquire-lock
150operations; the validator will still find whether these locks can be
151acquired in a circular fashion.
152
153Furthermore, the following usage based lock dependencies are not allowed
154between any two lock-classes::
155
156   <hardirq-safe>   ->  <hardirq-unsafe>
157   <softirq-safe>   ->  <softirq-unsafe>
158
159The first rule comes from the fact that a hardirq-safe lock could be
160taken by a hardirq context, interrupting a hardirq-unsafe lock - and
161thus could result in a lock inversion deadlock. Likewise, a softirq-safe
162lock could be taken by an softirq context, interrupting a softirq-unsafe
163lock.
164
165The above rules are enforced for any locking sequence that occurs in the
166kernel: when acquiring a new lock, the validator checks whether there is
167any rule violation between the new lock and any of the held locks.
168
169When a lock-class changes its state, the following aspects of the above
170dependency rules are enforced:
171
172- if a new hardirq-safe lock is discovered, we check whether it
173  took any hardirq-unsafe lock in the past.
174
175- if a new softirq-safe lock is discovered, we check whether it took
176  any softirq-unsafe lock in the past.
177
178- if a new hardirq-unsafe lock is discovered, we check whether any
179  hardirq-safe lock took it in the past.
180
181- if a new softirq-unsafe lock is discovered, we check whether any
182  softirq-safe lock took it in the past.
183
184(Again, we do these checks too on the basis that an interrupt context
185could interrupt _any_ of the irq-unsafe or hardirq-unsafe locks, which
186could lead to a lock inversion deadlock - even if that lock scenario did
187not trigger in practice yet.)
188
189Exception: Nested data dependencies leading to nested locking
190-------------------------------------------------------------
191
192There are a few cases where the Linux kernel acquires more than one
193instance of the same lock-class. Such cases typically happen when there
194is some sort of hierarchy within objects of the same type. In these
195cases there is an inherent "natural" ordering between the two objects
196(defined by the properties of the hierarchy), and the kernel grabs the
197locks in this fixed order on each of the objects.
198
199An example of such an object hierarchy that results in "nested locking"
200is that of a "whole disk" block-dev object and a "partition" block-dev
201object; the partition is "part of" the whole device and as long as one
202always takes the whole disk lock as a higher lock than the partition
203lock, the lock ordering is fully correct. The validator does not
204automatically detect this natural ordering, as the locking rule behind
205the ordering is not static.
206
207In order to teach the validator about this correct usage model, new
208versions of the various locking primitives were added that allow you to
209specify a "nesting level". An example call, for the block device mutex,
210looks like this::
211
212  enum bdev_bd_mutex_lock_class
213  {
214       BD_MUTEX_NORMAL,
215       BD_MUTEX_WHOLE,
216       BD_MUTEX_PARTITION
217  };
218
219mutex_lock_nested(&bdev->bd_contains->bd_mutex, BD_MUTEX_PARTITION);
220
221In this case the locking is done on a bdev object that is known to be a
222partition.
223
224The validator treats a lock that is taken in such a nested fashion as a
225separate (sub)class for the purposes of validation.
226
227Note: When changing code to use the _nested() primitives, be careful and
228check really thoroughly that the hierarchy is correctly mapped; otherwise
229you can get false positives or false negatives.
230
231Annotations
232-----------
233
234Two constructs can be used to annotate and check where and if certain locks
235must be held: lockdep_assert_held*(&lock) and lockdep_*pin_lock(&lock).
236
237As the name suggests, lockdep_assert_held* family of macros assert that a
238particular lock is held at a certain time (and generate a WARN() otherwise).
239This annotation is largely used all over the kernel, e.g. kernel/sched/
240core.c::
241
242  void update_rq_clock(struct rq *rq)
243  {
244	s64 delta;
245
246	lockdep_assert_held(&rq->lock);
247	[...]
248  }
249
250where holding rq->lock is required to safely update a rq's clock.
251
252The other family of macros is lockdep_*pin_lock(), which is admittedly only
253used for rq->lock ATM. Despite their limited adoption these annotations
254generate a WARN() if the lock of interest is "accidentally" unlocked. This turns
255out to be especially helpful to debug code with callbacks, where an upper
256layer assumes a lock remains taken, but a lower layer thinks it can maybe drop
257and reacquire the lock ("unwittingly" introducing races). lockdep_pin_lock()
258returns a 'struct pin_cookie' that is then used by lockdep_unpin_lock() to check
259that nobody tampered with the lock, e.g. kernel/sched/sched.h::
260
261  static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf)
262  {
263	rf->cookie = lockdep_pin_lock(&rq->lock);
264	[...]
265  }
266
267  static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf)
268  {
269	[...]
270	lockdep_unpin_lock(&rq->lock, rf->cookie);
271  }
272
273While comments about locking requirements might provide useful information,
274the runtime checks performed by annotations are invaluable when debugging
275locking problems and they carry the same level of details when inspecting
276code.  Always prefer annotations when in doubt!
277
278Proof of 100% correctness:
279--------------------------
280
281The validator achieves perfect, mathematical 'closure' (proof of locking
282correctness) in the sense that for every simple, standalone single-task
283locking sequence that occurred at least once during the lifetime of the
284kernel, the validator proves it with a 100% certainty that no
285combination and timing of these locking sequences can cause any class of
286lock related deadlock. [1]_
287
288I.e. complex multi-CPU and multi-task locking scenarios do not have to
289occur in practice to prove a deadlock: only the simple 'component'
290locking chains have to occur at least once (anytime, in any
291task/context) for the validator to be able to prove correctness. (For
292example, complex deadlocks that would normally need more than 3 CPUs and
293a very unlikely constellation of tasks, irq-contexts and timings to
294occur, can be detected on a plain, lightly loaded single-CPU system as
295well!)
296
297This radically decreases the complexity of locking related QA of the
298kernel: what has to be done during QA is to trigger as many "simple"
299single-task locking dependencies in the kernel as possible, at least
300once, to prove locking correctness - instead of having to trigger every
301possible combination of locking interaction between CPUs, combined with
302every possible hardirq and softirq nesting scenario (which is impossible
303to do in practice).
304
305.. [1]
306
307    assuming that the validator itself is 100% correct, and no other
308    part of the system corrupts the state of the validator in any way.
309    We also assume that all NMI/SMM paths [which could interrupt
310    even hardirq-disabled codepaths] are correct and do not interfere
311    with the validator. We also assume that the 64-bit 'chain hash'
312    value is unique for every lock-chain in the system. Also, lock
313    recursion must not be higher than 20.
314
315Performance:
316------------
317
318The above rules require **massive** amounts of runtime checking. If we did
319that for every lock taken and for every irqs-enable event, it would
320render the system practically unusably slow. The complexity of checking
321is O(N^2), so even with just a few hundred lock-classes we'd have to do
322tens of thousands of checks for every event.
323
324This problem is solved by checking any given 'locking scenario' (unique
325sequence of locks taken after each other) only once. A simple stack of
326held locks is maintained, and a lightweight 64-bit hash value is
327calculated, which hash is unique for every lock chain. The hash value,
328when the chain is validated for the first time, is then put into a hash
329table, which hash-table can be checked in a lockfree manner. If the
330locking chain occurs again later on, the hash table tells us that we
331don't have to validate the chain again.
332
333Troubleshooting:
334----------------
335
336The validator tracks a maximum of MAX_LOCKDEP_KEYS number of lock classes.
337Exceeding this number will trigger the following lockdep warning:
338
339	(DEBUG_LOCKS_WARN_ON(id >= MAX_LOCKDEP_KEYS))
340
341By default, MAX_LOCKDEP_KEYS is currently set to 8191, and typical
342desktop systems have less than 1,000 lock classes, so this warning
343normally results from lock-class leakage or failure to properly
344initialize locks.  These two problems are illustrated below:
345
3461.	Repeated module loading and unloading while running the validator
347	will result in lock-class leakage.  The issue here is that each
348	load of the module will create a new set of lock classes for
349	that module's locks, but module unloading does not remove old
350	classes (see below discussion of reuse of lock classes for why).
351	Therefore, if that module is loaded and unloaded repeatedly,
352	the number of lock classes will eventually reach the maximum.
353
3542.	Using structures such as arrays that have large numbers of
355	locks that are not explicitly initialized.  For example,
356	a hash table with 8192 buckets where each bucket has its own
357	spinlock_t will consume 8192 lock classes -unless- each spinlock
358	is explicitly initialized at runtime, for example, using the
359	run-time spin_lock_init() as opposed to compile-time initializers
360	such as __SPIN_LOCK_UNLOCKED().  Failure to properly initialize
361	the per-bucket spinlocks would guarantee lock-class overflow.
362	In contrast, a loop that called spin_lock_init() on each lock
363	would place all 8192 locks into a single lock class.
364
365	The moral of this story is that you should always explicitly
366	initialize your locks.
367
368One might argue that the validator should be modified to allow
369lock classes to be reused.  However, if you are tempted to make this
370argument, first review the code and think through the changes that would
371be required, keeping in mind that the lock classes to be removed are
372likely to be linked into the lock-dependency graph.  This turns out to
373be harder to do than to say.
374
375Of course, if you do run out of lock classes, the next thing to do is
376to find the offending lock classes.  First, the following command gives
377you the number of lock classes currently in use along with the maximum::
378
379	grep "lock-classes" /proc/lockdep_stats
380
381This command produces the following output on a modest system::
382
383	lock-classes:                          748 [max: 8191]
384
385If the number allocated (748 above) increases continually over time,
386then there is likely a leak.  The following command can be used to
387identify the leaking lock classes::
388
389	grep "BD" /proc/lockdep
390
391Run the command and save the output, then compare against the output from
392a later run of this command to identify the leakers.  This same output
393can also help you find situations where runtime lock initialization has
394been omitted.
395
396Recursive read locks:
397---------------------
398The whole of the rest document tries to prove a certain type of cycle is equivalent
399to deadlock possibility.
400
401There are three types of lockers: writers (i.e. exclusive lockers, like
402spin_lock() or write_lock()), non-recursive readers (i.e. shared lockers, like
403down_read()) and recursive readers (recursive shared lockers, like rcu_read_lock()).
404And we use the following notations of those lockers in the rest of the document:
405
406	W or E:	stands for writers (exclusive lockers).
407	r:	stands for non-recursive readers.
408	R:	stands for recursive readers.
409	S:	stands for all readers (non-recursive + recursive), as both are shared lockers.
410	N:	stands for writers and non-recursive readers, as both are not recursive.
411
412Obviously, N is "r or W" and S is "r or R".
413
414Recursive readers, as their name indicates, are the lockers allowed to acquire
415even inside the critical section of another reader of the same lock instance,
416in other words, allowing nested read-side critical sections of one lock instance.
417
418While non-recursive readers will cause a self deadlock if trying to acquire inside
419the critical section of another reader of the same lock instance.
420
421The difference between recursive readers and non-recursive readers is because:
422recursive readers get blocked only by a write lock *holder*, while non-recursive
423readers could get blocked by a write lock *waiter*. Considering the follow example:
424
425	TASK A:			TASK B:
426
427	read_lock(X);
428				write_lock(X);
429	read_lock_2(X);
430
431Task A gets the reader (no matter whether recursive or non-recursive) on X via
432read_lock() first. And when task B tries to acquire writer on X, it will block
433and become a waiter for writer on X. Now if read_lock_2() is recursive readers,
434task A will make progress, because writer waiters don't block recursive readers,
435and there is no deadlock. However, if read_lock_2() is non-recursive readers,
436it will get blocked by writer waiter B, and cause a self deadlock.
437
438Block conditions on readers/writers of the same lock instance:
439--------------------------------------------------------------
440There are simply four block conditions:
441
4421.	Writers block other writers.
4432.	Readers block writers.
4443.	Writers block both recursive readers and non-recursive readers.
4454.	And readers (recursive or not) don't block other recursive readers but
446	may block non-recursive readers (because of the potential co-existing
447	writer waiters)
448
449Block condition matrix, Y means the row blocks the column, and N means otherwise.
450
451	    | E | r | R |
452	+---+---+---+---+
453	  E | Y | Y | Y |
454	+---+---+---+---+
455	  r | Y | Y | N |
456	+---+---+---+---+
457	  R | Y | Y | N |
458
459	(W: writers, r: non-recursive readers, R: recursive readers)
460
461
462acquired recursively. Unlike non-recursive read locks, recursive read locks
463only get blocked by current write lock *holders* other than write lock
464*waiters*, for example:
465
466	TASK A:			TASK B:
467
468	read_lock(X);
469
470				write_lock(X);
471
472	read_lock(X);
473
474is not a deadlock for recursive read locks, as while the task B is waiting for
475the lock X, the second read_lock() doesn't need to wait because it's a recursive
476read lock. However if the read_lock() is non-recursive read lock, then the above
477case is a deadlock, because even if the write_lock() in TASK B cannot get the
478lock, but it can block the second read_lock() in TASK A.
479
480Note that a lock can be a write lock (exclusive lock), a non-recursive read
481lock (non-recursive shared lock) or a recursive read lock (recursive shared
482lock), depending on the lock operations used to acquire it (more specifically,
483the value of the 'read' parameter for lock_acquire()). In other words, a single
484lock instance has three types of acquisition depending on the acquisition
485functions: exclusive, non-recursive read, and recursive read.
486
487To be concise, we call that write locks and non-recursive read locks as
488"non-recursive" locks and recursive read locks as "recursive" locks.
489
490Recursive locks don't block each other, while non-recursive locks do (this is
491even true for two non-recursive read locks). A non-recursive lock can block the
492corresponding recursive lock, and vice versa.
493
494A deadlock case with recursive locks involved is as follow:
495
496	TASK A:			TASK B:
497
498	read_lock(X);
499				read_lock(Y);
500	write_lock(Y);
501				write_lock(X);
502
503Task A is waiting for task B to read_unlock() Y and task B is waiting for task
504A to read_unlock() X.
505
506Dependency types and strong dependency paths:
507---------------------------------------------
508Lock dependencies record the orders of the acquisitions of a pair of locks, and
509because there are 3 types for lockers, there are, in theory, 9 types of lock
510dependencies, but we can show that 4 types of lock dependencies are enough for
511deadlock detection.
512
513For each lock dependency:
514
515	L1 -> L2
516
517, which means lockdep has seen L1 held before L2 held in the same context at runtime.
518And in deadlock detection, we care whether we could get blocked on L2 with L1 held,
519IOW, whether there is a locker L3 that L1 blocks L3 and L2 gets blocked by L3. So
520we only care about 1) what L1 blocks and 2) what blocks L2. As a result, we can combine
521recursive readers and non-recursive readers for L1 (as they block the same types) and
522we can combine writers and non-recursive readers for L2 (as they get blocked by the
523same types).
524
525With the above combination for simplification, there are 4 types of dependency edges
526in the lockdep graph:
527
5281) -(ER)->: exclusive writer to recursive reader dependency, "X -(ER)-> Y" means
529	    X -> Y and X is a writer and Y is a recursive reader.
530
5312) -(EN)->: exclusive writer to non-recursive locker dependency, "X -(EN)-> Y" means
532	    X -> Y and X is a writer and Y is either a writer or non-recursive reader.
533
5343) -(SR)->: shared reader to recursive reader dependency, "X -(SR)-> Y" means
535	    X -> Y and X is a reader (recursive or not) and Y is a recursive reader.
536
5374) -(SN)->: shared reader to non-recursive locker dependency, "X -(SN)-> Y" means
538	    X -> Y and X is a reader (recursive or not) and Y is either a writer or
539	    non-recursive reader.
540
541Note that given two locks, they may have multiple dependencies between them, for example:
542
543	TASK A:
544
545	read_lock(X);
546	write_lock(Y);
547	...
548
549	TASK B:
550
551	write_lock(X);
552	write_lock(Y);
553
554, we have both X -(SN)-> Y and X -(EN)-> Y in the dependency graph.
555
556We use -(xN)-> to represent edges that are either -(EN)-> or -(SN)->, the
557similar for -(Ex)->, -(xR)-> and -(Sx)->
558
559A "path" is a series of conjunct dependency edges in the graph. And we define a
560"strong" path, which indicates the strong dependency throughout each dependency
561in the path, as the path that doesn't have two conjunct edges (dependencies) as
562-(xR)-> and -(Sx)->. In other words, a "strong" path is a path from a lock
563walking to another through the lock dependencies, and if X -> Y -> Z is in the
564path (where X, Y, Z are locks), and the walk from X to Y is through a -(SR)-> or
565-(ER)-> dependency, the walk from Y to Z must not be through a -(SN)-> or
566-(SR)-> dependency.
567
568We will see why the path is called "strong" in next section.
569
570Recursive Read Deadlock Detection:
571----------------------------------
572
573We now prove two things:
574
575Lemma 1:
576
577If there is a closed strong path (i.e. a strong circle), then there is a
578combination of locking sequences that causes deadlock. I.e. a strong circle is
579sufficient for deadlock detection.
580
581Lemma 2:
582
583If there is no closed strong path (i.e. strong circle), then there is no
584combination of locking sequences that could cause deadlock. I.e.  strong
585circles are necessary for deadlock detection.
586
587With these two Lemmas, we can easily say a closed strong path is both sufficient
588and necessary for deadlocks, therefore a closed strong path is equivalent to
589deadlock possibility. As a closed strong path stands for a dependency chain that
590could cause deadlocks, so we call it "strong", considering there are dependency
591circles that won't cause deadlocks.
592
593Proof for sufficiency (Lemma 1):
594
595Let's say we have a strong circle:
596
597	L1 -> L2 ... -> Ln -> L1
598
599, which means we have dependencies:
600
601	L1 -> L2
602	L2 -> L3
603	...
604	Ln-1 -> Ln
605	Ln -> L1
606
607We now can construct a combination of locking sequences that cause deadlock:
608
609Firstly let's make one CPU/task get the L1 in L1 -> L2, and then another get
610the L2 in L2 -> L3, and so on. After this, all of the Lx in Lx -> Lx+1 are
611held by different CPU/tasks.
612
613And then because we have L1 -> L2, so the holder of L1 is going to acquire L2
614in L1 -> L2, however since L2 is already held by another CPU/task, plus L1 ->
615L2 and L2 -> L3 are not -(xR)-> and -(Sx)-> (the definition of strong), which
616means either L2 in L1 -> L2 is a non-recursive locker (blocked by anyone) or
617the L2 in L2 -> L3, is writer (blocking anyone), therefore the holder of L1
618cannot get L2, it has to wait L2's holder to release.
619
620Moreover, we can have a similar conclusion for L2's holder: it has to wait L3's
621holder to release, and so on. We now can prove that Lx's holder has to wait for
622Lx+1's holder to release, and note that Ln+1 is L1, so we have a circular
623waiting scenario and nobody can get progress, therefore a deadlock.
624
625Proof for necessary (Lemma 2):
626
627Lemma 2 is equivalent to: If there is a deadlock scenario, then there must be a
628strong circle in the dependency graph.
629
630According to Wikipedia[1], if there is a deadlock, then there must be a circular
631waiting scenario, means there are N CPU/tasks, where CPU/task P1 is waiting for
632a lock held by P2, and P2 is waiting for a lock held by P3, ... and Pn is waiting
633for a lock held by P1. Let's name the lock Px is waiting as Lx, so since P1 is waiting
634for L1 and holding Ln, so we will have Ln -> L1 in the dependency graph. Similarly,
635we have L1 -> L2, L2 -> L3, ..., Ln-1 -> Ln in the dependency graph, which means we
636have a circle:
637
638	Ln -> L1 -> L2 -> ... -> Ln
639
640, and now let's prove the circle is strong:
641
642For a lock Lx, Px contributes the dependency Lx-1 -> Lx and Px+1 contributes
643the dependency Lx -> Lx+1, and since Px is waiting for Px+1 to release Lx,
644so it's impossible that Lx on Px+1 is a reader and Lx on Px is a recursive
645reader, because readers (no matter recursive or not) don't block recursive
646readers, therefore Lx-1 -> Lx and Lx -> Lx+1 cannot be a -(xR)-> -(Sx)-> pair,
647and this is true for any lock in the circle, therefore, the circle is strong.
648
649References:
650-----------
651[1]: https://en.wikipedia.org/wiki/Deadlock
652[2]: Shibu, K. (2009). Intro To Embedded Systems (1st ed.). Tata McGraw-Hill
653