1=====================
2Hugetlbfs Reservation
3=====================
4
5Overview
6========
7
8Huge pages as described at Documentation/mm/hugetlbpage.rst are typically
9preallocated for application use.  These huge pages are instantiated in a
10task's address space at page fault time if the VMA indicates huge pages are
11to be used.  If no huge page exists at page fault time, the task is sent
12a SIGBUS and often dies an unhappy death.  Shortly after huge page support
13was added, it was determined that it would be better to detect a shortage
14of huge pages at mmap() time.  The idea is that if there were not enough
15huge pages to cover the mapping, the mmap() would fail.  This was first
16done with a simple check in the code at mmap() time to determine if there
17were enough free huge pages to cover the mapping.  Like most things in the
18kernel, the code has evolved over time.  However, the basic idea was to
19'reserve' huge pages at mmap() time to ensure that huge pages would be
20available for page faults in that mapping.  The description below attempts to
21describe how huge page reserve processing is done in the v4.10 kernel.
22
23
24Audience
25========
26This description is primarily targeted at kernel developers who are modifying
27hugetlbfs code.
28
29
30The Data Structures
31===================
32
33resv_huge_pages
34	This is a global (per-hstate) count of reserved huge pages.  Reserved
35	huge pages are only available to the task which reserved them.
36	Therefore, the number of huge pages generally available is computed
37	as (``free_huge_pages - resv_huge_pages``).
38Reserve Map
39	A reserve map is described by the structure::
40
41		struct resv_map {
42			struct kref refs;
43			spinlock_t lock;
44			struct list_head regions;
45			long adds_in_progress;
46			struct list_head region_cache;
47			long region_cache_count;
48		};
49
50	There is one reserve map for each huge page mapping in the system.
51	The regions list within the resv_map describes the regions within
52	the mapping.  A region is described as::
53
54		struct file_region {
55			struct list_head link;
56			long from;
57			long to;
58		};
59
60	The 'from' and 'to' fields of the file region structure are huge page
61	indices into the mapping.  Depending on the type of mapping, a
62	region in the reserv_map may indicate reservations exist for the
63	range, or reservations do not exist.
64Flags for MAP_PRIVATE Reservations
65	These are stored in the bottom bits of the reservation map pointer.
66
67	``#define HPAGE_RESV_OWNER    (1UL << 0)``
68		Indicates this task is the owner of the reservations
69		associated with the mapping.
70	``#define HPAGE_RESV_UNMAPPED (1UL << 1)``
71		Indicates task originally mapping this range (and creating
72		reserves) has unmapped a page from this task (the child)
73		due to a failed COW.
74Page Flags
75	The PagePrivate page flag is used to indicate that a huge page
76	reservation must be restored when the huge page is freed.  More
77	details will be discussed in the "Freeing huge pages" section.
78
79
80Reservation Map Location (Private or Shared)
81============================================
82
83A huge page mapping or segment is either private or shared.  If private,
84it is typically only available to a single address space (task).  If shared,
85it can be mapped into multiple address spaces (tasks).  The location and
86semantics of the reservation map is significantly different for the two types
87of mappings.  Location differences are:
88
89- For private mappings, the reservation map hangs off the VMA structure.
90  Specifically, vma->vm_private_data.  This reserve map is created at the
91  time the mapping (mmap(MAP_PRIVATE)) is created.
92- For shared mappings, the reservation map hangs off the inode.  Specifically,
93  inode->i_mapping->private_data.  Since shared mappings are always backed
94  by files in the hugetlbfs filesystem, the hugetlbfs code ensures each inode
95  contains a reservation map.  As a result, the reservation map is allocated
96  when the inode is created.
97
98
99Creating Reservations
100=====================
101Reservations are created when a huge page backed shared memory segment is
102created (shmget(SHM_HUGETLB)) or a mapping is created via mmap(MAP_HUGETLB).
103These operations result in a call to the routine hugetlb_reserve_pages()::
104
105	int hugetlb_reserve_pages(struct inode *inode,
106				  long from, long to,
107				  struct vm_area_struct *vma,
108				  vm_flags_t vm_flags)
109
110The first thing hugetlb_reserve_pages() does is check if the NORESERVE
111flag was specified in either the shmget() or mmap() call.  If NORESERVE
112was specified, then this routine returns immediately as no reservations
113are desired.
114
115The arguments 'from' and 'to' are huge page indices into the mapping or
116underlying file.  For shmget(), 'from' is always 0 and 'to' corresponds to
117the length of the segment/mapping.  For mmap(), the offset argument could
118be used to specify the offset into the underlying file.  In such a case,
119the 'from' and 'to' arguments have been adjusted by this offset.
120
121One of the big differences between PRIVATE and SHARED mappings is the way
122in which reservations are represented in the reservation map.
123
124- For shared mappings, an entry in the reservation map indicates a reservation
125  exists or did exist for the corresponding page.  As reservations are
126  consumed, the reservation map is not modified.
127- For private mappings, the lack of an entry in the reservation map indicates
128  a reservation exists for the corresponding page.  As reservations are
129  consumed, entries are added to the reservation map.  Therefore, the
130  reservation map can also be used to determine which reservations have
131  been consumed.
132
133For private mappings, hugetlb_reserve_pages() creates the reservation map and
134hangs it off the VMA structure.  In addition, the HPAGE_RESV_OWNER flag is set
135to indicate this VMA owns the reservations.
136
137The reservation map is consulted to determine how many huge page reservations
138are needed for the current mapping/segment.  For private mappings, this is
139always the value (to - from).  However, for shared mappings it is possible that
140some reservations may already exist within the range (to - from).  See the
141section :ref:`Reservation Map Modifications <resv_map_modifications>`
142for details on how this is accomplished.
143
144The mapping may be associated with a subpool.  If so, the subpool is consulted
145to ensure there is sufficient space for the mapping.  It is possible that the
146subpool has set aside reservations that can be used for the mapping.  See the
147section :ref:`Subpool Reservations <sub_pool_resv>` for more details.
148
149After consulting the reservation map and subpool, the number of needed new
150reservations is known.  The routine hugetlb_acct_memory() is called to check
151for and take the requested number of reservations.  hugetlb_acct_memory()
152calls into routines that potentially allocate and adjust surplus page counts.
153However, within those routines the code is simply checking to ensure there
154are enough free huge pages to accommodate the reservation.  If there are,
155the global reservation count resv_huge_pages is adjusted something like the
156following::
157
158	if (resv_needed <= (resv_huge_pages - free_huge_pages))
159		resv_huge_pages += resv_needed;
160
161Note that the global lock hugetlb_lock is held when checking and adjusting
162these counters.
163
164If there were enough free huge pages and the global count resv_huge_pages
165was adjusted, then the reservation map associated with the mapping is
166modified to reflect the reservations.  In the case of a shared mapping, a
167file_region will exist that includes the range 'from' - 'to'.  For private
168mappings, no modifications are made to the reservation map as lack of an
169entry indicates a reservation exists.
170
171If hugetlb_reserve_pages() was successful, the global reservation count and
172reservation map associated with the mapping will be modified as required to
173ensure reservations exist for the range 'from' - 'to'.
174
175.. _consume_resv:
176
177Consuming Reservations/Allocating a Huge Page
178=============================================
179
180Reservations are consumed when huge pages associated with the reservations
181are allocated and instantiated in the corresponding mapping.  The allocation
182is performed within the routine alloc_huge_page()::
183
184	struct page *alloc_huge_page(struct vm_area_struct *vma,
185				     unsigned long addr, int avoid_reserve)
186
187alloc_huge_page is passed a VMA pointer and a virtual address, so it can
188consult the reservation map to determine if a reservation exists.  In addition,
189alloc_huge_page takes the argument avoid_reserve which indicates reserves
190should not be used even if it appears they have been set aside for the
191specified address.  The avoid_reserve argument is most often used in the case
192of Copy on Write and Page Migration where additional copies of an existing
193page are being allocated.
194
195The helper routine vma_needs_reservation() is called to determine if a
196reservation exists for the address within the mapping(vma).  See the section
197:ref:`Reservation Map Helper Routines <resv_map_helpers>` for detailed
198information on what this routine does.
199The value returned from vma_needs_reservation() is generally
2000 or 1.  0 if a reservation exists for the address, 1 if no reservation exists.
201If a reservation does not exist, and there is a subpool associated with the
202mapping the subpool is consulted to determine if it contains reservations.
203If the subpool contains reservations, one can be used for this allocation.
204However, in every case the avoid_reserve argument overrides the use of
205a reservation for the allocation.  After determining whether a reservation
206exists and can be used for the allocation, the routine dequeue_huge_page_vma()
207is called.  This routine takes two arguments related to reservations:
208
209- avoid_reserve, this is the same value/argument passed to alloc_huge_page()
210- chg, even though this argument is of type long only the values 0 or 1 are
211  passed to dequeue_huge_page_vma.  If the value is 0, it indicates a
212  reservation exists (see the section "Memory Policy and Reservations" for
213  possible issues).  If the value is 1, it indicates a reservation does not
214  exist and the page must be taken from the global free pool if possible.
215
216The free lists associated with the memory policy of the VMA are searched for
217a free page.  If a page is found, the value free_huge_pages is decremented
218when the page is removed from the free list.  If there was a reservation
219associated with the page, the following adjustments are made::
220
221	SetPagePrivate(page);	/* Indicates allocating this page consumed
222				 * a reservation, and if an error is
223				 * encountered such that the page must be
224				 * freed, the reservation will be restored. */
225	resv_huge_pages--;	/* Decrement the global reservation count */
226
227Note, if no huge page can be found that satisfies the VMA's memory policy
228an attempt will be made to allocate one using the buddy allocator.  This
229brings up the issue of surplus huge pages and overcommit which is beyond
230the scope reservations.  Even if a surplus page is allocated, the same
231reservation based adjustments as above will be made: SetPagePrivate(page) and
232resv_huge_pages--.
233
234After obtaining a new huge page, (page)->private is set to the value of
235the subpool associated with the page if it exists.  This will be used for
236subpool accounting when the page is freed.
237
238The routine vma_commit_reservation() is then called to adjust the reserve
239map based on the consumption of the reservation.  In general, this involves
240ensuring the page is represented within a file_region structure of the region
241map.  For shared mappings where the reservation was present, an entry
242in the reserve map already existed so no change is made.  However, if there
243was no reservation in a shared mapping or this was a private mapping a new
244entry must be created.
245
246It is possible that the reserve map could have been changed between the call
247to vma_needs_reservation() at the beginning of alloc_huge_page() and the
248call to vma_commit_reservation() after the page was allocated.  This would
249be possible if hugetlb_reserve_pages was called for the same page in a shared
250mapping.  In such cases, the reservation count and subpool free page count
251will be off by one.  This rare condition can be identified by comparing the
252return value from vma_needs_reservation and vma_commit_reservation.  If such
253a race is detected, the subpool and global reserve counts are adjusted to
254compensate.  See the section
255:ref:`Reservation Map Helper Routines <resv_map_helpers>` for more
256information on these routines.
257
258
259Instantiate Huge Pages
260======================
261
262After huge page allocation, the page is typically added to the page tables
263of the allocating task.  Before this, pages in a shared mapping are added
264to the page cache and pages in private mappings are added to an anonymous
265reverse mapping.  In both cases, the PagePrivate flag is cleared.  Therefore,
266when a huge page that has been instantiated is freed no adjustment is made
267to the global reservation count (resv_huge_pages).
268
269
270Freeing Huge Pages
271==================
272
273Huge page freeing is performed by the routine free_huge_page().  This routine
274is the destructor for hugetlbfs compound pages.  As a result, it is only
275passed a pointer to the page struct.  When a huge page is freed, reservation
276accounting may need to be performed.  This would be the case if the page was
277associated with a subpool that contained reserves, or the page is being freed
278on an error path where a global reserve count must be restored.
279
280The page->private field points to any subpool associated with the page.
281If the PagePrivate flag is set, it indicates the global reserve count should
282be adjusted (see the section
283:ref:`Consuming Reservations/Allocating a Huge Page <consume_resv>`
284for information on how these are set).
285
286The routine first calls hugepage_subpool_put_pages() for the page.  If this
287routine returns a value of 0 (which does not equal the value passed 1) it
288indicates reserves are associated with the subpool, and this newly free page
289must be used to keep the number of subpool reserves above the minimum size.
290Therefore, the global resv_huge_pages counter is incremented in this case.
291
292If the PagePrivate flag was set in the page, the global resv_huge_pages counter
293will always be incremented.
294
295.. _sub_pool_resv:
296
297Subpool Reservations
298====================
299
300There is a struct hstate associated with each huge page size.  The hstate
301tracks all huge pages of the specified size.  A subpool represents a subset
302of pages within a hstate that is associated with a mounted hugetlbfs
303filesystem.
304
305When a hugetlbfs filesystem is mounted a min_size option can be specified
306which indicates the minimum number of huge pages required by the filesystem.
307If this option is specified, the number of huge pages corresponding to
308min_size are reserved for use by the filesystem.  This number is tracked in
309the min_hpages field of a struct hugepage_subpool.  At mount time,
310hugetlb_acct_memory(min_hpages) is called to reserve the specified number of
311huge pages.  If they can not be reserved, the mount fails.
312
313The routines hugepage_subpool_get/put_pages() are called when pages are
314obtained from or released back to a subpool.  They perform all subpool
315accounting, and track any reservations associated with the subpool.
316hugepage_subpool_get/put_pages are passed the number of huge pages by which
317to adjust the subpool 'used page' count (down for get, up for put).  Normally,
318they return the same value that was passed or an error if not enough pages
319exist in the subpool.
320
321However, if reserves are associated with the subpool a return value less
322than the passed value may be returned.  This return value indicates the
323number of additional global pool adjustments which must be made.  For example,
324suppose a subpool contains 3 reserved huge pages and someone asks for 5.
325The 3 reserved pages associated with the subpool can be used to satisfy part
326of the request.  But, 2 pages must be obtained from the global pools.  To
327relay this information to the caller, the value 2 is returned.  The caller
328is then responsible for attempting to obtain the additional two pages from
329the global pools.
330
331
332COW and Reservations
333====================
334
335Since shared mappings all point to and use the same underlying pages, the
336biggest reservation concern for COW is private mappings.  In this case,
337two tasks can be pointing at the same previously allocated page.  One task
338attempts to write to the page, so a new page must be allocated so that each
339task points to its own page.
340
341When the page was originally allocated, the reservation for that page was
342consumed.  When an attempt to allocate a new page is made as a result of
343COW, it is possible that no free huge pages are free and the allocation
344will fail.
345
346When the private mapping was originally created, the owner of the mapping
347was noted by setting the HPAGE_RESV_OWNER bit in the pointer to the reservation
348map of the owner.  Since the owner created the mapping, the owner owns all
349the reservations associated with the mapping.  Therefore, when a write fault
350occurs and there is no page available, different action is taken for the owner
351and non-owner of the reservation.
352
353In the case where the faulting task is not the owner, the fault will fail and
354the task will typically receive a SIGBUS.
355
356If the owner is the faulting task, we want it to succeed since it owned the
357original reservation.  To accomplish this, the page is unmapped from the
358non-owning task.  In this way, the only reference is from the owning task.
359In addition, the HPAGE_RESV_UNMAPPED bit is set in the reservation map pointer
360of the non-owning task.  The non-owning task may receive a SIGBUS if it later
361faults on a non-present page.  But, the original owner of the
362mapping/reservation will behave as expected.
363
364
365.. _resv_map_modifications:
366
367Reservation Map Modifications
368=============================
369
370The following low level routines are used to make modifications to a
371reservation map.  Typically, these routines are not called directly.  Rather,
372a reservation map helper routine is called which calls one of these low level
373routines.  These low level routines are fairly well documented in the source
374code (mm/hugetlb.c).  These routines are::
375
376	long region_chg(struct resv_map *resv, long f, long t);
377	long region_add(struct resv_map *resv, long f, long t);
378	void region_abort(struct resv_map *resv, long f, long t);
379	long region_count(struct resv_map *resv, long f, long t);
380
381Operations on the reservation map typically involve two operations:
382
3831) region_chg() is called to examine the reserve map and determine how
384   many pages in the specified range [f, t) are NOT currently represented.
385
386   The calling code performs global checks and allocations to determine if
387   there are enough huge pages for the operation to succeed.
388
3892)
390  a) If the operation can succeed, region_add() is called to actually modify
391     the reservation map for the same range [f, t) previously passed to
392     region_chg().
393  b) If the operation can not succeed, region_abort is called for the same
394     range [f, t) to abort the operation.
395
396Note that this is a two step process where region_add() and region_abort()
397are guaranteed to succeed after a prior call to region_chg() for the same
398range.  region_chg() is responsible for pre-allocating any data structures
399necessary to ensure the subsequent operations (specifically region_add()))
400will succeed.
401
402As mentioned above, region_chg() determines the number of pages in the range
403which are NOT currently represented in the map.  This number is returned to
404the caller.  region_add() returns the number of pages in the range added to
405the map.  In most cases, the return value of region_add() is the same as the
406return value of region_chg().  However, in the case of shared mappings it is
407possible for changes to the reservation map to be made between the calls to
408region_chg() and region_add().  In this case, the return value of region_add()
409will not match the return value of region_chg().  It is likely that in such
410cases global counts and subpool accounting will be incorrect and in need of
411adjustment.  It is the responsibility of the caller to check for this condition
412and make the appropriate adjustments.
413
414The routine region_del() is called to remove regions from a reservation map.
415It is typically called in the following situations:
416
417- When a file in the hugetlbfs filesystem is being removed, the inode will
418  be released and the reservation map freed.  Before freeing the reservation
419  map, all the individual file_region structures must be freed.  In this case
420  region_del is passed the range [0, LONG_MAX).
421- When a hugetlbfs file is being truncated.  In this case, all allocated pages
422  after the new file size must be freed.  In addition, any file_region entries
423  in the reservation map past the new end of file must be deleted.  In this
424  case, region_del is passed the range [new_end_of_file, LONG_MAX).
425- When a hole is being punched in a hugetlbfs file.  In this case, huge pages
426  are removed from the middle of the file one at a time.  As the pages are
427  removed, region_del() is called to remove the corresponding entry from the
428  reservation map.  In this case, region_del is passed the range
429  [page_idx, page_idx + 1).
430
431In every case, region_del() will return the number of pages removed from the
432reservation map.  In VERY rare cases, region_del() can fail.  This can only
433happen in the hole punch case where it has to split an existing file_region
434entry and can not allocate a new structure.  In this error case, region_del()
435will return -ENOMEM.  The problem here is that the reservation map will
436indicate that there is a reservation for the page.  However, the subpool and
437global reservation counts will not reflect the reservation.  To handle this
438situation, the routine hugetlb_fix_reserve_counts() is called to adjust the
439counters so that they correspond with the reservation map entry that could
440not be deleted.
441
442region_count() is called when unmapping a private huge page mapping.  In
443private mappings, the lack of a entry in the reservation map indicates that
444a reservation exists.  Therefore, by counting the number of entries in the
445reservation map we know how many reservations were consumed and how many are
446outstanding (outstanding = (end - start) - region_count(resv, start, end)).
447Since the mapping is going away, the subpool and global reservation counts
448are decremented by the number of outstanding reservations.
449
450.. _resv_map_helpers:
451
452Reservation Map Helper Routines
453===============================
454
455Several helper routines exist to query and modify the reservation maps.
456These routines are only interested with reservations for a specific huge
457page, so they just pass in an address instead of a range.  In addition,
458they pass in the associated VMA.  From the VMA, the type of mapping (private
459or shared) and the location of the reservation map (inode or VMA) can be
460determined.  These routines simply call the underlying routines described
461in the section "Reservation Map Modifications".  However, they do take into
462account the 'opposite' meaning of reservation map entries for private and
463shared mappings and hide this detail from the caller::
464
465	long vma_needs_reservation(struct hstate *h,
466				   struct vm_area_struct *vma,
467				   unsigned long addr)
468
469This routine calls region_chg() for the specified page.  If no reservation
470exists, 1 is returned.  If a reservation exists, 0 is returned::
471
472	long vma_commit_reservation(struct hstate *h,
473				    struct vm_area_struct *vma,
474				    unsigned long addr)
475
476This calls region_add() for the specified page.  As in the case of region_chg
477and region_add, this routine is to be called after a previous call to
478vma_needs_reservation.  It will add a reservation entry for the page.  It
479returns 1 if the reservation was added and 0 if not.  The return value should
480be compared with the return value of the previous call to
481vma_needs_reservation.  An unexpected difference indicates the reservation
482map was modified between calls::
483
484	void vma_end_reservation(struct hstate *h,
485				 struct vm_area_struct *vma,
486				 unsigned long addr)
487
488This calls region_abort() for the specified page.  As in the case of region_chg
489and region_abort, this routine is to be called after a previous call to
490vma_needs_reservation.  It will abort/end the in progress reservation add
491operation::
492
493	long vma_add_reservation(struct hstate *h,
494				 struct vm_area_struct *vma,
495				 unsigned long addr)
496
497This is a special wrapper routine to help facilitate reservation cleanup
498on error paths.  It is only called from the routine restore_reserve_on_error().
499This routine is used in conjunction with vma_needs_reservation in an attempt
500to add a reservation to the reservation map.  It takes into account the
501different reservation map semantics for private and shared mappings.  Hence,
502region_add is called for shared mappings (as an entry present in the map
503indicates a reservation), and region_del is called for private mappings (as
504the absence of an entry in the map indicates a reservation).  See the section
505"Reservation cleanup in error paths" for more information on what needs to
506be done on error paths.
507
508
509Reservation Cleanup in Error Paths
510==================================
511
512As mentioned in the section
513:ref:`Reservation Map Helper Routines <resv_map_helpers>`, reservation
514map modifications are performed in two steps.  First vma_needs_reservation
515is called before a page is allocated.  If the allocation is successful,
516then vma_commit_reservation is called.  If not, vma_end_reservation is called.
517Global and subpool reservation counts are adjusted based on success or failure
518of the operation and all is well.
519
520Additionally, after a huge page is instantiated the PagePrivate flag is
521cleared so that accounting when the page is ultimately freed is correct.
522
523However, there are several instances where errors are encountered after a huge
524page is allocated but before it is instantiated.  In this case, the page
525allocation has consumed the reservation and made the appropriate subpool,
526reservation map and global count adjustments.  If the page is freed at this
527time (before instantiation and clearing of PagePrivate), then free_huge_page
528will increment the global reservation count.  However, the reservation map
529indicates the reservation was consumed.  This resulting inconsistent state
530will cause the 'leak' of a reserved huge page.  The global reserve count will
531be  higher than it should and prevent allocation of a pre-allocated page.
532
533The routine restore_reserve_on_error() attempts to handle this situation.  It
534is fairly well documented.  The intention of this routine is to restore
535the reservation map to the way it was before the page allocation.   In this
536way, the state of the reservation map will correspond to the global reservation
537count after the page is freed.
538
539The routine restore_reserve_on_error itself may encounter errors while
540attempting to restore the reservation map entry.  In this case, it will
541simply clear the PagePrivate flag of the page.  In this way, the global
542reserve count will not be incremented when the page is freed.  However, the
543reservation map will continue to look as though the reservation was consumed.
544A page can still be allocated for the address, but it will not use a reserved
545page as originally intended.
546
547There is some code (most notably userfaultfd) which can not call
548restore_reserve_on_error.  In this case, it simply modifies the PagePrivate
549so that a reservation will not be leaked when the huge page is freed.
550
551
552Reservations and Memory Policy
553==============================
554Per-node huge page lists existed in struct hstate when git was first used
555to manage Linux code.  The concept of reservations was added some time later.
556When reservations were added, no attempt was made to take memory policy
557into account.  While cpusets are not exactly the same as memory policy, this
558comment in hugetlb_acct_memory sums up the interaction between reservations
559and cpusets/memory policy::
560
561	/*
562	 * When cpuset is configured, it breaks the strict hugetlb page
563	 * reservation as the accounting is done on a global variable. Such
564	 * reservation is completely rubbish in the presence of cpuset because
565	 * the reservation is not checked against page availability for the
566	 * current cpuset. Application can still potentially OOM'ed by kernel
567	 * with lack of free htlb page in cpuset that the task is in.
568	 * Attempt to enforce strict accounting with cpuset is almost
569	 * impossible (or too ugly) because cpuset is too fluid that
570	 * task or memory node can be dynamically moved between cpusets.
571	 *
572	 * The change of semantics for shared hugetlb mapping with cpuset is
573	 * undesirable. However, in order to preserve some of the semantics,
574	 * we fall back to check against current free page availability as
575	 * a best attempt and hopefully to minimize the impact of changing
576	 * semantics that cpuset has.
577	 */
578
579Huge page reservations were added to prevent unexpected page allocation
580failures (OOM) at page fault time.  However, if an application makes use
581of cpusets or memory policy there is no guarantee that huge pages will be
582available on the required nodes.  This is true even if there are a sufficient
583number of global reservations.
584
585Hugetlbfs regression testing
586============================
587
588The most complete set of hugetlb tests are in the libhugetlbfs repository.
589If you modify any hugetlb related code, use the libhugetlbfs test suite
590to check for regressions.  In addition, if you add any new hugetlb
591functionality, please add appropriate tests to libhugetlbfs.
592
593--
594Mike Kravetz, 7 April 2017
595