hugetlbfs_reserv.rst 29 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596
  1. .. _hugetlbfs_reserve:
  2. =====================
  3. Hugetlbfs Reservation
  4. =====================
  5. Overview
  6. ========
  7. Huge pages as described at :ref:`hugetlbpage` are typically
  8. preallocated for application use. These huge pages are instantiated in a
  9. task's address space at page fault time if the VMA indicates huge pages are
  10. to be used. If no huge page exists at page fault time, the task is sent
  11. a SIGBUS and often dies an unhappy death. Shortly after huge page support
  12. was added, it was determined that it would be better to detect a shortage
  13. of huge pages at mmap() time. The idea is that if there were not enough
  14. huge pages to cover the mapping, the mmap() would fail. This was first
  15. done with a simple check in the code at mmap() time to determine if there
  16. were enough free huge pages to cover the mapping. Like most things in the
  17. kernel, the code has evolved over time. However, the basic idea was to
  18. 'reserve' huge pages at mmap() time to ensure that huge pages would be
  19. available for page faults in that mapping. The description below attempts to
  20. describe how huge page reserve processing is done in the v4.10 kernel.
  21. Audience
  22. ========
  23. This description is primarily targeted at kernel developers who are modifying
  24. hugetlbfs code.
  25. The Data Structures
  26. ===================
  27. resv_huge_pages
  28. This is a global (per-hstate) count of reserved huge pages. Reserved
  29. huge pages are only available to the task which reserved them.
  30. Therefore, the number of huge pages generally available is computed
  31. as (``free_huge_pages - resv_huge_pages``).
  32. Reserve Map
  33. A reserve map is described by the structure::
  34. struct resv_map {
  35. struct kref refs;
  36. spinlock_t lock;
  37. struct list_head regions;
  38. long adds_in_progress;
  39. struct list_head region_cache;
  40. long region_cache_count;
  41. };
  42. There is one reserve map for each huge page mapping in the system.
  43. The regions list within the resv_map describes the regions within
  44. the mapping. A region is described as::
  45. struct file_region {
  46. struct list_head link;
  47. long from;
  48. long to;
  49. };
  50. The 'from' and 'to' fields of the file region structure are huge page
  51. indices into the mapping. Depending on the type of mapping, a
  52. region in the reserv_map may indicate reservations exist for the
  53. range, or reservations do not exist.
  54. Flags for MAP_PRIVATE Reservations
  55. These are stored in the bottom bits of the reservation map pointer.
  56. ``#define HPAGE_RESV_OWNER (1UL << 0)``
  57. Indicates this task is the owner of the reservations
  58. associated with the mapping.
  59. ``#define HPAGE_RESV_UNMAPPED (1UL << 1)``
  60. Indicates task originally mapping this range (and creating
  61. reserves) has unmapped a page from this task (the child)
  62. due to a failed COW.
  63. Page Flags
  64. The PagePrivate page flag is used to indicate that a huge page
  65. reservation must be restored when the huge page is freed. More
  66. details will be discussed in the "Freeing huge pages" section.
  67. Reservation Map Location (Private or Shared)
  68. ============================================
  69. A huge page mapping or segment is either private or shared. If private,
  70. it is typically only available to a single address space (task). If shared,
  71. it can be mapped into multiple address spaces (tasks). The location and
  72. semantics of the reservation map is significantly different for the two types
  73. of mappings. Location differences are:
  74. - For private mappings, the reservation map hangs off the VMA structure.
  75. Specifically, vma->vm_private_data. This reserve map is created at the
  76. time the mapping (mmap(MAP_PRIVATE)) is created.
  77. - For shared mappings, the reservation map hangs off the inode. Specifically,
  78. inode->i_mapping->private_data. Since shared mappings are always backed
  79. by files in the hugetlbfs filesystem, the hugetlbfs code ensures each inode
  80. contains a reservation map. As a result, the reservation map is allocated
  81. when the inode is created.
  82. Creating Reservations
  83. =====================
  84. Reservations are created when a huge page backed shared memory segment is
  85. created (shmget(SHM_HUGETLB)) or a mapping is created via mmap(MAP_HUGETLB).
  86. These operations result in a call to the routine hugetlb_reserve_pages()::
  87. int hugetlb_reserve_pages(struct inode *inode,
  88. long from, long to,
  89. struct vm_area_struct *vma,
  90. vm_flags_t vm_flags)
  91. The first thing hugetlb_reserve_pages() does is check if the NORESERVE
  92. flag was specified in either the shmget() or mmap() call. If NORESERVE
  93. was specified, then this routine returns immediately as no reservations
  94. are desired.
  95. The arguments 'from' and 'to' are huge page indices into the mapping or
  96. underlying file. For shmget(), 'from' is always 0 and 'to' corresponds to
  97. the length of the segment/mapping. For mmap(), the offset argument could
  98. be used to specify the offset into the underlying file. In such a case,
  99. the 'from' and 'to' arguments have been adjusted by this offset.
  100. One of the big differences between PRIVATE and SHARED mappings is the way
  101. in which reservations are represented in the reservation map.
  102. - For shared mappings, an entry in the reservation map indicates a reservation
  103. exists or did exist for the corresponding page. As reservations are
  104. consumed, the reservation map is not modified.
  105. - For private mappings, the lack of an entry in the reservation map indicates
  106. a reservation exists for the corresponding page. As reservations are
  107. consumed, entries are added to the reservation map. Therefore, the
  108. reservation map can also be used to determine which reservations have
  109. been consumed.
  110. For private mappings, hugetlb_reserve_pages() creates the reservation map and
  111. hangs it off the VMA structure. In addition, the HPAGE_RESV_OWNER flag is set
  112. to indicate this VMA owns the reservations.
  113. The reservation map is consulted to determine how many huge page reservations
  114. are needed for the current mapping/segment. For private mappings, this is
  115. always the value (to - from). However, for shared mappings it is possible that
  116. some reservations may already exist within the range (to - from). See the
  117. section :ref:`Reservation Map Modifications <resv_map_modifications>`
  118. for details on how this is accomplished.
  119. The mapping may be associated with a subpool. If so, the subpool is consulted
  120. to ensure there is sufficient space for the mapping. It is possible that the
  121. subpool has set aside reservations that can be used for the mapping. See the
  122. section :ref:`Subpool Reservations <sub_pool_resv>` for more details.
  123. After consulting the reservation map and subpool, the number of needed new
  124. reservations is known. The routine hugetlb_acct_memory() is called to check
  125. for and take the requested number of reservations. hugetlb_acct_memory()
  126. calls into routines that potentially allocate and adjust surplus page counts.
  127. However, within those routines the code is simply checking to ensure there
  128. are enough free huge pages to accommodate the reservation. If there are,
  129. the global reservation count resv_huge_pages is adjusted something like the
  130. following::
  131. if (resv_needed <= (resv_huge_pages - free_huge_pages))
  132. resv_huge_pages += resv_needed;
  133. Note that the global lock hugetlb_lock is held when checking and adjusting
  134. these counters.
  135. If there were enough free huge pages and the global count resv_huge_pages
  136. was adjusted, then the reservation map associated with the mapping is
  137. modified to reflect the reservations. In the case of a shared mapping, a
  138. file_region will exist that includes the range 'from' - 'to'. For private
  139. mappings, no modifications are made to the reservation map as lack of an
  140. entry indicates a reservation exists.
  141. If hugetlb_reserve_pages() was successful, the global reservation count and
  142. reservation map associated with the mapping will be modified as required to
  143. ensure reservations exist for the range 'from' - 'to'.
  144. .. _consume_resv:
  145. Consuming Reservations/Allocating a Huge Page
  146. =============================================
  147. Reservations are consumed when huge pages associated with the reservations
  148. are allocated and instantiated in the corresponding mapping. The allocation
  149. is performed within the routine alloc_huge_page()::
  150. struct page *alloc_huge_page(struct vm_area_struct *vma,
  151. unsigned long addr, int avoid_reserve)
  152. alloc_huge_page is passed a VMA pointer and a virtual address, so it can
  153. consult the reservation map to determine if a reservation exists. In addition,
  154. alloc_huge_page takes the argument avoid_reserve which indicates reserves
  155. should not be used even if it appears they have been set aside for the
  156. specified address. The avoid_reserve argument is most often used in the case
  157. of Copy on Write and Page Migration where additional copies of an existing
  158. page are being allocated.
  159. The helper routine vma_needs_reservation() is called to determine if a
  160. reservation exists for the address within the mapping(vma). See the section
  161. :ref:`Reservation Map Helper Routines <resv_map_helpers>` for detailed
  162. information on what this routine does.
  163. The value returned from vma_needs_reservation() is generally
  164. 0 or 1. 0 if a reservation exists for the address, 1 if no reservation exists.
  165. If a reservation does not exist, and there is a subpool associated with the
  166. mapping the subpool is consulted to determine if it contains reservations.
  167. If the subpool contains reservations, one can be used for this allocation.
  168. However, in every case the avoid_reserve argument overrides the use of
  169. a reservation for the allocation. After determining whether a reservation
  170. exists and can be used for the allocation, the routine dequeue_huge_page_vma()
  171. is called. This routine takes two arguments related to reservations:
  172. - avoid_reserve, this is the same value/argument passed to alloc_huge_page()
  173. - chg, even though this argument is of type long only the values 0 or 1 are
  174. passed to dequeue_huge_page_vma. If the value is 0, it indicates a
  175. reservation exists (see the section "Memory Policy and Reservations" for
  176. possible issues). If the value is 1, it indicates a reservation does not
  177. exist and the page must be taken from the global free pool if possible.
  178. The free lists associated with the memory policy of the VMA are searched for
  179. a free page. If a page is found, the value free_huge_pages is decremented
  180. when the page is removed from the free list. If there was a reservation
  181. associated with the page, the following adjustments are made::
  182. SetPagePrivate(page); /* Indicates allocating this page consumed
  183. * a reservation, and if an error is
  184. * encountered such that the page must be
  185. * freed, the reservation will be restored. */
  186. resv_huge_pages--; /* Decrement the global reservation count */
  187. Note, if no huge page can be found that satisfies the VMA's memory policy
  188. an attempt will be made to allocate one using the buddy allocator. This
  189. brings up the issue of surplus huge pages and overcommit which is beyond
  190. the scope reservations. Even if a surplus page is allocated, the same
  191. reservation based adjustments as above will be made: SetPagePrivate(page) and
  192. resv_huge_pages--.
  193. After obtaining a new huge page, (page)->private is set to the value of
  194. the subpool associated with the page if it exists. This will be used for
  195. subpool accounting when the page is freed.
  196. The routine vma_commit_reservation() is then called to adjust the reserve
  197. map based on the consumption of the reservation. In general, this involves
  198. ensuring the page is represented within a file_region structure of the region
  199. map. For shared mappings where the reservation was present, an entry
  200. in the reserve map already existed so no change is made. However, if there
  201. was no reservation in a shared mapping or this was a private mapping a new
  202. entry must be created.
  203. It is possible that the reserve map could have been changed between the call
  204. to vma_needs_reservation() at the beginning of alloc_huge_page() and the
  205. call to vma_commit_reservation() after the page was allocated. This would
  206. be possible if hugetlb_reserve_pages was called for the same page in a shared
  207. mapping. In such cases, the reservation count and subpool free page count
  208. will be off by one. This rare condition can be identified by comparing the
  209. return value from vma_needs_reservation and vma_commit_reservation. If such
  210. a race is detected, the subpool and global reserve counts are adjusted to
  211. compensate. See the section
  212. :ref:`Reservation Map Helper Routines <resv_map_helpers>` for more
  213. information on these routines.
  214. Instantiate Huge Pages
  215. ======================
  216. After huge page allocation, the page is typically added to the page tables
  217. of the allocating task. Before this, pages in a shared mapping are added
  218. to the page cache and pages in private mappings are added to an anonymous
  219. reverse mapping. In both cases, the PagePrivate flag is cleared. Therefore,
  220. when a huge page that has been instantiated is freed no adjustment is made
  221. to the global reservation count (resv_huge_pages).
  222. Freeing Huge Pages
  223. ==================
  224. Huge page freeing is performed by the routine free_huge_page(). This routine
  225. is the destructor for hugetlbfs compound pages. As a result, it is only
  226. passed a pointer to the page struct. When a huge page is freed, reservation
  227. accounting may need to be performed. This would be the case if the page was
  228. associated with a subpool that contained reserves, or the page is being freed
  229. on an error path where a global reserve count must be restored.
  230. The page->private field points to any subpool associated with the page.
  231. If the PagePrivate flag is set, it indicates the global reserve count should
  232. be adjusted (see the section
  233. :ref:`Consuming Reservations/Allocating a Huge Page <consume_resv>`
  234. for information on how these are set).
  235. The routine first calls hugepage_subpool_put_pages() for the page. If this
  236. routine returns a value of 0 (which does not equal the value passed 1) it
  237. indicates reserves are associated with the subpool, and this newly free page
  238. must be used to keep the number of subpool reserves above the minimum size.
  239. Therefore, the global resv_huge_pages counter is incremented in this case.
  240. If the PagePrivate flag was set in the page, the global resv_huge_pages counter
  241. will always be incremented.
  242. .. _sub_pool_resv:
  243. Subpool Reservations
  244. ====================
  245. There is a struct hstate associated with each huge page size. The hstate
  246. tracks all huge pages of the specified size. A subpool represents a subset
  247. of pages within a hstate that is associated with a mounted hugetlbfs
  248. filesystem.
  249. When a hugetlbfs filesystem is mounted a min_size option can be specified
  250. which indicates the minimum number of huge pages required by the filesystem.
  251. If this option is specified, the number of huge pages corresponding to
  252. min_size are reserved for use by the filesystem. This number is tracked in
  253. the min_hpages field of a struct hugepage_subpool. At mount time,
  254. hugetlb_acct_memory(min_hpages) is called to reserve the specified number of
  255. huge pages. If they can not be reserved, the mount fails.
  256. The routines hugepage_subpool_get/put_pages() are called when pages are
  257. obtained from or released back to a subpool. They perform all subpool
  258. accounting, and track any reservations associated with the subpool.
  259. hugepage_subpool_get/put_pages are passed the number of huge pages by which
  260. to adjust the subpool 'used page' count (down for get, up for put). Normally,
  261. they return the same value that was passed or an error if not enough pages
  262. exist in the subpool.
  263. However, if reserves are associated with the subpool a return value less
  264. than the passed value may be returned. This return value indicates the
  265. number of additional global pool adjustments which must be made. For example,
  266. suppose a subpool contains 3 reserved huge pages and someone asks for 5.
  267. The 3 reserved pages associated with the subpool can be used to satisfy part
  268. of the request. But, 2 pages must be obtained from the global pools. To
  269. relay this information to the caller, the value 2 is returned. The caller
  270. is then responsible for attempting to obtain the additional two pages from
  271. the global pools.
  272. COW and Reservations
  273. ====================
  274. Since shared mappings all point to and use the same underlying pages, the
  275. biggest reservation concern for COW is private mappings. In this case,
  276. two tasks can be pointing at the same previously allocated page. One task
  277. attempts to write to the page, so a new page must be allocated so that each
  278. task points to its own page.
  279. When the page was originally allocated, the reservation for that page was
  280. consumed. When an attempt to allocate a new page is made as a result of
  281. COW, it is possible that no free huge pages are free and the allocation
  282. will fail.
  283. When the private mapping was originally created, the owner of the mapping
  284. was noted by setting the HPAGE_RESV_OWNER bit in the pointer to the reservation
  285. map of the owner. Since the owner created the mapping, the owner owns all
  286. the reservations associated with the mapping. Therefore, when a write fault
  287. occurs and there is no page available, different action is taken for the owner
  288. and non-owner of the reservation.
  289. In the case where the faulting task is not the owner, the fault will fail and
  290. the task will typically receive a SIGBUS.
  291. If the owner is the faulting task, we want it to succeed since it owned the
  292. original reservation. To accomplish this, the page is unmapped from the
  293. non-owning task. In this way, the only reference is from the owning task.
  294. In addition, the HPAGE_RESV_UNMAPPED bit is set in the reservation map pointer
  295. of the non-owning task. The non-owning task may receive a SIGBUS if it later
  296. faults on a non-present page. But, the original owner of the
  297. mapping/reservation will behave as expected.
  298. .. _resv_map_modifications:
  299. Reservation Map Modifications
  300. =============================
  301. The following low level routines are used to make modifications to a
  302. reservation map. Typically, these routines are not called directly. Rather,
  303. a reservation map helper routine is called which calls one of these low level
  304. routines. These low level routines are fairly well documented in the source
  305. code (mm/hugetlb.c). These routines are::
  306. long region_chg(struct resv_map *resv, long f, long t);
  307. long region_add(struct resv_map *resv, long f, long t);
  308. void region_abort(struct resv_map *resv, long f, long t);
  309. long region_count(struct resv_map *resv, long f, long t);
  310. Operations on the reservation map typically involve two operations:
  311. 1) region_chg() is called to examine the reserve map and determine how
  312. many pages in the specified range [f, t) are NOT currently represented.
  313. The calling code performs global checks and allocations to determine if
  314. there are enough huge pages for the operation to succeed.
  315. 2)
  316. a) If the operation can succeed, region_add() is called to actually modify
  317. the reservation map for the same range [f, t) previously passed to
  318. region_chg().
  319. b) If the operation can not succeed, region_abort is called for the same
  320. range [f, t) to abort the operation.
  321. Note that this is a two step process where region_add() and region_abort()
  322. are guaranteed to succeed after a prior call to region_chg() for the same
  323. range. region_chg() is responsible for pre-allocating any data structures
  324. necessary to ensure the subsequent operations (specifically region_add()))
  325. will succeed.
  326. As mentioned above, region_chg() determines the number of pages in the range
  327. which are NOT currently represented in the map. This number is returned to
  328. the caller. region_add() returns the number of pages in the range added to
  329. the map. In most cases, the return value of region_add() is the same as the
  330. return value of region_chg(). However, in the case of shared mappings it is
  331. possible for changes to the reservation map to be made between the calls to
  332. region_chg() and region_add(). In this case, the return value of region_add()
  333. will not match the return value of region_chg(). It is likely that in such
  334. cases global counts and subpool accounting will be incorrect and in need of
  335. adjustment. It is the responsibility of the caller to check for this condition
  336. and make the appropriate adjustments.
  337. The routine region_del() is called to remove regions from a reservation map.
  338. It is typically called in the following situations:
  339. - When a file in the hugetlbfs filesystem is being removed, the inode will
  340. be released and the reservation map freed. Before freeing the reservation
  341. map, all the individual file_region structures must be freed. In this case
  342. region_del is passed the range [0, LONG_MAX).
  343. - When a hugetlbfs file is being truncated. In this case, all allocated pages
  344. after the new file size must be freed. In addition, any file_region entries
  345. in the reservation map past the new end of file must be deleted. In this
  346. case, region_del is passed the range [new_end_of_file, LONG_MAX).
  347. - When a hole is being punched in a hugetlbfs file. In this case, huge pages
  348. are removed from the middle of the file one at a time. As the pages are
  349. removed, region_del() is called to remove the corresponding entry from the
  350. reservation map. In this case, region_del is passed the range
  351. [page_idx, page_idx + 1).
  352. In every case, region_del() will return the number of pages removed from the
  353. reservation map. In VERY rare cases, region_del() can fail. This can only
  354. happen in the hole punch case where it has to split an existing file_region
  355. entry and can not allocate a new structure. In this error case, region_del()
  356. will return -ENOMEM. The problem here is that the reservation map will
  357. indicate that there is a reservation for the page. However, the subpool and
  358. global reservation counts will not reflect the reservation. To handle this
  359. situation, the routine hugetlb_fix_reserve_counts() is called to adjust the
  360. counters so that they correspond with the reservation map entry that could
  361. not be deleted.
  362. region_count() is called when unmapping a private huge page mapping. In
  363. private mappings, the lack of a entry in the reservation map indicates that
  364. a reservation exists. Therefore, by counting the number of entries in the
  365. reservation map we know how many reservations were consumed and how many are
  366. outstanding (outstanding = (end - start) - region_count(resv, start, end)).
  367. Since the mapping is going away, the subpool and global reservation counts
  368. are decremented by the number of outstanding reservations.
  369. .. _resv_map_helpers:
  370. Reservation Map Helper Routines
  371. ===============================
  372. Several helper routines exist to query and modify the reservation maps.
  373. These routines are only interested with reservations for a specific huge
  374. page, so they just pass in an address instead of a range. In addition,
  375. they pass in the associated VMA. From the VMA, the type of mapping (private
  376. or shared) and the location of the reservation map (inode or VMA) can be
  377. determined. These routines simply call the underlying routines described
  378. in the section "Reservation Map Modifications". However, they do take into
  379. account the 'opposite' meaning of reservation map entries for private and
  380. shared mappings and hide this detail from the caller::
  381. long vma_needs_reservation(struct hstate *h,
  382. struct vm_area_struct *vma,
  383. unsigned long addr)
  384. This routine calls region_chg() for the specified page. If no reservation
  385. exists, 1 is returned. If a reservation exists, 0 is returned::
  386. long vma_commit_reservation(struct hstate *h,
  387. struct vm_area_struct *vma,
  388. unsigned long addr)
  389. This calls region_add() for the specified page. As in the case of region_chg
  390. and region_add, this routine is to be called after a previous call to
  391. vma_needs_reservation. It will add a reservation entry for the page. It
  392. returns 1 if the reservation was added and 0 if not. The return value should
  393. be compared with the return value of the previous call to
  394. vma_needs_reservation. An unexpected difference indicates the reservation
  395. map was modified between calls::
  396. void vma_end_reservation(struct hstate *h,
  397. struct vm_area_struct *vma,
  398. unsigned long addr)
  399. This calls region_abort() for the specified page. As in the case of region_chg
  400. and region_abort, this routine is to be called after a previous call to
  401. vma_needs_reservation. It will abort/end the in progress reservation add
  402. operation::
  403. long vma_add_reservation(struct hstate *h,
  404. struct vm_area_struct *vma,
  405. unsigned long addr)
  406. This is a special wrapper routine to help facilitate reservation cleanup
  407. on error paths. It is only called from the routine restore_reserve_on_error().
  408. This routine is used in conjunction with vma_needs_reservation in an attempt
  409. to add a reservation to the reservation map. It takes into account the
  410. different reservation map semantics for private and shared mappings. Hence,
  411. region_add is called for shared mappings (as an entry present in the map
  412. indicates a reservation), and region_del is called for private mappings (as
  413. the absence of an entry in the map indicates a reservation). See the section
  414. "Reservation cleanup in error paths" for more information on what needs to
  415. be done on error paths.
  416. Reservation Cleanup in Error Paths
  417. ==================================
  418. As mentioned in the section
  419. :ref:`Reservation Map Helper Routines <resv_map_helpers>`, reservation
  420. map modifications are performed in two steps. First vma_needs_reservation
  421. is called before a page is allocated. If the allocation is successful,
  422. then vma_commit_reservation is called. If not, vma_end_reservation is called.
  423. Global and subpool reservation counts are adjusted based on success or failure
  424. of the operation and all is well.
  425. Additionally, after a huge page is instantiated the PagePrivate flag is
  426. cleared so that accounting when the page is ultimately freed is correct.
  427. However, there are several instances where errors are encountered after a huge
  428. page is allocated but before it is instantiated. In this case, the page
  429. allocation has consumed the reservation and made the appropriate subpool,
  430. reservation map and global count adjustments. If the page is freed at this
  431. time (before instantiation and clearing of PagePrivate), then free_huge_page
  432. will increment the global reservation count. However, the reservation map
  433. indicates the reservation was consumed. This resulting inconsistent state
  434. will cause the 'leak' of a reserved huge page. The global reserve count will
  435. be higher than it should and prevent allocation of a pre-allocated page.
  436. The routine restore_reserve_on_error() attempts to handle this situation. It
  437. is fairly well documented. The intention of this routine is to restore
  438. the reservation map to the way it was before the page allocation. In this
  439. way, the state of the reservation map will correspond to the global reservation
  440. count after the page is freed.
  441. The routine restore_reserve_on_error itself may encounter errors while
  442. attempting to restore the reservation map entry. In this case, it will
  443. simply clear the PagePrivate flag of the page. In this way, the global
  444. reserve count will not be incremented when the page is freed. However, the
  445. reservation map will continue to look as though the reservation was consumed.
  446. A page can still be allocated for the address, but it will not use a reserved
  447. page as originally intended.
  448. There is some code (most notably userfaultfd) which can not call
  449. restore_reserve_on_error. In this case, it simply modifies the PagePrivate
  450. so that a reservation will not be leaked when the huge page is freed.
  451. Reservations and Memory Policy
  452. ==============================
  453. Per-node huge page lists existed in struct hstate when git was first used
  454. to manage Linux code. The concept of reservations was added some time later.
  455. When reservations were added, no attempt was made to take memory policy
  456. into account. While cpusets are not exactly the same as memory policy, this
  457. comment in hugetlb_acct_memory sums up the interaction between reservations
  458. and cpusets/memory policy::
  459. /*
  460. * When cpuset is configured, it breaks the strict hugetlb page
  461. * reservation as the accounting is done on a global variable. Such
  462. * reservation is completely rubbish in the presence of cpuset because
  463. * the reservation is not checked against page availability for the
  464. * current cpuset. Application can still potentially OOM'ed by kernel
  465. * with lack of free htlb page in cpuset that the task is in.
  466. * Attempt to enforce strict accounting with cpuset is almost
  467. * impossible (or too ugly) because cpuset is too fluid that
  468. * task or memory node can be dynamically moved between cpusets.
  469. *
  470. * The change of semantics for shared hugetlb mapping with cpuset is
  471. * undesirable. However, in order to preserve some of the semantics,
  472. * we fall back to check against current free page availability as
  473. * a best attempt and hopefully to minimize the impact of changing
  474. * semantics that cpuset has.
  475. */
  476. Huge page reservations were added to prevent unexpected page allocation
  477. failures (OOM) at page fault time. However, if an application makes use
  478. of cpusets or memory policy there is no guarantee that huge pages will be
  479. available on the required nodes. This is true even if there are a sufficient
  480. number of global reservations.
  481. Hugetlbfs regression testing
  482. ============================
  483. The most complete set of hugetlb tests are in the libhugetlbfs repository.
  484. If you modify any hugetlb related code, use the libhugetlbfs test suite
  485. to check for regressions. In addition, if you add any new hugetlb
  486. functionality, please add appropriate tests to libhugetlbfs.
  487. --
  488. Mike Kravetz, 7 April 2017