Skip to content

Commit 04ada09

Browse files
mjkravetzakpm00
authored andcommitted
hugetlb: don't delete vma_lock in hugetlb MADV_DONTNEED processing
madvise(MADV_DONTNEED) ends up calling zap_page_range() to clear page tables associated with the address range. For hugetlb vmas, zap_page_range will call __unmap_hugepage_range_final. However, __unmap_hugepage_range_final assumes the passed vma is about to be removed and deletes the vma_lock to prevent pmd sharing as the vma is on the way out. In the case of madvise(MADV_DONTNEED) the vma remains, but the missing vma_lock prevents pmd sharing and could potentially lead to issues with truncation/fault races. This issue was originally reported here [1] as a BUG triggered in page_try_dup_anon_rmap. Prior to the introduction of the hugetlb vma_lock, __unmap_hugepage_range_final cleared the VM_MAYSHARE flag to prevent pmd sharing. Subsequent faults on this vma were confused as VM_MAYSHARE indicates a sharable vma, but was not set so page_mapping was not set in new pages added to the page table. This resulted in pages that appeared anonymous in a VM_SHARED vma and triggered the BUG. Address issue by adding a new zap flag ZAP_FLAG_UNMAP to indicate an unmap call from unmap_vmas(). This is used to indicate the 'final' unmapping of a hugetlb vma. When called via MADV_DONTNEED, this flag is not set and the vm_lock is not deleted. [1] https://lore.kernel.org/lkml/CAO4mrfdLMXsao9RF4fUE8-Wfde8xmjsKrTNMNC9wjUb6JudD0g@mail.gmail.com/ Link: https://lkml.kernel.org/r/20221114235507.294320-3-mike.kravetz@oracle.com Fixes: 90e7e7f ("mm: enable MADV_DONTNEED for hugetlb mappings") Signed-off-by: Mike Kravetz <mike.kravetz@oracle.com> Reported-by: Wei Chen <harperchen1110@gmail.com> Cc: Axel Rasmussen <axelrasmussen@google.com> Cc: David Hildenbrand <david@redhat.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: Mina Almasry <almasrymina@google.com> Cc: Nadav Amit <nadav.amit@gmail.com> Cc: Naoya Horiguchi <naoya.horiguchi@linux.dev> Cc: Peter Xu <peterx@redhat.com> Cc: Rik van Riel <riel@surriel.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: <stable@vger.kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
1 parent 21b85b0 commit 04ada09

3 files changed

Lines changed: 19 additions & 12 deletions

File tree

include/linux/mm.h

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1868,6 +1868,8 @@ struct zap_details {
18681868
* default, the flag is not set.
18691869
*/
18701870
#define ZAP_FLAG_DROP_MARKER ((__force zap_flags_t) BIT(0))
1871+
/* Set in unmap_vmas() to indicate a final unmap call. Only used by hugetlb */
1872+
#define ZAP_FLAG_UNMAP ((__force zap_flags_t) BIT(1))
18711873

18721874
#ifdef CONFIG_MMU
18731875
extern bool can_do_mlock(void);

mm/hugetlb.c

Lines changed: 16 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -5206,17 +5206,22 @@ void __unmap_hugepage_range_final(struct mmu_gather *tlb,
52065206

52075207
__unmap_hugepage_range(tlb, vma, start, end, ref_page, zap_flags);
52085208

5209-
/*
5210-
* Unlock and free the vma lock before releasing i_mmap_rwsem. When
5211-
* the vma_lock is freed, this makes the vma ineligible for pmd
5212-
* sharing. And, i_mmap_rwsem is required to set up pmd sharing.
5213-
* This is important as page tables for this unmapped range will
5214-
* be asynchrously deleted. If the page tables are shared, there
5215-
* will be issues when accessed by someone else.
5216-
*/
5217-
__hugetlb_vma_unlock_write_free(vma);
5218-
5219-
i_mmap_unlock_write(vma->vm_file->f_mapping);
5209+
if (zap_flags & ZAP_FLAG_UNMAP) { /* final unmap */
5210+
/*
5211+
* Unlock and free the vma lock before releasing i_mmap_rwsem.
5212+
* When the vma_lock is freed, this makes the vma ineligible
5213+
* for pmd sharing. And, i_mmap_rwsem is required to set up
5214+
* pmd sharing. This is important as page tables for this
5215+
* unmapped range will be asynchrously deleted. If the page
5216+
* tables are shared, there will be issues when accessed by
5217+
* someone else.
5218+
*/
5219+
__hugetlb_vma_unlock_write_free(vma);
5220+
i_mmap_unlock_write(vma->vm_file->f_mapping);
5221+
} else {
5222+
i_mmap_unlock_write(vma->vm_file->f_mapping);
5223+
hugetlb_vma_unlock_write(vma);
5224+
}
52205225
}
52215226

52225227
void unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start,

mm/memory.c

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1711,7 +1711,7 @@ void unmap_vmas(struct mmu_gather *tlb, struct maple_tree *mt,
17111711
{
17121712
struct mmu_notifier_range range;
17131713
struct zap_details details = {
1714-
.zap_flags = ZAP_FLAG_DROP_MARKER,
1714+
.zap_flags = ZAP_FLAG_DROP_MARKER | ZAP_FLAG_UNMAP,
17151715
/* Careful - we need to zap private pages too! */
17161716
.even_cows = true,
17171717
};

0 commit comments

Comments
 (0)