mirror of https://github.com/torvalds/linux.git
mm: kill mm_wr_locked from unmap_vmas() and unmap_single_vma()
Kill mm_wr_locked since commit f8e97613fe ("mm: convert VM_PFNMAP
tracking to pfnmap_track() + pfnmap_untrack()") remove the user.
Link: https://lkml.kernel.org/r/20251104085709.2688433-1-wangkefeng.wang@huawei.com
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
Reviewed-by: Liam R. Howlett <Liam.Howlett@oracle.com>
Acked-by: David Hildenbrand (Red Hat) <david@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
3b12a53b64
commit
340b59816b
|
|
@ -2480,7 +2480,7 @@ static inline void zap_vma_pages(struct vm_area_struct *vma)
|
|||
}
|
||||
void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
|
||||
struct vm_area_struct *start_vma, unsigned long start,
|
||||
unsigned long end, unsigned long tree_end, bool mm_wr_locked);
|
||||
unsigned long end, unsigned long tree_end);
|
||||
|
||||
struct mmu_notifier_range;
|
||||
|
||||
|
|
|
|||
12
mm/memory.c
12
mm/memory.c
|
|
@ -2023,8 +2023,7 @@ void unmap_page_range(struct mmu_gather *tlb,
|
|||
|
||||
static void unmap_single_vma(struct mmu_gather *tlb,
|
||||
struct vm_area_struct *vma, unsigned long start_addr,
|
||||
unsigned long end_addr,
|
||||
struct zap_details *details, bool mm_wr_locked)
|
||||
unsigned long end_addr, struct zap_details *details)
|
||||
{
|
||||
unsigned long start = max(vma->vm_start, start_addr);
|
||||
unsigned long end;
|
||||
|
|
@ -2070,7 +2069,6 @@ static void unmap_single_vma(struct mmu_gather *tlb,
|
|||
* @start_addr: virtual address at which to start unmapping
|
||||
* @end_addr: virtual address at which to end unmapping
|
||||
* @tree_end: The maximum index to check
|
||||
* @mm_wr_locked: lock flag
|
||||
*
|
||||
* Unmap all pages in the vma list.
|
||||
*
|
||||
|
|
@ -2085,8 +2083,7 @@ static void unmap_single_vma(struct mmu_gather *tlb,
|
|||
*/
|
||||
void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
|
||||
struct vm_area_struct *vma, unsigned long start_addr,
|
||||
unsigned long end_addr, unsigned long tree_end,
|
||||
bool mm_wr_locked)
|
||||
unsigned long end_addr, unsigned long tree_end)
|
||||
{
|
||||
struct mmu_notifier_range range;
|
||||
struct zap_details details = {
|
||||
|
|
@ -2102,8 +2099,7 @@ void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
|
|||
unsigned long start = start_addr;
|
||||
unsigned long end = end_addr;
|
||||
hugetlb_zap_begin(vma, &start, &end);
|
||||
unmap_single_vma(tlb, vma, start, end, &details,
|
||||
mm_wr_locked);
|
||||
unmap_single_vma(tlb, vma, start, end, &details);
|
||||
hugetlb_zap_end(vma, &details);
|
||||
vma = mas_find(mas, tree_end - 1);
|
||||
} while (vma && likely(!xa_is_zero(vma)));
|
||||
|
|
@ -2139,7 +2135,7 @@ void zap_page_range_single_batched(struct mmu_gather *tlb,
|
|||
* unmap 'address-end' not 'range.start-range.end' as range
|
||||
* could have been expanded for hugetlb pmd sharing.
|
||||
*/
|
||||
unmap_single_vma(tlb, vma, address, end, details, false);
|
||||
unmap_single_vma(tlb, vma, address, end, details);
|
||||
mmu_notifier_invalidate_range_end(&range);
|
||||
if (is_vm_hugetlb_page(vma)) {
|
||||
/*
|
||||
|
|
|
|||
|
|
@ -1274,7 +1274,7 @@ void exit_mmap(struct mm_struct *mm)
|
|||
tlb_gather_mmu_fullmm(&tlb, mm);
|
||||
/* update_hiwater_rss(mm) here? but nobody should be looking */
|
||||
/* Use ULONG_MAX here to ensure all VMAs in the mm are unmapped */
|
||||
unmap_vmas(&tlb, &vmi.mas, vma, 0, ULONG_MAX, ULONG_MAX, false);
|
||||
unmap_vmas(&tlb, &vmi.mas, vma, 0, ULONG_MAX, ULONG_MAX);
|
||||
mmap_read_unlock(mm);
|
||||
|
||||
/*
|
||||
|
|
|
|||
5
mm/vma.c
5
mm/vma.c
|
|
@ -483,8 +483,7 @@ void unmap_region(struct ma_state *mas, struct vm_area_struct *vma,
|
|||
|
||||
tlb_gather_mmu(&tlb, mm);
|
||||
update_hiwater_rss(mm);
|
||||
unmap_vmas(&tlb, mas, vma, vma->vm_start, vma->vm_end, vma->vm_end,
|
||||
/* mm_wr_locked = */ true);
|
||||
unmap_vmas(&tlb, mas, vma, vma->vm_start, vma->vm_end, vma->vm_end);
|
||||
mas_set(mas, vma->vm_end);
|
||||
free_pgtables(&tlb, mas, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS,
|
||||
next ? next->vm_start : USER_PGTABLES_CEILING,
|
||||
|
|
@ -1228,7 +1227,7 @@ static inline void vms_clear_ptes(struct vma_munmap_struct *vms,
|
|||
tlb_gather_mmu(&tlb, vms->vma->vm_mm);
|
||||
update_hiwater_rss(vms->vma->vm_mm);
|
||||
unmap_vmas(&tlb, mas_detach, vms->vma, vms->start, vms->end,
|
||||
vms->vma_count, mm_wr_locked);
|
||||
vms->vma_count);
|
||||
|
||||
mas_set(mas_detach, 1);
|
||||
/* start and end may be different if there is no prev or next vma. */
|
||||
|
|
|
|||
|
|
@ -848,8 +848,7 @@ static inline void update_hiwater_vm(struct mm_struct *mm)
|
|||
|
||||
static inline void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
|
||||
struct vm_area_struct *vma, unsigned long start_addr,
|
||||
unsigned long end_addr, unsigned long tree_end,
|
||||
bool mm_wr_locked)
|
||||
unsigned long end_addr, unsigned long tree_end)
|
||||
{
|
||||
}
|
||||
|
||||
|
|
|
|||
Loading…
Reference in New Issue