mirror of https://github.com/torvalds/linux.git
mm: remove is_hugetlb_entry_[migration, hwpoisoned]()
We do not need to have explicit helper functions for these, it adds a level of confusion and indirection when we can simply use software leaf entry logic here instead and spell out the special huge_pte_none() case we must consider. No functional change intended. Link: https://lkml.kernel.org/r/0e92d6924d3de88cd014ce1c53e20edc08fc152e.1762812360.git.lorenzo.stoakes@oracle.com Signed-off-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Cc: Alexander Gordeev <agordeev@linux.ibm.com> Cc: Alistair Popple <apopple@nvidia.com> Cc: Al Viro <viro@zeniv.linux.org.uk> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Axel Rasmussen <axelrasmussen@google.com> Cc: Baolin Wang <baolin.wang@linux.alibaba.com> Cc: Baoquan He <bhe@redhat.com> Cc: Barry Song <baohua@kernel.org> Cc: Byungchul Park <byungchul@sk.com> Cc: Chengming Zhou <chengming.zhou@linux.dev> Cc: Chris Li <chrisl@kernel.org> Cc: Christian Borntraeger <borntraeger@linux.ibm.com> Cc: Christian Brauner <brauner@kernel.org> Cc: Claudio Imbrenda <imbrenda@linux.ibm.com> Cc: David Hildenbrand <david@redhat.com> Cc: Dev Jain <dev.jain@arm.com> Cc: Gerald Schaefer <gerald.schaefer@linux.ibm.com> Cc: Gregory Price <gourry@gourry.net> Cc: Heiko Carstens <hca@linux.ibm.com> Cc: "Huang, Ying" <ying.huang@linux.alibaba.com> Cc: Hugh Dickins <hughd@google.com> Cc: Jan Kara <jack@suse.cz> Cc: Jann Horn <jannh@google.com> Cc: Janosch Frank <frankja@linux.ibm.com> Cc: Jason Gunthorpe <jgg@ziepe.ca> Cc: Joshua Hahn <joshua.hahnjy@gmail.com> Cc: Kairui Song <kasong@tencent.com> Cc: Kemeng Shi <shikemeng@huaweicloud.com> Cc: Lance Yang <lance.yang@linux.dev> Cc: Leon Romanovsky <leon@kernel.org> Cc: Liam Howlett <liam.howlett@oracle.com> Cc: Mathew Brost <matthew.brost@intel.com> Cc: Matthew Wilcox (Oracle) <willy@infradead.org> Cc: Miaohe Lin <linmiaohe@huawei.com> Cc: Michal Hocko <mhocko@suse.com> Cc: Mike Rapoport <rppt@kernel.org> Cc: Muchun Song <muchun.song@linux.dev> Cc: Naoya Horiguchi <nao.horiguchi@gmail.com> Cc: Nhat Pham <nphamcs@gmail.com> Cc: Nico Pache <npache@redhat.com> Cc: Oscar Salvador <osalvador@suse.de> Cc: Pasha Tatashin <pasha.tatashin@soleen.com> Cc: Peter Xu <peterx@redhat.com> Cc: Rakie Kim <rakie.kim@sk.com> Cc: Rik van Riel <riel@surriel.com> Cc: Ryan Roberts <ryan.roberts@arm.com> Cc: SeongJae Park <sj@kernel.org> Cc: Suren Baghdasaryan <surenb@google.com> Cc: Sven Schnelle <svens@linux.ibm.com> Cc: Vasily Gorbik <gor@linux.ibm.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Wei Xu <weixugc@google.com> Cc: xu xin <xu.xin16@zte.com.cn> Cc: Yuanchu Xie <yuanchu@google.com> Cc: Zi Yan <ziy@nvidia.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
9ff30bb9ab
commit
03bfbc3ad6
|
|
@ -2500,22 +2500,23 @@ static void make_uffd_wp_huge_pte(struct vm_area_struct *vma,
|
||||||
unsigned long addr, pte_t *ptep,
|
unsigned long addr, pte_t *ptep,
|
||||||
pte_t ptent)
|
pte_t ptent)
|
||||||
{
|
{
|
||||||
unsigned long psize;
|
const unsigned long psize = huge_page_size(hstate_vma(vma));
|
||||||
|
softleaf_t entry;
|
||||||
|
|
||||||
if (is_hugetlb_entry_hwpoisoned(ptent) || pte_is_marker(ptent))
|
if (huge_pte_none(ptent))
|
||||||
return;
|
|
||||||
|
|
||||||
psize = huge_page_size(hstate_vma(vma));
|
|
||||||
|
|
||||||
if (is_hugetlb_entry_migration(ptent))
|
|
||||||
set_huge_pte_at(vma->vm_mm, addr, ptep,
|
|
||||||
pte_swp_mkuffd_wp(ptent), psize);
|
|
||||||
else if (!huge_pte_none(ptent))
|
|
||||||
huge_ptep_modify_prot_commit(vma, addr, ptep, ptent,
|
|
||||||
huge_pte_mkuffd_wp(ptent));
|
|
||||||
else
|
|
||||||
set_huge_pte_at(vma->vm_mm, addr, ptep,
|
set_huge_pte_at(vma->vm_mm, addr, ptep,
|
||||||
make_pte_marker(PTE_MARKER_UFFD_WP), psize);
|
make_pte_marker(PTE_MARKER_UFFD_WP), psize);
|
||||||
|
|
||||||
|
entry = softleaf_from_pte(ptent);
|
||||||
|
if (softleaf_is_hwpoison(entry) || softleaf_is_marker(entry))
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (softleaf_is_migration(entry))
|
||||||
|
set_huge_pte_at(vma->vm_mm, addr, ptep,
|
||||||
|
pte_swp_mkuffd_wp(ptent), psize);
|
||||||
|
else
|
||||||
|
huge_ptep_modify_prot_commit(vma, addr, ptep, ptent,
|
||||||
|
huge_pte_mkuffd_wp(ptent));
|
||||||
}
|
}
|
||||||
#endif /* CONFIG_HUGETLB_PAGE */
|
#endif /* CONFIG_HUGETLB_PAGE */
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -274,8 +274,6 @@ void hugetlb_vma_lock_release(struct kref *kref);
|
||||||
long hugetlb_change_protection(struct vm_area_struct *vma,
|
long hugetlb_change_protection(struct vm_area_struct *vma,
|
||||||
unsigned long address, unsigned long end, pgprot_t newprot,
|
unsigned long address, unsigned long end, pgprot_t newprot,
|
||||||
unsigned long cp_flags);
|
unsigned long cp_flags);
|
||||||
bool is_hugetlb_entry_migration(pte_t pte);
|
|
||||||
bool is_hugetlb_entry_hwpoisoned(pte_t pte);
|
|
||||||
void hugetlb_unshare_all_pmds(struct vm_area_struct *vma);
|
void hugetlb_unshare_all_pmds(struct vm_area_struct *vma);
|
||||||
void fixup_hugetlb_reservations(struct vm_area_struct *vma);
|
void fixup_hugetlb_reservations(struct vm_area_struct *vma);
|
||||||
void hugetlb_split(struct vm_area_struct *vma, unsigned long addr);
|
void hugetlb_split(struct vm_area_struct *vma, unsigned long addr);
|
||||||
|
|
|
||||||
91
mm/hugetlb.c
91
mm/hugetlb.c
|
|
@ -4846,32 +4846,6 @@ static void set_huge_ptep_maybe_writable(struct vm_area_struct *vma,
|
||||||
set_huge_ptep_writable(vma, address, ptep);
|
set_huge_ptep_writable(vma, address, ptep);
|
||||||
}
|
}
|
||||||
|
|
||||||
bool is_hugetlb_entry_migration(pte_t pte)
|
|
||||||
{
|
|
||||||
swp_entry_t swp;
|
|
||||||
|
|
||||||
if (huge_pte_none(pte) || pte_present(pte))
|
|
||||||
return false;
|
|
||||||
swp = pte_to_swp_entry(pte);
|
|
||||||
if (is_migration_entry(swp))
|
|
||||||
return true;
|
|
||||||
else
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
bool is_hugetlb_entry_hwpoisoned(pte_t pte)
|
|
||||||
{
|
|
||||||
swp_entry_t swp;
|
|
||||||
|
|
||||||
if (huge_pte_none(pte) || pte_present(pte))
|
|
||||||
return false;
|
|
||||||
swp = pte_to_swp_entry(pte);
|
|
||||||
if (is_hwpoison_entry(swp))
|
|
||||||
return true;
|
|
||||||
else
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
static void
|
||||||
hugetlb_install_folio(struct vm_area_struct *vma, pte_t *ptep, unsigned long addr,
|
hugetlb_install_folio(struct vm_area_struct *vma, pte_t *ptep, unsigned long addr,
|
||||||
struct folio *new_folio, pte_t old, unsigned long sz)
|
struct folio *new_folio, pte_t old, unsigned long sz)
|
||||||
|
|
@ -4900,6 +4874,7 @@ int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src,
|
||||||
unsigned long npages = pages_per_huge_page(h);
|
unsigned long npages = pages_per_huge_page(h);
|
||||||
struct mmu_notifier_range range;
|
struct mmu_notifier_range range;
|
||||||
unsigned long last_addr_mask;
|
unsigned long last_addr_mask;
|
||||||
|
softleaf_t softleaf;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
if (cow) {
|
if (cow) {
|
||||||
|
|
@ -4947,16 +4922,16 @@ int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src,
|
||||||
entry = huge_ptep_get(src_vma->vm_mm, addr, src_pte);
|
entry = huge_ptep_get(src_vma->vm_mm, addr, src_pte);
|
||||||
again:
|
again:
|
||||||
if (huge_pte_none(entry)) {
|
if (huge_pte_none(entry)) {
|
||||||
/*
|
/* Skip if src entry none. */
|
||||||
* Skip if src entry none.
|
goto next;
|
||||||
*/
|
}
|
||||||
;
|
|
||||||
} else if (unlikely(is_hugetlb_entry_hwpoisoned(entry))) {
|
softleaf = softleaf_from_pte(entry);
|
||||||
|
if (unlikely(softleaf_is_hwpoison(softleaf))) {
|
||||||
if (!userfaultfd_wp(dst_vma))
|
if (!userfaultfd_wp(dst_vma))
|
||||||
entry = huge_pte_clear_uffd_wp(entry);
|
entry = huge_pte_clear_uffd_wp(entry);
|
||||||
set_huge_pte_at(dst, addr, dst_pte, entry, sz);
|
set_huge_pte_at(dst, addr, dst_pte, entry, sz);
|
||||||
} else if (unlikely(is_hugetlb_entry_migration(entry))) {
|
} else if (unlikely(softleaf_is_migration(softleaf))) {
|
||||||
softleaf_t softleaf = softleaf_from_pte(entry);
|
|
||||||
bool uffd_wp = pte_swp_uffd_wp(entry);
|
bool uffd_wp = pte_swp_uffd_wp(entry);
|
||||||
|
|
||||||
if (!is_readable_migration_entry(softleaf) && cow) {
|
if (!is_readable_migration_entry(softleaf) && cow) {
|
||||||
|
|
@ -4975,7 +4950,6 @@ int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src,
|
||||||
entry = huge_pte_clear_uffd_wp(entry);
|
entry = huge_pte_clear_uffd_wp(entry);
|
||||||
set_huge_pte_at(dst, addr, dst_pte, entry, sz);
|
set_huge_pte_at(dst, addr, dst_pte, entry, sz);
|
||||||
} else if (unlikely(pte_is_marker(entry))) {
|
} else if (unlikely(pte_is_marker(entry))) {
|
||||||
const softleaf_t softleaf = softleaf_from_pte(entry);
|
|
||||||
const pte_marker marker = copy_pte_marker(softleaf, dst_vma);
|
const pte_marker marker = copy_pte_marker(softleaf, dst_vma);
|
||||||
|
|
||||||
if (marker)
|
if (marker)
|
||||||
|
|
@ -5033,9 +5007,7 @@ int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src,
|
||||||
}
|
}
|
||||||
hugetlb_install_folio(dst_vma, dst_pte, addr,
|
hugetlb_install_folio(dst_vma, dst_pte, addr,
|
||||||
new_folio, src_pte_old, sz);
|
new_folio, src_pte_old, sz);
|
||||||
spin_unlock(src_ptl);
|
goto next;
|
||||||
spin_unlock(dst_ptl);
|
|
||||||
continue;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (cow) {
|
if (cow) {
|
||||||
|
|
@ -5056,6 +5028,8 @@ int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src,
|
||||||
set_huge_pte_at(dst, addr, dst_pte, entry, sz);
|
set_huge_pte_at(dst, addr, dst_pte, entry, sz);
|
||||||
hugetlb_count_add(npages, dst);
|
hugetlb_count_add(npages, dst);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
next:
|
||||||
spin_unlock(src_ptl);
|
spin_unlock(src_ptl);
|
||||||
spin_unlock(dst_ptl);
|
spin_unlock(dst_ptl);
|
||||||
}
|
}
|
||||||
|
|
@ -6064,8 +6038,10 @@ vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
|
||||||
ret = 0;
|
ret = 0;
|
||||||
|
|
||||||
/* Not present, either a migration or a hwpoisoned entry */
|
/* Not present, either a migration or a hwpoisoned entry */
|
||||||
if (!pte_present(vmf.orig_pte)) {
|
if (!pte_present(vmf.orig_pte) && !huge_pte_none(vmf.orig_pte)) {
|
||||||
if (is_hugetlb_entry_migration(vmf.orig_pte)) {
|
const softleaf_t softleaf = softleaf_from_pte(vmf.orig_pte);
|
||||||
|
|
||||||
|
if (softleaf_is_migration(softleaf)) {
|
||||||
/*
|
/*
|
||||||
* Release the hugetlb fault lock now, but retain
|
* Release the hugetlb fault lock now, but retain
|
||||||
* the vma lock, because it is needed to guard the
|
* the vma lock, because it is needed to guard the
|
||||||
|
|
@ -6076,9 +6052,12 @@ vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma,
|
||||||
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
|
mutex_unlock(&hugetlb_fault_mutex_table[hash]);
|
||||||
migration_entry_wait_huge(vma, vmf.address, vmf.pte);
|
migration_entry_wait_huge(vma, vmf.address, vmf.pte);
|
||||||
return 0;
|
return 0;
|
||||||
} else if (is_hugetlb_entry_hwpoisoned(vmf.orig_pte))
|
}
|
||||||
|
if (softleaf_is_hwpoison(softleaf)) {
|
||||||
ret = VM_FAULT_HWPOISON_LARGE |
|
ret = VM_FAULT_HWPOISON_LARGE |
|
||||||
VM_FAULT_SET_HINDEX(hstate_index(h));
|
VM_FAULT_SET_HINDEX(hstate_index(h));
|
||||||
|
}
|
||||||
|
|
||||||
goto out_mutex;
|
goto out_mutex;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -6460,7 +6439,9 @@ long hugetlb_change_protection(struct vm_area_struct *vma,
|
||||||
i_mmap_lock_write(vma->vm_file->f_mapping);
|
i_mmap_lock_write(vma->vm_file->f_mapping);
|
||||||
last_addr_mask = hugetlb_mask_last_page(h);
|
last_addr_mask = hugetlb_mask_last_page(h);
|
||||||
for (; address < end; address += psize) {
|
for (; address < end; address += psize) {
|
||||||
|
softleaf_t entry;
|
||||||
spinlock_t *ptl;
|
spinlock_t *ptl;
|
||||||
|
|
||||||
ptep = hugetlb_walk(vma, address, psize);
|
ptep = hugetlb_walk(vma, address, psize);
|
||||||
if (!ptep) {
|
if (!ptep) {
|
||||||
if (!uffd_wp) {
|
if (!uffd_wp) {
|
||||||
|
|
@ -6492,15 +6473,23 @@ long hugetlb_change_protection(struct vm_area_struct *vma,
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
pte = huge_ptep_get(mm, address, ptep);
|
pte = huge_ptep_get(mm, address, ptep);
|
||||||
if (unlikely(is_hugetlb_entry_hwpoisoned(pte))) {
|
if (huge_pte_none(pte)) {
|
||||||
/* Nothing to do. */
|
if (unlikely(uffd_wp))
|
||||||
} else if (unlikely(is_hugetlb_entry_migration(pte))) {
|
/* Safe to modify directly (none->non-present). */
|
||||||
softleaf_t entry = softleaf_from_pte(pte);
|
set_huge_pte_at(mm, address, ptep,
|
||||||
|
make_pte_marker(PTE_MARKER_UFFD_WP),
|
||||||
|
psize);
|
||||||
|
goto next;
|
||||||
|
}
|
||||||
|
|
||||||
|
entry = softleaf_from_pte(pte);
|
||||||
|
if (unlikely(softleaf_is_hwpoison(entry))) {
|
||||||
|
/* Nothing to do. */
|
||||||
|
} else if (unlikely(softleaf_is_migration(entry))) {
|
||||||
struct folio *folio = softleaf_to_folio(entry);
|
struct folio *folio = softleaf_to_folio(entry);
|
||||||
pte_t newpte = pte;
|
pte_t newpte = pte;
|
||||||
|
|
||||||
if (is_writable_migration_entry(entry)) {
|
if (softleaf_is_migration_write(entry)) {
|
||||||
if (folio_test_anon(folio))
|
if (folio_test_anon(folio))
|
||||||
entry = make_readable_exclusive_migration_entry(
|
entry = make_readable_exclusive_migration_entry(
|
||||||
swp_offset(entry));
|
swp_offset(entry));
|
||||||
|
|
@ -6527,7 +6516,7 @@ long hugetlb_change_protection(struct vm_area_struct *vma,
|
||||||
if (pte_is_uffd_wp_marker(pte) && uffd_wp_resolve)
|
if (pte_is_uffd_wp_marker(pte) && uffd_wp_resolve)
|
||||||
/* Safe to modify directly (non-present->none). */
|
/* Safe to modify directly (non-present->none). */
|
||||||
huge_pte_clear(mm, address, ptep, psize);
|
huge_pte_clear(mm, address, ptep, psize);
|
||||||
} else if (!huge_pte_none(pte)) {
|
} else {
|
||||||
pte_t old_pte;
|
pte_t old_pte;
|
||||||
unsigned int shift = huge_page_shift(hstate_vma(vma));
|
unsigned int shift = huge_page_shift(hstate_vma(vma));
|
||||||
|
|
||||||
|
|
@ -6540,16 +6529,10 @@ long hugetlb_change_protection(struct vm_area_struct *vma,
|
||||||
pte = huge_pte_clear_uffd_wp(pte);
|
pte = huge_pte_clear_uffd_wp(pte);
|
||||||
huge_ptep_modify_prot_commit(vma, address, ptep, old_pte, pte);
|
huge_ptep_modify_prot_commit(vma, address, ptep, old_pte, pte);
|
||||||
pages++;
|
pages++;
|
||||||
} else {
|
|
||||||
/* None pte */
|
|
||||||
if (unlikely(uffd_wp))
|
|
||||||
/* Safe to modify directly (none->non-present). */
|
|
||||||
set_huge_pte_at(mm, address, ptep,
|
|
||||||
make_pte_marker(PTE_MARKER_UFFD_WP),
|
|
||||||
psize);
|
|
||||||
}
|
}
|
||||||
spin_unlock(ptl);
|
|
||||||
|
|
||||||
|
next:
|
||||||
|
spin_unlock(ptl);
|
||||||
cond_resched();
|
cond_resched();
|
||||||
}
|
}
|
||||||
/*
|
/*
|
||||||
|
|
|
||||||
|
|
@ -768,16 +768,21 @@ static int queue_folios_hugetlb(pte_t *pte, unsigned long hmask,
|
||||||
unsigned long flags = qp->flags;
|
unsigned long flags = qp->flags;
|
||||||
struct folio *folio;
|
struct folio *folio;
|
||||||
spinlock_t *ptl;
|
spinlock_t *ptl;
|
||||||
pte_t entry;
|
pte_t ptep;
|
||||||
|
|
||||||
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
|
ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte);
|
||||||
entry = huge_ptep_get(walk->mm, addr, pte);
|
ptep = huge_ptep_get(walk->mm, addr, pte);
|
||||||
if (!pte_present(entry)) {
|
if (!pte_present(ptep)) {
|
||||||
if (unlikely(is_hugetlb_entry_migration(entry)))
|
if (!huge_pte_none(ptep)) {
|
||||||
|
const softleaf_t entry = softleaf_from_pte(ptep);
|
||||||
|
|
||||||
|
if (unlikely(softleaf_is_migration(entry)))
|
||||||
qp->nr_failed++;
|
qp->nr_failed++;
|
||||||
|
}
|
||||||
|
|
||||||
goto unlock;
|
goto unlock;
|
||||||
}
|
}
|
||||||
folio = pfn_folio(pte_pfn(entry));
|
folio = pfn_folio(pte_pfn(ptep));
|
||||||
if (!queue_folio_required(folio, qp))
|
if (!queue_folio_required(folio, qp))
|
||||||
goto unlock;
|
goto unlock;
|
||||||
if (!(flags & (MPOL_MF_MOVE | MPOL_MF_MOVE_ALL)) ||
|
if (!(flags & (MPOL_MF_MOVE | MPOL_MF_MOVE_ALL)) ||
|
||||||
|
|
|
||||||
15
mm/migrate.c
15
mm/migrate.c
|
|
@ -515,16 +515,18 @@ void migration_entry_wait(struct mm_struct *mm, pmd_t *pmd,
|
||||||
void migration_entry_wait_huge(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep)
|
void migration_entry_wait_huge(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep)
|
||||||
{
|
{
|
||||||
spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep);
|
spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep);
|
||||||
|
softleaf_t entry;
|
||||||
pte_t pte;
|
pte_t pte;
|
||||||
|
|
||||||
hugetlb_vma_assert_locked(vma);
|
hugetlb_vma_assert_locked(vma);
|
||||||
spin_lock(ptl);
|
spin_lock(ptl);
|
||||||
pte = huge_ptep_get(vma->vm_mm, addr, ptep);
|
pte = huge_ptep_get(vma->vm_mm, addr, ptep);
|
||||||
|
|
||||||
if (unlikely(!is_hugetlb_entry_migration(pte))) {
|
if (huge_pte_none(pte))
|
||||||
spin_unlock(ptl);
|
goto fail;
|
||||||
hugetlb_vma_unlock_read(vma);
|
|
||||||
} else {
|
entry = softleaf_from_pte(pte);
|
||||||
|
if (softleaf_is_migration(entry)) {
|
||||||
/*
|
/*
|
||||||
* If migration entry existed, safe to release vma lock
|
* If migration entry existed, safe to release vma lock
|
||||||
* here because the pgtable page won't be freed without the
|
* here because the pgtable page won't be freed without the
|
||||||
|
|
@ -533,7 +535,12 @@ void migration_entry_wait_huge(struct vm_area_struct *vma, unsigned long addr, p
|
||||||
*/
|
*/
|
||||||
hugetlb_vma_unlock_read(vma);
|
hugetlb_vma_unlock_read(vma);
|
||||||
migration_entry_wait_on_locked(pte_to_swp_entry(pte), ptl);
|
migration_entry_wait_on_locked(pte_to_swp_entry(pte), ptl);
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fail:
|
||||||
|
spin_unlock(ptl);
|
||||||
|
hugetlb_vma_unlock_read(vma);
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue